일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | |||
5 | 6 | 7 | 8 | 9 | 10 | 11 |
12 | 13 | 14 | 15 | 16 | 17 | 18 |
19 | 20 | 21 | 22 | 23 | 24 | 25 |
26 | 27 | 28 | 29 | 30 | 31 |
- 특정 layer 추출
- 파라미터 수 확인
- Instruction dataset
- vsocde 익스텐션
- mqtt
- Foundation Transformers
- Video Understanding
- def train
- 모델 freeze
- 논문 작성 요령
- 강화학습
- def validation
- KOSMOS-2
- 모델 구조 변경
- 가상환경
- 논문리뷰
- mPLUG-2
- 주식 용어정리
- 가중치 없이 모델 로드
- layer 추출
- pretrained model layer
- Multimodal Large Language Model
- 특정 layer 동결
- 파라미터 수
- DeepNet
- 주식
- 모델 동결
- MLLM
- cnn
- 논문 작성
- Today
- Total
목록MLLM (2)
시작은 미약하였으나 , 그 끝은 창대하리라
논문링크 : https://arxiv.org/abs/2306.14824 Kosmos-2: Grounding Multimodal Large Language Models to the World We introduce Kosmos-2, a Multimodal Large Language Model (MLLM), enabling new capabilities of perceiving object descriptions (e.g., bounding boxes) and grounding text to the visual world. Specifically, we represent refer expressions as links in Markdown, i arxiv.org Published : 2023.07 (arXi..
논문 링크: https://arxiv.org/abs/2302.14045 Language Is Not All You Need: Aligning Perception with Language Models A big convergence of language, multimodal perception, action, and world modeling is a key step toward artificial general intelligence. In this work, we introduce Kosmos-1, a Multimodal Large Language Model (MLLM) that can perceive general modalities, learn arxiv.org Published : 2023.03 ..