일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | |||
5 | 6 | 7 | 8 | 9 | 10 | 11 |
12 | 13 | 14 | 15 | 16 | 17 | 18 |
19 | 20 | 21 | 22 | 23 | 24 | 25 |
26 | 27 | 28 | 29 | 30 | 31 |
- 주식
- 파라미터 수
- MLLM
- 3C4P
- def train
- 논문리뷰
- layer 추출
- Foundation Transformers
- 모델 구조 변경
- KOSMOS-2
- Instruction dataset
- 논문 작성
- Video Understanding
- pretrained model layer
- Multimodal Large Language Model
- 파라미터 수 확인
- 가상환경
- 특정 layer 동결
- 논문 작성 요령
- 특정 layer 추출
- DeepNet
- 모델 freeze
- 모델 동결
- mPLUG-2
- 가중치 없이 모델 로드
- cnn
- 강화학습
- mqtt
- def validation
- 주식 용어정리
- Today
- Total
목록논문리뷰 (2)
시작은 미약하였으나 , 그 끝은 창대하리라
논문 링크: https://www.mdpi.com/2504-446X/7/2/114 Semantic Scene Understanding with Large Language Models on Unmanned Aerial Vehicles Unmanned Aerial Vehicles (UAVs) are able to provide instantaneous visual cues and a high-level data throughput that could be further leveraged to address complex tasks, such as semantically rich scene understanding. In this work, we built on the use of Lar www.mdpi.co..
논문링크 : https://arxiv.org/abs/2302.00402 mPLUG-2: A Modularized Multi-modal Foundation Model Across Text, Image and Video Recent years have witnessed a big convergence of language, vision, and multi-modal pretraining. In this work, we present mPLUG-2, a new unified paradigm with modularized design for multi-modal pretraining, which can benefit from modality collaboration whil arxiv.org Published ..