![](http://i1.daumcdn.net/thumb/C148x148.fwebp.q85/?fname=https://blog.kakaocdn.net/dn/o1LuC/btsHiK82BFk/YkN8HvXkyW3CBBFeqBsUkK/img.png)
What is VLM (Vision Language Model)?VLM is a model with a multi-modal architecture that learns to associate information from image and text modalities. The focus of the multi-modal learning is to pre-train an model on vision and language task and improve the downstream task performance such as VQA (Vision Question Answering). Why VLM? What are the Use Cases?1. Image Search and Retrieval / 2. Rob..
Research (연구 관련)
2024. 5. 9. 07:00
공지사항
최근에 올라온 글
최근에 달린 댓글
- Total
- Today
- Yesterday
링크
TAG
- pyrender
- nohup
- Docker
- nerf
- pytorch
- Generative model
- deep learning
- 헬스
- 문경식
- 비전
- Machine Learning
- Interview
- Transformation
- 컴퓨터비전
- camera coordinate
- Pose2Mesh
- 2d pose
- densepose
- spin
- Virtual Camera
- demo
- 머신러닝
- 인터뷰
- 컴퓨터비젼
- VAE
- 에디톨로지
- focal length
- 피트니스
- part segmentation
- world coordinate
일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | |||
5 | 6 | 7 | 8 | 9 | 10 | 11 |
12 | 13 | 14 | 15 | 16 | 17 | 18 |
19 | 20 | 21 | 22 | 23 | 24 | 25 |
26 | 27 | 28 | 29 | 30 | 31 |
글 보관함