Llava-onevision: Easy Visual Task Transfer | Awesome LLM Papers Contribute to Awesome LLM Papers

Llava-onevision: Easy Visual Task Transfer

Bo Li, Yuanhan Zhang, Dong Guo, Renrui Zhang, Feng Li, Hao Zhang, Kaichen Zhang, Yanwei Li, Ziwei Liu, Chunyuan Li . No Venue 2024

[Paper] [Paper]   Search on Google Scholar   Search on Semantic Scholar
Fine Tuning

We present LLaVA-OneVision, a family of open large multimodal models (LMMs) developed by consolidating our insights into data, models, and visual representations in the LLaVA-NeXT blog series. Our experimental results demonstrate that LLaVA-OneVision is the first single model that can simultaneously push the performance boundaries of open LMMs in three important computer vision scenarios: single-image, multi-image, and video scenarios. Importantly, the design of LLaVA-OneVision allows strong transfer learning across different modalities/scenarios, yielding new emerging capabilities. In particular, strong video understanding and cross-scenario capabilities are demonstrated through task transfer from images to videos.

https://huggingface.co/discussions/paper/66b2deee8188b0d6f73f4e9f

Similar Work