ShareGPT4V Lin-Chen/ShareGPT4V-7B Image-Text-to-Text • Updated Jun 6, 2024 • 7.14k • 85 Lin-Chen/ShareGPT4V-13B Text Generation • Updated Jun 6, 2024 • 8 • 34 Lin-Chen/ShareCaptioner Feature Extraction • Updated Jun 6, 2024 • 40 • 57 Lin-Chen/ShareGPT4V Viewer • Updated Jun 6, 2024 • 1.35M • 2.07k • 306
MMStar An elite vision-indispensable multi-modal benchmark Lin-Chen/MMStar Viewer • Updated Apr 7, 2024 • 1.5k • 20.9k • 50 Are We on the Right Way for Evaluating Large Vision-Language Models? Paper • 2403.20330 • Published Mar 29, 2024 • 6
Are We on the Right Way for Evaluating Large Vision-Language Models? Paper • 2403.20330 • Published Mar 29, 2024 • 6
ShareGPT4Video ShareGPT4Video/ShareGPT4Video Viewer • Updated Mar 7, 2025 • 40.2k • 2.47k • 202 Lin-Chen/sharegpt4video-8b Visual Question Answering • Updated Jul 1, 2024 • 172 • 45 Lin-Chen/ShareCaptioner-Video Text Generation • Updated Jun 11, 2024 • 769 • 19 Runtime error 28 ShareGPT4Video 8B 🎬 28 Send video and text for explanation or action
Open-LLaVA-NeXT Open-source implementation of the LLaVA-NeXT series with Open-LLaVA-NeXT repositary Lin-Chen/open-llava-next-llama3-8b Image-Text-to-Text • 8B • Updated May 27, 2024 • 20 • 26 Lin-Chen/open-llava-next-vicuna-7b Image-Text-to-Text • 7B • Updated May 27, 2024 • 119 • 3 Lin-Chen/Open-LLaVA-NeXT-mix1M Updated Oct 25, 2024 • 40 • 15
ShareGPT4V Lin-Chen/ShareGPT4V-7B Image-Text-to-Text • Updated Jun 6, 2024 • 7.14k • 85 Lin-Chen/ShareGPT4V-13B Text Generation • Updated Jun 6, 2024 • 8 • 34 Lin-Chen/ShareCaptioner Feature Extraction • Updated Jun 6, 2024 • 40 • 57 Lin-Chen/ShareGPT4V Viewer • Updated Jun 6, 2024 • 1.35M • 2.07k • 306
ShareGPT4Video ShareGPT4Video/ShareGPT4Video Viewer • Updated Mar 7, 2025 • 40.2k • 2.47k • 202 Lin-Chen/sharegpt4video-8b Visual Question Answering • Updated Jul 1, 2024 • 172 • 45 Lin-Chen/ShareCaptioner-Video Text Generation • Updated Jun 11, 2024 • 769 • 19 Runtime error 28 ShareGPT4Video 8B 🎬 28 Send video and text for explanation or action
MMStar An elite vision-indispensable multi-modal benchmark Lin-Chen/MMStar Viewer • Updated Apr 7, 2024 • 1.5k • 20.9k • 50 Are We on the Right Way for Evaluating Large Vision-Language Models? Paper • 2403.20330 • Published Mar 29, 2024 • 6
Are We on the Right Way for Evaluating Large Vision-Language Models? Paper • 2403.20330 • Published Mar 29, 2024 • 6
Open-LLaVA-NeXT Open-source implementation of the LLaVA-NeXT series with Open-LLaVA-NeXT repositary Lin-Chen/open-llava-next-llama3-8b Image-Text-to-Text • 8B • Updated May 27, 2024 • 20 • 26 Lin-Chen/open-llava-next-vicuna-7b Image-Text-to-Text • 7B • Updated May 27, 2024 • 119 • 3 Lin-Chen/Open-LLaVA-NeXT-mix1M Updated Oct 25, 2024 • 40 • 15