InternVL2_5-8B
A multimodal large language model supporting interaction understanding between images and text.
CommonProductImageMultimodalLarge Language Model
InternVL2_5-8B is a multimodal large language model (MLLM) developed by OpenGVLab, significantly enhanced with training and testing strategies as well as data quality improvements based on InternVL 2.0. This model employs the 'ViT-MLP-LLM' architecture, integrating the newly pre-trained InternViT with various pre-trained language models, such as InternLM 2.5 and Qwen 2.5, utilizing a randomly initialized MLP projector. The InternVL 2.5 series models demonstrate outstanding performance on multimodal tasks, including image and video understanding and multilingual comprehension.
InternVL2_5-8B Visit Over Time
Monthly Visits
20899836
Bounce Rate
46.04%
Page per Visit
5.2
Visit Duration
00:04:57