MoE-LLaVA
An expert mixture model based on large-scale vision-language models
CommonProductImageLarge Scale ModelMulti-modal Learning
MoE-LLaVA is an expert mixture model based on large-scale vision-language models, demonstrating excellent performance in multi-modal learning. It has fewer parameters but exhibits high performance and can be trained in a short time. The model supports Gradio Web UI and CLI inference, and provides functions such as model library, requirements and installation, training and validation, customization, visualization, and API.
MoE-LLaVA Visit Over Time
Monthly Visits
515580771
Bounce Rate
37.20%
Page per Visit
5.8
Visit Duration
00:06:42