Recently, a Seattle-based startup, Moondream, has launched a compact visual language model named moondream2. Despite its small size, the model has performed exceptionally well in various benchmark tests, drawing significant attention. As an open-source model, moondream2 is expected to enable local image recognition capabilities on smartphones.
Moondream2 was officially released in March, capable of processing both text and image inputs, with abilities to answer questions, perform Optical Character Recognition (OCR), count objects, and categorize items. Since its launch, the Moondream team has continuously updated the model, enhancing its benchmark performance. The July version showed significant improvements in OCR and document comprehension, particularly in analyzing historical economic data. The model scored over 60% on DocVQA, TextVQA, and GQA, demonstrating its robust capabilities when executed locally.
One notable feature of moondream2 is its compact size: it has only 1.6 billion parameters, making it not only operable on cloud servers but also on local computers and even on some low-performance devices like smartphones or single-board computers.
Despite its small size, its performance is on par with some competitive models with billions of parameters, and even outperforms them in certain benchmark tests.
In a comparison of mobile device visual language models, researchers noted that although moondream2 has only 170 million parameters, its performance is comparable to models with 700 million parameters, only slightly less effective on the SQA dataset. This indicates that while small models perform well, they still face challenges in understanding specific contexts.
The developer of the model, Vikhyat Korrapati, stated that moondream2 was built based on datasets from other models such as SigLIP, Microsoft's Phi-1.5, and LLaVA. This open-source model is now freely available for download on GitHub and has a demo version showcased on Hugging Face. On coding platforms, moondream2 has garnered widespread attention from the developer community, receiving over 5,000 stars.
This success has caught the eye of investors: Moondream has successfully raised $4.5 million in a seed round led by Felicis Ventures, Microsoft's M12GitHub fund, and Ascend. The company's CEO, Jay Allen, who has extensive experience at Amazon Web Services (AWS), leads this growing startup.
The launch of moondream2 marks the emergence of a series of professionally optimized open-source models that require fewer resources while providing performance similar to larger, older models. Although there are already some small local models on the market, such as Apple's intelligent assistant and Google's Gemini Nano, these manufacturers still outsource more complex tasks to the cloud.
huggingface: https://huggingface.co/vikhyatk/moondream2
github: https://github.com/vikhyat/moondream
Key Points:
🌟 Moondream has introduced moondream2, a visual language model with only 160 million parameters, capable of running on small devices like smartphones.
📈 The model boasts robust capabilities in text and image processing, able to answer questions, perform OCR, count objects, and categorize items, with outstanding benchmark performance.
💰 Moondream has successfully raised $4.5 million in funding, with the CEO having worked at Amazon, and the team continuously updating and enhancing the model's performance.