Recently, Hugging Face introduced a brand-new AI tool—SmolLM. These are a series of high-performance small language models with parameters ranging from 135M to 1.7B, specifically designed for various devices and applications. Imagine these compact models running efficiently on smartphones and laptops—it's pretty cool!

The standout feature of SmolLM models is their small yet powerful nature. They perform exceptionally well with fewer computational resources, helping users protect their privacy. Hugging Face used a dataset called SmolLM-Corpus for training these models, which is meticulously curated and includes rich educational and synthetic data, ensuring the models learn a wide range of knowledge.

Specifically, SmolLM comes in three versions: 135M, 360M, and 1.7B parameters. These models can handle multiple tasks and operate flexibly based on users' hardware configurations. For instance, the SmolLM-135M model outperforms many同类产品, becoming a top choice among models with fewer than 200M parameters.

image.png

SmolLM models have been evaluated in various benchmark tests, assessing common sense reasoning and world knowledge. These models have demonstrated impressive performance, outperforming others in their respective size categories. For example, despite being trained on fewer tokens, the SmolLM-135M model surpasses the MobileLM-125M, which is currently the best model with fewer than 200M parameters. Similarly, the SmolLM-360M and SmolLM-1.7B models outperform all other models with fewer than 500M and 2B parameters, respectively.

In addition to their outstanding performance, SmolLM has been specially tuned to excel in understanding instructions and answering questions. Hugging Face also offers a WebGPU demo, allowing everyone to directly experience the capabilities of these models.

The release of SmolLM demonstrates that even small models can achieve amazing performance through high-quality training data.

Link: https://huggingface.co/collections/HuggingFaceTB/smollm-6695016cad7167254ce15966

Key Points:

1. 🚀 **Efficient Performance**: SmolLM models perform well with low computational resources, protecting user privacy.

2. 📚 **Rich Data**: Utilizes high-quality SmolLM-Corpus dataset, ensuring diverse knowledge acquisition.

3. 💻 **Versatile Applications**: Suitable for devices like smartphones and laptops, flexible operation to meet various needs.