2024-07-31 11:27:05.AIbase.10.7k
Zyphra Launches Small Language Model Zamba2-2.7B: Speed Doubled, Memory Cost Reduced by 27%
Zyphra has launched the Zamba2-2.7B language model, which is a milestone in the small language model domain. Its performance and efficiency have significantly improved, with a training dataset size of around 30 trillion tokens, reducing resource requirements during inference and making it an efficient solution for mobile applications. Key highlights include a twofold increase in response generation speed, a 27% reduction in memory usage, and a 1.29 times decrease in generation latency, particularly suited for applications requiring real-time interaction, such as virtual assistants and chatbots.