OLMoE-1B-7B
An efficient open-source large language model.
CommonProductProductivityLarge Language ModelOpen Source
OLMoE-1B-7B is a specialized mixture of experts large language model (LLM) with 100 million active parameters and a total of 700 million parameters, released in September 2024. This model outperforms similarly priced models and competes with larger models such as Llama2-13B. OLMoE is fully open-source and supports a variety of features, including text generation, model training, and deployment.
OLMoE-1B-7B Visit Over Time
Monthly Visits
20899836
Bounce Rate
46.04%
Page per Visit
5.2
Visit Duration
00:04:57