DBRX

A new efficient open-source large language model standard

EditorRecommendationProductivityArtificial IntelligenceLarge Language Model
DBRX is a general-purpose large language model (LLM) built by Databricks' Mosaic research team. It outperforms all existing open-source models in standard benchmark tests. It uses a Mixture-of-Experts (MoE) architecture with 36.2 billion parameters, boasting excellent language understanding, programming, mathematical, and logical reasoning capabilities. DBRX aims to promote the development of high-quality open-source LLMs and facilitates enterprise customization of the model based on their own data. Databricks provides enterprise users with the ability to interactively use DBRX, leverage its long context capabilities to build retrieval-enhanced systems, and build customized DBRX models based on their own data.
Visit

DBRX Visit Over Time

Monthly Visits

494758773

Bounce Rate

37.69%

Page per Visit

5.7

Visit Duration

00:06:29

DBRX Visit Trend

DBRX Visit Geography

DBRX Traffic Sources

DBRX Alternatives