ModernBERT-base

Efficient bidirectional encoder model for processing long texts.

CommonProductProgrammingBERTLong Text Processing
ModernBERT-base is a modern bidirectional encoder Transformer model pretrained on 2 trillion English and code samples, natively supporting up to 8192 tokens of context. The model incorporates cutting-edge architectural improvements such as Rotary Positional Embeddings (RoPE), Local-Global Alternating Attention, and Unpadding, showing exceptional performance on long-text processing tasks. It is ideal for processing long documents for tasks such as retrieval, classification, and semantic search within large corpuses. Since the training data is primarily in English and code, its performance may be reduced when handling other languages.
Visit

ModernBERT-base Visit Over Time

Monthly Visits

20899836

Bounce Rate

46.04%

Page per Visit

5.2

Visit Duration

00:04:57

ModernBERT-base Visit Trend

ModernBERT-base Visit Geography

ModernBERT-base Traffic Sources

ModernBERT-base Alternatives