Jais

13B Parameter Arabic-English Bilingual Isomorphic Model

CommonProductProductivityChatbotIntelligent Chat
This is a pre-trained bilingual large language model with 130 billion parameters, supporting Arabic and English. It is trained on a dataset of 72 billion Arabic tokens and 279 billion English/Code tokens. The Arabic data has been iterated 1.6 epochs compared to 1 epoch for English/Code, with a total of 395 billion tokens trained. The model is based on the Transformer decoder-specific architecture (GPT-3), utilizing the SwiGLU nonlinear activation function. It implements ALiBi positional embeddings, allowing extrapolation to long sequence lengths, providing enhanced context processing and model accuracy.
Visit

Jais Visit Over Time

Monthly Visits

17788201

Bounce Rate

44.87%

Page per Visit

5.4

Visit Duration

00:05:32

Jais Visit Trend

Jais Visit Geography

Jais Traffic Sources

Jais Alternatives