Falcon Mamba
The first 7B large-scale model that operates without an attention mechanism.
CommonProductProgrammingLarge ModelsNo Attention
Falcon Mamba is the first 7B large-scale model released by the Technology Innovation Institute (TII) in Abu Dhabi that does not use attention mechanisms. This model is free from the computational and storage costs that increase with longer sequences, while still maintaining performance on par with current state-of-the-art models.
Falcon Mamba Visit Over Time
Monthly Visits
26103677
Bounce Rate
43.69%
Page per Visit
5.5
Visit Duration
00:04:43