Moonlight-16B-A3B
Moonlight-16B-A3B is a 16B parameter Mixture-of-Experts (MoE) model trained with the Muon optimizer for efficient language generation.
Moonlight-16B-A3B Visit Over Time
Monthly Visits
27175375
Bounce Rate
44.30%
Page per Visit
5.8
Visit Duration
00:04:57