AIbase
Product LibraryTool NavigationMCP

NanoMoE

Public

A compact Transformer model for character-level generation with Mixture-of-Experts, Rotary Attention, and F-gram contextual augmentation.

Creat2025-02-19T19:25:20
Update2025-06-18T11:59:17
0
Stars
0
Stars Increase

Related projects