AIBase
Home
AI NEWS
AI Tools
AI Models
MCP
AI Services
AI Compute
AI Tutorial
Datasets
EN

AI News

View More

Global's First Pure AMD-Trained MoE Large Model ZAYA1 Launch: 14T Tokens + CCA Attention Performance Comparable to Qwen3

AMD, IBM, and Zyphra launch ZAYA1, the first MoE model trained entirely on AMD hardware. Pretrained on 14T tokens, it matches Qwen3 series performance with strong math reasoning. Uses 128 nodes × 8 MI300X GPUs (750 PFLOPs), CCA attention mechanism, and curriculum learning. Optimized versions to follow.....

5.9k 21 minutes ago
Global's First Pure AMD-Trained MoE Large Model ZAYA1 Launch: 14T Tokens + CCA Attention Performance Comparable to Qwen3
AIBase
Empowering the future, your artificial intelligence solution think tank
English简体中文繁體中文にほんご
FirendLinks:
AI Newsletters AI ToolsMCP ServersAI NewsAIBaseLLM LeaderboardAI Ranking
© 2025AIBase
Business CooperationSite Map