Baobao Algorithm Notes
Dec 25, 2025 · Artificial Intelligence
TeleChat3-105B: China’s First 100B‑Scale MoE Model and Its Technical Breakthroughs
The article analyzes TeleChat3-105B-A4.7-Thinking, the first domestically built 100‑billion‑parameter Mixture‑of‑Experts model, detailing its multi‑dimensional evaluation, three‑stage training pipeline, hardware‑level optimizations, fine‑grained architecture, and its significance for the evolving AI competition landscape.
AI trainingChinese AIMixture of Experts
0 likes · 6 min read
