NewBeeNLP
Oct 21, 2024 · Artificial Intelligence
Why Do MOE Experts Collapse? An In‑Depth Look at HOME’s Multi‑Task Architecture
This article analyzes the polarization issues in industrial Mixture‑of‑Experts (MoE) frameworks, explains expert collapse, degradation, and under‑fitting, and details the HOME model’s input types, architectural innovations, normalization, gating mechanisms, and related DICE‑BN insights.
Expert NormalizationGating MechanismsMixture of Experts
0 likes · 10 min read
