Bilibili Tech
Bilibili Tech
Apr 14, 2026 · Artificial Intelligence

Can 10% of Instruction Data Match Full-Scale Fine-Tuning? The SPICE Solution

The SPICE method leverages Fisher Information Matrix submodularity and a novel gradient‑conflict penalty to select a small, high‑quality subset of instruction‑tuning data, achieving comparable or superior performance to full‑data fine‑tuning while dramatically reducing training cost.

Fisher informationGradient ConflictInstruction Tuning
0 likes · 13 min read
Can 10% of Instruction Data Match Full-Scale Fine-Tuning? The SPICE Solution
Machine Learning Algorithms & Natural Language Processing
Machine Learning Algorithms & Natural Language Processing
Mar 26, 2026 · Artificial Intelligence

Can Uni‑X Eliminate Multimodal Gradient Conflict with a Pure Autoregressive Design?

The paper reveals that standard shared‑parameter Transformers suffer severe gradient conflict when jointly processing low‑entropy text and high‑entropy visual tokens, and proposes Uni‑X—a two‑end‑separated, middle‑shared autoregressive model that isolates modality‑specific layers, reduces conflict, improves efficiency, and achieves strong results on image generation and editing benchmarks.

Autoregressive ModelGradient ConflictICLR 2026
0 likes · 8 min read
Can Uni‑X Eliminate Multimodal Gradient Conflict with a Pure Autoregressive Design?
AI Frontier Lectures
AI Frontier Lectures
Mar 21, 2025 · Artificial Intelligence

How ConFIG Eliminates Gradient Conflicts for Faster Multi‑Task Deep Learning

The paper introduces ConFIG (Conflict‑Free Inverse Gradients), a mathematically proven method that resolves gradient conflicts among multiple loss terms in physics‑informed neural networks, multi‑task learning, and continual learning, and its momentum‑based variant M‑ConFIG that further accelerates training while maintaining accuracy.

CONFIGGradient ConflictM-ConFIG
0 likes · 11 min read
How ConFIG Eliminates Gradient Conflicts for Faster Multi‑Task Deep Learning