Machine Learning Algorithms & Natural Language Processing
Mar 7, 2026 · Artificial Intelligence
How Princeton’s DYSCO Decoder Boosts Long-Context Reasoning by 25% Without Fine‑Tuning
The DYSCO (Dynamic Attention‑Scaling Decoding) algorithm, introduced by Princeton’s Chen Danqi team and NYU, eliminates the need for fine‑tuning and restores performance on long‑context tasks, delivering up to a 25% relative gain on 128K token benchmarks while adding only about 3.8% extra FLOPs.
DYSCODynamic Attention ScalingLong-context LLM
0 likes · 10 min read
