DataFunTalk
DataFunTalk
Jan 13, 2026 · Artificial Intelligence

How Conditional Memory (Engram) Boosts Large Language Models Beyond MoE

DeepSeek's new paper introduces a conditional memory mechanism called Engram that complements Mixture‑of‑Experts, providing O(1) lookup, improving knowledge retrieval, reasoning, and long‑context performance while scaling efficiently on the same FLOPs budget.

Conditional MemoryEngramMemory retrieval
0 likes · 18 min read
How Conditional Memory (Engram) Boosts Large Language Models Beyond MoE
Kuaishou Tech
Kuaishou Tech
Aug 25, 2025 · Artificial Intelligence

How Context-as-Memory Enables Scene‑Consistent Long Video Generation

This article introduces the Context-as-Memory approach, which treats previously generated video frames as memory to achieve scene‑consistent interactive long video generation, and details a camera‑trajectory‑based memory retrieval mechanism that dramatically improves efficiency and performance over existing state‑of‑the‑art methods.

AIMemory retrievalVideo Generation
0 likes · 7 min read
How Context-as-Memory Enables Scene‑Consistent Long Video Generation
AI Frontier Lectures
AI Frontier Lectures
Jun 20, 2025 · Artificial Intelligence

How GCA Achieves 1000× Length Generalization in Large Language Models

Ant Research introduces GCA, a causal retrieval‑based grouped cross‑attention mechanism that end‑to‑end learns to fetch relevant past chunks, dramatically reducing memory usage and achieving over 1000× length generalization on long‑context language modeling tasks, with near‑constant inference memory and linear training cost.

AI researchGrouped Cross AttentionLLM efficiency
0 likes · 11 min read
How GCA Achieves 1000× Length Generalization in Large Language Models