DataFunSummit
DataFunSummit
Mar 21, 2026 · Artificial Intelligence

How Slidebatching Revolutionizes LLM Inference Scheduling for Faster, More Efficient AI Services

The article examines the memory and latency challenges of 1750‑billion‑parameter LLM inference, introduces the xLLM framework’s Slidebatching and PD‑separation scheduling strategies, and details how these techniques achieve up to 35% system‑throughput gains and 52% SLO compliance improvements in real‑world multi‑priority workloads.

AI performanceLLMPD separation
0 likes · 15 min read
How Slidebatching Revolutionizes LLM Inference Scheduling for Faster, More Efficient AI Services