Enabling Search Agents to Think While Waiting: Diffusion LLMs Deliver 15% Faster Inference Without Accuracy Loss
The paper introduces DLLM‑Searcher, which equips diffusion large language models with a two‑stage training pipeline and a P‑ReAct inference scheme, allowing the model to issue tool calls while simultaneously reasoning, yielding 14‑22% end‑to‑end speedup and matching or surpassing traditional autoregressive agents on multi‑hop QA benchmarks.
