Highlights of Meituan's ACL 2024 Papers: Speculative Decoding, Graph‑Structured Decoding, DolphCoder, and Instruction Fine‑tuning

Meituan showcases four ACL 2024 papers—Early‑Exiting Speculative Decoding with a Thompson‑sampling controller, Graph‑Structured Speculative Decoding that merges draft hypotheses in a DAG, DolphCoder, a code‑generation LLM improved by diverse multi‑objective instruction tuning, and a study of instruction fine‑tuning that finds it mainly aligns existing knowledge—while inviting attendees to its booth 11 and a live paper discussion on August 12.

Meituan Technology Team
Meituan Technology Team
Meituan Technology Team
Highlights of Meituan's ACL 2024 Papers: Speculative Decoding, Graph‑Structured Decoding, DolphCoder, and Instruction Fine‑tuning

This article selects four papers from Meituan's technical team that were accepted at ACL 2024, covering training‑cost optimization, speculative decoding, code‑generation optimization, and instruction fine‑tuning.

ACL is the premier international conference on computational linguistics and natural language processing, organized by the Association for Computational Linguistics. It is a CCF‑A class venue and the most influential conference in the field.

Meituan will have a booth (No. 11) at the ACL 2024 venue. A live stream and paper discussion will be held on August 12 at 17:00. Attendees are welcome to make appointments.

01. Speculative Decoding via Early‑exiting for Faster LLM Inference with Thompson Sampling Control Mechanism (Long Paper) – PDF

The rapid growth of large language models (LLMs) raises inference cost. The authors propose Early‑Exiting Speculative Decoding (EESD), which adds an early‑exit after N layers to generate draft tokens, uses self‑distillation to improve them, and employs a Thompson‑sampling controller to decide the number of draft tokens per step. Experiments on 13B and 70B models show significant speed‑up without quality loss.

02. Graph‑Structured Speculative Decoding (Long Paper) – PDF

This work extends speculative decoding by generating multiple draft hypotheses and organizing them in a directed acyclic graph (DAG). The graph allows efficient prediction and merging of repeated token sequences, reducing the draft model’s computation. Applied to models up to 70B parameters, the method (GSD) improves generation speed by 1.73‑1.96× over standard speculative decoding.

03. DolphCoder: Echo‑Locating Code Large Language Models with Diverse and Multi‑Objective Instruction Tuning (Long Paper) – PDF

To boost code‑generation performance of large code LLMs, the paper introduces DolphCoder, a diversified instruction‑tuned model that incorporates self‑assessment objectives. It learns varied instruction targets and combines code‑evaluation goals, achieving superior results on HumanEval and MBPP benchmarks. The study shows that diverse responses and better evaluation of code solutions enhance both code creation and correctness.

04. Learning or Self‑aligning? Rethinking Instruction Fine‑tuning (Long Paper) – PDF

The authors propose a knowledge‑perturbation analysis framework to disentangle the effects of instruction fine‑tuning (IFT) on large models. Experiments reveal that IFT mainly aligns existing internal knowledge rather than injecting new knowledge, and that preserving internal knowledge consistency is crucial for successful fine‑tuning.

Live broadcast will walk through the ACL papers. Meituan’s booth (No. 11) and technical experts will be present for交流 and Q&A.

code generationLLMSpeculative DecodingNLPInstruction TuningACL
Meituan Technology Team
Written by

Meituan Technology Team

Over 10,000 engineers powering China’s leading lifestyle services e‑commerce platform. Supporting hundreds of millions of consumers, millions of merchants across 2,000+ industries. This is the public channel for the tech teams behind Meituan, Dianping, Meituan Waimai, Meituan Select, and related services.

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.