Artificial Intelligence 12 min read

Xiaohongshu 2024 Large Model Frontier Paper Sharing Live Event

On June 27, 2024, Xiaohongshu’s technical team will livestream a two‑hour session across WeChat Channels, Bilibili, Douyin and Xiaohongshu, showcasing six top‑conference papers on large‑model advances—including early‑stopping and fine‑grained self‑consistency, novel evaluation methods, negative‑sample‑assisted distillation, and LLM‑based note recommendation—followed by a Q&A and recruitment briefing.

Xiaohongshu Tech REDtech
Xiaohongshu Tech REDtech
Xiaohongshu Tech REDtech
Xiaohongshu 2024 Large Model Frontier Paper Sharing Live Event

The Xiaohongshu technical team announces a live streaming session titled "Xiaohongshu 2024 Large Model Frontier Paper Sharing" scheduled for June 27, 2024, 19:00‑21:30 (UTC+8) on WeChat Channels, Bilibili, Douyin, and Xiaohongshu.

The session will feature six papers accepted at top international conferences (ICLR, ACL, CVPR, AAAI, SIGIR, WWW) that explore new opportunities and challenges at the intersection of large models and natural language processing, as well as effective evaluation methods and real‑world applications.

Early‑Stopping Self‑Consistency (ESC) – Speaker: Li Yiwei

Self‑Consistency (SC) is a widely used decoding strategy for chain‑of‑thought reasoning but incurs high computational cost due to multiple samples. The proposed Early‑Stopping Self‑Consistency (ESC) introduces a simple, scalable sampling process that stops early when sufficient consensus is reached, dramatically reducing the number of samples without sacrificing performance. Experiments on mathematical, commonsense, and symbolic reasoning tasks show a large reduction in average sampling steps while preserving accuracy. Paper: https://arxiv.org/abs/2401.10480

Fine‑Grained Self‑Consistency (FSC) – Speaker: Wang Xinglin

FSC addresses the limitation of coarse‑grained sample selection in existing self‑consistency methods for free‑form generation tasks. By leveraging fine‑grained commonality among sample fragments through a large‑model self‑fusion mechanism, FSC achieves significant improvements on code generation, summarization, and mathematical reasoning benchmarks while keeping computational overhead comparable. GitHub: https://github.com/WangXinglin/FSC

BatchEval – Speaker: Yuan Peiwen

BatchEval is a novel evaluation framework that attains near‑human performance on text assessment with lower cost. It analyzes the shortcomings of current metrics—uneven score distributions and limited perspective diversity—and introduces a batch‑wise comparison inspired by human evaluation practices. Empirical results show BatchEval outperforms state‑of‑the‑art methods in both efficiency and effectiveness. Paper: https://arxiv.org/abs/2401.00437

PEEM (Probabilistic EM‑based Evaluation Method) – Speaker: Yuan Peiwen

PEEM leverages inter‑model consistency to evaluate large language models that surpass human performance, where reliable human labels become scarce. The method derives a theoretical guarantee that, under independence between reference and target model predictions, consistency with the reference model serves as an accurate ability metric. An EM‑based algorithm mitigates practical violations of the independence assumption, enabling accurate assessment of super‑human LLMs. GitHub: https://github.com/ypw0102/PEEM

Negative‑Sample‑Assisted Distillation – Speaker: Li Yiwei

This work proposes a pioneering framework that incorporates negative (incorrect) samples into the knowledge distillation process for large language models. The three‑step pipeline—Negative‑Assisted Training (NAT), Negative‑Calibrated Enhancement (NCE), and Adaptive Self‑Consistency (ASC)—demonstrates that negative data can substantially improve the distilled model’s reasoning quality, especially on complex mathematical problems. Paper: https://arxiv.org/abs/2312.12832

LLM‑Based Note Content Representation for Recommendation – Speaker: Zhang Chao

To address the cold‑start problem of newly created notes on the Xiaohongshu app, this research builds a recommendation system that encodes note content using large language models. Two approaches are explored: generation‑enhanced representations and multimodal content embeddings. The system has been deployed in multiple product scenarios, delivering notable performance gains. Paper: https://arxiv.org/abs/2403.01744

The live stream will also include a Q&A session, a QR‑code for a community group, and recruitment information for the Xiaohongshu search team.

Large Language ModelsRecommendation systemsmodel evaluationAI researchknowledge distillationSelf-Consistency
Xiaohongshu Tech REDtech
Written by

Xiaohongshu Tech REDtech

Official account of the Xiaohongshu tech team, sharing tech innovations and problem insights, advancing together.

0 followers
Reader feedback

How this landed with the community

login Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.