Cloud Native 16 min read

ByteDance’s Cloud‑Native Tricks for 703 Billion Festival Red‑Envelope Clicks

In 2021, ByteDance leveraged its cloud‑native stack—including traffic governance, service mesh, and rapid deployment pipelines—to deliver 703 billion red‑envelope interactions during the Spring Festival Gala, overcoming a 27‑day preparation window, high‑concurrency demands, and extensive emergency rehearsals to ensure seamless user experience.

Volcano Engine Developer Services
Volcano Engine Developer Services
Volcano Engine Developer Services
ByteDance’s Cloud‑Native Tricks for 703 Billion Festival Red‑Envelope Clicks

In 2021, the CCTV Spring Festival Gala partnered with Douyin (TikTok) to distribute 12 billion red envelopes, generating 703 billion interactions, a feat made possible by Volcano Engine’s cloud‑native infrastructure.

Act 1: The Challenge Arrives

With only 27 days to prepare—far less than the typical 50‑day window—ByteDance faced a massive high‑concurrency challenge. The team identified nine critical milestones, including three stress tests, one disaster‑recovery drill, four script rehearsals, and a final live run.

Key technical goals were accurate traffic forecasting, optimal resource scheduling, and guaranteeing user experience under peak load.

Act 2: Eighteen Rehearsals

The project team conducted 18 technical rehearsals, covering traffic governance, service governance, and development efficiency.

Traffic Governance : A fusion of edge computing, dynamic CDN acceleration, and full‑link traffic scheduling provided high bandwidth, low latency, and automatic disaster recovery. Multi‑dimensional routing policies (APP ID, interface, user ID) achieved sub‑0.01% traffic switching precision and rapid configuration updates.

Service Governance : ServiceMesh enabled flexible intra‑cluster traffic routing, load balancing, overload control, and cross‑cluster resource scaling via an online‑offline unified scheduling system.

Development Efficiency : A one‑stop observability platform and public RPC components supported large‑scale pressure testing, real‑time monitoring, and automated performance optimization, allowing the system to handle millions of QPS without incidents.

Act 3: The Unseen Black Swan

Despite the intense pressure, the team avoided all 16 predefined emergency scenarios. Real‑time monitoring across hundreds of engineers ensured rapid issue detection and resolution, preventing any major outage during the live event.

Act 4: How ByteDance Moves Forward

The success stemmed from a mesh‑like collaborative model, where over 20 teams aligned on shared goals rather than rigid task assignments. Long‑term investment in foundational infrastructure—such as a unified storage matrix, hybrid offline/online compute, self‑built CDN, dynamic acceleration, and robust disaster‑recovery systems—provided the flexibility needed for large‑scale events.

Notably, the in‑house client dynamic engine framework Lynx reduced release costs and accelerated iteration, marking its first deployment at this scale.

Conclusion

The Spring Festival Gala red‑envelope project proved ByteDance’s cloud‑native capabilities—dynamic edge acceleration, massive hybrid deployment, and rapid traffic scheduling—are mature enough to handle extreme traffic spikes, turning a daunting challenge into a validated technical achievement.

cloud-nativehigh-concurrencyservice-meshtraffic-management
Volcano Engine Developer Services
Written by

Volcano Engine Developer Services

The Volcano Engine Developer Community, Volcano Engine's TOD community, connects the platform with developers, offering cutting-edge tech content and diverse events, nurturing a vibrant developer culture, and co-building an open-source ecosystem.

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.