How Tencent’s TNC Neural Codec Won 2025 Image & Video Compression Challenges

At the end of 2025, Tencent’s Shannon Lab’s neural codec TNC achieved top rankings in both the VCIP low‑complexity end‑to‑end image compression contest and the PCS high‑compression intelligent image compression challenge, demonstrating superior PSNR gains, low decoding complexity, and innovative VAE‑INR architecture across image and video tracks.

Tencent Cloud Developer
Tencent Cloud Developer
Tencent Cloud Developer
How Tencent’s TNC Neural Codec Won 2025 Image & Video Compression Challenges

Competition Results

In December 2025 the VCIP 2025 Low‑Complexity End‑to‑End Image Compression Challenge and the PCS 2025 High‑Compression Intelligent Image Compression Challenge were concluded. Tencent Shannon Lab’s Tencent Neural Codec (TNC) achieved the best visual quality at the same bitrate in both contests, winning the image track (average 0.51 bpp, +1.66 dB PSNR over BPG, up to +2.81 dB) and the video track (subjective MOS +1.20 over VTM, +0.57 over ECM, objective PSNR +3.07 dB over VTM).

TNC Image Codec Architecture

VAE‑INR Hybrid Encoder

TNC introduces a controllable frame‑level and block‑level rate‑distortion optimizer that selects either a Variational Auto‑Encoder (VAE) or an Implicit Neural Representation (INR) for each block. This hybrid design exploits the high‑fidelity reconstruction of INR when it outperforms the VAE.

VAE‑INR architecture diagram
VAE‑INR architecture diagram

Asymmetric Encoder‑Decoder Design

The encoder uses a deep, high‑capacity network to extract rich latent features, while the decoder employs a lightweight MobileNet‑like backbone with ShuffleNet‑style channel grouping and re‑parameterized 3×3 convolutions to keep decoding cost low. The activation function is replaced by WSiLU, a smooth, everywhere‑differentiable variant that avoids gradient truncation.

Hyper‑Prior and Context Model

A hyper‑prior network extracts side information from the latent features. A dual autoregressive context model predicts each latent element from already decoded spatial neighbors and from a non‑uniform channel partition, progressively fusing context to improve probability estimates for arithmetic coding.

Context model diagram
Context model diagram

Generalized Gaussian Entropy Model

The latent distribution is modeled by a generalized Gaussian with mean, scale, and shape parameters, providing a richer fit to diverse data and reducing symbol‑coding overhead.

Generalized Gaussian PDFs
Generalized Gaussian PDFs

INR Over‑fit Coding

Each image is split into eight blocks. An INR is over‑fitted to each block; if its reconstruction error is lower than the VAE’s, the learned latent variables and network weights are packed into the bitstream using a lottery‑ticket mask, enabling low‑complexity decoding while improving compression.

INR coding flow
INR coding flow

Subjective Quality Optimization Pipeline

Train a single‑rate model with L2 loss.

Fine‑tune for multiple rates using randomly sampled Lagrange multipliers as conditioning signals.

Add LPIPS to the loss to preserve semantic details.

Freeze encoder and entropy model, then apply a subjective aggregation loss in adversarial training to generate realistic textures.

Subjective optimization results
Subjective optimization results

AI‑Driven TNC Video Codec

The video codec builds on the image stack and adds:

ESRGAN‑based AI pre‑processing for detail enhancement.

Neural Network Loop Filtering (NNLF) for post‑processing.

A complexity‑aware CUTree adaptive QP calculation.

Scene detection with local complexity analysis to refine QP.

Video codec architecture
Video codec architecture

Frame‑Level Parallelism

NNLF is split into two stages: the first pass determines frame‑level parameters from a subset of rows; the second pass processes the remaining rows using those parameters, allowing dependent frames to start decoding earlier.

Parallel processing diagram
Parallel processing diagram

Algorithm‑Engine Joint Acceleration

Compute‑intensive kernels are accelerated with AVX2 intrinsics and GPU kernels. The SADL inference library gains ~45 % speedup, and adaptive block skipping based on texture and QP adds another ~45 % improvement, yielding an overall decoding speed 2.11× faster than the unoptimized baseline.

Performance breakdown
Performance breakdown

Results Summary

Image codec: average PSNR gain +1.66 dB over BPG at 0.51 bpp, peak gain +2.81 dB.

Video codec (subjective): MOS +1.20 over VTM, +0.57 over ECM.

Video codec (objective): PSNR gain +3.07 dB over VTM, +2.00 dB over ECM.

Future Outlook

While TNC demonstrates state‑of‑the‑art compression efficiency and visual quality, challenges remain for commercial deployment, including cross‑platform decoding consistency and strict terminal‑side complexity constraints. Ongoing work will focus on further algorithmic innovation and engineering optimization to bridge the gap to industry adoption.

AIVAEvideo compressionNeural CodecTNCCompression ChallengeINR
Tencent Cloud Developer
Written by

Tencent Cloud Developer

Official Tencent Cloud community account that brings together developers, shares practical tech insights, and fosters an influential tech exchange community.

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.