Artificial Intelligence 11 min read

Tencent AI Lab's Advances in High‑Fidelity 3D Face Digitization and Evaluation

This article presents Tencent AI Lab's recent research on efficient 3D face digitization—including single‑photo, multi‑photo, and RGB‑D selfie pipelines—describes a detailed production workflow, introduces a new evaluation benchmark (REALY), and shares insights from a technical Q&A session.

DataFunTalk
DataFunTalk
DataFunTalk
Tencent AI Lab's Advances in High‑Fidelity 3D Face Digitization and Evaluation

The presentation introduces Tencent AI Lab's latest work on high‑efficiency 3D facial digitization, covering methods that use a single photo, multiple photos, and RGB‑D selfie capture to reconstruct realistic 3D face models.

It outlines a complete production pipeline: light‑field scanning, high‑poly reconstruction, defect cleaning, UV mapping, low‑poly retopology, texture and normal map generation, followed by dynamic modeling steps such as expression capture, controller binding, and final rendering.

To improve production speed, three solution tiers (A, B, C) are explored. Tier A replaces costly 360° light‑field rigs with a simple camera array and AI‑driven generation; Tier B targets consumer‑grade devices; Tier C achieves reconstruction from one or a few photos, making the technology accessible to end users.

The RGB‑D selfie workflow is detailed in seven steps: automatic frame selection, initial model fitting, differentiable‑rendering optimization, texture/normal map synthesis, head/attachment mounting, AutoRigging, and text/voice‑driven animation. Each step is illustrated with accompanying images.

For evaluation, the authors identify shortcomings in traditional rigid‑alignment and nearest‑neighbor error metrics, then propose improvements: local mask‑based alignment and a reverse non‑rigid deformation to refine correspondence. They introduce the REALY benchmark, containing 100 image‑model pairs with per‑region masks, enabling fine‑grained assessment.

Extensive experiments on open‑source single‑image 3D face reconstruction methods demonstrate the effectiveness of the new metrics. The work has been published at ECCV and TOG, with code and data publicly released.

The session concludes with a Q&A covering lighting correction, hair modeling, oral cavity handling, and factors that make their digital humans more realistic than competing approaches.

computer vision3D face reconstructionEvaluation Benchmarkdigital humansAI LabDifferentiable RenderingRGBD
DataFunTalk
Written by

DataFunTalk

Dedicated to sharing and discussing big data and AI technology applications, aiming to empower a million data scientists. Regularly hosts live tech talks and curates articles on big data, recommendation/search algorithms, advertising algorithms, NLP, intelligent risk control, autonomous driving, and machine learning/deep learning.

0 followers
Reader feedback

How this landed with the community

login Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.