Artificial Intelligence 21 min read

Causal Inference and Bias Correction Methods in Ant Financial Risk Control

This article presents how Ant Group applies causal inference techniques—including confounding bias analysis, double‑difference methods, DiDTree, and shrinkage‑based causal trees—to correct biases in risk‑control scenarios, detailing the theoretical background, algorithmic designs, experimental validation, and practical deployment.

DataFunSummit
DataFunSummit
DataFunSummit
Causal Inference and Bias Correction Methods in Ant Financial Risk Control

Background of Causal Inference – Causal effect estimation aims to quantify how a treatment (e.g., issuing a coupon) changes an outcome (e.g., conversion rate) beyond what would happen without the treatment. Randomized experiments provide unbiased estimates but are costly; observational data are abundant but suffer from confounding bias, illustrated by Simpson’s paradox.

Confounding Bias and Its Mitigation – When a hidden variable (e.g., user payment frequency) influences both the treatment assignment and the outcome, the observed effect is distorted. Mitigating this bias requires either (1) using pre‑intervention differences to model the bias or (2) leveraging a small amount of unbiased experimental data together with large biased observational data.

Time‑Series De‑biasing: Double Difference (DiD) and DiDTree – The double‑difference approach subtracts pre‑intervention outcome differences to remove confounding bias, assuming a parallel trend. DiDTree extends this by embedding a parallel‑trend loss into a tree‑based model and using gradient‑boosting to enforce the assumption locally in each leaf, enabling individual‑level effect estimation.

Experiment‑Data De‑biasing: Shrinkage Causal Trees and FAST – By fusing unbiased experimental data with biased observational data through shrinkage estimation, a weighted combination of the two models yields lower mean‑squared error. This principle is applied to causal trees: leaf‑level effects are estimated via shrinkage, and a parallel‑trend loss modifies the splitting criterion. The resulting Fast Shrinkage Tree (FAST) and its ensemble (FAST‑RF) achieve robust performance across simulated and real datasets.

Experimental Evaluation – Extensive simulations varying the degree of parallel‑trend satisfaction and treatment‑group imbalance show that DiDTree and FAST methods remain stable, outperforming traditional DiD, meta‑learners, causal forests, and neural‑network baselines. Real‑world experiments on Ant’s risk‑control data confirm the superiority of FAST‑RF in estimating average treatment effects.

Applications in Ant Group – In credit‑limit adjustments, price reductions, and other risk‑control scenarios, these causal de‑biasing techniques enable accurate uplift estimation without costly large‑scale randomized trials, directly improving business outcomes.

Q&A Highlights – The audience asked about the relationship between DiDTree and causal forests, handling non‑monotonic uplift rankings, and the role of parallel assumptions in leaf nodes; the presenter clarified the methodological distinctions and practical remedies.

machine learningcausal inferencerisk controlbias correctionant financialdouble differenceshrinkage estimation
DataFunSummit
Written by

DataFunSummit

Official account of the DataFun community, dedicated to sharing big data and AI industry summit news and speaker talks, with regular downloadable resource packs.

0 followers
Reader feedback

How this landed with the community

login Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.