Evidence Mining for Explainable AI: Methods and Applications

The talk introduces evidence‑mining techniques that extract supporting information from input text to improve model explainability, discusses the shortcut‑learning pitfalls of existing methods, and presents a new approach that enhances reliability and integrates with large‑model chain‑of‑thought compression for more interpretable, efficient reasoning.

Machine Learning Algorithms & Natural Language Processing
Machine Learning Algorithms & Natural Language Processing
Machine Learning Algorithms & Natural Language Processing
Evidence Mining for Explainable AI: Methods and Applications

The MLNLP Academic Talk series promotes exchange among machine‑learning and natural‑language‑processing researchers. This session, held on April 18, 2026, features Associate Professor Yue Linan from Southeast University.

Evidence mining extracts key information from raw input text to support model predictions, serving as a crucial technique for explainability in high‑risk decision‑making. Existing methods often fall into the "shortcut learning" trap, over‑relying on superficial features, which reduces the reliability of the extracted evidence, and they are largely confined to small‑model paradigms.

The presentation introduces a shortcut‑avoidance evidence‑mining method that improves the faithfulness and reliability of explanations by addressing the shortcut learning issue. It also describes recent work that combines evidence mining with chain‑of‑thought compression for large models, aiming to enhance the interpretability of efficient reasoning processes in large‑model scenarios.

Speaker bio: Yue Linan is an associate professor in the School of Computer Science and Engineering at Southeast University, PhD graduate of the University of Science and Technology of China, and a member of the PALM lab. His research focuses on knowledge representation and reasoning, efficient large‑model inference, and trustworthy AI. He has authored over 30 papers in CCF‑A/B journals and conferences, received awards such as the Chinese Academy of Sciences President’s Special Award, and serves on editorial boards and AI committees.

Original Source

Signed-in readers can open the original source through BestHub's protected redirect.

Sign in to view source
Republication Notice

This article has been distilled and summarized from source material, then republished for learning and reference. If you believe it infringes your rights, please contactadmin@besthub.devand we will review it promptly.

large language modelsAI researchModel Interpretabilityexplainable AIevidence mining
Machine Learning Algorithms & Natural Language Processing
Written by

Machine Learning Algorithms & Natural Language Processing

Focused on frontier AI technologies, empowering AI researchers' progress.

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.