Wu Shixiong's Large Model Academy
Author

Wu Shixiong's Large Model Academy

We continuously share large‑model know‑how, helping you master core skills—LLM, RAG, fine‑tuning, deployment—from zero to job offer, tailored for career‑switchers, autumn recruiters, and those seeking stable large‑model positions.

107
Articles
0
Likes
33
Views
0
Comments
Recent Articles

Latest from Wu Shixiong's Large Model Academy

100 recent articles max
Wu Shixiong's Large Model Academy
Wu Shixiong's Large Model Academy
Feb 4, 2026 · Artificial Intelligence

Why LLM Agents Rush to Call Tools and How to Stop Them

The article explains that premature tool calls in LLM agents stem from a data‑distribution bias in fine‑tuning, and it presents practical fixes such as adding non‑tool samples, enforcing a Thought chain, and using negative sampling to teach the model when to think before acting.

AgentLLMThought Chain
0 likes · 10 min read
Why LLM Agents Rush to Call Tools and How to Stop Them
Wu Shixiong's Large Model Academy
Wu Shixiong's Large Model Academy
Feb 3, 2026 · Artificial Intelligence

Why Loss Masking Is the Hidden Key to Effective LLM Fine‑Tuning

The article explains how loss masking in supervised fine‑tuning of large language models prevents the model from learning irrelevant tokens such as user inputs, system prompts, tool outputs, and padding, thereby focusing training on the assistant’s responses and improving performance and generalization.

AI trainingFine-tuningLLM
0 likes · 10 min read
Why Loss Masking Is the Hidden Key to Effective LLM Fine‑Tuning
Wu Shixiong's Large Model Academy
Wu Shixiong's Large Model Academy
Dec 15, 2025 · Artificial Intelligence

Mastering Text2SQL: From Schema Design to Secure Multi‑Step LLM Pipelines

This article explains how Text2SQL works by teaching LLMs to understand a closed‑world database schema, constructing tightly constrained prompts, validating generated SQL, handling execution errors, and using a second LLM call to translate results into natural language, while highlighting common pitfalls and engineering best practices.

LLMSQL ValidationSchema Design
0 likes · 9 min read
Mastering Text2SQL: From Schema Design to Secure Multi‑Step LLM Pipelines
Wu Shixiong's Large Model Academy
Wu Shixiong's Large Model Academy
Dec 12, 2025 · Artificial Intelligence

Why Fixing Bad Cases Beats Adding More Data in RLHF

In industrial RLHF, repairing bad cases—structural error samples—provides explicit alignment signals that improve model capability far more efficiently than simply increasing data volume, because it teaches the model how to correct mistakes rather than just exposing it to more examples.

Bad CaseCapability ImprovementData Efficiency
0 likes · 9 min read
Why Fixing Bad Cases Beats Adding More Data in RLHF
Wu Shixiong's Large Model Academy
Wu Shixiong's Large Model Academy
Dec 11, 2025 · Artificial Intelligence

Why Reward Models Need Reasoning: From Scalar Scores to RM‑R1

Interviewers increasingly ask why modern reward models must go beyond scalar scores to incorporate reasoning, and this article explains the limitations of traditional scalar reward models, the benefits of the RM‑R1 framework, and how reasoning‑based rewards improve alignment, stability, and task performance in large language model training.

AI alignmentLLMRLHF
0 likes · 11 min read
Why Reward Models Need Reasoning: From Scalar Scores to RM‑R1
Wu Shixiong's Large Model Academy
Wu Shixiong's Large Model Academy
Dec 6, 2025 · Artificial Intelligence

How to Engineer Industrial‑Scale Function Call Data for AI Agents

This article explains why handcrafted SFT data fails for function‑call agents, introduces a controllable data‑sandbox approach, details label and variable design, shows code for seed generation and full dialogue synthesis, and demonstrates how the resulting dataset improves model routing, multi‑turn handling, tool sequencing, and error resilience.

AI-agentPrompt Designdata-engineering
0 likes · 12 min read
How to Engineer Industrial‑Scale Function Call Data for AI Agents
Wu Shixiong's Large Model Academy
Wu Shixiong's Large Model Academy
Dec 5, 2025 · Artificial Intelligence

Why Do LLM Function Calls Hallucinate Parameters and How to Prevent It?

This article explains the root causes of hallucinated parameters in LLM Function Calls, outlines five common failure patterns, and presents a systematic five‑step engineering framework—including schema design, prompt rules, dynamic routing, result validation, and clarification—to reliably eliminate such errors in real‑world AI agents.

AI AgentFunction CallLLM
0 likes · 11 min read
Why Do LLM Function Calls Hallucinate Parameters and How to Prevent It?