Wu Shixiong's Large Model Academy
Wu Shixiong's Large Model Academy
Oct 23, 2025 · Artificial Intelligence

Why the Transformer Core Structure Is the Key to AI Interview Success

This article explains the fundamental purpose, architecture, and variants of the Transformer model—including Encoder‑Decoder, Encoder‑only, and Decoder‑only designs—while detailing how attention mechanisms work and why modern large‑language models favor the Decoder‑only approach, providing a concise framework for answering interview questions.

AI InterviewEncoder-DecoderLarge Language Model
0 likes · 10 min read
Why the Transformer Core Structure Is the Key to AI Interview Success
Data Party THU
Data Party THU
Oct 6, 2025 · Artificial Intelligence

How OneCAT Redefines Multimodal AI with a Decoder‑Only Architecture

OneCAT introduces a unified decoder‑only transformer that eliminates separate visual encoders, employs a modality‑specific MoE, integrates multi‑scale visual generation, and achieves state‑of‑the‑art performance and efficiency across multimodal understanding, text‑to‑image synthesis, and image editing tasks.

AI modelOneCATdecoder-only
0 likes · 14 min read
How OneCAT Redefines Multimodal AI with a Decoder‑Only Architecture
Baobao Algorithm Notes
Baobao Algorithm Notes
Jul 5, 2023 · Artificial Intelligence

Session‑Level Sample Organization for Decoder‑Only LLM Fine‑Tuning

This article explains how to restructure multi‑turn dialogue data into single session‑level training samples for decoder‑only large language models, leveraging causal attention and simple position IDs, and provides a concrete implementation, performance results, and a gradient‑weight analysis.

ChatGLM2LLM fine-tuningdecoder-only
0 likes · 7 min read
Session‑Level Sample Organization for Decoder‑Only LLM Fine‑Tuning