Tag

knowledge transfer

1 views collected around this technical thread.

Model Perspective
Model Perspective
Jun 14, 2025 · Fundamentals

Unlocking the True Power of the Learning Compound Effect

The article explains the true nature of the compound effect in learning, outlining its financial origins, three essential elements, why many potential compounds fail, and four mechanisms—continuous use, transfer, feedback, and connection—that enable knowledge to generate lasting, exponential growth.

compound effectcontinuous improvementeducational design
0 likes · 9 min read
Unlocking the True Power of the Learning Compound Effect
Architect's Guide
Architect's Guide
May 13, 2025 · Artificial Intelligence

DeepSeek Model Distillation Technology: Overview, Innovations, Architecture, Training, Performance, and Challenges

This article provides a comprehensive overview of DeepSeek's model distillation technology, detailing its definition, key innovations, architecture, training methods, performance gains, and the remaining challenges such as the implicit performance ceiling and multimodal data distillation.

AI optimizationDeepSeekModel Distillation
0 likes · 14 min read
DeepSeek Model Distillation Technology: Overview, Innovations, Architecture, Training, Performance, and Challenges
Top Architect
Top Architect
Feb 14, 2025 · Artificial Intelligence

DeepSeek Model Distillation: Principles, Innovations, Architecture, and Performance

This article provides an in‑depth overview of DeepSeek’s model distillation technology, covering its definition, core principles, innovative data‑model distillation integration, architecture design, training strategies, performance gains, and the challenges of scaling to multimodal data.

AI optimizationDeepSeekModel Distillation
0 likes · 16 min read
DeepSeek Model Distillation: Principles, Innovations, Architecture, and Performance
IT Architects Alliance
IT Architects Alliance
Feb 10, 2025 · Artificial Intelligence

DeepSeek Distillation Technology: Principles, Innovations, Performance, and Future Outlook

The article explains DeepSeek's model distillation technique, covering its fundamental knowledge‑transfer principles, unique innovations such as data‑model fusion and task‑specific strategies, impressive benchmark results, practical applications in edge and online inference, existing challenges, and future research directions.

AI optimizationEdge ComputingModel Distillation
0 likes · 15 min read
DeepSeek Distillation Technology: Principles, Innovations, Performance, and Future Outlook
DataFunTalk
DataFunTalk
Nov 17, 2024 · Artificial Intelligence

Federated Learning and Data Security in the Era of Large Models: Research Overview and the FLAIR Platform

This presentation reviews recent research on data security and utilization in the large‑model era, covering privacy‑preserving federated learning, knowledge‑transfer techniques, prototype‑based modeling, multi‑model fusion methods such as FuseGen, and introduces the federated knowledge computing platform FLAIR for both horizontal and vertical federated scenarios.

FLAIRFederated LearningLarge Models
0 likes · 19 min read
Federated Learning and Data Security in the Era of Large Models: Research Overview and the FLAIR Platform
DataFunTalk
DataFunTalk
Jul 9, 2024 · Artificial Intelligence

Graph Knowledge Transfer and the Knowledge Bridge Learning Framework

This article presents an overview of graph knowledge transfer, discussing the data‑hungry problem, distribution shift in graph data, the Knowledge Bridge Learning (KBL) paradigm, the Bridged‑GNN implementation, experimental results across multiple scenarios, and future research directions.

Graph Neural Networksbridged-GNNdomain adaptation
0 likes · 19 min read
Graph Knowledge Transfer and the Knowledge Bridge Learning Framework
DataFunSummit
DataFunSummit
Apr 28, 2024 · Artificial Intelligence

Graph Knowledge Transfer: Methods, Practices, and the Knowledge Bridge Learning Framework

This article presents a comprehensive overview of graph knowledge transfer, covering its definition, the data‑hungry problem, distribution shift challenges, the Knowledge Bridge Learning (KBL) framework, the Bridged‑GNN model, extensive experiments on real‑world scenarios, and a concluding Q&A session.

Graph Neural Networksdomain adaptationgraph learning
0 likes · 22 min read
Graph Knowledge Transfer: Methods, Practices, and the Knowledge Bridge Learning Framework
DataFunTalk
DataFunTalk
Feb 2, 2024 · Artificial Intelligence

Utilizing Negative Samples for Knowledge Distillation of Large Language Models

This paper presents a novel framework that leverages negative samples during large language model distillation through three stages—Negative Assistive Training, Negative Calibration Enhancement, and Adaptive Self‑Consistency—demonstrating significant accuracy gains on challenging mathematical reasoning benchmarks and improved generalization to out‑of‑distribution tasks.

Chain-of-ThoughtLLM distillationknowledge transfer
0 likes · 13 min read
Utilizing Negative Samples for Knowledge Distillation of Large Language Models
DataFunTalk
DataFunTalk
May 24, 2023 · Artificial Intelligence

Graph Transfer Learning and VS-Graph: Knowledge Transferable Graph Neural Networks

This article reviews recent advances in graph transfer learning, introduces the novel VS-Graph scenario for knowledge transfer between dominant and silent nodes, and details the Knowledge Transferable Graph Neural Network (KTGNN) framework with domain‑adaptive feature completion, message passing, and transferable classifier modules, highlighting experimental results and future research directions.

AIGraph Neural NetworksNode Classification
0 likes · 27 min read
Graph Transfer Learning and VS-Graph: Knowledge Transferable Graph Neural Networks