Tag

pre‑training

1 views collected around this technical thread.

DataFunTalk
DataFunTalk
Jul 2, 2024 · Artificial Intelligence

Application of Large Language Models in Recommendation Systems: Overview and Future Directions

This article provides a comprehensive overview of how large language models (LLMs) are applied in recommendation systems, covering two main paradigms—LLM+RS as a component and LLM as a standalone recommender—detailing their impact on pre‑training, fine‑tuning, prompting, and future research challenges.

Fine-tuningFuture DirectionsLLM
0 likes · 6 min read
Application of Large Language Models in Recommendation Systems: Overview and Future Directions
DataFunSummit
DataFunSummit
Feb 17, 2024 · Artificial Intelligence

When to Pre‑Train Graph Neural Networks: Data‑Active Pre‑Training and a Graph Generator Framework

This article examines the conditions under which graph neural network pre‑training is beneficial, proposes a data‑centric generator framework to assess transferability, introduces a data‑active pre‑training strategy that selects informative graphs, and presents experimental results showing that using less, well‑chosen data can outperform full‑scale pre‑training.

Graph Neural Networksdata selectiongraph generator
0 likes · 16 min read
When to Pre‑Train Graph Neural Networks: Data‑Active Pre‑Training and a Graph Generator Framework
Rare Earth Juejin Tech Community
Rare Earth Juejin Tech Community
Jul 24, 2023 · Artificial Intelligence

Comprehensive Survey of Large Language Models: History, Key Technologies, Resources, and Future Directions

This article provides a detailed overview of large language models (LLMs), tracing their evolution from statistical and neural language models to modern pre‑trained transformers, discussing scaling, training, adaptation, utilization, evaluation methods, available resources, and outlining current challenges and future research directions.

AI researchModel ScalingPrompt Engineering
0 likes · 26 min read
Comprehensive Survey of Large Language Models: History, Key Technologies, Resources, and Future Directions
DataFunTalk
DataFunTalk
Feb 28, 2019 · Artificial Intelligence

A Comprehensive Introduction to BERT: Architecture, Pre‑training, and Implementation

This article provides an in‑depth overview of BERT, covering its NLP background, GLUE benchmark achievements, Transformer‑based architecture, pre‑training strategies (MLM and NSP), downstream fine‑tuning methods, and includes detailed PyTorch code implementations of its core components.

BERTNLPPyTorch
0 likes · 19 min read
A Comprehensive Introduction to BERT: Architecture, Pre‑training, and Implementation