Tag

GPT-3

0 views collected around this technical thread.

DataFunTalk
DataFunTalk
Aug 23, 2023 · Artificial Intelligence

Evaluating Large Language Model Item Encoders for Textual Collaborative Filtering in Recommendation Systems

This article investigates whether replacing traditional ID-based item encoders with massive LLMs such as GPT‑3 improves recommendation performance, by conducting extensive experiments on three real‑world datasets, analyzing performance limits, generality of item representations, and comparing against ID‑based and prompt‑based methods.

AIGPT-3LLM
0 likes · 15 min read
Evaluating Large Language Model Item Encoders for Textual Collaborative Filtering in Recommendation Systems
Architect
Architect
Apr 14, 2023 · Artificial Intelligence

Overview of Prominent Large Language Models and Instruction Fine‑Tuning Techniques

The article surveys major large language models—including GPT‑3, T5, LaMDA, Jurassic‑1, MT‑NLG, Gopher, Chinchilla, PaLM, U‑PaLM, OPT, LLaMA, BLOOM, GLM‑130B, and ERNIE 3.0 Titan—explains their architectures, scaling trade‑offs, and then details instruction‑fine‑tuned variants such as T0, FLAN, GPT‑3.5, ChatGPT, GPT‑4, Alpaca and ChatGLM, providing references for further study.

AIChatGPTGPT-3
0 likes · 27 min read
Overview of Prominent Large Language Models and Instruction Fine‑Tuning Techniques
360 Tech Engineering
360 Tech Engineering
Mar 17, 2023 · Artificial Intelligence

Understanding ChatGPT: OpenAI’s Development, Model Evolution, and Training Techniques

This article provides an overview of ChatGPT’s rapid rise, OpenAI’s founding, the evolution of GPT models up to GPT‑3, the data‑driven training process, model capabilities and limitations, and practical guidance for users, highlighting the interplay between open‑source research and commercial deployment.

ChatGPTGPT-3Machine Learning
0 likes · 14 min read
Understanding ChatGPT: OpenAI’s Development, Model Evolution, and Training Techniques
DataFunSummit
DataFunSummit
Feb 19, 2023 · Artificial Intelligence

Understanding In-Context Learning in Large Language Models: Experiments, Analysis, and Theoretical Insights

This article explains the concept of in‑context learning in large language models, presents experimental evaluations such as copy‑output, date‑formatting, and label‑remapping tasks, and discusses a recent theoretical analysis that links attention layers to implicit gradient‑based fine‑tuning, highlighting why model scale and data volume matter.

GPT-3In-Context LearningMachine Learning
0 likes · 15 min read
Understanding In-Context Learning in Large Language Models: Experiments, Analysis, and Theoretical Insights