Tag

language models

0 views collected around this technical thread.

DataFunSummit
DataFunSummit
Jul 22, 2024 · Artificial Intelligence

From BERT to LLM: Language Model Applications in 360 Advertising Recommendation

This talk explores how 360's advertising recommendation system leverages language models—from BERT to large‑scale LLMs—to improve user interest modeling, feature extraction, and conversion‑rate prediction, detailing practical challenges, engineering solutions, experimental results, and future research directions.

BERTLLMRecommendation systems
0 likes · 18 min read
From BERT to LLM: Language Model Applications in 360 Advertising Recommendation
Sohu Tech Products
Sohu Tech Products
Mar 20, 2024 · Artificial Intelligence

Comparison of Base LLM and Instruction Tuned LLM

The diagram contrasts a Base LLM, which merely predicts the next word from training data and can continue stories or answer simple facts but may generate unsafe text, with an Instruction‑Tuned LLM that is fine‑tuned via RLHF to understand and follow commands, delivering more accurate, useful, and safe responses.

AIAI applicationsBASE model
0 likes · 7 min read
Comparison of Base LLM and Instruction Tuned LLM
php中文网 Courses
php中文网 Courses
Nov 30, 2023 · Information Security

ChatGPT Repeat Prompt Vulnerability Exposes Sensitive Personal Information

Researchers discovered that prompting ChatGPT with repeated words can cause the model to leak private data such as phone numbers and email addresses, highlighting a serious repeat‑prompt vulnerability that reveals substantial personally identifiable information from its training corpus.

ChatGPTPIIarXiv
0 likes · 3 min read
ChatGPT Repeat Prompt Vulnerability Exposes Sensitive Personal Information
DataFunSummit
DataFunSummit
Nov 16, 2023 · Artificial Intelligence

Application of Language Models in Molecular Structure Prediction

This talk presents how large language models are leveraged for predicting protein, antibody, and RNA structures, covering background, model stability, generative approaches, antibody-specific models, RNA modeling, and protein‑RNA interaction prediction, along with experimental results and future research directions.

AI for biologyGenerative ModelsRNA modeling
0 likes · 17 min read
Application of Language Models in Molecular Structure Prediction
Architect
Architect
Oct 12, 2023 · Artificial Intelligence

Evolution of Language Models: From Statistical N‑grams to GPT‑4

This article provides a comprehensive overview of natural language processing and language‑model research, tracing the historical development from early rule‑based and statistical N‑gram models through neural network approaches such as RNN, LSTM, ELMo, and Transformer, and detailing the architectures, strengths, and limitations of the GPT series up to GPT‑4, while also discussing evaluation metrics, practical applications, and future challenges.

Artificial IntelligenceGPTNLP
0 likes · 34 min read
Evolution of Language Models: From Statistical N‑grams to GPT‑4
Zhuanzhuan Tech
Zhuanzhuan Tech
Sep 28, 2023 · Artificial Intelligence

Evolution of Language Models and an Overview of the GPT Series

This article surveys the development of natural language processing from early rule‑based systems through statistical n‑gram models, neural language models, RNNs, LSTMs, ELMo, Transformers and BERT, and then details the architecture, training methods, advantages and limitations of the GPT‑1, GPT‑2, GPT‑3, ChatGPT and GPT‑4 models, concluding with a discussion of future challenges and references.

Artificial IntelligenceGPTNLP
0 likes · 30 min read
Evolution of Language Models and an Overview of the GPT Series
Rare Earth Juejin Tech Community
Rare Earth Juejin Tech Community
Aug 1, 2023 · Artificial Intelligence

Do Language Models Learn Language in the Same Stages as Children? An Analysis of GPT‑2 Developmental Trajectories

This article reviews a study that compares the stage‑wise language acquisition of infants with the learning trajectory of GPT‑2, using linguistic probes and statistical tests to determine whether deep language models follow sequential or parallel learning patterns similar to children.

AI researchGPT-2developmental learning
0 likes · 17 min read
Do Language Models Learn Language in the Same Stages as Children? An Analysis of GPT‑2 Developmental Trajectories
360 Quality & Efficiency
360 Quality & Efficiency
Jul 21, 2023 · Artificial Intelligence

Prompt Engineering: Principles, Design Guidelines, and Practical Use Cases with ChatGPT

This article introduces prompt engineering for ChatGPT, explains key design principles, and demonstrates a series of practical applications such as text classification, summarization, role‑playing, terminal emulation, output formatting, temperature control, iterative fine‑tuning, and reverse‑engineering of prompts.

AI Prompt DesignChatGPTPrompt Engineering
0 likes · 8 min read
Prompt Engineering: Principles, Design Guidelines, and Practical Use Cases with ChatGPT
Sohu Tech Products
Sohu Tech Products
Jul 19, 2023 · Artificial Intelligence

Understanding the Inner Workings of ChatGPT and Neural Networks

This article explains how ChatGPT generates text by predicting the next token using large language models, describes the role of probability, temperature, and attention mechanisms in transformers, and discusses neural network training, embeddings, semantic spaces, and the broader implications for artificial intelligence research.

Artificial IntelligenceChatGPTlanguage models
0 likes · 79 min read
Understanding the Inner Workings of ChatGPT and Neural Networks
Airbnb Technology Team
Airbnb Technology Team
May 23, 2023 · Artificial Intelligence

Applying Text Generation Models to Scalable Customer Support at Airbnb

Airbnb replaced its XLM‑RoBERTa ranking with an MT5 encoder‑decoder for content recommendation, built a real‑time generative assistant for reply suggestions and intent detection, and deployed a T5‑based paraphrase chatbot, showing that large‑scale pre‑trained transformers improve relevance, agent efficiency, and user satisfaction.

AIAirbnbCustomer Support
0 likes · 12 min read
Applying Text Generation Models to Scalable Customer Support at Airbnb
Python Programming Learning Circle
Python Programming Learning Circle
Mar 17, 2023 · Artificial Intelligence

Analysis of New Bing’s Behavior Compared to ChatGPT: Issues, User Experiences, and Underlying AI Models

The article examines the public testing of the new Bing chatbot, contrasting its internet‑enabled, citation‑rich responses and occasional erratic, immature behavior with ChatGPT’s more stable output, while exploring user‑reported failures, speculative technical reasons, and the ethical implications of deploying advanced language models.

AI behaviorBingChatGPT
0 likes · 8 min read
Analysis of New Bing’s Behavior Compared to ChatGPT: Issues, User Experiences, and Underlying AI Models
Architect
Architect
Feb 13, 2023 · Artificial Intelligence

Understanding InstructGPT and ChatGPT: Architecture, Training Pipeline, and Performance Analysis

This article provides a comprehensive overview of the GPT series and explains how InstructGPT and ChatGPT are built by combining supervised fine‑tuning, reward modeling, and Proximal Policy Optimization, detailing their datasets, training pipeline, performance advantages, limitations, and future research directions.

AIChatGPTGPT
0 likes · 21 min read
Understanding InstructGPT and ChatGPT: Architecture, Training Pipeline, and Performance Analysis
Architect's Guide
Architect's Guide
Feb 9, 2023 · Artificial Intelligence

Why ChatGPT Is So Powerful: A Technical Overview of NLP Model Evolution

This article explains why ChatGPT performs so well by tracing the evolution of natural‑language processing from rule‑based grammars through statistical n‑gram models to neural architectures like RNNs, LSTMs, attention mechanisms, Transformers, and the massive data and training methods that power modern large language models.

AttentionChatGPTNLP
0 likes · 14 min read
Why ChatGPT Is So Powerful: A Technical Overview of NLP Model Evolution
Architect
Architect
Feb 6, 2023 · Artificial Intelligence

Understanding How ChatGPT Works: RLHF, PPO, and Consistency Challenges

This article explains the underlying mechanisms of ChatGPT, including its GPT‑3 foundation, the role of supervised fine‑tuning, human‑feedback reinforcement learning (RLHF), PPO optimization, consistency issues, evaluation metrics, and the limitations of these training strategies, with references to key research papers.

AI alignmentChatGPTPPO
0 likes · 16 min read
Understanding How ChatGPT Works: RLHF, PPO, and Consistency Challenges
DataFunSummit
DataFunSummit
Jan 14, 2023 · Artificial Intelligence

Key Transformer Model Papers Across Language, Vision, Speech, and Time‑Series Domains

This article surveys the most influential Transformer‑based research papers—from the original Attention Is All You Need work to recent models such as Autoformer and FEDformer—covering breakthroughs in natural language processing, computer vision, speech recognition, and long‑term series forecasting, and provides download links for each.

AISpeech RecognitionTransformer
0 likes · 17 min read
Key Transformer Model Papers Across Language, Vision, Speech, and Time‑Series Domains
DataFunSummit
DataFunSummit
Dec 28, 2022 · Artificial Intelligence

OpenAI’s Language Model Evolution Toward AGI

This article traces OpenAI’s progression from GPT‑1 through GPT‑3, Codex, InstructGPT, and ChatGPT, highlighting how increasing model scale, prompt‑based task integration, and human‑feedback alignment have driven the evolution toward more capable, generalizable language intelligence aimed at achieving artificial general intelligence.

AGIAIChatGPT
0 likes · 15 min read
OpenAI’s Language Model Evolution Toward AGI
政采云技术
政采云技术
Jul 5, 2022 · Artificial Intelligence

Overview of Natural Language Processing Techniques and Their Evolution

This article provides a comprehensive overview of natural language processing, covering its definition, historical development from one‑hot encoding to modern models such as word2vec, ELMo, GPT, and BERT, and discusses the advantages, limitations, and key concepts of each technique.

Artificial IntelligenceNLPWord Embedding
0 likes · 23 min read
Overview of Natural Language Processing Techniques and Their Evolution
DataFunSummit
DataFunSummit
Jan 25, 2022 · Artificial Intelligence

Intelligent Lyric Generation for Music: Techniques, Models, and Future Directions

This article explores how AI and natural language processing technologies are applied to music lyric creation, covering background challenges, rhyme retrieval methods, advanced language models such as SongNet, decoding strategies, style transfer, and a multi‑level generation platform that aims to streamline professional songwriting.

AI lyric generationMusic AINatural Language Processing
0 likes · 14 min read
Intelligent Lyric Generation for Music: Techniques, Models, and Future Directions
DataFunSummit
DataFunSummit
Nov 14, 2021 · Artificial Intelligence

Overview of Pre‑training Models and the UER‑py Framework for Natural Language Processing

This article introduces the importance of pre‑training in natural language processing, reviews classic pre‑training models such as Skip‑thoughts, BERT, GPT‑2 and T5, presents the modular UER‑py framework and its Chinese resources, compares it with Huggingface Transformers, and outlines practical deployment steps in industry.

NLPPretrainingTransformer
0 likes · 22 min read
Overview of Pre‑training Models and the UER‑py Framework for Natural Language Processing