Tag

transformers

0 views collected around this technical thread.

Rare Earth Juejin Tech Community
Rare Earth Juejin Tech Community
Feb 12, 2025 · Frontend Development

UnoCSS Installation, Basic Usage, Presets, Transformers, and Common Tips

This article provides a comprehensive guide to UnoCSS, covering installation in Vue 3 + Vite and Nuxt 3 projects, basic syntax and interactive documentation, Iconify SVG integration, various presets and transformers, as well as practical shortcuts, responsive design, safelist handling, custom rules, theming, and dark‑mode support.

CSS UtilityNuxtPresets
0 likes · 22 min read
UnoCSS Installation, Basic Usage, Presets, Transformers, and Common Tips
DaTaobao Tech
DaTaobao Tech
Nov 13, 2024 · Artificial Intelligence

Understanding Neural Networks and Transformers: Principles, Implementation, and Applications

The article surveys neural networks from basic neuron operations and loss functions through deep architectures to the Transformer model, detailing embeddings, positional encoding, self‑attention, multi‑head attention, residual links, and encoder‑decoder design, and includes PyTorch code examples for linear regression, translation, and fine‑tuning Hugging Face’s MiniRBT for text classification.

AINLPPyTorch
0 likes · 44 min read
Understanding Neural Networks and Transformers: Principles, Implementation, and Applications
AntTech
AntTech
Nov 13, 2024 · Artificial Intelligence

Nimbus: Secure and Efficient Two‑Party Inference for Transformers

The article introduces Nimbus, a novel two‑party privacy‑preserving inference framework for Transformer models that accelerates linear‑layer matrix multiplication and activation‑function evaluation through an outer‑product encoding and distribution‑aware polynomial approximation, achieving 2.7‑4.7× speedup over prior work while maintaining model accuracy.

Privacy-Preserving AIcryptographysecret sharing
0 likes · 6 min read
Nimbus: Secure and Efficient Two‑Party Inference for Transformers
DaTaobao Tech
DaTaobao Tech
May 27, 2024 · Artificial Intelligence

Sampling Strategies for Large Language Models: Greedy, Beam, Top‑K, Top‑p, and Temperature

The article explains how greedy search, beam search, Top‑K, Top‑p (nucleus) sampling, and temperature each shape large language model generation, comparing their effects on repetition, diversity, and creativity, and provides concise TensorFlow‑based code examples illustrating these inference‑time strategies.

AIGenerationLLM
0 likes · 15 min read
Sampling Strategies for Large Language Models: Greedy, Beam, Top‑K, Top‑p, and Temperature
Rare Earth Juejin Tech Community
Rare Earth Juejin Tech Community
Jul 22, 2023 · Artificial Intelligence

Building an Image Classification Model with Transformers and TensorFlow: Theory, Code, and Practice

This article explains how to leverage computer‑vision techniques and deep‑learning frameworks such as Transformers and TensorFlow to build a complete image‑classification pipeline, covering the underlying RGB and CNN principles, model architecture, data preparation, training, and inference with runnable Python code.

CNNPythonTensorFlow
0 likes · 15 min read
Building an Image Classification Model with Transformers and TensorFlow: Theory, Code, and Practice
Architect
Architect
Jul 1, 2023 · Artificial Intelligence

Comprehensive Guide to Text Generation Decoding Strategies with HuggingFace Transformers

This tutorial explores various text generation decoding methods—including greedy search, beam search, top‑k/top‑p sampling, sample‑and‑rank, and group beam search—explaining their principles, providing detailed Python code examples, and comparing their use in modern large language models.

HuggingFacebeam searchdecoding strategies
0 likes · 59 min read
Comprehensive Guide to Text Generation Decoding Strategies with HuggingFace Transformers
Tencent Cloud Developer
Tencent Cloud Developer
Jun 1, 2023 · Artificial Intelligence

A Comprehensive Guide to Decoding Strategies for Text Generation with HuggingFace Transformers

This guide thoroughly explains the major decoding strategies for neural text generation in HuggingFace Transformers—including greedy, beam, diverse beam, sampling, top‑k, top‑p, sample‑and‑rank, beam sampling, and group beam search—detailing their principles, Python implementations with LogitsProcessor components, workflow diagrams, comparative analysis, and references to original research.

HuggingFaceNatural Language Processingbeam search
0 likes · 60 min read
A Comprehensive Guide to Decoding Strategies for Text Generation with HuggingFace Transformers
Architect
Architect
Apr 24, 2023 · Artificial Intelligence

MOSS 003: Open‑Source Large Language Model Development, Training Data, and Plugin‑Enabled Deployment

The article details the evolution of the open‑source MOSS series—from OpenChat 001 to MOSS 003—covering data collection, fine‑tuning procedures, multilingual capabilities, plugin architecture, example code for inference, and upcoming releases, providing a comprehensive technical overview for AI practitioners.

AIMOSSlarge language model
0 likes · 11 min read
MOSS 003: Open‑Source Large Language Model Development, Training Data, and Plugin‑Enabled Deployment
Architect
Architect
Feb 19, 2023 · Artificial Intelligence

Training a Positive Review Generator with RLHF and PPO

This article demonstrates how to apply Reinforcement Learning from Human Feedback (RLHF) using a sentiment‑analysis model as a reward function and Proximal Policy Optimization (PPO) to fine‑tune a language model that generates positive product reviews, complete with code snippets and experimental results.

PPORLHFlanguage model
0 likes · 10 min read
Training a Positive Review Generator with RLHF and PPO
DataFunTalk
DataFunTalk
Jan 3, 2022 · Artificial Intelligence

Top AI Stories of 2021: Large‑Scale Pretrained Models, Transformers, Multimodal AI, and Emerging Challenges

The article reviews the 2021 AI landscape, highlighting the race for ever‑larger pretrained models, the dominance of Transformers across modalities, the promise and limits of large models, the rise of multimodal systems, regulatory considerations, and the still‑nascent progress in reinforcement learning.

AI Industryai governancelarge language models
0 likes · 12 min read
Top AI Stories of 2021: Large‑Scale Pretrained Models, Transformers, Multimodal AI, and Emerging Challenges