Tag

LoRA

1 views collected around this technical thread.

DataFunSummit
DataFunSummit
Jun 10, 2025 · Artificial Intelligence

How Quwan’s Kaitian Model Tackles Emotional AI for Social Apps – Architecture, Training Tricks, and Safety

Quwan Technology presents its Kaitian social large model, designed for personalized, emotionally rich, multimodal AI interactions, detailing its scene‑specific goals, CPT+SFT+RLHF training pipeline, data desensitization, LoRA fine‑tuning, evaluation methods, pruning, latency trade‑offs, safety mechanisms, and future feedback loops.

AI safetyLoRARLHF
0 likes · 13 min read
How Quwan’s Kaitian Model Tackles Emotional AI for Social Apps – Architecture, Training Tricks, and Safety
DaTaobao Tech
DaTaobao Tech
Jun 4, 2025 · Artificial Intelligence

Understanding Large Language Model Architecture, Parameters, Memory, Storage, and Fine‑Tuning Techniques

This article provides a comprehensive overview of large language models (LLMs), covering their transformer architecture, parameter counts, GPU memory and storage requirements, and detailed fine‑tuning methods such as prompt engineering, data construction, LoRA, PEFT, RLHF, and DPO, along with practical deployment and inference acceleration strategies.

DPOFine-tuningLLM
0 likes · 17 min read
Understanding Large Language Model Architecture, Parameters, Memory, Storage, and Fine‑Tuning Techniques
Architect
Architect
Apr 1, 2025 · Artificial Intelligence

When to Fine‑Tune Large Language Models vs. Relying on Prompting and RAG

The article explains why most projects should start with prompt engineering or simple agent workflows, outlines the scenarios where model fine‑tuning adds real value, compares fine‑tuning with Retrieval‑Augmented Generation, and offers practical criteria for deciding which approach to adopt.

AI deploymentLoRAPrompt Engineering
0 likes · 9 min read
When to Fine‑Tune Large Language Models vs. Relying on Prompting and RAG
Alibaba Cloud Infrastructure
Alibaba Cloud Infrastructure
Mar 18, 2025 · Cloud Native

Gray Release of LoRA and Base Models Using ACK Gateway with AI Extension on Kubernetes

This guide explains how to deploy large language model inference services on a GPU-enabled Kubernetes cluster, configure ACK Gateway with AI Extension for intelligent routing and load balancing, and perform gray releases for both LoRA fine‑tuned models and base models such as QwQ‑32B and DeepSeek‑R1, including step‑by‑step commands and validation procedures.

ACK GatewayAI inferenceCloud Native
0 likes · 25 min read
Gray Release of LoRA and Base Models Using ACK Gateway with AI Extension on Kubernetes
Cognitive Technology Team
Cognitive Technology Team
Feb 24, 2025 · Artificial Intelligence

Fine-Tuning Large Language Models with LoRA: A Step-by-Step Guide and Code Example

This article demonstrates the before-and-after effects of fine‑tuning a large language model, explains the concept with analogies, details hardware setup, dataset preparation, LoRA configuration, training arguments, and provides complete Python code for a pure‑framework fine‑tuning workflow.

HuggingFaceLLM fine-tuningLoRA
0 likes · 24 min read
Fine-Tuning Large Language Models with LoRA: A Step-by-Step Guide and Code Example
vivo Internet Technology
vivo Internet Technology
Feb 12, 2025 · Artificial Intelligence

Bidirectional Optimization of NLLB-200 and ChatGPT for Low-Resource Language Translation

The paper proposes a bidirectional optimization framework that fine‑tunes the low‑resource NLLB‑200 translation model with LoRA using data generated by ChatGPT, while also translating low‑resource prompts with NLLB before feeding them to LLMs, thereby improving multilingual translation quality yet requiring careful validation of noisy synthetic data.

Fine-tuningLLMLoRA
0 likes · 28 min read
Bidirectional Optimization of NLLB-200 and ChatGPT for Low-Resource Language Translation
360 Zhihui Cloud Developer
360 Zhihui Cloud Developer
Jan 9, 2025 · Artificial Intelligence

Unlocking Efficient Large Model Fine‑Tuning: LoRA, LoRA+, rsLoRA, DoRA & PiSSA Explained

This article introduces the fundamentals of large‑model fine‑tuning, compares popular parameter‑efficient methods such as LoRA and its variants, presents experimental results on the Qwen2.5‑7B model, and discusses current challenges and future research directions.

AI researchLoRAlarge model fine-tuning
0 likes · 17 min read
Unlocking Efficient Large Model Fine‑Tuning: LoRA, LoRA+, rsLoRA, DoRA & PiSSA Explained
DataFunSummit
DataFunSummit
Jan 6, 2025 · Artificial Intelligence

Efficient Large‑Model Training with LLaMA‑Factory: Overview, Techniques, and Applications

This article explains how to train large language models efficiently using LLaMA‑Factory, covering low‑resource training challenges, memory‑saving optimizations for parameters, gradients and activations, framework features, quick‑start guidance, performance tuning, real‑world case studies, and a detailed Q&A.

AIDeepSpeedLLaMA-Factory
0 likes · 10 min read
Efficient Large‑Model Training with LLaMA‑Factory: Overview, Techniques, and Applications
ZhongAn Tech Team
ZhongAn Tech Team
Nov 16, 2024 · Artificial Intelligence

Weekly AI Digest Issue 2: Video Generation, Large Models, AGI, and LoRA Fine‑Tuning

This weekly AI roundup discusses emerging video generation tools like PixelDance and Vidu 1.5, debates on scaling limits of large models, AGI geopolitical considerations, and a MIT study comparing LoRA with full fine‑tuning for domain adaptation.

AGIAIFine-tuning
0 likes · 8 min read
Weekly AI Digest Issue 2: Video Generation, Large Models, AGI, and LoRA Fine‑Tuning
Architecture and Beyond
Architecture and Beyond
Nov 2, 2024 · Artificial Intelligence

Step-by-Step Guide to Training a LoRA Model with Flux1_dev on ComfyUI

This tutorial walks programmers through preparing a GPU cloud environment, installing ComfyUI, downloading Flux1_dev models, integrating a custom LoRA, labeling generated images, and finally training the LoRA using ai‑toolkit, providing detailed commands, configuration tips, and practical cost estimates.

AI image generationComfyUIFlux
0 likes · 12 min read
Step-by-Step Guide to Training a LoRA Model with Flux1_dev on ComfyUI
DataFunSummit
DataFunSummit
Sep 23, 2024 · Artificial Intelligence

TransLLM: A Framework for Cross‑Language Transfer of Conversational Large Language Models

This article presents TransLLM, a cross‑language migration framework that enables high‑quality conversational LLMs to be transferred to low‑resource languages by preserving advanced capabilities through Recovery KD, LoRA‑based continual pre‑training, and a translation‑thinking‑chain, with extensive experiments showing superior performance and safety over ChatGPT and GPT‑4.

LoRASafetyconversation LLM
0 likes · 22 min read
TransLLM: A Framework for Cross‑Language Transfer of Conversational Large Language Models
JD Tech Talk
JD Tech Talk
Jul 9, 2024 · Artificial Intelligence

Getting Started with AI Image Generation Using Stable Diffusion for Promotional Posters

This guide introduces the fundamentals of AI image generation with Stable Diffusion, covering three main usage methods, the Draw Things desktop app, model types, samplers, prompts, and post‑processing techniques to create high‑quality promotional graphics for events like the 618 sale.

AI artDrawThingsLoRA
0 likes · 11 min read
Getting Started with AI Image Generation Using Stable Diffusion for Promotional Posters
Practical DevOps Architecture
Practical DevOps Architecture
Jun 28, 2024 · Artificial Intelligence

Large Model (LLM) Training Curriculum – Weekly Topics and Resources

This article outlines a five‑week large‑model training curriculum, detailing weekly topics such as transformer fundamentals, encoder‑decoder architectures, self‑attention, LoRA fine‑tuning, and quantization, along with associated video lectures and PDF slide decks for developers.

AILLMLoRA
0 likes · 3 min read
Large Model (LLM) Training Curriculum – Weekly Topics and Resources
58 Tech
58 Tech
Jun 3, 2024 · Artificial Intelligence

Parameter-Efficient Fine-Tuning (PEFT) Methods for Large Language Models: LoRA, QLoRA, AdaLoRA, SoRA, and Training Acceleration with Unsloth

This article systematically analyzes popular parameter‑efficient fine‑tuning (PEFT) techniques for large language models—including Adapter Tuning, Prefix Tuning, LoRA, QLoRA, AdaLoRA, and SoRA—detailing their principles, implementation code, experimental results on NLU tasks, and practical acceleration using the Unsloth library.

AdaLoRALoRAPEFT
0 likes · 39 min read
Parameter-Efficient Fine-Tuning (PEFT) Methods for Large Language Models: LoRA, QLoRA, AdaLoRA, SoRA, and Training Acceleration with Unsloth
Rare Earth Juejin Tech Community
Rare Earth Juejin Tech Community
May 31, 2024 · Artificial Intelligence

Generating Custom QA Datasets with Large Language Models and Fine‑Tuning via LoRA

This article explains how to use a large language model to automatically convert long‑form texts into Alpaca‑style question‑answer pairs, build a LangChain processing chain, and then fine‑tune a model such as Phi‑3‑mini‑4k‑instruct with LoRA, providing full Python code examples.

Dataset GenerationFine-tuningLLM
0 likes · 11 min read
Generating Custom QA Datasets with Large Language Models and Fine‑Tuning via LoRA
Sohu Tech Products
Sohu Tech Products
May 21, 2024 · Artificial Intelligence

OPPO Multimodal Pretrained Model Deployment in Cloud-Edge Scenarios: Practices and Optimizations

OPPO details how it deploys multimodal pretrained models on resource‑constrained edge devices by compressing CLIP‑based image‑text retrieval, adapting Chinese text‑to‑image generation with LoRA and adapters, and lightweighting diffusion models through layer pruning and progressive distillation, achieving sub‑3‑second generation while preserving cloud‑level quality.

ClipLoRAOPPO
0 likes · 18 min read
OPPO Multimodal Pretrained Model Deployment in Cloud-Edge Scenarios: Practices and Optimizations
OPPO Kernel Craftsman
OPPO Kernel Craftsman
Mar 29, 2024 · Artificial Intelligence

InternLM Model Research and XTuner Practical Guide (Part 1): DataLoader, Model Conversion, Merging, and Inference

The guide walks through fine‑tuning InternLM‑Chat‑7B with XTuner, showing how to build a DataLoader from a HuggingFace Dataset, convert a LoRA .pth checkpoint to HuggingFace format, merge the adapter into the base model, run inference, and adapt the process for custom datasets and 4‑bit quantization experiments.

DataLoaderFineTuningInternLM
0 likes · 27 min read
InternLM Model Research and XTuner Practical Guide (Part 1): DataLoader, Model Conversion, Merging, and Inference
DaTaobao Tech
DaTaobao Tech
Mar 6, 2024 · Artificial Intelligence

AI Clothing Graffiti Project: Implementation and Optimization of AIGC Technology in Taobao Life 2

The AI Clothing Graffiti Project in Taobao Life 2 leverages Stable Diffusion, ControlNet, and LoRA to let users generate and stylize clothing designs via text‑image prompts, employing parallel processing, face repair, and content filtering, and has launched successfully, inviting algorithm engineers to join the team.

AIAIGCComputer Vision
0 likes · 14 min read
AI Clothing Graffiti Project: Implementation and Optimization of AIGC Technology in Taobao Life 2
Ximalaya Technology Team
Ximalaya Technology Team
Feb 1, 2024 · Artificial Intelligence

Understanding AI Image Generation: Diffusion Models, CLIP, and Control Techniques

This guide explains how AI image generators such as Stable Diffusion and DALL·E 3 turn text prompts into pictures by using diffusion models, CLIP‑aligned embeddings, and optional controls like negative prompts, fine‑tuned LoRA checkpoints and ControlNet conditioning, highlighting their differences, workflow, and practical customization.

AI image generationClipControlNet
0 likes · 18 min read
Understanding AI Image Generation: Diffusion Models, CLIP, and Control Techniques