YOLO-Master v2026.02 Unveils Four Innovations for SOTA Object Detection

Tencent’s YOLO-Master v2026.02 adds a Mixture‑of‑Experts architecture, zero‑overhead LoRA fine‑tuning, Sparse SAHI inference for large images, and Cluster‑Weighted NMS, delivering 3‑5× faster inference, up to 70% reduced training resources, and markedly higher detection accuracy across diverse benchmarks.

AIWalker
AIWalker
AIWalker
YOLO-Master v2026.02 Unveils Four Innovations for SOTA Object Detection

Introduction

Tencent officially released YOLO-Master v2026.02, the latest milestone in its open‑source YOLO series. The update targets a better balance between model efficiency and architectural flexibility for computer‑vision researchers and engineers.

YOLO-Master v2026.02 banner
YOLO-Master v2026.02 banner

Four Core Innovations

Mixture of Experts (MoE) : dynamic expert activation expands model capacity without extra compute cost.

Low‑Rank Adaptation (LoRA) : fine‑tunes only 1‑5% of parameters while retaining >95% of full‑parameter performance.

Sparse SAHI : intelligent adaptive slicing skips empty regions, accelerating inference 3‑5× on 4K/8K images.

Cluster‑Weighted NMS (CW‑NMS) : Gaussian‑weighted box fusion improves localization precision over traditional NMS.

1. Mixture of Experts (MoE)

MoE introduces conditional computation via ES‑MoE blocks, enabling "compute‑on‑demand" routing. The implementation includes a dedicated MoE loss (Load Balancing Loss to keep expert usage even, Z‑Loss for logit stability) and an adaptive weight‑adjustment mechanism that balances primary and auxiliary losses.

Smart pruning ( MoEPruner) analyses validation‑set expert utilization and automatically removes experts with utilization below a default 15% threshold, yielding 20‑30% inference acceleration while preserving accuracy.

The MoE router, expert networks, and gating are fully decoupled, supporting routing strategies such as Top‑K, Soft Routing, and Expert Choice, which simplifies custom expert integration.

YOLO-Master introduces ES‑MoE blocks to achieve "compute‑on‑demand" via dynamic routing.
MoE architecture diagram
MoE architecture diagram

2. Low‑Rank Adaptation (LoRA)

LoRA provides a zero‑architecture‑overhead adaptation: LoRA adapters are applied at the parameter level rather than inserting new modules, so the original YOLO backbone remains unchanged. Activation is driven entirely by configuration parameters, allowing seamless switching between LoRA‑enabled and standard training modes.

The three‑step activation process is:

Dynamic Weight Interception : LoRA adapters modify weights directly.

Configuration‑Driven Activation : behavior is toggled via hyper‑parameters.

Backward Compatibility : the base model stays intact, enabling easy fallback.

Consequently, users only need to set a few config flags to start LoRA training without any code changes.

# Example LoRA configuration
lora_r = 8-16   # small models
lora_r = 16-32  # medium models
lora_r = 32-64  # large models
lora_alpha = 2 * lora_r   # typical
lora_alpha = 4 * lora_r   # aggressive fine‑tuning
LoRA adapter illustration
LoRA adapter illustration

3. Sparse SAHI

Sparse SAHI tackles ultra‑large image detection (4K/8K) by generating a low‑resolution objectness heatmap, adaptively slicing only regions with probability >0.15, running high‑resolution inference on those slices, and finally merging results with CW‑NMS.

Objectness Heatmap Generation : low‑res full‑image inference produces a probability map.

Adaptive Slicing : smart cuts skip low‑probability areas.

High‑Resolution Inference : detailed processing on selected patches.

Result Merging : CW‑NMS fuses detections across slices.

This pipeline yields a 3‑5× speed boost while preserving high detection precision, making it ideal for satellite imagery, autonomous‑driving perception, and industrial inspection.

Sparse SAHI achieves 3‑5× inference acceleration on large‑scale images.
Sparse SAHI workflow
Sparse SAHI workflow

4. Cluster‑Weighted NMS (CW‑NMS)

CW‑NMS replaces the hard‑threshold suppression of traditional NMS with a Gaussian‑weighted averaging of overlapping boxes. The weighted box is computed as:

weighted_box = Σ(box_i × w_i) / Σ(w_i)
where w_i = exp(-IoU²/2σ²) × conf_i

This approach retains more candidates in dense scenes, delivering higher mAP while adding only a modest computational overhead.

CW‑NMS formula illustration
CW‑NMS formula illustration

Performance Benchmarks

Extensive ablation on the YOLOv11 family demonstrates the efficiency of LoRA. For example, YOLO11x (56.9 M parameters, 114.6 MB) requires only a 14.1 MB adapter (≈6.20% of total parameters) to achieve comparable accuracy, cutting GPU memory usage by ~70%.

YOLO11n: 2.6 M params, LoRA 527 k (20.29%); 70% memory reduction.

YOLO11s: 9.4 M params, LoRA 1.02 M (10.81%); 70% memory reduction.

YOLO11m: 20.1 M params, LoRA 1.64 M (8.16%); 70% memory reduction.

YOLO11l: 25.3 M params, LoRA 2.35 M (9.29%); 70% memory reduction.

YOLO11x: 56.9 M params, LoRA 3.53 M (6.20%); 70% memory reduction.

Storage compression ratios reach up to 8.13× for the largest model (YOLO11x), enabling cloud deployment with only 14.1 MB adapters instead of the full 114.6 MB model, saving ~88% of storage and transfer costs.

Application Scenarios & Best Practices

Resource‑constrained environments : LoRA rank and alpha settings can be tuned per model size (see code snippet above) to balance performance and efficiency.

Ultra‑large image detection :

results = model.predict(
    source="large_aerial_image.jpg",
    sparse_sahi=True,
    slice_size=640,
    overlap_ratio=0.2,
    objectness_threshold=0.15,
    conf=0.25,
    iou=0.45
)

Dense object scenarios (crowd counting, cell detection) :

results = model.predict(
    source="dense_objects.jpg",
    cluster=True,   # enables CW‑NMS
    sigma=0.1,      # Gaussian weight std
    conf=0.25,
    max_det=300
)

Conclusion & Outlook

YOLO-Master v2026.02 marks a significant step forward for object detection, delivering scalable capacity via MoE, resource‑efficient fine‑tuning with LoRA, fast large‑image inference through Sparse SAHI, and superior localization with CW‑NMS. As the open‑source community around YOLO‑Master grows, these innovations are poised to become standard tools in the computer‑vision toolbox.

Model performance leaderboard
Model performance leaderboard
Additional benchmark charts
Additional benchmark charts
Further results
Further results

Related Links

GitHub repository: https://github.com/Tencent/YOLO-Master

Latest release: https://github.com/Tencent/YOLO-Master/releases/tag/YOLO-Master-v26.02

Hugging Face model hub: https://huggingface.co/gatilin/YOLO-Master-ckpts-v0

Additional HF checkpoint: https://huggingface.co/gatilin/YOLO-Master-ckpts-v0_1

computer visionModel OptimizationLoRAMixture of ExpertsYOLOSparse Inference
AIWalker
Written by

AIWalker

Focused on computer vision, image processing, color science, and AI algorithms; sharing hardcore tech, engineering practice, and deep insights as a diligent AI technology practitioner.

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.