How MiniCPM-V 4.6 Achieves Lightning‑Fast Multimodal AI on Smartphones (Open‑Source)
MiniCPM-V 4.6 combines a SigLIP2 visual encoder with a Qwen3.5 LLM, cuts FLOPs by over 50%, lowers token cost up to 43×, scores 13 on the Artificial Analysis Intelligence Index, and runs with 75 ms first‑token latency on 3136×3136 images across iOS, Android and HarmonyOS, all with fully open‑source code and extensive quantization support.
MiniCPM‑V 4.6 is an open‑source multimodal model that combines a SigLIP2‑400M visual encoder with a Qwen3.5‑0.8B large language model. The development team deeply re‑engineered the internal computation flow to cut compute demand.
Compute‑efficiency breakthrough
On the Artificial Analysis Intelligence Index benchmark the model scores 13 points, higher than the same‑size Qwen3.5‑0.8B (10) and its Thinking variant (11). Compared with those baselines, MiniCPM‑V 4.6 reduces token‑processing cost by 19× and 43× respectively and also outperforms the larger 3 B Ministral model (score 11). The improvement stems from adopting the latest LLaVA‑UHD v4 techniques, which lower visual‑encoding FLOPs by more than 50 %. The reduced arithmetic load translates into higher energy efficiency and a token‑throughput increase of roughly 1.5× over Qwen3.5‑0.8B. First‑token latency on a 3136×3136 image is 75 ms.
Multimodal capabilities
The model retains the series’ strength in single‑image analysis, multi‑image fusion, and dynamic video understanding. Across OpenCompass, RefCOCO, HallusionBench, MUIRBench and OCRBench the model consistently surpasses Qwen3.5‑0.8B and reaches the capability level of a 2 B‑scale model. To balance speed and visual fidelity, MiniCPM‑V 4.6 introduces a mixed visual‑token compression ratio ranging from 4× to 16×, allowing runtime selection between higher quality analysis and faster response.
Edge deployment
MiniCPM‑V 4.6 is fully compatible with iOS, Android and HarmonyOS. All adaptation code is released under an open‑source license, enabling developers to follow the provided guide and reproduce smooth AI interactions on phones or tablets.
Ecosystem integration and quantization
Inference support includes vLLM, SGLang, llama.cpp and Ollama. Fine‑tuning ecosystems such as SWIFT and LLaMA‑Factory are also compatible. To accommodate diverse hardware, the release provides quantized model files in GGUF, BNB, AWQ and GPTQ formats.
Model hub: https://huggingface.co/openbmb/MiniCPM-V-4.6
Source code: https://github.com/OpenBMB/MiniCPM-V
Mobile‑app repository: https://github.com/OpenBMB/MiniCPM-V-Apps
Signed-in readers can open the original source through BestHub's protected redirect.
This article has been distilled and summarized from source material, then republished for learning and reference. If you believe it infringes your rights, please contactand we will review it promptly.
SuanNi
A community for AI developers that aggregates large-model development services, models, and compute power.
How this landed with the community
Was this worth your time?
0 Comments
Thoughtful readers leave field notes, pushback, and hard-won operational detail here.
