Industry Insights 10 min read

How SGLang’s $100M Seed Funding Powers the Next‑Gen Open AI Infrastructure

RadixArk raised a $100 million seed round backed by top hardware and AI investors to turn the open‑source SGLang inference engine and the Miles RL framework into day‑0 standards, aiming to democratize AI infrastructure and eliminate bottlenecks from training to inference.

Machine Heart
Machine Heart
Machine Heart
How SGLang’s $100M Seed Funding Powers the Next‑Gen Open AI Infrastructure

On May 5, AI‑infra startup RadixArk announced a $100 million seed round, valuing the company at $400 million post‑money. The round was led by Accel and Spark Capital and included investors such as NVIDIA’s NVentures, AMD, MediaTek, Broadcom, Intel, as well as industry figures like Intel CEO Chen Liwu and OpenAI co‑founder John Schulman.

SGLang, the open‑source LLM inference project launched in 2023, quickly became a de‑facto standard: it has amassed over 27 k GitHub stars, runs on more than 400 k GPUs, and processes tens of trillions of tokens daily for users including Google, Microsoft, NVIDIA, Oracle, AMD, LinkedIn, xAI and Thinking Machines Lab. The project maintains day‑0 compatibility across major model evolutions (MoE, long‑context, reasoning, multimodal) through innovations such as the RadixAttention prefix‑cache that organizes shared prefixes into a tree and deduplicates KV storage.

The founding team combines deep system and algorithm expertise: CEO Ying Sheng (Stanford PhD, former Databricks and xAI engineer) and CTO Banghua Zhu (UC Berkeley PhD, former Nvidia Principal Research Scientist). A hardware‑vendor technical lead described the team as “the most valuable founder combination in AI‑infra history.”

Beyond inference, RadixArk released the Miles reinforcement‑learning framework in November 2025, targeting large‑scale RL stability and efficiency; it is already used by over 20 teams for MoE model training. In April 2025 the team demonstrated simultaneous support for DeepSeek‑V4 inference and RL training, leveraging system‑level optimizations such as the ShadowRadix prefix cache, a single‑chip Flash Compressor, and a Lightning TopK that reduces latency to 15 µs, while bridging FP8 inference to BF16 training.

The investor mix signals a strategic bet: hardware giants recognize that a hardware‑agnostic, open‑source inference system can unlock higher performance across heterogeneous platforms, while AI leaders anticipate a unified training‑to‑inference infrastructure. RadixArk’s vision is to make AI infrastructure a public utility—reliable, widely accessible, and not locked behind private systems.

To reward the community, RadixArk offers free usage credits on its hosted platform to anyone who stars the SGLang GitHub repository. The company is also hiring across system, model, compiler, kernel, scheduler and evaluation domains.

Original Source

Signed-in readers can open the original source through BestHub's protected redirect.

Sign in to view source
Republication Notice

This article has been distilled and summarized from source material, then republished for learning and reference. If you believe it infringes your rights, please contactadmin@besthub.devand we will review it promptly.

Reinforcement LearningAI infrastructureSGLangDeepSeek V4RadixArkHardware‑agnostic AIOpen-source inference
Machine Heart
Written by

Machine Heart

Professional AI media and industry service platform

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.