Machine Heart
Author

Machine Heart

Professional AI media and industry service platform

207
Articles
0
Likes
3
Views
0
Comments
Recent Articles

Latest from Machine Heart

100 recent articles max
Machine Heart
Machine Heart
Apr 18, 2026 · Artificial Intelligence

Why Embodied Data Is the Biggest Gold Mine: Inside the World’s First Hundred‑Billion‑Scale Multimodal Data Cloud Mall

Paxini, together with JD Cloud, Tencent Cloud, and Baidu Intelligent Cloud, launches the world’s first hundred‑billion‑scale, full‑modal, high‑degree‑of‑freedom embodied AI data cloud mall, offering instant online data procurement, end‑to‑end model training pipelines, and validated performance gains in both lab and real‑world robot tasks.

Roboticsdata cloud marketplaceembodied AI
0 likes · 13 min read
Why Embodied Data Is the Biggest Gold Mine: Inside the World’s First Hundred‑Billion‑Scale Multimodal Data Cloud Mall
Machine Heart
Machine Heart
Apr 17, 2026 · Artificial Intelligence

Can π0.7 Unlock Compositional Generalization and Cross‑Embodiment Transfer for VLA?

The new π0.7 model from Physical Intelligence demonstrates emergent compositional generalization and cross‑embodiment transfer in visual‑language‑action (VLA) robots by leveraging massive heterogeneous data and richly structured prompts, outperforming specialist Recap models on tasks such as air‑fryer cooking, clothing folding, and coffee making.

RoboticsVLAcompositional generalization
0 likes · 11 min read
Can π0.7 Unlock Compositional Generalization and Cross‑Embodiment Transfer for VLA?
Machine Heart
Machine Heart
Apr 17, 2026 · Artificial Intelligence

Combining Transformers and RNNs: Google’s Memory Caching Unlocks Ultra‑Long Context

Google Research introduces Memory Caching (MC), a technique that gives RNNs growing memory capacity, bridging the gap with Transformers to enable ultra‑long context processing while reducing memory demands, and demonstrates its effectiveness through extensive language‑modeling and recall experiments.

AI ArchitectureGoogle ResearchMemory Caching
0 likes · 7 min read
Combining Transformers and RNNs: Google’s Memory Caching Unlocks Ultra‑Long Context
Machine Heart
Machine Heart
Apr 17, 2026 · Artificial Intelligence

Can LLMs Truly Mimic Human Shopping Behavior? The OPeRA Dataset and Evaluation

The paper introduces OPeRA, a step‑wise online‑shopping dataset capturing observations, personas, rationales, and actions from real users, and uses it to benchmark LLMs on next‑action prediction, revealing that even top models like GPT‑4.1 achieve only about 20 % accuracy on fine‑grained actions, with persona information offering limited benefit while rationales prove crucial.

AIEvaluationLLM
0 likes · 9 min read
Can LLMs Truly Mimic Human Shopping Behavior? The OPeRA Dataset and Evaluation
Machine Heart
Machine Heart
Apr 17, 2026 · Artificial Intelligence

DeepSeek Introduces Mega MoE and FP4 Indexer – Inside the New GPU Fusion Kernel

DeepSeek's latest DeepGEMM update adds Mega MoE, a fused GPU kernel that collapses the entire Mixture‑of‑Experts pipeline and overlaps computation with NVLink communication, while also unveiling an FP4 indexer and FP8×FP4 precision experiments, signaling a push toward highly efficient large‑scale AI training.

DeepGEMMDeepSeekFP4 Indexer
0 likes · 5 min read
DeepSeek Introduces Mega MoE and FP4 Indexer – Inside the New GPU Fusion Kernel
Machine Heart
Machine Heart
Apr 17, 2026 · Artificial Intelligence

Can Table Modeling Scale? Rethinking the Tree Model Era Amid Compute Shifts

The article examines how a single NVIDIA H100 GPU delivers roughly 200‑fold more FP16 compute than a 96‑core CPU Hadoop node, explores the "Bitter Lesson" of scaling‑driven AI breakthroughs, and presents large‑scale pretraining experiments that show table and sequence models now exhibit clear scaling laws, challenging the dominance of traditional tree‑based approaches.

FOUNDKMLPScaling Law
0 likes · 10 min read
Can Table Modeling Scale? Rethinking the Tree Model Era Amid Compute Shifts
Machine Heart
Machine Heart
Apr 16, 2026 · Artificial Intelligence

CPL++: A Self‑Aware, Self‑Correcting Framework for Weakly Supervised Visual Grounding

The CPL++ framework equips weakly supervised visual grounding models with confidence‑aware pseudo‑label learning, self‑supervised association correction, and dynamic validation, enabling the model to detect and amend erroneous region‑query links during training, which yields absolute performance gains of 1–6 % across five benchmark datasets.

Visual Groundingcomputer visionconfidence-aware
0 likes · 9 min read
CPL++: A Self‑Aware, Self‑Correcting Framework for Weakly Supervised Visual Grounding
Machine Heart
Machine Heart
Apr 16, 2026 · Artificial Intelligence

Achieving 4.6× Faster Diffusion Model Training with FP4‑BF16 Dual‑Track Parallelism (Sol‑RL)

Sol‑RL, a framework from NVIDIA, Hong Kong University and MIT, integrates NVFP4 inference for large‑scale rollout exploration and BF16 precision for high‑fidelity regeneration, delivering up to 4.64× faster convergence at equivalent reward levels while preserving BF16 training fidelity across SANA, FLUX.1 and SD3.5‑L models.

BF16Diffusion ModelsFP4
0 likes · 9 min read
Achieving 4.6× Faster Diffusion Model Training with FP4‑BF16 Dual‑Track Parallelism (Sol‑RL)
Machine Heart
Machine Heart
Apr 16, 2026 · Industry Insights

Why Foreign Developers Rush to China’s GLM Coding Plan While Claude Demands Face Verification

Anthropic now forces Claude users to complete a five‑minute identity check with a government‑issued photo ID and live selfie, sparking backlash, while the cheaper Chinese GLM Coding Plan—priced far below overseas subscriptions—has foreign developers scrambling to buy it, highlighting stark market and pricing contrasts.

AIAnthropicClaude
0 likes · 6 min read
Why Foreign Developers Rush to China’s GLM Coding Plan While Claude Demands Face Verification