DeWu Technology
DeWu Technology
Mar 25, 2026 · Big Data

How Code LLM Transforms E‑commerce Data Warehouses: From Data Rights to AI‑Driven Automation

This article analyzes how large‑language models for code, exemplified by Claude Code, are integrated into an e‑commerce data‑warehouse ecosystem, defining data‑rights boundaries, introducing agentic workflows, decoupling cognitive and execution runtimes, and establishing standardized I/O contracts to achieve safe, scalable AI‑assisted development and governance.

Data WarehouseStandardized I/Obig data
0 likes · 24 min read
How Code LLM Transforms E‑commerce Data Warehouses: From Data Rights to AI‑Driven Automation
PaperAgent
PaperAgent
Dec 4, 2025 · Artificial Intelligence

From Code Foundations to AI Agents: A Deep Dive into Code LLMs and Their Applications

This article reviews a comprehensive 303‑page survey on code foundation models, tracing the evolution of code‑focused large language models from 2021 to 2025, comparing general‑purpose and specialized LLMs, and presenting extensive experiments on prompting, fine‑tuning, reinforcement learning, and autonomous coding agents.

AI codingcode LLMlarge language models
0 likes · 5 min read
From Code Foundations to AI Agents: A Deep Dive into Code LLMs and Their Applications
Volcano Engine Developer Services
Volcano Engine Developer Services
Dec 10, 2024 · Artificial Intelligence

Introducing FullStack Bench: Multi‑Language Code LLM Benchmark & SandboxFusion

The article presents FullStack Bench, a newly open‑sourced, multi‑language code‑LLM evaluation dataset covering over 11 real‑world programming scenarios and 16 languages, along with the SandboxFusion execution environment, and reports comprehensive benchmark results that highlight the superiority of closed‑source models over most open‑source alternatives.

AI evaluationFullStack BenchSandboxFusion
0 likes · 11 min read
Introducing FullStack Bench: Multi‑Language Code LLM Benchmark & SandboxFusion
Baobao Algorithm Notes
Baobao Algorithm Notes
Nov 14, 2024 · Artificial Intelligence

How OpenCoder’s RefineCode Dataset Powers Next‑Gen Code LLMs

The OpenCoder technical report details the creation of the RefineCode dataset, its multi‑stage preprocessing, filtering, and sampling pipelines, the pre‑training and fine‑tuning schedules for 1.5B and 8B models, and the autonomous data selection methods that together achieve performance comparable to Qwen2.5‑Coder.

Artificial IntelligenceAutoDSInstruction Tuning
0 likes · 18 min read
How OpenCoder’s RefineCode Dataset Powers Next‑Gen Code LLMs