Why MCP Still Matters: Finding the Optimal Path for Agents to Connect to External Systems

The article compares direct API calls, CLI tools, and the Model Context Protocol (MCP) for agent integration, explains MCP's token overhead, presents two token‑reduction strategies, and outlines design principles for building high‑availability MCP servers to maximize agent utility.

Code Mala Tang
Code Mala Tang
Code Mala Tang
Why MCP Still Matters: Finding the Optimal Path for Agents to Connect to External Systems

Three Integration Approaches and Their Capability Boundaries

The core distinction among the three methods is whether a unified middle layer connects agents to external services.

Approach 1: Direct API Calls

Lowest entry barrier—one API per use case—but scaling leads to an "M×N integration dilemma" where each new agent‑API pair requires separate authentication code, causing exponential maintenance costs.

Approach 2: Command‑Line Interface (CLI)

CLI reuses existing toolchains and works well locally, yet it struggles on mobile, web, or cloud environments without a shell, and relies on disk‑stored credentials, making it unsuitable for production deployments.

Approach 3: Model Context Protocol (MCP)

MCP standardizes the middle layer, encapsulating authentication, tool discovery, and semantic expression. A single remote server can serve Claude, ChatGPT, Cursor, VS Code, and other clients across any deployment environment. Mature systems often combine all three approaches for complementary strengths.

Key Pain Points of MCP

ScaleKit benchmarked GitHub’s MCP server against the gh CLI and found that MCP’s schema bloat consumes excessive tokens; a single tool definition can require hundreds to thousands of tokens, inflating context size.

Two Major Solutions to Reduce Token Consumption

Solution 1: On‑Demand Tool Loading (Tool Search)

Instead of loading all 43 tool definitions (≈55,000 tokens) at conversation start, the agent first declares its task, and the system searches for and loads only the most relevant tools.

Tests show >85% token reduction for tool definitions without hurting tool‑selection accuracy. Traditional MCP queries for repository language cost 44,026 tokens versus 1,365 tokens for CLI (≈32×). With Tool Search, MCP drops to ~10,000 tokens (≈7× CLI).

Solution 2: Programmatic Tool Invocation

Raw tool results are processed in a sandboxed code executor; only the final processed output is fed back to the model, cutting token usage by ~37% in multi‑step workflows. Combining both solutions further streamlines context and improves response speed.

Key Practices for Building High‑Availability MCP Servers

Remote Server Deployment – Enables access from web, mobile, and cloud agents; the only way to ensure universal reach.

Organize Tools by User Intent, Not API Endpoints – Fewer, well‑described tools are more practical than many granular APIs.

Code‑Orchestration Mode for Large APIs – Expose 1–2 code‑receiving tools; agents write scripts to handle hundreds of operations (e.g., Cloudflare’s MCP server uses only search and execute tools, covering ~2,500 endpoints with ~1 KB of definition).

MCP Apps for Rich Interaction – Allow tools to return interactive UI elements (charts, forms, dashboards) directly in the chat, improving adoption and retention.

Standardized Authentication (CIMD) – Reduces first‑login friction and avoids repeated credential prompts; supported by Claude SDKs.

Vault‑Managed Tokens – Automatic token injection and refresh eliminate manual secret handling.

Synergy Between MCP Servers and Skills

Skills provide procedural knowledge (“how to use tools”), while MCP supplies access (“what tools are available”). High‑quality agents combine both.

Plugin Packaging – Bundle skills, MCP servers, hooks, and LSP servers into a single plugin for one‑click deployment.

Direct Skill Distribution from MCP Servers – Servers can ship a matching skill set alongside tool definitions (e.g., Canva, Notion, Sentry), ensuring agents receive usage guidance automatically.

Conclusion

For local development, CLI + Skills offers lightweight, fast iteration. For cloud production, MCP + Skills provides standardized, cross‑platform integration. Simple use cases favor direct API calls. MCP is not dead; it is evolving into the standard integration layer for cloud‑native agents, with automatic capability upgrades as client support expands.

CLIcloud-nativeIntegrationAI agentsMCPToken Optimization
Code Mala Tang
Written by

Code Mala Tang

Read source code together, write articles together, and enjoy spicy hot pot together.

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.