How Model Context Protocol (MCP) Is Redefining AI Agent Development
This article explains how the Model Context Protocol (MCP) decouples tool providers from AI agents, introduces a USB‑C‑like standard for context and tool exchange, and demonstrates its impact on development paradigms, tool ecosystems, and future AI agent architectures.
Introduction
Recent discussions about MCP reveal a crucial point that is often overlooked: by standardising the protocol, tool providers and application developers become decoupled, which will shift the AI Agent development paradigm much like the front‑end/back‑end separation in web development.
Case Study: Agent TARS
The article uses the open‑source multimodal AI agent Agent TARS (https://agent-tars.com/) to illustrate MCP’s role in the development paradigm and tool‑ecosystem expansion.
Key Terminology
AI Agent : In the LLM context, an AI Agent can perceive, plan, and act autonomously, going beyond simple chatbots or Copilot assistants.
Copilot : An AI‑assisted tool that offers suggestions and automates tasks but does not act independently.
MCP (Model Context Protocol): An open protocol that standardises how applications provide context to LLMs, analogous to a USB‑C port for AI models.
Agent TARS : An open‑source multimodal AI agent that integrates seamlessly with real‑world tools.
RESTful API : A software architectural style for client‑server communication.
Background
AI has evolved from simple chatbots to Copilot assistants and now to autonomous agents that require richer context (Resources) and toolsets (Tools) to perform complex tasks.
Pain Points
High coupling : Tool developers must understand the internal implementation of agents, making tool development and debugging difficult.
Poor tool reusability : Tools are tightly bound to the agent code, preventing cross‑language reuse.
Ecosystem fragmentation : Each agent ecosystem provides its own OpenAPI, leading to incompatibility between tools.
Goal
Decouple tools from the agent layer by introducing an MCP Server that standardises context and tool calls. The MCP Server offers a uniform interface for agents to access resources, prompts, and tools.
Demonstrations
Four examples show how MCP is used in practice:
Example 1 : Technical analysis of a stock and market‑order purchase using a broker’s MCP server.
Example 2 : Querying CPU, memory, and network speed via a command‑line MCP server.
Example 3 : Retrieving the top‑5 most‑liked products on ProductHunt using a browser MCP server.
Example 4 : Comparing multiple agent frameworks and generating a markdown report.
What Is MCP?
MCP is an open protocol, similar to a USB‑C interface for AI applications, that defines a JSON‑RPC‑based message format for LLMs to interact with external resources and tools. It standardises three core concepts:
Resources : Data, documents, configuration, screenshots, etc., supplied to the model.
Prompts : Customisable prompts that guide tool usage.
Tools : Functions that the model can invoke to perform actions.
Supported language bindings include TypeScript, Python, Java, Kotlin, and C#.
Comparison with Function Call
MCP
Function Call
Definition
Standard interface for model‑tool integration, including Resources, Prompts, and Tools.
Flat list of functions without explicit resource or prompt concepts.
Protocol
JSON‑RPC, supports bidirectional communication and discovery.
JSON‑Schema, static function calls.
Invocation
Stdio, SSE, Streamable HTTP, in‑process call.
In‑process function call.
Use‑case
Dynamic, complex interactions.
Single, static function execution.
Integration difficulty
Higher.
Lower.
Engineering maturity
High.
Low.
Analogy to Front‑End/Back‑End Separation
Just as early web development suffered from tightly coupled front‑end and back‑end code, AI agents face similar coupling issues. MCP introduces a “tool layer” that separates tool developers from agent developers, enabling modular composition of capabilities.
Practical Implementation
Using the mcp‑server‑browser package as an example, the article walks through the creation of an npm package, the package.json configuration, and the implementation of entry points ( src/server.ts and src/index.ts). Key exported methods are: listTools: Enumerates all available functions. callTool: Invokes a specific function. close: Cleans up the server.
Development can be performed with the MCP Inspector (https://modelcontextprotocol.io/docs/tools/inspector), which provides a playground for testing resources, prompts, and tools. npm run dev Running the dev script launches a local server that prints connection logs such as: Starting MCP inspector... Tool definitions use zod schemas that are converted to JSON‑Schema for MCP compliance.
const toolsMap = {
browser_navigate: {
description: 'Navigate to a URL',
inputSchema: z.object({ url: z.string() }),
handle: async (args) => {
// implementation
return { content: [{ type: 'text', text: `Navigated to ${args.url}` }], isError: false };
}
}
};The MCP client aggregates internal and external servers, supporting both in‑process function calls and out‑of‑process transports (Stdio, SSE, Streamable HTTP). Example client initialisation:
import { MCPClient } from '@agent-infra/mcp-client';
import { client as mcpBrowserClient } from '@agent-infra/mcp-server-browser';
const client = new MCPClient([
{ name: 'browser', description: 'web browser tools', localClient: mcpBrowserClient }
]);
const tools = await client.listTools();Remote invocation (e.g., in a web app) can be achieved by wrapping Stdio servers with SSE or Streamable HTTP, effectively “cloud‑ifying” MCP servers.
async with MCPServerSse(name="fetch", params={"url": "https://{mcp_faas_id}.mcp.bytedance.net/sse"}) as mcp_server:
tools = await MCPUtil.get_function_tools(mcp_server)
# use tools in OpenAI chat completionFuture Directions
Remote MCP Support : Authentication, service discovery, and stateless design for production‑grade, Kubernetes‑native MCP services.
Agent Support : Enhancing complex agent workflows and human‑machine interaction.
Developer Ecosystem : Encouraging broader participation to expand AI Agent capabilities.
Exploring reinforcement‑learning integration to enable models to generalise across newly added MCP tools.
Standardising Agent‑to‑Agent communication protocols (e.g., Google’s A2A, ANP).
References
Detailed MCP analysis article (WeChat link)
a16z explanation of MCP and AI tools (WeChat link)
Medium tutorial: "How to Build an MCP Server Fast"
Zhihu article on AI agent internet trends
YouTube video: "With MCP, AI is complete"
Silicon Valley 101: Dawn of AI Agent explosion (YouTube)
Comprehensive MCP guide (guangzhengli.com)
OpenAI tweet confirming MCP support
Volcano Engine Developer Services
The Volcano Engine Developer Community, Volcano Engine's TOD community, connects the platform with developers, offering cutting-edge tech content and diverse events, nurturing a vibrant developer culture, and co-building an open-source ecosystem.
How this landed with the community
Was this worth your time?
0 Comments
Thoughtful readers leave field notes, pushback, and hard-won operational detail here.
