A16z: Deep understanding of the future of MCP and AI tool ecosystem

Written by
Audrey Miles
Updated on:July-10th-2025
Recommendation

Explore the new chapter of AI tool ecology, how MCP reshapes the interaction between AI agents and tools.

Core content:
1. The lack of standardized interfaces in the AI ​​field and MCP's solution
2. MCP protocol definition and comparison with LSP
3. MCP's practice and challenges in current application scenarios

Yang Fangxian
Founder of 53AI/Most Valuable Expert of Tencent Cloud (TVP)

Since OpenAI released the function calling feature in 2023, I have been thinking about how to truly unlock an agent- and tool-centric ecosystem. As the underlying models become increasingly intelligent, AI agents' ability to interact with external tools, data, and APIs becomes increasingly decentralized: developers have to customize specific business logic for each system in order for agents to complete operations and integration between different systems.

It is obvious that we need a standardized interface for executing instructions, obtaining data, and calling tools. In the early days of the Internet, API was the unified language for communication between software; but in the field of AI, we still lack a corresponding standard.

Since its release in November 2024, the Model Context Protocol (MCP) has quickly gained widespread attention among developers and the AI ​​community as a potential solution to the above problems. This article will explore what MCP is, why it will change the way AI interacts with tools, how developers use it today, and the challenges that still need to be addressed.

Let’s get started.




What is MCP?


MCP is an open protocol that allows various systems to provide context to AI models in a generalizable way. It defines how AI models call external tools, obtain data, and interact with services. For example, the following shows how the Resend MCP server works with multiple MCP clients:

This idea is not new; MCP is inspired by LSP (Language Server Protocol). In LSP, when the user types in the editor, the client requests auto-completion or diagnostic information from the language server.

MCP extends beyond LSP in that it adopts an "agent-centric" execution model. While LSP is primarily passive (responding to user input and IDE requests), MCP supports autonomous AI workflows. Based on a given context, AI agents are free to decide which tools to use, in what order to call them, and how to string these calls together to complete a task. MCP also introduces a "human intervention" link, allowing people to provide additional data and approve the execution process when necessary.




Current mainstream application scenarios


As long as they have a compatible MCP server, users can turn each MCP client into an "everything app".
Take Cursor as an example: although Cursor is a code editor, it is also a fully implemented MCP client. By installing the Slack MCP server, it can become a Slack client; by installing the Resend MCP server, it can send emails; by installing the Replicate MCP server, it can generate images. Even more powerful is to install multiple MCP servers on a client to unlock combined tasks: users can install a server to generate the front-end interface, and let the agent call the image generation MCP server to generate a hero image for the website.
Aside from Cursor, most current use cases fall roughly into two categories: localized workflows for developers and new experiences built with LLM clients .

1. Workflow for developers

For developers who are immersed in writing code every day, "I don't want to leave my IDE to do X" is a common voice. MCP server provides strong support for this demand.
For example, developers no longer need to switch to the Supabase interface to check the database status, they can directly execute read-only SQL commands in the IDE through the Postgres MCP server; they can also use the Upstash MCP server to create and manage cache indexes. When writing code, developers can also use Browsetools MCP in combination to provide AI agents with real-time environment access to obtain console logs and debugging information.

An example of how the Cursor proxy can use Browsertools to access console logs and other real-time data and debug more efficiently.


In addition, the MCP server can also provide developers with highly accurate code context. For example, by crawling a web page, or automatically generating a corresponding MCP server based on the documentation, the AI ​​agent can instantly obtain the external tools it needs. This means that developers can spend less time writing boilerplate code and focus on actually using these tools - whether it is getting real-time context, executing commands, or expanding new capabilities for AI assistants at any time.

2. New application experience

Not only IDEs (such as Cursor) can become MCP clients. Although they are more popular among technical people, clients for mass users are also on the rise. For example, for non-technical users, Claude Desktop is a friendly and easy-to-use entry point, allowing more people to easily use MCP-based tools. Next, we are also likely to see the emergence of MCP clients specifically for business scenarios such as customer support, marketing copywriting, design, and image editing - these fields are naturally in line with AI's expertise in pattern recognition and creative writing.
The design of the MCP client and the interaction modes it supports largely determine its functional boundaries. For example, a chat application will not usually provide a vector drawing canvas, and design tools are unlikely to allow you to remotely execute code. Therefore, the experience of the MCP client determines the experience of the end user interacting with MCP, and at this level, we still have a lot of room for innovation that has not been explored.
An interesting example is the “@command” implemented by Highlight to call any MCP server installed by its client. This allows the MCP client to stream generated content to any downstream application.

An example of a Notion MCP (plugin) implementation by Highlight.


Another example is using the Blender MCP server: even a novice who is not familiar with the basic operations of Blender can describe the 3D model he wants to create in natural language. We are witnessing the rise of text-to-3D workflows, and the community is also implementing similar MCP servers for tools such as Unity and Unreal Engine.

Example of using Claude Desktop with  Blender MCP Server .

Although we often refer to "servers" and "clients", the MCP ecosystem is gradually taking shape as the protocol evolves. The figure below shows the most active areas at present (of course there are still many gaps), and as MCP matures, we expect to see more new players join. (We will also explore more potential possibilities of MCP in the next section.)

Market map

As far as MCP clients are concerned, it is not surprising that most high-quality clients today are still centered on code development—developers are often early adopters of new technologies. However, as the protocol improves, more clients focusing on business scenarios will inevitably emerge.
Most existing MCP servers are "local first" and are mostly made by a single developer or team. This is mainly because MCP currently only supports SSE-based and imperative connections. However, we believe that as remote MCP becomes the first citizen and MCP gradually adopts streaming HTTP transmission, the entire ecosystem of servers will become more and more abundant.
At the same time, a number of new MCP markets and server hosting solutions are emerging to promote the discovery and distribution of MCP servers. Marketplaces such as Mintlify's mcpt, Smithery, and OpenTools are providing developers with a more convenient way to discover, share, and contribute new MCP servers - similar to npm for JavaScript package management or RapidAPI for API discovery. This level of marketization is crucial for the acquisition of standardized, high-quality MCP servers, and will also enable AI agents to dynamically select and integrate the required tools.
As MCP becomes more popular, infrastructure and development toolchains will be critical to scalability, reliability, and accessibility. Server generation tools such as Mintlify, Stainless, and Speakeasy are simplifying the process of creating MCP-compatible services, while hosting services such as Cloudflare and Smithery are trying to solve deployment and scalability issues. At the same time, connection management platforms such as Toolbase are also beginning to simplify processes such as key management and proxy for local MCP.



Future Possibilities


It is important to note that we are still in the early stages of the development of agent-native architectures. Although MCP is now widely sought after, there are still many unresolved problems in developing and launching MCP.
Here are some key issues that the next phase of the MCP protocol may need to unlock:

1. Hosting and multi-tenancy

MCP supports the scenario of "one AI agent for multiple tools", but multi-tenant architecture (such as SaaS products) requires high concurrency support of "multiple users for the same MCP server". Using remote servers by default may be a way to make MCP servers reach a wider range of users in the short term, but many companies still want to deploy MCP servers in their own environments and separate the data plane and control plane.

To deploy and maintain MCP servers at scale, a mature tool chain is needed, which is the next key link to widespread adoption.

2. Authentication

MCP currently does not define a standard authentication mechanism , nor does it specify how to securely manage and delegate authorization when interacting with third-party APIs. At this stage, authentication is usually solved by each implementation and deployment scenario. In fact, MCP is currently used more in local integration scenarios - and such scenarios often do not have a strong demand for authentication.

If remote MCP is to be truly popularized, a more complete authentication paradigm is needed. For developers, a unified authentication method should include:

  • Client authentication: Use standard methods such as OAuth or API Token for client-server interaction

  • Tool authentication: Provides auxiliary functions or wrappers for third-party API authentication

  • Multi-user authentication: tenant-level authentication for enterprise deployments

3. Authorization

Even if the tool is authenticated, who has the right to use it? How fine-grained is the permission? MCP does not yet have a built-in permission model, and current access control is almost entirely at the session level—a tool is either accessible or completely prohibited. Although there may be more fine-grained permission control mechanisms in the future, the current common practice is to use an authorization process based on OAuth 2.1. Once authenticated, the entire session has access to the tool.

As the number of agents and tools continues to grow, this session-level access management becomes increasingly complex. Currently, each agent typically requires a separate session and authorization credentials, forming an ever-expanding web of access management.

4. Gateway

As your MCP application scales, you may need a gateway layer to centrally handle authentication, authorization, traffic management, and tool selection. Similar to a traditional API gateway, this gateway can perform access control, route requests to the correct MCP server, and perform load balancing and caching. This is especially important for multi-tenant environments, where different users and agents often require different permissions. A standardized gateway can simplify client-server interactions, enhance security, and provide better observability, making MCP deployments easier to scale and manage.

5. Discoverability and availability of MCP servers

Currently, it is still cumbersome to find and set up an MCP server: developers need to locate the relevant endpoints or scripts, configure authentication, and ensure that the server is compatible with the client. Integrating new servers is time-consuming and laborious, and AI agents cannot dynamically discover or adapt to available servers.

But from Anthropic's speech at the AI ​​Engineer Conference last month, they are preparing the MCP server registry and discovery protocol. If it can be successfully launched, this is likely to become the next key node for the large-scale popularization of MCP.

6. Execution Environment

Most AI workflows require calling multiple tools in sequence, but MCP does not have a built-in "workflow" concept to manage these calling steps. If each client is required to implement its own resumability and retryability mechanisms, it will inevitably increase complexity. Although some developers are currently trying to use third-party solutions such as Inngest, if stateful execution can be upgraded to the core concept of the MCP protocol, it will greatly simplify the execution model of most developers.

7. Standardized client experience

A common question from the developer community is: How do we think about "tool selection" when building an MCP client? Does everyone need to implement their own tool selection logic based on search enhancement generation (RAG)? Or is there a chance that this logic can be standardized?

Not only that, the industry also lacks a unified UI/UX model to call tools (some are slash commands, some are purely natural language triggers). If unified tool discovery, sorting, and execution capabilities can be achieved at the client level, developers and users can get a more consistent experience.

8. Debugging

Many MCP server developers have found that it is not easy to make the same MCP server work properly on different clients. Most MCP clients have their own implementation details, and client-side debugging information is often missing or difficult to obtain, making the debugging process extremely difficult. As more MCP servers begin to "remote", we need a whole set of new tools to simplify the development experience in local and remote environments.




Impact of the AI ​​tool ecosystem



The developer experience of MCP reminds me of API development in the 2010s. The concept is new and exciting, but the surrounding toolchain is still in its infancy. If we look ahead a few years, assuming that MCP really becomes the de facto standard for AI-driven workflows, what will happen? Here are some possible trend predictions:
  1. The competitive advantage of developer-oriented companies will evolve from "best API design" to "best tool set" . If MCP allows agents to automatically discover tools, then API or SDK providers need to ensure that their tools can be easily retrieved and perform well in similar tasks, enough for agents to choose. And this will be more sophisticated and immediate than human developers.
  2. If every application becomes an MCP client and every API becomes an MCP server, we may see entirely new pricing models where agencies dynamically select tools based on speed, cost, and relevance. This more market-based adoption approach will allow the most “useful” tools to quickly stand out, breaking the dominance of tools based on user base or historical inertia.
  3. Documentation will become a critical component of the MCP infrastructure . Companies need to provide machine-readable (e.g., llms.txt) tool and API specifications and default MCP server outputs based on existing documentation so that AI can quickly understand and use these features.
  4. API is no longer the end point, but just the starting point . Developers will gradually realize that the mapping between API and tools is not one-to-one. Tools are a higher-level abstraction that makes the best call combination based on the task scenario - for example, proxy callsdraft_email_and_send()It may involve multiple underlying API calls, but it can significantly reduce latency. The design of the MCP server will focus more on specific scenarios and use cases rather than simple API encapsulation.
  5. If every software is an MCP client by default, then hosting will also evolve . The characteristics of the workload are very different from traditional websites. Each client may require multi-step execution and support for recoverable, retryable, and long-duration task management. Hosts will also need to perform real-time load balancing across different MCP servers to balance cost, latency, and performance, so that AI agents can dynamically select the best tool.
MCP is already reshaping the AI ​​agent ecosystem; the key to the next wave of progress will be how we solve these fundamental challenges. If MCP can become the default interface for AI to interact with tools, it is possible to give birth to a new generation of autonomous, multimodal, and deeply integrated AI experiences.
If MCP is widely adopted, the construction, use and business models of tools will be profoundly reshaped. We are full of expectations for the future, especially this year - this is a critical year: Will a unified MCP market emerge? Will AI agent authentication become seamless? Can multi-step execution be written into the agreement?