matiaslapolla.com

Model Context Protocol (MCP) - The USB-C for AI

Image might have been edited with AI.

Model Context Protocol (MCP) - The USB-C for AI

April 19, 2025
6 min read
Table of Contents
index

What is MCP?

The Model Context Protocol (MCP) is an open JSON‑RPC–based standard that lets a language‑model host (“host process”) connect through lightweight clients to multiple servers that expose prompts, resources and tools in a uniform way — conceptually, “USB‑C for AI.”

Its design borrows from earlier interface standards such as SQL and, more explicitly, the Language Server Protocol (LSP): like LSP decoupled IDEs from language back‑ends, MCP decouples LLMs from data/tool back‑ends, scaling better than bespoke integrations.

Why do we need another protocol?

Current approachKey Pain Point(s)
Prompt engineeringManual, fragile, bounded by context window
Classic RAG (vector DB + search)Model cannot execute actions; must paste raw text
Function‑calling / pluginsN × M grid of vendor‑specific adapters

MCP collapses that N × M grid into N + M: each LLM and each integration only implements MCP once.

The N × M Problem in Context

N = number of models or AI platforms (e.g., OpenAI GPT‑4, Anthropic Claude, Mistral, Gemini, etc.)

M = number of tools, APIs, or data sources (e.g., Slack, GitHub, Notion, Salesforce…)

const totalIntegrations = N * M;

Example:

AI PlatformSlackGitHubNotionSalesforce
GPT‑4
Claude
Gemini

Each cell represents a custom, handcrafted adapter, often with different logic, formats, and security models. This does not scale.

What MCP Does

MCP proposes a universal adapter model.

Each tool (M) implements the MCP server interface once.

Each model (N) only needs to implement the MCP client interface

That reduces the complexity from N × M integrations to:

const totalIntegrations = N + M;

The LLMs don’t care how Slack or GitHub work — they just speak MCP. This means that, for example, Slack’s MCP server doesn’t care if it’s being queried by Claude or GPT‑4. As long as both sides understand MCP, the integration works.

Because servers can stream structured context and expose invocable tools, the model can reason over fresh data, ground responses and trigger side‑effects — without leaking raw credentials into prompts.

It’s something like:

Without MCP: Every phone manufacturer must build a custom charger for every type of outlet.

With MCP: Everyone uses USB‑C. Plug in and go.

Reference Architectures

VariantSketchWhen to choose
A. Central “Context Hub” - Diagram AHost → single MCP client → hub server → N downstream servers. The hub deduplicates embeddings, enforces org‑wide policy and keeps audit logs.Enterprises that need a choke‑point for compliance and observability.
B. Agent Mesh - Diagram BEach autonomous agent embeds its own host+client pair; servers run wherever the data lives. Message‑bus (e.g., NATS) coordinates tasks.Edge/desktop assistants, offline workflows, or environments that favour horizontal scale and fault‑isolation.

Diagram A:

Diagram A

Key Concepts Mapped

A → B = the LLM process sending JSON-RPC calls through a client.

B → C = the single point of orchestration and governance (hub).

C → D1…D4 = actual tools, databases, services exposed as MCP-compatible endpoints.

Diagram B:

Diagram B

Highlights:

  • Each agent is self-contained, able to reason and act.
  • The Message Bus allows distributed coordination, pub/sub tasking, and optionally even logging or observability.
  • MCP servers live close to their respective data or services (e.g., GitHub, Salesforce, S3).

Both share identical MCP semantics, so teams can migrate from hub to mesh (or hybrid) without rewriting servers.

Domain-Specific Implementations

DomainServers in playWhat MCP adds
Personal AI desktopfilesystem, calendar, emailFine‑grained permission prompts; no global file‑tree dump in the prompt.
Technical copilots (Replit, Sourcegraph Cody)git, docker, k8sSecure access to code, logs and CLIs; devs swap Claude↔GPT with no new adapters. Hugging Face Source Swap.
Customer supportzendesk, pricing-db, refund-toolLLM can both answer queries and execute refunds atomically.
Strategic analyticssnowflake, newsfeed, slackModel cross‑joins BI queries with live market news, then pings stakeholders.
No‑code workflow automationEach “Zap” offered as an MCP server; an LLM chains themRemoves brittle REST glue, enabling natural‑language composition.

Benefits and Drawbacks

Key advantages:

  • Portability & vendor‑neutrality – hosts can swap Claude, GPT‑4o or Gemma without touching back‑end code.
  • Least‑privilege isolation – each server is sandboxed; the host mediates consent.
  • Rapid ecosystem growth – >1 000 community servers in six months.

Open challenges:

  • Security – OAuth‑token theft, indirect prompt injection and “tool‑poisoning” remain active research areas.
  • Granular auth scopes – today every server invents its own scope vocabulary.
  • Extra latency hops – mitigated with keep‑alive streams and server‑side batching.
  • Economic incentive – open standards create positive spill‑overs but capture little direct revenue, echoing USB and HTTP history.

Some takeaways for architects & strategists

  • Model context becomes an interface, not a hack. Design your data/products to expose MCP servers rather than sprinkling embeddings everywhere.
  • Choose topology consciously. Central hubs simplify control; meshes unlock autonomy.
  • Budget for security reviews. Token storage, permission scopes and prompt sanitation are the new attack surface.
  • Track the standard. Like USB‑C, once an interface tips, it becomes costly to ignore.

MCP is still young, but its blend of open specification, rapid adoption and clear pain‑killer story signals a likely tectonic shift. Experiment in sandbox projects now; by the time your next platform revamp starts, MCP support may be table‑stakes.