WebMCP vs MCP: Google’s New AI Agent Protocol Explained

WebMCP vs MCP: Google’s New AI Agent Protocol Explained

In 2026, two competing protocols are shaping how AI agents interact with the web: Anthropic’s Model Context Protocol (MCP) and Google’s WebMCP. While they share a naming convention, these protocols solve fundamentally different problems and have very different implications for web scraping, proxy usage, and the future of the open web.

Understanding both protocols — and how they interact — is essential for anyone building AI-powered applications that need web data.

The Protocol Landscape

Before diving into the comparison, let’s establish what each protocol is designed to do:

MCP (Model Context Protocol) — Created by Anthropic. A standard for connecting AI models to external tools and data sources. Think of it as a way for an AI to use tools: search the web, query databases, call APIs.

WebMCP — Proposed by Google. A standard for websites to expose structured tools and data directly to AI agents through the browser. Think of it as a way for websites to offer their functionality to AI agents, similar to how APIs let developers access web services.

The key distinction: MCP is agent-to-tool communication. WebMCP is website-to-agent communication.

What Is MCP (Model Context Protocol)?

MCP, introduced by Anthropic in late 2024 and widely adopted by 2025-2026, defines how an AI model communicates with external tool servers. It’s a client-server protocol where:

  • The AI application (Claude Desktop, Cursor, a custom app) acts as the MCP client
  • Tool servers (Firecrawl, Bright Data, custom servers) expose capabilities
  • Communication happens via JSON-RPC over stdio or HTTP

How MCP Works

User: "What's the cheapest flight from SFO to NRT next week?"
    ↓
AI Agent (Claude)
    ↓ MCP call: scrape_url("https://flights.example.com/search?from=SFO&to=NRT")
MCP Server (e.g., Firecrawl)
    ↓ HTTP request via proxy
flights.example.com
    ↓ response
MCP Server → parses HTML → returns markdown
    ↓
AI Agent → interprets results → generates answer

Key MCP Characteristics

  • Agent-initiated: The AI decides what tools to use and when
  • Tool-centric: Servers expose discrete tools (scrape, search, query)
  • Infrastructure-side: MCP servers run on the agent developer’s infrastructure
  • No website cooperation needed: The target website doesn’t know or care about MCP
  • Works today: Widely adopted, mature ecosystem

What Is WebMCP?

WebMCP is Google’s proposed standard for how websites can expose tools and structured data directly to AI agents visiting their pages. Rather than the agent scraping and parsing HTML, the website explicitly declares: “Here are the actions an AI agent can perform on my site.”

The Core Idea

Websites already have robots.txt (telling crawlers what they can access) and sitemap.xml (telling search engines what pages exist). WebMCP adds a new layer: a machine-readable declaration of what an AI agent can do on the site.

How WebMCP Works

A website that implements WebMCP would include something like this:

<!-- In the page's <head> -->
<link rel="ai-tools" href="/.well-known/webmcp.json" />

The webmcp.json file declares available tools:

{
  "name": "ExampleAirlines",
  "version": "1.0",
  "tools": [
    {
      "name": "search_flights",
      "description": "Search for available flights between two airports",
      "endpoint": "/api/ai/flights/search",
      "method": "POST",
      "parameters": {
        "type": "object",
        "properties": {
          "origin": {
            "type": "string",
            "description": "Origin airport code (IATA)"
          },
          "destination": {
            "type": "string",
            "description": "Destination airport code (IATA)"
          },
          "date": {
            "type": "string",
            "format": "date",
            "description": "Travel date (YYYY-MM-DD)"
          },
          "passengers": {
            "type": "integer",
            "default": 1
          }
        },
        "required": ["origin", "destination", "date"]
      },
      "response_schema": {
        "type": "array",
        "items": {
          "type": "object",
          "properties": {
            "flight_number": {"type": "string"},
            "departure_time": {"type": "string"},
            "arrival_time": {"type": "string"},
            "price": {"type": "number"},
            "currency": {"type": "string"}
          }
        }
      }
    },
    {
      "name": "book_flight",
      "description": "Book a specific flight",
      "endpoint": "/api/ai/flights/book",
      "method": "POST",
      "authentication": "required",
      "parameters": {
        "type": "object",
        "properties": {
          "flight_number": {"type": "string"},
          "date": {"type": "string"},
          "passenger_name": {"type": "string"},
          "passenger_email": {"type": "string"}
        }
      }
    }
  ],
  "authentication": {
    "type": "oauth2",
    "authorization_url": "/oauth/authorize",
    "token_url": "/oauth/token",
    "scopes": ["read:flights", "write:bookings"]
  },
  "rate_limits": {
    "requests_per_minute": 60,
    "requests_per_day": 10000
  },
  "terms_of_use": "https://example-airlines.com/ai-terms"
  }
}

WebMCP Flow

User: "Book the cheapest flight from SFO to NRT"
    ↓
AI Agent (Gemini / Chrome AI)
    ↓ discovers WebMCP on example-airlines.com
Reads /.well-known/webmcp.json
    ↓ calls search_flights API directly
example-airlines.com/api/ai/flights/search
    ↓ structured JSON response
AI Agent → selects best option → calls book_flight
    ↓
Booking confirmed → response to user

Key WebMCP Characteristics

  • Website-initiated: The website chooses what to expose
  • API-centric: Websites offer structured endpoints, not raw HTML
  • Cooperative: Requires website adoption to work
  • No scraping needed: Structured data replaces HTML parsing
  • Still emerging: Proposal stage, limited adoption

MCP vs WebMCP: Detailed Comparison

DimensionMCP (Anthropic)WebMCP (Google)
Who controls itAgent developerWebsite owner
DirectionAgent calls toolsWebsite offers tools
Website cooperationNot neededRequired
Data formatMarkdown/HTML (parsed)Structured JSON (native)
Scraping requiredYesNo
Proxy neededUsually yesUsually no
AuthenticationAgent handles itBuilt-in OAuth
Rate limitingMust be managedDeclared by website
CoverageAny websiteOnly participating sites
MaturityProduction-readyProposal/early adoption
EcosystemLarge (100+ servers)Small (early adopters)
Best forData collection at scaleDirect service interaction

Implications for Proxy Usage

This is where the two protocols diverge most sharply, and where the impact on the proxy industry is most significant.

MCP: Proxies Are Essential

MCP-based web access is fundamentally scraping. The AI agent uses MCP tools to fetch web pages, and the MCP server makes HTTP requests to target websites. These websites don’t know or care about MCP — they see regular HTTP traffic.

This means:

  • Anti-bot protection applies: Websites will block repeated requests from the same IP
  • Proxy rotation is critical: MCP servers need diverse IP pools
  • Geographic targeting matters: Data often varies by location
  • Session management: Some scraping requires persistent proxy sessions
  • Bandwidth costs: Full page fetching uses significant bandwidth

For MCP-based agents, proxy infrastructure isn’t optional — it’s a core requirement.

WebMCP: Proxies Are Mostly Unnecessary

WebMCP provides a direct, sanctioned API from the website to the AI agent. Because the website explicitly offers this endpoint:

  • No anti-bot bypass needed: The website welcomes AI traffic
  • No IP rotation needed: Authenticated API access
  • Structured responses: No bandwidth wasted on HTML/CSS/JS
  • Rate limits are explicit: The website declares what’s allowed

However, there are still scenarios where proxies matter with WebMCP:

  1. Geographic pricing: If the API returns different data based on IP location
  2. Competitive analysis: Accessing competitors’ WebMCP endpoints from different regions
  3. Reliability: Routing through proxies for failover and load balancing
  4. Privacy: Organizations that don’t want their IP addresses revealed to every website

The Hybrid Reality

In practice, AI agents in 2026 and beyond will use both protocols:

class HybridWebAgent:
    def __init__(self, proxy_manager, mcp_client):
        self.proxy_manager = proxy_manager
        self.mcp_client = mcp_client

    async def get_data(self, url: str) -> dict:
        # Check if the site supports WebMCP
        webmcp_tools = await self.discover_webmcp(url)

        if webmcp_tools:
            # Use WebMCP — structured API, no proxy needed
            return await self.call_webmcp_tool(url, webmcp_tools)
        else:
            # Fall back to MCP scraping — needs proxy
            proxy = self.proxy_manager.get_proxy(url)
            return await self.mcp_client.scrape(url, proxy=proxy)

    async def discover_webmcp(self, url: str) -> dict | None:
        """Check if a website exposes WebMCP tools."""
        from urllib.parse import urlparse
        parsed = urlparse(url)
        webmcp_url = f"{parsed.scheme}://{parsed.netloc}/.well-known/webmcp.json"

        try:
            response = await fetch(webmcp_url, timeout=5)
            if response.status_code == 200:
                return response.json()
        except Exception:
            return None
        return None

How Websites Can Expose Tools to AI Agents

For website owners considering WebMCP adoption, here’s what implementation looks like:

Step 1: Define Your AI-Accessible Tools

Identify which parts of your site’s functionality should be available to AI agents. Common candidates:

  • Search functionality
  • Product/pricing lookups
  • Availability checks
  • Account-related actions (with authentication)

Step 2: Create the WebMCP Manifest

Place a webmcp.json file at /.well-known/webmcp.json:

{
  "name": "YourService",
  "version": "1.0",
  "description": "AI-accessible tools for YourService",
  "tools": [
    {
      "name": "search_products",
      "description": "Search products by keyword, category, or price range",
      "endpoint": "/api/ai/search",
      "method": "GET",
      "parameters": {
        "type": "object",
        "properties": {
          "query": {"type": "string"},
          "category": {"type": "string"},
          "min_price": {"type": "number"},
          "max_price": {"type": "number"},
          "sort": {"type": "string", "enum": ["price_asc", "price_desc", "relevance"]}
        },
        "required": ["query"]
      }
    }
  ]
}

Step 3: Build the API Endpoints

Create dedicated endpoints for AI agent traffic. These should return structured JSON, not HTML:

# Flask example
@app.route("/api/ai/search")
@rate_limit(per_minute=60)
def ai_product_search():
    query = request.args.get("query")
    results = product_service.search(query)

    return jsonify({
        "results": [
            {
                "name": p.name,
                "price": p.price,
                "url": p.url,
                "availability": p.in_stock
            }
            for p in results
        ],
        "total_count": len(results)
    })

Step 4: Add the Link Header

Reference the WebMCP manifest in your HTML:

<link rel="ai-tools" href="/.well-known/webmcp.json" />

Security and Privacy Considerations

Both protocols raise important security questions:

MCP Security Concerns

  • Prompt injection: Scraped web content could contain instructions that manipulate the AI agent
  • Data exfiltration: A compromised MCP server could leak sensitive user queries
  • Over-permissioning: MCP servers with too many capabilities could be exploited
  • Supply chain risk: Third-party MCP servers introduce dependency risk

WebMCP Security Concerns

  • Authentication abuse: AI agents managing OAuth tokens at scale creates attack surface
  • Tool misuse: Automated agents could abuse exposed functionality (mass bookings, account creation)
  • Data poisoning: Malicious WebMCP endpoints could feed false data to AI agents
  • Privacy leakage: WebMCP tools might expose user preferences and behavior patterns
  • Scraping via API: WebMCP endpoints could be used for mass data collection, replacing traditional scraping

Mitigation Strategies

For MCP:

  • Validate and sanitize all scraped content before passing to the LLM
  • Use read-only MCP servers where possible
  • Implement proxy rotation to avoid IP-based tracking
  • Monitor MCP server traffic for anomalies

For WebMCP:

  • Implement robust rate limiting on all AI endpoints
  • Require authentication for sensitive actions
  • Log all AI agent interactions for audit
  • Use CAPTCHA challenges for high-value actions (bookings, purchases)

What This Means for the Proxy Industry

Short-Term (2026): Proxies Are More Important Than Ever

MCP adoption is driving massive growth in proxy demand. Every AI agent that browses the web needs proxy infrastructure. WebMCP adoption is too limited to change this.

Medium-Term (2027-2028): A Bifurcated Market

As WebMCP gains adoption among major websites, proxy demand for those specific sites will decrease. But the long tail of millions of websites that won’t implement WebMCP will still require proxy-backed scraping.

Expected split:

  • Top 1,000 websites: Many will adopt WebMCP → less proxy need
  • Remaining millions: Will still require scraping → proxy demand persists

Long-Term (2029+): Proxies Evolve, Don’t Disappear

Even in a world with widespread WebMCP adoption, proxies remain valuable for:

  1. Competitive intelligence: Accessing competitors’ public data (they won’t offer you a friendly API)
  2. Verification: Cross-checking WebMCP data against actual page content
  3. Coverage gaps: Not all data will be exposed via WebMCP
  4. Geographic access: Proxies for geo-specific data remain essential
  5. Privacy: Hiding your organization’s identity when researching competitors

New Proxy Use Cases

WebMCP creates new proxy use cases:

  • API endpoint monitoring: Proxied access to competitors’ WebMCP endpoints
  • Geographic price verification: Checking if WebMCP endpoints return different data by location
  • WebMCP discovery scanning: Crawling millions of sites to find WebMCP endpoints
  • Fallback infrastructure: When WebMCP endpoints are down, proxy-backed scraping serves as backup

The Competitive Dynamics

Google’s Incentive

Google benefits from WebMCP because:

  • It positions Chrome/Gemini as the primary AI agent interface
  • Structured data from WebMCP flows through Google’s ecosystem
  • It reduces the “wild west” of web scraping that Google’s own properties are subject to
  • It gives Google influence over how websites expose data to AI

Anthropic’s Incentive

Anthropic benefits from MCP because:

  • It makes Claude the most capable AI agent (with the richest tool ecosystem)
  • It doesn’t require website cooperation — works on any site
  • The open-source nature builds community and ecosystem lock-in
  • It positions Anthropic as the infrastructure standard, like how HTTP or JSON became universal

Developer Implications

For developers building AI agents today:

  • Implement MCP support first — it works everywhere, right now
  • Add WebMCP discovery as a future-proofing measure
  • Keep your proxy infrastructure — you’ll need it regardless
  • Design for both — your agent should gracefully use WebMCP when available and fall back to MCP scraping when not

Practical Recommendations

For AI Agent Developers

Priority 1: MCP integration (works now, covers all sites)
Priority 2: Proxy infrastructure (essential for MCP)
Priority 3: WebMCP discovery (future-proofing)
Priority 4: Hybrid routing (auto-select best protocol)

For Website Owners

Priority 1: Evaluate whether AI agent traffic benefits your business
Priority 2: If yes, implement WebMCP for your most-used endpoints
Priority 3: Set clear rate limits and terms of use
Priority 4: Monitor AI agent behavior for abuse

For Proxy Providers

Priority 1: Optimize for AI agent traffic patterns (sessions, speed)
Priority 2: Build MCP server integrations
Priority 3: Offer AI-specific proxy tiers
Priority 4: Prepare for WebMCP by adding API monitoring features

Testing Your Setup

Whichever protocol you adopt, verify your infrastructure is working correctly:

Conclusion

MCP and WebMCP aren’t competitors — they’re complementary protocols that serve different needs. MCP gives AI agents the ability to access any website through scraping and proxy infrastructure. WebMCP gives websites the ability to offer structured access to AI agents directly.

The smart approach is to build for both: use MCP with robust proxy support for universal web access today, and implement WebMCP discovery to take advantage of structured endpoints as they become available. The proxy industry will evolve alongside these protocols, but it’s not going anywhere — there will always be data that requires active retrieval rather than cooperative sharing.


Related Reading

Scroll to Top