Windsurf is a powerful AI-native coding environment that supports direct integration with Model Context Protocol (MCP) servers.  By connecting Windsurf to MCP, you enable rich, real-time context injection into AI agents, enabling them to reason over internal tools, documents, tickets, alerts, and more—with precision and permissioning built-in. This setup allows you to inject real-time, relevant context from tools like GitHub, Jira, and Asana directly into your AI workflows within Windsurf. It also enables dynamic access to codebases, tasks, and system data, enhancing your developer productivity with high-fidelity contextual understanding.

This step-by-step guide will walk you through three methods for integrating Windsurf with an MCP server using two examples: GitHub, Asana, and Datadog.

MCP Server Types Supported

Windsurf supports the following types of MCP servers:

  • Local Stdio: Runs on your machine and streams responses via standard input/output (stdio).

  • Streamable HTTP: The server operates as an independent process that can handle multiple client connections

Step 1: Enable MCP in Windsurf

  1. Open the Windsurf editor.

  2. Log into your Windsurf account

  3. Press Cmd/Ctrl + ,  to open "Windsurf Settings".

  4. Scroll to “Plugins (MCP servers)” under Casacade

  5. Click Manage Plugins

  6. Click on “View raw config”

  1. This will create or open the config file at one of the following paths:

macOS: `~/.codeium/windsurf/mcp_config.json`
Windows: %USERPROFILE%\.codeium\windsurf\mcp_config.json
  1. Add or modify your configuration as shown in the examples below.

Step 2: Choose Your MCP Server Setup

Windsurf requires an MCP-compliant server that can respond to context queries. You have three setup options:

Option A: Run the MCP Server Locally

Ideal for development, testing, and debugging.

Instructions:

  1. Install Docker.

  2. Ensure Docker is running.

    • If you encounter issues pulling the image, try running `docker logout ghcr.io` to refresh your credentials.

  3. Create a GitHub Personal Access Token.

    • Choose permissions based on what you’re comfortable sharing with the LLM.

    • Learn more in the GitHub documentation.

Config Example:

{
  "mcpServers": {
    "github": {
      "command": "docker",
      "args": [
        "run",
        "-i",
        "--rm",
        "-e",
        "GITHUB_PERSONAL_ACCESS_TOKEN",
        "ghcr.io/github/github-mcp-server"
      ],
      "env": {
        "GITHUB_PERSONAL_ACCESS_TOKEN": "<YOUR_TOKEN>"
      }
    }
  }
}

Save & click refresh

Option B: Use a Hosted MCP Server (e.g., Asana, Atlassian)

The easiest option—no setup required. Note that you won’t have central control or visibility into interactions between Windsurf and the integrated service.

Method 1: Via UI

Press Cmd/Ctrl + Shift + P to open the Command Palette. Search for and select "MCP: Add Server.". And then select "HTTP (server-sent events)" 

Method 2: Via Config
  1. Install Node and npm.

  2. Install the `mcp-remote` package.

Config Example:
{
  "mcpServers": {
    "asana": {
      "command": "npx",
      "args": ["mcp-remote", "https://mcp.asana.com/sse"]
    }
  }
}

Save & click refresh

Option C: Deploy Your Own Remote MCP Server

Recommended for teams and production use. This setup offers:

  • Full visibility and governance

  • Access control across employees and agents

You can deploy your MCP server using a platform like Natoma. Connect apps from the registry, and retrieve the configuration and paste it into WindSurf IDE. Save & click refresh

or Streamable HTTP endpoint, and paste it into Windsurf IDE 

Save & click refresh

Troubleshooting & Testing

To verify MCP integration:

  • Ask Windsurf a question that requires context from recent activity.

  • Windsurf will send a background request to the MCP server, which returns structured context.

  • Monitor Windsurf MCP logs at

    • Press Cmd/Ctrl + Shift + P to open the Command Palette.

    • Search for and select "Developer: Show Logs..".

    • Click Windsurf

  • Ensure your MCP server responds with valid JSON in the correct schema.

  • Use the MCP Inspector for debugging.

Wrap-Up

Integrating Windsurf with an MCP server—whether self-hosted, remote, or SaaS—takes just a few minutes. This unlocks powerful contextual capabilities, allowing you to embed the full history, structure, and semantics of your dev stack into every AI interaction.

About Natoma

Natoma enables enterprises to adopt AI agents securely. The secure agent access gateway empowers organizations to unlock the full power of AI, by connecting agents to their tools and data without compromising security.

Leveraging a hosted MCP platform, Natoma provides enterprise-grade authentication, fine-grained authorization, and governance for AI agents with flexible deployment models and out-of-the-box support for 100+ pre-built MCP servers.

You may also be interested in:

How AI-Data Integration Solves the Enterprise Workflow Bottleneck

AI workflow automation transforms manual data transfers into intelligent, automated processes through secure control points. Key requirements include unified data access and proper authentication infrastructure.

How AI-Data Integration Solves the Enterprise Workflow Bottleneck

AI workflow automation transforms manual data transfers into intelligent, automated processes through secure control points. Key requirements include unified data access and proper authentication infrastructure.

How AI-Data Integration Solves the Enterprise Workflow Bottleneck

AI workflow automation transforms manual data transfers into intelligent, automated processes through secure control points. Key requirements include unified data access and proper authentication infrastructure.

The Enterprise Guide to AI Data Integration

The path to successful AI data integration requires strategic thinking beyond technical implementation.

The Enterprise Guide to AI Data Integration

The path to successful AI data integration requires strategic thinking beyond technical implementation.

The Enterprise Guide to AI Data Integration

The path to successful AI data integration requires strategic thinking beyond technical implementation.

Enterprises are running more MCP servers than they know — Natoma finds an average of 225 already deployed. What are they doing, and why does it matter?

The Rise of MCPs: 225 MCP servers per organization

Enterprises are running more MCP servers than they know — Natoma finds an average of 225 already deployed. What are they doing, and why does it matter?

Enterprises are running more MCP servers than they know — Natoma finds an average of 225 already deployed. What are they doing, and why does it matter?

The Rise of MCPs: 225 MCP servers per organization

Enterprises are running more MCP servers than they know — Natoma finds an average of 225 already deployed. What are they doing, and why does it matter?

Enterprises are running more MCP servers than they know — Natoma finds an average of 225 already deployed. What are they doing, and why does it matter?

The Rise of MCPs: 225 MCP servers per organization

Enterprises are running more MCP servers than they know — Natoma finds an average of 225 already deployed. What are they doing, and why does it matter?