Back to feed
Fabric Recent Update·Apr 21, 2026·Hasan Abo Shally

Agentic Fabric: How MCP is turning your data platform into an AI-native operating system


Something fundamental is changing in how developers interact with data platforms. Not a feature update, not a UI refresh, but a shift in the interface itself.

For the past decade, the default way to work with a data platform has been to open a portal, navigate through menus, or write code against REST APIs. Each new tool was a new integration, which meant a new authentication flow, a new set of API wrappers, and hours of plumbing before you could even start building.

The Model Context Protocol (MCP) is an open standard created by Anthropic and now adopted across the industry by GitHub, Cloudflare, Stripe, and others, gives AI agents a universal way to discover, understand, and operate external systems through one protocol. No need for custom integrations or separate authentication stacks for each agent—just one unified standard.

Figure: The Microsoft MCP Server extension in VS Code is now available and ready to install.Figure: The Microsoft MCP Server extension in VS Code is now available and ready to install.

We’re advancing this standard to Microsoft Fabric with two major milestones:

  • Fabric Local MCP (Generally Available) an open-source server that gives AI assistants deep knowledge of Fabric’s APIs, enables local-to-cloud data. operations, and can serve as an execution layer through the Fabric CLI.
  • Fabric Remote MCP (Preview) a cloud-hosted server that lets AI agents perform real, authenticated operations in your Fabric environment—no local setup required.

These aren’t two sides of the same coin. They’re two entry points into the same platform, designed for different contexts, different users, and different levels of autonomy. A developer pair-programming with GitHub Copilot uses the Local MCP to write grounded code and move data between their machine and OneLake. An autonomous agent in Copilot Studio uses the Remote MCP to manage workspaces and permissions on behalf of a team. A CI/CD pipeline uses the Fabric CLI, which the Local MCP can wrap as tools, to deploy changes without any human in the loop.

The common thread: your AI tools now understand Fabric natively. They can write code against the correct APIs, operate on real infrastructure, and do it all within the security model, audit trail, and RBAC boundaries you already trust.

Why choose MCP now?

MCP is to AI what USB was to hardware: a universal connector that replaces a tangle of proprietary cables with a single standard. Rather than creating unique integrations for each AI tool, you can simply expose your platform as an MCP server. Any MCP-compatible client—such as GitHub Copilot, Claude, or Cursor—can then connect instantly.

The timing matters, enterprises are racing to adopt AI agents, but integration complexity remains a top barrier. Building an agent that can simply “create a workspace” requires a full OAuth2 stack, token management, rate-limiting logic, and API versioning. MCP eliminates this plumbing entirely.

The ecosystem is converging on a single protocol, and Fabric is now part of it.

Summary

Fabric local MCP (Generally Available)

In October, we introduced the Fabric Local MCP Server (Preview), the response was remarkable. The announcement became one of our most-read posts, approaching 100K views. Developers made clear they want their AI assistants to understand Fabric deeply and natively.

The Fabric Local MCP runs on your machine and serves two purposes: it helps you build on top of Fabric and enables local-to-cloud data operations.

API documentation and best practices

These tools give your AI agent access to Fabric’s entire API surface without connecting to your environment:

The result: grounded code generation. When your AI assistant writes code to create a workspace or query a lakehouse table, it reads from the actual API spec, not a stale training snapshot.

OneLake: Managing data from local to cloud

These tools connect directly to OneLake and let agents work with your data:

Plus, additional tools for DFS-level item listing, table namespace discovery, and table configuration.

Core: Fabric item operations

This means an agent can look up the correct API spec, generate the code, upload data to OneLake, create items, and inspect table schemas within a single conversation. The local MCP server becomes a bridge between your development and Fabric environment.

What’s new

  • Integrated authentication allows for seamless sign-in flow—no manual token management.
  • Error handling and auto-retry agents recover gracefully from transient failures.
  • Production support is backed by Microsoft’s standard support policies.
  • Telemetry and diagnostics offer visibility into tool usage and performance.

Get started

Recommended method — VS Code extension:
Install the Fabric MCP extension from the VS Code marketplace. It configures everything automatically.

Any MCP client — manual configuration:

Works in VS Code with GitHub Copilot, Cursor, Claude Desktop, and any MCP-compatible client.

Refer to Fabric Local MCP on GitHub to learn more.

Fabric Remote MCP (Preview)

The Fabric Remote MCP is a cloud-hosted MCP server that enables AI agents perform real operations in your Fabric environment. No local installation is required; simply point your AI client to the endpoint, sign in, and start working.

Fabric Remote MCP Server URL: https://api.fabric.microsoft.com/v1/mcp/core

How it works

Your AI Assistant + MCP Protocol (Streamable HTTP) + Fabric Remote MCP + Fabric REST APIs: Entra ID Auth + RBAC and Fabric Audit Logs

Every request flow through Entra ID authentication and the agent operates with your identity, and it can never exceed your permissions. Every action is recorded in Fabric Audit Logs, giving admins full visibility.

Agent capabilities

Pro tip: Compose multiple MCP servers.

The power of MCP is composability, add multiple servers to your agent for richer capabilities:

Quick setup: no installation

  1. In VS Code: Cmd+Shift+P → “MCP: Add Server” → choose HTTP
  2. Enter the URL: https://api.fabric.microsoft.com/v1/mcp/core
  3. Sign in via browser (your Microsoft account)
  4. Ask your AI assistant: “List all my Fabric workspaces”

What’s possible

The real value isn’t in any single tool. It’s meeting developers and agents where they already work and letting them interact with Fabric in the way that fits their context.

For the developer: Build this correctly

“I need a Python script that reads from my lakehouse, transforms the data, and loads it into a warehouse. Show me the right APIs.”

The agent queries the Local MCP for the correct OpenAPI specs, checks best practices for pagination and error handling, and generates code that’s grounded in the actual API. If you need to upload test data, the agent uses OneLake tools to push files directly from your machine.

For the team lead: Set up my project

“Create a new workspace called ‘Q2-Analytics’, add a lakehouse, upload these CSV files, and give read access to the analytics team.”

The agent uses Remote MCP to create the workspace, create the lakehouse, upload files to OneLake, and add workspace roles. All through natural language, all within your permissions, and all audited.

For the platform engineer: Automate this workflow

“Wrap the Fabric CLI as MCP tools in the Local MCP. Your agents can write CLI scripts, backup routines, or migration workflows, with human oversight or fully autonomous, depending on your trust level.”

# Agent generates; you review and run:
fab create Q2-Analytics.Workspace -P capacityname=$CAPACITY

fab create Q2-Analytics.Workspace/raw-data.Lakehouse
fab deploy –config deployment-config.yaml

For the organization: Operate Fabric from Teams

“@FabricBot create a new workspace for the Q3 marketing campaign and add the marketing team as contributors.”

Build a custom agent in Copilot Studio, connect it to the Fabric Remote MCP, and deploy it to Microsoft Teams.

This pattern extends to any channel: Slack, custom web apps, and internal portals. The Remote MCP becomes your organization’s natural-language interface for Fabric operations.

Security by design

Giving AI agents the ability to operate in production environments requires a security first approach. Here’s our approach:

  • RBAC enforcement: Agents operate with the signed-in user’s permissions. No elevation, no bypass.
  • Full audit trail: Every MCP operation is recorded in Fabric Audit Logs, the same infrastructure that tracks portal and API actions.
  • Consequential action controls: Destructive operations are flagged with is_consequential, and agents confirm before proceeding.
  • No bulk data exfiltration: Remote MCP returns metadata and schema information. Bulk data flows through OneLake with its own security boundaries.
  • Admin controls: Fabric administrators manage MCP access through existing tenant settings.

Future directions and objectives

Here are our plans for release and upcoming progress:

Features we’re exploring

  • Service Principal authentication for automated and headless workflows.
  • Deployment pipelines and Git integration via MCP.
  • Job scheduling and monitoring.
  • Domain and folder management.
  • OneLake shortcuts and data access security.
  • Dry run/simulation mode for safe experimentation.
  • Multi-workspace operations.

Questions under consideration

  • What does a fully agentic data platform look like? One where MCP isn’t just a feature, but the primary interface for automation, governance, and operations.
  • How should agents compose across services like Fabric + Graph + GitHub + custom services, and what patterns emerge?
  • What safety controls are required before we can trust agents with production-level operations?

The agentic era is still early, and we’re building in the open and learning as we go. If you try the preview and find gaps — or discover scenarios we haven’t thought of — let us know!

Get started

Local MCP (Generally Available) – Get started with Fabric Pro-Dev MCP Server

Remote MCP (Preview) – Get started with Fabric Core MCP Server

The era of agentic data platforms is here. Whether you’re a developer writing code with an AI assistant, a platform engineer automating deployments, or a team building agents that operate Fabric from Microsoft Teams, MCP is the open standard that connects it all. And Fabric is now part of that ecosystem.

Now your AI tools speak Fabric fluently.

Related blog posts

Agentic Fabric: How MCP is turning your data platform into an AI-native operating system

Evolving Agentic Applications on Microsoft Fabric: From Automated Deployment to Integrating Data Agents

In our previous post, Operationalizing Agentic Applications with Microsoft Fabric, we focused on a core challenge teams encounter once an agentic application moves beyond a proof of concept: operational reality. Specifically, how do you observe, govern, evaluate, and analyze what agents are doing once they interact with real users, data, and business processes at scale? … Continue reading “Evolving Agentic Applications on Microsoft Fabric: From Automated Deployment to Integrating Data Agents”

Answers to common questions about Fabric Data Factory

As the Data Integration Customer Advisory Team (CAT) lead, I spent a lot of time talking to customers at the recent FabCon/SQLCon about Fabric Data Factory, and I came away with a clear picture of what’s on customers’ minds when it comes to the future of data integration. Many of the same questions came up … Continue reading “Answers to common questions about Fabric Data Factory”

Microsoft Fabric

Accelerate your data potential with a unified analytics solution that connects it all. Microsoft Fabric enables you to manage your data in one place with a suite of analytics experiences that seamlessly work together, all hosted on a lake-centric SaaS solution for simplicity and to maintain a single source of truth.

Get the latest news from Microsoft Fabric Blog

This will prompt you to login with your Microsoft account to subscribe

Visit our product blogs

View articles by category

View articles by date

What's new

Microsoft Store

Education

Business

Developer & IT

Company

#AI#Announcements#Fabric platform#Fabric Public APIs#Microsoft Fabric