← All articles
Semantic LayerApril 21, 20268 min read

MCP and the Semantic Layer: How AI Agents Finally Get Governed Business Context

By Wesley Nitikromo

AI agents can reason, plan, and execute complex multi-step workflows faster than any human analyst. What they cannot do — without significant architectural work — is understand what "revenue" means at your company. Or why "active customer" means something different in your CRM than it does in your data warehouse. Or why gross margin calculations vary by business unit and product line in ways that exist only in a policy document nobody has updated in 18 months.

This is the governed business context problem. And it is why the majority of enterprise AI deployments underdeliver. The model is not the bottleneck. The absence of shared, machine-readable business meaning is.

Model Context Protocol (MCP) paired with a governed semantic layer is the first serious architectural answer to this problem at scale.

What MCP Actually Is

MCP is an open protocol, introduced by Anthropic in late 2024 and rapidly adopted across the AI industry, that standardizes how AI agents connect to external data sources, tools, and systems. Think of it as a universal interface layer between AI agents and the world they need to reason over.

Before MCP, connecting an AI agent to your data required custom engineering for every integration. Want your agent to query Snowflake? Build a connector. Want it to call dbt metrics? Build another connector. Want it to access Cube? Another connector. The combinatorial explosion of N agents times M data sources meant that most teams either over-engineered fragile bespoke plumbing or dramatically limited what their agents could access.

MCP collapses this. One protocol, one interface definition. Any AI agent that speaks MCP can talk to any MCP server — whether that server wraps a database, an API, a document store, or, most importantly for data teams, a semantic layer.

The Semantic Layer as an MCP Server

Here is where this becomes interesting for anyone building AI on enterprise data.

The semantic layer — the governed abstraction that translates raw warehouse tables into trusted business metrics — is the ideal MCP server. When a semantic layer exposes an MCP interface, AI agents can query governed metric definitions directly. Not raw SQL. Not vector embeddings of documentation. Actual business logic: defined, versioned, and approved by the teams who own the numbers.

An AI agent asking "what was Q1 revenue by region?" through an MCP-connected semantic layer gets back a number calculated using the exact formula your finance team approved. The same formula in your board deck. The same formula your CFO trusts. With full lineage showing which source tables contributed, how fresh the data is, and which version of the metric definition was applied.

This is not an incremental improvement. It is a categorical shift in what enterprise AI can reliably do.

Internal testing across vendors shows LLM accuracy on business questions in raw table access mode runs around 40%. With a governed semantic layer as the context source, that number exceeds 83%. The difference is not model capability. It is whether the agent has access to a trusted definition or is making an educated guess based on column names and sample data.

What Breaks Without It

The failures that MCP-grounded semantic access prevents are specific and expensive.

Without governed context, agents query raw tables and infer metric logic from whatever they find. An agent sees a column called rev_gross_adj and makes a reasonable inference about what it means. That inference is wrong in ways the agent cannot detect. The business loses trust after the third time a board-level number has to be restated. The AI initiative stalls — not because the model failed, but because nobody built the semantic foundation underneath it.

The second failure mode is governance invisibility. When an agent makes a wrong recommendation, can you trace it? With raw table access, the answer is usually no. You know which query ran but not which business logic should have been applied. With MCP and a semantic layer, every agent action is traceable: which metric version, which source data, which access controls were enforced at query time.

This traceability is not optional in financial services, healthcare, or any environment where AI decisions carry regulatory implications. But even in unregulated environments, it is the only reliable way to debug agents that return unexpected results — and restore the trust of the teams who need to act on them.

The Architecture That Works in Production

A production-ready model context protocol semantic layer architecture has four components working in sequence.

The semantic layer is where business logic lives — whether that is dbt's MetricFlow-powered semantic layer, Cube Cloud, AtScale, or platform-native options like Snowflake Semantic Views. Metric definitions, entity relationships, access controls, and join paths are defined here once and reused everywhere downstream. This is Layer 2 of the Intelligence Allocation Stack: the layer that translates raw data into business meaning.

The MCP server exposes the semantic layer to AI agents through the standardized protocol. AtScale has a production MCP server today. Cube is building one. dbt's MCP integration is in progress. If your semantic layer vendor does not have an MCP server yet, you are back to custom connectors — which works as a bridge, but not a long-term architecture.

The orchestration layer (Dagster, Airflow, Prefect) keeps the semantic layer fresh. This is the component most teams underinvest in. MCP-grounded agents query governed definitions, but those definitions are only trustworthy if the underlying data is current. Orchestration ensures semantic layer refreshes are triggered as part of the same pipeline that runs your transformations — not as a manual afterthought on someone's calendar.

The AI agent layer — LangChain, CrewAI, Claude via API, or platform-native agents — queries the semantic layer through MCP with identity-aware access controls. The agent authenticates, and the MCP server enforces row-level security and column masking using the same policies applied to human analysts. Governance does not have to be re-implemented at the prompt layer. It is enforced at the data layer, where it belongs.

The Vendor Lock-in Risk Nobody Is Talking About Enough

There is a strategic decision embedded in this architecture that deserves more attention than it gets: where does the MCP server live?

Snowflake's Cortex Analyst and Databricks' LakehouseIQ both provide AI agent access to data — and the semantic context is embedded in the platform. This is efficient. It is also a strategic trap.

When your business logic is encoded inside your cloud data warehouse, your organizational intelligence becomes a hostage to that vendor's infrastructure decisions. Changing clouds means rebuilding your semantic layer from scratch. Adopting a new AI framework means negotiating with your platform vendor's roadmap. The semantic context your company has spent years accumulating — all the nuanced definitions, the edge cases, the business rules that exist nowhere in documentation — gets locked inside someone else's infrastructure.

The semantic layer should be the most portable component in your entire data stack. An MCP server that wraps a headless, platform-agnostic semantic layer gives you the same governed access pattern while preserving the freedom to change your warehouse, your AI models, or your BI tooling without starting over. This is not a theoretical concern. It is the architectural decision that separates organizations that own their intelligence from those that have licensed it from a cloud provider.

This is why tool selection matters as much as the integration pattern. See the Semantic Layer Tools Compared breakdown for how dbt, Cube, and AtScale differ on portability — because that difference becomes critical the moment your MCP server is in production.

Where This Is Headed: Semantic Observability

The immediate value of connecting MCP to a semantic layer is governed, traceable AI access to business metrics. The architectural implications extend further than most teams are thinking about today.

As MCP adoption matures, semantic observability becomes possible: real-time monitoring of how AI agents interpret and apply business logic, with alerting when interpretations drift from approved definitions. Not just "did the agent query the right metric?" but "is the agent applying the metric in the way the business intends?"

Semantic drift — the gradual divergence between what a metric definition says and how it gets applied in practice — is the silent killer of enterprise AI trust. It existed before AI agents. Humans navigated it through institutional knowledge and quarterly reconciliation meetings. AI agents cannot attend reconciliation meetings. They apply whatever logic they can access at query time, consistently and at machine speed.

The semantic layer, connected through MCP, is the only layer positioned to detect and correct semantic drift in real time. It becomes not just the source of business definitions but the audit layer for AI behavior at scale. This is what enterprises that get AI right in 2027 will have built in 2026.

The Control Plane for Enterprise AI

MCP makes the semantic layer the control plane for enterprise AI. Not a nice-to-have feature in your BI tool. Not a backlog item for Q3. The connective tissue between your business definitions and every AI system that will reason over your data — today and for the next five years.

The organizations building this infrastructure now are building for every agent they will deploy in the future. The ones waiting are accumulating technical debt in the form of fragile bespoke integrations, ungoverned raw table access, and AI systems whose outputs cannot be fully explained or trusted.

That is what allocating intelligence to the right layer means in practice. The semantic layer is Layer 2 of the Intelligence Allocation Stack. MCP is the protocol that makes Layer 2 available to everything in Layer 4. Build the layer before you need it urgently. You will need it sooner than you think.

Wesley Nitikromo

Founder of Unwind Data. Previously co-founded DataBright (acquired 2023). Data architect, analytics engineering specialist, and builder of AI-ready data infrastructure. Based in Amsterdam.