AI Analytics Reference Architecture for Data Leaders
At a glance
- McKinsey's 2025 State of AI says AI use is broad, but scaling remains limited, which is exactly the pattern reference architectures are meant to address.
- NIST AI 600-1 treats governance, documentation, evaluation, and human oversight as part of the deployment system, not as afterthoughts.
- The dbt Semantic Layer lets teams define metrics once and reuse them downstream, which makes it a common semantic layer in an AI analytics stack.
- Snowflake Semantic Views store business concepts, metrics, and relationships directly in the warehouse.
- Databricks metric views centralize business metrics and let users query them across dimensions at runtime.
- The MCP specification standardizes how hosts, clients, and servers exchange resources, prompts, and tools across AI applications.
- BigQuery conversational analytics and the Conversational Analytics API show a production pattern where data agents use context, tools, and access controls rather than raw schema alone.
- Snowflake Cortex Analyst says schemas alone lack business-process definitions and metric handling, which is why semantic models and verified examples improve reliability.
McKinsey's 2025 State of AI says 88 percent of respondents report regular AI use in at least one business function, but only about one-third say their organizations have begun scaling AI across the enterprise. The technical reason is often architectural: many teams add an AI interface before they decide where definitions, context, permissions, and evaluation actually live.
What This Architecture Is For
This is not a component glossary. It is a stack-level blueprint for answering one question:
If a business user asks a natural-language question about company data, which layers determine whether the answer is correct, governed, and reusable?
The architecture below is designed for data leaders who want:
- self-serve answers without shadow BI
- reusable metric logic
- governed access paths
- portability across multiple agent interfaces
- an evaluation loop that survives beyond the pilot
If you only need a component definition, read what is a context layer or what is MCP. This post is about how those pieces fit together.
The Six Layers
1. Source and Storage Layer
This is the physical data foundation:
- warehouses and lakehouses
- operational systems
- transformation outputs
- raw access policies
At this layer, data is stored, transformed, and protected. But business users do not think in table names or physical joins. That translation happens higher in the stack.
The important architectural point is that governance should begin here, not here alone. Access policies, lineage primitives, and authoritative tables come from this layer, but they do not automatically become understandable to AI systems.
2. Semantic Layer
The semantic layer standardizes business logic:
- metric definitions
- dimensions
- joins
- aggregation rules
This is the layer represented by tools and patterns such as the dbt Semantic Layer, LookML, Snowflake Semantic Views, and Databricks metric views.
It is the first place where business definitions become executable and reusable, which is why it is foundational to trustworthy AI analytics.
3. Context Layer
The context layer extends the semantic layer into a broader machine-readable operating context.
In practice, that includes:
- semantic definitions
- lineage
- metadata
- business glossary terms
- documentation
- verified examples
- defaults and exclusions
- source-system context
This layer exists because AI systems need more than a formula. They need to know which entities matter, which synonyms map to which terms, which sources are canonical, and how to explain the result in a way a reviewer can check.
That is why context layer vs semantic layer is not a category battle. The semantic layer is a necessary subset. The context layer is the broader operational surface.
4. Policy and Control Layer
The policy layer carries security and governance into AI delivery:
- row and column-level access
- audit logging
- human review rules
- retention and deletion controls
- interface-specific permissions
NIST AI 600-1 reinforces that governance and monitoring belong inside deployment design. Snowflake Cortex Analyst explicitly ties question answering to RBAC and semantic access controls, and BigQuery conversational analytics documents dedicated IAM roles and permissions for agent operations.
If this layer is missing, the architecture might still answer questions. It just cannot be trusted.
5. Agent and Interface Layer
This is where users actually interact with the system:
- Slack bots
- embedded chat
- email digests
- BI copilots
- custom APIs
- general-purpose assistants
The key architectural decision here is whether those interfaces consume shared context or each build their own interpretation of the data.
BigQuery's Conversational Analytics API shows one production pattern: the system receives business context plus tools such as SQL and Python. The MCP specification shows another important pattern: a standardized protocol for exposing resources, prompts, and tools across AI applications.
This is the layer where portability matters. If context is only usable inside one tool, the architecture is brittle.
6. Evaluation and Improvement Layer
Every production AI analytics stack needs a loop for:
- testing real questions
- validating correctness
- capturing failures
- improving context
- tracking regressions
Snowflake Cortex Analyst includes evaluation of semantic views, while Snowflake's verified-query optimization flow uses verified examples to expand coverage. BigQuery authored context recommends structured context, example queries, glossary terms, and relationship definitions for exactly the same reason.
Without this layer, the stack looks complete but cannot improve safely.
How the Layers Work Together
The architecture works like this:
- data is stored and governed in source systems
- metric logic is standardized in a semantic layer
- context broadens those definitions with metadata, documentation, and lineage
- policies constrain what each interface can expose
- agents and applications consume that shared context
- evaluation feeds corrections back into the context and semantic layers
The stack fails when those responsibilities collapse into one layer. For example:
- if you keep definitions inside the chat app, you get shadow BI
- if you rely only on warehouse schemas, you lose business meaning
- if you rely only on semantic definitions, you still miss documentation, examples, and interface behavior
- if you skip evaluation, drift goes undetected
Common Anti-Patterns
The UI-First Stack
The team starts with a chatbot or copilot and tries to add governance later. This creates an attractive interface with no stable definition layer underneath it.
The Semantic-Layer-Only Stack
The team centralizes metric logic and assumes the job is done. That helps, but it leaves gaps around documentation, examples, lineage exposure, and portable agent access.
The Single-Agent Stack
The first agent works, so the team hardcodes context into it. The second interface then has to repeat the work. By the third interface, no two experiences behave the same way.
The Evaluation-Free Stack
The team goes from demo to rollout without a controlled question set or owner-approved baselines. Failures are discovered only after executives start relying on answers.
Design Principles for Data Leaders
If you are reviewing or redesigning your stack, use these principles:
- Keep business definitions separate from interfaces.
- Make context portable across multiple agent surfaces.
- Treat access and auditability as architectural requirements.
- Give the system verified examples and glossary context, not only schemas.
- Build a correction loop before expanding scope.
These principles sound basic, but they are what separate a pilot from a reusable operating model.
How Kaelio Fits
Kaelio auto-builds a governed context layer from your data stack. Its built-in data agent (and any MCP-compatible agent) can then deliver trusted, sourced answers to every team.
In the reference architecture above, Kaelio sits between the semantic layer and the agent layer. It ingests definitions, lineage, BI logic, and documentation from across the stack, organizes them into one governed layer, and exposes that layer to multiple AI interfaces.
That means the same approved context can support:
- Kaelio's built-in data agent
- custom applications
- Slack and email workflows
- any MCP-compatible assistant
The value is not only better answers. It is architectural reuse.
FAQ
What is an AI analytics reference architecture?
An AI analytics reference architecture is a practical blueprint that shows how warehouses, semantic definitions, context, access controls, agent interfaces, and evaluation loops fit together in a production AI analytics stack.
Why is a semantic layer not enough on its own?
A semantic layer standardizes metrics and joins, which is essential. But production AI analytics also needs documentation, lineage, access controls, interface-level delivery, and evaluation workflows that extend beyond metric definitions alone.
Where does MCP fit in the architecture?
MCP fits at the agent-interface layer. It standardizes how AI applications connect to external resources, prompts, and tools, which makes governed context more portable across different agent experiences.
What is the most common AI analytics architecture mistake?
The most common mistake is letting each agent or application invent its own definitions and access logic. That creates multiple analytics surfaces with different answers, which is how shadow BI forms.
How does Kaelio fit into an AI analytics reference architecture?
Kaelio sits in the context-layer role. It auto-builds a governed context layer from warehouses, BI tools, semantic systems, and documentation so its built-in data agent, and any MCP-compatible agent, can answer using shared definitions, lineage, and source context.
Sources
- https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai/
- https://doi.org/10.6028/NIST.AI.600-1
- https://modelcontextprotocol.io/specification/2025-06-18
- https://docs.getdbt.com/docs/use-dbt-semantic-layer/dbt-sl
- https://docs.snowflake.com/en/user-guide/views-semantic/overview
- https://docs.snowflake.com/en/en/user-guide/snowflake-cortex/cortex-analyst
- https://docs.cloud.google.com/gemini/data-agents/conversational-analytics-api/overview
- https://docs.cloud.google.com/gemini/data-agents/conversational-analytics-api/data-agent-authored-context-bq
- https://docs.cloud.google.com/bigquery/docs/conversational-analytics
- https://docs.databricks.com/aws/en/business-semantics/metric-views