Metadata, Lineage & Context

Technical definitions for metadata and lineage concepts: data catalogs, column-level lineage, knowledge graphs, ontologies, the semantic layer, OpenLineage, active metadata, impact analysis, data contracts, and the six layers of context that AI agents need to make grounded decisions.

Metadata, Lineage & Context

Metadata — data about data — is the nervous system of data infrastructure. Where did this data come from? Who modified it? What depends on it? Is it sensitive? How fresh is it? What does it mean in business terms? Who used it last, and for what?

For BI, metadata supports human discovery and governance. An analyst searches the catalog, finds a table, reads its description, checks its lineage, and decides whether to use it. The metadata serves a human decision-making process that operates on a timescale of minutes to hours.

For AI, metadata is the context that determines whether an agent can make reliable, safe, and compliant decisions — at machine speed. An agent querying customer data needs to know (in milliseconds) what this data represents, how fresh it is, what quality guarantees it carries, who owns it, what governance policies apply, what downstream systems depend on it, and whether the agent's intended use is authorized. The gap between what metadata systems capture today and what AI agents need is one of the largest unsolved problems in data infrastructure.

This cluster covers metadata types, lineage mechanisms, context frameworks, and the tools that attempt — with varying degrees of success — to make data understandable to both humans and machines.


Active Metadata

A category coined by Atlan describing metadata that doesn't just describe data passively but actively drives automation, recommendations, and governance actions. Active metadata systems observe how data is created, used, transformed, and consumed — then use those observations to automate cataloging, suggest improvements, enforce policies, and alert on anomalies.

Traditional metadata is documentation — a description field in a catalog entry, a column comment in a DDL statement, a wiki page maintained by a data steward. Active metadata is operational intelligence — automatically generated, continuously updated, and machine-actionable.

Examples of active metadata in practice: Atlan automatically propagates classification tags (PII, confidential) downstream through lineage — if a source column is tagged as PII, all derived columns inherit the tag without manual intervention. Unity Catalog's automatic lineage captures query-level column dependencies from Spark execution plans and surfaces them in the catalog without user configuration. Snowflake's CLASSIFICATION_PROFILE function uses pattern recognition to suggest sensitivity classifications for columns — active metadata that identifies PII without human review.

The shift from passive to active metadata mirrors the broader shift from manual to autonomous data operations. In a passive system, metadata helps humans make decisions. In an active system, metadata enables machines to make decisions — or at least to propose decisions for human approval. The Autonomous DataOps vision extends active metadata from the catalog layer (where Atlan operates) to the entire infrastructure stack: active metadata that not only describes the data estate but drives provisioning, quality monitoring, governance enforcement, and self-healing.


Business Glossary

A curated, authoritative dictionary of business terms with their precise definitions, owners, and relationships to technical data assets. The business glossary bridges the vocabulary gap between business stakeholders (who talk about "customers," "churn rate," "revenue") and technical systems (which contain dim_customer, fact_monthly_mrr, stg_stripe__charges).

Collibra, Atlan, and Alation all provide business glossary functionality. Unity Catalog supports tags and comments that can serve as lightweight glossary entries. The glossary typically maps business terms to technical assets: "Monthly Recurring Revenue" → defined as SUM(subscription_amount) WHERE status = 'active' AND billing_frequency = 'monthly' → computed in analytics.finance.fact_mrr → owned by the Finance Analytics team.

For AI, the business glossary is a critical context resource. When an agent is asked to "analyze customer churn," it needs to know what "churn" means in this organization — is it cancellation, non-renewal, downgrade, or inactivity? The business glossary provides this grounding. Without it, the agent either hallucinates a definition (using its training data's generic understanding of "churn") or asks the user — neither of which scales.

The gap: business glossaries in current tools are designed for human browsing, not machine consumption. They contain natural-language descriptions, not machine-executable definitions. An agent can read that "churn" is defined as "customers who cancel their subscription within 30 days," but it can't automatically translate that into the correct SQL query against the organization's specific schema without additional context about table structures, join paths, and filter conditions. The semantic layer addresses this gap by providing machine-executable metric definitions.


Column-Level Lineage

Lineage tracking at the granularity of individual columns rather than tables or datasets. Column-level lineage answers: "where did this specific column's values come from?" — tracing a column through every transformation, join, aggregation, and derivation back to its original source.

Databricks Unity Catalog captures column-level lineage automatically from Spark SQL execution plans. When a query executes INSERT INTO target SELECT source.a + source.b AS c FROM source, Unity Catalog records that target.c is derived from source.a and source.b. Lineage is retained for 1 year. The External Lineage API allows ingesting lineage events from non-Databricks systems (Airflow, custom pipelines) into the same lineage graph.

Snowflake provides column-level access tracking through ACCESS_HISTORY, which records which columns were read and written by each query. This is access lineage (who touched what) rather than transformation lineage (how values were derived), though the combination of access patterns across queries can be used to infer transformation lineage.

dbt provides compile-time column-level lineage through its model DAG — the ref() function and SQL parsing enable dbt to trace which source columns flow into which model columns. Tools like dbt-column-lineage and Elementary extend this with richer column-level tracking.

Column-level lineage is critical for AI in three ways. Impact analysis: if source column X changes type or semantics, which downstream features, models, and agent behaviors are affected? Table-level lineage would flag the entire downstream table; column-level lineage pinpoints the specific affected columns, enabling precise impact assessment. PII propagation: proving that a PII column was properly masked at every stage of processing requires column-level tracing — table-level lineage can't distinguish between PII and non-PII columns in the same table. Feature provenance: tracing an ML feature back to its source columns through every transformation step, establishing the complete derivation chain for auditability.


Context Grounding

The process of providing AI agents with sufficient contextual information to make accurate, reliable, and safe decisions. A grounded agent operates with full awareness of what the data represents, where it came from, how it should be used, and what constraints apply. An ungrounded agent operates on raw data without context — and is correspondingly more likely to hallucinate, violate policies, misinterpret semantics, or take inappropriate actions.

Six layers of context are needed for fully grounded agents:

Layer 1 — Metadata and lineage: what this data is, where it came from, how it was transformed, and what depends on it. Provided by catalogs (Unity Catalog, Atlan) and lineage systems (OpenLineage, dbt).

Layer 2 — Query history and usage: how this data has been used historically, by whom, how often, and for what purposes. Provided by query logs (Snowflake QUERY_HISTORY, Databricks audit logs) and usage analytics.

Layer 3 — Human annotations: domain expert knowledge, corrections, business context, and quality assessments attached directly to data assets. Provided by catalog comments, dbt descriptions, and collaboration features (Atlan's Slack-like data conversations).

Layer 4 — Code-level enrichment: which notebooks, SQL scripts, dbt models, and pipeline definitions generated or transformed this data. Links data assets to the logic that produced them. Provided by dbt's model definitions, notebook metadata, and Git integration.

Layer 5 — Access control context: who is authorized to use this data, under what conditions, for what purposes, and with what restrictions. Provided by RBAC/ABAC policies (Unity Catalog grants, Snowflake masking policies) and governance frameworks.

Layer 6 — Evaluation loops: continuous assessment of whether AI outputs derived from this data are correct, safe, and aligned with organizational goals. Provided by MLflow evaluation, LLM-as-judge, and human feedback collection.

No platform integrates all six layers into a coherent context engine. Databricks leads on layers 1, 3, 4, and 6 (through Unity Catalog + MLflow). Snowflake leads on layers 2 and 5 (through query history and granular access controls). But the layers evolved independently on both platforms and don't form a unified context surface that an agent can query holistically.


Data Catalog

A metadata management system that inventories, organizes, and makes discoverable all data assets in an organization. Data catalogs answer the fundamental discovery question: "what data do we have, where is it, who owns it, what does it mean, and can I use it?"

Atlan positions as "GitHub for data teams" — a collaborative catalog with social features (conversations, mentions, bookmarks), active metadata propagation, and AI-powered suggestions. Named a Gartner Magic Quadrant Leader for Metadata Management. Revenue grew 6x over two years, indicating strong market pull.

Alation recently rebranded as an "Agentic Data Intelligence Platform" — pivoting from passive catalog to active agent-building capabilities. The most aggressive repositioning in the catalog market.

Collibra is the governance heavyweight — focused on regulatory compliance (EU AI Act, BCBS 239, HIPAA), data stewardship workflows, and enterprise policy management. Strongest in heavily regulated industries (financial services, healthcare).

Unity Catalog (Databricks) is the platform-native catalog for the Databricks lakehouse — managing tables, volumes, models, and functions with built-in governance, lineage, and AI asset registration (MLflow models as catalog objects). Open-sourced in 2024 for cross-platform adoption.

Snowflake Horizon provides catalog capabilities within the Snowflake ecosystem — discovery, governance, classification, and access management for Snowflake objects.

For AI, catalogs must evolve from human-browsable inventories to machine-queryable knowledge bases. An analyst searching for "customer data" uses a search interface. An agent needing "customer data" requires a programmatic API that returns not just the table location but its schema, quality scores, freshness SLA, governance policies, usage patterns, owner contact, and semantic meaning — all in a structured format the agent can process in milliseconds.


Data Discovery

The process of finding relevant data assets within an organization's data estate — answering "does this data exist, and where?" Data discovery is the first step in any data consumption workflow: before you can query, transform, or feed data to a model, you need to find it.

Traditional discovery is keyword search over catalog metadata — searching table names, column names, descriptions, and tags. Modern discovery adds semantic search (finding tables by meaning, not just keywords — "customer complaints" matches support_tickets even though the words don't overlap), usage-based recommendations ("analysts who used this table also used..."), and lineage-aware discovery ("this table feeds into three production dashboards and two ML models").

Atlan's AI-powered search and Alation's behavioral analytics represent the current frontier for discovery. Unity Catalog provides search within the Databricks platform. Snowflake Horizon provides search within Snowflake.

For AI agents, discovery must be fully programmatic. An agent building a RAG pipeline doesn't "browse" a catalog — it queries an API: "Find all tables containing customer interaction data from the last 90 days, with freshness SLA under 1 hour, that I'm authorized to access, with quality score above 0.95." This query combines metadata search, quality assessment, governance evaluation, and access control — requiring the catalog to function as a multi-dimensional query engine over metadata, not a search interface over descriptions.


Data Lineage

The record of data's complete journey from source to destination: where it was created, every transformation applied, every system it passed through, and where it's ultimately consumed. Lineage answers the forward question ("if I change this source, what breaks downstream?") and the backward question ("where did this dashboard number come from?").

Lineage operates at multiple granularities with increasing value and complexity:

Table-level lineage: Table A feeds Table B via Pipeline C. The coarsest granularity — knows which tables are connected but not which columns or records. Provided by most orchestration tools (Airflow, Dagster) and catalogs.

Column-level lineage: Column A.x is used to compute Column B.y via transformation Z. Traces the derivation of individual columns through the pipeline. Provided by Unity Catalog, dbt, and specialized lineage tools.

Row-level lineage: Record 12345 in Table A contributed to Record 67890 in Table B. The finest granularity — traces individual records through the pipeline. Computationally expensive and rarely available in production systems. Some CDC-based systems (Debezium) can provide row-level tracking for specific pipelines.

Cross-system lineage: Data flows from Salesforce → Fivetran → Snowflake → dbt → Looker. Lineage that spans multiple tools and platforms, requiring either a common lineage standard (OpenLineage) or a lineage aggregator that collects from each tool's native lineage API.

For AI, lineage must extend beyond traditional data pipelines to cover the full AI chain: source data → ingestion → transformation → feature computation → training dataset → model training → model deployment → agent reasoning → action execution → outcome measurement. This chain crosses data infrastructure (source through features), ML infrastructure (training through deployment), and application infrastructure (agent through outcome) — three domains that no single lineage system spans today.


Data Stewardship

The organizational practice of assigning responsibility for data quality, definitions, and governance to specific individuals (data stewards) who serve as the authoritative owners of data domains. Stewardship is the human governance layer that defines what data means, sets quality standards, resolves ambiguities, and ensures compliance.

In practice, data stewardship manifests as: Domain ownership — the Sales Analytics team owns all customer and revenue data, defining its business meaning, quality thresholds, and access policies. Issue resolution — when downstream consumers find data quality problems, the steward investigates and coordinates fixes. Policy definition — stewards define masking rules, retention policies, and usage restrictions for their domain's data.

Collibra provides the most structured stewardship workflow — with defined steward roles, issue tracking, approval workflows, and policy management. Atlan provides lightweight stewardship through ownership assignment and collaborative features. Unity Catalog supports ownership at the catalog/schema/table level.

For AI, stewardship creates an interesting paradox. Human stewards can't review every agent data access at machine speed — there are too many accesses, too fast. But without steward-defined policies, agents have no governance framework to follow. The resolution: stewards define declarative policies and quality thresholds, and the AI control plane enforces them automatically. Stewardship shifts from hands-on data management to policy authoring and exception handling — a role transformation that mirrors how DevOps shifted operations from manual server management to infrastructure-as-code.


Impact Analysis

The process of determining what would be affected by a proposed change to a data asset — before the change is made. Impact analysis answers: "if I modify column X in table Y, which dashboards break, which models degrade, which pipelines fail, and which agents produce incorrect results?"

Impact analysis is the forward traversal of the lineage graph: starting from the changed asset, follow all downstream dependencies to identify every affected consumer. The quality of impact analysis depends directly on the completeness and granularity of the lineage graph.

Table-level impact analysis: identifies which downstream tables consume the changed table. Available from most lineage tools. Useful but imprecise — flagging an entire table as affected when only one column changed.

Column-level impact analysis: identifies which specific downstream columns are derived from the changed column. Available from Unity Catalog, dbt, and specialized tools. Much more precise — a column rename affects only the specific downstream derivations, not all columns in the downstream table.

Cross-system impact analysis: traces impact from the data layer through the ML layer to the application layer. If a source column changes, which features are affected? Which models use those features? Which agents rely on those models? This requires lineage spanning data, ML, and application infrastructure — the metadata knowledge graph provides this.

For autonomous data operations, impact analysis should be automated: before any schema change is committed, the system automatically evaluates downstream impact and either rejects changes that would break critical consumers, alerts stewards for review, or automatically adapts downstream transformations. No production system provides this level of automated impact-driven change management.


Knowledge Graph

A graph data structure storing entities (nodes) and their relationships (edges) as a network of interconnected, typed facts. Unlike relational databases where relationships are implicit in foreign keys, knowledge graphs make relationships first-class objects — queryable, traversable, weighted, and annotatable.

Knowledge graphs represent information as triples: subject → predicate → object. "Customer table → feeds → Revenue model." "Agent-007 → accessed → PII column." "Pipeline-run-42 → produced → anomaly-in-table-X." These triples compose into a rich network where multi-hop queries reveal non-obvious relationships.

In enterprise data, knowledge graphs are used for: Enterprise knowledge management (connecting organizational concepts, processes, and data — Neo4j, Amazon Neptune, TigerGraph). Catalog backends (Atlan uses a graph model internally to represent data assets and relationships). Recommendation engines (traversing user-item-attribute graphs for personalized suggestions).

In the AI-native infrastructure proposal, a metadata knowledge graph connects all entities in the data estate: data assets (tables, columns, files), compute resources (warehouses, clusters, endpoints), actors (humans, agents, services), actions (queries, writes, API calls), policies (access rules, quality thresholds, compliance requirements), and outcomes (model predictions, business metrics, errors). This graph is queryable via Cypher, SPARQL, or SQL through virtual tables.

The knowledge graph enables cross-domain reasoning impossible with relational metadata: "Show me all agents that accessed PII data in the last 24 hours, whose actions resulted in customer complaints, and trace back to which data quality issues in which source tables may have contributed." This query crosses governance, observability, quality, and lineage domains — a traversal requiring a unified graph, not isolated relational systems.


Metadata Knowledge Graph

The specific application of knowledge graph technology to data infrastructure metadata — the architectural backbone of the AI control plane. The metadata knowledge graph unifies metadata from across the data estate into a single queryable structure that enables reasoning impossible with isolated catalog, lineage, and governance systems.

The graph connects five entity types: Data assets (tables, columns, files, embeddings, features — with quality scores, freshness, classifications). Compute resources (warehouses, clusters, serving endpoints — with capacity, cost, performance characteristics). Actors (human users, AI agents, service accounts — with roles, capabilities, delegation chains). Actions (queries, writes, API calls, agent tool invocations — with timestamps, parameters, authorization decisions). Outcomes (model predictions, business metrics, errors, user feedback — with evaluation scores and causal links).

Edges between entities capture relationships with rich metadata: "Agent X (actor) → queried (action, timestamp: T, authorized: true, policy: P) → Customer table (data asset, freshness: 3min, quality: 0.97) → to compute (action) → churn prediction (outcome, confidence: 0.84, evaluation: correct)."

This structure enables queries that no current system can answer: "What is the quality-weighted data provenance chain for predictions that received negative user feedback?" "Which agents have the highest error rate, and does the error correlate with the freshness of their input data?" "If we deprecate this source system, which agent workflows break, and what is the business impact based on historical outcome data?"

No production system implements a comprehensive metadata knowledge graph. Unity Catalog provides a relational catalog with lineage. Atlan provides a graph-based catalog with social features. Neither connects data metadata, agent metadata, action metadata, and outcome metadata in a single structure. Building the metadata knowledge graph is the highest-leverage architectural investment for autonomous data operations.


Metrics Layer / Semantic Layer

A centralized, version-controlled definition of business metrics that ensures consistent calculation across all consumption tools, queries, and AI agents. Instead of every dashboard, report, and agent defining "monthly recurring revenue" with slightly different SQL logic, the semantic layer defines it once — and all consumers reference the shared, authoritative definition.

dbt Semantic Layer (formerly dbt Metrics): defines metrics in YAML within dbt projects, with semantic models specifying dimensions, measures, and entities. Metrics are compiled to SQL at query time, adapted to the target warehouse dialect. Integrations with Tableau, Hex, Mode, and other BI tools allow metric consumption without SQL.

Looker's LookML: a modeling language that defines dimensions, measures, and relationships in a centralized repository. LookML models compile to SQL, ensuring consistent metric computation across all Looker explores and dashboards.

Cube and AtScale: semantic layer platforms that sit between warehouses and consumption tools, providing a consistent metric API that abstracts warehouse-specific SQL.

For AI, the semantic layer is an underappreciated asset. When an agent is asked to "analyze customer churn trends," it needs a machine-executable definition of "churn" — not a natural-language description but a precise SQL computation with filters, time windows, and join paths. The semantic layer provides exactly this: structured metric definitions that an agent can programmatically resolve to SQL.

The gap: current semantic layers define metrics for human consumption — BI tools query the semantic layer to generate dashboards. Extending the semantic layer to serve as a context resource for AI agents — where agents query metric definitions, dimensional relationships, and business rules to ground their reasoning — requires API-first access patterns and richer semantic annotations than current implementations provide.


Ontology

A formal, explicit specification of the concepts, properties, and relationships within a domain — the shared vocabulary and structure that systems use to describe the world. In data infrastructure, an ontology defines what entities exist (tables, columns, pipelines, models, agents), what properties they have (freshness, quality score, owner, sensitivity), and how they relate to each other (feeds, derives, consumes, produces).

The "shared ontology problem" is one of the fundamental reasons tool assembly fails for AI infrastructure. Each tool in the modern data stack has its own ontology — its own model of what exists and how things relate:

Airflow models DAGs of tasks with dependencies, schedules, and execution states. dbt models SQL nodes in a dependency graph with materializations and tests. Monte Carlo models tables with quality metrics — freshness, volume, schema, distribution. MLflow models experiments with runs, each run having parameters, metrics, and artifacts. Unity Catalog models a hierarchy of catalogs, schemas, and tables with governance properties. lakeFS models repositories with branches, commits, and merges.

These six ontologies are incompatible. There's no shared definition of what a "data asset" is (a table? a file? a model?), what a "quality issue" means (a failed test? an anomaly? a distribution shift?), or how an "operation" connects cause to effect. When an alert fires in Monte Carlo about a freshness violation, correlating it with an Airflow DAG failure, a dbt model change, and an MLflow model degradation requires translating across four ontologies manually.

The metadata knowledge graph resolves this by defining a unified ontology that all components map to — a shared language for the entire data estate.


OpenLineage

An open standard for data lineage collection and interchange, enabling cross-system lineage tracking without vendor lock-in. OpenLineage defines a common event model: a Run (an instance of a job executing) reads from Input Datasets and writes to Output Datasets, with structured Facets carrying metadata about schemas, data quality, SQL queries, and custom properties.

The event lifecycle: START (job begins, declaring intended inputs), RUNNING (periodic updates during execution), COMPLETE (job finishes, declaring actual outputs with quality facets), or FAIL (job fails, with error facets). Events are emitted to a lineage backend (Marquez, Atlan, Databricks External Lineage API) via HTTP.

OpenLineage integrations: Airflow (OpenLineageListener captures DAG-level lineage), Spark (OpenLineageSparkListener captures job-level lineage from query plans), dbt (native OpenLineage emission), Flink (lineage from streaming jobs), and Great Expectations (quality results as lineage facets).

Databricks' External Lineage API can ingest OpenLineage events into Unity Catalog, enabling lineage from external systems (Airflow orchestrating Snowflake queries, for example) to appear alongside native Databricks lineage in a unified graph.

For AI, OpenLineage needs extension to cover AI-specific operations: model training runs (reading training data, producing model artifacts), feature computation (reading raw data, producing feature vectors), inference (reading features, producing predictions), and agent actions (reading context, producing decisions and actions). These extensions don't yet exist in the standard — and defining them is a prerequisite for end-to-end AI lineage that spans data infrastructure through agent behavior.


Provenance

The documented origin and transformation history of a data asset — a complete record of where data came from and everything that happened to it. Provenance is lineage with an emphasis on auditability and trust: not just "what is the data flow?" but "can we prove, to a regulator or auditor, the exact chain of custody from source to output?"

Data provenance encompasses: Source provenance (which external system, API, or file produced this data, at what time, with what credentials). Transformation provenance (which code, at which version, with which parameters, transformed the data at each step). Access provenance (who or what accessed this data, when, for what purpose, and whether the access was authorized). Output provenance (which reports, predictions, decisions, or actions were derived from this data).

For AI systems, provenance is a regulatory requirement. The EU AI Act requires providers of high-risk AI systems to document "the data sets used, including information about the data and data preparation processes, the training methodologies and techniques used." BCBS 239 requires banks to maintain complete data lineage for risk reporting. HIPAA requires audit trails for protected health information.

The gap: current provenance in data platforms is fragmented. Snowflake's ACCESS_HISTORY tracks access provenance. Unity Catalog's lineage tracks transformation provenance. MLflow tracks training provenance. But the full chain — from source system through transformation through training through deployment through agent action through business outcome — exists in no single system. Assembling complete provenance requires querying multiple systems and manually stitching the results.


Schema Registry

A centralized service that stores, versions, and enforces data schemas — ensuring that data producers and consumers agree on the structure of exchanged data. Schema registries are the enforcement mechanism for data contracts in streaming and event-driven architectures.

Confluent Schema Registry (the dominant implementation for Kafka ecosystems) stores Avro, Protobuf, and JSON Schema definitions for Kafka topics. Producers register schemas before publishing; consumers retrieve schemas for deserialization. The registry enforces compatibility rules: BACKWARD (new schema can read old data), FORWARD (old schema can read new data), FULL (both), or NONE (no compatibility guarantee). Schema changes that violate the compatibility rule are rejected.

AWS Glue Schema Registry provides similar functionality for AWS-native streaming architectures. Databricks uses Unity Catalog as a schema registry for Delta tables, with schema enforcement on write. Snowflake enforces schemas at table definition time.

For data infrastructure, schema registries provide the structural contract layer — ensuring that schema changes don't silently break downstream consumers. For AI workloads, schema registries are necessary but insufficient: they guarantee structural compatibility (the columns exist with the right types) but not semantic compatibility (the column still means the same thing) or statistical compatibility (the distribution hasn't shifted). A column can maintain its schema (STRING, not null) while completely changing its content — a semantic drift that schema registries can't detect.


Semantic Observability

A monitoring approach focused on understanding the meaning of AI system behavior — not just operational metrics (latency, throughput, error rate) but the reasoning chain that explains why an AI system made a specific decision.

Traditional observability tells you that something happened: "Agent queried table X at 14:23:07, latency 320ms, returned 1,247 rows, cost $0.03." Semantic observability tells you why it happened and whether it was correct: "Agent queried table X because it was resolving customer complaint #89012. It retrieved the customer's 3-year interaction history, identified a pattern of recurring billing errors, determined the complaint was justified based on company policy §4.2, and proposed a $150 credit — which was approved by the governance layer because the agent's delegation authority includes credits up to $500."

MLflow Tracing is the closest production implementation: hierarchical span trees with typed span types (AGENT, LLM, TOOL, RETRIEVER, EMBEDDING, RERANKER, PARSER, CHAT_MODEL), input/output capture at each span, and the ability to attach typed assessments (HUMAN, CODE, LLM_JUDGE) directly to traces. The OpenTelemetry foundation enables integration with existing observability infrastructure (Jaeger, Zipkin, Datadog, Grafana).

Semantic observability is the prerequisite for debugging, auditing, and improving AI systems. When an agent makes a bad decision, reviewing the query log tells you what data it accessed. Reviewing the semantic trace tells you what data it considered, what reasoning it applied, what alternatives it evaluated, why it chose the action it did, and where in that chain the error occurred. This level of introspection is the difference between "the agent was wrong" and "the agent was wrong because the retriever missed a relevant policy document due to an embedding quality issue in the governance knowledge base."


Shared Ontology Problem

The fundamental incompatibility between the metadata models used by different tools in the modern data stack — the reason that assembling autonomous data operations from best-of-breed tools fails even when each tool is excellent in isolation.

The problem manifests concretely: An Airflow DAG failure is recorded as a task instance with a state, an exception, and a timestamp. A Monte Carlo anomaly is recorded as a table-level alert with a severity, a type, and affected columns. A dbt test failure is recorded as a node with a status and a message. An MLflow model degradation is recorded as a metric change across experiment runs. These four events may all stem from the same root cause — a source schema change — but correlating them requires translating between four incompatible metadata models.

The problem compounds at the automation layer. An autonomous system that detects a quality issue, diagnoses its root cause, and remediates it must reason across multiple tools' metadata. But "reasoning across" requires a shared language — shared definitions of what entities exist, what events mean, and how causes relate to effects. Without this shared ontology, automation is limited to within-tool scope: Monte Carlo can auto-detect anomalies within its purview, Airflow can auto-retry within its DAGs, dbt can auto-test within its models — but no system can reason about the cross-tool chain from root cause to downstream impact.

The metadata knowledge graph resolves the shared ontology problem by defining a unified entity-relationship model that all tools map to. Each tool emits events in its native format; an integration layer translates those events into the shared ontology; the knowledge graph stores the unified representation; and automation reasons over the unified graph. This is architecturally analogous to how OpenTelemetry unifies observability across heterogeneous services — a common data model that each service's native telemetry maps to.


Part of The Sciencer Wiki. Maintained by The Sciencer Company.