Verified by Garnet Grid

Data Mesh: Decentralized Data Architecture

Implement data mesh principles. Covers domain ownership, data as a product, self-serve data platform, federated governance, and the organizational shift from centralized to decentralized data architecture.

Data mesh flips the centralized data team model on its head. Instead of one data engineering team building pipelines for the entire company, domain teams own their data end-to-end — from production databases through to analytical data products. The central team provides self-serve infrastructure, not data.

This guide covers how to implement data mesh practically, including where teams commonly fail and how to avoid it.


Four Principles

┌──────────────────────────────────────────────┐
│                Data Mesh                      │
│                                              │
│  ┌───────────┐        ┌──────────────────┐   │
│  │ Domain    │        │ Data as a        │   │
│  │ Ownership │        │ Product          │   │
│  │           │        │                  │   │
│  │ Teams own │        │ Discoverable,    │   │
│  │ their     │        │ addressable,     │   │
│  │ data      │        │ trustworthy      │   │
│  └───────────┘        └──────────────────┘   │
│                                              │
│  ┌───────────┐        ┌──────────────────┐   │
│  │ Self-Serve│        │ Federated        │   │
│  │ Data      │        │ Computational    │   │
│  │ Platform  │        │ Governance       │   │
│  │           │        │                  │   │
│  │ Infra as  │        │ Global standards,│   │
│  │ platform  │        │ local autonomy   │   │
│  └───────────┘        └──────────────────┘   │
└──────────────────────────────────────────────┘

Domain Ownership

Before Data Mesh: Centralized

Commerce Team → raw data → Central Data Team → dashboards
Product Team  → raw data → Central Data Team → ML features
Finance Team  → raw data → Central Data Team → reports

                           Bottleneck

After Data Mesh: Decentralized

Commerce Domain
  ├── Source: orders database
  ├── Data Product: "order_events" (streaming)
  ├── Data Product: "daily_order_metrics" (batch)
  └── Owner: Commerce Data Engineer

Product Domain
  ├── Source: product catalog, clickstream
  ├── Data Product: "product_engagement" (batch)
  └── Owner: Product Analytics Engineer

Finance Domain
  ├── Consumes: order_events, product_engagement
  ├── Data Product: "revenue_reports" (batch)
  └── Owner: Finance Data Analyst

Data as a Product

A data product has the same quality expectations as a software product:

data_product:
  name: "daily_order_metrics"
  domain: "commerce"
  owner: "commerce-analytics"
  
  metadata:
    description: "Daily aggregated order metrics by category and region"
    documentation: "https://data-catalog.internal/commerce/daily_order_metrics"
    schema_version: "2.3.0"
    update_frequency: "daily at 06:00 UTC"
    
  quality:
    sla_availability: "99.5%"
    sla_freshness: "< 2 hours from midnight UTC"
    tests:
      - row_count_minimum: 100
      - no_null_primary_keys: true
      - revenue_matches_source_within: "2%"
    
  access:
    discovery: "data-catalog"        # How to find it
    access_method: "BigQuery view"   # How to consume it
    output_ports:
      - type: "sql"
        location: "analytics.commerce.daily_order_metrics"
      - type: "api"
        endpoint: "https://data-api.internal/v1/order-metrics"
    
  lineage:
    sources:
      - "commerce.orders"
      - "commerce.order_items"
      - "product.products"
    consumers:
      - "finance.revenue_reports"
      - "executive.kpi_dashboard"

Self-Serve Data Platform

What the platform team provides:

CapabilityPlatform ProvidesDomain Teams Do
StorageData lake/warehouse infrastructureDefine schemas, manage data
PipelinesOrchestration framework (Airflow, Dagster)Write transformation logic
QualityValidation framework (Great Expectations, dbt)Define quality rules
CatalogData catalog (DataHub, Atlan)Register data products
AccessAccess management infrastructureDefine access policies
MonitoringObservability platformSet alerts and SLAs

Federated Governance

Global Standards (Central)Local Autonomy (Domains)
Naming conventionsTable and column names (within standards)
PII classification rulesWhich columns in their data are PII
Quality testing frameworkSpecific quality thresholds
Schema registrySchema design and evolution
Access control frameworkWho gets access to their data
Data retention policiesHow long to retain their specific data

Anti-Patterns

Anti-PatternProblemFix
Mesh without platformEvery team builds infrastructure from scratchBuild self-serve platform first
Domain teams without data skillsTeams can’t own what they can’t buildEmbedded data engineers or training
Too many data productsGovernance overhead exceeds valueStart with critical 10-20 products
No interoperability standardsDomain products can’t be joinedFederated governance: common identifiers
”Mesh” as renamingRename teams but keep centralized processActual ownership transfer with accountability
Ignoring the journeyJump to full mesh day oneIncremental: 1-2 domains first, scale after learning

Checklist

  • Domains identified with clear data ownership boundaries
  • Data products defined with SLAs, schema, quality guarantees
  • Self-serve platform: pipeline, storage, catalog, quality
  • Federated governance: global standards + local autonomy
  • Data catalog: all products discoverable and documented
  • Quality monitoring: automated tests per data product
  • Pilot: 1-2 domains fully implemented as proof of concept
  • Organizational alignment: leadership supports decentralization
  • Data literacy: domain teams skilled in data engineering

:::note[Source] This guide is derived from operational intelligence at Garnet Grid Consulting. For data mesh consulting, visit garnetgrid.com. :::

Jakub Dimitri Rezayev
Jakub Dimitri Rezayev
Founder & Chief Architect • Garnet Grid Consulting

Jakub holds an M.S. in Customer Intelligence & Analytics and a B.S. in Finance & Computer Science from Pace University. With deep expertise spanning D365 F&O, Azure, Power BI, and AI/ML systems, he architects enterprise solutions that bridge legacy systems and modern technology — and has led multi-million dollar ERP implementations for Fortune 500 supply chains.

View Full Profile →