A sociotechnical approach that distributes data ownership to domain teams, treats data as a first-class product, and enables organizations to scale data capabilities without central bottlenecks.
Traditional data lakes and warehouses accumulate three systemic failure modes as organizations grow. Data Mesh exists to eliminate each one.
A single central data team receives requests from every domain in the organization. As the company scales, pipeline queues grow, delivery slows, and the team becomes a chronic blocker for business decisions.
When the Orders domain's data is modeled by a central team, critical business nuance gets lost. The team owning the pipeline doesn't own the domain — they make assumptions, introduce drift, and produce data nobody fully trusts.
All data flows through a single lake or warehouse. Schema changes are global events. A bad migration breaks 40 downstream consumers. Coupling creates fragility — the larger the platform, the harder it is to change anything safely.
Just as monolithic applications struggled to scale (one team, one deploy, one failure domain), monolithic data platforms exhibit the same pathologies.
Microservices broke apps into independently-deployable services, each owned by a team. Data Mesh breaks data into independently-owned data products, each with its own pipeline, schema, and SLA.
Both paradigms accept distributed complexity in exchange for organizational scalability. Neither is a silver bullet — both require platform investment and team maturity to succeed.