By Andreas Paech
Head of Data Engineering
gaming ma 2024 cover photo

Data rarely fails because a warehouse cannot compute fast enough. It fails because organisations cannot keep meaning, accountability, and change aligned as teams scale. If you have ever watched a meeting stall on “which number is correct?”, you have seen the real bottleneck: the pipeline may be green, but confidence is red. Decisions slow down – not from lack of data, but from lack of shared understanding. In earlier posts we focused on the technical levers of speed: platform choices and automating data operations (performance optimisation, schema management). This post covers the organisational layer: why data breaks specifically between teams, and what an operating model must make explicit so data velocity remains trustworthy as the company grows.

_____________________

Where the break happens

The moment data crosses a boundary, team, domain, or function, it becomes an interface. And unlike APIs, data interfaces are often implicit. Semantics live in people’s heads, assumptions are undocumented, and changes ship without a predictable change discipline. That is why “more dashboards” rarely fixes the problem. The failure mode is structural. In practice, most cross‑team breakage reduces to three misalignments:

Accountability is ambiguous. When a dataset “belongs to the data team,” it typically belongs to nobody with the context to prevent semantic drift. Upstream teams move fast; downstream teams absorb consequences.

Meaning is unstable. Fields don’t have to change names to change meaning. Attribution logic changes, tracking payloads evolve, or business rules are refactored. The dashboard still updates, yet it may no longer measure the same thing.

Change is unmanaged. Every organisation changes. The difference is whether change is treated as a predictable event (versioned, communicated, validated) or as a surprise discovered after stakeholders complain.

A quick example, the kind that happens everywhere: we once had an upstream tracking adjustment that looked “safe” because the schema was unchanged. The field remained present, but its interpretation shifted (edge cases were reclassified). Downstream reporting stayed technically healthy until stakeholders noticed a discontinuity in performance trends. The issue wasn’t compute; it was an interface change without a shared change protocol.

 

The operating model: what must be explicit

A scalable data operating model is not a re-org chart. It is a set of repeatable decisions about:

  • Who is accountable for which key dataset,
  • How definitions are agreed and preserved across teams,
  • How changes are introduced without breaking data consumption,
  • What the central platform standardises so teams can move independently.

The goal is simple: independent movement without cumulative confusion. Below are three operating-model decisions that consistently prevent cross‑team breakage. They are intentionally framed as decisions, because the failure mode is governance-by-accident, not missing tooling.

1) Move accountability upstream (without creating bureaucracy)

At scale, central teams cannot be the final owners of correctness. They do not control upstream release cadence, and they often lack the deepest domain context. The most reliable pattern is to place accountability with the teams closest to data production, while the central team provides standards, guardrails, and reliability support. This is not “make product engineers do analytics.” It is expanding the definition of done for data-producing change so that shipping a feature includes shipping its data interface responsibly:

  • A clear intent/definition for key events and datasets,
  • Basic validations that run automatically,
  • And explicit expectations for what constitutes a breaking change.

When this becomes normal, the data function stops being a repair shop and becomes what it should be: a reliability and enablement layer that scales.

2) Treat datasets as interfaces and make meaning reviewable

Cross‑team speed requires explicit interfaces. In software we do this with versioning, review, and continuous checks. Data needs the same discipline, adapted to semantics. Call it a “contract” if you like, but the implementation detail matters less than the behaviour: meaning must be reviewable, and change must be predictable. In practice, an interface agreement for a dataset should answer, minimally:

  • What fields exist and which are required,
  • What those fields mean (units, edge cases, business rules),
  • What “acceptable quality” looks like for the use case,
  • What changes are compatible vs breaking,
  • And how consumers will be warned and migrated when breaking change is necessary.

Two hard-learned points: First, documentation alone does not prevent drift. It needs a small amount of automation and review discipline. And second: Perfection is not required. The system improves when agreements are easy to update and treated like code: reviewed, versioned, and continuously checked.

3) Standardise the path, not the decisions

Decentralised accountability without standardisation becomes decentralised chaos. The central platform function creates leverage by standardising the path teams run on, so teams can deliver safely without reinventing patterns. What this looks like in a mature organisation is not “approval gates,” but default workflows that make the right behaviour cheap:

  • Standard delivery patterns (how a dataset is created, tested, documented, and released),
  • Shared health signals (freshness, anomaly detection, change visibility, lineage/blast radius),
  • Access that scales (least-privilege by default, fast exceptions when needed, and clear escalation paths).

A useful gut-check: if teams need to ask permission for routine changes, you have recreated the bottleneck. If teams can change freely but no one can predict impact, you have recreated chaos. The operating model exists to sit between those extremes.

 

Running the data operating model in practice

Operating models fail when they remain philosophical. They succeed when they produce feedback loops that change behaviour. You do not need a large KPI program to start. A light, executive‑readable scorecard can be built from three questions:

  1. Are the most important datasets reliably available when needed?
  2. How quickly do we detect and recover when correctness breaks?
  3. How often do upstream changes break downstream consumers?

These three questions are enough to surface whether your issue is reliability, detection, or change discipline and they are actionable without over-instrumentation. Likewise, you do not need governance theatre. Two lightweight cadences are usually sufficient early on:

  • a short recurring review of upcoming changes that may affect shared datasets,
  • and blameless incident reviews when reporting integrity breaks (focused on prevention, not blame).

Data breaks between teams when accountability is assumed, meaning is implicit, and change is unmanaged. A scalable operating model makes those three things explicit and then runs them as a system. If you want a practical starting point: pick one high-impact dataset this month and make three things unambiguous: the accountable owner, the interface agreement, and the health signals. Then standardise the path so the next domain can adopt it quickly.

That is how you scale speed without scaling confusion.

Be sure to subscribe to our newsletter to stay up to date with the latest news.