KB article

Canonical Metrics: One Definition, Many Views

Canonical metrics standardize meaning while allowing flexible reporting views.

arf-kbsemantic-integritycanonical-metricmeasure-inventorysemantic-contract

TL;DR

  • Define each metric once and reuse it everywhere.
  • Views can vary, but the definition must not.

The problem

  • Different teams define the same metric differently.
  • AI cannot infer which definition is the “right” one.

Why it matters

  • Canonical metrics ensure consistency across dashboards and AI answers.
  • They reduce time spent reconciling numbers.

Symptoms

  • The same KPI shows different values in different reports.
  • Stakeholders debate definitions instead of insights.

Root causes

  • Metric definitions stored in docs but not embedded in the model.
  • Local optimization leads to local definitions.

What good looks like

  • One authoritative measure per KPI with clear scope and unit.
  • Variant measures explicitly reference the canonical base.

How to fix

  • Create a metric catalog with owners and definitions.
  • Build canonical measures and update reports to use them.
  • Allow variants only when they explicitly reference the base.

Pitfalls

  • Allowing “temporary” local measures to linger.
  • Hiding canonical definitions in external docs only.

Checklist

  • Canonical metric list published and owned.
  • All KPIs map to a canonical measure.
  • Variants explicitly documented.

Framework placement

Primary ARF layer: Semantic Integrity. Diagnostic bridge: semantic-reliability, change-reliability.