KB article
Why Multiple Measures for the Same Metric Break AI Answers
Multiple measures for the same metric create conflicting answers and undermine trust.
arf-kbsemantic-integritycanonical-metricmeasure-singularitymetric-sprawl
TL;DR
- AI needs one authoritative definition per metric.
- Duplicate measures lead to inconsistent answers and hidden assumptions.
The problem
- Teams often create new measures to solve local reporting needs.
- Over time, multiple definitions of the same metric coexist in the model.
Why it matters
- AI will pick a measure based on name or context, not intent.
- Conflicting numbers erode confidence in both BI and AI outputs.
Symptoms
- Revenue differs across dashboards with similar filters.
- Two users ask the same question and receive different values.
Root causes
- No canonical metric list or owner.
- Measures are copied and edited instead of reused.
What good looks like
- A single canonical measure per metric with controlled variants.
- Clear measure naming that encodes purpose and scope.
How to fix
- Inventory all measures that represent the same concept.
- Choose a canonical measure and map others to it.
- Deprecate duplicates and update reports to use the canonical version.
Pitfalls
- Renaming measures without updating reports.
- Leaving duplicates because “someone might need them.”
Checklist
- One canonical measure per metric.
- Deprecated measures marked and removed from new use.
- Reports migrated to canonical measures.
Framework placement
Primary ARF layer: Semantic Integrity. Diagnostic bridge: semantic-reliability, change-reliability.