KB article

Semantic Drift: How Definitions Quietly Change Over Time

Semantic drift happens when metric meaning changes without clear communication.

arf-kbsemantic-integritysemantic-driftsemantic-contractlineage

TL;DR

  • Drift is usually accidental and cumulative.
  • AI amplifies drift by spreading inconsistent definitions.

The problem

  • Metric definitions change as business rules evolve.
  • Old reports and AI prompts still assume previous meanings.

Why it matters

  • Drift makes trend analysis misleading.
  • AI may mix old and new definitions in a single answer.

Symptoms

  • Historical comparisons suddenly look “off.”
  • Same question yields different results before and after a model update.

Root causes

  • Changes to DAX logic without updating descriptions.
  • No change control for metric definitions.

What good looks like

  • Definition changes are versioned and documented.
  • Historical data is recomputed or clearly segmented.

How to fix

  • Add explicit version notes to metric descriptions.
  • Implement change review for KPI definitions.
  • Communicate drift events to report owners.

Pitfalls

  • Assuming “the data didn’t change” means the metric didn’t change.
  • Leaving old measures in place without deprecation.

Checklist

  • Definition change log exists.
  • Metrics have owners and review cycles.
  • Old definitions are deprecated or versioned.

Framework placement

Primary ARF layer: Semantic Integrity. Diagnostic bridge: semantic-reliability, change-reliability.