Definition
A process documentation data quality framework defines measurable rules for model and metadata health (completeness, timeliness, uniqueness, consistency) and turns them into scorecards with ownership and remediation—so the repository stays trustworthy over time.
- Don’t govern by policy documents—govern by **scorecards**.
- Every red metric must have an owner, SLA, and escalation path.
- Start with 4 dimensions: completeness, timeliness, uniqueness, consistency.
- Add drift detection: models that diverge from execution are risk, not documentation debt.
The hidden risk: process repositories decay silently
Process repositories rarely fail loudly. They fail quietly:
- models are “mostly right” but not reliable enough for audits
- duplicates and variants multiply
- metadata fields are empty or inconsistent
- owners change and accountability disappears
This creates a predictable outcome: stakeholders stop trusting the repository and process work becomes a one-off workshop activity again.
A data quality framework changes this by making health measurable and actionable.
The 4 core dimensions (and what they mean for processes)
1) Completeness
Are the required fields present?
- owner + reviewer
- scope (products, regions, systems)
- last reviewed date
- controls coverage where relevant
2) Timeliness
Is the model reviewed within your policy window?
- e.g. high-risk journeys: every 90 days
- lower-risk: every 180–365 days
3) Uniqueness
Do you have duplicates or overlapping variants?
- same process name, different content
- same content, different objects
4) Consistency
Do models follow conventions so teams can understand them?
- naming standards
- lane strategy
- gateway usage and exception patterns
Anti-pattern: measuring without remediation
Scorecards without owners become theatre. The framework only works when every red metric triggers a remediation workflow and has a clear accountable person.
Scorecards: RAG thresholds and “definition of done” for process models
Create a simple scorecard for each model and for the landscape overall.
A practical starting point:
- Green: required metadata complete; reviewed on time; no duplicates; conventions pass
- Yellow: minor gaps; fix within SLA
- Red: missing ownership, outdated review, or control-relevant gaps; escalate
Then define Definition of Done for publishing:
- minimum metadata filled
- owner assigned
- review + approval recorded
- exceptions modeled with a standard pattern
Make the score visible where work happens
If teams have to open a separate dashboard to see health, they won’t. Surface health in the repository UI and in publish workflows.
Ownership: model owner, control owner, data owner
Most quality programs fail because ownership is vague.
Define three distinct roles:
- Model owner: accountable for the BPMN lifecycle and semantic correctness
- Control owner (optional but critical in regulated ops): accountable for control design + evidence points
- Data owner: accountable for evidence sources (logs, dashboards, integrations)
Then attach SLAs:
- outdated review: owner notified, then escalated
- missing metadata: cannot publish
- duplicate detection: assigned remediation ticket
Drift detection: connecting “model health” to “reality health”
Completeness and timeliness aren’t enough.
If the model is complete but wrong, it is still risk.
Add drift signals:
- conformance checking (should vs is) where event logs exist
- spot checks via walkthrough recordings
- exception volumes (too many exceptions means the main path is wrong)
Related reading:
Rollout strategy: start small, scale with templates
A realistic rollout plan:
- Pick 20–30 high-value models (risk + volume).
- Implement the four dimensions + RAG thresholds.
- Add a remediation workflow that creates tasks automatically.
- Publish a weekly scorecard update (no meetings required).
- Scale via templates: metadata schemas, naming conventions, exception patterns.
Common mistakes to avoid
Learn from others so you don't repeat the same pitfalls.
Relying on annual reviews only
Repositories drift monthly, not yearly.
Use timeliness scorecards with escalations and lightweight cadence.
Making fields optional
Optional metadata becomes missing metadata.
Define required fields and block publishing when they’re missing.
Treating duplicates as harmless
Duplicates destroy trust and create inconsistent controls coverage.
Detect duplicates and assign remediation tickets with owners.
Expert insights
What the experts say
"Your repository is either a living system with scorecards and ownership—or it becomes a museum of outdated diagrams."
Process Governance Architect
Take action
Your action checklist
Apply what you've learned with this practical checklist.
Define required process metadata fields (CDEs)
Set RAG thresholds for completeness/timeliness/uniqueness/consistency
Assign owners and SLAs for each rule
Block publishing when required fields are missing
Publish weekly scorecards and auto-create remediation tasks
Add drift detection for high-risk journeys