Compliance Levels
OpenReason Protocol defines four compliance levels: three prospective levels (Basic, Standard, Full) for real-time decision-making, and one retrospective level (PostHoc) for analyzing past decisions. Choose based on your resource constraints, stakes, and audience.
Overview
| Level | Layers | Use Case | Typical Time Investment |
|---|---|---|---|
| ORP-Basic | Layer 1 only | Initial documentation, low-stakes decisions | 2–4 hours |
| ORP-PostHoc | Layer 1 only (post-hoc) | Retrospective analysis of past decisions/datasets | 1–3 days |
| ORP-Standard | Layers 1–3 | Policy proposals, contested decisions | 1–2 days |
| ORP-Full | Layers 1–5 | High-stakes, long-term, or highly contested | 3–5 days |
ORP-Basic (Layer 1 Only)
What it requires:
- Header metadata
- Data provenance for all datasets used
What it provides:
- Transparency about data sources
- Explicit documentation of exclusions and limitations
- Foundation for others to build on
Best for:
- Initial documentation before full analysis
- Low-stakes internal decisions
- Resource-constrained contexts
- Establishing data transparency as a baseline
Not sufficient for:
- Public policy proposals affecting >1000 people
- Decisions with distributional impacts across groups
- Contested proposals where stakeholders disagree
Example Use Cases
- Internal team decision on which dataset to use for analysis
- Blog post documenting data sources for public consumption
- Preliminary documentation before full proposal
- Academic research data provenance tracking
Creating an ORP-Basic Document
orp new my-doc.yaml --compliance basicThe tool will prompt you for:
- Document title and ID
- Author information
- Data provenance details (source, methodology, limitations)
Time investment: 2-4 hours, mostly spent documenting datasets accurately.
ORP-PostHoc (Layer 1, Post-Hoc Reconstruction)
What it requires:
- Header metadata (with
post_hoc: trueflag) - Data provenance for datasets (Layer 1)
- Analyst/reconstructor role (not original decision-maker)
- Accountability gap documentation
What it provides:
- Retrospective transparency for past decisions
- Documentation of what is unknown/unknowable
- Foundation for understanding historical choices
- Accountability gap identification
Best for:
- Analyzing AI training datasets created without ORP
- Reconstructing past policy decisions
- Academic analysis of historical choices
- Post-hoc transparency when original documentation is absent
Not sufficient for:
- Real-time decision-making (use ORP-Basic/Standard/Full instead)
- Prospective governance (requires full accountability)
When to Use ORP-PostHoc
Use ORP-PostHoc when:
- Decision already made — You’re analyzing past choices, not making new ones
- Original documentation absent — Creators didn’t document using ORP
- Accountability gaps exist — You cannot know what original decision-makers considered
- Transparency value remains — Even partial documentation better than none
Example Use Cases
- AI training dataset analysis (ImageNet, LAION-5B, Common Crawl, GitHub Copilot)
- Historical policy decision reconstruction
- Academic research on past governance choices
- Transparency audits of existing systems
- Post-mortem analysis of controversial decisions
Creating an ORP-PostHoc Document
# ORP-PostHoc documents typically created manually
# (post_hoc flag must be set explicitly in YAML)Example structure:
orp_version: '0.2'
post_hoc: true
authors:
- name: Analyst Name
role: analyst # or reconstructor
affiliation: Organization
# ... rest of documentKey fields:
post_hoc: true— Marks as retrospective analysisrole: analystorrole: reconstructor— Not original decision-makeraccountability_gap— Documents what’s unknowable
Time investment: 1-3 days, depending on available source material and research depth.
Differences from ORP-Basic
| Aspect | ORP-Basic | ORP-PostHoc |
|---|---|---|
| Timing | Prospective (before decision) | Retrospective (after decision) |
| Author role | Decision-maker or steward | Analyst or reconstructor |
| Accountability | Real-time documentation | Accountability gap documentation |
| Purpose | Govern ongoing decisions | Understand past decisions |
| Extensions | Optional | Often needed (orp-ai-training-v1, etc.) |
Real Examples
See our AI training datasets analysis for four comprehensive ORP-PostHoc reconstructions:
- ImageNet (1.28M images)
- LAION-5B (5.85B pairs)
- Common Crawl (250TB)
- GitHub Copilot (100M+ repos)
Time investment: 1-3 days of research and documentation per dataset.
ORP-Standard (Layers 1-3)
What it requires:
- Everything from ORP-Basic
- Consequence simulation (L2): Model outcomes with scenarios
- Empathy mapping (L3): Identify stakeholders, test minority impacts
What it provides:
- Transparent reasoning about expected outcomes
- Explicit acknowledgment of who is affected and how
- Systematic comparison of alternative scenarios
- Minority stress testing
Best for:
- Policy proposals affecting significant populations
- Decisions with distributional impacts
- Resource allocation with trade-offs
- AI training dataset documentation
- Proposals that will face public scrutiny
Not sufficient for:
- Highly contested proposals with multiple competing versions
- Long-term decisions requiring accountability tracking
- Cases where systematic forking and comparison is needed
Example Use Cases
- Municipal budget allocation
- Healthcare policy affecting multiple stakeholder groups
- Education funding reform
- Infrastructure projects with environmental impacts
- AI model training dataset selection and documentation
Creating an ORP-Standard Document
orp new my-doc.yaml --compliance standardAdditional prompts will include:
- Affected population description
- Variables to model (independent and dependent)
- Scenarios to compare
- Stakeholder identification
- Minority groups to stress-test
Time investment: 1-2 days, with most time spent on modeling consequences and identifying stakeholders.
ORP-Full (All 5 Layers)
What it requires:
- Everything from ORP-Standard
- Accountability ledger (L4): Track all methodological decisions
- Fork registry (L5): Document lineage, invite alternatives
What it provides:
- Complete transparency and accountability
- Systematic comparison with alternative proposals
- Clear attribution of decisions to individuals
- Forkable proposal that others can modify and improve
- Full audit trail of reasoning evolution
Best for:
- High-stakes policy decisions (national/regional impact)
- Highly contested proposals with multiple alternatives
- Long-term decisions affecting future generations
- Cases requiring systematic comparison of competing approaches
- Decisions that will be revisited and refined over time
Required for:
- Government policy proposals (recommended)
- Major infrastructure investments
- Climate policy with intergenerational impacts
- Contested resource allocation decisions
Example Use Cases
- National tax reform
- Climate policy with 20+ year horizons
- Healthcare system restructuring
- Major infrastructure projects (>$100M)
- AI governance frameworks
- Constitutional amendments or major legal reforms
Creating an ORP-Full Document
orp new my-doc.yaml --compliance fullAll prompts from Standard, plus:
- Decision tracking setup (who will maintain ledger)
- Forking invitation (license, contact)
- Known alternatives to document
Time investment: 3-5 days initially, with ongoing maintenance as decisions evolve.
Choosing the Right Level
Ask yourself:
Question 1: What are the stakes?
- Low (internal, <100 people affected) → Basic is sufficient
- Medium (public, 100-10,000 affected) → Standard recommended
- High (policy, >10,000 affected or long-term) → Full recommended
Question 2: How contested is this?
- Consensus (general agreement) → Lower level may suffice
- Debated (significant disagreement) → Standard minimum
- Highly contested (multiple competing proposals) → Full enables systematic comparison
Question 3: How long will this matter?
- Short-term (<1 year impact) → Lower level may suffice
- Medium-term (1-5 years) → Standard recommended
- Long-term (>5 years or intergenerational) → Full recommended
Question 4: What resources do you have?
- Limited (hours, not days) → Basic to establish baseline
- Moderate (1-2 days available) → Standard
- Adequate (3+ days available) → Full
Upgrading Between Levels
Documents can be upgraded incrementally:
# Start with Basic
orp new policy.yaml --compliance basic
# ... fill in data provenance ...
# Later upgrade to Standard
# (edit the file, add L2 and L3, change compliance_level)
orp validate policy.yaml
# Later upgrade to Full
# (add L4 and L5, change compliance_level)
orp validate policy.yamlTip: Start with Basic to establish data transparency, then upgrade as the proposal matures.
Compliance Checking
Check your document’s current compliance level:
orp check policy.yamlOutput will show:
- Which layers are present
- Current compliance level
- Guidance on reaching next level
Common Questions
”Must I use ORP-Full for all government decisions?”
No. Use the level appropriate to the stakes and contestation. But as a principle: the more people affected and the longer the impact, the higher the compliance level should be.
”Can I mix compliance levels?”
Yes. A proposal might use ORP-Full for the main analysis but ORP-Basic for minor supporting datasets. Declare the highest level used in the compliance_level field.
”What if I don’t have resources for Full but the decision warrants it?”
Document what you can with Standard, explicitly note in the summary that Full would be better, and invite others to fork and extend. Incomplete transparency is better than no transparency.
”Can I validate without full compliance?”
Yes. The validator will show warnings if you could achieve a higher compliance level, but it won’t fail validation if you’re correctly claiming Basic when you only have L1.
Next Steps
- Get started - Install the validator
- Create your first document - Step-by-step ORP-Standard example
- See examples - Real ORP-Full document (Danish Property Tax)
- Read the specification - Full technical details