Colorado Artificial Intelligence Act (CAIA) Compliance
Requirements, timelines, and a practical readiness checklist for developers and deployers of high-risk AI used in consequential decisions across hiring, lending, housing, healthcare, insurance, education, and essential services.
Who Must Comply
The Colorado Artificial Intelligence Act (CAIA) covers high-risk AI used in consequential decisions. Obligations apply to developers that build or modify such systems and deployers that use them in Colorado.
Developers
Create, license, or substantially modify high-risk AI used in consequential decisions (e.g., hiring, lending, housing, healthcare, insurance, education, essential services).
Deployers
Organizations in Colorado using high-risk AI to make or materially influence consequential decisions affecting consumers.

What Counts as “High-Risk” AI
High-risk AI under the Colorado Artificial Intelligence Act (CAIA) is defined by use case, not by model type. A system is in scope when its output makes or materially influences a consequential decision about a consumer.
Education access and outcomes
Employment screening, hiring, promotion, termination
Credit/financing eligibility and terms
Healthcare eligibility, triage, or benefits
Housing approval and tenant screening
Insurance underwriting, pricing, or claims handling
Essential government services and benefits
Legal services access or outcomes


Core Obligations for Developers (CAIA)
Reasonable Care & Bias Controls
Technical Documentation Package
Use Boundaries
Impact-Assessment Enablement
Public Transparency Statement
Incident Notification (90-Day)
Change Management & Versioning
Data Governance
Security & Access Controls
Operator & Monitoring Guidance
Core Obligations for Deployers (CAIA)
Scope and Inventory
AI Risk Management Program
Impact Assessments
Human Oversight
Data Governance
Transparency Statement
Recordkeeping and Auditability
Vendor & Contract Controls
Small-Deployer Relief Check
Governance Cadence

How We Work
A practical checklist to structure compliance work under the Colorado Artificial Intelligence Act (CAIA) for high-risk AI used in consequential decisions.
Step 1: Confirm Scope
Map AI use cases to consequential decisions; flag systems that make or materially influence eligibility, pricing, benefits, or rights.
Step 2: Assign Ownership
Name accountable leads across product, legal, data, and compliance, define decision rights
Step 3: Risk Management
Establish an AI RMP aligned with NIST AI RMF / ISO/IEC 42001, scaled to system impact.
Step 4: Technical Documentation
Intended use, data sources, limitations, performance, monitoring guidance, change history.
Step 5: Impact Assessments
Complete before deployment, on schedule (e.g., annually), and after material changes, record outcomes.
Step 6: Consumer Disclosures
Pre-use notice of AI involvement, adverse-decision explanations, and a human-review appeal path, ensure accessibility and language coverage.
Step 7: Data Governance & Testing
Provenance tracking, representativeness checks, bias/fairness testing, robustness, drift monitoring with thresholds.
Step 8: Periodic Reviews
Calendarized reviews and internal audits to verify controls, metrics, and disclosure accuracy.
CAIA Readiness and Implementation Support
Accelerate alignment with the Colorado Artificial Intelligence Act (CAIA) through a structured program for high-risk AI used in consequential decisions.
FAQs
Explore answers to pressing questions about CAIA risk tiers, duties, and documentation.
Yes, if products or services result in consequential decisions about Colorado consumers, or if high-risk AI is deployed in Colorado.
Certain deployer duties may not apply to organizations with fewer than 50 FTEs that use systems as intended and do not train them on their own data, provided equivalent impact assessments from the developer are available.
Pre-use notice of AI involvement, explanation of any adverse decision, and a clear path to human review/appeal. Disclosures should be accessible and available in appropriate languages.
No. Enforcement authority rests with the Colorado Attorney General.
Both regulate risk in AI, but CAIA centers on consequential decisions about consumers in Colorado, while the EU AI Act uses tiered risk categories and broader market-conformity duties.
Yes. Coverage depends on how the system is used in decisions, not on hosting model.
Outcomes that unlawfully differentiate or unfairly disadvantage protected classes or consumers through the AI system’s operation or data, considering intended and known uses.