A practical, risk-tiered checklist framework for evaluating AI release readiness — with a lightweight CLI evaluation tool.
AI systems need release readiness checks that go beyond ordinary software quality gates: model behaviour, fallback paths, observability, and accountability all require explicit verification.
Three risk tiers — choose based on safety impact, regulatory exposure, and reversibility:
| Tier | Use when |
|---|---|
| Low risk | Internal tools, no safety impact, easily reversible |
| Medium risk | Customer-facing, some regulatory context, limited fallback |
| High risk | Safety-critical, regulated environment, hard to reverse |
Higher tiers include all requirements from lower tiers, plus additional items.
git clone https://github.com/simaba/release-checklist.git
cd release-checklist
pip install -r requirements.txt
# Evaluate against a medium-risk configuration
python src/check_release.py configs/medium-risk-example.yamlExample output:
AI Release Readiness Evaluation
================================
Risk tier: medium | Checking 24 items...
✓ Model evaluation completed on held-out test set
✓ Baseline performance documented
✓ Fallback behaviour defined and tested
✗ Bias and fairness assessment not completed
...
Result: NOT READY — 3 items require attention
checklists/
low-risk.md # Checklist for low-risk AI features
medium-risk.md # Checklist for medium-risk AI features
high-risk.md # Checklist for high-risk AI features
configs/
medium-risk-example.yaml # Example YAML configuration
high-risk-example.yaml # Example YAML configuration
src/
check_release.py # CLI evaluation tool
requirements.txt
- Fork the repository
- Edit the checklist
.mdfiles to match your organisation's requirements - Update the YAML configs to reflect your feature's risk profile
- Run
check_release.pyas part of your release pipeline
- AI Release Governance Framework — the broader framework this checklist operationalises
- Enterprise AI Governance Playbook — where this checklist fits in the full operating lifecycle
This repository is part of a connected toolkit for responsible AI operations:
| Repository | Purpose |
|---|---|
| Enterprise AI Governance Playbook | End-to-end AI operating model from intake to improvement |
| AI Release Governance Framework | Risk-based release gates for AI systems |
| AI Release Readiness Checklist | Risk-tiered pre-release checklists with CLI tool |
| AI Accountability Design Patterns | Patterns for human oversight and escalation |
| Multi-Agent Governance Framework | Roles, authority, and escalation for agent systems |
| Multi-Agent Orchestration Patterns | Sequential, parallel, and feedback-loop patterns |
| AI Agent Evaluation Framework | System-level evaluation across 5 dimensions |
| Agent System Simulator | Runnable multi-agent simulator with governance controls |
| LLM-powered Lean Six Sigma | AI copilot for structured process improvement |
Shared in a personal capacity. Open to collaborations and feedback — connect on LinkedIn or Medium.