7.0 KiB
7.0 KiB
Test Design and Risk Assessment - Validation Checklist
Prerequisites
- Story markdown with clear acceptance criteria exists
- PRD or epic documentation available
- Architecture documents available (optional)
- Requirements are testable and unambiguous
Process Steps
Step 1: Context Loading
- PRD.md read and requirements extracted
- Epics.md or specific epic documentation loaded
- Story markdown with acceptance criteria analyzed
- Architecture documents reviewed (if available)
- Existing test coverage analyzed
- Knowledge base fragments loaded (risk-governance, probability-impact, test-levels, test-priorities)
Step 2: Risk Assessment
- Genuine risks identified (not just features)
- Risks classified by category (TECH/SEC/PERF/DATA/BUS/OPS)
- Probability scored (1-3 for each risk)
- Impact scored (1-3 for each risk)
- Risk scores calculated (probability × impact)
- High-priority risks (score ≥6) flagged
- Mitigation plans defined for high-priority risks
- Owners assigned for each mitigation
- Timelines set for mitigations
- Residual risk documented
Step 3: Coverage Design
- Acceptance criteria broken into atomic scenarios
- Test levels selected (E2E/API/Component/Unit)
- No duplicate coverage across levels
- Priority levels assigned (P0/P1/P2/P3)
- P0 scenarios meet strict criteria (blocks core + high risk + no workaround)
- Data prerequisites identified
- Tooling requirements documented
- Execution order defined (smoke → P0 → P1 → P2/P3)
Step 4: Deliverables Generation
- Risk assessment matrix created
- Coverage matrix created
- Execution order documented
- Resource estimates calculated
- Quality gate criteria defined
- Output file written to correct location
- Output file uses template structure
Output Validation
Risk Assessment Matrix
- All risks have unique IDs (R-001, R-002, etc.)
- Each risk has category assigned
- Probability values are 1, 2, or 3
- Impact values are 1, 2, or 3
- Scores calculated correctly (P × I)
- High-priority risks (≥6) clearly marked
- Mitigation strategies specific and actionable
Coverage Matrix
- All requirements mapped to test levels
- Priorities assigned to all scenarios
- Risk linkage documented
- Test counts realistic
- Owners assigned where applicable
- No duplicate coverage (same behavior at multiple levels)
Execution Order
- Smoke tests defined (<5 min target)
- P0 tests listed (<10 min target)
- P1 tests listed (<30 min target)
- P2/P3 tests listed (<60 min target)
- Order optimizes for fast feedback
Resource Estimates
- P0 hours calculated (count × 2 hours)
- P1 hours calculated (count × 1 hour)
- P2 hours calculated (count × 0.5 hours)
- P3 hours calculated (count × 0.25 hours)
- Total hours summed
- Days estimate provided (hours / 8)
- Estimates include setup time
Quality Gate Criteria
- P0 pass rate threshold defined (should be 100%)
- P1 pass rate threshold defined (typically ≥95%)
- High-risk mitigation completion required
- Coverage targets specified (≥80% recommended)
Quality Checks
Evidence-Based Assessment
- Risk assessment based on documented evidence
- No speculation on business impact
- Assumptions clearly documented
- Clarifications requested where needed
- Historical data referenced where available
Risk Classification Accuracy
- TECH risks are architecture/integration issues
- SEC risks are security vulnerabilities
- PERF risks are performance/scalability concerns
- DATA risks are data integrity issues
- BUS risks are business/revenue impacts
- OPS risks are deployment/operational issues
Priority Assignment Accuracy
- P0: Truly blocks core functionality
- P0: High-risk (score ≥6)
- P0: No workaround exists
- P1: Important but not blocking
- P2/P3: Nice-to-have or edge cases
Test Level Selection
- E2E used only for critical paths
- API tests cover complex business logic
- Component tests for UI interactions
- Unit tests for edge cases and algorithms
- No redundant coverage
Integration Points
Knowledge Base Integration
- risk-governance.md consulted
- probability-impact.md applied
- test-levels-framework.md referenced
- test-priorities-matrix.md used
- Additional fragments loaded as needed
Status File Integration
- bmm-workflow-status.md exists
- Test design logged in Quality & Testing Progress
- Epic number and scope documented
- Completion timestamp recorded
Workflow Dependencies
- Can proceed to
atddworkflow with P0 scenarios - Can proceed to
automateworkflow with full coverage plan - Risk assessment informs
gateworkflow criteria - Integrates with
ciworkflow execution order
Completion Criteria
All must be true:
- All prerequisites met
- All process steps completed
- All output validations passed
- All quality checks passed
- All integration points verified
- Output file complete and well-formatted
- Team review scheduled (if required)
Post-Workflow Actions
User must complete:
- Review risk assessment with team
- Prioritize mitigation for high-priority risks (score ≥6)
- Allocate resources per estimates
- Run
atddworkflow to generate P0 tests - Set up test data factories and fixtures
- Schedule team review of test design document
Recommended next workflows:
- Run
atddworkflow for P0 test generation - Run
frameworkworkflow if not already done - Run
ciworkflow to configure pipeline stages
Rollback Procedure
If workflow fails:
- Delete output file
- Review error logs
- Fix missing context (PRD, architecture docs)
- Clarify ambiguous requirements
- Retry workflow
Notes
Common Issues
Issue: Too many P0 tests
- Solution: Apply strict P0 criteria - must block core AND high risk AND no workaround
Issue: Risk scores all high
- Solution: Differentiate between high-impact (3) and degraded (2) impacts
Issue: Duplicate coverage across levels
- Solution: Use test pyramid - E2E for critical paths only
Issue: Resource estimates too high
- Solution: Invest in fixtures/factories to reduce per-test setup time
Best Practices
- Base risk assessment on evidence, not assumptions
- High-priority risks (≥6) require immediate mitigation
- P0 tests should cover <10% of total scenarios
- Avoid testing same behavior at multiple levels
- Include smoke tests (P0 subset) for fast feedback
Checklist Complete: Sign off when all items validated.
Completed by: **___** Date: **___** Epic: **___** Notes: **************___**************