5.1: SUP.1 Quality Assurance
Process Definition
Purpose
SUP.1 Purpose: To provide independent and objective assurance that work products and processes comply with defined criteria and that non-conformances are resolved and further prevented.
Why QA Matters in AI-Assisted Development: When AI tools generate or modify work products, the traditional assumption that a human author has verified compliance no longer holds by default. Quality assurance must explicitly verify that AI-generated outputs meet the same standards as human-authored ones, and that the AI tools themselves are used within their qualified boundaries.
Outcomes
| Outcome | Description | AI Relevance |
|---|---|---|
| O1 | Quality assurance is performed independently and objectively without conflicts of interest | AI tools can augment independence by providing automated, bias-free checks, but must not replace the organizational independence requirement |
| O2 | Criteria for the quality of work products and process performance are defined | Criteria must explicitly address AI-generated content, including provenance tracking, confidence thresholds, and hallucination detection |
| O3 | Conformance of work products and process performance with the defined criteria and targets is verified, documented and communicated to the relevant parties | AI enables continuous verification rather than periodic audits, surfacing deviations earlier in the lifecycle |
| O4 | Non-conformances are tracked, resolved, and further prevented | AI pattern analysis across projects identifies systemic issues and predicts future non-conformances |
| O5 | Non-conformances are escalated to appropriate levels of management | Automated escalation workflows triggered by severity thresholds and aging timers |
| O6 | Management ensures that escalated non-conformances are resolved | Dashboard visibility and AI-generated resolution recommendations accelerate management response |
Base Practices with AI Integration
AI Integration Levels: L1 = AI-assisted (templates, suggestions); L2 = AI-augmented (automated checking, human confirms); L3 = AI-automated (autonomous execution with human oversight).
| BP | Base Practice | Description | AI Level | AI Application | Human Responsibility |
|---|---|---|---|---|---|
| BP1 | Ensure independence of quality assurance | QA function must be organizationally independent from development | L1 | Independence verification through organizational chart analysis and conflict-of-interest detection | Define QA organizational structure; approve independence arrangements |
| BP2 | Define criteria for quality assurance | Establish measurable criteria for work products and processes | L1 | Criteria templates derived from ASPICE base practices, project standards, and historical quality data | Review and approve criteria; tailor to project-specific needs |
| BP3 | Assure quality of work products | Verify work products against defined criteria | L2 | Automated completeness, consistency, and compliance checking; AI-driven quality scoring of documents, code, and test artifacts | Validate AI findings; resolve ambiguous cases; sign off on quality status |
| BP4 | Assure quality of process activities | Verify process execution against defined processes | L2 | Process adherence monitoring via CI/CD pipeline analysis, workflow log mining, and milestone tracking | Interpret deviations in context; approve process waivers; confirm audit findings |
| BP5 | Summarize and communicate QA activities and results | Report QA status to stakeholders | L2 | Automated report generation with trend analysis, executive summaries, and actionable recommendations | Review reports for accuracy; present findings to stakeholders; contextualize results |
| BP6 | Ensure resolution of non-conformances | Track corrective actions to closure | L2 | Status monitoring, aging alerts, resolution effectiveness tracking, and recurrence prediction | Approve corrective actions; verify closure evidence; accept residual risk |
| BP7 | Escalate non-conformances | Escalate unresolved issues to management | L1 | Automated escalation notifications based on severity, age, and resolution deadline thresholds | Define escalation criteria; make escalation decisions for ambiguous cases; ensure management engagement |
AI-Powered Quality Assurance
Principle: AI transforms QA from periodic, sample-based auditing to continuous, comprehensive monitoring. However, the QA engineer retains authority over all compliance decisions and finding classifications.
Automated Compliance Checking Against Standards
AI-powered compliance checking continuously verifies work products and processes against applicable standards, project plans, and regulatory requirements.
| Compliance Area | What AI Checks | Standards Referenced | Automation Approach |
|---|---|---|---|
| Requirements quality | Completeness, testability, ambiguity, uniqueness | ASPICE SWE.1 BP3-BP6; ISO 26262-8 Clause 6 | NLP analysis of requirement text against quality rules; cross-reference with verification criteria |
| Design documentation | Architectural completeness, interface consistency, traceability | ASPICE SWE.2 BP1-BP5 | Structural analysis of design documents; automated traceability matrix validation |
| Code compliance | Coding standard adherence, MISRA violations, complexity metrics | MISRA C:2012; CERT C; project coding standards | Static analysis integration (SonarQube, Axivion); rule-based compliance scoring |
| Test coverage | Structural coverage, requirement coverage, boundary analysis | ASPICE SWE.4 BP4; ISO 26262-6 Table 12 | Coverage tool integration; gap identification and test case suggestions |
| Process compliance | Milestone gate criteria, review records, approval evidence | ASPICE SUP.1 BP3-BP4; project QA plan | Workflow log analysis; artifact existence and completeness verification |
| Traceability | Bidirectional links between requirements, design, code, tests | ASPICE SUP.1 BP3; ISO 26262-8 Clause 6.4.2 | Automated gap detection in traceability matrices; orphan and suspect link identification |
# Automated compliance check configuration (illustrative example)
compliance_check:
name: "ASPICE SWE.1 Requirements Compliance"
schedule: "on_commit" # Trigger on every commit to requirements repository
checks:
- id: CHK-REQ-001
name: "Requirement completeness"
rule: "Every requirement must have: ID, description, priority, verification criterion"
severity: HIGH
tool: "ai_nlp_analyzer"
- id: CHK-REQ-002
name: "Requirement testability"
rule: "Requirements must contain measurable acceptance criteria"
severity: MEDIUM
tool: "ai_nlp_analyzer"
- id: CHK-REQ-003
name: "Traceability completeness"
rule: "Every SW requirement must trace to at least one system requirement"
severity: HIGH
tool: "polarion_api"
- id: CHK-REQ-004
name: "No ambiguous language"
rule: "Flag use of 'should', 'may', 'might', 'could', 'appropriate', 'as needed'"
severity: LOW
tool: "ai_nlp_analyzer"
reporting:
format: "json"
destination: "qa_dashboard"
notify_on_failure: ["qa_lead@project.com"]
Process Adherence Monitoring
Continuous vs. Periodic QA: Traditional QA relies on scheduled audits that sample a subset of activities. AI-powered process monitoring checks every workflow execution, every artifact state transition, and every approval event in real time.
| Monitoring Aspect | Data Sources | AI Technique | Alert Conditions |
|---|---|---|---|
| Workflow compliance | CI/CD pipeline logs, ALM tool events | Sequence pattern matching against defined process models | Steps skipped, out-of-order execution, missing approvals |
| Review execution | Code review tool logs, meeting records | Review completeness scoring; participant coverage analysis | Reviews without documented findings, missing reviewer roles, insufficient review time |
| Milestone gate criteria | Project management tool, artifact repository | Gate readiness scoring against checklist criteria | Gate attempted with incomplete work products, missing sign-offs |
| Tool usage compliance | Tool access logs, license servers | Usage pattern analysis against qualified tool list | Unqualified tool versions, unauthorized tool substitutions |
| Configuration management | Git logs, baseline records | Branch policy compliance, merge request workflow validation | Direct commits to protected branches, baselines without required approvals |
Work Product Quality Scoring
AI assigns quality scores to work products based on multiple dimensions, providing QA engineers with a prioritized view of where human review effort is most needed.
| Quality Dimension | Scoring Method | Score Range | Threshold |
|---|---|---|---|
| Completeness | Mandatory field/section coverage analysis | 0-100% | >= 95% for release |
| Consistency | Cross-reference validation between related work products | 0-100% | >= 90% for release |
| Correctness | Rule-based checks against standards and templates | 0-100% | >= 90% for release |
| Clarity | NLP readability analysis (Flesch-Kincaid adapted for technical docs) | 0-100% | >= 80% for release |
| Traceability | Link coverage and bidirectionality verification | 0-100% | 100% for safety-relevant |
| Composite Score | Weighted average (configurable per project) | 0-100% | >= 90% for release |
AI Work Product Quality Report:
Document: SRS_BCM_v1.2.docx
Analysis Date: 2025-01-15
Dimension Score Status Details
─────────────────────────────────────────────────────────────
Completeness 87% WARNING 3 requirements missing verification criteria
Consistency 94% PASS 2 minor terminology inconsistencies
Correctness 91% PASS All structural rules satisfied
Clarity 83% PASS 2 requirements flagged for ambiguity
Traceability 90% WARNING 5 of 48 requirements missing upward traces
─────────────────────────────────────────────────────────────
Composite Score 89% WARNING Below 90% release threshold
Priority Review Areas:
1. [HIGH] Complete traceability for SWE-BCM-110..114
2. [MEDIUM] Add verification criteria to SWE-BCM-103, 107, 119
3. [LOW] Resolve terminology: "actuator" vs "motor driver" usage
Human Review Required:
- Address all HIGH priority items before release gate
- Confirm LOW items are acceptable terminology variations
Deviation Detection and Prediction
Predictive QA: Beyond detecting existing deviations, AI analyzes historical patterns to predict where quality problems are likely to emerge, enabling preventive action rather than corrective action.
| Detection Type | Method | Data Inputs | Lead Time |
|---|---|---|---|
| Real-time deviation | Rule-based comparison against baselines | Current artifact state vs. defined criteria | Immediate |
| Trend-based prediction | Statistical trend analysis on quality metrics | Historical quality scores, defect density, review finding rates | 1-2 sprints |
| Pattern-based prediction | ML classification trained on historical non-conformances | Past NCRs, root causes, project characteristics | Project phase |
| Cross-project learning | Similarity matching across project portfolio | Multi-project quality databases, common failure modes | Project kickoff |
# Deviation prediction configuration (illustrative example)
deviation_prediction:
model: "qa_deviation_predictor_v2"
training_data: "historical_ncr_database"
risk_indicators:
- name: "Requirements volatility"
threshold: "> 15% change rate after SWE.1 baseline"
prediction: "HIGH risk of downstream rework and traceability gaps"
- name: "Review finding density drop"
threshold: "< 0.5 findings per review session"
prediction: "MEDIUM risk of superficial reviews"
- name: "Test case growth stagnation"
threshold: "Test cases not growing proportionally to requirements"
prediction: "HIGH risk of insufficient test coverage"
- name: "Late integration defects"
threshold: "> 30% of defects found in integration testing"
prediction: "MEDIUM risk of inadequate unit testing"
actions:
- trigger: "HIGH risk prediction"
action: "Schedule focused QA audit within 5 business days"
notify: ["qa_lead", "project_manager"]
- trigger: "MEDIUM risk prediction"
action: "Add to next scheduled QA review agenda"
notify: ["qa_lead"]
AI-Assisted QA Framework
The diagram below presents the AI-assisted QA framework, showing how automated compliance checks, evidence collection, and trend analysis feed into the QA audit workflow while maintaining human decision authority.
QA Audit Automation
Note: AI-assisted audits do not replace human auditor judgment. AI handles evidence gathering, checklist pre-population, and compliance pre-screening. The auditor interprets findings, assesses context, and makes the final compliance determination.
Audit Preparation
AI accelerates audit preparation by automatically gathering evidence, pre-populating checklists, and identifying areas of concern before the auditor begins.
| Preparation Step | Manual Effort | AI-Assisted Effort | AI Contribution |
|---|---|---|---|
| Evidence gathering | 4-8 hours per process | 30-60 minutes | Automated retrieval from ALM, Git, review tools; completeness verification |
| Checklist population | 2-4 hours per process | 15-30 minutes | Pre-fill checklist items with evidence references and preliminary pass/fail status |
| Scope identification | 1-2 hours | 15 minutes | Risk-based audit scope recommendation using historical findings and change activity |
| Schedule coordination | 1-2 hours | 10 minutes | Calendar analysis, stakeholder availability matching, resource optimization |
| Previous findings review | 1-2 hours | 10 minutes | Historical finding summary, recurrence pattern analysis, closure status verification |
| Total per process | 9-18 hours | 1.5-2.5 hours | ~85% effort reduction in preparation |
Audit Execution Support
# AI-assisted audit execution workflow (illustrative example)
audit_execution:
phase: "preparation"
steps:
- step: "Auto-gather evidence"
ai_actions:
- "Query Polarion for all SWE.1 work products"
- "Retrieve review records from GitLab merge requests"
- "Extract test results from CI/CD pipeline artifacts"
- "Collect approval records from workflow engine"
output: "evidence_package.zip"
- step: "Pre-screen compliance"
ai_actions:
- "Check each checklist item against gathered evidence"
- "Mark items as PASS (evidence found and valid), FAIL (evidence missing or invalid), or REVIEW (ambiguous)"
- "Calculate preliminary compliance score"
output: "prescreened_checklist.yaml"
- step: "Risk-based focus areas"
ai_actions:
- "Identify checklist items marked REVIEW or FAIL"
- "Cross-reference with historical audit findings"
- "Rank focus areas by risk and impact"
output: "audit_focus_report.md"
phase: "execution"
steps:
- step: "Auditor reviews pre-screened results"
human_actions:
- "Validate AI PASS determinations (sample-based)"
- "Investigate all FAIL and REVIEW items"
- "Interview process participants as needed"
- "Document findings with objective evidence"
- step: "Finding classification"
ai_actions:
- "Suggest finding severity based on ASPICE rating impact"
- "Link findings to specific base practice gaps"
- "Propose corrective action categories"
human_actions:
- "Confirm or override severity classification"
- "Finalize finding descriptions"
- "Agree corrective actions with process owner"
Process Audit Example
# QA Process Audit Checklist (illustrative example)
audit:
id: QA-AUDIT-001
process: SWE.1
date: (audit date)
auditor: AI-Assisted + QA Lead
checklist:
work_products:
- id: WP-17-08
name: "SW Requirements Specification"
check: "Document exists and is under version control"
status: PASS
evidence: "SRS_BCM_v1.2.docx in Polarion"
- id: WP-17-11
name: "Traceability Record"
check: "All SW requirements traced to system requirements"
status: PARTIAL
finding: "5 of 48 requirements missing traces"
severity: MEDIUM
- id: WP-17-12
name: "Verification Criteria"
check: "All requirements have verification criteria"
status: PASS
evidence: "Verification matrix complete"
reviews:
- type: "Requirements Review"
check: "Review conducted with documented results"
status: PASS
evidence: "Review record RR-SWE1-001"
- type: "Stakeholder Agreement"
check: "Requirements approved by stakeholders"
status: FAIL
finding: "No formal approval record found"
severity: HIGH
tools:
- tool: "Polarion ALM"
check: "Tool configured per project standards"
status: PASS
findings_summary:
high: 1
medium: 1
low: 0
overall_status: "CONDITIONAL PASS - Address findings"
AI Quality Analysis
L2: Automated Quality Checking
AI Quality Analysis Report:
Document: SRS_BCM_v1.2.docx
Analysis Date: 2025-01-15
The diagram below visualizes the AI completeness analysis results, showing coverage percentages across requirement categories and flagging areas with insufficient specification.
Human Review Required:
- Address ambiguity in flagged requirements
- Complete missing interface requirements
- Add rationale to requirements missing justification
- Fill traceability gaps
Non-Conformance Tracking
Finding Management Workflow
# Non-conformance finding (illustrative example)
finding:
id: NCR-(year)-(number)
source: QA-AUDIT-001
date: (finding date)
severity: HIGH
category: process
description: |
No formal stakeholder approval record exists for software
requirements specification SRS_BCM_v1.2.
evidence: |
- Review record RR-SWE1-001 shows technical review completed
- No customer/stakeholder sign-off document found
- Requirements marked as "Draft" in Polarion
root_cause_analysis:
ai_suggestion: |
Similar pattern in 3 previous projects suggests process gap:
Review process does not include explicit stakeholder approval step.
confidence: medium
human_validation: pending
corrective_action:
immediate: "Obtain stakeholder approval for SRS_BCM_v1.2"
systemic: "Add stakeholder approval gate to requirements workflow"
owner: Project Manager
due_date: 2025-01-22
status: open
tracking:
opened: 2025-01-15
target_close: 2025-01-22
escalation_date: 2025-01-25
Quality Metrics Dashboard
Note: Metrics should be tailored to project size and ASPICE target capability level. The following represents a comprehensive set; select metrics appropriate to your context.
Core QA Metrics
| Metric | Formula | Target | Measurement Frequency | AI Contribution |
|---|---|---|---|---|
| Audit compliance rate | Passed items / Total items per audit | >= 90% | Per audit | Automated calculation from checklist data |
| NCR closure rate | Closed NCRs / Total NCRs (within target date) | >= 85% | Weekly | Aging analysis and closure trend prediction |
| NCR recurrence rate | Recurring NCRs / Total NCRs | < 10% | Monthly | Pattern matching to identify recurring root causes |
| Process adherence score | Compliant process executions / Total process executions | >= 95% | Continuous | Workflow log mining and compliance scoring |
| Work product quality score | Composite quality score (see scoring section) | >= 90% | Per work product update | Automated multi-dimensional scoring |
| Audit coverage | Processes audited / Total processes in scope | 100% per cycle | Quarterly | Audit schedule optimization and gap tracking |
| Finding density | Findings per audit / Work products audited | Trending downward | Per audit | Trend analysis with seasonal adjustment |
| Mean time to resolution (MTTR) | Average days from NCR opening to verified closure | < 15 business days | Monthly | Resolution time prediction and bottleneck identification |
QA Key Performance Indicators (KPIs)
| KPI | Description | Green | Yellow | Red |
|---|---|---|---|---|
| QA Plan Execution | Percentage of planned QA activities completed on schedule | >= 90% | 75-89% | < 75% |
| Audit Finding Trend | Direction of finding density over last 3 audits | Decreasing | Stable | Increasing |
| NCR Backlog Health | Open NCRs within resolution target vs. overdue | < 10% overdue | 10-25% overdue | > 25% overdue |
| Escalation Rate | Percentage of NCRs requiring management escalation | < 5% | 5-15% | > 15% |
| Preventive Action Effectiveness | Reduction in recurrence after systemic corrective action | > 80% reduction | 50-80% reduction | < 50% reduction |
| AI Finding Accuracy | Percentage of AI-flagged issues confirmed by human auditor | > 85% | 70-85% | < 70% |
AI-Driven Insights
AI augments the metrics dashboard by providing contextual analysis that goes beyond raw numbers.
| Insight Type | Description | Example |
|---|---|---|
| Trend explanation | AI correlates metric changes with project events | "NCR rate increased 40% coinciding with new team member onboarding and 3 concurrent change requests" |
| Cross-project benchmarking | Compare current project metrics against portfolio baselines | "Audit compliance rate is 8% below portfolio average for projects at this lifecycle phase" |
| Predictive alerts | Forecast metric threshold breaches before they occur | "At current trajectory, NCR backlog will exceed RED threshold within 2 sprints" |
| Root cause clustering | Group findings by common root causes for systemic action | "67% of traceability findings across 4 audits trace to the same process gap in requirements import workflow" |
| Resource optimization | Recommend audit schedule adjustments based on risk | "SWE.3 audit can be deferred 2 weeks; SWE.4 should be advanced due to high change activity" |
The dashboard below consolidates the AI-driven QA insights described above into a single monitoring view, tracking audit compliance, finding trends, and predictive alerts.
HITL Protocol for QA Decisions
Fundamental Principle: AI augments QA efficiency but must never replace human judgment for compliance determinations. ASPICE requires that quality assurance is performed by competent personnel; AI is a tool, not a qualified auditor.
Decision Authority Matrix
| QA Activity | AI Role | Human Role | Authority |
|---|---|---|---|
| Evidence gathering | Collect and organize artifacts from tools | Verify completeness of evidence package | Human confirms |
| Compliance pre-screening | Apply rules to flag pass/fail/review items | Validate all AI determinations, especially edge cases | Human decides |
| Finding classification | Suggest severity and category based on criteria | Confirm or override classification with contextual judgment | Human decides |
| Corrective action proposal | Recommend actions based on historical effectiveness | Define, approve, and assign corrective actions | Human decides |
| NCR escalation | Trigger escalation alerts based on thresholds | Make final escalation decision considering context | Human decides |
| Audit report generation | Draft report with findings, metrics, and trends | Review, edit, and approve report before distribution | Human approves |
| Process waiver | Flag waiver request against defined criteria | Evaluate justification and approve or deny waiver | Human decides |
| Quality gate decision | Calculate readiness score against gate criteria | Make pass/conditional pass/fail determination | Human decides |
Mandatory Human Checkpoints
- Audit finding confirmation: Every finding flagged by AI must be reviewed and confirmed or dismissed by a qualified QA auditor before it is formally recorded
- Severity classification: AI severity suggestions are advisory; the auditor determines the final severity based on project context, safety impact, and customer requirements
- Corrective action approval: All corrective actions, whether AI-suggested or human-defined, require approval by the responsible process owner and QA lead
- Audit report sign-off: The audit report is a controlled work product; only the designated QA auditor may sign off on the final version
- Quality gate decisions: Gate pass/fail determinations remain exclusively human decisions; AI provides the data, humans make the call
- Escalation authorization: While AI triggers escalation alerts, the QA lead determines whether formal escalation to management is warranted
Override and Feedback Protocol
# HITL override tracking (illustrative example)
hitl_override:
id: HITL-QA-001
date: (override date)
ai_recommendation: "FAIL - Missing stakeholder approval record"
human_decision: "CONDITIONAL PASS"
justification: |
Verbal approval obtained in steering committee meeting SC-2025-003.
Formal sign-off scheduled for 2025-01-20. Risk accepted for 5 business
days to avoid blocking downstream activities.
approved_by: "QA Lead"
conditions:
- "Formal approval must be obtained by 2025-01-20"
- "If not obtained, finding reverts to FAIL status"
feedback_to_ai:
category: "verbal_approval_acceptance"
learning: "Verbal approval with documented meeting minutes may be
accepted as interim evidence with time-bounded conditions"
Tool Integration
Note: Tool selection depends on project infrastructure, budget, and existing toolchain. The following integrations are representative; adapt to your specific environment.
SonarQube Integration
SonarQube provides continuous code quality inspection that feeds directly into the QA compliance framework.
| Integration Point | QA Application | Configuration |
|---|---|---|
| Quality Gate status | Automated pass/fail for code quality criteria | Define project-specific quality gate with MISRA compliance, coverage thresholds, and duplication limits |
| Issue tracking | Feed code quality findings into NCR workflow | Map SonarQube severity (BLOCKER, CRITICAL) to QA severity (HIGH, MEDIUM) |
| Technical debt | Track and report technical debt as quality metric | Set technical debt ratio thresholds per project phase |
| Coverage reporting | Verify structural coverage against ASPICE/ISO 26262 targets | Configure coverage thresholds per ASIL level (e.g., ASIL-B: MC/DC not required but statement coverage >= 80%) |
| Rule compliance | MISRA C:2012 rule adherence tracking | Enable MISRA C plugin; configure mandatory vs. advisory rule sets |
Axivion Suite Integration
Axivion provides architecture verification and code quality analysis particularly suited to embedded and safety-critical software.
| Integration Point | QA Application | Configuration |
|---|---|---|
| Architecture compliance | Verify implementation matches architectural design | Define architecture model; check for dependency violations on every build |
| Clone detection | Identify code duplication that may indicate design issues | Set clone thresholds; feed violations into QA finding workflow |
| Metric trending | Track complexity, coupling, and cohesion metrics over time | Define metric baselines at project start; alert on degradation |
| MISRA checking | Comprehensive MISRA C:2012 compliance analysis | Configure against project-specific deviation permits |
| Delta analysis | Focus review on changed code areas | Integrate with CI/CD to analyze only modified files per commit |
Custom AI QA Agents
Purpose-built AI agents extend QA automation beyond what commercial tools provide.
| Agent | Purpose | Inputs | Outputs | Integration |
|---|---|---|---|---|
| Traceability Auditor | Verify bidirectional traceability completeness | ALM export (requirements, design, test cases) | Gap report with missing links and orphaned items | Polarion/codebeamer API; scheduled daily |
| Review Quality Checker | Assess whether reviews meet minimum quality criteria | Review records, comments, time logs | Review quality score; findings for superficial reviews | GitLab/GitHub API; triggered per merge request |
| Process Compliance Monitor | Check workflow execution against defined process | CI/CD logs, ALM workflow events, approval records | Compliance score per process; deviation alerts | Event-driven; monitors tool webhooks |
| Document Quality Scorer | Score work products across quality dimensions | Document content, templates, standards | Multi-dimensional quality score with improvement suggestions | Triggered on document update in repository |
| NCR Pattern Analyzer | Identify recurring non-conformance patterns | Historical NCR database | Pattern report with systemic action recommendations | Scheduled monthly; on-demand for audit preparation |
# Custom AI QA agent configuration (illustrative example)
qa_agents:
traceability_auditor:
schedule: "daily at 02:00 UTC"
data_sources:
- type: "polarion"
query: "project:BCM AND type:requirement"
- type: "polarion"
query: "project:BCM AND type:test_case"
checks:
- "Every requirement has at least one test case linked"
- "Every test case traces to at least one requirement"
- "No orphaned requirements (no parent system requirement)"
- "No suspect links (requirement changed after test case baselined)"
output:
format: "json"
destination: "qa_dashboard"
alert_threshold: "any gap found"
review_quality_checker:
trigger: "merge_request_merged"
checks:
- name: "Minimum review time"
rule: "Review duration >= 30 minutes for changes > 100 lines"
- name: "Reviewer coverage"
rule: "At least 2 reviewers including domain expert"
- name: "Finding resolution"
rule: "All review findings marked resolved or deferred with justification"
output:
format: "markdown"
destination: "merge_request_comment"
Work Products
| WP ID | Work Product | Description | AI Role | AI Level | Human Responsibility |
|---|---|---|---|---|---|
| 15-01 | QA plan | Defines QA strategy, scope, schedule, criteria, and resources | Template generation from ASPICE base practices and project context | L1 | Tailor to project; approve scope and schedule |
| 15-02 | QA report | Documents audit findings, compliance status, and recommendations | Automated finding aggregation, trend analysis, and report drafting | L2 | Validate findings; approve report content |
| 08-28 | Non-conformance record | Documents individual non-conformance with evidence and corrective actions | Classification, severity suggestion, root cause pattern matching | L2 | Confirm classification; approve corrective actions |
| 13-25 | Corrective action log | Tracks corrective and preventive actions to closure | Status monitoring, aging alerts, effectiveness tracking | L2 | Approve actions; verify closure evidence |
| 15-03 | Audit checklist | Detailed checklist for process and work product audits | Pre-population from ASPICE base practices; evidence auto-gathering | L2 | Customize per audit scope; validate completions |
| 13-50 | QA metrics report | Periodic report on QA KPIs, trends, and predictions | Automated metric calculation, trend analysis, predictive insights | L2 | Interpret metrics in context; define actions |
| 08-50 | Process improvement proposal | Systemic improvement proposals based on QA findings | Pattern analysis across NCRs; improvement suggestions | L1 | Evaluate feasibility; prioritize and approve |
Common Quality Issues in AI-Assisted Development
Critical Awareness: AI-assisted development introduces new categories of quality issues that traditional QA processes may not detect. QA criteria and checklists must be explicitly updated to address these risks.
| Quality Issue | Description | Detection Method | Mitigation |
|---|---|---|---|
| AI hallucination in requirements | AI generates plausible but incorrect requirements or rationale | Cross-reference AI outputs against source documents; domain expert review | Mandatory human review of all AI-generated requirements; provenance tracking |
| Unverified code suggestions | AI-generated code accepted without adequate review or testing | Code review quality checking; coverage analysis for AI-generated code | Flag AI-generated code for enhanced review; require test evidence |
| Inconsistent terminology | AI uses different terms than project glossary or standards | NLP-based terminology consistency checking | Maintain and enforce project glossary; configure AI tools with glossary |
| Traceability gaps from AI changes | AI modifications break existing traceability links | Automated traceability audit after every AI-assisted change | Run traceability validation in CI/CD pipeline; block merge on gaps |
| Over-reliance on AI findings | QA team treats AI results as authoritative without validation | HITL compliance monitoring; audit of override rates | Enforce mandatory human confirmation; track and review override frequency |
| Tool qualification gaps | AI tools used beyond their qualified scope or version | Tool usage monitoring; qualification record maintenance | Maintain tool qualification matrix; alert on unqualified tool usage |
| Bias in AI quality scoring | AI scoring model trained on biased historical data | Periodic scoring model validation against expert assessments | Calibrate AI models quarterly; compare AI scores with manual audit results |
| Loss of process knowledge | Team relies on AI for process compliance without understanding rationale | Process competency assessments; training records review | Maintain training program; ensure QA team understands underlying standards |
AI-Specific QA Checklist Extension
# Additional QA checklist items for AI-assisted development
ai_qa_checks:
- id: AI-QA-001
question: "Are AI-generated work products clearly identified and labeled?"
evidence: "Provenance metadata in document properties or commit messages"
- id: AI-QA-002
question: "Has every AI-generated requirement been reviewed by a domain expert?"
evidence: "Review record with explicit AI-content review confirmation"
- id: AI-QA-003
question: "Are AI tools used within their qualified scope and version?"
evidence: "Tool qualification record; tool version log"
- id: AI-QA-004
question: "Is AI confidence level documented for AI-generated analyses?"
evidence: "Confidence scores in analysis reports; thresholds defined in QA plan"
- id: AI-QA-005
question: "Are HITL override decisions documented with justification?"
evidence: "HITL override log with rationale and approver"
- id: AI-QA-006
question: "Is the AI model/tool version recorded for reproducibility?"
evidence: "Tool version in CI/CD logs; model version in analysis reports"
- id: AI-QA-007
question: "Are AI-generated test cases validated for relevance and coverage?"
evidence: "Test case review record; coverage analysis report"
Implementation Checklist
Usage: This checklist supports incremental adoption of AI-powered QA. Organizations should target Level 1 items first, then progress to Level 2 and Level 3 as capability matures.
Level 1: Foundation (AI-Assisted)
| Item | Action | Responsible | Status |
|---|---|---|---|
| 1.1 | Define QA plan with explicit AI integration scope and boundaries | QA Lead | [ ] |
| 1.2 | Establish QA criteria that address AI-generated work products | QA Lead + Process Owner | [ ] |
| 1.3 | Create audit checklists with AI-specific items (AI-QA-001 through AI-QA-007) | QA Lead | [ ] |
| 1.4 | Define HITL protocol with mandatory human checkpoints | QA Lead + Project Manager | [ ] |
| 1.5 | Set up NCR tracking system with AI classification fields | QA Lead + Tool Admin | [ ] |
| 1.6 | Train QA team on AI tool capabilities and limitations | QA Lead + Training | [ ] |
| 1.7 | Establish tool qualification records for all AI tools used in QA | QA Lead + Safety Manager | [ ] |
| 1.8 | Define quality metrics and KPI thresholds | QA Lead + Project Manager | [ ] |
Level 2: Integration (AI-Augmented)
| Item | Action | Responsible | Status |
|---|---|---|---|
| 2.1 | Integrate SonarQube (or equivalent) quality gate into CI/CD pipeline | DevOps + QA Lead | [ ] |
| 2.2 | Deploy automated traceability auditor agent | DevOps + QA Lead | [ ] |
| 2.3 | Implement automated work product quality scoring | QA Lead + AI Engineer | [ ] |
| 2.4 | Configure process adherence monitoring on CI/CD workflows | DevOps + QA Lead | [ ] |
| 2.5 | Set up QA metrics dashboard with automated data collection | QA Lead + Tool Admin | [ ] |
| 2.6 | Implement automated audit evidence gathering from ALM and Git | QA Lead + Tool Admin | [ ] |
| 2.7 | Deploy review quality checker on merge request workflows | DevOps + QA Lead | [ ] |
| 2.8 | Configure NCR pattern analyzer for monthly systemic analysis | QA Lead + AI Engineer | [ ] |
| 2.9 | Establish AI finding accuracy tracking (target: > 85%) | QA Lead | [ ] |
| 2.10 | Integrate Axivion (or equivalent) for architecture compliance checking | DevOps + Architect | [ ] |
Level 3: Optimization (AI-Automated with Oversight)
| Item | Action | Responsible | Status |
|---|---|---|---|
| 3.1 | Deploy predictive deviation detection using historical NCR data | QA Lead + AI Engineer | [ ] |
| 3.2 | Implement cross-project benchmarking in QA dashboard | QA Lead + PMO | [ ] |
| 3.3 | Enable AI-driven audit scope optimization based on risk analysis | QA Lead | [ ] |
| 3.4 | Deploy continuous process compliance monitoring (real-time) | DevOps + QA Lead | [ ] |
| 3.5 | Implement AI-generated improvement proposals from NCR clustering | QA Lead + AI Engineer | [ ] |
| 3.6 | Establish feedback loop from HITL overrides to AI model refinement | QA Lead + AI Engineer | [ ] |
| 3.7 | Validate AI QA accuracy quarterly with independent manual audits | QA Lead + External Auditor | [ ] |
Summary
SUP.1 Quality Assurance:
- AI Level: L1-L2 (AI detects, human confirms)
- Primary AI Value: Automated compliance checking, continuous process monitoring, predictive deviation detection
- Human Essential: Finding validation, corrective action approval, quality gate decisions, audit sign-off
- Key Outputs: QA reports, non-conformance records, corrective action logs, quality metrics
- Focus: Process adherence, work product quality, AI-specific quality risks
- Tool Ecosystem: SonarQube, Axivion, custom AI QA agents, ALM integration
- HITL Principle: AI augments QA efficiency; humans retain all compliance decision authority