AI Controls, Monitoring & Incidents
Effective AI governance requires more than policies — it demands operationalized controls, continuous performance monitoring, and structured incident response capabilities.
Three Lines of Defense — Applied to AI
AI Control Types
| Assessment Type | Question Asked | Testing Approach | AI Example |
|---|---|---|---|
| Design Effectiveness | Is the control designed well enough to address the risk? | Walkthrough, inspection of control documentation, inquiry | Review the model validation policy — does it require bias testing for demographic parity? |
| Operating Effectiveness | Is the control actually operating as designed, consistently? | Re-performance, observation, sampling of evidence | Pull 25 model deployment records — was bias testing completed and documented for each? |
| Design gap | — | The control doesn't address the right risk | Policy requires accuracy testing but not fairness testing — bias risk is unaddressed |
| Operating gap | — | The control is well-designed but not followed | Policy requires bias testing but 12 of 25 deployments have no testing documentation |
AI Control Design & Key Control Areas
The critical control domains within AI systems that auditors assess — from data and model controls to human oversight and access management.
| Control Area | Key Controls to Audit | Evidence to Request |
|---|---|---|
| Data Quality | Data quality gates, completeness checks, representativeness assessment, data lineage tracking | Data quality reports, pipeline test logs, lineage documentation |
| Model Validation | Performance testing, fairness testing, robustness testing, independent review of validation results | Model validation reports, test datasets, approval records |
| Access Control | Least-privilege access to training data and model parameters, separation of duties (developer ≠ validator) | Access control logs, user access reviews, permission matrices |
| Change Management | Version control for models and data, change approval workflow, regression testing, rollback procedures | Change tickets, version history, deployment approvals |
| Human Oversight | HITL review queues, override capability, escalation thresholds, decision audit trails | Review queue logs, override records, SLA compliance reports |
| Monitoring | Performance dashboards, drift alerts, bias metric tracking, anomaly detection | Monitoring dashboards, alert logs, incident tickets from alerts |
| Explainability | LIME/SHAP outputs available for decisions, explanation quality review, regulatory disclosure capability | Explanation samples, disclosure templates, stakeholder comprehension testing |
AI Monitoring & Drift Detection
Continuous monitoring is the primary detective control for AI systems — catching silent degradation, drift, bias creep, and anomalous behavior in production.
Three Types of AI Drift
Key Model Performance Metrics
Explainability Tools (XAI)
| Component | What It Tracks | Audit Questions |
|---|---|---|
| Performance monitoring | Accuracy, precision, recall, F1 against ground truth labels | Is there a feedback loop to capture ground truth in production? How quickly? |
| Drift monitoring | Input distribution (PSI), output distribution, concept drift | What drift thresholds trigger alerts? What happens when a threshold is breached? |
| Bias monitoring | Fairness metrics across demographic groups over time | Are fairness metrics tracked for all protected groups? Is there a remediation SLA? |
| Anomaly detection | Unusual patterns in predictions, inputs, or volumes | Are there automated anomaly alerts? Who receives them? What is the response process? |
| Human override tracking | Rate, reasons, and patterns of human overrides of AI decisions | Is the override rate tracked? High override rates may indicate model degradation or poor calibration. |
| Audit logging | All AI decisions, inputs, outputs, model version, timestamp | Are logs immutable? Retained for required period? Sufficient to reconstruct any decision? |
AI Incidents & Adversarial Threats
Classifying, responding to, and learning from AI incidents — plus the emerging landscape of adversarial attacks specific to AI systems.
AI Incident Severity Classification
AI Incident Response Lifecycle
Adversarial AI Attack Types
Practice Quiz
10 AAIA-style questions on AI Controls, Monitoring & Incidents. Select an answer to see instant feedback.
Memory Hooks
High-yield mnemonics and patterns to lock in AI Controls, Monitoring & Incidents for the AAIA.
| Fact | Answer |
|---|---|
| The line of defense that provides independent AI audit assurance | Third line (Internal Audit) |
| Control that stops the risk before it occurs | Preventive control |
| Hardest drift to detect — requires ground truth labels | Concept drift (label drift) |
| PSI threshold indicating significant drift requiring investigation | PSI > 0.2 (significant); 0.1–0.2 slight shift; <0.1 stable |
| SHAP is based on which mathematical concept? | Shapley values from cooperative game theory |
| Adversarial attack that targets LLMs specifically | Prompt injection |
| Primary defense against membership inference attacks | Differential privacy in training; rate limiting; output perturbation |
| Design effectiveness vs. operating effectiveness | Design = control addresses the right risk; Operating = control is actually performed as designed |
| High override rate on an AI system signals what? | Potential model degradation, poor calibration, or inadequate model performance |
| First step in AI incident response | Detection & Alerting — identify, classify severity, open ticket |
Flashcards & Study Advisor
Click any card to flip it. Use the Study Advisor for targeted guidance by topic area.
In the three lines of defense, which line owns AI systems, and which provides independent audit assurance?
First line (AI development/business units) owns and operates AI systems. Third line (Internal Audit) provides independent assurance. Second line provides oversight and policy but does NOT own systems or run audits.
What is the difference between design effectiveness and operating effectiveness of an AI control?
Design effectiveness: Is the control well-designed to address the targeted risk? (Walkthrough, inspection). Operating effectiveness: Is the control actually being performed as designed, consistently over time? (Sampling, re-performance). Both must be tested in an AI audit.
What are the three types of AI drift, and which is hardest to detect?
Data drift: Input distributions shift (detected by PSI/KS tests). Concept drift: Input-output relationship changes — hardest, requires ground truth labels. Prediction drift: Output distributions shift — easiest, no labels needed. Concept drift is most dangerous.
What is PSI, and what value signals significant drift requiring investigation?
Population Stability Index (PSI) measures the shift in score distributions between training and production. PSI < 0.1 = stable; 0.1–0.2 = slight shift; > 0.2 = significant drift — investigate and consider retraining. Threshold-based monitoring trigger.
What is the key difference between LIME and SHAP, and when should an auditor use each?
LIME = local only (explains one prediction; faster but less consistent). SHAP = global + local (model-wide feature importance + per-prediction explanations; based on Shapley values). Use SHAP for bias audits across the whole model; LIME for quick local explanations of individual decisions.
What are the 6 steps of the AI incident response lifecycle?
① Detection & Alerting ② Containment (limit harm — rollback, suspend, route to humans) ③ Investigation & Root Cause Analysis ④ Remediation (retrain, fix data, patch) ⑤ Recovery & Validation ⑥ Post-Incident Review & Reporting.
What is prompt injection, and which AI systems are vulnerable to it?
Prompt injection = attacker embeds malicious instructions in user input to hijack an LLM's behavior — overriding system prompts, extracting private data, or generating prohibited content. Specific to generative AI and LLM systems. Controls: input sanitization, output filtering, privilege separation between system and user prompts.
A fraud detection model has 97% accuracy but 42% recall. What is the core problem and why?
The model is missing 58% of actual fraud cases (false negatives). High accuracy on an imbalanced dataset is misleading — predicting "no fraud" almost always yields high accuracy when fraud is rare. Recall (sensitivity) = TP/(TP+FN) is the critical metric for fraud detection because false negatives (missed fraud) are far more costly than false positives.
Ready for the Full AAIA Practice Tests?
Access hundreds of AI controls, monitoring, and incident flashcards and practice questions on FlashGenius.
Unlock Full Practice Tests on FlashGenius →Study Advisor
Three Lines of Defense Tips
- Line 2 ≠ Line 3: The risk management function (Line 2) sets AI policy and monitors the first line — it does NOT conduct independent audits. Internal audit (Line 3) is independent and reports to the board/audit committee.
- AI governance committee = Line 2: Chief AI Officers, AI risk managers, and compliance functions all sit in Line 2. They own the framework but not the AI systems.
- Separation is essential: If AI developers are also validating their own models with no independent review, the three-lines model has broken down. This is a governance finding.
- Exam trap: Questions may describe a scenario where "the data science team validates its own models." This is a Line 1 control gap — validation should have Line 2 or independent oversight.
- External audit = 4th line: Some organizations recognize external auditors as a 4th line. Know this if the exam asks about additional assurance beyond internal audit.
Control Design & Testing Tips
- Design before operating: Always assess design effectiveness first. If a control is poorly designed, testing operating effectiveness is irrelevant — it won't address the risk regardless of how well it's followed.
- Evidence is everything: A control that cannot be evidenced is treated as non-operating. If bias testing is required but there are no test reports, the control is not operating effectively regardless of what staff claim.
- HITL patterns on the exam: "Human-in-the-loop" = reviews every decision. "Human-on-the-loop" = monitors but AI acts autonomously. "Human-out-of-the-loop" = fully automated. High-stakes decisions (credit, employment) likely require HITL per GDPR Art. 22.
- Model change management: Automated retraining pipelines without a formal approval gate are a common gap. Even minor threshold changes should go through change management for high-risk AI systems.
- Compensating controls: When a primary control fails (e.g., model is a black box with no explainability), compensating controls (100% human review of high-confidence decisions) reduce residual risk. Document why the compensating control is equivalent.
Monitoring & Drift Tips
- Concept drift is the exam favorite: It's the trickiest — the model's learned mapping becomes invalid because the real-world relationship has changed. No code changes, no data pipeline failures — the world just changed. Requires labeled feedback to detect.
- PSI thresholds to memorize: <0.1 = stable, 0.1–0.2 = slight shift (investigate), >0.2 = significant drift (retrain). These exact thresholds appear in exam questions.
- High override rate is a red flag: If humans are frequently overriding AI decisions, this may indicate model degradation, poor calibration, or a mismatch between model scope and actual use case. Track and investigate override patterns.
- Audit logs must be immutable: For AI decision audit trails, logs must be tamper-proof, retained for the required period, and contain enough information to reconstruct any decision (input, output, model version, timestamp).
- SHAP for bias audits: When the question involves understanding which features drive outcomes across all predictions (not just one), SHAP's global summaries are the right tool. Look for protected-attribute proxies with high SHAP values.
Incidents & Attacks Tips
- Containment before root cause: In AI incident response, containment (stop the bleeding) always comes before deep investigation. A common exam trap presents options where investigation is listed before containment.
- Prompt injection = LLM-specific: This attack only applies to systems that take free-text instructions as inputs. Traditional classifiers cannot be prompt-injected. Know which attack types apply to which model architectures.
- Privacy attacks: Membership inference and model inversion both target training data privacy. Differential privacy is the most rigorous technical defense for both.
- Data poisoning in continuous training: Models that retrain automatically on production data (common in recommendation systems) are especially vulnerable to poisoning attacks — an adversary can inject data into the production stream to corrupt future model versions.
- Post-incident review is required: A complete incident response includes a post-mortem. Questions asking "what is the LAST step" in incident response should be answered with post-incident review and lessons learned — not recovery.
AAIA Exam Strategy
- Domain 2 is 46% of the exam: The combined controls/monitoring/operations domain is the single largest portion. Prioritize controls and monitoring material heavily in your prep.
- Auditor's perspective always: AAIA questions ask what the auditor should do, assess, or find — not what the AI team should build. "The auditor should recommend retraining" is usually wrong; "the auditor should assess whether the retraining process has appropriate controls" is right.
- Detective controls dominate AI: Because AI can degrade silently, detective controls (monitoring, drift detection, anomaly alerts) are the dominant control type for AI systems. Most AAIA control questions will relate to monitoring and detection.
- Risk-based approach: Not all AI systems need the same level of controls. High-risk AI (EU AI Act) requires more rigorous controls than minimal-risk AI. Always factor in risk level when assessing whether controls are appropriate.
- Recall for high-stakes classifiers: In any scenario about fraud detection, medical diagnosis, or safety-critical classification, recall (catching all positives) is the priority metric. False negatives are almost always the larger risk in these domains.