Foundations of AI Governance
Domain I establishes the conceptual vocabulary that every other AIGP domain builds on. You cannot govern what you do not understand — this domain covers what AI is, how it works, where it can go wrong, and the ethical principles that shape responsible governance.
Why Governance? AI systems make decisions at a scale, speed, and complexity that no individual human overseer can match. A hiring algorithm can screen thousands of applicants in seconds — and encode discrimination at the same scale. Governance exists to ensure AI serves human values, remains accountable, and causes no unjustified harm.
Understanding the difference between AI types, how machine learning models are trained, and what generative AI and foundation models actually do — the technical vocabulary the AIGP exam assumes you know.
Governance doesn't start at deployment — it must be embedded from the first line of problem definition through decommissioning. Each stage introduces different risks and requires different controls.
The ethical principles — fairness, transparency, accountability, explainability, safety, and human oversight — that form the normative foundation of every AI governance framework tested on the AIGP exam.
| Topic Area | Key Concepts |
|---|---|
| AI capability types | Narrow AI (ANI), Artificial General Intelligence (AGI), Superintelligence (ASI) |
| Machine learning approaches | Supervised, unsupervised, reinforcement learning; deep learning; foundation models |
| Generative AI | LLMs, diffusion models, hallucination, prompt engineering, fine-tuning |
| AI lifecycle | 7 stages from problem definition to retirement; governance touchpoints at each stage |
| AI risk categories | Safety, privacy, bias/fairness, security, accountability, transparency, autonomy |
| Responsible AI principles | Fairness, transparency, accountability, explainability, safety, human oversight |
| International AI principles | OECD AI Principles (5), UNESCO Recommendation, IEEE Ethically Aligned Design |
| Key terminology | Algorithm, model, training, inference, bias, drift, hallucination, interpretability |
AI governance sits at the intersection of technology, law, ethics, and organizational risk management. An AI governance professional must be fluent in all four domains — understanding enough technical detail to identify risk, enough law to ensure compliance, enough ethics to spot value misalignment, and enough risk management to prioritize controls.
AI is the broad field of systems that simulate human intelligence.
Machine Learning (ML) is a subset of AI — systems that learn from data without being explicitly programmed for each decision.
Deep Learning is a subset of ML using multi-layered neural networks — the engine behind image recognition, NLP, and generative AI.
AI & Machine Learning Fundamentals
The AIGP exam expects you to understand AI at a conceptual level — not to build models, but to govern them. That requires knowing what type of AI you're governing and how it learns.
AI Capability Tiers
Machine Learning Approaches
| Concept | Definition | Governance Implication |
|---|---|---|
| Large Language Model (LLM) | Foundation model trained on vast text to generate, summarize, translate, and reason in natural language | Risk of hallucination, copyright infringement, bias amplification, and misuse at scale |
| Hallucination | AI generates plausible-sounding but factually incorrect information with apparent confidence | High-stakes use cases (medical, legal) require human review; transparency obligations apply |
| Fine-tuning | Adapting a pretrained foundation model for a specific task using smaller, domain-specific datasets | Fine-tuning data quality, bias, and provenance must be governed — not just the base model |
| Prompt Engineering | Crafting inputs to guide model outputs toward desired behavior without retraining | Prompt injection attacks; adversarial prompting; output quality and safety depend on prompt design |
| Multimodal AI | AI that processes and generates multiple data types — text, images, audio, video — in combination | Expanded harm surface: deepfakes, synthetic media, cross-modal bias |
These terms are often confused but have distinct meanings in AI governance contexts. Both matter for accountability and regulatory compliance.
Simple models (linear regression, decision trees) are highly interpretable but less accurate for complex tasks. Complex models (deep neural networks, LLMs) achieve higher accuracy but are "black boxes" — difficult to interpret, requiring external explainability tools (SHAP, LIME).
The AI Lifecycle & Governance Touchpoints
Governance must be embedded throughout the entire AI lifecycle — not bolted on at deployment. Each stage introduces distinct risks and requires specific controls. The AIGP exam tests governance at every stage.
Key Exam Rule: Governance applies from Stage 1 — problem definition — not from deployment. The most consequential governance decisions (what the AI is for, what data it can use, what oversight is required) are made before a single line of code is written. Late-stage governance is reactive; early-stage governance is preventive.
| Stage | Key Governance Activity | Output / Artifact |
|---|---|---|
| Problem Definition | AI use case risk classification; define intended use and prohibited uses | Use Case Assessment, Risk Tier Classification |
| Data Collection | Privacy impact assessment; consent verification; bias audit on source data | Data Provenance Record, DPIA, Bias Assessment |
| Training | Document model architecture, hyperparameters, training data version | Model Card, Technical Documentation |
| Validation & Testing | Fairness testing across protected groups; adversarial testing; performance thresholds | Validation Report, Red Team Findings |
| Deployment | Human oversight design; user notification and disclosure; access controls | Deployment Authorization, User Notice |
| Monitoring | Drift detection; incident logging; periodic re-evaluation | Monitoring Dashboard, Incident Reports |
| Retirement | Safe data disposal; archive model documentation; communicate to affected users | Decommission Plan, Final Audit Report |
Data drift: The statistical distribution of input data changes after deployment (e.g., customer behavior shifts post-pandemic). The model wasn't trained on this new distribution.
Concept drift: The relationship between inputs and outputs changes (e.g., what constitutes fraudulent behavior evolves). The model's predictions become systematically wrong even on similar inputs.
Governance implication: Continuous monitoring with predefined performance thresholds and re-training triggers is mandatory for high-stakes AI systems.
Model Card: A short document accompanying a trained model that describes its intended use, performance across demographic groups, limitations, and known failure modes. Developed by Google.
Datasheet for Datasets: Standardized documentation of a training dataset's motivation, composition, collection process, preprocessing, and recommended uses. Promotes transparency and accountability in data-driven AI.
AI Risk Categories & Harms
AI risk is multi-dimensional. Governance professionals must understand each category, its causes, and the controls that mitigate it. The AIGP exam presents scenario-based questions that require mapping a harm to its risk type and appropriate response.
| Bias Type | Where It Originates | Example |
|---|---|---|
| Historical bias | Training data reflects past societal discrimination | Job ad algorithm deprioritizes women because historical hiring data was male-dominated |
| Representation bias | Training data underrepresents certain groups | Facial recognition trained mostly on lighter-skinned faces performs worse on darker-skinned faces |
| Measurement bias | Proxy variables introduced during data collection | Using zip code as a proxy for creditworthiness encodes racial segregation patterns |
| Aggregation bias | One model applied to groups with different characteristics | Medical model trained on predominantly male data applied equally to female patients |
| Deployment bias | Model used in a context different from what it was trained for | Resume-screening tool trained on tech roles applied to non-tech hiring |
| Feedback loop bias | Model outputs influence future training data | Predictive policing tool increases arrests in over-patrolled areas → more data → more predictions → more policing |
Adversarial Inputs: Carefully crafted inputs designed to fool the model — a stop sign with stickers that an autonomous vehicle misclassifies.
Data Poisoning: Injecting malicious data into the training set to corrupt model behavior — training a spam filter to pass certain phishing emails.
Model Inversion: Querying a model repeatedly to reconstruct sensitive training data — recovering personal information from a medical model.
Prompt Injection: Embedding malicious instructions in user inputs to override the LLM's intended behavior or safety controls.
Risk Tiering: Classify AI use cases by potential harm severity and scope. High-risk = more controls. This is how the EU AI Act structures its obligations.
AI Impact Assessment: Pre-deployment analysis of potential harms across population groups, similar to a Privacy Impact Assessment but broader in scope.
Red Teaming: Adversarial testing where a dedicated team attempts to make the AI system fail, produce harmful outputs, or be manipulated — identifies risks before deployment.
Responsible AI Principles
Responsible AI is the normative core of AI governance. These principles — drawn from international consensus — define what "good AI" looks like and provide the ethical baseline that laws and frameworks operationalize.
Principles ≠ Rules: Responsible AI principles are not checklists — they are values that must be interpreted and applied to specific contexts. The same principle (e.g., fairness) may require different technical implementations depending on the use case, affected population, and legal context.
The 5 OECD AI Principles (2019)
-
1Inclusive Growth, Sustainable Development & Well-beingAI should benefit people and the planet. Stakeholders should engage proactively in responsible stewardship of trustworthy AI in pursuit of beneficial outcomes — economic, social, and environmental.
-
2Human-Centered Values & FairnessAI must respect the rule of law, human rights, democratic values, and diversity. AI systems should not discriminate unjustly and should be designed to avoid or correct bias against individuals or groups.
-
3Transparency & ExplainabilityAI actors should commit to transparency and responsible disclosure. People interacting with AI systems should be able to understand AI-driven decisions and have access to meaningful explanations.
-
4Robustness, Security & SafetyAI systems must function reliably and safely throughout their lifecycle. They should be secure against attacks, resilient to errors, and developed with risk management mechanisms that meet safety standards.
-
5AccountabilityAI actors — developers, deployers, and users — must be accountable for outcomes and should enable mechanisms for human oversight, auditability, and redress when AI causes harm.
Significance: The OECD AI Principles (2019) were the first internationally agreed-upon, government-backed standards for responsible AI — adopted by 42+ countries. They are non-binding but highly influential, shaping the EU AI Act, US AI policy, and every major national AI strategy.
Core Responsible AI Principles — Exam Definitions
| Principle | Definition | Practical Application |
|---|---|---|
| Fairness | AI must not discriminate unjustifiably against individuals or groups based on protected characteristics | Bias audits across demographic groups; fairness metrics (equal opportunity, demographic parity) |
| Transparency | Organizations must be open about what AI systems do, how they work, and when AI is being used | User notices, model cards, public disclosures of AI use in consequential decisions |
| Accountability | Someone must be responsible for AI outcomes — humans cannot hide behind algorithmic decisions | Clear ownership of AI systems; audit trails; grievance and redress mechanisms |
| Explainability | AI decisions, especially consequential ones, must be explainable to affected individuals and regulators | Explainability tools (SHAP, LIME); human-readable justifications for automated decisions |
| Safety | AI must not cause physical or psychological harm; must be designed with fail-safes and risk mitigation | Safety testing, red teaming, human-in-the-loop for high-risk decisions |
| Human Oversight | Meaningful human control must be maintained, especially for high-impact AI decisions | Human-in-the-loop design; override mechanisms; prohibition on fully automated high-stakes decisions |
| Privacy | AI must respect individuals' rights to control their personal data and be free from undue surveillance | Data minimization, purpose limitation, consent management, anonymization |
| Beneficence | AI should produce positive outcomes for individuals, society, and the environment | Impact assessments; benefit analysis; equitable access to AI benefits |
Essential AI Governance Terminology
Practice Quiz — Foundations of AI Governance
10 AIGP-style scenario questions. Select your answer to see instant feedback and explanation.
Review explanations above for any missed questions.
Memory Hooks & AI Advisor
Lock in the most exam-tested foundations concepts. Use the Advisor for focused deep-dive guidance by category.
Flashcards
Click each card to flip and reveal the answer.
What distinguishes Narrow AI from AGI, and which exists today?
Narrow AI (ANI) handles ONE specific task — all AI today. AGI reasons across ALL domains like a human — theoretical only. No AGI exists commercially.
What are the 3 core ML approaches and their defining characteristic?
Supervised = labeled data → predictions. Unsupervised = unlabeled data → patterns. Reinforcement = rewards → optimized behavior over time.
At which lifecycle stage does governance FIRST apply, and why?
Stage 1: Problem Definition. The most consequential decisions — purpose, scope, oversight design — are made here. Late governance is reactive; early governance prevents harm.
What is historical bias and why is it common in AI hiring tools?
Historical bias: training data reflects past discrimination. If past hiring was male-dominated, the model learns to replicate that pattern — even without using gender as an explicit feature.
What is the difference between explainability and interpretability?
Explainability = WHY this specific decision was made (post-hoc, per-decision). Interpretability = HOW the model works overall (global, model-structural). SHAP/LIME provide explainability for black-box models.
What are the 5 OECD AI Principles and are they legally binding?
1. Inclusive Growth 2. Human-Centered Values & Fairness 3. Transparency & Explainability 4. Robustness, Security & Safety 5. Accountability. Non-binding but first internationally agreed AI standards (42+ countries, 2019).
What is AI hallucination and what governance control addresses it?
Hallucination: generative AI produces confident but factually wrong or fabricated outputs. Control: mandatory human review for high-stakes LLM outputs; user disclosures about AI limitations.
What is the difference between data drift and concept drift?
Data drift: statistical distribution of inputs changes after deployment. Concept drift: the underlying relationship between inputs and outputs changes. Both require continuous monitoring and re-training triggers.
AI Advisor
Select a category for focused exam guidance.
AI & ML Concepts
- All deployed AI is Narrow AI: Every real-world AI system — including GPT-4, Claude, and image recognition — is ANI. AGI is theoretical. The AIGP exam won't ask you to design AGI, but may ask what governance challenges it would pose.
- Supervised learning = most common: Labeled data + prediction task. If someone describes a model that was "trained with examples of correct outputs," that's supervised learning.
- Unsupervised ≠ uncontrolled: Unsupervised doesn't mean ungoverned. Clustering algorithms that segment customers without labels can still encode bias in their grouping criteria.
- Foundation models shift governance complexity: A single foundation model underlies thousands of applications. Governance failures upstream (in pretraining data) propagate to all downstream uses.
- Hallucination is structural, not a bug: LLMs generate text probabilistically — the next most likely token. They have no internal truth-checking mechanism. Governance must assume hallucination risk exists in all LLM outputs.
- Explainability vs Interpretability: Explainability answers "why THIS decision?" — post-hoc, per-instance. Interpretability answers "how does THIS model work?" — global, structural. Black-box models need external XAI tools (SHAP, LIME) for explainability.
- Simple models sacrifice accuracy for transparency: A linear regression is fully interpretable but may be less accurate than a deep neural network. This tradeoff is a governance decision — high-stakes decisions may require interpretable models even at accuracy cost.
AI Lifecycle & Governance
- Governance starts at Stage 1: Problem definition is where the highest-leverage governance decisions are made. If an AI use case is fundamentally harmful, no amount of post-deployment monitoring fixes it.
- Data stage is the most common bias entry point: Training data quality, representativeness, labeling accuracy, and consent are all governance obligations at Stage 2 — before any model is trained.
- Model Cards are governance artifacts: Not just technical documents. A Model Card that honestly describes limitations, performance across demographic groups, and known failure modes is a transparency and accountability tool.
- Retirement is a governance stage: Decommissioning an AI system requires data disposal plans, user notification, model archiving, and a final audit. The AIGP exam may test whether you recognize retirement as part of the governance lifecycle.
- Monitoring is never optional for high-risk AI: Concept drift and data drift can cause a well-designed model to produce harmful outcomes months after deployment. Continuous monitoring with predefined alerting thresholds is mandatory governance practice.
- Human-in-the-loop placement matters: Human review BEFORE a decision takes effect (meaningful oversight) differs fundamentally from human review after harm has occurred (reactive oversight). Governance frameworks require the former for high-stakes decisions.
AI Risk Categories
- Bias is both a technical and ethical failure: Algorithmic bias isn't just a bad prediction — it's a potential civil rights violation. Disparate impact on protected classes can create legal liability even when discrimination is unintentional.
- Historical bias is the most common: Most real-world bias originates in training data that reflects historical patterns of societal discrimination. Cleaning the data doesn't automatically remove the bias — the patterns are encoded in the statistical relationships.
- Feedback loops amplify bias over time: When a model's outputs influence future training data (predictive policing → more arrests in targeted areas → more data confirming predictions), bias compounds. Governance must break these loops.
- Adversarial attacks are AI-specific: Traditional software security doesn't address model inversion, data poisoning, or adversarial inputs. AI security governance requires AI-specific threat modeling.
- Accountability gaps are organizational: "The algorithm decided" is not a legal defense. Governance frameworks explicitly require organizations to identify accountable humans for all AI-driven outcomes.
- Privacy risk is broader than GDPR: AI can infer sensitive attributes (health, sexual orientation, political views) from innocuous data at scale. Privacy governance for AI goes beyond consent — it requires purpose limitation, data minimization, and re-identification risk assessment.
Responsible AI Principles
- OECD Principles are non-binding but foundational: 42+ countries adopted them in 2019. They provide the normative vocabulary used by the EU AI Act, US AI policy, and virtually every national AI strategy. The exam treats them as the shared baseline.
- Principles require interpretation: "Fairness" means different things statistically — demographic parity vs. equal opportunity vs. predictive parity can all be defined as fair but are mathematically incompatible. Context determines which applies.
- Accountability ≠ Transparency: Transparency means being open about what the AI does. Accountability means someone is responsible for what it does. A system can be transparent (fully documented) but still have no accountable owner.
- Human oversight has degrees: Human-in-the-loop (humans approve each decision) vs. human-on-the-loop (humans can override) vs. human-in-command (humans can shut down) represent different oversight levels. High-risk AI requires the first.
- UNESCO Recommendation (2021): Broader than OECD — 193 member states. Introduces environmental sustainability and right to privacy as explicit AI governance concerns alongside the OECD principles.
- Beneficence vs Non-maleficence: Responsible AI requires both — designing AI to produce positive outcomes AND to avoid harm. The absence of active harm is not sufficient; the system must also deliver genuine benefit.
Key Terminology
- Algorithm vs Model: An algorithm is the method (e.g., random forest). A model is the output of running that algorithm on specific training data. Multiple models can use the same algorithm but behave differently based on data.
- Inference: Using a trained model to make predictions on new data. Most AI harm occurs at inference time — during deployment — not during training. Governance at inference requires monitoring, logging, and access controls.
- Overfitting vs Generalization: An overfit model memorizes training data and fails on new inputs. Good generalization is a governance concern — a model that only works on training data provides false confidence and fails in production.
- Foundation model ≠ fine-tuned model: The base foundation model is pretrained at scale. Fine-tuning adapts it for a specific task. Governance must address both — base model risks AND risks introduced by fine-tuning data and process.
- Red teaming: Adversarial testing by a dedicated team attempting to break the AI system — finding safety failures, harmful outputs, bias, security vulnerabilities — before deployment. Now required by several AI frameworks and regulations.
- Model Card: Standardized documentation of a model's intended use, performance characteristics, limitations, and demographic performance disparities. Increasingly required by regulators and treated as a governance artifact, not just a technical document.
- Human-in-the-loop: A system design where a human must review and approve AI outputs before they take effect. Critical for high-stakes domains (hiring, lending, healthcare, criminal justice). Contrasts with fully automated decision-making.
Reinforce Your AIGP Foundations
Deepen your understanding with targeted AIGP flashcard decks on FlashGenius — covering all four AIGP domains.
Unlock Full Flashcard Deck on FlashGenius →