Domain 1. AI Governance and Risk
Learning Objectives:
Within this domain, the AI auditor should be able to:
- Evaluate impacts, opportunities, and risk when integrating AI solutions within the audit process.
 
- Evaluate AI solutions to advise on impact, opportunities, and risk to organization.
 
- Evaluate the impact of AI solutions on system interactions, environment, and humans.
 
- Evaluate the role and impact of AI decision-making systems on the organization and stakeholders.
 
- Evaluate the organization’s AI policies and procedures, including compliance with legal and regulatory requirements.
 
- Evaluate the monitoring and reporting of metrics (e.g., KPIs, KRIs) specific to AI.
 
- Evaluate whether the organization has defined ownership of AI-related risk, controls, procedures, decisions, and standards.
 
- Evaluate the organization’s data governance program specific to AI.
 
- Evaluate the organization’s privacy program specific to AI.
 
- Evaluate the organization’s problem and incident management programs specific to AI.
 
- Evaluate the organization’s change management program specific to AI.
 
- Evaluate the organization’s configuration management program specific to AI.
 
- Evaluate the organization’s threat and vulnerability management programs specific to AI.
 
- Evaluate the organization’s identity and access management program specific to AI.
 
- Evaluate vendors and supply chain management program specific to AI solutions.
 
- Evaluate the design and effectiveness of controls specific to AI.
 
- Evaluate data inputs requirements for AI models (e.g., data appropriateness, bias, and privacy).
 
- Evaluate system/business requirements for AI solutions to ensure alignment with enterprise architecture.
 
- Evaluate AI solution life cycle (e.g., design, development, deployment, monitoring, and decommissioning) and inputs/outputs for compliance and risk.
 
- Evaluate algorithms and models to ensure AI solutions are aligned to business objectives, policies, and procedures.
 
- Analyze the impact of AI on the workforce to advise stakeholders on how to address AI-related workforce impacts, training, and education.
 
- Evaluate that awareness programs align to the organization’s AI-related policies and procedures.
 
 
Section A. AI Models, Considerations, and Requirements
1. Types of AI
- Generative
 
- Predictive
 
- Narrow
 
- General
 
2. Machine learning/AI Models
- Basic models
 
- Neural networks
 
3. Algorithms
- Classes of Algorithms
 
- Additional AI Considerations (technical terms and concepts relevant to the IS auditor)
 
4. AI Lifecycle Overview
- Plan and Design
 
- Collect and Process Data
 
- Build and/or Adapt Model(s)
 
- Test, Evaluate, Verify, and Validate
 
- Make Available for Use/Deploy
 
- Operate and Monitor
 
- Retire/Decommission
 
5. Business Considerations
- Business Use Cases, Needs, Scope, and Objectives
 
- Cost-Benefit Analysis
 
- Return on Investment
 
- Internal vs. Cloud Hosting
 
- Vendors
 
- Shared Responsibility
 
 
Section B. AI Governance and Program Management
1. AI Strategy
- Strategies
 
- Opportunities
 
- Vision and Mission
 
- Value Alignment
 
2. AI-related Roles and Responsibilities
- Categories, Focuses, and Common Examples
 
3. AI-related Policies and Procedures
4. AI Training and Awareness
- Skills, Knowledge, and Competencies
 
5. Program metrics
- Examples of Metrics with Objectives and Definitions
 
 
Section C. AI Risk Management
1. AI-related Risk Identification
- AI Threat Landscape
 
- AI Risks
 
- Challenges for AI Risk Management
 
2. Risk Assessment
- Risk Assessment
 
- Risk Appetite and Tolerance
 
- Risk Mitigation and Prioritization
 
- Remediation Plans/Best Practices
 
3. Risk Monitoring
- Continuous Improvement
 
- Risk and Performance Metrics
 
 
Section D. Privacy and Data Governance Programs
1. Data Governance
- Data Classification
 
- Data Clustering
 
- Data Licensing
 
- Data Cleansing and Retention
 
2. Privacy Considerations
- Data Privacy
 
- Data Ownership (Governance and Privacy)
 
3. Privacy Regulatory Considerations
- Data Consent
 
- Collection, Use, and Disclosure
 
 
Section E. Leading Practices, Ethics, Regulations, and Standards for AI
1. Standards, Frameworks, and Regulations Related to AI
- Best Practices
 
- Industry Standards and Frameworks
 
- Laws and Regulations
 
2. Ethical Considerations
- Ethical Use
 
- Bias and Fairness
 
- Transparency and Explainability
 
- Trust and Safety
 
- IP Considerations
 
- Human Rights
 
 
Domain 2. AI Operations
Learning Objectives:
Within this domain, the AI auditor should be able to:
- Evaluate impacts, opportunities, and risk when integrating AI solutions within the audit process.
 
- Evaluate AI solutions to advise on impact, opportunities, and risk to organization.
 
- Evaluate the impact of AI solutions on system interactions, environment, and humans.
 
- Evaluate the role and impact of AI decision-making systems on the organization and stakeholders.
 
- Evaluate the organization’s AI policies and procedures, including compliance with legal and regulatory requirements.
 
- Evaluate the monitoring and reporting of metrics (e.g., KPIs, KRIs) specific to AI.
 
- Evaluate whether the organization has defined ownership of AI-related risk, controls, procedures, decisions, and standards.
 
- Evaluate the organization’s data governance program specific to AI.
 
- Evaluate the organization’s privacy program specific to AI.
 
- Evaluate the organization’s problem and incident management programs specific to AI.
 
- Evaluate the organization’s change management program specific to AI.
 
- Evaluate the organization’s configuration management program specific to AI.
 
- Evaluate the organization’s threat and vulnerability management programs specific to AI.
 
- Evaluate the organization’s identity and access management program specific to AI.
 
- Evaluate vendors and supply chain management program specific to AI solutions.
 
- Evaluate the design and effectiveness of controls specific to AI.
 
- Evaluate data inputs requirements for AI models (e.g., data appropriateness, bias, and privacy).
 
- Evaluate system/business requirements for AI solutions to ensure alignment with enterprise architecture.
 
- Evaluate AI solution life cycle (e.g., design, development, deployment, monitoring, and decommissioning) and inputs/outputs for compliance and risk.
 
- Evaluate algorithms and models to ensure AI solutions are aligned to business objectives, policies, and procedures.
 
- Analyze the impact of AI on workforce to advise stakeholders to address AI-related workforce impacts, training, and education.
 
- Evaluate that awareness programs align to the organization’s AI-related policies and procedures.
 
 
Section A. Data Management Specific to AI
1. Data Collection
- Consent
 
- Fit for Purpose
 
- Data Lag
 
2. Data Classification
3. Data Confidentiality
4. Data Quality
5. Data Balancing
6. Data Scarcity
7. Data Security
- Data Encoding
 
- Data Access
 
- Data Secrecy
 
- Data Replication
 
- Data Backup
 
 
Section B. AI Solution Development Methodologies and Lifecycle
1. AI Solution Development Life Cycle
- Use Case Development
 
- Design
 
- Development
 
- Deployment
 
- Monitoring and Maintenance
 
- Decommission
 
2. Privacy and Security by Design
- Explainability
 
- Robustness
 
 
Section C. Change Management Specific to AI
1. Change Management Considerations
- Data Dependency
 
- AI Model
 
- Regulatory and Societal Impact
 
- Emergency Changes
 
- Configuration Management
 
Section D. Supervision of AI Solutions
1. AI Agency
- Logging and Monitoring
 
- AI Observability
 
- Human in the Loop (HITL)
 
- Hallucination
 
Section E. Testing Techniques for AI Solutions
1. Conventional Software Testing Techniques
- A/B Testing
 
- Unit and Integration Testing
 
- Objective Verification
 
- Code Reviews
 
- Black Box Testing
 
2. AI-Specific Testing Techniques
- Model Cards
 
- Bias Testing
 
- Adversarial Testing
 
 
Section F. Threats and Vulnerabilities Specific to AI
1. Types of AI-related Threats
- Training Data Leakage
 
- Data Poisoning
 
- Model Poisoning
 
- Model Theft
 
- Prompt Injections
 
- Model Evasion
 
- Model Inversion
 
- Threats for Using Vendor Supplied AI
 
- AI Solution Disruption
 
2. Controls for AI-related Threats
- Threat and Vulnerability Identification
 
- Prompt Templates
 
- Defensive Distillation
 
- Regularization
 
 
Section G. Incident Response Management Specific to AI
1. Prepare
- Policies, Procedures, and Model Documentation
 
- Incident Response Team
 
- Tabletop Exercises
 
2. Identify and Report
3. Assess
4. Respond
- Containment
 
- Eradication
 
- Recovery
 
 
5. Post-Incident Review
Domain 3. AI Auditing Tools and Techniques
Learning Objectives:
Within this domain, the AI auditor should be able to:
- Evaluate impacts, opportunities, and risk when integrating AI solutions within the audit process.
 
- Utilize AI solutions to enhance audit processes, including planning, execution, and reporting.
 
- Evaluate the monitoring and reporting of metrics (e.g., KPIs, KRIs) specific to AI.
 
- Evaluate data input requirements for AI models (e.g., data appropriateness, bias, and privacy).
 
 
Section A. Audit Planning and Design
1. Identification of AI Assets and Controls
- Inventory Objective and Procedure
 
- Inventory and Data Gathering Methods
 
- Documentation
 
2. Types of AI Controls
- Examples including Control Categories, Controls, and Explanations
 
3. Audit Use Cases
- Large Language Models
 
- Audit Process Improvement
 
- Generative AI
 
- Audit-Specific AI Applications
 
4. Internal Training for AI Use
- Key Components for Auditor Knowledge
 
- Practical Skills Development
 
 
Section B. Audit Testing and Sampling Methodologies
1. Designing an AI Audit
- AI Audit Objectives
 
- Audit Scoping and Resources
 
2. AI Audit Testing Methodologies
- AI Systems Overall Testing
 
- Financial Models
 
3. AI Sampling
- Judgmental sampling
 
- AI sampling
 
4. Outcomes of AI testing
- Reduce false positives
 
- Reduce workforce needs
 
- Outliers
 
 
Section C. Audit Evidence Collection Techniques
1. Data Collection
- Training and Testing Data
 
- Unstructured and Structured Data Collection
 
- Extract, Transform, and Load
 
- Data Manipulation
 
- Scraping
 
2. Walkthroughs and interviews
- Design Interview Questions
 
3. AI Collection Tools
- Using AI to Collect Logs
 
- AI agents to create outputs
 
- Voice to Speech
 
- Optimal Character Recognition
 
 
Section D. Audit Data Quality and Data Analytics
1. Data Quality
2. Data Analytics
3. Data Reporting
 
Section E. AI Audit Outputs and Reports
1. Reports
- Report Types (examples and details)
 
- Advisory Reports
 
- Charts and Visualizations
 
2. Audit Follow-up
3. Quality Assurance and mitigate risk.