Dr. Dongyang

AI Governance & Ethics — Cheatsheet

Quick reference based on Alison: AI Governance and Ethics and some extra follow-up readings


Five Pillars of Responsible AI

Pillar Core Question
Fairness Does the system produce equitable outcomes across demographics?
Transparency Can stakeholders understand how decisions are made?
Accountability Is there a clear owner for AI outcomes and harms?
Privacy Is personal data protected throughout the AI lifecycle?
Societal Impact Are broader economic and social effects being managed?

1. Bias & Fairness

Where Bias Comes From

Source Example
Training data Historical hiring data encoding past discrimination
Model design Word embeddings reproducing stereotypes ("doctor" ↔ male)
Deployment context Predictive policing reinforcing over-policing of minority areas

Bias Mitigation Techniques

Key Evidence

Study Finding
Buolamwini & Gebru "Gender Shades" (2018) Facial recognition error: up to 34.7% for darker-skinned females vs 0.8% for lighter-skinned males
ProPublica on COMPAS (2016) Black defendants ~2x more likely to be falsely flagged as high recidivism risk
Obermeyer et al. (2019) Healthcare algorithm underestimated illness in Black patients by using cost as a proxy for health

2. Transparency & Explainability

The Problem

Deep learning models act as "black boxes" — high performance but low interpretability, undermining trust in high-stakes domains (healthcare, criminal justice, finance).

Solutions

Tool / Approach What It Does
LIME Approximates any model locally with an interpretable model to explain individual predictions
SHAP Uses Shapley values (game theory) to attribute each feature's contribution to a prediction
Interpretable algorithms Simpler models (decision trees, linear models) where appropriate
Explainability by design Build explanation capabilities into the system from the start

Why It Matters


3. Accountability

Key Principles

Challenges


4. Data Privacy & Security

Privacy Measures

Measure Description
Privacy by Design Embed privacy protections into every stage of AI development
Privacy Impact Assessments Identify data risks before deployment
Differential Privacy Statistical guarantees that individual records don't significantly affect outputs
Anonymization Remove personally identifiable information from datasets
Encryption + Access Controls Protect data at rest and in transit
Informed Consent Ensure individuals understand and agree to how their data is used

Security Priorities

Cultural Practices


5. Societal Impact


Regulatory Landscape

Regulation / Framework Scope
GDPR EU data protection: consent, right to explanation, data breach accountability
CCPA California: consumer rights to know, delete, opt-out of data sales
EU AI Act Risk-based AI regulation: prohibited, high-risk, limited-risk, minimal-risk tiers
IEEE Ethically Aligned Design 8 principles + P7000 standards series for autonomous/intelligent systems
Asilomar AI Principles 23 principles on research, ethics/values, and long-term AI issues
EU Ethics Guidelines for Trustworthy AI Trustworthy AI = lawful + ethical + robust; 7 key requirements
OECD AI Principles 5 values-based principles; adopted by 47 countries, endorsed by G20
ISO/IEC 42001:2023 First international standard for AI management systems
Google's AI Principles 7 objectives: beneficial, unbiased, safe, accountable, private, rigorous, available

Case Study Lessons

HealthCore (Healthcare AI)

MediTech (Diagnostic Tool Bias)

TechNova (Regulatory Compliance)

Cambridge Analytica (Data Misuse)

PredPol (Predictive Policing)

Watson for Oncology (Healthcare AI Failure)


Organizational Action Checklist

Regulations & Laws

Ethics Frameworks & Principles

Standards

Research Organizations & Initiatives

Explainability Tools

Key Research & Case Studies

Privacy Techniques