Week 3: Ethics & Bias
Week 3: Ethics & Bias#
Focus: Algorithmic bias and fairness
📚 Required Readings#
Primary Readings#
“Weapons of Math Destruction” — Chapter 1 (30 min)
- How algorithms can amplify inequality
- Case studies of biased AI systems
- The impact on vulnerable communities
“Defining and Detecting Algorithmic Bias” (25 min)
- Statistical definitions of fairness
- Types of bias: historical, measurement, representation, evaluation
- Methods for detecting and mitigating bias
Supplementary Resources#
- ProPublica: “Machine Bias” investigation — Interactive article
- “The Ethics of Artificial Intelligence” — Stanford Encyclopedia excerpt
💭 Discussion Prompts#
- How can an AI system be mathematically “accurate” yet fundamentally unfair?
- When is some level of bias acceptable in AI systems?
- Who should be responsible for the harms caused by biased algorithms?
📝 Preparation for Class#
- Read about a real-world case of algorithmic bias (e.g., Amazon hiring, COMPAS, facial recognition)
- Prepare a 2-minute summary: What went wrong? Who was harmed?
- Brainstorm: How could that system have been designed more fairly?
🔗 Related Assignment#
See Week 3 Assignment for this week’s task.