Spring 2025
By the end of this activity, you will be able to:
You will complete Part A individually and submit before class on Monday 5/4. Parts B and C will be done in pairs, beginning in class, with any remaining tasks completed by the following Monday, 5/11.
Automated prediction systems are often used to make or inform decisions that affect people’s lives. Often these decisions relate to the allocation of resources:
When these algorithms make systematically different allocative recommendations for different groups of people, awarding resources to some while denying those same resources to others for arbitrary reasons, we may have an instance of allocative harm or allocative bias.1
In this activity we’ll explore a famous case of allocative bias in the distribution of medical care, documented in the journal article Dissecting Racial Bias in an Algorithm Used to Manage the Health of Populations by Ziad Obermeyer, Brian Powers, Christine Vogeli, and Sendhil Mullainathan (Science, 2019).
In this study, the authors considered a healthcare recommendation system which assigned a medical risk score to patients; patients with higher scores were considered to face greater future risk to their health. Patients who received very high risk scores were recommended for an intensive care program intended to improve their outcomes. Here, it’s important to remember:
If a patient is indeed at high medical risk, then it is a good outcome for them to receive a high risk score and be recommended for intensive care.
Take ~30 minutes to read through the article. The PDF version of this article can be directly accessed at this link hosted by the FTC. You can focus on the first four pages of the text, especially including the abstract, the description of the data set, Fig. 1 and surrounding text, and the section “Mechanism of bias.”
In a text editor of your choice that is capable of exporting to pdf, answer these questions:
Submit your answers in PDF format to the Data Ethics 2 - Reading and Questions assignment on Canvas.
Download the allocative_bias.ipynb notebook and complete Parts B and C. Details of these tasks are given in the notebook.
Submit your completed notebook to the Data Ethics 2 - Activity assignment on Canvas.
Your answers to the questions from Part A are scored out of 4 points, and as usual will be graded on effort, thoughtfulness, and clarity.
Your solutions for Parts B and C are scored out of 6 points, one for each task in the notebook.
This assignment was adapted from an assignment generously shared by Phil Chodrow.
See this talk for a good primer on bias in ML and why it’s such a difficult problem: https://www.youtube.com/watch?v=fMym_BKWQzk↩︎