Lecture 38 - Bias in ML

Announcements:

Goals:

Now that you know the overall shape of how to train and use machine learning models, it's worth asking:

What could possibly go wrong?

The state of the art, especially in vision and language

Example: how-to-make-a-racist-ai-without-really-trying.ipynb

Further reading:

ConceptNet Numberbatch 17.04: better, less-stereotyped word vectors

Semantics derived automatically from language corpora contain human-like biases

Image Representations Learned With Unsupervised Pre-Training Contain Human-like Biases

Towards Fairer Datasets: Filtering and Balancing the Distribution of the People Subtree in the ImageNet Hierarchy

racist data destruction? a Boston housing dataset controversy