Linear Digressions

Debiasing Word Embeddings

Linear Digressions

When we covered the Word2Vec algorithm for embedding words, we mentioned parenthetically that the word embeddings it produces can sometimes be a little bit less than ideal--in particular, gender bias from our society can creep into the embeddings and give results that are sexist. For example, occupational words like "doctor" and "nurse" are more highly aligned with "man" or "woman," which can create problems because these word embeddings are used in algorithms that help people find information or make decisions. However, a group of researchers has released a new paper detailing ways to de-bias the embeddings, so we retain gender info that's not particularly problematic (for example, "king" vs. "queen") while correcting bias.

Next Episodes

Linear Digressions

The Kernel Trick and Support Vector Machines @ Linear Digressions

📆 2017-12-11 02:58 / 00:17:48


Linear Digressions

Maximal Margin Classifiers @ Linear Digressions

📆 2017-12-04 05:03 / 00:14:21


Linear Digressions

Re - Release: The Cocktail Party Problem @ Linear Digressions

📆 2017-11-27 03:11 / 00:13:43


Linear Digressions

Clustering with DBSCAN @ Linear Digressions

📆 2017-11-20 04:08 / 00:16:14


Linear Digressions

The Kaggle Survey on Data Science @ Linear Digressions

📆 2017-11-13 03:49 / 00:25:20