Fairness in Machine Learning: Detecting and Removing Gender Bias in Language Models

Introduction

This study addresses gender biases in language models by implementing and testing two approaches: debiasing the dataset and debiasing the corpus embeddings (hard debiasing), resulting in significant reductions of bias in the BERT model, with dataset debiasing reducing bias by 29.41% and hard debiasing by 11.76%, accompanied by minimal classification accuracy loss.

Aesha Shah
Aesha Shah
ex-Data Engineer Intern @ Amazon | Computer Science

Machine Learning and Data Science enthusiast.