Full Seminar Details
Angel Pavon Perez
Knowledge Media Institute
This event took place on Wednesday 12 October 2022 at 11:30
Demand for fairer machine learning models is rapidly growing due to their increasing use in many decision-making processes. Several methods have been developed to detect and mitigate the bias of these models. One common approach for addressing such bias is simply dropping the sensitive attribute from the training data (e.g. gender). Such an approach is limited by focusing on the fairness of the training process rather than on the output model. This kind of simplistic approaches overlook the fact that sensitive attributes can be indirectly represented by other attributes in the data (e.g. maternity leave taken). However, there is currently little research aiming at understanding how covariance in data can contribute to the propagation of bias in machine learning models. In this seminar, we use feature selection techniques and statistical tests to study the covariance of these attributes and show how this covariance can help explain model bias in credit risk data. We further demonstrate how fairness can be significantly improved by eliminating the related attributes and the subsequent impact on model accuracy.