Authors:
Yanchen Wang
1
;
Rex Bone
1
;
Will Fleisher
1
;
Carole Gresenz
1
;
Jean Mitchell
1
;
Wilbert van der Klaauw
2
;
Crystal Wang
2
and
Lisa Singh
1
Affiliations:
1
Georgetown University, Washington, DC, U.S.A.
;
2
Federal Reserve Bank New York, New York, NY, U.S.A.
Keyword(s):
Machine Learning Fairness, Disease Prediction, Multivariate Sensitive Attribute, Model Fine Tuning.
Abstract:
The role of artificial intelligence is growing in healthcare and disease prediction. Because of its potential impact and demographic disparities that have been identified in machine learning models for disease prediction, there are growing concerns about transparency, accountability and fairness of these predictive models. However, very little research has investigated methods for improving model fairness in disease prediction, particularly when the sensitive attribute is multivariate and when the distribution of sensitive attribute groups is highly skewed. In this work, we explore algorithmic fairness when predicting heart disease and Alzheimer’s Disease and Related Dementias (ADRD). We propose a fine tuning approach to improve model fairness that takes advantage of observations from the majority groups to build a pre-trained model and uses observations from each underrepresented subgroup to fine tune the pre-trained model, thereby incorporating additional specific knowledge about e
ach subgroup. We find that our fine tuning approach performs better than other algorithmic fairness fixing methods across all subgroups even if the subgroup distribution is very imbalanced and some subgroups are very small. This is an important step toward understanding approaches for improving fairness for healthcare and disease prediction.
(More)