CheXclusion: Fairness Gaps in Deep Chest X-ray Classifiers

Laleh Seyyed-Kalantari1,2,*, Guanxiong Liu1,2, Matthew McDermott3, Irene Y. Chen3, Marzyeh Ghassemi1,2


1Computer Science, University of Toronto
2Vector Institute
3Electrical Engineering and Computer Science, Massachusetts Institute of Technology
*Corresponding author
Email: laleh@cs.toronto.edu

Pacific Symposium on Biocomputing 26:232-243(2021)

© 2021 World Scientific
Open Access chapter published by World Scientific Publishing Company and distributed under the terms of the Creative Commons Attribution (CC BY) 4.0 License.


Abstract

Machine learning systems have received much attention recently for their ability to achieve expert-level performance on clinical tasks, particularly in medical imaging. Here, we examine the extent to which state-of-the-art deep learning classifiers trained to yield diagnostic labels from X-ray images are biased with respect to protected attributes. We train convolution neural networks to predict 14 diagnostic labels in 3 prominent public chest X-ray datasets: MIMIC-CXR, Chest-Xray8, CheXpert, as well as a multi-site aggregation of all those datasets. We evaluate the TPR disparity — the difference in true positive rates (TPR) — among different protected attributes such as patient sex, age, race, and insurance type as a proxy for socioeconomic status. We demonstrate that TPR disparities exist in the state-of-the-art classifiers in all datasets, for all clinical tasks, and all subgroups. A multi-source dataset corresponds to the smallest disparities, suggesting one way to reduce bias. We find that TPR disparities are not significantly correlated with a subgroup's proportional disease burden. As clinical models move from papers to products, we encourage clinical decision makers to carefully audit for algorithmic disparities prior to deployment. Our supplementary materials can be found at, http://www.marzyehghassemi.com/chexclusion-supp-3/.


[Full-Text PDF] [PSB Home Page]