Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Jan:2021:2512-2522.
doi: 10.1109/wacv48630.2021.00256. Epub 2021 Jun 14.

Representation Learning with Statistical Independence to Mitigate Bias

Affiliations

Representation Learning with Statistical Independence to Mitigate Bias

Ehsan Adeli et al. IEEE Winter Conf Appl Comput Vis. 2021 Jan.

Abstract

Presence of bias (in datasets or tasks) is inarguably one of the most critical challenges in machine learning applications that has alluded to pivotal debates in recent years. Such challenges range from spurious associations between variables in medical studies to the bias of race in gender or face recognition systems. Controlling for all types of biases in the dataset curation stage is cumbersome and sometimes impossible. The alternative is to use the available data and build models incorporating fair representation learning. In this paper, we propose such a model based on adversarial training with two competing objectives to learn features that have (1) maximum discriminative power with respect to the task and (2) minimal statistical mean dependence with the protected (bias) variable(s). Our approach does so by incorporating a new adversarial loss function that encourages a vanished correlation between the bias and the learned features. We apply our method to synthetic data, medical images (containing task bias), and a dataset for gender classification (containing dataset bias). Our results show that the learned features by our method not only result in superior prediction performance but also are unbiased.

PubMed Disclaimer

Figures

Figure 1:
Figure 1:
Average face images for each shade category (1st row), average saliency map of the trained baseline (2nd row), and BR-Net (3rd row) color-coded with the normalized saliency for each pixel. BR-Net results in more stable patterns across all 6 shades. The last column shows the tSNE projection of the learned representations by each method. Our method results in a better representation space invariant to the bias variable (shade) while the baseline shows a pattern influenced by the bias. Average accuracy of per-shade gender classification over 5 runs of 5-fold cross-validation (pre-trained on ImageNet, fine-tuned on GS-PPB) is shown on each average map. BR-Net not only obtains better accuracy for the darker shade but also regularizes the model to improve overall per-category accuracy.
Figure 2:
Figure 2:
BR-Net architecture: FE learns features, F, that successfully classify (C) the input while being invariant (statistically independent) to the protected variables, b, using BP and the adversarial loss, −λLbp (based on correlation coefficient). Forward arrows show forward paths while the backward dashed ones indicate back-propagation with the respective gradient () values.
Figure 3:
Figure 3:
BR-Net can remove direct dependency between F and b for both dataset or task bias.
Figure 4:
Figure 4:
Formation of synthetic dataset (a) and comparison of results for different methods (b).
Figure 5:
Figure 5:
tSNE projection of the learned features for different methods. Color indicates the value of σB.
Figure 6:
Figure 6:
Statistical dependence between the learned features and age for the CTRL cohort in the HIV experiment, which is quantitatively measured by dcor2.
Figure 7:
Figure 7:
Accuracy, TNR, and TPR of the HIV experiment, as a function of the # of iterations for (a) 3D CNN baseline, (b) BR-Net. Our method is robust against the imbalanced age distribution between HIV and CTRL.
Figure 8:
Figure 8:
Accuracy of gender prediction from face images across all shades (1 to 6) of the GS-PPB dataset with two backbones, (left) VGG16 and (right) ResNet50. BR-Net consistently results in more accurate predictions in all 6 shade categories.
Figure 9:
Figure 9:
Learned representations by different methods. Color encodes the 6 categories of skin shade.

Similar articles

Cited by

References

    1. Adeli Ehsan, Kwon Dongjin, Zhao Qingyu, Pfefferbaum Adolf, Zahr Natalie M, Sullivan Edith V, and Pohl Kilian M. Chained regularization for identifying brain patterns specific to HIV infection. NeuroImage, 183:425–437, 2018. - PMC - PubMed
    1. Akuzawa Kei, Iwasawa Yusuke, and Matsuo Yutaka. Adversarial invariant feature learning with accuracy constraint for domain generalization. arXiv preprint arXiv:1904.12543, 2019.
    1. Barocas Solon, Hardt Moritz, and Narayanan Arvind. Fairness in machine learning. Neural Information Processing Systems Tutorial, 2017.
    1. Barocas Solon and Selbst Andrew D. Big data’s disparate impact. Calif L. Rev, 104:671, 2016.
    1. Baron Reuben M and Kenny David A. The moderator–mediator variable distinction in social psychological research: Conceptual, strategic, and statistical considerations. Journal of personality and social psychology, 51(6):1173, 1986. - PubMed

LinkOut - more resources