Fair human-centric image dataset for ethical AI benchmarking
- PMID: 41193813
- PMCID: PMC12675298
- DOI: 10.1038/s41586-025-09716-2
Fair human-centric image dataset for ethical AI benchmarking
Abstract
Computer vision is central to many artificial intelligence (AI) applications, from autonomous vehicles to consumer devices. However, the data behind such technical innovations are often collected with insufficient consideration of ethical concerns1-3. This has led to a reliance on datasets that lack diversity, perpetuate biases and are collected without the consent of data rights holders. These datasets compromise the fairness and accuracy of AI models and disenfranchise stakeholders4-8. Although awareness of the problems of bias in computer vision technologies, particularly facial recognition, has become widespread9, the field lacks publicly available, consensually collected datasets for evaluating bias for most tasks3,10,11. In response, we introduce the Fair Human-Centric Image Benchmark (FHIBE, pronounced 'Feebee'), a publicly available human image dataset implementing best practices for consent, privacy, compensation, safety, diversity and utility. FHIBE can be used responsibly as a fairness evaluation dataset for many human-centric computer vision tasks, including pose estimation, person segmentation, face detection and verification, and visual question answering. By leveraging comprehensive annotations capturing demographic and physical attributes, environmental factors, instrument and pixel-level annotations, FHIBE can identify a wide variety of biases. The annotations also enable more nuanced and granular bias diagnoses, enabling practitioners to better understand sources of bias and mitigate potential downstream harms. FHIBE therefore represents an important step forward towards trustworthy AI, raising the bar for fairness benchmarks and providing a road map for responsible data curation in AI.
© 2025. The Author(s).
Conflict of interest statement
Competing interests: Sony Group Corporation, with inventors J.T.A.A. and A.X., has a pending US patent application US20240078839A1, filed on 14 August 2023, that is currently under examination. It covers aspects of the human-centric image dataset specification and annotation techniques that were used in this paper. The same application has also been filed in Europe (application number 23761605.7, filed on 15 January 2025) and China (application number 202380024486.X, filed on 30 August 2024) and the applications are pending.
Figures
References
-
- Sambasivan, N. et al. “Everyone wants to do the model work, not the data work”: data cascades in high-stakes AI. In Proc. ACM CHI Conference on Human Factors in Computing Systems (ACM, 2021).
-
- Birhane, A. & Prabhu, V. U. Large image datasets: a pyrrhic win for computer vision? In Proc. IEEE Winter Conference on Applications of Computer Vision (WACV) 1536–1546 (IEEE, 2021).
-
- Andrews, J. T. et al. Ethical considerations for collecting human-centric image datasets. In Proc. Advances in Neural Information Processing Systems Datasets and Benchmarks Track (NeurIPS D&B) 55320–55360 (Curran Associates, 2023).
-
- Hundt, A., Agnew, W., Zeng, V., Kacianka, S. & Gombolay, M. Robots enact malignant stereotypes. In Proc. ACM Conference on Fairness, Accountability, and Transparency (FAccT) 743–756 (ACM, 2022).
-
- Wilson, B., Hoffman, J. & Morgenstern, J. Predictive inequity in object detection. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (IEEE, 2019).
MeSH terms
LinkOut - more resources
Full Text Sources
