[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3461702.3462594acmconferencesArticle/Chapter ViewAbstractPublication PagesaiesConference Proceedingsconference-collections
research-article
Open access

A Step Toward More Inclusive People Annotations for Fairness

Published: 30 July 2021 Publication History

Abstract

The Open Images Dataset contains approximately 9 million images and is a widely accepted dataset for computer vision research. As is common practice for large datasets, the annotations are not exhaustive, with bounding boxes and attribute labels for only a subset of the classes in each image. In this paper, we present a new set of annotations on a subset of the Open Images dataset called the MIAP (More Inclusive Annotations for People) subset, containing bounding boxes and attributes for all of the people visible in those images. The attributes and labeling methodology for the MIAP subset were designed to enable research into model fairness. In addition, we analyze the original annotation methodology for the person class and its subclasses, discussing the resulting patterns in order to inform future annotation efforts. By considering both the original and exhaustive annotation sets, researchers can also now study how systematic patterns in training annotations affect modeling.

References

[1]
M. Alvi, A. Zisserman, and C. Nellaker. 2018. Turning a Blind Eye: Explicit Removal of Biases and Variation from Deep Neural Network Embeddings. In ECCV Workshop on Bias Estimation in Face Analytics .
[2]
James Atwood, Yoni Halpern, Pallavi Baljekar, Eric Breck, D. Sculley, Pavel Ostyakov, Sergey I. Nikolenko, Igor Ivanov, Roman Solovyev, Weimin Wang, and Miha Skalic. 2020. The Inclusive Images Competition. In The NeurIPS '18 Competition, Sergio Escalera and Ralf Herbrich (Eds.). Springer International Publishing, Cham, 155--186.
[3]
Alex Beutel, Jilin Chen, Tulsee Doshi, Hai Qian, Allison Woodruff, Christine Luu, Pierre Kreitmann, Jonathan Bischof, and Ed H Chi. 2019. Putting fairness principles into practice: Challenges, metrics, and improvements. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society. 453--459.
[4]
Gerlof Bouma. 2009. Normalized (pointwise) mutual information in collocation extraction. Proceedings of GSCL (2009), 31--40.
[5]
Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Fairness, Accountability, and Transparency (FAT*). PMLR, 77--91.
[6]
Terrance de Vries, Ishan Misra, Changhan Wang, and Laurens van der Maaten. 2019. Does Object Recognition Work for Everyone?. In CVPR Workshop on Fairness Accountability Transparency and Ethics in Computer Vision (FATE CV) .
[7]
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet: A large-scale hierarchical image database. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Ieee, 248--255.
[8]
DoD. 2018. 2018 Demographics Report: Profile of the Military Community. Technical Report. Department of Defence, Office of the Deputy Assistant Secretary of Defense for Military Community and Family Policy (ODASD (MC&FP)).
[9]
Chris Dulhanty and Alexander Wong. 2019. Auditing ImageNet: Towards a Model-driven Framework for Annotating Demographic Attributes of Large-Scale Image Datasets. In CVPR Workshop on Fairness Accountability Transparency and Ethics in Computer Vision (FATE CV) .
[10]
Timnit Gebru, Jamie Morgenstern, Briana Vecchione, Jennifer Wortman Vaughan, Hanna Wallach, Hal Daumé III, and Kate Crawford. 2018. Datasheets for datasets. arXiv preprint arXiv:1803.09010 (2018).
[11]
Patrick Grother, Mei Ngan, and Kayee Hanaoka. 2018. Ongoing Face Recognition Vendor Test (FRVT) Part 1: Verification. National Institute of Standards and Technology (2018).
[12]
Foad Hamidi, Morgan Klaus Scheuerman, and Stacy M. Branham. 2018. Gender Recognition or Gender Reductionism? The Social Implications of Embedded Gender Recognition Systems. In CHI Conference on Human Factors in Computing Systems (Montreal QC, Canada). 1--13. https://doi.org/10.1145/3173574.3173582
[13]
Moritz Hardt, Eric Price, Eric Price, and Nati Srebro. 2016. Equality of Opportunity in Supervised Learning. In Advances in Neural Information Processing Systems, D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett (Eds.), Vol. 29. Curran Associates, Inc., 3315--3323. https://proceedings.neurips.cc/paper/2016/file/9d2682367c3935defcb1f9e247a97c0d-Paper.pdf
[14]
Kimmo Karkkainen and Jungseock Joo. 2021. FairFace: Face Attribute Dataset for Balanced Race, Gender, and Age for Bias Measurement and Mitigation. In IEEE/CVF Winter Conference on Applications of Computer Vision (WACV). 1548--1558.
[15]
Os Keyes. 2018. The Misgendering Machines: Trans/HCI Implications of Automatic Gender Recognition. Proceedings of the ACM on Human-Computer Interaction, Article 88 (Nov. 2018), 22 pages. https://doi.org/10.1145/3274357
[16]
Alina Kuznetsova, Hassan Rom, Neil Alldrin, Jasper Uijlings, Ivan Krasin, Jordi Pont-Tuset, Shahab Kamali, Stefan Popov, Matteo Malloci, Alexander Kolesnikov, et al. 2020. The open images dataset v4: Unified image classification, object detection, and visual relationship detection at scale. International Journal of Computer Vision (IJCV) (2020), 1--26.
[17]
Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. 2014. Microsoft coco: Common objects in context. In European Conference on Computer Vision (ECCV). Springer, 740--755.
[18]
Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. 2015. Deep Learning Face Attributes in the Wild. In International Conference on Computer Vision (ICCV).
[19]
David Madras, Elliot Creager, Toniann Pitassi, and Richard Zemel. 2018. Learning adversarially fair and transferable representations. In International Conference on Machine Learning. PMLR, 3384--3393.
[20]
Michele Merler, Nalini Ratha, Rogerio S Feris, and John R Smith. 2019. Diversity in faces. arXiv preprint arXiv:1901.10436 (2019).
[21]
Margaret Mitchell, Simone Wu, Andrew Zaldivar, Parker Barnes, Lucy Vasserman, Ben Hutchinson, Elena Spitzer, Inioluwa Deborah Raji, and Timnit Gebru. 2019. Model Cards for Model Reporting. In Fairness, Accountability, and Transparency (FAT*). 220--229. https://doi.org/10.1145/3287560.3287596
[22]
Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. 2015. ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), Vol. 115, 3 (2015), 211--252. https://doi.org/10.1007/s11263-015-0816-y
[23]
Hee Jung Ryu, Hartwig Adam, and Margaret Mitchell. 2017. Inclusivefacenet: Improving face attribute detection with race and gender diversity. arXiv preprint arXiv:1712.00193 (2017).
[24]
Morgan Klaus Scheuerman, Katta Speil, Oliver L. Haimson, Foad Hamidi, and Stacy M. Branham. 2020. HCI Guidelines for Gender Equity and Inclusivity. http://morgan-klaus.com/gender-guidelines.html. Version 1.1.
[25]
Shreya Shankar, Yoni Halpern, Eric Breck, James Atwood, Jimbo Wilson, and D. Sculley. 2017. No Classification without Representation: Assessing Geodiversity Issues in Open Data Sets for the Developing World. In NeurIPS Workshop: Machine Learning for the Developing World.
[26]
Craig I Watson. 2016. NIST Special Database 18. NIST Mugshot Identification Database (MID).
[27]
Wikipedia contributors. 2020. Boy -- Wikipedia, The Free Encyclopedia. https://en.wikipedia.org/w/index.php?title=Boy&oldid=992668770. [Online; accessed 29-January-2021].
[28]
Wikipedia contributors. 2021. Girl -- Wikipedia, The Free Encyclopedia. https://en.wikipedia.org/w/index.php?title=Girl&oldid=1000129400. [Online; accessed 29-January-2021].
[29]
Benjamin Wilson, Judy Hoffman, and Jamie Morgenstern. 2019. Predictive Inequity in Object Detection. In CVPR Workshop on Fairness Accountability Transparency and Ethics in Computer Vision (FATE CV).
[30]
Zhe Wu, Navaneeth Bodla, Bharat Singh, Mahyar Najibi, Rama Chellappa, and Larry S Davis. 2018. Soft sampling for robust object detection. arXiv preprint arXiv:1806.06986 (2018).
[31]
Kaiyu Yang, Klint Qinami, Li Fei-Fei, Jia Deng, and Olga Russakovsky. 2020. Towards Fairer Datasets: Filtering and Balancing the Distribution of the People Subtree in the ImageNet Hierarchy. In Fairness, Accountability, and Transparency (FAT*).
[32]
Zhifei Zhang, Yang Song, and Hairong Qi. 2017. Age Progression/Regression by Conditional Adversarial Autoencoder. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE.

Cited By

View all
  • (2024)VideoPoetProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3693075(25105-25124)Online publication date: 21-Jul-2024
  • (2024)Attribute annotation and bias evaluation in visual datasets for autonomous drivingJournal of Big Data10.1186/s40537-024-00976-911:1Online publication date: 27-Sep-2024
  • (2024)Auditing Gender Presentation Differences in Text-to-Image ModelsProceedings of the 4th ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization10.1145/3689904.3694710(1-10)Online publication date: 29-Oct-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
AIES '21: Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society
July 2021
1077 pages
ISBN:9781450384735
DOI:10.1145/3461702
This work is licensed under a Creative Commons Attribution International 4.0 License.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 30 July 2021

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. computer vision
  2. datasets
  3. fairness

Qualifiers

  • Research-article

Conference

AIES '21
Sponsor:

Acceptance Rates

Overall Acceptance Rate 61 of 162 submissions, 38%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)218
  • Downloads (Last 6 weeks)27
Reflects downloads up to 09 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)VideoPoetProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3693075(25105-25124)Online publication date: 21-Jul-2024
  • (2024)Attribute annotation and bias evaluation in visual datasets for autonomous drivingJournal of Big Data10.1186/s40537-024-00976-911:1Online publication date: 27-Sep-2024
  • (2024)Auditing Gender Presentation Differences in Text-to-Image ModelsProceedings of the 4th ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization10.1145/3689904.3694710(1-10)Online publication date: 29-Oct-2024
  • (2024)Generalized People Diversity: Learning a Human Perception-Aligned Diversity Representation for People ImagesProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658940(797-821)Online publication date: 3-Jun-2024
  • (2024)On Scaling Up a Multilingual Vision and Language Model2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)10.1109/CVPR52733.2024.01368(14432-14444)Online publication date: 16-Jun-2024
  • (2024)Adversarial attacks and defenses in explainable artificial intelligenceInformation Fusion10.1016/j.inffus.2024.102303107:COnline publication date: 2-Jul-2024
  • (2024)For a semiotic AI: Bridging computer vision and visual semiotics for computational observation of large scale facial image archivesComputer Vision and Image Understanding10.1016/j.cviu.2024.104187249(104187)Online publication date: Dec-2024
  • (2023)Ethical considerations for responsible data curationProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3668537(55320-55360)Online publication date: 10-Dec-2023
  • (2023)Ground Truth Or Dare: Factors Affecting The Creation Of Medical Datasets For Training AIProceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society10.1145/3600211.3604766(351-362)Online publication date: 8-Aug-2023
  • (2023)Evaluating the Fairness of Discriminative Foundation Models in Computer VisionProceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society10.1145/3600211.3604720(809-833)Online publication date: 8-Aug-2023
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media