[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3630106.3658977acmotherconferencesArticle/Chapter ViewAbstractPublication PagesfacctConference Proceedingsconference-collections
research-article

The Impact of Differential Feature Under-reporting on Algorithmic Fairness

Published: 05 June 2024 Publication History

Abstract

Predictive risk models in the public sector are commonly developed using administrative data that is more complete for subpopulations that more greatly rely on public services. In the United States, for instance, information on health care utilization is routinely available to government agencies for individuals supported by Medicaid and Medicare, but not for the privately insured. Critiques of public sector algorithms have identified such “differential feature under-reporting” as a driver of disparities in algorithmic decision-making. Yet this form of data bias remains understudied from a technical viewpoint. While prior work has examined the fairness impacts of additive feature noise and features that are clearly marked as missing, little is known about the setting of data missingness absent indicators (i.e. differential feature under-reporting). In this work, we study an analytically tractable model of differential feature under-reporting to characterizethe impact of under-report on algorithmic fairness. We demonstrate how standard missing data methods typically fail to mitigate bias in this setting, and propose a new set of augmented loss and imputation methods. Our results show that, in real world data settings, under-reporting typically exacerbates disparities. The proposed solution methods show some success in mitigating disparities attributable to feature under-reporting.

References

[1]
Roy Adams, Yuelong Ji, Xiaobin Wang, and Suchi Saria. 2019. Learning Models from Data with Measurement Error: Tackling Underreporting. In International Conference on Machine Learning (IMCL ’20).
[2]
Muhammad Aurangzeb Ahmad, Carly Eckert, and Ankur Teredesai. 2019. The Challenge of Imputation in Explainable Artificial Intelligence Models. arXiv preprint, arXiv:1907.12669 (2019).
[3]
Dennis J. Aigner and Glen G. Cain. 1977. Statistical Theories of Discrimination in Labor Markets. Industrial and Labor Relations Review 30, 2 (Jan. 1977), 175.
[4]
Michelle Alexander. 2010. The new Jim Crow: Mass Incarceration in the Age of Colorblindness. New Press, New York, NY.
[5]
J D Angrist and Jorn-Steffen Pischke. 2008. Mostly harmless econometrics. Princeton University Press, Princeton, NJ.
[6]
Larson J. Mattu S. Angwin, J. and L. Kirchner. 2016. Machine Bias. There’s software used across the country to predict future criminals. And it’s biased against blacks.ProPublica (2016). https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
[7]
Nicholas C Arpey, Anne H Gaglioti, and Marcy E Rosenbaum. 2017. How socioeconomic status affects patient perceptions of health care: A qualitative study. J. Prim. Care Community Health 8, 3 (July 2017), 169–175.
[8]
T.C. Bailey, M.S. Carvalho, T.M. Lapa, W.V. Souza, and M.J. Brewer. 2005. Modeling of Under-detection of Cases in Disease Surveillance. Annals of Epidemiology 15, 5 (May 2005), 335–343.
[9]
Geoffrey C Barnes and Jordan M Hyatt. 2012. Classifying adult probationers by forecasting future offending. National Institute of Justice. Retrieved February 4 (2012), 2020.
[10]
Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, and Aaron Roth. 2018. Fairness in Criminal Justice Risk Assessments: The State of the Art. Sociological Methods & Research 50, 1 (July 2018), 3–44.
[11]
Emily Black, Hadi Elzayn, Alexandra Chouldechova, Jacob Goldin, and Daniel Ho. 2022. Algorithmic fairness and vertical equity: Income fairness with IRS tax audit models. In 2022 ACM Conference on Fairness, Accountability, and Transparency.
[12]
Hermann Brenner and Dana Loomis. 1994. Varied Forms of Bias Due to Nondifferential Error in Measuring Exposure. Epidemiology 5, 5 (1994), 510–517.
[13]
Irwin Bross. 1954. Misclassification in 2 X 2 Tables. Biometrics 10, 4 (Dec. 1954), 478.
[14]
Bradley Butcher, Chris Robinson, Miri Zilka, Riccardo Fogliato, Carolyn Ashurst, and Adrian Weller. 2022. Racial Disparities in the Enforcement of Marijuana Violations in the US. In Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society(AIES ’22). 130–143.
[15]
Danton S. Char, Nigam H. Shah, and David Magnus. 2018. Implementing Machine Learning in Health Care — Addressing Ethical Challenges. New England Journal of Medicine 378, 11 (March 2018), 981–983.
[16]
Irene Y. Chen, Fredrik D. Johansson, and David Sontag. 2018. Why is My Classifier Discriminatory?. In 32nd International Conference on Neural Information Processing Systems(NIPS’18).
[17]
Alexandra Chouldechova, Diana Benavides-Prado, Oleksandr Fialko, and Rhema Vaithianathan. 2018. A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. In Conference on Fairness, Accountability and Transparency (FAT*).
[18]
Haitao Chu, Zhaojie Wang, Stephen R. Cole, and Sander Greenland. 2006. Sensitivity Analysis of Misclassification: A Graphical and a Bayesian Approach. Annals of Epidemiology 16, 11 (2006), 834–841.
[19]
Federico Cismondi, André S. Fialho, Susana M. Vieira, Shane R. Reti, João M.C. Sousa, and Stan N. Finkelstein. 2013. Missing data in medical databases: Impute, delete or classify?Artificial Intelligence in Medicine 58, 1 (May 2013), 63–72.
[20]
Guilherme Lopes de Oliveira, Rosangela Helena Loschi, and Renato Martins Assunção. 2017. A random-censoring Poisson model for underreported data. Statistics in Medicine 36, 30 (Oct. 2017), 4873–4892.
[21]
Frances Ding, Moritz Hardt, John Miller, and Ludwig Schmidt. 2021. Retiring Adult: New Datasets for Fair Machine Learning. In Advances in Neural Information Processing Systems (Neurips ’21).
[22]
Mustafa Dosemeci, Sholom Wacholder, and Jay H. Lubin. 1990. Does Nondifferential Misclassification of Exposure Always Bias a True Effect Toward the Null Value?American Journal of Epidemiology 132, 4 (1990), 746–748.
[23]
Jessie K Edwards, Stephen R Cole, Melissa A Troester, and David B Richardson. 2013. Accounting for misclassified outcomes in binary regression models using multiple imputation with internal validation data. Am. J. Epidemiol. 177, 9 (May 2013), 904–912.
[24]
Charles Elkan and Keith Noto. 2008. Learning classifiers from only positive and unlabeled data. In Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM.
[25]
Zeynep Engin and Philip Treleaven. 2019. Algorithmic government: Automating public services and supporting civil servants in using data science technologies. Comput. J. 62, 3 (2019), 448–460.
[26]
Virginia Eubanks. 2018. A Child Abuse Prediction Model Fails Poor Families. https://www.wired.com/story/excerpt-from-automating-inequality/
[27]
Martínez-Plumed Fernando, Ferri Cèsar, Nieves David, and Hernández-Orallo José. 2021. Missing the missing values: The ugly duckling of fairness in machine learning. International Journal of Intelligent Systems 36, 7 (March 2021), 3217–3258.
[28]
Riccardo Fogliato, Alice Xiang, Zachary Lipton, Daniel Nagin, and Alexandra Chouldechova. 2021. On the Validity of Arrest as a Proxy for Offense: Race and the Likelihood of Arrest for Violent Crimes. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society(AIES ’21). 100–111.
[29]
Centre for Social Data Analytics at the Auckland University of Technology. 2020. Implementing the Hello Baby Prevention Program in Allegheny County. (2020). https://www.alleghenycountyanalytics.us/wp-content/uploads/2020/12/Hello-Baby-Methodology-v6.pdf
[30]
Christian Fricke. 2020. Missing Fairness: The Discriminatory Effect of Missing Values inDatasets on Fairness in Machine Learning. Master thesis (2020).
[31]
Wayne A Fuller. 1987. Measurement Error Models. John Wiley & Sons, Nashville, TN.
[32]
Andrew Gelman, John B Carlin, Hal S Stern, David B Dunson, Aki Vehtari, and Donald B Rubin. 2013. Bayesian Data Analysis (3 ed.). Chapman & Hall/CRC, Philadelphia, PA.
[33]
Milena A. Gianfrancesco, Suzanne Tamang, Jinoos Yazdany, and Gabriela Schmajuk. 2018. Potential Biases in Machine Learning Algorithms Using Electronic Health Record Data. JAMA Internal Medicine 178, 11 (Nov. 2018), 1544.
[34]
Sander Greenland. 2014. Sensitivity Analysis and Bias Analysis. In Handbook of Epidemiology. Springer New York, 685–706.
[35]
Rolf H H Groenwold and Olaf M Dekkers. 2020. Missing data: the impact of what is not there. European Journal of Endocrinology 183, 4 (Oct. 2020), E7–E9.
[36]
Jerry Hausman. 2001. Mismeasured Variables in Econometric Analysis: Problems from the Right and Problems from the Left. Journal of Economic Perspectives 15, 4 (2001), 57–67.
[37]
Kimberly A Houser and Debra Sanders. 2016. The use of big data analytics by the IRS: Efficient solutions or the end of privacy as we know it. Vand. J. Ent. & Tech. L. 19 (2016), 817.
[38]
Vincent Jeanselme, Maria De-Arteaga, Zhe Zhang, Jessica Barrett, and Brian Tom. 2022. Imputation Strategies Under Clinical Presence: Impact on Algorithmic Fairness. Machine Learning for Health (ML4H ’22).
[39]
Fereshte Khani and Percy Liang. 2020. Feature Noise Induces Loss Discrepancy Across Groups. In International Conference on Machine Learning (ICML ’20).
[40]
G. King, J. Honaker, A. Joseph, and K. Scheve. 2001. Analyzing incomplete political science data: an alternative algorithm for multipleimputation. Am. Polit. Sci. Rev.95 (2001), 49–69.
[41]
Chamari I Kithulgoda, Rhema Vaithianathan, and Dennis P Culhane. 2022. Predictive risk modeling to identify homeless clients at risk for prioritizing services using routinely collected data. Journal of Technology in Human Services 40, 2 (2022), 134–156.
[42]
Karen Levy, Kyla E Chasalow, and Sarah Riley. 2021. Algorithms and decision-making in the public sector. Annual Review of Law and Social Science 17 (2021), 309–334.
[43]
Sandra G Mayson. 2019. Bias in, bias out. The Yale Law Journal 128, 8 (2019), 2218–2300.
[44]
John F McCarthy, Robert M Bossarte, Ira R Katz, Caitlin Thompson, Janet Kemp, Claire M Hannemann, Christopher Nielson, and Michael Schoenbaum. 2015. Predictive modeling and concentration of the risk of suicide: implications for preventive interventions in the US Department of Veterans Affairs. American journal of public health 105, 9 (2015), 1935–1942.
[45]
P.E. McKnight, K.M. McKnight, S. Sidani, and A. J. Figueredo. 2007. Missing Data: A Gentle Introduction.
[46]
Nagarajan Natarajan, Inderjit S Dhillon, Pradeep K Ravikumar, and Ambuj Tewari. 2013. Learning with Noisy Labels. In Advances in Neural Information Processing Systems, C.J. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K.Q. Weinberger (Eds.). Vol. 26. Curran Associates, Inc.
[47]
A. Nemirovski, A. Juditsky, G. Lan, and A. Shapiro. 2009. Robust Stochastic Approximation Approach to Stochastic Programming. SIAM Journal on Optimization 19, 4 (2009), 1574–1609.
[48]
Edmund Phelps. 1972. The Statistical Theory of Racism and Sexism. American Economic Review 62, 4 (1972), 659–61.
[49]
Emma Pierson, Camelia Simoiu, Jan Overgoor, Sam Corbett-Davies, Daniel Jenson, Amy Shoemaker, Vignesh Ramachandran, Phoebe Barghouty, Cheryl Phillips, Ravi Shroff, and Sharad Goel. 2020. A large-scale analysis of racial disparities in police stops across the United States. Nat. Hum. Behav. 4, 7 (July 2020), 736–745.
[50]
Dewi Rahardja and Dean M. Young. 2021. Confidence Intervals for the Risk Ratio Using Double Sampling with Misclassified Binomial Data. Journal of Data Science 9, 4 (2021), 529–548.
[51]
Alvin Rajkomar, Michaela Hardt, Michael D. Howell, Greg Corrado, and Marshall H. Chin. 2018. Ensuring Fairness in Machine Learning to Advance Health Equity. Annals of Internal Medicine 169, 12 (Dec. 2018), 866.
[52]
UCI Machine Learning Repository. 1994. German Credit data. (1994). https://archive.ics.uci.edu/ml/datasets/statlog+(german+credit+data)
[53]
Rashida Richardson, Jason M Schultz, and Kate Crawford. 2019. Dirty data, bad predictions: How civil rights violations impact police data, predictive policing systems, and justice. NYUL Rev. Online 94 (2019), 15.
[54]
Donald B. Rubin. 1976. Inference and missing data. Biometrika 63, 3 (1976), 581–592.
[55]
Konstantinos Sechidis, Matthew Sperrin, Emily S. Petherick, Mikel Luján, and Gavin Brown. 2017. Dealing with under-reported variables: An information theoretic solution. International Journal of Approximate Reasoning 85 (2017), 159–177.
[56]
Oliver Stoner, Theo Economou, and Gabriela Drummond Marques da Silva. 2019. A Hierarchical Framework for Correcting Under-Reporting in Count Data. J. Amer. Statist. Assoc. 114, 528 (apr 2019), 1481–1492.
[57]
Rhema Vaithianathan, Emily Putnam-Hornstein, Nan Jiang, Parma Nand, and Tim Maloney. 2017. Developing predictive models to support child maltreatment hotline screening decisions: Allegheny County methodology and implementation. Center for Social data Analytics (2017).
[58]
Marvin Van Bekkum and Frederik Zuiderveen Borgesius. 2021. Digital welfare fraud detection and the Dutch SyRI judgment. European Journal of Social Security 23, 4 (2021), 323–340.
[59]
Yanchen Wang and Lisa Singh. 2021. Analyzing the impact of missing values and selection bias on fairness. International Journal of Data Science and Analytics 12, 2 (May 2021), 101–119.
[60]
Kevin Wu, Dominik Dahlem, Christopher Hane, Eran Halperin, and James Zou. 2023. Collecting data when missingness is unknown: a method for improving model performance given under-reporting in patient populations. In Proceedings of the Conference on Health, Inference, and Learning, Vol. 209. PMLR, 229–242.
[61]
Yiliang Zhang and Qi Long. 2021. Assessing Fairness in the Presence of Missing Data. In Advances in Neural Information Processing Systems (Neurips ’21).
[62]
Helen Zhou, Balakrishnan Sivaraman, and Zachary C. Lipton. 2022. Domain Adaptation Under Missingness Shift. arXiv preprint, arXiv:2211.02093 (2022).

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Other conferences
FAccT '24: Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency
June 2024
2580 pages
ISBN:9798400704505
DOI:10.1145/3630106
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 05 June 2024

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

FAccT '24

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 102
    Total Downloads
  • Downloads (Last 12 months)102
  • Downloads (Last 6 weeks)39
Reflects downloads up to 02 Mar 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media