Imperfect Inferences: A Practical Assessment

Aaron Rieke, Vincent Southerland, Dan Svirsky, Mingwei Hsu

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Measuring racial disparities is challenging, especially when demographic labels are unavailable. Recently, some researchers and advocates have argued that companies should infer race and other demographic factors to help them understand and address discrimination. Others have been more skeptical, emphasizing the inaccuracy of racial inferences, critiquing the conceptualization of demographic categories themselves, and arguing that the use of demographic data might encourage algorithmic tweaks where more radical interventions are needed. We conduct a novel empirical analysis that informs this debate, using a dataset of self-reported demographic information provided by users of the ride-hailing service Uber who consented to share this information for research purposes. As a threshold matter, we show how this data reflects the enduring power of racism in society. We find differences by race across a range of outcomes. For example, among self-reported African-American riders, we see racial differences on factors from iOS use to local pollution levels. We then turn to a practical assessment of racial inference methodologies and offer two key findings. First, every inference method we tested has significant errors, miscategorizing people relative to their self-reports (even as the self-reports themselves suffer from selection bias). Second, and most importantly, we found that the inference methods worked: they reliably confirmed directional racial disparities that we knew were reflected in our dataset. Our analysis also suggests that the choice of inference methods should be informed by the measurement task. For example, disparities that are geographic in nature might be best captured by inferences that rely on geography; discrimination based on a person's name might be best detected by inferences that rely on names. In conclusion, our analysis shows that common racial inference methods have real and practical utility in shedding light on aggregate, directional disparities, despite their imperfections. While the recent literature has identified notable challenges regarding the collection and use of this data, these challenges should not be seen as dispositive.

Original languageEnglish (US)
Title of host publicationProceedings of 2022 5th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2022
PublisherAssociation for Computing Machinery
Pages767-777
Number of pages11
ISBN (Electronic)9781450393522
DOIs
StatePublished - Jun 21 2022
Event5th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2022 - Virtual, Online, Korea, Republic of
Duration: Jun 21 2022Jun 24 2022

Publication series

NameACM International Conference Proceeding Series

Conference

Conference5th ACM Conference on Fairness, Accountability, and Transparency, FAccT 2022
Country/TerritoryKorea, Republic of
CityVirtual, Online
Period6/21/226/24/22

Keywords

  • civil rights
  • demographics
  • discrimination
  • fairness
  • inference
  • race

ASJC Scopus subject areas

  • Software
  • Human-Computer Interaction
  • Computer Vision and Pattern Recognition
  • Computer Networks and Communications

Fingerprint

Dive into the research topics of 'Imperfect Inferences: A Practical Assessment'. Together they form a unique fingerprint.

Cite this