TY - GEN
T1 - Resume Format, LinkedIn URLs and Other Unexpected Influences on AI Personality Prediction in Hiring
T2 - 5th AAAI/ACM Conference on Artificial Intelligence, Ethics, and Society, AIES 2022
AU - Rhea, Alene
AU - Markey, Kelsey
AU - D'Arinzo, Lauren
AU - Schellmann, Hilke
AU - Sloane, Mona
AU - Squires, Paul
AU - Stoyanovich, Julia
N1 - Funding Information:
This research was supported in part by NSF Awards No. 1934464, 1916505, and 1922658. We thank Dhara Mungra for her work on data collection and preliminary analysis, Daphna Harel and Joshua R. Loftus for their advice on statistical methods, and Falaah Arif Khan for her work on the generalization of the auditing framework.
Publisher Copyright:
© 2022 ACM.
PY - 2022/7/26
Y1 - 2022/7/26
N2 - Automated hiring systems are among the fastest-developing of all high-stakes AI systems. Among these are algorithmic personality tests that use insights from psychometric testing, and promise to surface personality traits indicative of future success based on job seekers' resumes or social media profiles. We interrogate the reliability of such systems using stability of the outputs they produce, noting that reliability is a necessary, but not a sufficient, condition for validity. We develop a methodology for an external audit of stability of algorithmic personality tests, and instantiate this methodology in an audit of two systems, Humantic AI and Crystal. Rather than challenging or affirming the assumptions made in psychometric testing-that personality traits are meaningful and measurable constructs, and that they are indicative of future success on the job-we frame our methodology around testing the underlying assumptions made by the vendors of the algorithmic personality tests themselves. In our audit of Humantic AI and Crystal, we find that both systems show substantial instability on key facets of measurement, and so cannot be considered valid testing instruments. For example, Crystal frequently computes different personality scores if the same resume is given in PDF vs. in raw text, violating the assumption that the output of an algorithmic personality test is stable across job-irrelevant input variations. Among other notable findings is evidence of persistent-and often incorrect-data linkage by Humantic AI. An open-source implementation of our auditing methodology, and of the audits of Humantic AI and Crystal, is available at https://github.com/DataResponsibly/hiring-stability-Audit.
AB - Automated hiring systems are among the fastest-developing of all high-stakes AI systems. Among these are algorithmic personality tests that use insights from psychometric testing, and promise to surface personality traits indicative of future success based on job seekers' resumes or social media profiles. We interrogate the reliability of such systems using stability of the outputs they produce, noting that reliability is a necessary, but not a sufficient, condition for validity. We develop a methodology for an external audit of stability of algorithmic personality tests, and instantiate this methodology in an audit of two systems, Humantic AI and Crystal. Rather than challenging or affirming the assumptions made in psychometric testing-that personality traits are meaningful and measurable constructs, and that they are indicative of future success on the job-we frame our methodology around testing the underlying assumptions made by the vendors of the algorithmic personality tests themselves. In our audit of Humantic AI and Crystal, we find that both systems show substantial instability on key facets of measurement, and so cannot be considered valid testing instruments. For example, Crystal frequently computes different personality scores if the same resume is given in PDF vs. in raw text, violating the assumption that the output of an algorithmic personality test is stable across job-irrelevant input variations. Among other notable findings is evidence of persistent-and often incorrect-data linkage by Humantic AI. An open-source implementation of our auditing methodology, and of the audits of Humantic AI and Crystal, is available at https://github.com/DataResponsibly/hiring-stability-Audit.
KW - algorithm audit
KW - hiring
KW - personality
KW - reliability
KW - stability
KW - validity
UR - http://www.scopus.com/inward/record.url?scp=85137167190&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85137167190&partnerID=8YFLogxK
U2 - 10.1145/3514094.3534189
DO - 10.1145/3514094.3534189
M3 - Conference contribution
AN - SCOPUS:85137167190
T3 - AIES 2022 - Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society
SP - 572
EP - 587
BT - AIES 2022 - Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society
PB - Association for Computing Machinery, Inc
Y2 - 1 August 2022 through 3 August 2022
ER -