Cited By
View all- Simson JFabris AKern C(2024)Lazy Data Practices Harm Fairness ResearchProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658931(642-659)Online publication date: 3-Jun-2024
\(x\in\mathcal{X}\) | A vector of non-sensitive attributes |
\(s\in\mathcal{S}\) | A vector of sensitive attributes |
\(s=g\) | A sensitive attribute value |
\(s={g}^{\mathsf{c}}\) | Complement of a sensitive attribute value |
\(y\in\mathcal{Y}\) | Target variable from domain \(\mathcal{Y}\) |
\(i=(x_{i},s_{i},y_{i})\) | A data point or item |
\(\hat{y}=f(x)\) | A classifier \(f:\mathcal{X}\rightarrow\mathcal{Y}\) issuing predictions in \(\mathcal{Y}\) for data points in \(\mathcal{X}\) |
\(f_{\text{soft}}(x)\) | A soft classifier, from which predictions \(\hat{y}=f(x)\) can be derived through thresholding |
\(\tau=\text{argsort}(f_{\text{soft}}(x))\) | Ranking of items, typically sorted by target estimates |
\(i=\tau(k)\) | Item at rank \(k\) in \(\tau\) |
\(N\) | Number of items |
\(N_{g}\) | Number of items in \(g\) |
\(N_{g}^{k}\) | Number of items in \(g\) in the top \(k\) positions of a ranking |
\(D_{g}\) | Desired representation for group \(g\) |
\(h(x)\) | A classifier \(h:\mathcal{X}\rightarrow\mathcal{S}\) issuing predictions in \(\mathcal{S}\) for data points |
\(h_{\text{soft}}(x)\) | A soft classifier for sensitive attributes |
Used in | Flavor | Cond. | Multi. | Gran. | Norm. | Interp. | |
---|---|---|---|---|---|---|---|
\(\operatorname{\text{skew@}k}\) | [103] | Outcome | No | No | \(\checkmark\) | D | |
\(\operatorname{NDKL}\) | [11, 103] | Outcome | \(\checkmark\) | \(\checkmark\) | \(\checkmark\) | No | |
\(\operatorname{DI}\) | [33, 38, 69, 148, 271] | Outcome | \(\checkmark\) | No | \(\checkmark\) | \(\checkmark\) | |
\(\operatorname{DD}\) | [222] | Outcome | No | No | \(\checkmark\) | \(\checkmark\) | |
\(\operatorname{RPP}\) | [4, 132, 199] | Outcome | No | No | No | \(\checkmark\) | |
\(\operatorname{DRD}\) | [283] | Outcome | No | \(\checkmark\) | \(\checkmark\) | \(\checkmark\) | |
\(\operatorname{SKL}\) | [199] | Outcome | No | \(\checkmark\) | \(\checkmark\) | No | |
\(\operatorname{LRR}\) | [48] | Outcome | \(x\) | No | \(\checkmark\) | No | D |
\(\operatorname{TPRD}\) | [63, 119] | Outcome | \(y\) | No | No | \(\checkmark\) | \(\checkmark\) |
\(\operatorname{FNRR}\) | [148] | Outcome | \(y\) | No | No | \(\checkmark\) | \(\checkmark\) |
\(\operatorname{MED}\) | [231] | Outcome | \(y\) | \(\checkmark\) | \(\checkmark\) | \(\checkmark\) | \(\checkmark\) |
\(\operatorname{RMS}\) | [119] | Outcome | \(y\) | \(\checkmark\) | No | No | No |
\(\operatorname{xEO}\) | [185] | Outcome | \(y\) | No | \(\checkmark\) | No | No |
\(\operatorname{MAE}\) | [231, 276] | Accuracy | \(y\) | No | \(\checkmark\) | \(\checkmark\) | No |
\(\operatorname{BCRD}\) | [148] | Accuracy | \(y\) | No | No | \(\checkmark\) | No |
\(\operatorname{MID}\) | [148] | Accuracy | \(y\) | No | No | \(\checkmark\) | No |
\(\operatorname{SD}\) | [222] | Impact | No | No | \(\checkmark\) | \(\checkmark\) | |
\(\operatorname{GBS}\) | [128] | Representational | No | n.a | No | \(\checkmark\) | |
\(\operatorname{sAUC}\) | [33, 119, 120, 194, 196, 222] | Process | No | n.a. | No | n.a. | |
\(\operatorname{GTR}\) | [196] | Process | \(x\) | No | n.a. | No | n.a. |
\(\operatorname{MIA}\) | [276] | Process | \(y\) | \(\checkmark\) | No | \(\checkmark\) | n.a. |
In | Family | Measures | s availability | Mode | Approach | Summary | |
---|---|---|---|---|---|---|---|
Rule-based Scraping or Substitution | [63, 194, 222] | PRE | sAUC, DD, TPRD, SD | None | Text | Proxy reduction | Remove or substitute gender identifiers based on hard-coded rules |
Importance-Based Scraping | [33, 194] | PRE | sAUC, DI | Train | Proxy reduction | Iteratively remove proxy features most predictive of sensitive attribute | |
Balanced Sampling | [11, 276] | PRE | NDKL, MIA, MAE | Train | Re-balancing | Resample training set with same group cardinality in each class | |
Group Norming | [33] | PRE | sAUC, DI | Train | Groupwise feature transform | \(z\) -score normalization of each feature within each group | |
Subspace Projection | [194] | PRE | sAUC | None | Text | Proxy reduction | Remove gender information via embedding projection |
Adversarial Inference | [120, 199, 222, 276] | IN | sAUC, DD, TPRD, SD, MIA, MAE, SKL | Train | Proxy reduction | Reduce sensitive info in latent representation through additional adv. loss | |
Face Decorrelation | [120] | IN | sAUC, DI | None | Image | Proxy reduction | Discourage intermediate representations correlated with face ID through adv. loss |
Name Decorrelation | [119] | IN | sAUC, TPRD, RMS | None | Text | Proxy reduction | Discourage intermediate representation correlated with names by minimizing MI |
Fair TF-IDF | [69] | IN | DI | Train | Text | Proxy reduction | Reduce feature weight according to its group-specificity |
DetGreedy | [11, 103, 240] | POST | NDKL | Runtime | Output re-ranking | Re-rank items enforcing desired group representation in each prefix | |
CDF Rescoring | [185] | POST | xEO | Runtime | Output re-rescoring | Re-score items with groupwise CDF trick | |
Spatial Partitioning | [38] | POST | DI | Train | Output re-ranking | Promote candidates based on group membership probability |
Dataset name | Used in | Type | Language | Geography | Target variable | Sensitive variable \({}^{\mathrm{a}}\) | Hiring stage |
---|---|---|---|---|---|---|---|
Chinese Bios | [283] | Textual | Chinese | Synthetic | Mention of job-specific skills in bio | Gender (B, A) | Sourcing |
Bias in Bios | [63, 119] | Textual | English | World | Candidate occupation | Gender (B, A) | Sourcing |
Engineers and Scientists | [58] | Textual | English | Unknown | Candidate received an offer | Unknown | Screening |
IT Resumes | [194, 196] | Textual | English | US | Candidate was called back | Gender (B, S) | Sourcing |
CVs from Singapore | [69] | Textual | English | Singapore | CV match with job description | Ethnicity (A) | Sourcing |
DPG Resumes | [222] | Textual | English, Dutch | The Netherlands | Candidate industry of interest | Gender (B, A) | Sourcing |
ChaLearn First Impression | [120, 148, 276] | Visual | English | Unknown | Speaker hireability and personality annotated by AMT workers | Gender (B, A), ethnicity (A) | Screening |
Student Interviews | [33] | Visual | English | Unknown | Speaker hireability annotated by research assistants | Gender (NB, S) | Screening |
SHL Interviews | [231] | Visual | English | US, UK, India | Speaker communication skills rated by experts | Gender (B), age, race, country of residence | Screening |
Oil Company Interviews | [139] | Visual | Dutch | The Netherlands | Candidate self-reported commitment | Gender, age | Screening |
FairCVs | [119, 199] | Visual | English | Synthetic | Synthetic score | Gender (B, A), race (A) | Sourcing |
Requirements and Candidates | [172] | Tabular | Synthetic | Synthetic | Synthetic score | Synthetic | Sourcing |
Jobs and Candidates | [11] | Tabular | Unknown | The Netherlands | Candidate recruited or shortlisted | Gender (B, S) | Sourcing |
Pymetrics Bias Group | [271] | Tabular | Unknown | Unknown | Similarity to current employees | Gender (S), race (S) | Screening |
IBM HR Analytics | [105] | Tabular | Unknown | Unknown | Employee resignation | Gender (B), age, marital status | Evaluation |
Resume Search Engines | [48] | Tabular | English | US | Unknown | Gender (B, S) | Sourcing |
Walmart Employees | [38] | Tabular | English | US | Employee tenure and performance ratings | Synthetic | Screening |
Web Developers Field Study | [13] | Tabular | English | US | Unknown | Gender (B,S), age (S), ethnicity (S) | Screening |
Chinese Job Recommendations | [282] | Tabular | Chinese | China | Unknown | Gender (B,A), age (A) | Sourcing |
Facebook Ads Audiences | [4] | Tabular | English | US | User clicks | Gender (B, S), race (B, A) | Sourcing |
Opportunities | Limitations | Recommendations | |
---|---|---|---|
Bias and Validity | Consider large candidate pools, reduce human biases, and attract minority candidates | Risk of encoding individual biases along with inevitable societal biases; invalid target variables | Focus on vulnerable populations beyond acceptance rates; study individual fairness and exploratory policies; carefully scrutinize new technologies |
Broader Context | Trigger positive feedback loops; consider tech-recruiter collaboration | Narrow focus on local outcomes can overlook fairness in entire hiring process; risk of repurposing for termination | Center on job seeker impacts; identify leaks in pipelines; design for recruiter-tech interaction; beware of performance and tenure prediction |
Data | Support evaluations of diversity and inclusion | Reduced geographical, linguistic, and sensitive attribute coverage | Design data collection for diversity; develop IP-friendly audits |
Law | Apply binding regulation to positively influence industry | Legal restrictions on fairness approaches: concerns about discrimination and data protection | Multidisciplinary research balancing fairness, privacy, and anti-discrimination; monitor EU AI Act |
The use of AI-enabled hiring software raises questions about the practice of Human Resource (HR) professionals' use of the software and its consequences. We interviewed 15 recruiters and HR professionals about their experiences around two decision-...
There has been rapidly growing interest in the use of algorithms in hiring, especially as a means to address or mitigate bias. Yet, to date, little is known about how these methods are used in practice. How are algorithmic assessments built, validated, ...
Algorithms and machine learning models, including the decisions made by these models, are becoming ubiquitous in our daily life, including hiring. We make no value judgment regarding this development; rather, we simply acknowledge that it is quickly ...
Association for Computing Machinery
New York, NY, United States
Check if you have access through your login credentials or your institution to get full access on this article.
Sign in