UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Fairer machine learning in the real world: Mitigating discrimination without collecting sensitive data

Veale, M; Binns, R; (2017) Fairer machine learning in the real world: Mitigating discrimination without collecting sensitive data. Big Data & Society , 4 (2) pp. 1-17. 10.1177/2053951717743530. Green open access

[thumbnail of Veale_2053951717743530.pdf]
Preview
Text
Veale_2053951717743530.pdf - Published Version

Download (251kB) | Preview

Abstract

Decisions based on algorithmic, machine learning models can be unfair, reproducing biases in historical data used to train them. While computational techniques are emerging to address aspects of these concerns through communities such as discrimination-aware data mining (DADM) and fairness, accountability and transparency machine learning (FATML), their practical implementation faces real-world challenges. For legal, institutional or commercial reasons, organisations might not hold the data on sensitive attributes such as gender, ethnicity, sexuality or disability needed to diagnose and mitigate emergent indirect discrimination-by-proxy, such as redlining. Such organisations might also lack the knowledge and capacity to identify and manage fairness issues that are emergent properties of complex sociotechnical systems. This paper presents and discusses three potential approaches to deal with such knowledge and information deficits in the context of fairer machine learning. Trusted third parties could selectively store data necessary for performing discrimination discovery and incorporating fairness constraints into model-building in a privacy-preserving manner. Collaborative online platforms would allow diverse organisations to record, share and access contextual and experiential knowledge to promote fairness in machine learning systems. Finally, unsupervised learning and pedagogically interpretable algorithms might allow fairness hypotheses to be built for further selective testing and exploration. Real-world fairness challenges in machine learning are not abstract, constrained optimisation problems, but are institutionally and contextually grounded. Computational fairness tools are useful, but must be researched and developed in and with the messy contexts that will shape their deployment, rather than just for imagined situations. Not doing so risks real, near-term algorithmic harm.

Type: Article
Title: Fairer machine learning in the real world: Mitigating discrimination without collecting sensitive data
Open access status: An open access version is available from UCL Discovery
DOI: 10.1177/2053951717743530
Publisher version: http://doi.org/10.1177/2053951717743530
Language: English
Additional information: © The Author(s) 2017. This article is distributed under the terms of the Creative Commons Attribution 4.0 License (http://www.creativecommons.org/licenses/by/4.0/) which permits any use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage).
Keywords: Algorithmic accountability, algorithms, discrimination, machine learning, personal data, privacy
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL SLASH
UCL > Provost and Vice Provost Offices > UCL SLASH > Faculty of Laws
URI: https://discovery.ucl.ac.uk/id/eprint/1574754
Downloads since deposit
186Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item