Verifiable Fairness: Privacy–preserving Computation of Fairness for Machine Learning Systems

  • Ehsan Toreini*
  • , Maryam Mehrnezhad
  • , Aad van Moorsel
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Fair machine learning is a thriving and vibrant research topic. In this paper, we propose Fairness as a Service (FaaS), a secure, verifiable and privacy-preserving protocol to computes and verify the fairness of any machine learning (ML) model. In the deisgn of FaaS, the data and outcomes are represented through cryptograms to ensure privacy. Also, zero knowledge proofs guarantee the well-formedness of the cryptograms and underlying data. FaaS is model–agnostic and can support various fairness metrics; hence, it can be used as a service to audit the fairness of any ML model. Our solution requires no trusted third party or private channels for the computation of the fairness metric. The security guarantees and commitments are implemented in a way that every step is securely transparent and verifiable from the start to the end of the process. The cryptograms of all input data are publicly available for everyone, e.g., auditors, social activists and experts, to verify the correctness of the process. We implemented FaaS to investigate performance and demonstrate the successful use of FaaS for a publicly available data set with thousands of entries.

Original languageEnglish
Title of host publicationComputer Security. ESORICS 2023 International Workshops
Subtitle of host publicationCPS4CIP, ADIoT, SecAssure, WASP, TAURIN, PriST-AI, and SECAI, The Hague, The Netherlands, September 25–29, 2023, Revised Selected Papers, Part II
EditorsSokratis Katsikas, Habtamu Abie, Silvio Ranise, Luca Verderame, Enrico Cambiaso, Rita Ugarelli, Isabel Praça, Wenjuan Li, Weizhi Meng, Steven Furnell, Basel Katt, Sandeep Pirbhulal, Ankur Shukla, Michele Ianni, Mila Dalla Preda, Kim-Kwang Raymond Choo, Miguel Pupo Correia, Abhishta Abhishta, Giovanni Sileno, Mina Alishahi, Harsha Kalutarage, Naoto Yanai
PublisherSpringer
Pages569-584
Number of pages16
Edition1
ISBN (Electronic)9783031541292
ISBN (Print)9783031541285
DOIs
Publication statusPublished - 12 Mar 2024
EventInternational Workshops which were held in conjunction with 28th European Symposium on Research in Computer Security, ESORICS 2023 - The Hague, Netherlands
Duration: 25 Sept 202329 Sept 2023

Publication series

NameLecture Notes in Computer Science
Volume14399
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

ConferenceInternational Workshops which were held in conjunction with 28th European Symposium on Research in Computer Security, ESORICS 2023
Country/TerritoryNetherlands
CityThe Hague
Period25/09/2329/09/23

Bibliographical note

Publisher Copyright: © The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.

Keywords

  • artificial intelligence
  • fairness computation
  • machine learning fairness
  • trustworthiness

ASJC Scopus subject areas

  • Theoretical Computer Science
  • General Computer Science

Fingerprint

Dive into the research topics of 'Verifiable Fairness: Privacy–preserving Computation of Fairness for Machine Learning Systems'. Together they form a unique fingerprint.

Cite this