Chrome Extension
WeChat Mini Program
Use on ChatGLM

Differential Privacy for Fair Deep Learning Models

2021 IEEE International Systems Conference (SysCon)(2021)

University of Houston

Cited 1|Views4
Abstract
Increasingly, we rely on deep learning to make decisions. Yet, these models may take unfair decisions due to bias in the training data-sets. Most of the time, this bias is discrimination due to sensitive attributes such as gender, race, or ethnicity. In fact, discrimination and inequities are a daily reality that continues to exist in many fields, e.g. hiring processes, tenures, and workplaces. Therefore, exiting data-sets may be biased which results in getting unfair learning models. In this paper, we empirically prove that learning models trained on biased datasets will produce unfair and discriminating models. To handle this problem, we propose a pre-processing approach that takes advantage of differential privacy properties to mitigate bias in sensitive attributes of data-sets. More precisely, in the learning process, we introduce the randomized response mechanism to mitigate the inequity and avoid discrimination from the training data-set. We evaluate our approach in a hiring process using a synthetic data-set of resumes of candidates. Simulation results show that our approach mitigates bias and takes more fair decisions compared to the TensorFlow differential privacy library or a learning model without our pre-processing approach. This is owing to the fact that the TensorFlow library applies differential privacy to all attributes, Whilst in our approach we only apply the differential privacy mechanism to sensitive attributes that are the source of the discrimination.
More
Translated text
Key words
Differential Privacy,Fair decision,Discrimination,Bias,Sensitive attribute,TensorFlow
求助PDF
上传PDF
Bibtex
AI Read Science
AI Summary
AI Summary is the key point extracted automatically understanding the full text of the paper, including the background, methods, results, conclusions, icons and other key content, so that you can get the outline of the paper at a glance.
Example
Background
Key content
Introduction
Methods
Results
Related work
Fund
Key content
  • Pretraining has recently greatly promoted the development of natural language processing (NLP)
  • We show that M6 outperforms the baselines in multimodal downstream tasks, and the large M6 with 10 parameters can reach a better performance
  • We propose a method called M6 that is able to process information of multiple modalities and perform both single-modal and cross-modal understanding and generation
  • The model is scaled to large model with 10 billion parameters with sophisticated deployment, and the 10 -parameter M6-large is the largest pretrained model in Chinese
  • Experimental results show that our proposed M6 outperforms the baseline in a number of downstream tasks concerning both single modality and multiple modalities We will continue the pretraining of extremely large models by increasing data to explore the limit of its performance
Upload PDF to Generate Summary
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Related Papers

Differentially Private Block Coordinate Descent

JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES 2023

被引用1

Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn
Chat Paper
GPU is busy, summary generation fails
Rerequest