Training Data Debugging for the Fairness of Machine Learning Software
Fri 13 May 2022 04:05 - 04:10 at ICSE room 4-even hours - Software Fairness Chair(s): Aldeida Aleti
With the widespread application of machine learning (ML) software, especially in high-risk tasks, the concern about their unfairness has been raised towards both developers and users of ML software. The unfairness of ML software indicates the software behavior affected by the sensitive features (e.g., sex), which leads to biased and illegal decisions and has become a worthy problem for the whole software engineering community.
According to the “data-driven” programming paradigm of ML software, we consider the root cause of the unfairness as biased features in training data. Inspired by software debugging, we propose a novel method, \textbf{L}inear-regression based \textbf{T}raining \textbf{D}ata \textbf{D}ebugging (LTDD), to \textbf{debug} feature values in training data, i.e., (a) identify which features and which parts of them are biased, and (b) exclude the biased parts of such features to recover as much valuable and unbiased information as possible to build fair ML software. We conduct an extensive study on nine data sets and three classifiers to evaluate the effect of our method LTDD compared with four baseline methods. Experimental results show that (a) LTDD can better improve the fairness of ML software with less or comparable damage to the performance, and (b) LTDD is more actionable for fairness improvement in realistic scenarios.
Mon 9 MayDisplayed time zone: Eastern Time (US & Canada) change
Fri 13 MayDisplayed time zone: Eastern Time (US & Canada) change
04:00 - 05:00 | Software FairnessTechnical Track at ICSE room 4-even hours Chair(s): Aldeida Aleti Monash University | ||
04:00 5mTalk | FairNeuron: Improving Deep Neural Network Fairness with Adversary Games on Selective Neurons Technical Track Xuanqi Gao Xi'an Jiaotong University, Juan Zhai Rutgers University, Shiqing Ma Rutgers University, Chao Shen Xi'an Jiaotong University, Yufei Chen Xi'an Jiaotong University, Qian Wang Wuhan University DOI Pre-print Media Attached | ||
04:05 5mTalk | Training Data Debugging for the Fairness of Machine Learning Software Technical Track Yanhui Li Department of Computer Science and Technology, Nanjing University, Linghan Meng Nanjing University, Lin Chen Department of Computer Science and Technology, Nanjing University, Li Yu Nanjing University, Di Wu Momenta, Yuming Zhou Nanjing University, Baowen Xu Nanjing University Pre-print Media Attached | ||
04:10 5mTalk | NeuronFair: Interpretable White-Box Fairness Testing through Biased Neuron Identification Technical Track haibin zheng Zhejiang University of Technology, Zhiqing Chen Zhejiang University of Technology, Tianyu Du Zhejiang University, Xuhong Zhang Zhejiang University, Yao Cheng Huawei International, Shouling Ji Zhejiang University, Jingyi Wang Zhejiang University, Yue Yu College of Computer, National University of Defense Technology, Changsha 410073, China, Jinyin Chen College of Information Engineering, Zhejiang University of Technology, Hangzhou 310023, China DOI Pre-print Media Attached | ||
04:15 5mTalk | Explanation-Guided Fairness Testing through Genetic Algorithm Technical Track Ming Fan Xi'an Jiaotong University, Wenying Wei Xi'an Jiaotong University, Wuxia Jin Xi'an Jiaotong University, Zijiang Yang Western Michigan University, Ting Liu Xi'an Jiaotong University DOI Pre-print |