首页 | 本学科首页   官方微博 | 高级检索  
     

深度学习模型中的公平性研究
引用本文:王昱颖,张敏,杨晶然,徐晟恺,陈仪香.深度学习模型中的公平性研究[J].软件学报,2023,34(9):4037-4055.
作者姓名:王昱颖  张敏  杨晶然  徐晟恺  陈仪香
作者单位:华东师范大学 软件工程学院, 上海 200062;上海市高可信重点实验室 华东师范大学, 上海 200062
基金项目:国家自然科学基金(61672012); 科技部重点研发项目(2020AAA0107800); 国家自然科学基金项目中以国际合作项目(62161146001)
摘    要:近几年深度神经网络正被广泛应用于现实决策系统,决策系统中的不公平现象会加剧社会不平等,造成社会危害.因此研究者们开始对深度学习系统的公平性展开大量研究,但大部分研究都从群体公平的角度切入,且这些缓解群体偏见的方法无法保证群体内部的公平.针对以上问题,本文定义了两种个体公平率计算方法,分别为基于输出标签的个体公平率(IFRb),即相似样本对在模型预测中标签相同的概率和基于输出分布的个体公平率(IFRp),即相似样本对的预测分布差异在阈值范围内的概率,后者是更严格的个体公平.更进一步,本文提出一种提高模型个体公平性的算法IIFR,该算法通过余弦相似度计算样本之间的差异程度,利用相似临界值筛选出满足条件的相似训练样本对,最后在训练过程中将相似训练样本对的输出差异作为个体公平损失项添加到目标函数中,惩罚模型输出差异过大的相似训练样本对,以达到提高模型个体公平性的目的.实验结果表明,IIFR算法在个体公平的提升上优于最先进的个体公平提升方法.此外IIFR算法能够在提高模型个体公平性的同时,较好地维持模型的群体公平性.

关 键 词:深度学习  模型偏见  个体公平  群体公平
收稿时间:2022/8/23 0:00:00
修稿时间:2022/12/14 0:00:00

Research on Fairness in Deep Learning Models
WANG Yu-Ying,ZHANG Min,YANG Jing-Ran,XU Sheng-Kai,CHEN Yi-Xiang.Research on Fairness in Deep Learning Models[J].Journal of Software,2023,34(9):4037-4055.
Authors:WANG Yu-Ying  ZHANG Min  YANG Jing-Ran  XU Sheng-Kai  CHEN Yi-Xiang
Affiliation:School of Software and Engineering, East China Normal University, Shanghai 200062, China;Shanghai Key Laboratory of Trustworthy Computing, East China Normal University, Shanghai 200062, China
Abstract:In recent years, deep neural networks have been widely used in real decision-making systems. Inequities in decision-making systems will exacerbate social inequality and cause social harm. Therefore, researchers have begun to carry out a lot of research on the fairness of deep learning systems, but most of them focus on group fairness, and they cannot guarantee the fairness within the group. In response to the above problems, we define two individual fairness calculation methods, which are individual fairness rate IFRb based on labels of output, that is the probability of having same predicted label for two similar samples, and individual fairness rate IFRp based on distributions of output, that is the probability of having similiar predicted output distribution for two similar samples, respectively, the latter being the stricter individual fairness. In addition, we also propose an algorithm IIFR to improve the individual fairness of these models. The algorithm uses the cosine similarity to measure the similarity between samples, and then selects the similar sample pairs by the similarity threshold decided by different applications, finally adds the output difference of the similar sample pairs to the objective function as an individual fairness loss item during the training process, which penalizes the similar training samples with large differences of model output in order to improve the individual fairness of the model. The experimental results show that our IIFR algorithm outperforms the state-of-the-art method on the improvement of individual fairness. In addition, IIFR can maintain group fairness of models while improving individual fairness.
Keywords:deep learning  model bias  individual fairness  group fairness
点击此处可从《软件学报》浏览原始摘要信息
点击此处可从《软件学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号