资源论文mdfa: Multi-Differential Fairness Auditor for Black Box Classifiers

mdfa: Multi-Differential Fairness Auditor for Black Box Classifiers

2019-10-10 | |  43 |   37 |   0
Abstract Machine learning algorithms are increasingly involved in sensitive decision-making processes with adverse implications on individuals. This paper presents mdfa, an approach that identifies the characteristics of the victims of a classifier’s discrimination. We measure discrimination as a violation of multi-differential fairness. Multi-differential fairness is a guarantee that a black box classifier’s outcomes do not leak information on the sensitive attributes of a small group of individuals. We reduce the problem of identifying worst-case violations to matching distributions and predicting where sensitive attributes and classifier’s outcomes coincide. We apply mdfa to a recidivism risk assessment classifier and demonstrate that for individuals with little criminal history, identified African-Americans are three-times more likely to be considered at high risk of violent recidivism than similar non-African-Americans

上一篇:Learning to Select Knowledge for Response Generation in Dialog Systems

下一篇:MINA: Multilevel Knowledge-Guided Attention for Modeling Electrocardiography Signals

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...