资源论文OVERLEARNING REVEALS SENSITIVE ATTRIBUTES

OVERLEARNING REVEALS SENSITIVE ATTRIBUTES

2020-01-02 | |  60 |   55 |   0

Abstract

“Overlearning” means that a model trained for a seemingly simple objective implicitly learns to recognize attributes and concepts that are (1) not part of the learning objective, and (2) sensitive from a privacy or bias perspective. For example, a binary gender classifier of facial images also learns to recognize races—even races that are not represented in the training data—and identities. We demonstrate overlearning in several vision and NLP models and analyze its harmful consequences. First, inference-time representations of an overlearned model reveal sensitive attributes of the input, breaking privacy protections such as model partitioning. Second, an overlearned model can be “re-purposed” for a different, privacy-violating task even in the absence of the original training data. We show that overlearning is intrinsic for some tasks and cannot be prevented by censoring unwanted attributes. Finally, we investigate where, when, and why overlearning happens during model training.

上一篇:KALEIDOSCOPE :A NE FFICIENT, LEARNABLE REPRE -SENTATION FOR ALL STRUCTURED LINEAR MAPS

下一篇:NEVER GIVE UP :L EARNING DIRECTED EXPLO -RATION STRATEGIES

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...