资源论文Seeing through the Human Reporting Bias: Visual Classifiers from Noisy Human-Centric Labels

Seeing through the Human Reporting Bias: Visual Classifiers from Noisy Human-Centric Labels

2019-12-23 | |  58 |   39 |   0

Abstract

When human annotators are given a choice about what to label in an image, they apply their own subjective judgments on what to ignore and what to mention. We refer to these noisy “human-centric” annotations as exhibiting human reporting bias. Examples of such annotations include image tags and keywords found on photo sharing sites, or in datasets containing image captions. In this paper, we use these noisy annotations for learning visually correct image classififiers. Such annotations do not use consistent vocabulary, and miss a signifificant amount of the information present in an image; however, we demonstrate that the noise in these annotations exhibits structure and can be modeled. We propose an algorithm to decouple the human reporting bias from the correct visually grounded labels. Our results are highly interpretable for reporting “what’s in the image” versus “what’s worth saying.” We demonstrate the algorithm’s effificacy along a variety of metrics and datasets, including MS COCO and Yahoo Flickr 100M. We show significant improvements over traditional algorithms for both image classifification and image captioning, doubling the performance of existing methods in some cases.

上一篇:Learned Binary Spectral Shape Descriptor for 3D Shape Correspondence

下一篇:Contour detection in unstructured 3D point clouds

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...