资源论文AnchorNet: A Weakly Supervised Network to Learn Geometry-sensitive Features For Semantic Matching

AnchorNet: A Weakly Supervised Network to Learn Geometry-sensitive Features For Semantic Matching

2019-12-06 | |  74 |   44 |   0

Abstract

Despite signifificant progress of deep learning in recent years, state-of-the-art semantic matching methods still rely on legacy features such as SIFT or HoG. We argue that the strong invariance properties that are key to the success of recent deep architectures on the classifification task make them unfifit for dense correspondence tasks, unless a large amount of supervision is used. In this work, we propose a deep network, termed AnchorNet, that produces image representations that are well-suited for semantic matching. It relies on a set of fifilters whose response is geometrically consistent across different object instances, even in the presence of strong intra-class, scale, or viewpoint variations. Trained only with weak image-level labels, the fifinal representation successfully captures information about the object structure and improves results of state-of-the-art semantic matching methods such as the deformable spatial pyramid or the proposal flflow methods. We show positive results on the cross-instance matching task where different instances of the same object category are matched as well as on a new cross-category semantic matching task aligning pairs of instances each from a different object class

上一篇:Weakly-supervised Visual Grounding of Phrases with Linguistic Structures

下一篇:Combining Bottom-Up, Top-Down, and Smoothness Cues for Weakly Supervised Image Segmentation

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...