资源论文Discovering Multipart Appearance Models from Captioned Images

Discovering Multipart Appearance Models from Captioned Images

2020-03-31 | |  49 |   36 |   0

Abstract

Even a relatively unstructured captioned image set depict- ing a variety of ob jects in cluttered scenes contains strong correlations between caption words and repeated visual structures. We exploit these correlations to discover named ob jects and learn hierarchical models of their appearance. Revising and extending a previous technique for finding small, distinctive configurations of local features, our method assembles these co-occurring parts into graphs with greater spatial extent and flex- ibility. The resulting multipart appearance models remain scale, transla- tion and rotation invariant, but are more reliable detectors and provide better localization. We demonstrate improved annotation precision and recall on datasets to which the non-hierarchical technique was previously applied and show extended spatial coverage of detected ob jects.

上一篇:Dense, Robust, and Accurate Motion Field Estimation from Stereo Image Sequences in Real-Time

下一篇:Efficient Highly Over-Complete Sparse Coding Using a Mixture Model

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...