资源论文Visual Semantic Search:Retrieving Videos via Complex Textual Queries

Visual Semantic Search:Retrieving Videos via Complex Textual Queries

2019-12-11 | |  84 |   44 |   0

Abstract

In this paper,we tackle the problem of retrieving videos using complex natural language queries.Towards this goal, we first parse the sentential descriptions into a semantic graph,which is then matched to visual concepts using a generalized bipartite matching algorithm.Our approach exploits object appearance,morion and spatial relations, and learns the importance of each term using structure pre- diction.We demonstrate the efectiveness of our approach on a new dataset designed for semantic search in the context of autonomous driving,which exhibits complex and highly dyamic scenes with many objects.We show thar our ap- proach is able to locate a major portion of the objects de- scribed in the query with high accuracy,and improve the relevance in video retrieval.

上一篇:SphereFlow: 6 DoF Scene Flow from RGB-D Pairs

下一篇:Interval Tracker: Tracking by Interval Analysis

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...