资源论文Multi-Cue Visual Tracking Using Robust Feature-Level Fusion Based on Joint Sparse Representation

Multi-Cue Visual Tracking Using Robust Feature-Level Fusion Based on Joint Sparse Representation

2019-12-16 | |  47 |   42 |   0

Abstract

The use of multiple features for tracking has been proved as an effective approach because limitation of each feature could be compensated. Since different types of variations such as illumination, occlusion and pose may happen in a video sequence, especially long sequence videos, how to dynamically select the appropriate features is one of the key problems in this approach. To address this issue in multicue visual tracking, this paper proposes a new joint sparse representation model for robust feature-level fusion. The proposed method dynamically removes unreliable features to be fused for tracking by using the advantages of sparse representation. As a result, robust tracking performance is obtained. Experimental results on publicly available videos show that the proposed method outperforms both existing sparse representation based and fusion-based trackers

上一篇:Continuous Manifold Based Adaptation for Evolving Visual Domains

下一篇:SeamSeg: Video Object Segmentation using Patch Seams

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...