scispace - formally typeset
Open AccessProceedings Article

Deep Attentive Tracking via Reciprocative Learning

Reads0
Chats0
TLDR
In this paper, a reciprocative learning algorithm was proposed to exploit visual attention for training deep classifiers, which consists of feed-forward and backward operations to generate attention maps, which serve as regularization terms coupled with the original classification loss function for training.
Abstract
Visual attention, derived from cognitive neuroscience, facilitates human perception on the most pertinent subset of the sensory data. Recently, significant efforts have been made to exploit attention schemes to advance computer vision systems. For visual tracking, it is often challenging to track target objects undergoing large appearance changes. Attention maps facilitate visual tracking by selectively paying attention to temporal robust features. Existing tracking-by-detection approaches mainly use additional attention modules to generate feature weights as the classifiers are not equipped with such mechanisms. In this paper, we propose a reciprocative learning algorithm to exploit visual attention for training deep classifiers. The proposed algorithm consists of feed-forward and backward operations to generate attention maps, which serve as regularization terms coupled with the original classification loss function for training. The deep classifier learns to attend to the regions of target objects robust to appearance changes. Extensive experiments on large-scale benchmark datasets show that the proposed attentive tracking method performs favorably against the state-of-the-art approaches.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

GOT-10k: A Large High-Diversity Benchmark for Generic Object Tracking in the Wild

TL;DR: A large tracking database that offers an unprecedentedly wide coverage of common moving objects in the wild, called GOT-10k, and the first video trajectory dataset that uses the semantic hierarchy of WordNet to guide class population, which ensures a comprehensive and relatively unbiased coverage of diverse moving objects.
Proceedings ArticleDOI

SiamCAR: Siamese Fully Convolutional Classification and Regression for Visual Tracking

TL;DR: A novel fully convolutional Siamese network to solve visual tracking end-to-end in a per-pixel manner by decomposing the visual tracking task into two subproblems as classification for pixel category and regression for object bounding box at this pixel is proposed.
Proceedings ArticleDOI

Target-Aware Deep Tracking

TL;DR: This paper develops a regression loss and a ranking loss to guide the generation of target-active and scale-sensitive features and proposes a novel scheme to learn target-aware features, which can better recognize the targets undergoing significant appearance variations than pre-trained deep features.
Proceedings ArticleDOI

Unsupervised Deep Tracking

TL;DR: The proposed unsupervised tracker achieves the baseline accuracy of fully supervised trackers, which require complete and accurate labels during training, and exhibits a potential in leveraging unlabeled or weakly labeled data to further improve the tracking accuracy.
Journal ArticleDOI

Deep Learning for Visual Tracking: A Comprehensive Survey

TL;DR: This survey aims to systematically investigate the current DL-based visual tracking methods, benchmark datasets, and evaluation metrics, and extensively evaluates and analyzes the leading visualtracking methods.
Related Papers (5)