High Performance Visual Object Tracking with Unified Convolutional
Networks
release_w5nhmig2hnewvl53ll7spb4pu4
by
Zheng Zhu, Wei Zou, Guan Huang, Dalong Du, Chang Huang
2019
Abstract
Convolutional neural networks (CNN) based tracking approaches have shown
favorable performance in recent benchmarks. Nonetheless, the chosen CNN
features are always pre-trained in different tasks and individual components in
tracking systems are learned separately, thus the achieved tracking performance
may be suboptimal. Besides, most of these trackers are not designed towards
real-time applications because of their time-consuming feature extraction and
complex optimization details. In this paper, we propose an end-to-end framework
to learn the convolutional features and perform the tracking process
simultaneously, namely, a unified convolutional tracker (UCT). Specifically,
the UCT treats feature extractor and tracking process both as convolution
operation and trains them jointly, which enables learned CNN features are
tightly coupled with tracking process. During online tracking, an efficient
model updating method is proposed by introducing peak-versus-noise ratio (PNR)
criterion, and scale changes are handled efficiently by incorporating a scale
branch into network. Experiments are performed on four challenging tracking
datasets: OTB2013, OTB2015, VOT2015 and VOT2016. Our method achieves leading
performance on these benchmarks while maintaining beyond real-time speed.
In text/plain
format
Archived Files and Locations
application/pdf 2.1 MB
file_xiayltdasraftmsjsc4oevg35i
|
arxiv.org (repository) web.archive.org (webarchive) |
1908.09445v1
access all versions, variants, and formats of this works (eg, pre-prints)