Robust appearance feature learning using pixel‐wise discrimination for visual tracking

Considering the high dimensions of video sequences, it is often challenging to acquire a sufficient dataset to train the tracking models. From this perspective, we propose to revisit the idea of hand‐crafted feature learning to avoid such a requirement from a dataset. The proposed tracking approach...

Full description

Saved in:
Bibliographic Details
Published inETRI journal Vol. 41; no. 4; pp. 483 - 493
Main Authors Kim, Minji, Kim, Sungchan
Format Journal Article
LanguageEnglish
Published Electronics and Telecommunications Research Institute (ETRI) 01.08.2019
한국전자통신연구원
Subjects
Online AccessGet full text
ISSN1225-6463
2233-7326
DOI10.4218/etrij.2018-0486

Cover

More Information
Summary:Considering the high dimensions of video sequences, it is often challenging to acquire a sufficient dataset to train the tracking models. From this perspective, we propose to revisit the idea of hand‐crafted feature learning to avoid such a requirement from a dataset. The proposed tracking approach is composed of two phases, detection and tracking, according to how severely the appearance of a target changes. The detection phase addresses severe and rapid variations by learning a new appearance model that classifies the pixels into foreground (or target) and background. We further combine the raw pixel features of the color intensity and spatial location with convolutional feature activations for robust target representation. The tracking phase tracks a target by searching for frame regions where the best pixel‐level agreement to the model learned from the detection phase is achieved. Our two‐phase approach results in efficient and accurate tracking, outperforming recent methods in various challenging cases of target appearance changes.
Bibliography:Funding information
This work was funded by the research funds of Chonbuk National University in 2014.
https://doi.org/10.4218/etrij.2018-0486
ISSN:1225-6463
2233-7326
DOI:10.4218/etrij.2018-0486