Visual tracking has attracted a significant attention in the last few decades. The recent surge in the number of publications on tracking-related problems have made it almost impossible to follow the developments in the field. One of the reasons is that there is a lack of commonly accepted annotated data-sets and standardized evaluation protocols that would allow objective comparison of different tracking methods. To address this issue, the Visual Object Tracking (VOT) workshop was organized in conjunction with ICCV2013. Researchers from academia as well as industry were invited to participate in the first VOT2013 challenge which aimed at single-object visual trackers that do not apply pre-learned models of object appearance (model-free). Presented here is the VOT2013 benchmark dataset for evaluation of single-object visual trackers as well as the results obtained by the trackers competing in the challenge. In contrast to related attempts in tracker benchmarking, the dataset is labeled per-frame by visual attributes that indicate occlusion, illumination change, motion change, size change and camera motion, offering a more systematic comparison of the trackers. Furthermore, we have designed an automated system for performing and evaluating the experiments. We present the evaluation protocol of the VOT2013 challenge and the results of a comparison of 27 trackers on the benchmark dataset. The dataset, the evaluation tools and the tracker rankings are publicly available from the challenge website.
ICCV Workshop on the VOT2013 Visual Object Tracking Challenge, 2013.
Download the PDF
Download the code from our GitHub repository.
More information is available at the Project page.
@inproceedings{Kristan2013VOT, author = {Kristan, Matej and Pflugfelder, R. and Leonardis, Ales and Matas, Jiri and Porikli, Fatih and Cehovin, Luka and Nebehay, Georg and Gustavo, Fernandez and Vojir, Tomas}, booktitle = {Workshop on the VOT2013 Visual Object Tracking Challenge}, month = dec, pages = {98--111}, title = {The Visual Object Tracking {VOT2013} challenge results}, year = {2013} }Back to publication list