2017
DOI: 10.48550/arxiv.1712.01358
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Long-Term Visual Object Tracking Benchmark

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
44
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 12 publications
(44 citation statements)
references
References 0 publications
0
44
0
Order By: Relevance
“…The CNN layer has a sigmoid based activation on the outputs. The videos from TLP dataset [16] and Basketball dataset [6] (described in Section 4.1) are used to train the model. We sub-sample from the full trajectory of the video at random frames and create a set of input trajectories each of a fixed length of 512 and use these as single instances of training data for the model.…”
Section: Implementation Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…The CNN layer has a sigmoid based activation on the outputs. The videos from TLP dataset [16] and Basketball dataset [6] (described in Section 4.1) are used to train the model. We sub-sample from the full trajectory of the video at random frames and create a set of input trajectories each of a fixed length of 512 and use these as single instances of training data for the model.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…Stage Performance dataset: We build a Stage Performance dataset that comprises of two wide-angle recordings of staged performances each of 12 and 10 minutes, respectively. The videos are selected from the Track Long and Prosper (TLP) Dataset [16]. The original recordings were done using a static wide-angle camera covering the entire action in the scene.…”
Section: Datasetsmentioning
confidence: 99%
“…The training datasets used in this paper for SALNet and GPGNet are TLP50 [33], DTB70 [23] and LaSOT [14] 2 dataset which totally contain 120 (50 + 70) and 1400 video sequences, respectively 3 . The LaSOT provide both the BBox and natural language annotations of target object, which is suitable for our natural language guided tracking task.…”
Section: Datasets and Evaluation Criterionmentioning
confidence: 99%
“…https://cis.temple.edu/lasot/3 The baseline method pyMDNet used in this paper is implemented based on PyTorch and pre-trained on two long-term dataset TLP[33] and DTB70 dataset[23] for all our experiments 4. https://github.com/QUVA-Lab/lang-tracker…”
mentioning
confidence: 99%
“…For instance, the most commonly used OTB dataset has an average length of about 20 seconds [35] per clip. Work by Moudgil and Gandhi [23] observed a sharp performance drop when the trackers were evaluated on long sequences. Following works [10,33,21] also make similar observations and suggest that we need alternate ways to evaluate and analyze long term tracking performance.…”
Section: Introductionmentioning
confidence: 99%