2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) 2019
DOI: 10.1109/iccvw.2019.00282
|View full text |Cite
|
Sign up to set email alerts
|

Visual Tracking by Means of Deep Reinforcement Learning and an Expert Demonstrator

Abstract: In the last decade many different algorithms have been proposed to track a generic object in videos. Their execution on recent large-scale video datasets can produce a great amount of various tracking behaviours. New trends in Reinforcement Learning showed that demonstrations of an expert agent can be efficiently used to speed-up the process of policy learning. Taking inspiration from such works and from the recent applications of Reinforcement Learning to visual tracking, we propose two novel trackers, A3CT, … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4
1

Relationship

3
6

Authors

Journals

citations
Cited by 37 publications
(18 citation statements)
references
References 45 publications
0
18
0
Order By: Relevance
“…Tracker. We follow the most recent advancements in deep regression tracking [7], [34], [9] to implement our tracker s(•|θ) as a deep neural network with weights θ. The network gets as input s t as two image patches which pass through two ResNet-18 [35] CNN branches with shared weights.…”
Section: Methodsmentioning
confidence: 99%
“…Tracker. We follow the most recent advancements in deep regression tracking [7], [34], [9] to implement our tracker s(•|θ) as a deep neural network with weights θ. The network gets as input s t as two image patches which pass through two ResNet-18 [35] CNN branches with shared weights.…”
Section: Methodsmentioning
confidence: 99%
“…An improvement to [45] for SOT was proposed by [84], where a visual tracker was formulated using DRL and an expert demonstrator. The authors treated the problem as an MDP, where the state consists of two consecutive frames that have been cropped using the bounding box corresponding to the former frame and used a scaling factor to control the offset while cropping.…”
Section: Drl In Object Trackingmentioning
confidence: 99%
“…The figure illustrates a general implementation of object tracking in videos using DRL, where the state consists of two consecutive frames (F t , F t+1 ) with a bounding box for the first frame produced by another algorithm for the first iteration or by the previous iterations of DRL agent. The actions corresponds to the moving the bounding on the image to fit the object in frame F t+1 , hence forming a new state with frame F t+1 and frame F t+2 along with the bounding box for frame F t+1 predicted by previous iteration and reward corresponds to whether IOU is greater then a given threshold as used by [118], [308], [45], [84], [307], [168], [169].…”
Section: Objectmentioning
confidence: 99%
“…< l a t e x i t s h a 1 _ b a s e 6 4 = " 0 q a i s a C W N S A d B x M e B n 5 Athlete Tracking. The first step of the pipeline is to ploit a visual object tracker [9,10] to track the motion of the athlete across all the frames up to F t . We used a tracker outputting a bounding-box b t = (x The number in the top-left corner of each image reports the frame index t in the video.…”
Section: Pipelinementioning
confidence: 99%