2018
DOI: 10.48550/arxiv.1807.11332
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Action Detection from a Robot-Car Perspective

Abstract: We present the new Road Event and Activity Detection (READ) dataset, designed and created from an autonomous vehicle perspective to take action detection challenges to autonomous driving. READ will give scholars in computer vision, smart cars and machine learning at large the opportunity to conduct research into exciting new problems such as understanding complex (road) activities, discerning the behaviour of sentient agents, and predicting both the label and the location of future actions and events, with the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 24 publications
(70 reference statements)
0
4
0
Order By: Relevance
“…Also, it should be able to detect and anticipate road users activities such as moving away, moving towards, crossing the road, and anomalous events in real-time to adjust the speed and handle the situation. Therefore, spatio-temporal action localization algorithms need to be developed to guarantee the safety of self-driving cars [205]. Yao et al [206] proposed a traffic anomaly detection with a when-where-what pipeline to detect, localize, and recognize anomalous events from egocentric videos.…”
Section: Action Detection In Autonomous Drivingmentioning
confidence: 99%
“…Also, it should be able to detect and anticipate road users activities such as moving away, moving towards, crossing the road, and anomalous events in real-time to adjust the speed and handle the situation. Therefore, spatio-temporal action localization algorithms need to be developed to guarantee the safety of self-driving cars [205]. Yao et al [206] proposed a traffic anomaly detection with a when-where-what pipeline to detect, localize, and recognize anomalous events from egocentric videos.…”
Section: Action Detection In Autonomous Drivingmentioning
confidence: 99%
“…Action Recognition A variety of datasets have been introduced for action recognition with a single action label [24,48,20,32,21] and multiple action labels [43,59,4] in videos. Recently released datasets such as AVA [16], READ [14], and EPIC-KITCHENS [12] contain actions with corresponding localization around a person or object. Our TITAN dataset is similar to AVA in the sense that it provides spatio-temporal localization for each agent with multiple action labels.…”
Section: Datasetsmentioning
confidence: 99%
“…We concatenate the embedded features through (11)(12), which are given from the hidden states of the bounding box encoder GRU (6), the hidden states of the ego encoder GRU (7), encoded interaction (10) and action embedding (3). We encode all information for 10 observation time steps from (14). We decode the future locations using decoder GRU for 20 future time steps (20).…”
Section: Future Object Localizationmentioning
confidence: 99%
“…Second, by splitting the process in two parts, i.e., the tagging and the scenario mining, the scenario mining can be applied to different types of data (e.g., data from a vehicle [16] or a measurement unit above the road [12], [17]). It is also possible to have manually tagged data, e.g., see [18]. Third, our approach is easily scalable because additional types of scenarios can be mined by describing them as a combination of (sequential) tags.…”
Section: Introductionmentioning
confidence: 99%