2019 IEEE Intelligent Vehicles Symposium (IV) 2019
DOI: 10.1109/ivs.2019.8813859
|View full text |Cite
|
Sign up to set email alerts
|

Deep Intersection Classification Using First and Third Person Views

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 43 publications
0
7
0
Order By: Relevance
“…For this reason, the most suitable comparisons concerning the work in [6] are those that consider sequences starting from 20m up to the intersection, see Figure 7a. A second comparison can be made with the results of [17]. In both cases, our approach improved the performances of these contributions.…”
Section: B Evaluation Methods and Resultsmentioning
confidence: 91%
See 1 more Smart Citation
“…For this reason, the most suitable comparisons concerning the work in [6] are those that consider sequences starting from 20m up to the intersection, see Figure 7a. A second comparison can be made with the results of [17]. In both cases, our approach improved the performances of these contributions.…”
Section: B Evaluation Methods and Resultsmentioning
confidence: 91%
“…Works in this category include the approach in [16] with a network called IntersectNet, where a sequence of 16 images was passed through an ensemble of Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN), combined to set up a three-types intersection classifier (four road-crossing and T-junctions) using a simple average-pooling fusion layer. A similar ensemble coupled with a more elaborated integration network was used in the work proposed in [17]. Here the authors suggest to use two sets of images relative to the intersection processed with DNN and RNN respectively.…”
Section: Related Workmentioning
confidence: 99%
“…They were able to recognize six types of intersections. Koji and Kanji [ 93 ] used two types of input. First, they used images before an intersection of Third-Person Vision (TPV) and sequences of images while an intersection is passed First-person vision (FPV).…”
Section: Environment Mappingmentioning
confidence: 99%
“…Together with the KITTI dataset, our proposal address this issue, as we believe that much more data and research is needed in this field. The work presented in [ 23 ] maintains a similar approach about temporal integration, but instead of detecting whether the intersection is reached, it tries to classify them into seven basic typologies. For this purpose, their system uses two data sources that the authors called Input-F and Input-T. Input-T is an RGB image taken just before reaching the intersection, and Input-F is a sequence of images taken while crossing the intersection.…”
Section: Related Workwordmentioning
confidence: 99%