2020
DOI: 10.48550/arxiv.2003.05982
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

LaserFlow: Efficient and Probabilistic Object Detection and Motion Forecasting

Abstract: In this work, we present LaserFlow, an efficient method for 3D object detection and motion forecasting from LiDAR. Unlike the previous work, our approach utilizes the native range view representation of the LiDAR, which enables our method to operate at the full range of the sensor in real-time without voxelization or compression of the data. We propose a new multi-sweep fusion architecture, which extracts and merges temporal features directly from the range images. Furthermore, we propose a novel technique for… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
5

Relationship

4
1

Authors

Journals

citations
Cited by 5 publications
(11 citation statements)
references
References 28 publications
0
11
0
Order By: Relevance
“…MotionNet [32] used a spatial-temporal pyramid network to jointly perform detection and motion prediction for each BEV grid cell. LaserFlow [7] proposed an end-to-end model using multi-frame RV lidar inputs, unlike the other methods which use BEV representations, which can also perform prediction on multiple actor types. Compared to our method, most of the above end-to-end methods do not consider motion prediction on diverse road actor types, and none of them addresses the multimodal nature of possible future trajectories.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…MotionNet [32] used a spatial-temporal pyramid network to jointly perform detection and motion prediction for each BEV grid cell. LaserFlow [7] proposed an end-to-end model using multi-frame RV lidar inputs, unlike the other methods which use BEV representations, which can also perform prediction on multiple actor types. Compared to our method, most of the above end-to-end methods do not consider motion prediction on diverse road actor types, and none of them addresses the multimodal nature of possible future trajectories.…”
Section: Related Workmentioning
confidence: 99%
“…However, these approaches have generally focused only on vehicles and produce a single trajectory rather than full motion distributions. More recent work has shown the ability to learn a continuous distribution directly from sensor data for multiple classes [7], but the distributions are not multimodal. Prediction methods that operate on detections rather than the raw sensor data have shown improved performance by introducing multiclass predictions [8], estimates of uncertainty [9], [10], or incorporating multiple modes [11].…”
Section: Introductionmentioning
confidence: 99%
“…Fig. 2), optical flow [43], [60], depth estimation [50], [57], [61], [62], visual odometry [35], [63], [64], image annotation [41], [65]- [67], visual tracking [52], trajectory prediction [49], [68]- [70] and end-to-end perception [47], [71], [72]. Tab.…”
Section: Development Operationmentioning
confidence: 99%
“…The model assumes a univariate Laplace distribution for each of the (c x , c y , sin θ , cos θ ) components at each of the n + 1 timepoints, i.e, (n + 1) × 4 × 2 values are regressed for the means (µ's) and diversity parameters (b's). By contrast, in models with the polynomial representation, the regression values are the N µ + 1 coefficients in (4) and the N b + 1 coefficients in (5), for the four components individually. If multimodal prediction is modeled, such as for the vehicles in MultiXNet, independent polynomial representation is applied to each separate mode.…”
Section: Applying the Representation In Supervised Learningmentioning
confidence: 99%
“…In robotics in general and self-driving vehicle (SDV) applications in particular, anticipating the motion of other actors around the robot plays a critical role in planning safe paths to navigate the environment [1]. Recently, significant improvements have come from exploring the input representation of the sensor data [2], [3], [4], [5], [6] and the neural network structures [7], [8], [9]. Likewise, the output representation for trajectories has seen extensions to account for multimodality [10], [11], [12] and for modeling probability distributions over their future locations [9], [13].…”
Section: Introductionmentioning
confidence: 99%