Abstract:Many advanced driver assistance systems (ADAS) are currently trying to utilise multi-sensor architectures, where the driver assistance algorithm receives data from a multitude of sensors. As mono-sensor systems cannot provide reliable and consistent readings under all circumstances because of errors and other limitations, fusing data from multiple sensors ensures that the environmental parameters are perceived correctly and reliably for most scenarios, thereby substantially improving the reliability of the mul… Show more
“…Sensor fusion is a popular approach in modern perception systems, where information from multiple sensors is combined to enhance the overall perception and understanding of the environment. In the context of object detection and tracking, integrating 3D LiDAR data with other sensors, such as cameras, can provide more accurate and comprehensive information about the surrounding objects [ 90 , 91 ].…”
Object detection and tracking are vital in computer vision and visual surveillance, allowing for the detection, recognition, and subsequent tracking of objects within images or video sequences. These tasks underpin surveillance systems, facilitating automatic video annotation, identification of significant events, and detection of abnormal activities. However, detecting and tracking small objects introduce significant challenges within computer vision due to their subtle appearance and limited distinguishing features, which results in a scarcity of crucial information. This deficit complicates the tracking process, often leading to diminished efficiency and accuracy. To shed light on the intricacies of small object detection and tracking, we undertook a comprehensive review of the existing methods in this area, categorizing them from various perspectives. We also presented an overview of available datasets specifically curated for small object detection and tracking, aiming to inform and benefit future research in this domain. We further delineated the most widely used evaluation metrics for assessing the performance of small object detection and tracking techniques. Finally, we examined the present challenges within this field and discussed prospective future trends. By tackling these issues and leveraging upcoming trends, we aim to push forward the boundaries in small object detection and tracking, thereby augmenting the functionality of surveillance systems and broadening their real-world applicability.
“…Sensor fusion is a popular approach in modern perception systems, where information from multiple sensors is combined to enhance the overall perception and understanding of the environment. In the context of object detection and tracking, integrating 3D LiDAR data with other sensors, such as cameras, can provide more accurate and comprehensive information about the surrounding objects [ 90 , 91 ].…”
Object detection and tracking are vital in computer vision and visual surveillance, allowing for the detection, recognition, and subsequent tracking of objects within images or video sequences. These tasks underpin surveillance systems, facilitating automatic video annotation, identification of significant events, and detection of abnormal activities. However, detecting and tracking small objects introduce significant challenges within computer vision due to their subtle appearance and limited distinguishing features, which results in a scarcity of crucial information. This deficit complicates the tracking process, often leading to diminished efficiency and accuracy. To shed light on the intricacies of small object detection and tracking, we undertook a comprehensive review of the existing methods in this area, categorizing them from various perspectives. We also presented an overview of available datasets specifically curated for small object detection and tracking, aiming to inform and benefit future research in this domain. We further delineated the most widely used evaluation metrics for assessing the performance of small object detection and tracking techniques. Finally, we examined the present challenges within this field and discussed prospective future trends. By tackling these issues and leveraging upcoming trends, we aim to push forward the boundaries in small object detection and tracking, thereby augmenting the functionality of surveillance systems and broadening their real-world applicability.
“…In this paper, for the purpose of gauging the performance of the tracker algorithm under multiple scenarios, we have used some standard off-the-shelf LiDAR, camera, and RaDAR detection algorithms, namely, YOLOv4 for camera object detection as implemented by Kumar et al [9], a RaDAR based object detection methodology as worked upon by Manjunath et al [10], and DBSCAN (Density-Based Spatial Clustering of Applications with Noise) for LiDAR object detection, which is similar to that developed by Deng et al [11]. As explained earlier in Section 1, we shall use the target-level sensor fusion model, as worked upon by Deo et al [5], for fusing the sensor data.…”
Section: Camera Radar and Lidar Sensor Fusionmentioning
confidence: 99%
“…Several studies on the topic of multi-object sensor fusion and tracking are available in the public domain [2,4,5]. In a typical target-level sensor fusion architecture, all target objects are independently tracked once the data from multiple sensors is associated or fused [6].…”
Section: Introductionmentioning
confidence: 99%
“…Sensor fusion techniques, on the other hand, leverage the output provided by multiple sensors; as a result, they are more reliable and error-free, compared to the techniques that employ a single sensor to gauge the environment [2,3]. Several studies on the topic of multi-object sensor fusion and tracking are available in the public domain [2,4,5]. In a typical target-level sensor fusion architecture, all target objects are independently tracked once the data from multiple sensors is associated or fused [6].…”
Section: Introductionmentioning
confidence: 99%
“…Considering this problem statement, in this paper, we build a logic which switches between the Extended Kalman Filter and the Unscented Kalman Filter, based on the realworld traffic density surrounding the ego-vehicle, and the ego-vehicle's type of Overall, there are a variety of algorithms, present in today's technological world, that may be used for object detection, fusion, and tracking. While there are multiple fusion architectures, for our work, we follow a target-level sensor fusion architecture (also known as centralised fusion architecture) for ease of application and experimentation, as described in the work presented by Deo et al [5]. There are multiple works which provide data regarding which tracking algorithms work best under specific external scenarios.…”
Modern cars utilise Advanced Driver Assistance Systems (ADAS) in several ways. In ADAS, the use of multiple sensors to gauge the environment surrounding the ego-vehicle offers numerous advantages, as fusing information from more than one sensor helps to provide highly reliable and error-free data. The fused data is typically then fed to a tracker algorithm, which helps to reduce noise and compensate for situations when received sensor data is temporarily absent or spurious, or to counter the offhand false positives and negatives. The performances of these constituent algorithms vary vastly under different scenarios. In this paper, we focus on the variation in the performance of tracker algorithms in sensor fusion due to the alteration in external conditions in different scenarios, and on the methods for countering that variation. We introduce a sensor fusion architecture, where the tracking algorithm is spontaneously switched to achieve the utmost performance under all scenarios. By employing a Real-time Traffic Density Estimation (RTDE) technique, we may understand whether the ego-vehicle is currently in dense or sparse traffic conditions. A highly dense traffic (or congested traffic) condition would mean that external circumstances are non-linear; similarly, sparse traffic conditions would mean that the probability of linear external conditions would be higher. We also employ a Traffic Sign Recognition (TSR) algorithm, which is able to monitor for construction zones, junctions, schools, and pedestrian crossings, thereby identifying areas which have a high probability of spontaneous, on-road occurrences. Based on the results received from the RTDE and TSR algorithms, we construct a logic which switches the tracker of the fusion architecture between an Extended Kalman Filter (for linear external scenarios) and an Unscented Kalman Filter (for non-linear scenarios). This ensures that the fusion model always uses the tracker that is best suited for its current needs, thereby yielding consistent accuracy across multiple external scenarios, compared to the fusion models that employ a fixed single tracker.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.