A generic and robust approach for the real-time detection of people and vehicles from an Unmanned Aerial Vehicle (UAV) is an important goal within the framework of fully autonomous UAV deployment for aerial reconnaissance and surveillance. Here we present an approach for the automatic detection of vehicles based on using multiple trained cascaded Haar classifiers with secondary confirmation in thermal imagery. Additionally we present a related approach for people detection in thermal imagery based on a similar cascaded classification technique combining additional multivariate Gaussian shape matching. The results presented show the successful detection of vehicle and people under varying conditions in both isolated rural and cluttered urban environments with minimal false positive detection. Performance of the detector is optimized to reduce the overall false positive rate by aiming at the detection of each object of interest (vehicle/person) at least once in the environment (i.e. per search patter flight path) rather than every object in each image frame. Currently the detection rate for people is ~70% and cars ~80% although the overall episodic object detection rate for each flight pattern exceeds 90%.
Generalised wide are search and surveillance is a common-place tasking for multi-sensory equipped autonomous systems. Here we present on a key supporting topic to this task -the automatic interpretation, fusion and detected target reporting from multi-modal sensor information received from multiple autonomous platforms deployed for wide-area environment search. We detail the realization of a real-time methodology for the automated detection of people and vehicles using combined visible-band (EO), thermal-band (IR) and radar sensing from a deployed network of multiple autonomous platforms (ground and aerial). This facilities real-time target detection, reported with varying levels of confidence, using information from both multiple sensors and multiple sensor platforms to provide environment-wide situational awareness. A range of automatic classification approaches are proposed, driven by underlying machine learning techniques, that facilitate the automatic detection of either target type with cross-modal target confirmation. Extended results are presented that show both the detection of people and vehicles under varying conditions in both isolated rural and cluttered urban environments with minimal false positive detection. Performance evaluation is presented at an episodic level with individual classifiers optimized for maximal each object of interest (vehicle/person) detection over a given search path/pattern of the environment, across all sensors and modalities, rather than on a per sensor sample basis. Episodic target detection, evaluated over a number of wide-area environment search and reporting tasks, generally exceeds 90%+ for the targets considered here.
Here we investigate the automatic detection of fire pixel regions in conventional video (or still) imagery within realtime bounds. As an extension to prior, established approaches within this field we specifically look to extend the primary use of threshold-driven colour spectroscopy to the combined use of colour-texture feature descriptors as an input to a trained classification approach that is independent of temporal information. We show the limitations of such spectroscopy driven approaches on simple, real-world examples and propose our novel extension as a robust, real-time solution within this field by combining simple texture descriptors to illustrate maximal 98% fire region detection.
We address the challenge of human behaviour analysis within automated image understanding. Whilst prior work concentrates on this task within visible-band (EO) imagery, by contrast we target basic human pose classification in thermal-band (infrared, IR) imagery. By leveraging the key advantages of limb localization this imagery offers we target two distinct human pose classification problems of varying complexity: 1) identifying passive or active individuals within the scene and 2) the identification of individuals potentially carrying weapons. Both approaches use a discrete set of features capturing body pose characteristics from which a range of machine learning techniques are then employed for final classification. Significant success is shown on these challenging tasks over a wide range of environmental conditions within the wider context of automated human target tracking in thermal-band (IR) imagery.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.