2013
DOI: 10.1142/s0129065713500214
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Integration of Micro-Doppler Sonar and Auditory Signals for Behavior Classification With Convolutional Networks

Abstract: The ability to recognize the behavior of individuals is of great interest in the general field of safety (e.g. building security, crowd control, transport analysis, independent living for the elderly). Here we report a new real-time acoustic system for human action and behavior recognition that integrates passive audio and active micro-Doppler sonar signatures over multiple time scales. The system architecture is based on a six-layer convolutional neural network, trained and evaluated using a dataset of 10 sub… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
4
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 19 publications
0
4
0
Order By: Relevance
“…Range sensing is another contactless ambient modality based on the remote measurement of distances. Commonly employed sensors are pyroelectric infrared (PIR) [ 38 , 39 , 40 ], sonar [ 41 ], Lidar/range camera [ 42 , 43 ], radar [ 44 ] and Wi-Fi systems [ 45 ]. Acceptability and performance are quite good, especially in the case of Range camera and Radar, since depth maps and radar scans are not able to capture privacy-sensitive information.…”
Section: Related Workmentioning
confidence: 99%
“…Range sensing is another contactless ambient modality based on the remote measurement of distances. Commonly employed sensors are pyroelectric infrared (PIR) [ 38 , 39 , 40 ], sonar [ 41 ], Lidar/range camera [ 42 , 43 ], radar [ 44 ] and Wi-Fi systems [ 45 ]. Acceptability and performance are quite good, especially in the case of Range camera and Radar, since depth maps and radar scans are not able to capture privacy-sensitive information.…”
Section: Related Workmentioning
confidence: 99%
“…According to the different types of data used for human action recognition, human action recognition is usually divided into the following three main categories: human action recognition based on vision, human action recognition based on acoustics, and human action recognition based on inertial sensors [ 13 ]. Visual action recognition extracts human action data from image or video data obtained by optical sensors [ 14 , 15 ], acoustic action recognition uses sound signals for high-precision hand action tracking and gesture recognition [ 16 , 17 ], inertial sensor action recognition focuses on extracting human action inertial data from wearable inertial sensors [ 18 , 19 ]. Liu Yutao [ 20 ] summarized the three methods, as shown in the following Table 1 .…”
Section: Introductionmentioning
confidence: 99%
“…If the object itself contains moving parts, each moving part will result in a modulation of the base Doppler frequency shift, known as the micro‐Doppler effect. A micro‐Doppler sonar was reported in [1] to produce signal signatures of human gait and subsequently employed in behaviour analysis and classification [2].…”
Section: Introductionmentioning
confidence: 99%
“…If the object itself contains moving parts, each moving part will result in a modulation of the base Doppler frequency shift, known as the micro-Doppler effect. A micro-Doppler sonar was reported in [1] to produce signal signatures of human gait and subsequently employed in behaviour analysis and classification [2].The sonar micro-Doppler system in [1] employs principles also found in echolocating animals such as bats and dolphins that emit mechanical waves that bounce onto objects in an environment and allow them to place themselves in space, navigate and even hunt [3,4]. Echolocating bats can fall into two different categories, the ones emitting a frequency-modulated ultrasound signal, known as FM, and the ones emitting a constant frequency signal, known as CF.…”
mentioning
confidence: 99%