2016
DOI: 10.1109/tpami.2016.2537323
|View full text |Cite
|
Sign up to set email alerts
|

Semantic Event Fusion of Different Visual Modality Concepts for Activity Recognition

Abstract: Combining multimodal concept streams from heterogeneous sensors is a problem superficially explored for activity recognition. Most studies explore simple sensors in nearly perfect conditions, where temporal synchronization is guaranteed. Sophisticated fusion schemes adopt problem-specific graphical representations of events that are generally deeply linked with their training data and focused on a single sensor. This paper proposes a hybrid framework between knowledge-driven and probabilistic-driven methods fo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
18
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 35 publications
(19 citation statements)
references
References 37 publications
1
18
0
Order By: Relevance
“…Various approaches have been proposed to use knowledge representation for video understanding such as semantic-visual knowledge bases like FrameNet and Imagenet for modeling rich event-centric concepts and their relationships for video event detection [43], a knowledge and probabilistic driven framework for activity recognition [44], semantic representations for event detection [45], [46]. Souza et al deploy objects, actions and their bonds into graphs and use simulated annealing for event inference using temporal connections [47], [48].…”
Section: Knowledge Representation For Video Understandingmentioning
confidence: 99%
“…Various approaches have been proposed to use knowledge representation for video understanding such as semantic-visual knowledge bases like FrameNet and Imagenet for modeling rich event-centric concepts and their relationships for video event detection [43], a knowledge and probabilistic driven framework for activity recognition [44], semantic representations for event detection [45], [46]. Souza et al deploy objects, actions and their bonds into graphs and use simulated annealing for event inference using temporal connections [47], [48].…”
Section: Knowledge Representation For Video Understandingmentioning
confidence: 99%
“…); Calibration and synchronization of multi-modal data; Multi-modal datasets and evaluation metrics; Leisure, security, health and energy applications based on multi-modal data; Multi-modal Affective Computing and Social Signal processing systems; Multi-modal algorithms designed for GPU, smart phones and game consoles. A total of 17 papers were published withiin this Special Issue at IEEE TPAMI [31], [32], [33], [34], [35], [36], [37], [38], [39], [40], [41], [42], [43], [44], [45], [46], [47]. 2017 TPAMI Faces: Currently we are organizing a Special Issue at IEEE Transactions on Pattern Analysis and Machine Intelligence journal in the topic of face analysis.…”
Section: Special Issuesmentioning
confidence: 99%
“…Considering the 100 houses deployment plan and its financial viability, the low cost consumer RGBD camera, Asus Xtion, was selected. Furthermore, as previously mentioned, RGBD devices now represent the state of the art for indoor activity monitoring [9,20,21,22,11]. For the SPHERE project, the camera needs to be coupled with a machine with suitable processing capacity, minimal intrusion on the user, and minimal cost.…”
Section: Hardware Platformmentioning
confidence: 99%
“…Video based systems are efficient for implementing alert systems to detect dangerous events like falls, as in [20]. Furthermore, video data analysis allows one to identify specific actions, long term activities, and behavioural patterns [9], with some exploiting contextual information [11]. While video based platforms offer the opportunity to extract unique, continuous, and rich information from the home environment, they also present a number of disadvantages, such as privacy issues [9], user acceptance and system cost and scalability.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation