Video based applications disclosed need for efficiently extracting and modeling the video contents. The video features can be classified into normal data, relative features and logic content. Semantic level understanding is required for core content of video. So to get video content automatic semantic content framework is proposed. In proposed system a semantic content extraction system that allows the user to query and regain objects, events, and concepts that are extracted automatically. VISCOM is a video semantic content model which contains classes and relations between classes. Objects and events are represented by some VISCOM classes and other classes are used in the automatic semantic content extraction framework. VISCOM classes collect the semantic content types and relations.Ontology based fuzzy video data semantic model which uses spatial and temporal relations in event and concept definition is proposed. Extracted objects from consecutive representative frames are processed to extract temporal relations. Additional rules to lower spatial relation computation cost and to define some difficult situations more successfully are used. To extract objects from video we apply k-means clustering algorithm. By, which we get the more relevant objects related to user query.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.