2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI) 2015
DOI: 10.1109/cbmi.2015.7153604
|View full text |Cite
|
Sign up to set email alerts
|

VSD2014: A dataset for violent scenes detection in hollywood movies and web videos

Abstract: In this paper, we introduce a violent scenes and violence-related concept detection dataset named VSD2014. It contains annotations as well as auditory and visual features of typical Hollywood movies and user-generated footage shared on the web. The dataset is the result of a joint annotation endeavor of different research institutions and responds to the real-world use case of parental guidance in selecting appropriate content for children. The dataset has been validated during the Violent Scenes Detection (VS… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
21
0
2

Year Published

2015
2015
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 23 publications
(23 citation statements)
references
References 22 publications
0
21
0
2
Order By: Relevance
“…At the beginning, the algorithm selects groups G 3 , G 4 and G 5 in practically 100% of the cases because of the low threshold imposed (1 MaxMFLOPS). When we increase this value to 3 MaxMFLOPS the spectral features appear.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…At the beginning, the algorithm selects groups G 3 , G 4 and G 5 in practically 100% of the cases because of the low threshold imposed (1 MaxMFLOPS). When we increase this value to 3 MaxMFLOPS the spectral features appear.…”
Section: Resultsmentioning
confidence: 99%
“…In this sense, violence can be detected through audio and video surveillance. Some works in the literature treat this problem using both audio and video processing, 3,4,5 and the results obtained with the combination of those sources seems to be efficient.…”
Section: Introductionmentioning
confidence: 99%
“…This approach outperformed the other methods on the second VSD sub-task (i.e., violence detection in user generated videos from YouTube). The most common features used by most of the participating teams were MFCC (audio) and dense trajectories (visual+temporal) [17].…”
Section: Introductionmentioning
confidence: 99%
“…The Affect Task of MediaEval has provided a common ground for researchers to work on this problem and compare their algorithms in an efficient way. A publicly available dataset provides a detailed annotation ground truth of multiple audio and visual concepts concerning violence [17]. In MediaEval 2014, many teams participated for the VSD task [17].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation