2019
DOI: 10.1007/s42452-019-1870-9
|View full text |Cite
|
Sign up to set email alerts
|

Semantic concept based video retrieval using convolutional neural network

Abstract: Retrieval of videos efficiently and effectively has become a challenging issue nowadays and dealing with multi-concept videos is the center of focus. The aim of the work presented here is to propose an improved semantic concept-based video retrieval method using a novel ranked intersection filtering technique and a foreground driven concept co-occurrence matrix. In the proposed ranked intersection filtering technique, an intersection of ranked concept probability scores is taken from key-frames associated with… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 27 publications
0
2
0
Order By: Relevance
“…Paper [15] proposed a nested invariant pool (NIP) method to obtain compact and discriminative convolutional neural network video retrieval descriptors, which mainly acts on the middle feature mapping of CNN in a nested manner, reducing the feature size of CNN, boosting the geometric invariance of depth descriptors, and significantly improving the performance of video retrieval. In [16] uses a classififier fusing asymmetric training depth CNNs to deal with data imbalance to achieve effificient video retrieval. The dual encoding network proposed in reference [17] encodes inputs in a similar manner.…”
Section: Introductionmentioning
confidence: 99%
“…Paper [15] proposed a nested invariant pool (NIP) method to obtain compact and discriminative convolutional neural network video retrieval descriptors, which mainly acts on the middle feature mapping of CNN in a nested manner, reducing the feature size of CNN, boosting the geometric invariance of depth descriptors, and significantly improving the performance of video retrieval. In [16] uses a classififier fusing asymmetric training depth CNNs to deal with data imbalance to achieve effificient video retrieval. The dual encoding network proposed in reference [17] encodes inputs in a similar manner.…”
Section: Introductionmentioning
confidence: 99%
“…Finally, the classification scores of each shot are fused to obtain the final class score. Similar studies are based on the shot boundary detection and keyframe extraction technique 11,12 . In such studies, video semantics are only considered on the frame level.…”
Section: Introductionmentioning
confidence: 99%
“…Similar studies are based on the shot boundary detection and keyframe extraction technique. 11,12 In such studies, video semantics are only considered on the frame level. There have been various studies considering spatio-temporal characteristics.…”
mentioning
confidence: 99%