2018
DOI: 10.3390/jimaging4070089
|View full text |Cite
|
Sign up to set email alerts
|

Digital Comics Image Indexing Based on Deep Learning

Abstract: Abstract:The digital comic book market is growing every year now, mixing digitized and digital-born comics. Digitized comics suffer from a limited automatic content understanding which restricts online content search and reading applications. This study shows how to combine state-of-the-art image analysis methods to encode and index images into an XML-like text file. Content description file can then be used to automatically split comic book images into sub-images corresponding to panels easily indexable with … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 41 publications
(21 citation statements)
references
References 67 publications
(150 reference statements)
0
21
0
Order By: Relevance
“…Four different object types have been detected in [27]. There are also studies on specialized network for comic face detection [28,29] or comic character detection [30].…”
Section: Related Workmentioning
confidence: 99%
“…Four different object types have been detected in [27]. There are also studies on specialized network for comic face detection [28,29] or comic character detection [30].…”
Section: Related Workmentioning
confidence: 99%
“…Another subset of the DCM was fully annotated by humans. Nguyen, Rigaud, and Burie () created DCM_772, a subset of 772 pages from the Digital Comic Museum stratified by publisher. Ground truth is available in the form of object bounding boxes around characters, which are further differentiated into four classes: human‐like, object‐like, animal‐like, and extra (supporting role characters), making this data set potentially suitable for differentiated character detection and recognition.…”
Section: Data Setsmentioning
confidence: 99%
“…Nguyen et al () used an off‐the‐shelf YOLOv2 region proposal network (Redmon & Farhadi, ) with some adjusted priors to detect bounding boxes around panels, characters, and faces. For panels, they achieved comparable or slightly better performance than feature engineering techniques, especially when the testing material was from the same set of comics books as the training material.…”
Section: Analysis Of Visual Structurementioning
confidence: 99%
See 1 more Smart Citation
“…The ever-popular Convolutional Neural Networks (CNN) and its derivations are often used in these said usages as they show great potential in dealing with images. The researches carried out in [20], [21] and [22] through the usage of the object detection models of YOLOv2 [23], a customised Faster R-CNN [24] model and Mask R-CNN [25] respectively, stand as a testimony for this fact.…”
Section: Introductionmentioning
confidence: 99%