2018
DOI: 10.1007/978-3-030-00692-1_25
|View full text |Cite
|
Sign up to set email alerts
|

Extracting Textual Overlays from Social Media Videos Using Neural Networks

Abstract: Textual overlays are often used in social media videos as people who watch them without the sound would otherwise miss essential information conveyed in the audio stream. This is why extraction of those overlays can serve as an important meta-data source, e.g. for content classification or retrieval tasks. In this work, we present a robust method for extracting textual overlays from videos that builds up on multiple neural network architectures. The proposed solution relies on several processing steps: keyfram… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 10 publications
0
1
0
Order By: Relevance
“…Therefore, visible banners that show up are annotated in parallel to Valence. In a later stage we want to detect, and measure the influence on visual features or exclude/ replace them [68], [69], [70].…”
Section: Bannersmentioning
confidence: 99%
“…Therefore, visible banners that show up are annotated in parallel to Valence. In a later stage we want to detect, and measure the influence on visual features or exclude/ replace them [68], [69], [70].…”
Section: Bannersmentioning
confidence: 99%