2017
DOI: 10.1016/j.eswa.2016.10.049
|View full text |Cite
|
Sign up to set email alerts
|

Combining heterogeneous sources in an interactive multimedia content retrieval model

Abstract: Interactive multimodal information retrieval systems (IMIR) increase the capabilities of traditional search systems, by adding the ability to retrieve information of different types (modes) and from different sources. This article describes a formal model for interactive multimodal information retrieval. This model includes formal and widespread definitions of each component of an IMIR system. A use case that focuses on information retrieval regarding sports validates the model, by developing a prototype that … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(9 citation statements)
references
References 45 publications
0
7
0
Order By: Relevance
“…The text contained in an image is extracted using Tesseract 14 . This is a standard Optical Character Recognition Tool used in many other information retrieval studies as [34][35][36]. The defects highlighted by the study [37] requiring prior training will have been limited by a prior filtering of the scanned files according to their qualities.…”
Section: Application With Textual Contentmentioning
confidence: 99%
“…The text contained in an image is extracted using Tesseract 14 . This is a standard Optical Character Recognition Tool used in many other information retrieval studies as [34][35][36]. The defects highlighted by the study [37] requiring prior training will have been limited by a prior filtering of the scanned files according to their qualities.…”
Section: Application With Textual Contentmentioning
confidence: 99%
“…These types are vertical, horizontal, diagonal with 45 and 135 degrees, and non-directional edges. Sometimes the sub-image can be categorized as non-edge block [9]. There are future hopes and attempts to develop a joint representation of EHD and CLD [10].…”
Section: Lucene Image Retrievalmentioning
confidence: 99%
“…The challenges of providing sufficient training examples and defining effective search queries in words has seen the investigation of alternatives to text-based queries which use richer forms of query data, and direct comparisons between query data and target videos via more general purpose feature-based similarity measures which bypass concept learning altogether. Examples include: query-by-example where users can provide an example image or video defining their query (e.g., Snoek et al (2007); Yang et al (2013)), queryby-sketch where users can draw their query (e.g., Hu et al (2007)), query-by-object where users can provide an image of an object defining their query (e.g., Sivic & Zisserman (2003)), and combinations thereof (e.g., Moreno-Schneider et al (2017)).…”
Section: Content-based Video Retrievalmentioning
confidence: 99%