2019
DOI: 10.1007/978-3-030-37734-2_32
|View full text |Cite
|
Sign up to set email alerts
|

Interactive Search and Exploration in Discussion Forums Using Multimodal Embeddings

Abstract: In this paper we present a novel interactive multimodal learning system, which facilitates search and exploration in large networks of social multimedia users. It allows the analyst to identify and select users of interest, and to find similar users in an interactive learning setting. Our approach is based on novel multimodal representations of users, words and concepts, which we simultaneously learn by deploying a general-purpose neural embedding model. We show these representations to be useful not only for … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 27 publications
0
1
0
Order By: Relevance
“…From an algorithmic point-of-view, a major issue in MMIS relates to the fusion of several modalities (e.g., text and image) to obtain a meaningful representation. Recent state-of-the-art techniques employ joint representation techniques to find a latent space in which multiple modality information can be projected and compared [26]. This can be a challenging task, e.g., while content data such as text and images might be well-aligned, obtaining a latent space to perform joint content and user-preference representation (e.g., ratings, clicks, social-media information) might not be trivial.…”
Section: Multi-modal Information Seekingmentioning
confidence: 99%
“…From an algorithmic point-of-view, a major issue in MMIS relates to the fusion of several modalities (e.g., text and image) to obtain a meaningful representation. Recent state-of-the-art techniques employ joint representation techniques to find a latent space in which multiple modality information can be projected and compared [26]. This can be a challenging task, e.g., while content data such as text and images might be well-aligned, obtaining a latent space to perform joint content and user-preference representation (e.g., ratings, clicks, social-media information) might not be trivial.…”
Section: Multi-modal Information Seekingmentioning
confidence: 99%