2021
DOI: 10.48550/arxiv.2109.05778
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Text is NOT Enough: Integrating Visual Impressions into Open-domain Dialogue Generation

Lei Shen,
Haolan Zhan,
Xin Shen
et al.

Abstract: Open-domain dialogue generation in natural language processing (NLP) is by default a pure-language task, which aims to satisfy human need for daily communication on open-ended topics by producing related and informative responses. In this paper, we point out that hidden images, named as visual impressions (VIs), can be explored from the text-only data to enhance dialogue understanding and help generate better responses. Besides, the semantic dependency between an dialogue post and its response is complicated, … Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 43 publications
(59 reference statements)
0
1
0
Order By: Relevance
“…However, the generated responses may suffer from some notorious problems, such as being generic, inconsistent, or unrelated to the given contexts. Previous studies tried to solve these issues by feeding extra information, e.g., topics [25], sentence types [36], personas [19], emotions [43,60], documents [30], multi-modal [46] or knowledge [17,27,55,56,61], augmenting the model itself [54,59], or modifying the loss function [18].…”
Section: Introductionmentioning
confidence: 99%
“…However, the generated responses may suffer from some notorious problems, such as being generic, inconsistent, or unrelated to the given contexts. Previous studies tried to solve these issues by feeding extra information, e.g., topics [25], sentence types [36], personas [19], emotions [43,60], documents [30], multi-modal [46] or knowledge [17,27,55,56,61], augmenting the model itself [54,59], or modifying the loss function [18].…”
Section: Introductionmentioning
confidence: 99%