2018
DOI: 10.48550/arxiv.1809.01984
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Training Millions of Personalized Dialogue Agents

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(29 citation statements)
references
References 0 publications
0
28
0
Order By: Relevance
“…To train the model, a cross entropy loss is used. Similar to Mazaré et al (2018), during training we consider the other elements of the batch as negatives.…”
Section: Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…To train the model, a cross entropy loss is used. Similar to Mazaré et al (2018), during training we consider the other elements of the batch as negatives.…”
Section: Modelsmentioning
confidence: 99%
“…Chit-chat agents, by contrast, might focus on coarse statistical regularities of dialogue data without accurately modeling the underlying "meaning"; but the data often covers a much wider space of natural language. For example, Twitter or Reddit chit-chat tasks (Li et al, 2016a;Yang et al, 2018;Mazaré et al, 2018) cover a huge spectrum of language and diverse topics. Chit-chat and goal-oriented dialogue are not mutually exclusive: when humans engage in chit-chat, their aim is to exchange information, or to elicit specific responses from their partners.…”
Section: Introductionmentioning
confidence: 99%
“…However, we use a straightforward strategy that directly concatenates the speaker's name with the corresponding utterance. This strategy is inspired by recent research in personalized dialogue modeling that use persona information to represent speak- ers (Li et al, 2016;Zhang et al, 2018b;Mazaré et al, 2018). In subsection 5.2, we will empirically demonstrate its superiority over the feature-based method in Lee et al (2017).…”
Section: Input Representationsmentioning
confidence: 99%
“…We compare our speaker modeling strategy (denoted by speaker as input), which directly concatenates the speaker's name with the corresponding utterance, with the strategy in Wiseman et al larger number of speakers. Compared with the coarse modeling of whether two utterances are from the same speaker, a speaker's name can be thought of as speaker ID in persona dialogue learning (Li et al, 2016;Zhang et al, 2018b;Mazaré et al, 2018). Representations learned for names have the potential to better generalize the global information of the speakers in the multi-party dialogue situation, leading to better context modeling and thus better results.…”
Section: Analyses On Speaker Modeling Strategiesmentioning
confidence: 99%
“…Further developments propose to use a speaker embedding vector in neural models to capture the implicit speaking style of an individual speaker [20,23,31,46,48], or the style of a group of speakers [45]. Other approaches also attempt to endow dialogue models with personae which are described by natural language sentences [26,47].…”
Section: Introductionmentioning
confidence: 99%