Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1361
|View full text |Cite
|
Sign up to set email alerts
|

Playing 20 Question Game with Policy-Based Reinforcement Learning

Abstract: The 20 Questions (Q20) game is a well known game which encourages deductive reasoning and creativity. In the game, the answerer first thinks of an object such as a famous person or a kind of animal. Then the questioner tries to guess the object by asking 20 questions. In a Q20 game system, the user is considered as the answerer while the system itself acts as the questioner which requires a good strategy of question selection to figure out the correct object and win the game. However, the optimal policy of que… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 24 publications
(23 citation statements)
references
References 8 publications
0
22
0
Order By: Relevance
“…There are several mainstream methods in the DRL framework including Deep Q-Network (Mnih et al 2015) and Policy Networks (Silver et al 2016). Besides, DRL is widely used in many NLP tasks (Wu, Li, and Wang 2018;Feng et al 2018;Li et al 2019;Narasimhan, Kulkarni, and Barzilay 2015;He et al 2015;Hu et al 2018a). These works prove the rationality and effectiveness of applying DRL to NLP tasks, which support our work on LJP.…”
Section: Related Work Deep Reinforcement Learningmentioning
confidence: 99%
“…There are several mainstream methods in the DRL framework including Deep Q-Network (Mnih et al 2015) and Policy Networks (Silver et al 2016). Besides, DRL is widely used in many NLP tasks (Wu, Li, and Wang 2018;Feng et al 2018;Li et al 2019;Narasimhan, Kulkarni, and Barzilay 2015;He et al 2015;Hu et al 2018a). These works prove the rationality and effectiveness of applying DRL to NLP tasks, which support our work on LJP.…”
Section: Related Work Deep Reinforcement Learningmentioning
confidence: 99%
“…Language-based interaction has been studied in the context of visual question answering (de Vries et al, 2017;Chattopadhyay et al, 2017;Lee et al, 2019;Shukla et al, 2019), SQL generation (Gur et al, 2018;Yao et al, 2019), information retrieval (Chung et al, 2018;Aliannejadi et al, 2019) and multi-turn textbased question answering (Rao and Daumé III, 2018;Reddy et al, 2019;Choi et al, 2018). Most methods require learning from recorded dialogues Hu et al, 2018;Rao and Daumé III, 2018) or conducting Wizard-of-Oz dialog annotations (Kelley, 1984;Wen et al, 2017). Instead, we limit the interaction to multiple-choice and binary questions.…”
Section: Related Workmentioning
confidence: 99%
“…This simplification allows us to reduce the complexity of data annotation while still achieving effective interaction. Our task can be viewed as an instance of the popular 20-question game (20Q), which has been applied to a celebrities knowledge base Hu et al, 2018). Our approach differs in using natural language descriptions of classification targets, questions and answers to compute our distributions, instead of treating them as categorical or structural data.…”
Section: Related Workmentioning
confidence: 99%
“…For different purposes, there are various question generation tasks. Hu et al (2018) aim to ask questions to play the 20 question game. Dhingra et al (2017) teach models to ask questions to limit the number of answer candidates in task-oriented dialogues.…”
Section: Question Generationmentioning
confidence: 99%