The 23rd IEEE International Symposium on Robot and Human Interactive Communication 2014
DOI: 10.1109/roman.2014.6926289
|View full text |Cite
|
Sign up to set email alerts
|

Applying affective feedback to reinforcement learning in ZOEI, a comic humanoid robot

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 11 publications
0
7
0
Order By: Relevance
“…We identify three types of evaluation methodologies: (1) an evaluation from the algorithm point of view, (2) evaluation and assessment of user experience-related subjective measures, and (3) evaluation of both learning algorithm-related factors and user experience-related factors. Several studies only reported the self-rated questionnaire results [45] or user opinions [55]. There are also studies which do not include any evaluation, and only a short discussion regarding the learned policy [53,57,100,101] The cumulative collected reward over time is the most commonly used evaluation method.…”
Section: Evaluation Methodsologiesmentioning
confidence: 99%
See 2 more Smart Citations
“…We identify three types of evaluation methodologies: (1) an evaluation from the algorithm point of view, (2) evaluation and assessment of user experience-related subjective measures, and (3) evaluation of both learning algorithm-related factors and user experience-related factors. Several studies only reported the self-rated questionnaire results [45] or user opinions [55]. There are also studies which do not include any evaluation, and only a short discussion regarding the learned policy [53,57,100,101] The cumulative collected reward over time is the most commonly used evaluation method.…”
Section: Evaluation Methodsologiesmentioning
confidence: 99%
“…Q-learning, along with its different variations, is the most commonly used RL method in social robotics. The studies using Q-learning are [ 3 , 13 , 34 , 52 , 53 , 54 , 55 , 56 , 57 , 58 , 59 , 60 , 61 ]. These comprise studies using standard Q-learning [ 3 , 54 , 55 , 58 , 60 , 62 ], studies modify Q-learning for dealing with delayed reward [ 52 ], tuning the parameters for Q-learning such as [ 13 , 34 , 52 ], dealing with decreasing human feedback over time [ 52 ], comparing their proposed algorithm with Q-learning [ 33 , 49 , 61 , 63 , 64 ], variation of Q-learning called Object Q-learning [ 64 , 65 , 66 ], combining Q-learning with fuzzy inference [ 67 ], SARSA [ 68 , 69 ], TD( ) [ 70 ], MAXQ [ 33 , 71 , 72 ], R-learning [ 32 ], and Deep Q-learning [ 35 , 36 , 73 , 74 ].…”
Section: Categorization Of Rl Approaches In Social Robotics Based mentioning
confidence: 99%
See 1 more Smart Citation
“…These metrics can be incorporated in creating the audience ratings for motions. Feedback from the audience ranges from using visual cues such as the audience holding colored markers such as paddles [Knight et al, 2011] to indicate their preference or audio feedback such as the applause or cheers from the audience or surveys at the end of the interaction [Addo and Ahamed, 2014]. These feedback can be converted into a noisy numerical value to model the distribution of the observed audience preference.…”
Section: Selection Of Motions Using Audience Preferences Of Motion Sementioning
confidence: 99%
“…The approach of Knight et al assumes that the audience preference on features do not vary over time, whereas we model the boredom of the audience and account for different weights assigned to the motions in different sequences. Addo and Ahamed also use a robot to tell jokes, but use reinforcement learning [Addo and Ahamed, 2014]. However, to learn a good policy, they have to explore all jokes in all the states, whereas we do not have to query all sequences to pick the best sequence.…”
Section: Selection Of Motions Using Audience Preferences Of Motion Sementioning
confidence: 99%