Proceedings of the 21th ACM International Conference on Intelligent Virtual Agents 2021
DOI: 10.1145/3472306.3478352
|View full text |Cite
|
Sign up to set email alerts
|

Agents United

Abstract: Figure 1: A group of ASAP (1, 3, 5) and GRETA (2, 4) agents in the same Unity environment. A user can participate in a multiparty group conversation by selecting a response from the menu on the right (in the same interface or on a separate device).

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 28 publications
(31 reference statements)
0
3
0
Order By: Relevance
“…For our persuasive ECA, we implemented the customizable Unity Multipurpose Avatar 2 (UMA2) and seated the model within the user’s social space (ie, 2 m). We implemented the Agents United Dialogue Platform (ie, dialog manager, including generation of behavior markup language [BML]) [ 42 ], multimodal BML realizer Artificial Social Agents Platform (ASAP) [ 43 , 44 ], and related ASAP-Unity Bridge [ 45 ]. This includes Microsoft’s text-to-speech (TTS) engine (ie, Dutch Frank) with body language (ie, lip sync, gestures, gaze), partially scripted using the WOOL platform (WOOL Foundation; ie, gesture, gaze) [ 42 ], with gestures and posture animations using Unity’s built-in Mecanim system.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…For our persuasive ECA, we implemented the customizable Unity Multipurpose Avatar 2 (UMA2) and seated the model within the user’s social space (ie, 2 m). We implemented the Agents United Dialogue Platform (ie, dialog manager, including generation of behavior markup language [BML]) [ 42 ], multimodal BML realizer Artificial Social Agents Platform (ASAP) [ 43 , 44 ], and related ASAP-Unity Bridge [ 45 ]. This includes Microsoft’s text-to-speech (TTS) engine (ie, Dutch Frank) with body language (ie, lip sync, gestures, gaze), partially scripted using the WOOL platform (WOOL Foundation; ie, gesture, gaze) [ 42 ], with gestures and posture animations using Unity’s built-in Mecanim system.…”
Section: Methodsmentioning
confidence: 99%
“…We implemented the Agents United Dialogue Platform (ie, dialog manager, including generation of behavior markup language [BML]) [ 42 ], multimodal BML realizer Artificial Social Agents Platform (ASAP) [ 43 , 44 ], and related ASAP-Unity Bridge [ 45 ]. This includes Microsoft’s text-to-speech (TTS) engine (ie, Dutch Frank) with body language (ie, lip sync, gestures, gaze), partially scripted using the WOOL platform (WOOL Foundation; ie, gesture, gaze) [ 42 ], with gestures and posture animations using Unity’s built-in Mecanim system. For user input, we integrated a dialog UI instead of speech recognition to avoid usability issues, given that the reliability of speech-to-text engines remains troublesome.…”
Section: Methodsmentioning
confidence: 99%
“…Relevant previous research projects in this field include Humaine (Online: ht tps://cordis.europa.eu/project/id/507422 (accessed on 30 June 2022)), which focused on emotional human-machine interaction and provided an extensive corpus of data on the forms emotion can take on during conversations [30], as well as Semaine (Online: https://cordis.europa.eu/project/id/211486 (accessed on 30 June 2022)), which explored the impact of nonverbal expressions such as head gestures [31] and laughter [32]. Technical artifacts resulting from these projects, such as GRETA [33] or the Agents United platform [34], help researchers and developers to setup their own multiagent applications. Furthermore, focusing more on the healthcare domain, the SimSensei system showed how CAs may be used as a tool to measure psychological distress in semi-structured interviews [35].…”
Section: Social Intelligence and Conversational Agentsmentioning
confidence: 99%
“…Some of these applications may favor fully automated approaches, such as when only an input/output emotion label is needed [13], [14]. Other applications might need a set of intuitive parameters associated with values for each expression, in order to edit or analyze them [15], [16], [17], [18], [19].…”
Section: Introductionmentioning
confidence: 99%