Skip to Main content Skip to Navigation
New interface
Poster communications

Multimodal modeling of expressiveness for human-machine interaction

Abstract : Myriad of applications involve the interaction of humans with machines, such as reception agents, home assistants, chatbots or autonomous vehicles’ agents. Humans can control the virtual agents by the mean of various modalities including sound, vision, and touch. As the number of these applications increases, a key problem is the requirement of integrating all modalities, to leverage the interaction’s quality, as well as the user’s experience in the virtual world. In this State-of-the-Art review paper, we discuss about designing engaging virtual agents with expressive gestures and prosody. This paper is part of a work that aims to review the mechanisms that govern multimodal interaction, such as the agent’s expressiveness and the adaptation of its behavior, to help remove technological barriers and develop a conversational agent capable of adapting naturally and coherently to its interlocutor.
Complete list of metadata

Cited literature [39 references]  Display  Hide  Download
Contributor : CCSD Connect in order to contact the contributor
Submitted on : Tuesday, September 8, 2020 - 2:35:43 PM
Last modification on : Monday, May 9, 2022 - 1:22:45 PM
Long-term archiving on: : Saturday, December 5, 2020 - 3:18:54 AM


Files produced by the author(s)


  • HAL Id : hal-02933482, version 1


Mireille Fares, Catherine I Pelachaud, Nicolas Obin. Multimodal modeling of expressiveness for human-machine interaction. Workshop sur les Affects, Compagnons artificiels et Interactions, Jun 2020, Saint Pierre d'Oléron, France. ⟨hal-02933482⟩



Record views


Files downloads