skip to main content
10.1145/3284432.3287190acmconferencesArticle/Chapter ViewAbstractPublication PageshaiConference Proceedingsconference-collections
abstract

Integration of DNN Generated Spontaneous Reactions with a Generic Multimodal Framework for Embodied Conversational Agents

Published: 04 December 2018 Publication History

Abstract

This paper describes the recent extensions of our previously proposed GECA Framework in incorporating the interconnection with GECA original networking library and ZeroMQ message passing library, the connection with Python based DNN developing API, Keras, and a character animator supporting FACS based detailed facial animations and lip-syncing with real-time generated voice tracks.

References

[1]
Tadas Baltrusaitis, Chaitanya Ahuja, and Louis-Philippe Morency. 2017. Multimodal Machine Learning: A Survey and Taxonomy. CoRR, Vol. abs/1705.09406 (2017). arxiv: 1705.09406 http://arxiv.org/abs/1705.09406
[2]
Paul Ekman, Wallace V. Friesen, and Joseph C. Hager. 2002. Facial Action Coding System (FACS). Website. http://www.face-and-emotion.com/dataface/facs/description.jsp
[3]
Hung-Hsuan Huang, Aleksandra Cerekovic, Yukiko Nakano, Igor S. Pandzic, and Toyoaki Nishida. 2008a. The Design of a Generic Framework for Integrating ECA Components. In The 7th International Conference of Autonomous Agents and Multiagent Systems (AAMAS'08), Lin Padgham, David Parkes, and Jorg P. Muller (Eds.). ACM, IFAAMAS, Estorial, Portugal, 128--135.
[4]
Hung-Hsuan Huang, Aleksandra Cerekovic, Kateryna Tarasenko, Vjekoslav Levacic, Goranka Zoric, Igor S. Pandzic, Yukiko Nakano, and Toyoaki Nishida. 2008b. Integrating Embodied Conversational Agent Components with a Generic Framework. Multiagent and Grid Systems, Vol. 4, 4 (Dec. 2008), 371--386.
[5]
David R. Traum and Staffan Larsson. 2003. The information state approach to dialogue management. Kluwer, Dordrecht, 325--353.
[6]
Stef van der Struijk, Hung-Hsuan Huang, Maryam Sadat Mirzaei, and Toyoaki Nishida. 2018. FACSvatar: An Open Source Modular Framework for Real-Time FACS based Facial Animation. In 18th ACM International Conference on Intelligent Virtual Agents (IVA 2018). Sydney, Australia.

Cited By

View all
  • (2019)An Investigation on the Effectiveness of Multimodal Fusion and Temporal Feature Extraction in Reactive and Spontaneous Behavior Generative RNN Models for Listener AgentsProceedings of the 7th International Conference on Human-Agent Interaction10.1145/3349537.3351908(89-96)Online publication date: 25-Sep-2019
  • (2019)Development of a Platform for RNN Driven Multimodal Interaction with Embodied Conversational AgentsProceedings of the 19th ACM International Conference on Intelligent Virtual Agents10.1145/3308532.3329448(200-202)Online publication date: 1-Jul-2019
  • (2019)Toward RNN Based Micro Non-verbal Behavior Generation for Virtual Listener AgentsSocial Computing and Social Media. Design, Human Behavior and Analytics10.1007/978-3-030-21902-4_5(53-63)Online publication date: 8-Jun-2019

Index Terms

  1. Integration of DNN Generated Spontaneous Reactions with a Generic Multimodal Framework for Embodied Conversational Agents

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      HAI '18: Proceedings of the 6th International Conference on Human-Agent Interaction
      December 2018
      402 pages
      ISBN:9781450359535
      DOI:10.1145/3284432
      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 04 December 2018

      Check for updates

      Author Tags

      1. embodied conversational agents
      2. facial expression
      3. facs
      4. multimodal interaction
      5. zeromq

      Qualifiers

      • Abstract

      Conference

      HAI '18
      Sponsor:
      HAI '18: 6th International Conference on Human-Agent Interaction
      December 15 - 18, 2018
      Southampton, United Kingdom

      Acceptance Rates

      HAI '18 Paper Acceptance Rate 40 of 92 submissions, 43%;
      Overall Acceptance Rate 121 of 404 submissions, 30%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)1
      • Downloads (Last 6 weeks)0
      Reflects downloads up to 29 Jan 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2019)An Investigation on the Effectiveness of Multimodal Fusion and Temporal Feature Extraction in Reactive and Spontaneous Behavior Generative RNN Models for Listener AgentsProceedings of the 7th International Conference on Human-Agent Interaction10.1145/3349537.3351908(89-96)Online publication date: 25-Sep-2019
      • (2019)Development of a Platform for RNN Driven Multimodal Interaction with Embodied Conversational AgentsProceedings of the 19th ACM International Conference on Intelligent Virtual Agents10.1145/3308532.3329448(200-202)Online publication date: 1-Jul-2019
      • (2019)Toward RNN Based Micro Non-verbal Behavior Generation for Virtual Listener AgentsSocial Computing and Social Media. Design, Human Behavior and Analytics10.1007/978-3-030-21902-4_5(53-63)Online publication date: 8-Jun-2019

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media