skip to main content
10.1145/3383652.3423877acmconferencesArticle/Chapter ViewAbstractPublication PagesivaConference Proceedingsconference-collections
research-article

A framework to co-optimize task and social dialogue policies using Reinforcement Learning

Published: 19 October 2020 Publication History

Editorial Notes

A Corrected Version of Record for this paper was published on August 29, 2022. Alankar Jain was left off the Version of Record as an author due to an error. For reference purposes, the VoR may still be accessed via the Supplemental Material section on this citation page.

Abstract

One of the main challenges for conversational agents is to select the optimal dialogue policy based on the state of the interaction. This challenge becomes even harder when the conversational agent not only has to achieve a specific task, but also aims at building rapport. Although some work already tried to tackle this challenge using a Reinforcement Learning (RL) approach, they tend to consider one single optimal policy for all the users, regardless of their conversational goals. In this work, we describe a framework that allows us to build a RL-based agent able to adapt its dialogue policy depending on its user's conversational goals. After we build a rule-based agent and a user simulator communicating at the dialog-act level, we crowdsource the surface sentences authoring for both the simulated users and the agent, which allow us to generate a dataset of interactions in natural language. Then, we annotate each of these interactions with a single rapport score and analyze the links between simulated users' conversational goals, agent conversational policies, and rapport. Our results show that rapport was higher when both or none of the interlocutors tried to build rapport. We use this result to inform the design of a social reward function, and we rely on this social reward function to train a RL-based agent using an hybrid approach of supervised learning and reinforcement learning. We evaluate our approach by comparing two different versions of our RL-based agent: one that takes users' conversational goals into account and another that does not. The results show that an agent adapting its dialogue policy depending on users' conversational goals performs better.

Supplementary Material

3423877-vor (3423877-vor.pdf)
Version of Record for "A framework to co-optimize task and social dialogue policies using Reinforcement Learning" by Pecune et al., Proceedings of the 20th ACM International Conference on Intelligent Virtual Agents (IVA '20).

References

[1]
Jaime C Acosta and Nigel G Ward. 2011. Achieving rapport with turn-by-turn, user-responsive emotional coloring. Speech Communication 53, 9-10 (2011).
[2]
Timothy W Bickmore, Laura M Pfeifer Vardoulakis, and Daniel Schulman. 2013. Tinker: a relational agent museum guide. Autonomous agents and multi-agent systems 27, 2 (2013), 254--276.
[3]
Justine Cassell and Timothy Bickmore. 2003. Negotiated collusion: Modeling social language and its relationship effects in intelligent agents. User modeling and user-adapted interaction 13, 1-2 (2003), 89--132.
[4]
Herbert H Clark and Edward F Schaefer. 1989. Contributing to discourse. Cognitive science 13, 2 (1989), 259--294.
[5]
David DeVault, Ron Artstein, Grace Benn, Teresa Dey, Ed Fast, Alesia Gainer, Kallirroi Georgila, Jon Gratch, Arno Hartholt, Margaux Lhommet, et al. 2014. SimSensei Kiosk: A virtual human interviewer for healthcare decision support. In International conference on Autonomous agents and multi-agent systems. International Foundation for Autonomous Agents and Multiagent Systems, 1061--1068.
[6]
Aimee L Drolet and Michael W Morris. 2000. Rapport in conflict resolution: Accounting for how face-to-face contact fosters mutual cooperation in mixed-motive conflicts. Journal of Experimental Social Psychology 36, 1 (2000), 26--50.
[7]
Emmanuel Ferreira and Fabrice Lefèvre. 2015. Reinforcement-learning based dialogue system for human-robot interactions with socially-inspired rewards. Computer Speech and Language 34, 1 (2015), 256--274.
[8]
Kallirroi Georgila, Maria K Wolters, and Johanna D Moore. 2010. Learning dialogue strategies from older and younger simulated users. In Proceedings of the 11th Annual Meeting of the Special Interest Group on Discourse and Dialogue. Association for Computational Linguistics, 103--106.
[9]
Goren Gordon, Samuel Spaulding, Jacqueline Kory Westlund, Jin Joo Lee, Luke Plummer, Marayna Martinez, Madhurima Das, and Cynthia Breazeal. 2016. Affective personalization of a social robot tutor for children's second language skills. In Proceedings of the Conference on Artificial Intelligence (AAAI-16). 3951--3957.
[10]
James Henderson, Oliver Lemon, and Kallirroi Georgila. 2008. Hybrid reinforcement/supervised learning of dialogue policies from fixed data sets. Computational Linguistics 34, 4 (2008), 487--511.
[11]
Alankar Jain, Florian Pecune, Yoichi Matsuyama, and Justine Cassell. 2018. A user simulator architecture for socially-aware conversational agents. In Proceedings of the 18th International Conference on Intelligent Virtual Agents. 133--140.
[12]
Matt Stuttle Steve Young Jost Schatzmann, Karl Weilhammer. 2006. A Survey of Statistical User Simulation Techniques for Reinforcement-Learning of Dialogue Management Strategies. The Knowledge Engineering Review 1, 24 (2006).
[13]
Michael Jugovac and Dietmar Jannach. 2017. Interacting with recom-menders---overview and research directions. ACM Transactions on Interactive Intelligent Systems (TiiS) 7, 3 (2017), 10.
[14]
Tejas D Kulkarni, Karthik Narasimhan, Ardavan Saeedi, and Josh Tenenbaum. 2016. Hierarchical deep reinforcement learning: Integrating temporal abstraction and intrinsic motivation. In Advances in neural information processing systems.
[15]
Iolanda Leite, André Pereira, Ginevra Castellano, Samuel Mascarenhas, Carlos Martinho, and Ana Paiva. 2011. Modelling Empathy in Social Robotic Companions. In International Conference on User Modeling, Adaptation, and Personalization.
[16]
Xiujun Li, Yun-Nung Chen, Lihong Li, Jianfeng Gao, and Asli Celikyilmaz. 2017. Investigation of language understanding impact for reinforcement learning based dialogue systems. arXiv preprint arXiv: 1703.07055 (2017).
[17]
Florian Pecune, Jingya Chen, Yoichi Matsuyama, and Justine Cassell. 2018. Field Trial Analysis of Socially Aware Robot Assistant. In Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems. International Foundation for Autonomous Agents and Multiagent Systems, 1241--1249.
[18]
Florian Pecune, Magalie Ochs, and Catherine Pelachaud. 2013. A formal model of social relations for artificial companions. In European Conference on Multi-Agent Systems.
[19]
Charles Rich and Candace L Sidner. 2012. Using collaborative discourse theory to partially automate dialogue tree authoring. In International Conference on Intelligent Virtual Agents. Springer, 327--340.
[20]
Hannes Ritschel, Tobias Baur, and Elisabeth Andre: 2017. Adapting a Robot's Linguistic Style Based on Socially-Aware Reinforcement Learning.
[21]
Jost Schatzmann, Blaise Thomson, Karl Weilhammer, Hui Ye, and Steve Young. 2007. Agenda-based user simulation for bootstrapping a POMDP dialogue system. In NAACL-Short '07 Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics; Companion Volume, Short Papers. Association for Computational Linguistics.
[22]
Pararth Shah, Dilek Hakkani-Tür, Gokhan Tür, Abhinav Rastogi, Ankur Bapna, Neha Nayak, and Larry Heck. 2018. Building a Conversational Agent Overnight with Dialogue Self-Play. arXiv preprint arXiv:1801.04871 (2018).
[23]
Weiyan Shi and Zhou Yu. 2018. Sentiment Adaptive End-to-End Dialog Systems. arXiv preprint arXiv: 1804.10731 (2018).
[24]
Heung-Yeung Shum, Xiao-dong He, and Di Li. 2018. From Eliza to Xiaolce: challenges and opportunities with social chatbots. Frontiers of Information Technology & Electronic Engineering 19, 1 (2018), 10--26.
[25]
Tanmay Sinha and Justine Cassell. 2015. We click, we align, we learn: Impact of influence and convergence processes on student learning and rapport building. In 1st Workshop on Modeling INTERPERsonal SynchrONy And infLuence.
[26]
Andreas Stolcke, Elizabeth Shriberg, Rebecca Bates, Noah Coccaro, Daniel Juraf-sky, Rachel Martin, Marie Meteer, Klaus Ries, Paul Taylor, Carol Van Ess-Dykema, et al. 1998. Dialog act modeling for conversational speech. In AAAI Spring Symposium on Applying Machine Learning to Discourse Processing. 98--105.
[27]
Linda Tickle-Degnen and Robert Rosenthal. 1990. The nature of rapport and its nonverbal correlates. Psychological inquiry 1, 4 (1990), 285---293.
[28]
Karen Tracy and Nikolas Coupland. 1990. Multiple goals in discourse: An overview of issues. Journal of Language and Social Psychology 9, 1-2 (1990).
[29]
Steve Young. 2010. Still talking to machines (cognitively speaking). In Eleventh Annual Conference of the International Speech Communication Association.
[30]
Zhou Yu, Alan W Black, and Alexander I. Rudnicky. 2017. Learning Conversational Systems that Interleave Task and Non-Task Content. In International Joint Conference on Artificial Intelligence. arXiv:1703.00099.
[31]
Ran Zhao, Alexandros Papangelis, and Justine Cassell. 2014. Towards a dyadic computational model of rapport management for human-virtual agent interaction. In International Conference on Intelligent Virtual Agents. Springer, 514--527.
[32]
Ran Zhao, Tanmay Sinha, Alan W Black, and Justine Cassell. 2016. Automatic recognition of conversational strategies in the service of a socially-aware dialog system. In Proceedings of the 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue. 381--392.
[33]
Zian Zhao, Michael Madaio, Florian Pecune, Yoichi Matsuyama, and Justine Cassell. 2018. Socially-Conditioned Task Reasoning for a Virtual Tutoring Agent. In 17th International Conference on Autonomous Agents and MultiAgent Systems.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
IVA '20: Proceedings of the 20th ACM International Conference on Intelligent Virtual Agents
October 2020
394 pages
ISBN:9781450375863
DOI:10.1145/3383652
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 19 October 2020

Permissions

Request permissions for this article.

Check for updates

Badges

  • Honorable Mention

Author Tags

  1. Conversational Agent
  2. Dialogue Manager
  3. Rapport
  4. Reinforcement Learning
  5. Socially-Aware

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

IVA '20
Sponsor:
IVA '20: ACM International Conference on Intelligent Virtual Agents
October 20 - 22, 2020
Scotland, Virtual Event, UK

Acceptance Rates

Overall Acceptance Rate 53 of 196 submissions, 27%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)64
  • Downloads (Last 6 weeks)1
Reflects downloads up to 28 Dec 2024

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media