skip to main content
10.1145/3562007.3562035acmotherconferencesArticle/Chapter ViewAbstractPublication PagesccrisConference Proceedingsconference-collections
research-article

An end-to-end information fusion model for task-oriented dialogue systems

Published: 12 October 2022 Publication History

Abstract

There are many aspects to consider when building an end-to-end task-oriented dialogue system, but recent studies have ignored the importance of dialogue history and textual representation. Although existing neural networks can effectively construct a linear relationship between dialogue history and knowledge base information, it is difficult to improve the dialogue quality. In this paper, we propose an end-to-end neural network model that reuses dialogue history information for task-oriented dialogue systems. We also reconstructed a large-scale word vector data set for use in NLP research. Our model integrates the implicit forgetting information and selection information of the dialogue history into the dialogue generation results through LSTM cell units, and uses Bert pre-trained word vectors for feature transformation to enhance entity representation capabilities. Finally, our model achieved a maximum increase of 2.4% in F1 value and a maximum 4% increase in Bleu value on the public data sets Incar and CamRest.

References

[1]
Z. Wei, Q. Liu, B. Peng, H. Tou, T. Chen, X. Huang, K.-F. Wong, X. Dai, Task-oriented dialogue system for automatic diagnosis, in: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Vol. 2, 2018, pp. 201–207.
[2]
Reddy R, Contractor D, Raghu D, Multi-level memory for task oriented dialogs[J]. arXiv preprint arXiv:1810.10647, 2018.
[3]
NNLM:Yoshua Bengio, Réjean Ducharme, and Pascal Vincent. A neural probabilistic language model. In Advances in Neural Information Processing Systems, pages 932–938, 2001.
[4]
Pennington J, Socher R, Manning C D. Glove: Global vectors for word representation[C]//Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). 2014: 1532-1543.
[5]
Devlin J, Chang M W, Lee K, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding[J]. arXiv preprint arXiv:1810.04805, 2018.
[6]
Wen T H, Vandyke D, Mrksic N, A network-based end-to-end trainable task-oriented dialogue system[J]. arXiv preprint arXiv:1604.04562, 2016.
[7]
Lei W, Jin X, Kan M Y, Sequicity: Simplifying task-oriented dialogue systems with single sequence-to-sequence architectures[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2018: 1437-1447.
[8]
M. Eric, C.D. Manning, A copy-augmented sequence-to-sequence architecture gives good performance on task-oriented dialogue, 2017, arXiv preprint arXiv:1701.04024.
[9]
Chen X, Xu J, Xu B. A working memory model for task-oriented dialog response generation[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019: 2687-2693.
[10]
Luong M T, Pham H, Manning C D. Effective approaches to attention-based neural machine translation[J]. arXiv preprint arXiv:1508.04025, 2015.
[11]
Wen T H, Vandyke D, Mrksic N, A network-based end-to-end trainable task-oriented dialogue system[J]. arXiv preprint arXiv:1604.04562, 2016.
[12]
Kingma D P, Ba J. Adam: A method for stochastic optimization[J]. arXiv preprint arXiv:1412.6980, 2014.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
CCRIS '22: Proceedings of the 2022 3rd International Conference on Control, Robotics and Intelligent System
August 2022
253 pages
ISBN:9781450396851
DOI:10.1145/3562007
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 12 October 2022

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

CCRIS'22

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 36
    Total Downloads
  • Downloads (Last 12 months)7
  • Downloads (Last 6 weeks)0
Reflects downloads up to 07 Nov 2024

Other Metrics

Citations

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media