Open Access
ARTICLE
Neural Dialogue Model with Retrieval Attention for Personalized Response Generation
Cong Xu1, 2, Zhenqi Sun2, 3, Qi Jia2, 3, Dezheng Zhang2, 3, Yonghong Xie2, 3,*, Alan Yang4
1 School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing,
100083, China.
2 Beijing Key Laboratory of Knowledge Engineering for Materials Science, University of Science and
Technology Beijing, Beijing, 100083, China.
3 School of Computer and Communication Engineering, University of Science and Technology Beijing,
Beijing, 100083, China.
4 Amphenol AssembleTech, Houston, TX 77070, US.
* Corresponding Author: Yonghong Xie. Email: .
Computers, Materials & Continua 2020, 62(1), 113-122. https://doi.org/10.32604/cmc.2020.05239
Abstract
With the success of new speech-based human-computer interfaces, there is a
great need for effective and friendly dialogue agents that can communicate with people
naturally and continuously. However, the lack of personality and consistency is one of
critical problems in neural dialogue systems. In this paper, we aim to generate consistent
response with fixed profile and background information for building a realistic dialogue
system. Based on the encoder-decoder model, we propose a retrieval mechanism to deliver
natural and fluent response with proper information from a profile database. Moreover, in
order to improve the efficiency of training the dataset related to profile information, we
adopt a method of pre-training and adjustment for general dataset and profile dataset. Our
model is trained by social dialogue data from Weibo. According to both automatic and
human evaluation metrics, the proposed model significantly outperforms standard
encoder-decoder model and other improved models on providing the correct profile and
high-quality responses.
Keywords
Cite This Article
C. Xu, Z. Sun, Q. Jia, D. Zhang, Y. Xie
et al., "Neural dialogue model with retrieval attention for personalized response generation,"
Computers, Materials & Continua, vol. 62, no.1, pp. 113–122, 2020. https://doi.org/10.32604/cmc.2020.05239
Citations