冯广敬,刘箴,刘婷婷,许根,庄寅,王媛怡,柴艳杰. 基于情感变量的二阶段对话生成模型[J]. 中文信息学报, 2022, 36(5): 102-111.
FENG Guangjing, LIU Zhen, LIU Tingting, XU Gen, ZHUANG Yin, WANG Yuanyi, CHAI Yanjie. A Two-Stage Dialogue Generation Model Based on Affective Variables. , 2022, 36(5): 102-111.
A Two-Stage Dialogue Generation Model Based on Affective Variables
FENG Guangjing1, LIU Zhen1, LIU Tingting2, XU Gen3, ZHUANG Yin1, WANG Yuanyi2, CHAI Yanjie1
1.Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo, Zhejiang 315211, China; 2.Faculty of Information Engineering,College of Science and Technology Ningbo University, Cixi, Zhejiang 315300, China; 3.Institute of Advanced Manufacturing Technology, Ningbo Institute of Materials Technology & Engineering, CAS, Ningbo, Zhejiang 315201, China
Abstract:Emotional dialogue generation has become one of the popular topics in natural language processing. It can improve the interaction between human and computer, but existing affective dialogue generation models only use a single affective variable and is easy to generate boring responses. To ensure the response sentences are not only semantically correct but also diversified, a two-stage dialogue generation model is proposed in this paper. In the first stage, DialoGPT with its powerful language understanding capabilities are used to ensure that responses with correct semantics can be generated. Main emotional variables and mixed emotional variables are fused to be global emotional variables to deal with the boring response. In the second stage, the global emotional variable is used to rewrite the response generated in the first stage, so as to polish the statement. Experimental results show that the proposed model performs better on the Empathetic Dialogues dataset than the baseline models.
[1] Xu Y, Zhao H, Zhang Z. Topic-Aware multi-turn dialogue modeling[C]//Proceedings of the AAAI Conference on Artificial Intelligence, 2021, 35(16): 14176-14184. [2] Xing C, Wu W, Wu Y, et al. Topic aware neural response generation[C]//Proceedings of the AAAI Conference on Artificial Intelligence, 2017. [3] 李少博, 孙承杰, 徐振, 等. 基于知识拷贝机制的生成式对话模型[J]. 中文信息学报, 2021,35(02): 107-115. [4] Wu W, Guo Z, Zhou X, et al. Proactive human-machine conversation with explicit conversation goal[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics,2019: 3794-3804. [5] Zheng Y, Zhang R, Huang M, et al. A pre-training based personalized dialogue generation model with persona-sparse data[C]//Proceedings of the AAAI Conference on Artificial Intelligence, 2020, 34(05): 9693-9700. [6] Song H, Wang Y, Zhang W, et al. Generate, delete and rewrite: a three-stage framework for improving persona consistency of dialogue generation[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020: 5821-5831. [7] Prendinger H, Mori J, Ishizuka M. Using human physiology to evaluate subtle expressivity of a virtual quizmaster in a mathematical game[J]. International Journal of Human-computer Studies,2005,62(2): 231-245. [8] Zech E, Rimé B. Is talking about an emotional experience helpful? Effects on emotional recovery and perceived benefits[J]. Clinical Psychology and Psychotherapy: An International Journal of Theory and Practice, 2005, 12(4): 270-287. [9] Sutskever I, Vinyals O, Le Q V. Sequence to sequence learning with neural networks[C]//Proceedings of the 27th International Conference on Neural Information Processing Systems, 2014: 3104-3112. [10] Zhou H, Huang M, Zhang T, et al. Emotional chatting machine: Emotional conversation generation with internal and external memory[C]//Proceedings of the AAAI Conference on Artificial Intelligence, 2018. [11] Zhang Y, Sun S, Galley M, et al. Dialogpt: large-scale generative pre-training for conversational response generation[J]. arXiv preprint arXiv:1911.00536, 2019. [12] Shen L, Feng Y. CDL: Curriculum dual learning for emotion-controllable response generation[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020: 556-566. [13] Song Z, Zheng X, Liu L, et al. Generating responses with a specific emotion in dialog[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 3685-3695. [14] Ma Z, Yang R, Du B, et al. A control unit for emotional conversation generation[J]. IEEE Access, 2020, (8): 43168-43176. [15] Li Q, Chen H, Ren Z, et al. EmpDG: multi-resolution interactive empathetic dialogue generation[C]//Proceedings of the 28th International Conference on Computational Linguistics, 2020: 4454-4466. [16] Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need[C]//Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017: 6000-6010. [17] Majumder N, Hong P, Peng S, et al. MIME: mimicking emotions for empathetic response generation[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing,2020: 8968-8979. [18] Nie L, Wang W, Hong R, et al. Multimodal dialog system: generating responses via adaptive decoders[C]//Proceedings of the 27th ACM International Conference on Multimedia, 2019: 1098-1106. [19] Radford A, Wu J, Child R, et al. Language models are unsupervised multitask learners[J].OpenAI blog, 2019, 1(8): 9. [20] Radford A, Narasimhan K, Salimans T, et al. Improving language understanding by generative pre-training[EB/OL]. https://www.cs.ubc.ca/~amuham01/LING530/papers/radford2018improving.pdf[2019-07-16]. [21] 孙五俊, 姜媛, 方平. 混合情绪能促进心理健康吗?[J]. 心理科学, 2021, 44(01): 230-236. [22] Serban I V, Sordoni A, Lowe R, et al. A hierarchical latent variable encoder-decoder model for generating dialogues[C]//Proceedings of the 31th AAAI Conference on Artificial Intelligence, 2017: 3295-3301. [23] Rashkin H, Smith E M, Li M, et al. Towards empathetic open-domain conversation models: a new benchmark and dataset[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 5370-5381. [24] Papineni K, Roukos S, Ward T, et al. Bleu: a method for automatic evaluation of machine translation[C]//Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics,2002: 311-318. [25] Bengio Y, Ducharme R, Vincent P, et al. A neural probabilistic language model[J]. Journal of Machine Learning Research, 2003, 3: 1137-1155. [26] Liu C-W, Lowe R, Serban I, et al. How not to evaluate your dialogue system: an empirical study of unsupervised evaluation metrics for dialogue response generation[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2016: 2122-2132. [27] Lin Z, Madotto A, Shin J, et al. MoEL: mixture of empathetic listeners[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019: 121-132.