基于强化学习的医疗问题诉求分类

吴昊,黄德根,林晓惠

PDF(1370 KB)
PDF(1370 KB)
中文信息学报 ›› 2021, Vol. 35 ›› Issue (3) : 100-106.
信息抽取与文本挖掘

基于强化学习的医疗问题诉求分类

  • 吴昊,黄德根,林晓惠
作者信息 +

Medical Question Appeal Classification Based on Reinforcement Learning

  • WU Hao, HUANG Degen, LIN Xiaohui
Author information +
History +

摘要

医疗问题诉求分类属于文本分类,是自然语言处理中的基础任务。该文提出一种基于强化学习的方法对医疗问题诉求进行分类。首先,通过强化学习自动识别出医疗问题中的关键词,并且对医疗问题中的关键词和非关键词赋予不同的值构成一个向量;其次,利用该向量作为attention机制的权重向量,对Bi-LSTM模型生成的隐含层状态序列加权求和得到问题表示;最后通过Softmax分类器对问题表示进行分类。实验结果表明,该方法比基于Bi-LSTM模型的分类结果准确率提高1.49%。

Abstract

Medical question appeal classification can be dealt with as a text classification issue. This paper presents a reinforcement learning method for medical question appeal classification. Firstly, the keywords in medical problems are automatically identified by reinforcement learning, and the keywords and non-keywords in medical problems are assigned different values to constitute a vector. Secondly, the vector is set as the weight vector of the attention mechanism, and the weighted sum of the hidden layer generated by the Bi-LSTM model constitute question representation. At last, the question representation is classified by softmax classifier. Experimental results show that the accuracy of this method outperforms Bi-LSTM model by 1.49%.

关键词

强化学习 / Bi-LSTM / 注意力机制

Key words

reinforcement learning / Bi-LSTM / attention mechanism

引用本文

导出引用
吴昊,黄德根,林晓惠. 基于强化学习的医疗问题诉求分类. 中文信息学报. 2021, 35(3): 100-106
WU Hao, HUANG Degen, LIN Xiaohui. Medical Question Appeal Classification Based on Reinforcement Learning. Journal of Chinese Information Processing. 2021, 35(3): 100-106

参考文献

[1] Socher R, Perelygin A, Wu Je, et al. Recursive deep models for semantic compositionality over a sentiment treebank[C]//Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing (EMNLP13), 2013: 1631-1642.
[2] Socher R, Huval B, Manning CD, et al. Semantic compositionality through recursive matrix-vector spaces[C]//Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, 2012: 1201-1211.
[3] Cho K, Merrinboer BV, Gulcehre C, et al. Learning phrase representations using RNN encoder-decoder for statistical machine translation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2012: 1724-1734,
[4] Hill F, Cho K, Korhonen A. Learning distributed re-presentations of sentences from unlabelled data[C]//Proceedings of Annual Conference of the North American Chapter of the Association for Computational Linguistics, 2016: 1367-1377.
[5] Yin W, Kann K, Yu M, et al. Comparative study of CNN and RNN for natural language processing[J]. arXiv preprint arXiv: 1702.01923, 2017.
[6] Wen Y, Zhang W, Luo R, et al. Learning text representation using recurrent convolutional neural network with high way layers[J]. arXiv preprintar Xiv: 1606.06905, 2016.
[7] He H, Gimpel K, Lin J. Multi-perspective sentence similarity modeling with convolutional neural networks[C]//Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 2015: 1576-1586.
[8] Lai SW, Xu LH, Liu K, et al. Recurrent convolutional neural networks for text classification[C]//Proceedings of the 29th AAAI Conference on Artificial Intelligence (AAAI'15), 2015: 2267-2273.
[9] Hochreiter S, Schmidhuber J. Long short-term memory[J]. Neural Computation, 1997: 1735-1780.
[10] Zhang Y, Liu Q, Song L. Sentence-state LSTM for text representation[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Long Papers), 2018: 317-327.
[11] Tai KS, Socher R, Manning CD. Improved semantic representations from tree-structured long short-term memory networks[C]//Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics, 2015: 1556-1566
[12] Zhang T, Huang M, Zhao L. Learning structured representation for text classification via reinforcement learning[C]//Proceedings of the 32nd AAAI Conference on Artificial Intelligence, 2018: 6053-6060.
[13] Yogatama D, Blunsom P, Dyer C, et al. Learning to compose words into sentences with reinforcement learning[C]//Proceedings of International Conference on Learning Representations,2017:1-10.
[14] Williams R J. Simple statistical gradient-following algorithms for connectionist reinforcement learning[J]. Machine Learning, 1992, 8(3-4): 229-256.
[15] Sutton R S, McAllester D, Singh S, et al. Policy gradient methods for reinforcement learning with function approximation[C]//Proceedings of the 12th International Conference on Neural Information Processing Systems,1999: 1057-1063.
[16] Graves A, Mohamed A R, Hinton G. Speech recognition with deep recurrent neural networks[C]//Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing, 2013: 6645-6649.
[17] Lin Z, Feng M, Santos CN, et al. A structured self-attentive sentence embedding[C]//Proceedings of the International Conference on Learning Representations, 2017.
[18] Mikolov T, Sutskever I, Chen K, et al. Distributed representations of words and phrases and their compositionality[C]//Proceedings of the 26th International Conference on Neural Information Processing Systems, 2013, 26: 3111-3119.

基金

国家自然科学基金(U1936109,61672127)
PDF(1370 KB)

1011

Accesses

0

Citation

Detail

段落导航
相关文章

/