Abstract:Medical question appeal classification can be dealt with as a text classification issue. This paper presents a reinforcement learning method for medical question appeal classification. Firstly, the keywords in medical problems are automatically identified by reinforcement learning, and the keywords and non-keywords in medical problems are assigned different values to constitute a vector. Secondly, the vector is set as the weight vector of the attention mechanism, and the weighted sum of the hidden layer generated by the Bi-LSTM model constitute question representation. At last, the question representation is classified by softmax classifier. Experimental results show that the accuracy of this method outperforms Bi-LSTM model by 1.49%.
[1] Socher R, Perelygin A, Wu Je, et al. Recursive deep models for semantic compositionality over a sentiment treebank[C]//Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing (EMNLP13), 2013: 1631-1642. [2] Socher R, Huval B, Manning CD, et al. Semantic compositionality through recursive matrix-vector spaces[C]//Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, 2012: 1201-1211. [3] Cho K, Merrinboer BV, Gulcehre C, et al. Learning phrase representations using RNN encoder-decoder for statistical machine translation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2012: 1724-1734, [4] Hill F, Cho K, Korhonen A. Learning distributed re-presentations of sentences from unlabelled data[C]//Proceedings of Annual Conference of the North American Chapter of the Association for Computational Linguistics, 2016: 1367-1377. [5] Yin W, Kann K, Yu M, et al. Comparative study of CNN and RNN for natural language processing[J]. arXiv preprint arXiv: 1702.01923, 2017. [6] Wen Y, Zhang W, Luo R, et al. Learning text representation using recurrent convolutional neural network with high way layers[J]. arXiv preprintar Xiv: 1606.06905, 2016. [7] He H, Gimpel K, Lin J. Multi-perspective sentence similarity modeling with convolutional neural networks[C]//Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 2015: 1576-1586. [8] Lai SW, Xu LH, Liu K, et al. Recurrent convolutional neural networks for text classification[C]//Proceedings of the 29th AAAI Conference on Artificial Intelligence (AAAI'15), 2015: 2267-2273. [9] Hochreiter S, Schmidhuber J. Long short-term memory[J]. Neural Computation, 1997: 1735-1780. [10] Zhang Y, Liu Q, Song L. Sentence-state LSTM for text representation[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Long Papers), 2018: 317-327. [11] Tai KS, Socher R, Manning CD. Improved semantic representations from tree-structured long short-term memory networks[C]//Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics, 2015: 1556-1566 [12] Zhang T, Huang M, Zhao L. Learning structured representation for text classification via reinforcement learning[C]//Proceedings of the 32nd AAAI Conference on Artificial Intelligence, 2018: 6053-6060. [13] Yogatama D, Blunsom P, Dyer C, et al. Learning to compose words into sentences with reinforcement learning[C]//Proceedings of International Conference on Learning Representations,2017:1-10. [14] Williams R J. Simple statistical gradient-following algorithms for connectionist reinforcement learning[J]. Machine Learning, 1992, 8(3-4): 229-256. [15] Sutton R S, McAllester D, Singh S, et al. Policy gradient methods for reinforcement learning with function approximation[C]//Proceedings of the 12th International Conference on Neural Information Processing Systems,1999: 1057-1063. [16] Graves A, Mohamed A R, Hinton G. Speech recognition with deep recurrent neural networks[C]//Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing, 2013: 6645-6649. [17] Lin Z, Feng M, Santos CN, et al. A structured self-attentive sentence embedding[C]//Proceedings of the International Conference on Learning Representations, 2017. [18] Mikolov T, Sutskever I, Chen K, et al. Distributed representations of words and phrases and their compositionality[C]//Proceedings of the 26th International Conference on Neural Information Processing Systems, 2013, 26: 3111-3119.