Fine-grained Named Entity Recognition for Multi-scenario
SHENG Jian1, XIANG Zhengpeng1, QIN Bing1, LIU Ming1, WANG Lifeng2
1.Research Center for Social Computing and Information Retrieval, Harbin Institute of Technology, Harbin, Heilongjiang 150001, China; 2.Tencent Technology (Shenzhen) CO., Ltd. Shenzhen, Guangdong 518000, China
Abstract:Name entity recognition is a classical research issue in data mining community. To recognize the entities in multi-domain with fine-grained labels, we propose a method of utilizes web thesaurus to annotate web data automatically to acquire large-scale training corpus. To minimize the influence of the noises in training corpus, we design a two-phase entity recognition method. First, the entity’s domain label is obtained. After that, the context of each recognized entity is used to determine the fine-grained label for one entity. Experimental results demonstrate that the proposed method can obtain high accuracy on entity recognition in multiple domains.
[1] Fine S,Singer Y,Tishby N.The hierarchical hidden Markov model:Analysis and applications[J].Machine learning,1998,32(1):41-62. [2] Borthwick A,Grishman R.A maximum entropy approach to named entity recognition[D].New York University,Graduate School of Arts and Science,1999. [3] McCallum A,Freitag D,Pereira F C N.Maximum entropy Markov models for information extraction and segmentation[C]//Proceedings of the 17th ICML,2000:591-598. [4] Lafferty J,Mc Callum A,Pereira F C N.Conditional random fields:Probabilistic models for segmenting and labeling sequence data[C]//Proceedings of the 18th ICML,2001:282-289. [5] Ratnaparkhi A.A maximum entropy model for part-of-speech tagging[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing,1996:133-142. [6] Collobert R,Weston J,Bootul L.Natural language processing(almost) from Scratch[J].arXiv preprine arXiv:1103.0398.2011. [7] Mikolov T,Sutskever I,Chen K,et al.Distributed representations of words and phrases and their compositionality[C]//Proceedings of the 26th International Conference on Neural Information Processing Systems,2013:3111-3119. [8] Gers F A,Schraudolph N N,Schmidhuber J.Learning precise timing with LSTM recurrent networks[J].Journal of Machine Learning Research,2002,3(Aug):115-143. [9] Huang Z,Xu W,Yu K.Bidirectional LSTM-CRF models for sequence tagging[J].arXiv preprint ar-Xiv:1508.01991,2015:1-10. [10] Ando R K,Zhang T.A framework for learning predictive structures from multiple tasks and unlabeled data[J].The Journal of Machine Learning Research,2005(6):1817-1853. [11] Jing H Y,Zhang T.Named entity recognition through classifier combination[C]//Proceedings of the 7th Conference on Natural Language Learning at HLT-NAACL-2003,2003:168-171. [12] Kim Yoon.Convolutional neural networks for sentence classification[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP),2014:1746-1751. [13] Zhang Y,Wallace B.A sensitivity analysis of (and practitioners' guide to) convolutional neural networks for sentence classification[J].arXiv preprint arXiv:1510.03820,2015. [14] M Boden.A guide to recurrent neural networks and back-propagation[R].SICS Technical Report J 2002:03,SICS. [15] Hammerton J.Named entity recognition with Long short-term memory[C]//Proceedings of the 7th Conference on Natural Language Learning at HLT-NAACL 2003-Volume 4.Association for Computational Linguistics,2003:172-175. [16] Tang D Qin B,Feng X.Effective LSTMs for target-dependent sentiment classification[C]//Proceedings of the COLING 2016,2016:3298-3307. [17] Yin Q,Zhang Y,Zhang W.Chinese zero pronoun resolution with deep memory network[C]//Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing,2017:1309-1318.