属性抽取是细粒度情感分析的子任务之一,其目标是从评论文本中抽取用户所评价的属性。在特定领域中,某些属性可能会频繁出现在不同的评论文本中,称之为高频属性。高频属性具有较高的领域表征能力,易被监督学习模型感知。相对地,低频属性出现频率低,可供训练的样本总量较为稀疏,使得神经网络模型难以充分学习相应的语言现象,从而使测试阶段的低频属性抽取难度较高。由于低频属性经常与高频属性同时出现在局部文字片段之中,该文根据这一特点,提出一种融合高频属性信息的属性抽取方法: 跟踪和记录模型识别的高频属性,使用卷积神经网络和注意力机制编码高频属性的上下文信息,并通过门控机制融入其他词项的表示学习过程中,辅助低频属性的识别。该文在国际语义评测大会2014和2016提供的笔记本电脑及餐馆领域数据集上进行了实验,相比于基线模型,该文方法在这两个英文数据集上F1值分别提升了2.33和1.44个百分点,并且总体性能高于现有前沿技术。
Abstract
Aspect extraction is one subtask of fine-grained sentiment analysis, which aims to extract the aspects that users express opinions on comments. Appearing in various comments, high-frequency aspects have strong domain representation ability and are easy to be perceived by the supervised learning model. We propose an aspect extraction method that integrates high-frequency aspects information. We track and record the high-frequency aspects recognized by model, encode the context information of high-frequency aspects by convolutional neural network and attention mechanism, and integrate the information into the representation learning process through the gated mechanism. Experiments on two benchmark datasets: Laptop of SemEval-14 and Restaurant of SemEval-16 demonstrate 2.33% and 1.44% improvement, respectively, compared with the baseline models.
关键词
属性抽取 /
高频属性 /
门控机制
{{custom_keyword}} /
Key words
aspect extraction /
high-frequency aspects /
gated mechanism
{{custom_keyword}} /
{{custom_sec.title}}
{{custom_sec.title}}
{{custom_sec.content}}
参考文献
[1] HU M,LIU B. Mining and summarizing customer reviews[C]//Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data mining, 2004: 168-177.
[2] ZHUANG L,FENG J,XIAO YAN Z. Movie review mining and summarization[C]//Proceedings of the 15th ACM International Conference on Information and Knowledge Management, 2006: 43-50.
[3] Wang B,WANG Houfeng. Bootstrapping both product features and opinion words from Chinese customer reviews with cross-inducing[C]//Proceedings of the 3rd International Joint Conference on Natural Language Processing, 2008: 289-295.
[4] QIU G,LIU B,BU J,et al. Opinion word expansion and target extraction through double propagation[J]. Computational Linguistics,2011,37(1): 9-27.
[5] MEI Q,ZHAI C. A mixture model for contextual text mining[C]//Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2006: 649-655.
[6] TITOV I,MCDONALD R. Modeling online reviews with multi-grain topic models[C]//Proceedings of the 17th International Conference on World Wide Web. ACM,2008: 111-120.
[7] MUKHERJEE A,LIU B. Aspect extraction through semi-supervised modeling[C]//Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics,2012: 339-348.
[8] JAKOB N,GUREVYCH I. Extracting opinion targets in a single-and cross-domain setting with conditional random fields[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics,2010: 1035-1045.
[9] LI F,HAN C,HUANG M,et al. Structure-aware review mining and summarization[C]//Proceedings of the 23rd International Conference on Computational Linguistics. Association for Computational Linguistics,2010: 653-661.
[10] LIU P,JOTY S,MENG H. Fine-grained opinion mining with recurrent neural networks andword embeddings[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2015: 1433-1443.
[11] TOH Z,SU J. NLANGP at semeval-2016 task 5: improving aspect based sentiment analysis using neural network features[C]//Proceedings of the 10th International Workshop on Semantic Evaluation, 2016: 282-288.
[12] XU H,LIU B,SHU L,et al. Double embeddings and CNN-based sequence labeling for aspect extraction[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 2018: 592-598.
[13] DAI H,SONG Y. Neural aspect and opinion term extraction with mined rules as weak supervision[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 5268-5277.
[14] MA D,LI S,WU F,et al. Exploring sequence-to-sequence learning in aspect term extraction[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 3538-3547.
[15] XU H,LIU B,SHU L,et al. BERT post-training for review reading comprehension and aspect-based sentiment analysis[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019: 2324-2335.
[16] WEI Z,HONG Y,ZOU B,et al. Don't eclipse your arts due to small discrepancies: Boundary repositioning with a pointer network for aspect extraction[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020: 3678-3684.
[17] WANG W,PAN S J,DAHLMEIER D,et al. Recursive neural conditional random fields for aspect-based sentiment analysis[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2016: 616-626.
[18] WANG W,PAN S J,DAHLMEIER D,et al. Coupled multi-layer attentions for co-extraction of aspect and opinion terms[C]//Proceedings of the 3st AAAI Conference on Artificial Intelligence, 2017: 3316-3322.
[19] WANG W,PAN S J,DAHLMEIER D. Multi-task memory networks for category-specific aspect and opinion terms co-extraction[J]. arXiv preprint arXiv: 1702.01776,2017.
[20] LI X,LAM W. Deep multi-task learning for aspect term extraction with memory interaction[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2017: 2886-2892.
[21] LI X,BING L,LI P,et al. Aspect term extraction with history attention and selective transformation[C]//Proceedings of the27th International Joint Conference on Artificial Intelligence. AAAI Press,2018: 4194-4200.
[22] VASWANI A,SHAZEER N,PARMAR N,et al. Attention is all you need[C]//Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017: 5998-6008.
[23] CUI Y,CHEN Z,WEI S,et al. Attention-over-attention neural networks for reading comprehension[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017: 593-602.
[24] JEFFREY P,RICHARDSOCHER,C D. Manning. Glove: global vectors for word representation[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2014: 1532-1543.
[25] DEVLIN J,CHANG M W,LEE K,et al. BERT: Pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies,2019: 4171-4186.
[26] ROTEM D,GILI B,SEGEV S,et al. The hitchhiker’s guide to testing statistical significance in natural language processing[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 2018: 1383-1392.
{{custom_fnGroup.title_cn}}
脚注
{{custom_fn.content}}
基金
国家自然科学基金(61672367,61751206,62076174)
{{custom_fund}}