Abstract:Entity relation extraction aims to extract semantic relations between entities from text. This task is well addressed for normative texts such as news reports and Wikipedia, but less touched for dialogue texts. Compared with standard text, dialogue is an interactive process, and such information hidden in the interaction challenges the entity relation extraction task. This paper proposes an entity relation extraction method that incorporates interactive information via cross-attention mechanisms. It also adopt the multitask learning to deal with the issue of unblance distribution. The experiments on DialogRE public dataset reveal a result of 54.1% F1 and 50.7% of F1c, which proves validity of the method.
[1] Golshan P N, Dashti H A R, Azizi S, et al. A study of recent contributions on information extraction[C]//Proceedings of the CoRR, 2018. [2] Asai A, Hashimoto K, Hajishirzi H, et al. Learning to retrieve reasoning paths over wikipedia graph for question answering[C]//Proceedings of the CoRR, 2019. [3] Zhou X, Li L, Dong D, et al. Multi-turn response selection for chatbots with deep attention matching network[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 2018: 1118-1127. [4] Yu D, Sun K, Cardie C, et al. Dialogue-based relation extraction[C]//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020: 4927-4940. [5] Guo Q, Qiu X, Liu P, et al. Star-Transformer[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019, 1315-1325. [6] Hao Y, Zhang Y, Liu K, et al. An end-to-end model for question answering over knowledge base with cross-attention combining global knowledge[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017: 221-231. [7] Zeng D, Liu K, Lai S, et al. Relation classification via convolutional deep neural network[C]//Proceedings of the 25th International Conference on Computational Linguistics, Technical Papers, 2014: 2335-2344. [8] Katiyar A, Cardie C. Going out on a limb: Joint extraction of entity mentions and relations without dependency trees[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2017: 917-928. [9] Wang L, Cao Z, De Melo G, et al. Relation classification via multi-level attention CNNS[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 2016: 1298-1307. [10] Zhang Y, Qi P, Manning C D. Graph convolution over pruned dependency trees improves relation extraction[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2018: 2205-2215. [11] Guo Z, Zhang Y, Lu W. Attention guided graph convolutional networks for relation extraction[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 241-251. [12] 李青青, 杨志豪, 罗凌,等. 基于多任务学习的生物医学实体关系抽取[J]. 中文信息学报, 2019, 33(8): 84-92. [13] 刘苏文, 邵一帆, 钱龙华,等. 基于联合学习的生物医学因果关系抽取[J]. 中文信息学报,2020, 34(4): 60-68. [14] Quirk C, Poon H. Distant supervision for relation extraction beyond the sentence boundary[C]//Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, 2017: 1171-1182. [15] Yao Y, Ye D, Li P, et al. DoCRED: A large-scale document-level relation extraction dataset[C]//Proceedings of the 57th Conference of the Association for Computational Linguistics, 2019: 764-777.[16] Devlin J, Chang M W, Lee K, et al.Bert: Pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019: 4171-4186. [17] 陈治纲, 何丕廉, 孙越恒, 等. 基于向量空间模型的文本分类系统的研究与实现[J]. 中文信息学报, 2005, 19(1): 37-42. [18] Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need[C]//Proceedings of the Advances in Neural Information Processing Systems. 2017: 6000-6010. [19] Yao L, Torabi A, Cho K, et al. Video description generation incorporating spatio-temporal features and a soft-attention mechanism[C]//Proceedings of the CoRR, 2015. [20] Pennington J, Socher R, Manning C D. Glove: Global vectors for word representation[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2014: 1532-1543. [21] Kingma D P, Ba J. Adam: A method for stochastic optimization[C]//Proceedings of the 3rd International Conference on Learning Representations, 2015.