多跳式文本阅读理解方法综述

倪艺函,兰艳艳,庞亮,程学旗

PDF(1914 KB)
PDF(1914 KB)
中文信息学报 ›› 2022, Vol. 36 ›› Issue (11) : 1-19.
综述

多跳式文本阅读理解方法综述

  • 倪艺函1,2,兰艳艳3,庞亮1,2,程学旗1,2
作者信息 +

A Survey of Multi-hop Reading Comprehension for Text

  • NI Yihan1,2, LAN Yanyan3, PANG Liang1,2, CHENG Xueqi1,2
Author information +
History +

摘要

多跳阅读理解成为近年来自然语言理解领域的研究热点,与简单阅读理解相比,它更加复杂,需要面对如下挑战: ①结合多处内容线索,如多文档阅读等; ②具有可解释性,如给出推理路径等。为应对这些挑战,出现了各类不同的工作。因此该文综述了多跳式文本阅读理解这一复杂阅读理解任务,首先给出了多跳文本阅读理解任务的定义;由于推理是多跳阅读理解模型的基础能力,根据推理方式的不同,多跳阅读理解模型可以分为三类: 基于结构化推理的多跳阅读理解模型、基于线索抽取的多跳阅读理解模型、基于问题拆分的多跳阅读理解模型,该文接下来比较分析了各类模型在常见多跳阅读理解模型任务数据集上的实验结果,发现这三类模型之间各有优劣。最后探讨了未来的研究方向。

Abstract

In recent years, multi-hop reading comprehension has become a hot topic in natural language understanding. Compared with single-hop reading comprehension, multi-hop reading comprehension is more challenging considering the involvement of multiple clues (e.g. multiple documents reading) the expectation for explicit reasoning paths. This paper summarizes the multi-hop reading comprehension task. This paper first gives the definition of multi-hop reading comprehension task. And then, and according to different reasoning methods, multi-hop reading comprehension models are can be divided into three categories: models based on structured reasoning, models based on evidence extraction, and models based on question decomposition. This paper analyzes the experimental results of these models on the common multi-hop reading comprehension datasets, revealing their advantages and disadvantages. Finally, the future research directions are discussed.

关键词

多跳阅读理解 / 图神经网络 / 迭代线索抽取 / 问题拆分

Key words

multi-hop reading comprehension / graph neural network / iterative evidence extraction / question decomposition

引用本文

导出引用
倪艺函,兰艳艳,庞亮,程学旗. 多跳式文本阅读理解方法综述. 中文信息学报. 2022, 36(11): 1-19
NI Yihan, LAN Yanyan, PANG Liang, CHENG Xueqi. A Survey of Multi-hop Reading Comprehension for Text. Journal of Chinese Information Processing. 2022, 36(11): 1-19

参考文献

[1] YANG Z, QI P, ZHANG S, et al. HotpotQA: A dataset for diverse, explainable multi-hop question answering[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2018: 2369-2380.
[2] ZHAO C, XIONG C, ROSSET C, et al. Transformer-Xh: Multi-evidence reasoning with extra hop attention[C]//Proceedings of the International Conference on Learning Representations, 2019.
[3] KIM D, KIM S,KWAK N. Textbook question answering with multi-modal context graph understanding and self-supervised open-set comprehension[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 3568-3584.
[4] TU M, HUANG K, WANG G et al. Select, answer and explain: Interpretable multi-hop reading comprehension over multiple documents[C]//Proceedings of the 34th AAAI Conference on Artificial Intelligence, 2020: 9073-9080.
[5] DE CAO N, AZIZ W,TITOV I. Question answering by reasoning across documents with graph convolutional networks[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA: Association for Computational, 2019: 2306-2317.
[6] SUN H, BEDRAXWEISS T,COHEN W. PullNET: Open domain question answering with iterative retrieval on knowledge bases and text[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019: 2380-2390.
[7] YE D, LIN Y, LIU Z, et al. Multi-paragraph reasoning with knowledge-enhanced graph neural network[J/OL]. arXiv preprint arXiv: 1911.02170, 2019.
[8] LONG A, MASON J, BLAIR A, et al. Multi-hop reading comprehension via deep reinforcement learning based document traversal[J/OL]. arXiv preprint arXiv: 1905.09438, 2019.
[9] TU M, WANG G, HUANG J, et al. Multi-hop reading comprehension across multiple documents by reasoning over heterogeneous graphs[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 2704-2713.
[10] TANG Z, SHEN Y, MA X, et al. Multi-hop reading comprehension across documents with path-based graph convolutional network[J/OL]. arXiv preprint arXiv: 2006.06478, 2020.
[11] THAYAPARAN M, VALENTINO M, SCHLEGEL V, et al. Identifying supporting facts for multi-hop question answering with document graph networks[C]//Proceedings of the 13th Workshop on Graph-Based Methods for Natural Language Processing, 2019: 42-51.
[12] FANG Y, SUN S, GAN Z, et al. Hierarchical graph network for multi-hop question answering[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing. Online: Association for Computational Linguistics, 2020: 8823-8838.
[13] TU M, HUANG J, HE X, et al. Graph sequential network for reasoning over sequences[J/OL]. arXiv preprint arXiv: 2004.02001, 2020.
[14] SONG L, WANG Z, YU M, et al. Exploring graph-structured passage representation for multi-hop reading comprehension with graph neural networks[J/OL]. arXiv preprint arXiv: 1809.02040, 2018.
[15] QIU L, XIAO Y, QU Y, et al. Dynamically fused graph network for multi-hop reasoning[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 6140-6150.
[16] ZHANG Y, NIE P, RAMAMURTHY A, et al. DDRQA: Dynamic document reranking for open-domain multi-hop question answering[J/OL]. arXiv preprint arXiv: 2009.07465, 2020.
[17] DING M, ZHOU C, CHEN Q, et al. Cognitive graph for multi-hop reading comprehension at scale[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 2694-2703.
[18] CAO Y, FANG M,TAO D. Bag: Bidirectional attention entity graph convolutional network for multi-hop reasoning question answering[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA: Association for Computational, 2019: 357-362.
[19] MA X, ZHU Q, ZHOU Y, et al. Asking complex questions with multi-hop answer-focused reasoning[J/OL]. arXiv preprint arXiv: 2009.07402, 2020.
[20] 舒冲, 欧阳智, 杜逆索, 等. 基于改进图节点的图神经多跳阅读理解研究[J/OL]. 计算机工程, 2020: 1-9 10.19678/j.issn.1000-3428.0059917[2021-05-07].
[21] KUNDU S, KHOT T, SABHARWAL A, et al. Exploiting explicit paths for multi-hop reading comprehension[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 2737-2747.
[22] JIANG Y, JOSHI N, CHEN Y C, et al. Explore, propose and assemble: An interpretable model for multi-hop reading comprehension[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 2714-2725.
[23] FENG Y, YU M, XIONG W, et al. Learning to recover reasoning chains for multi-hop question answering via cooperative games[J/OL]. arXiv preprint arXiv: 2004.02393, 2020.
[24] ASAI A, HASHIMOTO K, HAJISHIRZI H, et al. Learning to retrieve reasoning paths over wikipedia graph for question answering[C]//Proceedings of the International Conference on Learning Representations, 2019.
[25] CHEN J, LIN S T, DURRETT G. Multi-hop question answering via reasoning chains[J/OL]. arXiv preprint arXiv: 1910.02610, 2019.
[26] QI P, LIN X, MEHR L, et al. Answering complex open-domain questions through iterative query generation[C]//Proceedings of the International Joint Nonference on Natural Language Processing, 2019: 2590-2602.
[27] NISHIDA K, NISHIDA K, NAGATA M, et al. Answering while summarizing: Multi-task learning for multi-hop qa with evidence extraction[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 2335-2345.
[28] DHINGRA B, ZAHEER M, BALACHANDRAN V, et al. Differentiable reasoning over a virtual knowledge base[C]//Proceedings of the International Conference on Learning Representations, 2019.
[29] FELDMAN Y,ELYANIV R. Multi-hop paragraph retrieval for open-domain question answering[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 2296-2309.
[30] DAS R, GODBOLE A, KAVARTHAPU D, et al. Multi-step entity-centric information retrieval for multi-hop question answering[C]//Proceedings of the 2nd Workshop on Machine Reading for Question Answering, 2019: 113-118.
[31] DAS R, DHULIAWALA S, ZAHEER M, et al. Multi-step retriever-reader interaction for scalable open-domain question answering[C]//Proceedings of the International Conference on Learning Representations, 2018.
[32] XIONG W, YU M, GUO X, et al. Simple yet effective bridge reasoning for open-domain multi-hop question answering[C]//Proceedings of the 2nd Workshop on Machine Reading for Question Answering, 2019: 48-52.
[33] YADAV V, BETHARD S,SURDEANU M. Unsupervised alignment-based iterative evidence retrieval for multi-hop question answering[C]//Proceedings of the 58th Annual Meeting of the Association for Computational LinguisticsAssociation for Computational Linguistics, 2020: 4514-4525.
[34] GROENEVELD D, KHOT T, SABHARWAL A, et al. A simple yet strong pipeline for hotpotqa[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing. Online: Association for Computational Linguistics, 2020: 8839-8845.
[35] NIE Y, WANG S,BANSAL M. Revealing the importance of semantic retrieval for machine reading at scale[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019: 2553-2566.
[36] TRIVEDI H, KWON H, KHOT T, et al. Repurposing entailment for multi-hop question answering tasks[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Minneapolis, MN, USA: Association for Computational, 2019: 2948-2958.
[37] YADAV V, BETHARD S,SURDEANU M. Quick and (not so) dirty: Unsupervised selection of justification sentences for multi-hop question answering[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. Hong Kong, China: Association for Computational Linguistics, 2019: 2578-2589.
[38] 段艺文. 多层注意机制下阅读理解问答模型研究与应用[D]. 成都: 电子科技大学硕士学位论文, 2020.
[39] 李天仙. 基于多跳注意力的中文机器阅读理解[D]. 武汉: 华中师范大学硕士学位论文, 2020.
[40] 盛艺暄,兰曼. 利用外部知识辅助和多步推理的选择题型机器阅读理解模型[J]. 计算机系统应用, 2020, 29(4): 1-9.
[41] KUMAR A, IRSOY O, ONDRUSKA P, et al. Ask me anything: Dynamic memory networks for natural language processing[C]//Proceedings of the International Conference on Machine Learning, 2016: 1378-1387.
[42] ZHONG V, XIONG C, KESKAR N S, et al. Coarse-grain fine-grain coattention network for multi-evidence question answering[C]//Proceedings of the International Conference on Learning Representations, 2019.
[43] BAUER L, WANG Y,BANSAL M. Commonsense for generative multi-hop question answering tasks[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2018: 4220-4230.
[44] SHEN Y, LIU X, DUH K, et al. An empirical analysis of multiple-turn reasoning strategies in reading comprehension tasks[C]//Proceedings of the 8th International Joint Conference on Natural Language Processing, 2017: 957-966.
[45] SUKHBAATAR S, WESTON J,FERGUS R. End-to-end memory networks[C]//Proceedings of the 29th International Conference on Neural Information Processing Systems, 2015: 2440-2448.
[46] AINSLIE J, ONTANON S, ALBERTI C, et al. ETC: Encoding long and structured inputs in transformers[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2020: 268-284.
[47] BELTAGY I, PETERS M E,COHAN A. Longformer: the long-document transformer[J/OL]. arXiv preprint arXiv: 2004.05150, 2020.
[48] WESTON J, CHOPRA S,BORDES A. Memory networks[C]//Proceedings of the International Conference on Learning Representations, 2015.
[49] YU S, INDURTHI S R, BACK S, et al. A multi-stage memory augmented neural network for machine reading comprehension[C]//Proceedings of the Workshop on Machine Reading for Question Answering, 2018: 21-30.
[50] SHEN Y, HUANG P S, GAO J, et al. ReasoNet: Learning to stop reading in machine comprehension[C]//Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2017: 1047-1055.
[51] GONG Y,BOWMAN S. Ruminating Reader: Reasoning with gated multi-hop attention[C]//Proceedings of the Workshop on Machine Reading for Question Answering, 2018: 1-11.
[52] PEREZ E, LEWIS P, YIH W-T, et al. Unsupervised question decomposition for question answering[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing. Online: Association for Computational Linguistics,2020: 8864-8880.
[53] KHOT T, KHASHABI D, RICHARDSON K, et al. Text modular networks: Learning to decompose tasks in the language of existing models[J/OL]. arXiv preprint arXiv: 2009.00751, 2020.
[54] MIN S, ZHONG V, ZETTLEMOYER L, et al. Multi-hop reading comprehension through question decomposition and rescoring[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 6097-6109.
[55] TRAN N K,NIEDERE C. Multihop attention networks for question answer matching[C]//Proceedings of the 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, 2018: 325-334.
[56] JIANG Y,BANSAL M. Self-assembling modular networks for interpretable multi-hop reasoning[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019: 4464-4474.
[57] GRAIL Q, PEREZ J,GAUSSIER E. Latent question reformulation and information accumulation for multi-hop machine reading[DB/OL]. https://openreview.net/pdf?id=S1x63TEYvr[2021-05-16].
[58] GUPTA N, LIN K, ROTH D, et al. Neural module networks for reasoning over text[C]//Proceedings of the International Conference on Learning Representations, 2019.
[59] WEBER L, MINERVINI P, MNCHMEYER J, et al. NlProlog: Reasoning with weak unification for question answering in natural language[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 6151-6161.
[60] JIANG Y,BANSAL M. Avoiding reasoning shortcuts: Adversarial evaluation, training, and model development for multi-hop[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 2726-2736.
[61] MIN S, WALLACE E, SINGH S, et al. Compositional questions do not necessitate multi-hop reasoning[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics,2019: 4249-4257.
[62] TRIVEDI H, BALASUBRAMANIAN N, KHOT T, et al. Measuring and reducing non-multifact reasoning in multi-hop question answering[J/OL]. arXiv preprint arXiv: 2005.00789, 2020.
[63] TANG Y, NG H T,TUNG A. Do multi-hop question answering systems know how to answer the single-hop sub-questions?[C]//Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume. Online: Association for Computational Linguistics, 2021: 3244-3249.
[64] WANG H, YU M, GUO X, et al. Do multi-hop readers dream of reasoning chains?[C]//Proceedings of the 2nd Workshop on Machine Reading for Question Answering, 2019: 91-97.
[65] CHEN J,DURRETT G. How to learn (and how not to learn) multi-hop reasoning with memory networks[DB/OL]. https://openreview.net/pdf?id=B1lf43A5Y7[2021-05-07].
[66] SHAO N, CUI Y, LIU T, et al. Is graph structure necessary for multi-hop question answering?[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing. Online: Association for Computational Linguistics, 2020: 7187-7192.
[67] USTER S, SUSHIL M,DAELEMANS W. Why cant memory networks read effectively?[J/OL]. arXiv preprint arXiv: 1910.07350, 2019.
[68] WELBL J, STENETORP P,RIEDEL S. Constructing datasets for multi-hop reading comprehension across documents[J]. Transactions of the Association for Computational Linguistics, 2018, 6: 287-302.
[69] CHEN J,DURRETT G. Understanding dataset design choices for multi-hop reasoning[C]//Proceedings of the North American Chapter of the Association for Computational Linguistics, 2019: 4026-4032.
[70] WESTON J, BORDES A, CHOPRA S, et al. Towards ai-complete question answering: a set of prerequisite toy tasks[C]//Proceedings of the International Conference on Learning Representations, 2016.
[71] CHEN W, ZHA H, CHEN Z, et al. HybridQA: A dataset of multi-hop question answering over tabular and textual data[C]//Proceedings of the Findings of the Association for Computational Linguistics: EMNLP, 2020: 1026-1036.
[72] TALMOR A,BERANT J. The web as a knowledge-base for answering complex questions[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, Louisiana, USA: Association for Computational Linguistics, 2018: 641-651.
[73] ABUJABAL A, ROY R S, YAHYA M, et al. ComQA: A community-sourced dataset for complex factoid question answering with paraphrase clusters[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Minneapolis, MN, USA: Association for Computational, 2019: 307-317.
[74] KHOT T, CLARK P, GUERQUIN M, et al. QASC: A dataset for question answering via sentence composition[C]//Proceedings of the 34th AAAI Conference on Artificial Intelligence, 2020.
[75] WOLFSON T, GEVA M, GUPTA A, et al. Break it down: A question understanding benchmark[J]. Transactions of the Association for Computational Linguistics, 2020, 8: 183-198.
[76] KOCˇISKY' T, SCHWARZ J, BLUNSOM P, et al. The narrativeqa reading comprehension challenge[J]. Transactions of the Association for Computational Linguistics, 2018, 6: 317-328.
[77] KHASHABI D, CHATURVEDI S, ROTH M, et al. Looking beyond the surface: A challenge set for reading comprehension over multiple sentences[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. New Orleans, Louisiana, USA: Association for Computational Linguistics, 2018: 252-262.
[78] DUA D, WANG Y, DASIGI P, et al. DROP: A reading comprehension benchmark requiring discrete reasoning over paragraphs[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA: Association for Computational, 2019: 2368-2378.
[79] REDDY S, CHEN D,MANNING C D. COQA: A conversational question answering challenge[J]. Transactions of the Association for Computational Linguistics, 2019, 7: 249-266.
[80] CHOI E, HE H, IYYER M, et al. QUAC: Question answering in context[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2018: 2174-2184.
[81] ELGOHARY A, ZHAO C,BOYD-GRABER J. a dataset and baselines for sequential open-domain question answering[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2018: 1077-1083.
[82] MIHAYLOV T, CLARK P, KHOT T, et al. Can a suit of armor conduct electricity?: A new dataset for open book question answering[C]//Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2018: 2381-2391.
[83] SAHA A, ARALIKATTE R, KHAPRA M M, et al. DuoRC: Towards complex language understanding with paraphrased reading comprehension[C]//Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics. Melbourne, Australia: Association for Computational Linguistics, 2018: 1683-1693.
[84] SEO M, KEMBHAVI A, FARHADI A, et al. Bidirectional attention flow for machine comprehension[C]//Proceedings of the International Conference on Learning Representations, 2017.
[85] DEVLIN J, CHANG M W, LEE K, et al. BERT: Pretraining of deep bidirectional transformers for language understanding[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Minneapolis, MN, USA: Association for Computational, 2019: 4171-4186.
[86] TALMOR A,BERANT J. MultiQA: An empirical investigation of generalization and transfer in reading comprehension[C]//Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019: 4911-4921.
[87] WEISSENBORN D, WIESE G,SEIFFE L. Making neural QA as simple as possible but not simpler[C]//Proceedings of the 21st Conference on Computational Natural Language Learning, 2017: 271-280.
[88] DHINGRA B, JIN Q, YANG Z, et al. Neural models for reasoning over multiple mentions using coreference[C]//Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. New Orleans, Louisiana, USA: Association for Computational Linguistics, 2018: 42-48.
[89] RADFORD A, WU J, CHILD R, et al. Language models are unsupervised multitask learners[DB/OL]. https://cdn.openai.com/better-language-models/language_models_are_unsupervised_multitask_learners.pdf[2021-05-07].
[90] YANG H, WANG H, GUO S, et al. Learning to decompose compound questions with reinforcement learning[DB/OL]. https://openreview.net/pdf?id=SJl2ps0qKQ [2021-05-07].

基金

国家自然科学基金(61773362;61906180);国家重点研发计划(2016QY02D0405;2019QY2303);腾讯AILab犀牛鸟专项研究计划(JR202033)
PDF(1914 KB)

2534

Accesses

0

Citation

Detail

段落导航
相关文章

/