论文标题
使用图形注意网络进行文档建模
Document Modeling with Graph Attention Networks for Multi-grained Machine Reading Comprehension
论文作者
论文摘要
自然问题是一种具有两个粒度答案的新挑战的机器阅读理解基准,这是一个很长的答案(通常是段落)和一个简短的答案(长答案中的一个或多个实体)。尽管现有方法在此基准测试中具有有效性,但他们在培训期间分别对这两个子任务进行了对待,同时忽略了他们的依赖性。为了解决这个问题,我们提出了一个新颖的多元机器阅读理解框架,该框架着重于建模文档的层次结构性质,这是粒度的不同级别:文档,段落,句子,句子和代币。我们利用图形注意力网络获得不同级别的表示形式,以便可以同时学习它们。长和简短的答案可以分别从段落级别的表示和令牌级表示。通过这种方式,我们可以对两个粒度答案之间的依赖关系进行建模,以提供彼此的证据。我们共同培训这两个子任务,我们的实验表明,我们的方法在长期和简短的答案标准下都大大优于先前的系统。
Natural Questions is a new challenging machine reading comprehension benchmark with two-grained answers, which are a long answer (typically a paragraph) and a short answer (one or more entities inside the long answer). Despite the effectiveness of existing methods on this benchmark, they treat these two sub-tasks individually during training while ignoring their dependencies. To address this issue, we present a novel multi-grained machine reading comprehension framework that focuses on modeling documents at their hierarchical nature, which are different levels of granularity: documents, paragraphs, sentences, and tokens. We utilize graph attention networks to obtain different levels of representations so that they can be learned simultaneously. The long and short answers can be extracted from paragraph-level representation and token-level representation, respectively. In this way, we can model the dependencies between the two-grained answers to provide evidence for each other. We jointly train the two sub-tasks, and our experiments show that our approach significantly outperforms previous systems at both long and short answer criteria.