论文标题
不要对此读太多:自适应计算用于开放域问题回答
Don't Read Too Much into It: Adaptive Computation for Open-Domain Question Answering
论文作者
论文摘要
大多数开放域问题回答的方法包括一个轻巧的检索器,它选择了一组候选段落,以及一个计算昂贵的读取器,它检查了段落以识别正确的答案。先前的作品表明,随着检索段落的数量增加,读者的性能也会增加。但是,他们认为所有检索的段落都具有同等的重要性,并将相同数量的计算分配给它们,从而大大增加了计算成本。为了降低这一成本,我们建议使用自适应计算来控制分配的读取段落的计算预算。我们首先引入了一种在单个段落上运行的技术,该技术依赖于任何时间预测和对早期退出概率的每层估计。然后,我们介绍SkylineBuilder,这是一种基于通过强化学习训练的资源分配策略,在每个步骤中动态决定哪种段落的方法。我们对小队开放的结果表明,在几种强大的静态和适应性方法上,自适应计算有所改善,导致计算降低4.3倍,同时保留了95%的完整模型性能。
Most approaches to Open-Domain Question Answering consist of a light-weight retriever that selects a set of candidate passages, and a computationally expensive reader that examines the passages to identify the correct answer. Previous works have shown that as the number of retrieved passages increases, so does the performance of the reader. However, they assume all retrieved passages are of equal importance and allocate the same amount of computation to them, leading to a substantial increase in computational cost. To reduce this cost, we propose the use of adaptive computation to control the computational budget allocated for the passages to be read. We first introduce a technique operating on individual passages in isolation which relies on anytime prediction and a per-layer estimation of an early exit probability. We then introduce SkylineBuilder, an approach for dynamically deciding on which passage to allocate computation at each step, based on a resource allocation policy trained via reinforcement learning. Our results on SQuAD-Open show that adaptive computation with global prioritisation improves over several strong static and adaptive methods, leading to a 4.3x reduction in computation while retaining 95% performance of the full model.