论文标题

重新访问生成常识性推理:预订方法

Revisiting Generative Commonsense Reasoning: A Pre-Ordering Approach

论文作者

Zhao, Chao, Brahman, Faeze, Huang, Tenghao, Chaturvedi, Snigdha

论文摘要

预训练的模型(PTM)已导致自然语言产生(NLG)的大幅改善。但是,目前尚不清楚他们拥有多少常识性知识。为了评估NLG模型的常识性知识,最近的工作提出了生成常识性推理的问题,例如,给定一组无序的概念,构成了逻辑句子。解决此问题的现有方法假设PTM缺乏足够的参数知识,可以通过引入外部知识或特定于任务的预训练目标来克服。与这一趋势不同,我们认为PTM由于其输入的订单不足属性而低估了PTM的固有能力。特别是,我们假设输入概念的顺序会影响PTM利用其常识性知识的能力。为此,我们提出了一种预订方法,以精心操纵发电前给定概念的顺序。实验表明,我们的方法可以优于更复杂的模型,这些模型可以访问许多外部数据和资源。

Pre-trained models (PTMs) have lead to great improvements in natural language generation (NLG). However, it is still unclear how much commonsense knowledge they possess. With the goal of evaluating commonsense knowledge of NLG models, recent work has proposed the problem of generative commonsense reasoning, e.g., to compose a logical sentence given a set of unordered concepts. Existing approaches to this problem hypothesize that PTMs lack sufficient parametric knowledge for this task, which can be overcome by introducing external knowledge or task-specific pre-training objectives. Different from this trend, we argue that PTM's inherent ability for generative commonsense reasoning is underestimated due to the order-agnostic property of its input. In particular, we hypothesize that the order of the input concepts can affect the PTM's ability to utilize its commonsense knowledge. To this end, we propose a pre-ordering approach to elaborately manipulate the order of the given concepts before generation. Experiments show that our approach can outperform the more sophisticated models that have access to a lot of external data and resources.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源