论文标题
从大型预训练的模型中提取特定任务的逻辑规则
Distilling Task-specific Logical Rules from Large Pre-trained Models
论文作者
论文摘要
逻辑规则(无论是可转移还是可解释)被广泛用作许多下游任务(例如命名实体标签)的弱监督信号。为了减少人类写作规则的努力,以前的研究人员采用了一种迭代方法来自动从几种种子规则中学习逻辑规则。但是,获得更多种子规则只能通过额外的人类注释来实现,并以沉重的成本来完成。受种子规则的规模和质量的限制,以前系统的模型性能是有限的。在本文中,我们开发了一个新颖的框架流,以从大型预训练模型中提取特定于任务的逻辑规则。具体来说,我们借用了最新的基于及时的语言模型作为知识专家,以产生初始种子规则,并基于形成的高质量实例池(充当中介角色),我们一直教导专家适合我们的任务和学习特定于任务的逻辑规则。对三个公共实体标记基准测试基准的实验证明了我们提出的框架的有效性。有了几个预定义的及时模板,我们的系统比以前的最新方法获得了重大改进。
Logical rules, both transferable and explainable, are widely used as weakly supervised signals for many downstream tasks such as named entity tagging. To reduce the human effort of writing rules, previous researchers adopt an iterative approach to automatically learn logical rules from several seed rules. However, obtaining more seed rules can only be accomplished by extra human annotation with heavy costs. Limited by the size and quality of the seed rules, the model performance of previous systems is bounded. In this paper, we develop a novel framework STREAM to distill task-specific logical rules from large pre-trained models. Specifically, we borrow recent prompt-based language models as the knowledge expert to yield initial seed rules, and based on the formed high-quality instance pool that acts as an intermediary role, we keep teaching the expert to fit our task and learning task-specific logical rules. Experiments on three public named entity tagging benchmarks demonstrate the effectiveness of our proposed framework. With several predefined prompt templates, our system has gained significant improvements over previous state-of-the-art methods.