论文标题
使用反馈加权学习部署后改善对话式问答系统
Improving Conversational Question Answering Systems after Deployment using Feedback-Weighted Learning
论文作者
论文摘要
对话系统与用户的相互作用为部署后改善了它们带来了一个令人兴奋的机会,但是几乎没有证据表明其可行性。在大多数应用程序中,用户无法为系统提供正确的答案,但是他们能够提供二进制(正确,不正确)的反馈。在本文中,我们提出了基于重要性抽样的反馈加权学习,以改善使用二进制用户反馈的初始监督系统。我们对文档分类(用于开发)和对话问题进行了模拟实验,以回答Quac和doqa等数据集,其中二进制用户反馈是从黄金注释中得出的。结果表明,我们的方法能够在初始监督系统上改进,接近一个完全监督的系统,该系统可以访问内域实验中相同的标记示例(QUAC),甚至在室外实验(DOQA)中匹配。我们的工作打开了前景,以利用与真实用户的互动并改善部署后的对话系统。
The interaction of conversational systems with users poses an exciting opportunity for improving them after deployment, but little evidence has been provided of its feasibility. In most applications, users are not able to provide the correct answer to the system, but they are able to provide binary (correct, incorrect) feedback. In this paper we propose feedback-weighted learning based on importance sampling to improve upon an initial supervised system using binary user feedback. We perform simulated experiments on document classification (for development) and Conversational Question Answering datasets like QuAC and DoQA, where binary user feedback is derived from gold annotations. The results show that our method is able to improve over the initial supervised system, getting close to a fully-supervised system that has access to the same labeled examples in in-domain experiments (QuAC), and even matching in out-of-domain experiments (DoQA). Our work opens the prospect to exploit interactions with real users and improve conversational systems after deployment.