论文标题
医疗问题理解和回答知识基础和语义自我诉讼
Medical Question Understanding and Answering with Knowledge Grounding and Semantic Self-Supervision
论文作者
论文摘要
当前的医学问题答复系统很难处理长期,详细和非正式措辞的问题,该问题称为消费者健康问题(CHQ)。为了解决这个问题,我们通过知识基础和语义自学介绍了一个医学问题理解和回答系统。我们的系统是一条管道,首先使用监督的摘要损失来汇总长,医学,用户编写的问题。然后,我们的系统执行两步检索以返回答案。该系统首先将汇总的用户问题与来自受信任的医学知识库的常见问题解答匹配,然后从相应的答案文档中检索固定数量的相关句子。在没有标签以进行问题匹配或回答相关性的情况下,我们设计了3种新颖,自我监管和语义引导的损失。我们根据两个强大的基于检索的问题回答基线来评估我们的模型。评估人员提出自己的问题,并根据我们的基本线和自己的系统根据其相关性进行评分。他们发现我们的系统会检索更相关的答案,同时达到20倍的速度。我们的自我监督损失还有助于总结器在胭脂以及人类评估指标中获得更高的分数。我们发布我们的代码以鼓励进一步的研究。
Current medical question answering systems have difficulty processing long, detailed and informally worded questions submitted by patients, called Consumer Health Questions (CHQs). To address this issue, we introduce a medical question understanding and answering system with knowledge grounding and semantic self-supervision. Our system is a pipeline that first summarizes a long, medical, user-written question, using a supervised summarization loss. Then, our system performs a two-step retrieval to return answers. The system first matches the summarized user question with an FAQ from a trusted medical knowledge base, and then retrieves a fixed number of relevant sentences from the corresponding answer document. In the absence of labels for question matching or answer relevance, we design 3 novel, self-supervised and semantically-guided losses. We evaluate our model against two strong retrieval-based question answering baselines. Evaluators ask their own questions and rate the answers retrieved by our baselines and own system according to their relevance. They find that our system retrieves more relevant answers, while achieving speeds 20 times faster. Our self-supervised losses also help the summarizer achieve higher scores in ROUGE, as well as in human evaluation metrics. We release our code to encourage further research.