论文标题
深度多式模式指南
Deep Multimodal Guidance for Medical Image Classification
论文作者
论文摘要
医学成像是现代医学治疗和诊断的基石。但是,对于特定静脉局体任务的成像方式通常涉及使用特定模式的可行性(例如,短期等待时间,低成本,快速获取,放射线/侵入性降低)与临床任务(例如,诊断准确性,治疗计划的效率,治疗计划的效率)。在这项工作中,我们旨在运用从较不可行但表现更好(优越)模态中学到的知识,以指导利用更易于且表现不佳(劣等)模态,并将其转向提高性能。我们专注于在基于图像的诊断中应用深度学习。我们开发了一种轻巧的指导模型,该模型在训练仅消耗次要模态的模型时,利用从上级模式中学到的潜在表示。我们在两个临床应用的背景下检查了我们方法的优势:多任务皮肤病变分类来自临床和皮肤镜图像以及来自多序列磁共振成像(MRI)和组织病理学图像的脑肿瘤分类。对于这两种情况,我们在不需要出色的模态的情况下显示出劣质模式的诊断性能。此外,在脑肿瘤分类的情况下,我们的方法的表现优于在上级模态上训练的模型,同时产生与推理过程中使用两种方式的模型相当的结果。
Medical imaging is a cornerstone of therapy and diagnosis in modern medicine. However, the choice of imaging modality for a particular theranostic task typically involves trade-offs between the feasibility of using a particular modality (e.g., short wait times, low cost, fast acquisition, reduced radiation/invasiveness) and the expected performance on a clinical task (e.g., diagnostic accuracy, efficacy of treatment planning and guidance). In this work, we aim to apply the knowledge learned from the less feasible but better-performing (superior) modality to guide the utilization of the more-feasible yet under-performing (inferior) modality and steer it towards improved performance. We focus on the application of deep learning for image-based diagnosis. We develop a light-weight guidance model that leverages the latent representation learned from the superior modality, when training a model that consumes only the inferior modality. We examine the advantages of our method in the context of two clinical applications: multi-task skin lesion classification from clinical and dermoscopic images and brain tumor classification from multi-sequence magnetic resonance imaging (MRI) and histopathology images. For both these scenarios we show a boost in diagnostic performance of the inferior modality without requiring the superior modality. Furthermore, in the case of brain tumor classification, our method outperforms the model trained on the superior modality while producing comparable results to the model that uses both modalities during inference.