论文标题
通过多个实例学习的组织病理学图像分类和本地化的多分辨率模型
A Multi-resolution Model for Histopathology Image Classification and Localization with Multiple Instance Learning
论文作者
论文摘要
组织病理学图像为疾病诊断提供了丰富的信息。大量的组织病理学图像已被数字化为高分辨率的整个幻灯片图像,从而开发了开发计算图像分析工具以减少病理学家的工作量并有可能改善内部和内部观察者一致性的机会。整个幻灯片图像分析的大多数工作都集中在小型预选区域的分类或分割上,这需要细粒的注释,并且不宽容以扩展大规模的整个幻灯片分析。在本文中,我们提出了一个多分辨率的多个实例学习模型,该模型利用显着性图来检测可疑区域的细颗粒等级预测。我们的模型不依赖昂贵的区域或像素级注释,而是只能使用幻灯片级别的标签对端到端进行训练。该模型是在包含830名患者的20,229个幻灯片的大型前列腺活检数据集上开发的。该模型的准确性达到92.7%,Cohen的Kappa的良性为81.8%,低年级(即1年级)和高级(即成绩组> = 2)预测,接收器操作特征曲线(AUROC)下的面积为98.2%,平均精确性(AP)为97.4%,分别为97.4%。该模型在外部数据集上获得了99.4%的AUROC,AP为99.8%。
Histopathological images provide rich information for disease diagnosis. Large numbers of histopathological images have been digitized into high resolution whole slide images, opening opportunities in developing computational image analysis tools to reduce pathologists' workload and potentially improve inter- and intra- observer agreement. Most previous work on whole slide image analysis has focused on classification or segmentation of small pre-selected regions-of-interest, which requires fine-grained annotation and is non-trivial to extend for large-scale whole slide analysis. In this paper, we proposed a multi-resolution multiple instance learning model that leverages saliency maps to detect suspicious regions for fine-grained grade prediction. Instead of relying on expensive region- or pixel-level annotations, our model can be trained end-to-end with only slide-level labels. The model is developed on a large-scale prostate biopsy dataset containing 20,229 slides from 830 patients. The model achieved 92.7% accuracy, 81.8% Cohen's Kappa for benign, low grade (i.e. Grade group 1) and high grade (i.e. Grade group >= 2) prediction, an area under the receiver operating characteristic curve (AUROC) of 98.2% and an average precision (AP) of 97.4% for differentiating malignant and benign slides. The model obtained an AUROC of 99.4% and an AP of 99.8% for cancer detection on an external dataset.