论文标题
深层神经网络中嘈杂标签的嘈杂标签的元结构
Elucidating Meta-Structures of Noisy Labels in Semantic Segmentation by Deep Neural Networks
论文作者
论文摘要
嘈杂标签对深层神经网络(DNN)的监督培训已在图像分类中进行了广泛的研究,但在图像分段中却少得多。我们对嘈杂分割标签训练的DNN的学习行为的理解仍然有限。我们解决了生物学显微镜图像的二元分割和自然图像的多级分割的缺陷。我们根据其噪声过渡矩阵(NTM)对分割标签进行分类,并比较了由不同类型标签训练的DNN的性能。当我们随机采样一小部分(例如10%)或翻转大量的地面真实标签(例如90%)以训练DNN时,它们的分割性能在很大程度上保持不变。这表明DNN在标签中隐藏的结构,而不是像素级标签,本身是在其监督的语义分割培训中。我们称这些隐藏的结构元结构。当使用对元结构进行不同扰动的标签被用于训练DNN时,它们在特征提取和分割方面的性能始终如一。相比之下,添加元结构信息可大大提高二进制语义分割中无监督模型的性能。我们在数学上以空间密度分布为单位。我们在理论上和实验上展示了该公式如何解释DNN的关键学习行为。
Supervised training of deep neural networks (DNNs) by noisy labels has been studied extensively in image classification but much less in image segmentation. Our understanding of the learning behavior of DNNs trained by noisy segmentation labels remains limited. We address this deficiency in both binary segmentation of biological microscopy images and multi-class segmentation of natural images. We classify segmentation labels according to their noise transition matrices (NTMs) and compare performance of DNNs trained by different types of labels. When we randomly sample a small fraction (e.g., 10%) or flip a large fraction (e.g., 90%) of the ground-truth labels to train DNNs, their segmentation performance remains largely unchanged. This indicates that DNNs learn structures hidden in labels rather than pixel-level labels per se in their supervised training for semantic segmentation. We call these hidden structures meta-structures. When labels with different perturbations to the meta-structures are used to train DNNs, their performance in feature extraction and segmentation degrades consistently. In contrast, addition of meta-structure information substantially improves performance of an unsupervised model in binary semantic segmentation. We formulate meta-structures mathematically as spatial density distributions. We show theoretically and experimentally how this formulation explains key observed learning behavior of DNNs.