论文标题
通过深度蒸馏进行散焦模糊检测
Defocus Blur Detection via Depth Distillation
论文作者
论文摘要
Defocus Blur检测(DBD)的目的是将聚焦区和异常区域与单个图像以像素 - 智能为单位。由于广泛用于数码相机和智能手机摄影,因此这项任务已引起了很多关注。但是,在部分散焦图像中确定晦涩的均匀区域和边界过渡仍然具有挑战性。为了解决这些问题,我们首次将深度信息引入DBD。固定相机参数后,我们认为DBD的准确性与场景深度高度相关。因此,我们将深度信息视为DBD的近似软标签,并提出了一个受知识蒸馏启发的联合学习框架。详细说明,我们从地面真理和从训练有素的深度估计网络中提取的深度从地面真相中学习了散焦。因此,尖锐的区域将提供深度估计的强大先验,而模糊检测也从蒸馏深度获得了好处。此外,我们在完全卷积网络(FCN)中提出了一个新颖的解码器作为我们的网络结构。在解码器的每个级别中,我们设计了选择性接收场块(SRFB),以有效合并多尺度特征,并将侧面输出重用为监督引导的注意块(SAB)。与以前的方法不同,所提出的解码器构建了接收场金字塔,并简单有效地强调了显着区域。实验表明,我们的方法在两个流行数据集上的其他最先进方法优于11种其他最先进的方法。我们的方法在单个GPU上还以30 fps的速度运行,该GPU比以前的工作快2倍。该代码可在以下网址找到:https://github.com/vinthony/depth-distillation
Defocus Blur Detection(DBD) aims to separate in-focus and out-of-focus regions from a single image pixel-wisely. This task has been paid much attention since bokeh effects are widely used in digital cameras and smartphone photography. However, identifying obscure homogeneous regions and borderline transitions in partially defocus images is still challenging. To solve these problems, we introduce depth information into DBD for the first time. When the camera parameters are fixed, we argue that the accuracy of DBD is highly related to scene depth. Hence, we consider the depth information as the approximate soft label of DBD and propose a joint learning framework inspired by knowledge distillation. In detail, we learn the defocus blur from ground truth and the depth distilled from a well-trained depth estimation network at the same time. Thus, the sharp region will provide a strong prior for depth estimation while the blur detection also gains benefits from the distilled depth. Besides, we propose a novel decoder in the fully convolutional network(FCN) as our network structure. In each level of the decoder, we design the Selective Reception Field Block(SRFB) for merging multi-scale features efficiently and reuse the side outputs as Supervision-guided Attention Block(SAB). Unlike previous methods, the proposed decoder builds reception field pyramids and emphasizes salient regions simply and efficiently. Experiments show that our approach outperforms 11 other state-of-the-art methods on two popular datasets. Our method also runs at over 30 fps on a single GPU, which is 2x faster than previous works. The code is available at: https://github.com/vinthony/depth-distillation