论文标题
基于渠道关节和软杆的双流计算机生成的图像检测网络
Dual Stream Computer-Generated Image Detection Network Based On Channel Joint And Softpool
论文作者
论文摘要
随着计算机图形技术的开发,计算机软件合成的图像越来越接近照片。尽管计算机图形技术为我们带来了游戏和电影领域中的盛大视觉盛宴,但它也可以被不良意愿的人使用来指导公众意见并造成政治危机或社会动荡。因此,如何将计算机生成的图形(CG)与照片(PG)区分开已成为数字图像取证领域的重要主题。本文提出了基于通道关节和软杆的双流卷积神经网络。提出的网络体系结构包括一个用于提取图像噪声信息的残差模块和一个用于捕获图像浅的语义信息的联合通道信息提取模块。此外,我们还设计了一个残留结构,以增强特征提取并减少剩余流中信息的损失。联合通道信息提取模块可以获取输入图像的浅语义信息,该信息可以用作残差模块的信息补充块。整个网络使用Softpool来减少图像下采样的信息丢失。最后,我们融合了两个流以获得分类结果。 SPL2018和DSTOK上的实验表明,所提出的方法的表现优于现有方法,尤其是在DSTOK数据集上。例如,我们的模型的性能超过了最新的3%。
With the development of computer graphics technology, the images synthesized by computer software become more and more closer to the photographs. While computer graphics technology brings us a grand visual feast in the field of games and movies, it may also be utilized by someone with bad intentions to guide public opinions and cause political crisis or social unrest. Therefore, how to distinguish the computer-generated graphics (CG) from the photographs (PG) has become an important topic in the field of digital image forensics. This paper proposes a dual stream convolutional neural network based on channel joint and softpool. The proposed network architecture includes a residual module for extracting image noise information and a joint channel information extraction module for capturing the shallow semantic information of image. In addition, we also design a residual structure to enhance feature extraction and reduce the loss of information in residual flow. The joint channel information extraction module can obtain the shallow semantic information of the input image which can be used as the information supplement block of the residual module. The whole network uses SoftPool to reduce the information loss of down-sampling for image. Finally, we fuse the two flows to get the classification results. Experiments on SPL2018 and DsTok show that the proposed method outperforms existing methods, especially on the DsTok dataset. For example, the performance of our model surpasses the state-of-the-art by a large margin of 3%.