论文标题
知名度学习的源通道编码用于高保真图像语义传输
Perceptual Learned Source-Channel Coding for High-Fidelity Image Semantic Transmission
论文作者
论文摘要
作为一种实现端到端无线图像语义传输的一种新颖方法,基于深度学习的联合源通道编码(Deep JSCC)方法在深度学习和沟通社区中都出现了。但是,当前的深入JSCC图像传输系统通常针对传统的失真指标进行优化,例如峰值信噪比(PSNR)或多尺度结构相似性(MS-SSIM)。但是,由于不完善的无线通道,对于低传输速率,这些失真指标偏向于像素的保存而失去意义。为了说明人类语义通信中的视觉感知,开发超出传统PSNR和MS-SSIM指标的新的Deep JSCC系统非常重要。在本文中,我们介绍了对抗性损失,以优化Deep JSCC,该损失倾向于保留全球语义信息和本地纹理。我们新的Deep JSCC体系结构结合了编码器,无线通道,解码器/发电机和判别器,它们在感知和对抗性损失下共同学习。与最先进的工程图像编码传输系统和传统的Deep JSCC系统相比,我们的方法在视觉上产生的结果更加令人愉悦。一项用户研究证实,实现感知上相似的端到端图像传输质量,该建议的方法可以节省约50 \%的无线通道带宽成本。
As one novel approach to realize end-to-end wireless image semantic transmission, deep learning-based joint source-channel coding (deep JSCC) method is emerging in both deep learning and communication communities. However, current deep JSCC image transmission systems are typically optimized for traditional distortion metrics such as peak signal-to-noise ratio (PSNR) or multi-scale structural similarity (MS-SSIM). But for low transmission rates, due to the imperfect wireless channel, these distortion metrics lose significance as they favor pixel-wise preservation. To account for human visual perception in semantic communications, it is of great importance to develop new deep JSCC systems optimized beyond traditional PSNR and MS-SSIM metrics. In this paper, we introduce adversarial losses to optimize deep JSCC, which tends to preserve global semantic information and local texture. Our new deep JSCC architecture combines encoder, wireless channel, decoder/generator, and discriminator, which are jointly learned under both perceptual and adversarial losses. Our method yields human visually much more pleasing results than state-of-the-art engineered image coded transmission systems and traditional deep JSCC systems. A user study confirms that achieving the perceptually similar end-to-end image transmission quality, the proposed method can save about 50\% wireless channel bandwidth cost.