论文标题

重新思考对比度学习中的最小足够代表

Rethinking Minimal Sufficient Representation in Contrastive Learning

论文作者

Wang, Haoqing, Guo, Xun, Deng, Zhi-Hong, Lu, Yan

论文摘要

对数据之间的对比学习在自我监督的表示学习和学习表示的领域取得了杰出的成功,这在广泛的下游任务中很有用。由于一个视图的所有监督信息都来自另一种观点,因此对比度学习大致获得了最小的足够表示,其中包含共享信息,并消除了视图之间的非共享信息。考虑到下游任务的多样性,不能保证所有与任务相关的信息在观点之间共享。因此,我们假设不共享的与任务相关的信息不能忽略,从理论上则证明,对比度学习中的最小足够表示不足以使下游任务导致性能退化。这揭示了一个新问题,即对比度学习模型有过度适合观点之间共享信息的风险。为了减轻此问题,我们建议将表示形式和输入之间的相互信息提高,以大约引入更多与任务相关的信息,因为我们无法在培训过程中使用任何下游任务信息。广泛的实验验证了我们的分析的合理性和方法的有效性。它大大提高了下游任务中几种经典的对比学习模型的性能。我们的代码可在https://github.com/haoqing-wang/infocl上找到。

Contrastive learning between different views of the data achieves outstanding success in the field of self-supervised representation learning and the learned representations are useful in broad downstream tasks. Since all supervision information for one view comes from the other view, contrastive learning approximately obtains the minimal sufficient representation which contains the shared information and eliminates the non-shared information between views. Considering the diversity of the downstream tasks, it cannot be guaranteed that all task-relevant information is shared between views. Therefore, we assume the non-shared task-relevant information cannot be ignored and theoretically prove that the minimal sufficient representation in contrastive learning is not sufficient for the downstream tasks, which causes performance degradation. This reveals a new problem that the contrastive learning models have the risk of over-fitting to the shared information between views. To alleviate this problem, we propose to increase the mutual information between the representation and input as regularization to approximately introduce more task-relevant information, since we cannot utilize any downstream task information during training. Extensive experiments verify the rationality of our analysis and the effectiveness of our method. It significantly improves the performance of several classic contrastive learning models in downstream tasks. Our code is available at https://github.com/Haoqing-Wang/InfoCL.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源