论文标题

学习使用图形神经网络进行哈希进行推荐系统

Learning to Hash with Graph Neural Networks for Recommender Systems

论文作者

Tan, Qiaoyu, Liu, Ninghao, Zhao, Xing, Yang, Hongxia, Zhou, Jingren, Hu, Xia

论文摘要

图表学习在支持高质量候选搜索方面吸引了很多关注。尽管它可以在学习用户项目交互网络中学习对象的嵌入向量的有效性,但推断用户在连续嵌入空间中偏好的计算成本还是很大的。在这项工作中,我们研究了使用图神经网络(GNN)哈希的问题进行高质量检索,并提出了一个简单而有效的离散表示学习框架,以共同学习连续和离散的代码。具体而言,提出了带有GNNS(Hashgnn)的深度散列,其中包括两个组件,一个用于学习节点表示的GNN编码器,以及用于将表示表示形式编码为哈希代码的哈希层。通过共同优化两个损失,即重建观察到的链接的重建损失,并通过保留哈希代码的相对顺序排名损失,从而对整个架构进行了训练。提出了一种基于直通估算器(Ste)的新型离散优化策略,并提出了指导。主要思想是避免使用连续嵌入指导的STE反向传播的梯度放大,在这种情况下,我们从学习一个更轻松的网络开始,该网络模仿连续的嵌入,并在训练期间使其在训练过程中发展,直到最终回到Ste。对三个公开可用的全面实验和一个现实世界中的阿里巴巴公司数据集进行了全面的实验表明,我们的模型不仅可以与连续的同行相比,而且在推断期间的运行速度更快。

Graph representation learning has attracted much attention in supporting high quality candidate search at scale. Despite its effectiveness in learning embedding vectors for objects in the user-item interaction network, the computational costs to infer users' preferences in continuous embedding space are tremendous. In this work, we investigate the problem of hashing with graph neural networks (GNNs) for high quality retrieval, and propose a simple yet effective discrete representation learning framework to jointly learn continuous and discrete codes. Specifically, a deep hashing with GNNs (HashGNN) is presented, which consists of two components, a GNN encoder for learning node representations, and a hash layer for encoding representations to hash codes. The whole architecture is trained end-to-end by jointly optimizing two losses, i.e., reconstruction loss from reconstructing observed links, and ranking loss from preserving the relative ordering of hash codes. A novel discrete optimization strategy based on straight through estimator (STE) with guidance is proposed. The principal idea is to avoid gradient magnification in back-propagation of STE with continuous embedding guidance, in which we begin from learning an easier network that mimic the continuous embedding and let it evolve during the training until it finally goes back to STE. Comprehensive experiments over three publicly available and one real-world Alibaba company datasets demonstrate that our model not only can achieve comparable performance compared with its continuous counterpart but also runs multiple times faster during inference.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源