论文标题
具有高速网络的科学论文的多任务推荐系统
Multi-task recommendation system for scientific papers with high-way networks
论文作者
论文摘要
如今,从研究社区中撰写的大量论文中查找和选择最相关的科学论文是研究人员的关键挑战之一。众所周知,关于学者和院士的研究兴趣的许多信息属于他们阅读的论文。分析和从这些论文中提取上下文特征可以帮助我们向它们提出最相关的论文。在本文中,我们提出了一个多任务推荐系统(RS),该系统可以预测论文建议并生成其元数据(例如关键字)。该系统被实现为三阶段的深神经网络编码器,该编码器试图将更长的文本序列映射到嵌入矢量,并同时学习以预测特定用户和论文的关键字的推荐率。这种方法背后的动机是,本文表示为关键字的主题是研究人员偏好的有用预测指标。为了实现这一目标,我们使用RNN,高速公路和卷积神经网络的系统组合来训练端到端的上下文感知的协作矩阵。我们的应用程序使用高速公路网络对系统进行了非常深的训练,结合了RNN和CNN的好处,以找到最重要的因素并做出潜在的代表。高速公路网络使我们能够通过学习更复杂的语义结构表示来增强传统的RNN和CNN管道。使用此方法,我们还可以克服冷启动问题,并通过大量文本序列学习潜在特征。
Finding and selecting the most relevant scientific papers from a large number of papers written in a research community is one of the key challenges for researchers these days. As we know, much information around research interest for scholars and academicians belongs to papers they read. Analysis and extracting contextual features from these papers could help us to suggest the most related paper to them. In this paper, we present a multi-task recommendation system (RS) that predicts a paper recommendation and generates its meta-data such as keywords. The system is implemented as a three-stage deep neural network encoder that tries to maps longer sequences of text to an embedding vector and learns simultaneously to predict the recommendation rate for a particular user and the paper's keywords. The motivation behind this approach is that the paper's topics expressed as keywords are a useful predictor of preferences of researchers. To achieve this goal, we use a system combination of RNNs, Highway and Convolutional Neural Networks to train end-to-end a context-aware collaborative matrix. Our application uses Highway networks to train the system very deep, combine the benefits of RNN and CNN to find the most important factor and make latent representation. Highway Networks allow us to enhance the traditional RNN and CNN pipeline by learning more sophisticated semantic structural representations. Using this method we can also overcome the cold start problem and learn latent features over large sequences of text.