论文标题

从弱监督的学习到积极学习

From Weakly Supervised Learning to Active Learning

论文作者

Cabannes, Vivien

论文摘要

自从最近的监督学习成功以来,应用数学和机器计算已经引起了很多希望。许多行业的从业者一直在尝试从旧范式切换到机器学习。有趣的是,与微调模型相比,这些数据科学家花费更多的时间取消,注释和清洁数据。该论文是由以下问题激励的:我们可以比监督学习的一个更通用的框架来从混乱数据中学习吗? 假设数据收集的瓶颈在于注释,则通过弱监督学习的镜头来解决这个问题。我们将弱的监督建模为给予而不是独特的目标,即一组候选目标。我们认为,应该寻找与大多数观测值相匹配的``乐观''功能。这使我们能够得出一个原理来消除部分标签。我们还讨论了将无监督的学习技术纳入我们的框架的优势,特别是通过扩散技术接近的歧管正则化,为此,我们得出了一种新的算法,该算法通过输入维度比基线方法更好地扩展。 最后,我们从被动切换到主动弱监督的学习,引入了``主动标签''框架,其中从业人员可以查询有关所选数据的弱信息。除其他外,我们利用一个事实,即不需要完整的信息来访问随机梯度并执行随机梯度下降。

Applied mathematics and machine computations have raised a lot of hope since the recent success of supervised learning. Many practitioners in industries have been trying to switch from their old paradigms to machine learning. Interestingly, those data scientists spend more time scrapping, annotating and cleaning data than fine-tuning models. This thesis is motivated by the following question: can we derive a more generic framework than the one of supervised learning in order to learn from clutter data? This question is approached through the lens of weakly supervised learning, assuming that the bottleneck of data collection lies in annotation. We model weak supervision as giving, rather than a unique target, a set of target candidates. We argue that one should look for an ``optimistic'' function that matches most of the observations. This allows us to derive a principle to disambiguate partial labels. We also discuss the advantage to incorporate unsupervised learning techniques into our framework, in particular manifold regularization approached through diffusion techniques, for which we derived a new algorithm that scales better with input dimension then the baseline method. Finally, we switch from passive to active weakly supervised learning, introducing the ``active labeling'' framework, in which a practitioner can query weak information about chosen data. Among others, we leverage the fact that one does not need full information to access stochastic gradients and perform stochastic gradient descent.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源