论文标题

如何使深度RL在实践中起作用

How to Make Deep RL Work in Practice

论文作者

Rao, Nirnai, Aljalbout, Elie, Sauer, Axel, Haddadin, Sami

论文摘要

近年来,深入的加强学习(RL)可以解决挑战性的控制问题。为了能够将RL用于大规模的现实世界应用,其性能是必要的一定程度的可靠性。报告的最新算法的结果通常很难繁殖。原因之一是某些实施细节会对性能产生重大影响。通常,这些细节并未被强调为实现最新性能的重要技术。此外,默认情况下通常使用监督学习的技术,但以不同的方式不理解的方式影响强化学习环境中的算法。在本文中,我们研究了某些初始化,输入归一化和自适应学习技术对最新RL算法性能的影响。我们提出建议,默认情况下要使用哪种技术,并突出显示可能从专门针对RL量身定制的解决方案中受益的领域。

In recent years, challenging control problems became solvable with deep reinforcement learning (RL). To be able to use RL for large-scale real-world applications, a certain degree of reliability in their performance is necessary. Reported results of state-of-the-art algorithms are often difficult to reproduce. One reason for this is that certain implementation details influence the performance significantly. Commonly, these details are not highlighted as important techniques to achieve state-of-the-art performance. Additionally, techniques from supervised learning are often used by default but influence the algorithms in a reinforcement learning setting in different and not well-understood ways. In this paper, we investigate the influence of certain initialization, input normalization, and adaptive learning techniques on the performance of state-of-the-art RL algorithms. We make suggestions which of those techniques to use by default and highlight areas that could benefit from a solution specifically tailored to RL.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源