论文标题
动态多代理系统中强化学习模型的解释
Explanation of Reinforcement Learning Model in Dynamic Multi-Agent System
论文作者
论文摘要
最近,对深度加强学习(DRL)系统的透明度和可解释性的兴趣越来越大。作为我们日常生活中最自然的交流方式,口头解释值得更多关注,因为它们使用户能够更好地了解系统,最终可能导致高水平的信任和平稳的协作。本文报告了为DRL行为代理产生口头解释的新工作。基于规则的模型旨在使用一系列具有先验知识的规则来构建解释。然后,提出了一种学习模型,以通过采用基于规则的解释作为培训数据来扩展对一般情况产生口头解释的隐含逻辑。与基于静态规则的模型相比,学习模型具有更好的灵活性和可推广性。通过客观指标对两种模型的性能进行定量评估。结果表明,两种模型产生的口头解释都提高了用户对DRL系统解释性的主观满意度。此外,学习模型的七个变体旨在说明输入渠道,注意机制和提议的编码器在提高言语解释质量方面的贡献。
Recently, there has been increasing interest in transparency and interpretability in Deep Reinforcement Learning (DRL) systems. Verbal explanations, as the most natural way of communication in our daily life, deserve more attention, since they allow users to gain a better understanding of the system which ultimately could lead to a high level of trust and smooth collaboration. This paper reports a novel work in generating verbal explanations for DRL behaviors agent. A rule-based model is designed to construct explanations using a series of rules which are predefined with prior knowledge. A learning model is then proposed to expand the implicit logic of generating verbal explanation to general situations by employing rule-based explanations as training data. The learning model is shown to have better flexibility and generalizability than the static rule-based model. The performance of both models is evaluated quantitatively through objective metrics. The results show that verbal explanation generated by both models improve subjective satisfaction of users towards the interpretability of DRL systems. Additionally, seven variants of the learning model are designed to illustrate the contribution of input channels, attention mechanism, and proposed encoder in improving the quality of verbal explanation.