论文标题
信心和解释对AI辅助决策的准确性和信任校准的影响
Effect of Confidence and Explanation on Accuracy and Trust Calibration in AI-Assisted Decision Making
论文作者
论文摘要
如今,AI越来越多地帮助人类专家在高风险的情况下做出决定。在这些情况下,完全自动化通常是不受欢迎的,不仅是由于结果的重要性,而且还因为人类专家可以借鉴其领域知识与模型的补充以确保任务成功。我们将这些方案称为AI辅助决策,人类和人工智能的个人优势聚集在一起以优化联合决策结果。他们成功的关键是逐案适当地对AI进行适当的人类信任。知道何时信任或不信任AI可以使人类专家适当地运用其知识,从而在模型可能表现不佳的情况下改善决策结果。这项研究对人类和人工智能仅具有可比性的AI辅助决策进行了案例研究,并探讨了揭示特定案例模型信息的功能是否可以校准信任并改善人类和AI的共同绩效。具体而言,我们研究了显示特定预测的置信度评分和局部解释的效果。通过两个人类的实验,我们表明置信度得分可以帮助校准人们对AI模型的信任,但是仅靠信任校准就不足以改善AI辅助决策,这也可能取决于人类是否可以带来足够的独特知识来补充AI的错误。我们还强调了使用本地解释来制定AI辅助决策方案的问题,并邀请研究社区探索新的方法来解释人类对AI的信任。
Today, AI is being increasingly used to help human experts make decisions in high-stakes scenarios. In these scenarios, full automation is often undesirable, not only due to the significance of the outcome, but also because human experts can draw on their domain knowledge complementary to the model's to ensure task success. We refer to these scenarios as AI-assisted decision making, where the individual strengths of the human and the AI come together to optimize the joint decision outcome. A key to their success is to appropriately \textit{calibrate} human trust in the AI on a case-by-case basis; knowing when to trust or distrust the AI allows the human expert to appropriately apply their knowledge, improving decision outcomes in cases where the model is likely to perform poorly. This research conducts a case study of AI-assisted decision making in which humans and AI have comparable performance alone, and explores whether features that reveal case-specific model information can calibrate trust and improve the joint performance of the human and AI. Specifically, we study the effect of showing confidence score and local explanation for a particular prediction. Through two human experiments, we show that confidence score can help calibrate people's trust in an AI model, but trust calibration alone is not sufficient to improve AI-assisted decision making, which may also depend on whether the human can bring in enough unique knowledge to complement the AI's errors. We also highlight the problems in using local explanation for AI-assisted decision making scenarios and invite the research community to explore new approaches to explainability for calibrating human trust in AI.