论文标题

可解释的AI用于可解释的信用评分

Explainable AI for Interpretable Credit Scoring

论文作者

Demajo, Lara Marie, Vella, Vince, Dingli, Alexiei

论文摘要

随着人工智能(AI)的不断增长的成就以及最近在金融技术(Fintech)中增强的热情,信用评分等应用程序已获得了实质性的学术兴趣。信用评分有助于财务专家就是否接受贷款申请做出更好的决定,因此不接受具有很高违约可能性的贷款。除了此类信用评分模型所面临的嘈杂和高度不平衡的数据挑战外,诸如《通用数据保护法规》(GDPR)引入的“解释权”和《同等信用机会法》(ECOA)提出的“权利解释”(ECOA)都增加了模型解释性的需求,以确保算法的决定是可理解和相当的算法决定的。最近引入的一个有趣的概念是可解释的AI(XAI),该概念的重点是使黑框模型更加易于解释。在这项工作中,我们提出了一种既准确又可以解释的信用评分模型。对于分类,使用极端梯度提升(XGBoost)模型实现了家庭资产信贷(HELOC)和贷款俱乐部(LC)数据集的最新性能。然后,使用360度的解释框架进一步增强了该模型,该框架提供了不同的解释(即不同情况下不同的人所要求的不同的解释(即基于本地特征的全局,基于本地实例和本地实例)。通过使用具有功能固定的,具有应用的和人为接收的分析的评估表明,所提供的解释是简单的,一致的,并且满足了六个预定的假设测试,以实现正确性,有效性,易于理解,细节充足和可信赖性。

With the ever-growing achievements in Artificial Intelligence (AI) and the recent boosted enthusiasm in Financial Technology (FinTech), applications such as credit scoring have gained substantial academic interest. Credit scoring helps financial experts make better decisions regarding whether or not to accept a loan application, such that loans with a high probability of default are not accepted. Apart from the noisy and highly imbalanced data challenges faced by such credit scoring models, recent regulations such as the `right to explanation' introduced by the General Data Protection Regulation (GDPR) and the Equal Credit Opportunity Act (ECOA) have added the need for model interpretability to ensure that algorithmic decisions are understandable and coherent. An interesting concept that has been recently introduced is eXplainable AI (XAI), which focuses on making black-box models more interpretable. In this work, we present a credit scoring model that is both accurate and interpretable. For classification, state-of-the-art performance on the Home Equity Line of Credit (HELOC) and Lending Club (LC) Datasets is achieved using the Extreme Gradient Boosting (XGBoost) model. The model is then further enhanced with a 360-degree explanation framework, which provides different explanations (i.e. global, local feature-based and local instance-based) that are required by different people in different situations. Evaluation through the use of functionallygrounded, application-grounded and human-grounded analysis show that the explanations provided are simple, consistent as well as satisfy the six predetermined hypotheses testing for correctness, effectiveness, easy understanding, detail sufficiency and trustworthiness.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源