论文标题
无需解释模型的可解释AI
Explainable AI without Interpretable Model
论文作者
论文摘要
只要AI存在,解释性就一直是AI的挑战。随着AI在社会中的最近使用,AI系统将能够解释其结果背后的推理也比以往任何时候都变得更加重要。特别是如果使用机器学习对AI系统进行了培训,则它倾向于包含太多参数,无法分析和理解它们,这导致它们被称为“ Black-Box”系统。大多数可解释的AI(XAI)方法基于提取可解释的模型,该模型可用于产生解释。但是,可解释的模型不一定准确地映射到原始的黑框模型。此外,最终用户的可解释模型的可理解性仍然值得怀疑。本文提出的上下文重要性和效用(CIU)的概念使得可以直接对黑盒结果产生类似人类的解释,而无需创建可解释的模型。因此,CIU的解释准确地映射到黑框模型本身。 CIU是完全不合时宜的模型,可以与任何黑盒系统一起使用。除了重要性之外,与大多数现有的XAI方法相比,决策理论中众所周知的实用概念为解释提供了新的方面。最后,CIU可以在任何抽象的水平上产生解释,并使用不同的词汇和其他互动方式,这使得可以根据上下文和目标用户调整解释和互动。
Explainability has been a challenge in AI for as long as AI has existed. With the recently increased use of AI in society, it has become more important than ever that AI systems would be able to explain the reasoning behind their results also to end-users in situations such as being eliminated from a recruitment process or having a bank loan application refused by an AI system. Especially if the AI system has been trained using Machine Learning, it tends to contain too many parameters for them to be analysed and understood, which has caused them to be called `black-box' systems. Most Explainable AI (XAI) methods are based on extracting an interpretable model that can be used for producing explanations. However, the interpretable model does not necessarily map accurately to the original black-box model. Furthermore, the understandability of interpretable models for an end-user remains questionable. The notions of Contextual Importance and Utility (CIU) presented in this paper make it possible to produce human-like explanations of black-box outcomes directly, without creating an interpretable model. Therefore, CIU explanations map accurately to the black-box model itself. CIU is completely model-agnostic and can be used with any black-box system. In addition to feature importance, the utility concept that is well-known in Decision Theory provides a new dimension to explanations compared to most existing XAI methods. Finally, CIU can produce explanations at any level of abstraction and using different vocabularies and other means of interaction, which makes it possible to adjust explanations and interaction according to the context and to the target users.