暂无图片
暂无图片
暂无图片
暂无图片
暂无图片

特征属性之外的可解释性:用概念激活向量进行定量测试

北邮数据科学与商务智能实验室 2020-06-05
1592

原标题:Interpretability Beyond FeatureAttribution: Quantitative Testing with Concept Activation Vectors (TCAV)

作者:Been Kim  MartinWattenberg  Justin Gilmer  Carrie Cai  James Wexler

Fernanda Viegas  Rory Sayres

中文摘要:

因为深度学习模型的规模,复杂性,不透明的内部状态,其解释一直是一个挑战。此外,许多系统,如图像分类器,都是基于低级特征而不是高级概念。为了应对这些挑战,我们引入了概念激活向量(CAVs),它用人类友好的概念来解释神经网络的内部状态。关键的想法是把神经网络的高维内部状态看作是一种辅助,而不是一种障碍。我们展示了如何将CAVs作为一项技术的一部分,使用CAVsTCAV)进行测试,该技术使用方向导数来量化用户定义的概念对分类结果的重要性,例如,斑马的预测对条纹的存在有多敏感。以图像分类领域为试验场,我们描述了CAVs如何用于探索假设并生成标准图像分类网络和医学应用系统的理解。

英文摘要:

The interpretation of deep learningmodels is a challenge due to their size, complexity, and often opaque internalstate. In addition, many systems, such as image classifiers, operate onlow-level features rather than high-level concepts. To address thesechallenges, we introduce Concept Activation Vectors (CAVs), which provide aninterpretation of a neural net’s internal state in terms of human-friendlyconcepts. The key idea is to view the high-dimensional internal state of aneural net as an aid, not an obstacle. We show how to use CAVs as part of atechnique, Testing with CAVs (TCAV), that uses directional derivatives toquantify the degree to which a user-defined concept is important to aclassification result–for example, how sensitive a prediction of zebra is tothe presence of stripes. Using the domain of image classification as a testingground, we describe how CAVs may be used to explore hypotheses and generateinsights for a standard image classification network as well as a medicalapplication.

文献总结:

点击“阅读原文”,了解论文详情!

文章转载自北邮数据科学与商务智能实验室,如果涉嫌侵权,请发送邮件至:contact@modb.pro进行举报,并提供相关证据,一经查实,墨天轮将立刻删除相关内容。

评论