Network Dissection: Quantifying Interpretability of Deep Visual Representations

2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)(2017)

引用 1629|浏览187
暂无评分
摘要
We propose a general framework called Network Dissection for quantifying the interpretability of latent representations of CNNs by evaluating the alignment between individual hidden units and a set of semantic concepts. Given any CNN model, the proposed method draws on a broad data set of visual concepts to score the semantics of hidden units at each intermediate convolutional layer. The units with semantics are given labels across a range of objects, parts, scenes, textures, materials, and colors. We use the proposed method to test the hypothesis that interpretability of units is equivalent to random linear combinations of units, then we apply our method to compare the latent representations of various networks when trained to solve different supervised and self-supervised training tasks. We further analyze the effect of training iterations, compare networks trained with different initializations, examine the impact of network depth and width, and measure the effect of dropout and batch normalization on the interpretability of deep visual representations. We demonstrate that the proposed method can shed light on characteristics of CNN models and training methods that go beyond measurements of their discriminative power.
更多
查看译文
关键词
visual concepts,intermediate convolutional layer,semantic concepts,individual hidden units,Network Dissection,Network dissection,CNN model,deep visual representations,different classification problems,latent representations,representation space,axis-independent property
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要