Representative Interpretations [Eng]
Lam et al. / Finding Representative Interpretations on Convolutional Neural Networks / ICCV 2021
한국어로 쓰인 리뷰를 읽으려면 여기를 누르세요.
1. Problem definition
Despite the success of deep learning models on various tasks, there is a lack of interpretability to understand the decision logic behind deep learning models. In fields where decision-making is critical to results, such as process systems and healthcare, it is hard to use models that lack interpretability due to reliability issues. It requires sufficient interpretability to make deep learning models widely applicable.
In this paper, the authors propose a new framework to interpret the decision-making process of Deep convolutional neural networks(CNNs) which are the basic architecture of many deep learning models. The goal is to develop representative interpretations of a trained CNN to reveal the common semantics data that contribute to many closely related predictions.
How can we find such representative interpretations of a trained CNN? Before reviewing the details, I introduce the summary for the paper.
Consider a function that maps the feature map produced by the last convolutional layer to the logits that denote the final decision.
Since this function is a piecewise linear function, it applies different decision logics for regions separated by linear boundaries.
For each image, the authors propose to solve the optimization problem to construct a subset of linear boundaries that provides good representative interpretations.
[Opinion]
It is reasonable to determine a CNN model with ReLU activation functions as a target to interpret, since this architecture has been sufficiently validated to perform well. Furthermore, the proposed method use optimization rather than heuristics so that it can give trustworthy solutions.
2. Motivation
Related Work
There are various types of existing interpretation methods for CNNs.
Conceptual interpretation methods
identify a set of concepts that contribute to the predictions on a pre-defined group of conceptually similar images.
These methods require sophisticated customization on deep neural networks.
Example-based methods
Find exemplar images to interpret the decision of a deep neural network.
Prototype-based methods summarize the entire model using a small number of instances as prototypes.
The selection of prototypes considers little about the decision process of the model.
Idea
In this paper, the goal is to provide representative interpretations in a general CNN model by considering decision boundaries.
Find the linear decision boundaries of the convex polytopes that encode the decision logic of a trained CNN.
This problem can be formulated as a co-clustering problem. The co-clustering problem means that it finds one cluster for the set of similar images and the other cluster for the set of linear boundaries that cover the similar images.
Convert the co-clustering problem into a submodular cost submodular cover (SCSC) problem to make the problem feasible.
3. Method
Setting
Consider image classification using a CNN with ReLU activation functions.
Reference images denote unlabeled images that we want to interpret by this method.
Representative Interpretations
Before formulating our problem, we have to specify a goal to find representative interpretations.
[Representative interpretation]
[Linear boundaries]
[Goal]
Finding Representative Interpretations
What is a 'good' representative interpretation? It requires two conditions:
Avoid covering images in different classes.
It can be formulated as the following optimization problem. The authors call this problem as the co-clustering problem, since it finds one cluster for the set of similar images and the other cluster for the set of linear boundaries that cover the similar images simultaneously.
[Co-clustering Problem]
However, a set optimization problem such as the co-clustering problem is computationally complex to optimize. Therefore, in this paper, the authors:
define submodular optimization problem to make the problem feasible.
What is Submodular Optimization?
A set optimization problem that finds the optimal subset from candidates is computationally complex, since the computational cost increases exponentially as the number of candidates increases.
When the objective function satisfies submodularity, the greedy algorithm achieves at least a constant fraction of the objective value obtained by the optimal solution.
Therefore, submodular optimization makes a set optimization problem feasible with guaranteeing a sufficiently good performance.
Submodularity requires diminishing return property. You can check the details here.
[Opinion]
Submodular Cost Submodular Cover problem
[SCSC Problem]
This formulation satisfies conditions for submodular cost and submodular cover. You can check it in Appendix A of the paper.
Finally, the SCSC problem can be solved by iteratively selecting a linear boundary through the following greedy algorithm.
Ranking Similar Images
[Semantic Distance]
4. Experiment & Result
Experimental setup
The authors compare representative interpretation (RI) method with Automatic Concept-based Explanation(ACE), CAM-based methods(Grad-CAM, Grad-CAM++, Score-CAM).
Dataset: Gender Classification (GC), ASIRRA, Retinal OCT Images (RO), FOOD datasets
Target model: VGG-19
Result
Case Study
This experiment evaluates if each method provides a proper interpretation for similar images.
The first row shows the result retrieved by RI method. Unlike the other methods, the heat maps in images indicate consistent semantics in the images.
RI method can successfully find the interpretation for the input image, as well as a set of images sharing the same interpretation.
Quantitative Experiment
In this experiment, the authors quantitatively evaluate how computed interpretations can be used to classify unseen dataset. The following two measures:
[Average Drop (AD)]
[Average Increase (AI)]
5. Conclusion
In this paper, a co-clustering problem is formulated to interpret the decision-making process of CNN by considering decision boundaries.
To solve the co-clustering problem, the greedy algorithm can be applied by converting it into the SCSC problem.
It has been experimentally shown that proposed representative interpretations reflect common semantics in the unseen images.
Take home message
As deep neural networks have been widely used in various fields, it is more important to interpret a decision logic of DNNs. In this spirit, it is impressive to suggest representative interpretations by considering decision boundaries and I hope to extend such studies further.
Author / Reviewer information
Author
장원준 (Wonjoon Chang)
KAIST AI, Statistical Artificial Intelligence Lab.
one_jj@kaist.ac.kr
Research Topics: Explainable AI, Time series analysis.
https://github.com/onejoon
Reviewer
Reference & Additional materials
Lam, P. C. H., Chu, L., Torgonskiy, M., Pei, J., Zhang, Y., & Wang, L. (2021). Finding representative interpretations on convolutional neural networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision.
Ghorbani, A., Wexler, J., Zou, J., & Kim, B. (2019). Towards automatic concept-based explanations.
Kim, B., Khanna, R., & Koyejo, O. O. (2016). Examples are not enough, learn to criticize! criticism for interpretability. Advances in neural information processing systems, 29.
Cho, S., Chang, W., Lee, G., & Choi, J. (2021, August). Interpreting Internal Activation Patterns in Deep Temporal Neural Networks by Finding Prototypes. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining.
Chu, L., Hu, X., Hu, J., Wang, L., & Pei, J. (2018, July). Exact and consistent interpretation for piecewise linear neural networks: A closed form solution. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining.
Last updated