📝
Awesome reviews
  • Welcome
  • Paper review
    • [2022 Spring] Paper review
      • RobustNet [Eng]
      • DPT [Kor]
      • DALL-E [Kor]
      • VRT: A Video Restoration Transformer [Kor]
      • Barbershop [Kor]
      • Barbershop [Eng]
      • REFICS [ENG]
      • Deep texture manifold [Kor]
      • SlowFast Networks [Kor]
      • SCAN [Eng]
      • DPT [Kor]
      • Chaining a U-Net With a Residual U-Net for Retinal Blood Vessels Segmentation [Kor]
      • Chaining a U-Net With a Residual U-Net for Retinal Blood Vessels Segmentation [Eng]
      • Patch Cratf : Video Denoising by Deep Modeling and Patch Matching [Eng]
      • LAFITE: Towards Language-Free Training for Text-to-Image Generation [Kor]
      • RegSeg [Eng]
      • D-NeRF [Eng]
      • SimCLR [Kor]
      • LabOR [Kor]
      • LabOR [Eng]
      • SegFormer [Kor]
      • Self-Calibrating Neural Radiance Fields [Kor]
      • Self-Calibrating Neural Radiance Fields [Eng]
      • GIRAFFE [Kor]
      • GIRAFFE [Eng]
      • DistConv [Kor]
      • SCAN [Eng]
      • slowfastnetworks [Kor]
      • Nesterov and Scale-Invariant Attack [Kor]
      • OutlierExposure [Eng]
      • TSNs [Kor]
      • TSNs [Eng]
      • Improving the Transferability of Adversarial Samples With Adversarial Transformations [Kor]
      • VOS: OOD detection by Virtual Outlier Synthesis [Kor]
      • MultitaskNeuralProcess [Kor]
      • RSLAD [Eng]
      • Deep Learning for 3D Point Cloud Understanding: A Survey [Eng]
      • BEIT [Kor]
      • Divergence-aware Federated Self-Supervised Learning [Eng]
      • NeRF-W [Kor]
      • Learning Multi-Scale Photo Exposure Correction [Eng]
      • ReActNet: Towards Precise Binary Neural Network with Generalized Activation Functions [Eng]
      • ViT [Eng]
      • CrossTransformer [Kor]
      • NeRF [Kor]
      • RegNeRF [Kor]
      • Image Inpainting with External-internal Learning and Monochromic Bottleneck [Eng]
      • CLIP-NeRF [Kor]
      • CLIP-NeRF [Eng]
      • DINO: Emerging Properties in Self-Supervised Vision Transformers [Eng]
      • DINO: Emerging Properties in Self-Supervised Vision Transformers [Kor]
      • DatasetGAN [Eng]
      • MOS [Kor]
      • MOS [Eng]
      • PlaNet [Eng]
      • MAE [Kor]
      • Fair Attribute Classification through Latent Space De-biasing [Kor]
      • Fair Attribute Classification through Latent Space De-biasing [Eng]
      • Learning to Adapt in Dynamic, Real-World Environments Through Meta-Reinforcement Learning [Kor]
      • PointNet [Kor]
      • PointNet [Eng]
      • MSD AT [Kor]
      • MM-TTA [Kor]
      • MM-TTA [Eng]
      • M-CAM [Eng]
      • MipNerF [Kor]
      • The Emergence of Objectness: Learning Zero-Shot Segmentation from Videos [Eng]
      • Calibration [Eng]
      • CenterPoint [Kor]
      • YOLOX [Kor]
    • [2021 Fall] Paper review
      • DenseNet [Kor]
      • Time series as image [Kor]
      • mem3d [Kor]
      • GraSP [Kor]
      • DRLN [Kor]
      • VinVL: Revisiting Visual Representations in Vision-Language Models [Eng]
      • VinVL: Revisiting Visual Representations in Vision-Language Models [Kor]
      • NeSyXIL [Kor]
      • NeSyXIL [Eng]
      • RCAN [Kor]
      • RCAN [Eng]
      • MI-AOD [Kor]
      • MI-AOD [Eng]
      • DAFAS [Eng]
      • HyperGAN [Eng]
      • HyperGAN [Kor]
      • Scene Text Telescope: Text-focused Scene Image Super-Resolution [Eng]
      • Scene Text Telescope: Text-focused Scene Image Super-Resolution [Kor]
      • UPFlow [Eng]
      • GFP-GAN [Kor]
      • Federated Contrastive Learning [Kor]
      • Federated Contrastive Learning [Eng]
      • BGNN [Kor]
      • LP-KPN [Kor]
      • Feature Disruptive Attack [Kor]
      • Representative Interpretations [Kor]
      • Representative Interpretations [Eng]
      • Neural Discrete Representation Learning [KOR]
      • Neural Discrete Representation Learning [ENG]
      • Video Frame Interpolation via Adaptive Convolution [Kor]
      • Separation of hand motion and pose [kor]
      • pixelNeRF [Kor]
      • pixelNeRF [Eng]
      • SRResNet and SRGAN [Eng]
      • MZSR [Kor]
      • SANforSISR [Kor]
      • IPT [Kor]
      • Swin Transformer [kor]
      • CNN Cascade for Face Detection [Kor]
      • CapsNet [Kor]
      • Towards Better Generalization: Joint Depth-Pose Learning without PoseNet [Kor]
      • CSRNet [Kor]
      • ScrabbleGAN [Kor]
      • CenterTrack [Kor]
      • CenterTrack [Eng]
      • STSN [Kor]
      • STSN [Eng]
      • VL-BERT:Visual-Linguistic BERT [Kor]
      • VL-BERT:Visual-Linguistic BERT [Eng]
      • Squeeze-and-Attention Networks for Semantic segmentation [Kor]
      • Shot in the dark [Kor]
      • Noise2Self [Kor]
      • Noise2Self [Eng]
      • Dynamic Head [Kor]
      • PSPNet [Kor]
      • PSPNet [Eng]
      • CUT [Kor]
      • CLIP [Eng]
      • Local Implicit Image Function [Kor]
      • Local Implicit Image Function [Eng]
      • MetaAugment [Eng]
      • Show, Attend and Tell [Kor]
      • Transformer [Kor]
      • DETR [Eng]
      • Multimodal Versatile Network [Eng]
      • Multimodal Versatile Network [Kor]
      • BlockDrop [Kor]
      • MDETR [Kor]
      • MDETR [Eng]
      • FSCE [Kor]
      • waveletSR [Kor]
      • DAN-net [Eng]
      • Boosting Monocular Depth Estimation [Eng]
      • Progressively Complementary Network for Fisheye Image Rectification Using Appearance Flow [Kor]
      • Syn2real-generalization [Kor]
      • Syn2real-generalization [Eng]
      • GPS-Net [Kor]
      • Frustratingly Simple Few Shot Object Detection [Eng]
      • DCGAN [Kor]
      • RealSR [Kor]
      • AMP [Kor]
      • AMP [Eng]
      • RCNN [Kor]
      • MobileNet [Eng]
  • Author's note
    • [2022 Spring] Author's note
      • Pop-Out Motion [Kor]
    • [2021 Fall] Author's note
      • Standardized Max Logits [Eng]
      • Standardized Max Logits [Kor]
  • Dive into implementation
    • [2022 Spring] Implementation
      • Supervised Contrastive Replay [Kor]
      • Pose Recognition with Cascade Transformers [Eng]
    • [2021 Fall] Implementation
      • Diversity Input Method [Kor]
        • Source code
      • Diversity Input Method [Eng]
        • Source code
  • Contributors
    • [2022 Fall] Contributors
    • [2021 Fall] Contributors
  • How to contribute?
    • (Template) Paper review [Language]
    • (Template) Author's note [Language]
    • (Template) Implementation [Language]
  • KAIST AI
Powered by GitBook
On this page
  • 1. Problem definition
  • CLEVER-Hans
  • Related work
  • Motivating Example: Color-MNIST
  • Idea
  • 3. Method
  • 4. Experiment & Result
  • Dataset: CLEVER-Hans
  • Experimental setting
  • Result
  • 5. Conclusion
  • Take home message
  • Author / Reviewer information
  • Author
  • Reviewer
  • Reference & Additional materials

Was this helpful?

  1. Paper review
  2. [2021 Fall] Paper review

NeSyXIL [Eng]

Stammer and Schramowski et al. / Right for the Right Concept; Revising Neuro-Symbolic Concepts by Interacting with their Explanations / CVPR 2021

PreviousNeSyXIL [Kor]NextRCAN [Kor]

Last updated 3 years ago

Was this helpful?

한국어로 쓰인 리뷰를 읽으려면 를 누르세요.

1. Problem definition

  • Most explanation methods in deep learning map importance estimates for a model's prediction back to the original input space.

  • These "visual" explanations do not provide semantic explanations of how the model determines.

  • When a model learns wrong decisions, such as CLEVER-Hans, it is difficult to intervene in the model's behavior through semantic explanations.

  • Therefore, there is a need for a methodology that can intervene on the model's behavior via "semantic" feedbacks, not by giving feedbacks using "visual" explanations such as Grad-CAM.

CLEVER-Hans

  • https://simple.wikipedia.org/wiki/Clever_Hans

  • MotivationMachine learning models may show Clever-Hans like moments when solving a task by learning the “wrong” thing, e.g.making use of confounding factors within a data set.

Related work

  • Explainable AI (XAI)

    • In general, explainable artificial intelligence (XAI) approaches represent the description of the model in a way that humans can understand, and are used to verify the reason for the decision of the model (in the form of a black box).

    • There are various XAI approaches, but most of the explanations utilize visual explanations at the input space level.

      • [52, 2]: backpropagation-based approaches

      • [28]: prototype-based approaches

    • Several studies have investigated methods that produce explanations other than these visual explanations. None of the aforementioned studies, however, investigate using the explanations as a means of intervening on the model.

  • Explanatory interactive learning (XIL)

    • XIL = XAI + Active Learning

    • XIL incorporates XAI in the learning process by involving the human user in the training loop, interacting with the explanations.

    • The human user can query the model for explanations of individual predictions and respond by correcting the model if necessary, providing slightly improved feedback on the explanations.

  • Neuro-Symbolic architectures

    • This field of research has received increasing interest in recent years as a means of solving issues of individual subsystems by combining the advantages of symbolic and sub-symbolic models.

    • Neuro-symbolic architecture has the following characteristics.

      • data-driven

      • Sub-symbolic representation

      • Symbolic based reasoning system

Motivating Example: Color-MNIST

  • To motivate the problem setting, authors first revert to a well known confounded toy data set. ColorMNIST consists of colored MNIST digits. Within the training set, each number is confounded with a specific color, whereas in the test set, the color association is shuffled or inverted.

  • A simple CNN model can reach 100% accuracy on the training set, but only 23% on the test set, indicating that the model has learned to largely focus on the color for accurate prediction rather than the digits themselves.

  • Figure 2 depicts the visual explanation (here created using GradCAM) of a zero that is predicted as a nine. From the visual explanation it becomes clear that the model is focusing on the correct object, however why the model is predicting the wrong digit label does not become clear without an understanding of the underlying training data distribution.

  • Importantly, although the model is wrong for the right reason, it is a non-trivial problem of interacting with the model to revise its decision using XIL solely based on these explanations.

Idea

  • As shown in the above Color-MNIST example, it is difficult to interpret and further intervene in model decisions only with visual-level explanations such as Grad-CAM.

  • Therefore, authors propose a Neuro-Symbolic methodology that can fix the model at the semantic level.

  • This method enables feedback in the form of “never focus on the color to make your decision” as it would eliminate the confounding factor.

3. Method

  • Neuro-Symbolic Architecture (The top part of Figure 3)

    • The architecture consisting of two modules, a concept embedding and a reasoning module.

    • In the case of the concept module, it plays a role in mapping into symbolic expressions that humans can understand.

    • Reasoning module makes predictions based on this symbolic representation.

  • Retrieving Neuro-Symbolic Explanations (depicted by the gray arrows of Figure 3.)

    • Given these two modules, we can extract explanations for the separate tasks,

  • Neuro-Symbolic Concepts

    • explanatory loss term

  • Reasoning Module

    • As the output of the concept embedding module represents an unordered set.

    • Therefore, we require our reasoning module to handle such an input structure by using Set Transformer.

    • To generate the explanations of the Set Transformer given the symbolic representation, we make use of the gradient-based Integrated Gradients explanation method.

  • (Slot) Attention is All You Need (for object-based explanations)

(Here, we do not deal with Gradient-based Integrated Gradients explanation, Set Transformer, and Slot Attention in details.)

4. Experiment & Result

Dataset: CLEVER-Hans

  • Dataset

    • CLEVER-Hans3

      • CLEVR-Hans3 contains three classes, of which two are confounded.

      • Images of the first class contain a large cube and large cylinder. The large cube has the color gray in every image of the train and validation set. Within the test set, the color of the large cube is shuffled randomly. Images of the second class contain a small sphere and small metal cube. The small sphere is made of metal in all training and validation set images, however, can be made of either rubber or metal in the test set. Images of the third class contain a large blue sphere and a small yellow sphere in all images of the data set. This class is not confounded.

    • CLEVER-Hans7

      • CLEVR-Hans7 contains seven classes, of which four are confounded.

Experimental setting

  • Dataset

    • CLEVER-Hans3

    • CLEVER-Hans7

  • Baselines

    • CNN (Default): ResNet-based CNN

    • CNN (XIL): ResNet-based CNN + XIL

      • For visual explanations, use GradCAM

    • NeSy (Default): Neuro Symbolic Architecture

    • NeSy (XIL): Neuro Symbolic Architecture + XIL

      • Note that this proposed method utilizes both visual and semantic explanations.

  • Training setup

    • Default: A standard classification setting using the crossentropy loss

    • XIL: Using additional explanatory loss, we intervenes on the model's behavior.

  • Evaluation metric

    • Classification accuracy

Result

Table 2: Experimental results on Clevr-Hans3 and ClevrHans7.

  • Observation 1: CNN produces Clever-Hans moment.

    • Evidence: As Table 2 indicates, the default CNN reaches near perfect classification accuracies in the confounded validation sets but much lower accuracy in the non-confounded test sets.

  • Observation 2: Visual Explanations also produce Clever-Hans moment.

    • Evidence: As Table 2 indicates, the CNN (XIL) also reaches near perfect classification accuracies in the confounded validation sets but much lower accuracy in the non-confounded test sets.

  • Observation 3: Compared to the standard CNN settings, we can observe an increased test accuracy for NeSy model.

    • Evidence: As Table 2 indicates, the test performance of the NeSy model shows better than CNN model.

  • Observation 4: Neuro-Symbolic model is not resilient against confounding factors.

    • Evidence: As Table 2 indicates, even with a higher test accuracy than the CNN model could achieve, this accuracy is still considerably below the again near perfect validation accuracy.

  • Observation 5: Neuro-Symbolic XIL to the Rescue!

    • Evidence: The XIL revised NeuroSymbolic model produces test accuracies much higher than was previously possible in all other settings, including the XIL revised CNN.

    • => The proposed method is effective.

    • => In other words, semantic explanations are effective in moments like CLEVER-Hans.

Figure 5: Qualitative results of different model and training settings on CLEVER-Hans3

  • Figure 5 shows how the explanations change for two different models using default setting (Default) and human intervention (XIL) setting.

  • We can observe that CNN(Default), CNN(XIL), and NeSY(Default) all do not predict correctly, and only NeSY(XIL) predicts correctly.

  • In the case of CNN models, it is difficult to semantically interpret model's decisions.

  • In the case of the NeSY models, we can semantically interpret model's decisions.

5. Conclusion

  • The authors propose a methodology for intervening in the neuro-symbolic scene representation space, which allows modification of the model at the semantic level.

  • To prove the effectiveness, they constructed new confounded visual scene datasets, CLEVER-Hans3 and CLEVER-Hans7.

  • The results of experiments on CLEVER-Hans demonstrate that "semantic" explanations can identify confounders that are not identifiable using "visual" expalantions only.

  • As a result, feedback on this semantic level makes it possible to revise the model from focusing on these factors.

Take home message

Through the motivated example of Color-MNIST, this paper shows why it is important to use symbolic expressions when intervening in a model.

Authors first proposed the Neural Symbolic Architecture to XIL, showing through experimental results that it is more effective to intervene in the model than visual descriptions such as Grad-CAM.

If we can make good use of symbolic expressions as well as non-symbolic expressions in deep learning, it will be of great help in utilizing human's cognitive ability.

Author / Reviewer information

Author

Seongsu Bae

  • KAIST AI

  • contact: seongsu@kaist.ac.kr

Reviewer

  1. Korean name (English name): Affiliation / Contact information

  2. Korean name (English name): Affiliation / Contact information

  3. ...

Reference & Additional materials

Given an input image xi∈Xx_i \in Xxi​∈X, concept module encodes it into a symbolic representation, h(xi)=z^ih(x_i) = \hat{z}_ih(xi​)=z^i​.

z^i∈[0,1]N×D\hat{z}_i \in [0,1]^{N\times D}z^i​∈[0,1]N×D : the symoblic representation

Given an input z^i\hat{z}_iz^i​, reasoning module produces the prediction output, g(z^i)=y^ig(\hat{z}_i)=\hat{y}_ig(z^i​)=y^​i​.

y^i∈[0,1]N×Nc\hat{y}_i \in [0,1]^{N\times N_c}y^​i​∈[0,1]N×Nc​ : the prediction output given the symbolic representation

X:=[x1,…,xN]∈RN×MX := [x_1, \dots, x_N] \in \mathbb{R}^{N\times M}X:=[x1​,…,xN​]∈RN×M, with XXX being divided into subsets of NcN_cNc​ classes.

We write an explanation function in a general notation as E(m(⋅),o,s)E(m(\cdot), o, s)E(m(⋅),o,s), which retrieves the explanation of a specific module, m(⋅)m(\cdot)m(⋅), given the module’s input sss, and the module’s output if it is the final module or the explanation of the following module if it is not, both summarized as ooo here.

For reasoning module, we can consider its explanations as Eg(g(⋅),y^i,zi)=:e^igE^g(g(\cdot), \hat{y}_i, z_i) = : \hat{e}^g_iEg(g(⋅),y^​i​,zi​)=:e^ig​.

e^ig\hat{e}^g_ie^ig​ represents the explanation of the reasoning module given the final predicted output y^i\hat{y}_iy^​i​.

For concept module, we can consider its explanations as Eh(h(⋅),e^ig,xi)=:e^ihE^h(h(\cdot), \hat{e}^g_i, x_i) = : \hat{e}^h_iEh(h(⋅),e^ig​,xi​)=:e^ih​.

e^ih\hat{e}^h_ie^ih​ presents the explanation of the concept module given the explanation of the reasoning module e^ig\hat{e}^g_ie^ig​

Lexpl=λ∑i=1Nr(Aiv,e^ih)+(1−λ)∑i=1Nr(Ais,e^ig)L_{expl} = \lambda \sum_{i=1}^N r(A_i^v, \hat{e}^h_i) + (1-\lambda) \sum_{i=1}^N r(A_i^s, \hat{e}^g_i)Lexpl​=λ∑i=1N​r(Aiv​,e^ih​)+(1−λ)∑i=1N​r(Ais​,e^ig​)

r(⋅,⋅)r(\cdot, \cdot)r(⋅,⋅): regularization function (e.g. RRR, HINT)

AivA^v_iAiv​: visual feedback, in the form of binary image masks

AisA^s_iAis​: semantic feedback, in the form of binary mask for symbolic space

By using the Slot Attention module, we can botain the attention maps BiB_iBi​ for a sample xix_ixi​.

By using BiB_iBi​ and e^ig\hat{e}^g_ie^ig​, we can represent the exaplanation Eh(h(⋅),e^ig,xi)E^h(h(\cdot), \hat{e}^g_i, x_i)Eh(h(⋅),e^ig​,xi​).

Korea version
Right for the Right Concept: Revising Neuro-Symbolic Concepts by Interacting with their Explanations