Hide metadata

dc.contributor.authorHeggen, Mona
dc.date.accessioned2021-09-07T22:45:42Z
dc.date.available2021-09-07T22:45:42Z
dc.date.issued2021
dc.identifier.citationHeggen, Mona. An investigation of different interpretability methods used to evaluate a prediction from a CNN model. Master thesis, University of Oslo, 2021
dc.identifier.urihttp://hdl.handle.net/10852/87832
dc.description.abstractIn this thesis we investigate different interpretability methods for evaluating predictions from Convolutional Neural Networks. We look at research on several explanation methods with a focus on Local Interpretable Model-agnostic Explanations (LIME) and Layer-wise Relevance Propagation (LRP). Our goal is to investigate different interpretability methods and how robust they are in comparison to each other. We do initial experiments by testing a set of images with Guided Backpropagation, Gradient-weighted Class Activation Mapping (Grad-CAM), LIME and LRP. In the next set of experiments we focus on LRP and LIME. The models we use are VGG16 with and without batchnorm layers. We use rotation and Gaussian noise to transform the input images. To measure the robustness we use Root Mean Square Error (RMSE). The transformation is added to the input and sent through the model. The output from the model is sent through the interpretability method. The resulting heatmap for the transformed image is then compared with the original heatmap to measure the RMSE score. We use a set of small transformations and a set of more extreme transformations. The transformations we use for rotation are between 0.5-10 degrees and 15-40 degrees. For the Gaussian noise we use $\sigma$ between 0.01-0.10 and 0.25-10.0. We observe that LIME focuses on super pixels and will therefore be less robust for transformations compared to LRP. We find that methods which emphasises on both positive and negative contributions, such as LRP and Grad-CAM are more helpful since they highlight the regions that contribute and work against the prediction in the image. When implementing LRP with models using batchnorm layers we find that this give unreliable results. We handle this by merging the batchnorm layers with the corresponding convolutional layer before backpropagating LRP. Our experiments show that the explanation from the interpretability method correlates significantly with the models robustness. Though in some cases the robustness of the model is not reflected in the interpretability method and this is especially noticeable when Gaussian noise are applied to the input in the LIME experiments.eng
dc.language.isoeng
dc.subjectAttribution Methods
dc.subjectXAI
dc.subjectCNN
dc.subjectInterpretability
dc.subjectLIME
dc.subjectLRP
dc.subjectMachine Learning
dc.subjectDeep Learning
dc.titleAn investigation of different interpretability methods used to evaluate a prediction from a CNN modeleng
dc.typeMaster thesis
dc.date.updated2021-09-07T22:45:42Z
dc.creator.authorHeggen, Mona
dc.identifier.urnURN:NBN:no-90473
dc.type.documentMasteroppgave
dc.identifier.fulltextFulltext https://www.duo.uio.no/bitstream/handle/10852/87832/1/MasterThesis04.pdf


Files in this item

Appears in the following Collection

Hide metadata