Interpretable Explanations of Black Boxes by Meaningful Perturbation - CAP6412 Spring 2021
University of Central Florida via YouTube
Overview
Syllabus
Intro
Content
Abstract Image Saliency Methods Summary Attention Map Limited by heuristic properties and architectural constraints
Introduction Current Problems The interpretation for the black box predictor The intuitive visualization method is only heuristic, and the meaning remains unclear.
Contribution Develop principles and methods to explain any black box function By determine mapping attributes - Internal mechanisms is used to implement these attributes
Related Work Gradient-based method -Backpropagates the gradient for a class label to the image layer Other methods: DeConvNet, Guided Backprop
Related Work - CAM
Related Work Comparison
Comparison with other saliency methods
Principle Black bax is a mapping function
Explanations as meta-predictors Rules are used to explain a robin classifier
Advantages of Explanations as Meta-predictors The faithfulness of images can be measured as prediction accuracy To find the explanations automatically
Local Explanations
Saliency Deleting parts of image x, as the perturbations for the whole image X
A Meaningful Image Perturbation 11
Deletion and Preservation
Artifacts Reduction
Experiment-Interpretability
Experiment Testing hypotheses: animal part saliency
Experiment-Adversarial defense
Experiment localization and pointing
Conclusion
Questions?
Taught by
UCF CRCV