Network Interpretation
9 papers with code • 0 benchmarks • 0 datasets
Benchmarks
These leaderboards are used to track progress in Network Interpretation
Most implemented papers
Fooling Neural Network Interpretations via Adversarial Model Manipulation
We ask whether the neural network interpretation methods can be fooled via adversarial model manipulation, which is defined as a model fine-tuning step that aims to radically alter the explanations without hurting the accuracy of the original models, e. g., VGG19, ResNet50, and DenseNet121.
Plug-and-Play VQA: Zero-shot VQA by Conjoining Large Pretrained Models with Zero Training
Visual question answering (VQA) is a hallmark of vision and language reasoning and a challenging task under the zero-shot setting.
An Empirical Study on the Relation between Network Interpretability and Adversarial Robustness
We demonstrate that training the networks to have interpretable gradients improves their robustness to adversarial perturbations.
Proper Network Interpretability Helps Adversarial Robustness in Classification
Recent works have empirically shown that there exist adversarial examples that can be hidden from neural network interpretability (namely, making network interpretation maps visually similar), or interpretability is itself susceptible to adversarial attacks.
Attribution Preservation in Network Compression for Reliable Network Interpretation
Neural networks embedded in safety-sensitive applications such as self-driving cars and wearable health monitors rely on two important techniques: input attribution for hindsight analysis and network compression to reduce its size for edge-computing.
DeepEverest: Accelerating Declarative Top-K Queries for Deep Neural Network Interpretation
We design, implement, and evaluate DeepEverest, a system for the efficient execution of interpretation by example queries over the activation values of a deep neural network.
Reflash Dropout in Image Super-Resolution
Dropout is designed to relieve the overfitting problem in high-level vision tasks but is rarely applied in low-level vision tasks, like image super-resolution (SR).
Towards More Robust Interpretation via Local Gradient Alignment
However, the lack of considering the normalization of the attributions, which is essential in their visualizations, has been an obstacle to understanding and improving the robustness of feature attribution methods.
Perturbation on Feature Coalition: Towards Interpretable Deep Neural Networks
The inherent "black box" nature of deep neural networks (DNNs) compromises their transparency and reliability.