Search Results for author: Thomas Fel

Found 19 papers, 10 papers with code

Feature Accentuation: Revealing 'What' Features Respond to in Natural Images

no code implementations15 Feb 2024 Chris Hamblin, Thomas Fel, Srijani Saha, Talia Konkle, George Alvarez

Most research has primarily centered around attribution methods, which provide explanations in the form of heatmaps, showing where the model directs its attention for a given feature.

On the Foundations of Shortcut Learning

no code implementations24 Oct 2023 Katherine L. Hermann, Hossein Mobahi, Thomas Fel, Michael C. Mozer

Deep-learning models can extract a rich assortment of features from data.

Saliency strikes back: How filtering out high frequencies improves white-box explanations

no code implementations18 Jul 2023 Sabine Muzellec, Thomas Fel, Victor Boutin, Léo Andéol, Rufin VanRullen, Thomas Serre

Attribution methods correspond to a class of explainability methods (XAI) that aim to assess how individual inputs contribute to a model's decision-making process.

Computational Efficiency Decision Making

Unlocking Feature Visualization for Deeper Networks with MAgnitude Constrained Optimization

1 code implementation11 Jun 2023 Thomas Fel, Thibaut Boissin, Victor Boutin, Agustin Picard, Paul Novello, Julien Colin, Drew Linsley, Tom Rousseau, Rémi Cadène, Laurent Gardes, Thomas Serre

However, its widespread adoption has been limited due to a reliance on tricks to generate interpretable images, and corresponding challenges in scaling it to deeper neural networks.

Adversarial alignment: Breaking the trade-off between the strength of an attack and its relevance to human perception

no code implementations5 Jun 2023 Drew Linsley, Pinyuan Feng, Thibaut Boissin, Alekh Karkada Ashok, Thomas Fel, Stephanie Olaiya, Thomas Serre

Harmonized DNNs achieve the best of both worlds and experience attacks that are detectable and affect features that humans find diagnostic for recognition, meaning that attacks on these models are more likely to be rendered ineffective by inducing similar effects on human perception.

Adversarial Attack Adversarial Robustness +2

COCKATIEL: COntinuous Concept ranKed ATtribution with Interpretable ELements for explaining neural net classifiers on NLP tasks

1 code implementation11 May 2023 Fanny Jourdan, Agustin Picard, Thomas Fel, Laurent Risser, Jean Michel Loubes, Nicholas Asher

COCKATIEL is a novel, post-hoc, concept-based, model-agnostic XAI technique that generates meaningful explanations from the last layer of a neural net model trained on an NLP classification task by using Non-Negative Matrix Factorization (NMF) to discover the concepts the model leverages to make predictions and by exploiting a Sensitivity Analysis to estimate accurately the importance of each of these concepts for the model.

Explainable Artificial Intelligence (XAI) Sentiment Analysis

Confident Object Detection via Conformal Prediction and Conformal Risk Control: an Application to Railway Signaling

no code implementations12 Apr 2023 Léo Andéol, Thomas Fel, Florence De Grancey, Luca Mossina

Deploying deep learning models in real-world certified systems requires the ability to provide confidence estimates that accurately reflect their uncertainty.

Conformal Prediction object-detection +1

Diffusion Models as Artists: Are we Closing the Gap between Humans and Machines?

1 code implementation27 Jan 2023 Victor Boutin, Thomas Fel, Lakshya Singhal, Rishav Mukherji, Akash Nagaraj, Julien Colin, Thomas Serre

An important milestone for AI is the development of algorithms that can produce drawings that are indistinguishable from those of humans.

Conformal Prediction for Trustworthy Detection of Railway Signals

no code implementations26 Jan 2023 Léo Andéol, Thomas Fel, Florence De Grancey, Luca Mossina

We present an application of conformal prediction, a form of uncertainty quantification with guarantees, to the detection of railway signals.

Conformal Prediction Uncertainty Quantification

CRAFT: Concept Recursive Activation FacTorization for Explainability

1 code implementation CVPR 2023 Thomas Fel, Agustin Picard, Louis Bethune, Thibaut Boissin, David Vigouroux, Julien Colin, Rémi Cadène, Thomas Serre

However, recent research has exposed the limited practical value of these methods, attributed in part to their narrow focus on the most prominent regions of an image -- revealing "where" the model looks, but failing to elucidate "what" the model sees in those areas.

Harmonizing the object recognition strategies of deep neural networks with humans

3 code implementations8 Nov 2022 Thomas Fel, Ivan Felipe, Drew Linsley, Thomas Serre

Across 84 different DNNs trained on ImageNet and three independent datasets measuring the where and the how of human visual strategies for object recognition on those images, we find a systematic trade-off between DNN categorization accuracy and alignment with human visual strategies for object recognition.

Object Object Recognition

On the explainable properties of 1-Lipschitz Neural Networks: An Optimal Transport Perspective

no code implementations NeurIPS 2023 Mathieu Serrurier, Franck Mamalet, Thomas Fel, Louis Béthune, Thibaut Boissin

Input gradients have a pivotal role in a variety of applications, including adversarial attack algorithms for evaluating model robustness, explainable AI techniques for generating Saliency Maps, and counterfactual explanations. However, Saliency Maps generated by traditional neural networks are often noisy and provide limited insights.

Adversarial Attack counterfactual +1

Making Sense of Dependence: Efficient Black-box Explanations Using Dependence Measure

1 code implementation13 Jun 2022 Paul Novello, Thomas Fel, David Vigouroux

HSIC measures the dependence between regions of an input image and the output of a model based on kernel embeddings of distributions.

object-detection Object Detection

What I Cannot Predict, I Do Not Understand: A Human-Centered Evaluation Framework for Explainability Methods

1 code implementation6 Dec 2021 Julien Colin, Thomas Fel, Remi Cadene, Thomas Serre

A multitude of explainability methods and associated fidelity performance metrics have been proposed to help better understand how modern AI systems make decisions.

How Good is your Explanation? Algorithmic Stability Measures to Assess the Quality of Explanations for Deep Neural Networks

no code implementations7 Sep 2020 Thomas Fel, David Vigouroux, Rémi Cadène, Thomas Serre

A plethora of methods have been proposed to explain how deep neural networks reach their decisions but comparatively, little effort has been made to ensure that the explanations produced by these methods are objectively relevant.

Image Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.