Search Results for author: Xavier Boix

Found 31 papers, 12 papers with code

Multi-domain improves out-of-distribution and data-limited scenarios for medical image analysis

no code implementations10 Oct 2023 Ece Ozkan, Xavier Boix

We refer to this approach as multi-domain model and compare its performance to that of specialized models.

D3: Data Diversity Design for Systematic Generalization in Visual Question Answering

1 code implementation15 Sep 2023 Amir Rahimi, Vanessa D'Amario, Moyuru Yamada, Kentaro Takemoto, Tomotake Sasaki, Xavier Boix

We demonstrate that this result is independent of the similarity between the training and testing data and applies to well-known families of neural network architectures for VQA (i. e. monolithic architectures and neural module networks).

Question Answering Systematic Generalization +1

Modularity Trumps Invariance for Compositional Robustness

1 code implementation15 Jun 2023 Ian Mason, Anirban Sarkar, Tomotake Sasaki, Xavier Boix

In this work we develop a compositional image classification task where, given a few elemental corruptions, models are asked to generalize to compositions of these corruptions.

Domain Generalization Image Classification

Transformer Module Networks for Systematic Generalization in Visual Question Answering

1 code implementation27 Jan 2022 Moyuru Yamada, Vanessa D'Amario, Kentaro Takemoto, Xavier Boix, Tomotake Sasaki

We reveal that Neural Module Networks (NMNs), i. e., question-specific compositions of modules that tackle a sub-task, achieve better or similar systematic generalization performance than the conventional Transformers, even though NMNs' modules are CNN-based.

Question Answering Systematic Generalization +1

Robust Upper Bounds for Adversarial Training

1 code implementation17 Dec 2021 Dimitris Bertsimas, Xavier Boix, Kimberly Villalobos Carballo, Dick den Hertog

We introduce a new approach to adversarial training by minimizing an upper bound of the adversarial loss that is based on a holistic expansion of the network instead of separate bounds for each layer.

Symmetry Perception by Deep Networks: Inadequacy of Feed-Forward Architectures and Improvements with Recurrent Connections

1 code implementation8 Dec 2021 Shobhita Sundaram, Darius Sinha, Matthew Groth, Tomotake Sasaki, Xavier Boix

Symmetry is omnipresent in nature and perceived by the visual system of many species, as it facilitates detecting ecologically important classes of objects in our environment.

Three approaches to facilitate DNN generalization to objects in out-of-distribution orientations and illuminations

1 code implementation30 Oct 2021 Akira Sakai, Taro Sunagawa, Spandan Madan, Kanata Suzuki, Takashi Katoh, Hiromichi Kobashi, Hanspeter Pfister, Pawan Sinha, Xavier Boix, Tomotake Sasaki

While humans have a remarkable capability of recognizing objects in out-of-distribution (OoD) orientations and illuminations, Deep Neural Networks (DNNs) severely suffer in this case, even when large amounts of training examples are available.

Emergent Neural Network Mechanisms for Generalization to Objects in Novel Orientations

no code implementations28 Sep 2021 Avi Cooper, Xavier Boix, Daniel Harari, Spandan Madan, Hanspeter Pfister, Tomotake Sasaki, Pawan Sinha

The capability of Deep Neural Networks (DNNs) to recognize objects in orientations outside the distribution of the training data is not well understood.

The Foes of Neural Network's Data Efficiency Among Unnecessary Input Dimensions

no code implementations13 Jul 2021 Vanessa D'Amario, Sanjana Srivastava, Tomotake Sasaki, Xavier Boix

Datasets often contain input dimensions that are unnecessary to predict the output label, e. g. background in object recognition, which lead to more trainable parameters.

Object Recognition

Adversarial examples within the training distribution: A widespread challenge

1 code implementation30 Jun 2021 Spandan Madan, Tomotake Sasaki, Hanspeter Pfister, Tzu-Mao Li, Xavier Boix

This result provides evidence supporting theories attributing adversarial examples to the proximity of data to ground-truth class boundaries, and calls into question other theories which do not account for this more stringent definition of adversarial attacks.

Object Recognition Open-Ended Question Answering

On the Capability of CNNs to Generalize to Unseen Category-Viewpoint Combinations

no code implementations1 Jan 2021 Spandan Madan, Timothy Henry, Jamell Arthur Dozier, Helen Ho, Nishchal Bhandari, Tomotake Sasaki, Fredo Durand, Hanspeter Pfister, Xavier Boix

We find that learning category and viewpoint in separate networks compared to a shared one leads to an increase in selectivity and invariance, as separate networks are not forced to preserve information about both category and viewpoint.

Object Recognition Viewpoint Estimation

The Foes of Neural Network’s Data Efficiency Among Unnecessary Input Dimensions

no code implementations1 Jan 2021 Vanessa D'Amario, Sanjana Srivastava, Tomotake Sasaki, Xavier Boix

In this paper, we investigate the impact of unnecessary input dimensions on one of the central issues of machine learning: the number of training examples needed to achieve high generalization performance, which we refer to as the network's data efficiency.

Foveation Image Classification +3

When and how CNNs generalize to out-of-distribution category-viewpoint combinations

2 code implementations15 Jul 2020 Spandan Madan, Timothy Henry, Jamell Dozier, Helen Ho, Nishchal Bhandari, Tomotake Sasaki, Frédo Durand, Hanspeter Pfister, Xavier Boix

In this paper, we investigate when and how such OOD generalization may be possible by evaluating CNNs trained to classify both object category and 3D viewpoint on OOD combinations, and identifying the neural mechanisms that facilitate such OOD generalization.

Object Recognition Viewpoint Estimation

Robustness to Transformations Across Categories: Is Robustness To Transformations Driven by Invariant Neural Representations?

no code implementations30 Jun 2020 Hojin Jang, Syed Suleman Abbas Zaidi, Xavier Boix, Neeraj Prasad, Sharon Gilad-Gutnick, Shlomit Ben-Ami, Pawan Sinha

Our results with state-of-the-art DCNNs indicate that invariant neural representations do not always drive robustness to transformations, as networks show robustness for categories seen transformed during training even in the absence of invariant neural representations.

Frivolous Units: Wider Networks Are Not Really That Wide

1 code implementation10 Dec 2019 Stephen Casper, Xavier Boix, Vanessa D'Amario, Ling Guo, Martin Schrimpf, Kasper Vinken, Gabriel Kreiman

We identify two distinct types of "frivolous" units that proliferate when the network's width is increased: prunable units which can be dropped out of the network without significant change to the output and redundant units whose activities can be expressed as a linear combination of others.

Do Deep Neural Networks for Segmentation Understand Insideness?

no code implementations25 Sep 2019 Kimberly M Villalobos, Vilim Stih, Amineh Ahmadinejad, Jamell Dozier, Andrew Francl, Frederico Azevedo, Tomotake Sasaki, Xavier Boix

At the heart of image segmentation lies the problem of determining whether a pixel is inside or outside a region, which we denote as the "insideness" problem.

Image Segmentation Segmentation +1

Minimal Images in Deep Neural Networks: Fragile Object Recognition in Natural Images

no code implementations ICLR 2019 Sanjana Srivastava, Guy Ben-Yosef, Xavier Boix

Ullman et al. 2016 show that a slight modification of the location and size of the visible region of the minimal image produces a sharp drop in human recognition accuracy.

Object Recognition

Theory IIIb: Generalization in Deep Networks

no code implementations29 Jun 2018 Tomaso Poggio, Qianli Liao, Brando Miranda, Andrzej Banburski, Xavier Boix, Jack Hidary

Here we prove a similar result for nonlinear multilayer DNNs near zero minima of the empirical loss.

Binary Classification

Theory of Deep Learning III: explaining the non-overfitting puzzle

no code implementations30 Dec 2017 Tomaso Poggio, Kenji Kawaguchi, Qianli Liao, Brando Miranda, Lorenzo Rosasco, Xavier Boix, Jack Hidary, Hrushikesh Mhaskar

In this note, we show that the dynamics associated to gradient descent minimization of nonlinear networks is topologically equivalent, near the asymptotically stable minima of the empirical error, to linear gradient system in a quadratic potential with a degenerate (for square loss) or almost degenerate (for logistic or crossentropy loss) Hessian.

General Classification

Herding Generalizes Diverse M -Best Solutions

no code implementations14 Nov 2016 Ece Ozkan, Gemma Roig, Orcun Goksel, Xavier Boix

We show that the algorithm to extract diverse M -solutions from a Conditional Random Field (called divMbest [1]) takes exactly the form of a Herding procedure [2], i. e. a deterministic dynamical system that produces a sequence of hypotheses that respect a set of observed moment constraints.

Semantic Segmentation

Predicting When Saliency Maps Are Accurate and Eye Fixations Consistent

no code implementations CVPR 2016 Anna Volokitin, Michael Gygli, Xavier Boix

Many computational models of visual attention use image features and machine learning techniques to predict eye fixation locations as saliency maps.

Object Object Recognition

Foveation-based Mechanisms Alleviate Adversarial Examples

no code implementations19 Nov 2015 Yan Luo, Xavier Boix, Gemma Roig, Tomaso Poggio, Qi Zhao

To see this, first, we report results in ImageNet that lead to a revision of the hypothesis that adversarial perturbations are a consequence of CNNs acting as a linear classifier: CNNs act locally linearly to changes in the image regions with objects recognized by the CNN, and in other regions the CNN may act non-linearly.

Foveation Translation

Self-Adaptable Templates for Feature Coding

no code implementations NeurIPS 2014 Xavier Boix, Gemma Roig, Salomon Diether, Luc V. Gool

Within this processing pipeline, the common trend is to learn the feature coding templates, often referred as codebook entries, filters, or over-complete basis.

Image Classification Object Recognition +1

Comment on "Ensemble Projection for Semi-supervised Image Classification"

no code implementations29 Aug 2014 Xavier Boix, Gemma Roig, Luc van Gool

In a series of papers by Dai and colleagues [1, 2], a feature map (or kernel) was introduced for semi- and unsupervised learning.

Classification General Classification +1

SEEDS: Superpixels Extracted via Energy-Driven Sampling

1 code implementation16 Sep 2013 Michael Van den Bergh, Xavier Boix, Gemma Roig, Luc van Gool

We define a robust and fast to evaluate energy function, based on enforcing color similarity between the bound- aries and the superpixel color histogram.

Superpixels

Random Binary Mappings for Kernel Learning and Efficient SVM

no code implementations19 Jul 2013 Gemma Roig, Xavier Boix, Luc van Gool

SVMs suffer from various drawbacks in terms of selecting the right kernel, which depends on the image descriptors, as well as computational and memory efficiency.

Attribute Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.