Search Results for author: Mark Ibrahim

Found 24 papers, 11 papers with code

Uncertainty-Based Abstention in LLMs Improves Safety and Reduces Hallucinations

no code implementations16 Apr 2024 Christian Tomani, Kamalika Chaudhuri, Ivan Evtimov, Daniel Cremers, Mark Ibrahim

A major barrier towards the practical deployment of large language models (LLMs) is their lack of reliability.

Question Answering

The Bias of Harmful Label Associations in Vision-Language Models

no code implementations11 Feb 2024 Caner Hazirbas, Alicia Sun, Yonathan Efroni, Mark Ibrahim

Despite the remarkable performance of foundation vision-language models, the shared representation space for text and vision can also encode harmful label associations detrimental to fairness.

Fairness

WorldSense: A Synthetic Benchmark for Grounded Reasoning in Large Language Models

1 code implementation27 Nov 2023 Youssef Benchekroun, Megi Dervishi, Mark Ibrahim, Jean-Baptiste Gaya, Xavier Martinet, Grégoire Mialon, Thomas Scialom, Emmanuel Dupoux, Dieuwke Hupkes, Pascal Vincent

We propose WorldSense, a benchmark designed to assess the extent to which LLMs are consistently able to sustain tacit world models, by testing how they draw simple inferences from descriptions of simple arrangements of entities.

In-Context Learning

Battle of the Backbones: A Large-Scale Comparison of Pretrained Models across Computer Vision Tasks

2 code implementations NeurIPS 2023 Micah Goldblum, Hossein Souri, Renkun Ni, Manli Shu, Viraj Prabhu, Gowthami Somepalli, Prithvijit Chattopadhyay, Mark Ibrahim, Adrien Bardes, Judy Hoffman, Rama Chellappa, Andrew Gordon Wilson, Tom Goldstein

Battle of the Backbones (BoB) makes this choice easier by benchmarking a diverse suite of pretrained models, including vision-language models, those trained via self-supervised learning, and the Stable Diffusion backbone, across a diverse set of computer vision tasks ranging from classification to object detection to OOD generalization and more.

Benchmarking object-detection +2

VPA: Fully Test-Time Visual Prompt Adaptation

no code implementations26 Sep 2023 Jiachen Sun, Mark Ibrahim, Melissa Hall, Ivan Evtimov, Z. Morley Mao, Cristian Canton Ferrer, Caner Hazirbas

Inspired by the success of textual prompting, several studies have investigated the efficacy of visual prompt tuning.

Pseudo Label Test-time Adaptation +3

PUG: Photorealistic and Semantically Controllable Synthetic Data for Representation Learning

1 code implementation NeurIPS 2023 Florian Bordes, Shashank Shekhar, Mark Ibrahim, Diane Bouchacourt, Pascal Vincent, Ari S. Morcos

Synthetic image datasets offer unmatched advantages for designing and evaluating deep neural networks: they make it possible to (i) render as many data samples as needed, (ii) precisely control each scene and yield granular ground truth labels (and captions), (iii) precisely control distribution shifts between training and testing to isolate variables of interest for sound experimentation.

Representation Learning

Does Progress On Object Recognition Benchmarks Improve Real-World Generalization?

no code implementations24 Jul 2023 Megan Richards, Polina Kirichenko, Diane Bouchacourt, Mark Ibrahim

Second, we study model generalization across geographies by measuring the disparities in performance across regions, a more fine-grained measure of real world generalization.

Object Recognition

Pinpointing Why Object Recognition Performance Degrades Across Income Levels and Geographies

1 code implementation11 Apr 2023 Laura Gustafson, Megan Richards, Melissa Hall, Caner Hazirbas, Diane Bouchacourt, Mark Ibrahim

As an example, we show that mitigating a model's vulnerability to texture can improve performance on the lower income level.

Object Recognition

A Whac-A-Mole Dilemma: Shortcuts Come in Multiples Where Mitigating One Amplifies Others

1 code implementation CVPR 2023 Zhiheng Li, Ivan Evtimov, Albert Gordo, Caner Hazirbas, Tal Hassner, Cristian Canton Ferrer, Chenliang Xu, Mark Ibrahim

Key to advancing the reliability of vision systems is understanding whether existing methods can overcome multiple shortcuts or struggle in a Whac-A-Mole game, i. e., where mitigating one shortcut amplifies reliance on others.

Domain Generalization Image Classification +1

ImageNet-X: Understanding Model Mistakes with Factor of Variation Annotations

no code implementations3 Nov 2022 Badr Youbi Idrissi, Diane Bouchacourt, Randall Balestriero, Ivan Evtimov, Caner Hazirbas, Nicolas Ballas, Pascal Vincent, Michal Drozdzal, David Lopez-Paz, Mark Ibrahim

Equipped with ImageNet-X, we investigate 2, 200 current recognition models and study the types of mistakes as a function of model's (1) architecture, e. g. transformer vs. convolutional, (2) learning paradigm, e. g. supervised vs. self-supervised, and (3) training procedures, e. g., data augmentation.

Data Augmentation

Robust Self-Supervised Learning with Lie Groups

no code implementations24 Oct 2022 Mark Ibrahim, Diane Bouchacourt, Ari Morcos

Our approach applies the formalism of Lie groups to capture continuous transformations to improve models' robustness to distributional shifts.

Self-Supervised Learning

The Robustness Limits of SoTA Vision Models to Natural Variation

no code implementations24 Oct 2022 Mark Ibrahim, Quentin Garrido, Ari Morcos, Diane Bouchacourt

We study not only how robust recent state-of-the-art models are, but also the extent to which models can generalize variation in factors when they're present during training.

Disentanglement of Correlated Factors via Hausdorff Factorized Support

1 code implementation13 Oct 2022 Karsten Roth, Mark Ibrahim, Zeynep Akata, Pascal Vincent, Diane Bouchacourt

We show that the use of HFS consistently facilitates disentanglement and recovery of ground-truth factors across a variety of correlation settings and benchmarks, even under severe training correlations and correlation shifts, with in parts over $+60\%$ in relative improvement over existing disentanglement methods.

Disentanglement

CrypTen: Secure Multi-Party Computation Meets Machine Learning

1 code implementation NeurIPS 2021 Brian Knott, Shobha Venkataraman, Awni Hannun, Shubho Sengupta, Mark Ibrahim, Laurens van der Maaten

To foster adoption of secure MPC in machine learning, we present CrypTen: a software framework that exposes popular secure MPC primitives via abstractions that are common in modern machine-learning frameworks, such as tensor computations, automatic differentiation, and modular neural networks.

BIG-bench Machine Learning Image Classification +4

Grounding inductive biases in natural images:invariance stems from variations in data

1 code implementation NeurIPS 2021 Diane Bouchacourt, Mark Ibrahim, Ari S. Morcos

While prior work has focused on synthetic data, we attempt here to characterize the factors of variation in a real dataset, ImageNet, and study the invariance of both standard residual networks and the recently proposed vision transformer with respect to changes in these factors.

Data Augmentation Translation

Grounding inductive biases in natural images: invariance stems from variations in data

1 code implementation NeurIPS 2021 Diane Bouchacourt, Mark Ibrahim, Ari S. Morcos

While prior work has focused on synthetic data, we attempt here to characterize the factors of variation in a real dataset, ImageNet, and study the invariance of both standard residual networks and the recently proposed vision transformer with respect to changes in these factors.

Data Augmentation Translation

Addressing the Topological Defects of Disentanglement via Distributed Operators

1 code implementation10 Feb 2021 Diane Bouchacourt, Mark Ibrahim, Stéphane Deny

A core challenge in Machine Learning is to learn to disentangle natural factors of variation in data (e. g. object shape vs. pose).

Disentanglement

Addressing the Topological Defects of Disentanglement

no code implementations1 Jan 2021 Diane Bouchacourt, Mark Ibrahim, Stephane Deny

A core challenge in Machine Learning is to disentangle natural factors of variation in data (e. g. object shape vs pose).

Disentanglement

Global Explanations of Neural Networks: Mapping the Landscape of Predictions

1 code implementation6 Feb 2019 Mark Ibrahim, Melissa Louie, Ceena Modarres, John Paisley

A barrier to the wider adoption of neural networks is their lack of interpretability.

Mixed Membership Recurrent Neural Networks

no code implementations23 Dec 2018 Ghazal Fazelnia, Mark Ibrahim, Ceena Modarres, Kevin Wu, John Paisley

Models for sequential data such as the recurrent neural network (RNN) often implicitly model a sequence as having a fixed time interval between observations and do not account for group-level effects when multiple sequences are observed.

Dynamic Topic Modeling

Towards Explainable Deep Learning for Credit Lending: A Case Study

no code implementations15 Nov 2018 Ceena Modarres, Mark Ibrahim, Melissa Louie, John Paisley

Deep learning adoption in the financial services industry has been limited due to a lack of model interpretability.

Cannot find the paper you are looking for? You can Submit a new open access paper.