no code implementations • 17 Oct 2024 • Mazda Moayeri, Vidhisha Balachandran, Varun Chandrasekaran, Safoora Yousefi, Thomas Fel, Soheil Feizi, Besmira Nushi, Neel Joshi, Vibhav Vineet
With models getting stronger, evaluations have grown more complex, testing multiple skills in one benchmark and even in the same instance at once.
1 code implementation • 12 Jun 2024 • Arman Zarei, Keivan Rezaei, Samyadeep Basu, Mehrdad Saberi, Mazda Moayeri, Priyatham Kattakinda, Soheil Feizi
We also show that re-weighting the erroneous attention contributions in CLIP can also lead to improved compositional performances, however these improvements are often less significant than those achieved by solely learning a linear projection head, highlighting erroneous attentions to be only a minor error source.
no code implementations • 27 May 2024 • Florian Bordes, Richard Yuanzhe Pang, Anurag Ajay, Alexander C. Li, Adrien Bardes, Suzanne Petryk, Oscar Mañas, Zhiqiu Lin, Anas Mahmoud, Bargav Jayaraman, Mark Ibrahim, Melissa Hall, Yunyang Xiong, Jonathan Lebensold, Candace Ross, Srihari Jayakumar, Chuan Guo, Diane Bouchacourt, Haider Al-Tahan, Karthik Padthe, Vasu Sharma, Hu Xu, Xiaoqing Ellen Tan, Megan Richards, Samuel Lavoie, Pietro Astolfi, Reyhane Askari Hemmat, Jun Chen, Kushal Tirumala, Rim Assouel, Mazda Moayeri, Arjang Talattof, Kamalika Chaudhuri, Zechun Liu, Xilun Chen, Quentin Garrido, Karen Ullrich, Aishwarya Agrawal, Kate Saenko, Asli Celikyilmaz, Vikas Chandra
Then, we present and discuss approaches to evaluate VLMs.
no code implementations • 25 Apr 2024 • Mazda Moayeri, Michael Rabbat, Mark Ibrahim, Diane Bouchacourt
We propose a method to encode and account for diversity within a class using inferred attributes, still in the zero-shot setting without retraining.
no code implementations • 11 Apr 2024 • Mazda Moayeri, Samyadeep Basu, Sriram Balasubramanian, Priyatham Kattakinda, Atoosa Chengini, Robert Brauneis, Soheil Feizi
Recent text-to-image generative models such as Stable Diffusion are extremely adept at mimicking and generating copyrighted content, raising concerns amongst artists that their unique styles may be improperly copied.
no code implementations • 29 Sep 2023 • Keivan Rezaei, Mehrdad Saberi, Mazda Moayeri, Soheil Feizi
To improve on these shortcomings, we propose a novel approach that prioritizes interpretability in this problem: we start by obtaining human-understandable concepts (tags) of images in the dataset and then analyze the model's behavior based on the presence or absence of combinations of these tags.
1 code implementation • 10 May 2023 • Mazda Moayeri, Keivan Rezaei, Maziar Sanjabi, Soheil Feizi
We observe that the mapping between an image's representation in one model to its representation in another can be learned surprisingly well with just a linear layer, even across diverse models.
no code implementations • 17 Nov 2022 • Sahil Singla, Atoosa Malemir Chegini, Mazda Moayeri, Soheil Feiz
Our Data-Centric Debugging (DCD) framework carefully creates a debug-train set by selecting images from $\mathcal{F}$ that are perceptually similar to the images in $\mathcal{E}_{sample}$.
no code implementations • 15 Sep 2022 • Mazda Moayeri, Kiarash Banihashem, Soheil Feizi
In this setting, through theoretical and empirical analysis, we show that (i) adversarial training with $\ell_1$ and $\ell_2$ norms increases the model reliance on spurious features; (ii) For $\ell_\infty$ adversarial training, spurious reliance only occurs when the scale of the spurious features is larger than that of the core features; (iii) adversarial training can have an unintended consequence in reducing distributional robustness, specifically when spurious correlations are changed in the new test domain.
no code implementations • 28 Mar 2022 • Sahil Singla, Mazda Moayeri, Soheil Feizi
Deep neural networks can be unreliable in the real world especially when they heavily use spurious features for their predictions.
1 code implementation • CVPR 2022 • Mazda Moayeri, Phillip Pope, Yogesh Balaji, Soheil Feizi
While datasets with single-label supervision have propelled rapid advances in image classification, additional annotations are necessary in order to quantitatively assess how models make predictions.
no code implementations • ICCV 2021 • Mazda Moayeri, Soheil Feizi
In this paper, we propose a self-supervised method to detect adversarial attacks and classify them to their respective threat models, based on a linear model operating on the embeddings from a pre-trained self-supervised encoder.