no code implementations • 21 Mar 2023 • Jefferson Hernandez, Ruben Villegas, Vicente Ordonez
We show that visual representations learned under ViC-MAE generalize well to both video classification and image classification tasks.
no code implementations • 14 Mar 2023 • Jaspreet Ranjit, Tianlu Wang, Baishakhi Ray, Vicente Ordonez
We also find that (2) models finetuned on larger scale datasets are more likely to introduce new biased associations.
1 code implementation • 22 Nov 2022 • Paola Cascante-Bonilla, Leonid Karlinsky, James Seale Smith, Yanjun Qi, Vicente Ordonez
Generalized Zero-Shot Learning (GZSL) aims to train a classifier that can generalize to unseen classes, using a set of attributes as auxiliary information, and the visual features extracted from a pre-trained convolutional neural network.
no code implementations • 30 Jun 2022 • Ziyan Yang, Kushal Kafle, Franck Dernoncourt, Vicente Ordonez
We propose a margin-based loss for vision-language model pretraining that encourages gradient-based explanations that are consistent with region-level annotations.
1 code implementation • LREC 2022 • Samhita Honnavalli, Aesha Parekh, Lily Ou, Sophie Groenwold, Sharon Levy, Vicente Ordonez, William Yang Wang
Our results show that GPT-2 amplifies bias by considering women as junior and men as senior more often than the ground truth in both domains.
no code implementations • CVPR 2022 • Paola Cascante-Bonilla, Hui Wu, Letao Wang, Rogerio Feris, Vicente Ordonez
By exploiting 3D and physics simulation platforms, we provide a pipeline to generate synthetic data to expand and replace type-specific questions and answers without risking the exposure of sensitive or personal data that might be present in real images.
1 code implementation • 24 Mar 2022 • Ziyuan Zhong, Yuchi Tian, Conor J. Sweeney, Vicente Ordonez, Baishakhi Ray
In particular, it can repair confusion error and bias error of DNN models for both single-label and multi-label image classifications.
1 code implementation • 14 Dec 2021 • Aman Shrivastava, Ramprasaath R. Selvaraju, Nikhil Naik, Vicente Ordonez
We propose CLIP-Lite, an information efficient method for visual representation learning by feature alignment with textual annotations.
no code implementations • 29 Oct 2021 • Aman Shrivastava, Yanjun Qi, Vicente Ordonez
Our empirical results show that MIMKD outperforms competing approaches across a wide range of student-teacher pairs with different capacities, with different architectures, and when student networks are with extremely low capacity.
no code implementations • 16 Jun 2021 • Paola Cascante-Bonilla, Arshdeep Sekhon, Yanjun Qi, Vicente Ordonez
This paper proposes PatchMix, a data augmentation method that creates new samples by composing patches from pairs of images in a grid-like pattern.
1 code implementation • ICCV 2021 • Fuwen Tan, Jiangbo Yuan, Vicente Ordonez
Instance-level image retrieval is the task of searching in a large database for images that match an object in a query image.
1 code implementation • 2 Dec 2020 • Leticia Pinto-Alva, Ian K. Torres, Rosangel Garcia, Ziyan Yang, Vicente Ordonez
We aim for ChairSegments to be the equivalent of the CIFAR-10 dataset but for quickly designing and iterating over novel model architectures for segmentation.
2 code implementations • CVPR 2021 • Jack Lanchantin, Tianlu Wang, Vicente Ordonez, Yanjun Qi
Multi-label image classification is the task of predicting a set of labels corresponding to objects, attributes or other entities present in an image.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Ziyan Yang, Leticia Pinto-Alva, Franck Dernoncourt, Vicente Ordonez
Our work aims to leverage visual feature space to pass information across languages.
1 code implementation • EMNLP 2021 • Fuxiao Liu, Yinghan Wang, Tianlu Wang, Vicente Ordonez
We propose Visual News Captioner, an entity-aware model for the task of news image captioning.
2 code implementations • CVPR 2021 • Vitali Petsiuk, Rajiv Jain, Varun Manjunatha, Vlad I. Morariu, Ashutosh Mehra, Vicente Ordonez, Kate Saenko
We propose D-RISE, a method for generating visual explanations for the predictions of object detectors.
1 code implementation • ACL 2020 • Tianlu Wang, Xi Victoria Lin, Nazneen Fatema Rajani, Bryan McCann, Vicente Ordonez, Caiming Xiong
Word embeddings derived from human-generated corpora inherit strong gender bias which can be further amplified by downstream models.
1 code implementation • 16 Jan 2020 • Paola Cascante-Bonilla, Fuwen Tan, Yanjun Qi, Vicente Ordonez
Pseudo-labeling works by applying pseudo-labels to samples in the unlabeled set by using a model trained on the combination of the labeled samples and any previously pseudo-labeled samples, and iteratively repeating this process in a self-training cycle.
2 code implementations • ICCV 2021 • Sonia Baee, Erfan Pakdamanian, Inki Kim, Lu Feng, Vicente Ordonez, Laura Barnes
Inspired by human visual attention, we propose a novel inverse reinforcement learning formulation using Maximum Entropy Deep Inverse Reinforcement Learning (MEDIRL) for predicting the visual attention of drivers in accident-prone situations.
1 code implementation • NeurIPS 2019 • Fuwen Tan, Paola Cascante-Bonilla, Xiaoxiao Guo, Hui Wu, Song Feng, Vicente Ordonez
We show that using multiple rounds of natural language queries as input can be surprisingly effective to find arbitrarily specific images of complex scenes.
2 code implementations • 8 Aug 2019 • Paola Cascante-Bonilla, Kalpathy Sitaraman, Mengjia Luo, Vicente Ordonez
Film media is a rich form of artistic expression.
1 code implementation • 20 May 2019 • Yuchi Tian, Ziyuan Zhong, Vicente Ordonez, Gail Kaiser, Baishakhi Ray
We found that many of the reported erroneous cases in popular DNN image classifiers occur because the trained models confuse one class with another or show biases towards some classes over others.
1 code implementation • NAACL 2019 • Jieyu Zhao, Tianlu Wang, Mark Yatskar, Ryan Cotterell, Vicente Ordonez, Kai-Wei Chang
In this paper, we quantify, analyze and mitigate gender bias exhibited in ELMo's contextualized word vectors.
no code implementations • NAACL 2019 • Paola Cascante-Bonilla, Xuwang Yin, Vicente Ordonez, Song Feng
In this paper we introduce Chat-crowd, an interactive environment for visual layout composition via conversational interactions.
2 code implementations • ICCV 2019 • Tianlu Wang, Jieyu Zhao, Mark Yatskar, Kai-Wei Chang, Vicente Ordonez
In this work, we present a framework to measure and mitigate intrinsic biases with respect to protected variables --such as gender-- in visual recognition tasks.
3 code implementations • CVPR 2019 • Fuwen Tan, Song Feng, Vicente Ordonez
In this paper, we propose Text2Scene, a model that generates various forms of compositional scene representations from natural language descriptions.
1 code implementation • 22 May 2018 • Shanmin Pang, Jin Ma, Jianru Xue, Jihua Zhu, Vicente Ordonez
We show that by considering each deep feature as a heat source, our unsupervised aggregation method is able to avoid over-representation of \emph{bursty} features.
2 code implementations • NAACL 2018 • Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang
We introduce a new benchmark, WinoBias, for coreference resolution focused on gender bias.
1 code implementation • CVPR 2018 • Tianlu Wang, Kota Yamaguchi, Vicente Ordonez
We propose an inference procedure for deep convolutional neural networks (CNNs) when partial evidence is available.
3 code implementations • EMNLP 2017 • Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang
Language is increasingly being used to define rich visual recognition problems with supporting image collections sourced from the web.
no code implementations • EMNLP 2017 • Xuwang Yin, Vicente Ordonez
Generating captions for images is a task that has recently received considerable attention.
no code implementations • 4 Jun 2017 • Fuwen Tan, Crispin Bernier, Benjamin Cohen, Vicente Ordonez, Connelly Barnes
Image compositing is a method used to generate realistic yet fake imagery by inserting contents from one image to another.
2 code implementations • CVPR 2017 • Mark Yatskar, Vicente Ordonez, Luke Zettlemoyer, Ali Farhadi
Semantic sparsity is a common challenge in structured visual classification problems; when the output space is complex, the vast majority of the possible predictions are rarely, if ever, seen in the training set.
Ranked #10 on
Grounded Situation Recognition
on SWiG
17 code implementations • 16 Mar 2016 • Mohammad Rastegari, Vicente Ordonez, Joseph Redmon, Ali Farhadi
We propose two efficient approximations to standard convolutional neural networks: Binary-Weight-Networks and XNOR-Networks.
Ranked #10 on
Classification with Binary Neural Network
on ImageNet
no code implementations • TACL 2014 • Polina Kuznetsova, Vicente Ordonez, Tamara L. Berg, Yejin Choi
We present a new tree based approach to composing expressive image descriptions that makes use of naturally occuring web images with captions.
no code implementations • NeurIPS 2011 • Vicente Ordonez, Girish Kulkarni, Tamara L. Berg
We develop and demonstrate automatic image description methods using a large captioned photo collection.