no code implementations • 10 Jan 2025 • Oindrila Saha, Logan Lawrence, Grant van Horn, Subhransu Maji
Transductive zero-shot learning with vision-language models leverages image-image similarities within the dataset to achieve better classification accuracy compared to the inductive setting.
no code implementations • 19 Dec 2024 • Rangel Daroya, Elijah Cole, Oisin Mac Aodha, Grant van Horn, Subhransu Maji
WildSAT uses a contrastive learning framework to combine information from species distribution maps with text descriptions that capture habitat and range details, alongside satellite images, to train or fine-tune models.
no code implementations • 11 Dec 2024 • Rangel Daroya, Luisa Vieira Lucchese, Travis Simmons, Punwath Prum, Tamlin Pavelsky, John Gardner, Colin J. Gleason, Subhransu Maji
We experiment with recent advances in deep network architectures and show that masking models can benefit from these, especially when combined with pre-training on large satellite imagery datasets.
1 code implementation • 14 Oct 2024 • Max Hamilton, Christian Lange, Elijah Cole, Alexander Shepard, Samuel Heinrich, Oisin Mac Aodha, Grant van Horn, Subhransu Maji
Species range maps (SRMs) are essential tools for research and policy-making in ecology, conservation, and environmental management.
no code implementations • 27 May 2024 • Yuefeng Peng, Jaechul Roh, Subhransu Maji, Amir Houmansadr
The core idea is that a member sample exhibits more resistance to adversarial perturbations than a non-member.
1 code implementation • CVPR 2024 • Rangel Daroya, Aaron Sun, Subhransu Maji
Modeling and visualizing relationships between tasks or datasets is an important step towards solving various meta-tasks such as dataset discovery, multi-tasking, and transfer learning.
1 code implementation • CVPR 2024 • Oindrila Saha, Grant van Horn, Subhransu Maji
By prompting LLMs in various ways, we generate descriptions that capture visual appearance, habitat, and geographic regions and pair them with existing attributes such as the taxonomic structure of the categories.
1 code implementation • 8 Dec 2023 • Gustavo Perez, Daniel Sheldon, Grant van Horn, Subhransu Maji
We propose a human-in-the-loop approach for estimating population size driven by a pairwise similarity derived from an off-the-shelf Re-ID system.
1 code implementation • 25 Sep 2023 • Oindrila Saha, Subhransu Maji
For example, under a linear-evaluation scheme, the classification accuracy of a ResNet50 trained on ImageNet using DetCon, a self-supervised learning approach, improves from 35. 4% to 42. 0% on the Caltech-UCSD Birds, from 35. 5% to 44. 1% on the FGVC Aircraft, and from 29. 7% to 37. 4% on the Stanford Cars.
1 code implementation • 20 Sep 2023 • Rangel Daroya, Aaron Sun, Subhransu Maji
We present a set of metrics that utilize vision priors to effectively assess the performance of saliency methods on image classification tasks.
no code implementations • ICCV 2023 • Zezhou Cheng, Carlos Esteves, Varun Jampani, Abhishek Kar, Subhransu Maji, Ameesh Makadia
Consequently, there is growing interest in extending NeRF models to jointly optimize camera poses and scene representation, which offers an alternative to off-the-shelf SfM pipelines which have well-understood failure modes.
no code implementations • 5 Jun 2023 • Gustavo Perez, Subhransu Maji, Daniel Sheldon
Many modern applications use computer vision to detect and count objects in massive image collections.
no code implementations • 13 Dec 2022 • Zezhou Cheng, Matheus Gadelha, Subhransu Maji
We propose a technique for learning single-view 3D object pose estimation models by utilizing a new source of data -- in-the-wild videos where objects turn.
2 code implementations • 18 Aug 2022 • Gopal Sharma, Kangxue Yin, Subhransu Maji, Evangelos Kalogerakis, Or Litany, Sanja Fidler
As a result, the learned 2D representations are view-invariant and geometrically consistent, leading to better generalization when trained on a limited number of labeled shapes compared to alternatives that utilize self-supervision in 2D or 3D alone.
no code implementations • 24 Jul 2022 • Zezhou Cheng, Menglei Chai, Jian Ren, Hsin-Ying Lee, Kyle Olszewski, Zeng Huang, Subhransu Maji, Sergey Tulyakov
In this paper, we propose a generic multi-modal generative model that couples the 2D modalities and implicit 3D representations through shared latent spaces.
no code implementations • 11 Apr 2022 • Oindrila Saha, Zezhou Cheng, Subhransu Maji
A significant bottleneck in training deep networks for part segmentation is the cost of obtaining detailed annotations.
1 code implementation • 22 Mar 2022 • Chenyun Wu, Subhransu Maji
We investigate how well CLIP understands texture in natural images described by natural language.
1 code implementation • 27 Dec 2021 • Gopal Sharma, Bidya Dash, Aruni RoyChowdhury, Matheus Gadelha, Marios Loizou, Liangliang Cao, Rui Wang, Erik Learned-Miller, Subhransu Maji, Evangelos Kalogerakis
We present PriFit, a semi-supervised approach for label-efficient learning of 3D point cloud segmentation networks.
no code implementations • CVPR 2022 • Oindrila Saha, Zezhou Cheng, Subhransu Maji
Motivated by this we present an alternative approach based on contrastive learning and compare their performance on standard few-shot part segmentation benchmarks.
1 code implementation • 23 Nov 2021 • Jong-Chyi Su, Subhransu Maji
We propose techniques to incorporate coarse taxonomic labels to train image classifiers in fine-grained domains.
no code implementations • 3 Aug 2021 • Gustavo Perez, Subhransu Maji
We consider the problem of adapting a network trained on three-channel color images to a hyperspectral domain with a large number of channels.
1 code implementation • 2 Jul 2021 • Zenglin Shi, Pascal Mettes, Subhransu Maji, Cees G. M. Snoek
The deep image prior showed that a randomly initialized network with a suitable architecture can be trained to solve inverse imaging problems by simply optimizing it's parameters to reconstruct a single degraded image.
2 code implementations • 2 Jun 2021 • Jong-Chyi Su, Subhransu Maji
Semi-iNat is a challenging dataset for semi-supervised classification with a long-tailed distribution of classes, fine-grained categories, and domain shifts between labeled and unlabeled data.
1 code implementation • CVPR 2021 • Jong-Chyi Su, Zezhou Cheng, Subhransu Maji
We evaluate the effectiveness of semi-supervised learning (SSL) on a realistic benchmark where data exhibits considerable class imbalance and contains images from novel classes.
2 code implementations • 11 Mar 2021 • Jong-Chyi Su, Subhransu Maji
From this collection, we sample a subset of classes and their labels, while adding the images from the remaining classes to the unlabeled set of images.
no code implementations • 26 Jan 2021 • Orchid Majumder, Avinash Ravichandran, Subhransu Maji, Alessandro Achille, Marzia Polito, Stefano Soatto
In this work we investigate the complementary roles of these two sources of information by combining instance-discriminative contrastive learning and supervised learning in a single framework called Supervised Momentum Contrastive learning (SUPMOCO).
1 code implementation • CVPR 2021 • Zhaowei Cai, Avinash Ravichandran, Subhransu Maji, Charless Fowlkes, Zhuowen Tu, Stefano Soatto
We present a plug-in replacement for batch normalization (BN) called exponential moving average normalization (EMAN), which improves the performance of existing student-teacher based self- and semi-supervised learning techniques.
Self-Supervised Learning
Semi-Supervised Image Classification
1 code implementation • 16 Dec 2020 • Gustavo Perez, Matteo Messa, Daniela Calzetti, Subhransu Maji, Dooseok Jung, Angela Adamo, Mattia Siressi
We present a machine learning (ML) pipeline to identify star clusters in the multi{color images of nearby galaxies, from observations obtained with the Hubble Space Telescope as part of the Treasury Project LEGUS (Legacy ExtraGalactic Ultraviolet Survey).
no code implementations • 6 Oct 2020 • Zitian Chen, Subhransu Maji, Erik Learned-Miller
To alleviate problems caused by the distribution shift, previous research has explored the use of unlabeled examples from the novel classes, in addition to labeled examples of the base classes, which is known as the transductive setting.
1 code implementation • CVPR 2020 • Chenyun Wu, Zhe Lin, Scott Cohen, Trung Bui, Subhransu Maji
We consider the problem of segmenting image regions given a natural language phrase, and study it on a novel dataset of 77, 262 images and 345, 486 phrase-region pairs.
Ranked #4 on
Referring Expression Segmentation
on PhraseCut
no code implementations • ECCV 2020 • Chenyun Wu, Mikayla Timm, Subhransu Maji
Textures in natural images can be characterized by color, shape, periodicity of elements within them, and other attributes that can be described using natural language.
1 code implementation • ICCV 2021 • Zezhou Cheng, Jong-Chyi Su, Subhransu Maji
Given a collection of images, humans are able to discover landmarks by modeling the shared geometric structure across instances.
1 code implementation • EMNLP 2020 • Tu Vu, Tong Wang, Tsendsuren Munkhdalai, Alessandro Sordoni, Adam Trischler, Andrew Mattarella-Micke, Subhransu Maji, Mohit Iyyer
We also develop task embeddings that can be used to predict the most transferable source tasks for a given target task, and we validate their effectiveness in experiments controlled for source and target data size.
no code implementations • 24 Apr 2020 • Zezhou Cheng, Saadia Gabriel, Pankaj Bhambhani, Daniel Sheldon, Subhransu Maji, Andrew Laughlin, David Winkler
The US weather radar archive holds detailed information about biological phenomena in the atmosphere over the last 20 years.
no code implementations • 8 Apr 2020 • Matheus Gadelha, Rui Wang, Subhransu Maji
We present a prior for manifold structured data, such as surfaces of 3D shapes, where deep neural networks are adopted to reconstruct a target shape using gradient descent starting from a random initialization.
no code implementations • CVPR 2020 • Matheus Gadelha, Giorgio Gori, Duygu Ceylan, Radomir Mech, Nathan Carr, Tamy Boubekeur, Rui Wang, Subhransu Maji
We present a generative model to synthesize 3D shapes as sets of handles -- lightweight proxies that approximate the original 3D shape -- for applications in interactive editing, shape parsing, and building compact 3D representations.
1 code implementation • ECCV 2020 • Matheus Gadelha, Aruni RoyChowdhury, Gopal Sharma, Evangelos Kalogerakis, Liangliang Cao, Erik Learned-Miller, Rui Wang, Subhransu Maji
The problems of shape classification and part segmentation from 3D point clouds have garnered increasing attention in the last few years.
2 code implementations • ECCV 2020 • Gopal Sharma, Difan Liu, Subhransu Maji, Evangelos Kalogerakis, Siddhartha Chaudhuri, Radomír Měch
We propose a novel, end-to-end trainable, deep network called ParSeNet that decomposes a 3D point cloud into parametric surface patches, including B-spline patches as well as basic geometric primitives.
no code implementations • 22 Dec 2019 • Gopal Sharma, Rishabh Goyal, Difan Liu, Evangelos Kalogerakis, Subhransu Maji
We investigate two architectures for this task --- a vanilla encoder (CNN) - decoder (RNN) and another architecture that augments the encoder with an explicit memory module based on the program execution stack.
1 code implementation • 18 Dec 2019 • Prithvijit Chakrabarty, Subhransu Maji
The "deep image prior" proposed by Ulyanov et al. is an intriguing property of neural nets: a convolutional encoder-decoder network can be used as a prior for natural images.
2 code implementations • ECCV 2020 • Jong-Chyi Su, Subhransu Maji, Bharath Hariharan
We investigate the role of self-supervised learning (SSL) in the context of few-shot learning.
no code implementations • 3 Oct 2019 • Gopal Sharma, Evangelos Kalogerakis, Subhransu Maji
We present a framework for learning representations of 3D shapes that reflect the information present in this meta data and show that it leads to improved generalization for semantic segmentation tasks.
no code implementations • 2 Jul 2019 • Tsung-Yu Lin, Mikayla Timm, Chenyun Wu, Subhransu Maji
We analyze how categories from recent FGVC challenges can be described by their textural content.
no code implementations • 17 Jun 2019 • Jong-Chyi Su, Subhransu Maji, Bharath Hariharan
We present a technique to improve the transferability of deep representations learned on small labeled datasets by introducing self-supervised tasks as auxiliary loss functions.
no code implementations • 11 Jun 2019 • Matheus Gadelha, Aartika Rai, Subhransu Maji, Rui Wang
To this end, we present new differentiable projection operators that can be used by PrGAN to learn better 3D generative models.
1 code implementation • CVPR 2019 • Zezhou Cheng, Matheus Gadelha, Subhransu Maji, Daniel Sheldon
The deep image prior was recently introduced as a prior for natural images.
no code implementations • 16 Apr 2019 • Jong-Chyi Su, Yi-Hsuan Tsai, Kihyuk Sohn, Buyu Liu, Subhransu Maji, Manmohan Chandraker
Our approach, active adversarial domain adaptation (AADA), explores a duality between two related problems: adversarial domain alignment and importance sampling for adapting models across domains.
7 code implementations • CVPR 2019 • Kwonjoon Lee, Subhransu Maji, Avinash Ravichandran, Stefano Soatto
We propose to use these predictors as base learners to learn representations for few-shot learning and show they offer better tradeoffs between feature size and performance across a range of few-shot recognition benchmarks.
Ranked #12 on
Few-Shot Image Classification
on FC100 5-way (1-shot)
1 code implementation • ICCV 2019 • Alessandro Achille, Michael Lam, Rahul Tewari, Avinash Ravichandran, Subhransu Maji, Charless Fowlkes, Stefano Soatto, Pietro Perona
We demonstrate that this embedding is capable of predicting task similarities that match our intuition about semantic and taxonomic relations between different visual tasks (e. g., tasks based on classifying different types of plants are similar) We also demonstrate the practical value of this framework for the meta-task of selecting a pre-trained feature extractor for a new task.
no code implementations • 7 Sep 2018 • Jong-Chyi Su, Matheus Gadelha, Rui Wang, Subhransu Maji
We investigate the role of representations and architectures for classifying 3D shapes in terms of their computational efficiency, generalization, and robustness to adversarial transformations.
no code implementations • ECCV 2018 • Tsung-Yu Lin, Subhransu Maji, Piotr Koniusz
In this paper, we study a class of orderless aggregation functions designed to minimize interference or equalize contributions in the context of second-order features and we show that they can be computed just as efficiently as their first-order counterparts and they have favorable properties over aggregation by summation.
1 code implementation • ECCV 2018 • Matheus Gadelha, Rui Wang, Subhransu Maji
We present multiresolution tree-structured networks to process point clouds for 3D shape understanding and generation tasks.
no code implementations • 24 May 2018 • Yang Zhou, Zhan Xu, Chris Landreth, Evangelos Kalogerakis, Subhransu Maji, Karan Singh
We present a novel deep-learning based approach to producing animator-centric speech motion curves that drive a JALI or standard FACS-based production face-rig, directly from input audio.
Graphics
2 code implementations • CVPR 2018 • Hang Su, Varun Jampani, Deqing Sun, Subhransu Maji, Evangelos Kalogerakis, Ming-Hsuan Yang, Jan Kautz
We present a network architecture for processing point clouds that directly operates on a collection of points represented as a sparse set of samples in a high-dimensional lattice.
Ranked #33 on
Semantic Segmentation
on ScanNet
(test mIoU metric)
1 code implementation • CVPR 2018 • Gopal Sharma, Rishabh Goyal, Difan Liu, Evangelos Kalogerakis, Subhransu Maji
In contrast, our model uses a recurrent neural network that parses the input shape in a top-down manner, which is significantly faster and yields a compact and easy-to-interpret sequence of modeling instructions.
no code implementations • ICCV 2017 • Jong-Chyi Su, Chenyun Wu, Huaizu Jiang, Subhransu Maji
We collect a large dataset of such phrases by asking annotators to describe several visual differences between a pair of instances within a category.
no code implementations • 21 Jul 2017 • Tsung-Yu Lin, Subhransu Maji
We present an alternative scheme for computing gradients that is faster and yet it offers improvements over the baseline model.
3 code implementations • 20 Jul 2017 • Zhaoliang Lun, Matheus Gadelha, Evangelos Kalogerakis, Subhransu Maji, Rui Wang
The decoder converts this representation into depth and normal maps capturing the underlying surface from several output viewpoints.
no code implementations • 19 Jul 2017 • Matheus Gadelha, Subhransu Maji, Rui Wang
We propose to use the expressive power of neural networks to learn a distribution over the shape coefficients in a generative-adversarial framework.
no code implementations • 18 Dec 2016 • Matheus Gadelha, Subhransu Maji, Rui Wang
In this paper we investigate the problem of inducing a distribution over three-dimensional structures given two-dimensional views of multiple objects taken from unknown viewpoints.
1 code implementation • CVPR 2017 • Evangelos Kalogerakis, Melinos Averkiou, Subhransu Maji, Siddhartha Chaudhuri
Our architecture combines image-based Fully Convolutional Networks (FCNs) and surface-based Conditional Random Fields (CRFs) to yield coherent segmentations of 3D shapes.
no code implementations • 1 Apr 2016 • Jong-Chyi Su, Subhransu Maji
Model compression and knowledge distillation have been successfully applied for cross-architecture and cross-domain transfer learning.
no code implementations • 10 Feb 2016 • Tamir Hazan, Francesco Orabona, Anand D. Sarwate, Subhransu Maji, Tommi Jaakkola
This paper shows that the expected value of perturb-max inference with low dimensional perturbations can be used sequentially to generate unbiased samples from the Gibbs distribution.
no code implementations • ICCV 2015 • Tsung-Yu Lin, Aruni RoyChowdhury, Subhransu Maji
We propose bilinear models, a recognition architecture that consists of two feature extractors whose outputs are multiplied using outer product at each location of the image and pooled to obtain an image descriptor.
Ranked #68 on
Fine-Grained Image Classification
on CUB-200-2011
Fine-Grained Image Classification
Fine-Grained Visual Recognition
no code implementations • CVPR 2016 • Tsung-Yu Lin, Subhransu Maji
A number of recent approaches have used deep convolutional neural networks (CNNs) to build texture representations.
no code implementations • 9 Jul 2015 • Mircea Cimpoi, Subhransu Maji, Iasonas Kokkinos, Andrea Vedaldi
Visual textures have played a key role in image understanding because they convey important semantics of images, and because texture representations that pool local image descriptors in an orderless manner have had a tremendous impact in diverse applications.
no code implementations • 3 Jun 2015 • Aruni RoyChowdhury, Tsung-Yu Lin, Subhransu Maji, Erik Learned-Miller
We demonstrate the performance of the B-CNN model beginning from an AlexNet-style network pre-trained on ImageNet.
1 code implementation • CVPR 2015 • Mircea Cimpoi, Subhransu Maji, Andrea Vedaldi
Research in texture recognition often concentrates on the problem of material recognition in uncluttered conditions, an assumption rarely met by applications.
no code implementations • ICCV 2015 • Hang Su, Subhransu Maji, Evangelos Kalogerakis, Erik Learned-Miller
A longstanding question in computer vision concerns the representation of 3D shapes for recognition: should 3D shapes be represented with descriptors operating on their native 3D formats, such as voxel grid or polygon mesh, or can they be effectively represented with view-based descriptors?
Ranked #103 on
3D Point Cloud Classification
on ModelNet40
4 code implementations • 29 Apr 2015 • Tsung-Yu Lin, Aruni RoyChowdhury, Subhransu Maji
We then present a systematic analysis of these networks and show that (1) the bilinear features are highly redundant and can be reduced by an order of magnitude in size without significant loss in accuracy, (2) are also effective for other image classification tasks such as texture and scene recognition, and (3) can be trained from scratch on the ImageNet dataset offering consistent improvements over the baseline architecture.
Ranked #27 on
Fine-Grained Image Classification
on NABirds
Fine-Grained Image Classification
Fine-Grained Visual Recognition
+1
no code implementations • 5 Mar 2015 • Liwen Zhang, Subhransu Maji, Ryota Tomioka
Similarity between objects is multi-faceted and it can be easier for human annotators to measure it when the focus is on a specific aspect.
no code implementations • 25 Nov 2014 • Mircea Cimpoi, Subhransu Maji, Andrea Vedaldi
Research in texture recognition often concentrates on the problem of material recognition in uncluttered conditions, an assumption rarely met by applications.
no code implementations • CVPR 2014 • Rashmi Tonge, Subhransu Maji, C. V. Jawahar
We propose an approach for segmenting the individual buildings in typical skyline images.
no code implementations • CVPR 2014 • Andrea Vedaldi, Siddharth Mahendran, Stavros Tsogkas, Subhransu Maji, Ross Girshick, Juho Kannala, Esa Rahtu, Iasonas Kokkinos, Matthew B. Blaschko, David Weiss, Ben Taskar, Karen Simonyan, Naomi Saphra, Sammy Mohamed
We show that the collected data can be used to study the relation between part detection and attribute prediction by diagnosing the performance of classifiers that pool information from different parts of an object.
no code implementations • CVPR 2014 • Catherine Wah, Grant van Horn, Steve Branson, Subhransu Maji, Pietro Perona, Serge Belongie
Current human-in-the-loop fine-grained visual categorization systems depend on a predefined vocabulary of attributes and parts, usually determined by experts.
no code implementations • NeurIPS 2013 • Tamir Hazan, Subhransu Maji, Joseph Keshet, Tommi Jaakkola
In this work we develop efficient methods for learning random MAP predictors for structured label problems.
14 code implementations • CVPR 2014 • Mircea Cimpoi, Subhransu Maji, Iasonas Kokkinos, Sammy Mohamed, Andrea Vedaldi
Patterns and textures are defining characteristics of many natural objects: a shirt can be striped, the wings of a butterfly can be veined, and the skin of an animal can be scaly.
no code implementations • NeurIPS 2013 • Tamir Hazan, Subhransu Maji, Tommi Jaakkola
In this paper we describe how MAP inference can be used to sample efficiently from Gibbs distributions.
no code implementations • 21 Jun 2013 • Subhransu Maji, Esa Rahtu, Juho Kannala, Matthew Blaschko, Andrea Vedaldi
This paper introduces FGVC-Aircraft, a new dataset containing 10, 000 images of aircraft spanning 100 aircraft models, organised in a three-level hierarchy.
no code implementations • CVPR 2013 • Subhransu Maji, Gregory Shakhnarovich
We study the problem of part discovery when partial correspondence between instances of a category are available.