Search Results for author: Rakshith Subramanyam

Found 4 papers, 3 papers with code

CREPE: Learnable Prompting With CLIP Improves Visual Relationship Prediction

1 code implementation10 Jul 2023 Rakshith Subramanyam, T. S. Jayram, Rushil Anirudh, Jayaraman J. Thiagarajan

In this paper, we explore the potential of Vision-Language Models (VLMs), specifically CLIP, in predicting visual object relationships, which involves interpreting visual features from images into language-based relations.

Object Relation

Target-Aware Generative Augmentations for Single-Shot Adaptation

1 code implementation22 May 2023 Kowshik Thopalli, Rakshith Subramanyam, Pavan Turaga, Jayaraman J. Thiagarajan

We argue that augmentations utilized by existing methods are insufficient to handle large distribution shifts, and hence propose a new approach SiSTA, which first fine-tunes a generative model from the source domain using a single-shot target, and then employs novel sampling strategies for curating synthetic target data.

Attribute Object Recognition +1

Single-Shot Domain Adaptation via Target-Aware Generative Augmentation

1 code implementation29 Oct 2022 Rakshith Subramanyam, Kowshik Thopalli, Spring Berman, Pavan Turaga, Jayaraman J. Thiagarajan

The problem of adapting models from a source domain using data from any target domain of interest has gained prominence, thanks to the brittle generalization in deep neural networks.

Attribute Test-time Adaptation

Contrastive Knowledge-Augmented Meta-Learning for Few-Shot Classification

no code implementations25 Jul 2022 Rakshith Subramanyam, Mark Heimann, Jayram Thathachar, Rushil Anirudh, Jayaraman J. Thiagarajan

Model agnostic meta-learning algorithms aim to infer priors from several observed tasks that can then be used to adapt to a new task with few examples.

Classification Few-Shot Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.