Search Results for author: Ruchika Chavhan

Found 7 papers, 3 papers with code

Fool Your (Vision and) Language Model With Embarrassingly Simple Permutations

1 code implementation2 Oct 2023 Yongshuo Zong, Tingyang Yu, Bingchen Zhao, Ruchika Chavhan, Timothy Hospedales

Large language and vision-language models are rapidly being deployed in practice thanks to their impressive capabilities in instruction following, in-context learning, and so on.

In-Context Learning Instruction Following +3

Meta Omnium: A Benchmark for General-Purpose Learning-to-Learn

1 code implementation CVPR 2023 Ondrej Bohdal, Yinbing Tian, Yongshuo Zong, Ruchika Chavhan, Da Li, Henry Gouk, Li Guo, Timothy Hospedales

Meta-learning and other approaches to few-shot learning are widely studied for image recognition, and are increasingly applied to other vision tasks such as pose estimation and dense prediction.

Few-Shot Learning Pose Estimation +1

Amortised Invariance Learning for Contrastive Self-Supervision

1 code implementation24 Feb 2023 Ruchika Chavhan, Henry Gouk, Jan Stuehmer, Calum Heggan, Mehrdad Yaghoobi, Timothy Hospedales

Contrastive self-supervised learning methods famously produce high quality transferable representations by learning invariances to different data augmentations.

Contrastive Learning Representation Learning +1

Quality Diversity for Visual Pre-Training

no code implementations ICCV 2023 Ruchika Chavhan, Henry Gouk, Da Li, Timothy Hospedales

Notably, the augmentations used in both supervised and self-supervised training lead to features with high invariance to spatial and appearance transformations.

Inductive Bias Transfer Learning

HyperInvariances: Amortizing Invariance Learning

no code implementations17 Jul 2022 Ruchika Chavhan, Henry Gouk, Jan Stühmer, Timothy Hospedales

Providing invariances in a given learning task conveys a key inductive bias that can lead to sample-efficient learning and good generalisation, if correctly specified.

Inductive Bias

FRIDA -- Generative Feature Replay for Incremental Domain Adaptation

no code implementations28 Dec 2021 Sayan Rakshit, Anwesh Mohanty, Ruchika Chavhan, Biplab Banerjee, Gemma Roig, Subhasis Chaudhuri

Inspired by the notion of generative feature replay, we propose a novel framework called Feature Replay based Incremental Domain Adaptation (FRIDA) which leverages a new incremental generative adversarial network (GAN) called domain-generic auxiliary classification GAN (DGAC-GAN) for producing domain-specific feature representations seamlessly.

Generative Adversarial Network Unsupervised Domain Adaptation

A Novel Actor Dual-Critic Model for Remote Sensing Image Captioning

no code implementations5 Oct 2020 Ruchika Chavhan, Biplab Banerjee, Xiao Xiang Zhu, Subhasis Chaudhuri

We deal with the problem of generating textual captions from optical remote sensing (RS) images using the notion of deep reinforcement learning.

Image Captioning Sentence

Cannot find the paper you are looking for? You can Submit a new open access paper.