Search Results for author: Yashaswi Verma

Found 5 papers, 2 papers with code

Recurrent Image Annotation With Explicit Inter-Label Dependencies

1 code implementation ECCV 2020 Ayushi Dutta, Yashaswi Verma, C. V. Jawahar

Additionally, it provides a new perspecitve of looking at an unordered set of labels as equivalent to a collection of different permutations (sequences) of those labels, thus naturally aligning with the image annotation task.

Image Captioning

Action-based Early Autism Diagnosis Using Contrastive Feature Learning

1 code implementation12 Sep 2022 Asha Rani, Pankaj Yadav, Yashaswi Verma

To address this, we adopt contrastive feature learning in both self supervised and supervised learning frameworks, and show that these can lead to a significant increase in the prediction accuracy of a binary classifier on this task.

Contrastive Learning of Visual-Semantic Embeddings

no code implementations17 Oct 2021 Anurag Jain, Yashaswi Verma

Contrastive learning is a powerful technique to learn representations that are semantically distinctive and geometrically invariant.

Contrastive Learning Image Classification +2

On-the-fly Global Embeddings Using Random Projections for Extreme Multi-label Classification

no code implementations17 Dec 2019 Yashaswi Verma

The goal of eXtreme Multi-label Learning (XML) is to automatically annotate a given data point with the most relevant subset of labels from an extremely large vocabulary of labels (e. g., a million labels).

Extreme Multi-Label Classification Multi-Label Learning

Relative Parts: Distinctive Parts for Learning Relative Attributes

no code implementations CVPR 2014 Ramachandruni N. Sandeep, Yashaswi Verma, C. V. Jawahar

The notion of relative attributes as introduced by Parikh and Grauman (ICCV, 2011) provides an appealing way of comparing two images based on their visual properties (or attributes) such as "smiling" for face images, "naturalness" for outdoor images, etc.

Attribute Image Retrieval

Cannot find the paper you are looking for? You can Submit a new open access paper.