Search Results for author: Vineeth Balasubramanian

Found 7 papers, 1 papers with code

Fiducial Focus Augmentation for Facial Landmark Detection

no code implementations23 Feb 2024 Purbayan Kar, Vishal Chudasama, Naoyuki Onoe, Pankaj Wasnik, Vineeth Balasubramanian

To effectively utilize the newly proposed augmentation technique, we employ a Siamese architecture-based training mechanism with a Deep Canonical Correlation Analysis (DCCA)-based loss to achieve collective learning of high-level feature representations from two different views of the input images.

Face Alignment Facial Landmark Detection +1

Class-Incremental Learning with Cross-Space Clustering and Controlled Transfer

1 code implementation7 Aug 2022 Arjun Ashok, K J Joseph, Vineeth Balasubramanian

This allows the model to learn classes in such a way that it maximizes positive forward transfer from similar prior classes, thus increasing plasticity, and minimizes negative backward transfer on dissimilar prior classes, whereby strengthening stability.

Class Incremental Learning Clustering +1

Learning Modular Structures That Generalize Out-of-Distribution

no code implementations7 Aug 2022 Arjun Ashok, Chaitanya Devaguptapu, Vineeth Balasubramanian

generalization remains to be a key challenge for real-world machine learning systems.

Retrospective Loss: Looking Back to Improve Training of Deep Neural Networks

no code implementations24 Jun 2020 Surgan Jandial, Ayush Chopra, Mausoom Sarkar, Piyush Gupta, Balaji Krishnamurthy, Vineeth Balasubramanian

Deep neural networks (DNNs) are powerful learning machines that have enabled breakthroughs in several domains.

Training Autoencoders by Alternating Minimization

no code implementations ICLR 2018 Sneha Kudugunta, Adepu Shankar, Surya Chavali, Vineeth Balasubramanian, Purushottam Kar

We present DANTE, a novel method for training neural networks, in particular autoencoders, using the alternating minimization principle.

DAiSEE: Towards User Engagement Recognition in the Wild

no code implementations7 Sep 2016 Abhay Gupta, Arjun D'Cunha, Kamal Awasthi, Vineeth Balasubramanian

We introduce DAiSEE, the first multi-label video classification dataset comprising of 9068 video snippets captured from 112 users for recognizing the user affective states of boredom, confusion, engagement, and frustration in the wild.

General Classification Video Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.