Search Results for author: Dhruva Sahrawat

Found 4 papers, 2 papers with code

"Notic My Speech" -- Blending Speech Patterns With Multimedia

no code implementations12 Jun 2020 Dhruva Sahrawat, Yaman Kumar, Shashwat Aggarwal, Yifang Yin, Rajiv Ratn Shah, Roger Zimmermann

To close the gap between speech understanding and multimedia video applications, in this paper, we show the initial experiments by modelling the perception on visual speech and showing its use case on video compression.

speech-recognition Video Compression +1

Heterogeneity Loss to Handle Intersubject and Intrasubject Variability in Cancer

1 code implementation6 Mar 2020 Shubham Goswami, Suril Mehta, Dhruva Sahrawat, Anubha Gupta, Ritu Gupta

We also propose an unorthodox ensemble strategy that helps us in providing improved classification over models trained on 7-folds giving a weighted-$F_1$ score of 95. 26% on unseen (test) subjects' data that are, so far, the best results on the C-NMC 2019 dataset for B-ALL classification.

Keyphrase Extraction from Scholarly Articles as Sequence Labeling using Contextualized Embeddings

no code implementations19 Oct 2019 Dhruva Sahrawat, Debanjan Mahata, Mayank Kulkarni, Haimin Zhang, Rakesh Gosangi, Amanda Stent, Agniv Sharma, Yaman Kumar, Rajiv Ratn Shah, Roger Zimmermann

In this paper, we formulate keyphrase extraction from scholarly articles as a sequence labeling task solved using a BiLSTM-CRF, where the words in the input text are represented using deep contextualized embeddings.

Keyphrase Extraction Word Embeddings

Harnessing GANs for Zero-shot Learning of New Classes in Visual Speech Recognition

1 code implementation29 Jan 2019 Yaman Kumar, Dhruva Sahrawat, Shubham Maheshwari, Debanjan Mahata, Amanda Stent, Yifang Yin, Rajiv Ratn Shah, Roger Zimmermann

To solve this problem, we present a novel approach to zero-shot learning by generating new classes using Generative Adversarial Networks (GANs), and show how the addition of unseen class samples increases the accuracy of a VSR system by a significant margin of 27% and allows it to handle speaker-independent out-of-vocabulary phrases.

speech-recognition Visual Speech Recognition +1

Cannot find the paper you are looking for? You can Submit a new open access paper.