no code implementations • 31 Oct 2023 • Gabrielle Cohn, Rishika Agarwal, Deepanshu Gupta, Siddharth Patwardhan
We introduce EELBERT, an approach for compression of transformer-based models (e. g., BERT), with minimal impact on the accuracy of downstream tasks.
no code implementations • 18 Aug 2020 • Rishika Agarwal, Xiaochuan Niu, Pranay Dighe, Srikanth Vishnubhotla, Sameer Badaskar, Devang Naik
In this paper, we propose a novel solution to the FTM problem by introducing a parallel ASR decoding process with a special language model trained from "out-of-domain" data sources.