no code implementations • 2 Feb 2024 • David Grangier, Angelos Katharopoulos, Pierre Ablin, Awni Hannun
In the first scenario, we propose an effective solution based on importance sampling: we resample the pretraining set to imitate the specialization data and train a small model on it.
Ranked #1 on
Language Modelling
on The Pile
(Test perplexity metric)
no code implementations • 20 Nov 2023 • David Grangier, Pierre Ablin, Awni Hannun
Large neural networks pretrained on web-scale corpora are central to modern machine learning.
no code implementations • 10 Nov 2023 • Lucio Dery, David Grangier, Awni Hannun
We propose a framework which combines structured pruning with transfer learning to reduce the need for task-specific data.
2 code implementations • 29 Jan 2022 • Jacob Kahn, Vineel Pratap, Tatiana Likhomanenko, Qiantong Xu, Awni Hannun, Jeff Cai, Paden Tomasello, Ann Lee, Edouard Grave, Gilad Avidov, Benoit Steiner, Vitaliy Liptchinsky, Gabriel Synnaeve, Ronan Collobert
This is in part due to the difficulties involved in prototyping new computational paradigms with existing frameworks.
1 code implementation • 28 Jan 2022 • Vineel Pratap, Awni Hannun, Gabriel Synnaeve, Ronan Collobert
These experiments show that STC can recover most of the performance of supervised baseline when up to 70% of the labels are missing.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+3
no code implementations • 6 Oct 2021 • Shubho Sengupta, Vineel Pratap, Awni Hannun
We benchmark our parallel algorithm on the composition of random graphs and the composition of graphs commonly used in speech recognition.
1 code implementation • NeurIPS 2021 • Brian Knott, Shobha Venkataraman, Awni Hannun, Shubho Sengupta, Mark Ibrahim, Laurens van der Maaten
To foster adoption of secure MPC in machine learning, we present CrypTen: a software framework that exposes popular secure MPC primitives via abstractions that are common in modern machine-learning frameworks, such as tensor computations, automatic differentiation, and modular neural networks.
1 code implementation • 30 Jul 2021 • Awni Hannun
Given the remarkable changes in the state of speech recognition over the previous decade, what can we expect over the coming decade?
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • 21 Jun 2021 • Awni Hannun
I argue that the alternative, evolution, is important to the development of machine intelligence and underinvested in terms of research allocation.
2 code implementations • ICLR 2022 • Yuge Shi, Jeffrey Seely, Philip H. S. Torr, N. Siddharth, Awni Hannun, Nicolas Usunier, Gabriel Synnaeve
We perform experiments on both the Wilds benchmark, which captures distribution shift in the real world, as well as datasets in DomainBed benchmark that focuses more on synthetic-to-real transfer.
Ranked #6 on
Image Classification
on iWildCam2020-WILDS
1 code implementation • NeurIPS 2021 • Ruihan Wu, Chuan Guo, Awni Hannun, Laurens van der Maaten
Machine-learning systems such as self-driving cars or virtual assistants are composed of a large number of machine-learning models that recognize image content, transcribe speech, analyze natural language, infer preferences, rank options, etc.
1 code implementation • 23 Feb 2021 • Awni Hannun, Chuan Guo, Laurens van der Maaten
This information leaks either through the model itself or through predictions made by the model.
no code implementations • 11 Dec 2020 • Mimee Xu, Laurens van der Maaten, Awni Hannun
We show that in private, forward influence functions provide an appealing trade-off between high quality appraisal and required computation, in spite of label noise, class imbalance, and missing data.
1 code implementation • 2 Oct 2020 • Awni Hannun, Vineel Pratap, Jacob Kahn, Wei-Ning Hsu
We introduce a framework for automatic differentiation with weighted finite-state transducers (WFSTs) allowing them to be used dynamically at training time.
1 code implementation • 9 Jul 2020 • Laurens van der Maaten, Awni Hannun
This is problematic when the training data needs to remain private.
no code implementations • 6 Jul 2020 • Vineel Pratap, Anuroop Sriram, Paden Tomasello, Awni Hannun, Vitaliy Liptchinsky, Gabriel Synnaeve, Ronan Collobert
We study training a single acoustic model for multiple languages with the aim of improving automatic speech recognition (ASR) performance on low-resource languages, and over-all simplifying deployment of ASR systems that support diverse languages.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
1 code implementation • 19 May 2020 • Qiantong Xu, Tatiana Likhomanenko, Jacob Kahn, Awni Hannun, Gabriel Synnaeve, Ronan Collobert
In particular, IPL fine-tunes an existing model at each iteration using both labeled data and a subset of unlabeled data.
Ranked #13 on
Speech Recognition
on LibriSpeech test-other
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+4
1 code implementation • 24 Feb 2020 • Wei-Ning Hsu, Ann Lee, Gabriel Synnaeve, Awni Hannun
For sequence transduction tasks like speech recognition, a strong structured prior model encodes rich information about the target space, implicitly ruling out invalid sequences by assigning them low probability.
Ranked #51 on
Speech Recognition
on LibriSpeech test-other
no code implementations • 27 Jan 2020 • Vineel Pratap, Qiantong Xu, Jacob Kahn, Gilad Avidov, Tatiana Likhomanenko, Awni Hannun, Vitaliy Liptchinsky, Gabriel Synnaeve, Ronan Collobert
We design an online end-to-end speech recognition system based on Time-Depth Separable (TDS) convolutions and Connectionist Temporal Classification (CTC).
no code implementations • 9 Jan 2020 • Chuan Guo, Awni Hannun, Brian Knott, Laurens van der Maaten, Mark Tygert, Ruiyu Zhu
Secure multiparty computations enable the distribution of so-called shares of sensitive data to multiple parties such that the multiple parties can effectively process the data while being unable to glean much information about the data (at least not without collusion among all parties to put back together all the shares).
1 code implementation • ICML 2020 • Chuan Guo, Tom Goldstein, Awni Hannun, Laurens van der Maaten
Good data stewardship requires removal of data at the request of the data's owner.
no code implementations • 16 Oct 2019 • Adrien Dufraux, Emmanuel Vincent, Awni Hannun, Armelle Brun, Matthijs Douze
The transcriptions used to train an Automatic Speech Recognition (ASR) system may contain errors.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • 11 Oct 2019 • Awni Hannun, Brian Knott, Shubho Sengupta, Laurens van der Maaten
This paper considers a learning setting in which multiple parties aim to train a contextual bandit together in a private way: the parties aim to maximize the total reward but do not want to share any of the relevant information they possess with the other parties.
no code implementations • 25 Sep 2019 • Wei-Ning Hsu, Ann Lee, Gabriel Synnaeve, Awni Hannun
We propose local prior matching (LPM), a self-supervised objective for speech recognition.
no code implementations • 19 Sep 2019 • Jacob Kahn, Ann Lee, Awni Hannun
We revisit self-training in the context of end-to-end speech recognition.
no code implementations • ICML 2020 • Ronan Collobert, Awni Hannun, Gabriel Synnaeve
We propose a direct-to-word sequence model which uses a word network to learn word embeddings from letters.
no code implementations • 4 Apr 2019 • Awni Hannun, Ann Lee, Qiantong Xu, Ronan Collobert
Coupled with a convolutional language model, our time-depth separable convolution architecture improves by more than 22% relative WER over the best previously reported sequence-to-sequence results on the noisy LibriSpeech test set.
1 code implementation • 16 Feb 2019 • Ronan Collobert, Awni Hannun, Gabriel Synnaeve
We demonstrate our approach scales by applying it to speech recognition, jointly training acoustic and word-level language models.
8 code implementations • 18 Dec 2018 • Vineel Pratap, Awni Hannun, Qiantong Xu, Jeff Cai, Jacob Kahn, Gabriel Synnaeve, Vitaliy Liptchinsky, Ronan Collobert
This paper introduces wav2letter++, the fastest open-source deep learning speech recognition framework.
3 code implementations • 28 Nov 2016 • Chris Lengerich, Awni Hannun
We propose a single neural network architecture for two tasks: on-line keyword spotting and voice activity detection.
no code implementations • 31 Mar 2016 • Zhenyao Zhu, Jesse H. Engel, Awni Hannun
Deep learning has dramatically improved the performance of speech recognition systems through learning hierarchies of features optimized for the task at hand.
36 code implementations • 8 Dec 2015 • Dario Amodei, Rishita Anubhai, Eric Battenberg, Carl Case, Jared Casper, Bryan Catanzaro, Jingdong Chen, Mike Chrzanowski, Adam Coates, Greg Diamos, Erich Elsen, Jesse Engel, Linxi Fan, Christopher Fougner, Tony Han, Awni Hannun, Billy Jun, Patrick LeGresley, Libby Lin, Sharan Narang, Andrew Ng, Sherjil Ozair, Ryan Prenger, Jonathan Raiman, Sanjeev Satheesh, David Seetapun, Shubho Sengupta, Yi Wang, Zhiqian Wang, Chong Wang, Bo Xiao, Dani Yogatama, Jun Zhan, Zhenyao Zhu
We show that an end-to-end deep learning approach can be used to recognize either English or Mandarin Chinese speech--two vastly different languages.
24 code implementations • 17 Dec 2014 • Awni Hannun, Carl Case, Jared Casper, Bryan Catanzaro, Greg Diamos, Erich Elsen, Ryan Prenger, Sanjeev Satheesh, Shubho Sengupta, Adam Coates, Andrew Y. Ng
We present a state-of-the-art speech recognition system developed using end-to-end deep learning.