no code implementations • ICLR 2020 • Raden Mu'az Mun'im, Jie Lin, Vijay Chandrasekhar, Koichi Shinoda
(4) Fast, it is observed that the number of training epochs required by MaskConvNet is close to training a baseline without pruning.
no code implementations • 12 Nov 2018 • Raden Mu'az Mun'im, Nakamasa Inoue, Koichi Shinoda
We investigate the feasibility of sequence-level knowledge distillation of Sequence-to-Sequence (Seq2Seq) models for Large Vocabulary Continuous Speech Recognition (LVSCR).