no code implementations • 12 Jun 2017 • Baskaran Sankaran, Markus Freitag, Yaser Al-Onaizan
Usually, the candidate lists are a combination of external word-to-word aligner, phrase table entries or most frequent words.
no code implementations • 6 Feb 2017 • Markus Freitag, Yaser Al-Onaizan, Baskaran Sankaran
Knowledge distillation describes a method for training a student network to perform better by learning from a stronger teacher network.
no code implementations • 9 Aug 2016 • Baskaran Sankaran, Haitao Mi, Yaser Al-Onaizan, Abe Ittycheriah
Attention-based Neural Machine Translation (NMT) models suffer from attention deficiency issues as has been observed in recent research.
no code implementations • EMNLP 2016 • Orhan Firat, Baskaran Sankaran, Yaser Al-Onaizan, Fatos T. Yarman Vural, Kyunghyun Cho
In this paper, we propose a novel finetuning algorithm for the recently introduced multi-way, mulitlingual neural machine translate that enables zero-resource machine translation.
no code implementations • EMNLP 2016 • Haitao Mi, Baskaran Sankaran, Zhiguo Wang, Abe Ittycheriah
In this paper, we enhance the attention-based neural machine translation (NMT) by adding explicit coverage embedding models to alleviate issues of repeating and dropping translations in NMT.