no code implementations • WAT 2022 • Alberto Poncelas, Johanes Effendi, Ohnmar Htun, Sunil Yadav, Dongzhe Wang, Saurabh Jain
This paper introduces our neural machine translation system’s participation in the WAT 2022 shared translation task (team ID: sakura).
no code implementations • ACL (WAT) 2021 • Raymond Hendy Susanto, Dongzhe Wang, Sunil Yadav, Mausam Jain, Ohnmar Htun
This paper introduces our neural machine translation systems’ participation in the WAT 2021 shared translation tasks (team ID: sakura).
no code implementations • AACL (WAT) 2020 • Dongzhe Wang, Ohnmar Htun
This paper introduces our neural machine translation systems’ participation in the WAT 2020 (team ID: goku20).
1 code implementation • COLING 2020 • Yuxi Xie, Liangming Pan, Dongzhe Wang, Min-Yen Kan, Yansong Feng
Recent question generation (QG) approaches often utilize the sequence-to-sequence framework (Seq2Seq) to optimize the log-likelihood of ground-truth questions using teacher forcing.