1 code implementation • 4 Oct 2024 • Haoran Xu, Kenton Murray, Philipp Koehn, Hieu Hoang, Akiko Eriguchi, Huda Khayrallah
In this paper, we prioritize quality over scaling number of languages, with a focus on multilingual machine translation task, and introduce X-ALMA, a model designed with a commitment to ensuring top-tier performance across 50 diverse languages, regardless of their resource levels.
no code implementations • 14 Nov 2023 • Hieu Hoang, Huda Khayrallah, Marcin Junczys-Dowmunt
We propose the on-the-fly ensembling of a machine translation model with an LLM, prompted on the same task and input.
no code implementations • 16 Mar 2022 • Viet-Khoa Vo-Ho, Kashu Yamazaki, Hieu Hoang, Minh-Triet Tran, Ngan Le
To address such limitations, meta-learning has been adopted in the scenarios of few-shot learning and multiple tasks.
no code implementations • 9 Nov 2021 • Thanh Nguyen, Hieu Hoang, Chang D. Yoo
Single Image Super-Resolution (SISR) is a very active research field.
2 code implementations • ACL 2020 • Marta Ba{\~n}{\'o}n, Pin-zhen Chen, Barry Haddow, Kenneth Heafield, Hieu Hoang, Miquel Espl{\`a}-Gomis, Mikel L. Forcada, Amir Kamran, Faheem Kirefu, Philipp Koehn, Sergio Ortiz Rojas, Leopoldo Pla Sempere, Gema Ram{\'\i}rez-S{\'a}nchez, Elsa Sarr{\'\i}as, Marek Strelec, Brian Thompson, William Waites, Dion Wiggins, Jaume Zaragoza
We report on methods to create the largest publicly available parallel corpora by crawling the web, using open source software.
no code implementations • WS 2018 • Marcin Junczys-Dowmunt, Kenneth Heafield, Hieu Hoang, Roman Grundkiewicz, Anthony Aue
This paper describes the submissions of the "Marian" team to the WNMT 2018 shared task.
no code implementations • WS 2018 • Hieu Hoang, Tomasz Dwojak, Rihards Krislauks, Daniel Torregrosa, Kenneth Heafield
This paper describes the submissions to the efficiency track for GPUs at the Workshop for Neural Machine Translation and Generation by members of the University of Edinburgh, Adam Mickiewicz University, Tilde and University of Alicante.
no code implementations • 5 May 2018 • Robert Lim, Kenneth Heafield, Hieu Hoang, Mark Briers, Allen Malony
Neural machine translation (NMT) has been accelerated by deep learning neural networks over statistical-based approaches, due to the plethora and programmability of commodity heterogeneous computing architectures such as FPGAs and GPUs and the massive amount of training corpuses generated from news outlets, government agencies and social media.
3 code implementations • ACL 2018 • Marcin Junczys-Dowmunt, Roman Grundkiewicz, Tomasz Dwojak, Hieu Hoang, Kenneth Heafield, Tom Neckermann, Frank Seide, Ulrich Germann, Alham Fikri Aji, Nikolay Bogoychev, André F. T. Martins, Alexandra Birch
We present Marian, an efficient and self-contained Neural Machine Translation framework with an integrated automatic differentiation engine based on dynamic computation graphs.
no code implementations • EACL 2017 • Nizar Habash, Nasser Zalmout, Dima Taji, Hieu Hoang, Maverick Alzate
We present Arab-Acquis, a large publicly available dataset for evaluating machine translation between 22 European languages and Arabic.
no code implementations • AMTA 2016 • Hieu Hoang, Nikolay Bogoychev, Lane Schwartz, Marcin Junczys-Dowmunt
The utilization of statistical machine translation (SMT) has grown enormously over the last decade, many using open-source software developed by the NLP community.
2 code implementations • IWSLT 2016 • Marcin Junczys-Dowmunt, Tomasz Dwojak, Hieu Hoang
In this paper we provide the largest published comparison of translation quality for phrase-based SMT and neural machine translation across 30 translation directions.