1 code implementation • EAMT 2020 • Felix Hieber, Tobias Domhan, Michael Denkowski, David Vilar
We present Sockeye 2, a modernized and streamlined version of the Sockeye neural machine translation (NMT) toolkit.
no code implementations • Findings (ACL) 2022 • Markus Freitag, David Vilar, David Grangier, Colin Cherry, George Foster
In this work we propose a method for training MT systems to achieve a more natural style, i. e. mirroring the style of text originally written in the target language.
no code implementations • Findings (EMNLP) 2021 • Julia Kreutzer, David Vilar, Artem Sokolov
Training data for machine translation (MT) is often sourced from a multitude of large corpora that are multi-faceted in nature, e. g. containing contents from multiple domains or different levels of quality or complexity.
1 code implementation • ACL (IWSLT) 2021 • David Vilar, Marcello Federico
Sub-word segmentation is currently a standard tool for training neural machine translation (MT) systems and other NLP tasks.
no code implementations • EMNLP 2021 • Andrea Schioppa, David Vilar, Artem Sokolov, Katja Filippova
Fine-grained control of machine translation (MT) outputs along multiple attributes is critical for many modern MT applications and is a requirement for gaining users’ trust.
no code implementations • 16 Nov 2022 • David Vilar, Markus Freitag, Colin Cherry, Jiaming Luo, Viresh Ratnakar, George Foster
Large language models (LLMs) that have been trained on multilingual but not parallel text exhibit a remarkable ability to translate between languages.
1 code implementation • 6 Dec 2021 • Andrea Schioppa, Polina Zablotskaia, David Vilar, Artem Sokolov
We address efficient calculation of influence functions for tracking predictions back to the training data.
no code implementations • 13 Oct 2021 • Julia Kreutzer, David Vilar, Artem Sokolov
Training data for machine translation (MT) is often sourced from a multitude of large corpora that are multi-faceted in nature, e. g. containing contents from multiple domains or different levels of quality or complexity.
1 code implementation • AMTA 2020 • Tobias Domhan, Michael Denkowski, David Vilar, Xing Niu, Felix Hieber, Kenneth Heafield
We present Sockeye 2, a modernized and streamlined version of the Sockeye neural machine translation (NMT) toolkit.
no code implementations • NAACL 2018 • David Vilar
In this paper we explore the use of Learning Hidden Unit Contribution for the task of neural machine translation.
no code implementations • NAACL 2018 • Matt Post, David Vilar
The end-to-end nature of neural machine translation (NMT) removes many ways of manually guiding the translation process that were available in older paradigms.
16 code implementations • 15 Dec 2017 • Felix Hieber, Tobias Domhan, Michael Denkowski, David Vilar, Artem Sokolov, Ann Clifton, Matt Post
Written in Python and built on MXNet, the toolkit offers scalable training and inference for the three most prominent encoder-decoder architectures: attentional recurrent neural networks, self-attentional transformers, and fully convolutional networks.
no code implementations • LREC 2014 • Eleftherios Avramidis, Aljoscha Burchardt, Sabine Hunsicker, Maja Popovi{\'c}, Cindy Tscherwinka, David Vilar, Hans Uszkoreit
Human translators are the key to evaluating machine translation (MT) quality and also to addressing the so far unanswered question when and how to use MT in professional translation workflows.
no code implementations • LREC 2012 • Eleftherios Avramidis, Aljoscha Burchardt, Christian Federmann, Maja Popovi{\'c}, Cindy Tscherwinka, David Vilar
Significant breakthroughs in machine translation only seem possible if human translators are taken into the loop.