no code implementations • EMNLP (Eval4NLP) 2020 • Jacob Bremerman, Huda Khayrallah, Douglas Oard, Matt Post
The first and principal contribution is an evaluation measure that characterizes the translation quality of an entire n-best list by asking whether many of the valid translations are placed near the top of the list.
1 code implementation • 12 Oct 2024 • Hyojung Han, Akiko Eriguchi, Haoran Xu, Hieu Hoang, Marine Carpuat, Huda Khayrallah
We propose VocADT, a novel method for vocabulary adaptation using adapter modules that are trained to learn the optimal linear combination of existing embeddings while keeping the model's weights fixed.
no code implementations • 4 Oct 2024 • Haoran Xu, Kenton Murray, Philipp Koehn, Hieu Hoang, Akiko Eriguchi, Huda Khayrallah
In this paper, we prioritize quality over scaling number of languages, with a focus on multilingual machine translation task, and introduce X-ALMA, a model designed with a commitment to ensuring top-tier performance across 50 diverse languages, regardless of their resource levels.
no code implementations • 15 Sep 2024 • Brian Thompson, Nitika Mathur, Daniel Deutsch, Huda Khayrallah
Selecting an automatic metric that best emulates human annotators is often non-trivial, because there is no clear definition of "best emulates."
no code implementations • 14 Nov 2023 • Hieu Hoang, Huda Khayrallah, Marcin Junczys-Dowmunt
We propose the on-the-fly ensembling of a machine translation model with an LLM, prompted on the same task and input.
1 code implementation • 14 Aug 2023 • Matt Post, Thamme Gowda, Roman Grundkiewicz, Huda Khayrallah, Rohit Jain, Marcin Junczys-Dowmunt
Many machine translation toolkits make use of a data preparation step wherein raw data is transformed into a tensor format that can be used directly by the trainer.
no code implementations • 23 May 2023 • Huda Khayrallah, Zuhaib Akhtar, Edward Cohen, Jyothir S V, João Sedoc
We release MMSMR, a Massively Multi-System MultiReference dataset to enable future work on metrics and evaluation for dialog.
1 code implementation • AMTA 2022 • Weiting Tan, Shuoyang Ding, Huda Khayrallah, Philipp Koehn
Neural Machine Translation (NMT) models are known to suffer from noisy inputs.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Huda Khayrallah, João Sedoc
Non-task-oriented dialog models suffer from poor quality and non-diverse responses.
no code implementations • NAACL 2021 • Huda Khayrallah, João Sedoc
We consider the intrinsic evaluation of neural generative dialog models through the lens of Grice's Maxims of Conversation (1975).
no code implementations • WS 2020 • Huda Khayrallah, Jacob Bremerman, Arya D. McCarthy, Kenton Murray, Winston Wu, Matt Post
This paper presents the Johns Hopkins University submission to the 2020 Duolingo Shared Task on Simultaneous Translation and Paraphrase for Language Education (STAPLE).
1 code implementation • EMNLP 2020 • Huda Khayrallah, Brian Thompson, Matt Post, Philipp Koehn
Many valid translations exist for a given sentence, yet machine translation (MT) is trained with a single reference translation, exacerbating data sparsity in low-resource settings.
no code implementations • IJCNLP 2019 • Brian Thompson, Rebecca Knowles, Xuan Zhang, Huda Khayrallah, Kevin Duh, Philipp Koehn
Bilingual lexicons are valuable resources used by professional human translators.
no code implementations • NAACL 2019 • Brian Thompson, Jeremy Gwinnup, Huda Khayrallah, Kevin Duh, Philipp Koehn
Continued training is an effective method for domain adaptation in neural machine translation.
1 code implementation • NAACL 2019 • J. Edward Hu, Huda Khayrallah, Ryan Culkin, Patrick Xia, Tongfei Chen, Matt Post, Benjamin Van Durme
Lexically-constrained sequence decoding allows for explicit positive or negative phrase-based constraints to be placed on target output strings in generation tasks such as machine translation or monolingual text rewriting.
1 code implementation • 2 Nov 2018 • Xuan Zhang, Gaurav Kumar, Huda Khayrallah, Kenton Murray, Jeremy Gwinnup, Marianna J. Martindale, Paul McNamee, Kevin Duh, Marine Carpuat
Machine translation systems based on deep neural networks are expensive to train.
no code implementations • WS 2018 • Huda Khayrallah, Hainan Xu, Philipp Koehn
This work describes our submission to the WMT18 Parallel Corpus Filtering shared task.
no code implementations • WS 2018 • Philipp Koehn, Huda Khayrallah, Kenneth Heafield, Mikel L. Forcada
We posed the shared task of assigning sentence-level quality scores for a very noisy corpus of sentence pairs crawled from the web, with the goal of sub-selecting 1{\%} and 10{\%} of high-quality data to be used to train machine translation systems.
1 code implementation • WS 2018 • Brian Thompson, Huda Khayrallah, Antonios Anastasopoulos, Arya D. McCarthy, Kevin Duh, Rebecca Marvin, Paul McNamee, Jeremy Gwinnup, Tim Anderson, Philipp Koehn
To better understand the effectiveness of continued training, we analyze the major components of a neural machine translation system (the encoder, decoder, and each embedding space) and consider each component's contribution to, and capacity for, domain adaptation.
1 code implementation • WS 2018 • Huda Khayrallah, Brian Thompson, Kevin Duh, Philipp Koehn
Supervised domain adaptation{---}where a large generic corpus and a smaller in-domain corpus are both available for training{---}is a challenge for neural machine translation (NMT).
1 code implementation • WS 2018 • Huda Khayrallah, Philipp Koehn
We examine how various types of noise in the parallel training data impact the quality of neural machine translation systems.
no code implementations • IJCNLP 2017 • Huda Khayrallah, Gaurav Kumar, Kevin Duh, Matt Post, Philipp Koehn
Domain adaptation is a major challenge for neural machine translation (NMT).
no code implementations • EMNLP 2017 • Ryan Cotterell, Ekaterina Vylomova, Huda Khayrallah, Christo Kirov, David Yarowsky
The generation of complex derived word forms has been an overlooked problem in NLP; we fill this gap by applying neural sequence-to-sequence models to the task.
3 code implementations • WS 2019 • Adrian Benton, Huda Khayrallah, Biman Gujral, Dee Ann Reisinger, Sheng Zhang, Raman Arora
We present Deep Generalized Canonical Correlation Analysis (DGCCA) -- a method for learning nonlinear transformations of arbitrarily many views of data, such that the resulting transformations are maximally informative of each other.