no code implementations • 7 Mar 2024 • Suzanna Sia, David Mueller, Kevin Duh
Self-supervised large language models have demonstrated the ability to perform Machine Translation (MT) via in-context learning, but little is known about where the model performs the task with respect to prompt instructions and demonstration examples.
1 code implementation • 13 Dec 2022 • David Mueller, Nicholas Andrews, Mark Dredze
Learning these models often requires specialized training algorithms that address task-conflict in the shared parameter updates, which otherwise can lead to negative transfer.
no code implementations • EMNLP 2020 • Steven Reich, David Mueller, Nicholas Andrews
However, extending these methods to structured prediction is not always straightforward or effective; furthermore, a held-out calibration set may not always be available.
1 code implementation • ACL 2020 • David Mueller, Nicholas Andrews, Mark Dredze
However, a straightforward implementation of this simple idea does not always work in practice: naive training of NER models using annotated data drawn from multiple languages consistently underperforms models trained on monolingual data alone, despite having access to more training data.
Multilingual Named Entity Recognition named-entity-recognition +2
1 code implementation • EMNLP 2018 • David Mueller, Greg Durrett
To disambiguate between closely related concepts, entity linking systems need to effectively distill cues from their context, which may be quite noisy.