Search Results for author: Ben Peters

Found 9 papers, 4 papers with code

Did Translation Models Get More Robust Without Anyone Even Noticing?

no code implementations6 Mar 2024 Ben Peters, André F. T. Martins

Neural machine translation (MT) models achieve strong results across a variety of settings, but it is widely believed that they are highly sensitive to "noisy" inputs, such as spelling errors, abbreviations, and other formatting issues.

Machine Translation Translation

Tower: An Open Multilingual Large Language Model for Translation-Related Tasks

1 code implementation27 Feb 2024 Duarte M. Alves, José Pombal, Nuno M. Guerreiro, Pedro H. Martins, João Alves, Amin Farajian, Ben Peters, Ricardo Rei, Patrick Fernandes, Sweta Agrawal, Pierre Colombo, José G. C. de Souza, André F. T. Martins

While general-purpose large language models (LLMs) demonstrate proficiency on multiple tasks within the domain of translation, approaches based on open LLMs are competitive only when specializing on a single task.

Language Modelling Large Language Model +1

Smoothing and Shrinking the Sparse Seq2Seq Search Space

1 code implementation NAACL 2021 Ben Peters, André F. T. Martins

Current sequence-to-sequence models are trained to minimize cross-entropy and use softmax to compute the locally normalized probabilities over target sequences.

Machine Translation Morphological Inflection +1

One-Size-Fits-All Multilingual Models

no code implementations WS 2020 Ben Peters, Andr{\'e} F. T. Martins

For both tasks, we present multilingual models, training jointly on data in all languages.

LEMMA

IT--IST at the SIGMORPHON 2019 Shared Task: Sparse Two-headed Models for Inflection

no code implementations WS 2019 Ben Peters, Andr{\'e} F. T. Martins

This paper presents the Instituto de Telecomunica{\c{c}}{\~o}es{--}Instituto Superior T{\'e}cnico submission to Task 1 of the SIGMORPHON 2019 Shared Task.

LEMMA

Interpretable Structure Induction via Sparse Attention

no code implementations WS 2018 Ben Peters, Vlad Niculae, Andr{\'e} F. T. Martins

Neural network methods are experiencing wide adoption in NLP, thanks to their empirical performance on many tasks.

Cannot find the paper you are looking for? You can Submit a new open access paper.