Search Results for author: Raúl Vázquez

Found 10 papers, 3 papers with code

Multilingual NMT with a language-independent attention bridge

1 code implementation WS 2019 Raúl Vázquez, Alessandro Raganato, Jörg Tiedemann, Mathias Creutz

In this paper, we propose a multilingual encoder-decoder architecture capable of obtaining multilingual sentence representations by means of incorporating an intermediate {\em attention bridge} that is shared across all languages.

NMT Sentence +2

The University of Helsinki submissions to the WMT19 news translation task

no code implementations WS 2019 Aarne Talman, Umut Sulubacak, Raúl Vázquez, Yves Scherrer, Sami Virpioja, Alessandro Raganato, Arvi Hurskainen, Jörg Tiedemann

In this paper, we present the University of Helsinki submissions to the WMT 2019 shared task on news translation in three language pairs: English-German, English-Finnish and Finnish-English.

Sentence Translation

An Empirical Investigation of Word Alignment Supervision for Zero-Shot Multilingual Neural Machine Translation

no code implementations EMNLP 2021 Alessandro Raganato, Raúl Vázquez, Mathias Creutz, Jörg Tiedemann

In this paper, we investigate the benefits of an explicit alignment to language labels in Transformer-based MNMT models in the zero-shot context, by jointly training one cross attention head with word alignment supervision to stress the focus on the target language label.

Machine Translation Translation +1

A Closer Look at Parameter Contributions When Training Neural Language and Translation Models

no code implementations COLING 2022 Raúl Vázquez, Hande Celikkanat, Vinit Ravishankar, Mathias Creutz, Jörg Tiedemann

We analyze the learning dynamics of neural language and translation models using Loss Change Allocation (LCA), an indicator that enables a fine-grained analysis of parameter updates when optimizing for the loss function.

Causal Language Modeling Language Modelling +3

Why bother with geometry? On the relevance of linear decompositions of Transformer embeddings

1 code implementation10 Oct 2023 Timothee Mickus, Raúl Vázquez

A recent body of work has demonstrated that Transformer embeddings can be linearly decomposed into well-defined sums of factors, that can in turn be related to specific network inputs or components.

Machine Translation Sentence

SemEval-2024 Shared Task 6: SHROOM, a Shared-task on Hallucinations and Related Observable Overgeneration Mistakes

no code implementations12 Mar 2024 Timothee Mickus, Elaine Zosa, Raúl Vázquez, Teemu Vahtola, Jörg Tiedemann, Vincent Segonne, Alessandro Raganato, Marianna Apidianaki

This paper presents the results of the SHROOM, a shared task focused on detecting hallucinations: outputs from natural language generation (NLG) systems that are fluent, yet inaccurate.

Machine Translation Paraphrase Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.