Search Results for author: José A. R. Fonollosa

Found 23 papers, 10 papers with code

Pushing the Limits of Zero-shot End-to-End Speech Translation

1 code implementation16 Feb 2024 Ioannis Tsiamas, Gerard I. Gállego, José A. R. Fonollosa, Marta R. Costa-jussà

The speech encoder seamlessly integrates with the MT model at inference, enabling direct translation from speech to text, across all languages supported by the MT model.

Speech-to-Text Translation Translation

Promoting Generalized Cross-lingual Question Answering in Few-resource Scenarios via Self-knowledge Distillation

1 code implementation29 Sep 2023 Casimiro Pio Carrino, Carlos Escolano, José A. R. Fonollosa

Our approach seeks to enhance cross-lingual QA transfer using a high-performing multilingual model trained on a large-scale dataset, complemented by a few thousand aligned QA examples across languages.

Cross-Lingual Question Answering Cross-Lingual Transfer +3

Efficient Speech Translation with Dynamic Latent Perceivers

1 code implementation28 Oct 2022 Ioannis Tsiamas, Gerard I. Gállego, José A. R. Fonollosa, Marta R. Costa-jussà

Transformers have been the dominant architecture for Speech Translation in recent years, achieving significant improvements in translation quality.

Speech-to-Text Translation Translation

SHAS: Approaching optimal Segmentation for End-to-End Speech Translation

2 code implementations9 Feb 2022 Ioannis Tsiamas, Gerard I. Gállego, José A. R. Fonollosa, Marta R. Costa-jussà

Speech translation datasets provide manual segmentations of the audios, which are not available in real-world scenarios, and existing segmentation methods usually significantly reduce translation quality at inference time.

Segmentation Speech-to-Text Translation +1

End-to-End Speech Translation with Pre-trained Models and Adapters: UPC at IWSLT 2021

1 code implementation ACL (IWSLT) 2021 Gerard I. Gállego, Ioannis Tsiamas, Carlos Escolano, José A. R. Fonollosa, Marta R. Costa-jussà

Our submission also uses a custom segmentation algorithm that employs pre-trained Wav2Vec 2. 0 for identifying periods of untranscribable text and can bring improvements of 2. 5 to 3 BLEU score on the IWSLT 2019 test set, as compared to the result with the given segmentation.

Ranked #2 on Speech-to-Text Translation on MuST-C EN->DE (using extra training data)

Segmentation Speech-to-Text Translation +1

Training Multilingual Machine Translation by Alternately Freezing Language-Specific Encoders-Decoders

no code implementations29 May 2020 Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa, Mikel Artetxe

We propose a modular architecture of language-specific encoder-decoders that constitutes a multilingual machine translation system that can be incrementally extended to new languages without the need for retraining the existing system when adding new languages.

Machine Translation Natural Language Inference +2

From Bilingual to Multilingual Neural Machine Translation by Incremental Training

no code implementations ACL 2019 Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa

Multilingual Neural Machine Translation approaches are based on the use of task-specific models and the addition of one more language can only be done by retraining the whole system.

Machine Translation Translation

Joint Source-Target Self Attention with Locality Constraints

2 code implementations16 May 2019 José A. R. Fonollosa, Noe Casas, Marta R. Costa-jussà

The dominant neural machine translation models are based on the encoder-decoder structure, and many of them rely on an unconstrained receptive field over source and target sequences.

Language Modelling Machine Translation +1

Towards Interlingua Neural Machine Translation

no code implementations15 May 2019 Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa

By adding and forcing this interlingual loss, we are able to train multiple encoders and decoders for each language, sharing a common intermediate representation.

Machine Translation Translation

(Self-Attentive) Autoencoder-based Universal Language Representation for Machine Translation

no code implementations15 Oct 2018 Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa

Preliminary results on the WMT 2017 Turkish/English task shows that the proposed architecture is capable of learning a universal language representation and simultaneously training both translation directions with state-of-the-art results.

Machine Translation Sentence +1

Conditional distribution variability measures for causality detection

no code implementations25 Jan 2016 José A. R. Fonollosa

In this paper we derive variability measures for the conditional probability distributions of a pair of random variables, and we study its application in the inference of causal-effect relationships.

Cannot find the paper you are looking for? You can Submit a new open access paper.