Search Results for author: José A. R. Fonollosa

Found 24 papers, 10 papers with code

Language Modelling for Speaker Diarization in Telephonic Interviews

no code implementations28 Jan 2025 Miquel India, Javier Hernando, José A. R. Fonollosa

In this study we analyze how an appropriate fusion of both kind of features is able to obtain good results in these cases.

Acoustic Modelling Language Modelling +4

Pushing the Limits of Zero-shot End-to-End Speech Translation

1 code implementation16 Feb 2024 Ioannis Tsiamas, Gerard I. Gállego, José A. R. Fonollosa, Marta R. Costa-jussà

The speech encoder seamlessly integrates with the MT model at inference, enabling direct translation from speech to text, across all languages supported by the MT model.

Speech-to-Text Translation Translation

Promoting Generalized Cross-lingual Question Answering in Few-resource Scenarios via Self-knowledge Distillation

1 code implementation29 Sep 2023 Casimiro Pio Carrino, Carlos Escolano, José A. R. Fonollosa

Our approach seeks to enhance cross-lingual QA transfer using a high-performing multilingual model trained on a large-scale dataset, complemented by a few thousand aligned QA examples across languages.

Cross-Lingual Question Answering Cross-Lingual Transfer +3

Efficient Speech Translation with Dynamic Latent Perceivers

1 code implementation28 Oct 2022 Ioannis Tsiamas, Gerard I. Gállego, José A. R. Fonollosa, Marta R. Costa-jussà

Transformers have been the dominant architecture for Speech Translation in recent years, achieving significant improvements in translation quality.

Speech-to-Text Translation Translation

SHAS: Approaching optimal Segmentation for End-to-End Speech Translation

2 code implementations9 Feb 2022 Ioannis Tsiamas, Gerard I. Gállego, José A. R. Fonollosa, Marta R. Costa-jussà

Speech translation datasets provide manual segmentations of the audios, which are not available in real-world scenarios, and existing segmentation methods usually significantly reduce translation quality at inference time.

Segmentation Speech-to-Text Translation +1

End-to-End Speech Translation with Pre-trained Models and Adapters: UPC at IWSLT 2021

1 code implementation ACL (IWSLT) 2021 Gerard I. Gállego, Ioannis Tsiamas, Carlos Escolano, José A. R. Fonollosa, Marta R. Costa-jussà

Our submission also uses a custom segmentation algorithm that employs pre-trained Wav2Vec 2. 0 for identifying periods of untranscribable text and can bring improvements of 2. 5 to 3 BLEU score on the IWSLT 2019 test set, as compared to the result with the given segmentation.

Ranked #2 on Speech-to-Text Translation on MuST-C EN->DE (using extra training data)

Decoder Segmentation +2

Training Multilingual Machine Translation by Alternately Freezing Language-Specific Encoders-Decoders

no code implementations29 May 2020 Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa, Mikel Artetxe

We propose a modular architecture of language-specific encoder-decoders that constitutes a multilingual machine translation system that can be incrementally extended to new languages without the need for retraining the existing system when adding new languages.

Decoder Machine Translation +3

From Bilingual to Multilingual Neural Machine Translation by Incremental Training

no code implementations ACL 2019 Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa

Multilingual Neural Machine Translation approaches are based on the use of task-specific models and the addition of one more language can only be done by retraining the whole system.

Decoder Machine Translation +1

Joint Source-Target Self Attention with Locality Constraints

2 code implementations16 May 2019 José A. R. Fonollosa, Noe Casas, Marta R. Costa-jussà

The dominant neural machine translation models are based on the encoder-decoder structure, and many of them rely on an unconstrained receptive field over source and target sequences.

Decoder Language Modeling +3

Towards Interlingua Neural Machine Translation

no code implementations15 May 2019 Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa

By adding and forcing this interlingual loss, we are able to train multiple encoders and decoders for each language, sharing a common intermediate representation.

Machine Translation Translation

(Self-Attentive) Autoencoder-based Universal Language Representation for Machine Translation

no code implementations15 Oct 2018 Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa

Preliminary results on the WMT 2017 Turkish/English task shows that the proposed architecture is capable of learning a universal language representation and simultaneously training both translation directions with state-of-the-art results.

Machine Translation Sentence +1

Conditional distribution variability measures for causality detection

no code implementations25 Jan 2016 José A. R. Fonollosa

In this paper we derive variability measures for the conditional probability distributions of a pair of random variables, and we study its application in the inference of causal-effect relationships.

Cannot find the paper you are looking for? You can Submit a new open access paper.