Modular Speech-to-Text Translation for Zero-Shot Cross-Modal Transfer

5 Oct 2023  ·  Paul-Ambroise Duquenne, Holger Schwenk, Benoît Sagot ·

Recent research has shown that independently trained encoders and decoders, combined through a shared fixed-size representation, can achieve competitive performance in speech-to-text translation. In this work, we show that this type of approach can be further improved with multilingual training. We observe significant improvements in zero-shot cross-modal speech translation, even outperforming a supervised approach based on XLSR for several languages.

PDF Abstract

Datasets


Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods