Search Results for author: Stig-Arne Grönroos

Found 15 papers, 4 papers with code

Silo NLP’s Participation at WAT2022

no code implementations WAT 2022 Shantipriya Parida, Subhadarshi Panda, Stig-Arne Grönroos, Mark Granroth-Wilding, Mika Koistinen

This paper provides the system description of “Silo NLP’s” submission to the Workshop on Asian Translation (WAT2022).

Translation

The University of Helsinki and Aalto University submissions to the WMT 2020 news and low-resource translation tasks

no code implementations WMT (EMNLP) 2020 Yves Scherrer, Stig-Arne Grönroos, Sami Virpioja

This paper describes the joint participation of University of Helsinki and Aalto University to two shared tasks of WMT 2020: the news translation between Inuktitut and English and the low-resource translation between German and Upper Sorbian.

Multi-Task Learning Translation

Isotropy, Clusters, and Classifiers

no code implementations5 Feb 2024 Timothee Mickus, Stig-Arne Grönroos, Joseph Attieh

Whether embedding spaces use all their dimensions equally, i. e., whether they are isotropic, has been a recent subject of discussion.

Democratizing Neural Machine Translation with OPUS-MT

no code implementations4 Dec 2022 Jörg Tiedemann, Mikko Aulamo, Daria Bakshandaeva, Michele Boggia, Stig-Arne Grönroos, Tommi Nieminen, Alessandro Raganato, Yves Scherrer, Raul Vazquez, Sami Virpioja

This paper presents the OPUS ecosystem with a focus on the development of open machine translation models and tools, and their integration into end-user applications, development platforms and professional workflows.

Machine Translation Translation

Silo NLP's Participation at WAT2022

1 code implementation2 Aug 2022 Shantipriya Parida, Subhadarshi Panda, Stig-Arne Grönroos, Mark Granroth-Wilding, Mika Koistinen

This paper provides the system description of "Silo NLP's" submission to the Workshop on Asian Translation (WAT2022).

Translation

Transfer learning and subword sampling for asymmetric-resource one-to-many neural translation

1 code implementation8 Apr 2020 Stig-Arne Grönroos, Sami Virpioja, Mikko Kurimo

There are several approaches for improving neural machine translation for low-resource languages: Monolingual data can be exploited via pretraining or data augmentation; Parallel corpora on related language pairs can be used via parameter sharing or transfer learning in multilingual models; Subword segmentation and regularization techniques can be applied to ensure high coverage of the vocabulary.

Data Augmentation Denoising +3

Finnish Language Modeling with Deep Transformer Models

no code implementations14 Mar 2020 Abhilash Jain, Aku Ruohe, Stig-Arne Grönroos, Mikko Kurimo

Transformers have recently taken the center stage in language modeling after LSTM's were considered the dominant model architecture for a long time.

Language Modelling

Morfessor EM+Prune: Improved Subword Segmentation with Expectation Maximization and Pruning

1 code implementation LREC 2020 Stig-Arne Grönroos, Sami Virpioja, Mikko Kurimo

Using English, Finnish, North Sami, and Turkish data sets, we show that this approach is able to find better solutions to the optimization problem defined by the Morfessor Baseline model than its original recursive training algorithm.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +4

Multimodal Machine Translation through Visuals and Speech

no code implementations28 Nov 2019 Umut Sulubacak, Ozan Caglayan, Stig-Arne Grönroos, Aku Rouhe, Desmond Elliott, Lucia Specia, Jörg Tiedemann

Multimodal machine translation involves drawing information from more than one modality, based on the assumption that the additional modalities will contain useful alternative views of the input data.

Image Captioning Multimodal Machine Translation +4

Cannot find the paper you are looking for? You can Submit a new open access paper.