no code implementations • WMT (EMNLP) 2020 • Jindřich Libovický, Viktor Hangya, Helmut Schmid, Alexander Fraser
We present our systems for the WMT20 Very Low Resource MT Task for translation between German and Upper Sorbian.
no code implementations • Findings (ACL) 2022 • Jindřich Libovický, Helmut Schmid, Alexander Fraser
We present a literature and empirical survey that critically assesses the state of the art in character-level modeling for machine translation (MT).
no code implementations • IWSLT (EMNLP) 2018 • Viktor Hangya, Fabienne Braune, Yuliya Kalasouskaya, Alexander Fraser
We show that our approach is effective, on three language-pairs, without the use of any bilingual signal which is important because parallel sentence mining is most useful in low resource scenarios.
no code implementations • Findings (EMNLP) 2021 • Denis Peskov, Viktor Hangya, Jordan Boyd-Graber, Alexander Fraser
He is associated with founding a company in the United States, so perhaps the German founder Carl Benz could stand in for Gates in those contexts.
no code implementations • EMNLP (MRL) 2021 • Lisa Woller, Viktor Hangya, Alexander Fraser
In contrast to previous approaches which leverage independently pre-trained embeddings of languages, we (i) train CLWEs for the low-resource and a related language jointly and (ii) map them to the target language to build the final multilingual space.
Bilingual Lexicon Induction Cross-Lingual Word Embeddings +1
no code implementations • WMT (EMNLP) 2020 • Alexander Fraser
We describe the WMT 2020 Shared Tasks in Unsupervised MT and Very Low Resource Supervised MT.
no code implementations • WMT (EMNLP) 2021 • Wen Lai, Jindřich Libovický, Alexander Fraser
This paper describes the submission of LMU Munich to the WMT 2021 multilingual machine translation task for small track #1, which studies translation between 6 languages (Croatian, Hungarian, Estonian, Serbian, Macedonian, English) in 30 directions.
no code implementations • WMT (EMNLP) 2021 • Viktor Hangya, Qianchu Liu, Dario Stojanovski, Alexander Fraser, Anna Korhonen
The performance of NMT systems has improved drastically in the past few years but the translation of multi-sense words still poses a challenge.
no code implementations • WMT (EMNLP) 2021 • Jindřich Libovický, Alexander Fraser
We present the findings of the WMT2021 Shared Tasks in Unsupervised MT and Very Low Resource Supervised MT.
no code implementations • WMT (EMNLP) 2021 • Jindřich Libovický, Alexander Fraser
We present our submissions to the WMT21 shared task in Unsupervised and Very Low Resource machine translation between German and Upper Sorbian, German and Lower Sorbian, and Russian and Chuvash.
no code implementations • LREC (BUCC) 2022 • Silvia Severini, Viktor Hangya, Masoud Jalili Sabet, Alexander Fraser, Hinrich Schütze
The two approaches we find most effective are: 1) using identical words as seed lexicons (which unsupervised approaches incorrectly assume are not available for orthographically distinct language pairs) and 2) combining such lexicons with pairs extracted by matching romanized versions of words with an edit distance threshold.
no code implementations • EACL (LTEDI) 2021 • Irina Bigoulaeva, Viktor Hangya, Alexander Fraser
Rather than collecting and annotating new hate speech data, we show how to use cross-lingual transfer learning to leverage already existing data from higher-resource languages.
no code implementations • 28 Oct 2024 • Lukas Edman, Lisa Bylinina, Faeze Ghorbanpour, Alexander Fraser
This paper describes a linguistically-motivated approach to the 2024 edition of the BabyLM Challenge (Warstadt et al. 2023).
1 code implementation • 3 Oct 2024 • Sarah Masud, Sahajpreet Singh, Viktor Hangya, Alexander Fraser, Tanmoy Chakraborty
For subjective tasks such as hate detection, where people perceive hate differently, the Large Language Model's (LLM) ability to represent diverse groups is unclear.
1 code implementation • 1 Oct 2024 • Wen Lai, Viktor Hangya, Alexander Fraser
Text style transfer (TST) aims to modify the style of a text without altering its original meaning.
1 code implementation • 23 Sep 2024 • Lukas Edman, Helmut Schmid, Alexander Fraser
Large Language Models (LLMs) show remarkable performance on a wide variety of tasks.
no code implementations • 3 Jun 2024 • Wen Lai, Mohsen Mesgar, Alexander Fraser
We perform multilingual instruction tuning on the constructed instruction data and further align the LLMs with human feedback using the DPO algorithm on our cross-lingual human feedback dataset.
no code implementations • EMNLP 2015 • Thomas Muller, Ryan Cotterell, Alexander Fraser, Hinrich Schütze
We present LEMMING, a modular log-linear model that jointly models lemmatization and tagging and supports the integration of arbitrary global features.
no code implementations • CONLL 2015 • Ryan Cotterell, Thomas Müller, Alexander Fraser, Hinrich Schütze
We present labeled morphological segmentation, an alternative view of morphological processing that unifies several tasks.
no code implementations • 9 Apr 2024 • Katharina Hämmerl, Jindřich Libovický, Alexander Fraser
Cross-lingual alignment, the meaningful similarity of representations across languages in multilingual language models, has been an active field of research in recent years.
1 code implementation • 29 Jan 2024 • Felix Friedrich, Katharina Hämmerl, Patrick Schramowski, Manuel Brack, Jindrich Libovicky, Kristian Kersting, Alexander Fraser
Our results show that not only do models exhibit strong gender biases but they also behave differently across languages.
no code implementations • 21 Nov 2023 • Viktor Hangya, Silvia Severini, Radoslav Ralev, Alexander Fraser, Hinrich Schütze
In this paper, we propose to build multilingual word embeddings (MWEs) via a novel language chain-based approach, that incorporates intermediate related languages to bridge the gap between the distant source and target.
no code implementations • 14 Nov 2023 • Wen Lai, Viktor Hangya, Alexander Fraser
Despite the growing variety of languages supported by existing multilingual neural machine translation (MNMT) models, most of the world's languages are still being left behind.
1 code implementation • 1 Jun 2023 • Katharina Hämmerl, Alina Fastowski, Jindřich Libovický, Alexander Fraser
We investigate outlier dimensions and their relationship to anisotropy in multiple pre-trained multilingual language models.
1 code implementation • 26 May 2023 • Yihong Liu, Alexandra Chronopoulou, Hinrich Schütze, Alexander Fraser
By conducting extensive experiments on different language pairs, including similar and distant, high and low-resource languages, we find that our method alleviates the copying problem, thus improving the translation performance on low-resource languages.
no code implementations • 23 May 2023 • Viktor Hangya, Alexander Fraser
Our experiments show that using already existing datasets and only a few-shots of the target task the performance of models improve both monolingually and across languages.
1 code implementation • 22 May 2023 • Wen Lai, Alexandra Chronopoulou, Alexander Fraser
Despite advances in multilingual neural machine translation (MNMT), we argue that there are still two major challenges in this area: data imbalance and representation degeneration.
no code implementations • 14 Feb 2023 • Alexandra Chronopoulou, Matthew E. Peters, Alexander Fraser, Jesse Dodge
We also explore weight averaging of adapters trained on the same domain with different hyper-parameters, and show that it preserves the performance of a PLM on new domains while obtaining strong in-domain results.
1 code implementation • 14 Nov 2022 • Katharina Hämmerl, Björn Deiseroth, Patrick Schramowski, Jindřich Libovický, Constantin A. Rothkopf, Alexander Fraser, Kristian Kersting
Do the models capture moral norms from English and impose them on other languages?
1 code implementation • 21 Oct 2022 • Wen Lai, Alexandra Chronopoulou, Alexander Fraser
We consider a very challenging scenario: adapting the MNMT model both to a new domain and to a new language pair at the same time.
no code implementations • 10 Oct 2022 • Sophie Henning, William Beluch, Alexander Fraser, Annemarie Friedrich
With this survey, the first overview on class imbalance in deep-learning based NLP, we provide guidance for NLP researchers and practitioners dealing with imbalanced data.
no code implementations • 30 Sep 2022 • Alexandra Chronopoulou, Dario Stojanovski, Alexander Fraser
Training a new adapter on each language pair or training a single adapter on all language pairs without updating the pretrained model has been proposed as a parameter-efficient alternative.
no code implementations • 31 May 2022 • Silvia Severini, Viktor Hangya, Masoud Jalili Sabet, Alexander Fraser, Hinrich Schütze
The two approaches we find most effective are: 1) using identical words as seed lexicons (which unsupervised approaches incorrectly assume are not available for orthographically distinct language pairs) and 2) combining such lexicons with pairs extracted by matching romanized versions of words with an edit distance threshold.
no code implementations • 26 Mar 2022 • Marius Gassen, Benjamin Hättasch, Benjamin Hilprecht, Nadja Geisler, Alexander Fraser, Carsten Binnig
However, developing a conversational agent (i. e., a chatbot-like interface) to allow end-users to interact with an application using natural language requires both immense amounts of training data and NLP expertise.
no code implementations • 25 Mar 2022 • Marion Weller-Di Marco, Matthias Huck, Alexander Fraser
Key challenges of rich target-side morphology in data-driven machine translation include: (1) A large amount of differently inflected word surface forms entails a larger vocabulary and thus data sparsity.
no code implementations • 18 Mar 2022 • Katharina Hämmerl, Björn Deiseroth, Patrick Schramowski, Jindřich Libovický, Alexander Fraser, Kristian Kersting
Massively multilingual sentence representations are trained on large corpora of uncurated data, with a very imbalanced proportion of languages included in the training.
1 code implementation • Findings (ACL) 2022 • Katharina Hämmerl, Jindřich Libovický, Alexander Fraser
We combine the strengths of static and contextual models to improve multilingual representations.
no code implementations • 15 Jan 2022 • Irina Bigoulaeva, Viktor Hangya, Iryna Gurevych, Alexander Fraser
The goal of hate speech detection is to filter negative online content aiming at certain groups of people.
1 code implementation • COLING 2022 • Wen Lai, Jindřich Libovický, Alexander Fraser
First, we want to reach domain robustness, i. e., we want to reach high quality on both domains seen in the training data and unseen domains.
no code implementations • 15 Oct 2021 • Jindřich Libovický, Helmut Schmid, Alexander Fraser
We present a literature and empirical survey that critically assesses the state of the art in character-level modeling for machine translation (MT).
1 code implementation • spnlp (ACL) 2022 • Jindřich Libovický, Alexander Fraser
We propose the neural string edit distance model for string-pair matching and string transduction based on learnable string edit distance.
1 code implementation • NAACL 2021 • Alexandra Chronopoulou, Dario Stojanovski, Alexander Fraser
Successful methods for unsupervised neural machine translation (UNMT) employ crosslingual pretraining via self-supervision, often in the form of a masked language modeling or a sequence generation task, which requires the model to align the lexical- and high-level representations of the two languages.
no code implementations • COLING 2020 • Silvia Severini, Viktor Hangya, Alexander Fraser, Hinrich Sch{\"u}tze
In this paper, we enrich BWE-based BDI with transliteration information by using Bilingual Orthography Embeddings (BOEs).
no code implementations • COLING 2020 • Dario Stojanovski, Benno Krojer, Denis Peskov, Alexander Fraser
Recent high scores on pronoun translation using context-aware neural machine translation have suggested that current approaches work well.
1 code implementation • WMT (EMNLP) 2020 • Alexandra Chronopoulou, Dario Stojanovski, Viktor Hangya, Alexander Fraser
Our core unsupervised neural machine translation (UNMT) system follows the strategy of Chronopoulou et al. (2020), using a monolingual pretrained language generation model (on German) and fine-tuning it on both German and Upper Sorbian, before initializing a UNMT model, which is trained with online backtranslation.
no code implementations • ACL 2021 • Tobias Eder, Viktor Hangya, Alexander Fraser
For low resource languages training MWEs monolingually results in MWEs of poor quality, and thus poor bilingual word embeddings (BWEs) as well.
1 code implementation • EMNLP 2020 • Alexandra Chronopoulou, Dario Stojanovski, Alexander Fraser
Using a language model (LM) pretrained on two languages with large monolingual data in order to initialize an unsupervised neural machine translation (UNMT) system yields state-of-the-art results.
no code implementations • 10 Jul 2020 • Anita Ramm, Ekaterina Lapshinova-Koltunski, Alexander Fraser
Grammatical tense and mood are important linguistic phenomena to consider in natural language processing (NLP) research.
no code implementations • EACL (AdaptNLP) 2021 • Dario Stojanovski, Alexander Fraser
Achieving satisfying performance in machine translation on domains for which there is no training data is challenging.
2 code implementations • EMNLP 2020 • Jindřich Libovický, Alexander Fraser
Applying the Transformer architecture on the character level usually requires very deep architectures that are difficult and slow to train.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Jindřich Libovický, Rudolf Rosa, Alexander Fraser
Multilingual contextual embeddings, such as multilingual BERT and XLM-RoBERTa, have proved useful for many multi-lingual tasks.
1 code implementation • 8 Nov 2019 • Jindřich Libovický, Rudolf Rosa, Alexander Fraser
Multilingual BERT (mBERT) provides sentence representations for 104 languages, which are useful for many multi-lingual tasks.
no code implementations • ACL 2018 • Philipp Dufter, Mengjie Zhao, Martin Schmitt, Alexander Fraser, Hinrich Schütze
We present a new method for estimating vector space representations of words: embedding learning by concept induction.
no code implementations • WS 2017 • Aleš Tamchyna, Marion Weller-Di Marco, Alexander Fraser
NMT systems have problems with large vocabulary sizes.
no code implementations • ACL 2016 • Aleš Tamchyna, Alexander Fraser, Ondřej Bojar, Marcin Junczys-Dowmunt
Discriminative translation models utilizing source context have been shown to help statistical machine translation performance.