Search Results for author: Chantal Amrhein

Found 13 papers, 10 papers with code

Machine Translation Meta Evaluation through Translation Accuracy Challenge Sets

1 code implementation29 Jan 2024 Nikita Moghe, Arnisa Fazla, Chantal Amrhein, Tom Kocmi, Mark Steedman, Alexandra Birch, Rico Sennrich, Liane Guillou

We benchmark metric performance, assess their incremental performance over successive campaigns, and measure their sensitivity to a range of linguistic phenomena.

Benchmarking Machine Translation +3

A Benchmark for Evaluating Machine Translation Metrics on Dialects Without Standard Orthography

1 code implementation28 Nov 2023 Noëmi Aepli, Chantal Amrhein, Florian Schottmann, Rico Sennrich

For sensible progress in natural language processing, it is important that we are aware of the limitations of the evaluation metrics we use.

Machine Translation Text Generation

ACES: Translation Accuracy Challenge Sets at WMT 2023

no code implementations2 Nov 2023 Chantal Amrhein, Nikita Moghe, Liane Guillou

We benchmark the performance of segmentlevel metrics submitted to WMT 2023 using the ACES Challenge Set (Amrhein et al., 2022).

Translation World Knowledge

Evaluating the Effectiveness of Natural Language Inference for Hate Speech Detection in Languages with Limited Labeled Data

1 code implementation6 Jun 2023 Janis Goldzycher, Moritz Preisig, Chantal Amrhein, Gerold Schneider

In this paper, we test whether natural language inference (NLI) models which perform well in zero- and few-shot settings can benefit hate speech detection performance in scenarios where only a limited amount of labeled data is available in the target language.

Hate Speech Detection Natural Language Inference

Exploiting Biased Models to De-bias Text: A Gender-Fair Rewriting Model

1 code implementation18 May 2023 Chantal Amrhein, Florian Schottmann, Rico Sennrich, Samuel Läubli

We hypothesise that creating training data in the reverse direction, i. e. starting from gender-fair text, is easier for morphologically complex languages and show that it matches the performance of state-of-the-art rewriting models for English.

Fairness Machine Translation +2

ACES: Translation Accuracy Challenge Sets for Evaluating Machine Translation Metrics

1 code implementation27 Oct 2022 Chantal Amrhein, Nikita Moghe, Liane Guillou

As machine translation (MT) metrics improve their correlation with human judgement every year, it is crucial to understand the limitations of such metrics at the segment level.

Machine Translation Translation +1

Don't Discard Fixed-Window Audio Segmentation in Speech-to-Text Translation

1 code implementation24 Oct 2022 Chantal Amrhein, Barry Haddow

For real-life applications, it is crucial that end-to-end spoken language translation models perform well on continuous audio, without relying on human-supplied segmentation.

Segmentation Speech-to-Text Translation +1

Identifying Weaknesses in Machine Translation Metrics Through Minimum Bayes Risk Decoding: A Case Study for COMET

1 code implementation10 Feb 2022 Chantal Amrhein, Rico Sennrich

Neural metrics have achieved impressive correlation with human judgements in the evaluation of machine translation systems, but before we can safely optimise towards such metrics, we should be aware of (and ideally eliminate) biases toward bad translations that receive high scores.

Machine Translation Translation

How Suitable Are Subword Segmentation Strategies for Translating Non-Concatenative Morphology?

1 code implementation Findings (EMNLP) 2021 Chantal Amrhein, Rico Sennrich

Data-driven subword segmentation has become the default strategy for open-vocabulary machine translation and other NLP tasks, but may not be sufficiently generic for optimal learning of non-concatenative morphology.

Machine Translation Segmentation +1

On Biasing Transformer Attention Towards Monotonicity

1 code implementation NAACL 2021 Annette Rios, Chantal Amrhein, Noëmi Aepli, Rico Sennrich

Many sequence-to-sequence tasks in natural language processing are roughly monotonic in the alignment between source and target sequence, and previous work has facilitated or enforced learning of monotonic attention behavior via specialized attention functions or pretraining.

Morphological Inflection Transliteration

On Romanization for Model Transfer Between Scripts in Neural Machine Translation

no code implementations Findings of the Association for Computational Linguistics 2020 Chantal Amrhein, Rico Sennrich

Our results show that romanization entails information loss and is thus not always superior to simpler vocabulary transfer methods, but can improve the transfer between related languages with different scripts.

Machine Translation Transfer Learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.