Search Results for author: Arianna Bisazza

Found 47 papers, 20 papers with code

Using Confidential Data for Domain Adaptation of Neural Machine Translation

1 code implementation NAACL (PrivateNLP) 2021 Sohyung Kim, Arianna Bisazza, Fatih Turkmen

We study the problem of domain adaptation in Neural Machine Translation (NMT) when domain-specific data cannot be shared due to confidentiality or copyright issues.

Domain Adaptation Machine Translation +2

Evaluating Pre-training Objectives for Low-Resource Translation into Morphologically Rich Languages

no code implementations LREC 2022 Prajit Dhar, Arianna Bisazza, Gertjan van Noord

We conduct our evaluation on four typologically diverse target MRLs, and find that PT-Inflect surpasses NMT systems trained only on parallel data.

Machine Translation NMT +1

Evaluating Text Generation from Discourse Representation Structures

1 code implementation ACL (GEM) 2021 Chunliu Wang, Rik van Noord, Arianna Bisazza, Johan Bos

We present an end-to-end neural approach to generate English sentences from formal meaning representations, Discourse Representation Structures (DRSs).

Text Generation

Encoding of lexical tone in self-supervised models of spoken language

no code implementations25 Mar 2024 Gaofei Shen, Michaela Watkins, Afra Alishahi, Arianna Bisazza, Grzegorz Chrupała

Interpretability research has shown that self-supervised Spoken Language Models (SLMs) encode a wide variety of features in human speech from the acoustic, phonetic, phonological, syntactic and semantic levels, to speaker characteristics.

Cross-Lingual Consistency of Factual Knowledge in Multilingual Language Models

1 code implementation16 Oct 2023 Jirui Qi, Raquel Fernández, Arianna Bisazza

Finally, we conduct a case study on CLC when new factual associations are inserted in the PLMs via model editing.

Model Editing

Quantifying the Plausibility of Context Reliance in Neural Machine Translation

2 code implementations2 Oct 2023 Gabriele Sarti, Grzegorz Chrupała, Malvina Nissim, Arianna Bisazza

Establishing whether language models can use contextual information in a human-plausible way is important to ensure their trustworthiness in real-world settings.

Machine Translation Translation

Wave to Syntax: Probing spoken language models for syntax

1 code implementation30 May 2023 Gaofei Shen, Afra Alishahi, Arianna Bisazza, Grzegorz Chrupała

Understanding which information is encoded in deep models of spoken and written language has been the focus of much research in recent years, as it is crucial for debugging and improving these architectures.

Are Character-level Translations Worth the Wait? Comparing ByT5 and mT5 for Machine Translation

1 code implementation28 Feb 2023 Lukas Edman, Gabriele Sarti, Antonio Toral, Gertjan van Noord, Arianna Bisazza

Pretrained character-level and byte-level language models have been shown to be competitive with popular subword models across a range of Natural Language Processing (NLP) tasks.

Machine Translation NMT +1

Inseq: An Interpretability Toolkit for Sequence Generation Models

2 code implementations27 Feb 2023 Gabriele Sarti, Nils Feldhus, Ludwig Sickert, Oskar van der Wal, Malvina Nissim, Arianna Bisazza

Past work in natural language processing interpretability focused mainly on popular classification tasks while largely overlooking generation settings, partly due to a lack of dedicated tools.

Feature Importance Machine Translation +2

Communication Drives the Emergence of Language Universals in Neural Agents: Evidence from the Word-order/Case-marking Trade-off

1 code implementation30 Jan 2023 Yuchen Lian, Arianna Bisazza, Tessa Verhoef

Artificial learners often behave differently from human learners in the context of neural agent-based simulations of language emergence and change.

DivEMT: Neural Machine Translation Post-Editing Effort Across Typologically Diverse Languages

1 code implementation24 May 2022 Gabriele Sarti, Arianna Bisazza, Ana Guerberof Arenas, Antonio Toral

We publicly release the complete dataset including all collected behavioral data, to foster new research on the translation capabilities of NMT systems for typologically diverse languages.

Machine Translation NMT +1

On the Difficulty of Translating Free-Order Case-Marking Languages

2 code implementations13 Jul 2021 Arianna Bisazza, Ahmet Üstün, Stephan Sportel

Identifying factors that make certain languages harder to model than others is essential to reach language equality in future Natural Language Processing technologies.

Machine Translation NMT +1

The Effect of Efficient Messaging and Input Variability on Neural-Agent Iterated Language Learning

no code implementations EMNLP 2021 Yuchen Lian, Arianna Bisazza, Tessa Verhoef

Natural languages display a trade-off among different strategies to convey syntactic structure, such as word order or inflection.

UDapter: Language Adaptation for Truly Universal Dependency Parsing

1 code implementation EMNLP 2020 Ahmet Üstün, Arianna Bisazza, Gosse Bouma, Gertjan van Noord

The resulting parser, UDapter, outperforms strong monolingual and multilingual baselines on the majority of both high-resource and low-resource (zero-shot) languages, showing the success of the proposed adaptation approach.

Dependency Parsing Transfer Learning

Understanding Cross-Lingual Syntactic Transfer in Multilingual Recurrent Neural Networks

no code implementations NoDaLiDa 2021 Prajit Dhar, Arianna Bisazza

It is now established that modern neural language models can be successfully trained on multiple languages simultaneously without changes to the underlying architecture.

Cross-Lingual Transfer

BERTje: A Dutch BERT Model

2 code implementations19 Dec 2019 Wietse de Vries, Andreas van Cranenburgh, Arianna Bisazza, Tommaso Caselli, Gertjan van Noord, Malvina Nissim

The transformer-based pre-trained language model BERT has helped to improve state-of-the-art performance on many natural language processing (NLP) tasks.

Language Modelling named-entity-recognition +5

Zero-shot Dependency Parsing with Pre-trained Multilingual Sentence Representations

no code implementations WS 2019 Ke Tran, Arianna Bisazza

We investigate whether off-the-shelf deep bidirectional sentence representations trained on a massively multilingual corpus (multilingual BERT) enable the development of an unsupervised universal dependency parser.

Dependency Parsing Sentence +1

The Lazy Encoder: A Fine-Grained Analysis of the Role of Morphology in Neural Machine Translation

no code implementations EMNLP 2018 Arianna Bisazza, Clara Tump

Neural sequence-to-sequence models have proven very effective for machine translation, but at the expense of model interpretability.

Machine Translation NMT +2

The Importance of Being Recurrent for Modeling Hierarchical Structure

1 code implementation EMNLP 2018 Ke Tran, Arianna Bisazza, Christof Monz

Recent work has shown that recurrent neural networks (RNNs) can implicitly capture and exploit hierarchical information when trained to solve common natural language processing tasks such as language modeling (Linzen et al., 2016) and neural machine translation (Shi et al., 2016).

Language Modelling Machine Translation +1

Examining the Tip of the Iceberg: A Data Set for Idiom Translation

1 code implementation LREC 2018 Marzieh Fadaee, Arianna Bisazza, Christof Monz

Neural Machine Translation (NMT) has been widely used in recent years with significant improvements for many language pairs.

Machine Translation NMT +1

Dynamic Data Selection for Neural Machine Translation

1 code implementation EMNLP 2017 Marlies van der Wees, Arianna Bisazza, Christof Monz

Intelligent selection of training data has proven a successful technique to simultaneously increase training efficiency and translation performance for phrase-based machine translation (PBMT).

Machine Translation NMT +1

Learning Topic-Sensitive Word Representations

1 code implementation ACL 2017 Marzieh Fadaee, Arianna Bisazza, Christof Monz

Distributed word representations are widely used for modeling words in NLP tasks.

Measuring the Effect of Conversational Aspects on Machine Translation Quality

no code implementations COLING 2016 Marlies van der Wees, Arianna Bisazza, Christof Monz

Finally, we find that male speakers are harder to translate and use more vulgar language than female speakers, and that vulgarity is often not preserved during translation.

Machine Translation Translation

A Simple but Effective Approach to Improve Arabizi-to-English Statistical Machine Translation

no code implementations WS 2016 Marlies van der Wees, Arianna Bisazza, Christof Monz

A major challenge for statistical machine translation (SMT) of Arabic-to-English user-generated text is the prevalence of text written in Arabizi, or Romanized Arabic.

Translation Transliteration

Neural versus Phrase-Based Machine Translation Quality: a Case Study

no code implementations EMNLP 2016 Luisa Bentivogli, Arianna Bisazza, Mauro Cettolo, Marcello Federico

Within the field of Statistical Machine Translation (SMT), the neural approach (NMT) has recently emerged as the first technology able to challenge the long-standing dominance of phrase-based approaches (PBMT).

Machine Translation NMT +1

Recurrent Memory Networks for Language Modeling

2 code implementations NAACL 2016 Ke Tran, Arianna Bisazza, Christof Monz

In this paper, we propose Recurrent Memory Network (RMN), a novel RNN architecture, that not only amplifies the power of RNN but also facilitates our understanding of its internal functioning and allows us to discover underlying patterns in data.

Language Modelling Sentence +1

A Survey of Word Reordering in Statistical Machine Translation: Computational Models and Language Phenomena

no code implementations17 Feb 2015 Arianna Bisazza, Marcello Federico

Word reordering is one of the most difficult aspects of statistical machine translation (SMT), and an important factor of its quality and efficiency.

Machine Translation Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.