Search Results for author: Alireza Mohammadshahi

Found 13 papers, 11 papers with code

Leeroo Orchestrator: Elevating LLMs Performance Through Model Integration

1 code implementation25 Jan 2024 Alireza Mohammadshahi, Ali Shaikh, Majid Yazdani

In this paper, we propose an architecture to harness the collective knowledge of multiple trained LLMs to create a new state-of-the-art.

Ranked #4 on Multi-task Language Understanding on MMLU (using extra training data)

Multi-task Language Understanding

Investigating Multi-Pivot Ensembling with Massively Multilingual Machine Translation Models

1 code implementation13 Nov 2023 Alireza Mohammadshahi, Jannis Vamvas, Rico Sennrich

Massively multilingual machine translation models allow for the translation of a large number of languages with a single model, but have limited performance on low- and very-low-resource translation directions.

Hallucination Machine Translation +1

Transformers as Graph-to-Graph Models

1 code implementation27 Oct 2023 James Henderson, Alireza Mohammadshahi, Andrei C. Coman, Lesly Miculicich

We argue that Transformers are essentially graph-to-graph models, with sequences just being a special case.

Mitigating Hallucinations and Off-target Machine Translation with Source-Contrastive and Language-Contrastive Decoding

1 code implementation13 Sep 2023 Rico Sennrich, Jannis Vamvas, Alireza Mohammadshahi

Experiments on the massively multilingual models M2M-100 (418M) and SMaLL-100 show that these methods suppress hallucinations and off-target translations, reducing the number of translations with segment-level chrF2 below 10 by 67-83% on average, and the number of translations with oscillatory hallucinations by 75-92% on average, across 57 tested translation directions.

Machine Translation Translation

SMaLL-100: Introducing Shallow Multilingual Machine Translation Model for Low-Resource Languages

3 code implementations20 Oct 2022 Alireza Mohammadshahi, Vassilina Nikoulina, Alexandre Berard, Caroline Brun, James Henderson, Laurent Besacier

In recent years, multilingual machine translation models have achieved promising performance on low-resource language pairs by sharing information between similar languages, thus enabling zero-shot translation.

Machine Translation Translation

What Do Compressed Multilingual Machine Translation Models Forget?

1 code implementation22 May 2022 Alireza Mohammadshahi, Vassilina Nikoulina, Alexandre Berard, Caroline Brun, James Henderson, Laurent Besacier

In this work, we assess the impact of compression methods on Multilingual Neural Machine Translation models (MNMT) for various language groups, gender, and semantic biases by extensive analysis of compressed models on different machine translation benchmarks, i. e. FLORES-101, MT-Gender, and DiBiMT.

Machine Translation Memorization +1

The DCU-EPFL Enhanced Dependency Parser at the IWPT 2021 Shared Task

1 code implementation ACL (IWPT) 2021 James Barry, Alireza Mohammadshahi, Joachim Wagner, Jennifer Foster, James Henderson

The task involves parsing Enhanced UD graphs, which are an extension of the basic dependency trees designed to be more facilitative towards representing semantic structure.

valid

Syntax-Aware Graph-to-Graph Transformer for Semantic Role Labelling

no code implementations15 Apr 2021 Alireza Mohammadshahi, James Henderson

Recent models have shown that incorporating syntactic knowledge into the semantic role labelling (SRL) task leads to a significant improvement.

Ranked #6 on Semantic Role Labeling on CoNLL 2005 (using extra training data)

Semantic Role Labeling

Recursive Non-Autoregressive Graph-to-Graph Transformer for Dependency Parsing with Iterative Refinement

1 code implementation29 Mar 2020 Alireza Mohammadshahi, James Henderson

We propose the Recursive Non-autoregressive Graph-to-Graph Transformer architecture (RNGTr) for the iterative refinement of arbitrary graphs through the recursive application of a non-autoregressive Graph-to-Graph Transformer and apply it to syntactic dependency parsing.

Dependency Parsing

Graph-to-Graph Transformer for Transition-based Dependency Parsing

1 code implementation Findings of the Association for Computational Linguistics 2020 Alireza Mohammadshahi, James Henderson

We propose the Graph2Graph Transformer architecture for conditioning on and predicting arbitrary graphs, and apply it to the challenging task of transition-based dependency parsing.

Structured Prediction Transition-Based Dependency Parsing

Aligning Multilingual Word Embeddings for Cross-Modal Retrieval Task

1 code implementation EMNLP (WS) 2019 Alireza Mohammadshahi, Remi Lebret, Karl Aberer

In this paper, we propose a new approach to learn multimodal multilingual embeddings for matching images and their relevant captions in two languages.

Cross-Modal Retrieval Image-to-Text Retrieval +3

Cannot find the paper you are looking for? You can Submit a new open access paper.