Search Results for author: Peyman Passban

Found 16 papers, 2 papers with code

What is Lost in Knowledge Distillation?

no code implementations7 Nov 2023 Manas Mohanty, Tanya Roosta, Peyman Passban

Deep neural networks (DNNs) have improved NLP tasks significantly, but training and maintaining such networks could be costly.

Knowledge Distillation Model Compression

Training Mixed-Domain Translation Models via Federated Learning

no code implementations NAACL 2022 Peyman Passban, Tanya Roosta, Rahul Gupta, Ankit Chadha, Clement Chung

Training mixed-domain translation models is a complex task that demands tailored architectures and costly data preparation techniques.

Benchmarking Federated Learning +3

Dynamic Position Encoding for Transformers

no code implementations COLING 2022 Joyce Zheng, Mehdi Rezagholizadeh, Peyman Passban

To solve this problem, position embeddings are defined exclusively for each time step to enrich word information.

Machine Translation NMT +1

From Fully Trained to Fully Random Embeddings: Improving Neural Machine Translation with Compact Word Embedding Tables

no code implementations18 Apr 2021 Krtin Kumar, Peyman Passban, Mehdi Rezagholizadeh, Yiu Sing Lau, Qun Liu

Embedding matrices are key components in neural natural language processing (NLP) models that are responsible to provide numerical representations of input tokens.\footnote{In this paper words and subwords are referred to as \textit{tokens} and the term \textit{embedding} only refers to embeddings of inputs.}

Machine Translation NMT +2

Robust Embeddings Via Distributions

no code implementations17 Apr 2021 Kira A. Selby, Yinong Wang, Ruizhe Wang, Peyman Passban, Ahmad Rashid, Mehdi Rezagholizadeh, Pascal Poupart

Despite recent monumental advances in the field, many Natural Language Processing (NLP) models still struggle to perform adequately on noisy domains.

Revisiting Robust Neural Machine Translation: A Transformer Case Study

no code implementations Findings (EMNLP) 2021 Peyman Passban, Puneeth S. M. Saladi, Qun Liu

There is a large body of work in the NMT literature on analyzing the behavior of conventional models for the problem of noise but Transformers are relatively understudied in this context.

Denoising Machine Translation +2

ALP-KD: Attention-Based Layer Projection for Knowledge Distillation

no code implementations27 Dec 2020 Peyman Passban, Yimeng Wu, Mehdi Rezagholizadeh, Qun Liu

Knowledge distillation is considered as a training and compression strategy in which two neural networks, namely a teacher and a student, are coupled together during training.

Knowledge Distillation

Tailoring Neural Architectures for Translating from Morphologically Rich Languages

no code implementations COLING 2018 Peyman Passban, Andy Way, Qun Liu

A morphologically complex word (MCW) is a hierarchical constituent with meaning-preserving subunits, so word-based models which rely on surface forms might not be powerful enough to translate such structures.

Machine Translation NMT +2

Investigating Backtranslation in Neural Machine Translation

no code implementations17 Apr 2018 Alberto Poncelas, Dimitar Shterionov, Andy Way, Gideon Maillette de Buy Wenniger, Peyman Passban

A prerequisite for training corpus-based machine translation (MT) systems -- either Statistical MT (SMT) or Neural MT (NMT) -- is the availability of high-quality parallel data.

Machine Translation NMT +1

Enriching Phrase Tables for Statistical Machine Translation Using Mixed Embeddings

no code implementations COLING 2016 Peyman Passban, Qun Liu, Andy Way

PBSMT engines by default provide four probability scores in phrase tables which are considered as the main set of bilingual features.

Document Classification Machine Translation +3

Cannot find the paper you are looking for? You can Submit a new open access paper.