Search Results for author: Mitchell A. Gordon

Found 3 papers, 1 papers with code

Distill, Adapt, Distill: Training Small, In-Domain Models for Neural Machine Translation

no code implementations WS 2020 Mitchell A. Gordon, Kevin Duh

We explore best practices for training small, memory efficient machine translation models with sequence-level knowledge distillation in the domain adaptation setting.

Domain Adaptation Knowledge Distillation +2

Cannot find the paper you are looking for? You can Submit a new open access paper.