no code implementations • 17 Apr 2024 • Shaomu Tan, Di wu, Christof Monz
Training a unified multilingual model promotes knowledge transfer but inevitably introduces negative interference.
1 code implementation • 22 Jan 2024 • Di wu, Shaomu Tan, Yan Meng, David Stap, Christof Monz
Zero-shot translation aims to translate between language pairs not seen during training in Multilingual Machine Translation (MMT) and is largely considered an open problem.
1 code implementation • 16 Oct 2023 • Shaomu Tan, Christof Monz
Our findings highlight that the target side translation quality is the most influential factor, with vocabulary overlap consistently impacting ZS performance.
no code implementations • 15 Oct 2023 • Di wu, Shaomu Tan, David Stap, Ali Araabi, Christof Monz
This paper describes the UvA-MT's submission to the WMT 2023 shared task on general machine translation.
1 code implementation • 29 Aug 2023 • Sotirios Kastanas, Shaomu Tan, Yi He
In this study, we aim to fill these gaps by conducting a comparative evaluation of state-of-the-art models in document layout analysis and investigating the potential of cross-lingual layout analysis by utilizing machine translation techniques.
1 code implementation • NeurIPS 2023 • Baohao Liao, Shaomu Tan, Christof Monz
One effective way to reduce the activation memory is to apply a reversible model, so the intermediate activations are not necessary to be cached and can be recomputed.
no code implementations • 17 Dec 2022 • Shaomu Tan, Denis Paperno
In many real-world scenarios, the absence of external knowledge source like Wikipedia restricts question answering systems to rely on latent internal knowledge in limited dialogue data.