Counterfactual Data Augmentation for Neural Machine Translation

NAACL 2021  ·  Qi Liu, Matt Kusner, Phil Blunsom ·

We propose a data augmentation method for neural machine translation. It works by interpreting language models and phrasal alignment causally. Specifically, it creates augmented parallel translation corpora by generating (path-specific) counterfactual aligned phrases. We generate these by sampling new source phrases from a masked language model, then sampling an aligned counterfactual target phrase by noting that a translation language model can be interpreted as a Gumbel-Max Structural Causal Model (Oberst and Sontag, 2019). Compared to previous work, our method takes both context and alignment into account to maintain the symmetry between source and target sequences. Experiments on IWSLT{'}15 English → Vietnamese, WMT{'}17 English → German, WMT{'}18 English → Turkish, and WMT{'}19 robust English → French show that the method can improve the performance of translation, backtranslation and translation robustness.

PDF Abstract

Datasets


Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods


No methods listed for this paper. Add relevant methods here