We propose a simple yet accurate method for dependency parsing that treats parsing as tagging (PaT). That is, our approach addresses the parsing of dependency trees with a sequence model implemented with a bidirectional LSTM over BERT embeddings, where the {``}tag{''} to be predicted at each token position is the relative position of the corresponding head. For example, for the sentence John eats cake, the tag to be predicted for the token cake is -1 because its head (eats) occurs one token to the left. Despite its simplicity, our approach performs well. For example, our approach outperforms the state-of-the-art method of (Fern{\'a}ndez-Gonz{\'a}lez and G{\'o}mez-Rodr{\'\i}guez, 2019) on Universal Dependencies (UD) by 1.76{\%} unlabeled attachment score (UAS) for English, 1.98{\%} UAS for French, and 1.16{\%} UAS for German. On average, on 12 UD languages, our method with minimal tuning performs comparably with this state-of-the-art approach: better by 0.11{\%} UAS, and worse by 0.58{\%} LAS.

PDF Abstract
No code implementations yet. Submit your code now


  Add Datasets introduced or used in this paper

Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.