2 code implementations • PoliticalNLP (LREC) 2022 • Luise Dürlich, Sebastian Reimann, Gustav Finnveden, Joakim Nivre, Sara Stymne
Causality detection is the task of extracting information about causal relations from text.
no code implementations • LaTeCHCLfL (COLING) 2022 • Ellinor Lindqvist, Eva Pettersson, Joakim Nivre
We also test how text pre-processing and different feature representations affect the result, and we examine feature importance for our main class of interest - petitions.
no code implementations • UDW (COLING) 2020 • Marsida Toska, Joakim Nivre, Daniel Zeman
In this paper, we introduce the first Universal Dependencies (UD) treebank for standard Albanian, consisting of 60 sentences collected from the Albanian Wikipedia, annotated with lemmas, universal part-of-speech tags, morphological features and syntactic dependencies.
1 code implementation • ACL 2022 • Fredrik Carlsson, Joey Öhman, Fangyu Liu, Severine Verlinden, Joakim Nivre, Magnus Sahlgren
We propose a resource-efficient method for converting a pre-trained CLM into this architecture, and demonstrate its potential on various experiments, including the novel task of contextualized word inclusion.
no code implementations • 25 Dec 2024 • Lifeng Han, Kilian Evang, Archna Bhatia, Gosse Bouma, A. Seza Doğruöz, Marcos Garcia, Voula Giouli, Joakim Nivre, Alexandre Rademacher
Starting in 2003 when the first MWE workshop was held with ACL in Sapporo, Japan, this year, the joint workshop of MWE-UD co-located with the LREC-COLING 2024 conference marked the 20th anniversary of MWE workshop events over the past nearly two decades.
no code implementations • 5 Dec 2024 • Fredrik Carlsson, Fangyu Liu, Daniel Ward, Murathan Kurfali, Joakim Nivre
This paper introduces the counter-intuitive generalization results of overfitting pre-trained large language models (LLMs) on very small datasets.
1 code implementation • 26 Mar 2024 • Leonie Weissweiler, Nina Böbel, Kirian Guiller, Santiago Herrera, Wesley Scivetti, Arthur Lorenzi, Nurit Melnik, Archna Bhatia, Hinrich Schütze, Lori Levin, Amir Zeldes, Joakim Nivre, William Croft, Nathan Schneider
The Universal Dependencies (UD) project has created an invaluable collection of treebanks with contributions in over 140 languages.
no code implementations • 2 Nov 2023 • Evangelia Gogoulou, Timothée Lesort, Magnus Boman, Joakim Nivre
The recent increase in data and model scale for language model pre-training has led to huge training costs.
no code implementations • 17 Oct 2021 • Artur Kulmizev, Joakim Nivre
In the last half-decade, the field of natural language processing (NLP) has undergone two major transitions: the switch to neural networks as the primary modeling paradigm and the homogenization of the training regime (pre-train, then fine-tune).
1 code implementation • 26 Jul 2021 • Gongbo Tang, Philipp Rönchen, Rico Sennrich, Joakim Nivre
In this paper, we evaluate the translation of negation both automatically and manually, in English--German (EN--DE) and English--Chinese (EN--ZH).
no code implementations • EACL 2021 • Ali Basirat, Joakim Nivre
Standard models for syntactic dependency parsing take words to be the elementary units that enter into dependency relations.
no code implementations • EACL 2021 • Vinit Ravishankar, Artur Kulmizev, Mostafa Abdou, Anders Søgaard, Joakim Nivre
Since the popularization of the Transformer as a general-purpose feature encoder for NLP, many studies have attempted to decode linguistic structure from its novel multi-head attention mechanism.
no code implementations • COLING 2020 • Gongbo Tang, Rico Sennrich, Joakim Nivre
The attention distribution pattern shows that separators attract a lot of attention and we explore a sparse word-level attention to enforce character hidden states to capture the full word-level information.
no code implementations • 9 Jul 2020 • Ali Basirat, Joakim Nivre
We study the effect of rich supertag features in greedy transition-based dependency parsing.
no code implementations • 9 Jul 2020 • Ali Basirat, Christian Hardmeier, Joakim Nivre
The effect of these generalizations on the word vectors is intrinsically studied with regard to the spread and the discriminability of the word vectors.
no code implementations • WS 2020 • Daniel Hershcovich, Miryam de Lhoneux, Artur Kulmizev, Elham Pejhan, Joakim Nivre
We present K{\o}psala, the Copenhagen-Uppsala system for the Enhanced Universal Dependencies Shared Task at IWPT 2020.
1 code implementation • 25 May 2020 • Daniel Hershcovich, Miryam de Lhoneux, Artur Kulmizev, Elham Pejhan, Joakim Nivre
We present K{\o}psala, the Copenhagen-Uppsala system for the Enhanced Universal Dependencies Shared Task at IWPT 2020.
no code implementations • LREC 2020 • Maja Buljan, Joakim Nivre, Stephan Oepen, Lilja {\O}vrelid
We discuss methodological choices in contrastive and diagnostic evaluation in meaning representation parsing, i. e. mapping from natural language utterances to graph-based encodings of its semantic structure.
no code implementations • ACL 2020 • Artur Kulmizev, Vinit Ravishankar, Mostafa Abdou, Joakim Nivre
Recent work on the interpretability of deep neural language models has concluded that many properties of natural language syntax are encoded in their representational spaces.
no code implementations • LREC 2020 • Joakim Nivre, Marie-Catherine de Marneffe, Filip Ginter, Jan Hajič, Christopher D. Manning, Sampo Pyysalo, Sebastian Schuster, Francis Tyers, Daniel Zeman
Universal Dependencies is an open community effort to create cross-linguistically consistent treebank annotation for many languages within a dependency-based lexicalist framework.
no code implementations • IJCNLP 2019 • Artur Kulmizev, Miryam de Lhoneux, Johannes Gontrum, Elena Fano, Joakim Nivre
Transition-based and graph-based dependency parsers have previously been shown to have complementary strengths and weaknesses: transition-based parsers exploit rich structural features but suffer from error propagation, while graph-based parsers benefit from global optimization but have restricted feature scope.
no code implementations • IJCNLP 2019 • Gongbo Tang, Rico Sennrich, Joakim Nivre
We find that encoder hidden states outperform word embeddings significantly which indicates that encoders adequately encode relevant information for disambiguation into hidden states.
no code implementations • 20 Aug 2019 • Artur Kulmizev, Miryam de Lhoneux, Johannes Gontrum, Elena Fano, Joakim Nivre
Transition-based and graph-based dependency parsers have previously been shown to have complementary strengths and weaknesses: transition-based parsers exploit rich structural features but suffer from error propagation, while graph-based parsers benefit from global optimization but have restricted feature scope.
1 code implementation • CL (ACL) 2020 • Miryam de Lhoneux, Sara Stymne, Joakim Nivre
We find that the parser learns different information about AVCs and FMVs if only sequential models (BiLSTMs) are used in the architecture but similar information when a recursive layer is used.
no code implementations • RANLP 2019 • Gongbo Tang, Rico Sennrich, Joakim Nivre
In this paper, we try to understand neural machine translation (NMT) via simplifying NMT architectures and training encoder-free NMT models.
1 code implementation • NAACL 2019 • Miryam de Lhoneux, Miguel Ballesteros, Joakim Nivre
When ablating the forward LSTM, performance drops less dramatically and composition recovers a substantial part of the gap, indicating that a forward LSTM and composition capture similar information.
1 code implementation • WS 2018 • Gosse Bouma, Jan Hajic, Dag Haug, Joakim Nivre, Per Erik Solberg, Lilja {\O}vrelid
Although treebanks annotated according to the guidelines of Universal Dependencies (UD) now exist for many languages, the goal of annotating the same phenomena in a cross-linguistically consistent fashion is not always met.
no code implementations • WS 2018 • Joakim Nivre, Paola Marongiu, Filip Ginter, Jenna Kanerva, Simonetta Montemagni, Sebastian Schuster, Maria Simi
We evaluate two cross-lingual techniques for adding enhanced dependencies to existing treebanks in Universal Dependencies.
no code implementations • WS 2018 • Gongbo Tang, Rico Sennrich, Joakim Nivre
Recent work has shown that the encoder-decoder attention mechanisms in neural machine translation (NMT) are different from the word alignment in statistical machine translation.
no code implementations • CONLL 2018 • Daniel Zeman, Jan Haji{\v{c}}, Martin Popel, Martin Potthast, Milan Straka, Filip Ginter, Joakim Nivre, Slav Petrov
Every year, the Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which participants train and test their learning systems on the same data sets.
no code implementations • CONLL 2018 • Aaron Smith, Bernd Bohnet, Miryam de Lhoneux, Joakim Nivre, Yan Shao, Sara Stymne
We present the Uppsala system for the CoNLL 2018 Shared Task on universal dependency parsing.
no code implementations • EMNLP 2018 • Aaron Smith, Miryam de Lhoneux, Sara Stymne, Joakim Nivre
We provide a comprehensive analysis of the interactions between pre-trained word embeddings, character models and POS tags in a transition-based dependency parser.
1 code implementation • TACL 2018 • Yan Shao, Christian Hardmeier, Joakim Nivre
Word segmentation is a low-level NLP task that is non-trivial for a considerable number of languages.
1 code implementation • COLING 2018 • Gongbo Tang, Fabienne Cap, Eva Pettersson, Joakim Nivre
In this paper, we apply different NMT models to the problem of historical spelling normalization for five languages: English, German, Hungarian, Icelandic, and Swedish.
1 code implementation • ACL 2018 • Sara Stymne, Miryam de Lhoneux, Aaron Smith, Joakim Nivre
How to make the most of multiple heterogeneous treebanks when training a monolingual dependency parser is an open question.
2 code implementations • NAACL 2018 • Sebastian Schuster, Joakim Nivre, Christopher D. Manning
Sentences with gapping, such as Paul likes coffee and Mary tea, lack an overt predicate to indicate the relation between two or more arguments.
no code implementations • IJCNLP 2017 • Yan Shao, Christian Hardmeier, Joakim Nivre
We extensively analyse the correlations and drawbacks of conventionally employed evaluation metrics for word segmentation.
1 code implementation • WS 2017 • Miryam de Lhoneux, Sara Stymne, Joakim Nivre
In this paper, we extend the arc-hybrid system for transition-based parsing with a swap transition that enables reordering of the words and construction of non-projective trees.
no code implementations • CONLL 2017 • Daniel Zeman, Martin Popel, Milan Straka, Jan Haji{\v{c}}, Joakim Nivre, Filip Ginter, Juhani Luotolahti, Sampo Pyysalo, Slav Petrov, Martin Potthast, Francis Tyers, Elena Badmaeva, Memduh Gokirmak, Anna Nedoluzhko, Silvie Cinkov{\'a}, Jan Haji{\v{c}} jr., Jaroslava Hlav{\'a}{\v{c}}ov{\'a}, V{\'a}clava Kettnerov{\'a}, Zde{\v{n}}ka Ure{\v{s}}ov{\'a}, Jenna Kanerva, Stina Ojala, Anna Missil{\"a}, Christopher D. Manning, Sebastian Schuster, Siva Reddy, Dima Taji, Nizar Habash, Herman Leung, Marie-Catherine de Marneffe, Manuela Sanguinetti, Maria Simi, Hiroshi Kanayama, Valeria de Paiva, Kira Droganova, H{\'e}ctor Mart{\'\i}nez Alonso, {\c{C}}a{\u{g}}r{\i} {\c{C}}{\"o}ltekin, Umut Sulubacak, Hans Uszkoreit, Vivien Macketanz, Aljoscha Burchardt, Kim Harris, Katrin Marheinecke, Georg Rehm, Tolga Kayadelen, Mohammed Attia, Ali Elkahky, Zhuoran Yu, Emily Pitler, Saran Lertpradit, M, Michael l, Jesse Kirchner, Hector Fern Alcalde, ez, Jana Strnadov{\'a}, Esha Banerjee, Ruli Manurung, Antonio Stella, Atsuko Shimada, Sookyoung Kwak, Gustavo Mendon{\c{c}}a, L, Tatiana o, Rattima Nitisaroj, Josie Li
The Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which participants train and test their learning systems on the same data sets.
no code implementations • CONLL 2017 • Miryam de Lhoneux, Yan Shao, Ali Basirat, Eliyahu Kiperwasser, Sara Stymne, Yoav Goldberg, Joakim Nivre
We present the Uppsala submission to the CoNLL 2017 shared task on parsing from raw text to universal dependencies.
1 code implementation • IJCNLP 2017 • Yan Shao, Christian Hardmeier, Jörg Tiedemann, Joakim Nivre
We present a character-based model for joint segmentation and POS tagging for Chinese.
no code implementations • CL (ACL) 2021 • Joakim Nivre, Daniel Zeman, Filip Ginter, Francis Tyers
Universal Dependencies (UD) is a project that seeks to develop cross-linguistically consistent treebank annotation for many languages.
no code implementations • COLING 2016 • Umut Sulubacak, Memduh Gokirmak, Francis Tyers, {\c{C}}a{\u{g}}r{\i} {\c{C}}{\"o}ltekin, Joakim Nivre, G{\"u}l{\c{s}}en Eryi{\u{g}}it
The Universal Dependencies (UD) project was conceived after the substantial recent interest in unifying annotation schemes across languages.
no code implementations • WS 2016 • Joakim Nivre
Universal Dependencies is an initiative to develop cross-linguistically consistent grammatical annotation for many languages, with the goal of facilitating multilingual parser development, cross-lingual learning and parsing research from a language typology perspective.
no code implementations • LREC 2016 • Joakim Nivre, Marie-Catherine de Marneffe, Filip Ginter, Yoav Goldberg, Jan Haji{\v{c}}, Christopher D. Manning, Ryan Mcdonald, Slav Petrov, Sampo Pyysalo, Natalia Silveira, Reut Tsarfaty, Daniel Zeman
Cross-linguistically consistent annotation is necessary for sound comparative evaluation and cross-lingual learning experiments.
no code implementations • LREC 2016 • Kaja Dobrovoljc, Joakim Nivre
This paper presents the construction of an open-source dependency treebank of spoken Slovenian, the first syntactically annotated collection of spontaneous speech in Slovenian.
no code implementations • LREC 2016 • Mojgan Seraji, Filip Ginter, Joakim Nivre
The Persian Universal Dependency Treebank (Persian UD) is a recent effort of treebanking Persian with Universal Dependencies (UD), an ongoing project that designs unified and cross-linguistically valid grammatical representations including part-of-speech tags, morphological features, and dependency relations.
no code implementations • 21 Mar 2016 • Bernd Bohnet, Miguel Ballesteros, Ryan Mcdonald, Joakim Nivre
Experiments on five languages show that feature selection can result in more compact models as well as higher accuracy under all conditions, but also that a dynamic ordering works better than a static ordering and that joint systems benefit more than standalone taggers.
no code implementations • LREC 2014 • Marie-Catherine de Marneffe, Timothy Dozat, Natalia Silveira, Katri Haverinen, Filip Ginter, Joakim Nivre, Christopher D. Manning
Revisiting the now de facto standard Stanford dependency representation, we propose an improved taxonomy to capture grammatical relations across languages, including morphologically rich ones.
no code implementations • LREC 2014 • Mojgan Seraji, Carina Jahani, Be{\'a}ta Megyesi, Joakim Nivre
We present the Uppsala Persian Dependency Treebank (UPDT) with a syntactic annotation scheme based on Stanford Typed Dependencies.
no code implementations • WS 2013 • Djam{\'e} Seddah, Reut Tsarfaty, S K{\"u}bler, ra, C, Marie ito, Jinho D. Choi, Rich{\'a}rd Farkas, Jennifer Foster, Iakes Goenaga, Koldo Gojenola Galletebeitia, Yoav Goldberg, Spence Green, Nizar Habash, Marco Kuhlmann, Wolfgang Maier, Joakim Nivre, Adam Przepi{\'o}rkowski, Ryan Roth, Wolfgang Seeker, Yannick Versley, Veronika Vincze, Marcin Woli{\'n}ski, Alina Wr{\'o}blewska, Eric Villemonte de la Clergerie
no code implementations • ACL 2013 • Ryan McDonald, Joakim Nivre, Yvonne Quirmbach-Brundage, Yoav Goldberg, Dipanjan Das, Kuzman Ganchev, Keith Hall, Slav Petrov, Hao Zhang, Oscar T{\"a}ckstr{\"o}m, Claudia Bedini, N{\'u}ria Bertomeu Castell{\'o}, Jungmee Lee
no code implementations • TACL 2013 • Oscar T{\"a}ckstr{\"o}m, Dipanjan Das, Slav Petrov, Ryan Mcdonald, Joakim Nivre
We consider the construction of part-of-speech taggers for resource-poor languages.
no code implementations • TACL 2013 • Yoav Goldberg, Joakim Nivre
This problem is aggravated by the fact that they are normally trained using oracles that are deterministic and incomplete in the sense that they assume a unique canonical path through the transition system and are only valid as long as the parser does not stray from this path.
no code implementations • TACL 2013 • Bernd Bohnet, Joakim Nivre, Igor Boguslavsky, Rich{\'a}rd Farkas, Filip Ginter, Jan Haji{\v{c}}
Joint morphological and syntactic analysis has been proposed as a way of improving parsing accuracy for richly inflected languages.
no code implementations • LREC 2012 • Mojgan Seraji, Be{\'a}ta Megyesi, Joakim Nivre
As for resources, we describe the Uppsala PErsian Corpus (UPEC) which is a modified version of the Bijankhan corpus with additional sentence segmentation and consistent tokenization modified for more appropriate syntactic annotation.
no code implementations • LREC 2012 • Miguel Ballesteros, Joakim Nivre
Freely available statistical parsers often require careful optimization to produce state-of-the-art results, which can be a non-trivial task especially for application developers who are not interested in parsing research for its own sake.