1 code implementation • NAACL 2022 • Jiarui Yao, Nianwen Xue, Bonan Min
The task of modal dependency parsing aims to parse a text into its modal dependency structure, which is a representation for the factuality of events in the text.
no code implementations • LILT 2019 • Bin Li, Yuan Wen, Li Song, Weiguang Qu, Nianwen Xue
One significant change we have made to the AMR annotation methodology is the inclusion of the alignment between word tokens in the sentence and the concepts/relations in the CAMR annotation to make it easier for automatic parsers to model the correspondence between a sentence and its meaning representation.
no code implementations • EMNLP (ACL) 2021 • Jin Zhao, Nianwen Xue, Jens Van Gysel, Jinho D. Choi
We present UMR-Writer, a web-based application for annotating Uniform Meaning Representations (UMR), a graph-based, cross-linguistically applicable semantic representation developed recently to support the development of interpretable natural language applications that require deep semantic analysis of texts.
no code implementations • EMNLP 2020 • Jiarui Yao, Haoling Qiu, Bonan Min, Nianwen Xue
We present the construction of a corpus of 500 Wikinews articles annotated with temporal dependency graphs (TDGs) that can be used to train systems to understand temporal relations in text.
no code implementations • 18 Aug 2022 • Annemarie Friedrich, Nianwen Xue, Alexis Palmer
This includes whether a situation is described as a state or as an event, whether the situation is finished or ongoing, and whether it is viewed as a whole or with a focus on a particular phase.
1 code implementation • ACL 2021 • Jiarui Yao, Haoling Qiu, Jin Zhao, Bonan Min, Nianwen Xue
In this paper, we frame factuality assessment as a modal dependency parsing task that identifies the events and their sources, formally known as conceivers, and then determine the level of certainty that the sources are asserting with respect to the events.
1 code implementation • ACL 2021 • Jingxuan Yang, Kerui Xu, Jun Xu, Si Li, Sheng Gao, Jun Guo, Nianwen Xue, Ji-Rong Wen
A second (multi-relational) GCN is then applied to the utterance states to produce a discourse relation-augmented representation for the utterances, which are then fused together with token states in each utterance as input to a dropped pronoun recovery layer.
Ranked #5 on Discourse Parsing on STAC
no code implementations • CONLL 2020 • Stephan Oepen, Omri Abend, Lasha Abzianidze, Johan Bos, Jan Hajic, Daniel Hershcovich, Bin Li, Tim O{'}Gorman, Nianwen Xue, Daniel Zeman
Extending a similar setup from the previous year, five distinct approaches to the representation of sentence meaning in the form of directed graphs were represented in the English training and evaluation data for the task, packaged in a uniform graph abstraction and serialization; for four of these representation frameworks, additional training and evaluation data was provided for one additional language per framework.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Jingxuan Yang, Kerui Xu, Jun Xu, Si Li, Sheng Gao, Jun Guo, Ji-Rong Wen, Nianwen Xue
Exploratory analysis also demonstrates that the GCRF did help to capture the dependencies between pronouns in neighboring utterances, thus contributes to the performance improvements.
no code implementations • CONLL 2019 • Stephan Oepen, Omri Abend, Jan Hajic, Daniel Hershcovich, Marco Kuhlmann, Tim O{'}Gorman, Nianwen Xue, Jayeol Chun, Milan Straka, Zdenka Uresova
The 2019 Shared Task at the Conference for Computational Language Learning (CoNLL) was devoted to Meaning Representation Parsing (MRP) across frameworks.
no code implementations • WS 2019 • Zi Lin, Nianwen Xue
The parsing accuracy varies a great deal for different meaning representations.
no code implementations • WS 2019 • James Pustejovsky, Ken Lai, Nianwen Xue
In this paper, we propose an extension to Abstract Meaning Representations (AMRs) to encode scope information of quantifiers and negation, in a way that overcomes the semantic gaps of the schema while maintaining its cognitive simplicity.
no code implementations • SEMEVAL 2019 • Yuchen Zhang, Nianwen Xue
Temporal Dependency Trees are a structured temporal representation that represents temporal relations among time expressions and events in a text as a dependency tree structure.
1 code implementation • NAACL 2019 • Jingxuan Yang, Jianzhuo Tong, Si Li, Sheng Gao, Jun Guo, Nianwen Xue
Pronouns are often dropped in Chinese sentences, and this happens more frequently in conversational genres as their referents can be easily understood from context.
2 code implementations • EMNLP 2018 • Yuchen Zhang, Nianwen Xue
In a parsing-only evaluation setup where gold time expressions and events are provided, our parser reaches 0. 81 and 0. 70 f-score on unlabeled and labeled parsing respectively, a result that is very competitive against alternative approaches.
2 code implementations • LREC 2018 • Yuchen Zhang, Nianwen Xue
Temporal relations between events and time expressions in a document are often modeled in an unstructured manner where relations between individual pairs of time expressions and events are considered in isolation.
no code implementations • NAACL 2018 • Chuan Wang, Bin Li, Nianwen Xue
This paper presents the first AMR parser built on the Chinese AMR bank.
no code implementations • EMNLP 2017 • Chuan Wang, Nianwen Xue
This paper proposes to tackle the AMR parsing bottleneck by improving two components of an AMR parser: concept identification and alignment.
Ranked #7 on AMR Parsing on LDC2014T12
no code implementations • CL 2017 • Dun Deng, Nianwen Xue
In this article, we conduct an empirical investigation of translation divergences between Chinese and English relying on a parallel treebank.
no code implementations • EACL 2017 • Attapol Rutherford, Vera Demberg, Nianwen Xue
Here, we propose neural network models that are based on feedforward and long-short term memory architecture and systematically study the effects of varying structures.
no code implementations • EACL 2017 • Xiaochang Peng, Chuan Wang, Daniel Gildea, Nianwen Xue
Neural attention models have achieved great success in different NLP tasks.
no code implementations • 30 Nov 2016 • Si Li, Nianwen Xue
A patent is a property right for an invention granted by the government to the inventor.
no code implementations • 7 Jun 2016 • Attapol T. Rutherford, Vera Demberg, Nianwen Xue
Inferring implicit discourse relations in natural language text is the most difficult subtask in discourse parsing.
no code implementations • LREC 2016 • Xuansong Li, Martha Palmer, Nianwen Xue, Lance Ramshaw, Mohamed Maamouri, Ann Bies, Kathryn Conger, Stephen Grimes, Stephanie Strassel
High accuracy for automated translation and information retrieval calls for linguistic annotations at various language levels.
no code implementations • LREC 2014 • Nianwen Xue, Yuchen Zhang
We describe a {``}distant annotation{''} method where we mark up the semantic tense, event type, and modality of Chinese events via a word-aligned parallel corpus.
no code implementations • LREC 2014 • Nianwen Xue, Ond{\v{r}}ej Bojar, Jan Haji{\v{c}}, Martha Palmer, Zde{\v{n}}ka Ure{\v{s}}ov{\'a}, Xiuhong Zhang
Abstract Meaning Representations (AMRs) are rooted, directional and labeled graphs that abstract away from morpho-syntactic idiosyncrasies such as word category (verbs and nouns), word order, and function words (determiners, some prepositions).
no code implementations • LREC 2012 • Xuansong Li, Stephanie Strassel, Stephen Grimes, Safa Ismael, Mohamed Maamouri, Ann Bies, Nianwen Xue
Parallel aligned treebanks (PAT) are linguistic corpora annotated with morphological and syntactic structures that are aligned at sentence as well as sub-sentence levels.
no code implementations • LREC 2012 • Elizabeth Baran, Yaqin Yang, Nianwen Xue
These abstract pronouns are identified as ''''''``unspecified'''''''', ''''''``pleonastic'''''''', ''''''``event'''''''', and ''''''``existential'''''''' and are argued to exist cross-linguistically.