no code implementations • CODI 2021 • Frances Yung, Merel Scholman, Vera Demberg
In the current contribution, we analyse whether a sophisticated connective generation module is necessary to select a connective, or whether this can be solved with simple methods (such as random choice between connectives that are known to express a given relation, or usage of a generic language model).
no code implementations • COLING (CODI, CRAC) 2022 • Frances Yung, Kaveri Anuranjana, Merel Scholman, Vera Demberg
Implicit discourse relations can convey more than one relation sense, but much of the research on discourse relations has focused on single relation senses.
no code implementations • COLING 2022 • Marian Marchal, Merel Scholman, Frances Yung, Vera Demberg
In many linguistic fields requiring annotated data, multiple interpretations of a single item are possible.
1 code implementation • LREC 2022 • Merel Scholman, Tianai Dong, Frances Yung, Vera Demberg
Both the corpus and the dataset can facilitate a multitude of applications and research purposes, for example to function as training data to improve the performance of automatic discourse relation parsers, as well as facilitate research into non-connective signals of discourse relations.
no code implementations • LREC 2022 • Merel Scholman, Valentina Pyatkin, Frances Yung, Ido Dagan, Reut Tsarfaty, Vera Demberg
The current contribution studies the effect of worker selection and training on the agreement on implicit relation labels between workers and gold labels, for both the DC and the QA method.
no code implementations • CODI 2021 • Merel Scholman, Tianai Dong, Frances Yung, Vera Demberg
Existing parse methods use varying approaches to identify explicit discourse connectives, but their performance has not been consistently evaluated in comparison to each other, nor have they been evaluated consistently on text other than newspaper articles.
no code implementations • 3 Apr 2023 • Valentina Pyatkin, Frances Yung, Merel C. J. Scholman, Reut Tsarfaty, Ido Dagan, Vera Demberg
Disagreement in natural language annotation has mostly been studied from a perspective of biases introduced by the annotators and the annotation frameworks.
no code implementations • WS 2019 • Frances Yung, Vera Demberg, Merel Scholman
The perspective of being able to crowd-source coherence relations bears the promise of acquiring annotations for new texts quickly, which could then increase the size and variety of discourse-annotated corpora.
no code implementations • WS 2019 • Wei Shi, Frances Yung, Vera Demberg
Implicit discourse relation classification is one of the most challenging and important tasks in discourse parsing, due to the lack of connective as strong linguistic cues.
no code implementations • WS 2018 • Frances Yung, Vera Demberg
A number of different discourse connectives can be used to mark the same discourse relation, but it is unclear what factors affect connective choice.
no code implementations • IJCNLP 2017 • Wei Shi, Frances Yung, Raphael Rubino, Vera Demberg
Implicit discourse relation recognition is an extremely challenging task due to the lack of indicative connectives.
General Classification
Implicit Discourse Relation Classification
+3
no code implementations • IJCNLP 2017 • Frances Yung, Hiroshi Noji, Yuji Matsumoto
Humans process language word by word and construct partial linguistic structures on the fly before the end of the sentence is perceived.