1 code implementation • ACL 2022 • Kevin Stowe, Prasetya Utama, Iryna Gurevych
Natural language inference (NLI) has been widely used as a task to train and evaluate models for language understanding.
no code implementations • NAACL (DaSH) 2021 • Jan-Christoph Klie, Richard Eckart de Castilho, Iryna Gurevych
Entity linking (EL) is concerned with disambiguating entity mentions in a text against knowledge bases (KB).
1 code implementation • Findings (EMNLP) 2021 • Kexin Wang, Nils Reimers, Iryna Gurevych
Learning sentence embeddings often requires a large amount of labeled data.
1 code implementation • Findings (EMNLP) 2021 • Mohsen Mesgar, Leonardo F. R. Ribeiro, Iryna Gurevych
Entity grids and entity graphs are two frameworks for modeling local coherence.
1 code implementation • COLING (CRAC) 2022 • Haixia Chai, Nafise Sadat Moosavi, Iryna Gurevych, Michael Strube
The results of our extrinsic evaluation show that while there is a significant difference between the performance of the rule-based system vs. state-of-the-art neural model on coreference resolution datasets, we do not observe a considerable difference on their impact on downstream models.
1 code implementation • AKBC 2021 • Michael Bugert, Iryna Gurevych
Cross-document event coreference resolution (CDCR) is the task of identifying which event mentions refer to the same events throughout a collection of documents.
1 code implementation • Findings (ACL) 2022 • Andreas Waldis, Tilman Beck, Iryna Gurevych
Identifying the relation between two sentences requires datasets with pairwise annotations.
1 code implementation • CoNLL (EMNLP) 2021 • Kevin Stowe, Nils Beck, Iryna Gurevych
Metaphor generation is a difficult task, and has seen tremendous improvement with the advent of deep pretrained models.
no code implementations • 12 Apr 2024 • Ji-Ung Lee, Marc E. Pfetsch, Iryna Gurevych
This work proposes a novel method to generate C-Tests; a deviated form of cloze tests (a gap filling exercise) where only the last part of a word is turned into a gap.
no code implementations • 22 Mar 2024 • Chen Cecilia Liu, Iryna Gurevych
Prior research has found that differences in the early period of neural network training significantly impact the performance of in-distribution (ID) tasks.
no code implementations • 6 Mar 2024 • Jiahui Geng, Yova Kementchedjhieva, Preslav Nakov, Iryna Gurevych
To the best of our knowledge, we are the first to evaluate MLLMs for real-world fact-checking.
1 code implementation • 6 Mar 2024 • Indraneil Paul, Goran Glavaš, Iryna Gurevych
In particular, most mainstream Code-LMs have been pre-trained on source code files alone.
no code implementations • 5 Mar 2024 • Anmol Goel, Nico Daheim, Iryna Gurevych
In this work, we address this gap by augmenting open-source datasets for positive text rewriting with synthetically-generated Socratic rationales using a novel framework called \textsc{SocraticReframe}.
1 code implementation • 27 Feb 2024 • Yuesong Shen, Nico Daheim, Bai Cong, Peter Nickl, Gian Maria Marconi, Clement Bazan, Rio Yokota, Iryna Gurevych, Daniel Cremers, Mohammad Emtiyaz Khan, Thomas Möllenhoff
We give extensive empirical evidence against the common belief that variational learning is ineffective for large neural networks.
1 code implementation • 19 Feb 2024 • Justus-Jonas Erker, Florian Mai, Nils Reimers, Gerasimos Spanakis, Iryna Gurevych
Search-based dialog models typically re-encode the dialog history at every turn, incurring high cost.
no code implementations • 17 Feb 2024 • Yuxia Wang, Jonibek Mansurov, Petar Ivanov, Jinyan Su, Artem Shelmanov, Akim Tsvigun, Osama Mohanned Afzal, Tarek Mahmoud, Giovanni Puccetti, Thomas Arnold, Alham Fikri Aji, Nizar Habash, Iryna Gurevych, Preslav Nakov
The advent of Large Language Models (LLMs) has brought an unprecedented surge in machine-generated text (MGT) across diverse channels.
no code implementations • 3 Feb 2024 • Fajri Koto, Tilman Beck, Zeerak Talat, Iryna Gurevych, Timothy Baldwin
Improving multilingual language models capabilities in low-resource languages is generally difficult due to the scarcity of large-scale data in those languages.
1 code implementation • 2 Feb 2024 • Andreas Waldis, Yufang Hou, Iryna Gurevych
Pre-trained language models (LMs) perform well in In-Topic setups, where training and testing data come from the same topics.
no code implementations • 31 Jan 2024 • Jan Buchmann, Max Eichler, Jan-Micha Bodensohn, Ilia Kuznetsov, Iryna Gurevych
Long documents often exhibit structure with hierarchically organized elements of different functions, such as section headers and paragraphs.
1 code implementation • 18 Jan 2024 • Haritz Puerto, Martin Tutek, Somak Aditya, Xiaodan Zhu, Iryna Gurevych
Reasoning is a fundamental component of language understanding.
1 code implementation • 17 Jan 2024 • Dominic Petrak, Thy Thy Tran, Iryna Gurevych
The success of task-oriented and document-grounded dialogue systems depends on users accepting and enjoying using them.
1 code implementation • 18 Nov 2023 • Clifton Poth, Hannah Sterz, Indraneil Paul, Sukannya Purkayastha, Leon Engländer, Timo Imhof, Ivan Vulić, Sebastian Ruder, Iryna Gurevych, Jonas Pfeiffer
We introduce Adapters, an open-source library that unifies parameter-efficient and modular transfer learning in large language models.
1 code implementation • 15 Nov 2023 • Yuxia Wang, Revanth Gangi Reddy, Zain Muhammad Mujahid, Arnav Arora, Aleksandr Rubashevskii, Jiahui Geng, Osama Mohammed Afzal, Liangming Pan, Nadav Borenstein, Aditya Pillai, Isabelle Augenstein, Iryna Gurevych, Preslav Nakov
The increased use of large language models (LLMs) across a variety of real-world applications calls for mechanisms to verify the factual accuracy of their outputs.
no code implementations • 14 Nov 2023 • Jiahui Geng, Fengyu Cai, Yuxia Wang, Heinz Koeppl, Preslav Nakov, Iryna Gurevych
Assessing their confidence and calibrating them across different tasks can help mitigate risks and enable LLMs to produce better generations.
no code implementations • 13 Nov 2023 • Sheng Lu, Hendrik Schuff, Iryna Gurevych
In-context learning (ICL) has become one of the most popular learning paradigms.
1 code implementation • 11 Nov 2023 • Luke Bates, Peter Ebert Christensen, Preslav Nakov, Iryna Gurevych
Here, to aid understanding of memes, we release a knowledge base of memes and information found on www. knowyourmeme. com, which we call the Know Your Meme Knowledge Base (KYMKB), composed of more than 54, 000 images.
1 code implementation • 7 Nov 2023 • Sukannya Purkayastha, Anne Lauscher, Iryna Gurevych
In this work, we are the first to explore Jiu-Jitsu argumentation for peer review by proposing the novel task of attitude and theme-guided rebuttal generation.
1 code implementation • 1 Nov 2023 • Yongxin Huang, Kexin Wang, Sourav Dutta, Raj Nath Patel, Goran Glavaš, Iryna Gurevych
As a solution, we propose AdaSent, which decouples SEPT from DAPT by training a SEPT adapter on the base PLM.
1 code implementation • 24 Oct 2023 • Dominic Petrak, Nafise Sadat Moosavi, Ye Tian, Nikolai Rozanov, Iryna Gurevych
Learning from free-text human feedback is essential for dialog systems, but annotated data is scarce and usually covers only a small fraction of error types known in conversational AI.
no code implementations • 19 Oct 2023 • Nico Daheim, Thomas Möllenhoff, Edoardo Maria Ponti, Iryna Gurevych, Mohammad Emtiyaz Khan
Models trained on different datasets can be merged by a weighted-averaging of their parameters, but why does it work and when can it fail?
1 code implementation • 18 Oct 2023 • Sheng Lu, Shan Chen, Yingya Li, Danielle Bitterman, Guergana Savova, Iryna Gurevych
In-context learning (ICL) is a new learning paradigm that has gained popularity along with the development of large language models.
no code implementations • 15 Sep 2023 • Andreas Waldis, Yufang Hou, Iryna Gurevych
Our findings challenge the previously asserted general superiority of in-context learning (ICL) for OOD.
1 code implementation • 15 Sep 2023 • Chen Cecilia Liu, Fajri Koto, Timothy Baldwin, Iryna Gurevych
Large language models (LLMs) are highly adept at question answering and reasoning tasks, but when reasoning in a situational context, human expectations vary depending on the relevant cultural common ground.
1 code implementation • 14 Sep 2023 • Rachneet Sachdeva, Martin Tutek, Iryna Gurevych
In recent years, large language models (LLMs) have shown remarkable capabilities at scale, particularly at generating text conditioned on a prompt.
1 code implementation • 13 Sep 2023 • Tilman Beck, Hendrik Schuff, Anne Lauscher, Iryna Gurevych
However, the available NLP literature disagrees on the efficacy of this technique - it remains unclear for which tasks and scenarios it can help, and the role of the individual factors in sociodemographic prompting is still unexplored.
1 code implementation • 4 Sep 2023 • Sheng Lu, Irina Bigoulaeva, Rachneet Sachdeva, Harish Tayyar Madabushi, Iryna Gurevych
Large language models have exhibited emergent abilities, demonstrating exceptional performance across diverse tasks for which they were not explicitly trained, including those that require complex reasoning abilities.
1 code implementation • 19 Jul 2023 • Nandan Thakur, Kexin Wang, Iryna Gurevych, Jimmy Lin
In this work, we provide SPRINT, a unified Python toolkit based on Pyserini and Lucene, supporting a common interface for evaluating neural sparse retrieval.
2 code implementations • 16 Jul 2023 • Jan-Christoph Klie, Richard Eckart de Castilho, Iryna Gurevych
A majority of the annotated publications apply good or excellent quality management.
no code implementations • 29 Jun 2023 • Ji-Ung Lee, Haritz Puerto, Betty van Aken, Yuki Arase, Jessica Zosa Forde, Leon Derczynski, Andreas Rücklé, Iryna Gurevych, Roy Schwartz, Emma Strubell, Jesse Dodge
Many recent improvements in NLP stem from the development and use of large pre-trained language models (PLMs) with billions of parameters.
1 code implementation • 31 May 2023 • Haishuo Fang, Haritz Puerto, Iryna Gurevych
To evaluate the effectiveness of UKP-SQuARE in teaching scenarios, we adopted it in a postgraduate NLP course and surveyed the students after the course.
1 code implementation • 24 May 2023 • Tianyu Yang, Thy Thy Tran, Iryna Gurevych
These models also suffer from posterior collapse, i. e., the decoder tends to ignore latent variables and directly access information captured in the encoder through the cross-attention mechanism.
2 code implementations • 24 May 2023 • Yuxia Wang, Jonibek Mansurov, Petar Ivanov, Jinyan Su, Artem Shelmanov, Akim Tsvigun, Chenxi Whitehouse, Osama Mohammed Afzal, Tarek Mahmoud, Toru Sasaki, Thomas Arnold, Alham Fikri Aji, Nizar Habash, Iryna Gurevych, Preslav Nakov
These results show that the problem is far from solved and that there is a lot of room for improvement.
1 code implementation • 23 May 2023 • Jakub Macina, Nico Daheim, Sankalan Pal Chowdhury, Tanmay Sinha, Manu Kapur, Iryna Gurevych, Mrinmaya Sachan
While automatic dialogue tutors hold great potential in making education personalized and more accessible, research on such systems has been hampered by a lack of sufficiently large and high-quality datasets.
2 code implementations • 23 May 2023 • Kexin Wang, Nils Reimers, Iryna Gurevych
This drives us to build a benchmark for this task including multiple datasets from heterogeneous domains.
no code implementations • 22 May 2023 • Aniket Pramanick, Yufang Hou, Saif M. Mohammad, Iryna Gurevych
In this study, we propose a systematic framework for analyzing the evolution of research topics in a scientific field using causal discovery and inference techniques.
no code implementations • 12 May 2023 • Georgia Chalvatzaki, Ali Younes, Daljeet Nandha, An Le, Leonardo F. R. Ribeiro, Iryna Gurevych
Long-horizon task planning is essential for the development of intelligent assistive and service robots.
1 code implementation • 25 Apr 2023 • Jan-Christoph Klie, Ji-Ung Lee, Kevin Stowe, Gözde Gül Şahin, Nafise Sadat Moosavi, Luke Bates, Dominic Petrak, Richard Eckart de Castilho, Iryna Gurevych
Citizen Science is an alternative to crowdsourcing that is relatively unexplored in the context of NLP.
no code implementations • 18 Apr 2023 • Sukannya Purkayastha, Sebastian Ruder, Jonas Pfeiffer, Iryna Gurevych, Ivan Vulić
In order to boost the capacity of mPLMs to deal with low-resource and unseen languages, we explore the potential of leveraging transliteration on a massive scale.
1 code implementation • 31 Mar 2023 • Haritz Puerto, Tim Baumgärtner, Rachneet Sachdeva, Haishuo Fang, Hao Zhang, Sewin Tariverdian, Kexin Wang, Iryna Gurevych
To ease research in multi-agent models, we extend UKP-SQuARE, an online platform for QA research, to support three families of multi-agent systems: i) agent selection, ii) early-fusion of agents, and iii) late-fusion of agents.
1 code implementation • 30 Mar 2023 • Nico Daheim, Nouha Dziri, Mrinmaya Sachan, Iryna Gurevych, Edoardo M. Ponti
We evaluate our method -- using different variants of Flan-T5 as a backbone language model -- on multiple datasets for information-seeking dialogue generation and compare our method with state-of-the-art techniques for faithfulness, such as CTRL, Quark, DExperts, and Noisy Channel reranking.
3 code implementations • 13 Mar 2023 • Ulf A. Hamster, Ji-Ung Lee, Alexander Geyken, Iryna Gurevych
Training and inference on edge devices often requires an efficient setup due to computational limitations.
1 code implementation • 24 Feb 2023 • Dennis Zyska, Nils Dycke, Jan Buchmann, Ilia Kuznetsov, Iryna Gurevych
Recent years have seen impressive progress in AI-assisted writing, yet the developments in AI-assisted reading are lacking.
1 code implementation • 17 Feb 2023 • Luke Bates, Iryna Gurevych
Few-shot text classification systems have impressive capabilities but are infeasible to deploy and use reliably due to their dependence on prompting and billion-parameter language models.
1 code implementation • 24 Jan 2023 • Jakub Macina, Nico Daheim, Lingzhi Wang, Tanmay Sinha, Manu Kapur, Iryna Gurevych, Mrinmaya Sachan
Designing dialog tutors has been challenging as it involves modeling the diverse and complex pedagogical strategies employed by human tutors.
1 code implementation • 13 Jan 2023 • Chen Cecilia Liu, Jonas Pfeiffer, Ivan Vulić, Iryna Gurevych
Our experiments reveal that scheduled unfreezing induces different learning dynamics compared to standard fine-tuning, and provide evidence that the dynamics of Fisher Information during training correlate with cross-lingual generalization performance.
1 code implementation • 19 Dec 2022 • Haau-Sing Li, Mohsen Mesgar, André F. T. Martins, Iryna Gurevych
We hypothesize that the under-specification of a natural language description can be resolved by asking clarification questions.
1 code implementation • 19 Dec 2022 • Martin Funkquist, Ilia Kuznetsov, Yufang Hou, Iryna Gurevych
To address this challenge, we propose CiteBench: a benchmark for citation text generation that unifies multiple diverse datasets and enables standardized evaluation of citation text generation models across task designs and domains.
no code implementations • 22 Nov 2022 • Tobias Mayer, Neha Warikoo, Oliver Grimm, Andreas Reif, Iryna Gurevych
While these conversations are part of the daily routine of clinicians, gathering them is usually hindered by various ethical (purpose of data usage), legal (data privacy) and technical (data formatting) limitations.
no code implementations • 22 Nov 2022 • Neha Warikoo, Tobias Mayer, Dana Atzil-Slonim, Amir Eliassaf, Shira Haimovitz, Iryna Gurevych
No study has examined EC between the subjective experience of emotions and emotion expression in therapy or whether this coherence is associated with clients' well being.
no code implementations • 14 Nov 2022 • Anxo Pérez, Neha Warikoo, Kexin Wang, Javier Parapar, Iryna Gurevych
Depressive disorders constitute a severe public health issue worldwide.
1 code implementation • 12 Nov 2022 • Nils Dycke, Ilia Kuznetsov, Iryna Gurevych
Peer review constitutes a core component of scholarly publishing; yet it demands substantial expertise and training, and is susceptible to errors and biases.
no code implementations • 10 Nov 2022 • Ilia Kuznetsov, Iryna Gurevych
Natural language processing (NLP) researchers develop models of grammar, meaning and communication based on written text.
1 code implementation • 3 Nov 2022 • Tilman Beck, Andreas Waldis, Iryna Gurevych
Stance detection deals with identifying an author's stance towards a target.
1 code implementation • 31 Oct 2022 • Irina Bigoulaeva, Rachneet Sachdeva, Harish Tayyar Madabushi, Aline Villavicencio, Iryna Gurevych
We compare sequential fine-tuning with a model for multi-task learning in the context where we are interested in boosting performance on two tasks, one of which depends on the other.
1 code implementation • 25 Oct 2022 • Max Glockner, Yufang Hou, Iryna Gurevych
In our analysis, we show that, by design, existing NLP task definitions for fact-checking cannot refute misinformation as professional fact-checkers do for the majority of claims.
1 code implementation • 19 Oct 2022 • Tim Baumgärtner, Leonardo F. R. Ribeiro, Nils Reimers, Iryna Gurevych
Pairing a lexical retriever with a neural re-ranking model has set state-of-the-art performance on large-scale information retrieval datasets.
no code implementations • 12 Oct 2022 • Mohsen Mesgar, Thy Thy Tran, Goran Glavas, Iryna Gurevych
First, the unexplored combination of the cross-encoder architecture (with parameterized similarity scoring function) and episodic meta-learning consistently yields the best FSIC performance.
no code implementations • 12 Oct 2022 • Gregor Geigle, Chen Cecilia Liu, Jonas Pfeiffer, Iryna Gurevych
While many VEs -- of different architectures, trained on different data and objectives -- are publicly available, they are not designed for the downstream V+L tasks.
no code implementations • 31 Aug 2022 • Marcos Treviso, Ji-Ung Lee, Tianchu Ji, Betty van Aken, Qingqing Cao, Manuel R. Ciosici, Michael Hassid, Kenneth Heafield, Sara Hooker, Colin Raffel, Pedro H. Martins, André F. T. Martins, Jessica Zosa Forde, Peter Milder, Edwin Simpson, Noam Slonim, Jesse Dodge, Emma Strubell, Niranjan Balasubramanian, Leon Derczynski, Iryna Gurevych, Roy Schwartz
Recent work in natural language processing (NLP) has yielded appealing results from scaling model parameters and training data; however, using only scale to improve performance means that resource consumption also grows.
2 code implementations • 30 Aug 2022 • Haishuo Fang, Ji-Ung Lee, Nafise Sadat Moosavi, Iryna Gurevych
In contrast to conventional, predefined activation functions, RAFs can adaptively learn optimal activation functions during training according to input data.
1 code implementation • 19 Aug 2022 • Rachneet Sachdeva, Haritz Puerto, Tim Baumgärtner, Sewin Tariverdian, Hao Zhang, Kexin Wang, Hossain Shaikh Saadi, Leonardo F. R. Ribeiro, Iryna Gurevych
In this paper, we introduce SQuARE v2, the new version of SQuARE, to provide an explainability infrastructure for comparing models based on methods such as saliency maps and graph-based explanations.
1 code implementation • 16 Aug 2022 • Lorenz Stangier, Ji-Ung Lee, Yuxi Wang, Marvin Müller, Nicholas Frick, Joachim Metternich, Iryna Gurevych
We evaluate TexPrax in a user-study with German factory employees who ask their colleagues for solutions on problems that arise during their daily work.
1 code implementation • 12 Aug 2022 • Ivan Habernal, Daniel Faber, Nicola Recchia, Sebastian Bretthauer, Iryna Gurevych, Indra Spiecker genannt Döhmann, Christoph Burchard
Identifying, classifying, and analyzing arguments in legal discourse has been a prominent area of research since the inception of the argument mining field.
1 code implementation • 5 Jun 2022 • Jan-Christoph Klie, Bonnie Webber, Iryna Gurevych
While researchers show that their approaches work well on their newly introduced datasets, they rarely compare their methods to previous work or on the same datasets.
no code implementations • 23 May 2022 • Benjamin Schiller, Johannes Daxenberger, Iryna Gurevych
The task of Argument Mining, that is extracting argumentative sentences for a specific topic from large document sources, is an inherently difficult task for machine learning models and humans alike, as large Argument Mining datasets are rare and recognition of argumentative sentences requires expert knowledge.
2 code implementations • 13 May 2022 • Dominic Petrak, Nafise Sadat Moosavi, Iryna Gurevych
In this paper, we propose a new extended pretraining approach called Arithmetic-Based Pretraining that jointly addresses both in one extended pretraining step without requiring architectural changes or pretraining from scratch.
1 code implementation • NAACL 2022 • Prasetya Ajie Utama, Joshua Bambrick, Nafise Sadat Moosavi, Iryna Gurevych
In this work, we show that NLI models can be effective for this task when the training data is augmented with high-quality task-oriented examples.
Abstractive Text Summarization Natural Language Inference +1
1 code implementation • NAACL 2022 • Nafise Sadat Moosavi, Quentin Delfosse, Kristian Kersting, Iryna Gurevych
The resulting adapters (a) contain about 50% of the learning parameters of the standard adapter and are therefore more efficient at training and inference, and require less storage space, and (b) achieve considerably higher performances in low-data settings.
1 code implementation • 22 Apr 2022 • Ilia Kuznetsov, Jan Buchmann, Max Eichler, Iryna Gurevych
While existing NLP studies focus on the analysis of individual texts, editorial assistance often requires modeling interactions between pairs of texts -- yet general frameworks and datasets to support this scenario are missing.
2 code implementations • NAACL 2022 • Leonardo F. R. Ribeiro, Mengwen Liu, Iryna Gurevych, Markus Dreyer, Mohit Bansal
Despite recent improvements in abstractive summarization, most current approaches generate summaries that are not factually consistent with the source document, severely restricting their trust and usage in real-world applications.
1 code implementation • ACL 2022 • Tim Baumgärtner, Kexin Wang, Rachneet Sachdeva, Max Eichler, Gregor Geigle, Clifton Poth, Hannah Sterz, Haritz Puerto, Leonardo F. R. Ribeiro, Jonas Pfeiffer, Nils Reimers, Gözde Gül Şahin, Iryna Gurevych
Recent advances in NLP and information retrieval have given rise to a diverse set of question answering tasks that are of different formats (e. g., extractive, abstractive), require different model architectures (e. g., generative, discriminative), and setups (e. g., with or without retrieval).
1 code implementation • 15 Feb 2022 • Chen Liu, Jonas Pfeiffer, Anna Korhonen, Ivan Vulić, Iryna Gurevych
2) We analyze cross-lingual VQA across different question types of varying complexity for different multilingual multimodal Transformers, and identify question types that are the most difficult to improve on.
2 code implementations • 14 Feb 2022 • Federico Ruggeri, Mohsen Mesgar, Iryna Gurevych
The applications of conversational agents for scientific disciplines (as expert domains) are understudied due to the lack of dialogue data to train such agents.
Ranked #1 on Fact Selection on ArgSciChat
1 code implementation • 27 Jan 2022 • Nils Dycke, Ilia Kuznetsov, Iryna Gurevych
The shift towards publicly available text sources has enabled language processing at unprecedented scale, yet leaves under-serviced the domains where public and openly licensed data is scarce.
no code implementations • 15 Jan 2022 • Irina Bigoulaeva, Viktor Hangya, Iryna Gurevych, Alexander Fraser
The goal of hate speech detection is to filter negative online content aiming at certain groups of people.
5 code implementations • NAACL 2022 • Kexin Wang, Nandan Thakur, Nils Reimers, Iryna Gurevych
This limits the usage of dense retrieval approaches to only a few domains with large training datasets.
Ranked #9 on Zero-shot Text Search on BEIR
1 code implementation • 3 Dec 2021 • Haritz Puerto, Gözde Gül Şahin, Iryna Gurevych
The recent explosion of question answering (QA) datasets and models has increased the interest in the generalization of models across multiple domains and formats by either training on multiple datasets or by combining multiple models.
no code implementations • 29 Nov 2021 • Iryna Gurevych, Michael Kohler, Gözde Gül Sahin
Pattern recognition based on a high-dimensional predictor is considered.
1 code implementation • Findings (ACL) 2022 • Jonas Pfeiffer, Gregor Geigle, Aishwarya Kamath, Jan-Martin O. Steitz, Stefan Roth, Ivan Vulić, Iryna Gurevych
In this work, we address this gap and provide xGQA, a new multilingual evaluation benchmark for the visual question answering task.
no code implementations • 9 Sep 2021 • Jan-Martin O. Steitz, Jonas Pfeiffer, Iryna Gurevych, Stefan Roth
Reasoning over multiple modalities, e. g. in Visual Question Answering (VQA), requires an alignment of semantic concepts across domains.
1 code implementation • EMNLP 2021 • Prasetya Ajie Utama, Nafise Sadat Moosavi, Victor Sanh, Iryna Gurevych
Recent prompt-based approaches allow pretrained language models to achieve strong performances on few-shot finetuning by reformulating downstream tasks as a language modeling problem.
1 code implementation • EMNLP 2021 • Leonardo F. R. Ribeiro, Jonas Pfeiffer, Yue Zhang, Iryna Gurevych
Recent work on multilingual AMR-to-text generation has exclusively focused on data augmentation strategies that utilize silver AMR.
no code implementations • 2 Sep 2021 • Nils Dycke, Edwin Simpson, Ilia Kuznetsov, Iryna Gurevych
Peer review is the primary means of quality control in academia; as an outcome of a peer review process, program and area chairs make acceptance decisions for each paper based on the review reports and scores they received.
1 code implementation • ACL 2022 • Tilman Beck, Bela Bohlender, Christina Viehmann, Vincent Hane, Yanik Adamson, Jaber Khuri, Jonas Brossmann, Jonas Pfeiffer, Iryna Gurevych
The open-access dissemination of pretrained language models through online repositories has led to a democratization of state-of-the-art natural language processing (NLP) research.
no code implementations • 1 Jul 2021 • Anne Lauscher, Henning Wachsmuth, Iryna Gurevych, Goran Glavaš
Despite extensive research efforts in recent years, computational argumentation (CA) remains one of the most challenging areas of natural language processing.
1 code implementation • CL (ACL) 2022 • Ji-Ung Lee, Jan-Christoph Klie, Iryna Gurevych
Annotation studies often require annotators to familiarize themselves with the task, its annotation scheme, and the data domain.
1 code implementation • ACL 2021 • Kevin Stowe, Tuhin Chakrabarty, Nanyun Peng, Smaranda Muresan, Iryna Gurevych
Guided by conceptual metaphor theory, we propose to control the generation process by encoding conceptual mappings between cognitive domains to generate meaningful metaphoric expressions.
1 code implementation • ACL 2021 • Tilman Beck, Ji-Ung Lee, Christina Viehmann, Marcus Maurer, Oliver Quiring, Iryna Gurevych
This work investigates the use of interactively updated label suggestions to improve upon the efficiency of gathering annotations on the task of opinion mining in German Covid-19 social media data.
2 code implementations • 17 Apr 2021 • Nandan Thakur, Nils Reimers, Andreas Rücklé, Abhishek Srivastava, Iryna Gurevych
To address this, and to facilitate researchers to broadly evaluate the effectiveness of their models, we introduce Benchmarking-IR (BEIR), a robust and heterogeneous evaluation benchmark for information retrieval.
Ranked #1 on Argument Retrieval on ArguAna (BEIR)
1 code implementation • 17 Apr 2021 • Aniket Pramanick, Tilman Beck, Kevin Stowe, Iryna Gurevych
Language use changes over time, and this impacts the effectiveness of NLP systems.
1 code implementation • 16 Apr 2021 • Nafise Sadat Moosavi, Andreas Rücklé, Dan Roth, Iryna Gurevych
In this paper, we introduce SciGen, a new challenge dataset for the task of reasoning-aware data-to-text generation consisting of tables from scientific articles and their corresponding descriptions.
1 code implementation • EMNLP 2021 • Clifton Poth, Jonas Pfeiffer, Andreas Rücklé, Iryna Gurevych
Our best methods achieve an average Regret@3 of less than 1% across all target tasks, demonstrating that we are able to efficiently identify the best datasets for intermediate training.
1 code implementation • 14 Apr 2021 • Gregor Geigle, Nils Reimers, Andreas Rücklé, Iryna Gurevych
We argue that there exist a wide range of specialized QA agents in literature.
7 code implementations • 14 Apr 2021 • Kexin Wang, Nils Reimers, Iryna Gurevych
Learning sentence embeddings often requires a large amount of labeled data.
Ranked #1 on Re-Ranking on AskUbuntu
2 code implementations • 1 Apr 2021 • Max Glockner, Ieva Staliūnaitė, James Thorne, Gisela Vallejo, Andreas Vlachos, Iryna Gurevych
Automated fact-checking systems verify claims against evidence to predict their veracity.
1 code implementation • 22 Mar 2021 • Gregor Geigle, Jonas Pfeiffer, Nils Reimers, Ivan Vulić, Iryna Gurevych
Current state-of-the-art approaches to cross-modal retrieval process text and visual input jointly, relying on Transformer-based architectures with cross-attention mechanisms that attend over all words and objects in an image.
1 code implementation • EMNLP 2021 • Leonardo F. R. Ribeiro, Yue Zhang, Iryna Gurevych
Pretrained language models (PLM) have recently advanced graph-to-text generation, where the input graph is linearized into a sequence and fed into the PLM to obtain its representation.
Ranked #1 on Data-to-Text Generation on AMR3.0
no code implementations • 3 Feb 2021 • Patrick Abels, Zahra Ahmadi, Sophie Burkhardt, Benjamin Schiller, Iryna Gurevych, Stefan Kramer
We use a topic model to extract topic- and sentence-specific evidence from the structured knowledge base Wikidata, building a graph based on the cosine similarity between the entity word vectors of Wikidata and the vector of the given sentence.
no code implementations • 15 Jan 2021 • Yevgeniy Puzikov, Simoes Stanley, Iryna Gurevych, Immanuel Schweizer
In this work we empirically compare the dominant optimization paradigms which provide supervision signals during training: backtranslation, adversarial training and reinforcement learning.
2 code implementations • EMNLP 2021 • Jonas Pfeiffer, Ivan Vulić, Iryna Gurevych, Sebastian Ruder
The ultimate challenge is dealing with under-resourced languages not covered at all by the models and written in scripts unseen during pretraining.
1 code implementation • ACL 2021 • Phillip Rust, Jonas Pfeiffer, Ivan Vulić, Sebastian Ruder, Iryna Gurevych
In this work, we provide a systematic and comprehensive empirical comparison of pretrained multilingual language models versus their monolingual counterparts with regard to their monolingual task performance.
1 code implementation • ACL 2021 • Mingzhu Wu, Nafise Sadat Moosavi, Dan Roth, Iryna Gurevych
We propose a methodology for creating MRC datasets that better reflect the challenges of coreference reasoning and use it to create a sample evaluation set.
no code implementations • ACL 2021 • Nils Reimers, Iryna Gurevych
Information Retrieval using dense low-dimensional representations recently became popular and showed out-performance to traditional sparse-representations like BM25.
1 code implementation • CL (ACL) 2021 • Michael Bugert, Nils Reimers, Iryna Gurevych
This raises strong concerns on their generalizability -- a must-have for downstream applications where the magnitude of domains or event mentions is likely to exceed those found in a curated corpus.
no code implementations • 23 Oct 2020 • Julia Siekiera, Marius Köppel, Edwin Simpson, Kevin Stowe, Iryna Gurevych, Stefan Kramer
We therefore adapt the DirectRanker to provide a new deep model for ranking creative language with small data.
no code implementations • 23 Oct 2020 • Nafise Sadat Moosavi, Marcel de Boer, Prasetya Ajie Utama, Iryna Gurevych
Existing approaches to improve robustness against dataset biases mostly focus on changing the training objective so that models learn less from biased examples.
1 code implementation • EMNLP 2021 • Andreas Rücklé, Gregor Geigle, Max Glockner, Tilman Beck, Jonas Pfeiffer, Nils Reimers, Iryna Gurevych
Massively pre-trained transformer models are computationally expensive to fine-tune, slow for inference, and have large storage requirements.
1 code implementation • NAACL 2021 • Nandan Thakur, Nils Reimers, Johannes Daxenberger, Iryna Gurevych
Bi-encoders, on the other hand, require substantial training data and fine-tuning over the target task to achieve competitive performance.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Max Glockner, Ivan Habernal, Iryna Gurevych
We propose a differentiable training-framework to create models which output faithful rationales on a sentence level, by solely applying supervision on the target task.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Mingzhu Wu, Nafise Sadat Moosavi, Andreas Rücklé, Iryna Gurevych
Our framework weights each example based on the biases it contains and the strength of those biases in the training data.
1 code implementation • EMNLP 2020 • Andreas Rücklé, Jonas Pfeiffer, Iryna Gurevych
We investigate the model performances on nine benchmarks of answer selection and question similarity tasks, and show that all 140 models transfer surprisingly well, where the large majority of models substantially outperforms common IR baselines.
1 code implementation • EMNLP 2020 • Prasetya Ajie Utama, Nafise Sadat Moosavi, Iryna Gurevych
Recently proposed debiasing methods are shown to be effective in mitigating this tendency.
1 code implementation • 3 Aug 2020 • Tristan Miller, Erik-Lân Do Dinh, Edwin Simpson, Iryna Gurevych
Most humour processing systems to date make at best discrete, coarse-grained distinctions between the comical and the conventional, yet such notions are better conceptualized as a broad spectrum.
3 code implementations • EMNLP (NLP4ConvAI) 2021 • Leonardo F. R. Ribeiro, Martin Schmitt, Hinrich Schütze, Iryna Gurevych
We show that the PLMs BART and T5 achieve new state-of-the-art results and that task-adaptive pretraining strategies improve their performance even further.
Ranked #1 on KG-to-Text Generation on WebNLG (All)
8 code implementations • EMNLP 2020 • Jonas Pfeiffer, Andreas Rücklé, Clifton Poth, Aishwarya Kamath, Ivan Vulić, Sebastian Ruder, Kyunghyun Cho, Iryna Gurevych
We propose AdapterHub, a framework that allows dynamic "stitching-in" of pre-trained adapters for different tasks and languages.
1 code implementation • ACL 2020 • Jan-Christoph Klie, Richard Eckart de Castilho, Iryna Gurevych
Entity linking (EL) is concerned with disambiguating entity mentions in a text against knowledge bases (KB).
1 code implementation • CONLL 2020 • Steffen Eger, Johannes Daxenberger, Iryna Gurevych
We then probe embeddings in a multilingual setup with design choices that lie in a 'stable region', as we identify for English, and find that results on English do not transfer to other languages.
no code implementations • NAACL (TextGraphs) 2021 • Martin Schmitt, Leonardo F. R. Ribeiro, Philipp Dufter, Iryna Gurevych, Hinrich Schütze
We present Graformer, a novel Transformer-based encoder-decoder architecture for graph-to-text generation.
Ranked #5 on KG-to-Text Generation on AGENDA
1 code implementation • EMNLP (DeeLIO) 2020 • Anne Lauscher, Olga Majewska, Leonardo F. R. Ribeiro, Iryna Gurevych, Nikolai Rozanov, Goran Glavaš
Following the major success of neural language models (LMs) such as BERT or GPT-2 on a variety of language understanding tasks, recent work focused on injecting (structured) knowledge from external resources into these models.
1 code implementation • ACL 2020 • Ji-Ung Lee, Christian M. Meyer, Iryna Gurevych
Existing approaches to active learning maximize the system performance by sampling unlabeled instances for annotation that yield the most efficient training.
3 code implementations • EACL 2021 • Jonas Pfeiffer, Aishwarya Kamath, Andreas Rücklé, Kyunghyun Cho, Iryna Gurevych
We show that by separating the two stages, i. e., knowledge extraction and knowledge composition, the classifier can effectively exploit the representations learned from multiple tasks in a non-destructive manner.
no code implementations • 1 May 2020 • Jonas Pfeiffer, Edwin Simpson, Iryna Gurevych
We compare different models for low resource multi-task sequence tagging that leverage dependencies between label sequences for different tasks.
1 code implementation • ACL 2020 • Prasetya Ajie Utama, Nafise Sadat Moosavi, Iryna Gurevych
Models for natural language understanding (NLU) tasks often rely on the idiosyncratic biases of the dataset, which make them brittle against test cases outside the training distribution.
no code implementations • EACL 2021 • Mohsen Mesgar, Edwin Simpson, Iryna Gurevych
Neural models for response generation produce responses that are semantically plausible but not necessarily factually consistent with facts describing the speaker's persona.
3 code implementations • EMNLP 2020 • Jonas Pfeiffer, Ivan Vulić, Iryna Gurevych, Sebastian Ruder
The main goal behind state-of-the-art pre-trained multilingual models such as multilingual BERT and XLM-R is enabling and bootstrapping NLP applications in low-resource languages through zero-shot or few-shot cross-lingual transfer.
Ranked #5 on Cross-Lingual Transfer on XCOPA (using extra training data)
no code implementations • EMNLP 2020 • Ilia Kuznetsov, Iryna Gurevych
Deep pre-trained contextualized encoders like BERT (Delvin et al., 2019) demonstrate remarkable performance on a range of downstream tasks.
1 code implementation • NAACL 2021 • Benjamin Schiller, Johannes Daxenberger, Iryna Gurevych
In this work, we train a language model for argument generation that can be controlled on a fine-grained level to generate sentence-level arguments for a given topic, stance, and aspect.
no code implementations • ACL 2020 • Gözde Gül Şahin, Yova Kementchedjhieva, Phillip Rust, Iryna Gurevych
To expose this problem in a new light, we introduce a challenge on learning from small data, PuzzLing Machines, which consists of Rosetta Stone puzzles from Linguistic Olympiads for high school students.
11 code implementations • EMNLP 2020 • Nils Reimers, Iryna Gurevych
The training is based on the idea that a translated sentence should be mapped to the same location in the vector space as the original sentence.
no code implementations • 28 Feb 2020 • Kevin Stowe, Leonardo Ribeiro, Iryna Gurevych
This work describes the task of metaphoric paraphrase generation, in which we are given a literal sentence and are charged with generating a metaphoric paraphrase.
1 code implementation • ICLR 2020 • Shweta Mahajan, Iryna Gurevych, Stefan Roth
Therefore, we propose a novel semi-supervised framework, which models shared information between domains and domain-specific information separately.
1 code implementation • 29 Jan 2020 • Leonardo F. R. Ribeiro, Yue Zhang, Claire Gardent, Iryna Gurevych
Recent graph-to-text models generate text from graph-based data using either global or local aggregation to learn node representations.
Ranked #1 on Graph-to-Sequence on WebNLG
1 code implementation • 6 Jan 2020 • Benjamin Schiller, Johannes Daxenberger, Iryna Gurevych
Stance Detection (StD) aims to detect an author's stance towards a certain topic or claim and has become a key component in applications like fake news detection, claim validation, and argument search.
1 code implementation • 17 Dec 2019 • Andreas Hanselowski, Iryna Gurevych
Word embeddings are rich word representations, which in combination with deep neural networks, lead to large performance gains for many NLP tasks.
no code implementations • 11 Dec 2019 • Gözde Gül Şahin, Iryna Gurevych
We show that they are able to recover the morphological input parameters, i. e., predicting the lemma (e. g., cat) or the morphological tags (e. g., Plural) when run in the reverse direction, without any significant performance drop in the forward direction, i. e., predicting the surface form (e. g., cats).
1 code implementation • 4 Dec 2019 • Edwin Simpson, Iryna Gurevych
As previous solutions based on Gaussian processes do not scale to large numbers of users, items or pairwise labels, we propose a stochastic variational inference approach that limits computational and memory costs.
no code implementations • 23 Nov 2019 • Mohsen Mesgar, Paul Youssef, Lin Li, Dominik Bierwirth, Yihao Li, Christian M. Meyer, Iryna Gurevych
Our conversational agent UKP-ATHENA assists NLP researchers in finding and exploring scientific literature, identifying relevant authors, planning or post-processing conference visits, and preparing paper submissions using a unified interface based on natural language inputs and responses.
1 code implementation • 22 Nov 2019 • Edwin Simpson, Yang Gao, Iryna Gurevych
For many NLP applications, such as question answering and summarisation, the goal is to select the best solution from a large space of candidates to meet a particular user's needs.
1 code implementation • IJCNLP 2019 • Andreas Rücklé, Nafise Sadat Moosavi, Iryna Gurevych
We show that our proposed approaches are more effective in many cases because they can utilize larger amounts of unlabeled data from cQA forums.
2 code implementations • CONLL 2019 • Andreas Hanselowski, Christian Stab, Claudia Schulz, Zile Li, Iryna Gurevych
Automated fact-checking based on machine learning is a promising approach to identify false information distributed on the web.
no code implementations • WS 2019 • Yevgeniy Puzikov, Claire Gardent, Ido Dagan, Iryna Gurevych
End-to-end neural approaches have achieved state-of-the-art performance in many natural language processing (NLP) tasks.
no code implementations • 19 Sep 2019 • Nafise Sadat Moosavi, Prasetya Ajie Utama, Andreas Rücklé, Iryna Gurevych
Finally, we show that using the coverage information is not only beneficial for improving the performance across different datasets of the same task.
no code implementations • 14 Sep 2019 • Shweta Mahajan, Teresa Botschen, Iryna Gurevych, Stefan Roth
One of the key challenges in learning joint embeddings of multiple modalities, e. g. of images and text, is to ensure coherent cross-modal semantics that generalize across datasets.
no code implementations • 10 Sep 2019 • Jonas Pfeiffer, Aishwarya Kamath, Iryna Gurevych, Sebastian Ruder
Recent research towards understanding neural networks probes models in a top-down manner, but is only able to identify model tendencies that are known a priori.
2 code implementations • IJCNLP 2019 • Florian Böhm, Yang Gao, Christian M. Meyer, Ori Shapira, Ido Dagan, Iryna Gurevych
Human evaluation experiments show that, compared to the state-of-the-art supervised-learning systems and ROUGE-as-rewards RL summarisation systems, the RL systems using our learned rewards during training generate summarieswith higher human ratings.
1 code implementation • IJCNLP 2019 • Leonardo F. R. Ribeiro, Claire Gardent, Iryna Gurevych
Generating text from graph-based data, such as Abstract Meaning Representation (AMR), is a challenging task due to the inherent difficulty in how to properly encode the structure of a graph with labeled edges.
Ranked #11 on AMR-to-Text Generation on LDC2017T10
no code implementations • IJCNLP 2019 • Jonas Pfeiffer, Christian M. Meyer, Claudia Schulz, Jan Kiesewetter, Jan Zottmann, Michael Sailer, Elisabeth Bauer, Frank Fischer, Martin R. Fischer, Iryna Gurevych
Our proposed system FAMULUS helps students learn to diagnose based on automatic feedback in virtual patient simulations, and it supports instructors in labeling training data.
60 code implementations • IJCNLP 2019 • Nils Reimers, Iryna Gurevych
However, it requires that both sentences are fed into the network, which causes a massive computational overhead: Finding the most similar pair in a collection of 10, 000 sentences requires about 50 million inference computations (~65 hours) with BERT.
Ranked #5 on Linear-Probe Classification on SentEval
1 code implementation • ACL 2020 • Mohsen Mesgar, Sebastian Bücker, Iryna Gurevych
Recent dialogue coherence models use the coherence features designed for monologue texts, e. g. nominal entities, to represent utterances and then explicitly augment them with dialogue-relevant features, e. g., dialogue act labels.
1 code implementation • 30 Jul 2019 • Yang Gao, Christian M. Meyer, Mohsen Mesgar, Iryna Gurevych
The predominant RL paradigm for summarisation learns a cross-input policy, which requires considerable time, data and parameter tuning due to the huge search spaces and the delayed rewards.
1 code implementation • IJCNLP 2019 • Max Eichler, Gözde Gül Şahin, Iryna Gurevych
We present LINSPECTOR WEB, an open source multilingual inspector to analyze word representations.
no code implementations • ACL 2019 • Tobias Falke, Leonardo F. R. Ribeiro, Prasetya Ajie Utama, Ido Dagan, Iryna Gurevych
While recent progress on abstractive summarization has led to remarkably fluent summaries, factual errors in generated summaries still severely limit their use in practice.
1 code implementation • ACL 2019 • Edwin Simpson, Erik-L{\^a}n Do Dinh, Tristan Miller, Iryna Gurevych
The inability to quantify key aspects of creative language is a frequent obstacle to natural language understanding.
2 code implementations • ACL 2019 • Nils Reimers, Benjamin Schiller, Tilman Beck, Johannes Daxenberger, Christian Stab, Iryna Gurevych
We experiment with two recent contextualized word embedding methods (ELMo and BERT) in the context of open-domain argument search.
1 code implementation • 7 Jun 2019 • Yang Gao, Christian M. Meyer, Iryna Gurevych
Interactive NLP is a promising paradigm to close the gap between automatic NLP systems and the human upper bound.
no code implementations • ACL 2019 • Claudia Schulz, Christian M. Meyer, Jan Kiesewetter, Michael Sailer, Elisabeth Bauer, Martin R. Fischer, Frank Fischer, Iryna Gurevych
Many complex discourse-level tasks can aid domain experts in their work but require costly expert annotations for data creation.
no code implementations • WS 2019 • Steffen Eger, Andreas Rücklé, Iryna Gurevych
Our motivation is to challenge the current evaluation of sentence embeddings and to provide an easy-to-access reference for future research.
1 code implementation • NAACL 2019 • Tobias Falke, Iryna Gurevych
Concept map-based multi-document summarization has recently been proposed as a variant of the traditional summarization task with graph-structured summaries.
1 code implementation • NAACL 2019 • Tristan Miller, Maria Sukhareva, Iryna Gurevych
The study of argumentation and the development of argument mining tools depends on the availability of annotated data, which is challenging to obtain in sufficient quantity and quality.
1 code implementation • 22 Apr 2019 • Dietrich Trautmann, Johannes Daxenberger, Christian Stab, Hinrich Schütze, Iryna Gurevych
In this work, we argue that the task should be performed on a more fine-grained level of sequence labeling.
1 code implementation • 5 Apr 2019 • Nils Reimers, Iryna Gurevych
We evaluate different methods that combine the three vectors from the language model in order to achieve the best possible performance in downstream NLP tasks.
1 code implementation • NAACL 2019 • Steffen Eger, Gözde Gül Şahin, Andreas Rücklé, Ji-Ung Lee, Claudia Schulz, Mohsen Mesgar, Krishnkant Swarnkar, Edwin Simpson, Iryna Gurevych
Visual modifications to text are often used to obfuscate offensive comments in social media (e. g., "! d10t") or as a writing style ("1337" in "leet speak"), among other scenarios.
1 code implementation • NAACL 2019 • Yang Gao, Steffen Eger, Ilia Kuznetsov, Iryna Gurevych, Yusuke Miyao
We then focus on the role of the rebuttal phase, and propose a novel task to predict after-rebuttal (i. e., final) scores from initial reviews and author responses.
3 code implementations • CL 2020 • Gözde Gül Şahin, Clara Vania, Ilia Kuznetsov, Iryna Gurevych
We present a reusable methodology for creation and evaluation of such tests in a multilingual setting.
1 code implementation • 7 Mar 2019 • Steffen Eger, Chao Li, Florian Netzer, Iryna Gurevych
By extrapolation, we predict that these topics will remain lead problems/approaches in their fields in the short- and mid-term.
1 code implementation • EMNLP 2018 • Steffen Eger, Paul Youssef, Iryna Gurevych
Activation functions play a crucial role in neural networks because they are the nonlinearities which have been attributed to the success story of deep learning.
1 code implementation • 26 Nov 2018 • Claudia Schulz, Christian M. Meyer, Michael Sailer, Jan Kiesewetter, Elisabeth Bauer, Frank Fischer, Martin R. Fischer, Iryna Gurevych
We aim to enable the large-scale adoption of diagnostic reasoning analysis and feedback by automating the epistemic activity identification.
1 code implementation • IJCNLP 2019 • Edwin Simpson, Iryna Gurevych
Current methods for sequence tagging, a core task in NLP, are data hungry, which motivates the use of crowdsourcing as a cheap way to obtain labelled data.
1 code implementation • WS 2018 • Yevgeniy Puzikov, Iryna Gurevych
E2E NLG Challenge is a shared task on generating restaurant descriptions from sets of key-value pairs.
Ranked #11 on Data-to-Text Generation on E2E NLG Challenge
1 code implementation • EMNLP 2018 • Daniil Sorokin, Iryna Gurevych
Most approaches to Knowledge Base Question Answering are based on semantic parsing.
no code implementations • EMNLP 2018 • Beto Boullosa, Richard Eckart de Castilho, Naveen Kumar, Jan-Christoph Klie, Iryna Gurevych
Annotating entity mentions and linking them to a knowledge resource are essential tasks in many domains.
1 code implementation • WS 2018 • Gil Rocha, Christian Stab, Henrique Lopes Cardoso, Iryna Gurevych
Argument mining aims to detect and identify argument structures from textual resources.
1 code implementation • WS 2018 • Teresa Botschen, Daniil Sorokin, Iryna Gurevych
Common-sense argumentative reasoning is a challenging task that requires holistic understanding of the argumentation where external knowledge about the world is hypothesized to play a key role.
1 code implementation • WS 2018 • Steffen Eger, Andreas R{\"u}ckl{\'e}, Iryna Gurevych
We consider unsupervised cross-lingual transfer on two tasks, viz., sentence-level argumentation mining and standard POS tagging.
no code implementations • CONLL 2018 • Ilia Kuznetsov, Iryna Gurevych
Thematic role hierarchy is a widely used linguistic tool to describe interactions between semantic roles and their syntactic realizations.
no code implementations • EMNLP 2018 • Christian Stab, Tristan Miller, Benjamin Schiller, Pranav Rai, Iryna Gurevych
Argument mining is a core technology for automating argument search in large document collections.
1 code implementation • WS 2018 • Andreas Hanselowski, Hao Zhang, Zile Li, Daniil Sorokin, Benjamin Schiller, Claudia Schulz, Iryna Gurevych
The Fact Extraction and VERification (FEVER) shared task was launched to support the development of systems able to verify claims by extracting supporting or refuting facts from raw text.
1 code implementation • EMNLP 2018 • Yang Gao, Christian M. Meyer, Iryna Gurevych
The merit of preference-based interactive summarisation is that preferences are easier for users to provide than reference summaries.
1 code implementation • COLING 2018 • Daniil Sorokin, Iryna Gurevych
The most approaches to Knowledge Base Question Answering are based on semantic parsing.
Ranked #1 on Knowledge Base Question Answering on WebQSP-WD
1 code implementation • COLING 2018 • Erik-L{\^a}n Do Dinh, Steffen Eger, Iryna Gurevych
In this paper we investigate multi-task learning for related non-literal language phenomena.
no code implementations • COLING 2018 • Erik-L{\^a}n Do Dinh, Steffen Eger, Iryna Gurevych
In this paper, we tackle four different tasks of non-literal language classification: token and construction level metaphor detection, classification of idiomatic use of infinitive-verb compounds, and classification of non-literal particle verbs.
1 code implementation • COLING 2018 • Ilia Kuznetsov, Iryna Gurevych
We examine the effect of lemmatization and POS typing on word embedding performance in a novel resource-based evaluation scenario, as well as on standard similarity benchmarks.
1 code implementation • COLING 2018 • Jan-Christoph Klie, Michael Bugert, Beto Boullosa, Richard Eckart de Castilho, Iryna Gurevych
We introduce INCEpTION, a new annotation platform for tasks including interactive and semantic annotation (e. g., concept linking, fact linking, knowledge base population, semantic frame annotation).
1 code implementation • COLING 2018 • Andreas Hanselowski, Avinesh PVS, Benjamin Schiller, Felix Caspelherr, Debanjan Chaudhuri, Christian M. Meyer, Iryna Gurevych
To date, there is no in-depth analysis paper to critically discuss FNC-1{'}s experimental setup, reproduce the results, and draw conclusions for next-generation stance classification methods.
1 code implementation • COLING 2018 • Steffen Eger, Johannes Daxenberger, Christian Stab, Iryna Gurevych
Argumentation mining (AM) requires the identification of complex discourse structures and has lately been applied with success monolingually.
no code implementations • WS 2018 • Yevgeniy Puzikov, Iryna Gurevych
Surface Realization Shared Task 2018 is a workshop on generating sentences from lemmatized sets of dependency triples.
1 code implementation • COLING 2018 • Lisa Beinborn, Teresa Botschen, Iryna Gurevych
This survey discusses how recent developments in multimodal processing facilitate conceptual grounding of language.
7 code implementations • 13 Jun 2018 • Andreas Hanselowski, Avinesh PVS, Benjamin Schiller, Felix Caspelherr, Debanjan Chaudhuri, Christian M. Meyer, Iryna Gurevych
To date, there is no in-depth analysis paper to critically discuss FNC-1's experimental setup, reproduce the results, and draw conclusions for next-generation stance classification methods.
1 code implementation • TACL 2018 • Edwin Simpson, Iryna Gurevych
We introduce a scalable Bayesian preference learning method for identifying convincing arguments in the absence of gold-standard rat- ings or rankings.