Search Results for author: Anne Lauscher

Found 45 papers, 29 papers with code

An Argument-Annotated Corpus of Scientific Publications

no code implementations WS 2018 Anne Lauscher, Goran Glava{\v{s}}, Simone Paolo Ponzetto

We analyze the annotated argumentative structures and investigate the relations between argumentation and other rhetorical aspects of scientific writing, such as discourse roles and citation contexts.

Argument Mining

Are We Consistently Biased? Multidimensional Analysis of Biases in Distributional Word Vectors

1 code implementation SEMEVAL 2019 Anne Lauscher, Goran Glavaš

In this work, we present a systematic study of biases encoded in distributional word vector spaces: we analyze how consistent the bias effects are across languages, corpora, and embedding models.

Cross-Lingual Transfer Word Embeddings

MinScIE: Citation-centered Open Information Extraction

1 code implementation Joint Conference on Digital Libraries (JCDL) 2019 Anne Lauscher, Yide Song, Kiril Gashteovski

Acknowledging the importance of citations in scientific literature, in this work we present MinScIE, an Open Information Extraction system which provides structured knowledge enriched with semantic information about citations.

Open Information Extraction

Specializing Unsupervised Pretraining Models for Word-Level Semantic Similarity

1 code implementation COLING 2020 Anne Lauscher, Ivan Vulić, Edoardo Maria Ponti, Anna Korhonen, Goran Glavaš

In this work, we complement such distributional knowledge with external lexical knowledge, that is, we integrate the discrete knowledge on word-level semantic similarity into pretraining.

Language Modelling Lexical Simplification +7

A General Framework for Implicit and Explicit Debiasing of Distributional Word Vector Spaces

4 code implementations13 Sep 2019 Anne Lauscher, Goran Glavaš, Simone Paolo Ponzetto, Ivan Vulić

Moreover, we successfully transfer debiasing models, by means of cross-lingual embedding spaces, and remove or attenuate biases in distributional word vector spaces of languages that lack readily available bias specifications.

Word Embeddings

From Zero to Hero: On the Limitations of Zero-Shot Cross-Lingual Transfer with Multilingual Transformers

no code implementations1 May 2020 Anne Lauscher, Vinit Ravishankar, Ivan Vulić, Goran Glavaš

Massively multilingual transformers pretrained with language modeling objectives (e. g., mBERT, XLM-R) have become a de facto default transfer paradigm for zero-shot cross-lingual transfer in NLP, offering unmatched transfer performance.

Cross-Lingual Word Embeddings Dependency Parsing +6

Common Sense or World Knowledge? Investigating Adapter-Based Knowledge Injection into Pretrained Transformers

1 code implementation EMNLP (DeeLIO) 2020 Anne Lauscher, Olga Majewska, Leonardo F. R. Ribeiro, Iryna Gurevych, Nikolai Rozanov, Goran Glavaš

Following the major success of neural language models (LMs) such as BERT or GPT-2 on a variety of language understanding tasks, recent work focused on injecting (structured) knowledge from external resources into these models.

Common Sense Reasoning World Knowledge

The OpenCitations Data Model

1 code implementation25 May 2020 Marilena Daquino, Silvio Peroni, David Shotton, Giovanni Colavizza, Behnam Ghavimi, Anne Lauscher, Philipp Mayr, Matteo Romanello, Philipp Zumstein

A variety of schemas and ontologies are currently used for the machine-readable description of bibliographic entities and citations.

Digital Libraries

Rhetoric, Logic, and Dialectic: Advancing Theory-based Argument Quality Assessment in Natural Language Processing

1 code implementation COLING 2020 Anne Lauscher, Lily Ng, Courtney Napoles, Joel Tetreault

Though preceding work in computational argument quality (AQ) mostly focuses on assessing overall AQ, researchers agree that writers would benefit from feedback targeting individual dimensions of argumentation theory.

Creating a Domain-diverse Corpus for Theory-based Argument Quality Assessment

1 code implementation COLING (ArgMining) 2020 Lily Ng, Anne Lauscher, Joel Tetreault, Courtney Napoles

Computational models of argument quality (AQ) have focused primarily on assessing the overall quality or just one specific characteristic of an argument, such as its convincingness or its clarity.

AraWEAT: Multidimensional Analysis of Biases in Arabic Word Embeddings

no code implementations COLING (WANLP) 2020 Anne Lauscher, Rafik Takieddin, Simone Paolo Ponzetto, Goran Glavaš

Our analysis yields several interesting findings, e. g., that implicit gender bias in embeddings trained on Arabic news corpora steadily increases over time (between 2007 and 2017).

Word Embeddings

Self-Supervised Learning for Visual Summary Identification in Scientific Publications

no code implementations21 Dec 2020 Shintaro Yamamoto, Anne Lauscher, Simone Paolo Ponzetto, Goran Glavaš, Shigeo Morishima

Providing visual summaries of scientific publications can increase information access for readers and thereby help deal with the exponential growth in the number of scientific publications.

Self-Supervised Learning

Scientia Potentia Est -- On the Role of Knowledge in Computational Argumentation

no code implementations1 Jul 2021 Anne Lauscher, Henning Wachsmuth, Iryna Gurevych, Goran Glavaš

Despite extensive research efforts in recent years, computational argumentation (CA) remains one of the most challenging areas of natural language processing.

Common Sense Reasoning Natural Language Understanding

Diachronic Analysis of German Parliamentary Proceedings: Ideological Shifts through the Lens of Political Biases

1 code implementation13 Aug 2021 Tobias Walter, Celina Kirschner, Steffen Eger, Goran Glavaš, Anne Lauscher, Simone Paolo Ponzetto

We analyze bias in historical corpora as encoded in diachronic distributional semantic models by focusing on two specific forms of bias, namely a political (i. e., anti-communism) and racist (i. e., antisemitism) one.

Diachronic Word Embeddings Word Embeddings

Sustainable Modular Debiasing of Language Models

no code implementations Findings (EMNLP) 2021 Anne Lauscher, Tobias Lüken, Goran Glavaš

Unfair stereotypical biases (e. g., gender, racial, or religious biases) encoded in modern pretrained language models (PLMs) have negative ethical implications for widespread adoption of state-of-the-art language technology.

Fairness Language Modelling

DS-TOD: Efficient Domain Specialization for Task Oriented Dialog

1 code implementation15 Oct 2021 Chia-Chien Hung, Anne Lauscher, Simone Paolo Ponzetto, Goran Glavaš

Recent work has shown that self-supervised dialog-specific pretraining on large conversational datasets yields substantial gains over traditional language modeling (LM) pretraining in downstream task-oriented dialog (TOD).

dialog state tracking Language Modelling +2

Welcome to the Modern World of Pronouns: Identity-Inclusive Natural Language Processing beyond Gender

no code implementations COLING 2022 Anne Lauscher, Archie Crowley, Dirk Hovy

Based on our observations and ethical considerations, we define a series of desiderata for modeling pronouns in language technology.

Fair and Argumentative Language Modeling for Computational Argumentation

1 code implementation ACL 2022 Carolin Holtermann, Anne Lauscher, Simone Paolo Ponzetto

We employ our resource to assess the effect of argumentative fine-tuning and debiasing on the intrinsic bias found in transformer-based language models using a lightweight adapter-based approach that is more sustainable and parameter-efficient than full fine-tuning.

Language Modelling

Multi2WOZ: A Robust Multilingual Dataset and Conversational Pretraining for Task-Oriented Dialog

1 code implementation NAACL 2022 Chia-Chien Hung, Anne Lauscher, Ivan Vulić, Simone Paolo Ponzetto, Goran Glavaš

We then introduce a new framework for multilingual conversational specialization of pretrained language models (PrLMs) that aims to facilitate cross-lingual transfer for arbitrary downstream TOD tasks.

Cross-Lingual Transfer dialog state tracking +1

On the Limitations of Sociodemographic Adaptation with Transformers

1 code implementation1 Aug 2022 Chia-Chien Hung, Anne Lauscher, Dirk Hovy, Simone Paolo Ponzetto, Goran Glavaš

We adapt the language representations for the sociodemographic dimensions of gender and age, using continuous language modeling and dynamic multi-task learning for adaptation, where we couple language modeling with the prediction of a sociodemographic class.

Language Modelling Multi-Task Learning

Back to the Future: On Potential Histories in NLP

no code implementations12 Oct 2022 Zeerak Talat, Anne Lauscher

Machine learning and NLP require the construction of datasets to train and fine-tune models.

Can Demographic Factors Improve Text Classification? Revisiting Demographic Adaptation in the Age of Transformers

1 code implementation13 Oct 2022 Chia-Chien Hung, Anne Lauscher, Dirk Hovy, Simone Paolo Ponzetto, Goran Glavaš

Previous work showed that incorporating demographic factors can consistently improve performance for various NLP tasks with traditional NLP models.

Language Modelling Multi-Task Learning +2

SocioProbe: What, When, and Where Language Models Learn about Sociodemographics

1 code implementation8 Nov 2022 Anne Lauscher, Federico Bianchi, Samuel Bowman, Dirk Hovy

Our results show that PLMs do encode these sociodemographics, and that this knowledge is sometimes spread across the layers of some of the tested PLMs.

Bridging Fairness and Environmental Sustainability in Natural Language Processing

no code implementations8 Nov 2022 Marius Hessenthaler, Emma Strubell, Dirk Hovy, Anne Lauscher

Fairness and environmental impact are important research directions for the sustainable development of artificial intelligence.

Dimensionality Reduction Fairness +4

What about em? How Commercial Machine Translation Fails to Handle (Neo-)Pronouns

no code implementations25 May 2023 Anne Lauscher, Debora Nozza, Archie Crowley, Ehm Miltersen, Dirk Hovy

As 3rd-person pronoun usage shifts to include novel forms, e. g., neopronouns, we need more research on identity-inclusive NLP.

Machine Translation Translation

Stereotypes and Smut: The (Mis)representation of Non-cisgender Identities by Text-to-Image Models

no code implementations26 May 2023 Eddie L. Ungless, Björn Ross, Anne Lauscher

Cutting-edge image generation has been praised for producing high-quality images, suggesting a ubiquitous future in a variety of applications.

Image Generation

Sensitivity, Performance, Robustness: Deconstructing the Effect of Sociodemographic Prompting

1 code implementation13 Sep 2023 Tilman Beck, Hendrik Schuff, Anne Lauscher, Iryna Gurevych

However, the available NLP literature disagrees on the efficacy of this technique - it remains unclear for which tasks and scenarios it can help, and the role of the individual factors in sociodemographic prompting is still unexplored.

Hate Speech Detection Zero-Shot Learning

AutomaTikZ: Text-Guided Synthesis of Scientific Vector Graphics with TikZ

1 code implementation30 Sep 2023 Jonas Belouadi, Anne Lauscher, Steffen Eger

To address this, we propose the use of TikZ, a well-known abstract graphics language that can be compiled to vector graphics, as an intermediate representation of scientific figures.

Language Modelling Large Language Model +2

ScaLearn: Simple and Highly Parameter-Efficient Task Transfer by Learning to Scale

1 code implementation2 Oct 2023 Markus Frohmann, Carolin Holtermann, Shahed Masoudian, Anne Lauscher, Navid Rekabsaz

We introduce ScaLearn, a simple and highly parameter-efficient two-stage MTL method that capitalizes on the knowledge of the source tasks by learning a minimal set of scaling parameters that enable effective knowledge transfer to a target task.

Multi-Task Learning

Values, Ethics, Morals? On the Use of Moral Concepts in NLP Research

no code implementations21 Oct 2023 Karina Vida, Judith Simon, Anne Lauscher

For instance, we analyse what ethical theory an approach is based on, how this decision is justified, and what implications it entails.

Ethics Philosophy

Exploring Jiu-Jitsu Argumentation for Writing Peer Review Rebuttals

1 code implementation7 Nov 2023 Sukannya Purkayastha, Anne Lauscher, Iryna Gurevych

In this work, we are the first to explore Jiu-Jitsu argumentation for peer review by proposing the novel task of attitude and theme-guided rebuttal generation.

Sentence

What the Weight?! A Unified Framework for Zero-Shot Knowledge Composition

1 code implementation23 Jan 2024 Carolin Holtermann, Markus Frohmann, Navid Rekabsaz, Anne Lauscher

The knowledge encapsulated in a model is the core factor determining its final performance on downstream tasks.

Benchmarking

Evaluating the Elementary Multilingual Capabilities of Large Language Models with MultiQ

1 code implementation6 Mar 2024 Carolin Holtermann, Paul Röttger, Timm Dill, Anne Lauscher

Therefore, in this paper, we investigate the basic multilingual capabilities of state-of-the-art open LLMs beyond their intended use.

Open-Ended Question Answering

Argument Quality Assessment in the Age of Instruction-Following Large Language Models

no code implementations24 Mar 2024 Henning Wachsmuth, Gabriella Lapesa, Elena Cabrio, Anne Lauscher, Joonsuk Park, Eva Maria Vecchi, Serena Villata, Timon Ziegenbein

The computational treatment of arguments on controversial issues has been subject to extensive NLP research, due to its envisioned impact on opinion formation, decision making, writing education, and the like.

Decision Making Instruction Following

Robust Pronoun Use Fidelity with English LLMs: Are they Reasoning, Repeating, or Just Biased?

1 code implementation4 Apr 2024 Vagrant Gautam, Eileen Bingert, Dawei Zhu, Anne Lauscher, Dietrich Klakow

We find that while models can mostly faithfully reuse previously-specified pronouns in the presence of no distractors, they are significantly worse at processing she/her/her, singular they and neopronouns.

Sentence

DS-TOD: Efficient Domain Specialization for Task-Oriented Dialog

1 code implementation Findings (ACL) 2022 Chia-Chien Hung, Anne Lauscher, Simone Ponzetto, Goran Glavaš

Recent work has shown that self-supervised dialog-specific pretraining on large conversational datasets yields substantial gains over traditional language modeling (LM) pretraining in downstream task-oriented dialog (TOD).

dialog state tracking Language Modelling +2

From Zero to Hero: On the Limitations of Zero-Shot Language Transfer with Multilingual Transformers

no code implementations EMNLP 2020 Anne Lauscher, Vinit Ravishankar, Ivan Vuli{\'c}, Goran Glava{\v{s}}

Massively multilingual transformers (MMTs) pretrained via language modeling (e. g., mBERT, XLM-R) have become a default paradigm for zero-shot language transfer in NLP, offering unmatched transfer performance.

Cross-Lingual Word Embeddings Dependency Parsing +5

Cannot find the paper you are looking for? You can Submit a new open access paper.