no code implementations • INLG (ACL) 2020 • Robin Rojowiec, Jana Götze, Philipp Sadler, Henrik Voigt, Sina Zarrieß, David Schlangen
We find that it is, and investigate several simple baselines, taking these from the related task of image captioning.
1 code implementation • COLING 2022 • Sharid Loáiciga, Anne Beyer, David Schlangen
Recent research shows that pre-trained language models, built to generate text conditioned on some context, learn to encode syntactic knowledge to a certain degree.
no code implementations • ACL (mmsr, IWCS) 2021 • Casey Kennington, David Schlangen
We offer a fine-grained information state annotation scheme that follows directly from the Incremental Unit abstract model of dialogue processing when used within a multimodal, co-located, interactive setting.
no code implementations • ACL (mmsr, IWCS) 2021 • Sharid Loáiciga, Simon Dobnik, David Schlangen
With this paper, we intend to start a discussion on the annotation of referential phenomena in situated dialogue.
no code implementations • NAACL (ALVR) 2021 • Sharid Loáiciga, Simon Dobnik, David Schlangen
We argue that there is still significant room for corpora that increase the complexity of both visual and linguistic domains and which capture different varieties of perceptual and conversational contexts.
no code implementations • COLING (CRAC) 2022 • Sharid Loáiciga, Simon Dobnik, David Schlangen
We present a first release of 500 documents from the multimodal corpus Tell-me-more (Ilinykh et al., 2019) annotated with coreference information according to the ARRAU guidelines (Poesio et al., 2021).
no code implementations • 11 Apr 2025 • Nicola Horst, Davide Mazzaccara, Antonia Schmidt, Michael Sullivan, Filippo Momentè, Luca Franceschetti, Philipp Sadler, Sherzod Hakimov, Alberto Testoni, Raffaella Bernardi, Raquel Fernández, Alexander Koller, Oliver Lemon, David Schlangen, Mario Giulianelli, Alessandro Suglia
Are we running out of learning signal?
no code implementations • 20 Feb 2025 • Filippo Momentè, Alessandro Suglia, Mario Giulianelli, Ambra Ferrari, Alexander Koller, Oliver Lemon, David Schlangen, Raquel Fernández, Raffaella Bernardi
We examine three evaluation paradigms: large question-answering benchmarks (e. g., MMLU and BBH), interactive games (e. g., Signalling Games or Taboo), and cognitive tests (e. g., for working memory or theory of mind).
no code implementations • 17 Feb 2025 • Sherzod Hakimov, Lara Pfennigschmidt, David Schlangen
This study utilizes the game Codenames as a benchmarking tool to evaluate large language models (LLMs) with respect to specific linguistic and cognitive skills.
no code implementations • 17 Feb 2025 • Jonathan Jordan, Sherzod Hakimov, David Schlangen
Large language models (LLMs) have risen to prominence as 'chatbots' for users to interact via natural language.
no code implementations • 1 Jan 2025 • Casey Kennington, Pierre Lison, David Schlangen
Efforts towards endowing robots with the ability to speak have benefited from recent advancements in NLP, in particular large language models.
no code implementations • 17 Sep 2024 • Chalamalasetti Kranti, Sherzod Hakimov, David Schlangen
While there has been a lot of research recently on robots in household environments, at the present time, most robots in existence can be found on shop floors, and most interactions between humans and robots happen there.
1 code implementation • 29 Aug 2024 • Luka Borec, Philipp Sadler, David Schlangen
This work analyses the text memorization behavior of large language models (LLMs) when subjected to nucleus sampling.
no code implementations • 18 Jul 2024 • David Schlangen
Natural Language Processing has moved rather quickly from modelling specific tasks to taking more general pre-trained models and fine-tuning them for specific tasks, to a point where we now have what appear to be inherently generalist models.
1 code implementation • 26 Jun 2024 • Anna Bavaresco, Raffaella Bernardi, Leonardo Bertolazzi, Desmond Elliott, Raquel Fernández, Albert Gatt, Esam Ghaleb, Mario Giulianelli, Michael Hanna, Alexander Koller, André F. T. Martins, Philipp Mondorf, Vera Neplenbroek, Sandro Pezzelle, Barbara Plank, David Schlangen, Alessandro Suglia, Aditya K Surikuchi, Ece Takmaz, Alberto Testoni
There is an increasing trend towards evaluating NLP models with LLMs instead of human judgments, raising questions about the validity of these evaluations, as well as their reproducibility in the case of proprietary models.
no code implementations • 25 Jun 2024 • Chalamalasetti Kranti, Sherzod Hakimov, David Schlangen
In the Minecraft Collaborative Building Task, two players collaborate: an Architect (A) provides instructions to a Builder (B) to assemble a specified structure using 3D blocks.
no code implementations • 20 Jun 2024 • Nidhir Bhavsar, Jonathan Jordan, Sherzod Hakimov, David Schlangen
But what makes the model perform well?
no code implementations • 20 Jun 2024 • Sherzod Hakimov, Yerkezhan Abdullayeva, Kushal Koshti, Antonia Schmidt, Yan Weiser, Anne Beyer, David Schlangen
On further analysis, we find that the exceptional deep captioning capabilities of the largest models drive some of the performance.
1 code implementation • 12 Jun 2024 • Isidora Jeknić, David Schlangen, Alexander Koller
Collaboration is an integral part of human dialogue.
no code implementations • 31 May 2024 • Anne Beyer, Kranti Chalamalasetti, Sherzod Hakimov, Brielen Madureira, Philipp Sadler, David Schlangen
In this paper, we take one of the proposed frameworks for setting up such game-play environments, and further test its usefulness as an evaluation instrument, along a number of dimensions: We show that it can easily keep up with new developments while avoiding data contamination, we show that the tests implemented within it are not yet saturated (human performance is substantially higher than that of even the best models), and we show that it lends itself to investigating additional questions, such as the impact of the prompting language on performance.
no code implementations • 2 May 2024 • Brielen Madureira, David Schlangen
Active participation in a conversation is key to building common ground, since understanding is jointly tailored by producers and recipients.
1 code implementation • 26 Mar 2024 • Philipp Sadler, Sherzod Hakimov, David Schlangen
In collaborative goal-oriented settings, the participants are not only interested in achieving a successful outcome, but do also implicitly negotiate the effort they put into the interaction (by adapting to each other).
1 code implementation • 20 Feb 2024 • Brielen Madureira, Patrick Kahardipraja, David Schlangen
Incremental models that process sentences one token at a time will sometimes encounter points where more than one interpretation is possible.
no code implementations • 7 Feb 2024 • Philipp Sadler, Sherzod Hakimov, David Schlangen
Albrecht and Stone (2018) state that modeling of changing behaviors remains an open problem "due to the essentially unconstrained nature of what other agents may do".
1 code implementation • 30 Jan 2024 • Brielen Madureira, David Schlangen
Clarification requests are a mechanism to help solve communication problems, e. g. due to ambiguity or underspecification, in instruction-following interactions.
no code implementations • 27 Oct 2023 • David Schlangen
Natural Language Processing prides itself to be an empirically-minded, if not outright empiricist field, and yet lately it seems to get itself into essentialist debates on issues of meaning and measurement ("Do Large Language Models Understand Language, And If So, How Much?").
1 code implementation • 27 Oct 2023 • Brielen Madureira, Pelin Çelikkol, David Schlangen
In NLP, incremental processors produce output in instalments, based on incoming prefixes of the linguistic input.
no code implementations • 11 Aug 2023 • Fabian Galetzka, Anne Beyer, David Schlangen
In this survey, we interpret Grice's maxims of cooperative conversation from the perspective of this specific research area and systematize the literature under the aspect of what makes a contribution appropriate: A neural conversation model has to be fluent, informative, consistent, coherent, and follow social norms.
1 code implementation • 28 Jul 2023 • Brielen Madureira, Patrick Kahardipraja, David Schlangen
Incremental dialogue model components produce a sequence of output prefixes based on incoming input.
no code implementations • 4 Jun 2023 • Brielen Madureira, David Schlangen
Instruction Clarification Requests are a mechanism to solve communication problems, which is very functional in instruction-following interactions.
1 code implementation • 24 May 2023 • Philipp Sadler, David Schlangen
NLP tasks are typically defined extensionally through datasets containing example instantiations (e. g., pairs of image i and text t), but motivated intensionally through capabilities invoked in verbal descriptions of the task (e. g., "t is a description of i, for which the content of i needs to be recognised and understood").
1 code implementation • 23 May 2023 • Sherzod Hakimov, David Schlangen
Specifically, we investigate the performance of open-source, open-access language models against GPT-3 on five vision-language tasks when given textually-encoded visual information.
1 code implementation • 22 May 2023 • Kranti Chalamalasetti, Jana Götze, Sherzod Hakimov, Brielen Madureira, Philipp Sadler, David Schlangen
Recent work has proposed a methodology for the systematic evaluation of "Situated Language Understanding Agents"-agents that operate in rich linguistic and non-linguistic contexts-through testing them in carefully constructed interactive settings.
1 code implementation • 22 May 2023 • Philipp Sadler, Sherzod Hakimov, David Schlangen
The ability to pick up on language signals in an ongoing interaction is crucial for future machine learning models to collaborate and interact with humans naturally.
1 code implementation • 18 May 2023 • Patrick Kahardipraja, Brielen Madureira, David Schlangen
RNNs are fast but monotonic (cannot correct earlier output, which can be necessary in incremental processing).
no code implementations • 14 Apr 2023 • David Schlangen
I argue that such tests need to be complemented with tests of language use embedded in a practice, to arrive at a more comprehensive evaluation of "artificial language understanding".
1 code implementation • 28 Feb 2023 • Brielen Madureira, David Schlangen
In visual instruction-following dialogue games, players can engage in repair mechanisms in face of an ambiguous or underspecified instruction that cannot be fully mapped to actions in the world.
no code implementations • 16 Feb 2023 • David Schlangen
Even in our increasingly text-intensive times, the primary site of language use is situated, co-present interaction.
no code implementations • CLASP 2022 • David Schlangen
The striking recent advances in eliciting seemingly meaningful language behaviour from language-only machine learning models have only made more apparent, through the surfacing of clear limitations, the need to go beyond the language-only mode and to ground these models "in the world".
1 code implementation • ACL 2022 • Brielen Madureira, David Schlangen
Our conclusion is that the ability to make the distinction between shared and privately known statements along the dialogue is moderately present in the analysed models, but not always incrementally consistent, which may partially be due to the limited need for grounding interactions in the original task.
no code implementations • LREC 2022 • Jana Götze, Maike Paetzel-Prüsmann, Wencke Liermann, Tim Diekmann, David Schlangen
This paper presents the slurk software, a lightweight interaction server for setting up dialog data collections and running experiments.
1 code implementation • EMNLP 2021 • Patrick Kahardipraja, Brielen Madureira, David Schlangen
In this work, we examine the feasibility of LT for incremental NLU in English.
1 code implementation • ACL 2021 • Fabian Galetzka, Jewgeni Rose, David Schlangen, Jens Lehmann
To improve the coherence and knowledge retrieval capabilities of non-task-oriented dialogue systems, recent Transformer-based models aim to integrate fixed background context.
1 code implementation • NAACL 2021 • Anne Beyer, Sharid Loáiciga, David Schlangen
Coherent discourse is distinguished from a mere collection of utterances by the satisfaction of a diverse set of constraints, for example choice of expression, logical relation between denoted events, and implicit compatibility with world-knowledge.
1 code implementation • EMNLP 2020 • Brielen Madureira, David Schlangen
While humans process language incrementally, the best language encoders currently used in NLP do not.
no code implementations • ICML Workshop LaReL 2020 • Brielen Madureira, David Schlangen
A suitable state representation is a fundamental part of the learning process in Reinforcement Learning.
no code implementations • ACL 2021 • David Schlangen
It has become a common pattern in our field: One group introduces a language task, exemplified by a dataset, which they argue is challenging enough to serve as a benchmark.
1 code implementation • LREC 2020 • Fabian Galetzka, Chukwuemeka U. Eneh, David Schlangen
Fully data driven Chatbots for non-goal oriented dialogues are known to suffer from inconsistent behaviour across their turns, stemming from a general difficulty in controlling parameters like their assumed background personality and knowledge of facts.
no code implementations • WS 2019 • Philipp Sadler, Tatjana Scheffler, David Schlangen
Learned dynamic weighting of the conditioning signal (attention) has been shown to improve neural language generation in a variety of settings.
no code implementations • WS 2019 • Nikolai Ilinykh, Sina Zarrie{\ss}, David Schlangen
We present a dataset consisting of what we call image description sequences, which are multi-sentence descriptions of the contents of an image.
no code implementations • WS 2019 • Nazia Attari, Martin Heckmann, David Schlangen
Despite recent attempts in the field of explainable AI to go beyond black box prediction models, typically already the training data for supervised machine learning is collected in a manner that treats the annotator as a {``}black box{''}, the internal workings of which remains unobserved.
no code implementations • 29 Aug 2019 • David Schlangen
Where early work on dialogue in Computational Linguistics put much emphasis on dialogue structure and its relation to the mental states of the dialogue participants (e. g., Allen 1979, Grosz & Sidner 1986), current work mostly reduces dialogue to the task of producing at any one time a next utterance; e. g. in neural chatbot or Visual Dialogue settings.
no code implementations • 28 Aug 2019 • David Schlangen
"This paper introduces a new task and a new dataset", "we improve the state of the art in X by Y" -- it is rare to find a current natural language processing paper (or AI paper more generally) that does not contain such statements.
no code implementations • 11 Jul 2019 • Nikolai Ilinykh, Sina Zarrieß, David Schlangen
Building computer systems that can converse about their visual environment is one of the oldest concerns of research in Artificial Intelligence and Computational Linguistics (see, for example, Winograd's 1972 SHRDLU system).
no code implementations • ACL 2019 • Sina Zarrieß, David Schlangen
Zero-shot learning in Language & Vision is the task of correctly labelling (or naming) objects of novel categories.
no code implementations • WS 2019 • David Schlangen
Propelling, and propelled by, the {``}deep learning revolution{''}, recent years have seen the introduction of ever larger corpora of images annotated with natural language expressions.
no code implementations • 15 Apr 2019 • David Schlangen
Propelling, and propelled by, the "deep learning revolution", recent years have seen the introduction of ever larger corpora of images annotated with natural language expressions.
no code implementations • WS 2018 • Sina Zarrie{\ss}, David Schlangen
In this work, we assess decoding strategies for referring expression generation with neural models.
no code implementations • WS 2018 • Nikolai Ilinykh, Sina Zarrie{\ss}, David Schlangen
Image captioning models are typically trained on data that is collected from people who are asked to describe an image, without being given any further task context.
no code implementations • WS 2018 • Sina Zarrie{\ss}, David Schlangen
Modeling traditional NLG tasks with data-driven techniques has been a major focus of research in NLG in the past decade.
no code implementations • IJCNLP 2017 • Ting Han, David Schlangen
While language conveys meaning largely symbolically, actual communication acts typically contain iconic elements as well: People gesture while they speak, or may even draw sketches while explaining something.
no code implementations • IJCNLP 2017 • Ting Han, Julian Hough, David Schlangen
When giving descriptions, speakers often signify object shape or size with hand gestures.
no code implementations • WS 2017 • Sina Zarrie{\ss}, M. Soledad L{\'o}pez Gambino, David Schlangen
Current referring expression generation systems mostly deliver their output as one-shot, written expressions.
no code implementations • EMNLP 2017 • Sina Zarrie{\ss}, David Schlangen
Corpora of referring expressions paired with their visual referents are a good source for learning word meanings directly grounded in visual representations.
no code implementations • WS 2017 • Soledad L{\'o}pez Gambino, Sina Zarrie{\ss}, David Schlangen
A common convention in graphical user interfaces is to indicate a {``}wait state{''}, for example while a program is preparing a response, through a changed cursor state or a progress bar.
no code implementations • ACL 2017 • Sina Zarrie{\ss}, David Schlangen
We present a model that learns individual predictors for object names that link visual and distributional aspects of word meaning during training.
no code implementations • EACL 2017 • Sina Zarrie{\ss}, David Schlangen
There has recently been a lot of work trying to use images of referents of words for improving vector space meaning representations derived from text.
no code implementations • EACL 2017 • Ting Han, David Schlangen
Grounded semantics is typically learnt from utterance-level meaning representations (e. g., successful database retrievals, denoted objects in images, moves in a game).
no code implementations • EACL 2017 • Julian Hough, David Schlangen
We present the joint task of incremental disfluency detection and utterance segmentation and a simple deep learning system which performs it on transcripts and ASR results.
no code implementations • LREC 2016 • Patrick Holthaus, Christian Leichsenring, Jasmin Bernotat, Viktor Richter, Marian Pohling, Birte Carlmeyer, Norman K{\"o}ster, Sebastian Meyer zu Borgsen, Ren{\'e} Zorn, Birte Schiffhauer, Kai Frederic Engelmann, Florian Lier, Simon Schulz, Philipp Cimiano, Friederike Eyssel, Thomas Hermann, Franz Kummert, David Schlangen, Sven Wachsmuth, Petra Wagner, Britta Wrede, Sebastian Wrede
In order to explore intuitive verbal and non-verbal interfaces in smart environments we recorded user interactions with an intelligent apartment.
no code implementations • LREC 2016 • Julian Hough, Ye Tian, Laura de Ruiter, Simon Betz, Spyros Kousidis, David Schlangen, Jonathan Ginzburg
We present the DUEL corpus, consisting of 24 hours of natural, face-to-face, loosely task-directed dialogue in German, French and Mandarin Chinese.
no code implementations • LREC 2016 • Sina Zarrie{\ss}, Julian Hough, Casey Kennington, Ramesh Manuvinakurike, David DeVault, Raquel Fern{\'a}ndez, David Schlangen
PentoRef is a corpus of task-oriented dialogues collected in systematically manipulated settings.
1 code implementation • ACL 2016 • David Schlangen, Sina Zarriess, Casey Kennington
A common use of language is to refer to visually present objects.