1 code implementation • EMNLP (BlackboxNLP) 2021 • Ionut-Teodor Sorodoc, Gemma Boleda, Marco Baroni
In recent years, the NLP community has shown increasing interest in analysing how deep learning models work.
no code implementations • 24 May 2024 • Emily Cheng, Marco Baroni, Carmen Amo Alonso
The increasing prevalence of Large Language Models (LMs) in critical applications highlights the need for controlled language generation strategies that are not only computationally efficient but that also enjoy performance guarantees.
no code implementations • 24 May 2024 • Emily Cheng, Diego Doimo, Corentin Kervadec, Iuri Macocco, Jade Yu, Alessandro Laio, Marco Baroni
A language model (LM) is a mapping from a linguistic context to an output token.
1 code implementation • 23 Feb 2024 • Nathanaël Carraz Rakotonirina, Marco Baroni
Transformer-based language models (LMs) track contextual information through large, hard-coded input windows.
no code implementations • 24 Oct 2023 • Corentin Kervadec, Francesca Franzon, Marco Baroni
Language model prompt optimization research has shown that semantically and grammatically well-formed manually crafted prompts are routinely outperformed by automatically generated token sequences with no apparent meaning or syntactic structure, including sequences of vectors from a model's embedding space.
1 code implementation • 20 Oct 2023 • Emily Cheng, Corentin Kervadec, Marco Baroni
For a language model (LM) to faithfully model human language, it must compress vast, potentially infinite information into relatively few dimensions.
1 code implementation • CVPR 2023 • Roberto Dessì, Michele Bevilacqua, Eleonora Gualdoni, Nathanael Carraz Rakotonirina, Francesca Franzon, Marco Baroni
However, when the model is used without further tuning to generate captions for out-of-domain datasets, our discriminatively-finetuned captioner generates descriptions that resemble human references more than those produced by the same captioner without finetuning.
1 code implementation • 20 Feb 2023 • Nathanaël Carraz Rakotonirina, Roberto Dessì, Fabio Petroni, Sebastian Riedel, Marco Baroni
We study whether automatically-induced prompts that effectively extract information from a language model can also be used, out-of-the-box, to probe other language models for the same information.
1 code implementation • 4 Feb 2023 • Matéo Mahaut, Francesca Franzon, Roberto Dessì, Marco Baroni
As a first step in this direction, we systematically explore the task of referential communication in a community of heterogeneous state-of-the-art pre-trained visual networks, showing that they can develop, in a self-supervised way, a shared protocol to refer to a target object among a set of candidates.
1 code implementation • 20 Oct 2022 • Roberto Dessì, Eleonora Gualdoni, Francesca Franzon, Gemma Boleda, Marco Baroni
We compare the 0-shot performance of a neural caption-based image retriever when given as input either human-produced captions or captions generated by a neural captioner.
no code implementations • 6 Oct 2021 • Eugene Kharitonov, Marco Baroni, Dieuwke Hupkes
In this work, we demonstrate that the size of the subword vocabulary learned by Byte-Pair Encoding (BPE) greatly affects both ability and tendency of standard Transformer models to memorize training data, even when we control for the number of learned parameters.
no code implementations • 16 Jun 2021 • Marco Baroni
A lively research field has recently emerged that uses experimental methods to probe the linguistic behavior of modern deep networks.
1 code implementation • NeurIPS 2021 • Roberto Dessì, Eugene Kharitonov, Marco Baroni
As deep networks begin to be deployed as autonomous agents, the issue of how they can communicate with each other becomes important.
no code implementations • 19 Jun 2020 • Yair Lakretz, Dieuwke Hupkes, Alessandra Vergallito, Marco Marelli, Marco Baroni, Stanislas Dehaene
We studied whether a modern artificial neural network trained with "deep learning" methods mimics a central aspect of human sentence processing, namely the storing of grammatical number and gender information in working memory and its use in long-distance agreement (e. g., capturing the correct number agreement between subject and verb when they are separated by other phrases).
1 code implementation • 3 Jun 2020 • Angeliki Lazaridou, Marco Baroni
The ability to cooperate through language is a defining feature of humans.
1 code implementation • ICLR 2020 • Jonathan Gordon, David Lopez-Paz, Marco Baroni, Diane Bouchacourt
Humans understand novel sentences by composing meanings and roles of core language components.
no code implementations • 22 Apr 2020 • Tal Linzen, Marco Baroni
Modern deep neural networks achieve impressive performance in engineering applications that require extensive linguistic skills, such as machine translation.
1 code implementation • ACL 2020 • Rahma Chaabouni, Eugene Kharitonov, Diane Bouchacourt, Emmanuel Dupoux, Marco Baroni
Third, while compositionality is not necessary for generalization, it provides an advantage in terms of language transmission: The more compositional a language is, the more easily it will be picked up by new learners, even when the latter differ in architecture from the original agents.
1 code implementation • EMNLP (BlackboxNLP) 2020 • Eugene Kharitonov, Marco Baroni
Studies of discrete languages emerging when neural agents communicate to solve a joint task often look for evidence of compositional structure.
no code implementations • 17 Mar 2020 • Marco Baroni
Deep-agent communities developing their own language-like communication protocol are a hot (or at least warm) topic in AI.
4 code implementations • NeurIPS 2020 • Laura Ruis, Jacob Andreas, Marco Baroni, Diane Bouchacourt, Brenden M. Lake
In this paper, we introduce a new benchmark, gSCAN, for evaluating compositional generalization in situated language understanding.
no code implementations • 5 Nov 2019 • Roberto Dessì, Diane Bouchacourt, Davide Crepaldi, Marco Baroni
Research in multi-agent cooperation has shown that artificial agents are able to learn to play a simple referential game while developing a shared lexicon.
no code implementations • ACL 2019 • Damian Blasi, Ryan Cotterell, Lawrence Wolf-Sonkin, Sabine Stoll, Balthasar Bickel, Marco Baroni
Embedding a clause inside another ({``}the girl [who likes cars [that run fast]] has arrived{''}) is a fundamental resource that has been argued to be a key driver of linguistic expressiveness.
no code implementations • IJCNLP 2019 • Eugene Kharitonov, Rahma Chaabouni, Diane Bouchacourt, Marco Baroni
There is renewed interest in simulating language emergence among deep neural agents that communicate to jointly solve a task, spurred by the practical aim to develop language-enabled interactive AIs, as well as by theoretical questions about the evolution of human language.
1 code implementation • TACL 2019 • Michael Hahn, Marco Baroni
Recurrent neural networks (RNNs) have reached striking performance in many natural language processing tasks.
1 code implementation • ICML 2020 • Eugene Kharitonov, Rahma Chaabouni, Diane Bouchacourt, Marco Baroni
There is growing interest in studying the languages that emerge when neural agents are jointly trained to solve tasks requiring communication through a discrete channel.
1 code implementation • ACL 2019 • Rahma Chaabouni, Eugene Kharitonov, Alessandro Lazaric, Emmanuel Dupoux, Marco Baroni
We train models to communicate about paths in a simple gridworld, using miniature languages that reflect or violate various natural language trends, such as the tendency to avoid redundancy or to minimize long-distance dependencies.
1 code implementation • NeurIPS 2019 • Rahma Chaabouni, Eugene Kharitonov, Emmanuel Dupoux, Marco Baroni
Despite renewed interest in emergent language simulations with neural networks, little is known about the basic properties of the induced code, and how they compare to human language.
1 code implementation • ACL 2019 • Diane Bouchacourt, Marco Baroni
Recent research studies communication emergence in communities of deep network agents assigned a joint task, hoping to gain insights on human language evolution.
no code implementations • ACL 2019 • Roberto Dessì, Marco Baroni
Lake and Baroni (2018) introduced the SCAN dataset probing the ability of seq2seq models to capture compositional generalizations, such as inferring the meaning of "jump around" 0-shot from the component words.
no code implementations • 30 Mar 2019 • Marco Baroni
In the last decade, deep artificial neural networks have achieved astounding performance in many natural language processing tasks.
1 code implementation • NAACL 2019 • Yair Lakretz, German Kruszewski, Theo Desbordes, Dieuwke Hupkes, Stanislas Dehaene, Marco Baroni
Importantly, the behaviour of these units is partially controlled by other units independently shown to track syntactic structure.
2 code implementations • 14 Jan 2019 • Brenden M. Lake, Tal Linzen, Marco Baroni
There have been striking recent improvements in machine learning for natural language processing, yet the best algorithms require vast amounts of experience and struggle to generalize new concepts in compositional ways.
1 code implementation • WS 2018 • Jasmijn Bastings, Marco Baroni, Jason Weston, Kyunghyun Cho, Douwe Kiela
Lake and Baroni (2018) recently introduced the SCAN data set, which consists of simple commands paired with action sequences and is intended to test the strong generalization abilities of recurrent sequence-to-sequence models.
no code implementations • EMNLP 2018 • Diane Bouchacourt, Marco Baroni
There is growing interest in the language developed by agents interacting in emergent-communication settings.
no code implementations • WS 2018 • João Loula, Marco Baroni, Brenden M. Lake
Systematic compositionality is the ability to recombine meaningful units with regular and predictable outcomes, and it's seen as key to humans' capacity for generalization in language.
no code implementations • ACL 2018 • Alexis Conneau, German Kruszewski, Guillaume Lample, Lo{\"\i}c Barrault, Marco Baroni
Although much effort has recently been devoted to training high-quality sentence embeddings, we still have a poor understanding of what they are capturing.
6 code implementations • 3 May 2018 • Alexis Conneau, German Kruszewski, Guillaume Lample, Loïc Barrault, Marco Baroni
Although much effort has recently been devoted to training high-quality sentence embeddings, we still have a poor understanding of what they are capturing.
2 code implementations • NAACL 2018 • Kristina Gulordava, Piotr Bojanowski, Edouard Grave, Tal Linzen, Marco Baroni
Recurrent neural networks (RNNs) have achieved impressive results in a variety of linguistic processing tasks, suggesting that they can induce non-trivial properties of language.
1 code implementation • 18 Feb 2018 • Adam Liška, Germán Kruszewski, Marco Baroni
Neural networks are very powerful learning systems, but they do not readily generalize from one task to the other.
no code implementations • ICLR 2018 • Brenden Lake, Marco Baroni
Humans can understand and produce new utterances effortlessly, thanks to their systematic compositional skills.
7 code implementations • ICML 2018 • Brenden M. Lake, Marco Baroni
Humans can understand and produce new utterances effortlessly, thanks to their compositional skills.
1 code implementation • EMNLP 2017 • Aurelie Herbelot, Marco Baroni
Distributional semantics models are known to struggle with small data.
no code implementations • 23 Feb 2017 • Mateo Rojas-Carulla, Marco Baroni, David Lopez-Paz
In this paper, we develop a framework to estimate the cause-effect relation between two static entities $x$ and $y$: for instance, an art masterpiece $x$ and its fraudulent copy $y$.
no code implementations • 6 Feb 2017 • Gemma Boleda, Sebastian Padó, Nghia The Pham, Marco Baroni
Reference is a crucial property of language that allows us to connect linguistic expressions to the world.
no code implementations • 31 Jan 2017 • Marco Baroni, Armand Joulin, Allan Jabri, Germàn Kruszewski, Angeliki Lazaridou, Klemen Simonic, Tomas Mikolov
With machine learning successfully applied to new daunting problems almost every day, general AI starts looking like an attainable goal.
1 code implementation • 21 Dec 2016 • Angeliki Lazaridou, Alexander Peysakhovich, Marco Baroni
The sender is told one of them is the target and is allowed to send a message from a fixed, arbitrary vocabulary to the receiver.
no code implementations • 28 Jun 2016 • Gemma Boleda, Sebastian Padó, Marco Baroni
One of the most basic functions of language is to refer to objects in a shared scene.
3 code implementations • ACL 2016 • Denis Paperno, Germán Kruszewski, Angeliki Lazaridou, Quan Ngoc Pham, Raffaella Bernardi, Sandro Pezzelle, Marco Baroni, Gemma Boleda, Raquel Fernández
We introduce LAMBADA, a dataset to evaluate the capabilities of computational models for text understanding by means of a word prediction task.
no code implementations • 23 May 2016 • Angeliki Lazaridou, Nghia The Pham, Marco Baroni
We propose an interactive multimodal framework for language learning.
no code implementations • 8 Mar 2016 • Angeliki Lazaridou, Nghia The Pham, Marco Baroni
As a first step towards agents learning to communicate about their visual environment, we propose a system that, given visual representations of a referent (cat) and a context (sofa), identifies their discriminative attributes, i. e., properties that distinguish them (has_tail).
1 code implementation • 25 Nov 2015 • Tomas Mikolov, Armand Joulin, Marco Baroni
The development of intelligent machines is one of the biggest unsolved challenges in computer science.
no code implementations • 10 Jun 2015 • Angeliki Lazaridou, Dat Tien Nguyen, Raffaella Bernardi, Marco Baroni
We introduce language-driven image generation, the task of generating an image visualizing the semantic contents of a word embedding, e. g., given the word embedding of grasshopper, we generate a natural image of a grasshopper.
no code implementations • TACL 2015 • Angeliki Lazaridou, Georgiana Dinu, Adam Liska, Marco Baroni
By building on the recent "zero-shot learning" approach, and paying attention to the linguistic nature of attributes as noun modifiers, and specifically adjectives, we show that it is possible to tag images with attribute-denoting adjectives even when no training data containing the relevant annotation are available.
no code implementations • HLT 2015 • Angeliki Lazaridou, Nghia The Pham, Marco Baroni
We extend the SKIP-GRAM model of Mikolov et al. (2013a) by taking visual information into account.
no code implementations • TACL 2015 • German Kruszewski, Denis Paperno, Marco Baroni
Corpus-based distributional semantic models capture degrees of semantic relatedness among the words of very large vocabularies, but have problems with logical phenomena such as entailment, that are instead elegantly handled by model-theoretic approaches, which, in turn, do not scale up.
5 code implementations • 20 Dec 2014 • Georgiana Dinu, Angeliki Lazaridou, Marco Baroni
The zero-shot paradigm exploits vector-based word representations extracted from text corpora with unsupervised methods to learn general mapping functions from other feature spaces onto word space, where the words associated to the nearest neighbours of the mapped vectors are used as their linguistic labels.
no code implementations • LREC 2014 • Marco Marelli, Stefano Menini, Marco Baroni, Luisa Bentivogli, Raffaella Bernardi, Roberto Zamparelli
Shared and internationally recognized benchmarks are fundamental for the development of any computational system.