1 code implementation • 3 Oct 2024 • Hosein Mohebbi, Grzegorz Chrupała, Willem Zuidema, Afra Alishahi, Ivan Titov
Neural speech models build deeply entangled internal representations, which capture a variety of features (e. g., fundamental frequency, loudness, syntactic category, or semantic content of a word) in a distributed encoding.
no code implementations • 25 Mar 2024 • Gaofei Shen, Michaela Watkins, Afra Alishahi, Arianna Bisazza, Grzegorz Chrupała
Interpretability research has shown that self-supervised Spoken Language Models (SLMs) encode a wide variety of features in human speech from the acoustic, phonetic, phonological, syntactic and semantic levels, to speaker characteristics.
1 code implementation • 15 Oct 2023 • Hosein Mohebbi, Grzegorz Chrupała, Willem Zuidema, Afra Alishahi
Transformers have become a key architecture in speech processing, but our understanding of how they build up representations of acoustic and linguistic structure is limited.
2 code implementations • 2 Oct 2023 • Gabriele Sarti, Grzegorz Chrupała, Malvina Nissim, Arianna Bisazza
Establishing whether language models can use contextual information in a human-plausible way is important to ensure their trustworthiness in real-world settings.
1 code implementation • 30 May 2023 • Gaofei Shen, Afra Alishahi, Arianna Bisazza, Grzegorz Chrupała
Understanding which information is encoded in deep models of spoken and written language has been the focus of much research in recent years, as it is crucial for debugging and improving these architectures.
no code implementations • 8 May 2023 • Grzegorz Chrupała
Human language is firstly spoken and only secondarily written.
1 code implementation • 30 Jan 2023 • Hosein Mohebbi, Willem Zuidema, Grzegorz Chrupała, Afra Alishahi
Self-attention weights and their transformed variants have been the main source of information for analyzing token-to-token interactions in Transformer-based models.
1 code implementation • 25 Feb 2022 • Mitja Nikolaus, Afra Alishahi, Grzegorz Chrupała
In the real world the coupling between the linguistic and the visual modality is loose, and often confounded by correlations with non-semantic aspects of the speech signal.
1 code implementation • LREC 2022 • Chris Emmery, Ákos Kádár, Grzegorz Chrupała, Walter Daelemans
The perturbed data, models, and code are available for reproduction at https://github. com/cmry/augtox
1 code implementation • 14 Jul 2021 • Afra Alishahia, Grzegorz Chrupała, Alejandrina Cristia, Emmanuel Dupoux, Bertrand Higy, Marvin Lavechin, Okko Räsänen, Chen Yu
We present the visually-grounded language modelling track that was introduced in the Zero-Resource Speech challenge, 2021 edition, 2nd round.
1 code implementation • EMNLP (BlackboxNLP) 2021 • Bertrand Higy, Lieke Gelderloos, Afra Alishahi, Grzegorz Chrupała
The distributed and continuous representations used by neural networks are at odds with representations employed in linguistics, which are typically symbolic.
no code implementations • 27 Apr 2021 • Grzegorz Chrupała
This survey provides an overview of the evolution of visually grounded models of spoken language over the last 20 years.
1 code implementation • EACL 2021 • Chris Emmery, Ákos Kádár, Grzegorz Chrupała
Written language contains stylistic cues that can be exploited to automatically infer a variety of potentially sensitive author information.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Bertrand Higy, Desmond Elliott, Grzegorz Chrupała
Visually-grounded models of spoken language understanding extract semantic information directly from speech, without relying on transcriptions.
no code implementations • ACL 2020 • Lieke Gelderloos, Grzegorz Chrupała, Afra Alishahi
Speech directed to children differs from adult-directed speech in linguistic aspects such as repetition, word choice, and sentence length, as well as in aspects of the speech signal itself, such as prosodic and phonemic variation.
1 code implementation • ACL 2020 • Grzegorz Chrupała, Bertrand Higy, Afra Alishahi
Given the fast development of analysis techniques for NLP and speech processing systems, few systematic studies have been conducted to compare the strengths and weaknesses of each method.
no code implementations • 9 Nov 2019 • Ákos Kádár, Grzegorz Chrupała, Afra Alishahi, Desmond Elliott
However, we do find that using an external machine translation model to generate the synthetic data sets results in better performance.
1 code implementation • ACL 2019 • Grzegorz Chrupała, Afra Alishahi
Analysis methods which enable us to better understand the representations and functioning of neural models of language are increasingly needed as deep learning becomes the dominant approach in NLP.
no code implementations • 5 Apr 2019 • Afra Alishahi, Grzegorz Chrupała, Tal Linzen
The EMNLP 2018 workshop BlackboxNLP was dedicated to resources and techniques specifically developed for analyzing and understanding the inner-workings and representations acquired by neural models of language.
1 code implementation • 21 Dec 2018 • Grzegorz Chrupała
A widespread approach to processing spoken language is to first automatically transcribe it into text.
1 code implementation • CONLL 2018 • Ákos Kádár, Desmond Elliott, Marc-Alexandre Côté, Grzegorz Chrupała, Afra Alishahi
Recent work has shown how to learn better visual-semantic embeddings by leveraging image descriptions in more than one language.
no code implementations • COLING 2018 • Ákos Kádár, Marc-Alexandre Côté, Grzegorz Chrupała, Afra Alishahi
Hierarchical Multiscale LSTM (Chung et al., 2016a) is a state-of-the-art language model that learns interpretable structure from character-level input.
1 code implementation • COLING 2018 • Chris Emmery, Enrique Manjavacas, Grzegorz Chrupała
The task of obfuscating writing style using sequence models has previously been investigated under the framework of obfuscation-by-transfer, where the input text is explicitly rewritten in another style.
no code implementations • WS 2019 • Grzegorz Chrupała, Lieke Gelderloos, Ákos Kádár, Afra Alishahi
In the domain of unsupervised learning most work on speech has focused on discovering low-level constructs such as phoneme inventories or word-like units.
1 code implementation • CONLL 2017 • Afra Alishahi, Marie Barking, Grzegorz Chrupała
We study the representation and encoding of phonemes in a recurrent neural network model of grounded speech.
2 code implementations • ACL 2017 • Grzegorz Chrupała, Lieke Gelderloos, Afra Alishahi
We present a visually grounded model of speech perception which projects spoken utterances and images to a joint semantic space.
no code implementations • COLING 2016 • Lieke Gelderloos, Grzegorz Chrupała
We present a model of visually-grounded language learning based on stacked gated recurrent neural networks which learns to predict visual features given an image description in the form of a sequence of phonemes.
1 code implementation • CL 2017 • Ákos Kádár, Grzegorz Chrupała, Afra Alishahi
We present novel methods for analyzing the activation patterns of RNNs from a linguistic point of view and explore the types of linguistic structure they learn.
1 code implementation • IJCNLP 2015 • Grzegorz Chrupała, Ákos Kádár, Afra Alishahi
We propose Imaginet, a model of learning visually grounded representations of language from coupled textual and visual input.
no code implementations • 18 Sep 2013 • Grzegorz Chrupała
To demonstrate the usefulness of the learned text embeddings, we use them as features in a supervised character level text segmentation and labeling task: recognizing spans of text containing programming language code.