no code implementations • NAACL (SIGTYP) 2022 • Qingxia Guo, Nathaniel Imel, Shane Steinert-Threlkeld
This paper introduces a database for crosslinguistic modal semantics.
no code implementations • 3 Dec 2024 • Leroy Z. Wang, R. Thomas McCoy, Shane Steinert-Threlkeld
What factors contribute to the relative success and corresponding difficulties of in-context learning for Large Language Models (LLMs)?
1 code implementation • 24 May 2024 • Abhinav Patil, Jaap Jumelet, Yu Ying Chiu, Andy Lapastora, Peter Shen, Lexie Wang, Clevis Willrich, Shane Steinert-Threlkeld
This paper introduces Filtered Corpus Training, a method that trains language models (LMs) on corpora with certain linguistic constructions filtered out from the training data, and uses it to measure the ability of LMs to perform linguistic generalization on the basis of indirect evidence.
no code implementations • 20 May 2024 • C. M. Downey, Terra Blevins, Dhwani Serai, Dwija Parikh, Shane Steinert-Threlkeld
To test this approach more rigorously, we systematically study best practices for adapting a pre-trained model to a language family.
no code implementations • 26 Mar 2024 • Nicolas Guerin, Shane Steinert-Threlkeld, Emmanuel Chemla
Unsupervised on-the-fly back-translation, in conjunction with multilingual pretraining, is the dominant method for unsupervised neural machine translation.
1 code implementation • 9 Sep 2023 • C. M. Downey, Terra Blevins, Nora Goldfine, Shane Steinert-Threlkeld
Pre-trained multilingual language models underpin a large portion of modern NLP tools outside of English.
no code implementations • 2 Sep 2023 • Shunjie Wang, Shane Steinert-Threlkeld
Despite the fact that Transformers perform well in NLP tasks, recent studies suggest that self-attention is theoretically limited in learning even some regular and context-free languages.
1 code implementation • 16 May 2023 • Yifan Jiang, Shane Steinert-Threlkeld
Feature attribution aims to explain the reasoning behind a black-box model's prediction by identifying the impact of each feature on the prediction.
1 code implementation • 11 Sep 2022 • Pangbo Ban, Yifan Jiang, Tianran Liu, Shane Steinert-Threlkeld
To what extent do pre-trained language models grasp semantic knowledge regarding the phenomenon of distributivity?
no code implementations • 11 Sep 2022 • David K. Yi, James V. Bruno, Jiayu Han, Peter Zukerman, Shane Steinert-Threlkeld
We investigate the extent to which verb alternation classes, as described by Levin (1993), are encoded in the embeddings of Large Pre-trained Language Models (PLMs) such as BERT, RoBERTa, ELECTRA, and DeBERTa using selectively constructed diagnostic classifiers for word and sentence-level prediction tasks.
1 code implementation • 14 Jul 2022 • C. M. Downey, Xuhui Zhou, Leo Z. Liu, Shane Steinert-Threlkeld
We formulate and test a technique to use Emergent Communication (EC) with a pre-trained multilingual model to improve on modern Unsupervised NMT systems, especially for low-resource languages.
1 code implementation • Findings (ACL) 2021 • Jaap Jumelet, Milica Denić, Jakub Szymanik, Dieuwke Hupkes, Shane Steinert-Threlkeld
We investigate the semantic knowledge of language models (LMs), focusing on (1) whether these LMs create categories of linguistic environments based on their semantic monotonicity properties, and (2) whether these categories play a similar role in LMs as in human language understanding, using negative polarity item licensing as a case study.
no code implementations • Findings (EMNLP) 2021 • Naomi Tachikawa Shapiro, Amandalynne Paullada, Shane Steinert-Threlkeld
We introduce a multilabel probing task to assess the morphosyntactic representations of word embeddings from multilingual language models.
1 code implementation • NAACL (SIGMORPHON) 2022 • C. M. Downey, Fei Xia, Gina-Anne Levow, Shane Steinert-Threlkeld
Segmentation remains an important preprocessing step both in languages where "words" or other important syntactic/semantic units (like morphemes) are not clearly delineated by white space, as well as when dealing with continuous speech data, where there is often no meaningful pause between words.
1 code implementation • ACL 2022 • Chih-chan Tien, Shane Steinert-Threlkeld
To study this theory, we design unsupervised models trained on unpaired sentences and single-pair supervised models trained on bitexts, both based on the unsupervised language model XLM-R with its parameters frozen.
no code implementations • EMNLP (BlackboxNLP) 2020 • Chuanrong Li, Lin Shengshuo, Leo Z. Liu, Xinyi Wu, Xuhui Zhou, Shane Steinert-Threlkeld
Although large-scale pretrained language models, such as BERT and RoBERTa, have achieved superhuman performance on in-distribution test sets, their performance suffers on out-of-distribution test sets (e. g., on contrast sets).
1 code implementation • ACL 2020 • Nur Geffen Lan, Emmanuel Chemla, Shane Steinert-Threlkeld
We propose a general framework to study language emergence through signaling games with neural agents.
1 code implementation • 24 Sep 2019 • Shane Steinert-Threlkeld
All natural languages exhibit a distinction between content words (like nouns and adjectives) and function words (like determiners, auxiliaries, prepositions).
1 code implementation • WS 2019 • Lewis O'Sullivan, Shane Steinert-Threlkeld
How are the meanings of linguistic expressions related to their use in concrete cognitive tasks?
1 code implementation • ACL 2018 • Sandro Pezzelle, Shane Steinert-Threlkeld, Raffaela Bernardi, Jakub Szymanik
We study the role of linguistic context in predicting quantifiers (`few', `all').