1 code implementation • 23 Oct 2024 • Peter Shaw, James Cohan, Jacob Eisenstein, Kenton Lee, Jonathan Berant, Kristina Toutanova
We propose a new programming language called ALTA and a compiler that can map ALTA programs to Transformer weights.
no code implementations • 12 Mar 2024 • Shikhar Murty, Christopher Manning, Peter Shaw, Mandar Joshi, Kenton Lee
Unfortunately, LM agents often fail to generalize to new environments without human demonstrations.
1 code implementation • 14 Dec 2023 • Jacob Eisenstein, Chirag Nagpal, Alekh Agarwal, Ahmad Beirami, Alex D'Amour, DJ Dvijotham, Adam Fisch, Katherine Heller, Stephen Pfohl, Deepak Ramachandran, Peter Shaw, Jonathan Berant
However, even pretrain reward ensembles do not eliminate reward hacking: we show several qualitative reward hacking phenomena that are not mitigated by ensembling because all reward models in the ensemble exhibit similar error patterns.
1 code implementation • NeurIPS 2023 • Peter Shaw, Mandar Joshi, James Cohan, Jonathan Berant, Panupong Pasupat, Hexiang Hu, Urvashi Khandelwal, Kenton Lee, Kristina Toutanova
Much of the previous work towards digital agents for graphical user interfaces (GUIs) has relied on text-based representations (derived from HTML or other structured data sources), which are not always readily available.
1 code implementation • 19 May 2023 • Chaitanya Malaviya, Peter Shaw, Ming-Wei Chang, Kenton Lee, Kristina Toutanova
To study the ability of retrieval systems to meet such information needs, we construct QUEST, a dataset of 3357 natural language queries with implicit set operations, that map to a set of entities corresponding to Wikipedia documents.
4 code implementations • 7 Oct 2022 • Kenton Lee, Mandar Joshi, Iulia Turc, Hexiang Hu, Fangyu Liu, Julian Eisenschlos, Urvashi Khandelwal, Peter Shaw, Ming-Wei Chang, Kristina Toutanova
Visually-situated language is ubiquitous -- sources range from textbooks with diagrams to web pages with images and tables, to mobile apps with buttons and forms.
Ranked #18 on Visual Question Answering (VQA) on InfographicVQA
no code implementations • COLING 2022 • Yury Zemlyanskiy, Michiel de Jong, Joshua Ainslie, Panupong Pasupat, Peter Shaw, Linlu Qiu, Sumit Sanghai, Fei Sha
Then, it retrieves exemplars with outputs similar to the preliminary prediction which are used to generate a final prediction.
no code implementations • 24 May 2022 • Linlu Qiu, Peter Shaw, Panupong Pasupat, Tianze Shi, Jonathan Herzig, Emily Pitler, Fei Sha, Kristina Toutanova
Meanwhile, recent work has shown considerable improvements on many NLP tasks from model scaling.
2 code implementations • NAACL 2022 • Linlu Qiu, Peter Shaw, Panupong Pasupat, Paweł Krzysztof Nowak, Tal Linzen, Fei Sha, Kristina Toutanova
Generic unstructured neural networks have been shown to struggle on out-of-distribution compositional generalization.
1 code implementation • 9 Nov 2021 • Wang Zhu, Peter Shaw, Tal Linzen, Fei Sha
Neural network models often generalize poorly to mismatched domains or distributions.
no code implementations • 6 Oct 2021 • Joseph R. Barr, Peter Shaw, Marcus Sobel
It's regarded as an axiom that a good model is one that compromises between bias and variance.
no code implementations • Findings (EMNLP) 2021 • BoWen Zhang, Hexiang Hu, Linlu Qiu, Peter Shaw, Fei Sha
We investigate ways to compose complex concepts in texts from primitive ones while grounding them in images.
2 code implementations • EMNLP 2021 • Linlu Qiu, Hexiang Hu, BoWen Zhang, Peter Shaw, Fei Sha
We analyze the grounded SCAN (gSCAN) benchmark, which was recently proposed to study systematic generalization for grounded language understanding.
Ranked #4 on Compositional Generalization (AVG) on ReaSCAN
Compositional Generalization (AVG) Systematic Generalization
no code implementations • Findings (EMNLP) 2021 • Jeremy R. Cole, Nanjiang Jiang, Panupong Pasupat, Luheng He, Peter Shaw
The dominant paradigm for semantic parsing in recent years is to formulate parsing as a sequence-to-sequence task, generating predictions with auto-regressive sequence decoders.
2 code implementations • 15 Apr 2021 • Jonathan Herzig, Peter Shaw, Ming-Wei Chang, Kelvin Guu, Panupong Pasupat, Yuan Zhang
Sequence-to-sequence (seq2seq) models are prevalent in semantic parsing, but have been found to struggle at out-of-distribution compositional generalization.
Ranked #3 on Semantic Parsing on CFQ
1 code implementation • ACL 2021 • Peter Shaw, Ming-Wei Chang, Panupong Pasupat, Kristina Toutanova
This has motivated new specialized architectures with stronger compositional biases, but most of these approaches have only been evaluated on synthetically-generated datasets, which are not representative of natural language variation.
no code implementations • ACL 2020 • Alane Suhr, Ming-Wei Chang, Peter Shaw, Kenton Lee
We study the task of cross-database semantic parsing (XSP), where a system that maps natural language utterances to executable SQL queries is evaluated on databases unseen during training.
no code implementations • 4 Oct 2019 • Zhen Zeng, Yuefeng Lu, Judong Shen, Wei Zheng, Peter Shaw, Mary Beth Dorr
Precision medicine is becoming a focus in medical research recently, as its implementation brings values to all stakeholders in the healthcare system.
no code implementations • IJCNLP 2019 • Thomas Müller, Francesco Piccinno, Massimo Nicosia, Peter Shaw, Yasemin Altun
We present a novel approach to answering sequential questions based on structured objects such as knowledge bases or tables without using a logical form as an intermediate representation.
no code implementations • ACL 2019 • Peter Shaw, Philip Massey, Angelica Chen, Francesco Piccinno, Yasemin Altun
Structured information about entities is critical for many semantic parsing tasks.
12 code implementations • NAACL 2018 • Peter Shaw, Jakob Uszkoreit, Ashish Vaswani
On the WMT 2014 English-to-German and English-to-French translation tasks, this approach yields improvements of 1. 3 BLEU and 0. 3 BLEU over absolute position representations, respectively.
Ranked #22 on Machine Translation on WMT2014 English-French