no code implementations • 22 Jun 2024 • Yash Kumar Lal, Vanya Cohen, Nathanael Chambers, Niranjan Balasubramanian, Raymond Mooney
A fundamental aspect of plans is the temporal order in which their steps needs to be executed, which reflects the underlying causal dependencies between them.
2 code implementations • 10 Jun 2024 • Jordan Voas, Raymond Mooney, David Harwath
We introduce Semantic Parsing in Contextual Environments (SPICE), a task designed to enhance artificial agents' contextual awareness by integrating multimodal inputs with prior contexts.
no code implementations • 21 May 2024 • Vanya Cohen, Jason Xinyu Liu, Raymond Mooney, Stefanie Tellex, David Watkins
With large language models, robots can understand language more flexibly and more capable than ever before.
no code implementations • 16 May 2024 • Albert Yu, Adeline Foote, Raymond Mooney, Roberto Martín-Martín
We demonstrate that training the image encoder to predict the language description or the distance between descriptions of a sim or real image serves as a useful, data-efficient pretraining step that helps learn a domain-invariant image representation.
no code implementations • 11 Apr 2024 • Jierui Li, Raymond Mooney
More specifically, we employ an LLM to generate explanations for a set of <problem, solution-program> pairs, then use <problem, explanation> pairs to fine-tune a smaller language model, which we refer to as the Reasoner, to learn algorithmic reasoning that can generate "how-to-solve" hints for unseen problems.
no code implementations • 1 Apr 2024 • Casey Kennington, Malihe Alikhani, Heather Pon-Barry, Katherine Atwell, Yonatan Bisk, Daniel Fried, Felix Gervits, Zhao Han, Mert Inan, Michael Johnston, Raj Korpan, Diane Litman, Matthew Marge, Cynthia Matuszek, Ross Mead, Shiwali Mohan, Raymond Mooney, Natalie Parde, Jivko Sinapov, Angela Stewart, Matthew Stone, Stefanie Tellex, Tom Williams
The ability to interact with machines using natural human language is becoming not just commonplace, but expected.
1 code implementation • 16 Feb 2024 • Ziru Chen, Michael White, Raymond Mooney, Ali Payani, Yu Su, Huan Sun
In this paper, we examine how large language models (LLMs) solve multi-step problems under a language agent framework with three components: a generator, a discriminator, and a planning method.
no code implementations • 8 Jan 2024 • Priyanka Mandikal, Raymond Mooney
Traditional information retrieval is based on sparse bag-of-words vector representations of documents and queries.
no code implementations • 19 Sep 2023 • Jordan Voas, Yili Wang, QiXing Huang, Raymond Mooney
Our findings indicate that none of the metrics currently used for this task show even a moderate correlation with human judgments on a sample level.
no code implementations • 11 Jul 2023 • Jierui Li, Szymon Tworkowski, Yingying Wu, Raymond Mooney
In this paper, we approach competitive-level programming problem-solving as a composite task of reasoning and code generation.
1 code implementation • 22 May 2023 • Ziru Chen, Shijie Chen, Michael White, Raymond Mooney, Ali Payani, Jayanth Srinivasa, Yu Su, Huan Sun
Thus, we propose a novel representation for SQL queries and their edits that adheres more closely to the pre-training corpora of language models of code.
1 code implementation • Findings (ACL) 2021 • Yash Kumar Lal, Nathanael Chambers, Raymond Mooney, Niranjan Balasubramanian
They are especially worse on questions whose answers are external to the narrative, thus providing a challenge for future QA and narrative understanding research.
no code implementations • NAACL 2018 • Nazneen Fatema Rajani, Raymond Mooney
We propose four categories of auxiliary features for ensembling for VQA.
no code implementations • IJCNLP 2017 • Rodolfo Corona, Jesse Thomason, Raymond Mooney
Speech is a natural channel for human-computer interaction in robotics and consumer applications.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +5
no code implementations • WS 2017 • Jesse Thomason, Jivko Sinapov, Raymond Mooney
Multi-modal grounded language learning connects language predicates to physical properties of objects in the world.
1 code implementation • CVPR 2017 • Subhashini Venugopalan, Lisa Anne Hendricks, Marcus Rohrbach, Raymond Mooney, Trevor Darrell, Kate Saenko
We propose minimizing a joint objective which can learn from these diverse data sources and leverage distributional semantic embeddings, enabling the model to generalize and describe novel objects outside of image-caption datasets.
3 code implementations • EMNLP 2016 • Subhashini Venugopalan, Lisa Anne Hendricks, Raymond Mooney, Kate Saenko
This paper investigates how linguistic knowledge mined from large text corpora can aid the generation of natural language descriptions of videos.
no code implementations • ICCV 2015 • Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko
Our LSTM model is trained on video-sentence pairs and learns to associate a sequence of video frames to a sequence of words in order to generate a description of the event in the video clip.
1 code implementation • CVPR 2016 • Lisa Anne Hendricks, Subhashini Venugopalan, Marcus Rohrbach, Raymond Mooney, Kate Saenko, Trevor Darrell
Current deep caption models can only describe objects contained in paired image-sentence corpora, despite the fact that they are pre-trained with large object recognition datasets, namely ImageNet.
4 code implementations • 3 May 2015 • Subhashini Venugopalan, Marcus Rohrbach, Jeff Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko
Our LSTM model is trained on video-sentence pairs and learns to associate a sequence of video frames to a sequence of words in order to generate a description of the event in the video clip.
1 code implementation • HLT 2015 • Subhashini Venugopalan, Huijuan Xu, Jeff Donahue, Marcus Rohrbach, Raymond Mooney, Kate Saenko
Solving the visual symbol grounding problem has long been a goal of artificial intelligence.