Search Results for author: Arkil Patel

Found 8 papers, 6 papers with code

Evaluating In-Context Learning of Libraries for Code Generation

no code implementations16 Nov 2023 Arkil Patel, Siva Reddy, Dzmitry Bahdanau, Pradeep Dasigi

Contemporary Large Language Models (LLMs) exhibit a high degree of code generation and comprehension capability.

Code Generation In-Context Learning

MAGNIFICo: Evaluating the In-Context Learning Ability of Large Language Models to Generalize to Novel Interpretations

1 code implementation18 Oct 2023 Arkil Patel, Satwik Bhattamishra, Siva Reddy, Dzmitry Bahdanau

Additionally, our analysis uncovers the semantic predispositions in LLMs and reveals the impact of recency bias for information presented in long contexts.

In-Context Learning Semantic Parsing +1

Understanding In-Context Learning in Transformers and LLMs by Learning to Learn Discrete Functions

no code implementations4 Oct 2023 Satwik Bhattamishra, Arkil Patel, Phil Blunsom, Varun Kanade

In this work, we take a step towards answering these questions by demonstrating the following: (a) On a test-bed with a variety of Boolean function classes, we find that Transformers can nearly match the optimal learning algorithm for 'simpler' tasks, while their performance deteriorates on more 'complex' tasks.

In-Context Learning

Simplicity Bias in Transformers and their Ability to Learn Sparse Boolean Functions

1 code implementation22 Nov 2022 Satwik Bhattamishra, Arkil Patel, Varun Kanade, Phil Blunsom

(ii) When trained on Boolean functions, both Transformers and LSTMs prioritize learning functions of low sensitivity, with Transformers ultimately converging to functions of lower sensitivity.

When Can Transformers Ground and Compose: Insights from Compositional Generalization Benchmarks

1 code implementation23 Oct 2022 Ankur Sikarwar, Arkil Patel, Navin Goyal

On analyzing the task, we find that identifying the target location in the grid world is the main challenge for the models.

Revisiting the Compositional Generalization Abilities of Neural Sequence Models

1 code implementation ACL 2022 Arkil Patel, Satwik Bhattamishra, Phil Blunsom, Navin Goyal

Compositional generalization is a fundamental trait in humans, allowing us to effortlessly combine known phrases to form novel sentences.

Are NLP Models really able to Solve Simple Math Word Problems?

3 code implementations NAACL 2021 Arkil Patel, Satwik Bhattamishra, Navin Goyal

Since existing solvers achieve high performance on the benchmark datasets for elementary level MWPs containing one-unknown arithmetic word problems, such problems are often considered "solved" with the bulk of research attention moving to more complex MWPs.

Math Math Word Problem Solving +1

Cannot find the paper you are looking for? You can Submit a new open access paper.