Search Results for author: Namrata Shivagunde

Found 5 papers, 5 papers with code

ReLoRA: High-Rank Training Through Low-Rank Updates

3 code implementations11 Jul 2023 Vladislav Lialin, Namrata Shivagunde, Sherin Muckatira, Anna Rumshisky

Despite the dominance and effectiveness of scaling, resulting in large networks with hundreds of billions of parameters, the necessity to train overparameterized models remains poorly understood, while training costs grow exponentially.

Life after BERT: What do Other Muppets Understand about Language?

1 code implementation ACL 2022 Vladislav Lialin, Kevin Zhao, Namrata Shivagunde, Anna Rumshisky

Existing pre-trained transformer analysis works usually focus only on one or two model families at a time, overlooking the variability of the architecture and pre-training objectives.

Down and Across: Introducing Crossword-Solving as a New NLP Benchmark

1 code implementation ACL 2022 Saurabh Kulshreshtha, Olga Kovaleva, Namrata Shivagunde, Anna Rumshisky

Solving crossword puzzles requires diverse reasoning capabilities, access to a vast amount of knowledge about language and the world, and the ability to satisfy the constraints imposed by the structure of the puzzle.

Natural Language Understanding Open-Domain Question Answering +1

Larger Probes Tell a Different Story: Extending Psycholinguistic Datasets Via In-Context Learning

1 code implementation29 Mar 2023 Namrata Shivagunde, Vladislav Lialin, Anna Rumshisky

Finally, we observe that while GPT3 has generated all the examples in ROLE-1500 is only able to solve 24. 6% of them during probing.

In-Context Learning Language Modelling +2

Deconstructing In-Context Learning: Understanding Prompts via Corruption

1 code implementation2 Apr 2024 Namrata Shivagunde, Vladislav Lialin, Sherin Muckatira, Anna Rumshisky

In contrast, the underlying pre-trained LLMs they use as a backbone are known to be brittle in this respect.

In-Context Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.