Search Results for author: Atharva Naik

Found 14 papers, 6 papers with code

PBEBench: A Multi-Step Programming by Examples Reasoning Benchmark inspired by Historical Linguistics

no code implementations29 May 2025 Atharva Naik, Darsh Agrawal, Manav Kapadnis, Yuwei An, Yash Mathur, Carolyn Rose, David Mortensen

Recently, long chain of thought (LCoT), Large Language Models (LLMs), have taken the machine learning world by storm with their breathtaking reasoning capabilities.

Math

An Empirical Study on Strong-Weak Model Collaboration for Repo-level Code Generation

1 code implementation26 May 2025 Shubham Gandhi, Atharva Naik, Yiqing Xie, Carolyn Rose

We study cost-efficient collaboration between strong and weak language models for repository-level code generation, where the weak model handles simpler tasks at lower cost, and the most challenging tasks are delegated to the strong model.

GitHub issue resolution

Programming by Examples Meets Historical Linguistics: A Large Language Model Based Approach to Sound Law Induction

no code implementations27 Jan 2025 Atharva Naik, Darsh Agrawal, Hong Sng, Clayton Marr, Kexun Zhang, Nathaniel R Robinson, Kalvin Chang, Rebecca Byrnes, Aravind Mysore, Carolyn Rose, David R Mortensen

Historical linguists have long written "programs" that convert reconstructed words in an ancestor language into their attested descendants via ordered string rewrite functions (called sound laws) However, writing these programs is time-consuming, motivating the development of automated Sound Law Induction (SLI) which we formulate as Programming by Examples (PBE) with Large Language Models (LLMs) in this paper.

Code Generation Inductive Bias +4

CRScore: Grounding Automated Evaluation of Code Review Comments in Code Claims and Smells

no code implementations29 Sep 2024 Atharva Naik, Marcus Alenius, Daniel Fried, Carolyn Rose

The task of automated code review has recently gained a lot of attention from the machine learning community.

valid

Can Large Language Models Code Like a Linguist?: A Case Study in Low Resource Sound Law Induction

no code implementations18 Jun 2024 Atharva Naik, Kexun Zhang, Nathaniel Robinson, Aravind Mysore, Clayton Marr, Hong Sng Rebecca Byrnes, Anna Cai, Kalvin Chang, David Mortensen

Historical linguists have long written a kind of incompletely formalized ''program'' that converts reconstructed words in an ancestor language into words in one of its attested descendants that consist of a series of ordered string rewrite functions (called sound laws).

Generating Situated Reflection Triggers about Alternative Solution Paths: A Case Study of Generative AI for Computer-Supported Collaborative Learning

no code implementations28 Apr 2024 Atharva Naik, Jessica Ruhan Yin, Anusha Kamath, Qianou Ma, Sherry Tongshuang Wu, Charles Murray, Christopher Bogart, Majd Sakr, Carolyn P. Rose

An advantage of Large Language Models (LLMs) is their contextualization capability - providing different responses based on student inputs like solution strategy or prior discussion, to potentially better engage students than standard feedback.

Cloud Computing

On the Limitations of Embedding Based Methods for Measuring Functional Correctness for Code Generation

no code implementations26 Apr 2024 Atharva Naik

The task of code generation from natural language (NL2Code) has become extremely popular, especially with the advent of Large Language Models (LLMs).

Code Generation HumanEval

Data Augmentation for Code Translation with Comparable Corpora and Multiple References

1 code implementation1 Nov 2023 Yiqing Xie, Atharva Naik, Daniel Fried, Carolyn Rose

One major challenge of translating code between programming languages is that parallel training data is often limited.

Code Translation Data Augmentation +1

Tricking LLMs into Disobedience: Formalizing, Analyzing, and Detecting Jailbreaks

1 code implementation24 May 2023 Abhinav Rao, Sachin Vashistha, Atharva Naik, Somak Aditya, Monojit Choudhury

Recent explorations with commercial Large Language Models (LLMs) have shown that non-expert users can jailbreak LLMs by simply manipulating their prompts; resulting in degenerate output behavior, privacy and security breaches, offensive outputs, and violations of content regulator policies.

Weakly Supervised Explainable Phrasal Reasoning with Neural Fuzzy Logic

1 code implementation18 Sep 2021 Zijun Wu, Zi Xuan Zhang, Atharva Naik, Zhijian Mei, Mauajama Firdaus, Lili Mou

In this work, we address the explainability of NLI by weakly supervised logical reasoning, and propose an Explainable Phrasal Reasoning (EPR) approach.

Explanation Generation Logical Reasoning +2

Understanding the Role of Affect Dimensions in Detecting Emotions from Tweets: A Multi-task Approach

1 code implementation9 May 2021 Rajdeep Mukherjee, Atharva Naik, Sriyash Poddar, Soham Dasgupta, Niloy Ganguly

For the regression task, VADEC, when trained with SenWave, achieves 7. 6% and 16. 5% gains in Pearson Correlation scores over the current state-of-the-art on the EMOBANK dataset for the Valence (V) and Dominance (D) affect dimensions respectively.

Emotion Classification regression +1

How Have We Reacted To The COVID-19 Pandemic? Analyzing Changing Indian Emotions Through The Lens of Twitter

no code implementations20 Aug 2020 Rajdeep Mukherjee, Sriyash Poddar, Atharva Naik, Soham Dasgupta

Since its outbreak, the ongoing COVID-19 pandemic has caused unprecedented losses to human lives and economies around the world.

Emotion Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.