no code implementations • 5 Mar 2025 • Tushar Aggarwal, Swayam Singh, Abhijeet Awasthi, Aditya Kanade, Nagarajan Natarajan
Software engineering activities frequently involve edits to existing code.
no code implementations • 28 Oct 2024 • Prakhar Verma, Sukruta Prakash Midigeshi, Gaurav Sinha, Arno Solin, Nagarajan Natarajan, Amit Sharma
We introduce Planning-guided Retrieval Augmented Generation (Plan$\times$RAG), a novel framework that augments the \emph{retrieve-then-reason} paradigm of existing RAG frameworks to \emph{plan-then-retrieve}.
no code implementations • 30 Sep 2024 • Sonu Mehta, Jayashree Mohan, Nagarajan Natarajan, Ramachandran Ramjee, Manik Varma
State-of-the-art XC techniques that demonstrate high accuracies (e. g., DEXML, Ren\'ee, DEXA) on standard datasets have per-epoch training time that scales as $O(L)$ or employ expensive negative sampling strategies, which are prohibitive in XC scenarios.
no code implementations • 15 Sep 2024 • Bhawna Paliwal, Deepak Saini, Mudit Dhawan, Siddarth Asokan, Nagarajan Natarajan, Surbhi Aggarwal, Pankaj Malhotra, Jian Jiao, Manik Varma
In response, we propose Cross-encoders with Joint Efficient Modeling (CROSS-JEM), a novel ranking approach that enables transformer-based models to jointly score multiple items for a query, maximizing parameter utilization.
no code implementations • 17 Jun 2024 • Daman Arora, Atharv Sonwane, Nalin Wadhwa, Abhav Mehrotra, Saiteja Utpala, Ramakrishna Bairi, Aditya Kanade, Nagarajan Natarajan
A common method to solve complex problems in software engineering, is to divide the problem into multiple sub-problems.
no code implementations • 15 Jun 2024 • Gurusha Juneja, Nagarajan Natarajan, Hua Li, Jian Jiao, Amit Sharma
Given a task in the form of a basic description and its training examples, prompt optimization is the problem of synthesizing the given information into a text prompt for a large language model (LLM).
no code implementations • 1 Mar 2024 • Sayak Ray Chowdhury, Anush Kini, Nagarajan Natarajan
Our experiments on IMDb sentiment generation and Anthropic's helpful-harmless dataset show that rDPO is robust to noise in preference labels compared to vanilla DPO and other heuristics proposed by practitioners.
no code implementations • 29 Jan 2024 • Manav Singhal, Tushar Aggarwal, Abhijeet Awasthi, Nagarajan Natarajan, Aditya Kanade
We propose a new benchmark NoFunEval to evaluate code LMs on non-functional requirements and simple classification instances for both functional and non-functional requirements.
no code implementations • 31 Oct 2023 • Daman Arora, Anush Kini, Sayak Ray Chowdhury, Nagarajan Natarajan, Gaurav Sinha, Amit Sharma
Given a query and a document corpus, the information retrieval (IR) task is to output a ranked list of relevant documents.
no code implementations • 30 Oct 2023 • Sayak Ray Chowdhury, Xingyu Zhou, Nagarajan Natarajan
Within a standard minimax estimation framework, we provide tight upper and lower bounds on the error in estimating $\theta^*$ under both local and central models of DP.
no code implementations • 22 Sep 2023 • Nalin Wadhwa, Jui Pradhan, Atharv Sonwane, Surya Prakash Sahu, Nagarajan Natarajan, Aditya Kanade, Suresh Parthasarathy, Sriram Rajamani
We present a tool, CORE (short for COde REvisions), architected using a pair of LLMs organized as a duo comprised of a proposer and a ranker.
no code implementations • 23 Feb 2022 • Divyam Anshumaan, Sriram Balasubramanian, Shubham Tiwari, Nagarajan Natarajan, Sundararajan Sellamanickam, Venkata N. Padmanabhan
Simulating physical network paths (e. g., Internet) is a cornerstone research problem in the emerging sub-field of AI-for-networking.
no code implementations • 15 Feb 2021 • Aadirupa Saha, Nagarajan Natarajan, Praneeth Netrapalli, Prateek Jain
We study online learning with bandit feedback (i. e. learner has access to only zeroth-order oracle) where cost/reward functions $\f_t$ admit a "pseudo-1d" structure, i. e. $\f_t(\w) = \loss_t(\pred_t(\w))$ where the output of $\pred_t$ is one-dimensional.
3 code implementations • 15 Feb 2021 • Ajaykrishna Karthikeyan, Naman jain, Nagarajan Natarajan, Prateek Jain
Decision trees provide a rich family of highly non-linear but efficient models, due to which they continue to be the go-to family of predictive models by practitioners across domains.
no code implementations • NeurIPS Workshop CAP 2020 • Sahil Bhatia, Saswat Padhi, Nagarajan Natarajan, Rahul Sharma, Prateek Jain
Automated synthesis of inductive invariants is an important problem in software verification.
no code implementations • 14 Jul 2020 • Nagarajan Natarajan, Ajaykrishna Karthikeyan, Prateek Jain, Ivan Radicek, Sriram Rajamani, Sumit Gulwani, Johannes Gehrke
The goal of the synthesizer is to synthesize a "decision function" $f$ which transforms the features to a decision value for the black-box component so as to maximize the expected reward $E[r \circ f (x)]$ for executing decisions $f(x)$ for various values of $x$.
no code implementations • 26 Nov 2019 • Sahil Bhatia, Saswat Padhi, Nagarajan Natarajan, Rahul Sharma, Prateek Jain
Automated synthesis of inductive invariants is an important problem in software verification.
no code implementations • 18 Sep 2017 • Rahul Wadbude, Vivek Gupta, Piyush Rai, Nagarajan Natarajan, Harish Karnick, Prateek Jain
Our approach is novel in that it highlights interesting connections between label embedding methods used for multi-label learning and paragraph/document embedding methods commonly used for learning representations of text data.
no code implementations • ICML 2017 • Krzysztof Dembczyński, Wojciech Kotłowski, Oluwasanmi Koyejo, Nagarajan Natarajan
Statistical learning theory is at an inflection point enabled by recent advances in understanding and optimizing a wide range of metrics.
no code implementations • ICML 2017 • Kamalika Chaudhuri, Prateek Jain, Nagarajan Natarajan
In this work, we consider a theoretical analysis of the label requirement of active learning for regression under a heteroscedastic noise model, where the noise depends on the instance.
no code implementations • NeurIPS 2016 • Prateek Jain, Nagarajan Natarajan
We consider the problem of recommending relevant labels (items) for a given data point (user).
no code implementations • 3 May 2016 • Aditya Krishna Menon, Brendan van Rooyen, Nagarajan Natarajan
Suppose we have a sample of instances paired with binary labels corrupted by arbitrary instance- and label-dependent noise.
no code implementations • NeurIPS 2015 • Oluwasanmi O. Koyejo, Nagarajan Natarajan, Pradeep K. Ravikumar, Inderjit S. Dhillon
In particular, we show that for multilabel metrics constructed as instance-, micro- and macro-averages, the population optimal classifier can be decomposed into binary classifiers based on the marginal instance-conditional distribution of each label, with a weak association between labels via the threshold.
no code implementations • NeurIPS 2015 • Prateek Jain, Nagarajan Natarajan, Ambuj Tewari
We offer a general framework to derive mistake driven online algorithms and associated loss bounds.
no code implementations • 7 May 2015 • Nagarajan Natarajan, Oluwasanmi Koyejo, Pradeep Ravikumar, Inderjit S. Dhillon
We provide a general theoretical analysis of expected out-of-sample utility, also referred to as decision-theoretic classification, for non-decomposable binary classification metrics such as F-measure and Jaccard coefficient.
no code implementations • NeurIPS 2014 • Oluwasanmi O. Koyejo, Nagarajan Natarajan, Pradeep K. Ravikumar, Inderjit S. Dhillon
We consider a fairly large family of performance metrics given by ratios of linear combinations of the four fundamental population quantities.
no code implementations • 22 Nov 2014 • Cho-Jui Hsieh, Nagarajan Natarajan, Inderjit S. Dhillon
For the first case, we propose a "shifted matrix completion" method that recovers M using only a subset of indices corresponding to ones, while for the second case, we propose a "biased matrix completion" method that recovers the (thresholded) binary matrix.
no code implementations • NeurIPS 2013 • Nagarajan Natarajan, Inderjit S. Dhillon, Pradeep K. Ravikumar, Ambuj Tewari
In this paper, we theoretically study the problem of binary classification in the presence of random classification noise --- the learner, instead of seeing the true labels, sees labels that have independently been flipped with some small probability.