no code implementations • 28 Oct 2024 • Prakhar Verma, Sukruta Prakash Midigeshi, Gaurav Sinha, Arno Solin, Nagarajan Natarajan, Amit Sharma
We introduce Planning-guided Retrieval Augmented Generation (Plan$\times$RAG), a novel framework that augments the \emph{retrieve-then-reason} paradigm of existing RAG frameworks to \emph{plan-then-retrieve}.
1 code implementation • 18 Aug 2024 • Jatin Prakash, Anirudh Buvanesh, Bishal Santra, Deepak Saini, Sachin Yadav, Jian Jiao, Yashoteja Prabhu, Amit Sharma, Manik Varma
Extreme Classification (XC) aims to map a query to the most relevant documents from a very large document set.
no code implementations • 10 Jul 2024 • Aniket Vashishtha, Abhinav Kumar, Abbavaram Gowtham Reddy, Vineeth N Balasubramanian, Amit Sharma
Specifically, we consider an axiomatic training setup where an agent learns from multiple demonstrations of a causal axiom (or rule), rather than incorporating the axiom as an inductive bias or inferring it from data values.
no code implementations • 20 Jun 2024 • Amit Sharma, Hua Li, Xue Li, Jian Jiao
We evaluate the proposed algorithm on improving novelty for a query-ad recommendation task on a large-scale search engine.
no code implementations • 15 Jun 2024 • Gurusha Juneja, Nagarajan Natarajan, Hua Li, Jian Jiao, Amit Sharma
Given a task in the form of a basic description and its training examples, prompt optimization is the problem of synthesizing the given information into a text prompt for a large language model (LLM).
no code implementations • 12 Apr 2024 • Amit Sharma, Teodor-Dumitru Ene, Kishor Kunal, Mingjie Liu, Zafar Hasan, Haoxing Ren
This paper presents a comparative analysis of total cost of ownership (TCO) and performance between domain-adapted large language models (LLM) and state-of-the-art (SoTA) LLMs , with a particular emphasis on tasks related to coding assistance for chip design.
no code implementations • 24 Feb 2024 • Ekansh Chauhan, Amit Sharma, Megha S Uppin, C. V. Jawahar, P. K. Vinod
It establishes new performance benchmarks in glioma subtype classification across multiple datasets, including a novel dataset focused on the Indian demographic (IPD- Brain), providing a valuable resource for existing research.
1 code implementation • 9 Feb 2024 • Pragya Srivastava, Satvik Golechha, Amit Deshpande, Amit Sharma
Recent work shows that in-context learning and optimization of in-context examples (ICE) can significantly improve the accuracy of large language models (LLMs) on a wide range of tasks, leading to an apparent consensus that ICE optimization is crucial for better performance.
no code implementations • 30 Nov 2023 • Jake M. Hofman, Angelos Chatzimparmpas, Amit Sharma, Duncan J. Watts, Jessica Hullman
Amid rising concerns of reproducibility and generalizability in predictive modeling, we explore the possibility and potential benefits of introducing pre-registration to the field.
no code implementations • 31 Oct 2023 • Daman Arora, Anush Kini, Sayak Ray Chowdhury, Nagarajan Natarajan, Gaurav Sinha, Amit Sharma
Given a query and a document corpus, the information retrieval (IR) task is to output a ranked list of relevant documents.
no code implementations • 23 Oct 2023 • Aniket Vashishtha, Abbavaram Gowtham Reddy, Abhinav Kumar, Saketh Bachu, Vineeth N Balasubramanian, Amit Sharma
At the core of causal inference lies the challenge of determining reliable causal graphs solely based on observational data.
no code implementations • 1 Oct 2023 • Yair Gat, Nitay Calderon, Amir Feder, Alexander Chapanin, Amit Sharma, Roi Reichart
We hence present a second approach based on matching, and propose a method that is guided by an LLM at training-time and learns a dedicated embedding space.
no code implementations • 27 Jun 2023 • Parikshit Bansal, Amit Sharma
Instead, we propose a sampling strategy based on the difference in prediction scores between the base model and the finetuned NLP model, utilizing the fact that most NLP models are finetuned from a base model.
no code implementations • 19 Jun 2023 • Abhinav Kumar, Amit Deshpande, Amit Sharma
We prove that our method only requires that the ranking of estimated causal effects is correct across attributes to select the correct classifier.
no code implementations • 29 May 2023 • Abbavaram Gowtham Reddy, Saketh Bachu, Saloni Dash, Charchit Sharma, Amit Sharma, Vineeth N Balasubramanian
Counterfactual data augmentation has recently emerged as a method to mitigate confounding biases in the training data.
no code implementations • 26 May 2023 • Parikshit Bansal, Amit Sharma
Therefore, using methods from the causal inference literature, we propose an algorithm to regularize the learnt effect of the features on the model's prediction to the estimated effect of feature on label.
1 code implementation • 28 Apr 2023 • Emre Kiciman, Robert Ness, Amit Sharma, Chenhao Tan
The causal capabilities of large language models (LLMs) are a matter of significant debate, with critical implications for the use of LLMs in societally impactful domains such as medicine, science, law, and policy.
no code implementations • 22 Oct 2022 • Abbavaram Gowtham Reddy, Saloni Dash, Amit Sharma, Vineeth N Balasubramanian
Given a causal generative process, we formally characterize the adverse effects of confounding on any downstream tasks and show that the correlation between generative factors (attributes) can be used to quantitatively measure confounding between generative factors.
no code implementations • 7 Oct 2022 • Parikshit Bansal, Yashoteja Prabhu, Emre Kiciman, Amit Sharma
To explain this generalization failure, we consider an intervention-based importance metric, which shows that a fine-tuned model captures spurious correlations and fails to learn the causal features that determine the relevance between any two text inputs.
no code implementations • 17 Aug 2022 • Amit Sharma, Hua Li, Jian Jiao
Specifically, we propose a method to estimate counterfactuals using time-series predictive models and construct an attribution score, CF-Shapley, that is consistent with desirable axioms for attributing an observed change in the output metric.
no code implementations • 8 Jul 2022 • Abhinav Kumar, Chenhao Tan, Amit Sharma
Even under the most favorable conditions for learning a probing classifier when a concept's relevant features in representation space alone can provide 100% accuracy, we prove that a probing classifier is likely to use non-concept features and thus post-hoc or adversarial methods will fail to remove the concept correctly.
no code implementations • 15 Jun 2022 • Jivat Neet Kaur, Emre Kiciman, Amit Sharma
Based on the relationship between spurious attributes and the classification label, we obtain realizations of the canonical causal graph that characterize common distribution shifts and show that each shift entails different independence constraints over observed variables.
1 code implementation • 8 Feb 2022 • Chacha Chen, Shi Feng, Amit Sharma, Chenhao Tan
Our key result is that without assumptions about task-specific intuitions, explanations may potentially improve human understanding of model decision boundary, but they cannot improve human understanding of task decision boundary or model error.
1 code implementation • 4 Feb 2022 • Tomas Geffner, Javier Antoran, Adam Foster, Wenbo Gong, Chao Ma, Emre Kiciman, Amit Sharma, Angus Lamb, Martin Kukla, Nick Pawlowski, Miltiadis Allamanis, Cheng Zhang
Causal inference is essential for data-driven decision making across domains such as business engagement, medical treatment and policy making.
1 code implementation • 26 Dec 2021 • Victor Chernozhukov, Carlos Cinelli, Whitney Newey, Amit Sharma, Vasilis Syrgkanis
We develop a general theory of omitted variable bias for a wide range of common causal parameters, including (but not limited to) averages of potential outcomes, average treatment effects, average causal derivatives, and policy effects from covariate shifts.
no code implementations • 24 Nov 2021 • Sai Srinivas Kancheti, Abbavaram Gowtham Reddy, Vineeth N Balasubramanian, Amit Sharma
A trained neural network can be interpreted as a structural causal model (SCM) that provides the effect of changing input variables on the model's output.
no code implementations • 28 Oct 2021 • Mathias Lécuyer, Sang Hoon Kim, Mihir Nanavati, Junchen Jiang, Siddhartha Sen, Amit Sharma, Aleksandrs Slivkins
We develop a methodology, called Sayer, that leverages implicit feedback to evaluate and train new system policies.
1 code implementation • 7 Oct 2021 • Divyat Mahajan, Shruti Tople, Amit Sharma
Through extensive evaluation on a synthetic dataset and image datasets like MNIST, Fashion-MNIST, and Chest X-rays, we show that a lower OOD generalization gap does not imply better robustness to MI attacks.
1 code implementation • 27 Aug 2021 • Amit Sharma, Vasilis Syrgkanis, Cheng Zhang, Emre Kiciman
Estimation of causal effects involves crucial assumptions about the data-generating process, such as directionality of effect, presence of instrumental variables or mediators, and whether all relevant confounders are observed.
1 code implementation • 23 Aug 2021 • Jason Lequyer, Reuben Philip, Amit Sharma, Laurence Pelletier
Recent approaches have allowed for the denoising of single noisy images without access to any training data aside from that very image.
no code implementations • 27 May 2021 • Varun Chandrasekaran, Darren Edge, Somesh Jha, Amit Sharma, Cheng Zhang, Shruti Tople
However for real-world applications, the privacy of data is critical.
no code implementations • 24 Feb 2021 • Amit Sharma
In this note we introduce a notion of free cofibrations of permutative categories.
Category Theory Algebraic Topology
1 code implementation • 11 Jan 2021 • Alexander Lavin, Ciarán M. Gilligan-Lee, Alessya Visnjic, Siddha Ganju, Dava Newman, Atılım Güneş Baydin, Sujoy Ganguly, Danny Lange, Amit Sharma, Stephan Zheng, Eric P. Xing, Adam Gibson, James Parr, Chris Mattmann, Yarin Gal
The development and deployment of machine learning (ML) systems can be executed easily with modern tools, but the process is typically rushed and means-to-an-end.
no code implementations • 21 Dec 2020 • Naman Goel, Alfonso Amayuelas, Amit Deshpande, Amit Sharma
For example, in multi-stage settings where decisions are made in multiple screening rounds, we use our framework to derive the minimal distributions required to design a fair algorithm.
no code implementations • 11 Nov 2020 • Yanbo Xu, Divyat Mahajan, Liz Manrao, Amit Sharma, Emre Kiciman
For many kinds of interventions, such as a new advertisement, marketing intervention, or feature recommendation, it is important to target a specific subset of people for maximizing its benefits at minimum cost or potential harm.
2 code implementations • 10 Nov 2020 • Ramaravind Kommiya Mothilal, Divyat Mahajan, Chenhao Tan, Amit Sharma
In addition, by restricting the features that can be modified for generating counterfactual examples, we find that the top-k features from LIME or SHAP are often neither necessary nor sufficient explanations of a model's prediction.
4 code implementations • 9 Nov 2020 • Amit Sharma, Emre Kiciman
In addition to efficient statistical estimators of a treatment's effect, successful application of causal inference requires specifying assumptions about the mechanisms underlying observed data and testing whether they are valid, and to what extent.
no code implementations • 17 Sep 2020 • Saloni Dash, Vineeth N Balasubramanian, Amit Sharma
We present a method for generating counterfactuals by incorporating a structural causal model (SCM) in an improved variant of Adversarially Learned Inference (ALI), that generates counterfactuals in accordance with the causal relationships between attributes of an image.
1 code implementation • arXiv 2020 • Divyat Mahajan, Shruti Tople, Amit Sharma
In the domain generalization literature, a common objective is to learn representations independent of the domain after conditioning on the class label.
Ranked #1 on Domain Generalization on Rotated Fashion-MNIST
no code implementations • LREC 2020 • Devansh Mehta, Sebastin Santy, Ramaravind Kommiya Mothilal, Brij Mohan Lal Srivastava, Alok Sharma, Anurag Shukla, Vishnu Prasad, Venkanna U, Amit Sharma, Kalika Bali
The primary obstacle to developing technologies for low-resource languages is the lack of usable data.
3 code implementations • 6 Dec 2019 • Divyat Mahajan, Chenhao Tan, Amit Sharma
For explanations of ML models in critical domains such as healthcare and finance, counterfactual examples are useful for an end-user only to the extent that perturbation of feature inputs is feasible in the real world.
1 code implementation • ICML 2020 • Shruti Tople, Amit Sharma, Aditya Nori
Such privacy risks are exacerbated when a model's predictions are used on an unseen data distribution.
no code implementations • 3 Sep 2019 • Arpita Biswas, Siddharth Barman, Amit Deshpande, Amit Sharma
To quantify this bias, we propose a general notion of $\eta$-infra-marginality that can be used to evaluate the extent of this bias.
no code implementations • 10 Jul 2019 • Rathin Desai, Amit Sharma
We show that many popular methods, including back-door methods can be considered as weighting or representation learning algorithms, and provide general error bounds for their causal estimates.
7 code implementations • 19 May 2019 • Ramaravind Kommiya Mothilal, Amit Sharma, Chenhao Tan
Post-hoc explanations of machine learning models are crucial for people to understand and act on algorithmic predictions.
no code implementations • 5 Feb 2019 • Jackson A. Killian, Bryan Wilder, Amit Sharma, Daksha Shah, Vinod Choudhary, Bistra Dilkina, Milind Tambe
Digital Adherence Technologies (DATs) are an increasingly popular method for verifying patient adherence to many medications.
1 code implementation • 28 Nov 2016 • Amit Sharma, Jake M. Hofman, Duncan J. Watts
We present a method for estimating causal effects in time series data when fine-grained information about the outcome of interest is available.
no code implementations • 23 Jan 2014 • Priyankar Ghosh, Amit Sharma, P. P. Chakrabarti, Pallab Dasgupta
The proposed algorithms use a best first search technique and report the solutions using an implicit representation ordered by cost.