Search Results for author: Karthikeyan Shanmugam

Found 65 papers, 21 papers with code

A resource-constrained stochastic scheduling algorithm for homeless street outreach and gleaning edible food

no code implementations15 Mar 2024 Conor M. Artman, Aditya Mate, Ezinne Nwankwo, Aliza Heching, Tsuyoshi Idé, Jiří\, Navrátil, Karthikeyan Shanmugam, Wei Sun, Kush R. Varshney, Lauri Goldkind, Gidi Kroch, Jaclyn Sawyer, Ian Watson

We developed a common algorithmic solution addressing the problem of resource-constrained outreach encountered by social change organizations with different missions and operations: Breaking Ground -- an organization that helps individuals experiencing homelessness in New York transition to permanent housing and Leket -- the national food bank of Israel that rescues food from farms and elsewhere to feed the hungry.

Scheduling Thompson Sampling

General Identifiability and Achievability for Causal Representation Learning

1 code implementation24 Oct 2023 Burak Varici, Emre Acartürk, Karthikeyan Shanmugam, Ali Tajer

For identifiability, the paper establishes that perfect recovery of the latent causal model and variables is guaranteed under uncoupled interventions.

Representation Learning

Fairness under Covariate Shift: Improving Fairness-Accuracy tradeoff with few Unlabeled Test Samples

1 code implementation11 Oct 2023 Shreyas Havaldar, Jatin Chauhan, Karthikeyan Shanmugam, Jay Nandy, Aravindan Raghuveer

Our third contribution is theoretical, where we show that our weighted entropy term along with prediction loss on the training set approximates test loss under covariate shift.

Fairness Out-of-Distribution Generalization

Optimal Best-Arm Identification in Bandits with Access to Offline Data

no code implementations15 Jun 2023 Shubhada Agrawal, Sandeep Juneja, Karthikeyan Shanmugam, Arun Sai Suggala

Learning paradigms based purely on offline data as well as those based solely on sequential online learning have been well-studied in the literature.

InfoNCE Loss Provably Learns Cluster-Preserving Representations

no code implementations15 Feb 2023 Advait Parulekar, Liam Collins, Karthikeyan Shanmugam, Aryan Mokhtari, Sanjay Shakkottai

The goal of contrasting learning is to learn a representation that preserves underlying clusters by keeping samples with similar content, e. g. the ``dogness'' of a dog, close to each other in the space generated by the representation.

Score-based Causal Representation Learning with Interventions

no code implementations19 Jan 2023 Burak Varici, Emre Acarturk, Karthikeyan Shanmugam, Abhishek Kumar, Ali Tajer

The objectives are: (i) recovering the unknown linear transformation (up to scaling) and (ii) determining the directed acyclic graph (DAG) underlying the latent variables.

Representation Learning valid

Optimal Algorithms for Latent Bandits with Cluster Structure

no code implementations17 Jan 2023 Soumyabrata Pal, Arun Sai Suggala, Karthikeyan Shanmugam, Prateek Jain

Instead, we propose LATTICE (Latent bAndiTs via maTrIx ComplEtion) which allows exploitation of the latent cluster structure to provide the minimax optimal regret of $\widetilde{O}(\sqrt{(\mathsf{M}+\mathsf{N})\mathsf{T}})$, when the number of clusters is $\widetilde{O}(1)$.

Matrix Completion Recommendation Systems

Selective classification using a robust meta-learning approach

no code implementations12 Dec 2022 Nishant Jain, Karthikeyan Shanmugam, Pradeep Shenoy

Predictive uncertainty-a model's self awareness regarding its accuracy on an input-is key for both building robust models via training interventions and for test-time applications such as selective classification.

Bilevel Optimization Classification +3

Causal Bandits for Linear Structural Equation Models

1 code implementation26 Aug 2022 Burak Varici, Karthikeyan Shanmugam, Prasanna Sattigeri, Ali Tajer

Two linear mechanisms, one soft intervention and one observational, are assumed for each node, giving rise to $2^N$ possible interventions.

Thompson Sampling

Causal Graphs Underlying Generative Models: Path to Learning with Limited Data

no code implementations14 Jul 2022 Samuel C. Hoffman, Kahini Wadhawan, Payel Das, Prasanna Sattigeri, Karthikeyan Shanmugam

In this work, we provide a simple algorithm that relies on perturbation experiments on latent codes of a pre-trained generative autoencoder to uncover a causal graph that is implied by the generative model.

Attribute

PAC Generalization via Invariant Representations

no code implementations30 May 2022 Advait Parulekar, Karthikeyan Shanmugam, Sanjay Shakkottai

These are representations of the covariates such that the best model on top of the representation is invariant across training environments.

Out-of-Distribution Generalization PAC learning

Fourier Representations for Black-Box Optimization over Categorical Variables

no code implementations8 Feb 2022 Hamid Dadkhahi, Jesus Rios, Karthikeyan Shanmugam, Payel Das

In order to improve the performance and sample efficiency of such algorithms, we propose to use existing methods in conjunction with a surrogate model for the black-box evaluations over purely categorical variables.

regression Thompson Sampling

Auto-Transfer: Learning to Route Transferrable Representations

1 code implementation2 Feb 2022 Keerthiram Murugesan, Vijay Sadashivaiah, Ronny Luss, Karthikeyan Shanmugam, Pin-Yu Chen, Amit Dhurandhar

Knowledge transfer between heterogeneous source and target networks and tasks has received a lot of attention in recent times as large amounts of quality labeled data can be difficult to obtain in many applications.

Transfer Learning

CoFrNets: Interpretable Neural Architecture Inspired by Continued Fractions

no code implementations NeurIPS 2021 Isha Puri, Amit Dhurandhar, Tejaswini Pedapati, Karthikeyan Shanmugam, Dennis Wei, Kush R. Varshney

We experiment on nonlinear synthetic functions and are able to accurately model as well as estimate feature attributions and even higher order terms in some cases, which is a testament to the representational power as well as interpretability of such architectures.

Scalable Intervention Target Estimation in Linear Models

1 code implementation NeurIPS 2021 Burak Varici, Karthikeyan Shanmugam, Prasanna Sattigeri, Ali Tajer

This paper considers the problem of estimating the unknown intervention targets in a causal directed acyclic graph from observational and interventional data.

Auto-Transfer: Learning to Route Transferable Representations

no code implementations ICLR 2022 Keerthiram Murugesan, Vijay Sadashivaiah, Ronny Luss, Karthikeyan Shanmugam, Pin-Yu Chen, Amit Dhurandhar

Knowledge transfer between heterogeneous source and target networks and tasks has received a lot of attention in recent times as large amounts of quality labelled data can be difficult to obtain in many applications.

Transfer Learning

Episodic Bandits with Stochastic Experts

no code implementations7 Jul 2021 Nihal Sharma, Soumya Basu, Karthikeyan Shanmugam, Sanjay Shakkottai

The agent interacts with the environment over episodes, with each episode having different context distributions; this results in the `best expert' changing across episodes.

Finite-Sample Analysis of Off-Policy TD-Learning via Generalized Bellman Operators

no code implementations NeurIPS 2021 Zaiwei Chen, Siva Theja Maguluri, Sanjay Shakkottai, Karthikeyan Shanmugam

Our key step is to show that the generalized Bellman operator is simultaneously a contraction mapping with respect to a weighted $\ell_p$-norm for each $p$ in $[1,\infty)$, with a common contraction factor.

Finding Valid Adjustments under Non-ignorability with Minimal DAG Knowledge

1 code implementation22 Jun 2021 Abhin Shah, Karthikeyan Shanmugam, Kartik Ahuja

Our main result strengthens these prior results by showing that under a different expert-driven structural knowledge -- that one variable is a direct causal parent of treatment variable -- remarkably, testing for subsets (not involving the known parent variable) that are valid back-doors is equivalent to an invariance test.

Causal Inference Representation Learning +1

Treatment Effect Estimation using Invariant Risk Minimization

2 code implementations13 Mar 2021 Abhin Shah, Kartik Ahuja, Karthikeyan Shanmugam, Dennis Wei, Kush Varshney, Amit Dhurandhar

Inferring causal individual treatment effect (ITE) from observational data is a challenging problem whose difficulty is exacerbated by the presence of treatment assignment bias.

Domain Generalization regression

Efficient Encrypted Inference on Ensembles of Decision Trees

no code implementations5 Mar 2021 Kanthi Sarpatwar, Karthik Nandakumar, Nalini Ratha, James Rayfield, Karthikeyan Shanmugam, Sharath Pankanti, Roman Vaculin

In this work, we propose a framework to transfer knowledge extracted by complex decision tree ensembles to shallow neural networks (referred to as DTNets) that are highly conducive to encrypted inference.

BIG-bench Machine Learning

A Lyapunov Theory for Finite-Sample Guarantees of Asynchronous Q-Learning and TD-Learning Variants

no code implementations2 Feb 2021 Zaiwei Chen, Siva Theja Maguluri, Sanjay Shakkottai, Karthikeyan Shanmugam

As a by-product, by analyzing the convergence bounds of $n$-step TD and TD$(\lambda)$, we provide theoretical insights into the bias-variance trade-off, i. e., efficiency of bootstrapping in RL.

Q-Learning Reinforcement Learning (RL)

Active Structure Learning of Causal DAGs via Directed Clique Trees

1 code implementation NeurIPS 2020 Chandler Squires, Sara Magliacane, Kristjan Greenewald, Dmitriy Katz, Murat Kocaoglu, Karthikeyan Shanmugam

Most existing works focus on \textit{worst-case} or \textit{average-case} lower bounds for the number of interventions required to orient a DAG.

Selection bias

Causal Discovery from Soft Interventions with Unknown Targets: Characterization and Learning

no code implementations NeurIPS 2020 Amin Jaber, Murat Kocaoglu, Karthikeyan Shanmugam, Elias Bareinboim

One fundamental problem in the empirical sciences is of reconstructing the causal structure that underlies a phenomenon of interest through observation and experimentation.

Causal Discovery

Stochastic Linear Bandits with Protected Subspace

no code implementations2 Nov 2020 Advait Parulekar, Soumya Basu, Aditya Gopalan, Karthikeyan Shanmugam, Sanjay Shakkottai

We study a variant of the stochastic linear bandit problem wherein we optimize a linear objective function but rewards are accrued only orthogonal to an unknown subspace (which we interpret as a \textit{protected space}) given only zero-order stochastic oracle access to both the objective itself and protected subspace.

Active Structure Learning of Causal DAGs via Directed Clique Tree

4 code implementations1 Nov 2020 Chandler Squires, Sara Magliacane, Kristjan Greenewald, Dmitriy Katz, Murat Kocaoglu, Karthikeyan Shanmugam

Most existing works focus on worst-case or average-case lower bounds for the number of interventions required to orient a DAG.

Selection bias

Empirical or Invariant Risk Minimization? A Sample Complexity Perspective

3 code implementations ICLR 2021 Kartik Ahuja, Jun Wang, Amit Dhurandhar, Karthikeyan Shanmugam, Kush R. Varshney

Recently, invariant risk minimization (IRM) was proposed as a promising solution to address out-of-distribution (OOD) generalization.

Linear Regression Games: Convergence Guarantees to Approximate Out-of-Distribution Solutions

3 code implementations28 Oct 2020 Kartik Ahuja, Karthikeyan Shanmugam, Amit Dhurandhar

In Ahuja et al., it was shown that solving for the Nash equilibria of a new class of "ensemble-games" is equivalent to solving IRM.

regression

Causal Feature Selection for Algorithmic Fairness

no code implementations10 Jun 2020 Sainyam Galhotra, Karthikeyan Shanmugam, Prasanna Sattigeri, Kush R. Varshney

In this work, we consider fairness in the integration component of data management, aiming to identify features that improve prediction without adding any bias to the dataset.

Data Integration Fairness +2

A Multi-Channel Neural Graphical Event Model with Negative Evidence

no code implementations21 Feb 2020 Tian Gao, Dharmashankar Subramanian, Karthikeyan Shanmugam, Debarun Bhattacharjya, Nicholas Mattei

Event datasets are sequences of events of various types occurring irregularly over the time-line, and they are increasingly prevalent in numerous domains.

Learning Global Transparent Models Consistent with Local Contrastive Explanations

no code implementations NeurIPS 2020 Tejaswini Pedapati, Avinash Balakrishnan, Karthikeyan Shanmugam, Amit Dhurandhar

Based on a key insight we propose a novel method where we create custom boolean features from sparse local contrastive explanations of the black-box model and then train a globally transparent model on just these, and showcase empirically that such models have higher local consistency compared with other known strategies, while still being close in performance to models that are trained with access to the original data.

counterfactual

On Under-exploration in Bandits with Mean Bounds from Confounded Data

no code implementations19 Feb 2020 Nihal Sharma, Soumya Basu, Karthikeyan Shanmugam, Sanjay Shakkottai

We study a variant of the multi-armed bandit problem where side information in the form of bounds on the mean of each arm is provided.

Invariant Risk Minimization Games

3 code implementations ICML 2020 Kartik Ahuja, Karthikeyan Shanmugam, Kush R. Varshney, Amit Dhurandhar

The standard risk minimization paradigm of machine learning is brittle when operating in environments whose test distributions are different from the training distribution due to spurious correlations.

BIG-bench Machine Learning Image Classification

Characterization and Learning of Causal Graphs with Latent Variables from Soft Interventions

no code implementations NeurIPS 2019 Murat Kocaoglu, Amin Jaber, Karthikeyan Shanmugam, Elias Bareinboim

We introduce a novel notion of interventional equivalence class of causal graphs with latent variables based on these invariances, which associates each graphical structure with a set of interventional distributions that respect the do-calculus rules.

Differentially Private Distributed Data Summarization under Covariate Shift

no code implementations NeurIPS 2019 Kanthi Sarpatwar, Karthikeyan Shanmugam, Venkata Sitaramagiridharganesh Ganapavarapu, Ashish Jagmohan, Roman Vaculin

Our central result is a novel protocol that (a) ensures the curator accesses at most $O(K^{\frac{1}{3}}|D_s| + |D_v|)$ points (b) has formal privacy guarantees on the leakage of information between the data owners and (c) closely matches the best known non-private greedy algorithm.

Data Summarization Prototype Selection

Leveraging Simple Model Predictions for Enhancing its Performance

no code implementations25 Sep 2019 Amit Dhurandhar, Karthikeyan Shanmugam, Ronny Luss

Our method also leverages the per sample hardness estimate of the simple model which is not the case with the prior works which primarily consider the complex model's confidences/predictions and is thus conceptually novel.

Mix and Match: An Optimistic Tree-Search Approach for Learning Models from Mixture Distributions

1 code implementation NeurIPS 2020 Matthew Faw, Rajat Sen, Karthikeyan Shanmugam, Constantine Caramanis, Sanjay Shakkottai

We consider a covariate shift problem where one has access to several different training datasets for the same learning problem and a small validation set which possibly differs from all the individual training distributions.

Model Agnostic Contrastive Explanations for Structured Data

no code implementations31 May 2019 Amit Dhurandhar, Tejaswini Pedapati, Avinash Balakrishnan, Pin-Yu Chen, Karthikeyan Shanmugam, Ruchir Puri

Recently, a method [7] was proposed to generate contrastive explanations for differentiable models such as deep neural networks, where one has complete access to the model.

Enhancing Simple Models by Exploiting What They Already Know

no code implementations ICML 2020 Amit Dhurandhar, Karthikeyan Shanmugam, Ronny Luss

Our method also leverages the per sample hardness estimate of the simple model which is not the case with the prior works which primarily consider the complex model's confidences/predictions and is thus conceptually novel.

Small Data Image Classification

Leveraging Latent Features for Local Explanations

2 code implementations29 May 2019 Ronny Luss, Pin-Yu Chen, Amit Dhurandhar, Prasanna Sattigeri, Yunfeng Zhang, Karthikeyan Shanmugam, Chun-Chen Tu

As the application of deep neural networks proliferates in numerous areas such as medical imaging, video surveillance, and self driving cars, the need for explaining the decisions of these models has become a hot research topic, both at the global and local level.

General Classification Open-Ended Question Answering +1

Size of Interventional Markov Equivalence Classes in Random DAG Models

no code implementations5 Mar 2019 Dmitriy Katz, Karthikeyan Shanmugam, Chandler Squires, Caroline Uhler

For constant density, we show that the expected $\log$ observational MEC size asymptotically (in the number of vertices) approaches a constant.

Causal Inference Experimental Design

Improving Simple Models with Confidence Profiles

no code implementations NeurIPS 2018 Amit Dhurandhar, Karthikeyan Shanmugam, Ronny Luss, Peder Olsen

Our transfer method involves a theoretically justified weighting of samples during the training of the simple model using confidence scores of these intermediate layers.

Mimic and Classify : A meta-algorithm for Conditional Independence Testing

1 code implementation25 Jun 2018 Rajat Sen, Karthikeyan Shanmugam, Himanshu Asnani, Arman Rahimzamani, Sreeram Kannan

Given independent samples generated from the joint distribution $p(\mathbf{x},\mathbf{y},\mathbf{z})$, we study the problem of Conditional Independence (CI-Testing), i. e., whether the joint equals the CI distribution $p^{CI}(\mathbf{x},\mathbf{y},\mathbf{z})= p(\mathbf{z}) p(\mathbf{y}|\mathbf{z})p(\mathbf{x}|\mathbf{z})$ or not.

Structure Learning from Time Series with False Discovery Control

no code implementations24 May 2018 Bernat Guillen Pegueroles, Bhanukiran Vinzamuri, Karthikeyan Shanmugam, Steve Hedden, Jonathan D. Moyer, Kush R. Varshney

Almost all existing Granger causal algorithms condition on a large number of variables (all but two variables) to test for effects between a pair of variables.

Time Series Time Series Analysis

Confidence Scoring Using Whitebox Meta-models with Linear Classifier Probes

no code implementations14 May 2018 Tongfei Chen, Jiří Navrátil, Vijay Iyengar, Karthikeyan Shanmugam

We propose a novel confidence scoring mechanism for deep neural networks based on a two-model paradigm involving a base model and a meta-model.

Contextual Bandits with Stochastic Experts

1 code implementation23 Feb 2018 Rajat Sen, Karthikeyan Shanmugam, Nihal Sharma, Sanjay Shakkottai

We consider the problem of contextual bandits with stochastic experts, which is a variation of the traditional stochastic contextual bandit with experts problem.

Multi-Armed Bandits

Experimental Design for Learning Causal Graphs with Latent Variables

no code implementations NeurIPS 2017 Murat Kocaoglu, Karthikeyan Shanmugam, Elias Bareinboim

Next, we propose an algorithm that uses only O(d^2 log n) interventions that can learn the latents between both non-adjacent and adjacent variables.

Experimental Design

A Formal Framework to Characterize Interpretability of Procedures

no code implementations12 Jul 2017 Amit Dhurandhar, Vijay Iyengar, Ronny Luss, Karthikeyan Shanmugam

We provide a novel notion of what it means to be interpretable, looking past the usual association with human understanding.

TIP: Typifying the Interpretability of Procedures

no code implementations9 Jun 2017 Amit Dhurandhar, Vijay Iyengar, Ronny Luss, Karthikeyan Shanmugam

This leads to the insight that the improvement in the target model is not only a function of the oracle model's performance, but also its relative complexity with respect to the target model.

Knowledge Distillation

Sparse Quadratic Logistic Regression in Sub-quadratic Time

no code implementations8 Mar 2017 Karthikeyan Shanmugam, Murat Kocaoglu, Alexandros G. Dimakis, Sujay Sanghavi

We consider support recovery in the quadratic logistic regression setting - where the target depends on both p linear terms $x_i$ and up to $p^2$ quadratic terms $x_i x_j$.

regression

Identifying Best Interventions through Online Importance Sampling

no code implementations ICML 2017 Rajat Sen, Karthikeyan Shanmugam, Alexandros G. Dimakis, Sanjay Shakkottai

Motivated by applications in computational advertising and systems biology, we consider the problem of identifying the best out of several possible soft interventions at a source node $V$ in an acyclic causal directed graph, to maximize the expected value of a target node $Y$ (located downstream of $V$).

Contextual Bandits with Latent Confounders: An NMF Approach

no code implementations1 Jun 2016 Rajat Sen, Karthikeyan Shanmugam, Murat Kocaoglu, Alexandros G. Dimakis, Sanjay Shakkottai

Our algorithm achieves a regret of $\mathcal{O}\left(L\mathrm{poly}(m, \log K) \log T \right)$ at time $T$, as compared to $\mathcal{O}(LK\log T)$ for conventional contextual bandits, assuming a constant gap between the best arm and the rest for each context.

Matrix Completion Multi-Armed Bandits

Learning Causal Graphs with Small Interventions

2 code implementations NeurIPS 2015 Karthikeyan Shanmugam, Murat Kocaoglu, Alexandros G. Dimakis, Sriram Vishwanath

We prove that any deterministic adaptive algorithm needs to be a separating system in order to learn complete graphs in the worst case.

On the Information Theoretic Limits of Learning Ising Models

no code implementations NeurIPS 2014 Karthikeyan Shanmugam, Rashish Tandon, Alexandros G. Dimakis, Pradeep Ravikumar

We provide a general framework for computing lower-bounds on the sample complexity of recovering the underlying graphs of Ising models, given i. i. d samples.

Sparse Polynomial Learning and Graph Sketching

no code implementations NeurIPS 2014 Murat Kocaoglu, Karthikeyan Shanmugam, Alexandros G. Dimakis, Adam Klivans

We give an algorithm for exactly reconstructing f given random examples from the uniform distribution on $\{-1, 1\}^n$ that runs in time polynomial in $n$ and $2s$ and succeeds if the function satisfies the unique sign property: there is one output value which corresponds to a unique set of values of the participating parities.

Cannot find the paper you are looking for? You can Submit a new open access paper.