no code implementations • 11 Feb 2025 • Konstantin Donhauser, Charles Arnal, Mohammad Pezeshki, Vivien Cabannes, David Lopez-Paz, Kartik Ahuja
We demonstrate that it's possible to predict which heads are crucial for long-context processing using only local keys.
no code implementations • 8 Oct 2024 • Divyat Mahajan, Mohammad Pezeshki, Ioannis Mitliagkas, Kartik Ahuja, Pascal Vincent
In this work, we tackle a challenging and extreme form of subpopulation shift, which is termed compositional shift.
1 code implementation • 8 Apr 2024 • Artem Vysogorets, Kartik Ahuja, Julia Kempe
In the era of exceptionally data-hungry models, careful selection of the training data is essential to mitigate the extensive costs of deep learning.
1 code implementation • 7 Feb 2024 • Kartik Ahuja, Amin Mansouri
In the first part, we study structured limited capacity variants of different architectures and arrive at the generalization guarantees with limited diversity requirements on the training distribution.
no code implementations • 4 Oct 2023 • Kartik Ahuja, Amin Mansouri, Yixin Wang
Causal representation learning has emerged as the center of action in causal machine learning research.
no code implementations • 18 Sep 2023 • Sharut Gupta, Stefanie Jegelka, David Lopez-Paz, Kartik Ahuja
Two lines of work are taking the central stage in AI research.
1 code implementation • 28 Jun 2023 • Vitória Barin-Pacela, Kartik Ahuja, Simon Lacoste-Julien, Pascal Vincent
We introduce this novel form of identifiability, termed quantized factor identifiability, and provide a comprehensive proof of the recovery of the quantized factors.
1 code implementation • 26 May 2023 • Kartik Ahuja, David Lopez-Paz
In-context learning, a capability that enables a model to learn from input examples on the fly without necessitating weight updates, is a defining characteristic of large language models.
1 code implementation • 20 Dec 2022 • Alexandre Ramé, Kartik Ahuja, Jianyu Zhang, Matthieu Cord, Léon Bottou, David Lopez-Paz
In this paper, we thus propose model ratatouille, a new strategy to recycle the multiple fine-tunings of the same foundation model on diverse auxiliary tasks.
Ranked #19 on
Domain Generalization
on TerraIncognita
1 code implementation • 15 Nov 2022 • Hiroki Naganuma, Kartik Ahuja, Shiro Takagi, Tetsuya Motokawa, Rio Yokota, Kohta Ishikawa, Ikuro Sato, Ioannis Mitliagkas
Modern deep learning systems do not generalize well when the test data distribution is slightly different to the training data distribution.
no code implementations • 31 Oct 2022 • Sharut Gupta, Kartik Ahuja, Mohammad Havaei, Niladri Chatterjee, Yoshua Bengio
Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server.
1 code implementation • 24 Sep 2022 • Kartik Ahuja, Divyat Mahajan, Yixin Wang, Yoshua Bengio
Can interventional data facilitate causal representation learning?
1 code implementation • 2 Jun 2022 • Kartik Ahuja, Jason Hartford, Yoshua Bengio
We show that if the perturbations are applied only on mutually exclusive blocks of latents, we identify the latents up to those blocks.
no code implementations • 23 May 2022 • Sharut Gupta, Kartik Ahuja, Mohammad Havaei, Niladri Chatterjee, Yoshua Bengio
Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server.
no code implementations • 23 May 2022 • Kamalika Chaudhuri, Kartik Ahuja, Martin Arjovsky, David Lopez-Paz
When facing data with imbalanced classes or groups, practitioners follow an intriguing strategy to achieve best results.
1 code implementation • 10 Apr 2022 • Kartik Ahuja, Divyat Mahajan, Vasilis Syrgkanis, Ioannis Mitliagkas
In this work, we depart from these assumptions and ask: a) How can we get disentanglement when the auxiliary information does not provide conditional independence over the factors of variation?
1 code implementation • 18 Mar 2022 • Jean-Christophe Gagnon-Audet, Kartik Ahuja, Mohammad-Javad Darvishi-Bayazi, Pooneh Mousavi, Guillaume Dumas, Irina Rish
We revise the existing OOD generalization algorithms for time series tasks and evaluate them using our systematic framework.
no code implementations • ICLR 2022 • Kartik Ahuja, Jason Hartford, Yoshua Bengio
These results suggest that by exploiting inductive biases on mechanisms, it is possible to design a range of new identifiable representation learning approaches.
no code implementations • 29 Sep 2021 • Amit Dhurandhar, Karthikeyan Natesan Ramamurthy, Kartik Ahuja, Vijay Arya
Locally interpretable model agnostic explanations (LIME) method is one of the most popular methods used to explain black-box models at a per example level.
1 code implementation • 22 Jun 2021 • Abhin Shah, Karthikeyan Shanmugam, Kartik Ahuja
Our main result strengthens these prior results by showing that under a different expert-driven structural knowledge -- that one variable is a direct causal parent of treatment variable -- remarkably, testing for subsets (not involving the known parent variable) that are valid back-doors is equivalent to an invariance test.
2 code implementations • NeurIPS 2021 • Kartik Ahuja, Ethan Caballero, Dinghuai Zhang, Jean-Christophe Gagnon-Audet, Yoshua Bengio, Ioannis Mitliagkas, Irina Rish
To answer these questions, we revisit the fundamental assumptions in linear regression tasks, where invariance-based approaches were shown to provably generalize OOD.
no code implementations • 5 Jun 2021 • Dinghuai Zhang, Kartik Ahuja, Yilun Xu, Yisen Wang, Aaron Courville
Can models with particular structure avoid being biased towards spurious correlation in out-of-distribution (OOD) generalization?
2 code implementations • 4 Jun 2021 • Soroosh Shahtalebi, Jean-Christophe Gagnon-Audet, Touraj Laleh, Mojtaba Faramarzi, Kartik Ahuja, Irina Rish
A major bottleneck in the real-world applications of machine learning models is their failure in generalizing to unseen domains whose data distribution is not i. i. d to the training domains.
2 code implementations • 13 Mar 2021 • Abhin Shah, Kartik Ahuja, Karthikeyan Shanmugam, Dennis Wei, Kush Varshney, Amit Dhurandhar
Inferring causal individual treatment effect (ITE) from observational data is a challenging problem whose difficulty is exacerbated by the presence of treatment assignment bias.
no code implementations • 22 Dec 2020 • Kartik Ahuja, Amit Dhurandhar, Kush R. Varshney
Non-convex optimization problems are challenging to solve; the success and computational expense of a gradient descent algorithm or variant depend heavily on the initialization strategy.
3 code implementations • ICLR 2021 • Kartik Ahuja, Jun Wang, Amit Dhurandhar, Karthikeyan Shanmugam, Kush R. Varshney
Recently, invariant risk minimization (IRM) was proposed as a promising solution to address out-of-distribution (OOD) generalization.
3 code implementations • 28 Oct 2020 • Kartik Ahuja, Karthikeyan Shanmugam, Amit Dhurandhar
In Ahuja et al., it was shown that solving for the Nash equilibria of a new class of "ensemble-games" is equivalent to solving IRM.
1 code implementation • NeurIPS 2021 • Pouya Bashivan, Reza Bayat, Adam Ibrahim, Kartik Ahuja, Mojtaba Faramarzi, Touraj Laleh, Blake Aaron Richards, Irina Rish
Our method, called Adversarial Feature Desensitization (AFD), aims at learning features that are invariant towards adversarial perturbations of the inputs.
3 code implementations • ICML 2020 • Kartik Ahuja, Karthikeyan Shanmugam, Kush R. Varshney, Amit Dhurandhar
The standard risk minimization paradigm of machine learning is brittle when operating in environments whose test distributions are different from the training distribution due to spurious correlations.
1 code implementation • 2 May 2019 • Kartik Ahuja
Recently, a method called the Mutual Information Neural Estimator (MINE) that uses neural networks has been proposed to estimate mutual information and more generally the Kullback-Leibler (KL) divergence between two distributions.
no code implementations • 2 Nov 2018 • Kartik Ahuja, Mihaela van der Schaar
A clinician desires to use a risk-stratification method that achieves confident risk-stratification - the risk estimates of the different patients reflect the true risks with a high probability.
1 code implementation • 26 Oct 2018 • Kartik Ahuja, Mihaela van der Schaar
We use the new metric to develop a variable importance ranking approach.
Methodology
1 code implementation • 27 Jun 2018 • Kartik Ahuja, William Zame, Mihaela van der Schaar
Piecewise local-linear models provide a natural way to extend local-linear models to explain the global behavior of the model.
no code implementations • NeurIPS 2017 • Kartik Ahuja, William Zame, Mihaela van der Schaar
However, there has been limited work to address the personalized screening for these different diseases.