no code implementations • 15 Feb 2021 • Adam Kalai, Varun Kanade
We give an efficient algorithm for learning a binary function in a given class C of bounded VC dimension, with training data distributed according to P and test data according to Q, where P and Q may be arbitrary distributions over X.
no code implementations • 13 Feb 2020 • Vikas K. Garg, Adam Kalai, Katrina Ligett, Zhiwei Steven Wu
Domain generalization is the problem of machine learning when the training data and the test data come from different data domains.
no code implementations • 20 Feb 2018 • Steve Hanneke, Adam Kalai, Gautam Kamath, Christos Tzamos
A generative model may generate utter nonsense when it is fit to maximize the likelihood of observed data.
no code implementations • NeurIPS 2018 • Vikas K. Garg, Adam Kalai
We introduce a framework to leverage knowledge acquired from a repository of (heterogeneous) supervised datasets to new unsupervised datasets.
no code implementations • WS 2017 • Shyam Upadhyay, Kai-Wei Chang, Matt Taddy, Adam Kalai, James Zou
We present a multi-view Bayesian non-parametric algorithm which improves multi-sense word embeddings by (a) using multilingual (i. e., more than two languages) corpora to significantly improve sense embeddings beyond what one achieves with bilingual information, and (b) uses a principled approach to learn a variable number of senses per word, in a data-driven manner.
8 code implementations • NeurIPS 2016 • Tolga Bolukbasi, Kai-Wei Chang, James Zou, Venkatesh Saligrama, Adam Kalai
Geometrically, gender bias is first shown to be captured by a direction in the word embedding.
no code implementations • 20 Jun 2016 • Tolga Bolukbasi, Kai-Wei Chang, James Zou, Venkatesh Saligrama, Adam Kalai
Machine learning algorithms are optimized to model statistical properties of the training data.
no code implementations • 18 Feb 2013 • Sivan Sabato, Adam Kalai
When dealing with subjective, noisy, or otherwise nebulous features, the "wisdom of crowds" suggests that one may benefit from multiple judgments of the same feature on the same object.
no code implementations • NeurIPS 2011 • Sham M. Kakade, Varun Kanade, Ohad Shamir, Adam Kalai
In this paper, we provide algorithms for learning GLMs and SIMs, which are both computationally and statistically efficient.
no code implementations • NeurIPS 2009 • Varun Kanade, Adam Kalai
We prove strong noise-tolerance properties of a potential-based boosting algorithm, similar to MadaBoost (Domingo and Watanabe, 2000) and SmoothBoost (Servedio, 2003).
1 code implementation • 15 Oct 2000 • Avrim Blum, Adam Kalai, Hal Wasserman
Hence this natural extension to the statistical query model does not increase the set of weakly learnable functions.