1 code implementation • 17 Feb 2024 • Kyoungseok Jang, Chicheng Zhang, Kwang-Sung Jun
Assuming access to the distribution of the covariates, we propose a novel low-rank matrix estimation method called LowPopArt and provide its recovery guarantee that depends on a novel quantity denoted by B(Q) that characterizes the hardness of the problem, where Q is the covariance matrix of the measurement distribution.
1 code implementation • 28 Dec 2023 • Yichen Li, Chicheng Zhang
We study interactive imitation learning, where a learner interactively queries a demonstrating expert for action annotations, aiming to learn a policy that has performance competitive with the expert, using as few annotations as possible.
no code implementations • 23 Oct 2023 • Yinan Li, Chicheng Zhang
We study the problem of computationally and label efficient PAC active learning $d$-dimensional halfspaces with Tsybakov Noise~\citep{tsybakov2004optimal} under structured unlabeled data distributions.
1 code implementation • NeurIPS 2023 • Hao Qin, Kwang-Sung Jun, Chicheng Zhang
Maillard sampling \cite{maillard13apprentissage}, an attractive alternative to Thompson sampling, has recently been shown to achieve competitive regret guarantees in the sub-Gaussian reward setting \cite{bian2022maillard} while maintaining closed-form action probabilities, which is useful for offline policy evaluation.
1 code implementation • 25 Oct 2022 • Kyoungseok Jang, Chicheng Zhang, Kwang-Sung Jun
In this paper, we propose a simple and computationally efficient sparse linear estimation method called PopArt that enjoys a tighter $\ell_1$ recovery guarantee compared to Lasso (Tibshirani, 1996) in many problems.
no code implementations • 26 Sep 2022 • Yichen Li, Chicheng Zhang
We make the following contributions: (1) we show that in the $\textbf{COIL}$ problem, any proper online learning algorithm cannot guarantee a sublinear regret in general; (2) we propose $\textbf{Logger}$, an improper online learning algorithmic framework, that reduces $\textbf{COIL}$ to online linear optimization, by utilizing a new definition of mixed policy class; (3) we design two oracle-efficient algorithms within the $\textbf{Logger}$ framework that enjoy different sample and interaction round complexity tradeoffs, and conduct finite-sample analyses to show their improvements over naive behavior cloning; (4) we show that under the standard complexity-theoretic assumptions, efficient dynamic regret minimization is infeasible in the $\textbf{Logger}$ framework.
1 code implementation • 17 Jun 2022 • Zhi Wang, Chicheng Zhang, Kamalika Chaudhuri
We study the problem of online multi-task learning where the tasks are performed within similar but not necessarily identical multi-armed bandit environments.
no code implementations • 16 Jun 2022 • Tom Yan, Chicheng Zhang
The fast spreading adoption of machine learning (ML) by companies across industries poses significant regulatory challenges.
no code implementations • 23 Feb 2022 • Tom Yan, Chicheng Zhang
The growing use of machine learning models in consequential settings has highlighted an important and seemingly irreconcilable tension between transparency and vulnerability to gaming.
no code implementations • 28 Oct 2021 • Dharma KC, Chicheng Zhang, Chris Gniady, Parth Sandeep Agarwal, Sushil Sharma
Heart disease is the number one killer, and ECGs can assist in the early diagnosis and prevention of deadly outcomes.
1 code implementation • 15 Aug 2021 • Dharma KC, Chicheng Zhang
We study utilizing auxiliary information in training data to improve the trustworthiness of machine learning models.
no code implementations • NeurIPS 2021 • Chicheng Zhang, Zhi Wang
We study multi-task reinforcement learning (RL) in tabular episodic Markov decision processes (MDPs).
no code implementations • 10 Feb 2021 • Chicheng Zhang, Yinan Li
We give a computationally-efficient PAC active learning algorithm for $d$-dimensional homogeneous halfspaces that can tolerate Massart noise (Massart and N\'ed\'elec, 2006) and Tsybakov noise (Tsybakov, 2004).
1 code implementation • 29 Oct 2020 • Zhi Wang, Chicheng Zhang, Manish Kumar Singh, Laurel D. Riek, Kamalika Chaudhuri
In many real-world applications, multiple agents seek to learn how to perform highly related yet slightly different tasks in an online bandit learning protocol.
no code implementations • 25 Jun 2020 • Yining Chen, Haipeng Luo, Tengyu Ma, Chicheng Zhang
We propose a surprisingly simple algorithm that adaptively balances its regret and its number of label queries in settings where the data streams are from a mixture of hidden domains.
no code implementations • NeurIPS 2020 • Kwang-Sung Jun, Chicheng Zhang
In this paper, we focus on the finite hypothesis case and ask if one can achieve the asymptotic optimality while enjoying bounded regret whenever possible.
no code implementations • ICML Workshop LifelongML 2020 • Yining Chen, Haipeng Luo, Tengyu Ma, Chicheng Zhang
We propose a surprisingly simple algorithm that adaptively balances its regret and its number of label queries in settings where the data streams are from a mixture of hidden domains.
1 code implementation • NeurIPS 2020 • Maryam Majzoubi, Chicheng Zhang, Rajan Chari, Akshay Krishnamurthy, John Langford, Aleksandrs Slivkins
We create a computationally tractable algorithm for contextual bandits with continuous actions having unknown structure.
no code implementations • 6 Jun 2020 • Jie Shen, Chicheng Zhang
We answer this question in the affirmative by designing a computationally efficient active learning algorithm with near-optimal label complexity of $\tilde{O}\big({s \log^4 \frac d \epsilon} \big)$ and noise tolerance $\eta = \Omega(\epsilon)$, where $\epsilon \in (0, 1)$ is the target error rate, under the assumption that the distribution over (uncorrupted) unlabeled examples is isotropic log-concave.
no code implementations • NeurIPS 2020 • Chicheng Zhang, Jie Shen, Pranjal Awasthi
Even in the presence of mild label noise, i. e. $\eta$ is a small constant, this is a challenging problem and only recently have label complexity bounds of the form $\tilde{O}\big(s \cdot \mathrm{polylog}(d, \frac{1}{\epsilon})\big)$ been established in [Zhang, 2018] for computationally efficient algorithms.
6 code implementations • ICLR 2020 • Jordan T. Ash, Chicheng Zhang, Akshay Krishnamurthy, John Langford, Alekh Agarwal
We design a new algorithm for batch active learning with deep neural network models.
no code implementations • 6 Feb 2019 • Alina Beygelzimer, Dávid Pál, Balázs Szörényi, Devanathan Thiruvenkatachari, Chen-Yu Wei, Chicheng Zhang
Under the more challenging weak linear separability condition, we design an efficient algorithm with a mistake bound of $\min (2^{\widetilde{O}(K \log^2 (1/\gamma))}, 2^{\widetilde{O}(\sqrt{1/\gamma} \log K)})$.
no code implementations • 5 Feb 2019 • Akshay Krishnamurthy, John Langford, Aleksandrs Slivkins, Chicheng Zhang
We study contextual bandit learning with an abstract policy class and continuous action space.
1 code implementation • 2 Jan 2019 • Chicheng Zhang, Alekh Agarwal, Hal Daumé III, John Langford, Sahand N. Negahban
We investigate the feasibility of learning from a mix of both fully-labeled supervised data and contextual bandit data.
no code implementations • 7 May 2018 • Chicheng Zhang
We study the problem of efficient PAC active learning of homogeneous linear classifiers (halfspaces) in $\mathbb{R}^d$, where the goal is to learn a halfspace with low error using as few label queries as possible.
no code implementations • 7 Feb 2018 • Chicheng Zhang, Eran A. Mukamel, Kamalika Chaudhuri
We consider learning parameters of Binomial Hidden Markov Models, which may be used to model DNA methylation data.
no code implementations • ICML 2017 • Alina Beygelzimer, Francesco Orabona, Chicheng Zhang
An efficient bandit algorithm for $\sqrt{T}$-regret in online multiclass prediction?
no code implementations • 25 Feb 2017 • Alina Beygelzimer, Francesco Orabona, Chicheng Zhang
The regret bound holds simultaneously with respect to a family of loss functions parameterized by $\eta$, for a range of $\eta$ restricted by the norm of the competitor.
no code implementations • NeurIPS 2017 • Songbai Yan, Chicheng Zhang
It has been a long-standing problem to efficiently learn a halfspace using as few labels as possible in the presence of noise.
no code implementations • 21 Apr 2016 • Chicheng Zhang, Kamalika Chaudhuri
In this paper, we address both challenges.
no code implementations • NeurIPS 2016 • Alina Beygelzimer, Daniel Hsu, John Langford, Chicheng Zhang
We investigate active learning with access to two distinct oracles: Label (which is standard) and Search (which is not).
no code implementations • NeurIPS 2015 • Chicheng Zhang, Kamalika Chaudhuri
This work addresses active learning with labels obtained from strong and weak labelers, where in addition to the standard active learning setting, we have an extra weak labeler which may occasionally provide incorrect labels.
no code implementations • NeurIPS 2015 • Chicheng Zhang, Jimin Song, Kevin C Chen, Kamalika Chaudhuri
We develop a latent variable model and an efficient spectral algorithm motivated by the recent emergence of very large data sets of chromatin marks from multiple human cell types.
no code implementations • NeurIPS 2014 • Chicheng Zhang, Kamalika Chaudhuri
We study agnostic active learning, where the goal is to learn a classifier in a pre-specified hypothesis class interactively with as few label queries as possible, while making no assumptions on the true function generating the labels.