Search Results for author: Guy Blanc

Found 19 papers, 1 papers with code

Lifting uniform learners via distributional decomposition

no code implementations27 Mar 2023 Guy Blanc, Jane Lange, Ali Malik, Li-Yang Tan

We show how any PAC learning algorithm that works under the uniform distribution can be transformed, in a blackbox fashion, into one that works under an arbitrary and unknown distribution $\mathcal{D}$.

LEMMA PAC learning +1

Subsampling Suffices for Adaptive Data Analysis

no code implementations17 Feb 2023 Guy Blanc

Ensuring that analyses performed on a dataset are representative of the entire population is one of the central problems in statistics.

Multitask Learning via Shared Features: Algorithms and Hardness

no code implementations7 Sep 2022 Konstantina Bairaktari, Guy Blanc, Li-Yang Tan, Jonathan Ullman, Lydia Zakynthinou

We investigate the computational efficiency of multitask learning of Boolean functions over the $d$-dimensional hypercube, that are related by means of a feature representation of size $k \ll d$ shared across all tasks.

Attribute Computational Efficiency

A Query-Optimal Algorithm for Finding Counterfactuals

no code implementations14 Jul 2022 Guy Blanc, Caleb Koch, Jane Lange, Li-Yang Tan

Here $S(f)$ is the sensitivity of $f$, a discrete analogue of the Lipschitz constant, and $\Delta_f(x^\star)$ is the distance from $x^\star$ to its nearest counterfactuals.

counterfactual

Open Problem: Properly learning decision trees in polynomial time?

no code implementations29 Jun 2022 Guy Blanc, Jane Lange, Mingda Qiao, Li-Yang Tan

The previous fastest algorithm for this problem ran in $n^{O(\log n)}$ time, a consequence of Ehrenfeucht and Haussler (1989)'s classic algorithm for the distribution-free setting.

Popular decision tree algorithms are provably noise tolerant

no code implementations17 Jun 2022 Guy Blanc, Jane Lange, Ali Malik, Li-Yang Tan

Using the framework of boosting, we prove that all impurity-based decision tree learning algorithms, including the classic ID3, C4. 5, and CART, are highly noise tolerant.

On the power of adaptivity in statistical adversaries

no code implementations19 Nov 2021 Guy Blanc, Jane Lange, Ali Malik, Li-Yang Tan

Specifically, can the behavior of an algorithm $\mathcal{A}$ in the presence of oblivious adversaries always be well-approximated by that of an algorithm $\mathcal{A}'$ in the presence of adaptive adversaries?

Provably efficient, succinct, and precise explanations

no code implementations NeurIPS 2021 Guy Blanc, Jane Lange, Li-Yang Tan

We consider the problem of explaining the predictions of an arbitrary blackbox model $f$: given query access to $f$ and an instance $x$, output a small set of $x$'s features that in conjunction essentially determines $f(x)$.

Learning Theory

Properly learning decision trees in almost polynomial time

no code implementations1 Sep 2021 Guy Blanc, Jane Lange, Mingda Qiao, Li-Yang Tan

We give an $n^{O(\log\log n)}$-time membership query algorithm for properly and agnostically learning decision trees under the uniform distribution over $\{\pm 1\}^n$.

Decision tree heuristics can fail, even in the smoothed setting

no code implementations2 Jul 2021 Guy Blanc, Jane Lange, Mingda Qiao, Li-Yang Tan

Greedy decision tree learning heuristics are mainstays of machine learning practice, but theoretical justification for their empirical success remains elusive.

Learning stochastic decision trees

no code implementations8 May 2021 Guy Blanc, Jane Lange, Li-Yang Tan

Given an $\eta$-corrupted set of uniform random samples labeled by a size-$s$ stochastic decision tree, our algorithm runs in time $n^{O(\log(s/\varepsilon)/\varepsilon^2)}$ and returns a hypothesis with error within an additive $2\eta + \varepsilon$ of the Bayes optimal.

Reconstructing decision trees

no code implementations16 Dec 2020 Guy Blanc, Jane Lange, Li-Yang Tan

We give the first {\sl reconstruction algorithm} for decision trees: given queries to a function $f$ that is $\mathrm{opt}$-close to a size-$s$ decision tree, our algorithm provides query access to a decision tree $T$ where: $\circ$ $T$ has size $S = s^{O((\log s)^2/\varepsilon^3)}$; $\circ$ $\mathrm{dist}(f, T)\le O(\mathrm{opt})+\varepsilon$; $\circ$ Every query to $T$ is answered with $\mathrm{poly}((\log s)/\varepsilon)\cdot \log n$ queries to $f$ and in $\mathrm{poly}((\log s)/\varepsilon)\cdot n\log n$ time.

Learning Theory

Estimating decision tree learnability with polylogarithmic sample complexity

no code implementations NeurIPS 2020 Guy Blanc, Neha Gupta, Jane Lange, Li-Yang Tan

We show that top-down decision tree learning heuristics are amenable to highly efficient learnability estimation: for monotone target functions, the error of the decision tree hypothesis constructed by these heuristics can be estimated with polylogarithmically many labeled examples, exponentially smaller than the number necessary to run these heuristics, and indeed, exponentially smaller than information-theoretic minimum required to learn a good decision tree.

Universal guarantees for decision tree induction via a higher-order splitting criterion

no code implementations NeurIPS 2020 Guy Blanc, Neha Gupta, Jane Lange, Li-Yang Tan

We propose a simple extension of top-down decision tree learning heuristics such as ID3, C4. 5, and CART.

Efficient hyperparameter optimization by way of PAC-Bayes bound minimization

1 code implementation14 Aug 2020 John J. Cherian, Andrew G. Taube, Robert T. McGibbon, Panagiotis Angelikopoulos, Guy Blanc, Michael Snarski, Daniel D. Richman, John L. Klepeis, David E. Shaw

Identifying optimal values for a high-dimensional set of hyperparameters is a problem that has received growing attention given its importance to large-scale machine learning applications such as neural architecture search.

Hyperparameter Optimization Neural Architecture Search

Provable guarantees for decision tree induction: the agnostic setting

no code implementations ICML 2020 Guy Blanc, Jane Lange, Li-Yang Tan

We give strengthened provable guarantees on the performance of widely employed and empirically successful {\sl top-down decision tree learning heuristics}.

Top-down induction of decision trees: rigorous guarantees and inherent limitations

no code implementations18 Nov 2019 Guy Blanc, Jane Lange, Li-Yang Tan

We analyze the quality of this heuristic, obtaining near-matching upper and lower bounds: $\circ$ Upper bound: For every $f$ with decision tree size $s$ and every $\varepsilon \in (0,\frac1{2})$, this heuristic builds a decision tree of size at most $s^{O(\log(s/\varepsilon)\log(1/\varepsilon))}$.

Implicit regularization for deep neural networks driven by an Ornstein-Uhlenbeck like process

no code implementations19 Apr 2019 Guy Blanc, Neha Gupta, Gregory Valiant, Paul Valiant

We characterize the behavior of the training dynamics near any parameter vector that achieves zero training error, in terms of an implicit regularization term corresponding to the sum over the data points, of the squared $\ell_2$ norm of the gradient of the model with respect to the parameter vector, evaluated at each data point.

Adaptive Sampled Softmax with Kernel Based Sampling

no code implementations ICML 2018 Guy Blanc, Steffen Rendle

We empirically study the trade-off of bias, sampling distribution and sample size and show that kernel based sampling results in low bias with few samples.

Cannot find the paper you are looking for? You can Submit a new open access paper.