Search Results for author: Jane Lange

Found 15 papers, 0 papers with code

Agnostic proper learning of monotone functions: beyond the black-box correction barrier

no code implementations5 Apr 2023 Jane Lange, Arsen Vasilyan

Given $2^{\tilde{O}(\sqrt{n}/\varepsilon)}$ uniformly random examples of an unknown function $f:\{\pm 1\}^n \rightarrow \{\pm 1\}$, our algorithm outputs a hypothesis $g:\{\pm 1\}^n \rightarrow \{\pm 1\}$ that is monotone and $(\mathrm{opt} + \varepsilon)$-close to $f$, where $\mathrm{opt}$ is the distance from $f$ to the closest monotone function.

Lifting uniform learners via distributional decomposition

no code implementations27 Mar 2023 Guy Blanc, Jane Lange, Ali Malik, Li-Yang Tan

We show how any PAC learning algorithm that works under the uniform distribution can be transformed, in a blackbox fashion, into one that works under an arbitrary and unknown distribution $\mathcal{D}$.

LEMMA PAC learning +1

A Query-Optimal Algorithm for Finding Counterfactuals

no code implementations14 Jul 2022 Guy Blanc, Caleb Koch, Jane Lange, Li-Yang Tan

Here $S(f)$ is the sensitivity of $f$, a discrete analogue of the Lipschitz constant, and $\Delta_f(x^\star)$ is the distance from $x^\star$ to its nearest counterfactuals.

counterfactual

Open Problem: Properly learning decision trees in polynomial time?

no code implementations29 Jun 2022 Guy Blanc, Jane Lange, Mingda Qiao, Li-Yang Tan

The previous fastest algorithm for this problem ran in $n^{O(\log n)}$ time, a consequence of Ehrenfeucht and Haussler (1989)'s classic algorithm for the distribution-free setting.

Popular decision tree algorithms are provably noise tolerant

no code implementations17 Jun 2022 Guy Blanc, Jane Lange, Ali Malik, Li-Yang Tan

Using the framework of boosting, we prove that all impurity-based decision tree learning algorithms, including the classic ID3, C4. 5, and CART, are highly noise tolerant.

On the power of adaptivity in statistical adversaries

no code implementations19 Nov 2021 Guy Blanc, Jane Lange, Ali Malik, Li-Yang Tan

Specifically, can the behavior of an algorithm $\mathcal{A}$ in the presence of oblivious adversaries always be well-approximated by that of an algorithm $\mathcal{A}'$ in the presence of adaptive adversaries?

Provably efficient, succinct, and precise explanations

no code implementations NeurIPS 2021 Guy Blanc, Jane Lange, Li-Yang Tan

We consider the problem of explaining the predictions of an arbitrary blackbox model $f$: given query access to $f$ and an instance $x$, output a small set of $x$'s features that in conjunction essentially determines $f(x)$.

Learning Theory

Properly learning decision trees in almost polynomial time

no code implementations1 Sep 2021 Guy Blanc, Jane Lange, Mingda Qiao, Li-Yang Tan

We give an $n^{O(\log\log n)}$-time membership query algorithm for properly and agnostically learning decision trees under the uniform distribution over $\{\pm 1\}^n$.

Decision tree heuristics can fail, even in the smoothed setting

no code implementations2 Jul 2021 Guy Blanc, Jane Lange, Mingda Qiao, Li-Yang Tan

Greedy decision tree learning heuristics are mainstays of machine learning practice, but theoretical justification for their empirical success remains elusive.

Learning stochastic decision trees

no code implementations8 May 2021 Guy Blanc, Jane Lange, Li-Yang Tan

Given an $\eta$-corrupted set of uniform random samples labeled by a size-$s$ stochastic decision tree, our algorithm runs in time $n^{O(\log(s/\varepsilon)/\varepsilon^2)}$ and returns a hypothesis with error within an additive $2\eta + \varepsilon$ of the Bayes optimal.

Reconstructing decision trees

no code implementations16 Dec 2020 Guy Blanc, Jane Lange, Li-Yang Tan

We give the first {\sl reconstruction algorithm} for decision trees: given queries to a function $f$ that is $\mathrm{opt}$-close to a size-$s$ decision tree, our algorithm provides query access to a decision tree $T$ where: $\circ$ $T$ has size $S = s^{O((\log s)^2/\varepsilon^3)}$; $\circ$ $\mathrm{dist}(f, T)\le O(\mathrm{opt})+\varepsilon$; $\circ$ Every query to $T$ is answered with $\mathrm{poly}((\log s)/\varepsilon)\cdot \log n$ queries to $f$ and in $\mathrm{poly}((\log s)/\varepsilon)\cdot n\log n$ time.

Learning Theory

Estimating decision tree learnability with polylogarithmic sample complexity

no code implementations NeurIPS 2020 Guy Blanc, Neha Gupta, Jane Lange, Li-Yang Tan

We show that top-down decision tree learning heuristics are amenable to highly efficient learnability estimation: for monotone target functions, the error of the decision tree hypothesis constructed by these heuristics can be estimated with polylogarithmically many labeled examples, exponentially smaller than the number necessary to run these heuristics, and indeed, exponentially smaller than information-theoretic minimum required to learn a good decision tree.

Universal guarantees for decision tree induction via a higher-order splitting criterion

no code implementations NeurIPS 2020 Guy Blanc, Neha Gupta, Jane Lange, Li-Yang Tan

We propose a simple extension of top-down decision tree learning heuristics such as ID3, C4. 5, and CART.

Provable guarantees for decision tree induction: the agnostic setting

no code implementations ICML 2020 Guy Blanc, Jane Lange, Li-Yang Tan

We give strengthened provable guarantees on the performance of widely employed and empirically successful {\sl top-down decision tree learning heuristics}.

Top-down induction of decision trees: rigorous guarantees and inherent limitations

no code implementations18 Nov 2019 Guy Blanc, Jane Lange, Li-Yang Tan

We analyze the quality of this heuristic, obtaining near-matching upper and lower bounds: $\circ$ Upper bound: For every $f$ with decision tree size $s$ and every $\varepsilon \in (0,\frac1{2})$, this heuristic builds a decision tree of size at most $s^{O(\log(s/\varepsilon)\log(1/\varepsilon))}$.

Cannot find the paper you are looking for? You can Submit a new open access paper.