Search Results for author: Alon Brutzkus

Found 14 papers, 1 papers with code

How Uniform Random Weights Induce Non-uniform Bias: Typical Interpolating Neural Networks Generalize with Narrow Teachers

no code implementations9 Feb 2024 Gon Buzaglo, Itamar Harel, Mor Shpigel Nacson, Alon Brutzkus, Nathan Srebro, Daniel Soudry

We prove that such a random NN interpolator typically generalizes well if there exists an underlying narrow ``teacher NN" that agrees with the labels.

A Theoretical Analysis of Fine-tuning with Linear Teachers

no code implementations NeurIPS 2021 Gal Shachaf, Alon Brutzkus, Amir Globerson

We further present results on shallow ReLU models, and analyze the dependence of sample complexity there on source and target tasks.

Inductive Bias regression

On Learning Read-once DNFs With Neural Networks

no code implementations1 Jan 2021 Ido Bronstein, Alon Brutzkus, Amir Globerson

Here we focus on this setting where the functions are learned by a convex neural network and gradient descent.

Inductive Bias

On the Inductive Bias of a CNN for Distributions with Orthogonal Patterns

no code implementations1 Jan 2021 Alon Brutzkus, Amir Globerson

In this work, we consider a simplified image classification task where images contain orthogonal patches and are learned with a 3-layer overparameterized convolutional network and stochastic gradient descent (SGD).

Image Classification Inductive Bias

On the Optimality of Trees Generated by ID3

no code implementations11 Jul 2019 Alon Brutzkus, Amit Daniely, Eran Malach

Since its inception in the 1980s, ID3 has become one of the most successful and widely used algorithms for learning decision trees.

ID3 Learns Juntas for Smoothed Product Distributions

no code implementations20 Jun 2019 Alon Brutzkus, Amit Daniely, Eran Malach

In recent years, there are many attempts to understand popular heuristics.

Over-parameterization Improves Generalization in the XOR Detection Problem

no code implementations ICLR 2019 Alon Brutzkus, Amir Globerson

Empirical evidence suggests that neural networks with ReLU activations generalize better with over-parameterization.

Low Latency Privacy Preserving Inference

1 code implementation ICLR 2019 Alon Brutzkus, Oren Elisha, Ran Gilad-Bachrach

When applying machine learning to sensitive data, one has to find a balance between accuracy, information security, and computational-complexity.

Privacy Preserving Transfer Learning

Why do Larger Models Generalize Better? A Theoretical Perspective via the XOR Problem

no code implementations ICLR 2019 Alon Brutzkus, Amir Globerson

Empirical evidence suggests that neural networks with ReLU activations generalize better with over-parameterization.

Clustering

SGD Learns Over-parameterized Networks that Provably Generalize on Linearly Separable Data

no code implementations ICLR 2018 Alon Brutzkus, Amir Globerson, Eran Malach, Shai Shalev-Shwartz

Neural networks exhibit good generalization behavior in the over-parameterized regime, where the number of network parameters exceeds the number of observations.

Generalization Bounds

Globally Optimal Gradient Descent for a ConvNet with Gaussian Inputs

no code implementations ICML 2017 Alon Brutzkus, Amir Globerson

Deep learning models are often successfully trained using gradient descent, despite the worst case hardness of the underlying non-convex optimization problem.

Cannot find the paper you are looking for? You can Submit a new open access paper.