Search Results for author: Hanna Tseran

Found 3 papers, 2 papers with code

Mildly Overparameterized ReLU Networks Have a Favorable Loss Landscape

no code implementations31 May 2023 Kedar Karhadkar, Michael Murray, Hanna Tseran, Guido Montúfar

We study the loss landscape of both shallow and deep, mildly overparameterized ReLU neural networks on a generic finite input dataset for the squared error loss.

Expected Gradients of Maxout Networks and Consequences to Parameter Initialization

1 code implementation17 Jan 2023 Hanna Tseran, Guido Montúfar

We study the gradients of a maxout network with respect to inputs and parameters and obtain bounds for the moments depending on the architecture and the parameter distribution.

On the Expected Complexity of Maxout Networks

1 code implementation NeurIPS 2021 Hanna Tseran, Guido Montúfar

Learning with neural networks relies on the complexity of the representable functions, but more importantly, the particular assignment of typical parameters to functions of different complexity.

Cannot find the paper you are looking for? You can Submit a new open access paper.