Search Results for author: Guido Montúfar

Found 30 papers, 11 papers with code

Fisher-Rao Gradient Flows of Linear Programs and State-Action Natural Policy Gradients

no code implementations28 Mar 2024 Johannes Müller, Semih Çaycı, Guido Montúfar

Kakade's natural policy gradient method has been studied extensively in the last years showing linear convergence with and without regularization.

The Real Tropical Geometry of Neural Networks

no code implementations18 Mar 2024 Marie-Charlotte Brandenburg, Georg Loho, Guido Montúfar

The parameter space of ReLU neural networks is contained as a semialgebraic set inside the parameter space of tropical rational functions.

Benign overfitting in leaky ReLU networks with moderate input dimension

no code implementations11 Mar 2024 Kedar Karhadkar, Erin George, Michael Murray, Guido Montúfar, Deanna Needell

The problem of benign overfitting asks whether it is possible for a model to perfectly fit noisy training data and still generalize well.

Attribute Binary Classification

Mildly Overparameterized ReLU Networks Have a Favorable Loss Landscape

no code implementations31 May 2023 Kedar Karhadkar, Michael Murray, Hanna Tseran, Guido Montúfar

We study the loss landscape of both shallow and deep, mildly overparameterized ReLU neural networks on a generic finite input dataset for the squared error loss.

Function Space and Critical Points of Linear Convolutional Networks

no code implementations12 Apr 2023 Kathlén Kohn, Guido Montúfar, Vahid Shahverdi, Matthew Trager

We study the geometry of linear networks with one-dimensional convolutional layers.

Critical Points and Convergence Analysis of Generative Deep Linear Networks Trained with Bures-Wasserstein Loss

1 code implementation6 Mar 2023 Pierre Bréchet, Katerina Papagiannouli, Jing An, Guido Montúfar

We consider a deep matrix factorization model of covariance matrices trained with the Bures-Wasserstein distance.

Expected Gradients of Maxout Networks and Consequences to Parameter Initialization

1 code implementation17 Jan 2023 Hanna Tseran, Guido Montúfar

We study the gradients of a maxout network with respect to inputs and parameters and obtain bounds for the moments depending on the architecture and the parameter distribution.

Geometry and convergence of natural policy gradient methods

no code implementations3 Nov 2022 Johannes Müller, Guido Montúfar

We study the convergence of several natural policy gradient (NPG) methods in infinite-horizon discounted Markov decision processes with regular policy parametrizations.

Policy Gradient Methods

FoSR: First-order spectral rewiring for addressing oversquashing in GNNs

1 code implementation21 Oct 2022 Kedar Karhadkar, Pradeep Kr. Banerjee, Guido Montúfar

On the other hand, adding edges to the message-passing graph can lead to increasingly similar node representations and a problem known as oversmoothing.

Graph Classification

Enumeration of max-pooling responses with generalized permutohedra

no code implementations29 Sep 2022 Laura Escobar, Patricio Gallardo, Javier González-Anaya, José L. González, Guido Montúfar, Alejandro H. Morales

We investigate the combinatorics of max-pooling layers, which are functions that downsample input arrays by taking the maximum over shifted windows of input coordinates, and which are commonly used in convolutional neural networks.

Oversquashing in GNNs through the lens of information contraction and graph expansion

1 code implementation6 Aug 2022 Pradeep Kr. Banerjee, Kedar Karhadkar, Yu Guang Wang, Uri Alon, Guido Montúfar

We compare the spectral expansion properties of our algorithm with that of an existing curvature-based non-local rewiring strategy.

graph construction

On the effectiveness of persistent homology

1 code implementation21 Jun 2022 Renata Turkeš, Guido Montúfar, Nina Otter

The goal of this work is to identify some types of problems where PH performs well or even better than other methods in data analysis.

Topological Data Analysis

Solving infinite-horizon POMDPs with memoryless stochastic policies in state-action space

1 code implementation27 May 2022 Johannes Müller, Guido Montúfar

Reward optimization in fully observable Markov decision processes is equivalent to a linear program over the polytope of state-action frequencies.

Training Wasserstein GANs without gradient penalties

no code implementations27 Oct 2021 Dohyun Kwon, Yeoneung Kim, Guido Montúfar, Insoon Yang

We propose a stable method to train Wasserstein generative adversarial networks.

Learning curves for Gaussian process regression with power-law priors and targets

no code implementations ICLR 2022 Hui Jin, Pradeep Kr. Banerjee, Guido Montúfar

We characterize the power-law asymptotics of learning curves for Gaussian process regression (GPR) under the assumption that the eigenspectrum of the prior and the eigenexpansion coefficients of the target function follow a power law.

GPR regression

The Geometry of Memoryless Stochastic Policy Optimization in Infinite-Horizon POMDPs

2 code implementations ICLR 2022 Johannes Müller, Guido Montúfar

We then describe the optimization problem as a linear optimization problem in the space of feasible state-action frequencies subject to polynomial constraints that we characterize explicitly.

Geometry of Linear Convolutional Networks

no code implementations3 Aug 2021 Kathlén Kohn, Thomas Merkh, Guido Montúfar, Matthew Trager

We study the family of functions that are represented by a linear convolutional neural network (LCN).

On the Expected Complexity of Maxout Networks

1 code implementation NeurIPS 2021 Hanna Tseran, Guido Montúfar

Learning with neural networks relies on the complexity of the representable functions, but more importantly, the particular assignment of typical parameters to functions of different complexity.

Information Complexity and Generalization Bounds

no code implementations4 May 2021 Pradeep Kr. Banerjee, Guido Montúfar

We present a unifying picture of PAC-Bayesian and mutual information-based upper bounds on the generalization error of randomized learning algorithms.

Generalization Bounds

Sharp bounds for the number of regions of maxout networks and vertices of Minkowski sums

no code implementations16 Apr 2021 Guido Montúfar, Yue Ren, Leon Zhang

We present results on the number of linear regions of the functions that can be represented by artificial feedforward neural networks with maxout units.

Can neural networks learn persistent homology features?

no code implementations NeurIPS Workshop TDA_and_Beyond 2020 Guido Montúfar, Nina Otter, Yuguang Wang

Topological data analysis uses tools from topology -- the mathematical area that studies shapes -- to create representations of data.

Topological Data Analysis

Distributed Learning via Filtered Hyperinterpolation on Manifolds

no code implementations18 Jul 2020 Guido Montúfar, Yu Guang Wang

Learning mappings of data on manifolds is an important topic in contemporary machine learning, with applications in astrophysics, geophysics, statistical physics, medical diagnosis, biochemistry, 3D object analysis.

Geophysics Medical Diagnosis

Stochastic Feedforward Neural Networks: Universal Approximation

no code implementations22 Oct 2019 Thomas Merkh, Guido Montúfar

We investigate different types of shallow and deep architectures, and the minimal number of layers and units per layer that are sufficient and necessary in order for the network to be able to approximate any given stochastic mapping from the set of inputs to the set of outputs arbitrarily well.

How Well Do WGANs Estimate the Wasserstein Metric?

no code implementations9 Oct 2019 Anton Mallasto, Guido Montúfar, Augusto Gerolin

Generative modelling is often cast as minimizing a similarity measure between a data distribution and a model distribution.

The Variational Deficiency Bottleneck

no code implementations27 Oct 2018 Pradeep Kr. Banerjee, Guido Montúfar

We introduce a bottleneck method for learning data representations based on information deficiency, rather than the more traditional information sufficiency.

General Classification

On the Number of Linear Regions of Deep Neural Networks

no code implementations NeurIPS 2014 Guido Montúfar, Razvan Pascanu, Kyunghyun Cho, Yoshua Bengio

We study the complexity of functions computable by deep feedforward neural networks with piecewise linear activations in terms of the symmetries and the number of linear regions that they have.

Cannot find the paper you are looking for? You can Submit a new open access paper.