Search Results for author: Berfin Simsek

Found 6 papers, 0 papers with code

Learning Associative Memories with Gradient Descent

no code implementations28 Feb 2024 Vivien Cabannes, Berfin Simsek, Alberto Bietti

This work focuses on the training dynamics of one associative memory module storing outer products of token embeddings.

Memorization

The Loss Landscape of Shallow ReLU-like Neural Networks: Stationary Points, Saddle Escaping, and Network Embedding

no code implementations8 Feb 2024 Zhengqing Wu, Berfin Simsek, Francois Ged

Additionally, we show that, if a stationary point does not contain "escape neurons", which are defined with first-order conditions, then it must be a local minimum.

Network Embedding

Expand-and-Cluster: Parameter Recovery of Neural Networks

no code implementations25 Apr 2023 Flavio Martinelli, Berfin Simsek, Wulfram Gerstner, Johanni Brea

Can we identify the parameters of a neural network by probing its input-output mapping?

Clustering

Understanding out-of-distribution accuracies through quantifying difficulty of test samples

no code implementations28 Mar 2022 Berfin Simsek, Melissa Hall, Levent Sagun

Existing works show that although modern neural networks achieve remarkable generalization performance on the in-distribution (ID) dataset, the accuracy drops significantly on the out-of-distribution (OOD) datasets \cite{recht2018cifar, recht2019imagenet}.

Weight-space symmetry in neural network loss landscapes revisited

no code implementations25 Sep 2019 Berfin Simsek, Johanni Brea, Bernd Illing, Wulfram Gerstner

In a network of $d-1$ hidden layers with $n_k$ neurons in layers $k = 1, \ldots, d$, we construct continuous paths between equivalent global minima that lead through a `permutation point' where the input and output weight vectors of two neurons in the same hidden layer $k$ collide and interchange.

Weight-space symmetry in deep networks gives rise to permutation saddles, connected by equal-loss valleys across the loss landscape

no code implementations5 Jul 2019 Johanni Brea, Berfin Simsek, Bernd Illing, Wulfram Gerstner

The permutation symmetry of neurons in each layer of a deep neural network gives rise not only to multiple equivalent global minima of the loss function, but also to first-order saddle points located on the path between the global minima.

Cannot find the paper you are looking for? You can Submit a new open access paper.