Search Results for author: Jonas Kohler

Found 17 papers, 5 papers with code

Escaping Saddles with Stochastic Gradients

no code implementations ICML 2018 Hadi Daneshmand, Jonas Kohler, Aurelien Lucchi, Thomas Hofmann

We analyze the variance of stochastic gradients along negative curvature directions in certain non-convex machine learning models and show that stochastic gradients exhibit a strong component along these directions.

Adaptive norms for deep learning with regularized Newton methods

no code implementations22 May 2019 Jonas Kohler, Leonard Adolphs, Aurelien Lucchi

We investigate the use of regularized Newton methods with adaptive norms for optimizing neural networks.

The Role of Memory in Stochastic Optimization

no code implementations2 Jul 2019 Antonio Orvieto, Jonas Kohler, Aurelien Lucchi

We first derive a general continuous-time model that can incorporate arbitrary types of memory, for both deterministic and stochastic settings.

Stochastic Optimization

Ellipsoidal Trust Region Methods for Neural Network Training

no code implementations25 Sep 2019 Leonard Adolphs, Jonas Kohler, Aurelien Lucchi

We investigate the use of ellipsoidal trust region constraints for second-order optimization of neural networks.

A Sub-sampled Tensor Method for Non-convex Optimization

no code implementations23 Nov 2019 Aurelien Lucchi, Jonas Kohler

We present a stochastic optimization method that uses a fourth-order regularized model to find local minima of smooth and potentially non-convex objective functions with a finite-sum structure.

Stochastic Optimization

Batch Normalization Provably Avoids Rank Collapse for Randomly Initialised Deep Networks

no code implementations3 Mar 2020 Hadi Daneshmand, Jonas Kohler, Francis Bach, Thomas Hofmann, Aurelien Lucchi

Randomly initialized neural networks are known to become harder to train with increasing depth, unless architectural enhancements like residual connections and batch normalization are used.

Two-Level K-FAC Preconditioning for Deep Learning

no code implementations1 Nov 2020 Nikolaos Tselepidis, Jonas Kohler, Antonio Orvieto

In the context of deep learning, many optimization methods use gradient covariance information in order to accelerate the convergence of Stochastic Gradient Descent.

Vocal Bursts Valence Prediction

Batch normalization provably avoids ranks collapse for randomly initialised deep networks

no code implementations NeurIPS 2020 Hadi Daneshmand, Jonas Kohler, Francis Bach, Thomas Hofmann, Aurelien Lucchi

Randomly initialized neural networks are known to become harder to train with increasing depth, unless architectural enhancements like residual connections and batch normalization are used.

Learning Generative Models of Textured 3D Meshes from Real-World Images

1 code implementation ICCV 2021 Dario Pavllo, Jonas Kohler, Thomas Hofmann, Aurelien Lucchi

Recent advances in differentiable rendering have sparked an interest in learning generative models of textured 3D meshes from image collections.

Pose Estimation

This Looks Like That... Does it? Shortcomings of Latent Space Prototype Interpretability in Deep Networks

1 code implementation5 May 2021 Adrian Hoffmann, Claudio Fanconi, Rahul Rade, Jonas Kohler

Deep neural networks that yield human interpretable decisions by architectural design have lately become an increasingly popular alternative to post hoc interpretation of traditional black-box models.

Explainable artificial intelligence Image Classification +1

Vanishing Curvature and the Power of Adaptive Methods in Randomly Initialized Deep Networks

no code implementations7 Jun 2021 Antonio Orvieto, Jonas Kohler, Dario Pavllo, Thomas Hofmann, Aurelien Lucchi

This paper revisits the so-called vanishing gradient phenomenon, which commonly occurs in deep randomly initialized neural networks.

Adaptive Guidance: Training-free Acceleration of Conditional Diffusion Models

1 code implementation19 Dec 2023 Angela Castillo, Jonas Kohler, Juan C. Pérez, Juan Pablo Pérez, Albert Pumarola, Bernard Ghanem, Pablo Arbeláez, Ali Thabet

Our findings provide insights into the efficiency of the conditional denoising process that contribute to more practical and swift deployment of text-conditioned diffusion models.

Denoising Neural Architecture Search

fMPI: Fast Novel View Synthesis in the Wild with Layered Scene Representations

no code implementations26 Dec 2023 Jonas Kohler, Nicolas Griffiths Sanchez, Luca Cavalli, Catherine Herold, Albert Pumarola, Alberto Garcia Garcia, Ali Thabet

In this study, we propose two novel input processing paradigms for novel view synthesis (NVS) methods based on layered scene representations that significantly improve their runtime without compromising quality.

Novel View Synthesis

Cannot find the paper you are looking for? You can Submit a new open access paper.