Search Results for author: Ingo Steinwart

Found 33 papers, 8 papers with code

Conditioning of Banach Space Valued Gaussian Random Variables: An Approximation Approach Based on Martingales

no code implementations4 Apr 2024 Ingo Steinwart

In this paper we investigate the conditional distributions of two Banach space valued, jointly Gaussian random variables.

Gaussian Processes

Mind the spikes: Benign overfitting of kernels and neural networks in fixed dimension

1 code implementation NeurIPS 2023 Moritz Haas, David Holzmüller, Ulrike Von Luxburg, Ingo Steinwart

In this paper, we show that the smoothness of the estimators, and not the dimension, is the key: benign overfitting is possible if and only if the estimator's derivatives are large enough.

regression

Physics-Informed Gaussian Process Regression Generalizes Linear PDE Solvers

1 code implementation23 Dec 2022 Marvin Pförtner, Ingo Steinwart, Philipp Hennig, Jonathan Wenger

Crucially, this probabilistic viewpoint allows to (1) quantify the inherent discretization error; (2) propagate uncertainty about the model parameters to the solution; and (3) condition on noisy measurements.

Bayesian Inference regression

Utilizing Expert Features for Contrastive Learning of Time-Series Representations

1 code implementation23 Jun 2022 Manuel Nonnenmacher, Lukas Oldenburg, Ingo Steinwart, David Reeb

We therefore devise ExpCLR, a novel contrastive learning approach built on an objective that utilizes expert features to encourage both properties for the learned representation.

Contrastive Learning Representation Learning +2

A Framework and Benchmark for Deep Batch Active Learning for Regression

2 code implementations17 Mar 2022 David Holzmüller, Viktor Zaverkin, Johannes Kästner, Ingo Steinwart

We provide open-source code that includes efficient implementations of all kernels, kernel transformations, and selection methods, and can be used for reproducing our results.

Active Learning regression +1

SOSP: Efficiently Capturing Global Correlations by Second-Order Structured Pruning

1 code implementation NeurIPS 2021 Manuel Nonnenmacher, Thomas Pfeil, Ingo Steinwart, David Reeb

We validate SOSP-H by comparing it to our second method SOSP-I that uses a well-established Hessian approximation, and to numerous state-of-the-art methods.

Fast and Sample-Efficient Interatomic Neural Network Potentials for Molecules and Materials Based on Gaussian Moments

1 code implementation20 Sep 2021 Viktor Zaverkin, David Holzmüller, Ingo Steinwart, Johannes Kästner

Artificial neural networks (NNs) are one of the most frequently used machine learning approaches to construct interatomic potentials and enable efficient large-scale atomistic simulations with almost ab initio accuracy.

Active Learning

Which Minimizer Does My Neural Network Converge To?

no code implementations4 Nov 2020 Manuel Nonnenmacher, David Reeb, Ingo Steinwart

The loss surface of an overparameterized neural network (NN) possesses many global minima of zero training error.

Training Two-Layer ReLU Networks with Gradient Descent is Inconsistent

1 code implementation12 Feb 2020 David Holzmüller, Ingo Steinwart

We prove that two-layer (Leaky)ReLU networks initialized by e. g. the widely used method proposed by He et al. (2015) and trained using gradient descent on a least-squares loss are not universally consistent.

Vocal Bursts Valence Prediction

Reproducing Kernel Hilbert Spaces Cannot Contain all Continuous Functions on a Compact Metric Space

no code implementations8 Feb 2020 Ingo Steinwart

Given an uncountable, compact metric space, we show that there exists no reproducing kernel Hilbert space that contains the space of all continuous functions on this compact space.

Wide Neural Networks are Interpolating Kernel Methods: Impact of Initialization on Generalization

no code implementations25 Sep 2019 Manuel Nonnenmacher, David Reeb, Ingo Steinwart

The recently developed link between strongly overparametrized neural networks (NNs) and kernel methods has opened a new way to understand puzzling features of NNs, such as their convergence and generalization behaviors.

Generalization Bounds

Best-scored Random Forest Classification

no code implementations27 May 2019 Hanyuan Hang, Xiaoyu Liu, Ingo Steinwart

We propose an algorithm named best-scored random forest for binary classification problems.

Binary Classification Classification +1

Empirical Risk Minimization in the Interpolating Regime with Application to Neural Network Learning

no code implementations25 May 2019 Nicole Mücke, Ingo Steinwart

Moreover, we show that the same phenomenon occurs for DNNs with zero training error and sufficiently large architectures.

Learning Theory

A Sober Look at Neural Network Initializations

no code implementations27 Mar 2019 Ingo Steinwart

Initializing the weights and the biases is a key part of the training process of a neural network.

Optimal Learning with Anisotropic Gaussian SVMs

no code implementations4 Oct 2018 Hanyuan Hang, Ingo Steinwart

This paper investigates the nonparametric regression problem using SVMs with anisotropic Gaussian RBF kernels.

regression

Strictly proper kernel scores and characteristic kernels on compact spaces

no code implementations14 Dec 2017 Ingo Steinwart, Johanna F. Ziegel

Strictly proper kernel scores are well-known tool in probabilistic forecasting, while characteristic kernels have been extensively investigated in the machine learning literature.

Translation

Adaptive Clustering Using Kernel Density Estimators

no code implementations17 Aug 2017 Ingo Steinwart, Bharath K. Sriperumbudur, Philipp Thomann

We derive and analyze a generic, recursive algorithm for estimating all splits in a finite cluster tree as well as the corresponding clusters.

Clustering

Learning Rates for Kernel-Based Expectile Regression

no code implementations24 Feb 2017 Muhammad Farooq, Ingo Steinwart

Conditional expectiles are becoming an increasingly important tool in finance as well as in other areas of applications.

regression

Sobolev Norm Learning Rates for Regularized Least-Squares Algorithm

no code implementations23 Feb 2017 Simon Fischer, Ingo Steinwart

In this paper we extend these rates to norms stronger than the $L_2$-norm without requiring the regression function to be contained in the hypothesis space.

regression

liquidSVM: A Fast and Versatile SVM package

1 code implementation22 Feb 2017 Ingo Steinwart, Philipp Thomann

liquidSVM is a package written in C++ that provides SVM-type solvers for various classification and regression tasks.

General Classification regression

Learning with Hierarchical Gaussian Kernels

no code implementations2 Dec 2016 Ingo Steinwart, Philipp Thomann, Nico Schmid

We investigate iterated compositions of weighted sums of Gaussian kernels and provide an interpretation of the construction that shows some similarities with the architectures of deep neural networks.

Spatial Decompositions for Large Scale SVMs

no code implementations1 Dec 2016 Philipp Thomann, Ingrid Blaschzyk, Mona Meister, Ingo Steinwart

Our contributions are two fold: On the theoretical side we establish an oracle inequality for the overall learning method using the hinge loss, and show that the resulting rates match those known for SVMs solving the complete optimization problem with Gaussian kernels.

Kernel Density Estimation for Dynamical Systems

no code implementations13 Jul 2016 Hanyuan Hang, Ingo Steinwart, Yunlong Feng, Johan A. K. Suykens

We study the density estimation problem with observations generated by certain dynamical systems that admit a unique underlying invariant Lebesgue density.

Density Estimation

Learning theory estimates with observations from general stationary stochastic processes

no code implementations10 May 2016 Hanyuan Hang, Yunlong Feng, Ingo Steinwart, Johan A. K. Suykens

We show that when the stochastic processes satisfy a generalized Bernstein-type inequality, a unified treatment on analyzing the learning schemes with various mixing processes can be conducted and a sharp oracle inequality for generic regularized empirical risk minimization schemes can be established.

Learning Theory

Representation of Quasi-Monotone Functionals by Families of Separating Hyperplanes

no code implementations21 Aug 2015 Ingo Steinwart

We characterize when the level sets of a continuous quasi-monotone functional defined on a suitable convex subset of a normed space can be uniquely represented by a family of bounded continuous functionals.

BIG-bench Machine Learning

Towards an Axiomatic Approach to Hierarchical Clustering of Measures

no code implementations15 Aug 2015 Philipp Thomann, Ingo Steinwart, Nico Schmid

We propose some axioms for hierarchical clustering of probability measures and investigate their ramifications.

Clustering

Optimal Learning Rates for Localized SVMs

no code implementations23 Jul 2015 Mona Eberts, Ingo Steinwart

In addition, their motivation was always based on computational requirements.

regression

An SVM-like Approach for Expectile Regression

no code implementations14 Jul 2015 Muhammad Farooq, Ingo Steinwart

Expectile regression is a nice tool for investigating conditional distributions beyond the conditional mean.

regression

Fully adaptive density-based clustering

no code implementations30 Sep 2014 Ingo Steinwart

The clusters of a distribution are often defined by the connected components of a density level set.

Clustering

Optimal learning rates for least squares SVMs using Gaussian kernels

no code implementations NeurIPS 2011 Mona Eberts, Ingo Steinwart

We prove a new oracle inequality for support vector machines with Gaussian RBF kernels solving the regularized least squares regression problem.

regression

Universal Kernels on Non-Standard Input Spaces

no code implementations NeurIPS 2010 Andreas Christmann, Ingo Steinwart

We apply this technique for the following special cases: universal kernels on the set of probability measures, universal kernels based on Fourier transforms, and universal kernels for signal processing.

text-classification Text Classification

Fast Learning from Non-i.i.d. Observations

no code implementations NeurIPS 2009 Ingo Steinwart, Andreas Christmann

We prove an oracle inequality for generic regularized empirical risk minimization algorithms learning from $\a$-mixing processes.

Sparsity of SVMs that use the epsilon-insensitive loss

no code implementations NeurIPS 2008 Ingo Steinwart, Andreas Christmann

In this paper lower and upper bounds for the number of support vectors are derived for support vector machines (SVMs) based on the epsilon-insensitive loss function.

Cannot find the paper you are looking for? You can Submit a new open access paper.