Search Results for author: Robert Bamler

Found 26 papers, 11 papers with code

Your Finetuned Large Language Model is Already a Powerful Out-of-distribution Detector

no code implementations7 Apr 2024 Andi Zhang, Tim Z. Xiao, Weiyang Liu, Robert Bamler, Damon Wischik

We revisit the likelihood ratio between a pretrained large language model (LLM) and its finetuned variant as a criterion for out-of-distribution (OOD) detection.

Language Modelling Large Language Model +3

Predictive, scalable and interpretable knowledge tracing on structured domains

1 code implementation19 Mar 2024 Hanqi Zhou, Robert Bamler, Charley M. Wu, Álvaro Tejero-Cantero

This requires estimates of both the learner's progress (''knowledge tracing''; KT), and the prerequisite structure of the learning domain (''knowledge mapping'').

Bayesian Inference Continual Learning +1

A Compact Representation for Bayesian Neural Networks By Removing Permutation Symmetry

1 code implementation31 Dec 2023 Tim Z. Xiao, Weiyang Liu, Robert Bamler

Bayesian neural networks (BNNs) are a principled approach to modeling predictive uncertainties in deep learning, which are important in safety-critical applications.

Bayesian Inference Variational Inference

Upgrading VAE Training With Unlimited Data Plans Provided by Diffusion Models

no code implementations30 Oct 2023 Tim Z. Xiao, Johannes Zenn, Robert Bamler

Variational autoencoders (VAEs) are popular models for representation learning but their encoders are susceptible to overfitting (Cremer et al., 2018) because they are trained on a finite training set instead of the true (continuous) data distribution $p_{\mathrm{data}}(\mathbf{x})$.

Data Augmentation Representation Learning

The SVHN Dataset Is Deceptive for Probabilistic Generative Models Due to a Distribution Mismatch

no code implementations30 Oct 2023 Tim Z. Xiao, Johannes Zenn, Robert Bamler

However, with this work, we aim to warn the community about an issue of the SVHN dataset as a benchmark for generative modeling tasks: we discover that the official split into training set and test set of the SVHN dataset are not drawn from the same distribution.

Classification

Resampling Gradients Vanish in Differentiable Sequential Monte Carlo Samplers

1 code implementation27 Apr 2023 Johannes Zenn, Robert Bamler

Annealed Importance Sampling (AIS) moves particles along a Markov chain from a tractable initial distribution to an intractable target distribution.

Trading Information between Latents in Hierarchical Variational Autoencoders

1 code implementation9 Feb 2023 Tim Z. Xiao, Robert Bamler

Variational Autoencoders (VAEs) were originally motivated (Kingma & Welling, 2014) as probabilistic generative models in which one performs approximate Bayesian inference.

Bayesian Inference Data Compression +1

Hybridizing Physical and Data-driven Prediction Methods for Physicochemical Properties

no code implementations17 Feb 2022 Fabian Jirasek, Robert Bamler, Stephan Mandt

We apply the new approach to predict activity coefficients at infinite dilution and obtain significant improvements compared to the data-driven and physical baselines and established ensemble methods from the machine learning literature.

Bayesian Inference BIG-bench Machine Learning

User-Dependent Neural Sequence Models for Continuous-Time Event Data

1 code implementation NeurIPS 2020 Alex Boyd, Robert Bamler, Stephan Mandt, Padhraic Smyth

Modeling such data can be very challenging, in particular for applications with many different types of events, since it requires a model to predict the event types as well as the time of occurrence.

Variational Inference

Variational Bayesian Quantization

2 code implementations ICML 2020 Yibo Yang, Robert Bamler, Stephan Mandt

Our experimental results demonstrate the importance of taking into account posterior uncertainties, and show that image compression with the proposed algorithm outperforms JPEG over a wide range of bit rates using only a single standard VAE.

Image Compression Model Compression +2

Extreme Classification via Adversarial Softmax Approximation

1 code implementation ICLR 2020 Robert Bamler, Stephan Mandt

Training a classifier over a large number of classes, known as 'extreme classification', has become a topic of major interest with applications in technology, science, and e-commerce.

Classification General Classification

Machine Learning in Thermodynamics: Prediction of Activity Coefficients by Matrix Completion

no code implementations29 Jan 2020 Fabian Jirasek, Rodrigo A. S. Alves, Julie Damay, Robert A. Vandermeulen, Robert Bamler, Michael Bortz, Stephan Mandt, Marius Kloft, Hans Hasse

Activity coefficients, which are a measure of the non-ideality of liquid mixtures, are a key property in chemical engineering with relevance to modeling chemical and phase equilibria as well as transport processes.

BIG-bench Machine Learning Matrix Completion

A Quantum Field Theory of Representation Learning

no code implementations4 Jul 2019 Robert Bamler, Stephan Mandt

Continuous symmetries and their breaking play a prominent role in contemporary physics.

Representation Learning Time Series +1

Augmenting and Tuning Knowledge Graph Embeddings

1 code implementation1 Jul 2019 Robert Bamler, Farnood Salehi, Stephan Mandt

Knowledge graph embeddings rank among the most successful methods for link prediction in knowledge graphs, i. e., the task of completing an incomplete collection of relational facts.

Knowledge Graph Embeddings Knowledge Graphs +1

Probabilistic Knowledge Graph Embeddings

no code implementations27 Sep 2018 Farnood Salehi, Robert Bamler, Stephan Mandt

We develop a probabilistic extension of state-of-the-art embedding models for link prediction in relational knowledge graphs.

Knowledge Graph Embeddings Knowledge Graphs +2

Improving Optimization in Models With Continuous Symmetry Breaking

no code implementations ICML 2018 Robert Bamler, Stephan Mandt

We show that representation learning models for time series possess an approximate continuous symmetry that leads to slow convergence of gradient descent.

Representation Learning Time Series +2

Improving Optimization for Models With Continuous Symmetry Breaking

no code implementations8 Mar 2018 Robert Bamler, Stephan Mandt

We show that representation learning models for time series possess an approximate continuous symmetry that leads to slow convergence of gradient descent.

Representation Learning Time Series +2

Bayesian Paragraph Vectors

no code implementations10 Nov 2017 Geng Ji, Robert Bamler, Erik B. Sudderth, Stephan Mandt

Word2vec (Mikolov et al., 2013) has proven to be successful in natural language processing by capturing the semantic relationships between different words.

Sentiment Analysis Word Embeddings

Perturbative Black Box Variational Inference

no code implementations NeurIPS 2017 Robert Bamler, Cheng Zhang, Manfred Opper, Stephan Mandt

Black box variational inference (BBVI) with reparameterization gradients triggered the exploration of divergence measures other than the Kullback-Leibler (KL) divergence, such as alpha divergences.

Gaussian Processes Variational Inference

Structured Black Box Variational Inference for Latent Time Series Models

no code implementations4 Jul 2017 Robert Bamler, Stephan Mandt

Black box variational inference with reparameterization gradients (BBVI) allows us to explore a rich new class of Bayesian non-conjugate latent time series models; however, a naive application of BBVI to such structured variational models would scale quadratically in the number of time steps.

Collaborative Filtering Time Series +3

Dynamic Word Embeddings

1 code implementation ICML 2017 Robert Bamler, Stephan Mandt

We present a probabilistic language model for time-stamped text data which tracks the semantic evolution of individual words over time.

Language Modelling Variational Inference +1

Cannot find the paper you are looking for? You can Submit a new open access paper.