no code implementations • 2 Jun 2024 • Hlynur Davíð Hlynsson
Doubly robust learning offers a robust framework for causal inference from observational data by integrating propensity score and outcome modeling.
no code implementations • 26 Aug 2022 • Hlynur Davíð Hlynsson
This method needs only state-reward pairs from the environment for learning the representation.
no code implementations • 7 May 2021 • Hlynur Davíð Hlynsson, Laurenz Wiskott
One of the fundamental challenges in reinforcement learning (RL) is the one of data efficiency: modern algorithms require a very large number of training samples, especially compared to humans, for solving environments with high-dimensional observations.
no code implementations • 20 Sep 2020 • Hlynur Davíð Hlynsson, Merlin Schüler, Robin Schiewer, Tobias Glasmachers, Laurenz Wiskott
The prediction function is used as a forward model for search on a graph in a viewpoint-matching task and the representation learned to maximize predictability is found to outperform a pre-trained representation.
no code implementations • 3 Jul 2019 • Hlynur Davíð Hlynsson, Alberto N. Escalante-B., Laurenz Wiskott
The algorithms are trained on different-sized subsets of the MNIST and Omniglot data sets.
no code implementations • 22 Apr 2019 • Hlynur Davíð Hlynsson, Laurenz Wiskott
Several methods of estimating the mutual information of random variables have been developed in recent years.
no code implementations • 27 Aug 2018 • Merlin Schüler, Hlynur Davíð Hlynsson, Laurenz Wiskott
We propose Power Slow Feature Analysis, a gradient-based method to extract temporally slow features from a high-dimensional input stream that varies on a faster time-scale, as a variant of Slow Feature Analysis (SFA) that allows end-to-end training of arbitrary differentiable architectures and thereby significantly extends the class of models that can effectively be used for slow feature extraction.