Search Results for author: Hlynur Davíð Hlynsson

Found 7 papers, 0 papers with code

A Tutorial on Doubly Robust Learning for Causal Inference

no code implementations2 Jun 2024 Hlynur Davíð Hlynsson

Doubly robust learning offers a robust framework for causal inference from observational data by integrating propensity score and outcome modeling.

Causal Inference

Visual processing in context of reinforcement learning

no code implementations26 Aug 2022 Hlynur Davíð Hlynsson

This method needs only state-reward pairs from the environment for learning the representation.

reinforcement-learning Reinforcement Learning +2

Reward prediction for representation learning and reward shaping

no code implementations7 May 2021 Hlynur Davíð Hlynsson, Laurenz Wiskott

One of the fundamental challenges in reinforcement learning (RL) is the one of data efficiency: modern algorithms require a very large number of training samples, especially compared to humans, for solving environments with high-dimensional observations.

Reinforcement Learning (RL) Representation Learning

Latent Representation Prediction Networks

no code implementations20 Sep 2020 Hlynur Davíð Hlynsson, Merlin Schüler, Robin Schiewer, Tobias Glasmachers, Laurenz Wiskott

The prediction function is used as a forward model for search on a graph in a viewpoint-matching task and the representation learned to maximize predictability is found to outperform a pre-trained representation.

Navigate

Learning gradient-based ICA by neurally estimating mutual information

no code implementations22 Apr 2019 Hlynur Davíð Hlynsson, Laurenz Wiskott

Several methods of estimating the mutual information of random variables have been developed in recent years.

blind source separation

Gradient-based Training of Slow Feature Analysis by Differentiable Approximate Whitening

no code implementations27 Aug 2018 Merlin Schüler, Hlynur Davíð Hlynsson, Laurenz Wiskott

We propose Power Slow Feature Analysis, a gradient-based method to extract temporally slow features from a high-dimensional input stream that varies on a faster time-scale, as a variant of Slow Feature Analysis (SFA) that allows end-to-end training of arbitrary differentiable architectures and thereby significantly extends the class of models that can effectively be used for slow feature extraction.

Cannot find the paper you are looking for? You can Submit a new open access paper.