Search Results for author: Bernhard K. Aichernig

Found 6 papers, 1 papers with code

On the Relationship Between RNN Hidden State Vectors and Semantic Ground Truth

1 code implementation29 Jun 2023 Edi Muškardin, Martin Tappler, Ingo Pill, Bernhard K. Aichernig, Thomas Pock

We examine the assumption that the hidden-state vectors of recurrent neural networks (RNNs) tend to form clusters of semantically similar vectors, which we dub the clustering hypothesis.

Clustering

Learning Environment Models with Continuous Stochastic Dynamics

no code implementations29 Jun 2023 Martin Tappler, Edi Muškardin, Bernhard K. Aichernig, Bettina Könighofer

We aim to provide insights into the decisions faced by the agent by learning an automaton model of environmental behavior under the control of an agent.

Acrobot Benchmarking +2

Reinforcement Learning under Partial Observability Guided by Learned Environment Models

no code implementations23 Jun 2022 Edi Muskardin, Martin Tappler, Bernhard K. Aichernig, Ingo Pill

In practical applications, we can rarely assume full observability of a system's environment, despite such knowledge being important for determining a reactive control system's precise interaction with its environment.

Q-Learning reinforcement-learning +1

Search-Based Testing of Reinforcement Learning

no code implementations7 May 2022 Martin Tappler, Filip Cano Córdoba, Bernhard K. Aichernig, Bettina Könighofer

We present a search-based testing framework that enables a wide range of novel analysis capabilities for evaluating the safety and performance of deep RL agents.

reinforcement-learning Reinforcement Learning (RL)

Learning a Behavior Model of Hybrid Systems Through Combining Model-Based Testing and Machine Learning (Full Version)

no code implementations10 Jul 2019 Bernhard K. Aichernig, Roderick Bloem, Masoud Ebrahimi, Martin Horn, Franz Pernkopf, Wolfgang Roth, Astrid Rupp, Martin Tappler, Markus Tranninger

Therefore, there is considerable interest in learning such hybrid behavior by means of machine learning which requires sufficient and representative training data covering the behavior of the physical system adequately.

BIG-bench Machine Learning

L*-Based Learning of Markov Decision Processes (Extended Version)

no code implementations28 Jun 2019 Martin Tappler, Bernhard K. Aichernig, Giovanni Bacci, Maria Eichlseder, Kim G. Larsen

In this work, we study L*-based learning of deterministic Markov decision processes, first assuming an ideal setting with perfect information.

Active Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.