Search Results for author: Nils Jansen

Found 40 papers, 14 papers with code

A Stability-Based Abstraction Framework for Reach-Avoid Control of Stochastic Dynamical Systems with Unknown Noise Distributions

no code implementations2 Apr 2024 Thom Badings, Licio Romao, Alessandro Abate, Nils Jansen

To address this issue, we propose a novel abstraction scheme for stochastic linear systems that exploits the system's stability to obtain significantly smaller abstract models.

Robust Active Measuring under Model Uncertainty

1 code implementation18 Dec 2023 Merlijn Krale, Thiago D. Simão, Jana Tumova, Nils Jansen

Partial observability and uncertainty are common problems in sequential decision-making that particularly impede the use of formal models such as Markov decision processes (MDPs).

Decision Making

Factored Online Planning in Many-Agent POMDPs

no code implementations18 Dec 2023 Maris F. L. Galesloot, Thiago D. Simão, Sebastian Junges, Nils Jansen

However, the challenges of value estimation and belief estimation have only been tackled individually, which prevents existing methods from scaling to settings with many agents.

Reinforcement Learning by Guided Safe Exploration

no code implementations26 Jul 2023 Qisong Yang, Thiago D. Simão, Nils Jansen, Simon H. Tindemans, Matthijs T. J. Spaan

Drawing from transfer learning, we also regularize a target policy (the student) towards the guide while the student is unreliable and gradually eliminate the influence of the guide as training progresses.

reinforcement-learning Reinforcement Learning (RL) +2

More for Less: Safe Policy Improvement With Stronger Performance Guarantees

1 code implementation13 May 2023 Patrick Wienhöft, Marnix Suilen, Thiago D. Simão, Clemens Dubslaff, Christel Baier, Nils Jansen

In an offline reinforcement learning setting, the safe policy improvement (SPI) problem aims to improve the performance of a behavior policy according to which sample data has been generated.

Efficient Sensitivity Analysis for Parametric Robust Markov Chains

no code implementations1 May 2023 Thom Badings, Sebastian Junges, Ahmadreza Marandi, Ufuk Topcu, Nils Jansen

As our main contribution, we present an efficient method to compute these partial derivatives.

Act-Then-Measure: Reinforcement Learning for Partially Observable Environments with Active Measuring

1 code implementation14 Mar 2023 Merlijn Krale, Thiago D. Simão, Nils Jansen

In these models, actions consist of two components: a control action that affects the environment, and a measurement action that affects what the agent can observe.

reinforcement-learning Reinforcement Learning (RL)

Decision-Making Under Uncertainty: Beyond Probabilities

no code implementations10 Mar 2023 Thom Badings, Thiago D. Simão, Marnix Suilen, Nils Jansen

In this paper, the focus is on the uncertainty that goes beyond this classical interpretation, particularly by employing a clear distinction between aleatoric and epistemic uncertainty.

Decision Making Decision Making Under Uncertainty

Safe Policy Improvement for POMDPs via Finite-State Controllers

no code implementations12 Jan 2023 Thiago D. Simão, Marnix Suilen, Nils Jansen

In our novel approach to the SPI problem for POMDPs, we assume that a finite-state controller (FSC) represents the behavior policy and that finite memory is sufficient to derive optimal policies.

Reinforcement Learning (RL)

Robust Control for Dynamical Systems With Non-Gaussian Noise via Formal Abstractions

1 code implementation4 Jan 2023 Thom Badings, Licio Romao, Alessandro Abate, David Parker, Hasan A. Poonawala, Marielle Stoelinga, Nils Jansen

This iMDP is, with a user-specified confidence probability, robust against uncertainty in the transition probabilities, and the tightness of the probability intervals can be controlled through the number of samples.

Continuous Control

Probabilities Are Not Enough: Formal Controller Synthesis for Stochastic Dynamical Models with Epistemic Uncertainty

1 code implementation12 Oct 2022 Thom Badings, Licio Romao, Alessandro Abate, Nils Jansen

Stochastic noise causes aleatoric uncertainty, whereas imprecise knowledge of model parameters leads to epistemic uncertainty.

COOL-MC: A Comprehensive Tool for Reinforcement Learning and Model Checking

3 code implementations15 Sep 2022 Dennis Gross, Nils Jansen, Sebastian Junges, Guillermo A. Perez

This paper presents COOL-MC, a tool that integrates state-of-the-art reinforcement learning (RL) and model checking.

OpenAI Gym reinforcement-learning +1

Robust Anytime Learning of Markov Decision Processes

1 code implementation31 May 2022 Marnix Suilen, Thiago D. Simão, David Parker, Nils Jansen

Markov decision processes (MDPs) are formal models commonly used in sequential decision-making.

Bayesian Inference Decision Making

Safe Reinforcement Learning via Shielding under Partial Observability

no code implementations2 Apr 2022 Steven Carr, Nils Jansen, Sebastian Junges, Ufuk Topcu

Safe exploration is a common problem in reinforcement learning (RL) that aims to prevent agents from making disastrous decisions while exploring their environment.

reinforcement-learning Reinforcement Learning (RL) +2

Sampling-Based Robust Control of Autonomous Systems with Non-Gaussian Noise

no code implementations25 Oct 2021 Thom S. Badings, Alessandro Abate, Nils Jansen, David Parker, Hasan A. Poonawala, Marielle Stoelinga

We use state-of-the-art verification techniques to provide guarantees on the iMDP, and compute a controller for which these guarantees carry over to the autonomous system.

Convex Optimization for Parameter Synthesis in MDPs

1 code implementation30 Jun 2021 Murat Cubuktepe, Nils Jansen, Sebastian Junges, Joost-Pieter Katoen, Ufuk Topcu

The parameter synthesis problem is to compute an instantiation of these unspecified parameters such that the resulting MDP satisfies the temporal logic specification.

Collision Avoidance

Correct-by-construction reach-avoid control of partially observable linear stochastic systems

1 code implementation3 Mar 2021 Thom Badings, Hasan A. Poonawala, Marielle Stoelinga, Nils Jansen

By construction, any policy on the abstraction can be refined into a piecewise linear feedback controller for the LTI system.

Continuous Control

Damage detection using in-domain and cross-domain transfer learning

no code implementations7 Feb 2021 Zaharah A. Bukhsh, Nils Jansen, Aaqib Saeed

We, therefore, evaluate a combination of in-domain and cross-domain transfer learning strategies for damage detection in bridges.

Transfer Learning

Robust Finite-State Controllers for Uncertain POMDPs

no code implementations24 Sep 2020 Murat Cubuktepe, Nils Jansen, Sebastian Junges, Ahmadreza Marandi, Marnix Suilen, Ufuk Topcu

(3) We linearize this dual problem and (4) solve the resulting finite linear program to obtain locally optimal solutions to the original problem.

Collision Avoidance Motion Planning

Strengthening Deterministic Policies for POMDPs

no code implementations16 Jul 2020 Leonore Winterer, Ralf Wimmer, Nils Jansen, Bernd Becker

Second, based on the results of the original MILP, we employ a preprocessing of the POMDP to encompass memory-based decisions.

Enforcing Almost-Sure Reachability in POMDPs

1 code implementation30 Jun 2020 Sebastian Junges, Nils Jansen, Sanjit A. Seshia

Partially-Observable Markov Decision Processes (POMDPs) are a well-known stochastic model for sequential decision making under limited information.

Decision Making reinforcement-learning +2

Robustness Verification for Classifier Ensembles

no code implementations12 May 2020 Dennis Gross, Nils Jansen, Guillermo A. Pérez, Stephan Raaijmakers

The robustness-checking problem consists of assessing, given a set of classifiers and a labelled data set, whether there exists a randomized attack that induces a certain expected loss against all classifiers.

Image Classification

Verifiable RNN-Based Policies for POMDPs Under Temporal Logic Constraints

no code implementations13 Feb 2020 Steven Carr, Nils Jansen, Ufuk Topcu

Recurrent neural networks (RNNs) have emerged as an effective representation of control policies in sequential decision-making problems.

Decision Making

Neural Simplex Architecture

no code implementations1 Aug 2019 Dung T. Phan, Radu Grosu, Nils Jansen, Nicola Paoletti, Scott A. Smolka, Scott D. Stoller

NSA not only provides safety assurances in the presence of a possibly unsafe neural controller, but can also improve the safety of such a controller in an online setting via retraining, without overly degrading its performance.

Continuous Control

Synthesis of Provably Correct Autonomy Protocols for Shared Control

no code implementations15 May 2019 Murat Cubuktepe, Nils Jansen, Mohammed Alsiekh, Ufuk Topcu

We design the autonomy protocol to ensure that the resulting robot behavior satisfies given safety and performance specifications in probabilistic temporal logic.

Counterexample-Guided Strategy Improvement for POMDPs Using Recurrent Neural Networks

no code implementations20 Mar 2019 Steven Carr, Nils Jansen, Ralf Wimmer, Alexandru C. Serban, Bernd Becker, Ufuk Topcu

The particular problem is to determine strategies that provably adhere to (probabilistic) temporal logic constraints.

Shepherding Hordes of Markov Chains

1 code implementation15 Feb 2019 Milan Ceska, Nils Jansen, Sebastian Junges, Joost-Pieter Katoen

This paper considers large families of Markov chains (MCs) that are defined over a set of parameters with finite discrete domains.

The Partially Observable Games We Play for Cyber Deception

no code implementations28 Sep 2018 Mohamadreza Ahmadi, Murat Cubuktepe, Nils Jansen, Sebastian Junges, Joost-Pieter Katoen, Ufuk Topcu

Then, the deception problem is to compute a strategy for the deceiver that minimizes the expected cost of deception against all strategies of the infiltrator.

Safe Reinforcement Learning via Probabilistic Shields

no code implementations16 Jul 2018 Nils Jansen, Bettina Könighofer, Sebastian Junges, Alexandru C. Serban, Roderick Bloem

This paper targets the efficient construction of a safety shield for decision making in scenarios that incorporate uncertainty.

Decision Making reinforcement-learning +3

Synthesis in pMDPs: A Tale of 1001 Parameters

no code implementations5 Mar 2018 Murat Cubuktepe, Nils Jansen, Sebastian Junges, Joost-Pieter Katoen, Ufuk Topcu

This paper considers parametric Markov decision processes (pMDPs) whose transitions are equipped with affine functions over a finite set of parameters.

Human-in-the-Loop Synthesis for Partially Observable Markov Decision Processes

no code implementations27 Feb 2018 Steven Carr, Nils Jansen, Ralf Wimmer, Jie Fu, Ufuk Topcu

The efficient verification of this MC gives quantitative insights into the quality of the inferred human strategy by proving or disproving given system specifications.

Strategy Synthesis in POMDPs via Game-Based Abstractions

no code implementations14 Aug 2017 Leonore Winterer, Sebastian Junges, Ralf Wimmer, Nils Jansen, Ufuk Topcu, Joost-Pieter Katoen, Bernd Becker

We study synthesis problems with constraints in partially observable Markov decision processes (POMDPs), where the objective is to compute a strategy for an agent that is guaranteed to satisfy certain safety and performance specifications.

Motion Planning

Synthesis of Shared Control Protocols with Provable Safety and Performance Guarantees

no code implementations26 Oct 2016 Nils Jansen, Murat Cubuktepe, Ufuk Topcu

We formalize synthesis of shared control protocols with correctness guarantees for temporal logic specifications.

Cannot find the paper you are looking for? You can Submit a new open access paper.