Search Results for author: David Parker

Found 17 papers, 4 papers with code

HSVI-based Online Minimax Strategies for Partially Observable Stochastic Games with Neural Perception Mechanisms

no code implementations16 Apr 2024 Rui Yan, Gabriel Santos, Gethin Norman, David Parker, Marta Kwiatkowska

For the partially-informed agent, we propose a continual resolving approach which uses lower bounds, pre-computed offline with heuristic search value iteration (HSVI), instead of opponent counterfactual values.

counterfactual

Learning Algorithms for Verification of Markov Decision Processes

no code implementations14 Mar 2024 Tomáš Brázdil, Krishnendu Chatterjee, Martin Chmelik, Vojtěch Forejt, Jan Křetínský, Marta Kwiatkowska, Tobias Meggendorfer, David Parker, Mateusz Ujma

The presented framework focuses on probabilistic reachability, which is a core problem in verification, and is instantiated in two distinct scenarios.

Safe POMDP Online Planning via Shielding

no code implementations19 Sep 2023 Shili Sheng, David Parker, Lu Feng

POMDP online planning algorithms such as Partially Observable Monte-Carlo Planning (POMCP) can solve very large POMDPs with the goal of maximizing the expected return.

Autonomous Driving Decision Making +1

Using Reed-Muller Codes for Classification with Rejection and Recovery

1 code implementation12 Sep 2023 Daniel Fentham, David Parker, Mark Ryan

When deploying classifiers in the real world, users expect them to respond to inputs appropriately.

Multi-Agent Verification and Control with Probabilistic Model Checking

no code implementations5 Aug 2023 David Parker

Probabilistic model checking is a technique for formal automated reasoning about software or hardware systems that operate in the context of uncertainty or stochasticity.

Point-based Value Iteration for Neuro-Symbolic POMDPs

no code implementations30 Jun 2023 Rui Yan, Gabriel Santos, Gethin Norman, David Parker, Marta Kwiatkowska

This requires functions over continuous-state beliefs, for which we propose a novel piecewise linear and convex representation (P-PWLC) in terms of polyhedra covering the continuous-state space and value vectors, and extend Bellman backups to this representation.

Collision Avoidance Decision Making +1

Robust Control for Dynamical Systems With Non-Gaussian Noise via Formal Abstractions

1 code implementation4 Jan 2023 Thom Badings, Licio Romao, Alessandro Abate, David Parker, Hasan A. Poonawala, Marielle Stoelinga, Nils Jansen

This iMDP is, with a user-specified confidence probability, robust against uncertainty in the transition probabilities, and the tightness of the probability intervals can be controlled through the number of samples.

Continuous Control

Robust Anytime Learning of Markov Decision Processes

1 code implementation31 May 2022 Marnix Suilen, Thiago D. Simão, David Parker, Nils Jansen

Markov decision processes (MDPs) are formal models commonly used in sequential decision-making.

Bayesian Inference Decision Making

Strategy Synthesis for Zero-Sum Neuro-Symbolic Concurrent Stochastic Games

no code implementations13 Feb 2022 Rui Yan, Gabriel Santos, Gethin Norman, David Parker, Marta Kwiatkowska

Second, we introduce two novel representations for the value functions and strategies, constant-piecewise-linear (CON-PWL) and constant-piecewise-constant (CON-PWC) respectively, and propose Minimax-action-free PI by extending a recent PI method based on alternating player choices for finite state spaces to Borel state spaces, which does not require normal-form games to be solved.

Verified Probabilistic Policies for Deep Reinforcement Learning

1 code implementation10 Jan 2022 Edoardo Bacci, David Parker

Deep reinforcement learning is an increasingly popular technique for synthesising policies to control an agent's interaction with its environment.

reinforcement-learning Reinforcement Learning (RL)

Sampling-Based Robust Control of Autonomous Systems with Non-Gaussian Noise

no code implementations25 Oct 2021 Thom S. Badings, Alessandro Abate, Nils Jansen, David Parker, Hasan A. Poonawala, Marielle Stoelinga

We use state-of-the-art verification techniques to provide guarantees on the iMDP, and compute a controller for which these guarantees carry over to the autonomous system.

Multi-Objective Controller Synthesis with Uncertain Human Preferences

no code implementations10 May 2021 Shenghui Chen, Kayla Boggess, David Parker, Lu Feng

Complex real-world applications of cyber-physical systems give rise to the need for multi-objective controller synthesis, which concerns the problem of computing an optimal controller subject to multiple (possibly conflicting) criteria.

Canted antiferromagnetic order and spin dynamics in the honeycomb-lattice Tb2Ir3Ga9

no code implementations4 Mar 2021 Feng Ye, Zachary Morgan, Wei Tian, Songxue Chi, XiaoPing Wang, Michael E. Manley, David Parker, Mojammel A. Khan, J. F. Mitchell, Randy Fishman

Despite the $J_{{\rm eff}}=1/2$ moments, the spin Hamiltonian is denominated by a large in-plane anisotropy $K_z \sim -1$ meV.

Strongly Correlated Electrons Materials Science

Probabilistic Guarantees for Safe Deep Reinforcement Learning

no code implementations14 May 2020 Edoardo Bacci, David Parker

Deep reinforcement learning has been successfully applied to many control tasks, but the application of such agents in safety-critical scenarios has been limited due to safety concerns.

reinforcement-learning Reinforcement Learning (RL)

Simultaneous Task Allocation and Planning Under Uncertainty

no code implementations7 Mar 2018 Fatma Faruq, Bruno Lacerda, Nick Hawes, David Parker

We propose novel techniques for task allocation and planning in multi-robot systems operating in uncertain environments.

Verification and Control of Partially Observable Probabilistic Real-Time Systems

no code implementations21 Jun 2015 Gethin Norman, David Parker, Xueyi Zou

We then propose techniques to either verify that such a property holds or to synthesise a controller for the model which makes it true.

Logic in Computer Science Systems and Control

Cannot find the paper you are looking for? You can Submit a new open access paper.