Search Results for author: Tom Bewley

Found 9 papers, 3 papers with code

Zero-Shot Reinforcement Learning from Low Quality Data

1 code implementation26 Sep 2023 Scott Jeen, Tom Bewley, Jonathan M. Cullen

Zero-shot reinforcement learning (RL) promises to provide agents that can perform any task in an environment after an offline, reward-free pre-training phase.

Offline RL reinforcement-learning +1

Learning Interpretable Models of Aircraft Handling Behaviour by Reinforcement Learning from Human Feedback

no code implementations26 May 2023 Tom Bewley, Jonathan Lawry, Arthur Richards

We propose a method to capture the handling abilities of fast jet pilots in a software model via reinforcement learning (RL) from human preference feedback.

Reinforcement Learning (RL)

Reward Learning with Trees: Methods and Evaluation

no code implementations3 Oct 2022 Tom Bewley, Jonathan Lawry, Arthur Richards, Rachel Craddock, Ian Henderson

Recent efforts to learn reward functions from human feedback have tended to use deep neural networks, whose lack of transparency hampers our ability to explain agent behaviour or verify alignment.

Summarising and Comparing Agent Dynamics with Contrastive Spatiotemporal Abstraction

no code implementations17 Jan 2022 Tom Bewley, Jonathan Lawry, Arthur Richards

We introduce a data-driven, model-agnostic technique for generating a human-interpretable summary of the salient points of contrast within an evolving dynamical system, such as the learning process of a control agent.

reinforcement-learning Reinforcement Learning (RL)

Interpretable Preference-based Reinforcement Learning with Tree-Structured Reward Functions

no code implementations20 Dec 2021 Tom Bewley, Freddy Lecue

The potential of reinforcement learning (RL) to deliver aligned and performant agents is partially bottlenecked by the reward engineering problem.

reinforcement-learning Reinforcement Learning (RL)

TripleTree: A Versatile Interpretable Representation of Black Box Agents and their Environments

1 code implementation10 Sep 2020 Tom Bewley, Jonathan Lawry

In explainable artificial intelligence, there is increasing interest in understanding the behaviour of autonomous agents to build trust and validate performance.

Explainable artificial intelligence reinforcement-learning +1

Am I Building a White Box Agent or Interpreting a Black Box Agent?

no code implementations2 Jul 2020 Tom Bewley

The rule extraction literature contains the notion of a fidelity-accuracy dilemma: when building an interpretable model of a black box function, optimising for fidelity is likely to reduce performance on the underlying task, and vice versa.

Explainable artificial intelligence

Modelling Agent Policies with Interpretable Imitation Learning

no code implementations19 Jun 2020 Tom Bewley, Jonathan Lawry, Arthur Richards

As we deploy autonomous agents in safety-critical domains, it becomes important to develop an understanding of their internal mechanisms and representations.

Imitation Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.