General Reinforcement Learning

35 papers with code • 6 benchmarks • 7 datasets

This task has no description! Would you like to contribute one?

Libraries

Use these libraries to find General Reinforcement Learning models and implementations

Latest papers with no code

Dynamic Knowledge Injection for AIXI Agents

no code yet • 18 Dec 2023

Prior approximations of AIXI, a Bayesian optimality notion for general reinforcement learning, can only approximate AIXI's Bayesian environment model using an a-priori defined set of models.

Dropout Strategy in Reinforcement Learning: Limiting the Surrogate Objective Variance in Policy Optimization Methods

no code yet • 31 Oct 2023

Then, we introduced a general reinforcement learning framework applicable to mainstream policy optimization methods, and applied the dropout technique to the PPO algorithm to obtain the D-PPO variant.

Image Transformation Sequence Retrieval with General Reinforcement Learning

no code yet • 13 Jul 2023

In this work, the novel Image Transformation Sequence Retrieval (ITSR) task is presented, in which a model must retrieve the sequence of transformations between two given images that act as source and target, respectively.

L-SA: Learning Under-Explored Targets in Multi-Target Reinforcement Learning

no code yet • 23 May 2023

Tasks that involve interaction with various targets are called multi-target tasks.

Computably Continuous Reinforcement-Learning Objectives are PAC-learnable

no code yet • 9 Mar 2023

In particular, for the analysis that considers only sample complexity, we prove that if an objective given as an oracle is uniformly continuous, then it is PAC-learnable.

Policy Mirror Descent Inherently Explores Action Space

no code yet • 8 Mar 2023

SPMD with the second evaluation operator, namely truncated on-policy Monte Carlo (TOMC), attains an $\tilde{\mathcal{O}}(\mathcal{H}_{\mathcal{D}}/\epsilon^2)$ sample complexity, where $\mathcal{H}_{\mathcal{D}}$ mildly depends on the effective horizon and the size of the action space with properly chosen Bregman divergence (e. g., Tsallis divergence).

Computational Dualism and Objective Superintelligence

no code yet • 2 Feb 2023

This undermines claims regarding the behaviour of theorised, software superintelligence.

Accuracy-Guaranteed Collaborative DNN Inference in Industrial IoT via Deep Reinforcement Learning

no code yet • 31 Dec 2022

Specifically, sampling rate adaption, inference task offloading and edge computing resource allocation are jointly considered to minimize the average service delay while guaranteeing the long-term accuracy requirements of different inference services.

AcceRL: Policy Acceleration Framework for Deep Reinforcement Learning

no code yet • 28 Nov 2022

Inspired by the redundancy of neural networks, we propose a lightweight parallel training framework based on neural network compression, AcceRL, to accelerate the policy learning while ensuring policy quality.

Computable Artificial General Intelligence

no code yet • 21 May 2022

AIXI appears to be the only such formalism supported by proof that its behaviour is optimal, a consequence of its use of compression as a proxy for intelligence.