Search Results for author: Riad Akrour

Found 11 papers, 2 papers with code

Interpretable Decision Tree Search as a Markov Decision Process

1 code implementation22 Sep 2023 Hector Kohler, Riad Akrour, Philippe Preux

Finding an optimal decision tree for a supervised learning task is a challenging combinatorial problem to solve at scale.

Optimal Interpretability-Performance Trade-off of Classification Trees with Black-Box Reinforcement Learning

no code implementations11 Apr 2023 Hector Kohler, Riad Akrour, Philippe Preux

A given supervised classification task is modeled as a Markov decision problem (MDP) and then augmented with additional actions that gather information about the features, equivalent to building a DT.

reinforcement-learning Reinforcement Learning (RL)

Convex Optimization with an Interpolation-based Projection and its Application to Deep Learning

no code implementations13 Nov 2020 Riad Akrour, Asma Atamna, Jan Peters

We then propose an optimization algorithm that follows the gradient of the composition of the objective and the projection and prove its convergence for linear objectives and arbitrary convex and Lipschitz domain defining inequality constraints.

Continuous Action Reinforcement Learning from a Mixture of Interpretable Experts

1 code implementation10 Jun 2020 Riad Akrour, Davide Tateo, Jan Peters

Reinforcement learning (RL) has demonstrated its ability to solve high dimensional tasks by leveraging non-linear function approximators.

reinforcement-learning Reinforcement Learning (RL)

An Upper Bound of the Bias of Nadaraya-Watson Kernel Regression under Lipschitz Assumptions

no code implementations29 Jan 2020 Samuele Tosatto, Riad Akrour, Jan Peters

The Nadaraya-Watson kernel estimator is among the most popular nonparameteric regression technique thanks to its simplicity.

regression valid

Local Bayesian Optimization of Motor Skills

no code implementations ICML 2017 Riad Akrour, Dmitry Sorokin, Jan Peters, Gerhard Neumann

Bayesian optimization is renowned for its sample efficiency but its application to higher dimensional tasks is impeded by its focus on global optimization.

Bayesian Optimization Imitation Learning

Model-Free Trajectory-based Policy Optimization with Monotonic Improvement

no code implementations29 Jun 2016 Riad Akrour, Abbas Abdolmaleki, Hany Abdulsamad, Jan Peters, Gerhard Neumann

In order to show the monotonic improvement of our algorithm, we additionally conduct a theoretical analysis of our policy update scheme to derive a lower bound of the change in policy return between successive iterations.

Cannot find the paper you are looking for? You can Submit a new open access paper.