Search Results for author: Isaac J. Sledge

Found 13 papers, 1 papers with code

Adapting the Exploration Rate for Value-of-Information-Based Reinforcement Learning

no code implementations20 Dec 2022 Isaac J. Sledge, Jose C. Principe

We then illustrate that these trends hold for deep, value-of-information-based agents that learn to play ten simple games and over forty more complicated games for the Nintendo GameBoy system.

reinforcement-learning Reinforcement Learning (RL)

Estimating Rényi's $α$-Cross-Entropies in a Matrix-Based Way

1 code implementation24 Sep 2021 Isaac J. Sledge, Jose C. Principe

This yields matrix-based estimators of R\'enyi's $\alpha$-cross-entropies.

An Information-Theoretic Approach for Automatically Determining the Number of States when Aggregating Markov Chains

no code implementations5 Jul 2021 Isaac J. Sledge, Jose C. Principe

A fundamental problem when aggregating Markov chains is the specification of the number of state groups.

Annotating Motion Primitives for Simplifying Action Search in Reinforcement Learning

no code implementations24 Feb 2021 Isaac J. Sledge, Darshan W. Bryner, Jose C. Principe

We have previously developed a means of constraining, and hence speeding up, the search process through the use of motion primitives; motion primitives are sequences of pre-specified actions taken across a state series.

reinforcement-learning Reinforcement Learning (RL)

Faster Convergence in Deep-Predictive-Coding Networks to Learn Deeper Representations

no code implementations18 Jan 2021 Isaac J. Sledge, Jose C. Principe

It yields unsupervised object recognition that surpass convolutional autoencoders and are on par with convolutional networks trained in a supervised manner.

Object Recognition

An Exact Reformulation of Feature-Vector-based Radial-Basis-Function Networks for Graph-based Observations

no code implementations22 Jan 2019 Isaac J. Sledge, Jose C. Principe

An important property of our radial basis function networks is that they are guaranteed to yield the same responses as conventional radial-basis networks trained on a corresponding vector realization of the relationships encoded by the adjacency-matrix.

Guided Policy Exploration for Markov Decision Processes using an Uncertainty-Based Value-of-Information Criterion

no code implementations5 Feb 2018 Isaac J. Sledge, Matthew S. Emigh, Jose C. Principe

The value of information yields a stochastic routine for choosing actions during learning that can explore the policy space in a coarse to fine manner.

Partitioning Relational Matrices of Similarities or Dissimilarities using the Value of Information

no code implementations28 Oct 2017 Isaac J. Sledge, Jose C. Principe

In this paper, we provide an approach to clustering relational matrices whose entries correspond to either similarities or dissimilarities between objects.

Clustering

An Analysis of the Value of Information when Exploring Stochastic, Discrete Multi-Armed Bandits

no code implementations8 Oct 2017 Isaac J. Sledge, Jose C. Principe

High amounts of policy information are associated with exploration-dominant searches of the space and yield high rewards.

Multi-Armed Bandits

Analysis of Agent Expertise in Ms. Pac-Man using Value-of-Information-based Policies

no code implementations28 Feb 2017 Isaac J. Sledge, Jose C. Principe

This cost function is the value of information, which provides the optimal trade-off between the expected return of a policy and the policy's complexity; policy complexity is measured by number of bits and controlled by a single hyperparameter on the cost function.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.