Search Results for author: Jonathan Baxter

Found 12 papers, 0 papers with code

Theoretical Models of Learning to Learn

no code implementations27 Feb 2020 Jonathan Baxter

A Machine can only learn if it is biased in some way.

Some observations concerning Off Training Set (OTS) error

no code implementations18 Nov 2019 Jonathan Baxter

A form of generalisation error known as Off Training Set (OTS) error was recently introduced in [Wolpert, 1996b], along with a theorem showing that small training set error does not guarantee small OTS error, unless assumptions are made about the target function.

BIG-bench Machine Learning

General Matrix-Matrix Multiplication Using SIMD features of the PIII

no code implementations18 Nov 2019 Douglas Aberdeen, Jonathan Baxter

Generalised matrix-matrix multiplication forms the kernel of many mathematical algorithms.

Hebbian Synaptic Modifications in Spiking Neurons that Learn

no code implementations17 Nov 2019 Peter L. Bartlett, Jonathan Baxter

In this paper, we derive a new model of synaptic plasticity, based on recent algorithms for reinforcement learning (in which an agent attempts to learn appropriate actions to maximize its long-term average reward).

reinforcement-learning Reinforcement Learning (RL)

Learning Model Bias

no code implementations14 Nov 2019 Jonathan Baxter

In this paper the problem of {\em learning} appropriate domain-specific bias is addressed.

A Bayesian/Information Theoretic Model of Bias Learning

no code implementations14 Nov 2019 Jonathan Baxter

In this paper the problem of learning appropriate bias for an environment of related tasks is examined from a Bayesian perspective.

Bayesian Inference

The Canonical Distortion Measure for Vector Quantization and Function Approximation

no code implementations14 Nov 2019 Jonathan Baxter

In this paper it is shown how an {\em environment} of functions on an input space $X$ induces a {\em canonical distortion measure} (CDM) on X.

Quantization

Learning Internal Representations (COLT 1995)

no code implementations13 Nov 2019 Jonathan Baxter

It is proved that the number of examples $m$ {\em per task} required to ensure good generalisation from a representation learner obeys $m = O(a+b/n)$ where $n$ is the number of tasks being learnt and $a$ and $b$ are constants.

Representation Learning

Learning Internal Representations (PhD Thesis)

no code implementations9 Nov 2019 Jonathan Baxter

In this thesis it is argued that in general there is insufficient information in a single task for a learner to generalise well and that what is required for good generalisation is information about many similar learning tasks.

Learning Theory Representation Learning

Infinite-Horizon Policy-Gradient Estimation

no code implementations3 Jun 2011 Jonathan Baxter, Peter L. Bartlett

In this paper we introduce GPOMDP, a simulation-based algorithm for generating a {\em biased} estimate of the gradient of the {\em average reward} in Partially Observable Markov Decision Processes (POMDPs) controlled by parameterized stochastic policies.

KnightCap: A chess program that learns by combining TD(lambda) with game-tree search

no code implementations10 Jan 1999 Jonathan Baxter, Andrew Tridgell, Lex Weaver

In this paper we present TDLeaf(lambda), a variation on the TD(lambda) algorithm that enables it to be used in conjunction with game-tree search.

Cannot find the paper you are looking for? You can Submit a new open access paper.