no code implementations • NeurIPS 2012 • Tim V. Erven, Peter Grünwald, Mark D. Reid, Robert C. Williamson
We show that, in the special case of log-loss, stochastic mixability reduces to a well-known (but usually unnamed) martingale condition, which is used in existing convergence theorems for minimum description length and Bayesian inference.
no code implementations • NeurIPS 2011 • Tim V. Erven, Wouter M. Koolen, Steven D. Rooij, Peter Grünwald
In most previous analyses the learning rate was carefully tuned to obtain optimal worst-case performance, leading to suboptimal performance on easy instances, for example when there exists an action that is significantly better than all others.