no code implementations • 10 Mar 2021 • Shuofeng Zhang, Isaac Reid, Guillermo Valle Pérez, Ard Louis
As an alternative to flatness measures, we use a function based picture and propose using the log of Bayesian prior upon initialization, $\log P(f)$, as a predictor of the generalization when a DNN converges on function $f$ after training to zero error.