1 code implementation • 28 Dec 2023 • Tim G. J. Rudner, Sanyam Kapoor, Shikai Qiu, Andrew Gordon Wilson
In this work, we approach regularization in neural networks from a probabilistic perspective and show that by viewing parameter-space regularization as specifying an empirical prior distribution over the model parameters, we can derive a probabilistically well-motivated regularization technique that allows explicitly encoding information about desired predictive functions into neural network training.
1 code implementation • NeurIPS 2023 • Shikai Qiu, Tim G. J. Rudner, Sanyam Kapoor, Andrew Gordon Wilson
Moreover, the most likely parameters under the parameter posterior do not generally correspond to the most likely function induced by the parameter posterior.
1 code implementation • NeurIPS 2023 • Nate Gruver, Marc Finzi, Shikai Qiu, Andrew Gordon Wilson
By encoding time series as a string of numerical digits, we can frame time series forecasting as next-token prediction in text.
1 code implementation • 19 Jun 2023 • Shikai Qiu, Andres Potapczynski, Pavel Izmailov, Andrew Gordon Wilson
A major challenge to out-of-distribution generalization is reliance on spurious features -- patterns that are predictive of the class label in the training data distribution, but not causally related to the target.