no code implementations • NeurIPS 2020 • Marine Le Morvan, Julie Josses, Thomas Moreau, Erwan Scornet, Gael Varoquaux
We provide an upper bound on the Bayes risk of NeuMiss networks, and show that they have good predictive accuracy with both a number of parameters and a computational complexity independent of the number of missing data patterns.
no code implementations • NeurIPS 2020 • Aude Sportisse, Claire Boyer, Aymeric Dieuleveut, Julie Josses
Stochastic gradient algorithm is a key ingredient of many machine learning methods, particularly appropriate for large-scale learning.