no code implementations • 19 Nov 2023 • Ekaterina Lobacheva, Eduard Pockonechnyy, Maxim Kodryan, Dmitry Vetrov
Inspired by recent research that recommends starting neural networks training with large learning rates (LRs) to achieve the best generalization, we explore this hypothesis in detail.
1 code implementation • 8 Sep 2022 • Maxim Kodryan, Ekaterina Lobacheva, Maksim Nakhodnov, Dmitry Vetrov
In this work, we investigate the properties of training scale-invariant neural networks directly on the sphere using a fixed ELR.
1 code implementation • NeurIPS 2021 • Ekaterina Lobacheva, Maxim Kodryan, Nadezhda Chirkova, Andrey Malinin, Dmitry Vetrov
Training neural networks with batch normalization and weight decay has become a common practice in recent years.
1 code implementation • NeurIPS 2020 • Ekaterina Lobacheva, Nadezhda Chirkova, Maxim Kodryan, Dmitry Vetrov
Ensembles of deep neural networks are known to achieve state-of-the-art performance in uncertainty estimation and lead to accuracy improvement.
1 code implementation • 18 Jun 2020 • Maxim Kodryan, Dmitry Kropotov, Dmitry Vetrov
Tensor decomposition methods have proven effective in various applications, including compression and acceleration of neural networks.
no code implementations • WS 2019 • Maxim Kodryan, Artem Grachev, Dmitry Ignatov, Dmitry Vetrov
Reduction of the number of parameters is one of the most important goals in Deep Learning.