no code implementations • 29 Feb 2024 • Alexander Shabalin, Viacheslav Meshchaninov, Tingir Badmaev, Dmitry Molchanov, Grigory Bartosh, Sergey Markov, Dmitry Vetrov
Drawing inspiration from the success of diffusion models in various domains, numerous research papers proposed methods for adapting them to text data.
1 code implementation • NeurIPS 2023 • Andrey Okhotin, Dmitry Molchanov, Vladimir Arkhipkin, Grigory Bartosh, Viktor Ohanesian, Aibek Alanov, Dmitry Vetrov
In the case of Gaussian distributions, SS-DDPM is equivalent to DDPM.
1 code implementation • 21 Feb 2020 • Dmitry Molchanov, Alexander Lyzhov, Yuliya Molchanova, Arsenii Ashukha, Dmitry Vetrov
Test-time data augmentation$-$averaging the predictions of a machine learning model across multiple augmented samples of data$-$is a widely used technique that improves the predictive performance.
2 code implementations • ICLR 2020 • Arsenii Ashukha, Alexander Lyzhov, Dmitry Molchanov, Dmitry Vetrov
Uncertainty estimation and ensembling methods go hand-in-hand.
no code implementations • pproximateinference AABI Symposium 2019 • Iuliia Molchanova, Dmitry Molchanov, Novi Quadrianto, Dmitry Vetrov
In this work we construct flexible joint distributions from low-dimensional conditional semi-implicit distributions.
1 code implementation • 1 Nov 2018 • Valery Kharitonov, Dmitry Molchanov, Dmitry Vetrov
We study the Automatic Relevance Determination procedure applied to deep neural networks.
no code implementations • 5 Oct 2018 • Dmitry Molchanov, Valery Kharitonov, Artem Sobolev, Dmitry Vetrov
Unlike discriminator-based and kernel-based approaches to implicit variational inference, DSIVI optimizes a proper lower bound on ELBO that is asymptotically exact.
2 code implementations • ICLR 2019 • Kirill Neklyudov, Dmitry Molchanov, Arsenii Ashukha, Dmitry Vetrov
Ordinary stochastic neural networks mostly rely on the expected values of their weights to make predictions, whereas the induced noise is mostly used to capture the uncertainty, prevent overfitting and slightly boost the performance through test-time averaging.
no code implementations • 20 Feb 2018 • Max Kochurov, Timur Garipov, Dmitry Podoprikhin, Dmitry Molchanov, Arsenii Ashukha, Dmitry Vetrov
In industrial machine learning pipelines, data often arrive in parts.
1 code implementation • 13 Feb 2018 • Andrei Atanov, Arsenii Ashukha, Dmitry Molchanov, Kirill Neklyudov, Dmitry Vetrov
In this work, we investigate Batch Normalization technique and propose its probabilistic interpretation.
5 code implementations • NeurIPS 2017 • Kirill Neklyudov, Dmitry Molchanov, Arsenii Ashukha, Dmitry Vetrov
In the paper, we propose a new Bayesian model that takes into account the computational structure of neural networks and provides structured sparsity, e. g. removes neurons and/or convolutional channels in CNNs.
15 code implementations • ICML 2017 • Dmitry Molchanov, Arsenii Ashukha, Dmitry Vetrov
We explore a recently proposed Variational Dropout technique that provided an elegant Bayesian interpretation to Gaussian Dropout.