no code implementations • 2 Dec 2024 • Anton Voronov, Denis Kuznedelev, Mikhail Khoroshikh, Valentin Khrulkov, Dmitry Baranchuk
This work presents Switti, a scale-wise transformer for text-to-image generation.
no code implementations • 8 Apr 2024 • Sergey Kastryulin, Artem Konev, Alexander Shishenya, Eugene Lyapustin, Artem Khurshudov, Alexander Tselousov, Nikita Vinokurov, Denis Kuznedelev, Alexander Markovich, Grigoriy Livshits, Alexey Kirillov, Anastasiia Tabisheva, Liubov Chubarova, Marina Kaminskaia, Alexander Ustyuzhanin, Artemii Shvetsov, Daniil Shlenskii, Valerii Startsev, Dmitrii Kornilov, Mikhail Romanov, Artem Babenko, Sergei Ovcharenko, Valentin Khrulkov
In the rapidly progressing field of generative models, the development of efficient and high-fidelity text-to-image diffusion systems represents a significant frontier.
1 code implementation • 10 Apr 2023 • Nikita Starodubcev, Dmitry Baranchuk, Valentin Khrulkov, Artem Babenko
Finally, we show that our approach can adapt the pretrained model to the user-specified image and text description on the fly just for 4 seconds.
2 code implementations • CVPR 2022 • Aleksandr Ermolov, Leyla Mirvakhabova, Valentin Khrulkov, Nicu Sebe, Ivan Oseledets
Following this line of work, we propose a new hyperbolic-based model for metric learning.
Ranked #1 on Metric Learning on CUB-200-2011
no code implementations • 14 Feb 2022 • Valentin Khrulkov, Gleb Ryzhakov, Andrei Chertkov, Ivan Oseledets
Diffusion models have recently outperformed alternative approaches to model the distribution of natural images, such as GANs.
1 code implementation • ICLR 2022 • Dmitry Baranchuk, Ivan Rubachev, Andrey Voynov, Valentin Khrulkov, Artem Babenko
Denoising diffusion probabilistic models have recently received much research attention since they outperform alternative approaches, such as GANs, and currently provide state-of-the-art generative performance.
1 code implementation • ICCV 2021 • Valentin Khrulkov, Leyla Mirvakhabova, Ivan Oseledets, Artem Babenko
Recent advances in high-fidelity semantic image editing heavily rely on the presumably disentangled latent spaces of the state-of-the-art generative models, such as StyleGAN.
11 code implementations • NeurIPS 2021 • Yury Gorishniy, Ivan Rubachev, Valentin Khrulkov, Artem Babenko
The existing literature on deep learning for tabular data proposes a wide range of novel architectures and reports competitive results on various datasets.
1 code implementation • CVPR 2021 • Valentin Khrulkov, Artem Babenko
Given the dataset and the labels, we trained a CNN model that obtains a pair of images and for each image predicts a probability of being more preferable than its counterpart.
no code implementations • 11 Feb 2021 • Valentin Khrulkov, Leyla Mirvakhabova, Ivan Oseledets, Artem Babenko
Constructing disentangled representations is known to be a difficult task, especially in the unsupervised scenario.
no code implementations • 8 Feb 2021 • Valentin Khrulkov, Artem Babenko, Ivan Oseledets
Recent work demonstrated the benefits of studying continuous-time dynamics governing the GAN training.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Oleksii Hrinchuk, Valentin Khrulkov, Leyla Mirvakhabova, Elena Orlova, Ivan Oseledets
The embedding layers transforming input words into real vectors are the key components of deep neural networks used in natural language processing.
1 code implementation • 15 Aug 2020 • Leyla Mirvakhabova, Evgeny Frolov, Valentin Khrulkov, Ivan Oseledets, Alexander Tuzhilin
We introduce a simple autoencoder based on hyperbolic geometry for solving standard collaborative filtering problem.
2 code implementations • 29 Mar 2020 • Sergey Kolesnikov, Valentin Khrulkov
We present Catalyst. RL, an open-source PyTorch framework for reproducible and sample efficient reinforcement learning (RL) research.
no code implementations • 27 May 2019 • Valentin Khrulkov, Ivan Oseledets
Despite the fact that generative models are extremely successful in practice, the theory underlying this phenomenon is only starting to catch up with practice.
3 code implementations • CVPR 2020 • Valentin Khrulkov, Leyla Mirvakhabova, Evgeniya Ustinova, Ivan Oseledets, Victor Lempitsky
Computer vision tasks such as image classification, image retrieval and few-shot learning are currently dominated by Euclidean and spherical embeddings, so that the final decisions about class belongings or the degree of similarity are made using linear hyperplanes, Euclidean distances, or spherical geodesic distances (cosine similarity).
1 code implementation • 30 Jan 2019 • Oleksii Hrinchuk, Valentin Khrulkov, Leyla Mirvakhabova, Elena Orlova, Ivan Oseledets
The embedding layers transforming input words into real vectors are the key components of deep neural networks used in natural language processing.
no code implementations • ICLR 2019 • Valentin Khrulkov, Oleksii Hrinchuk, Ivan Oseledets
Such networks, however, are not very often applied to real life tasks.
1 code implementation • ICML 2018 • Valentin Khrulkov, Ivan Oseledets
One of the biggest challenges in the research of generative adversarial networks (GANs) is assessing the quality of generated samples and detecting various levels of mode collapse.
2 code implementations • 5 Jan 2018 • Alexander Novikov, Pavel Izmailov, Valentin Khrulkov, Michael Figurnov, Ivan Oseledets
Tensor Train decomposition is used across many branches of machine learning.
Mathematical Software Numerical Analysis
2 code implementations • ICLR 2018 • Valentin Khrulkov, Alexander Novikov, Ivan Oseledets
In this paper, we prove the expressive power theorem (an exponential lower bound on the width of the equivalent shallow network) for a class of recurrent neural networks -- ones that correspond to the Tensor Train (TT) decomposition.
no code implementations • CVPR 2018 • Valentin Khrulkov, Ivan Oseledets
Vulnerability of Deep Neural Networks (DNNs) to adversarial attacks has been attracting a lot of attention in recent studies.