no code implementations • 25 Nov 2024 • Egor Sevriugov, Ivan Oseledets
Non-autoregressive language models are emerging as effective alternatives to autoregressive models in the field of natural language processing, facilitating simultaneous token generation.
no code implementations • 23 Oct 2024 • Artem Basharin, Andrei Chertkov, Ivan Oseledets
We propose a new model for multi-token prediction in transformers, aiming to enhance sampling efficiency without compromising accuracy.
no code implementations • 23 Oct 2024 • Alexey Dontsov, Dmitrii Korzh, Alexey Zhavoronkin, Boris Mikheev, Denis Bobkov, Aibek Alanov, Oleg Y. Rogov, Ivan Oseledets, Elena Tutubalina
To address this, we introduce CLEAR, a new benchmark designed to evaluate MMU methods.
1 code implementation • 9 Oct 2024 • Viktoriia Chekalina, Anna Rudenko, Gleb Mezentsev, Alexander Mikhalev, Alexander Panchenko, Ivan Oseledets
The performance of Transformer models has been enhanced by increasing the number of parameters and the length of the processed text.
no code implementations • 6 Oct 2024 • Georgii Novikov, Alexander Gneushev, Alexey Kadeishvili, Ivan Oseledets
Nearest-neighbor search in large vector databases is crucial for various machine learning applications.
1 code implementation • 5 Oct 2024 • Tianchi Yu, Jingwei Qiu, Jiang Yang, Ivan Oseledets
In this paper, we propose to use Sinc interpolation in the context of Kolmogorov-Arnold Networks, neural networks with learnable activation functions, which recently gained attention as alternatives to multilayer perceptron.
no code implementations • 2 Oct 2024 • Vladimir Fanaskov, Ivan Oseledets
We perform a direct analysis of the dynamical system and show how to resolve problems caused by flat directions corresponding to dead neurons: (i) all information about the state vector at a fixed point can be extracted from the energy and Hessian matrix (of Lagrange function), (ii) it is enough to analyze stability in the range of Hessian matrix, (iii) if steady state touching flat region is stable the whole flat region is the basin of attraction.
1 code implementation • 27 Sep 2024 • Gleb Mezentsev, Danil Gusak, Ivan Oseledets, Evgeny Frolov
It approximates the CE loss for datasets with large-size catalogs, enhancing both time efficiency and memory usage without compromising recommendations quality.
Ranked #1 on Sequential Recommendation on Amazon Beauty
1 code implementation • 16 Sep 2024 • Mikhail Goncharov, Valentin Samokhin, Eugenia Soboleva, Roman Sokolov, Boris Shirokikh, Mikhail Belyaev, Anvar Kurmukov, Ivan Oseledets
We train our APE model on 8400 publicly available CT images of abdomen and chest regions.
no code implementations • 29 Aug 2024 • Tianchi Yu, Yiming Qi, Ivan Oseledets, Shiyi Chen
With growing investigations into solving partial differential equations by physics-informed neural networks (PINNs), more accurate and efficient PINNs are required to meet the practical demands of scientific computing.
1 code implementation • 5 Aug 2024 • Danil Gusak, Gleb Mezentsev, Ivan Oseledets, Evgeny Frolov
Scalability is a major challenge in modern recommender systems.
1 code implementation • 22 Jul 2024 • Georgii Novikov, Ivan Oseledets
A significant challenge in neural network training is the memory footprint associated with activation tensors, particularly in pointwise nonlinearity layers that traditionally save the entire input tensor for the backward pass, leading to substantial memory consumption.
1 code implementation • 7 Jun 2024 • Vladislav Trifonov, Alexander Rudikov, Oleg Iliev, Ivan Oseledets, Ekaterina Muravleva
We present ConDiff, a novel dataset for scientific machine learning.
no code implementations • 4 Jun 2024 • Vladimir Fanaskov, Tianchi Yu, Alexander Rudikov, Ivan Oseledets
The primal approach to physics-informed learning is a residual minimization.
no code implementations • 24 May 2024 • Vladislav Trifonov, Alexander Rudikov, Oleg Iliev, Ivan Oseledets, Ekaterina Muravleva
In our work, we recall well-established preconditioners from linear algebra and use them as a starting point for training the GNN.
1 code implementation • 19 May 2024 • Anton Razzhigaev, Matvey Mikhalchuk, Elizaveta Goncharova, Nikolai Gerasimenko, Ivan Oseledets, Denis Dimitrov, Andrey Kuznetsov
This regularization improves performance metrics on benchmarks like Tiny Stories and SuperGLUE and as well successfully decreases the linearity of the models.
1 code implementation • 13 May 2024 • Andrey V. Galichin, Mikhail Pautov, Alexey Zhavoronkin, Oleg Y. Rogov, Ivan Oseledets
We observe that the knowledge distillation significantly improves the efficiency of likelihood ratio of membership inference attack, especially in the black-box setting, i. e., when the architecture of the target model is unknown to the attacker.
no code implementations • 29 Apr 2024 • Dmitrii Korzh, Elvir Karimov, Mikhail Pautov, Oleg Y. Rogov, Ivan Oseledets
In this paper, we pioneer applying robustness certification techniques to speaker recognition, originally developed for the image domain.
no code implementations • 15 Apr 2024 • Daniil Merkulov, Daria Cherniuk, Alexander Rudikov, Ivan Oseledets, Ekaterina Muravleva, Aleksandr Mikhalev, Boris Kashin
In this paper, we introduce an algorithm for data quantization based on the principles of Kashin representation.
no code implementations • 9 Apr 2024 • Elizaveta Goncharova, Anton Razzhigaev, Matvey Mikhalchuk, Maxim Kurkin, Irina Abdullaeva, Matvey Skripkin, Ivan Oseledets, Denis Dimitrov, Andrey Kuznetsov
We propose an \textit{OmniFusion} model based on a pretrained LLM and adapters for visual modality.
Ranked #88 on Visual Question Answering on MM-Vet
no code implementations • 5 Feb 2024 • Gleb Ryzhakov, Andrei Chertkov, Artem Basharin, Ivan Oseledets
We develop a new method HTBB for the multidimensional black-box approximation and gradient-free optimization, which is based on the low-rank hierarchical Tucker decomposition with the use of the MaxVol indices selection procedure.
no code implementations • 5 Feb 2024 • Gleb Ryzhakov, Svetlana Pavlova, Egor Sevriugov, Ivan Oseledets
In addition, we also investigated simple cases of diffusion generative models by adding a stochastic term and obtained an explicit form of the expression for score.
no code implementations • 2 Feb 2024 • Daniel Bershatsky, Daria Cherniuk, Talgat Daulbaev, Aleksandr Mikhalev, Ivan Oseledets
In this paper we generalize and extend an idea of low-rank adaptation (LoRA) of large language models (LLMs) based on Transformer architecture.
no code implementations • 25 Jan 2024 • Kseniia Kuvshinova, Olga Tsymboi, Ivan Oseledets
The research in the field of adversarial attacks and models' vulnerability is one of the fundamental directions in modern machine learning.
no code implementations • 16 Jan 2024 • Mikhail Pautov, Nikita Bogdanov, Stanislav Pyatkin, Oleg Rogov, Ivan Oseledets
As deep learning (DL) models are widely and effectively used in Machine Learning as a Service (MLaaS) platforms, there is a rapidly growing interest in DL watermarking techniques that can be used to confirm the ownership of a particular model.
1 code implementation • 28 Dec 2023 • Nikita Pospelov, Andrei Chertkov, Maxim Beketov, Ivan Oseledets, Konstantin Anokhin
For a living system, such as a neuron, whose response to a stimulus is unknown and not differentiable, the only way to reveal these features is through a feedback loop that exposes it to a large set of different stimuli.
no code implementations • 6 Dec 2023 • Daria Cherniuk, Aleksandr Mikhalev, Ivan Oseledets
LoRA is a technique that reduces the number of trainable parameters in a neural network by introducing low-rank adapters to linear layers.
1 code implementation • 4 Dec 2023 • Albert Saiapin, Ivan Oseledets, Evgeny Frolov
In production applications of recommender systems, a continuous data flow is employed to update models in real-time.
no code implementations • 10 Nov 2023 • Anton Razzhigaev, Matvey Mikhalchuk, Elizaveta Goncharova, Ivan Oseledets, Denis Dimitrov, Andrey Kuznetsov
In this study, we present an investigation into the anisotropy dynamics and intrinsic dimension of embeddings in transformer architectures, focusing on the dichotomy between encoders and decoders.
no code implementations • 2 Oct 2023 • Roman Korkin, Ivan Oseledets, Aleksandr Katrutsa
This study proposes a novel memory-efficient recurrent neural network (RNN) architecture specified to solve the object localization problem.
no code implementations • 31 Aug 2023 • Egor Sevriugov, Ivan Oseledets
Evaluation metrics are essential for assessing the performance of generative models in image synthesis.
no code implementations • 31 Aug 2023 • Egor Sevriugov, Ivan Oseledets
Recent advancements in real image editing have been attributed to the exploration of Generative Adversarial Networks (GANs) latent space.
no code implementations • 17 Aug 2023 • Dmitrii Korzh, Mikhail Pautov, Olga Tsymboi, Ivan Oseledets
Randomized smoothing is the state-of-the-art approach to construct image classifiers that are provably robust against additive adversarial perturbations of bounded magnitude.
no code implementations • 8 Aug 2023 • Daria Cherniuk, Stanislav Abukhovich, Anh-Huy Phan, Ivan Oseledets, Andrzej Cichocki, Julia Gusak
Tensor decomposition of convolutional and fully-connected layers is an effective way to reduce parameters and FLOP in neural networks.
no code implementations • 5 Jun 2023 • Viktoriia Chekalina, Georgii Novikov, Julia Gusak, Ivan Oseledets, Alexander Panchenko
On the downstream tasks, including language understanding and text summarization, the model performs similarly to the original GPT-2 model.
no code implementations • 31 May 2023 • Marina Munkhoeva, Ivan Oseledets
Self-supervised methods received tremendous attention thanks to their seemingly heuristic approach to learning representations that respect the semantics of the data without any apparent supervision in the form of labels.
1 code implementation • 20 Mar 2023 • Andrei Chertkov, Olga Tsymboi, Mikhail Pautov, Ivan Oseledets
Neural networks are deployed widely in natural language processing tasks on the industrial scale, and perhaps the most often they are used as compounds of automatic machine translation systems.
no code implementations • 14 Mar 2023 • Roman Korkin, Ivan Oseledets, Aleksandr Katrutsa
Two well-known classes of filtering algorithms to solve the localization problem are Kalman filter-based methods and particle filter-based methods.
no code implementations • 5 Feb 2023 • Albert Sayapin, Gleb Balitskiy, Daniel Bershatsky, Aleksandr Katrutsa, Evgeny Frolov, Alexey Frolov, Ivan Oseledets, Vitaliy Kharin
Since the data about user experience are distributed among devices, the federated learning setup is used to train the proposed sequential matrix factorization model.
no code implementations • 12 Jan 2023 • Daria Fokina, Pavel Toktaliev, Oleg Iliev, Ivan Oseledets
Reactive flows in porous media play an important role in our life and are crucial for many industrial, environmental and biomedical applications.
no code implementations • 8 Jan 2023 • Yuliya Tukmacheva, Ivan Oseledets, Evgeny Frolov
We introduce the novel approach towards fake text reviews detection in collaborative filtering recommender systems.
no code implementations • 25 Dec 2022 • Daria Sushnikova, Pavel Kharyuk, Ivan Oseledets
In this paper, we propose a new neural network architecture based on the H2 matrix.
1 code implementation • 12 Dec 2022 • Evgeny Frolov, Ivan Oseledets
Self-attentive transformer models have recently been shown to solve the next item recommendation task very efficiently.
2 code implementations • 29 Sep 2022 • Valentin Leplat, Daniil Merkulov, Aleksandr Katrutsa, Daniel Bershatsky, Olga Tsymboi, Ivan Oseledets
Classical machine learning models such as deep neural networks are usually trained by using Stochastic Gradient Descent-based (SGD) algorithms.
1 code implementation • 29 Sep 2022 • Shakir Showkat Sofi, Ivan Oseledets
Numerous studies have proposed time series-based models as a viable alternative to numerical forecasts.
2 code implementations • 2 Aug 2022 • Mikhail Usvyatsov, Rafael Ballester-Rippoll, Lina Bashaeva, Konrad Schindler, Gonzalo Ferrer, Ivan Oseledets
We show that low-rank tensor compression is extremely compact to store and query time-varying signed distance functions.
1 code implementation • 31 Jul 2022 • Semen Budennyy, Vladimir Lazarev, Nikita Zakharenko, Alexey Korovin, Olga Plosskaya, Denis Dimitrov, Vladimir Arkhipkin, Ivan Oseledets, Ivan Barsola, Ilya Egorov, Aleksandra Kosterina, Leonid Zhukov
The size and complexity of deep neural networks continue to grow exponentially, significantly increasing energy consumption for training and inference by these models.
no code implementations • 6 Jul 2022 • Richik Sengupta, Soumik Adhikary, Ivan Oseledets, Jacob Biamonte
In this survey we recover the basics of tensor networks and explain the ongoing effort to develop the theory of tensor networks in machine learning.
1 code implementation • 10 May 2022 • Nikita Marin, Elizaveta Makhneva, Maria Lysyuk, Vladimir Chernyy, Ivan Oseledets, Evgeny Frolov
Conventional collaborative filtering techniques don't take into consideration the effect of discrepancy in users' rating perception.
1 code implementation • 9 May 2022 • Artyom Nikitin, Andrei Chertkov, Rafael Ballester-Ripoll, Ivan Oseledets, Evgeny Frolov
The problem is formulated as a Quadratic Unconstrained Binary Optimization (QUBO) which, due to its NP-hard complexity, is solved using Quantum Annealing on a quantum computer provided by D-Wave.
1 code implementation • 30 Apr 2022 • Konstantin Sozykin, Andrei Chertkov, Roman Schutski, Anh-Huy Phan, Andrzej Cichocki, Ivan Oseledets
We present a novel procedure for optimization based on the combination of efficient quantized tensor train representation and a generalized maximum matrix volume principle.
no code implementations • 25 Apr 2022 • Daria Fokina, Oleg Iliev, Pavel Toktaliev, Ivan Oseledets, Felix Schindler
Reactive flows are important part of numerous technical and environmental processes.
2 code implementations • CVPR 2022 • Aleksandr Ermolov, Leyla Mirvakhabova, Valentin Khrulkov, Nicu Sebe, Ivan Oseledets
Following this line of work, we propose a new hyperbolic-based model for metric learning.
Ranked #1 on Metric Learning on CUB-200-2011
no code implementations • 23 Feb 2022 • Aleksandr Katrutsa, Sergey Utyuzhnikov, Ivan Oseledets
The Dynamic Mode Decomposition has proved to be a very efficient technique to study dynamic data.
no code implementations • 21 Feb 2022 • Julia Gusak, Daria Cherniuk, Alena Shilova, Alexander Katrutsa, Daniel Bershatsky, Xunyi Zhao, Lionel Eyraud-Dubois, Oleg Shlyazhko, Denis Dimitrov, Ivan Oseledets, Olivier Beaumont
Modern Deep Neural Networks (DNNs) require significant memory to store weight, activations, and other intermediate tensors during training.
no code implementations • 14 Feb 2022 • Valentin Khrulkov, Gleb Ryzhakov, Andrei Chertkov, Ivan Oseledets
Diffusion models have recently outperformed alternative approaches to model the distribution of natural images, such as GANs.
1 code implementation • 2 Feb 2022 • Mikhail Pautov, Olesya Kuznetsova, Nurislam Tursynbek, Aleksandr Petiushko, Ivan Oseledets
In this work, we extend randomized smoothing to few-shot learning models that map inputs to normalized embeddings.
2 code implementations • 1 Feb 2022 • Georgii Novikov, Daniel Bershatsky, Julia Gusak, Alex Shonenkov, Denis Dimitrov, Ivan Oseledets
Every modern neural network model has quite a few pointwise nonlinearities in its architecture, and such operation induces additional memory costs which -- as we show -- can be significantly reduced by quantization of the gradients.
2 code implementations • 31 Jan 2022 • Daniel Bershatsky, Aleksandr Mikhalev, Alexandr Katrutsa, Julia Gusak, Daniil Merkulov, Ivan Oseledets
Also, we investigate the variance of the gradient estimate induced by the randomized matrix multiplication.
1 code implementation • ICCV 2021 • Valentin Khrulkov, Leyla Mirvakhabova, Ivan Oseledets, Artem Babenko
Recent advances in high-fidelity semantic image editing heavily rely on the presumably disentangled latent spaces of the state-of-the-art generative models, such as StyleGAN.
1 code implementation • 22 Sep 2021 • Mikhail Pautov, Nurislam Tursynbek, Marina Munkhoeva, Nikita Muravev, Aleksandr Petiushko, Ivan Oseledets
In safety-critical machine learning applications, it is crucial to defend models against adversarial attacks -- small modifications of the input that change the predictions.
no code implementations • 13 Jun 2021 • Svetlana Illarionova, Dmitrii Shadrin, Alexey Trekin, Vladimir Ignatiev, Ivan Oseledets
The near-infrared (NIR) spectral range (from 780 to 2500 nm) of the multispectral remote sensing imagery provides vital information for the landcover classification, especially concerning the vegetation assessment.
no code implementations • 12 May 2021 • Svetlana Illarionova, Sergey Nesteruk, Dmitrii Shadrin, Vladimir Ignatiev, Mariia Pukalchik, Ivan Oseledets
To overcome these data restrictions, it is common to consider various approaches including data augmentation techniques.
1 code implementation • 11 Apr 2021 • Oluwafemi Olaleke, Ivan Oseledets, Evgeny Frolov
In domains where users tend to develop long-term preferences that do not change too frequently, the stability of recommendations is an important factor of the perceived quality of a recommender system.
no code implementations • 27 Mar 2021 • Alexander Novikov, Maxim Rakhuba, Ivan Oseledets
In scientific computing and machine learning applications, matrices and more general multidimensional arrays (tensors) can often be approximated with the help of low-rank decompositions.
1 code implementation • 15 Mar 2021 • Julia Gusak, Alexandr Katrutsa, Talgat Daulbaev, Andrzej Cichocki, Ivan Oseledets
Moreover, we show that the right choice of solver parameterization can significantly affect neural ODEs models in terms of robustness to adversarial attacks.
no code implementations • 11 Feb 2021 • Valentin Khrulkov, Leyla Mirvakhabova, Ivan Oseledets, Artem Babenko
Constructing disentangled representations is known to be a difficult task, especially in the unsupervised scenario.
no code implementations • 8 Feb 2021 • Valentin Khrulkov, Artem Babenko, Ivan Oseledets
Recent work demonstrated the benefits of studying continuous-time dynamics governing the GAN training.
no code implementations • 22 Jan 2021 • Charlie Vanaret, Philipp Seufert, Jan Schwientek, Gleb Karpov, Gleb Ryzhakov, Ivan Oseledets, Norbert Asprion, Michael Bortz
Model-based experimental design is attracting increasing attention in chemical process engineering.
Optimization and Control
no code implementations • 13 Jan 2021 • Tsimboy Olga, Yermek Kapushev, Evgeny Burnaev, Ivan Oseledets
In this work we derive analytical expression for the Denoising Score matching using the Kernel Exponential Family as a model distribution.
no code implementations • 14 Dec 2020 • Nurislam Tursynbek, Aleksandr Petiushko, Ivan Oseledets
Differential privacy (DP) is a gold-standard concept of measuring and guaranteeing privacy in data analysis.
1 code implementation • 18 Nov 2020 • Nurislam Tursynbek, Ilya Vilkoviskiy, Maria Sindeeva, Ivan Oseledets
Furthermore, we propose to use Turing patterns, generated by cellular automata, as universal perturbations, and experimentally show that they significantly degrade the performance of deep learning models.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Oleksii Hrinchuk, Valentin Khrulkov, Leyla Mirvakhabova, Elena Orlova, Ivan Oseledets
The embedding layers transforming input words into real vectors are the key components of deep neural networks used in natural language processing.
no code implementations • NeurIPS Workshop LMCA 2020 • Taras Khakhulin, Roman Schutski, Ivan Oseledets
We propose a Reinforcement Learning-based approach to approximately solve the Tree Decomposition problem.
no code implementations • 6 Oct 2020 • Evgeny Ponomarev, Sergey Matveev, Ivan Oseledets
In this work, we consider latency approximation on mobile GPU as a data and hardware-specific problem.
1 code implementation • 15 Aug 2020 • Leyla Mirvakhabova, Evgeny Frolov, Valentin Khrulkov, Ivan Oseledets, Alexander Tuzhilin
We introduce a simple autoencoder based on hyperbolic geometry for solving standard collaborative filtering problem.
no code implementations • ECCV 2020 • Anh-Huy Phan, Konstantin Sobolev, Konstantin Sozykin, Dmitry Ermilov, Julia Gusak, Petr Tichavsky, Valeriy Glukhov, Ivan Oseledets, Andrzej Cichocki
Most state of the art deep neural networks are overparameterized and exhibit a high computational cost.
1 code implementation • 14 Jul 2020 • Alexandr Katrutsa, Daniil Merkulov, Nurislam Tursynbek, Ivan Oseledets
This descent direction is based on the normalized gradients of the individual losses.
no code implementations • 28 Jun 2020 • Nurislam Tursynbek, Aleksandr Petiushko, Ivan Oseledets
The brittleness of deep image classifiers to small adversarial input perturbations has been extensively studied in the last several years.
no code implementations • 8 Jun 2020 • Anton Tsitsulin, Marina Munkhoeva, Davide Mottin, Panagiotis Karras, Ivan Oseledets, Emmanuel Müller
Low-dimensional representations, or embeddings, of a graph's nodes facilitate several practical data science and data engineering tasks.
no code implementations • 5 Jun 2020 • Anna Shalova, Ivan Oseledets
The identification of nonlinear dynamics from observations is essential for the alignment of the theoretical ideas and experimental data.
1 code implementation • ICLR Workshop DeepDiffEq 2019 • Julia Gusak, Larisa Markeeva, Talgat Daulbaev, Alexandr Katrutsa, Andrzej Cichocki, Ivan Oseledets
Normalization is an important and vastly investigated technique in deep learning.
no code implementations • ICLR Workshop DeepDiffEq 2019 • Daniil Merkulov, Ivan Oseledets
We present a different view on stochastic optimization, which goes back to the splitting schemes for approximate solutions of ODE.
no code implementations • 7 Apr 2020 • Ivan Matvienko, Mikhail Gasanov, Anna Petrovskaia, Raghavendra Belur Jana, Maria Pukalchik, Ivan Oseledets
We present a comparison between the classical ML approaches and U-Net NN for classifying crops with a single satellite image.
1 code implementation • NeurIPS 2020 • Talgat Daulbaev, Alexandr Katrutsa, Larisa Markeeva, Julia Gusak, Andrzej Cichocki, Ivan Oseledets
We propose a simple interpolation-based method for the efficient approximation of gradients in neural ODE models.
1 code implementation • 26 Feb 2020 • Evgeny Ponomarev, Ivan Oseledets, Andrzej Cichocki
A system for trading the fixed volume of a financial instrument is proposed and experimentally tested; this is based on the asynchronous advantage actor-critic method with the use of several neural network architectures.
no code implementations • 10 Feb 2020 • Anna Shalova, Ivan Oseledets
Proper states' representations are the key to the successful dynamics modeling of chaotic systems.
no code implementations • 11 Dec 2019 • Yermek Kapushev, Ivan Oseledets, Evgeny Burnaev
In this paper, we consider the tensor completion problem representing the solution in the tensor train (TT) format.
1 code implementation • 29 Oct 2019 • Chunfeng Cui, Kaiqi Zhang, Talgat Daulbaev, Julia Gusak, Ivan Oseledets, Zheng Zhang
Secondly, we propose analyzing the vulnerability of a neural network using active subspace and finding an additive universal adversarial attack vector that can misclassify a dataset with a high probability.
1 code implementation • 28 Oct 2019 • Daria Fokina, Ivan Oseledets
We propose a new method for learning deep neural network models that is based on a greedy learning approach: we add one basis function at a time, and a new basis function is generated as a non-linear activation function applied to a linear combination of the previous basis functions.
no code implementations • 18 Oct 2019 • Taras Khakhulin, Roman Schutski, Ivan Oseledets
We show that the agent builton GCN and trained on a single graph using an Actor-Critic method can efficiently generalize to real-world TD problem instances.
no code implementations • 15 Oct 2019 • Julia Gusak, Talgat Daulbaev, Evgeny Ponomarev, Andrzej Cichocki, Ivan Oseledets
We introduce a new method for speeding up the inference of deep neural networks.
no code implementations • 11 Oct 2019 • Artem Chashchin, Mikhail Botchev, Ivan Oseledets, George Ovchinnikov
We show how by training neural networks with ResNet-like architecture on the solution samples, models can be developed to predict the ODE system solution further in time.
no code implementations • 14 Jun 2019 • Daniil Merkulov, Ivan Oseledets
In this paper we propose a method of obtaining points of extreme overfitting - parameters of modern neural networks, at which they demonstrate close to 100 % training accuracy, simultaneously with almost zero accuracy on the test sample.
2 code implementations • ICLR 2020 • Anton Tsitsulin, Marina Munkhoeva, Davide Mottin, Panagiotis Karras, Alex Bronstein, Ivan Oseledets, Emmanuel Müller
The ability to represent and compare machine learning models is crucial in order to quantify subtle model changes, evaluate generative models, and gather insights on neural network architectures.
no code implementations • 27 May 2019 • Valentin Khrulkov, Ivan Oseledets
Despite the fact that generative models are extremely successful in practice, the theory underlying this phenomenon is only starting to catch up with practice.
3 code implementations • CVPR 2020 • Valentin Khrulkov, Leyla Mirvakhabova, Evgeniya Ustinova, Ivan Oseledets, Victor Lempitsky
Computer vision tasks such as image classification, image retrieval and few-shot learning are currently dominated by Euclidean and spherical embeddings, so that the final decisions about class belongings or the degree of similarity are made using linear hyperplanes, Euclidean distances, or spherical geodesic distances (cosine similarity).
3 code implementations • 24 Mar 2019 • Julia Gusak, Maksym Kholiavchenko, Evgeny Ponomarev, Larisa Markeeva, Ivan Oseledets, Andrzej Cichocki
The low-rank tensor approximation is very promising for the compression of deep neural networks.
1 code implementation • 5 Mar 2019 • Alexandr Katrutsa, Ivan Oseledets
Therefore, to reduce this complexity, we use random sketching and compare it with the Kaczmarz method without preconditioning.
Numerical Analysis
1 code implementation • 30 Jan 2019 • Oleksii Hrinchuk, Valentin Khrulkov, Leyla Mirvakhabova, Elena Orlova, Ivan Oseledets
The embedding layers transforming input words into real vectors are the key components of deep neural networks used in natural language processing.
no code implementations • ICLR 2019 • Valentin Khrulkov, Oleksii Hrinchuk, Ivan Oseledets
Such networks, however, are not very often applied to real life tasks.
no code implementations • 8 Jan 2019 • Pavel Temirchev, Maxim Simonov, Ruslan Kostoev, Evgeny Burnaev, Ivan Oseledets, Alexey Akhmetov, Andrey Margarit, Alexander Sitnikov, Dmitry Koroteev
We present a novel technique for assessing the dynamics of multiphase fluid flow in the oil reservoir.
no code implementations • 18 Dec 2018 • Tsui-Wei Weng, Pin-Yu Chen, Lam M. Nguyen, Mark S. Squillante, Ivan Oseledets, Luca Daniel
With deep neural networks providing state-of-the-art machine learning models for numerous machine learning tasks, quantifying the robustness of these models has become an important area of research.
no code implementations • 20 Nov 2018 • Sergei Divakov, Ivan Oseledets
We present a novel approach to point set registration which is based on one-shot adversarial learning.
no code implementations • 7 Aug 2018 • Pavel Kharyuk, Ivan Oseledets
This work is devoted to elaboration on the idea to use block term decomposition for group data analysis and to raise the possibility of modelling group activity with (Lr, 1) and Tucker blocks.
no code implementations • 27 Jul 2018 • Evgeny Frolov, Ivan Oseledets
We propose a tensor-based model that fuses a more granular representation of user preferences with the ability to take additional side information into account.
no code implementations • 18 Jul 2018 • Pavel Kharyuk, Dmitry Nazarenko, Ivan Oseledets
Fourier-transform infra-red (FTIR) spectra of samples from 7 plant species were used to explore the influence of preprocessing and feature extraction on efficiency of machine learning algorithms.
3 code implementations • 18 Feb 2018 • Evgeny Frolov, Ivan Oseledets
We propose a new hybrid algorithm that allows incorporating both user and item side information within the standard collaborative filtering technique.
2 code implementations • ICLR 2018 • Marina Munkhoeva, Yermek Kapushev, Evgeny Burnaev, Ivan Oseledets
We consider the problem of improving kernel approximation via randomized feature maps.
1 code implementation • ICML 2018 • Valentin Khrulkov, Ivan Oseledets
One of the biggest challenges in the research of generative adversarial networks (GANs) is assessing the quality of generated samples and detecting various levels of mode collapse.
2 code implementations • 5 Jan 2018 • Alexander Novikov, Pavel Izmailov, Valentin Khrulkov, Michael Figurnov, Ivan Oseledets
Tensor Train decomposition is used across many branches of machine learning.
Mathematical Software Numerical Analysis
1 code implementation • 10 Nov 2017 • Alexandr Katrutsa, Talgat Daulbaev, Ivan Oseledets
This paper proposes the method to optimize restriction and prolongation operators in the two-grid method.
Numerical Analysis
2 code implementations • ICLR 2018 • Valentin Khrulkov, Alexander Novikov, Ivan Oseledets
In this paper, we prove the expressive power theorem (an exponential lower bound on the width of the equivalent shallow network) for a class of recurrent neural networks -- ones that correspond to the Tensor Train (TT) decomposition.
3 code implementations • 27 Sep 2017 • Ivan Sosnovik, Ivan Oseledets
The main novelty of this work is to state the problem as an image segmentation task.
no code implementations • CVPR 2018 • Valentin Khrulkov, Ivan Oseledets
Vulnerability of Deep Neural Networks (DNNs) to adversarial attacks has been attracting a lot of attention in recent studies.
1 code implementation • ACL 2017 • Alexander Fonarev, Oleksii Hrinchuk, Gleb Gusev, Pavel Serdyukov, Ivan Oseledets
Skip-Gram Negative Sampling (SGNS) word embedding model, well known by its implementation in "word2vec" software, is usually optimized by stochastic gradient descent.
no code implementations • 16 Oct 2016 • Alexander Fonarev, Alexander Mikhalev, Pavel Serdyukov, Gleb Gusev, Ivan Oseledets
Cold start problem in Collaborative Filtering can be solved by asking new users to rate a small seed set of representative items or by asking representative users to rate a new item.
2 code implementations • 14 Jul 2016 • Evgeny Frolov, Ivan Oseledets
In order to resolve this problem we propose to treat user feedback as a categorical variable and model it with users and items in a ternary way.
3 code implementations • 12 May 2016 • Alexander Novikov, Mikhail Trofimov, Ivan Oseledets
Modeling interactions between features improves the performance of machine learning solutions in many domains (e. g. recommender systems or sentiment analysis).
no code implementations • 19 Mar 2016 • Evgeny Frolov, Ivan Oseledets
A substantial progress in development of new and efficient tensor factorization techniques has led to an extensive research of their applicability in recommender systems field.
no code implementations • 24 Feb 2015 • Ben Usman, Ivan Oseledets
We propose a generalization of SimRank similarity measure for heterogeneous information networks.
10 code implementations • 19 Dec 2014 • Vadim Lebedev, Yaroslav Ganin, Maksim Rakhuba, Ivan Oseledets, Victor Lempitsky
We propose a simple two-step approach for speeding up convolution layers within large convolutional neural networks based on tensor decomposition and discriminative fine-tuning.
3 code implementations • 21 Aug 2014 • Jutho Haegeman, Christian Lubich, Ivan Oseledets, Bart Vandereycken, Frank Verstraete
In particular, our method is compatible with any Hamiltonian for which DMRG can be implemented efficiently and DMRG is obtained as a special case of imaginary time evolution with infinite time step.
Quantum Physics Strongly Correlated Electrons
1 code implementation • 6 Jan 2013 • Christian Lubich, Ivan Oseledets
The dynamical low-rank approximation of time-dependent matrices is a low-rank factorization updating technique.
Numerical Analysis 65F30, 65L05, 65L20, 15A23