Search Results for author: Ivan Oseledets

Found 106 papers, 45 papers with code

Black-Box Approximation and Optimization with Hierarchical Tucker Decomposition

no code implementations5 Feb 2024 Gleb Ryzhakov, Andrei Chertkov, Artem Basharin, Ivan Oseledets

We develop a new method HTBB for the multidimensional black-box approximation and gradient-free optimization, which is based on the low-rank hierarchical Tucker decomposition with the use of the MaxVol indices selection procedure.

Smart Flow Matching: On The Theory of Flow Matching Algorithms with Applications

no code implementations5 Feb 2024 Gleb Ryzhakov, Svetlana Pavlova, Egor Sevriugov, Ivan Oseledets

The paper presents the exact formula for the vector field that minimizes the loss for the standard flow.

LoTR: Low Tensor Rank Weight Adaptation

no code implementations2 Feb 2024 Daniel Bershatsky, Daria Cherniuk, Talgat Daulbaev, Aleksandr Mikhalev, Ivan Oseledets

In this paper we generalize and extend an idea of low-rank adaptation (LoRA) of large language models (LLMs) based on Transformer architecture.

Tensor Decomposition

Sparse and Transferable Universal Singular Vectors Attack

no code implementations25 Jan 2024 Kseniia Kuvshinova, Olga Tsymboi, Ivan Oseledets

The research in the field of adversarial attacks and models' vulnerability is one of the fundamental directions in modern machine learning.

Adversarial Attack

Probabilistically Robust Watermarking of Neural Networks

no code implementations16 Jan 2024 Mikhail Pautov, Nikita Bogdanov, Stanislav Pyatkin, Oleg Rogov, Ivan Oseledets

As deep learning (DL) models are widely and effectively used in Machine Learning as a Service (MLaaS) platforms, there is a rapidly growing interest in DL watermarking techniques that can be used to confirm the ownership of a particular model.

Fast gradient-free activation maximization for neurons in spiking neural networks

1 code implementation28 Dec 2023 Nikita Pospelov, Andrei Chertkov, Maxim Beketov, Ivan Oseledets, Konstantin Anokhin

Neural networks (NNs), both living and artificial, work due to being complex systems of neurons, each having its own specialization.

Tensor Decomposition

Run LoRA Run: Faster and Lighter LoRA Implementations

no code implementations6 Dec 2023 Daria Cherniuk, Aleksandr Mikhalev, Ivan Oseledets

LoRA is a technique that reduces the number of trainable parameters in a neural network by introducing low-rank adapters to linear layers.

Dynamic Collaborative Filtering for Matrix- and Tensor-based Recommender Systems

1 code implementation4 Dec 2023 Albert Saiapin, Ivan Oseledets, Evgeny Frolov

In production applications of recommender systems, a continuous data flow is employed to update models in real-time.

Collaborative Filtering Recommendation Systems

The Shape of Learning: Anisotropy and Intrinsic Dimensions in Transformer-Based Models

no code implementations10 Nov 2023 Anton Razzhigaev, Matvey Mikhalchuk, Elizaveta Goncharova, Ivan Oseledets, Denis Dimitrov, Andrey Kuznetsov

In this study, we present an investigation into the anisotropy dynamics and intrinsic dimension of embeddings in transformer architectures, focusing on the dichotomy between encoders and decoders.

Memory-efficient particle filter recurrent neural network for object localization

no code implementations2 Oct 2023 Roman Korkin, Ivan Oseledets, Aleksandr Katrutsa

This study proposes a novel memory-efficient recurrent neural network (RNN) architecture specified to solve the object localization problem.

Object Localization

Unsupervised evaluation of GAN sample quality: Introducing the TTJac Score

no code implementations31 Aug 2023 Egor Sevriugov, Ivan Oseledets

Evaluation metrics are essential for assessing the performance of generative models in image synthesis.

Image Generation

Robust GAN inversion

no code implementations31 Aug 2023 Egor Sevriugov, Ivan Oseledets

Recent advancements in real image editing have been attributed to the exploration of Generative Adversarial Networks (GANs) latent space.

Computational Efficiency

General Lipschitz: Certified Robustness Against Resolvable Semantic Transformations via Transformation-Dependent Randomized Smoothing

no code implementations17 Aug 2023 Dmitrii Korzh, Mikhail Pautov, Olga Tsymboi, Ivan Oseledets

Randomized smoothing is the state-of-the-art approach to construct image classifiers that are provably robust against additive adversarial perturbations of bounded magnitude.

Translation

Quantization Aware Factorization for Deep Neural Network Compression

no code implementations8 Aug 2023 Daria Cherniuk, Stanislav Abukhovich, Anh-Huy Phan, Ivan Oseledets, Andrzej Cichocki, Julia Gusak

Tensor decomposition of convolutional and fully-connected layers is an effective way to reduce parameters and FLOP in neural networks.

Neural Network Compression Quantization +1

Efficient GPT Model Pre-training using Tensor Train Matrix Representation

no code implementations5 Jun 2023 Viktoriia Chekalina, Georgii Novikov, Julia Gusak, Ivan Oseledets, Alexander Panchenko

On the downstream tasks, including language understanding and text summarization, the model performs similarly to the original GPT-2 model.

Language Modelling Text Summarization

Spectal Harmonics: Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning

no code implementations31 May 2023 Marina Munkhoeva, Ivan Oseledets

Self-supervised methods received tremendous attention thanks to their seemingly heuristic approach to learning representations that respect the semantics of the data without any apparent supervision in the form of labels.

Inductive Bias Low-Rank Matrix Completion +2

Translate your gibberish: black-box adversarial attack on machine translation systems

1 code implementation20 Mar 2023 Andrei Chertkov, Olga Tsymboi, Mikhail Pautov, Ivan Oseledets

Neural networks are deployed widely in natural language processing tasks on the industrial scale, and perhaps the most often they are used as compounds of automatic machine translation systems.

Adversarial Attack Machine Translation +1

Multiparticle Kalman filter for object localization in symmetric environments

no code implementations14 Mar 2023 Roman Korkin, Ivan Oseledets, Aleksandr Katrutsa

Two well-known classes of filtering algorithms to solve the localization problem are Kalman filter-based methods and particle filter-based methods.

Object Localization

Federated Privacy-preserving Collaborative Filtering for On-Device Next App Prediction

no code implementations5 Feb 2023 Albert Sayapin, Gleb Balitskiy, Daniel Bershatsky, Aleksandr Katrutsa, Evgeny Frolov, Alexey Frolov, Ivan Oseledets, Vitaliy Kharin

Since the data about user experience are distributed among devices, the federated learning setup is used to train the proposed sequential matrix factorization model.

Collaborative Filtering Federated Learning +1

Machine learning methods for prediction of breakthrough curves in reactive porous media

no code implementations12 Jan 2023 Daria Fokina, Pavel Toktaliev, Oleg Iliev, Ivan Oseledets

Reactive flows in porous media play an important role in our life and are crucial for many industrial, environmental and biomedical applications.

Gaussian Processes

Mitigating Human and Computer Opinion Fraud via Contrastive Learning

no code implementations8 Jan 2023 Yuliya Tukmacheva, Ivan Oseledets, Evgeny Frolov

We introduce the novel approach towards fake text reviews detection in collaborative filtering recommender systems.

Collaborative Filtering Contrastive Learning +1

FMM-Net: neural network architecture based on the Fast Multipole Method

no code implementations25 Dec 2022 Daria Sushnikova, Pavel Kharyuk, Ivan Oseledets

In this paper, we propose a new neural network architecture based on the H2 matrix.

Tensor-based Sequential Learning via Hankel Matrix Representation for Next Item Recommendations

1 code implementation12 Dec 2022 Evgeny Frolov, Ivan Oseledets

Self-attentive transformer models have recently been shown to solve the next item recommendation task very efficiently.

A case study of spatiotemporal forecasting techniques for weather forecasting

no code implementations29 Sep 2022 Shakir Showkat Sofi, Ivan Oseledets

The majority of real-world processes are spatiotemporal, and the data generated by them exhibits both spatial and temporal evolution.

Time Series Time Series Analysis +1

NAG-GS: Semi-Implicit, Accelerated and Robust Stochastic Optimizer

2 code implementations29 Sep 2022 Valentin Leplat, Daniil Merkulov, Aleksandr Katrutsa, Daniel Bershatsky, Olga Tsymboi, Ivan Oseledets

Classical machine learning models such as deep neural networks are usually trained by using Stochastic Gradient Descent-based (SGD) algorithms.

T4DT: Tensorizing Time for Learning Temporal 3D Visual Data

2 code implementations2 Aug 2022 Mikhail Usvyatsov, Rafael Ballester-Rippoll, Lina Bashaeva, Konrad Schindler, Gonzalo Ferrer, Ivan Oseledets

We show that low-rank tensor compression is extremely compact to store and query time-varying signed distance functions.

Eco2AI: carbon emissions tracking of machine learning models as the first step towards sustainable AI

1 code implementation31 Jul 2022 Semen Budennyy, Vladimir Lazarev, Nikita Zakharenko, Alexey Korovin, Olga Plosskaya, Denis Dimitrov, Vladimir Arkhipkin, Ivan Oseledets, Ivan Barsola, Ilya Egorov, Aleksandra Kosterina, Leonid Zhukov

The size and complexity of deep neural networks continue to grow exponentially, significantly increasing energy consumption for training and inference by these models.

Tensor networks in machine learning

no code implementations6 Jul 2022 Richik Sengupta, Soumik Adhikary, Ivan Oseledets, Jacob Biamonte

In this survey we recover the basics of tensor networks and explain the ongoing effort to develop the theory of tensor networks in machine learning.

BIG-bench Machine Learning Tensor Decomposition +1

Tensor-based Collaborative Filtering With Smooth Ratings Scale

1 code implementation10 May 2022 Nikita Marin, Elizaveta Makhneva, Maria Lysyuk, Vladimir Chernyy, Ivan Oseledets, Evgeny Frolov

Conventional collaborative filtering techniques don't take into consideration the effect of discrepancy in users' rating perception.

Collaborative Filtering Recommendation Systems

Are Quantum Computers Practical Yet? A Case for Feature Selection in Recommender Systems using Tensor Networks

1 code implementation9 May 2022 Artyom Nikitin, Andrei Chertkov, Rafael Ballester-Ripoll, Ivan Oseledets, Evgeny Frolov

The problem is formulated as a Quadratic Unconstrained Binary Optimization (QUBO) which, due to its NP-hard complexity, is solved using Quantum Annealing on a quantum computer provided by D-Wave.

Collaborative Filtering Feature Engineering +3

TTOpt: A Maximum Volume Quantized Tensor Train-based Optimization and its Application to Reinforcement Learning

1 code implementation30 Apr 2022 Konstantin Sozykin, Andrei Chertkov, Roman Schutski, Anh-Huy Phan, Andrzej Cichocki, Ivan Oseledets

We present a novel procedure for optimization based on the combination of efficient quantized tensor train representation and a generalized maximum matrix volume principle.

reinforcement-learning Reinforcement Learning (RL)

Survey on Large Scale Neural Network Training

no code implementations21 Feb 2022 Julia Gusak, Daria Cherniuk, Alena Shilova, Alexander Katrutsa, Daniel Bershatsky, Xunyi Zhao, Lionel Eyraud-Dubois, Oleg Shlyazhko, Denis Dimitrov, Ivan Oseledets, Olivier Beaumont

Modern Deep Neural Networks (DNNs) require significant memory to store weight, activations, and other intermediate tensors during training.

Understanding DDPM Latent Codes Through Optimal Transport

no code implementations14 Feb 2022 Valentin Khrulkov, Gleb Ryzhakov, Andrei Chertkov, Ivan Oseledets

Diffusion models have recently outperformed alternative approaches to model the distribution of natural images, such as GANs.

Smoothed Embeddings for Certified Few-Shot Learning

1 code implementation2 Feb 2022 Mikhail Pautov, Olesya Kuznetsova, Nurislam Tursynbek, Aleksandr Petiushko, Ivan Oseledets

In this work, we extend randomized smoothing to few-shot learning models that map inputs to normalized embeddings.

Adversarial Robustness Few-Shot Learning

Few-Bit Backward: Quantized Gradients of Activation Functions for Memory Footprint Reduction

2 code implementations1 Feb 2022 Georgii Novikov, Daniel Bershatsky, Julia Gusak, Alex Shonenkov, Denis Dimitrov, Ivan Oseledets

Every modern neural network model has quite a few pointwise nonlinearities in its architecture, and such operation induces additional memory costs which -- as we show -- can be significantly reduced by quantization of the gradients.

Neural Network Compression Quantization

Memory-Efficient Backpropagation through Large Linear Layers

2 code implementations31 Jan 2022 Daniel Bershatsky, Aleksandr Mikhalev, Alexandr Katrutsa, Julia Gusak, Daniil Merkulov, Ivan Oseledets

Also, we investigate the variance of the gradient estimate induced by the randomized matrix multiplication.

Model Compression

Latent Transformations via NeuralODEs for GAN-based Image Editing

1 code implementation ICCV 2021 Valentin Khrulkov, Leyla Mirvakhabova, Ivan Oseledets, Artem Babenko

Recent advances in high-fidelity semantic image editing heavily rely on the presumably disentangled latent spaces of the state-of-the-art generative models, such as StyleGAN.

Attribute

CC-Cert: A Probabilistic Approach to Certify General Robustness of Neural Networks

1 code implementation22 Sep 2021 Mikhail Pautov, Nurislam Tursynbek, Marina Munkhoeva, Nikita Muravev, Aleksandr Petiushko, Ivan Oseledets

In safety-critical machine learning applications, it is crucial to defend models against adversarial attacks -- small modifications of the input that change the predictions.

Adversarial Robustness

Generation of the NIR spectral Band for Satellite Images with Convolutional Neural Networks

no code implementations13 Jun 2021 Svetlana Illarionova, Dmitrii Shadrin, Alexey Trekin, Vladimir Ignatiev, Ivan Oseledets

The near-infrared (NIR) spectral range (from 780 to 2500 nm) of the multispectral remote sensing imagery provides vital information for the landcover classification, especially concerning the vegetation assessment.

Colorization Generative Adversarial Network +1

Dynamic Modeling of User Preferences for Stable Recommendations

1 code implementation11 Apr 2021 Oluwafemi Olaleke, Ivan Oseledets, Evgeny Frolov

In domains where users tend to develop long-term preferences that do not change too frequently, the stability of recommendations is an important factor of the perceived quality of a recommender system.

Incremental Learning Recommendation Systems

Automatic differentiation for Riemannian optimization on low-rank matrix and tensor-train manifolds

no code implementations27 Mar 2021 Alexander Novikov, Maxim Rakhuba, Ivan Oseledets

In scientific computing and machine learning applications, matrices and more general multidimensional arrays (tensors) can often be approximated with the help of low-rank decompositions.

Riemannian optimization

Meta-Solver for Neural Ordinary Differential Equations

1 code implementation15 Mar 2021 Julia Gusak, Alexandr Katrutsa, Talgat Daulbaev, Andrzej Cichocki, Ivan Oseledets

Moreover, we show that the right choice of solver parameterization can significantly affect neural ODEs models in terms of robustness to adversarial attacks.

Disentangled Representations from Non-Disentangled Models

no code implementations11 Feb 2021 Valentin Khrulkov, Leyla Mirvakhabova, Ivan Oseledets, Artem Babenko

Constructing disentangled representations is known to be a difficult task, especially in the unsupervised scenario.

Disentanglement Fairness

Functional Space Analysis of Local GAN Convergence

no code implementations8 Feb 2021 Valentin Khrulkov, Artem Babenko, Ivan Oseledets

Recent work demonstrated the benefits of studying continuous-time dynamics governing the GAN training.

Data Augmentation

Denoising Score Matching with Random Fourier Features

no code implementations13 Jan 2021 Tsimboy Olga, Yermek Kapushev, Evgeny Burnaev, Ivan Oseledets

In this work we derive analytical expression for the Denoising Score matching using the Kernel Exponential Family as a model distribution.

Denoising Density Estimation

Robustness Threats of Differential Privacy

no code implementations14 Dec 2020 Nurislam Tursynbek, Aleksandr Petiushko, Ivan Oseledets

Differential privacy (DP) is a gold-standard concept of measuring and guaranteeing privacy in data analysis.

Adversarial Turing Patterns from Cellular Automata

1 code implementation18 Nov 2020 Nurislam Tursynbek, Ilya Vilkoviskiy, Maria Sindeeva, Ivan Oseledets

Furthermore, we propose to use Turing patterns, generated by cellular automata, as universal perturbations, and experimentally show that they significantly degrade the performance of deep learning models.

Tensorized Embedding Layers

no code implementations Findings of the Association for Computational Linguistics 2020 Oleksii Hrinchuk, Valentin Khrulkov, Leyla Mirvakhabova, Elena Orlova, Ivan Oseledets

The embedding layers transforming input words into real vectors are the key components of deep neural networks used in natural language processing.

Performance of Hyperbolic Geometry Models on Top-N Recommendation Tasks

1 code implementation15 Aug 2020 Leyla Mirvakhabova, Evgeny Frolov, Valentin Khrulkov, Ivan Oseledets, Alexander Tuzhilin

We introduce a simple autoencoder based on hyperbolic geometry for solving standard collaborative filtering problem.

Collaborative Filtering

Geometry-Inspired Top-k Adversarial Perturbations

no code implementations28 Jun 2020 Nurislam Tursynbek, Aleksandr Petiushko, Ivan Oseledets

The brittleness of deep image classifiers to small adversarial input perturbations has been extensively studied in the last several years.

FREDE: Anytime Graph Embeddings

no code implementations8 Jun 2020 Anton Tsitsulin, Marina Munkhoeva, Davide Mottin, Panagiotis Karras, Ivan Oseledets, Emmanuel Müller

Low-dimensional representations, or embeddings, of a graph's nodes facilitate several practical data science and data engineering tasks.

Graph Embedding

Tensorized Transformer for Dynamical Systems Modeling

no code implementations5 Jun 2020 Anna Shalova, Ivan Oseledets

The identification of nonlinear dynamics from observations is essential for the alignment of the theoretical ideas and experimental data.

Language Modelling

Stochastic gradient algorithms from ODE splitting perspective

no code implementations ICLR Workshop DeepDiffEq 2019 Daniil Merkulov, Ivan Oseledets

We present a different view on stochastic optimization, which goes back to the splitting schemes for approximate solutions of ODE.

regression Stochastic Optimization

Using Reinforcement Learning in the Algorithmic Trading Problem

1 code implementation26 Feb 2020 Evgeny Ponomarev, Ivan Oseledets, Andrzej Cichocki

A system for trading the fixed volume of a financial instrument is proposed and experimentally tested; this is based on the asynchronous advantage actor-critic method with the use of several neural network architectures.

Algorithmic Trading reinforcement-learning +1

Deep Representation Learning for Dynamical Systems Modeling

no code implementations10 Feb 2020 Anna Shalova, Ivan Oseledets

Proper states' representations are the key to the successful dynamics modeling of chaotic systems.

Representation Learning

Tensor Completion via Gaussian Process Based Initialization

no code implementations11 Dec 2019 Yermek Kapushev, Ivan Oseledets, Evgeny Burnaev

In this paper, we consider the tensor completion problem representing the solution in the tensor train (TT) format.

regression

Active Subspace of Neural Networks: Structural Analysis and Universal Attacks

1 code implementation29 Oct 2019 Chunfeng Cui, Kaiqi Zhang, Talgat Daulbaev, Julia Gusak, Ivan Oseledets, Zheng Zhang

Secondly, we propose analyzing the vulnerability of a neural network using active subspace and finding an additive universal adversarial attack vector that can misclassify a dataset with a high probability.

Adversarial Attack Uncertainty Quantification

Growing axons: greedy learning of neural networks with application to function approximation

1 code implementation28 Oct 2019 Daria Fokina, Ivan Oseledets

We propose a new method for learning deep neural network models that is based on a greedy learning approach: we add one basis function at a time, and a new basis function is generated as a non-linear activation function applied to a linear combination of the previous basis functions.

Graph Convolutional Policy for Solving Tree Decomposition via Reinforcement Learning Heuristics

no code implementations18 Oct 2019 Taras Khakhulin, Roman Schutski, Ivan Oseledets

We show that the agent builton GCN and trained on a single graph using an Actor-Critic method can efficiently generalize to real-world TD problem instances.

reinforcement-learning Reinforcement Learning (RL) +1

Reduced-Order Modeling of Deep Neural Networks

no code implementations15 Oct 2019 Julia Gusak, Talgat Daulbaev, Evgeny Ponomarev, Andrzej Cichocki, Ivan Oseledets

We introduce a new method for speeding up the inference of deep neural networks.

Predicting dynamical system evolution with residual neural networks

no code implementations11 Oct 2019 Artem Chashchin, Mikhail Botchev, Ivan Oseledets, George Ovchinnikov

We show how by training neural networks with ResNet-like architecture on the solution samples, models can be developed to predict the ODE system solution further in time.

Time Series Time Series Analysis

Empirical study of extreme overfitting points of neural networks

no code implementations14 Jun 2019 Daniil Merkulov, Ivan Oseledets

In this paper we propose a method of obtaining points of extreme overfitting - parameters of modern neural networks, at which they demonstrate close to 100 % training accuracy, simultaneously with almost zero accuracy on the test sample.

The Shape of Data: Intrinsic Distance for Data Distributions

2 code implementations ICLR 2020 Anton Tsitsulin, Marina Munkhoeva, Davide Mottin, Panagiotis Karras, Alex Bronstein, Ivan Oseledets, Emmanuel Müller

The ability to represent and compare machine learning models is crucial in order to quantify subtle model changes, evaluate generative models, and gather insights on neural network architectures.

Universality Theorems for Generative Models

no code implementations27 May 2019 Valentin Khrulkov, Ivan Oseledets

Despite the fact that generative models are extremely successful in practice, the theory underlying this phenomenon is only starting to catch up with practice.

Hyperbolic Image Embeddings

3 code implementations CVPR 2020 Valentin Khrulkov, Leyla Mirvakhabova, Evgeniya Ustinova, Ivan Oseledets, Victor Lempitsky

Computer vision tasks such as image classification, image retrieval and few-shot learning are currently dominated by Euclidean and spherical embeddings, so that the final decisions about class belongings or the degree of similarity are made using linear hyperplanes, Euclidean distances, or spherical geodesic distances (cosine similarity).

Few-Shot Learning General Classification +3

Preconditioning Kaczmarz method by sketching

1 code implementation5 Mar 2019 Alexandr Katrutsa, Ivan Oseledets

Therefore, to reduce this complexity, we use random sketching and compare it with the Kaczmarz method without preconditioning.

Numerical Analysis

Tensorized Embedding Layers for Efficient Model Compression

1 code implementation30 Jan 2019 Oleksii Hrinchuk, Valentin Khrulkov, Leyla Mirvakhabova, Elena Orlova, Ivan Oseledets

The embedding layers transforming input words into real vectors are the key components of deep neural networks used in natural language processing.

Language Modelling Machine Translation +2

PROVEN: Certifying Robustness of Neural Networks with a Probabilistic Approach

no code implementations18 Dec 2018 Tsui-Wei Weng, Pin-Yu Chen, Lam M. Nguyen, Mark S. Squillante, Ivan Oseledets, Luca Daniel

With deep neural networks providing state-of-the-art machine learning models for numerous machine learning tasks, quantifying the robustness of these models has become an important area of research.

BIG-bench Machine Learning

Adversarial point set registration

no code implementations20 Nov 2018 Sergei Divakov, Ivan Oseledets

We present a novel approach to point set registration which is based on one-shot adversarial learning.

Modelling hidden structure of signals in group data analysis with modified (Lr, 1) and block-term decompositions

no code implementations7 Aug 2018 Pavel Kharyuk, Ivan Oseledets

This work is devoted to elaboration on the idea to use block term decomposition for group data analysis and to raise the possibility of modelling group activity with (Lr, 1) and Tucker blocks.

Clustering General Classification +1

Revealing the Unobserved by Linking Collaborative Behavior and Side Knowledge

no code implementations27 Jul 2018 Evgeny Frolov, Ivan Oseledets

We propose a tensor-based model that fuses a more granular representation of user preferences with the ability to take additional side information into account.

Comparative study of Discrete Wavelet Transforms and Wavelet Tensor Train decomposition to feature extraction of FTIR data of medicinal plants

no code implementations18 Jul 2018 Pavel Kharyuk, Dmitry Nazarenko, Ivan Oseledets

Fourier-transform infra-red (FTIR) spectra of samples from 7 plant species were used to explore the influence of preprocessing and feature extraction on efficiency of machine learning algorithms.

Clustering General Classification

HybridSVD: When Collaborative Information is Not Enough

3 code implementations18 Feb 2018 Evgeny Frolov, Ivan Oseledets

We propose a new hybrid algorithm that allows incorporating both user and item side information within the standard collaborative filtering technique.

Collaborative Filtering Model Selection

Geometry Score: A Method For Comparing Generative Adversarial Networks

1 code implementation ICML 2018 Valentin Khrulkov, Ivan Oseledets

One of the biggest challenges in the research of generative adversarial networks (GANs) is assessing the quality of generated samples and detecting various levels of mode collapse.

Tensor Train decomposition on TensorFlow (T3F)

2 code implementations5 Jan 2018 Alexander Novikov, Pavel Izmailov, Valentin Khrulkov, Michael Figurnov, Ivan Oseledets

Tensor Train decomposition is used across many branches of machine learning.

Mathematical Software Numerical Analysis

Deep Multigrid: learning prolongation and restriction matrices

1 code implementation10 Nov 2017 Alexandr Katrutsa, Talgat Daulbaev, Ivan Oseledets

This paper proposes the method to optimize restriction and prolongation operators in the two-grid method.

Numerical Analysis

Expressive power of recurrent neural networks

2 code implementations ICLR 2018 Valentin Khrulkov, Alexander Novikov, Ivan Oseledets

In this paper, we prove the expressive power theorem (an exponential lower bound on the width of the equivalent shallow network) for a class of recurrent neural networks -- ones that correspond to the Tensor Train (TT) decomposition.

Tensor Decomposition

Neural networks for topology optimization

2 code implementations27 Sep 2017 Ivan Sosnovik, Ivan Oseledets

The main novelty of this work is to state the problem as an image segmentation task.

Image Segmentation Semantic Segmentation

Art of singular vectors and universal adversarial perturbations

no code implementations CVPR 2018 Valentin Khrulkov, Ivan Oseledets

Vulnerability of Deep Neural Networks (DNNs) to adversarial attacks has been attracting a lot of attention in recent studies.

Image Classification

Riemannian Optimization for Skip-Gram Negative Sampling

1 code implementation ACL 2017 Alexander Fonarev, Oleksii Hrinchuk, Gleb Gusev, Pavel Serdyukov, Ivan Oseledets

Skip-Gram Negative Sampling (SGNS) word embedding model, well known by its implementation in "word2vec" software, is usually optimized by stochastic gradient descent.

Riemannian optimization

Efficient Rectangular Maximal-Volume Algorithm for Rating Elicitation in Collaborative Filtering

no code implementations16 Oct 2016 Alexander Fonarev, Alexander Mikhalev, Pavel Serdyukov, Gleb Gusev, Ivan Oseledets

Cold start problem in Collaborative Filtering can be solved by asking new users to rate a small seed set of representative items or by asking representative users to rate a new item.

Collaborative Filtering

Fifty Shades of Ratings: How to Benefit from a Negative Feedback in Top-N Recommendations Tasks

2 code implementations14 Jul 2016 Evgeny Frolov, Ivan Oseledets

In order to resolve this problem we propose to treat user feedback as a categorical variable and model it with users and items in a ternary way.

Collaborative Filtering Recommendation Systems

Exponential Machines

3 code implementations12 May 2016 Alexander Novikov, Mikhail Trofimov, Ivan Oseledets

Modeling interactions between features improves the performance of machine learning solutions in many domains (e. g. recommender systems or sentiment analysis).

Recommendation Systems Riemannian optimization +1

Tensor Methods and Recommender Systems

no code implementations19 Mar 2016 Evgeny Frolov, Ivan Oseledets

A substantial progress in development of new and efficient tensor factorization techniques has led to an extensive research of their applicability in recommender systems field.

Collaborative Filtering Recommendation Systems

Tensor SimRank for Heterogeneous Information Networks

no code implementations24 Feb 2015 Ben Usman, Ivan Oseledets

We propose a generalization of SimRank similarity measure for heterogeneous information networks.

Speeding-up Convolutional Neural Networks Using Fine-tuned CP-Decomposition

10 code implementations19 Dec 2014 Vadim Lebedev, Yaroslav Ganin, Maksim Rakhuba, Ivan Oseledets, Victor Lempitsky

We propose a simple two-step approach for speeding up convolution layers within large convolutional neural networks based on tensor decomposition and discriminative fine-tuning.

General Classification Tensor Decomposition

Unifying time evolution and optimization with matrix product states

3 code implementations21 Aug 2014 Jutho Haegeman, Christian Lubich, Ivan Oseledets, Bart Vandereycken, Frank Verstraete

In particular, our method is compatible with any Hamiltonian for which DMRG can be implemented efficiently and DMRG is obtained as a special case of imaginary time evolution with infinite time step.

Quantum Physics Strongly Correlated Electrons

A projector-splitting integrator for dynamical low-rank approximation

1 code implementation6 Jan 2013 Christian Lubich, Ivan Oseledets

The dynamical low-rank approximation of time-dependent matrices is a low-rank factorization updating technique.

Numerical Analysis 65F30, 65L05, 65L20, 15A23

Cannot find the paper you are looking for? You can Submit a new open access paper.