Search Results for author: Aleksandr Mikhalev

Found 4 papers, 1 papers with code

LoTR: Low Tensor Rank Weight Adaptation

no code implementations2 Feb 2024 Daniel Bershatsky, Daria Cherniuk, Talgat Daulbaev, Aleksandr Mikhalev, Ivan Oseledets

In this paper we generalize and extend an idea of low-rank adaptation (LoRA) of large language models (LLMs) based on Transformer architecture.

Tensor Decomposition

Run LoRA Run: Faster and Lighter LoRA Implementations

no code implementations6 Dec 2023 Daria Cherniuk, Aleksandr Mikhalev, Ivan Oseledets

LoRA is a technique that reduces the number of trainable parameters in a neural network by introducing low-rank adapters to linear layers.

Llama

Memory-Efficient Backpropagation through Large Linear Layers

2 code implementations31 Jan 2022 Daniel Bershatsky, Aleksandr Mikhalev, Alexandr Katrutsa, Julia Gusak, Daniil Merkulov, Ivan Oseledets

Also, we investigate the variance of the gradient estimate induced by the randomized matrix multiplication.

Model Compression

Cannot find the paper you are looking for? You can Submit a new open access paper.