1 code implementation • 30 Jul 2023 • Aleksey Romanov, Oleg Lashinin, Marina Ananyeva, Sergey Kolesnikov
In addition, we show the results of an ablation study and a case study of a few items.
no code implementations • 25 Jun 2023 • Veronika Ivanova, Oleg Lashinin, Marina Ananyeva, Sergey Kolesnikov
To solve this problem, we have collected and published a dataset containing information about the recommender models used in 903 papers, both as baselines and as proposed approaches.
1 code implementation • 14 Jun 2023 • Vladislav Kurenkov, Alexander Nikulin, Denis Tarasov, Sergey Kolesnikov
NetHack is known as the frontier of reinforcement learning research where learning-based methods still need to catch up to rule-based solutions.
no code implementations • 19 May 2023 • Stanislav Dereka, Ivan Karpukhin, Maksim Zhdanov, Sergey Kolesnikov
Deep ensembles are capable of achieving state-of-the-art results in classification and out-of-distribution (OOD) detection.
3 code implementations • 16 May 2023 • Denis Tarasov, Vladislav Kurenkov, Alexander Nikulin, Sergey Kolesnikov
In this work, we aim to bridge this gap by conducting a retrospective analysis of recent works in offline RL and propose ReBRAC, a minimalistic algorithm that integrates such design elements built on top of the TD3+BC method.
2 code implementations • 31 Jan 2023 • Alexander Nikulin, Vladislav Kurenkov, Denis Tarasov, Sergey Kolesnikov
Despite the success of Random Network Distillation (RND) in various domains, it was shown as not discriminative enough to be used as an uncertainty estimator for penalizing out-of-distribution actions in offline reinforcement learning.
2 code implementations • 20 Nov 2022 • Dmitriy Akimov, Vladislav Kurenkov, Alexander Nikulin, Denis Tarasov, Sergey Kolesnikov
This Normalizing Flows action encoder is pre-trained in a supervised manner on the offline dataset, and then an additional policy model - controller in the latent space - is trained via reinforcement learning.
2 code implementations • 20 Nov 2022 • Alexander Nikulin, Vladislav Kurenkov, Denis Tarasov, Dmitry Akimov, Sergey Kolesnikov
Training large neural networks is known to be time-consuming, with the learning duration taking days or even weeks.
2 code implementations • 13 Oct 2022 • Denis Tarasov, Alexander Nikulin, Dmitry Akimov, Vladislav Kurenkov, Sergey Kolesnikov
CORL is an open-source library that provides thoroughly benchmarked single-file implementations of both deep offline and offline-to-online reinforcement learning algorithms.
no code implementations • 23 May 2022 • Stanislav Dereka, Ivan Karpukhin, Sergey Kolesnikov
Large-scale datasets are essential for the success of deep learning in image retrieval.
1 code implementation • 19 May 2022 • Ivan Karpukhin, Stanislav Dereka, Sergey Kolesnikov
Classification tasks are usually evaluated in terms of accuracy.
Ranked #2 on
Image Classification
on SVHN
(Percentage correct metric)
no code implementations • 11 May 2022 • Mikhail Andronov, Sergey Kolesnikov
The evaluation of recommender systems from a practical perspective is a topic of ongoing discourse within the research community.
1 code implementation • 14 Feb 2022 • Ivan Karpukhin, Stanislav Dereka, Sergey Kolesnikov
We thus provide a new confidence evaluation benchmark and establish a baseline for future confidence prediction research.
no code implementations • 11 Oct 2021 • Sergey Kolesnikov, Oleg Lashinin, Michail Pechatov, Alexander Kosov
In this article, we aim to fill the gap in RecSys methods evaluation on the NPR task using publicly available datasets and (1) introduce the TTRS, a large-scale financial transactions dataset suitable for RecSys methods evaluation; (2) benchmark popular RecSys approaches on several datasets for the NPR task.
no code implementations • 8 Oct 2021 • Vladislav Kurenkov, Sergey Kolesnikov
In this work, we argue for the importance of an online evaluation budget for a reliable comparison of deep offline RL algorithms.
no code implementations • 14 Sep 2021 • Evgeniy Egorov, Vasily Kostyumov, Mikhail Konyk, Sergey Kolesnikov
Lipreading, also known as visual speech recognition, aims to identify the speech content from videos by analyzing the visual deformations of lips and nearby areas.
2 code implementations • 29 Mar 2020 • Sergey Kolesnikov, Valentin Khrulkov
We present Catalyst. RL, an open-source PyTorch framework for reproducible and sample efficient reinforcement learning (RL) research.
1 code implementation • 28 Feb 2019 • Sergey Kolesnikov, Oleksii Hrinchuk
Despite the recent progress in deep reinforcement learning field (RL), and, arguably because of it, a large body of work remains to be done in reproducing and carefully comparing different RL algorithms.
1 code implementation • 7 Feb 2019 • Łukasz Kidziński, Carmichael Ong, Sharada Prasanna Mohanty, Jennifer Hicks, Sean F. Carroll, Bo Zhou, Hongsheng Zeng, Fan Wang, Rongzhong Lian, Hao Tian, Wojciech Jaśkowski, Garrett Andersen, Odd Rune Lykkebø, Nihat Engin Toklu, Pranav Shyam, Rupesh Kumar Srivastava, Sergey Kolesnikov, Oleksii Hrinchuk, Anton Pechenko, Mattias Ljungström, Zhen Wang, Xu Hu, Zehong Hu, Minghui Qiu, Jun Huang, Aleksei Shpilman, Ivan Sosin, Oleg Svidchenko, Aleksandra Malysheva, Daniel Kudenko, Lance Rane, Aditya Bhatt, Zhengfei Wang, Penghui Qi, Zeyang Yu, Peng Peng, Quan Yuan, Wenxin Li, Yunsheng Tian, Ruihan Yang, Pingchuan Ma, Shauharda Khadka, Somdeb Majumdar, Zach Dwiel, Yinyin Liu, Evren Tumer, Jeremy Watson, Marcel Salathé, Sergey Levine, Scott Delp
In the NeurIPS 2018 Artificial Intelligence for Prosthetics challenge, participants were tasked with building a controller for a musculoskeletal model with a goal of matching a given time-varying velocity vector.
2 code implementations • 2 Apr 2018 • Łukasz Kidziński, Sharada Prasanna Mohanty, Carmichael Ong, Zhewei Huang, Shuchang Zhou, Anton Pechenko, Adam Stelmaszczyk, Piotr Jarosik, Mikhail Pavlov, Sergey Kolesnikov, Sergey Plis, Zhibo Chen, Zhizheng Zhang, Jiale Chen, Jun Shi, Zhuobin Zheng, Chun Yuan, Zhihui Lin, Henryk Michalewski, Piotr Miłoś, Błażej Osiński, Andrew Melnik, Malte Schilling, Helge Ritter, Sean Carroll, Jennifer Hicks, Sergey Levine, Marcel Salathé, Scott Delp
In the NIPS 2017 Learning to Run challenge, participants were tasked with building a controller for a musculoskeletal model to make it run as fast as possible through an obstacle course.
1 code implementation • 18 Nov 2017 • Mikhail Pavlov, Sergey Kolesnikov, Sergey M. Plis
In this paper, we present our approach to solve a physics-based reinforcement learning challenge "Learning to Run" with objective to train physiologically-based human model to navigate a complex obstacle course as quickly as possible.