Search Results for author: Mateusz Ostaszewski

Found 14 papers, 3 papers with code

A Case for Validation Buffer in Pessimistic Actor-Critic

no code implementations1 Mar 2024 Michal Nauman, Mateusz Ostaszewski, Marek Cygan

VPL uses a small validation buffer to adjust the levels of pessimism throughout the agent training, with the pessimism set such that the approximation error of the critic targets is minimized.

Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem

no code implementations5 Feb 2024 Maciej Wołczyk, Bartłomiej Cupiał, Mateusz Ostaszewski, Michał Bortkiewicz, Michał Zając, Razvan Pascanu, Łukasz Kuciński, Piotr Miłoś

Fine-tuning is a widespread technique that allows practitioners to transfer pre-trained capabilities, as recently showcased by the successful applications of foundation models.

Montezuma's Revenge NetHack +2

Curriculum reinforcement learning for quantum architecture search under hardware errors

no code implementations5 Feb 2024 Yash J. Patel, Akash Kundu, Mateusz Ostaszewski, Xavier Bonet-Monroig, Vedran Dunjko, Onur Danaci

In the case of parameter optimization alone, noise effects have been observed to dramatically influence the performance of the optimizer and final outcomes, which is a key line of study.

3D Architecture Computational Efficiency +2

Enhancing variational quantum state diagonalization using reinforcement learning techniques

1 code implementation19 Jun 2023 Akash Kundu, Przemysław Bedełek, Mateusz Ostaszewski, Onur Danaci, Yash J. Patel, Vedran Dunjko, Jarosław A. Miszczak

We demonstrate that the circuits proposed by the reinforcement learning methods are shallower than the standard variational quantum state diagonalization algorithm and thus can be used in situations where hardware capabilities limit the depth of quantum circuits.

Quantum Machine Learning reinforcement-learning +1

Emergency action termination for immediate reaction in hierarchical reinforcement learning

no code implementations11 Nov 2022 Michał Bortkiewicz, Jakub Łyskawa, Paweł Wawrzyński, Mateusz Ostaszewski, Artur Grudkowski, Tomasz Trzciński

In this paper, we address this gap in the state-of-the-art approaches and propose a method in which the validity of higher-level actions (thus lower-level goals) is constantly verified at the higher level.

Hierarchical Reinforcement Learning reinforcement-learning +1

Reinforcement learning with experience replay and adaptation of action dispersion

no code implementations30 Jul 2022 Paweł Wawrzyński, Wojciech Masarczyk, Mateusz Ostaszewski

To that end, the dispersion should be tuned to assure a sufficiently high probability (densities) of the actions in the replay buffer and the modes of the distributions that generated them, yet this dispersion should not be higher.

reinforcement-learning Reinforcement Learning (RL)

Reinforcement learning for optimization of variational quantum circuit architectures

no code implementations NeurIPS 2021 Mateusz Ostaszewski, Lea M. Trenkwalder, Wojciech Masarczyk, Eleanor Scerri, Vedran Dunjko

The study of Variational Quantum Eigensolvers (VQEs) has been in the spotlight in recent times as they may lead to real-world applications of near-term quantum devices.

reinforcement-learning Reinforcement Learning (RL)

Structure optimization for parameterized quantum circuits

no code implementations23 May 2019 Mateusz Ostaszewski, Edward Grant, Marcello Benedetti

We demonstrate the method for optimizing a variational quantum eigensolver for finding the ground states of Lithium Hydride and the Heisenberg model in simulation, and for finding the ground state of Hydrogen gas on the IBM Melbourne quantum computer.

Quantum Physics

QSWalk.jl: Julia package for quantum stochastic walks analysis

6 code implementations4 Jan 2018 Adam Glos, Jarosław Adam Miszczak, Mateusz Ostaszewski

The presented paper describes QSWalk. jl package for Julia programming language, developed for the purpose of simulating the evolution of open quantum systems.

Quantum Physics

Cannot find the paper you are looking for? You can Submit a new open access paper.