1 code implementation • NeurIPS 2021 • Maciej Wołczyk, Michał Zając, Razvan Pascanu, Łukasz Kuciński, Piotr Miłoś
Continual learning (CL) -- the ability to continuously learn, building on previously acquired knowledge -- is a natural requirement for long-lived autonomous reinforcement learning (RL) agents.
1 code implementation • 1 Jun 2022 • Michał Zawalski, Michał Tyrolski, Konrad Czechowski, Tomasz Odrzygóźdź, Damian Stachura, Piotr Piękos, Yuhuai Wu, Łukasz Kuciński, Piotr Miłoś
Complex reasoning problems contain states that vary in the computational cost required to determine a good action plan.
1 code implementation • NeurIPS 2021 • Konrad Czechowski, Tomasz Odrzygóźdź, Marek Zbysiński, Michał Zawalski, Krzysztof Olejnik, Yuhuai Wu, Łukasz Kuciński, Piotr Miłoś
In this paper, we implement kSubS using a transformer-based subgoal module coupled with the classical best-first search framework.
1 code implementation • 30 Nov 2021 • Piotr Januszewski, Mateusz Olko, Michał Królikowski, Jakub Świątkowski, Marcin Andrychowicz, Łukasz Kuciński, Piotr Miłoś
The growth of deep reinforcement learning (RL) has brought multiple exciting tools and methods to the field.
1 code implementation • 6 Mar 2024 • Bartosz Cywiński, Kamil Deja, Tomasz Trzciński, Bartłomiej Twardowski, Łukasz Kuciński
We introduce GUIDE, a novel continual learning approach that directs diffusion models to rehearse samples at risk of being forgotten.
1 code implementation • 19 Dec 2019 • Piotr Miłoś, Łukasz Kuciński, Konrad Czechowski, Piotr Kozakowski, Maciek Klimek
The former manifests itself through the use of value function, while the latter is powered by a tree search planner.
1 code implementation • NeurIPS Workshop LMCA 2020 • Piotr Kozakowski, Piotr Januszewski, Konrad Czechowski, Łukasz Kuciński, Piotr Miłoś
Planning in large state spaces inevitably needs to balance depth and breadth of the search.
1 code implementation • 25 Sep 2019 • Piotr Miłoś, Łukasz Kuciński, Konrad Czechowski, Piotr Kozakowski, Maciej Klimek
Notably, our method performs well in environments with sparse rewards where standard $TD(1)$ backups fail.
1 code implementation • 4 Oct 2019 • Tomasz Korbak, Julian Zubek, Łukasz Kuciński, Piotr Miłoś, Joanna Rączaszek-Leonardi
This paper explores a novel approach to achieving emergent compositional communication in multi-agent systems.
1 code implementation • NeurIPS Workshop LMCA 2020 • Konrad Czechowski, Tomasz Odrzygóźdź, Michał Izworski, Marek Zbysiński, Łukasz Kuciński, Piotr Miłoś
We propose $\textit{trust-but-verify}$ (TBV) mechanism, a new method which uses model uncertainty estimates to guide exploration.
Model-based Reinforcement Learning reinforcement-learning +1
no code implementations • NeurIPS 2021 • Łukasz Kuciński, Tomasz Korbak, Paweł Kołodziej, Piotr Miłoś
Communication is compositional if complex signals can be represented as a combination of simpler subparts.
no code implementations • ICML Workshop LaReL 2020 • Łukasz Kuciński, Paweł Kołodziej, Piotr Miłoś
In this paper, we investigate how communication through a noisy channel can lead to the emergence of compositional language.
no code implementations • 28 Sep 2022 • Maciej Wołczyk, Michał Zając, Razvan Pascanu, Łukasz Kuciński, Piotr Miłoś
The ability of continual learning systems to transfer knowledge from previously seen tasks in order to maximize performance on new tasks is a significant challenge for the field, limiting the applicability of continual learning solutions to realistic scenarios.
no code implementations • NeurIPS 2023 • Mateusz Olko, Michał Zając, Aleksandra Nowak, Nino Scherrer, Yashas Annadani, Stefan Bauer, Łukasz Kuciński, Piotr Miłoś
In this work, we propose a novel Gradient-based Intervention Targeting method, abbreviated GIT, that 'trusts' the gradient estimator of a gradient-based causal discovery framework to provide signals for the intervention acquisition function.
no code implementations • 8 Mar 2023 • Maciej Mikuła, Szymon Tworkowski, Szymon Antoniak, Bartosz Piotrowski, Albert Qiaochu Jiang, Jin Peng Zhou, Christian Szegedy, Łukasz Kuciński, Piotr Miłoś, Yuhuai Wu
By combining \method with a language-model-based automated theorem prover, we further improve the state-of-the-art proof success rate from $57. 0\%$ to $71. 0\%$ on the PISA benchmark using $4$x fewer parameters.
no code implementations • 28 Dec 2023 • Konrad Staniszewski, Szymon Tworkowski, Yu Zhao, Sebastian Jaszczur, Henryk Michalewski, Łukasz Kuciński, Piotr Miłoś
Recent developments in long-context large language models have attracted considerable attention.
no code implementations • 5 Feb 2024 • Maciej Wołczyk, Bartłomiej Cupiał, Mateusz Ostaszewski, Michał Bortkiewicz, Michał Zając, Razvan Pascanu, Łukasz Kuciński, Piotr Miłoś
Fine-tuning is a widespread technique that allows practitioners to transfer pre-trained capabilities, as recently showcased by the successful applications of foundation models.
no code implementations • 8 Mar 2024 • Łukasz Kuciński, Witold Drzewakowski, Mateusz Olko, Piotr Kozakowski, Łukasz Maziarka, Marta Emilia Nowakowska, Łukasz Kaiser, Piotr Miłoś
Time series methods are of fundamental importance in virtually any field of science that deals with temporally structured data.