Search Results for author: Piotr Miłoś

Found 37 papers, 22 papers with code

Beyond Recognition: Evaluating Visual Perspective Taking in Vision Language Models

no code implementations3 May 2025 Gracjan Góral, Alicja Ziarko, Piotr Miłoś, Michał Nauman, Maciej Wołczyk, Michał Kosiński

We investigate the ability of Vision Language Models (VLMs) to perform visual perspective taking using a novel set of visual tasks inspired by established human tests.

Diagnostic Object Recognition +2

Joint MoE Scaling Laws: Mixture of Experts Can Be Memory Efficient

no code implementations7 Feb 2025 Jan Ludziejewski, Maciej Pióro, Jakub Krajewski, Maciej Stefaniak, Michał Krutul, Jan Małaśnicki, Marek Cygan, Piotr Sankowski, Kamil Adamczewski, Piotr Miłoś, Sebastian Jaszczur

Mixture of Experts (MoE) architectures have significantly increased computational efficiency in both research and real-world applications of large-scale machine learning models.

Computational Efficiency Mixture-of-Experts

RapidDock: Unlocking Proteome-scale Molecular Docking

no code implementations16 Oct 2024 Rafał Powalski, Bazyli Klockiewicz, Maciej Jaśkowski, Bartosz Topolski, Paweł Dąbrowski-Tumański, Maciej Wiśniewski, Łukasz Kuciński, Piotr Miłoś, Dariusz Plewczynski

Accelerating molecular docking -- the process of predicting how molecules bind to protein targets -- could boost small-molecule drug discovery and revolutionize medicine.

Drug Discovery Molecular Docking +1

Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe

1 code implementation6 Jun 2024 Alicja Ziarko, Albert Q. Jiang, Bartosz Piotrowski, Wenda Li, Mateja Jamnik, Piotr Miłoś

Text embeddings are essential for many tasks, such as document retrieval, clustering, and semantic similarity assessment.

Decoder Retrieval +2

What Matters in Hierarchical Search for Combinatorial Reasoning Problems?

1 code implementation5 Jun 2024 Michał Zawalski, Gracjan Góral, Michał Tyrolski, Emilia Wiśnios, Franciszek Budrowski, Łukasz Kuciński, Piotr Miłoś

Efficiently tackling combinatorial reasoning problems, particularly the notorious NP-hard tasks, remains a significant challenge for AI research.

tsGT: Stochastic Time Series Modeling With Transformer

no code implementations8 Mar 2024 Łukasz Kuciński, Witold Drzewakowski, Mateusz Olko, Piotr Kozakowski, Łukasz Maziarka, Marta Emilia Nowakowska, Łukasz Kaiser, Piotr Miłoś

Time series methods are of fundamental importance in virtually any field of science that deals with temporally structured data.

Time Series

Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning

no code implementations1 Mar 2024 Michal Nauman, Michał Bortkiewicz, Piotr Miłoś, Tomasz Trzciński, Mateusz Ostaszewski, Marek Cygan

We tested these agents across 14 diverse tasks from 2 simulation benchmarks, measuring training metrics related to overestimation, overfitting, and plasticity loss -- issues that motivate the examined regularization techniques.

Reinforcement Learning (RL)

Analysing The Impact of Sequence Composition on Language Model Pre-Training

1 code implementation21 Feb 2024 Yu Zhao, Yuanbin Qu, Konrad Staniszewski, Szymon Tworkowski, Wei Liu, Piotr Miłoś, Yuxiang Wu, Pasquale Minervini

In this work, we find that applying causal masking can lead to the inclusion of distracting information from previous documents during pre-training, which negatively impacts the performance of the models on language modelling and downstream tasks.

In-Context Learning Language Modeling +2

Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem

1 code implementation5 Feb 2024 Maciej Wołczyk, Bartłomiej Cupiał, Mateusz Ostaszewski, Michał Bortkiewicz, Michał Zając, Razvan Pascanu, Łukasz Kuciński, Piotr Miłoś

Fine-tuning is a widespread technique that allows practitioners to transfer pre-trained capabilities, as recently showcased by the successful applications of foundation models.

Montezuma's Revenge NetHack +2

Structured Packing in LLM Training Improves Long Context Utilization

no code implementations28 Dec 2023 Konrad Staniszewski, Szymon Tworkowski, Sebastian Jaszczur, Yu Zhao, Henryk Michalewski, Łukasz Kuciński, Piotr Miłoś

Recent advancements in long-context large language models have attracted significant attention, yet their practical applications often suffer from suboptimal context utilization.

Information Retrieval Retrieval

Exploring Continual Learning of Diffusion Models

no code implementations27 Mar 2023 Michał Zając, Kamil Deja, Anna Kuzina, Jakub M. Tomczak, Tomasz Trzciński, Florian Shkurti, Piotr Miłoś

Diffusion models have achieved remarkable success in generating high-quality images thanks to their novel training procedures applied to unprecedented amounts of data.

Benchmarking Continual Learning +1

Magnushammer: A Transformer-Based Approach to Premise Selection

no code implementations8 Mar 2023 Maciej Mikuła, Szymon Tworkowski, Szymon Antoniak, Bartosz Piotrowski, Albert Qiaochu Jiang, Jin Peng Zhou, Christian Szegedy, Łukasz Kuciński, Piotr Miłoś, Yuhuai Wu

By combining \method with a language-model-based automated theorem prover, we further improve the state-of-the-art proof success rate from $57. 0\%$ to $71. 0\%$ on the PISA benchmark using $4$x fewer parameters.

Automated Theorem Proving Language Modeling +2

Trust Your $\nabla$: Gradient-based Intervention Targeting for Causal Discovery

no code implementations NeurIPS 2023 Mateusz Olko, Michał Zając, Aleksandra Nowak, Nino Scherrer, Yashas Annadani, Stefan Bauer, Łukasz Kuciński, Piotr Miłoś

In this work, we propose a novel Gradient-based Intervention Targeting method, abbreviated GIT, that 'trusts' the gradient estimator of a gradient-based causal discovery framework to provide signals for the intervention acquisition function.

Causal Discovery Experimental Design

Disentangling Transfer in Continual Reinforcement Learning

no code implementations28 Sep 2022 Maciej Wołczyk, Michał Zając, Razvan Pascanu, Łukasz Kuciński, Piotr Miłoś

The ability of continual learning systems to transfer knowledge from previously seen tasks in order to maximize performance on new tasks is a significant challenge for the field, limiting the applicability of continual learning solutions to realistic scenarios.

Continual Learning continuous-control +4

Thor: Wielding Hammers to Integrate Language Models and Automated Theorem Provers

no code implementations22 May 2022 Albert Q. Jiang, Wenda Li, Szymon Tworkowski, Konrad Czechowski, Tomasz Odrzygóźdź, Piotr Miłoś, Yuhuai Wu, Mateja Jamnik

Thor increases a language model's success rate on the PISA dataset from $39\%$ to $57\%$, while solving $8. 2\%$ of problems neither language models nor automated theorem provers are able to solve on their own.

Automated Theorem Proving

Continual World: A Robotic Benchmark For Continual Reinforcement Learning

1 code implementation NeurIPS 2021 Maciej Wołczyk, Michał Zając, Razvan Pascanu, Łukasz Kuciński, Piotr Miłoś

Continual learning (CL) -- the ability to continuously learn, building on previously acquired knowledge -- is a natural requirement for long-lived autonomous reinforcement learning (RL) agents.

Continual Learning reinforcement-learning +2

Planning and Learning Using Adaptive Entropy Tree Search

1 code implementation12 Feb 2021 Piotr Kozakowski, Mikołaj Pacek, Piotr Miłoś

We present Adaptive Entropy Tree Search (ANTS) - a novel algorithm combining planning and learning in the maximum entropy paradigm.

Emergence of compositional language in communication through noisy channel

no code implementations ICML Workshop LaReL 2020 Łukasz Kuciński, Paweł Kołodziej, Piotr Miłoś

In this paper, we investigate how communication through a noisy channel can lead to the emergence of compositional language.

Uncertainty - sensitive learning and planning with ensembles

1 code implementation25 Sep 2019 Piotr Miłoś, Łukasz Kuciński, Konrad Czechowski, Piotr Kozakowski, Maciej Klimek

Notably, our method performs well in environments with sparse rewards where standard $TD(1)$ backups fail.

Montezuma's Revenge Sokoban

Expert-augmented actor-critic for ViZDoom and Montezumas Revenge

2 code implementations10 Sep 2018 Michał Garmulewicz, Henryk Michalewski, Piotr Miłoś

We propose an expert-augmented actor-critic algorithm, which we evaluate on two environments with sparse rewards: Montezumas Revenge and a demanding maze from the ViZDoom suite.

Cannot find the paper you are looking for? You can Submit a new open access paper.