no code implementations • 14 Apr 2024 • Branislav Bosansky, Lada Hospodkova, Michal Najman, Maria Rigaki, Elnaz Babayeva, Viliam Lisy
We use GANs to learn changes in data distributions within different time periods of training data and then apply these changes to generate samples that could be in testing data.
1 code implementation • 6 Sep 2022 • Branislav Bosansky, Dominik Kouba, Ondrej Manhal, Thorsten Sick, Viliam Lisy, Jakub Kroustek, Petr Somol
The benefit of using dynamic sandboxes is the realistic simulation of file execution in the target machine and obtaining a log of such execution.
no code implementations • 22 Oct 2021 • Marek Galovic, Branislav Bosansky, Viliam Lisy
In malware behavioral analysis, the list of accessed and created files very often indicates whether the examined file is malicious or benign.
4 code implementations • 19 May 2021 • Simon Mandlik, Matej Racinsky, Viliam Lisy, Tomas Pevny
Learning from raw data input, thus limiting the need for manual feature engineering, is one of the key components of many successful applications of machine learning methods.
2 code implementations • 27 Feb 2021 • Michael Walton, Viliam Lisy
In this report, we present results reproductions for several core algorithms implemented in the OpenSpiel framework for learning in games.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 22 Apr 2020 • Olga Petrova, Karel Durkota, Galina Alperovich, Karel Horak, Michal Najman, Branislav Bosansky, Viliam Lisy
Experiments show that both algorithms are applicable for cases with low feature space dimensions but the learning-based method produces less exploitable strategies and it is scalable to higher dimensions.
no code implementations • 22 Dec 2016 • Viliam Lisy, Michael Bowling
Approximating a Nash equilibrium is currently the best performing approach for creating poker-playing programs.
Computer Science and Game Theory
no code implementations • NeurIPS 2013 • Viliam Lisy, Vojta Kovarik, Marc Lanctot, Branislav Bosansky
In this paper, we study Monte Carlo tree search (MCTS) in zero-sum extensive-form games with perfect information and simultaneous moves.