1 code implementation • NeurIPS 2023 • Leonard Papenmeier, Luigi Nardi, Matthias Poloczek
Impactful applications such as materials discovery, hardware design, neural architecture search, or portfolio optimization require optimizing high-dimensional black-box functions with mixed and combinatorial input spaces.
2 code implementations • 22 Apr 2023 • Leonard Papenmeier, Luigi Nardi, Matthias Poloczek
Recent advances have extended the scope of Bayesian optimization (BO) to expensive-to-evaluate black-box functions with dozens of dimensions, aspiring to unlock impactful applications, for example, in the life sciences, neural architecture search, and robotics.
no code implementations • 20 Feb 2020 • David Eriksson, Matthias Poloczek
The global optimization of a high-dimensional black-box function under black-box constraints is a pervasive task in machine learning, control, and engineering.
no code implementations • 21 Oct 2019 • Michael Pearce, Matthias Poloczek, Juergen Branke
Bayesian optimization is a powerful tool for expensive stochastic black-box optimization problems such as simulation-based optimization or machine learning hyperparameter tuning.
1 code implementation • 21 Oct 2019 • Yijia Wang, Matthias Poloczek, Daniel R. Jiang
Reinforcement learning in sparse-reward navigation environments with expensive and limited interactions is challenging and poses a need for effective exploration.
2 code implementations • NeurIPS 2019 • David Eriksson, Michael Pearce, Jacob R. Gardner, Ryan Turner, Matthias Poloczek
This motivates the design of a local probabilistic approach for global optimization of large-scale high-dimensional problems.
2 code implementations • ICML 2018 • Ricardo Baptista, Matthias Poloczek
The optimization of expensive-to-evaluate black-box functions over combinatorial structures is an ubiquitous task in machine learning, engineering and the natural sciences.
1 code implementation • NeurIPS 2017 • Jian Wu, Matthias Poloczek, Andrew Gordon Wilson, Peter I. Frazier
Bayesian optimization has been successful at global optimization of expensive-to-evaluate multimodal objective functions.
no code implementations • 11 Aug 2016 • Matthias Poloczek, Jialei Wang, Peter I. Frazier
We develop a framework for warm-starting Bayesian optimization, that reduces the solution time required to solve an optimization problem that is one in a sequence of related problems.
no code implementations • NeurIPS 2017 • Matthias Poloczek, Jialei Wang, Peter I. Frazier
We consider Bayesian optimization of an expensive-to-evaluate black-box objective function, where we also have access to cheaper approximations of the objective.