Search Results for author: Alexander Tornede

Found 12 papers, 8 papers with code

Interactive Hyperparameter Optimization in Multi-Objective Problems via Preference Learning

1 code implementation7 Sep 2023 Joseph Giovanelli, Alexander Tornede, Tanja Tornede, Marius Lindauer

In an experimental study targeting the environmental impact of ML, we demonstrate that our approach leads to substantially better Pareto fronts compared to optimizing based on a wrong indicator pre-selected by the user, and performs comparable in the case of an advanced user knowing which indicator to pick.

Hyperparameter Optimization

PyExperimenter: Easily distribute experiments and track results

1 code implementation16 Jan 2023 Tanja Tornede, Alexander Tornede, Lukas Fehring, Lukas Gehring, Helena Graf, Jonas Hanselle, Felix Mohr, Marcel Wever

PyExperimenter is a tool to facilitate the setup, documentation, execution, and subsequent evaluation of results from an empirical study of algorithms and in particular is designed to reduce the involved manual effort significantly.

HARRIS: Hybrid Ranking and Regression Forests for Algorithm Selection

1 code implementation31 Oct 2022 Lukas Fehring, Jonas Hanselle, Alexander Tornede

It is well known that different algorithms perform differently well on an instance of an algorithmic problem, motivating algorithm selection (AS): Given an instance of an algorithmic problem, which is the most suitable algorithm to solve it?

regression

A Survey of Methods for Automated Algorithm Configuration

no code implementations3 Feb 2022 Elias Schede, Jasmin Brandt, Alexander Tornede, Marcel Wever, Viktor Bengs, Eyke Hüllermeier, Kevin Tierney

We review existing AC literature within the lens of our taxonomies, outline relevant design choices of configuration approaches, contrast methods and problem variants against each other, and describe the state of AC in industry.

Machine Learning for Online Algorithm Selection under Censored Feedback

1 code implementation13 Sep 2021 Alexander Tornede, Viktor Bengs, Eyke Hüllermeier

In online algorithm selection (OAS), instances of an algorithmic problem class are presented to an agent one after another, and the agent has to quickly select a presumably best algorithm from a fixed set of candidate algorithms.

BIG-bench Machine Learning Thompson Sampling

Automated Machine Learning, Bounded Rationality, and Rational Metareasoning

no code implementations10 Sep 2021 Eyke Hüllermeier, Felix Mohr, Alexander Tornede, Marcel Wever

The notion of bounded rationality originated from the insight that perfectly rational behavior cannot be realized by agents with limited cognitive or computational resources.

AutoML BIG-bench Machine Learning

Algorithm Selection on a Meta Level

1 code implementation20 Jul 2021 Alexander Tornede, Lukas Gehring, Tanja Tornede, Marcel Wever, Eyke Hüllermeier

The problem of selecting an algorithm that appears most suitable for a specific instance of an algorithmic problem class, such as the Boolean satisfiability problem, is called instance-specific algorithm selection.

Ensemble Learning Meta-Learning

Towards Meta-Algorithm Selection

1 code implementation17 Nov 2020 Alexander Tornede, Marcel Wever, Eyke Hüllermeier

Instance-specific algorithm selection (AS) deals with the automatic selection of an algorithm from a fixed set of candidates most suitable for a specific instance of an algorithmic problem class, where "suitability" often refers to an algorithm's runtime.

Run2Survive: A Decision-theoretic Approach to Algorithm Selection based on Survival Analysis

1 code implementation6 Jul 2020 Alexander Tornede, Marcel Wever, Stefan Werner, Felix Mohr, Eyke Hüllermeier

In an extensive experimental study with the standard benchmark ASlib, our approach is shown to be highly competitive and in many cases even superior to state-of-the-art AS approaches.

Survival Analysis

Extreme Algorithm Selection With Dyadic Feature Representation

1 code implementation29 Jan 2020 Alexander Tornede, Marcel Wever, Eyke Hüllermeier

Algorithm selection (AS) deals with selecting an algorithm from a fixed set of candidate algorithms most suitable for a specific instance of an algorithmic problem, e. g., choosing solvers for SAT problems.

Hyperparameter Optimization Meta-Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.