Hyperparameter Optimization

277 papers with code • 1 benchmarks • 3 datasets

Hyperparameter Optimization is the problem of choosing a set of optimal hyperparameters for a learning algorithm. Whether the algorithm is suitable for the data directly depends on hyperparameters, which directly influence overfitting or underfitting. Each model requires different assumptions, weights or training speeds for different types of data under the conditions of a given loss function.

Source: Data-driven model for fracturing design optimization: focus on building digital database and production forecast

Libraries

Use these libraries to find Hyperparameter Optimization models and implementations
3 papers
7,373
3 papers
1,416
See all 14 libraries.

Latest papers with no code

Universal Link Predictor By In-Context Learning on Graphs

no code yet • 12 Feb 2024

In this work, we introduce the Universal Link Predictor (UniLP), a novel model that combines the generalizability of heuristic approaches with the pattern learning capabilities of parametric models.

Poisson Process for Bayesian Optimization

no code yet • 5 Feb 2024

BayesianOptimization(BO) is a sample-efficient black-box optimizer, and extensive methods have been proposed to build the absolute function response of the black-box function through a probabilistic surrogate model, including Tree-structured Parzen Estimator (TPE), random forest (SMAC), and Gaussian process (GP).

Glocal Hypergradient Estimation with Koopman Operator

no code yet • 5 Feb 2024

Through numerical experiments of hyperparameter optimization, including optimization of optimizers, we demonstrate the effectiveness of the glocal hypergradient estimation.

Breaking MLPerf Training: A Case Study on Optimizing BERT

no code yet • 4 Feb 2024

Speeding up the large-scale distributed training is challenging in that it requires improving various components of training including load balancing, communication, optimizers, etc.

Regularized boosting with an increasing coefficient magnitude stop criterion as meta-learner in hyperparameter optimization stacking ensemble

no code yet • 2 Feb 2024

This paper explores meta-learners for stacking ensemble in HPO, free of hyperparameter tuning, able to reduce the effects of multicollinearity and considering the ensemble learning process generalization power.

Large Language Model Agent for Hyper-Parameter Optimization

no code yet • 2 Feb 2024

Hyperparameter optimization is critical in modern machine learning, requiring expert knowledge, numerous trials, and high computational and human resources.

Scilab-RL: A software framework for efficient reinforcement learning and cognitive modeling research

no code yet • 25 Jan 2024

One problem with researching cognitive modeling and reinforcement learning (RL) is that researchers spend too much time on setting up an appropriate computational framework for their experiments.

A Unified Gaussian Process for Branching and Nested Hyperparameter Optimization

no code yet • 19 Jan 2024

To capture the conditional dependence between branching and nested parameters, a unified Bayesian optimization framework is proposed.

Adaptive Regret for Bandits Made Possible: Two Queries Suffice

no code yet • 17 Jan 2024

In this paper, we give query and regret optimal bandit algorithms under the strict notion of strongly adaptive regret, which measures the maximum regret over any contiguous interval $I$.

Hypercomplex neural network in time series forecasting of stock data

no code yet • 9 Jan 2024

The goal of this paper is to test three classes of neural network (NN) architectures based on four-dimensional (4D) hypercomplex algebras for time series prediction.