Hyperparameter Optimization

279 papers with code • 1 benchmarks • 3 datasets

Hyperparameter Optimization is the problem of choosing a set of optimal hyperparameters for a learning algorithm. Whether the algorithm is suitable for the data directly depends on hyperparameters, which directly influence overfitting or underfitting. Each model requires different assumptions, weights or training speeds for different types of data under the conditions of a given loss function.

Source: Data-driven model for fracturing design optimization: focus on building digital database and production forecast

Libraries

Use these libraries to find Hyperparameter Optimization models and implementations
3 papers
7,415
3 papers
1,428
See all 14 libraries.

HomOpt: A Homotopy-Based Hyperparameter Optimization Method

jeffkinnison/shadho 7 Aug 2023

Traditional methods, like grid search and Bayesian optimization, often struggle to quickly adapt and efficiently search the loss landscape.

19
07 Aug 2023

Is One Epoch All You Need For Multi-Fidelity Hyperparameter Optimization?

deephyper/benchmark 28 Jul 2023

Hyperparameter optimization (HPO) is crucial for fine-tuning machine learning models but can be computationally expensive.

3
28 Jul 2023

Shrink-Perturb Improves Architecture Mixing during Population Based Training for Neural Architecture Search

awesomelemon/pbt-nas 28 Jul 2023

In this work, we show that simultaneously training and mixing neural networks is a promising way to conduct Neural Architecture Search (NAS).

3
28 Jul 2023

Multi-Objective Optimization of Performance and Interpretability of Tabular Supervised Machine Learning Models

slds-lmu/paper_2023_eagga 17 Jul 2023

Efficient optimization is achieved via augmentation of the search space of the learning algorithm by incorporating feature selection, interaction and monotonicity constraints into the hyperparameter search space.

1
17 Jul 2023

PriorBand: Practical Hyperparameter Optimization in the Age of Deep Learning

automl/neps NeurIPS 2023

Hyperparameters of Deep Learning (DL) pipelines are crucial for their downstream performance.

34
21 Jun 2023

Does Long-Term Series Forecasting Need Complex Attention and Extra Long Inputs?

anoise/periodformer 8 Jun 2023

MABO allocates a process to each GPU via a queue mechanism, and then creates multiple trials at a time for asynchronous parallel search, which greatly reduces the search time.

11
08 Jun 2023

Improving Hyperparameter Learning under Approximate Inference in Gaussian Process Models

aaltoml/improved-hyperparameter-learning 7 Jun 2023

Approximate inference in Gaussian process (GP) models with non-conjugate likelihoods gets entangled with the learning of the model hyperparameters.

0
07 Jun 2023

Quick-Tune: Quickly Learning Which Pretrained Model to Finetune and How

releaunifreiburg/quicktune 6 Jun 2023

With the ever-increasing number of pretrained models, machine learning practitioners are continuously faced with which pretrained model to use, and how to finetune it for a new dataset.

17
06 Jun 2023

Stochastic Marginal Likelihood Gradients using Neural Tangent Kernels

aleximmer/ntk-marglik 6 Jun 2023

Recent works show that Bayesian model selection with Laplace approximations can allow to optimize such hyperparameters just like standard neural network parameters using gradients and on the training data.

5
06 Jun 2023