1 code implementation • 3 May 2024 • Aaron Klein, Jacek Golebiowski, Xingchen Ma, Valerio Perrone, Cedric Archambeau
Pre-trained language models (PLM), for example BERT or RoBERTa, mark the state-of-the-art for natural language understanding task when fine-tuned on labeled data.
no code implementations • 8 Dec 2023 • Lukas Balles, Cedric Archambeau, Giovanni Zappella
With increasing scale in model and dataset size, the training of deep neural networks becomes a massive computational burden.
1 code implementation • 5 May 2023 • David Salinas, Jacek Golebiowski, Aaron Klein, Matthias Seeger, Cedric Archambeau
Many state-of-the-art hyperparameter optimization (HPO) algorithms rely on model-based optimizers that learn surrogate models of the target function to guide the search.
1 code implementation • 24 Apr 2023 • Martin Wistuba, Martin Ferianc, Lukas Balles, Cedric Archambeau, Giovanni Zappella
We discuss requirements for the use of continual learning algorithms in practice, from which we derive design principles for Renate.
1 code implementation • 8 Feb 2023 • Gianluca Detommaso, Alberto Gasparin, Michele Donini, Matthias Seeger, Andrew Gordon Wilson, Cedric Archambeau
We present Fortuna, an open-source library for uncertainty quantification in deep learning.
2 code implementations • 15 Sep 2022 • Giuseppe Vietri, Cedric Archambeau, Sergul Aydore, William Brown, Michael Kearns, Aaron Roth, Ankit Siva, Shuai Tang, Zhiwei Steven Wu
A key innovation in our algorithm is the ability to directly handle numerical features, in contrast to a number of related prior approaches which require numerical features to be first converted into {high cardinality} categorical features via {a binning strategy}.
no code implementations • 17 Jul 2022 • Gianluca Detommaso, Alberto Gasparin, Andrew Wilson, Cedric Archambeau
As we move away from the data, the predictive uncertainty should increase, since a great variety of explanations are consistent with the little available information.
no code implementations • 28 Jun 2022 • Beyza Ermis, Giovanni Zappella, Martin Wistuba, Aditya Rawal, Cedric Archambeau
This phenomenon is known as catastrophic forgetting and it is often difficult to prevent due to practical constraints, such as the amount of data that can be stored or the limited computation sources that can be used.
no code implementations • 21 Mar 2022 • Deborah Sulem, Michele Donini, Muhammad Bilal Zafar, Francois-Xavier Aubet, Jan Gasthaus, Tim Januschowski, Sanjiv Das, Krishnaram Kenthapadi, Cedric Archambeau
In this work we propose a model-agnostic algorithm that generates counterfactual ensemble explanations for time series anomaly detection models.
no code implementations • 9 Mar 2022 • Beyza Ermis, Giovanni Zappella, Martin Wistuba, Aditya Rawal, Cedric Archambeau
Moreover, applications increasingly rely on large pre-trained neural networks, such as pre-trained Transformers, since the resources or data might not be available in sufficiently large quantities to practitioners to train the model from scratch.
no code implementations • 5 Nov 2021 • Riccardo Grazzi, Valentin Flunkert, David Salinas, Tim Januschowski, Matthias Seeger, Cedric Archambeau
While classical time series forecasting considers individual time series in isolation, recent advances based on deep learning showed that jointly learning from a large pool of related time series can boost the forecasting accuracy.
no code implementations • 10 Jun 2021 • David Salinas, Valerio Perrone, Olivier Cruchant, Cedric Archambeau
In three benchmarks where hardware is selected in addition to hyperparameters, we obtain runtime and cost reductions of at least 5. 8x and 8. 8x, respectively.
no code implementations • ICML Workshop AutoML 2021 • Julien Niklas Siems, Aaron Klein, Cedric Archambeau, Maren Mahsereci
Dynamic sparsity pruning undoes this limitation and allows to adapt the structure of the sparse neural network during training.
1 code implementation • 16 Apr 2021 • Anastasia Makarova, Huibin Shen, Valerio Perrone, Aaron Klein, Jean Baptiste Faddoul, Andreas Krause, Matthias Seeger, Cedric Archambeau
Across an extensive range of real-world HPO problems and baselines, we show that our termination criterion achieves a better trade-off between the test performance and optimization time.
1 code implementation • 17 Feb 2021 • Louis C. Tiao, Aaron Klein, Matthias Seeger, Edwin V. Bonilla, Cedric Archambeau, Fabio Ramos
Bayesian optimization (BO) is among the most effective and widely-used blackbox optimization methods.
3 code implementations • 24 Mar 2020 • Aaron Klein, Louis C. Tiao, Thibaut Lienart, Cedric Archambeau, Matthias Seeger
We introduce a model-based asynchronous multi-fidelity method for hyperparameter and neural architecture search that combines the strengths of asynchronous Hyperband and Gaussian process-based Bayesian optimization.
no code implementations • 22 Mar 2020 • Eric Hans Lee, Valerio Perrone, Cedric Archambeau, Matthias Seeger
Bayesian optimization (BO) is a class of global optimization algorithms, suitable for minimizing an expensive objective function in as few function evaluations as possible.
no code implementations • ICML 2020 • Cuong V. Nguyen, Tal Hassner, Matthias Seeger, Cedric Archambeau
We introduce a new measure to evaluate the transferability of representations learned by classifiers.
Ranked #4 on Transferability on classification benchmark
no code implementations • 15 Oct 2019 • Valerio Perrone, Iaroslav Shcherbatyi, Rodolphe Jenatton, Cedric Archambeau, Matthias Seeger
We propose constrained Max-value Entropy Search (cMES), a novel information theoretic-based acquisition function implementing this formulation.
no code implementations • NeurIPS 2019 • Valerio Perrone, Huibin Shen, Matthias Seeger, Cedric Archambeau, Rodolphe Jenatton
Despite its simplicity, we show that our approach considerably boosts BO by reducing the size of the search space, thus accelerating the optimization of a variety of black-box optimization problems.
no code implementations • NeurIPS 2018 • Valerio Perrone, Rodolphe Jenatton, Matthias W. Seeger, Cedric Archambeau
Bayesian optimization (BO) is a model-based approach for gradient-free black-box function optimization, such as hyperparameter optimization.
no code implementations • 8 Dec 2017 • Valerio Perrone, Rodolphe Jenatton, Matthias Seeger, Cedric Archambeau
Bayesian optimization (BO) is a model-based approach for gradient-free black-box function optimization.
no code implementations • ICML 2017 • Rodolphe Jenatton, Cedric Archambeau, Javier González, Matthias Seeger
The benefit of leveraging this structure is twofold: we explore the search space more efficiently and posterior inference scales more favorably with the number of observations than Gaussian Process-based approaches published in the literature.
no code implementations • 17 Feb 2016 • Rodolphe Jenatton, Jim Huang, Dominik Csiba, Cedric Archambeau
We consider online optimization in the 1-lookahead setting, where the objective does not decompose additively over the rounds of the online game.
no code implementations • 17 Jul 2015 • Cedric Archambeau, Beyza Ermis
We introduce incremental variational inference and apply it to latent Dirichlet allocation (LDA).
no code implementations • 24 Apr 2014 • Behrouz Behmardi, Cedric Archambeau, Guillaume Bouchard
Multi-view learning leverages correlations between different sources of data to make predictions in one view based on observations in another view.