no code implementations • ICLR 2019 • Nozomu Yoshinari, Kento Uchida, Shota Saito, Shinichi Shirakawa, Youhei Akimoto
The experimental results show that the proposed architecture search method is fast and can achieve comparable performance to the existing methods.
no code implementations • 2 Dec 2024 • Kento Uchida, Teppei Yamaguchi, Shinichi Shirakawa
Despite the state-of-the-art performance of the covariance matrix adaptation evolution strategy (CMA-ES), high-dimensional black-box optimization problems are challenging tasks.
no code implementations • 23 Aug 2024 • Kento Uchida, Ryoki Hamano, Masahiro Nomura, Shota Saito, Shinichi Shirakawa
Discrete and mixed-variable optimization problems have appeared in several real-world applications.
no code implementations • 10 Jul 2024 • Ryoki Hamano, Kento Uchida, Shinichi Shirakawa, Daiki Morinaga, Youhei Akimoto
We term this specific algorithm the categorical compact genetic algorithm (ccGA).
1 code implementation • 24 Jun 2024 • Ryoki Hamano, Shinichi Shirakawa, Masahiro Nomura
While the rank-one update makes the covariance matrix to increase the likelihood of generating a solution in the direction of the evolution path, this idea has been difficult to formulate and interpret as a natural gradient method unlike the rank-$\mu$ update.
no code implementations • 19 May 2024 • Kento Uchida, Kenta Nishihara, Shinichi Shirakawa
We derive that the set of maximizers of the noise-independent utility, which is used in the reevaluation technique, certainly contains the optimal solution, while the noise-dependent utility, which is used in the population size and leaning rate adaptations, does not satisfy it under multiplicative noise.
no code implementations • 17 May 2024 • Kento Uchida, Ryoki Hamano, Masahiro Nomura, Shota Saito, Shinichi Shirakawa
This optimization setting is known as safe optimization and formulated as a specialized type of constrained optimization problem with constraints for safety functions.
1 code implementation • 16 May 2024 • Ryoki Hamano, Shota Saito, Masahiro Nomura, Kento Uchida, Shinichi Shirakawa
CatCMA updates the parameters of the joint probability distribution in the natural gradient direction.
no code implementations • 1 May 2023 • Yohei Watanabe, Kento Uchida, Ryoki Hamano, Shota Saito, Masahiro Nomura, Shinichi Shirakawa
The margin correction has been applied to ($\mu/\mu_\mathrm{w}$,$\lambda$)-CMA-ES, while this paper introduces the margin correction into (1+1)-CMA-ES, an elitist version of CMA-ES.
1 code implementation • 31 Mar 2023 • Masashi Noguchi, Shinichi Shirakawa
In real-world applications, a machine learning model is required to handle an open-set recognition (OSR), where unknown classes appear during the inference, in addition to a domain shift, where the distribution of data differs between the training and inference phases.
1 code implementation • 19 Dec 2022 • Ryoki Hamano, Shota Saito, Masahiro Nomura, Shinichi Shirakawa
If the CMA-ES is applied to the MI-BBO with straightforward discretization, however, the variance corresponding to the integer variables becomes much smaller than the granularity of the discretization before reaching the optimal solution, which leads to the stagnation of the optimization.
no code implementations • 30 Aug 2022 • Shoma Shimizu, Takayuki Nishio, Shota Saito, Yoichi Hirose, Chen Yen-Hsiu, Shinichi Shirakawa
This paper proposes a neural architecture search (NAS) method for split computing.
no code implementations • 21 Jul 2022 • Yuhei Noda, Shota Saito, Shinichi Shirakawa
The proposed method allows us to obtain multiple architectures with different complexities in a single architecture search, resulting in reducing the search cost.
3 code implementations • 26 May 2022 • Ryoki Hamano, Shota Saito, Masahiro Nomura, Shinichi Shirakawa
If the CMA-ES is applied to the MI-BBO with straightforward discretization, however, the variance corresponding to the integer variables becomes much smaller than the granularity of the discretization before reaching the optimal solution, which leads to the stagnation of the optimization.
1 code implementation • 29 Mar 2022 • Ryoji Tanabe, Youhei Akimoto, Ken Kobayashi, Hiroshi Umeki, Shinichi Shirakawa, Naoki Hamada
The first phase in TPB aims to approximate a few Pareto optimal solutions by optimizing a sequence of single-objective scalar problems.
1 code implementation • 19 Oct 2021 • Yoichi Hirose, Nozomu Yoshinari, Shinichi Shirakawa
Building the benchmark dataset for joint optimization of architecture and training hyperparameters is essential to further NAS research.
no code implementations • 15 Jul 2019 • Shota Saito, Shinichi Shirakawa
We focus on the probabilistic model-based dynamic neural network structure optimization that considers the probability distribution of structure parameters and simultaneously optimizes both the distribution parameters and connection weights based on gradient methods.
1 code implementation • 21 May 2019 • Youhei Akimoto, Shinichi Shirakawa, Nozomu Yoshinari, Kento Uchida, Shota Saito, Kouhei Nishida
It accepts arbitrary search space (widely-applicable) and enables to employ a gradient-based simultaneous optimization of weights and architecture (fast).
no code implementations • 18 Sep 2018 • Kouhei Nishida, Hernan Aguirre, Shota Saito, Shinichi Shirakawa, Youhei Akimoto
This paper proposes a parameterless BBDO algorithm based on information geometric optimization, a recent framework for black box optimization using stochastic natural gradient.
no code implementations • 31 May 2018 • Shinichi Shirakawa, Youhei Akimoto, Kazuki Ouchi, Kouzou Ohara
The experimental results show that the sample reuse helps to reduce the number of function evaluations on many benchmark functions for both the PBIL and the pure rank-$\mu$ update CMA-ES.
no code implementations • 23 Jan 2018 • Shinichi Shirakawa, Yasushi Iwata, Youhei Akimoto
We consider a probability distribution that generates network structures, and optimize the parameters of the distribution instead of directly optimizing the network structure.
5 code implementations • 3 Apr 2017 • Masanori Suganuma, Shinichi Shirakawa, Tomoharu Nagao
To evaluate the proposed method, we constructed a CNN architecture for the image classification task with the CIFAR-10 dataset.