no code implementations • 15 Mar 2024 • Shin'ya Yamaguchi, Sekitoshi Kanai, Kazuki Adachi, Daiki Chijiwa
To this end, AdaRand minimizes the gap between feature vectors and random reference vectors that are sampled from class conditional Gaussian distributions.
no code implementations • 20 Feb 2024 • Hikari Otsuka, Daiki Chijiwa, Ángel López García-Arias, Yasuyuki Okoshi, Kazushi Kawamura, Thiem Van Chu, Daichi Fujiki, Susumu Takeuchi, Masato Motomura
In addition to reducing search space, the random freezing pattern can also be exploited to reduce model size in inference.
no code implementations • 9 Jun 2023 • Masanori Yamada, Tomoya Yamashita, Shin'ya Yamaguchi, Daiki Chijiwa
We also show that merged models require datasets for merging in order to achieve a high accuracy.
no code implementations • 23 May 2023 • Daiki Chijiwa
Training deep neural networks (DNNs) is computationally expensive, which is problematic especially when performing duplicated or similar training runs in model ensemble or fine-tuning pre-trained models, for example.
1 code implementation • 31 May 2022 • Daiki Chijiwa, Shin'ya Yamaguchi, Atsutoshi Kumagai, Yasutoshi Ida
Few-shot learning for neural networks (NNs) is an important problem that aims to train NNs with a few data.
no code implementations • 27 Apr 2022 • Shin'ya Yamaguchi, Sekitoshi Kanai, Atsutoshi Kumagai, Daiki Chijiwa, Hisashi Kashima
To transfer source knowledge without these assumptions, we propose a transfer learning method that uses deep generative models and is composed of the following two stages: pseudo pre-training (PP) and pseudo semi-supervised learning (P-SSL).
1 code implementation • NeurIPS 2021 • Daiki Chijiwa, Shin'ya Yamaguchi, Yasutoshi Ida, Kenji Umakoshi, Tomohiro Inoue
Pruning the weights of randomly initialized neural networks plays an important role in the context of lottery ticket hypothesis.