no code implementations • 19 Apr 2023 • Soichiro Nishimori, Sotetsu Koyamada, Shin Ishii
We proposed an RL algorithm that estimates the hidden states by end-to-end training, and visualize the estimation as a state-transition graph.
1 code implementation • NeurIPS 2023 • Sotetsu Koyamada, Shinri Okano, Soichiro Nishimori, Yu Murata, Keigo Habara, Haruka Kita, Shin Ishii
We propose Pgx, a suite of board game reinforcement learning (RL) environments written in JAX and optimized for GPU/TPU accelerators.
no code implementations • 30 Mar 2020 • Junjie Li, Sotetsu Koyamada, Qiwei Ye, Guoqing Liu, Chao Wang, Ruihan Yang, Li Zhao, Tao Qin, Tie-Yan Liu, Hsiao-Wuen Hon
Artificial Intelligence (AI) has achieved great success in many domains, and game AI is widely regarded as its beachhead since the dawn of AI.
no code implementations • ICLR 2018 • Sotetsu Koyamada, Yuta Kikuchi, Atsunori Kanemura, Shin-ichi Maeda, Shin Ishii
Neural sequence generation is commonly approached by using maximum- likelihood (ML) estimation or reinforcement learning (RL).
1 code implementation • 30 Jun 2017 • Sotetsu Koyamada, Yuta Kikuchi, Atsunori Kanemura, Shin-ichi Maeda, Shin Ishii
We propose a new neural sequence model training method in which the objective function is defined by $\alpha$-divergence.
no code implementations • 31 Jan 2015 • Sotetsu Koyamada, Yumi Shikauchi, Ken Nakae, Masanori Koyama, Shin Ishii
Our PSA successfully visualized the subject-independent features contributing to the subject-transferability of the trained decoder.
no code implementations • 21 Dec 2014 • Sotetsu Koyamada, Masanori Koyama, Ken Nakae, Shin Ishii
We then visualize the PSMs to demonstrate the PSA's ability to decompose the knowledge acquired by the trained classifiers.