no code implementations • NeurIPS Workshop DL-IG 2020 • Mahito Sugiyama, Koji Tsuda, Hiroyuki Nakahara
We present a lightweight variant of Boltzmann machines via sample space truncation, called a truncated Boltzmann machine (TBM), which has not been investigated before while can be naturally introduced from the log-linear model viewpoint.
no code implementations • 21 May 2018 • Mahito Sugiyama, Koji Tsuda, Hiroyuki Nakahara
We present transductive Boltzmann machines (TBMs), which firstly achieve transductive learning of the Gibbs distribution.
1 code implementation • NeurIPS 2018 • Mahito Sugiyama, Hiroyuki Nakahara, Koji Tsuda
We present a novel nonnegative tensor decomposition method, called Legendre decomposition, which factorizes an input tensor into a multiplicative combination of parameters.
no code implementations • ICLR 2018 • Mahito Sugiyama, Koji Tsuda, Hiroyuki Nakahara
We achieve bias-variance decomposition for Boltzmann machines using an information geometric formulation.
1 code implementation • ICML 2017 • Mahito Sugiyama, Hiroyuki Nakahara, Koji Tsuda
To theoretically prove the correctness of the algorithm, we model tensors as probability distributions in a statistical manifold and realize tensor balancing as projection onto a submanifold.