no code implementations • ICLR 2019 • Masahiro Suzuki, Yusuke Iwasawa, Yutaka Matsuo
To solve this, we propose a concept to learn a mapping that embeds both images and attributes to the shared representation space that can be generalized even for unseen classes by interpolating from the information of seen classes, which we refer to shared manifold learning.
2 code implementations • 5 Nov 2024 • Shohei Taniguchi, Keno Harada, Gouki Minegishi, Yuta Oshima, Seong Cheol Jeong, Go Nagahara, Tomoshi Iiyama, Masahiro Suzuki, Yusuke Iwasawa, Yutaka Matsuo
Adam is one of the most popular optimization algorithms in deep learning.
1 code implementation • 9 Oct 2024 • Fumiya Uchiyama, Takeshi Kojima, Andrew Gambardella, Qi Cao, Yusuke Iwasawa, Yutaka Matsuo
In addition, we found that models trained with programming languages exhibit a better ability to follow instructions compared to those trained with natural languages.
no code implementations • 1 Oct 2024 • Shota Takashiro, Takeshi Kojima, Andrew Gambardella, Qi Cao, Yusuke Iwasawa, Yutaka Matsuo
As large language models (LLMs) are applied across diverse domains, the ability to selectively unlearn specific information has become increasingly essential.
no code implementations • 4 Jun 2024 • Andrew Gambardella, Yusuke Iwasawa, Yutaka Matsuo
The ability (and inability) of large language models (LLMs) to perform arithmetic tasks has been the subject of much theoretical and practical debate.
1 code implementation • 3 Apr 2024 • Takeshi Kojima, Itsuki Okimura, Yusuke Iwasawa, Hitomi Yanaka, Yutaka Matsuo
Additionally, we tamper with less than 1% of the total neurons in each model during inference and demonstrate that tampering with a few language-specific neurons drastically changes the probability of target language occurrence in text generation.
1 code implementation • 26 Feb 2024 • Hiroki Furuta, Gouki Minegishi, Yusuke Iwasawa, Yutaka Matsuo
Grokking on modular addition has been known to implement Fourier representation and its calculation circuits with trigonometric identities in Transformers.
1 code implementation • 30 Nov 2023 • Qi Cao, Takeshi Kojima, Yutaka Matsuo, Yusuke Iwasawa
While Large Language Models (LLMs) have achieved remarkable performance in many tasks, much about their inner workings remains unclear.
1 code implementation • 30 Oct 2023 • Gouki Minegishi, Yusuke Iwasawa, Yutaka Matsuo
We aim to analyze the mechanism of grokking from the lottery ticket hypothesis, identifying the process to find the lottery tickets (good sparse subnetworks) as the key to describing the transitional phase between memorization and generalization.
1 code implementation • 29 Sep 2023 • Jiaxian Guo, Bo Yang, Paul Yoo, Bill Yuchen Lin, Yusuke Iwasawa, Yutaka Matsuo
Unlike perfect information games, where all elements are known to every player, imperfect information games emulate the real-world complexities of decision-making under uncertain or incomplete information.
no code implementations • 16 Sep 2023 • So Kuroki, Jiaxian Guo, Tatsuya Matsushima, Takuya Okubo, Masato Kobayashi, Yuya Ikeda, Ryosuke Takanami, Paul Yoo, Yutaka Matsuo, Yusuke Iwasawa
Due to the inherent uncertainty in their deformability during motion, previous methods in deformable object manipulation, such as rope and cloth, often required hundreds of real-world demonstrations to train a manipulation policy for each object, which hinders their applications in our ever-changing world.
no code implementations • 14 Jun 2023 • So Kuroki, Jiaxian Guo, Tatsuya Matsushima, Takuya Okubo, Masato Kobayashi, Yuya Ikeda, Ryosuke Takanami, Paul Yoo, Yutaka Matsuo, Yusuke Iwasawa
To achieve this, we augment the policy by conditioning it on deformable rope parameters and training it with a diverse range of simulated deformable ropes so that the policy can adjust actions based on different rope parameters.
no code implementations • 13 Jun 2023 • Xin Zhang, Jiaxian Guo, Paul Yoo, Yutaka Matsuo, Yusuke Iwasawa
To guarantee the visual coherence of the generated or edited image, we introduce an inpainting and harmonizing module to guide the pre-trained diffusion model to seamlessly blend the inserted subject into the scene naturally.
1 code implementation • 31 May 2023 • Shohei Taniguchi, Masahiro Suzuki, Yusuke Iwasawa, Yutaka Matsuo
We address the problem of biased gradient estimation in deep Boltzmann machines (DBMs).
no code implementations • 29 Dec 2022 • Kei Akuzawa, Yusuke Iwasawa, Yutaka Matsuo
This paper proposes using multimodal generative models for semi-supervised learning in the instruction following tasks.
no code implementations • 28 Nov 2022 • Xinrui Wang, Zhuoru Li, Xiao Zhou, Yusuke Iwasawa, Yutaka Matsuo
Previous learning based stylization methods suffer from the geometric and semantic gaps between portrait domain and style domain, which obstacles the style information to be correctly transferred to the portrait images, leading to poor stylization quality.
1 code implementation • 25 Nov 2022 • Hiroki Furuta, Yusuke Iwasawa, Yutaka Matsuo, Shixiang Shane Gu
The rise of generalist large-scale models in natural language and vision has made us expect that a massive data-driven approach could achieve broader generalization in other domains such as continuous control.
1 code implementation • 15 Sep 2022 • Shohei Taniguchi, Yusuke Iwasawa, Wataru Kumagai, Yutaka Matsuo
Based on the ALD, we also present a new deep latent variable model named the Langevin autoencoder (LAE).
no code implementations • 20 Jul 2022 • Tatsuya Matsushima, Yuki Noguchi, Jumpei Arima, Toshiki Aoki, Yuki Okita, Yuya Ikeda, Koki Ishimoto, Shohei Taniguchi, Yuki Yamashita, Shoichi Seto, Shixiang Shane Gu, Yusuke Iwasawa, Yutaka Matsuo
Tidying up a household environment using a mobile manipulator poses various challenges in robotics, such as adaptation to large real-world environmental variations, and safe and robust deployment in the presence of humans. The Partner Robot Challenge in World Robot Challenge (WRC) 2020, a global competition held in September 2021, benchmarked tidying tasks in the real home environments, and importantly, tested for full system performances. For this challenge, we developed an entire household service robot system, which leverages a data-driven approach to adapt to numerous edge cases that occur during the execution, instead of classical manual pre-programmed solutions.
1 code implementation • 28 Jun 2022 • Takeshi Kojima, Yutaka Matsuo, Yusuke Iwasawa
Vision Transformer (ViT) is becoming more popular in image processing.
3 code implementations • 24 May 2022 • Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, Yusuke Iwasawa
Pretrained large language models (LLMs) are widely used in many sub-fields of natural language processing (NLP) and generally known as excellent few-shot learners with task-specific exemplars.
Ranked #1 on Arithmetic Reasoning on MultiArith
no code implementations • NeurIPS 2021 • Yusuke Iwasawa, Yutaka Matsuo
This paper presents a new algorithm for domain generalization (DG), \textit{test-time template adjuster (T3A)}, aiming to robustify a model to unknown distribution shift.
1 code implementation • 25 Nov 2021 • Xin Zhang, Shixiang Shane Gu, Yutaka Matsuo, Yusuke Iwasawa
We propose Domain Prompt Learning (DPL) as a novel approach for domain inference in the form of conditional prompt generation.
Ranked #1 on Transfer Learning on Office-Home
no code implementations • 13 Oct 2021 • Kazutoshi Shinoda, Yuki Takezawa, Masahiro Suzuki, Yusuke Iwasawa, Yutaka Matsuo
An interactive instruction following task has been proposed as a benchmark for learning to map natural language instructions and first-person vision into sequences of actions to interact with objects in 3D environments.
no code implementations • 14 May 2021 • Kei Akuzawa, Yusuke Iwasawa, Yutaka Matsuo
Therefore, the meta-RL agent faces the challenge of specifying both the hidden task and states based on small amount of experience.
no code implementations • EACL 2021 • Takeshi Kojima, Yusuke Iwasawa, Yutaka Matsuo
In this paper, we propose a GAN model that aims to improve the approach to generating diverse texts conditioned by the latent space.
no code implementations • ICLR 2021 • Makoto Kawano, Wataru Kumagai, Akiyoshi Sannai, Yusuke Iwasawa, Yutaka Matsuo
We present the group equivariant conditional neural process (EquivCNP), a meta-learning method with permutation invariance in a data set as in conventional conditional neural processes (CNPs), and it also has transformation equivariance in data space.
no code implementations • 11 Jan 2021 • Takumi Watanabe, Hiroki Takahashi, Goh Sato, Yusuke Iwasawa, Yutaka Matsuo, Ikuko Eguchi Yairi
This paper introduces our methodology to estimate sidewalk accessibilities from wheelchair behavior via a triaxial accelerometer in a smartphone installed under a wheelchair seat.
no code implementations • 1 Jan 2021 • Kei Akuzawa, Yusuke Iwasawa, Yutaka Matsuo
However, by analyzing the sequential VAEs from the information theoretic perspective, we can claim that simply maximizing the MI encourages the latent variables to have redundant information and prevents the disentanglement of global and local features.
no code implementations • 1 Jan 2021 • Shohei Taniguchi, Yusuke Iwasawa, Yutaka Matsuo
Developing a latent variable model and an inference model with neural networks, yields Langevin autoencoders (LAEs), a novel Langevin-based framework for deep generative models.
no code implementations • 6 Apr 2020 • Kenya Sakka, Kotaro Nakayama, Nisei Kimura, Taiki Inoue, Yusuke Iwasawa, Ryohei Yamaguchi, Yosimasa Kawazoe, Kazuhiko Ohe, Yutaka Matsuo
And, we were confirmed from the generated findings that the proposed method was able to consider the orthographic variants.
1 code implementation • ACM 2019 • Hiromi Nakagawa, Yusuke Iwasawa, Yutaka Matsuo
Inspired by the recent successes of the graph neural network (GNN), we herein propose a GNN-based knowledge tracing method, i. e., graph-based knowledge tracing.
no code implementations • 25 Sep 2019 • Yusuke Iwasawa, Kei Akuzawa, Yutaka Matsuo
An adversarial invariance induction (AII) shows its power on this purpose, which maximizes the proxy of the conditional entropy between representations and attributes by adversarial training between an attribute discriminator and feature extractor.
no code implementations • 29 Apr 2019 • Kei Akuzawa, Yusuke Iwasawa, Yutaka Matsuo
However, previous domain-invariance-based methods overlooked the underlying dependency of classes on domains, which is responsible for the trade-off between classification accuracy and domain invariance.
no code implementations • ICLR Workshop LLD 2019 • Yusuke Iwasawa, Kei Akuzawa, Yutaka Matsuo
An adversarial feature learning (AFL) is a powerful framework to learn representations invariant to a nuisance attribute, which uses an adversarial game between a feature extractor and a categorical attribute classifier.
no code implementations • ICLR Workshop LLD 2019 • Kei Akuzawa, Yusuke Iwasawa, Yutaka Matsuo
Learning domain-invariant representation is a dominant approach for domain generalization.
no code implementations • 27 Sep 2018 • Kei Akuzawa, Yusuke Iwasawa, Yutaka Matsuo
Learning domain-invariant representation is a dominant approach for domain generalization, where we need to build a classifier that is robust toward domain shifts induced by change of users, acoustic or lighting conditions, etc.
no code implementations • 6 Apr 2018 • Kei Akuzawa, Yusuke Iwasawa, Yutaka Matsuo
Recent advances in neural autoregressive models have improve the performance of speech synthesis (SS).
no code implementations • ICLR 2018 • Yusuke Iwasawa, Kotaro Nakayama, Yutaka Matsuo
AFL learn such a representations by training the networks to deceive the adversary that predict the sensitive information from the network, and therefore, the success of the AFL heavily relies on the choice of the adversary.
no code implementations • ICLR 2018 • Joji Toyama, Yusuke Iwasawa, Kotaro Nakayama, Yutaka Matsuo
The partial reward function is a reward function for a partial sequence of a certain length.
no code implementations • ICLR 2018 • Shohei Ohsawa, Kei Akuzawa, Tatsuya Matsushima, Gustavo Bezerra, Yusuke Iwasawa, Hiroshi Kajino, Seiya Takenaka, Yutaka Matsuo
Existing multi-agent reinforcement learning (MARL) communication methods have relied on a trusted third party (TTP) to distribute reward to agents, leaving them inapplicable in peer-to-peer environments.