1 code implementation • 6 Dec 2024 • Kaiyan Zhao, Tsuguchika Tabaru, Kenichi Kobayashi, Takumi Honda, Masafumi Yamazaki, Yoshimasa Tsuruoka
Although recent quantized Large Language Models (LLMs), such as BitNet, have paved the way for significant reduction in memory usage during deployment with binary or ternary weights, training these models still demands substantial memory footprints.
no code implementations • 13 Oct 2024 • Megha Sharma, Muhammad Taimoor Haseeb, Gus Xia, Yoshimasa Tsuruoka
To address these challenges, we propose an automated music generation pipeline that produces background music for an input manga book.
1 code implementation • 24 Jul 2024 • Hayato Watahiki, Ryo Iwase, Ryosuke Unno, Yoshimasa Tsuruoka
Our approach leverages multi-domain behavioral cloning on unaligned trajectories of proxy tasks and employs maximum mean discrepancy (MMD) as a regularization term to encourage cross-domain alignment.
1 code implementation • 25 Jun 2024 • Zhongtao Miao, Kaiyan Zhao, Yoshimasa Tsuruoka
Specifically, we use relation tuples, which are not only human-readable but also machine-friendly and easier to verify than natural language.
2 code implementations • 23 May 2024 • Takuya Hiraoka, Guanquan Wang, Takashi Onishi, Yoshimasa Tsuruoka
We evaluate how accurately PIToD estimates the influence of experiences and its efficiency compared to LOO.
no code implementations • 15 May 2024 • Qiyu Wu, Masaaki Nagata, Zhongtao Miao, Yoshimasa Tsuruoka
In this work, we mitigate the problem in an LLM-based MT model by guiding it to better word alignment.
no code implementations • 3 Apr 2024 • Zhongtao Miao, Qiyu Wu, Kaiyan Zhao, Zilong Wu, Yoshimasa Tsuruoka
The field of cross-lingual sentence embeddings has recently experienced significant advancements, but research concerning low-resource languages has lagged due to the scarcity of parallel corpora.
no code implementations • 16 Sep 2023 • Kaiyan Zhao, Qiyu Wu, Xin-Qiang Cai, Yoshimasa Tsuruoka
Learning multi-lingual sentence embeddings is a fundamental task in natural language processing.
2 code implementations • 9 Jun 2023 • Qiyu Wu, Masaaki Nagata, Yoshimasa Tsuruoka
Most existing word alignment methods rely on manual alignment datasets or parallel corpora, which limits their usefulness.
no code implementations • 21 May 2023 • Takahisa Imagawa, Takuya Hiraoka, Yoshimasa Tsuruoka
However, most of the existing methods learn a finite number of discrete skills, and thus the variety of behaviors that can be exhibited with the learned skills is limited.
1 code implementation • 26 Jan 2023 • Takuya Hiraoka, Takashi Onishi, Yoshimasa Tsuruoka
In reinforcement learning (RL) with experience replay, experiences stored in a replay buffer influence the RL agent's performance.
no code implementations • 8 Aug 2022 • Shumpei Kubosawa, Takashi Onishi, Yoshimasa Tsuruoka
During the operation of a chemical plant, product quality must be consistently maintained, and the production of off-specification products should be minimized.
1 code implementation • NAACL 2022 • Sosuke Nishikawa, Ryokan Ri, Ikuya Yamada, Yoshimasa Tsuruoka, Isao Echizen
We present EASE, a novel method for learning sentence embeddings via contrastive learning between sentences and their related entities.
no code implementations • ACL 2022 • Ryokan Ri, Yoshimasa Tsuruoka
We investigate what kind of structural knowledge learned in neural network encoders is transferable to processing natural language.
no code implementations • 17 Jan 2022 • Shumpei Kubosawa, Takashi Onishi, Makoto Sakahara, Yoshimasa Tsuruoka
The system leverages reinforcement learning and a dynamic simulator that can simulate the railway traffic and passenger flow of a whole line.
4 code implementations • ACL 2022 • Ryokan Ri, Ikuya Yamada, Yoshimasa Tsuruoka
We train a multilingual language model with 24 languages with entity representations and show the model consistently outperforms word-based pretrained models in various cross-lingual transfer tasks.
Ranked #1 on
Cross-Lingual Question Answering
on XQuAD
(Average F1 metric, using extra
training data)
no code implementations • 15 Oct 2021 • Sosuke Nishikawa, Ikuya Yamada, Yoshimasa Tsuruoka, Isao Echizen
We present a multilingual bag-of-entities model that effectively boosts the performance of zero-shot cross-lingual text classification by extending a multilingual pre-trained language model (e. g., M-BERT).
2 code implementations • ICLR 2022 • Takuya Hiraoka, Takahisa Imagawa, Taisei Hashimoto, Takashi Onishi, Yoshimasa Tsuruoka
To make REDQ more computationally efficient, we propose a method of improving computational efficiency called DroQ, which is a variant of REDQ that uses a small ensemble of dropout Q-functions.
no code implementations • ACL (WAT) 2021 • Ryokan Ri, Toshiaki Nakazawa, Yoshimasa Tsuruoka
For Japanese-to-English translation, zero pronouns in Japanese pose a challenge, since the model needs to infer and produce the corresponding pronoun in the target side of the English sentence.
1 code implementation • MTSummit 2021 • Ryokan Ri, Toshiaki Nakazawa, Yoshimasa Tsuruoka
Placeholder translation systems enable the users to specify how a specific phrase is translated in the output sentence.
no code implementations • 7 May 2021 • Taisei Hashimoto, Yoshimasa Tsuruoka
The key idea of our method is making the transition between action-decision points usable as training data by considering pseudo-actions.
no code implementations • ICML Workshop LifelongML 2020 • Takahisa Imagawa, Takuya Hiraoka, Yoshimasa Tsuruoka
It learns a belief model over the embedding space and a belief-conditional policy and Q-function.
no code implementations • 4 Jun 2020 • Takuya Hiraoka, Takahisa Imagawa, Voot Tangkaratt, Takayuki Osa, Takashi Onishi, Yoshimasa Tsuruoka
Model-based meta-reinforcement learning (RL) methods have recently been shown to be a promising approach to improving the sample efficiency of RL in multi-task settings.
no code implementations • ACL 2021 • Sosuke Nishikawa, Ryokan Ri, Yoshimasa Tsuruoka
Unsupervised cross-lingual word embedding (CLWE) methods learn a linear transformation matrix that maps two monolingual embedding spaces that are separately trained with monolingual corpora.
no code implementations • ACL 2020 • Ryokan Ri, Yoshimasa Tsuruoka
Existing approaches to mapping-based cross-lingual word embeddings are based on the assumption that the source and target embedding spaces are structurally similar.
Bilingual Lexicon Induction
Cross-Lingual Word Embeddings
+1
no code implementations • ACL 2019 • Go Yasui, Yoshimasa Tsuruoka, Masaaki Nagata
Traditional model training for sentence generation employs cross-entropy loss as the loss function.
no code implementations • 25 Jun 2019 • Takahisa Imagawa, Takuya Hiraoka, Yoshimasa Tsuruoka
Reinforcement Learning, a machine learning framework for training an autonomous agent based on rewards, has shown outstanding results in various domains.
no code implementations • 5 Jun 2019 • Shiqi Gao, Fuminori Okuya, Yoshihiro Kawahara, Yoshimasa Tsuruoka
The evaluation function for imperfect information games is always hard to define but owns a significant impact on the playing strength of a program.
no code implementations • CL 2019 • Akiko Eriguchi, Kazuma Hashimoto, Yoshimasa Tsuruoka
In those NMT models, sentences are simply treated as sequences of words without any internal structure.
1 code implementation • NeurIPS 2019 • Takuya Hiraoka, Takahisa Imagawa, Tatsuya Mori, Takashi Onishi, Yoshimasa Tsuruoka
While there are several methods to learn options that are robust against the uncertainty of model parameters, these methods only consider either the worst case or the average (ordinary) case for learning options.
no code implementations • 6 Mar 2019 • Shumpei Kubosawa, Takashi Onishi, Yoshimasa Tsuruoka
Chemical plants are complex and dynamical systems consisting of many components for manipulation and sensing, whose state transitions depend on various factors such as time, disturbance, and operation procedures.
no code implementations • 6 Feb 2019 • Keigo Kawamura, Yoshimasa Tsuruoka
Despite the notable successes in video games such as Atari 2600, current AI is yet to defeat human champions in the domain of real-time strategy (RTS) games.
no code implementations • 30 Dec 2018 • Naoya Taguchi, Yoshimasa Tsuruoka
Memory-Augmented Neural Networks (MANNs) are a class of neural networks equipped with an external memory, and are reported to be effective for tasks requiring a large long-term memory and its selective use.
no code implementations • 29 Sep 2018 • Takuya Hiraoka, Takashi Onishi, Takahisa Imagawa, Yoshimasa Tsuruoka
In this paper, we propose a framework that can automatically refine symbol grounding functions and a high-level planner to reduce human effort for designing these modules.
1 code implementation • 10 Sep 2018 • Akari Asai, Akiko Eriguchi, Kazuma Hashimoto, Yoshimasa Tsuruoka
Given a target language without RC training data and a pivot language with RC training data (e. g. English), our method leverages existing RC resources in the pivot language by combining a competitive RC model in the pivot language with an attentive Neural Machine Translation (NMT) model.
no code implementations • 7 Sep 2018 • Seydou Ba, Takuya Hiraoka, Takashi Onishi, Toru Nakata, Yoshimasa Tsuruoka
The evaluation results show that, with variable simulation times, the proposed approach outperforms the conventional MCTS in the evaluated continuous decision space tasks and improves the performance of MCTS in most of the ALE tasks.
1 code implementation • NAACL 2019 • Kazuma Hashimoto, Yoshimasa Tsuruoka
A major obstacle in reinforcement learning-based sentence generation is the large action space whose size is equal to the vocabulary size of the target-side language.
no code implementations • 28 Jun 2018 • Kazeto Yamamoto, Takashi Onishi, Yoshimasa Tsuruoka
One potential solution to this problem is to combine reinforcement learning with automated symbol planning and utilize prior knowledge on the domain.
Hierarchical Reinforcement Learning
reinforcement-learning
+2
1 code implementation • ACL 2017 • Akiko Eriguchi, Yoshimasa Tsuruoka, Kyunghyun Cho
There has been relatively little attention to incorporating linguistic prior to neural machine translation.
no code implementations • EMNLP 2017 • Kazuma Hashimoto, Yoshimasa Tsuruoka
This paper presents a novel neural machine translation model which jointly learns translation and source-side latent graph representations of sentences.
no code implementations • WS 2016 • Akiko Eriguchi, Kazuma Hashimoto, Yoshimasa Tsuruoka
This paper reports our systems (UT-AKY) submitted in the 3rd Workshop of Asian Translation 2016 (WAT{'}16) and their results in the English-to-Japanese translation task.
no code implementations • WS 2016 • Kazuma Hashimoto, Akiko Eriguchi, Yoshimasa Tsuruoka
This paper describes our UT-KAY system that participated in the Workshop on Asian Translation 2016.
2 code implementations • EMNLP 2017 • Kazuma Hashimoto, Caiming Xiong, Yoshimasa Tsuruoka, Richard Socher
Transfer and multi-task learning have traditionally focused on either a single source-target pair or very few, similar tasks.
Ranked #3 on
Chunking
on Penn Treebank
no code implementations • WS 2016 • Yusuke Watanabe, Kazuma Hashimoto, Yoshimasa Tsuruoka
Recently, recurrent neural networks have been shown to be successful on a variety of NLP tasks such as caption generation; however, the existing domain adaptation techniques are limited to (1) tune the model parameters by the target dataset after the training by the source dataset, or (2) design the network to have dual output, one for the source domain and the other for the target domain.
no code implementations • 8 May 2016 • Yun-Ching Liu, Yoshimasa Tsuruoka
We develop the Asymmetric-MCTS algorithm, which is an MCTS variant that applies a simple regret algorithm on max nodes, and the UCB algorithm on min nodes.
no code implementations • LREC 2016 • Shinsuke Mori, John Richardson, Atsushi Ushiku, Tetsuro Sasada, Hirotaka Kameko, Yoshimasa Tsuruoka
We describe a detailed definition of named entities and show some statistics of our game commentary corpus.
no code implementations • ACL 2016 • Kazuma Hashimoto, Yoshimasa Tsuruoka
We present a novel method for jointly learning compositional and non-compositional phrase embeddings by adaptively weighting both types of embeddings using a compositionality scoring function.
1 code implementation • ACL 2016 • Akiko Eriguchi, Kazuma Hashimoto, Yoshimasa Tsuruoka
Most of the existing Neural Machine Translation (NMT) models focus on the conversion of sequential data and do not directly use syntactic information.
no code implementations • 11 May 2015 • Yun-Ching Liu, Yoshimasa Tsuruoka
The UCT algorithm, which combines the UCB algorithm and Monte-Carlo Tree Search (MCTS), is currently the most widely used variant of MCTS.
no code implementations • CONLL 2015 • Kazuma Hashimoto, Pontus Stenetorp, Makoto Miwa, Yoshimasa Tsuruoka
We present a novel learning method for word embeddings designed for relation classification.