no code implementations • COLING 2022 • Yuto Kuroda, Tomoyuki Kajiwara, Yuki Arase, Takashi Ninomiya
We propose a method to distill language-agnostic meaning embeddings from multilingual sentence encoders for unsupervised quality estimation of machine translation.
no code implementations • EAMT 2020 • Takeshi Hayakawa, Yuki Arase
We performed a detailed error analysis in domain-specific neural machine translation (NMT) for the English and Japanese language pair with fine-grained manual annotation.
no code implementations • EMNLP 2021 • Han Huang, Tomoyuki Kajiwara, Yuki Arase
Definition generation techniques aim to generate a definition of a target word or phrase given a context.
1 code implementation • EMNLP 2021 • Nattapong Tiyajamorn, Tomoyuki Kajiwara, Yuki Arase, Makoto Onizuka
Experimental results on both quality estimation of machine translation and cross-lingual semantic textual similarity tasks reveal that our method consistently outperforms the strong baselines using the original multilingual embedding.
Cross-Lingual Semantic Textual Similarity
Machine Translation
+3
no code implementations • EMNLP 2020 • Yuki Arase, Jun{'}ichi Tsujii
Most phrase alignments are compositional processes such that an alignment of a phrase pair is constructed based on the alignments of their child phrases.
no code implementations • LREC 2022 • Han Huang, Tomoyuki Kajiwara, Yuki Arase
This study investigated and released the JADE, a corpus for Japanese definition modelling, which is a technique that automatically generates definitions of a given target word and phrase.
1 code implementation • Findings (EMNLP) 2021 • Junya Takayama, Tomoyuki Kajiwara, Yuki Arase
We create a large-scale dialogue corpus that provides pragmatic paraphrases to advance technology for understanding the underlying intentions of users.
1 code implementation • Findings (EMNLP) 2021 • Yuki Arase, Tomoyuki Kajiwara
The results confirm that our representations exhibited a competitive performance compared to that of the state-of-the-art method transforming contextualised representations for the context-aware lexical semantic tasks and outperformed it for STS estimation.
1 code implementation • 17 Feb 2025 • Guanlin Li, Yuki Arase, Noel Crespi
Experiment results on CEFR-SP and TurkCorpus datasets show that the proposed method can effectively increase the frequency and diversity of vocabulary of the target level by more than $20\%$ compared to baseline models, while maintaining high simplification quality.
1 code implementation • 28 Sep 2024 • Tatsuya Zetsu, Yuki Arase, Tomoyuki Kajiwara
We propose edit operation based lexically constrained decoding for sentence simplification.
1 code implementation • 13 Sep 2024 • Yuki Arase, Tomoyuki Kajiwara
The results on the monolingual tasks confirmed that our representations exhibited a competitive performance compared to that of the previous study for the context-aware lexical semantic tasks and outperformed it for STS estimation.
no code implementations • 8 Mar 2024 • Xuanxin Wu, Yuki Arase
Second, current human evaluation approaches in sentence simplification often fall into two extremes: they are either too superficial, failing to offer a clear understanding of the models' performance, or overly detailed, making the annotation process complex and prone to inconsistency, which in turn affects the evaluation's reliability.
no code implementations • 29 Jun 2023 • Ji-Ung Lee, Haritz Puerto, Betty van Aken, Yuki Arase, Jessica Zosa Forde, Leon Derczynski, Andreas Rücklé, Iryna Gurevych, Roy Schwartz, Emma Strubell, Jesse Dodge
Many recent improvements in NLP stem from the development and use of large pre-trained language models (PLMs) with billions of parameters.
1 code implementation • 7 Jun 2023 • Yuki Arase, Han Bao, Sho Yokoi
Monolingual word alignment is crucial to model semantic interactions between sentences.
1 code implementation • 21 Oct 2022 • Yuki Arase, Satoru Uchida, Tomoyuki Kajiwara
Controllable text simplification is a crucial assistive technique for language learning and teaching.
1 code implementation • ACL 2021 • Sora Ohashi, Junya Takayama, Tomoyuki Kajiwara, Yuki Arase
Few-shot text classification aims to classify inputs whose label has only a few examples.
no code implementations • ACL 2021 • Sora Kadotani, Tomoyuki Kajiwara, Yuki Arase, Makoto Onizuka
Curriculum learning has improved the quality of neural machine translation, where only source-side features are considered in the metrics to determine the difficulty of translation.
no code implementations • COLING 2020 • Sora Ohashi, Mao Isogawa, Tomoyuki Kajiwara, Yuki Arase
We reduce the model size of pre-trained word embeddings by a factor of 200 while preserving its quality.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Akifumi Nakamachi, Tomoyuki Kajiwara, Yuki Arase
We optimize rewards of reinforcement learning in text simplification using metrics that are highly correlated with human-perspectives.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Junya Takayama, Yuki Arase
To control the specificity of generated responses, we add the distant supervision based on the co-occurrence degree and a PMI-based word prediction mechanism to a sequence-to-sequence model.
no code implementations • 17 Oct 2020 • Andrew Merritt, Chenhui Chu, Yuki Arase
Multimodal neural machine translation (NMT) has become an increasingly important area of research over the years because additional modalities, such as image data, can provide more context to textual data.
no code implementations • 11 Oct 2020 • Vipul Mishra, Chenhui Chu, Yuki Arase
Lexically cohesive translations preserve consistency in word choices in document-level translation.
no code implementations • ACL 2020 • Sora Ohashi, Junya Takayama, Tomoyuki Kajiwara, Chenhui Chu, Yuki Arase
Advanced pre-trained models for text representation have achieved state-of-the-art performance on various text classification tasks.
no code implementations • LREC 2020 • Yuki Arase, Tomoyuki Kajiwara, Chenhui Chu
The dataset we present in this paper is unique for the richness of annotated information, including detailed descriptions of drug reactions with full context.
no code implementations • LREC 2020 • Masato Yoshinaka, Tomoyuki Kajiwara, Yuki Arase
To estimate the likelihood of phrase alignments, SAPPHIRE uses phrase embeddings that are hierarchically composed of word embeddings.
Natural Language Inference
Natural Language Understanding
+2
no code implementations • WS 2019 • Kazuki Ashihara, Tomoyuki Kajiwara, Yuki Arase, Satoru Uchida
Herein we propose a method that combines these two approaches to contextualize word embeddings for lexical substitution.
1 code implementation • IJCNLP 2019 • Yuki Arase, Junichi Tsujii
It constitutes a set of tasks crucial for research on natural language understanding.
no code implementations • WS 2019 • Junya Takayama, Yuki Arase
A sequence-to-sequence model tends to generate generic responses with little information for input utterances.
1 code implementation • WS 2019 • Kozo Chikai, Junya Takayama, Yuki Arase
Specifically, our model generates domain-aware and sentiment-rich responses.
no code implementations • ACL 2019 • Koji Tanaka, Junya Takayama, Yuki Arase
One significant drawback of such a neural network based approach is that the response generation process is a black-box, and how a specific response is generated is unclear.
no code implementations • ACL 2019 • Daiki Nishihara, Tomoyuki Kajiwara, Yuki Arase
Our text simplification method succeeds in translating an input into a specific grade level by considering levels of both sentences and words.
no code implementations • ACL 2018 • Yuki Kawara, Chenhui Chu, Yuki Arase
Experiments show that the proposed method achieves comparable gain in translation quality to the state-of-the-art method but without a manual feature design.
no code implementations • EMNLP 2017 • Yuki Arase, Junichi Tsujii
We propose an efficient method to conduct phrase alignment on parse forests for paraphrase detection.