Search Results for author: Tomoyuki Kajiwara

Found 45 papers, 17 papers with code

Language-agnostic Representation from Multilingual Sentence Encoders for Cross-lingual Similarity Estimation

1 code implementation EMNLP 2021 Nattapong Tiyajamorn, Tomoyuki Kajiwara, Yuki Arase, Makoto Onizuka

Experimental results on both quality estimation of machine translation and cross-lingual semantic textual similarity tasks reveal that our method consistently outperforms the strong baselines using the original multilingual embedding.

Cross-Lingual Semantic Textual Similarity Machine Translation +3

Definition Modelling for Appropriate Specificity

no code implementations EMNLP 2021 Han Huang, Tomoyuki Kajiwara, Yuki Arase

Definition generation techniques aim to generate a definition of a target word or phrase given a context.

Definition Modelling Re-Ranking +1

Distilling Word Meaning in Context from Pre-trained Language Models

1 code implementation Findings (EMNLP) 2021 Yuki Arase, Tomoyuki Kajiwara

The results confirm that our representations exhibited a competitive performance compared to that of the state-of-the-art method transforming contextualised representations for the context-aware lexical semantic tasks and outperformed it for STS estimation.

Language Modelling Self-Supervised Learning +3

DIRECT: Direct and Indirect Responses in Conversational Text Corpus

1 code implementation Findings (EMNLP) 2021 Junya Takayama, Tomoyuki Kajiwara, Yuki Arase

We create a large-scale dialogue corpus that provides pragmatic paraphrases to advance technology for understanding the underlying intentions of users.

Adversarial Training on Disentangling Meaning and Language Representations for Unsupervised Quality Estimation

no code implementations COLING 2022 Yuto Kuroda, Tomoyuki Kajiwara, Yuki Arase, Takashi Ninomiya

We propose a method to distill language-agnostic meaning embeddings from multilingual sentence encoders for unsupervised quality estimation of machine translation.

Machine Translation Sentence +1

JADE: Corpus for Japanese Definition Modelling

no code implementations LREC 2022 Han Huang, Tomoyuki Kajiwara, Yuki Arase

This study investigated and released the JADE, a corpus for Japanese definition modelling, which is a technique that automatically generates definitions of a given target word and phrase.

Definition Modelling

A Benchmark Dataset for Multi-Level Complexity-Controllable Machine Translation

1 code implementation LREC 2022 Kazuki Tani, Ryoya Yuasa, Kazuki Takikawa, Akihiro Tamura, Tomoyuki Kajiwara, Takashi Ninomiya, Tsuneo Kato

Therefore, we create a benchmark test dataset for Japanese-to-English MLCC-MT from the Newsela corpus by introducing an automatic filtering of data with inappropriate sentence-level complexity, manual check for parallel target language sentences with different complexity levels, and manual translation.

Machine Translation NMT +2

Unsupervised Translation Quality Estimation Exploiting Synthetic Data and Pre-trained Multilingual Encoder

no code implementations9 Nov 2023 Yuto Kuroda, Atsushi Fujita, Tomoyuki Kajiwara, Takashi Ninomiya

In this paper, we extensively investigate the usefulness of synthetic TQE data and pre-trained multilingual encoders in unsupervised sentence-level TQE, both of which have been proven effective in the supervised training scenarios.

Sentence Translation

CEFR-Based Sentence Difficulty Annotation and Assessment

1 code implementation21 Oct 2022 Yuki Arase, Satoru Uchida, Tomoyuki Kajiwara

Controllable text simplification is a crucial assistive technique for language learning and teaching.

Sentence Text Simplification

Edit Distance Based Curriculum Learning for Paraphrase Generation

no code implementations ACL 2021 Sora Kadotani, Tomoyuki Kajiwara, Yuki Arase, Makoto Onizuka

Curriculum learning has improved the quality of neural machine translation, where only source-side features are considered in the metrics to determine the difficulty of translation.

Machine Translation Paraphrase Generation +1

WRIME: A New Dataset for Emotional Intensity Estimation with Subjective and Objective Annotations

1 code implementation NAACL 2021 Tomoyuki Kajiwara, Chenhui Chu, Noriko Takemura, Yuta Nakashima, Hajime Nagahara

We annotate 17, 000 SNS posts with both the writer{'}s subjective emotional intensity and the reader{'}s objective one to construct a Japanese emotion analysis dataset.

Emotion Recognition

Double Attention-based Multimodal Neural Machine Translation with Semantic Image Regions

1 code implementation EAMT 2020 YuTing Zhao, Mamoru Komachi, Tomoyuki Kajiwara, Chenhui Chu

In contrast, we propose the application of semantic image regions for MNMT by integrating visual and textual features using two individual attention mechanisms (double attention).

Machine Translation Translation

Text Classification with Negative Supervision

no code implementations ACL 2020 Sora Ohashi, Junya Takayama, Tomoyuki Kajiwara, Chenhui Chu, Yuki Arase

Advanced pre-trained models for text representation have achieved state-of-the-art performance on various text classification tasks.

General Classification Semantic Similarity +4

Annotation of Adverse Drug Reactions in Patients' Weblogs

no code implementations LREC 2020 Yuki Arase, Tomoyuki Kajiwara, Chenhui Chu

The dataset we present in this paper is unique for the richness of annotated information, including detailed descriptions of drug reactions with full context.

Word Complexity Estimation for Japanese Lexical Simplification

no code implementations LREC 2020 Daiki Nishihara, Tomoyuki Kajiwara

We introduce three language resources for Japanese lexical simplification: 1) a large-scale word complexity lexicon, 2) the first synonym lexicon for converting complex words to simpler ones, and 3) the first toolkit for developing and benchmarking Japanese lexical simplification system.

Benchmarking Lexical Simplification

Contextualized context2vec

no code implementations WS 2019 Kazuki Ashihara, Tomoyuki Kajiwara, Yuki Arase, Satoru Uchida

Herein we propose a method that combines these two approaches to contextualize word embeddings for lexical substitution.

Sentence Word Embeddings

Machine Translation Evaluation with BERT Regressor

no code implementations29 Jul 2019 Hiroki Shimanaka, Tomoyuki Kajiwara, Mamoru Komachi

We introduce the metric using BERT (Bidirectional Encoder Representations from Transformers) (Devlin et al., 2019) for automatic machine translation evaluation.

Machine Translation Translation

Controllable Text Simplification with Lexical Constraint Loss

no code implementations ACL 2019 Daiki Nishihara, Tomoyuki Kajiwara, Yuki Arase

Our text simplification method succeeds in translating an input into a specific grade level by considering levels of both sentences and words.

Sentence Text Simplification +1

Using Natural Language Processing to Develop an Automated Orthodontic Diagnostic System

no code implementations31 May 2019 Tomoyuki Kajiwara, Chihiro Tanikawa, Yuujin Shimizu, Chenhui Chu, Takashi Yamashiro, Hajime Nagahara

We work on the task of automatically designing a treatment plan from the findings included in the medical certificate written by the dentist.

TMU System for SLAM-2018

1 code implementation WS 2018 Masahiro Kaneko, Tomoyuki Kajiwara, Mamoru Komachi

We introduce the TMU systems for the second language acquisition modeling shared task 2018 (Settles et al., 2018).

Language Acquisition

Metric for Automatic Machine Translation Evaluation based on Universal Sentence Representations

no code implementations NAACL 2018 Hiroki Shimanaka, Tomoyuki Kajiwara, Mamoru Komachi

Sentence representations can capture a wide range of information that cannot be captured by local features based on character or word N-grams.

Machine Translation Sentence +1

Improving Japanese-to-English Neural Machine Translation by Paraphrasing the Target Language

no code implementations WS 2017 Yuuki Sekizawa, Tomoyuki Kajiwara, Mamoru Komachi

Neural machine translation (NMT) produces sentences that are more fluent than those produced by statistical machine translation (SMT).

Machine Translation NMT +1

MIPA: Mutual Information Based Paraphrase Acquisition via Bilingual Pivoting

1 code implementation IJCNLP 2017 Tomoyuki Kajiwara, Mamoru Komachi, Daichi Mochihashi

We present a pointwise mutual information (PMI)-based approach to formalize paraphrasability and propose a variant of PMI, called MIPA, for the paraphrase acquisition.

Learning Word Embeddings Semantic Textual Similarity +1

Building a Monolingual Parallel Corpus for Text Simplification Using Sentence Similarity Based on Alignment between Word Embeddings

no code implementations COLING 2016 Tomoyuki Kajiwara, Mamoru Komachi

To obviate the need for human annotation, we propose an unsupervised method that automatically builds the monolingual parallel corpus for text simplification using sentence similarity based on word embeddings.

Machine Translation Sentence +4

Cannot find the paper you are looking for? You can Submit a new open access paper.