no code implementations • ACL (WAT) 2021 • Hwichan Kim, Mamoru Komachi
In this paper, we introduce our TMU Neural Machine Translation (NMT) system submitted for the Patent task (Korean Japanese and English Japanese) of 8th Workshop on Asian Translation (Nakazawa et al., 2021).
no code implementations • ACL (WAT) 2021 • YuTing Zhao, Mamoru Komachi, Tomoyuki Kajiwara, Chenhui Chu
We introduce our TMEKU system submitted to the English-Japanese Multimodal Translation Task for WAT 2021.
no code implementations • ACL (WAT) 2021 • Seiichiro Kondo, Aomi Koyama, Tomoshige Kiyuna, Tosho Hirasawa, Mamoru Komachi
We introduce our TMU Japanese-to-English system, which employs a semi-autoregressive model, to tackle the WAT 2021 restricted translation task.
no code implementations • WAT 2022 • Seiichiro Kondo, Mamoru Komachi
One is lexical-constraint-aware neural machine translation (LeCA) (Chen et al., 2020), which is a method of adding RTVs at the end of input sentences.
no code implementations • AACL (WAT) 2020 • Hwichan Kim, Tosho Hirasawa, Mamoru Komachi
In this paper, we describe our TMU neural machine translation (NMT) system submitted for the Patent task (Korean→Japanese) of the 7th Workshop on Asian Translation (WAT 2020, Nakazawa et al., 2020).
no code implementations • AACL (WAT) 2020 • Hiroto Tamura, Tosho Hirasawa, Masahiro Kaneko, Mamoru Komachi
Subsequently, we pretrain a translation model on the augmented noisy data, and then fine-tune it on the clean data.
1 code implementation • AACL (WAT) 2020 • Zizheng Zhang, Tosho Hirasawa, Wei Houjing, Masahiro Kaneko, Mamoru Komachi
New things are being created and new words are constantly being added to languages worldwide.
no code implementations • LREC 2022 • Yujin Takahashi, Masahiro Kaneko, Masato Mita, Mamoru Komachi
This study investigates how supervised quality estimation (QE) models of grammatical error correction (GEC) are affected by the learners’ proficiency with the data.
no code implementations • WMT (EMNLP) 2020 • Akifumi Nakamachi, Hiroki Shimanaka, Tomoyuki Kajiwara, Mamoru Komachi
We introduce the TMUOU submission for the WMT20 Quality Estimation Shared Task 1: Sentence-Level Direct Assessment.
no code implementations • AACL (NLP-TEA) 2020 • Hongfei Wang, Mamoru Komachi
In this paper, we introduce our system for NLPTEA 2020 shared task of Chinese Grammatical Error Diagnosis (CGED).
1 code implementation • NAACL (WOAH) 2022 • Mana Ashida, Mamoru Komachi
With the widespread use of social media, online hate is increasing, and microaggressions are receiving attention.
no code implementations • NAACL (ACL) 2022 • Xiaomeng Pan, Hongfei Wang, Teruaki Oka, Mamoru Komachi
Creation of an ancient Chinese dataset is considered a significant challenge because determining the most appropriate sense in a context is difficult and time-consuming owing to the different usages in ancient and modern Chinese.
no code implementations • AACL (NLP-TEA) 2020 • Hiroki Homma, Mamoru Komachi
There are several problems in applying grammatical error correction (GEC) to a writing support system.
1 code implementation • 16 Jan 2025 • Hajime Kiyama, Taichi Aida, Mamoru Komachi, Toshinobu Ogiso, Hiroya Takamura, Daichi Mochihashi
To address those issues, we propose a simple yet intuitive framework for how semantic shifts occur over multiple time periods by leveraging a similarity matrix between the embeddings of the same word through time.
1 code implementation • 25 Sep 2024 • Hwichan Kim, Jun Suzuki, Tosho Hirasawa, Mamoru Komachi
This study explores how to enhance the zero-shot performance of MLLMs in non-English languages by leveraging their alignment capability between English and non-English languages.
no code implementations • 26 Mar 2024 • Masamune Kobayashi, Masato Mita, Mamoru Komachi
Large Language Models (LLMs) have been reported to outperform existing automatic evaluation metrics in some tasks, such as text summarization and machine translation.
2 code implementations • 5 Mar 2024 • Masamune Kobayashi, Masato Mita, Mamoru Komachi
The results of improved correlations by aligning the granularity in the sentence-level meta-evaluation, suggest that edit-based metrics may have been underestimated in existing studies.
1 code implementation • 10 May 2023 • Kenichiro Ando, Satoshi Sekine, Mamoru Komachi
Here, we propose WikiSQE, the first large-scale dataset for sentence quality estimation in Wikipedia.
no code implementations • 10 Mar 2023 • Kenichiro Ando, Mamoru Komachi, Takashi Okumura, Hiromasa Horiguchi, Yuji Matsumoto
During the patient's hospitalization, the physician must record daily observations of the patient and summarize them into a brief document called "discharge summary" when the patient is discharged.
no code implementations • 20 Sep 2022 • Kenichiro Ando, Takashi Okumura, Mamoru Komachi, Hiromasa Horiguchi, Yuji Matsumoto
We first defined three types of summarization units with different granularities to compare the performance of the discharge summary generation: whole sentences, clinical segments, and clauses.
no code implementations • LREC 2022 • Hwichan Kim, Sangwhan Moon, Naoaki Okazaki, Mamoru Komachi
Training a model using North Korean data is the most straightforward approach to solving this problem, but there is insufficient data to train NMT models.
no code implementations • LREC 2022 • Daisuke Suzuki, Yujin Takahashi, Ikumi Yamashita, Taichi Aida, Tosho Hirasawa, Michitaka Nakatsuji, Masato Mita, Mamoru Komachi
Therefore, in this study, we created a quality estimation dataset with manual evaluation to build an automatic evaluation model for Japanese GEC.
no code implementations • 17 Jan 2022 • Yujin Takahashi, Masahiro Kaneko, Masato Mita, Mamoru Komachi
This study investigates how supervised quality estimation (QE) models of grammatical error correction (GEC) are affected by the learners' proficiency with the data.
no code implementations • ACL 2021 • Seiichi Inoue, Taichi Aida, Mamoru Komachi, Manabu Asai
In this study, we propose a model that extends the continuous space topic model (CSTM), which flexibly controls word probability in a document, using pre-trained word embeddings.
1 code implementation • Findings (ACL) 2021 • Zhousi Chen, Longtu Zhang, Aizhan Imankulova, Mamoru Komachi
We propose two fast neural combinatory models for constituency parsing: binary and multi-branching.
2 code implementations • NAACL 2021 • Rob van der Goot, Ibrahim Sharaf, Aizhan Imankulova, Ahmet Üstün, Marija Stepanović, Alan Ramponi, Siti Oryza Khairunnisa, Mamoru Komachi, Barbara Plank
To tackle the challenge, we propose a joint learning approach, with English SLU training data and non-English auxiliary tasks from raw text, syntax and translation for transfer.
no code implementations • NAACL 2021 • Seiichiro Kondo, Kengo Hotate, Masahiro Kaneko, Mamoru Komachi
It is assumed that this issue is caused by insufficient number of long sentences in the training data.
no code implementations • NAACL 2021 • Aomi Koyama, Kengo Hotate, Masahiro Kaneko, Mamoru Komachi
Therefore, GEC studies have developed various methods to generate pseudo data, which comprise pairs of grammatical and artificially produced ungrammatical sentences.
no code implementations • COLING 2020 • Ikumi Yamashita, Satoru Katsumata, Masahiro Kaneko, Aizhan Imankulova, Mamoru Komachi
Cross-lingual transfer learning from high-resource languages (the source models) is effective for training models of low-resource languages (the target models) for various tasks.
1 code implementation • COLING 2020 • Ryoma Yoshimura, Masahiro Kaneko, Tomoyuki Kajiwara, Mamoru Komachi
We propose a reference-less metric trained on manual evaluations of system outputs for grammatical error correction (GEC).
no code implementations • COLING 2020 • Kengo Hotate, Masahiro Kaneko, Mamoru Komachi
In this study, we propose a beam search method to obtain diverse outputs in a local sequence transduction task where most of the tokens in the source and target sentences overlap, such as in grammatical error correction (GEC).
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Satoru Katsumata, Mamoru Komachi
In this study, we explored the utility of bidirectional and auto-regressive transformers (BART) as a generic pretrained encoder-decoder model for GEC.
1 code implementation • Asian Chapter of the Association for Computational Linguistics 2020 • Siti Oryza Khairunnisa, Aizhan Imankulova, Mamoru Komachi
In recent years, named entity recognition (NER) tasks in the Indonesian language have undergone extensive development.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Hongfei Wang, Michiki Kurosawa, Satoru Katsumata, Mamoru Komachi
In recent years, pre-trained models have been extensively studied, and several downstream tasks have benefited from their utilization.
1 code implementation • EAMT 2020 • YuTing Zhao, Mamoru Komachi, Tomoyuki Kajiwara, Chenhui Chu
In contrast, we propose the application of semantic image regions for MNMT by integrating visual and textual features using two individual attention mechanisms (double attention).
no code implementations • ACL 2020 • Hwichan Kim, Tosho Hirasawa, Mamoru Komachi
The primary limitation of North Korean to English translation is the lack of a parallel corpus; therefore, high translation accuracy cannot be achieved.
no code implementations • WS 2020 • Masahiro Kaneko, Aizhan Imankulova, Tosho Hirasawa, Mamoru Komachi
We introduce our TMU system that is submitted to The 4th Workshop on Neural Generation and Translation (WNGT2020) to English-to-Japanese (En→Ja) track on Simultaneous Translation And Paraphrase for Language Education (STAPLE) shared task.
no code implementations • ACL 2020 • Yujin Takahashi, Satoru Katsumata, Mamoru Komachi
To address the limitations of language and computational resources, we assume that introducing pseudo errors into sentences similar to those written by the language learners is more efficient, rather than incorporating random pseudo errors into monolingual data.
no code implementations • 23 Jun 2020 • Tosho Hirasawa, Zhishen Yang, Mamoru Komachi, Naoaki Okazaki
Video-guided machine translation as one of multimodal neural machine translation tasks targeting on generating high-quality text translation by tangibly engaging both video and text.
2 code implementations • 24 May 2020 • Satoru Katsumata, Mamoru Komachi
In this study, we explore the utility of bidirectional and auto-regressive transformers (BART) as a generic pretrained encoder-decoder model for GEC.
Ranked #16 on
Grammatical Error Correction
on CoNLL-2014 Shared Task
no code implementations • LREC 2020 • Aomi Koyama, Tomoshige Kiyuna, Kenji Kobayashi, Mio Arai, Mamoru Komachi
As our corpus has less noise and its annotation scheme reflects the characteristics of the dataset, it is ideal as an evaluation corpus for correcting grammatical errors in sentences written by JSL learners.
no code implementations • LREC 2020 • Reo Hirao, Mio Arai, Hiroki Shimanaka, Satoru Katsumata, Mamoru Komachi
In this study, we created an automated essay scoring (AES) system for nonnative Japanese learners using an essay dataset with annotations for a holistic score and multiple trait scores, including content, organization, and language scores.
1 code implementation • WMT (EMNLP) 2020 • Aizhan Imankulova, Masahiro Kaneko, Tosho Hirasawa, Mamoru Komachi
Simultaneous translation involves translating a sentence before the speaker's utterance is completed in order to realize real-time understanding in multiple languages.
no code implementations • WS 2019 • Aizhan Imankulova, Masahiro Kaneko, Mamoru Komachi
We introduce our system that is submitted to the News Commentary task (Japanese{\textless}-{\textgreater}Russian) of the 6th Workshop on Asian Translation.
no code implementations • 11 Sep 2019 • Michiki Kurosawa, Mamoru Komachi
In prior investigations, two models have been used: a translation model and a language model.
no code implementations • 2 Sep 2019 • Hayahide Yamagishi, Mamoru Komachi
We propose a weight sharing method wherein NMT saves decoder states and calculates an attention vector using the saved states when translating a current sentence.
no code implementations • WS 2019 • Ryoma Yoshimura, Hiroki Shimanaka, Yukio Matsumura, Hayahide Yamagishi, Mamoru Komachi
We use the outputs of off-the-shelf MT systems as pseudo-references filtered by paraphrasing in addition to a single human reference (gold reference).
no code implementations • WS 2019 • Mio Arai, Masahiro Kaneko, Mamoru Komachi
Existing example retrieval systems do not include grammatically incorrect examples or present only a few examples, if any.
no code implementations • WS 2019 • Masahiro Kaneko, Kengo Hotate, Satoru Katsumata, Mamoru Komachi
Thus, it is not straightforward to utilize language representations trained from a large corpus, such as Bidirectional Encoder Representations from Transformers (BERT), in a form suitable for the learner{'}s grammatical errors.
no code implementations • WS 2019 • Satoru Katsumata, Mamoru Komachi
We introduce unsupervised techniques based on phrase-based statistical machine translation for grammatical error correction (GEC) trained on a pseudo learner corpus created by Google Translation.
no code implementations • 29 Jul 2019 • Hiroki Shimanaka, Tomoyuki Kajiwara, Mamoru Komachi
We introduce the metric using BERT (Bidirectional Encoder Representations from Transformers) (Devlin et al., 2019) for automatic machine translation evaluation.
no code implementations • 23 Jul 2019 • Satoru Katsumata, Mamoru Komachi
We introduce unsupervised techniques based on phrase-based statistical machine translation for grammatical error correction (GEC) trained on a pseudo learner corpus created by Google Translation.
no code implementations • ACL 2019 • Mio Arai, Tomonori Kodaira, Mamoru Komachi
This study develops an incorrect example retrieval system, called Sakura, using a large-scale Lang-8 dataset for Japanese language learners.
no code implementations • ACL 2019 • Kengo Hotate, Masahiro Kaneko, Satoru Katsumata, Mamoru Komachi
In this paper, we propose a method for neural grammar error correction (GEC) that can control the degree of correction.
no code implementations • WS 2019 • Tosho Hirasawa, Mamoru Komachi
In this study, we examine various kinds of word embeddings and introduce two debiasing techniques for three multimodal NMT models and two language pairs -- English-German translation and English-French translation.
no code implementations • 15 Apr 2019 • Masahiro Kaneko, Mamoru Komachi
In this work, we investigate the effect of utilizing information not only from the final layer but also from intermediate layers of a pre-trained language representation model to detect grammatical errors.
no code implementations • NAACL 2019 • Hikaru Omori, Mamoru Komachi
An event-noun is a noun that has an argument structure similar to a predicate.
1 code implementation • NAACL 2019 • Tosho Hirasawa, Hayahide Yamagishi, Yukio Matsumura, Mamoru Komachi
Multimodal machine translation is an attractive application of neural machine translation (NMT).
no code implementations • 1 Mar 2019 • Longtu Zhang, Mamoru Komachi
Unsupervised neural machine translation (UNMT) requires only monolingual data of similar language pairs during training and can produce bi-directional translation models with relatively good performance on alphabetic languages (Lample et al., 2018).
no code implementations • 29 Jan 2019 • Tomoya Ogata, Mamoru Komachi, Tomoya Takatani
We propose a task to generate a complex sentence from a simple sentence in order to amplify various kinds of responses in the database.
1 code implementation • WS 2018 • Hiroki Shimanaka, Tomoyuki Kajiwara, Mamoru Komachi
We introduce the RUSE metric for the WMT18 metrics shared task.
no code implementations • WS 2018 • Longtu Zhang, Mamoru Komachi
Recent neural machine translation (NMT) systems have been greatly improved by encoder-decoder models with attention mechanisms and sub-word units.
no code implementations • PACLIC 2018 • Tomonori Kodaira, Mamoru Komachi
Neural network-based approaches have become widespread for abstractive text summarization.
no code implementations • 7 Sep 2018 • Longtu Zhang, Mamoru Komachi
Recent neural machine translation (NMT) systems have been greatly improved by encoder-decoder models with attention mechanisms and sub-word units.
no code implementations • WS 2018 • Tomoyuki Kajiwara, Mamoru Komachi
We introduce the TMU systems for the Complex Word Identification (CWI) Shared Task 2018.
1 code implementation • WS 2018 • Masahiro Kaneko, Tomoyuki Kajiwara, Mamoru Komachi
We introduce the TMU systems for the second language acquisition modeling shared task 2018 (Settles et al., 2018).
1 code implementation • ACL 2018 • Satoru Katsumata, Yukio Matsumura, Hayahide Yamagishi, Mamoru Komachi
For Japanese-to-English translation, this method achieves a BLEU score that is 0. 56 points more than that of a baseline.
no code implementations • NAACL 2018 • Michiki Kurosawa, Yukio Matsumura, Hayahide Yamagishi, Mamoru Komachi
Neural machine translation (NMT) has a drawback in that can generate only high-frequency words owing to the computational costs of the softmax function in the output layer.
no code implementations • NAACL 2018 • Hiroki Shimanaka, Tomoyuki Kajiwara, Mamoru Komachi
Sentence representations can capture a wide range of information that cannot be captured by local features based on character or word N-grams.
no code implementations • WS 2017 • Kent Shioda, Mamoru Komachi, Rue Ikeya, Daichi Mochihashi
This method implicitly models latent intent of query and sentences, and alleviates the problem of noisy alignment.
no code implementations • WS 2017 • Yukio Matsumura, Mamoru Komachi
We implemented beam search and ensemble decoding in the NMT system.
1 code implementation • IJCNLP 2017 • Tomoyuki Kajiwara, Mamoru Komachi, Daichi Mochihashi
We present a pointwise mutual information (PMI)-based approach to formalize paraphrasability and propose a variant of PMI, called MIPA, for the paraphrase acquisition.
1 code implementation • WS 2017 • Aizhan Imankulova, Takayuki Sato, Mamoru Komachi
Large-scale parallel corpora are indispensable to train highly accurate machine translators.
no code implementations • IJCNLP 2017 • Hayahide Yamagishi, Shin Kanouchi, Takayuki Sato, Mamoru Komachi
This study reports an attempt to predict the voice of reference using the information from the input sentences or previous input/output sentences.
1 code implementation • IJCNLP 2017 • Masahiro Kaneko, Yuya Sakaizawa, Mamoru Komachi
In this study, we improve grammatical error detection by learning word embeddings that consider grammaticality and error patterns.
Ranked #6 on
Grammatical Error Detection
on FCE
no code implementations • WS 2017 • Yuuki Sekizawa, Tomoyuki Kajiwara, Mamoru Komachi
Neural machine translation (NMT) produces sentences that are more fluent than those produced by statistical machine translation (SMT).
no code implementations • PACLIC 2018 • Yoshiaki Kitagawa, Mamoru Komachi
This study presents a Long Short-Term Memory (LSTM) neural network approach to Japanese word segmentation (JWS).
Ranked #3 on
Japanese Word Segmentation
on BCCWJ
1 code implementation • 26 Jun 2017 • Yukio Matsumura, Takayuki Sato, Mamoru Komachi
We confirm that our re-implementation also shows the same tendency and alleviates the problem of repeating and missing words in the translation on a English-Japanese task too.
no code implementations • PACLIC 2018 • Ryosuke Miyazaki, Mamoru Komachi
Previous approaches to training syntax-based sentiment classification models required phrase-level annotated corpora, which are not readily available in many languages other than English.
no code implementations • 2 Apr 2017 • Junki Matsuo, Mamoru Komachi, Katsuhito Sudoh
One of the most important problems in machine translation (MT) evaluation is to evaluate the similarity between translation hypotheses with different surface forms from the reference, especially at the segment level.
2 code implementations • LREC 2018 • Yuya Sakaizawa, Mamoru Komachi
An evaluation of distributed word representation is generally conducted using a word similarity task and/or a word analogy task.
no code implementations • 15 Mar 2017 • Ai Hirata, Mamoru Komachi
Named entity classification is the task of classifying text-based elements into various categories, including places, names, dates, times, and monetary values.
no code implementations • COLING 2016 • Tomoyuki Kajiwara, Mamoru Komachi
To obviate the need for human annotation, we propose an unsupervised method that automatically builds the monolingual parallel corpus for text simplification using sentence similarity based on word embeddings.
no code implementations • WS 2016 • Yasunobu Asakura, Masatsugu Hangyo, Mamoru Komachi
Information extraction from user-generated text has gained much attention with the growth of the Web. Disaster analysis using information from social media provides valuable, real-time, geolocation information for helping people caught up these in disasters.
no code implementations • WS 2016 • Takayuki Sato, Jun Harashima, Mamoru Komachi
However, little work has been done on machine translation of recipe texts.
Cultural Vocal Bursts Intensity Prediction
Information Retrieval
+2
no code implementations • WS 2016 • Shin Kanouchi, Katsuhito Sudoh, Mamoru Komachi
This paper presents an improved lexicalized reordering model for phrase-based statistical machine translation using a deep neural network.
no code implementations • WS 2016 • Hayahide Yamagishi, Shin Kanouchi, Takayuki Sato, Mamoru Komachi
The results showed that, we could control the voice of the generated sentence with 85. 0{\%} accuracy on average.
no code implementations • LREC 2016 • Ryuichi Tachibana, Mamoru Komachi
Therefore, we propose a new correctable word-typing game that is more similar to the actual writing process.
no code implementations • LREC 2012 • Toshinobu Ogiso, Mamoru Komachi, Yasuharu Den, Yuji Matsumoto
In order to construct an annotated diachronic corpus of Japanese, we propose to create a new dictionary for morphological analysis of Early Middle Japanese (Classical Japanese) based on UniDic, a dictionary for Contemporary Japanese.