1 code implementation • COLING 2022 • Koki Maeda, Masahiro Kaneko, Naoaki Okazaki
Correlations between IMPARA and human scores indicate that IMPARA is comparable or better than existing evaluation methods.
no code implementations • COLING 2022 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
We study the relationship between task-agnostic intrinsic and task-specific extrinsic social bias evaluation measures for MLMs, and find that there exists only a weak correlation between these two types of evaluation measures.
no code implementations • LREC 2022 • Sangwhan Moon, Won Ik Cho, Hye Joo Han, Naoaki Okazaki, Nam Soo Kim
As this problem originates from the conventional scheme used when creating a POS tagging corpus, we propose an improvement to the existing scheme, which makes it friendlier to generative tasks.
1 code implementation • spnlp (ACL) 2022 • Youmi Ma, Tatsuya Hiraoka, Naoaki Okazaki
We adopt table representations to model the entities and relations, casting the entity and relation extraction as a table-labeling problem.
1 code implementation • Findings (ACL) 2022 • Tatsuya Hiraoka, Sho Takase, Kei Uchiumi, Atsushi Keyaki, Naoaki Okazaki
We present two simple modifications for word-level perturbation: Word Replacement considering Length (WR-L) and Compositional Word Replacement (CWR). In conventional word replacement, a word in an input is replaced with a word sampled from the entire vocabulary, regardless of the length and context of the target word. WR-L considers the length of a target word by sampling words from the Poisson distribution. CWR considers the compositional candidates by restricting the source of sampling to related words that appear in subword regularization. Experimental results showed that the combination of WR-L and CWR improved the performance of text classification and machine translation.
no code implementations • EMNLP 2020 • Sangwhan Moon, Naoaki Okazaki
Large scale pre-trained language models have shown groundbreaking performance improvements for transfer learning in the domain of natural language processing.
no code implementations • INLG (ACL) 2021 • Ayana Niwa, Keisuke Nishiguchi, Naoaki Okazaki
We address the task of antonym prediction in a context, which is a fill-in-the-blanks problem.
no code implementations • MTSummit 2021 • Keiji Yasuda, Ichiro Yamada, Naoaki Okazaki, Hideki Tanaka, Hidehiro Asaka, Takeshi Anzai, Fumiaki Sugaya
The second technology is machine translation (MT), which enables users to read foreign news articles in their mother language.
no code implementations • 19 May 2023 • Masahiro Kaneko, Graham Neubig, Naoaki Okazaki
Humans work together to solve common problems by having discussions, explaining, and agreeing or disagreeing with each other.
no code implementations • 19 May 2023 • Masahiro Kaneko, Naoaki Okazaki
This paper proposes to predict a set of edit operations for the source text for local sequence transduction tasks.
no code implementations • 22 Apr 2023 • Sakae Mizuki, Naoaki Okazaki
A promising approach for knowledge-based Word Sense Disambiguation (WSD) is to select the sense whose contextualized embeddings computed for its definition sentence are closest to those computed for a target word in a given sentence.
1 code implementation • 17 Feb 2023 • Youmi Ma, An Wang, Naoaki Okazaki
First, we propose DREEAM, a memory-efficient approach that adopts evidence information as the supervisory signal, thereby guiding the attention modules of the DocRE system to assign high weights to evidence.
Ranked #1 on
Relation Extraction
on ReDocRED
no code implementations • 28 Jan 2023 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
Prior works have relied on human annotated examples to compare existing intrinsic bias evaluation measures.
no code implementations • 8 Nov 2022 • Hiroki Iida, Naoaki Okazaki
We conducted experiments using our method on datasets with a large vocabulary gap from a source domain.
no code implementations • 6 Oct 2022 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
We study the relationship between task-agnostic intrinsic and task-specific extrinsic social bias evaluation measures for Masked Language Models (MLMs), and find that there exists only a weak correlation between these two types of evaluation measures.
no code implementations • 26 Aug 2022 • Ayana Niwa, Sho Takase, Naoaki Okazaki
In addition, the proposed method outperforms an NAR baseline on the WMT'14 En-De dataset.
no code implementations • 27 Jul 2022 • Mengsay Loem, Sho Takase, Masahiro Kaneko, Naoaki Okazaki
Impressive performance of Transformer has been attributed to self-attention, where dependencies between entire input in a sequence are considered at every position.
1 code implementation • 25 May 2022 • Ao Liu, Haoyu Dong, Naoaki Okazaki, Shi Han, Dongmei Zhang
However, directly learning the logical inference knowledge from table-text pairs is very difficult for neural models because of the ambiguity of natural language and the scarcity of parallel data.
no code implementations • 19 May 2022 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
Different methods have been proposed to develop meta-embeddings from a given set of source embeddings.
1 code implementation • NAACL 2022 • Masahiro Kaneko, Aizhan Imankulova, Danushka Bollegala, Naoaki Okazaki
Unfortunately, it was reported that MLMs also learn discriminative biases regarding attributes such as gender and race.
no code implementations • Findings (ACL) 2022 • Sho Takase, Tatsuya Hiraoka, Naoaki Okazaki
Subword regularizations use multiple subword segmentations during training to improve the robustness of neural machine translation models.
1 code implementation • ACL 2022 • Ao Liu, An Wang, Naoaki Okazaki
In this work, we propose a simple yet effective semi-supervised framework to better utilize source-side unlabeled sentences based on consistency training.
Ranked #1 on
Formality Style Transfer
on GYAFC
(using extra training data)
1 code implementation • ACL 2022 • Masahiro Kaneko, Sho Takase, Ayana Niwa, Naoaki Okazaki
In this study, we introduce an Example-Based GEC (EB-GEC) that presents examples to language learners as a basis for a correction result.
no code implementations • LREC 2022 • Hwichan Kim, Sangwhan Moon, Naoaki Okazaki, Mamoru Komachi
Training a model using North Korean data is the most straightforward approach to solving this problem, but there is insufficient data to train NMT models.
no code implementations • NAACL (ACL) 2022 • Mengsay Loem, Sho Takase, Masahiro Kaneko, Naoaki Okazaki
Through experiments, we show that ExtraPhrase improves the performance of abstractive summarization tasks by more than 0. 50 points in ROUGE scores compared to the setting without data augmentation.
no code implementations • 12 Dec 2021 • Ao Liu, Congjian Luo, Naoaki Okazaki
We further introduce logical form generation (LG), a dual task of Logic2text that requires generating a valid logical form based on a text description of a table.
1 code implementation • EMNLP 2021 • Kosuke Yamada, Yuta Hitomi, Hideaki Tamori, Ryohei Sasano, Naoaki Okazaki, Kentaro Inui, Koichi Takeda
We also consider a new headline generation strategy that takes advantage of the controllable generation order of Transformer.
no code implementations • AKBC 2021 • Wiem Ben Rim, Carolin Lawrence, Kiril Gashteovski, Mathias Niepert, Naoaki Okazaki
With an extensive set of experiments, we perform and analyze these tests for several KGE models.
2 code implementations • Findings (ACL) 2021 • Tatsuya Hiraoka, Sho Takase, Kei Uchiumi, Atsushi Keyaki, Naoaki Okazaki
Since traditional tokenizers are isolated from a downstream task and model, they cannot output an appropriate tokenization depending on the task and model, although recent studies imply that the appropriate tokenization improves the performance.
no code implementations • SEMEVAL 2020 • Zhishen Yang, Lars Wolfsteller, Naoaki Okazaki
This paper describes the emphasis selection system of the team TextLearner for SemEval 2020 Task 10: Emphasis Selection For Written Text in Visual Media.
1 code implementation • COLING 2020 • Zhishen Yang, Naoaki Okazaki
In this paper, we address the task of news-image captioning, which generates a description of an image given the image and its article body as input.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Shin Kanouchi, Masato Neishi, Yuta Hayashibe, Hiroki Ouchi, Naoaki Okazaki
To tackle the challenge, this paper proposes a novel and practical task to explain evidences in recommending hotels given vague requests expressed freely in natural language.
no code implementations • SEMEVAL 2020 • Wiem Ben Rim, Naoaki Okazaki
We describe the system submitted by the SWAGex team to the SemEval-2020 Commonsense Validation and Explanation Task.
2 code implementations • 30 Nov 2020 • Emanuele Bugliarello, Ryan Cotterell, Naoaki Okazaki, Desmond Elliott
Large-scale pretraining and task-specific fine-tuning is now the standard methodology for many tasks in computer vision and natural language processing.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Tatsuya Hiraoka, Sho Takase, Kei Uchiumi, Atsushi Keyaki, Naoaki Okazaki
In traditional NLP, we tokenize a given sentence as a preprocessing, and thus the tokenization is unrelated to a target downstream task.
1 code implementation • Journal of Natural Language Processing 2022 • Youmi Ma, Tatsuya Hiraoka, Naoaki Okazaki
In this study, a novel method for extracting named entities and relations from unstructured text based on the table representation is presented.
Ranked #3 on
Relation Extraction
on CoNLL04
(NER Micro F1 metric)
no code implementations • LREC 2022 • Sho Takase, Naoaki Okazaki
Experimental results indicate that Transum improves the performance from the strong baseline, Transformer, in Chinese-English, Arabic-English, and English-Japanese translation datasets.
Abstractive Text Summarization
Cross-Lingual Abstractive Summarization
+3
no code implementations • 23 Jun 2020 • Tosho Hirasawa, Zhishen Yang, Mamoru Komachi, Naoaki Okazaki
Video-guided machine translation as one of multimodal neural machine translation tasks targeting on generating high-quality text translation by tangibly engaging both video and text.
1 code implementation • ACL 2020 • Emanuele Bugliarello, Sabrina J. Mielke, Antonios Anastasopoulos, Ryan Cotterell, Naoaki Okazaki
The performance of neural machine translation systems is commonly evaluated in terms of BLEU.
1 code implementation • ACL 2020 • Kazuki Matsumaru, Sho Takase, Naoaki Okazaki
Building a binary classifier that predicts an entailment relation between an article and its headline, we filter out untruthful instances from the supervision data.
no code implementations • LREC 2020 • Sho Shimazu, Sho Takase, Toshiaki Nakazawa, Naoaki Okazaki
Therefore, we present a hand-crafted dataset to evaluate whether translation models can resolve the zero pronoun problems in Japanese to English translations.
no code implementations • LREC 2020 • Sangwhan Moon, Naoaki Okazaki
In the context of multilingual language model pre-training, vocabulary size for languages with a broad set of potential characters is an unsolved problem.
no code implementations • WS 2019 • Yuichi Sasazawa, Sho Takase, Naoaki Okazaki
One of the key requirements of QG is to generate a question such that it results in a target answer.
1 code implementation • ACL 2020 • Emanuele Bugliarello, Naoaki Okazaki
Most neural machine translation models only rely on pairs of parallel sentences, assuming syntactic information is automatically learned by an attention mechanism.
2 code implementations • ACL 2019 • Hayate Iso, Yui Uehara, Tatsuya Ishigaki, Hiroshi Noji, Eiji Aramaki, Ichiro Kobayashi, Yusuke Miyao, Naoaki Okazaki, Hiroya Takamura
We propose a data-to-text generation model with two modules, one for tracking and the other for text generation.
no code implementations • SEMEVAL 2019 • Zhishen Yang, Sam Vijlbrief, Naoaki Okazaki
This paper presents our contextual emotion detection system in approaching the SemEval2019 shared task 3: EmoContext: Contextual Emotion Detection in Text.
1 code implementation • NAACL 2019 • Sho Takase, Naoaki Okazaki
Neural encoder-decoder models have been successful in natural language generation tasks.
Ranked #2 on
Text Summarization
on DUC 2004 Task 1
no code implementations • WS 2019 • Yuta Hitomi, Yuya Taguchi, Hideaki Tamori, Ko Kikuta, Jiro Nishitoba, Naoaki Okazaki, Kentaro Inui, Manabu Okumura
However, because there is no corpus of headlines of multiple lengths for a given article, previous research on controlling output length in headline generation has not discussed whether the system outputs could be adequately evaluated without multiple references of different lengths.
no code implementations • WS 2018 • Shun Kiyono, Sho Takase, Jun Suzuki, Naoaki Okazaki, Kentaro Inui, Masaaki Nagata
Developing a method for understanding the inner workings of black-box neural methods is an important research endeavor.
no code implementations • WS 2018 • Diana Galvan, Naoaki Okazaki, Koji Matsuda, Kentaro Inui
Temporal reasoning remains as an unsolved task for Natural Language Processing (NLP), particularly demonstrated in the clinical domain.
no code implementations • COLING 2018 • Akira Sasaki, Kazuaki Hanawa, Naoaki Okazaki, Kentaro Inui
This paper presents an approach to detect the stance of a user toward a topic based on their stances toward other topics and the social media posts of the user.
no code implementations • 22 Dec 2017 • Shun Kiyono, Sho Takase, Jun Suzuki, Naoaki Okazaki, Kentaro Inui, Masaaki Nagata
The encoder-decoder model is widely used in natural language generation tasks.
no code implementations • 7 Dec 2017 • Paul Reisert, Naoya Inoue, Naoaki Okazaki, Kentaro Inui
Our coverage result of 74. 6% indicates that argumentative relations can reasonably be explained by our small pattern set.
no code implementations • IJCNLP 2017 • Yuta Hitomi, Hideaki Tamori, Naoaki Okazaki, Kentaro Inui
This paper explores the idea of robot editors, automated proofreaders that enable journalists to improve the quality of their articles.
1 code implementation • IJCNLP 2017 • Sosuke Kobayashi, Naoaki Okazaki, Kentaro Inui
This study addresses the problem of identifying the meaning of unknown words or entities in a discourse with respect to the word embedding approaches used in neural language models.
no code implementations • WS 2017 • Hideaki Tamori, Yuta Hitomi, Naoaki Okazaki, Kentaro Inui
We address the issue of the quality of journalism and analyze daily article revision logs from a Japanese newspaper company.
1 code implementation • ACL 2016 • Sho Takase, Naoaki Okazaki, Kentaro Inui
Learning distributed representations for relation instances is a central technique in downstream NLP applications.
no code implementations • ACL 2017 • Akira Sasaki, Kazuaki Hanawa, Naoaki Okazaki, Kentaro Inui
We present in this paper our approach for modeling inter-topic preferences of Twitter users: for example, those who agree with the Trans-Pacific Partnership (TPP) also agree with free trade.
no code implementations • COLING 2016 • Kento Watanabe, Yuichiroh Matsubayashi, Naho Orita, Naoaki Okazaki, Kentaro Inui, Satoru Fukayama, Tomoyasu Nakano, Jordan Smith, Masataka Goto
This study proposes a computational model of the discourse segments in lyrics to understand and to model the structure of lyrics.
no code implementations • COLING 2016 • Natsuda Laokulrat, Sang Phan, Noriki Nishida, Raphael Shu, Yo Ehara, Naoaki Okazaki, Yusuke Miyao, Hideki Nakayama
Automatic video description generation has recently been getting attention after rapid advancement in image caption generation.
no code implementations • COLING 2016 • Naoya Inoue, Yuichiroh Matsubayashi, Masayuki Ono, Naoaki Okazaki, Kentaro Inui
This paper proposes a novel problem setting of selectional preference (SP) between a predicate and its arguments, called as context-sensitive SP (CSP).
1 code implementation • ACL 2016 • Ran Tian, Naoaki Okazaki, Kentaro Inui
This paper connects a vector-based composition model to a formal semantics, the Dependency-based Compositional Semantics (DCS).
no code implementations • 26 Nov 2015 • Ran Tian, Naoaki Okazaki, Kentaro Inui
Additive composition (Foltz et al, 1998; Landauer and Dumais, 1997; Mitchell and Lapata, 2010) is a widely used method for computing meanings of phrases, which takes the average of vector representations of the constituent words.