no code implementations • INLG (ACL) 2021 • Ayana Niwa, Keisuke Nishiguchi, Naoaki Okazaki
We address the task of antonym prediction in a context, which is a fill-in-the-blanks problem.
no code implementations • COLING 2022 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
We study the relationship between task-agnostic intrinsic and task-specific extrinsic social bias evaluation measures for MLMs, and find that there exists only a weak correlation between these two types of evaluation measures.
no code implementations • LREC 2022 • Sangwhan Moon, Won Ik Cho, Hye Joo Han, Naoaki Okazaki, Nam Soo Kim
As this problem originates from the conventional scheme used when creating a POS tagging corpus, we propose an improvement to the existing scheme, which makes it friendlier to generative tasks.
no code implementations • MTSummit 2021 • Keiji Yasuda, Ichiro Yamada, Naoaki Okazaki, Hideki Tanaka, Hidehiro Asaka, Takeshi Anzai, Fumiaki Sugaya
The second technology is machine translation (MT), which enables users to read foreign news articles in their mother language.
no code implementations • EMNLP 2020 • Sangwhan Moon, Naoaki Okazaki
Large scale pre-trained language models have shown groundbreaking performance improvements for transfer learning in the domain of natural language processing.
1 code implementation • COLING 2022 • Koki Maeda, Masahiro Kaneko, Naoaki Okazaki
Correlations between IMPARA and human scores indicate that IMPARA is comparable or better than existing evaluation methods.
1 code implementation • spnlp (ACL) 2022 • Youmi Ma, Tatsuya Hiraoka, Naoaki Okazaki
We adopt table representations to model the entities and relations, casting the entity and relation extraction as a table-labeling problem.
1 code implementation • Findings (ACL) 2022 • Tatsuya Hiraoka, Sho Takase, Kei Uchiumi, Atsushi Keyaki, Naoaki Okazaki
We present two simple modifications for word-level perturbation: Word Replacement considering Length (WR-L) and Compositional Word Replacement (CWR). In conventional word replacement, a word in an input is replaced with a word sampled from the entire vocabulary, regardless of the length and context of the target word. WR-L considers the length of a target word by sampling words from the Poisson distribution. CWR considers the compositional candidates by restricting the source of sampling to related words that appear in subword regularization. Experimental results showed that the combination of WR-L and CWR improved the performance of text classification and machine translation.
no code implementations • 21 Aug 2024 • Marco Cognetta, Vilém Zouhar, Naoaki Okazaki
Subword regularization, used widely in NLP, improves model performance by reducing the dependency on exact tokenizations, augmenting the training corpus, and exposing the model to more unique contexts during training.
no code implementations • 20 Aug 2024 • An Wang, Xingwu Sun, Ruobing Xie, Shuaipeng Li, Jiaqi Zhu, Zhen Yang, Pinxue Zhao, J. N. Han, Zhanhui Kang, Di Wang, Naoaki Okazaki, Cheng-Zhong Xu
To address the imbalance in expert activation, we propose a novel training objective that encourages the frequent activation of smaller experts, enhancing computational efficiency and parameter utilization.
no code implementations • 4 Jul 2024 • LLM-jp, :, Akiko Aizawa, Eiji Aramaki, Bowen Chen, Fei Cheng, Hiroyuki Deguchi, Rintaro Enomoto, Kazuki Fujii, Kensuke Fukumoto, Takuya Fukushima, Namgi Han, Yuto Harada, Chikara Hashimoto, Tatsuya Hiraoka, Shohei Hisada, Sosuke Hosokawa, Lu Jie, Keisuke Kamata, Teruhito Kanazawa, Hiroki Kanezashi, Hiroshi Kataoka, Satoru Katsumata, Daisuke Kawahara, Seiya Kawano, Atsushi Keyaki, Keisuke Kiryu, Hirokazu Kiyomaru, Takashi Kodama, Takahiro Kubo, Yohei Kuga, Ryoma Kumon, Shuhei Kurita, Sadao Kurohashi, Conglong Li, Taiki Maekawa, Hiroshi Matsuda, Yusuke Miyao, Kentaro Mizuki, Sakae Mizuki, Yugo Murawaki, Ryo Nakamura, Taishi Nakamura, Kouta Nakayama, Tomoka Nakazato, Takuro Niitsuma, Jiro Nishitoba, Yusuke Oda, Hayato Ogawa, Takumi Okamoto, Naoaki Okazaki, Yohei Oseki, Shintaro Ozaki, Koki Ryu, Rafal Rzepka, Keisuke Sakaguchi, Shota Sasaki, Satoshi Sekine, Kohei Suda, Saku Sugawara, Issa Sugiura, Hiroaki Sugiyama, Hisami Suzuki, Jun Suzuki, Toyotaro Suzumura, Kensuke Tachibana, Yu Takagi, Kyosuke Takami, Koichi Takeda, Masashi Takeshita, Masahiro Tanaka, Kenjiro Taura, Arseny Tolmachev, Nobuhiro Ueda, Zhen Wan, Shuntaro Yada, Sakiko Yahata, Yuya Yamamoto, Yusuke Yamauchi, Hitomi Yanaka, Rio Yokota, Koichiro Yoshino
This paper introduces LLM-jp, a cross-organizational project for the research and development of Japanese large language models (LLMs).
1 code implementation • 3 Jul 2024 • Rem Hida, Masahiro Kaneko, Naoaki Okazaki
In this paper, we investigate the sensitivity of LLMs when changing prompt variations (task instruction and prompt, few-shot examples, debias-prompt) by analyzing task performance and social bias of LLMs.
no code implementations • 27 Apr 2024 • Naoaki Okazaki, Kakeru Hattori, Hirai Shota, Hiroki Iida, Masanari Ohi, Kazuki Fujii, Taishi Nakamura, Mengsay Loem, Rio Yokota, Sakae Mizuki
Open Japanese large language models (LLMs) have been trained on the Japanese portions of corpora such as CC-100, mC4, and OSCAR.
no code implementations • 27 Apr 2024 • Kazuki Fujii, Taishi Nakamura, Mengsay Loem, Hiroki Iida, Masanari Ohi, Kakeru Hattori, Hirai Shota, Sakae Mizuki, Rio Yokota, Naoaki Okazaki
The results showed that the efficiency gained through vocabulary expansion had no negative impact on performance, except for the summarization task, and that the combined use of parallel corpora enhanced translation ability.
no code implementations • 25 Apr 2024 • Youmi Ma, An Wang, Naoaki Okazaki
In our proposal, annotators edit relation predictions from a model trained on the transferred dataset.
1 code implementation • 17 Apr 2024 • Masahiro Kaneko, Youmi Ma, Yuki Wata, Naoaki Okazaki
In this study, we propose a Sampling-based Pseudo-Likelihood (\textbf{SPL}) method for MIA (\textbf{SaMIA}) that calculates SPL using only the text generated by an LLM to detect leaks.
no code implementations • 30 Mar 2024 • Marco Cognetta, Tatsuya Hiraoka, Naoaki Okazaki, Rico Sennrich, Yuval Pinter
We explore threshold vocabulary trimming in Byte-Pair Encoding subword tokenization, a postprocessing step that replaces rare subwords with their component subwords.
no code implementations • 28 Feb 2024 • Koki Maeda, Shuhei Kurita, Taiki Miyanishi, Naoaki Okazaki
Given the accelerating progress of vision and language modeling, accurate evaluation of machine-generated image captions remains critical.
no code implementations • 25 Feb 2024 • Masanari Ohi, Masahiro Kaneko, Ryuto Koike, Mengsay Loem, Naoaki Okazaki
In this paper, we investigate the presence and impact of likelihood bias in LLM-based evaluators.
no code implementations • 22 Feb 2024 • Marco Cognetta, Vilém Zouhar, Sangwhan Moon, Naoaki Okazaki
In Tokenization and the Noiseless Channel (Zouhar et al., 2023a), R\'enyi efficiency is suggested as an intrinsic mechanism for evaluating a tokenizer: for NLP tasks, the tokenizer which leads to the highest R\'enyi efficiency of the unigram distribution should be chosen.
no code implementations • 15 Feb 2024 • Tatsuya Hiraoka, Naoaki Okazaki
Do pretrained language models have knowledge regarding the surface information of tokens?
no code implementations • 28 Jan 2024 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki, Timothy Baldwin
In this study, we examine the impact of LLMs' step-by-step predictions on gender bias in unscalable tasks.
no code implementations • 14 Nov 2023 • Mengsay Loem, Masahiro Kaneko, Naoaki Okazaki
Large Language Models (LLMs) can justify or critique their predictions through discussions with other models or humans, thereby enriching their intrinsic understanding of instances.
1 code implementation • 14 Nov 2023 • Ryuto Koike, Masahiro Kaneko, Naoaki Okazaki
Finally, our analysis indicates that the high instruction-following ability of LLMs fosters the large impact of such constraints on detection performance.
no code implementations • 9 Oct 2023 • Trang Nguyen, Naoaki Okazaki
Besides, diverse interpretations of the input lead to various modes of answer generation, highlighting the role of causal reasoning between interpreting and answering steps in VQA.
1 code implementation • 20 Sep 2023 • Masahiro Kaneko, Naoaki Okazaki
Generating explanations for GEC corrections involves aligning input and output tokens, identifying correction points, and presenting corresponding explanations consistently.
1 code implementation • 18 Sep 2023 • Panatchakorn Anantaprayoon, Masahiro Kaneko, Naoaki Okazaki
In Natural Language Inference (NLI), existing bias evaluation methods have focused on the prediction results of one specific label out of three labels, such as neutral.
no code implementations • 16 Sep 2023 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
In this study, we compare the impact of debiasing on performance across multiple downstream tasks using a wide-range of benchmark datasets that containing female, male, and stereotypical words.
1 code implementation • 21 Jul 2023 • Ryuto Koike, Masahiro Kaneko, Naoaki Okazaki
Experiments in the domain of student essays show that the proposed detector improves the detection performance on the attacker-generated texts by up to +41. 3 points F1-score.
1 code implementation • starsem 2023 • An Wang, Junfeng Jiang, Youmi Ma, Ao Liu, Naoaki Okazaki
Aspect sentiment quad prediction (ASQP) analyzes the aspect terms, opinion terms, sentiment polarity, and aspect categories in a text.
Ranked #3 on Aspect-Based Sentiment Analysis (ABSA) on ASQP
1 code implementation • 6 Jun 2023 • Zhishen Yang, Raj Dabre, Hideki Tanaka, Naoaki Okazaki
Automating figure caption generation helps move model understandings of scientific documents beyond text and will help authors write informative captions that facilitate communicating scientific findings.
no code implementations • 29 May 2023 • Mengsay Loem, Masahiro Kaneko, Sho Takase, Naoaki Okazaki
Large-scale pre-trained language models such as GPT-3 have shown remarkable performance across various natural language processing tasks.
1 code implementation • 19 May 2023 • Masahiro Kaneko, Graham Neubig, Naoaki Okazaki
Humans work together to solve common problems by having discussions, explaining, and agreeing or disagreeing with each other.
no code implementations • 19 May 2023 • Masahiro Kaneko, Naoaki Okazaki
Experiments show that the proposed method achieves comparable performance to the baseline in four tasks, paraphrasing, formality style transfer, GEC, and text simplification, despite reducing the length of the target text by as small as 21%.
1 code implementation • 22 Apr 2023 • Sakae Mizuki, Naoaki Okazaki
A promising approach for knowledge-based Word Sense Disambiguation (WSD) is to select the sense whose contextualized embeddings computed for its definition sentence are closest to those computed for a target word in a given sentence.
1 code implementation • 17 Feb 2023 • Youmi Ma, An Wang, Naoaki Okazaki
First, we propose DREEAM, a memory-efficient approach that adopts evidence information as the supervisory signal, thereby guiding the attention modules of the DocRE system to assign high weights to evidence.
Ranked #1 on Relation Extraction on ReDocRED
no code implementations • 28 Jan 2023 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
Prior works have relied on human annotated examples to compare existing intrinsic bias evaluation measures.
1 code implementation • 8 Nov 2022 • Hiroki Iida, Naoaki Okazaki
We conducted experiments using our method on datasets with a large vocabulary gap from a source domain.
no code implementations • 6 Oct 2022 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
We study the relationship between task-agnostic intrinsic and task-specific extrinsic social bias evaluation measures for Masked Language Models (MLMs), and find that there exists only a weak correlation between these two types of evaluation measures.
no code implementations • 26 Aug 2022 • Ayana Niwa, Sho Takase, Naoaki Okazaki
In addition, the proposed method outperforms an NAR baseline on the WMT'14 En-De dataset.
no code implementations • 27 Jul 2022 • Mengsay Loem, Sho Takase, Masahiro Kaneko, Naoaki Okazaki
Impressive performance of Transformer has been attributed to self-attention, where dependencies between entire input in a sequence are considered at every position.
1 code implementation • 25 May 2022 • Ao Liu, Haoyu Dong, Naoaki Okazaki, Shi Han, Dongmei Zhang
However, directly learning the logical inference knowledge from table-text pairs is very difficult for neural models because of the ambiguity of natural language and the scarcity of parallel data.
no code implementations • 19 May 2022 • Masahiro Kaneko, Danushka Bollegala, Naoaki Okazaki
Different methods have been proposed to develop meta-embeddings from a given set of source embeddings.
1 code implementation • NAACL 2022 • Masahiro Kaneko, Aizhan Imankulova, Danushka Bollegala, Naoaki Okazaki
Unfortunately, it was reported that MLMs also learn discriminative biases regarding attributes such as gender and race.
no code implementations • Findings (ACL) 2022 • Sho Takase, Tatsuya Hiraoka, Naoaki Okazaki
Subword regularizations use multiple subword segmentations during training to improve the robustness of neural machine translation models.
1 code implementation • ACL 2022 • Ao Liu, An Wang, Naoaki Okazaki
In this work, we propose a simple yet effective semi-supervised framework to better utilize source-side unlabeled sentences based on consistency training.
Ranked #1 on Formality Style Transfer on GYAFC (using extra training data)
1 code implementation • ACL 2022 • Masahiro Kaneko, Sho Takase, Ayana Niwa, Naoaki Okazaki
In this study, we introduce an Example-Based GEC (EB-GEC) that presents examples to language learners as a basis for a correction result.
no code implementations • LREC 2022 • Hwichan Kim, Sangwhan Moon, Naoaki Okazaki, Mamoru Komachi
Training a model using North Korean data is the most straightforward approach to solving this problem, but there is insufficient data to train NMT models.
no code implementations • NAACL (ACL) 2022 • Mengsay Loem, Sho Takase, Masahiro Kaneko, Naoaki Okazaki
Through experiments, we show that ExtraPhrase improves the performance of abstractive summarization tasks by more than 0. 50 points in ROUGE scores compared to the setting without data augmentation.
no code implementations • 12 Dec 2021 • Ao Liu, Congjian Luo, Naoaki Okazaki
We further introduce logical form generation (LG), a dual task of Logic2text that requires generating a valid logical form based on a text description of a table.
1 code implementation • EMNLP 2021 • Kosuke Yamada, Yuta Hitomi, Hideaki Tamori, Ryohei Sasano, Naoaki Okazaki, Kentaro Inui, Koichi Takeda
We also consider a new headline generation strategy that takes advantage of the controllable generation order of Transformer.
no code implementations • AKBC 2021 • Wiem Ben Rim, Carolin Lawrence, Kiril Gashteovski, Mathias Niepert, Naoaki Okazaki
With an extensive set of experiments, we perform and analyze these tests for several KGE models.
2 code implementations • Findings (ACL) 2021 • Tatsuya Hiraoka, Sho Takase, Kei Uchiumi, Atsushi Keyaki, Naoaki Okazaki
Since traditional tokenizers are isolated from a downstream task and model, they cannot output an appropriate tokenization depending on the task and model, although recent studies imply that the appropriate tokenization improves the performance.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Shin Kanouchi, Masato Neishi, Yuta Hayashibe, Hiroki Ouchi, Naoaki Okazaki
To tackle the challenge, this paper proposes a novel and practical task to explain evidences in recommending hotels given vague requests expressed freely in natural language.
no code implementations • SEMEVAL 2020 • Zhishen Yang, Lars Wolfsteller, Naoaki Okazaki
This paper describes the emphasis selection system of the team TextLearner for SemEval 2020 Task 10: Emphasis Selection For Written Text in Visual Media.
1 code implementation • COLING 2020 • Zhishen Yang, Naoaki Okazaki
In this paper, we address the task of news-image captioning, which generates a description of an image given the image and its article body as input.
no code implementations • SEMEVAL 2020 • Wiem Ben Rim, Naoaki Okazaki
We describe the system submitted by the SWAGex team to the SemEval-2020 Commonsense Validation and Explanation Task.
3 code implementations • 30 Nov 2020 • Emanuele Bugliarello, Ryan Cotterell, Naoaki Okazaki, Desmond Elliott
Large-scale pretraining and task-specific fine-tuning is now the standard methodology for many tasks in computer vision and natural language processing.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Tatsuya Hiraoka, Sho Takase, Kei Uchiumi, Atsushi Keyaki, Naoaki Okazaki
In traditional NLP, we tokenize a given sentence as a preprocessing, and thus the tokenization is unrelated to a target downstream task.
no code implementations • LREC 2022 • Sho Takase, Naoaki Okazaki
Experimental results indicate that Transum improves the performance from the strong baseline, Transformer, in Chinese-English, Arabic-English, and English-Japanese translation datasets.
Abstractive Text Summarization Cross-Lingual Abstractive Summarization +4
1 code implementation • Journal of Natural Language Processing 2022 • Youmi Ma, Tatsuya Hiraoka, Naoaki Okazaki
In this study, a novel method for extracting named entities and relations from unstructured text based on the table representation is presented.
Ranked #3 on Relation Extraction on CoNLL04 (NER Micro F1 metric)
no code implementations • 23 Jun 2020 • Tosho Hirasawa, Zhishen Yang, Mamoru Komachi, Naoaki Okazaki
Video-guided machine translation as one of multimodal neural machine translation tasks targeting on generating high-quality text translation by tangibly engaging both video and text.
1 code implementation • ACL 2020 • Emanuele Bugliarello, Sabrina J. Mielke, Antonios Anastasopoulos, Ryan Cotterell, Naoaki Okazaki
The performance of neural machine translation systems is commonly evaluated in terms of BLEU.
1 code implementation • ACL 2020 • Kazuki Matsumaru, Sho Takase, Naoaki Okazaki
Building a binary classifier that predicts an entailment relation between an article and its headline, we filter out untruthful instances from the supervision data.
no code implementations • LREC 2020 • Sho Shimazu, Sho Takase, Toshiaki Nakazawa, Naoaki Okazaki
Therefore, we present a hand-crafted dataset to evaluate whether translation models can resolve the zero pronoun problems in Japanese to English translations.
no code implementations • LREC 2020 • Sangwhan Moon, Naoaki Okazaki
In the context of multilingual language model pre-training, vocabulary size for languages with a broad set of potential characters is an unsolved problem.
no code implementations • WS 2019 • Yuichi Sasazawa, Sho Takase, Naoaki Okazaki
One of the key requirements of QG is to generate a question such that it results in a target answer.
1 code implementation • ACL 2020 • Emanuele Bugliarello, Naoaki Okazaki
Most neural machine translation models only rely on pairs of parallel sentences, assuming syntactic information is automatically learned by an attention mechanism.
2 code implementations • ACL 2019 • Hayate Iso, Yui Uehara, Tatsuya Ishigaki, Hiroshi Noji, Eiji Aramaki, Ichiro Kobayashi, Yusuke Miyao, Naoaki Okazaki, Hiroya Takamura
We propose a data-to-text generation model with two modules, one for tracking and the other for text generation.
no code implementations • SEMEVAL 2019 • Zhishen Yang, Sam Vijlbrief, Naoaki Okazaki
This paper presents our contextual emotion detection system in approaching the SemEval2019 shared task 3: EmoContext: Contextual Emotion Detection in Text.
1 code implementation • NAACL 2019 • Sho Takase, Naoaki Okazaki
Neural encoder-decoder models have been successful in natural language generation tasks.
Ranked #2 on Text Summarization on DUC 2004 Task 1
no code implementations • WS 2019 • Yuta Hitomi, Yuya Taguchi, Hideaki Tamori, Ko Kikuta, Jiro Nishitoba, Naoaki Okazaki, Kentaro Inui, Manabu Okumura
However, because there is no corpus of headlines of multiple lengths for a given article, previous research on controlling output length in headline generation has not discussed whether the system outputs could be adequately evaluated without multiple references of different lengths.
no code implementations • WS 2018 • Shun Kiyono, Sho Takase, Jun Suzuki, Naoaki Okazaki, Kentaro Inui, Masaaki Nagata
Developing a method for understanding the inner workings of black-box neural methods is an important research endeavor.
no code implementations • WS 2018 • Diana Galvan, Naoaki Okazaki, Koji Matsuda, Kentaro Inui
Temporal reasoning remains as an unsolved task for Natural Language Processing (NLP), particularly demonstrated in the clinical domain.
no code implementations • COLING 2018 • Akira Sasaki, Kazuaki Hanawa, Naoaki Okazaki, Kentaro Inui
This paper presents an approach to detect the stance of a user toward a topic based on their stances toward other topics and the social media posts of the user.
no code implementations • 22 Dec 2017 • Shun Kiyono, Sho Takase, Jun Suzuki, Naoaki Okazaki, Kentaro Inui, Masaaki Nagata
The encoder-decoder model is widely used in natural language generation tasks.
no code implementations • 7 Dec 2017 • Paul Reisert, Naoya Inoue, Naoaki Okazaki, Kentaro Inui
Our coverage result of 74. 6% indicates that argumentative relations can reasonably be explained by our small pattern set.
no code implementations • IJCNLP 2017 • Yuta Hitomi, Hideaki Tamori, Naoaki Okazaki, Kentaro Inui
This paper explores the idea of robot editors, automated proofreaders that enable journalists to improve the quality of their articles.
1 code implementation • IJCNLP 2017 • Sosuke Kobayashi, Naoaki Okazaki, Kentaro Inui
This study addresses the problem of identifying the meaning of unknown words or entities in a discourse with respect to the word embedding approaches used in neural language models.
no code implementations • WS 2017 • Hideaki Tamori, Yuta Hitomi, Naoaki Okazaki, Kentaro Inui
We address the issue of the quality of journalism and analyze daily article revision logs from a Japanese newspaper company.
1 code implementation • ACL 2016 • Sho Takase, Naoaki Okazaki, Kentaro Inui
Learning distributed representations for relation instances is a central technique in downstream NLP applications.
no code implementations • ACL 2017 • Akira Sasaki, Kazuaki Hanawa, Naoaki Okazaki, Kentaro Inui
We present in this paper our approach for modeling inter-topic preferences of Twitter users: for example, those who agree with the Trans-Pacific Partnership (TPP) also agree with free trade.
no code implementations • COLING 2016 • Natsuda Laokulrat, Sang Phan, Noriki Nishida, Raphael Shu, Yo Ehara, Naoaki Okazaki, Yusuke Miyao, Hideki Nakayama
Automatic video description generation has recently been getting attention after rapid advancement in image caption generation.
no code implementations • COLING 2016 • Kento Watanabe, Yuichiroh Matsubayashi, Naho Orita, Naoaki Okazaki, Kentaro Inui, Satoru Fukayama, Tomoyasu Nakano, Jordan Smith, Masataka Goto
This study proposes a computational model of the discourse segments in lyrics to understand and to model the structure of lyrics.
no code implementations • COLING 2016 • Naoya Inoue, Yuichiroh Matsubayashi, Masayuki Ono, Naoaki Okazaki, Kentaro Inui
This paper proposes a novel problem setting of selectional preference (SP) between a predicate and its arguments, called as context-sensitive SP (CSP).
1 code implementation • ACL 2016 • Ran Tian, Naoaki Okazaki, Kentaro Inui
This paper connects a vector-based composition model to a formal semantics, the Dependency-based Compositional Semantics (DCS).
no code implementations • 26 Nov 2015 • Ran Tian, Naoaki Okazaki, Kentaro Inui
Additive composition (Foltz et al, 1998; Landauer and Dumais, 1997; Mitchell and Lapata, 2010) is a widely used method for computing meanings of phrases, which takes the average of vector representations of the constituent words.