no code implementations • Findings (ACL) 2022 • Zuchao Li, Yiran Wang, Masao Utiyama, Eiichiro Sumita, Hai Zhao, Taro Watanabe
Inspired by this discovery, we then propose approaches to improving it, with respect to model structure and model training, to make the deep decoder practical in NMT.
no code implementations • NAACL (CALCS) 2021 • Chihiro Taguchi, Yusuke Sakai, Taro Watanabe
Given this situation, we proposed a transliteration method based on subword-level language identification.
no code implementations • SpaNLP (ACL) 2022 • Van-Hien Tran, Hiroki Ouchi, Taro Watanabe, Yuji Matsumoto
Zero-shot relation extraction (ZSRE) aims to predict target relations that cannot be observed during training.
Ranked #3 on
Zero-shot Relation Classification
on FewRel
no code implementations • WNUT (ACL) 2021 • Shohei Higashiyama, Masao Utiyama, Taro Watanabe, Eiichiro Sumita
Lexical normalization, in addition to word segmentation and part-of-speech tagging, is a fundamental task for Japanese user-generated text processing.
no code implementations • EURALI (LREC) 2022 • Chihiro Taguchi, Sei Iwata, Taro Watanabe
Experimenting on NMCTT and the Turkish-German CS treebank (SAGT), we demonstrate that the proposed annotation scheme introduced in NMCTT can improve the performance of the subword-level language identification.
1 code implementation • 13 Feb 2025 • Takumi Goto, Yusuke Sakai, Taro Watanabe
One of the goals of automatic evaluation metrics in grammatical error correction (GEC) is to rank GEC systems such that it matches human preferences.
1 code implementation • 29 Jan 2025 • Haruki Sakajo, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe
In this study, we created video and image datasets from the existing real-time MRI dataset and investigated whether LMs can understand vowel articulation based on tongue positions using vision-based information.
1 code implementation • 12 Jan 2025 • Justin Vasselli, Adam Nohejl, Taro Watanabe
Advancements in dialogue systems powered by large language models (LLMs) have outpaced the development of reliable evaluation metrics, particularly for diverse and creative responses.
1 code implementation • 11 Jan 2025 • Adam Nohejl, Taro Watanabe
Various measures of dispersion have been proposed to paint a fuller picture of a word's distribution in a corpus, but only little has been done to validate them externally.
3 code implementations • 6 Jan 2025 • Zhi Qu, Yiran Wang, Jiannan Mao, Chenchen Ding, Hideki Tanaka, Masao Utiyama, Taro Watanabe
We further scale up and collect 9. 3 billion sentence pairs across 24 languages from public datasets to pre-train two models, namely MITRE (multilingual translation with registers).
no code implementations • 29 Dec 2024 • Shintaro Ozaki, Yuta Kato, Siyuan Feng, Masayo Tomita, Kazuki Hayashi, Ryoma Obara, Masafumi Oyamada, Katsuhiko Hayashi, Hidetaka Kamigaito, Taro Watanabe
Retrieval Augmented Generation (RAG) complements the knowledge of Large Language Models (LLMs) by leveraging external information to enhance response accuracy for queries.
no code implementations • 24 Dec 2024 • Yusuke Ide, Joshua Tanner, Adam Nohejl, Jacob Hoffman, Justin Vasselli, Hidetaka Kamigaito, Taro Watanabe
MWEs in CoAM are tagged with MWE types, such as Noun and Verb, to enable fine-grained error analysis.
1 code implementation • 17 Dec 2024 • Takumi Goto, Justin Vasselli, Taro Watanabe
Various evaluation metrics have been proposed for Grammatical Error Correction (GEC), but many, particularly reference-free metrics, lack explainability.
1 code implementation • 3 Dec 2024 • Zhi Qu, Yiran Wang, Chenchen Ding, Hideki Tanaka, Masao Utiyama, Taro Watanabe
We propose dividing the decoding process into two stages so that target tokens are explicitly excluded in the first stage to implicitly boost the transfer capability across languages.
1 code implementation • 24 Oct 2024 • Adam Nohejl, Akio Hayakawa, Yusuke Ide, Taro Watanabe
The tasks of lexical complexity prediction (LCP) and complex word identification (CWI) commonly presuppose that difficult to understand words are shared by the target population.
no code implementations • 22 Oct 2024 • Aitaro Yamamoto, Hiroyuki Otomo, Hiroki Ouchi, Shohei Higashiyama, Hiroki Teranishi, Hiroyuki Shindo, Taro Watanabe
Previous studies on sequence-based extraction of human movement trajectories have an issue of inadequate trajectory representation.
no code implementations • 19 Oct 2024 • Hidetaka Kamigaito, Hiroyuki Deguchi, Yusuke Sakai, Katsuhiko Hayashi, Taro Watanabe
We also introduce a new MBR approach, Metric-augmented MBR (MAMBR), which increases diversity by adjusting the behavior of utility functions without altering the pseudo-references.
no code implementations • 17 Oct 2024 • Shintaro Ozaki, Kazuki Hayashi, Miyu Oba, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe
To address this, we propose a dataset, BQA, a body language question answering dataset, to validate whether the model can correctly interpret emotions from short clips of body language comprising 26 emotion labels of videos of body language.
1 code implementation • 16 Oct 2024 • Genta Indra Winata, Frederikus Hudi, Patrick Amadeus Irawan, David Anugraha, Rifki Afina Putri, Yutong Wang, Adam Nohejl, Ubaidillah Ariq Prathama, Nedjma Ousidhoum, Afifa Amriani, Anar Rzayev, Anirban Das, Ashmari Pramodya, Aulia Adila, Bryan Wilie, Candy Olivia Mawalim, Ching Lam Cheng, Daud Abolade, Emmanuele Chersoni, Enrico Santus, Fariz Ikhwantri, Garry Kuwanto, Hanyang Zhao, Haryo Akbarianto Wibowo, Holy Lovenia, Jan Christian Blaise Cruz, Jan Wira Gotama Putra, Junho Myung, Lucky Susanto, Maria Angelica Riera Machin, Marina Zhukova, Michael Anugraha, Muhammad Farid Adilazuarda, Natasha Santosa, Peerat Limkonchotiwat, Raj Dabre, Rio Alexander Audino, Samuel Cahyawijaya, Shi-Xiong Zhang, Stephanie Yulia Salim, Yi Zhou, Yinxuan Gui, David Ifeoluwa Adelani, En-Shiun Annie Lee, Shogo Okada, Ayu Purwarianti, Alham Fikri Aji, Taro Watanabe, Derry Tanti Wijaya, Alice Oh, Chong-Wah Ngo
This benchmark includes a visual question answering (VQA) dataset with text-image pairs across 30 languages and dialects, spanning 9 language families and featuring over 1 million data points, making it the largest multicultural VQA benchmark to date.
no code implementations • 8 Oct 2024 • Miyu Oba, Yohei Oseki, Akiyo Fukatsu, Akari Haga, Hiroki Ouchi, Taro Watanabe, Saku Sugawara
What kinds of and how much data is necessary for language models to induce grammatical knowledge to judge sentence acceptability?
no code implementations • 4 Oct 2024 • Huayang Li, Pat Verga, Priyanka Sen, Bowen Yang, Vijay Viswanathan, Patrick Lewis, Taro Watanabe, Yixuan Su
The context window of large language models (LLMs) has been extended significantly in recent years.
1 code implementation • 4 Oct 2024 • Adam Nohejl, Frederikus Hudi, Eunike Andriani Kardinata, Shintaro Ozaki, Maria Angelica Riera Machin, Hongyu Sun, Justin Vasselli, Taro Watanabe
Word frequency is a key variable in psycholinguistics, useful for modeling human familiarity with words even in the era of large language models (LLMs).
1 code implementation • 8 Sep 2024 • Zhe Cao, Zhi Qu, Hidetaka Kamigaito, Taro Watanabe
Furthermore, we propose architecture learning techniques and introduce a gradual pruning schedule during fine-tuning to exhaustively explore the optimal setting and the minimal intrinsic subspaces for each language, resulting in a lightweight yet effective fine-tuning procedure.
no code implementations • 3 Sep 2024 • Shintaro Ozaki, Kazuki Hayashi, Yusuke Sakai, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe
As the performance of Large-scale Vision Language Models (LVLMs) improves, they are increasingly capable of responding in multiple languages, and there is an expectation that the demand for explanations generated by LVLMs will grow.
no code implementations • 22 Aug 2024 • Yusuke Sakai, Adam Nohejl, Jiangnan Hang, Hidetaka Kamigaito, Taro Watanabe
In this study, we provide English and Japanese cross-lingual datasets for evaluating the NLU performance of LLMs, which include multiple instruction templates for fair evaluation of each task, along with regular expressions to constrain the output format.
no code implementations • 19 Aug 2024 • Yusuke Ide, Yuto Nishida, Miyu Oba, Yusuke Sakai, Justin Vasselli, Hidetaka Kamigaito, Taro Watanabe
The grammatical knowledge of language models (LMs) is often measured using a benchmark of linguistic minimal pairs, where LMs are presented with a pair of acceptable and unacceptable sentences and required to judge which is acceptable.
1 code implementation • 12 Aug 2024 • Peinan Zhang, Yusuke Sakai, Masato Mita, Hiroki Ouchi, Taro Watanabe
With the increase in the more fluent ad texts automatically created by natural language generation technology, it is in the high demand to verify the quality of these creatives in a real-world setting.
1 code implementation • 8 Aug 2024 • Hiroyuki Deguchi, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe
We published our mbrs as an MIT-licensed open-source project, and the code is available on GitHub.
1 code implementation • 5 Jul 2024 • Xincan Feng, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe
This paper provides theoretical interpretations of the smoothing methods for the NS loss in KGE and induces a new NS loss, Triplet Adaptive Negative Sampling (TANS), that can cover the characteristics of the conventional smoothing methods.
1 code implementation • 2 Jul 2024 • Wataru Hashimoto, Hidetaka Kamigaito, Taro Watanabe
This work investigates the impact of data augmentation on confidence calibration and uncertainty estimation in Named Entity Recognition (NER) tasks.
no code implementations • 2 Jul 2024 • Wataru Hashimoto, Hidetaka Kamigaito, Taro Watanabe
Trustworthy prediction in Deep Neural Networks (DNNs), including Pre-trained Language Models (PLMs) is important for safety-critical applications in the real world.
no code implementations • 2 Jul 2024 • Arturo Martínez Peguero, Taro Watanabe
Recent work in reframing, within the scope of text style transfer, has so far made use of out-of-context, task-prompted utterances in order to produce neutralizing or optimistic reframes.
no code implementations • 27 Jun 2024 • Ryo Tsujimoto, Hiroki Ouchi, Hidetaka Kamigaito, Taro Watanabe
Explaining temporal changes between satellite images taken at different times is important for urban planning and environmental monitoring.
no code implementations • 24 Jun 2024 • Deng Cai, Huayang Li, Tingchen Fu, Siheng Li, Weiwen Xu, Shuaiyi Li, Bowen Cao, Zhisong Zhang, Xinting Huang, Leyang Cui, Yan Wang, Lemao Liu, Taro Watanabe, Shuming Shi
Despite the general capabilities of pre-trained large language models (LLMs), they still need further adaptation to better serve practical applications.
1 code implementation • 18 Jun 2024 • Zhiyu Guo, Hidetaka Kamigaito, Taro Watanabe
Scaling the context size of large language models (LLMs) enables them to perform various new tasks, e. g., book summarization.
1 code implementation • 12 Jun 2024 • Zhi Qu, Chenchen Ding, Taro Watanabe
Understanding representation transfer in multilingual neural machine translation can reveal the representational issue causing the zero-shot translation deficiency.
no code implementations • 6 Jun 2024 • Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe
Constructed dataset is a benchmark for cross-lingual language-transfer capabilities of multilingual LMs, and experimental results showed high language-transfer capabilities for questions that LMs could easily solve, but lower transfer capabilities for questions requiring deep knowledge or commonsense.
1 code implementation • 30 Apr 2024 • Huy Hien Vu, Hidetaka Kamigaito, Taro Watanabe
Despite significant improvements in enhancing the quality of translation, context-aware machine translation (MT) models underperform in many cases.
1 code implementation • COLING (TextGraphs) 2022 • Xincan Feng, Zhi Qu, Yuchang Cheng, Taro Watanabe, Nobuhiro Yugami
A Knowledge Graph (KG) is the directed graphical representation of entities and relations in the real world.
no code implementations • 18 Apr 2024 • Yusuke Sakai, Mana Makinae, Hidetaka Kamigaito, Taro Watanabe
In Simultaneous Machine Translation (SiMT) systems, training with a simultaneous interpretation (SI) corpus is an effective method for achieving high-quality yet low-latency systems.
no code implementations • 29 Mar 2024 • Jesse Atuhurra, Iqra Ali, Tatsuya Hiraoka, Hidetaka Kamigaito, Tomoya Iwakura, Taro Watanabe
Our contribution is four-fold: 1) we introduced nine vision-and-language (VL) tasks (including object recognition, image-text matching, and more) and constructed multilingual visual-text datasets in four languages: English, Japanese, Swahili, and Urdu through utilizing templates containing \textit{questions} and prompting GPT4-V to generate the \textit{answers} and the \textit{rationales}, 2) introduced a new VL task named \textit{unrelatedness}, 3) introduced rationales to enable human understanding of the VLM reasoning process, and 4) employed human evaluation to measure the suitability of proposed datasets for VL tasks.
1 code implementation • 28 Mar 2024 • Eri Onami, Shuhei Kurita, Taiki Miyanishi, Taro Watanabe
Document question answering is a task of question answering on given documents such as reports, slides, pamphlets, and websites, and it is a truly demanding task as paper and electronic forms of documents are so common in our society.
no code implementations • 26 Mar 2024 • Jesse Atuhurra, Hiroyuki Shindo, Hidetaka Kamigaito, Taro Watanabe
Tokenization is one such technique because it allows for the words to be split based on characters or subwords, creating word embeddings that best represent the structure of the language.
1 code implementation • 25 Mar 2024 • Huayang Li, Deng Cai, Zhi Qu, Qu Cui, Hidetaka Kamigaito, Lemao Liu, Taro Watanabe
In our work, we propose a new task formulation of dense retrieval, cross-lingual contextualized phrase retrieval, which aims to augment cross-lingual applications by addressing polysemy using context information.
no code implementations • 13 Mar 2024 • Jesse Atuhurra, Seiveright Cargill Dujohn, Hidetaka Kamigaito, Hiroyuki Shindo, Taro Watanabe
Natural language processing (NLP) practitioners are leveraging large language models (LLM) to create structured datasets from semi-structured and unstructured data sources such as patents, papers, and theses, without having domain-specific knowledge.
no code implementations • 29 Feb 2024 • Kazuki Hayashi, Yusuke Sakai, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe
To address this issue, we propose a new task: the artwork explanation generation task, along with its evaluation dataset and metric for quantitatively assessing the understanding and utilization of knowledge about artworks.
1 code implementation • 22 Feb 2024 • Seiji Gobara, Hidetaka Kamigaito, Taro Watanabe
Experimental results on the Stack-Overflow dataset and the TSCC dataset, including multi-turn conversation show that LLMs can implicitly handle text difficulty between user input and its generated response.
no code implementations • 19 Feb 2024 • Kazuki Hayashi, Kazuma Onishi, Toma Suzuki, Yusuke Ide, Seiji Gobara, Shigeki Saito, Yusuke Sakai, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe
We validate it using a dataset of images from 15 categories, each with five critic review texts and annotated rankings in both English and Japanese, totaling over 2, 000 data instances.
1 code implementation • 17 Feb 2024 • Hiroyuki Deguchi, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe, Hideki Tanaka, Masao Utiyama
Minimum Bayes risk (MBR) decoding achieved state-of-the-art translation performance by using COMET, a neural metric that has a high correlation with human evaluation.
no code implementations • 14 Feb 2024 • Yuto Nishida, Makoto Morishita, Hidetaka Kamigaito, Taro Watanabe
Generating multiple translation candidates would enable users to choose the one that satisfies their needs.
no code implementations • 15 Nov 2023 • Yusuke Sakai, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe
Knowledge Graph Completion (KGC) is a task that infers unseen relationships between entities in a KG.
1 code implementation • 18 Oct 2023 • Hiroyuki Deguchi, Hayate Hirano, Tomoki Hoshino, Yuto Nishida, Justin Vasselli, Taro Watanabe
We publish our knn-seq as an MIT-licensed open-source project and the code is available on https://github. com/naist-nlp/knn-seq .
1 code implementation • 17 Sep 2023 • Xincan Feng, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe
Subsampling is effective in Knowledge Graph Embedding (KGE) for reducing overfitting caused by the sparsity in Knowledge Graph (KG) datasets.
1 code implementation • 14 Sep 2023 • Huayang Li, Siheng Li, Deng Cai, Longyue Wang, Lemao Liu, Taro Watanabe, Yujiu Yang, Shuming Shi
We release our dataset, model, and demo to foster future research in the area of multimodal instruction following.
Ranked #228 on
Visual Question Answering
on MM-Vet
2 code implementations • 30 Jun 2023 • Yusuke Ide, Masato Mita, Adam Nohejl, Hiroki Ouchi, Taro Watanabe
Lexical complexity prediction (LCP) is the task of predicting the complexity of words in a text on a continuous scale.
1 code implementation • Journal of Natural Language Processing 2023 • Van-Hien Tran, Hiroki Ouchi, Hiroyuki Shindo, Yuji Matsumoto, Taro Watanabe
This study argues that enhancing the semantic correlation between instances and relations is key to effectively solving the zero-shot relation extraction task.
Ranked #1 on
Zero-shot Relation Classification
on FewRel
1 code implementation • 5 Jun 2023 • Miyu Oba, Tatsuki Kuribayashi, Hiroki Ouchi, Taro Watanabe
With the success of neural language models (LMs), their language acquisition has gained much attention.
1 code implementation • 3 Jun 2023 • Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe
This task consists of two parts: the first is to generate a table containing knowledge about an entity and its related image, and the second is to generate an image from an entity with a caption and a table containing related knowledge of the entity.
1 code implementation • 23 May 2023 • Shohei Higashiyama, Hiroki Ouchi, Hiroki Teranishi, Hiroyuki Otomo, Yusuke Ide, Aitaro Yamamoto, Hiroyuki Shindo, Yuki Matsuda, Shoko Wakamiya, Naoya Inoue, Ikuya Yamada, Taro Watanabe
Geoparsing is a fundamental technique for analyzing geo-entity information in text.
no code implementations • 19 May 2023 • Hiroki Ouchi, Hiroyuki Shindo, Shoko Wakamiya, Yuki Matsuda, Naoya Inoue, Shohei Higashiyama, Satoshi Nakamura, Taro Watanabe
We have constructed Arukikata Travelogue Dataset and released it free of charge for academic research.
1 code implementation • 6 Dec 2022 • Ukyo Honda, Taro Watanabe, Yuji Matsumoto
Discriminativeness is a desirable feature of image captions: captions should describe the characteristic details of input images.
1 code implementation • 26 Oct 2022 • Huayang Li, Deng Cai, Jin Xu, Taro Watanabe
The combination of $n$-gram and neural LMs not only allows the neural part to focus on the deeper understanding of language but also provides a flexible way to customize an LM by switching the underlying $n$-gram model without changing the neural model.
1 code implementation • COLING 2022 • Zhi Qu, Taro Watanabe
Multilingual neural machine translation can translate unseen language pairs during training, i. e. zero-shot translation.
no code implementations • Findings (ACL) 2022 • Jiannan Xiang, Huayang Li, Defu Lian, Guoping Huang, Taro Watanabe, Lemao Liu
To this end, we study the dynamic relationship between the encoded linguistic information and task performance from the viewpoint of Pareto Optimality.
no code implementations • Journal of Natural Language Processing 2021 • Van-Hien Tran, Van-Thuy Phi, Akihiko Kato, Hiroyuki Shindo, Taro Watanabe, Yuji Matsumoto
A recent study (Yu et al. 2020) proposed a novel decomposition strategy that splits the task into two interrelated subtasks: detection of the head-entity (HE) and identification of the corresponding tail-entity and relation (TER) for each extracted head-entity.
no code implementations • 1 Nov 2021 • Yushi Hirose, Masashi Shimbo, Taro Watanabe
For knowledge graph completion, two major types of prediction models exist: one based on graph embeddings, and the other based on relation path rule induction.
no code implementations • Joint Conference on Lexical and Computational Semantics 2021 • Yuki Yamamoto, Yuji Matsumoto, Taro Watanabe
Abstract Meaning Representation (AMR) is a sentence-level meaning representation based on predicate argument structure.
1 code implementation • ACL 2021 • Yiran Wang, Hiroyuki Shindo, Yuji Matsumoto, Taro Watanabe
This paper presents a novel method for nested named entity recognition.
Ranked #13 on
Nested Named Entity Recognition
on ACE 2005
no code implementations • ACL 2021 • Sei Iwata, Taro Watanabe, Masaaki Nagata
In the experiments, our model surpassed the sequence labeling baseline.
no code implementations • ACL 2021 • Shintaro Harada, Taro Watanabe
It is reported that grammatical information is useful for machine translation (MT) task.
1 code implementation • EACL 2021 • Ukyo Honda, Yoshitaka Ushiku, Atsushi Hashimoto, Taro Watanabe, Yuji Matsumoto
Unsupervised image captioning is a challenging task that aims at generating captions without the supervision of image-sentence pairs, but only with images and sentences drawn from different sources and object labels detected from the images.
1 code implementation • NAACL 2021 • Shohei Higashiyama, Masao Utiyama, Taro Watanabe, Eiichiro Sumita
Morphological analysis (MA) and lexical normalization (LN) are both important tasks for Japanese user-generated text (UGT).
no code implementations • 1 Jan 2021 • Guanlin Li, Lemao Liu, Taro Watanabe, Conghui Zhu, Tiejun Zhao
Unsupervised Neural Machine Translation or UNMT has received great attention in recent years.
no code implementations • COLING 2020 • Yuya Sawada, Takashi Wada, Takayoshi Shibahara, Hiroki Teranishi, Shuhei Kondo, Hiroyuki Shindo, Taro Watanabe, Yuji Matsumoto
We propose a simple method for nominal coordination boundary identification.
no code implementations • WS 2018 • Wei Wang, Taro Watanabe, Macduff Hughes, Tetsuji Nakagawa, Ciprian Chelba
Measuring domain relevance of data and identifying or selecting well-fit domain data for machine translation (MT) is a well-studied topic, but denoising is not yet.
no code implementations • COLING 2016 • Yusuke Oda, Taku Kudo, Tetsuji Nakagawa, Taro Watanabe
In this paper, we propose a new decoding method for phrase-based statistical machine translation which directly uses multiple preordering candidates as a graph structure.