Search Results for author: Taro Watanabe

Found 96 papers, 38 papers with code

What Works and Doesn’t Work, A Deep Decoder for Neural Machine Translation

no code implementations Findings (ACL) 2022 Zuchao Li, Yiran Wang, Masao Utiyama, Eiichiro Sumita, Hai Zhao, Taro Watanabe

Inspired by this discovery, we then propose approaches to improving it, with respect to model structure and model training, to make the deep decoder practical in NMT.

Decoder Language Modelling +3

Universal Dependencies Treebank for Tatar: Incorporating Intra-Word Code-Switching Information

no code implementations EURALI (LREC) 2022 Chihiro Taguchi, Sei Iwata, Taro Watanabe

Experimenting on NMCTT and the Turkish-German CS treebank (SAGT), we demonstrate that the proposed annotation scheme introduced in NMCTT can improve the performance of the subword-level language identification.

Language Identification POS +1

Rethinking Evaluation Metrics for Grammatical Error Correction: Why Use a Different Evaluation Process than Human?

1 code implementation13 Feb 2025 Takumi Goto, Yusuke Sakai, Taro Watanabe

One of the goals of automatic evaluation metrics in grammatical error correction (GEC) is to rank GEC systems such that it matches human preferences.

Grammatical Error Correction Sentence

Tonguescape: Exploring Language Models Understanding of Vowel Articulation

1 code implementation29 Jan 2025 Haruki Sakajo, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe

In this study, we created video and image datasets from the existing real-time MRI dataset and investigated whether LMs can understand vowel articulation based on tongue positions using vision-based information.

Measuring the Robustness of Reference-Free Dialogue Evaluation Systems

1 code implementation12 Jan 2025 Justin Vasselli, Adam Nohejl, Taro Watanabe

Advancements in dialogue systems powered by large language models (LLMs) have outpaced the development of reliable evaluation metrics, particularly for diverse and creative responses.

Dialogue Evaluation TAG

Dispersion Measures as Predictors of Lexical Decision Time, Word Familiarity, and Lexical Complexity

1 code implementation11 Jan 2025 Adam Nohejl, Taro Watanabe

Various measures of dispersion have been proposed to paint a fuller picture of a word's distribution in a corpus, but only little has been done to validate them externally.

Registering Source Tokens to Target Language Spaces in Multilingual Neural Machine Translation

3 code implementations6 Jan 2025 Zhi Qu, Yiran Wang, Jiannan Mao, Chenchen Ding, Hideki Tanaka, Masao Utiyama, Taro Watanabe

We further scale up and collect 9. 3 billion sentence pairs across 24 languages from public datasets to pre-train two models, namely MITRE (multilingual translation with registers).

Machine Translation Translation

Understanding the Impact of Confidence in Retrieval Augmented Generation: A Case Study in the Medical Domain

no code implementations29 Dec 2024 Shintaro Ozaki, Yuta Kato, Siyuan Feng, Masayo Tomita, Kazuki Hayashi, Ryoma Obara, Masafumi Oyamada, Katsuhiko Hayashi, Hidetaka Kamigaito, Taro Watanabe

Retrieval Augmented Generation (RAG) complements the knowledge of Large Language Models (LLMs) by leveraging external information to enhance response accuracy for queries.

RAG

Improving Explainability of Sentence-level Metrics via Edit-level Attribution for Grammatical Error Correction

1 code implementation17 Dec 2024 Takumi Goto, Justin Vasselli, Taro Watanabe

Various evaluation metrics have been proposed for Grammatical Error Correction (GEC), but many, particularly reference-free metrics, lack explainability.

Attribute Grammatical Error Correction +1

Improving Language Transfer Capability of Decoder-only Architecture in Multilingual Neural Machine Translation

1 code implementation3 Dec 2024 Zhi Qu, Yiran Wang, Chenchen Ding, Hideki Tanaka, Masao Utiyama, Taro Watanabe

We propose dividing the decoding process into two stages so that target tokens are explicitly excluded in the first stage to implicitly boost the transfer capability across languages.

Attribute Contrastive Learning +3

Difficult for Whom? A Study of Japanese Lexical Complexity

1 code implementation24 Oct 2024 Adam Nohejl, Akio Hayakawa, Yusuke Ide, Taro Watanabe

The tasks of lexical complexity prediction (LCP) and complex word identification (CWI) commonly presuppose that difficult to understand words are shared by the target population.

Complex Word Identification Lexical Complexity Prediction

Graph-Structured Trajectory Extraction from Travelogues

no code implementations22 Oct 2024 Aitaro Yamamoto, Hiroyuki Otomo, Hiroki Ouchi, Shohei Higashiyama, Hiroki Teranishi, Hiroyuki Shindo, Taro Watanabe

Previous studies on sequence-based extraction of human movement trajectories have an issue of inadequate trajectory representation.

Theoretical Aspects of Bias and Diversity in Minimum Bayes Risk Decoding

no code implementations19 Oct 2024 Hidetaka Kamigaito, Hiroyuki Deguchi, Yusuke Sakai, Katsuhiko Hayashi, Taro Watanabe

We also introduce a new MBR approach, Metric-augmented MBR (MAMBR), which increases diversity by adjusting the behavior of utility functions without altering the pseudo-references.

Diversity Text Generation

BQA: Body Language Question Answering Dataset for Video Large Language Models

no code implementations17 Oct 2024 Shintaro Ozaki, Kazuki Hayashi, Miyu Oba, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe

To address this, we propose a dataset, BQA, a body language question answering dataset, to validate whether the model can correctly interpret emotions from short clips of body language comprising 26 emotion labels of videos of body language.

Question Answering

Can Language Models Induce Grammatical Knowledge from Indirect Evidence?

no code implementations8 Oct 2024 Miyu Oba, Yohei Oseki, Akiyo Fukatsu, Akari Haga, Hiroki Ouchi, Taro Watanabe, Saku Sugawara

What kinds of and how much data is necessary for language models to induce grammatical knowledge to judge sentence acceptability?

Language Acquisition Sentence

Exploring Intrinsic Language-specific Subspaces in Fine-tuning Multilingual Neural Machine Translation

1 code implementation8 Sep 2024 Zhe Cao, Zhi Qu, Hidetaka Kamigaito, Taro Watanabe

Furthermore, we propose architecture learning techniques and introduce a gradual pruning schedule during fine-tuning to exhaustively explore the optimal setting and the minimal intrinsic subspaces for each language, resulting in a lightweight yet effective fine-tuning procedure.

Machine Translation

Towards Cross-Lingual Explanation of Artwork in Large-scale Vision Language Models

no code implementations3 Sep 2024 Shintaro Ozaki, Kazuki Hayashi, Yusuke Sakai, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe

As the performance of Large-scale Vision Language Models (LVLMs) improves, they are increasingly capable of responding in multiple languages, and there is an expectation that the demand for explanations generated by LVLMs will grow.

Machine Translation Translation

Toward the Evaluation of Large Language Models Considering Score Variance across Instruction Templates

no code implementations22 Aug 2024 Yusuke Sakai, Adam Nohejl, Jiangnan Hang, Hidetaka Kamigaito, Taro Watanabe

In this study, we provide English and Japanese cross-lingual datasets for evaluating the NLU performance of LLMs, which include multiple instruction templates for fair evaluation of each task, along with regular expressions to constrain the output format.

Natural Language Understanding

How to Make the Most of LLMs' Grammatical Knowledge for Acceptability Judgments

no code implementations19 Aug 2024 Yusuke Ide, Yuto Nishida, Miyu Oba, Yusuke Sakai, Justin Vasselli, Hidetaka Kamigaito, Taro Watanabe

The grammatical knowledge of language models (LMs) is often measured using a benchmark of linguistic minimal pairs, where LMs are presented with a pair of acceptable and unacceptable sentences and required to judge which is acceptable.

AdTEC: A Unified Benchmark for Evaluating Text Quality in Search Engine Advertising

1 code implementation12 Aug 2024 Peinan Zhang, Yusuke Sakai, Masato Mita, Hiroki Ouchi, Taro Watanabe

With the increase in the more fluent ad texts automatically created by natural language generation technology, it is in the high demand to verify the quality of these creatives in a real-world setting.

Text Generation

mbrs: A Library for Minimum Bayes Risk Decoding

1 code implementation8 Aug 2024 Hiroyuki Deguchi, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe

We published our mbrs as an MIT-licensed open-source project, and the code is available on GitHub.

Text Generation

Unified Interpretation of Smoothing Methods for Negative Sampling Loss Functions in Knowledge Graph Embedding

1 code implementation5 Jul 2024 Xincan Feng, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe

This paper provides theoretical interpretations of the smoothing methods for the NS loss in KGE and induces a new NS loss, Triplet Adaptive Negative Sampling (TANS), that can cover the characteristics of the conventional smoothing methods.

Knowledge Graph Embedding Knowledge Graphs +1

Are Data Augmentation Methods in Named Entity Recognition Applicable for Uncertainty Estimation?

1 code implementation2 Jul 2024 Wataru Hashimoto, Hidetaka Kamigaito, Taro Watanabe

This work investigates the impact of data augmentation on confidence calibration and uncertainty estimation in Named Entity Recognition (NER) tasks.

Data Augmentation named-entity-recognition +2

Efficient Nearest Neighbor based Uncertainty Estimation for Natural Language Processing Tasks

no code implementations2 Jul 2024 Wataru Hashimoto, Hidetaka Kamigaito, Taro Watanabe

Trustworthy prediction in Deep Neural Networks (DNNs), including Pre-trained Language Models (PLMs) is important for safety-critical applications in the real world.

Dimensionality Reduction named-entity-recognition +4

Change My Frame: Reframing in the Wild in r/ChangeMyView

no code implementations2 Jul 2024 Arturo Martínez Peguero, Taro Watanabe

Recent work in reframing, within the scope of text style transfer, has so far made use of out-of-context, task-prompted utterances in order to produce neutralizing or optimistic reframes.

Style Transfer Text Style Transfer

Towards Temporal Change Explanations from Bi-Temporal Satellite Images

no code implementations27 Jun 2024 Ryo Tsujimoto, Hiroki Ouchi, Hidetaka Kamigaito, Taro Watanabe

Explaining temporal changes between satellite images taken at different times is important for urban planning and environmental monitoring.

Image Captioning

On the Transformations across Reward Model, Parameter Update, and In-Context Prompt

no code implementations24 Jun 2024 Deng Cai, Huayang Li, Tingchen Fu, Siheng Li, Weiwen Xu, Shuaiyi Li, Bowen Cao, Zhisong Zhang, Xinting Huang, Leyang Cui, Yan Wang, Lemao Liu, Taro Watanabe, Shuming Shi

Despite the general capabilities of pre-trained large language models (LLMs), they still need further adaptation to better serve practical applications.

Attention Score is not All You Need for Token Importance Indicator in KV Cache Reduction: Value Also Matters

1 code implementation18 Jun 2024 Zhiyu Guo, Hidetaka Kamigaito, Taro Watanabe

Scaling the context size of large language models (LLMs) enables them to perform various new tasks, e. g., book summarization.

All Book summarization

Languages Transferred Within the Encoder: On Representation Transfer in Zero-Shot Multilingual Translation

1 code implementation12 Jun 2024 Zhi Qu, Chenchen Ding, Taro Watanabe

Understanding representation transfer in multilingual neural machine translation can reveal the representational issue causing the zero-shot translation deficiency.

Contrastive Learning Decoder +3

mCSQA: Multilingual Commonsense Reasoning Dataset with Unified Creation Strategy by Language Models and Humans

no code implementations6 Jun 2024 Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe

Constructed dataset is a benchmark for cross-lingual language-transfer capabilities of multilingual LMs, and experimental results showed high language-transfer capabilities for questions that LMs could easily solve, but lower transfer capabilities for questions requiring deep knowledge or commonsense.

Common Sense Reasoning Natural Language Understanding

Context-Aware Machine Translation with Source Coreference Explanation

1 code implementation30 Apr 2024 Huy Hien Vu, Hidetaka Kamigaito, Taro Watanabe

Despite significant improvements in enhancing the quality of translation, context-aware machine translation (MT) models underperform in many cases.

Machine Translation Translation

Simultaneous Interpretation Corpus Construction by Large Language Models in Distant Language Pair

no code implementations18 Apr 2024 Yusuke Sakai, Mana Makinae, Hidetaka Kamigaito, Taro Watanabe

In Simultaneous Machine Translation (SiMT) systems, training with a simultaneous interpretation (SI) corpus is an effective method for achieving high-quality yet low-latency systems.

Machine Translation Translation

Constructing Multilingual Visual-Text Datasets Revealing Visual Multilingual Ability of Vision Language Models

no code implementations29 Mar 2024 Jesse Atuhurra, Iqra Ali, Tatsuya Hiraoka, Hidetaka Kamigaito, Tomoya Iwakura, Taro Watanabe

Our contribution is four-fold: 1) we introduced nine vision-and-language (VL) tasks (including object recognition, image-text matching, and more) and constructed multilingual visual-text datasets in four languages: English, Japanese, Swahili, and Urdu through utilizing templates containing \textit{questions} and prompting GPT4-V to generate the \textit{answers} and the \textit{rationales}, 2) introduced a new VL task named \textit{unrelatedness}, 3) introduced rationales to enable human understanding of the VLM reasoning process, and 4) employed human evaluation to measure the suitability of proposed datasets for VL tasks.

Image-text matching Object Recognition +1

JDocQA: Japanese Document Question Answering Dataset for Generative Language Models

1 code implementation28 Mar 2024 Eri Onami, Shuhei Kurita, Taiki Miyanishi, Taro Watanabe

Document question answering is a task of question answering on given documents such as reports, slides, pamphlets, and websites, and it is a truly demanding task as paper and electronic forms of documents are so common in our society.

Hallucination Question Answering +1

Introducing Syllable Tokenization for Low-resource Languages: A Case Study with Swahili

no code implementations26 Mar 2024 Jesse Atuhurra, Hiroyuki Shindo, Hidetaka Kamigaito, Taro Watanabe

Tokenization is one such technique because it allows for the words to be split based on characters or subwords, creating word embeddings that best represent the structure of the language.

Multilingual NLP Text Generation +1

Cross-lingual Contextualized Phrase Retrieval

1 code implementation25 Mar 2024 Huayang Li, Deng Cai, Zhi Qu, Qu Cui, Hidetaka Kamigaito, Lemao Liu, Taro Watanabe

In our work, we propose a new task formulation of dense retrieval, cross-lingual contextualized phrase retrieval, which aims to augment cross-lingual applications by addressing polysemy using context information.

Contrastive Learning Language Modelling +4

Distilling Named Entity Recognition Models for Endangered Species from Large Language Models

no code implementations13 Mar 2024 Jesse Atuhurra, Seiveright Cargill Dujohn, Hidetaka Kamigaito, Hiroyuki Shindo, Taro Watanabe

Natural language processing (NLP) practitioners are leveraging large language models (LLM) to create structured datasets from semi-structured and unstructured data sources such as patents, papers, and theses, without having domain-specific knowledge.

In-Context Learning Knowledge Distillation +5

Artwork Explanation in Large-scale Vision Language Models

no code implementations29 Feb 2024 Kazuki Hayashi, Yusuke Sakai, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe

To address this issue, we propose a new task: the artwork explanation generation task, along with its evaluation dataset and metric for quantitatively assessing the understanding and utilization of knowledge about artworks.

Explanation Generation Text Generation

Do LLMs Implicitly Determine the Suitable Text Difficulty for Users?

1 code implementation22 Feb 2024 Seiji Gobara, Hidetaka Kamigaito, Taro Watanabe

Experimental results on the Stack-Overflow dataset and the TSCC dataset, including multi-turn conversation show that LLMs can implicitly handle text difficulty between user input and its generated response.

Question Answering

IRR: Image Review Ranking Framework for Evaluating Vision-Language Models

no code implementations19 Feb 2024 Kazuki Hayashi, Kazuma Onishi, Toma Suzuki, Yusuke Ide, Seiji Gobara, Shigeki Saito, Yusuke Sakai, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe

We validate it using a dataset of images from 15 categories, each with five critic review texts and annotated rankings in both English and Japanese, totaling over 2, 000 data instances.

Diversity Image Captioning

Centroid-Based Efficient Minimum Bayes Risk Decoding

1 code implementation17 Feb 2024 Hiroyuki Deguchi, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe, Hideki Tanaka, Masao Utiyama

Minimum Bayes risk (MBR) decoding achieved state-of-the-art translation performance by using COMET, a neural metric that has a high correlation with human evaluation.

de-en Translation

Generating Diverse Translation with Perturbed kNN-MT

no code implementations14 Feb 2024 Yuto Nishida, Makoto Morishita, Hidetaka Kamigaito, Taro Watanabe

Generating multiple translation candidates would enable users to choose the one that satisfies their needs.

Diversity Machine Translation +1

knn-seq: Efficient, Extensible kNN-MT Framework

1 code implementation18 Oct 2023 Hiroyuki Deguchi, Hayate Hirano, Tomoki Hoshino, Yuto Nishida, Justin Vasselli, Taro Watanabe

We publish our knn-seq as an MIT-licensed open-source project and the code is available on https://github. com/naist-nlp/knn-seq .

Machine Translation NMT +1

Model-based Subsampling for Knowledge Graph Completion

1 code implementation17 Sep 2023 Xincan Feng, Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe

Subsampling is effective in Knowledge Graph Embedding (KGE) for reducing overfitting caused by the sparsity in Knowledge Graph (KG) datasets.

Knowledge Graph Completion Knowledge Graph Embedding +1

Japanese Lexical Complexity for Non-Native Readers: A New Dataset

2 code implementations30 Jun 2023 Yusuke Ide, Masato Mita, Adam Nohejl, Hiroki Ouchi, Taro Watanabe

Lexical complexity prediction (LCP) is the task of predicting the complexity of words in a text on a continuous scale.

Lexical Complexity Prediction

Second Language Acquisition of Neural Language Models

1 code implementation5 Jun 2023 Miyu Oba, Tatsuki Kuribayashi, Hiroki Ouchi, Taro Watanabe

With the success of neural language models (LMs), their language acquisition has gained much attention.

Cross-Lingual Transfer Language Acquisition

Table and Image Generation for Investigating Knowledge of Entities in Pre-trained Vision and Language Models

1 code implementation3 Jun 2023 Hidetaka Kamigaito, Katsuhiko Hayashi, Taro Watanabe

This task consists of two parts: the first is to generate a table containing knowledge about an entity and its related image, and the second is to generate an image from an entity with a caption and a table containing related knowledge of the entity.

Image Generation

Arukikata Travelogue Dataset

no code implementations19 May 2023 Hiroki Ouchi, Hiroyuki Shindo, Shoko Wakamiya, Yuki Matsuda, Naoya Inoue, Shohei Higashiyama, Satoshi Nakamura, Taro Watanabe

We have constructed Arukikata Travelogue Dataset and released it free of charge for academic research.

Switching to Discriminative Image Captioning by Relieving a Bottleneck of Reinforcement Learning

1 code implementation6 Dec 2022 Ukyo Honda, Taro Watanabe, Yuji Matsumoto

Discriminativeness is a desirable feature of image captions: captions should describe the characteristic details of input images.

Image Captioning reinforcement-learning +1

$N$-gram Is Back: Residual Learning of Neural Text Generation with $n$-gram Language Model

1 code implementation26 Oct 2022 Huayang Li, Deng Cai, Jin Xu, Taro Watanabe

The combination of $n$-gram and neural LMs not only allows the neural part to focus on the deeper understanding of language but also provides a flexible way to customize an LM by switching the underlying $n$-gram model without changing the neural model.

Domain Adaptation Language Modeling +3

Adapting to Non-Centered Languages for Zero-shot Multilingual Translation

1 code implementation COLING 2022 Zhi Qu, Taro Watanabe

Multilingual neural machine translation can translate unseen language pairs during training, i. e. zero-shot translation.

Machine Translation Translation

Improved Decomposition Strategy for Joint Entity and Relation Extraction

no code implementations Journal of Natural Language Processing 2021 Van-Hien Tran, Van-Thuy Phi, Akihiko Kato, Hiroyuki Shindo, Taro Watanabe, Yuji Matsumoto

A recent study (Yu et al. 2020) proposed a novel decomposition strategy that splits the task into two interrelated subtasks: detection of the head-entity (HE) and identification of the corresponding tail-entity and relation (TER) for each extracted head-entity.

Joint Entity and Relation Extraction Relation +1

Transductive Data Augmentation with Relational Path Rule Mining for Knowledge Graph Embedding

no code implementations1 Nov 2021 Yushi Hirose, Masashi Shimbo, Taro Watanabe

For knowledge graph completion, two major types of prediction models exist: one based on graph embeddings, and the other based on relation path rule induction.

Data Augmentation Knowledge Graph Completion +2

Removing Word-Level Spurious Alignment between Images and Pseudo-Captions in Unsupervised Image Captioning

1 code implementation EACL 2021 Ukyo Honda, Yoshitaka Ushiku, Atsushi Hashimoto, Taro Watanabe, Yuji Matsumoto

Unsupervised image captioning is a challenging task that aims at generating captions without the supervision of image-sentence pairs, but only with images and sentences drawn from different sources and object labels detected from the images.

Image Captioning image-sentence alignment +2

Denoising Neural Machine Translation Training with Trusted Data and Online Data Selection

no code implementations WS 2018 Wei Wang, Taro Watanabe, Macduff Hughes, Tetsuji Nakagawa, Ciprian Chelba

Measuring domain relevance of data and identifying or selecting well-fit domain data for machine translation (MT) is a well-studied topic, but denoising is not yet.

Denoising Machine Translation +2

Phrase-based Machine Translation using Multiple Preordering Candidates

no code implementations COLING 2016 Yusuke Oda, Taku Kudo, Tetsuji Nakagawa, Taro Watanabe

In this paper, we propose a new decoding method for phrase-based statistical machine translation which directly uses multiple preordering candidates as a graph structure.

Decoder Machine Translation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.