no code implementations • WAT 2022 • Toshiaki Nakazawa, Hideya Mino, Isao Goto, Raj Dabre, Shohei Higashiyama, Shantipriya Parida, Anoop Kunchukuttan, Makoto Morishita, Ondřej Bojar, Chenhui Chu, Akiko Eriguchi, Kaori Abe, Yusuke Oda, Sadao Kurohashi
This paper presents the results of the shared tasks from the 9th workshop on Asian translation (WAT2022).
1 code implementation • EMNLP (WNUT) 2020 • Sora Ohashi, Tomoyuki Kajiwara, Chenhui Chu, Noriko Takemura, Yuta Nakashima, Hajime Nagahara
We introduce the IDSOU submission for the WNUT-2020 task 2: identification of informative COVID-19 English Tweets.
no code implementations • ACL (WAT) 2021 • YuTing Zhao, Mamoru Komachi, Tomoyuki Kajiwara, Chenhui Chu
We introduce our TMEKU system submitted to the English-Japanese Multimodal Translation Task for WAT 2021.
no code implementations • CLIB 2022 • Iglika Nikolova-Stoupak, Shuichiro Shimizu, Chenhui Chu, Sadao Kurohashi
The corpus utilised to train machine translation models in the study is CCMatrix, provided by OPUS.
1 code implementation • ACL 2022 • Yongmin Kim, Chenhui Chu, Sadao Kurohashi
Existing visual grounding datasets are artificially made, where every query regarding an entity must be able to be grounded to a corresponding image region, i. e., answerable.
no code implementations • AACL (WAT) 2020 • Zhuoyuan Mao, Yibin Shen, Chenhui Chu, Sadao Kurohashi, Cheqing Jin
This paper describes the Japanese-Chinese Neural Machine Translation (NMT) system submitted by the joint team of Kyoto University and East China Normal University (Kyoto-U+ECNU) to WAT 2020 (Nakazawa et al., 2020).
no code implementations • ACL (WAT) 2021 • Toshiaki Nakazawa, Hideki Nakayama, Chenchen Ding, Raj Dabre, Shohei Higashiyama, Hideya Mino, Isao Goto, Win Pa Pa, Anoop Kunchukuttan, Shantipriya Parida, Ondřej Bojar, Chenhui Chu, Akiko Eriguchi, Kaori Abe, Yusuke Oda, Sadao Kurohashi
This paper presents the results of the shared tasks from the 8th workshop on Asian translation (WAT2021).
no code implementations • 20 Aug 2024 • Chengzhi Zhong, Fei Cheng, Qianying Liu, Junfeng Jiang, Zhen Wan, Chenhui Chu, Yugo Murawaki, Sadao Kurohashi
We examine the latent language of three typical categories of models for Japanese processing: Llama2, an English-centric model; Swallow, an English-centric model with continued pre-training in Japanese; and LLM-jp, a model pre-trained on balanced English and Japanese corpora.
no code implementations • 5 Aug 2024 • Yahui Fu, Chenhui Chu, Tatsuya Kawahara
Recent approaches for empathetic response generation mainly focus on emotional resonance and user understanding, without considering the system's personality.
no code implementations • 21 May 2024 • Sirou Chen, Sakiko Yahata, Shuichiro Shimizu, Zhengdong Yang, Yihang Li, Chenhui Chu, Sadao Kurohashi
Emotion plays a crucial role in human conversation.
no code implementations • 23 Mar 2024 • Hao Wang, Tang Li, Chenhui Chu, Nengjun Zhu, Rui Wang, Pinpin Zhu
This approach aims to generate relation representations that are more aware of the spatial context and unseen relation in a manner similar to human perception.
2 code implementations • 6 Mar 2024 • Yikun Sun, Zhen Wan, Nobuhiro Ueda, Sakiko Yahata, Fei Cheng, Chenhui Chu, Sadao Kurohashi
The creation of instruction data and evaluation benchmarks for serving Large language models often involves enormous human annotation.
no code implementations • 24 Jan 2024 • Wangjin Zhou, Zhengdong Yang, Chenhui Chu, Sheng Li, Raj Dabre, Yi Zhao, Tatsuya Kawahara
We propose MOS-FAD, where MOS can be leveraged at two key points in FAD: training data selection and model fusion.
no code implementations • 24 Jan 2024 • Duzhen Zhang, Yahan Yu, Jiahua Dong, Chenxing Li, Dan Su, Chenhui Chu, Dong Yu
In the past year, MultiModal Large Language Models (MM-LLMs) have undergone substantial advancements, augmenting off-the-shelf LLMs to support MM inputs or outputs via cost-effective training strategies.
no code implementations • 21 Nov 2023 • Wenqing Wei, Zhengdong Yang, Yuan Gao, Jiyi Li, Chenhui Chu, Shogo Okada, Sheng Li
The early-stage Alzheimer's disease (AD) detection has been considered an important field of medical studies.
1 code implementation • 7 Nov 2023 • Haiyue Song, Raj Dabre, Chenhui Chu, Atsushi Fujita, Sadao Kurohashi
To create the parallel corpora, we propose a dynamic programming based sentence alignment algorithm which leverages the cosine similarity of machine-translated sentences.
1 code implementation • 31 Oct 2023 • Yihang Li, Shuichiro Shimizu, Chenhui Chu, Sadao Kurohashi, Wei Li
In addition to the extensive training set, EVA contains a video-helpful evaluation set in which subtitles are ambiguous, and videos are guaranteed helpful for disambiguation.
1 code implementation • 23 Oct 2023 • Hao Wang, Qingxuan Wang, Yue Li, Changqing Wang, Chenhui Chu, Rui Wang
The use of visually-rich documents (VRDs) in various fields has created a demand for Document AI models that can read and comprehend documents like humans, which requires the overcoming of technical, linguistic, and cognitive barriers.
no code implementations • 23 Oct 2023 • Hao Wang, Xiahua Chen, Rui Wang, Chenhui Chu
Extracting meaningful entities belonging to predefined categories from Visually-rich Form-like Documents (VFDs) is a challenging task.
no code implementations • 31 Jul 2023 • Haiyue Song, Raj Dabre, Chenhui Chu, Sadao Kurohashi, Eiichiro Sumita
Sub-word segmentation is an essential pre-processing step for Neural Machine Translation (NMT).
no code implementations • 28 Jul 2023 • Yahui Fu, Koji Inoue, Chenhui Chu, Tatsuya Kawahara
We enhance ChatGPT's ability to reason for the system's perspective by integrating in-context learning with commonsense knowledge.
no code implementations • 17 May 2023 • Zhuoyuan Mao, Haiyue Song, Raj Dabre, Chenhui Chu, Sadao Kurohashi
The language-independency of encoded representations within multilingual neural machine translation (MNMT) models is crucial for their generalization ability on zero-shot translation.
1 code implementation • 16 May 2023 • Shuichiro Shimizu, Chenhui Chu, Sheng Li, Sadao Kurohashi
We present a new task, speech dialogue translation mediating speakers of different languages.
no code implementations • 16 May 2023 • Zhuoyuan Mao, Raj Dabre, Qianying Liu, Haiyue Song, Chenhui Chu, Sadao Kurohashi
This paper studies the impact of layer normalization (LayerNorm) on zero-shot translation (ZST).
no code implementations • 23 Aug 2022 • Tianwei Chen, Noa Garcia, Mayu Otani, Chenhui Chu, Yuta Nakashima, Hajime Nagahara
Is more data always better to train vision-and-language models?
1 code implementation • 31 May 2022 • Zhuoyuan Mao, Chenhui Chu, Sadao Kurohashi
Massively multilingual sentence representation models, e. g., LASER, SBERT-distill, and LaBSE, help significantly improve cross-lingual downstream tasks.
no code implementations • Findings (NAACL) 2022 • Zhuoyuan Mao, Chenhui Chu, Raj Dabre, Haiyue Song, Zhen Wan, Sadao Kurohashi
Meanwhile, the contrastive objective can implicitly utilize automatically learned word alignment, which has not been explored in many-to-many NMT.
no code implementations • 11 Apr 2022 • Zhengdong Yang, Wangjin Zhou, Chenhui Chu, Sheng Li, Raj Dabre, Raphael Rubino, Yi Zhao
This challenge aims to predict MOS scores of synthetic speech on two tracks, the main track and a more challenging sub-track: out-of-domain (OOD).
1 code implementation • 8 Apr 2022 • Qianying Liu, Zhuo Gong, Zhengdong Yang, Yuhang Yang, Sheng Li, Chenchen Ding, Nobuaki Minematsu, Hao Huang, Fei Cheng, Chenhui Chu, Sadao Kurohashi
Low-resource speech recognition has been long-suffering from insufficient training data.
1 code implementation • 20 Jan 2022 • Zhuoyuan Mao, Chenhui Chu, Sadao Kurohashi
In the present study, we propose novel sequence-to-sequence pre-training objectives for low-resource machine translation (NMT): Japanese-specific sequence to sequence (JASS) for language pairs involving Japanese as the source or target language, and English-specific sequence to sequence (ENSS) for language pairs involving English.
Low Resource Neural Machine Translation Low-Resource Neural Machine Translation +2
1 code implementation • LREC 2022 • Yihang Li, Shuichiro Shimizu, Weiqi Gu, Chenhui Chu, Sadao Kurohashi
Existing multimodal machine translation (MMT) datasets consist of images and video captions or general subtitles, which rarely contain linguistic ambiguity, making visual information not so effective to generate appropriate translations.
no code implementations • 26 Oct 2021 • Tianran Wu, Noa Garcia, Mayu Otani, Chenhui Chu, Yuta Nakashima, Haruo Takemura
Video question answering (VideoQA) is designed to answer a given question based on a relevant video clip.
no code implementations • ACL 2021 • Jules Samaran, Noa Garcia, Mayu Otani, Chenhui Chu, Yuta Nakashima
The impressive performances of pre-trained visually grounded language models have motivated a growing body of research investigating what has been learned during the pre-training.
no code implementations • ACL 2021 • Weiqi Gu, Haiyue Song, Chenhui Chu, Sadao Kurohashi
Video-guided machine translation, as one type of multimodal machine translations, aims to engage video contents as auxiliary information to address the word sense ambiguity problem in machine translation.
no code implementations • 25 Jun 2021 • Yusuke Hirota, Noa Garcia, Mayu Otani, Chenhui Chu, Yuta Nakashima, Ittetsu Taniguchi, Takao Onoye
This paper delves into the effectiveness of textual representations for image understanding in the specific context of VQA.
1 code implementation • NAACL 2021 • Tomoyuki Kajiwara, Chenhui Chu, Noriko Takemura, Yuta Nakashima, Hajime Nagahara
We annotate 17, 000 SNS posts with both the writer{'}s subjective emotional intensity and the reader{'}s objective one to construct a Japanese emotion analysis dataset.
1 code implementation • ACL 2021 • Zhuoyuan Mao, Prakhar Gupta, Pei Wang, Chenhui Chu, Martin Jaggi, Sadao Kurohashi
Large-scale models for learning fixed-dimensional cross-lingual sentence representations like LASER (Artetxe and Schwenk, 2019b) lead to significant improvement in performance on downstream tasks.
no code implementations • 25 May 2021 • Cheikh Brahim El Vaigh, Noa Garcia, Benjamin Renoust, Chenhui Chu, Yuta Nakashima, Hajime Nagahara
In this paper, we propose a novel use of a knowledge graph, that is constructed on annotated data and pseudo-labeled data.
no code implementations • 14 Jan 2021 • Vinay Damodaran, Sharanya Chakravarthy, Akshay Kumar, Anjana Umapathy, Teruko Mitamura, Yuta Nakashima, Noa Garcia, Chenhui Chu
Visual Question Answering (VQA) is of tremendous interest to the research community with important applications such as aiding visually impaired users and image-based search.
no code implementations • COLING 2020 • Raj Dabre, Chenhui Chu, Anoop Kunchukuttan
The advent of neural machine translation (NMT) has opened up exciting research in building multilingual translation systems i. e. translation models that can handle more than one language pair.
1 code implementation • EAMT 2020 • YuTing Zhao, Mamoru Komachi, Tomoyuki Kajiwara, Chenhui Chu
In contrast, we propose the application of semantic image regions for MNMT by integrating visual and textual features using two individual attention mechanisms (double attention).
no code implementations • 17 Oct 2020 • Andrew Merritt, Chenhui Chu, Yuki Arase
Multimodal neural machine translation (NMT) has become an increasingly important area of research over the years because additional modalities, such as image data, can provide more context to textual data.
no code implementations • 11 Oct 2020 • Vipul Mishra, Chenhui Chu, Yuki Arase
Lexically cohesive translations preserve consistency in word choices in document-level translation.
no code implementations • 11 Oct 2020 • Chenhui Chu, Yuto Takebayashi, Mishra Vipul, Yuta Nakashima
Visual relationship detection is crucial for scene understanding in images.
1 code implementation • 28 Aug 2020 • Noa Garcia, Chentao Ye, Zihua Liu, Qingtao Hu, Mayu Otani, Chenhui Chu, Yuta Nakashima, Teruko Mitamura
Our dataset inherently consists of visual (painting-based) and knowledge (comment-based) questions.
no code implementations • ACL 2020 • Sora Ohashi, Junya Takayama, Tomoyuki Kajiwara, Chenhui Chu, Yuki Arase
Advanced pre-trained models for text representation have achieved state-of-the-art performance on various text classification tasks.
no code implementations • LREC 2020 • Yuki Arase, Tomoyuki Kajiwara, Chenhui Chu
The dataset we present in this paper is unique for the richness of annotated information, including detailed descriptions of drug reactions with full context.
no code implementations • LREC 2020 • Koji Tanaka, Chenhui Chu, Haolin Ren, Benjamin Renoust, Yuta Nakashima, Noriko Takemura, Hajime Nagahara, Takao Fujikawa
In this paper, we propose a full pipeline of analysis of a large corpus about a century of public meeting in historical Australian news papers, from construction to visual exploration.
no code implementations • 17 Apr 2020 • Noa Garcia, Mayu Otani, Chenhui Chu, Yuta Nakashima
We propose a novel video understanding task by fusing knowledge-based and video question answering.
no code implementations • 4 Jan 2020 • Raj Dabre, Chenhui Chu, Anoop Kunchukuttan
We present a survey on multilingual neural machine translation (MNMT), which has gained a lot of traction in the recent years.
no code implementations • IJCNLP 2019 • Raj Dabre, Atsushi Fujita, Chenhui Chu
This paper highlights the impressive utility of multi-parallel corpora for transfer learning in a one-to-many low-resource neural machine translation (NMT) setting.
Low Resource Neural Machine Translation Low-Resource Neural Machine Translation +3
no code implementations • 23 Oct 2019 • Noa Garcia, Mayu Otani, Chenhui Chu, Yuta Nakashima
We propose a novel video understanding task by fusing knowledge-based and video question answering.
no code implementations • 26 Aug 2019 • Haipeng Sun, Rui Wang, Kehai Chen, Masao Utiyama, Eiichiro Sumita, Tiejun Zhao, Chenhui Chu
However, it has not been well-studied for unsupervised neural machine translation (UNMT), although UNMT has recently achieved remarkable results in several domain-specific language pairs.
no code implementations • 19 Jun 2019 • Chenhui Chu, Raj Dabre
In this paper, we propose two novel methods for domain adaptation for the attention-only neural machine translation (NMT) model, i. e., the Transformer.
no code implementations • 31 May 2019 • Tomoyuki Kajiwara, Chihiro Tanikawa, Yuujin Shimizu, Chenhui Chu, Takashi Yamashiro, Hajime Nagahara
We work on the task of automatically designing a treatment plan from the findings included in the medical certificate written by the dentist.
no code implementations • 14 May 2019 • Raj Dabre, Chenhui Chu, Anoop Kunchukuttan
We present a survey on multilingual neural machine translation (MNMT), which has gained a lot of traction in the recent years.
1 code implementation • COLING 2018 • Chenhui Chu, Mayu Otani, Yuta Nakashima
These extracted VGPs have the potential to improve language and image multimodal tasks such as visual question answering and image captioning.
no code implementations • COLING 2018 • Chenhui Chu, Rui Wang
Neural machine translation (NMT) is a deep learning based approach for machine translation, which yields the state-of-the-art translation performance in scenarios where large-scale parallel corpora are available.
no code implementations • ACL 2018 • Yuki Kawara, Chenhui Chu, Yuki Arase
Experiments show that the proposed method achieves comparable gain in translation quality to the state-of-the-art method but without a manual feature design.
no code implementations • ACL 2017 • Chenhui Chu, Raj Dabre, Sadao Kurohashi
In this paper, we propose a novel domain adaptation method named {``}mixed fine tuning{''} for neural machine translation (NMT).
no code implementations • 12 Jan 2017 • Chenhui Chu, Raj Dabre, Sadao Kurohashi
In this paper, we propose a novel domain adaptation method named "mixed fine tuning" for neural machine translation (NMT).
1 code implementation • WS 2016 • Fabien Cromieres, Chenhui Chu, Toshiaki Nakazawa, Sadao Kurohashi
We report very good translation results, especially when using neural MT for Chinese-to-Japanese translation.
no code implementations • WS 2016 • Chenhui Chu, Toshiaki Nakazawa, Daisuke Kawahara, Sadao Kurohashi
Treebanks are curial for natural language processing (NLP).
no code implementations • COLING 2016 • Mo Shen, Wingmui Li, HyunJeong Choe, Chenhui Chu, Daisuke Kawahara, Sadao Kurohashi
In this paper, we propose a new annotation approach to Chinese word segmentation, part-of-speech (POS) tagging and dependency labelling that aims to overcome the two major issues in traditional morphology-based annotation: Inconsistency and data sparsity.
no code implementations • 7 Jun 2016 • Chenhui Chu, Sadao Kurohashi
As alignment links are not given between English sentences and Abstract Meaning Representation (AMR) graphs in the AMR annotation, automatic alignment becomes indispensable for training an AMR parser.
no code implementations • LREC 2016 • Chenhui Chu, Sadao Kurohashi
Out-of-vocabulary (OOV) word is a crucial problem in statistical machine translation (SMT) with low resources.
no code implementations • LREC 2016 • Chenhui Chu, Raj Dabre, Sadao Kurohashi
Parallel corpora are crucial for machine translation (MT), however they are quite scarce for most language pairs and domains.
no code implementations • LREC 2016 • Antoine Bourlon, Chenhui Chu, Toshiaki Nakazawa, Sadao Kurohashi
Sentence alignment is a task that consists in aligning the parallel sentences in a translated article pair.
no code implementations • LREC 2014 • Chenhui Chu, Toshiaki Nakazawa, Sadao Kurohashi
Using the system, we construct a Chinese―Japanese parallel corpus with more than 126k highly accurate parallel sentences from Wikipedia.
no code implementations • LREC 2012 • Chenhui Chu, Toshiaki Nakazawa, Sadao Kurohashi
Chinese characters are used both in Japanese and Chinese, which are called Kanji and Hanzi respectively.