no code implementations • COLING 2022 • Hiroki Teranishi, Yuji Matsumoto
Generating synthetic data for supervised learning from large-scale pre-trained language models has enhanced performances across several NLP tasks, especially in low-resource scenarios.
no code implementations • AMTA 2016 • Ander Martinez, Yuji Matsumoto
This article combines three different ideas (splitting words into smaller units, using an extra dataset of a related language pair and using monolingual data) for improving the performance of NMT models on language pairs with limited data.
no code implementations • SpaNLP (ACL) 2022 • Van-Hien Tran, Hiroki Ouchi, Taro Watanabe, Yuji Matsumoto
Zero-shot relation extraction (ZSRE) aims to predict target relations that cannot be observed during training.
Ranked #3 on Zero-shot Relation Classification on FewRel
no code implementations • 28 May 2024 • Yi-Pei Chen, Noriki Nishida, Hideki Nakayama, Yuji Matsumoto
Enhancing user engagement through personalization in conversational agents has gained significance, especially with the advent of large language models that generate fluent responses.
2 code implementations • 27 Mar 2024 • Lisa Raithel, Hui-Syuan Yeh, Shuntaro Yada, Cyril Grouin, Thomas Lavergne, Aurélie Névéol, Patrick Paroubek, Philippe Thomas, Tomohiro Nishiyama, Sebastian Möller, Eiji Aramaki, Yuji Matsumoto, Roland Roller, Pierre Zweigenbaum
User-generated data sources have gained significance in uncovering Adverse Drug Reactions (ADRs), with an increasing number of discussions occurring in the digital world.
1 code implementation • Journal of Natural Language Processing 2023 • Van-Hien Tran, Hiroki Ouchi, Hiroyuki Shindo, Yuji Matsumoto, Taro Watanabe
This study argues that enhancing the semantic correlation between instances and relations is key to effectively solving the zero-shot relation extraction task.
Ranked #1 on Zero-shot Relation Classification on FewRel
1 code implementation • 2 Jun 2023 • Takashi Wada, Yuji Matsumoto, Timothy Baldwin, Jey Han Lau
We propose an unsupervised approach to paraphrasing multiword expressions (MWEs) in context.
no code implementations • 10 Mar 2023 • Kenichiro Ando, Mamoru Komachi, Takashi Okumura, Hiromasa Horiguchi, Yuji Matsumoto
During the patient's hospitalization, the physician must record daily observations of the patient and summarize them into a brief document called "discharge summary" when the patient is discharged.
1 code implementation • 6 Dec 2022 • Ukyo Honda, Taro Watanabe, Yuji Matsumoto
Discriminativeness is a desirable feature of image captions: captions should describe the characteristic details of input images.
no code implementations • 20 Sep 2022 • Kenichiro Ando, Takashi Okumura, Mamoru Komachi, Hiromasa Horiguchi, Yuji Matsumoto
We first defined three types of summarization units with different granularities to compare the performance of the discharge summary generation: whole sentences, clinical segments, and clauses.
1 code implementation • COLING 2022 • Takashi Wada, Timothy Baldwin, Yuji Matsumoto, Jey Han Lau
We propose a new unsupervised method for lexical substitution using pre-trained language models.
no code implementations • Journal of Natural Language Processing 2021 • Van-Hien Tran, Van-Thuy Phi, Akihiko Kato, Hiroyuki Shindo, Taro Watanabe, Yuji Matsumoto
A recent study (Yu et al. 2020) proposed a novel decomposition strategy that splits the task into two interrelated subtasks: detection of the head-entity (HE) and identification of the corresponding tail-entity and relation (TER) for each extracted head-entity.
no code implementations • Joint Conference on Lexical and Computational Semantics 2021 • Yuki Yamamoto, Yuji Matsumoto, Taro Watanabe
Abstract Meaning Representation (AMR) is a sentence-level meaning representation based on predicate argument structure.
1 code implementation • ACL 2021 • Yiran Wang, Hiroyuki Shindo, Yuji Matsumoto, Taro Watanabe
This paper presents a novel method for nested named entity recognition.
Ranked #13 on Nested Named Entity Recognition on ACE 2005
1 code implementation • EACL 2021 • Ukyo Honda, Yoshitaka Ushiku, Atsushi Hashimoto, Taro Watanabe, Yuji Matsumoto
Unsupervised image captioning is a challenging task that aims at generating captions without the supervision of image-sentence pairs, but only with images and sentences drawn from different sources and object labels detected from the images.
no code implementations • EACL 2021 • Vu Tran, Van-Hien Tran, Phuong Nguyen, Chau Nguyen, Ken Satoh, Yuji Matsumoto, Minh Nguyen
This paper presents CovRelex, a scientific paper retrieval system targeting entities and relations via relation extraction on COVID-19 scientific papers.
no code implementations • COLING 2020 • Yuya Sawada, Takashi Wada, Takayoshi Shibahara, Hiroki Teranishi, Shuhei Kondo, Hiroyuki Shindo, Taro Watanabe, Yuji Matsumoto
We propose a simple method for nominal coordination boundary identification.
1 code implementation • EMNLP (MRL) 2021 • Takashi Wada, Tomoharu Iwata, Yuji Matsumoto, Timothy Baldwin, Jey Han Lau
We propose a new approach for learning contextualised cross-lingual word embeddings based on a small parallel corpus (e. g. a few hundred sentence pairs).
Bilingual Lexicon Induction Cross-Lingual Word Embeddings +5
9 code implementations • EMNLP 2020 • Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, Yuji Matsumoto
In this paper, we propose new pretrained contextualized representations of words and entities based on the bidirectional transformer.
Ranked #1 on Entity Typing on Open Entity
no code implementations • 21 Jan 2020 • Itsumi Saito, Kyosuke Nishida, Kosuke Nishida, Atsushi Otsuka, Hisako Asano, Junji Tomita, Hiroyuki Shindo, Yuji Matsumoto
Unlike the previous models, our length-controllable abstractive summarization model incorporates a word-level extractive module in the encoder-decoder model instead of length embeddings.
1 code implementation • NAACL 2022 • Ikuya Yamada, Koki Washio, Hiroyuki Shindo, Yuji Matsumoto
We propose a global entity disambiguation (ED) model based on BERT.
Ranked #1 on Entity Disambiguation on MSNBC
no code implementations • 31 Aug 2019 • Hiroyuki Shindo, Yuji Matsumoto
Molecule property prediction is a fundamental problem for computer-aided drug discovery and materials science.
no code implementations • 15 Aug 2019 • Hamada A. Nayel, H. L. Shashirekha, Hiroyuki Shindo, Yuji Matsumoto
The proposed SR model, FROBES, improves the representation of multi-word entities.
no code implementations • ACL 2019 • Tatsuya Hiraoka, Hiroyuki Shindo, Yuji Matsumoto
To make the model robust against infrequent tokens, we sampled segmentation for each sentence stochastically during training, which resulted in improved performance of text classification.
1 code implementation • ACL 2019 • Takashi Wada, Tomoharu Iwata, Yuji Matsumoto
Recently, a variety of unsupervised methods have been proposed that map pre-trained word embeddings of different languages into the same space without any parallel data.
no code implementations • NAACL 2019 • Van-Hien Tran, Van-Thuy Phi, Hiroyuki Shindo, Yuji Matsumoto
Recently, relation classification has gained much success by exploiting deep neural networks.
no code implementations • NAACL 2019 • Hiroki Teranishi, Hiroyuki Shindo, Yuji Matsumoto
We propose a simple and accurate model for coordination boundary identification.
no code implementations • EMNLP 2020 • Ikuya Yamada, Akari Asai, Jin Sakuma, Hiroyuki Shindo, Hideaki Takeda, Yoshiyasu Takefuji, Yuji Matsumoto
The embeddings of entities in a large knowledge base (e. g., Wikipedia) are highly beneficial for solving various natural language tasks that involve real world knowledge.
1 code implementation • WS 2019 • Ronen Tamari, Hiroyuki Shindo, Dafna Shahaf, Yuji Matsumoto
Understanding procedural text requires tracking entities, actions and effects as the narrative unfolds.
no code implementations • WS 2018 • Hiroshi Kanayama, Na-Rae Han, Masayuki Asahara, Jena D. Hwang, Yusuke Miyao, Jinho D. Choi, Yuji Matsumoto
This paper discusses the representation of coordinate structures in the Universal Dependencies framework for two head-final languages, Japanese and Korean.
1 code implementation • PACLIC 2018 • Tomoki Matsuno, Katsuhiko Hayashi, Takahiro Ishihara, Hitoshi Manabe, Yuji Matsumoto
Currently, the biaffine classifier has been attracting attention as a method to introduce an attention mechanism into the modeling of binary relations.
2 code implementations • EMNLP 2018 • Hiroki Ouchi, Hiroyuki Shindo, Yuji Matsumoto
We present a simple and accurate span-based model for semantic role labeling (SRL).
Ranked #7 on Semantic Role Labeling on CoNLL 2005
no code implementations • COLING 2018 • Ryosuke Kohita, Hiroshi Noji, Yuji Matsumoto
One main challenge for incremental transition-based parsers, when future inputs are invisible, is to extract good features from a limited local context.
no code implementations • COLING 2018 • Yuji Matsumoto, Akihiko Kato, Hiroyuki Shindo, Toshio Morita
Those two tools cooperate so that the words and multi-word expressions stored in Cradle are directly referred to by ChaKi in conducting corpus annotation, and the words and expressions annotated in ChaKi can be output as a list of lexical entities that are to be stored in Cradle.
no code implementations • ACL 2018 • Jun Liu, Hiroyuki Shindo, Yuji Matsumoto
We present a computer-assisted learning system, Jastudy, which is particularly designed for Chinese-speaking learners of Japanese as a second language (JSL) to learn Japanese functional expressions with suggestion of appropriate example sentences.
1 code implementation • ACL 2018 • Van-Thuy Phi, Joan Santoso, Masashi Shimbo, Yuji Matsumoto
This paper addresses the tasks of automatic seed selection for bootstrapping relation extraction, and noise reduction for distantly supervised relation extraction.
no code implementations • 11 Jun 2018 • Yutaro Shigeto, Masashi Shimbo, Yuji Matsumoto
This paper proposes an inexpensive way to learn an effective dissimilarity function to be used for $k$-nearest neighbor ($k$-NN) classification.
2 code implementations • 8 May 2018 • Motoki Sato, Jun Suzuki, Hiroyuki Shindo, Yuji Matsumoto
This paper restores interpretability to such methods by restricting the directions of perturbations toward the existing words in the input embedding space.
no code implementations • IJCNLP 2017 • Itsumi Saito, Jun Suzuki, Kyosuke Nishida, Kugatsu Sadamitsu, Satoshi Kobashikawa, Ryo Masumura, Yuji Matsumoto, Junji Tomita
In this study, we investigated the effectiveness of augmented data for encoder-decoder-based neural normalization models.
no code implementations • IJCNLP 2017 • Motoki Sato, Hiroyuki Shindo, Ikuya Yamada, Yuji Matsumoto
We present Segment-level Neural CRF, which combines neural networks with a linear chain CRF for segment-level sequence modeling tasks such as named entity recognition (NER) and syntactic chunking.
no code implementations • IJCNLP 2017 • Frances Yung, Hiroshi Noji, Yuji Matsumoto
Humans process language word by word and construct partial linguistic structures on the fly before the end of the sentence is perceived.
no code implementations • IJCNLP 2017 • Hiroki Teranishi, Hiroyuki Shindo, Yuji Matsumoto
We propose a neural network model for coordination boundary detection.
no code implementations • IJCNLP 2017 • An Nguyen Le, Ander Martinez, Akifumi Yoshimoto, Yuji Matsumoto
In order to assess the performance, we construct model based on an attention mechanism encoder-decoder model in which the source language is input to the encoder as a sequence and the decoder generates the target language as a linearized dependency tree structure.
no code implementations • WS 2017 • Ryosuke Kohita, Hiroshi Noji, Yuji Matsumoto
We present a new transition system with word reordering for unrestricted non-projective dependency parsing.
no code implementations • CONLL 2017 • Motoki Sato, Hitoshi Manabe, Hiroshi Noji, Yuji Matsumoto
We describe our submission to the CoNLL 2017 shared task, which exploits the shared common knowledge of a language across different domains via a domain adaptation technique.
no code implementations • CONLL 2017 • Go Inoue, Hiroyuki Shindo, Yuji Matsumoto
One reason for this is that in the tagging scheme for such languages, a complete POS tag is formed by combining tags from multiple tag sets defined for each morphosyntactic category.
no code implementations • ACL 2017 • Hiroki Ouchi, Hiroyuki Shindo, Yuji Matsumoto
The performance of Japanese predicate argument structure (PAS) analysis has improved in recent years thanks to the joint modeling of interactions between multiple predicates.
no code implementations • ACL 2017 • Akihiko Kato, Hiroyuki Shindo, Yuji Matsumoto
Because syntactic structures and spans of multiword expressions (MWEs) are independently annotated in many English syntactic corpora, they are generally inconsistent with respect to one another, which is harmful to the implementation of an aggregate system.
1 code implementation • 18 Jun 2017 • Takuo Hamaguchi, Hidekazu Oiwa, Masashi Shimbo, Yuji Matsumoto
Knowledge base completion (KBC) aims to predict missing information in a knowledge base. In this paper, we address the out-of-knowledge-base (OOKB) entity problem in KBC:how to answer queries concerning test entities not observed at training time.
1 code implementation • ACL 2017 • Masashi Yoshikawa, Hiroshi Noji, Yuji Matsumoto
Our model achieves the state-of-the-art results on English and Japanese CCG parsing.
1 code implementation • EACL 2017 • Ryosuke Kohita, Hiroshi Noji, Yuji Matsumoto
Universal Dependencies (UD) is becoming a standard annotation scheme cross-linguistically, but it is argued that this scheme centering on content words is harder to parse than the conventional one centering on function words.
no code implementations • 22 Feb 2017 • Ai Azuma, Masashi Shimbo, Yuji Matsumoto
back propagation) on computation graphs with addition and multiplication, and so on.
no code implementations • WS 2016 • Jun Liu, Yuji Matsumoto
Learning functional expressions is one of the difficulties for language learners, since functional expressions tend to have multiple meanings and complicated usages in various situations.
no code implementations • COLING 2016 • Masayuki Asahara, Yuji Matsumoto, Toshio Morita
ChaKi. NET is a corpus management system for dependency structure annotated corpora.
no code implementations • WS 2016 • Masaru Fuji, Masao Utiyama, Eiichiro Sumita, Yuji Matsumoto
When translating formal documents, capturing the sentence structure specific to the sublanguage is extremely necessary to obtain high-quality translations.
no code implementations • WS 2016 • Muhaimin Hading, Yuji Matsumoto, Maki Sakamoto
This paper introduces Japanese lexical simplification.
no code implementations • WS 2016 • Ayaka Morimoto, Akifumi Yoshimoto, Akihiko Kato, Hiroyuki Shindo, Yuji Matsumoto
This paper presents our ongoing work on compilation of English multi-word expression (MWE) lexicon.
no code implementations • WS 2016 • Masayuki Asahara, Yuji Matsumoto
Paratactic syntactic structures are difficult to represent in syntactic dependency tree structures.
no code implementations • WS 2016 • Taishi Ikeda, Hiroyuki Shindo, Yuji Matsumoto
Text normalization is the task of transforming lexical variants to their canonical forms.
no code implementations • LREC 2016 • Akihiko Kato, Hiroyuki Shindo, Yuji Matsumoto
Nevertheless, this method often leads to the following problem: A node derived from an MWE could have multiple heads and the whole dependency structure including MWE might be cyclic.
no code implementations • LREC 2016 • Takaaki Tanaka, Yusuke Miyao, Masayuki Asahara, Sumire Uematsu, Hiroshi Kanayama, Shinsuke Mori, Yuji Matsumoto
We present an attempt to port the international syntactic annotation scheme, Universal Dependencies, to the Japanese language in this paper.
no code implementations • 22 Apr 2016 • Adhiguna Kuncoro, Yuichiro Sawai, Kevin Duh, Yuji Matsumoto
We propose a transition-based dependency parser using Recurrent Neural Networks with Long Short-Term Memory (LSTM) units.
no code implementations • 3 Jul 2015 • Yutaro Shigeto, Ikumi Suzuki, Kazuo Hara, Masashi Shimbo, Yuji Matsumoto
This paper discusses the effect of hubness in zero-shot learning, when ridge regression is used to find a mapping between the example space to the label space.
no code implementations • LREC 2014 • Fei Cheng, Kevin Duh, Yuji Matsumoto
Synthetic word analysis is a potentially important but relatively unexplored problem in Chinese natural language processing.
no code implementations • LREC 2014 • Lis Pereira, Elga Strafella, Yuji Matsumoto
This work presents an initial investigation on how to distinguish collocations from free combinations.
no code implementations • TACL 2013 • Katsuhiko Hayashi, Shuhei Kondo, Yuji Matsumoto
This paper proposes a discriminative forest reranking algorithm for dependency parsing that can be seen as a form of efficient stacked parsing.
no code implementations • LREC 2012 • Toshinobu Ogiso, Mamoru Komachi, Yasuharu Den, Yuji Matsumoto
In order to construct an annotated diachronic corpus of Japanese, we propose to create a new dictionary for morphological analysis of Early Middle Japanese (Classical Japanese) based on UniDic, a dictionary for Contemporary Japanese.