no code implementations • BioNLP (ACL) 2022 • Taiki Watanabe, Tomoya Ichikawa, Akihiro Tamura, Tomoya Iwakura, Chunpeng Ma, Tsuneo Kato
As one of the NER improvement approaches, multi-task learning that learns a model from multiple training data has been used.
no code implementations • ALTA 2021 • Qian Sun, Aili Shen, Hiyori Yoshikawa, Chunpeng Ma, Daniel Beck, Tomoya Iwakura, Timothy Baldwin
Hierarchical document categorisation is a special case of multi-label document categorisation, where there is a taxonomic hierarchy among the labels.
no code implementations • RANLP 2021 • Satoshi Hiai, Kazutaka Shimada, Taiki Watanabe, Akiva Miura, Tomoya Iwakura
In addition, our method shows approximately three times faster extraction speed than the BERT-based models on the ChemProt corpus and reduces the memory size to one sixth of the BERT ones.
no code implementations • RANLP 2021 • Hiyori Yoshikawa, Tomoya Iwakura, Kimi Kaneko, Hiroaki Yoshida, Yasutaka Kumano, Kazutaka Shimada, Rafal Rzepka, Patrycja Swieczkowska
To address the issue, we propose a method to estimate the domain expertise of each annotator before the annotation process using information easily available from the annotators beforehand.
no code implementations • RANLP 2021 • An Nguyen Le, Hajime Morita, Tomoya Iwakura
Biomedical Named Entities are complex, so approximate matching has been used to improve entity coverage.
no code implementations • RANLP 2021 • Kyoumoto Matsushita, Takuya Makino, Tomoya Iwakura
Most neural-based NLP models receive only vectors encoded from a sequence of subwords obtained from an input text.
no code implementations • 10 Sep 2024 • Kohei Tsuji, Tatsuya Hiraoka, Yuchang Cheng, Tomoya Iwakura
We propose a novel method to reduce the time of error detection.
no code implementations • 29 Mar 2024 • Jesse Atuhurra, Iqra Ali, Tatsuya Hiraoka, Hidetaka Kamigaito, Tomoya Iwakura, Taro Watanabe
Our contribution is four-fold: 1) we introduced nine vision-and-language (VL) tasks (including object recognition, image-text matching, and more) and constructed multilingual visual-text datasets in four languages: English, Japanese, Swahili, and Urdu through utilizing templates containing \textit{questions} and prompting GPT4-V to generate the \textit{answers} and the \textit{rationales}, 2) introduced a new VL task named \textit{unrelatedness}, 3) introduced rationales to enable human understanding of the VLM reasoning process, and 4) employed human evaluation to measure the suitability of proposed datasets for VL tasks.
no code implementations • 21 Apr 2023 • Tatsuya Hiraoka, Tomoya Iwakura
This paper proposes an example of the BiLSTM-based tokenizer with vocabulary restriction, which can capture wider contextual information for the tokenization process than non-neural-based tokenization methods used in existing work.
no code implementations • 21 Apr 2023 • Tatsuya Hiraoka, Tomoya Iwakura
Is preferred tokenization for humans also preferred for machine-learning (ML) models?
no code implementations • EACL 2021 • Chunpeng Ma, Aili Shen, Hiyori Yoshikawa, Tomoya Iwakura, Daniel Beck, Timothy Baldwin
Images are core components of multi-modal learning in natural language processing (NLP), and results have varied substantially as to whether images improve NLP tasks or not.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Yutaro Omote, Kyoumoto Matsushita, Tomoya Iwakura, Akihiro Tamura, Takashi Ninomiya
Instead of handcrafted rules, we propose Transformer-based models that predict SMILES strings from chemical compound names.
no code implementations • IJCNLP 2019 • Taiki Watanabe, Akihiro Tamura, Takashi Ninomiya, Takuya Makino, Tomoya Iwakura
We propose a method to improve named entity recognition (NER) for chemical compounds using multi-task learning by jointly training a chemical NER model and a chemical com- pound paraphrase model.
no code implementations • RANLP 2019 • Hajime Morita, Tomoya Iwakura
This paper proposes a partially deterministic morphological analysis method for improved processing speed.
no code implementations • ACL 2019 • Takuya Makino, Tomoya Iwakura, Hiroya Takamura, Manabu Okumura
The experimental results show that a state-of-the-art neural summarization model optimized with GOLC generates fewer overlength summaries while maintaining the fastest processing speed; only 6. 70{\%} overlength summaries on CNN/Daily and 7. 8{\%} on long summary of Mainichi, compared to the approximately 20{\%} to 50{\%} on CNN/Daily Mail and 10{\%} to 30{\%} on Mainichi with the other optimization methods.
no code implementations • CONLL 2018 • Hiyori Yoshikawa, Tomoya Iwakura
Instead of learning the individual classification layers for the support and target schemes, the proposed method converts the class label of each example on the support scheme into a set of candidate class labels on the target scheme via the class correspondence table, and then uses the candidate labels to learn the classification layer for the target scheme.
no code implementations • COLING 2018 • Tomoya Iwakura, Seiji Okajima, Nobuyuki Igata, Kunihiro Takeda, Yuzuru Yamakage, Naoshi Morita
We present our system that assists to detect heavy rain disaster, which is being used in real world in Japan.
no code implementations • RANLP 2017 • Takenobu Tokunaga, Hitoshi Nishikawa, Tomoya Iwakura
Utilising effective features in machine learning-based natural language processing (NLP) is crucial in achieving good performance for a given NLP task.
no code implementations • WS 2016 • Tomoya Iwakura, Tetsuro Takahashi, Akihiro Ohtani, Kunio Matsui
This paper introduces the NIFTY-Serve corpus, a large data archive collected from Japanese discussion forums that operated via a Bulletin Board System (BBS) between 1987 and 2006.