Search Results for author: Tomoya Iwakura

Found 24 papers, 0 papers with code

Evaluating Hierarchical Document Categorisation

no code implementations ALTA 2021 Qian Sun, Aili Shen, Hiyori Yoshikawa, Chunpeng Ma, Daniel Beck, Tomoya Iwakura, Timothy Baldwin

Hierarchical document categorisation is a special case of multi-label document categorisation, where there is a taxonomic hierarchy among the labels.

Relation Extraction Using Multiple Pre-Training Models in Biomedical Domain

no code implementations RANLP 2021 Satoshi Hiai, Kazutaka Shimada, Taiki Watanabe, Akiva Miura, Tomoya Iwakura

In addition, our method shows approximately three times faster extraction speed than the BERT-based models on the ChemProt corpus and reduces the memory size to one sixth of the BERT ones.

Relation Relation Extraction

Tell Me What You Read: Automatic Expertise-Based Annotator Assignment for Text Annotation in Expert Domains

no code implementations RANLP 2021 Hiyori Yoshikawa, Tomoya Iwakura, Kimi Kaneko, Hiroaki Yoshida, Yasutaka Kumano, Kazutaka Shimada, Rafal Rzepka, Patrycja Swieczkowska

To address the issue, we propose a method to estimate the domain expertise of each annotator before the annotation process using information easily available from the annotators beforehand.

text annotation

Learning Entity-Likeness with Multiple Approximate Matches for Biomedical NER

no code implementations RANLP 2021 An Nguyen Le, Hajime Morita, Tomoya Iwakura

Biomedical Named Entities are complex, so approximate matching has been used to improve entity coverage.

NER

Constructing Multilingual Visual-Text Datasets Revealing Visual Multilingual Ability of Vision Language Models

no code implementations29 Mar 2024 Jesse Atuhurra, Iqra Ali, Tatsuya Hiraoka, Hidetaka Kamigaito, Tomoya Iwakura, Taro Watanabe

Our contribution is four-fold: 1) we introduced nine vision-and-language (VL) tasks (including object recognition, image-text matching, and more) and constructed multilingual visual-text datasets in four languages: English, Japanese, Swahili, and Urdu through utilizing templates containing \textit{questions} and prompting GPT4-V to generate the \textit{answers} and the \textit{rationales}, 2) introduced a new VL task named \textit{unrelatedness}, 3) introduced rationales to enable human understanding of the VLM reasoning process, and 4) employed human evaluation to measure the suitability of proposed datasets for VL tasks.

Image-text matching Object Recognition +1

Downstream Task-Oriented Neural Tokenizer Optimization with Vocabulary Restriction as Post Processing

no code implementations21 Apr 2023 Tatsuya Hiraoka, Tomoya Iwakura

This paper proposes an example of the BiLSTM-based tokenizer with vocabulary restriction, which can capture wider contextual information for the tokenization process than non-neural-based tokenization methods used in existing work.

text-classification Text Classification

On the (In)Effectiveness of Images for Text Classification

no code implementations EACL 2021 Chunpeng Ma, Aili Shen, Hiyori Yoshikawa, Tomoya Iwakura, Daniel Beck, Timothy Baldwin

Images are core components of multi-modal learning in natural language processing (NLP), and results have varied substantially as to whether images improve NLP tasks or not.

text-classification Text Classification

Multi-Task Learning for Chemical Named Entity Recognition with Chemical Compound Paraphrasing

no code implementations IJCNLP 2019 Taiki Watanabe, Akihiro Tamura, Takashi Ninomiya, Takuya Makino, Tomoya Iwakura

We propose a method to improve named entity recognition (NER) for chemical compounds using multi-task learning by jointly training a chemical NER model and a chemical com- pound paraphrase model.

Multi-Task Learning named-entity-recognition +2

Global Optimization under Length Constraint for Neural Text Summarization

no code implementations ACL 2019 Takuya Makino, Tomoya Iwakura, Hiroya Takamura, Manabu Okumura

The experimental results show that a state-of-the-art neural summarization model optimized with GOLC generates fewer overlength summaries while maintaining the fastest processing speed; only 6. 70{\%} overlength summaries on CNN/Daily and 7. 8{\%} on long summary of Mainichi, compared to the approximately 20{\%} to 50{\%} on CNN/Daily Mail and 10{\%} to 30{\%} on Mainichi with the other optimization methods.

Document Summarization

Model Transfer with Explicit Knowledge of the Relation between Class Definitions

no code implementations CONLL 2018 Hiyori Yoshikawa, Tomoya Iwakura

Instead of learning the individual classification layers for the support and target schemes, the proposed method converts the class label of each example on the support scheme into a set of candidate class labels on the target scheme via the class correspondence table, and then uses the candidate labels to learn the classification layer for the target scheme.

General Classification Multi-class Classification +4

An Eye-tracking Study of Named Entity Annotation

no code implementations RANLP 2017 Takenobu Tokunaga, Hitoshi Nishikawa, Tomoya Iwakura

Utilising effective features in machine learning-based natural language processing (NLP) is crucial in achieving good performance for a given NLP task.

Active Learning Coreference Resolution +4

Big Community Data before World Wide Web Era

no code implementations WS 2016 Tomoya Iwakura, Tetsuro Takahashi, Akihiro Ohtani, Kunio Matsui

This paper introduces the NIFTY-Serve corpus, a large data archive collected from Japanese discussion forums that operated via a Bulletin Board System (BBS) between 1987 and 2006.

Cannot find the paper you are looking for? You can Submit a new open access paper.