1 code implementation • ECCV 2020 • Soichiro Fujita, Tsutomu Hirao, Hidetaka Kamigaito, Manabu Okumura, Masaaki Nagata
This paper proposes a new evaluation framework, Story Oriented Dense video cAptioning evaluation framework (SODA), for measuring the performance of video story description systems.
1 code implementation • 13 Apr 2024 • Hayato Tsukagoshi, Tsutomu Hirao, Makoto Morishita, Katsuki Chousa, Ryohei Sasano, Koichi Takeda
The task of Split and Rephrase, which splits a complex sentence into multiple simple sentences with the same meaning, improves readability and enhances the performance of downstream tasks in natural language processing (NLP).
1 code implementation • 8 Mar 2024 • Aru Maekawa, Tsutomu Hirao, Hidetaka Kamigaito, Manabu Okumura
Recently, decoder-only pre-trained large language models (LLMs), with several tens of billion parameters, have significantly impacted a wide range of natural language processing (NLP) tasks.
1 code implementation • 15 Oct 2022 • Naoki Kobayashi, Tsutomu Hirao, Hidetaka Kamigaito, Manabu Okumura, Masaaki Nagata
To promote and further develop RST-style discourse parsing models, we need a strong baseline that can be regarded as a reference for reporting reliable experimental results.
Ranked #1 on Discourse Parsing on Instructional-DT (Instr-DT)
no code implementations • NAACL 2021 • Naoki Kobayashi, Tsutomu Hirao, Hidetaka Kamigaito, Manabu Okumura, Masaaki Nagata
We then pre-train a neural RST parser with the obtained silver data and fine-tune it on the RST-DT.
Ranked #2 on Discourse Parsing on RST-DT (using extra training data)
no code implementations • Findings of the Association for Computational Linguistics 2020 • Kosuke Yamada, Tsutomu Hirao, Ryohei Sasano, Koichi Takeda, Masaaki Nagata
Dividing biomedical abstracts into several segments with rhetorical roles is essential for supporting researchers{'} information access in the biomedical domain.
1 code implementation • 3 Apr 2020 • Naoki Kobayashi, Tsutomu Hirao, Hidetaka Kamigaito, Manabu Okumura, Masaaki Nagata
To obtain better discourse dependency trees, we need to improve the accuracy of RST trees at the upper parts of the structures.
Ranked #3 on Discourse Parsing on RST-DT
no code implementations • 24 Mar 2020 • Hiroki Ikeuchi, Akio Watanabe, Tsutomu Hirao, Makoto Morishita, Masaaki Nishino, Yoichi Matsuo, Keishiro Watanabe
With the increase in scale and complexity of ICT systems, their operation increasingly requires automatic recovery from failures.
no code implementations • IJCNLP 2019 • Masaaki Nishino, Sho Takase, Tsutomu Hirao, Masaaki Nagata
An anagram is a sentence or a phrase that is made by permutating the characters of an input sentence or a phrase.
no code implementations • IJCNLP 2019 • Naoki Kobayashi, Tsutomu Hirao, Kengo Nakamura, Hidetaka Kamigaito, Manabu Okumura, Masaaki Nagata
The first one builds the optimal tree in terms of a dissimilarity score function that is defined for splitting a text span into smaller ones.
no code implementations • WS 2019 • Soichiro Murakami, Makoto Morishita, Tsutomu Hirao, Masaaki Nagata
This paper describes NTT's submission to the WMT19 robustness task.
no code implementations • EMNLP 2018 • Tsutomu Hirao, Hidetaka Kamigaito, Masaaki Nagata
This paper tackles automation of the pyramid method, a reliable manual evaluation framework.
no code implementations • NAACL 2018 • Hidetaka Kamigaito, Katsuhiko Hayashi, Tsutomu Hirao, Masaaki Nagata
To solve this problem, we propose a higher-order syntactic attention network (HiSAN) that can handle higher-order dependency features as an attention distribution on LSTM hidden states.
Ranked #3 on Sentence Compression on Google Dataset
no code implementations • NAACL 2018 • Ukyo Honda, Tsutomu Hirao, Masaaki Nagata
We propose a simple but highly effective automatic evaluation measure of summarization, pruned Basic Elements (pBE).
no code implementations • NAACL 2018 • Shinsaku Sakaue, Tsutomu Hirao, Masaaki Nishino, Masaaki Nagata
This approach is known to have three advantages: its applicability to many useful submodular objective functions, the efficiency of the greedy algorithm, and the provable performance guarantee.
no code implementations • IJCNLP 2017 • Hidetaka Kamigaito, Katsuhiko Hayashi, Tsutomu Hirao, Hiroya Takamura, Manabu Okumura, Masaaki Nagata
The sequence-to-sequence (Seq2Seq) model has been successfully applied to machine translation (MT).
no code implementations • ACL 2017 • Tsutomu Hirao, Masaaki Nishino, Masaaki Nagata
This paper derives an Integer Linear Programming (ILP) formulation to obtain an oracle summary of the compressive summarization paradigm in terms of ROUGE.
no code implementations • EACL 2017 • Tsutomu Hirao, Masaaki Nishino, Jun Suzuki, Masaaki Nagata
To analyze the limitations and the future directions of the extractive summarization paradigm, this paper proposes an Integer Linear Programming (ILP) formulation to obtain extractive oracle summaries in terms of ROUGE-N. We also propose an algorithm that enumerates all of the oracle summaries for a set of reference summaries to exploit F-measures that evaluate which system summaries contain how many sentences that are extracted as an oracle summary.
no code implementations • COLING 2016 • Xun Wang, Masaaki Nishino, Tsutomu Hirao, Katsuhito Sudoh, Masaaki Nagata
Existing methods focus on the extraction of key information, but often neglect coherence.