no code implementations • ACL 2022 • Shumpei Miyawaki, Taku Hasegawa, Kyosuke Nishida, Takuma Kato, Jun Suzuki
We tackle the tasks of image and text retrieval using a dual-encoder model in which images and text are encoded independently.
no code implementations • EMNLP 2021 • Shun Kiyono, Sosuke Kobayashi, Jun Suzuki, Kentaro Inui
Position representation is crucial for building position-aware representations in Transformers.
no code implementations • LANTERN (COLING) 2020 • Diana Galvan-Sosa, Jun Suzuki, Kyosuke Nishida, Koji Matsuda, Kentaro Inui
Despite recent achievements in natural language understanding, reasoning over commonsense knowledge still represents a big challenge to AI systems.
no code implementations • WMT (EMNLP) 2020 • Shun Kiyono, Takumi Ito, Ryuto Konno, Makoto Morishita, Jun Suzuki
In this paper, we describe the submission of Tohoku-AIP-NTT to the WMT’20 news translation task.
no code implementations • EMNLP (IWSLT) 2019 • Hirofumi Inaguma, Shun Kiyono, Nelson Enrique Yalta Soplin, Jun Suzuki, Kevin Duh, Shinji Watanabe
In this year, we mainly build our systems based on Transformer architectures in all tasks and focus on the end-to-end speech translation (E2E-ST).
no code implementations • 19 Nov 2022 • Shiki Sato, Yosuke Kishinami, Hiroaki Sugiyama, Reina Akama, Ryoko Tokuhisa, Jun Suzuki
Automation of dialogue system evaluation is a driving force for the efficient development of dialogue systems.
no code implementations • 28 Oct 2022 • Makoto Morishita, Jun Suzuki, Masaaki Nagata
With the collected parallel data, we can quickly adapt a machine translation model to the target domain.
1 code implementation • COLING 2022 • Yosuke Kishinami, Reina Akama, Shiki Sato, Ryoko Tokuhisa, Jun Suzuki, Kentaro Inui
Prior studies addressing target-oriented conversational tasks lack a crucial notion that has been intensively studied in the context of goal-oriented artificial intelligence agents, namely, planning.
1 code implementation • SIGDIAL (ACL) 2022 • Shiki Sato, Reina Akama, Hiroki Ouchi, Ryoko Tokuhisa, Jun Suzuki, Kentaro Inui
In this scenario, the quality of the n-best list considerably affects the occurrence of contradictions because the final response is chosen from this n-best list.
no code implementations • 16 Jun 2022 • Hiroaki Funayama, Tasuku Sato, Yuichiroh Matsubayashi, Tomoya Mizumoto, Jun Suzuki, Kentaro Inui
Towards guaranteeing high-quality predictions, we present the first study of exploring the use of human-in-the-loop framework for minimizing the grading cost while guaranteeing the grading quality by allowing a SAS model to share the grading task with a human grader.
no code implementations • 1 Jun 2022 • Sho Takase, Shun Kiyono, Sosuke Kobayashi, Jun Suzuki
In the perspective of a layer normalization (LN) position, the architecture of Transformers can be categorized into two types: Post-LN and Pre-LN.
no code implementations • BigScience (ACL) 2022 • Sosuke Kobayashi, Shun Kiyono, Jun Suzuki, Kentaro Inui
Ensembling is a popular method used to improve performance as a last resort.
1 code implementation • 23 May 2022 • Masato Mita, Keisuke Sakaguchi, Masato Hagiwara, Tomoya Mizumoto, Jun Suzuki, Kentaro Inui
Natural language processing technology has rapidly improved automated grammatical error correction tasks, and the community begins to explore document-level revision as one of the next challenges.
no code implementations • LREC 2022 • Makoto Morishita, Katsuki Chousa, Jun Suzuki, Masaaki Nagata
Most current machine translation models are mainly trained with parallel corpora, and their translation accuracy largely depends on the quality and quantity of the corpora.
no code implementations • 28 Sep 2021 • Hiroki Ouchi, Jun Suzuki, Sosuke Kobayashi, Sho Yokoi, Tatsuki Kuribayashi, Masashi Yoshikawa, Kentaro Inui
Interpretable rationales for model predictions are crucial in practical applications.
1 code implementation • 13 Sep 2021 • Shun Kiyono, Sosuke Kobayashi, Jun Suzuki, Kentaro Inui
Position representation is crucial for building position-aware representations in Transformers.
1 code implementation • EACL 2021 • Makoto Morishita, Jun Suzuki, Tomoharu Iwata, Masaaki Nagata
It is crucial to provide an inter-sentence context in Neural Machine Translation (NMT) models for higher-quality translation.
1 code implementation • 3 Feb 2021 • Yasufumi Taniguchi, Hiroki Nakayama, Kubo Takahiro, Jun Suzuki
Text-to-SQL is a crucial task toward developing methods for understanding natural language by computers.
no code implementations • 1 Jan 2021 • Sewon Min, Jordan Boyd-Graber, Chris Alberti, Danqi Chen, Eunsol Choi, Michael Collins, Kelvin Guu, Hannaneh Hajishirzi, Kenton Lee, Jennimaria Palomaki, Colin Raffel, Adam Roberts, Tom Kwiatkowski, Patrick Lewis, Yuxiang Wu, Heinrich Küttler, Linqing Liu, Pasquale Minervini, Pontus Stenetorp, Sebastian Riedel, Sohee Yang, Minjoon Seo, Gautier Izacard, Fabio Petroni, Lucas Hosseini, Nicola De Cao, Edouard Grave, Ikuya Yamada, Sonse Shimaoka, Masatoshi Suzuki, Shumpei Miyawaki, Shun Sato, Ryo Takahashi, Jun Suzuki, Martin Fajcik, Martin Docekal, Karel Ondrej, Pavel Smrz, Hao Cheng, Yelong Shen, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao, Barlas Oguz, Xilun Chen, Vladimir Karpukhin, Stan Peshterliev, Dmytro Okhonko, Michael Schlichtkrull, Sonal Gupta, Yashar Mehdad, Wen-tau Yih
We review the EfficientQA competition from NeurIPS 2020.
no code implementations • 22 Dec 2020 • Jun Suzuki
We then extend a classical result in the design problem, the Kiefer-Wolfowitz theorem, to a qubit system showing the D-optimal design is equivalent to a certain type of the A-optimal design.
Quantum Physics
no code implementations • EMNLP (sustainlp) 2020 • Sosuke Kobayashi, Sho Yokoi, Jun Suzuki, Kentaro Inui
Understanding the influence of a training instance on a neural network model leads to improving interpretability.
1 code implementation • COLING 2020 • Ryo Fujii, Masato Mita, Kaori Abe, Kazuaki Hanawa, Makoto Morishita, Jun Suzuki, Kentaro Inui
Neural Machine Translation (NMT) has shown drastic improvement in its quality when translating clean input, such as text from the news domain.
no code implementations • EMNLP 2020 • Takumi Ito, Tatsuki Kuribayashi, Masatoshi Hidaka, Jun Suzuki, Kentaro Inui
Despite the current diversity and inclusion initiatives in the academic community, researchers with a non-native command of English still face significant obstacles when writing papers in English.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Masato Mita, Shun Kiyono, Masahiro Kaneko, Jun Suzuki, Kentaro Inui
Existing approaches for grammatical error correction (GEC) largely rely on supervised learning with manually created GEC datasets.
no code implementations • ACL 2020 • Hiroaki Funayama, Shota Sasaki, Yuichiroh Matsubayashi, Tomoya Mizumoto, Jun Suzuki, Masato Mita, Kentaro Inui
We introduce a new task formulation of SAS that matches the actual usage.
1 code implementation • ACL 2020 • Takuma Kato, Kaori Abe, Hiroki Ouchi, Shumpei Miyawaki, Jun Suzuki, Kentaro Inui
In general, the labels used in sequence labeling consist of different types of elements.
1 code implementation • ACL 2020 • Masahiro Kaneko, Masato Mita, Shun Kiyono, Jun Suzuki, Kentaro Inui
The answer to this question is not as straightforward as one might expect because the previous common methods for incorporating a MLM into an EncDec model have potential drawbacks when applied to GEC.
Ranked #2 on
Grammatical Error Correction
on JFLEG
no code implementations • ACL 2020 • Ryosuke Kuwabara, Jun Suzuki, Hideki Nakayama
Model ensemble techniques often increase task performance in neural networks; however, they require increased time, memory, and management effort.
1 code implementation • ACL 2020 • Tatsuki Kuribayashi, Takumi Ito, Jun Suzuki, Kentaro Inui
We examine a methodology using neural language models (LMs) for analyzing the word order of language.
1 code implementation • EMNLP 2020 • Sho Yokoi, Ryo Takahashi, Reina Akama, Jun Suzuki, Kentaro Inui
Accordingly, we propose a method that first decouples word vectors into their norm and direction, and then computes alignment-based similarity using earth mover's distance (i. e., optimal transport cost), which we refer to as word rotator's distance.
1 code implementation • ACL 2020 • Shiki Sato, Reina Akama, Hiroki Ouchi, Jun Suzuki, Kentaro Inui
Existing automatic evaluation metrics for open-domain dialogue response generation systems correlate poorly with human evaluation.
1 code implementation • ACL 2020 • Hiroki Ouchi, Jun Suzuki, Sosuke Kobayashi, Sho Yokoi, Tatsuki Kuribayashi, Ryuto Konno, Kentaro Inui
Interpretable rationales for model predictions play a critical role in practical applications.
1 code implementation • EMNLP 2020 • Reina Akama, Sho Yokoi, Jun Suzuki, Kentaro Inui
Large-scale dialogue datasets have recently become available for training neural dialogue agents.
no code implementations • LREC 2020 • Makoto Morishita, Jun Suzuki, Masaaki Nagata
We constructed a parallel corpus for English-Japanese, for which the amount of publicly available parallel corpora is still limited.
no code implementations • WS 2019 • Makoto Morishita, Jun Suzuki, Masaaki Nagata
In this paper, we describe our systems that were submitted to the translation shared tasks at WAT 2019.
no code implementations • IJCNLP 2019 • Hiroki Ouchi, Jun Suzuki, Kentaro Inui
In transductive learning, an unlabeled test set is used for model training.
1 code implementation • WS 2019 • Takumi Ito, Tatsuki Kuribayashi, Hayato Kobayashi, Ana Brassard, Masato Hagiwara, Jun Suzuki, Kentaro Inui
The writing process consists of several stages such as drafting, revising, editing, and proofreading.
1 code implementation • IJCNLP 2019 • Xiaoyu Shen, Jun Suzuki, Kentaro Inui, Hui Su, Dietrich Klakow, Satoshi Sekine
As a result, the content to be described in the text cannot be explicitly controlled.
1 code implementation • IJCNLP 2019 • Masato Hagiwara, Takumi Ito, Tatsuki Kuribayashi, Jun Suzuki, Kentaro Inui
Language technologies play a key role in assisting people with their writing.
1 code implementation • IJCNLP 2019 • Shun Kiyono, Jun Suzuki, Masato Mita, Tomoya Mizumoto, Kentaro Inui
The incorporation of pseudo data in the training of grammatical error correction models has been one of the main factors in improving the performance of such models.
no code implementations • WS 2019 • Hiroki Asano, Masato Mita, Tomoya Mizumoto, Jun Suzuki
We introduce the AIP-Tohoku grammatical error correction (GEC) system for the BEA-2019 shared task in Track 1 (Restricted Track) and Track 2 (Unrestricted Track) using the same system architecture.
1 code implementation • ACL 2019 • Motoki Sato, Jun Suzuki, Shun Kiyono
A regularization technique based on adversarial perturbation, which was initially developed in the field of image processing, has been successfully applied to text classification tasks and has yielded attractive improvements.
no code implementations • ACL 2019 • Tatsuki Kuribayashi, Hiroki Ouchi, Naoya Inoue, Paul Reisert, Toshinori Miyoshi, Jun Suzuki, Kentaro Inui
For several natural language processing (NLP) tasks, span representation design is attracting considerable attention as a promising new technique; a common basis for an effective design has been established.
no code implementations • 13 Jun 2019 • Sho Takase, Jun Suzuki, Masaaki Nagata
This paper proposes a novel Recurrent Neural Network (RNN) language model that takes advantage of character information.
1 code implementation • NAACL 2019 • Shota Sasaki, Jun Suzuki, Kentaro Inui
The idea of subword-based word embeddings has been proposed in the literature, mainly for solving the out-of-vocabulary (OOV) word problem observed in standard word-based word embeddings.
1 code implementation • WS 2019 • Hono Shirai, Naoya Inoue, Jun Suzuki, Kentaro Inui
Specifically, we show how to adapt the targeted sentiment analysis task for pros/cons extraction in computer science papers and conduct an annotation study.
no code implementations • SEMEVAL 2019 • Kazuaki Hanawa, Shota Sasaki, Hiroki Ouchi, Jun Suzuki, Kentaro Inui
Our system achieved 80. 9{\%} accuracy on the test set for the formal run and got the 3rd place out of 42 teams.
no code implementations • WS 2018 • Shun Kiyono, Sho Takase, Jun Suzuki, Naoaki Okazaki, Kentaro Inui, Masaaki Nagata
Developing a method for understanding the inner workings of black-box neural methods is an important research endeavor.
no code implementations • 13 Oct 2018 • Shun Kiyono, Jun Suzuki, Kentaro Inui
We also demonstrate that our method has the more data, better performance property with promising scalability to the amount of unlabeled data.
no code implementations • WS 2018 • Makoto Morishita, Jun Suzuki, Masaaki Nagata
This paper describes NTT{'}s neural machine translation systems submitted to the WMT 2018 English-German and German-English news translation tasks.
no code implementations • EMNLP 2018 • Sho Yokoi, Sosuke Kobayashi, Kenji Fukumizu, Jun Suzuki, Kentaro Inui
As well as deriving PMI from mutual information, we derive this new measure from the Hilbert--Schmidt independence criterion (HSIC); thus, we call the new measure the pointwise HSIC (PHSIC).
1 code implementation • EMNLP 2018 • Sho Takase, Jun Suzuki, Masaaki Nagata
This paper proposes a state-of-the-art recurrent neural network (RNN) language model that combines probability distributions computed not only from a final RNN layer but also from middle layers.
Ranked #8 on
Language Modelling
on Penn Treebank (Word Level)
no code implementations • COLING 2018 • Makoto Morishita, Jun Suzuki, Masaaki Nagata
We hypothesize that in the NMT model, the appropriate subword units for the following three modules (layers) can differ: (1) the encoder embedding layer, (2) the decoder embedding layer, and (3) the decoder output layer.
1 code implementation • ACL 2018 • Jun Suzuki, Sho Takase, Hidetaka Kamigaito, Makoto Morishita, Masaaki Nagata
This paper investigates the construction of a strong baseline based on general purpose sequence-to-sequence models for constituency parsing.
Ranked #16 on
Constituency Parsing
on Penn Treebank
2 code implementations • 8 May 2018 • Motoki Sato, Jun Suzuki, Hiroyuki Shindo, Yuji Matsumoto
This paper restores interpretability to such methods by restricting the directions of perturbations toward the existing words in the input embedding space.
no code implementations • 22 Dec 2017 • Shun Kiyono, Sho Takase, Jun Suzuki, Naoaki Okazaki, Kentaro Inui, Masaaki Nagata
The encoder-decoder model is widely used in natural language generation tasks.
no code implementations • WS 2017 • Makoto Morishita, Jun Suzuki, Masaaki Nagata
In this year, we participated in four translation subtasks at WAT 2017.
no code implementations • IJCNLP 2017 • Itsumi Saito, Jun Suzuki, Kyosuke Nishida, Kugatsu Sadamitsu, Satoshi Kobashikawa, Ryo Masumura, Yuji Matsumoto, Junji Tomita
In this study, we investigated the effectiveness of augmented data for encoder-decoder-based neural normalization models.
1 code implementation • IJCNLP 2017 • Sho Takase, Jun Suzuki, Masaaki Nagata
This paper proposes a reinforcing method that refines the output layers of existing Recurrent Neural Network (RNN) language models.
no code implementations • EACL 2017 • Tsutomu Hirao, Masaaki Nishino, Jun Suzuki, Masaaki Nagata
To analyze the limitations and the future directions of the extractive summarization paradigm, this paper proposes an Integer Linear Programming (ILP) formulation to obtain extractive oracle summaries in terms of ROUGE-N. We also propose an algorithm that enumerates all of the oracle summaries for a set of reference summaries to exploit F-measures that evaluate which system summaries contain how many sentences that are extracted as an oracle summary.
no code implementations • EACL 2017 • Jun Suzuki, Masaaki Nagata
This paper tackles the reduction of redundant repeating generation that is often observed in RNN-based encoder-decoder models.
Ranked #4 on
Text Summarization
on DUC 2004 Task 1