no code implementations • NAACL (CMCL) 2021 • Nora Hollenstein, Emmanuele Chersoni, Cassandra L. Jacobs, Yohei Oseki, Laurent Prévot, Enrico Santus
The goal of the task is to predict 5 different token- level eye-tracking metrics of the Zurich Cognitive Language Processing Corpus (ZuCo).
no code implementations • CMCL (ACL) 2022 • Nora Hollenstein, Emmanuele Chersoni, Cassandra Jacobs, Yohei Oseki, Laurent Prévot, Enrico Santus
We present the second shared task on eye-tracking data prediction of the Cognitive Modeling and Computational Linguistics Workshop (CMCL).
1 code implementation • 9 Mar 2025 • Tatsuro Inaba, Kentaro Inui, Yusuke Miyao, Yohei Oseki, Benjamin Heinzerling, Yu Takagi
Large Language Models (LLMs) demonstrate remarkable multilingual capabilities and broad knowledge.
no code implementations • 17 Feb 2025 • Ryo Yoshida, Shinnosuke Isono, Kohei Kajikawa, Taiga Someya, Yushi Sugimito, Yohei Oseki
Recent work in computational psycholinguistics has revealed intriguing parallels between attention mechanisms and human memory retrieval, focusing primarily on Transformer architectures that operate on token-level representations.
no code implementations • 7 Feb 2025 • Masato Mita, Ryo Yoshida, Yohei Oseki
Large language models exhibit general linguistic abilities but significantly differ from humans in their efficiency of language acquisition.
no code implementations • 14 Nov 2024 • Akari Haga, Akiyo Fukatsu, Miyu Oba, Arianna Bisazza, Yohei Oseki
While current large language models have achieved a remarkable success, their data efficiency remains a challenge to overcome.
1 code implementation • 14 Oct 2024 • Kohei Kajikawa, Yusuke Kubota, Yohei Oseki
The results demonstrate that the language with the structure-dependent reduction operation is significantly more communicatively efficient than the counterfactual languages.
no code implementations • 8 Oct 2024 • Miyu Oba, Yohei Oseki, Akiyo Fukatsu, Akari Haga, Hiroki Ouchi, Taro Watanabe, Saku Sugawara
What kinds of and how much data is necessary for language models to induce grammatical knowledge to judge sentence acceptability?
no code implementations • 4 Jul 2024 • LLM-jp, :, Akiko Aizawa, Eiji Aramaki, Bowen Chen, Fei Cheng, Hiroyuki Deguchi, Rintaro Enomoto, Kazuki Fujii, Kensuke Fukumoto, Takuya Fukushima, Namgi Han, Yuto Harada, Chikara Hashimoto, Tatsuya Hiraoka, Shohei Hisada, Sosuke Hosokawa, Lu Jie, Keisuke Kamata, Teruhito Kanazawa, Hiroki Kanezashi, Hiroshi Kataoka, Satoru Katsumata, Daisuke Kawahara, Seiya Kawano, Atsushi Keyaki, Keisuke Kiryu, Hirokazu Kiyomaru, Takashi Kodama, Takahiro Kubo, Yohei Kuga, Ryoma Kumon, Shuhei Kurita, Sadao Kurohashi, Conglong Li, Taiki Maekawa, Hiroshi Matsuda, Yusuke Miyao, Kentaro Mizuki, Sakae Mizuki, Yugo Murawaki, Akim Mousterou, Ryo Nakamura, Taishi Nakamura, Kouta Nakayama, Tomoka Nakazato, Takuro Niitsuma, Jiro Nishitoba, Yusuke Oda, Hayato Ogawa, Takumi Okamoto, Naoaki Okazaki, Yohei Oseki, Shintaro Ozaki, Koki Ryu, Rafal Rzepka, Keisuke Sakaguchi, Shota Sasaki, Satoshi Sekine, Kohei Suda, Saku Sugawara, Issa Sugiura, Hiroaki Sugiyama, Hisami Suzuki, Jun Suzuki, Toyotaro Suzumura, Kensuke Tachibana, Yu Takagi, Kyosuke Takami, Koichi Takeda, Masashi Takeshita, Masahiro Tanaka, Kenjiro Taura, Arseny Tolmachev, Nobuhiro Ueda, Zhen Wan, Shuntaro Yada, Sakiko Yahata, Yuya Yamamoto, Yusuke Yamauchi, Hitomi Yanaka, Rio Yokota, Koichiro Yoshino
This paper introduces LLM-jp, a cross-organizational project for the research and development of Japanese large language models (LLMs).
no code implementations • 20 Feb 2024 • Ryo Yoshida, Taiga Someya, Yohei Oseki
Syntactic Language Models (SLMs) can be trained efficiently to reach relatively high performance; however, they have trouble with inference efficiency due to the explicit generation of syntactic structures.
1 code implementation • 19 Feb 2024 • Tatsuki Kuribayashi, Ryo Ueda, Ryo Yoshida, Yohei Oseki, Ted Briscoe, Timothy Baldwin
The world's languages exhibit certain so-called typological or implicational universals; for example, Subject-Object-Verb (SOV) languages typically use postpositions.
1 code implementation • 13 Nov 2023 • Tatsuki Kuribayashi, Yohei Oseki, Timothy Baldwin
In other words, pure next-word probability remains a strong predictor for human reading behavior, even in the age of LLMs.
2 code implementations • 22 Sep 2023 • Taiga Someya, Yushi Sugimoto, Yohei Oseki
In this paper, we introduce JCoLA (Japanese Corpus of Linguistic Acceptability), which consists of 10, 020 sentences annotated with binary acceptability judgments.
1 code implementation • 24 Oct 2022 • Ryo Yoshida, Yohei Oseki
In this paper, we propose a novel architecture called Composition Attention Grammars (CAGs) that recursively compose subtrees into a single vector representation with a composition function, and selectively attend to previous structural information with a self-attention mechanism.
1 code implementation • 23 May 2022 • Tatsuki Kuribayashi, Yohei Oseki, Ana Brassard, Kentaro Inui
Language models (LMs) have been used in cognitive modeling as well as engineering studies -- they compute information-theoretic complexity metrics that simulate humans' cognitive load during reading.
2 code implementations • EMNLP 2021 • Ryo Yoshida, Hiroshi Noji, Yohei Oseki
In computational linguistics, it has been shown that hierarchical structures make language models (LMs) more human-like.
1 code implementation • ACL 2021 • Tatsuki Kuribayashi, Yohei Oseki, Takumi Ito, Ryo Yoshida, Masayuki Asahara, Kentaro Inui
Overall, our results suggest that a cross-lingual evaluation will be necessary to construct human-like computational models.
1 code implementation • Findings (ACL) 2021 • Hiroshi Noji, Yohei Oseki
However, RNNGs are known to be harder to scale due to the difficulty of batched training.
no code implementations • LREC 2020 • Yohei Oseki, Masayuki Asahara
Importantly, this inter-fertilization between NLP, on one hand, and the cognitive (neuro)science of language, on the other, has been driven by the language resources annotated with human language processing data.
no code implementations • WS 2019 • Yohei Oseki, Yasutada Sudo, Hiromu Sakai, Alec Marantz
Previous {``}wug{''} tests (Berko, 1958) on Japanese verbal inflection have demonstrated that Japanese speakers, both adults and children, cannot inflect novel present tense forms to {``}correct{''} past tense forms predicted by rules of existent verbs (de Chene, 1982; Vance, 1987, 1991; Klafehn, 2003, 2013), indicating that Japanese verbs are merely stored in the mental lexicon.
no code implementations • WS 2019 • Yohei Oseki, Charles Yang, Alec Marantz
Sentences are represented as hierarchical syntactic structures, which have been successfully modeled in sentence processing.