Search Results for author: Yohei Oseki

Found 23 papers, 10 papers with code

CMCL 2021 Shared Task on Eye-Tracking Prediction

no code implementations NAACL (CMCL) 2021 Nora Hollenstein, Emmanuele Chersoni, Cassandra L. Jacobs, Yohei Oseki, Laurent Prévot, Enrico Santus

The goal of the task is to predict 5 different token- level eye-tracking metrics of the Zurich Cognitive Language Processing Corpus (ZuCo).

Prediction

How LLMs Learn: Tracing Internal Representations with Sparse Autoencoders

1 code implementation9 Mar 2025 Tatsuro Inaba, Kentaro Inui, Yusuke Miyao, Yohei Oseki, Benjamin Heinzerling, Yu Takagi

Large Language Models (LLMs) demonstrate remarkable multilingual capabilities and broad knowledge.

If Attention Serves as a Cognitive Model of Human Memory Retrieval, What is the Plausible Memory Representation?

no code implementations17 Feb 2025 Ryo Yoshida, Shinnosuke Isono, Kohei Kajikawa, Taiga Someya, Yushi Sugimito, Yohei Oseki

Recent work in computational psycholinguistics has revealed intriguing parallels between attention mechanisms and human memory retrieval, focusing primarily on Transformer architectures that operate on token-level representations.

Retrieval Sentence

Developmentally-plausible Working Memory Shapes a Critical Period for Language Acquisition

no code implementations7 Feb 2025 Masato Mita, Ryo Yoshida, Yohei Oseki

Large language models exhibit general linguistic abilities but significantly differ from humans in their efficiency of language acquisition.

Language Acquisition

BabyLM Challenge: Exploring the Effect of Variation Sets on Language Model Training Efficiency

no code implementations14 Nov 2024 Akari Haga, Akiyo Fukatsu, Miyu Oba, Arianna Bisazza, Yohei Oseki

While current large language models have achieved a remarkable success, their data efficiency remains a challenge to overcome.

Language Modeling Language Modelling

Is Structure Dependence Shaped for Efficient Communication?: A Case Study on Coordination

1 code implementation14 Oct 2024 Kohei Kajikawa, Yusuke Kubota, Yohei Oseki

The results demonstrate that the language with the structure-dependent reduction operation is significantly more communicatively efficient than the counterfactual languages.

counterfactual

Can Language Models Induce Grammatical Knowledge from Indirect Evidence?

no code implementations8 Oct 2024 Miyu Oba, Yohei Oseki, Akiyo Fukatsu, Akari Haga, Hiroki Ouchi, Taro Watanabe, Saku Sugawara

What kinds of and how much data is necessary for language models to induce grammatical knowledge to judge sentence acceptability?

Language Acquisition Sentence

LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs

no code implementations4 Jul 2024 LLM-jp, :, Akiko Aizawa, Eiji Aramaki, Bowen Chen, Fei Cheng, Hiroyuki Deguchi, Rintaro Enomoto, Kazuki Fujii, Kensuke Fukumoto, Takuya Fukushima, Namgi Han, Yuto Harada, Chikara Hashimoto, Tatsuya Hiraoka, Shohei Hisada, Sosuke Hosokawa, Lu Jie, Keisuke Kamata, Teruhito Kanazawa, Hiroki Kanezashi, Hiroshi Kataoka, Satoru Katsumata, Daisuke Kawahara, Seiya Kawano, Atsushi Keyaki, Keisuke Kiryu, Hirokazu Kiyomaru, Takashi Kodama, Takahiro Kubo, Yohei Kuga, Ryoma Kumon, Shuhei Kurita, Sadao Kurohashi, Conglong Li, Taiki Maekawa, Hiroshi Matsuda, Yusuke Miyao, Kentaro Mizuki, Sakae Mizuki, Yugo Murawaki, Akim Mousterou, Ryo Nakamura, Taishi Nakamura, Kouta Nakayama, Tomoka Nakazato, Takuro Niitsuma, Jiro Nishitoba, Yusuke Oda, Hayato Ogawa, Takumi Okamoto, Naoaki Okazaki, Yohei Oseki, Shintaro Ozaki, Koki Ryu, Rafal Rzepka, Keisuke Sakaguchi, Shota Sasaki, Satoshi Sekine, Kohei Suda, Saku Sugawara, Issa Sugiura, Hiroaki Sugiyama, Hisami Suzuki, Jun Suzuki, Toyotaro Suzumura, Kensuke Tachibana, Yu Takagi, Kyosuke Takami, Koichi Takeda, Masashi Takeshita, Masahiro Tanaka, Kenjiro Taura, Arseny Tolmachev, Nobuhiro Ueda, Zhen Wan, Shuntaro Yada, Sakiko Yahata, Yuya Yamamoto, Yusuke Yamauchi, Hitomi Yanaka, Rio Yokota, Koichiro Yoshino

This paper introduces LLM-jp, a cross-organizational project for the research and development of Japanese large language models (LLMs).

Tree-Planted Transformers: Unidirectional Transformer Language Models with Implicit Syntactic Supervision

no code implementations20 Feb 2024 Ryo Yoshida, Taiga Someya, Yohei Oseki

Syntactic Language Models (SLMs) can be trained efficiently to reach relatively high performance; however, they have trouble with inference efficiency due to the explicit generation of syntactic structures.

Continual Learning

Emergent Word Order Universals from Cognitively-Motivated Language Models

1 code implementation19 Feb 2024 Tatsuki Kuribayashi, Ryo Ueda, Ryo Yoshida, Yohei Oseki, Ted Briscoe, Timothy Baldwin

The world's languages exhibit certain so-called typological or implicational universals; for example, Subject-Object-Verb (SOV) languages typically use postpositions.

Psychometric Predictive Power of Large Language Models

1 code implementation13 Nov 2023 Tatsuki Kuribayashi, Yohei Oseki, Timothy Baldwin

In other words, pure next-word probability remains a strong predictor for human reading behavior, even in the age of LLMs.

JCoLA: Japanese Corpus of Linguistic Acceptability

2 code implementations22 Sep 2023 Taiga Someya, Yushi Sugimoto, Yohei Oseki

In this paper, we introduce JCoLA (Japanese Corpus of Linguistic Acceptability), which consists of 10, 020 sentences annotated with binary acceptability judgments.

Linguistic Acceptability

Composition, Attention, or Both?

1 code implementation24 Oct 2022 Ryo Yoshida, Yohei Oseki

In this paper, we propose a novel architecture called Composition Attention Grammars (CAGs) that recursively compose subtrees into a single vector representation with a composition function, and selectively attend to previous structural information with a self-attention mechanism.

Context Limitations Make Neural Language Models More Human-Like

1 code implementation23 May 2022 Tatsuki Kuribayashi, Yohei Oseki, Ana Brassard, Kentaro Inui

Language models (LMs) have been used in cognitive modeling as well as engineering studies -- they compute information-theoretic complexity metrics that simulate humans' cognitive load during reading.

Modeling Human Sentence Processing with Left-Corner Recurrent Neural Network Grammars

2 code implementations EMNLP 2021 Ryo Yoshida, Hiroshi Noji, Yohei Oseki

In computational linguistics, it has been shown that hierarchical structures make language models (LMs) more human-like.

Sentence

Lower Perplexity is Not Always Human-Like

1 code implementation ACL 2021 Tatsuki Kuribayashi, Yohei Oseki, Takumi Ito, Ryo Yoshida, Masayuki Asahara, Kentaro Inui

Overall, our results suggest that a cross-lingual evaluation will be necessary to construct human-like computational models.

Language Modeling Language Modelling

Design of BCCWJ-EEG: Balanced Corpus with Human Electroencephalography

no code implementations LREC 2020 Yohei Oseki, Masayuki Asahara

Importantly, this inter-fertilization between NLP, on one hand, and the cognitive (neuro)science of language, on the other, has been driven by the language resources annotated with human language processing data.

EEG

Inverting and Modeling Morphological Inflection

no code implementations WS 2019 Yohei Oseki, Yasutada Sudo, Hiromu Sakai, Alec Marantz

Previous {``}wug{''} tests (Berko, 1958) on Japanese verbal inflection have demonstrated that Japanese speakers, both adults and children, cannot inflect novel present tense forms to {``}correct{''} past tense forms predicted by rules of existent verbs (de Chene, 1982; Vance, 1987, 1991; Klafehn, 2003, 2013), indicating that Japanese verbs are merely stored in the mental lexicon.

Morphological Inflection

Modeling Hierarchical Syntactic Structures in Morphological Processing

no code implementations WS 2019 Yohei Oseki, Charles Yang, Alec Marantz

Sentences are represented as hierarchical syntactic structures, which have been successfully modeled in sentence processing.

Sentence

Cannot find the paper you are looking for? You can Submit a new open access paper.