Search Results for author: Takeshi Kojima

Found 5 papers, 4 papers with code

On the Multilingual Ability of Decoder-based Pre-trained Language Models: Finding and Controlling Language-Specific Neurons

1 code implementation3 Apr 2024 Takeshi Kojima, Itsuki Okimura, Yusuke Iwasawa, Hitomi Yanaka, Yutaka Matsuo

Additionally, we tamper with less than 1% of the total neurons in each model during inference and demonstrate that tampering with a few language-specific neurons drastically changes the probability of target language occurrence in text generation.

Text Generation

Unnatural Error Correction: GPT-4 Can Almost Perfectly Handle Unnatural Scrambled Text

1 code implementation30 Nov 2023 Qi Cao, Takeshi Kojima, Yutaka Matsuo, Yusuke Iwasawa

While Large Language Models (LLMs) have achieved remarkable performance in many tasks, much about their inner workings remains unclear.

Large Language Models are Zero-Shot Reasoners

2 code implementations24 May 2022 Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, Yusuke Iwasawa

Pretrained large language models (LLMs) are widely used in many sub-fields of natural language processing (NLP) and generally known as excellent few-shot learners with task-specific exemplars.

Arithmetic Reasoning Common Sense Reasoning +4

Cannot find the paper you are looking for? You can Submit a new open access paper.