Search Results for author: Ryohei Sasano

Found 38 papers, 10 papers with code

Leveraging Three Types of Embeddings from Masked Language Models in Idiom Token Classification

no code implementations *SEM (NAACL) 2022 Ryosuke Takahashi, Ryohei Sasano, Koichi Takeda

Recent research has shown that contextualized word embeddings derived from masked language models (MLMs) can give promising results for idiom token classification.

Classification token-classification +2

Cross-lingual Linking of Automatically Constructed Frames and FrameNet

no code implementations LREC 2022 Ryohei Sasano

A semantic frame is a conceptual structure describing an event, relation, or object along with its participants.

Cross-Lingual Word Embeddings Word Embeddings

On Representational Dissociation of Language and Arithmetic in Large Language Models

no code implementations17 Feb 2025 Riku Kisako, Tatsuki Kuribayashi, Ryohei Sasano

The association between language and (non-linguistic) thinking ability in humans has long been debated, and recently, neuroscientific evidence of brain activity patterns has been considered.

Arithmetic Reasoning

CiMaTe: Citation Count Prediction Effectively Leveraging the Main Text

no code implementations6 Oct 2024 Jun Hirako, Ryohei Sasano, Koichi Takeda

Prediction of the future citation counts of papers is increasingly important to find interesting papers among an ever-growing number of papers.

Prediction

Ruri: Japanese General Text Embeddings

1 code implementation12 Sep 2024 Hayato Tsukagoshi, Ryohei Sasano

In this report, we provide a detailed account of the development process of Ruri.

Knowledge Distillation

Simplifying Translations for Children: Iterative Simplification Considering Age of Acquisition with LLMs

1 code implementation8 Aug 2024 Masashi Oshika, Makoto Morishita, Tsutomu Hirao, Ryohei Sasano, Koichi Takeda

In this study, we propose a method that replaces words with high Age of Acquisitions (AoA) in translations with simpler words to match the translations to the user's level.

Machine Translation NMT +2

Are Social Sentiments Inherent in LLMs? An Empirical Study on Extraction of Inter-demographic Sentiments

no code implementations8 Aug 2024 Kunitomo Tanaka, Ryohei Sasano, Koichi Takeda

Large language models (LLMs) are supposed to acquire unconscious human knowledge and feelings, such as social common sense and biases, by training models from large amounts of text.

Common Sense Reasoning Sentiment Analysis

WikiSplit++: Easy Data Refinement for Split and Rephrase

1 code implementation13 Apr 2024 Hayato Tsukagoshi, Tsutomu Hirao, Makoto Morishita, Katsuki Chousa, Ryohei Sasano, Koichi Takeda

The task of Split and Rephrase, which splits a complex sentence into multiple simple sentences with the same meaning, improves readability and enhances the performance of downstream tasks in natural language processing (NLP).

Decoder Sentence +2

Verifying Claims About Metaphors with Large-Scale Automatic Metaphor Identification

no code implementations1 Apr 2024 Kotaro Aono, Ryohei Sasano, Koichi Takeda

There are several linguistic claims about situations where words are more likely to be used as metaphors.

Japanese SimCSE Technical Report

1 code implementation30 Oct 2023 Hayato Tsukagoshi, Ryohei Sasano, Koichi Takeda

We report the development of Japanese SimCSE, Japanese sentence embedding models fine-tuned with SimCSE.

Sentence Sentence Embedding +1

Transformer-based Live Update Generation for Soccer Matches from Microblog Posts

no code implementations25 Oct 2023 Masashi Oshika, Kosuke Yamada, Ryohei Sasano, Koichi Takeda

It has been known to be difficult to generate adequate sports updates from a sequence of vast amounts of diverse live tweets, although the live sports viewing experience with tweets is gaining the popularity.

Language Modeling Language Modelling

Acquiring Frame Element Knowledge with Deep Metric Learning for Semantic Frame Induction

no code implementations23 May 2023 Kosuke Yamada, Ryohei Sasano, Koichi Takeda

The semantic frame induction tasks are defined as a clustering of words into the frames that they evoke, and a clustering of their arguments according to the frame element roles that they should fill.

Clustering Language Modeling +2

Sentence Representations via Gaussian Embedding

1 code implementation22 May 2023 Shohei Yoda, Hayato Tsukagoshi, Ryohei Sasano, Koichi Takeda

Recent progress in sentence embedding, which represents the meaning of a sentence as a point in a vector space, has achieved high performance on tasks such as a semantic textual similarity (STS) task.

Contrastive Learning Natural Language Inference +5

Semantic Frame Induction with Deep Metric Learning

no code implementations27 Apr 2023 Kosuke Yamada, Ryohei Sasano, Koichi Takeda

Recent studies have demonstrated the usefulness of contextualized word embeddings in unsupervised semantic frame induction.

Metric Learning Word Embeddings

Cross-Modal Similarity-Based Curriculum Learning for Image Captioning

no code implementations14 Dec 2022 Hongkuan Zhang, Saku Sugawara, Akiko Aizawa, Lei Zhou, Ryohei Sasano, Koichi Takeda

Moreover, the higher model performance on difficult examples and unseen data also demonstrates the generalization ability.

Image Captioning Language Modeling +1

Transformer-based Lexically Constrained Headline Generation

1 code implementation EMNLP 2021 Kosuke Yamada, Yuta Hitomi, Hideaki Tamori, Ryohei Sasano, Naoaki Okazaki, Kentaro Inui, Koichi Takeda

We also consider a new headline generation strategy that takes advantage of the controllable generation order of Transformer.

Headline Generation

Semantic Frame Induction using Masked Word Embeddings and Two-Step Clustering

no code implementations ACL 2021 Kosuke Yamada, Ryohei Sasano, Koichi Takeda

Recent studies on semantic frame induction show that relatively high performance has been achieved by using clustering-based methods with contextualized word embeddings.

Clustering Vocal Bursts Valence Prediction +1

Verb Sense Clustering using Contextualized Word Representations for Semantic Frame Induction

no code implementations Findings (ACL) 2021 Kosuke Yamada, Ryohei Sasano, Koichi Takeda

Furthermore, we examine the extent to which the contextualized representation of a verb can estimate the number of frames that the verb can evoke.

Clustering

Self-Guided Curriculum Learning for Neural Machine Translation

no code implementations ACL (IWSLT) 2021 Lei Zhou, Liang Ding, Kevin Duh, Shinji Watanabe, Ryohei Sasano, Koichi Takeda

In the field of machine learning, the well-trained model is assumed to be able to recover the training labels, i. e. the synthetic labels predicted by the model should be as close to the ground-truth labels as possible.

Machine Translation NMT +2

DefSent: Sentence Embeddings using Definition Sentences

1 code implementation ACL 2021 Hayato Tsukagoshi, Ryohei Sasano, Koichi Takeda

However, these methods are only available for limited languages due to relying heavily on the large NLI datasets.

Natural Language Inference Sentence +3

Investigating Word-Class Distributions in Word Vector Spaces

no code implementations ACL 2020 Ryohei Sasano, Anna Korhonen

This paper presents an investigation on the distribution of word vectors belonging to a certain word class in a pre-trained word vector space.

Development of a Medical Incident Report Corpus with Intention and Factuality Annotation

no code implementations LREC 2020 Hongkuan Zhang, Ryohei Sasano, Koichi Takeda, Zoie Shui-Yee Wong

In this paper, we present our annotation scheme with respect to the definition of medication entities that we take into account, the method to annotate the relations between entities, and the details of the intention and factuality annotation.

Incorporating Textual Information on User Behavior for Personality Prediction

no code implementations ACL 2019 Kosuke Yamada, Ryohei Sasano, Koichi Takeda

Our experiments on the personality prediction of Twitter users show that the textual information of user behaviors is more useful than the co-occurrence information of the user behaviors.

Prediction

An Empirical Study on Fine-Grained Named Entity Recognition

no code implementations COLING 2018 Khai Mai, Thai-Hoang Pham, Minh Trung Nguyen, Tuan Duc Nguyen, Danushka Bollegala, Ryohei Sasano, Satoshi Sekine

However, there is little research on fine-grained NER (FG-NER), in which hundreds of named entity categories must be recognized, especially for non-English languages.

Chatbot named-entity-recognition +3

Distinguishing Japanese Non-standard Usages from Standard Ones

no code implementations EMNLP 2017 Tatsuya Aoki, Ryohei Sasano, Hiroya Takamura, Manabu Okumura

Our experimental results show that the model leveraging the context embedding outperforms other methods and provide us with findings, for example, on how to construct context embeddings and which corpus to use.

Machine Translation Word Embeddings

Cannot find the paper you are looking for? You can Submit a new open access paper.