Search Results for author: Yuqi Ren

Found 8 papers, 3 papers with code

CoDoNMT: Modeling Cohesion Devices for Document-Level Neural Machine Translation

1 code implementation COLING 2022 Yikun Lei, Yuqi Ren, Deyi Xiong

In this paper, we propose a document-level neural machine translation framework, CoDoNMT, which models cohesion devices from two perspectives: Cohesion Device Masking (CoDM) and Cohesion Attention Focusing (CoAF).

Machine Translation NMT +2

LHMKE: A Large-scale Holistic Multi-subject Knowledge Evaluation Benchmark for Chinese Large Language Models

no code implementations19 Mar 2024 Chuang Liu, Renren Jin, Yuqi Ren, Deyi Xiong

Current datasets collect questions from Chinese examinations across different subjects and educational levels to address this issue.

Multiple-choice

Do Large Language Models Mirror Cognitive Language Processing?

no code implementations28 Feb 2024 Yuqi Ren, Renren Jin, Tongxuan Zhang, Deyi Xiong

We employ Representational Similarity Analysis (RSA) to mearsure the alignment between 16 mainstream LLMs and fMRI signals of the brain.

Chatbot Logical Reasoning +1

AI-driven emergence of frequency information non-uniform distribution via THz metasurface spectrum prediction

no code implementations5 Dec 2023 Xiaohua Xing, Yuqi Ren, Die Zou, Qiankun Zhang, Bingxuan Mao, Jianquan Yao, Deyi Xiong, Shuang Zhang, Liang Wu

Recently, artificial intelligence has been extensively deployed across various scientific disciplines, optimizing and guiding the progression of experiments through the integration of abundant datasets, whilst continuously probing the vast theoretical space encapsulated within the data.

Bridging between Cognitive Processing Signals and Linguistic Features via a Unified Attentional Network

no code implementations16 Dec 2021 Yuqi Ren, Deyi Xiong

The proposed framework only requires cognitive processing signals recorded under natural reading as inputs, and can be used to detect a wide range of linguistic features with a single cognitive dataset.

Sentence

CogAlign: Learning to Align Textual Neural Representations to Cognitive Language Processing Signals

1 code implementation ACL 2021 Yuqi Ren, Deyi Xiong

Most previous studies integrate cognitive language processing signals (e. g., eye-tracking or EEG data) into neural models of natural language processing (NLP) just by directly concatenating word embeddings with cognitive features, ignoring the gap between the two modalities (i. e., textual vs. cognitive) and noise in cognitive features.

EEG named-entity-recognition +5

Cannot find the paper you are looking for? You can Submit a new open access paper.