Search Results for author: Yequan Wang

Found 37 papers, 13 papers with code

If an LLM Were a Character, Would It Know Its Own Story? Evaluating Lifelong Learning in LLMs

no code implementations30 Mar 2025 Siqi Fan, Xiusheng Huang, Yiqun Yao, Xuezhi Fang, Kang Liu, Peng Han, Shuo Shang, Aixin Sun, Yequan Wang

However, during multi-turn, multi-agent interactions, LLMs begin to exhibit consistent, character-like behaviors, hinting at a form of emergent lifelong learning.

Fact Checking Lifelong learning

SeniorTalk: A Chinese Conversation Dataset with Rich Annotations for Super-Aged Seniors

no code implementations20 Mar 2025 Yang Chen, Hui Wang, Shiyao Wang, Junyang Chen, Jiabei He, Jiaming Zhou, Xi Yang, Yequan Wang, Yonghua Lin, Yong Qin

While voice technologies increasingly serve aging populations, current systems exhibit significant performance gaps due to inadequate training data capturing elderly-specific vocal characteristics like presbyphonia and dialectal variations.

speaker-diarization Speaker Diarization +3

Position-Aware Depth Decay Decoding ($D^3$): Boosting Large Language Model Inference Efficiency

no code implementations11 Mar 2025 Siqi Fan, Xuezhi Fang, Xingrun Xing, Peng Han, Shuo Shang, Yequan Wang

Experiments on large language models (\ie the Llama) with $7 \sim 70$ billion parameters show that $D^3$ can achieve an average 1. 5x speedup compared with the full-inference pipeline while maintaining comparable performance with nearly no performance drop ($<1\%$) on the GSM8K and BBH benchmarks.

GSM8K Language Modeling +4

Few-Shot Learner Generalizes Across AI-Generated Image Detection

no code implementations15 Jan 2025 Shiyu Wu, Jing Liu, Jing Li, Yequan Wang

Current fake image detectors trained on large synthetic image datasets perform satisfactorily on limited studied generative models.

Knowledge Editing with Dynamic Knowledge Graphs for Multi-Hop Question Answering

no code implementations18 Dec 2024 Yifan Lu, Yigeng Zhou, Jing Li, Yequan Wang, Xuebo Liu, Daojing He, Fangming Liu, Min Zhang

Multi-hop question answering (MHQA) poses a significant challenge for large language models (LLMs) due to the extensive knowledge demands involved.

graph construction knowledge editing +4

Commonsense Knowledge Editing Based on Free-Text in LLMs

1 code implementation31 Oct 2024 Xiusheng Huang, Yequan Wang, Jun Zhao, Kang Liu

Knowledge editing technology is crucial for maintaining the accuracy and timeliness of large language models (LLMs) .

knowledge editing

Reasons and Solutions for the Decline in Model Performance after Editing

1 code implementation31 Oct 2024 Xiusheng Huang, Jiaxiang Liu, Yequan Wang, Kang Liu

In order to investigate the reasons for the performance decline of the edited model and optimize the editing method, this work explores the underlying reasons from both data and model perspectives.

knowledge editing

Sketch: A Toolkit for Streamlining LLM Operations

no code implementations5 Sep 2024 Xin Jiang, Xiang Li, Wenjia Ma, Xuezhi Fang, Yiqun Yao, Naitong Yu, Xuying Meng, Peng Han, Jing Li, Aixin Sun, Yequan Wang

Sketch comprises the following components: (1) a suite of task description schemas and prompt templates encompassing various NLP tasks; (2) a user-friendly, interactive process for building structured output LLM services tailored to various NLP tasks; (3) an open-source dataset for output format control, along with tools for dataset construction; and (4) an open-source model based on LLaMA3-8B-Instruct that adeptly comprehends and adheres to output formatting instructions.

Open-domain Implicit Format Control for Large Language Model Generation

1 code implementation8 Aug 2024 Yiqun Yao, Wenjia Ma, Xuezhi Fang, Xin Jiang, Xiang Li, Xuying Meng, Peng Han, Jing Li, Aixin Sun, Yequan Wang

Controlling the format of outputs generated by large language models (LLMs) is a critical functionality in various applications.

Language Modeling Language Modelling +1

Improving Zero-shot LLM Re-Ranker with Risk Minimization

no code implementations19 Jun 2024 Xiaowei Yuan, Zhao Yang, Yequan Wang, Jun Zhao, Kang Liu

In the Retrieval-Augmented Generation (RAG) system, advanced Large Language Models (LLMs) have emerged as effective Query Likelihood Models (QLMs) in an unsupervised way, which re-rank documents based on the probability of generating the query given the content of a document.

RAG Re-Ranking +1

Multimodal Reasoning with Multimodal Knowledge Graph

no code implementations4 Jun 2024 Junlin Lee, Yequan Wang, Jing Li, Min Zhang

Multimodal reasoning with large language models (LLMs) often suffers from hallucinations and the presence of deficient or outdated knowledge within LLMs.

cross-modal alignment Graph Attention +3

Spectral-Based Graph Neural Networks for Complementary Item Recommendation

no code implementations4 Jan 2024 Haitong Luo, Xuying Meng, Suhang Wang, Hanyun Cao, Weiyao Zhang, Yequan Wang, Yujun Zhang

In this study, we present a novel approach called Spectral-based Complementary Graph Neural Networks (SComGNN) that utilizes the spectral properties of complementary item graphs.

Attribute Recommendation Systems

Quantifying and Attributing the Hallucination of Large Language Models via Association Analysis

no code implementations11 Sep 2023 Li Du, Yequan Wang, Xingrun Xing, Yiqun Ya, Xiang Li, Xin Jiang, Xuezhi Fang

Although demonstrating superb performance on various NLP tasks, large language models (LLMs) still suffer from the hallucination problem, which threatens the reliability of LLMs.

Hallucination Instruction Following +2

FLM-101B: An Open LLM and How to Train It with $100K Budget

no code implementations7 Sep 2023 Xiang Li, Yiqun Yao, Xin Jiang, Xuezhi Fang, Xuying Meng, Siqi Fan, Peng Han, Jing Li, Li Du, Bowen Qin, Zheng Zhang, Aixin Sun, Yequan Wang

Large language models (LLMs) are considered important approaches towards foundational machine intelligence, achieving remarkable success in Natural Language Processing and multimodal tasks, among others.

Memorization

Rethinking Document-Level Relation Extraction: A Reality Check

no code implementations15 Jun 2023 Jing Li, Yequan Wang, Shuai Zhang, Min Zhang

Recently, numerous efforts have continued to push up performance boundaries of document-level relation extraction (DocRE) and have claimed significant progress in DocRE.

Document-level Relation Extraction Relation

Masked Structural Growth for 2x Faster Language Model Pre-training

1 code implementation4 May 2023 Yiqun Yao, Zheng Zhang, Jing Li, Yequan Wang

In terms of growth schedule, the impact of each single dimension on a schedule's efficiency is under-explored by existing work.

Language Modeling Language Modelling +2

FreeLM: Fine-Tuning-Free Language Model

no code implementations2 May 2023 Xiang Li, Xin Jiang, Xuying Meng, Aixin Sun, Yequan Wang

FreeLM outperforms large models e. g., GPT-3 and InstructGPT, on a range of language understanding tasks in experiments.

Language Modeling Language Modelling +1

NetGPT: Generative Pretrained Transformer for Network Traffic

1 code implementation19 Apr 2023 Xuying Meng, Chungang Lin, Yequan Wang, Yujun Zhang

Pretrained models for network traffic can utilize large-scale raw data to learn the essential characteristics of network traffic, and generate distinguishable results for input traffic without considering specific downstream tasks.

Scheduling Traffic Classification

DialogPaint: A Dialog-based Image Editing Model

no code implementations17 Mar 2023 Jingxuan Wei, Shiyu Wu, Xin Jiang, Yequan Wang

We introduce DialogPaint, a novel framework that bridges conversational interactions with image editing, enabling users to modify images through natural dialogue.

model Style Transfer

GCRE-GPT: A Generative Model for Comparative Relation Extraction

no code implementations15 Mar 2023 Yequan Wang, Hengran Zhang, Aixin Sun, Xuying Meng

Given comparative text, comparative relation extraction aims to extract two targets (\eg two cameras) in comparison and the aspect they are compared for (\eg image quality).

Relation Relation Extraction

PoKE: Prior Knowledge Enhanced Emotional Support Conversation with Latent Variable

no code implementations23 Oct 2022 Xiaohan Xu, Xuying Meng, Yequan Wang

Further experiments prove that abundant prior knowledge is conducive to high-quality emotional support, and a well-learned latent variable is critical to the diversity of generations.

Diversity

Perplexity from PLM Is Unreliable for Evaluating Text Quality

no code implementations12 Oct 2022 Yequan Wang, Jiawen Deng, Aixin Sun, Xuying Meng

Recently, amounts of works utilize perplexity~(PPL) to evaluate the quality of the generated text.

Common Sense Reasoning

CofeNet: Context and Former-Label Enhanced Net for Complicated Quotation Extraction

1 code implementation COLING 2022 Yequan Wang, Xiang Li, Aixin Sun, Xuying Meng, Huaming Liao, Jiafeng Guo

CofeNet is able to extract complicated quotations with components of variable lengths and complicated structures.

Chat-Capsule: A Hierarchical Capsule for Dialog-level Emotion Analysis

no code implementations23 Mar 2022 Yequan Wang, Xuying Meng, Yiyi Liu, Aixin Sun, Yao Wang, Yinhe Zheng, Minlie Huang

These models hence are not optimized for dialog-level emotion detection, i. e. to predict the emotion category of a dialog as a whole.

Emotion Recognition

Cannot find the paper you are looking for? You can Submit a new open access paper.