Search Results for author: Feiyu Xiong

Found 51 papers, 37 papers with code

Adaptive Preconditioners Trigger Loss Spikes in Adam

no code implementations5 Jun 2025 Zhiwei Bai, Zhangchen Zhou, Jiajie Zhao, Xiaolong Li, Zhiyu Li, Feiyu Xiong, Hongkang Yang, Yaoyu Zhang, Zhi-Qin John Xu

Loss spikes emerge commonly during training across neural networks of varying architectures and scales when using the Adam optimizer.

Attribute

Adversarial Preference Learning for Robust LLM Alignment

no code implementations30 May 2025 Yuanfu Wang, Pengyu Wang, Chenyang Xi, Bo Tang, Junyi Zhu, Wenqiang Wei, Chen Chen, Chao Yang, Jingfeng Zhang, Chaochao Lu, Yijun Niu, Keming Mao, Zhiyu Li, Feiyu Xiong, Jie Hu, MingChuan Yang

However, they remain vulnerable to adversarial attacks due to three key limitations: (1) the inefficiency and high cost of human annotation, (2) the vast diversity of potential adversarial attacks, and (3) the risk of feedback bias and reward hacking.

Jigsaw-R1: A Study of Rule-based Visual Reinforcement Learning with Jigsaw Puzzles

1 code implementation29 May 2025 Zifu Wang, Junyi Zhu, Bo Tang, Zhiyu Li, Feiyu Xiong, Jiaqian Yu, Matthew B. Blaschko

Although these observations are based on jigsaw puzzles and may vary across other visual tasks, this research contributes a valuable piece of jigsaw to the larger puzzle of collective understanding rule-based visual RL and its potential in multimodal learning.

Reinforcement Learning (RL)

Scalable Complexity Control Facilitates Reasoning Ability of LLMs

no code implementations29 May 2025 Liangkai Hang, Junjie Yao, Zhiwei Bai, Tianyi Chen, Yang Chen, Rongjie Diao, Hezhou Li, Pengxiao Lin, Zhiwei Wang, Cheng Xu, Zhongwang Zhang, Zhangchen Zhou, Zhiyu Li, Zehao Lin, Kai Chen, Feiyu Xiong, Yaoyu Zhang, Weinan E, Hongkang Yang, Zhi-Qin John Xu

The reasoning ability of large language models (LLMs) has been rapidly advancing in recent years, attracting interest in more fundamental approaches that can reliably enhance their generalizability.

Token-level Accept or Reject: A Micro Alignment Approach for Large Language Models

1 code implementation26 May 2025 Yang Zhang, Yu Yu, Bo Tang, Yu Zhu, Chuxiong Sun, Wenqiang Wei, Jie Hu, Zipeng Xie, Zhiyu Li, Feiyu Xiong, Edward Chung

With the rapid development of Large Language Models (LLMs), aligning these models with human preferences and values is critical to ensuring ethical and safe applications.

Binary Classification Sentence

CodeFlowBench: A Multi-turn, Iterative Benchmark for Complex Code Generation

1 code implementation30 Apr 2025 Sizhe Wang, Zhengren Wang, Dongsheng Ma, Yongan Yu, Rui Ling, Zhiyu Li, Feiyu Xiong, Wentao Zhang

Modern software development demands code that is maintainable, testable, and scalable by organizing the implementation into modular components with iterative reuse of existing codes.

Code Generation

xVerify: Efficient Answer Verifier for Reasoning Model Evaluations

1 code implementation14 Apr 2025 Ding Chen, Qingchen Yu, Pengyuan Wang, Wentao Zhang, Bo Tang, Feiyu Xiong, Xinchi Li, Minchuan Yang, Zhiyu Li

To train and evaluate xVerify, we construct the VAR dataset by collecting question-answer pairs generated by multiple LLMs across various datasets, leveraging multiple reasoning models and challenging evaluation sets designed specifically for reasoning model assessment.

MaintainCoder: Maintainable Code Generation Under Dynamic Requirements

1 code implementation31 Mar 2025 Zhengren Wang, Rui Ling, Chufan Wang, Yongan Yu, Sizhe Wang, Zhiyu Li, Feiyu Xiong, Wentao Zhang

Modern code generation has made significant strides in functional correctness and execution efficiency.

Code Generation

RARE: Retrieval-Augmented Reasoning Modeling

1 code implementation30 Mar 2025 Zhengren Wang, Jiayang Yu, Dongsheng Ma, Zhe Chen, Yu Wang, Zhiyu Li, Feiyu Xiong, Yanfeng Wang, Weinan E, Linpeng Tang, Wentao Zhang

Domain-specific intelligence demands specialized knowledge and sophisticated reasoning for problem-solving, posing significant challenges for large language models (LLMs) that struggle with knowledge hallucination and inadequate reasoning capabilities under constrained parameter budgets.

Hallucination Memorization +1

MoC: Mixtures of Text Chunking Learners for Retrieval-Augmented Generation System

1 code implementation12 Mar 2025 Jihao Zhao, Zhiyuan Ji, Zhaoxin Fan, Hanyu Wang, Simin Niu, Bo Tang, Feiyu Xiong, Zhiyu Li

Retrieval-Augmented Generation (RAG), while serving as a viable complement to large language models (LLMs), often overlooks the crucial aspect of text chunking within its pipeline.

Chunking Computational Efficiency +3

SEAP: Training-free Sparse Expert Activation Pruning Unlock the Brainpower of Large Language Models

1 code implementation10 Mar 2025 Xun Liang, Hanyu Wang, Huayi Lai, Simin Niu, Shichao Song, Jiawei Yang, Jihao Zhao, Feiyu Xiong, Bo Tang, Zhiyu Li

Notably, at 50% pruning, SEAP surpasses both WandA and FLAP by over 20%, and at 20% pruning, it incurs only a 2. 2% performance drop compared to the dense model.

Computational Efficiency

SurveyX: Academic Survey Automation via Large Language Models

1 code implementation20 Feb 2025 Xun Liang, Jiawei Yang, Yezhaohui Wang, Chen Tang, Zifan Zheng, Simin Niu, Shichao Song, Hanyu Wang, Bo Tang, Feiyu Xiong, Keming Mao, Zhiyu Li

Large Language Models (LLMs) have demonstrated exceptional comprehension capabilities and a vast knowledge base, suggesting that LLMs can serve as efficient tools for automated survey generation.

Survey

HopRAG: Multi-Hop Reasoning for Logic-Aware Retrieval-Augmented Generation

no code implementations18 Feb 2025 Hao liu, Zhengren Wang, Xi Chen, Zhiyu Li, Feiyu Xiong, Qinhan Yu, Wentao Zhang

Retrieval-Augmented Generation (RAG) systems often struggle with imperfect retrieval, as traditional retrievers focus on lexical or semantic similarity rather than logical relevance.

Logical Reasoning RAG +4

SafeRAG: Benchmarking Security in Retrieval-Augmented Generation of Large Language Model

1 code implementation28 Jan 2025 Xun Liang, Simin Niu, Zhiyu Li, Sensen Zhang, Hanyu Wang, Feiyu Xiong, Jason Zhaoxin Fan, Bo Tang, Shichao Song, Mengwei Wang, Jiawei Yang

However, the incorporation of external and unverified knowledge increases the vulnerability of LLMs because attackers can perform attack tasks by manipulating knowledge.

Benchmarking Language Modeling +4

GRAPHMOE: Amplifying Cognitive Depth of Mixture-of-Experts Network via Introducing Self-Rethinking Mechanism

no code implementations14 Jan 2025 Chen Tang, Bo Lv, Zifan Zheng, Bohao Yang, Kun Zhao, Ning Liao, Xiaoxing Wang, Feiyu Xiong, Zhiyu Li, Nayu Liu, Jingchi Jiang

Additionally, this study explores a novel recurrent routing strategy that may inspire further advancements in enhancing the reasoning capabilities of language models.

Mixture-of-Experts

Meta-Chunking: Learning Text Segmentation and Semantic Completion via Logical Perception

1 code implementation16 Oct 2024 Jihao Zhao, Zhiyuan Ji, Yuchen Feng, Pengnian Qi, Simin Niu, Bo Tang, Feiyu Xiong, Zhiyu Li

While Retrieval-Augmented Generation (RAG) has emerged as a promising paradigm for boosting large language models (LLMs) in knowledge-intensive tasks, it often overlooks the crucial aspect of text chunking within its workflow.

Binary Classification Chunking +6

FTII-Bench: A Comprehensive Multimodal Benchmark for Flow Text with Image Insertion

1 code implementation16 Oct 2024 Jiacheng Ruan, Yebin Yang, Zehao Lin, Yuchen Feng, Feiyu Xiong, Zeyun Tang, Zhiyu Li

Based on this, we introduce the Flow Text with Image Insertion Benchmark (FTII-Bench), which includes 318 high-quality Chinese image-text news articles and 307 high-quality English image-text news articles, covering 10 different news domains.

Articles Image Comprehension

MM-CamObj: A Comprehensive Multimodal Dataset for Camouflaged Object Scenarios

1 code implementation24 Sep 2024 Jiacheng Ruan, Wenzhen Yuan, Zehao Lin, Ning Liao, Zhiyu Li, Feiyu Xiong, Ting Liu, Yuzhuo Fu

CamObj-Instruct is collected for fine-tuning the LVLMs with improved instruction-following capabilities, and it includes 11, 363 images and 68, 849 conversations with diverse instructions.

Instruction Following

Attention Heads of Large Language Models: A Survey

1 code implementation5 Sep 2024 Zifan Zheng, Yezhaohui Wang, Yuxin Huang, Shichao Song, MingChuan Yang, Bo Tang, Feiyu Xiong, Zhiyu Li

Since the advent of ChatGPT, Large Language Models (LLMs) have excelled in various tasks but remain as black-box systems.

Survey

Controllable Text Generation for Large Language Models: A Survey

1 code implementation22 Aug 2024 Xun Liang, Hanyu Wang, Yezhaohui Wang, Shichao Song, Jiawei Yang, Simin Niu, Jie Hu, Dan Liu, Shunyu Yao, Feiyu Xiong, Zhiyu Li

This paper systematically reviews the latest advancements in CTG for LLMs, offering a comprehensive definition of its core concepts and clarifying the requirements for control conditions and text quality.

Attribute Prompt Engineering +2

Xinyu: An Efficient LLM-based System for Commentary Generation

no code implementations21 Aug 2024 Yiquan Wu, Bo Tang, Chenyang Xi, Yu Yu, Pengyu Wang, Yifei Liu, Kun Kuang, Haiying Deng, Zhiyu Li, Feiyu Xiong, Jie Hu, Peng Cheng, Zhonghao Wang, Yi Wang, Yi Luo, MingChuan Yang

To address the advanced requirements, we present an argument ranking model for arguments and establish a comprehensive evidence database that includes up-to-date events and classic books, thereby strengthening the substantiation of the evidence with retrieval augmented generation (RAG) technology.

RAG Retrieval-augmented Generation +1

Internal Consistency and Self-Feedback in Large Language Models: A Survey

1 code implementation19 Jul 2024 Xun Liang, Shichao Song, Zifan Zheng, Hanyu Wang, Qingchen Yu, Xunkai Li, Rong-Hua Li, Yi Wang, Zhonghao Wang, Feiyu Xiong, Zhiyu Li

In this paper, we use a unified perspective of internal consistency, offering explanations for reasoning deficiencies and hallucinations.

$\text{Memory}^3$: Language Modeling with Explicit Memory

no code implementations1 Jul 2024 Hongkang Yang, Zehao Lin, Wenjin Wang, Hao Wu, Zhiyu Li, Bo Tang, Wenqiang Wei, Jinbo Wang, Zeyun Tang, Shichao Song, Chenyang Xi, Yu Yu, Kai Chen, Feiyu Xiong, Linpeng Tang, Weinan E

The model is named $\text{Memory}^3$, since explicit memory is the third form of memory in LLMs after implicit memory (model parameters) and working memory (context key-values).

Language Modeling Language Modelling +3

FastMem: Fast Memorization of Prompt Improves Context Awareness of Large Language Models

1 code implementation23 Jun 2024 Junyi Zhu, Shuochen Liu, Yu Yu, Bo Tang, Yibo Yan, Zhiyu Li, Feiyu Xiong, Tong Xu, Matthew B. Blaschko

Large language models (LLMs) excel in generating coherent text, but they often struggle with context awareness, leading to inaccuracies in tasks requiring faithful adherence to provided information.

Memorization Reading Comprehension +1

Improving Generalization and Convergence by Enhancing Implicit Regularization

1 code implementation31 May 2024 Mingze Wang, Jinbo Wang, Haotian He, Zilin Wang, Guanhua Huang, Feiyu Xiong, Zhiyu Li, Weinan E, Lei Wu

In this work, we propose an Implicit Regularization Enhancement (IRE) framework to accelerate the discovery of flat solutions in deep learning, thereby improving generalization and convergence.

image-classification Image Classification

Empowering Large Language Models to Set up a Knowledge Retrieval Indexer via Self-Learning

1 code implementation27 May 2024 Xun Liang, Simin Niu, Zhiyu Li, Sensen Zhang, Shichao Song, Hanyu Wang, Jiawei Yang, Feiyu Xiong, Bo Tang, Chenyang Xi

Retrieval-Augmented Generation (RAG) offers a cost-effective approach to injecting real-time knowledge into large language models (LLMs).

Question Answering RAG +3

xFinder: Robust and Pinpoint Answer Extraction for Large Language Models

1 code implementation20 May 2024 Qingchen Yu, Zifan Zheng, Shichao Song, Zhiyu Li, Feiyu Xiong, Bo Tang, Ding Chen

The continuous advancement of large language models (LLMs) has brought increasing attention to the critical issue of developing fair and reliable methods for evaluating their performance.

NewsBench: A Systematic Evaluation Framework for Assessing Editorial Capabilities of Large Language Models in Chinese Journalism

1 code implementation29 Feb 2024 Miao Li, Ming-Bin Chen, Bo Tang, Shengbin Hou, Pengyu Wang, Haiying Deng, Zhiyu Li, Feiyu Xiong, Keming Mao, Peng Cheng, Yi Luo

We present NewsBench, a novel evaluation framework to systematically assess the capabilities of Large Language Models (LLMs) for editorial capabilities in Chinese journalism.

Ethics Multiple-choice

Controlled Text Generation for Large Language Model with Dynamic Attribute Graphs

1 code implementation17 Feb 2024 Xun Liang, Hanyu Wang, Shichao Song, Mengting Hu, Xunzhi Wang, Zhiyu Li, Feiyu Xiong, Bo Tang

In this study, we introduce a pluggable CTG framework for Large Language Models (LLMs) named Dynamic Attribute Graphs-based controlled text generation (DATG).

Attribute Language Modeling +3

Grimoire is All You Need for Enhancing Large Language Models

1 code implementation7 Jan 2024 Ding Chen, Shichao Song, Qingchen Yu, Zhiyu Li, Wenjin Wang, Feiyu Xiong, Bo Tang

In this paper, we propose a method SLEICL that involves learning from examples using strong language models and then summarizing and transferring these learned skills to weak language models for inference and application.

All In-Context Learning

UHGEval: Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation

1 code implementation26 Nov 2023 Xun Liang, Shichao Song, Simin Niu, Zhiyu Li, Feiyu Xiong, Bo Tang, Yezhaohui Wang, Dawei He, Peng Cheng, Zhonghao Wang, Haiying Deng

These techniques encompass the use of directed hallucination induction and strategies that deliberately alter authentic text to produce hallucinations.

Benchmarking Hallucination +2

CodeKGC: Code Language Model for Generative Knowledge Graph Construction

2 code implementations18 Apr 2023 Zhen Bi, Jing Chen, Yinuo Jiang, Feiyu Xiong, Wei Guo, Huajun Chen, Ningyu Zhang

However, large generative language model trained on structured data such as code has demonstrated impressive capability in understanding natural language for structural prediction and reasoning tasks.

Code Completion graph construction +2

BCRLSP: An Offline Reinforcement Learning Framework for Sequential Targeted Promotion

no code implementations16 Jul 2022 Fanglin Chen, Xiao Liu, Bo Tang, Feiyu Xiong, Serim Hwang, Guomian Zhuang

During deployment, we combine the offline RL model with the LP model to generate a robust policy under the budget constraints.

Offline RL reinforcement-learning +2

Disentangled Ontology Embedding for Zero-shot Learning

1 code implementation8 Jun 2022 Yuxia Geng, Jiaoyan Chen, Wen Zhang, Yajing Xu, Zhuo Chen, Jeff Z. Pan, Yufeng Huang, Feiyu Xiong, Huajun Chen

In this paper, we focus on ontologies for augmenting ZSL, and propose to learn disentangled ontology embeddings guided by ontology properties to capture and utilize more fine-grained class relationships in different aspects.

image-classification Image Classification +3

Bridging the Gap between Reality and Ideality of Entity Matching: A Revisiting and Benchmark Re-Construction

no code implementations12 May 2022 Tianshu Wang, Hongyu Lin, Cheng Fu, Xianpei Han, Le Sun, Feiyu Xiong, Hui Chen, Minlong Lu, Xiuwen Zhu

Experimental results demonstrate that the assumptions made in the previous benchmark construction process are not coincidental with the open environment, which conceal the main challenges of the task and therefore significantly overestimate the current progress of entity matching.

Entity Resolution

Ontology-enhanced Prompt-tuning for Few-shot Learning

no code implementations27 Jan 2022 Hongbin Ye, Ningyu Zhang, Shumin Deng, Xiang Chen, Hui Chen, Feiyu Xiong, Xi Chen, Huajun Chen

Specifically, we develop the ontology transformation based on the external knowledge graph to address the knowledge missing issue, which fulfills and converts structure knowledge to text.

Event Extraction Few-Shot Learning +1

SQUIRE: A Sequence-to-sequence Framework for Multi-hop Knowledge Graph Reasoning

1 code implementation17 Jan 2022 Yushi Bai, Xin Lv, Juanzi Li, Lei Hou, Yincen Qu, Zelin Dai, Feiyu Xiong

Multi-hop knowledge graph (KG) reasoning has been widely studied in recent years to provide interpretable predictions on missing links with evidential paths.

Decoder Navigate +1

Knowledge Graph Embedding in E-commerce Applications: Attentive Reasoning, Explanations, and Transferable Rules

no code implementations16 Dec 2021 Wen Zhang, Shumin Deng, Mingyang Chen, Liang Wang, Qiang Chen, Feiyu Xiong, Xiangwen Liu, Huajun Chen

We first identity three important desiderata for e-commerce KG systems: 1) attentive reasoning, reasoning over a few target relations of more concerns instead of all; 2) explanation, providing explanations for a prediction to help both users and business operators understand why the prediction is made; 3) transferable rules, generating reusable rules to accelerate the deployment of a KG to new systems.

Entity Embeddings Graph Attention +5

Interpretable performance analysis towards offline reinforcement learning: A dataset perspective

no code implementations12 May 2021 Chenyang Xi, Bo Tang, Jiajun Shen, Xinfu Liu, Feiyu Xiong, Xueying Li

We make it open-source for fair and comprehensive competitions between offline RL algorithms with complete datasets and checkpoints being provided.

Offline RL Q-Learning +3

Studying Product Competition Using Representation Learning

no code implementations21 May 2020 Fanglin Chen, Xiao Liu, Davide Proserpio, Isamar Troncoso, Feiyu Xiong

We show that, compared with state-of-the-art models, our approach is faster, and can produce more accurate demand forecasts and price elasticities.

Causal Inference Decision Making +1

Cannot find the paper you are looking for? You can Submit a new open access paper.