Search Results for author: Aiwei Liu

Found 38 papers, 26 papers with code

A Call for Collaborative Intelligence: Why Human-Agent Systems Should Precede AI Autonomy

1 code implementation11 Jun 2025 Henry Peng Zou, Wei-Chieh Huang, Yaozu Wu, Chunyu Miao, Dongyuan Li, Aiwei Liu, Yue Zhou, Yankai Chen, Weizhi Zhang, Yangning Li, Liancheng Fang, Renhe Jiang, Philip S. Yu

This paper argues that progress in AI should not be measured by how independent systems become, but by how well they can work with humans.

SSR: Speculative Parallel Scaling Reasoning in Test-time

no code implementations21 May 2025 Yuanlin Chu, Bo wang, Xiang Liu, Hong Chen, Aiwei Liu, Xuming Hu

Large language models (LLMs) have achieved impressive results on multi-step mathematical reasoning, yet at the cost of high computational overhead.

Diversity Math +1

Mark Your LLM: Detecting the Misuse of Open-Source Large Language Models via Watermarking

no code implementations6 Mar 2025 Yijie Xu, Aiwei Liu, Xuming Hu, Lijie Wen, Hui Xiong

Our experiments reveal that backdoor watermarking could effectively detect IP Violation, while inference-time watermark distillation is applicable in both scenarios but less robust to further fine-tuning and has a more significant impact on LLM performance compared to backdoor watermarking.

Scaling Laws for Many-Shot In-Context Learning with Self-Generated Annotations

no code implementations4 Mar 2025 Zhengyao Gu, Henry Peng Zou, Yankai Chen, Aiwei Liu, Weizhi Zhang, Philip S. Yu

The high cost of obtaining high-quality annotated data for in-context learning (ICL) has motivated the development of methods that use self-generated annotations in place of ground-truth labels.

In-Context Learning

Can LLM Watermarks Robustly Prevent Unauthorized Knowledge Distillation?

1 code implementation17 Feb 2025 Leyi Pan, Aiwei Liu, Shiyu Huang, Yijian Lu, Xuming Hu, Lijie Wen, Irwin King, Philip S. Yu

The radioactive nature of Large Language Model (LLM) watermarking enables the detection of watermarks inherited by student models when trained on the outputs of watermarked teacher models, making it a promising tool for preventing unauthorized knowledge distillation.

Knowledge Distillation Language Modeling +3

Exploring Response Uncertainty in MLLMs: An Empirical Evaluation under Misleading Scenarios

no code implementations5 Nov 2024 Yunkai Dang, Mengxi Gao, Yibo Yan, Xin Zou, Yanggan Gu, Aiwei Liu, Xuming Hu

By calculating the misleading rate, and capturing both correct-to-incorrect and incorrect-to-correct shifts between the two sets of responses, we can effectively metric the model's response uncertainty.

Less is More: Extreme Gradient Boost Rank-1 Adaption for Efficient Finetuning of LLMs

no code implementations25 Oct 2024 Yifei Zhang, Hao Zhu, Aiwei Liu, Han Yu, Piotr Koniusz, Irwin King

This work advances parameter-efficient fine-tuning for LLMs, and offers a promising solution for adapting LLMs to downstream tasks while optimizing performance and efficiency.

Computational Efficiency Ensemble Learning +1

Mitigating Modality Prior-Induced Hallucinations in Multimodal Large Language Models via Deciphering Attention Causality

1 code implementation7 Oct 2024 Guanyu Zhou, Yibo Yan, Xin Zou, Kun Wang, Aiwei Liu, Xuming Hu

These biases arise from the visual encoder and the Large Language Model (LLM) backbone, affecting the attention mechanism responsible for aligning multimodal inputs.

Causal Inference counterfactual +6

Recent Advances of Multimodal Continual Learning: A Comprehensive Survey

1 code implementation7 Oct 2024 Dianzhi Yu, Xinni Zhang, Yankai Chen, Aiwei Liu, Yifei Zhang, Philip S. Yu, Irwin King

Continual learning (CL) aims to empower machine learning models to learn continually from new data, while building upon previously acquired knowledge without forgetting.

Continual Learning Survey

TIS-DPO: Token-level Importance Sampling for Direct Preference Optimization With Estimated Weights

2 code implementations6 Oct 2024 Aiwei Liu, Haoping Bai, Zhiyun Lu, Yanchao Sun, Xiang Kong, Simon Wang, Jiulong Shan, Albin Madappally Jose, Xiaojiang Liu, Lijie Wen, Philip S. Yu, Meng Cao

In this work, we propose that the optimal data for DPO has equal expected rewards for each token in winning and losing responses, as there is no difference in token importance.

Can Watermarked LLMs be Identified by Users via Crafted Prompts?

1 code implementation4 Oct 2024 Aiwei Liu, Sheng Guan, Yiming Liu, Leyi Pan, Yifei Zhang, Liancheng Fang, Lijie Wen, Philip S. Yu, Xuming Hu

Finally, we propose that the key to enhancing the imperceptibility of watermarked LLMs is to increase the randomness of watermark key selection.

Interpretable Contrastive Monte Carlo Tree Search Reasoning

1 code implementation2 Oct 2024 Zitian Gao, Boye Niu, Xuzheng He, Haotian Xu, Hongzhang Liu, Aiwei Liu, Xuming Hu, Lijie Wen

Thus, we conducted extensive ablation studies and quantitative analysis on components of MCTS, revealing the impact of each component on the MCTS reasoning performance of LLMs.

Entropy-Based Decoding for Retrieval-Augmented Large Language Models

1 code implementation25 Jun 2024 Zexuan Qiu, Zijing Ou, Bin Wu, Jingjing Li, Aiwei Liu, Irwin King

Augmenting Large Language Models (LLMs) with retrieved external knowledge has proven effective for improving the factual accuracy of generated responses.

Open-Domain Question Answering Retrieval

Refiner: Restructure Retrieval Content Efficiently to Advance Question-Answering Capabilities

1 code implementation17 Jun 2024 Zhonghao Li, Xuming Hu, Aiwei Liu, Kening Zheng, Sirui Huang, Hui Xiong

Experiments show that a trained $\textit{Refiner}$ (with 7B parameters) exhibits significant gain to downstream LLM in improving answer accuracy, and outperforms other state-of-the-art advanced RAG and concurrent compressing approaches in various single-hop and multi-hop QA tasks.

Question Answering RAG +2

MarkLLM: An Open-Source Toolkit for LLM Watermarking

1 code implementation16 May 2024 Leyi Pan, Aiwei Liu, Zhiwei He, Zitian Gao, Xuandong Zhao, Yijian Lu, Binglin Zhou, Shuliang Liu, Xuming Hu, Lijie Wen, Irwin King, Philip S. Yu

However, the abundance of LLM watermarking algorithms, their intricate mechanisms, and the complex evaluation procedures and perspectives pose challenges for researchers and the community to easily experiment with, understand, and assess the latest advancements.

An Entropy-based Text Watermarking Detection Method

2 code implementations20 Mar 2024 Yijian Lu, Aiwei Liu, Dianzhi Yu, Jingjing Li, Irwin King

From the experiments, we demonstrate that our EWD can achieve better detection performance in low-entropy scenarios, and our method is also general and can be applied to texts with different entropy distributions.

ChatCite: LLM Agent with Human Workflow Guidance for Comparative Literature Summary

no code implementations5 Mar 2024 Yutong Li, Lu Chen, Aiwei Liu, Kai Yu, Lijie Wen

In this work, we firstly focus on the independent literature summarization step and introduce ChatCite, an LLM agent with human workflow guidance for comparative literature summary.

Retrieval

Can Watermarks Survive Translation? On the Cross-lingual Consistency of Text Watermark for Large Language Models

1 code implementation21 Feb 2024 Zhiwei He, Binglin Zhou, Hongkun Hao, Aiwei Liu, Xing Wang, Zhaopeng Tu, Zhuosheng Zhang, Rui Wang

Furthermore, we analyze two key factors that contribute to the cross-lingual consistency in text watermarking and propose X-SIR as a defense method against CWRA.

TAG

Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation

1 code implementation19 Feb 2024 Aiwei Liu, Haoping Bai, Zhiyun Lu, Xiang Kong, Simon Wang, Jiulong Shan, Meng Cao, Lijie Wen

In this paper, we propose a method to evaluate the response preference by using the output probabilities of response pairs under contrastive prompt pairs, which could achieve better performance on LLaMA2-7B and LLaMA2-13B compared to RLAIF.

Language Modeling Language Modelling +1

A Survey of Text Watermarking in the Era of Large Language Models

no code implementations13 Dec 2023 Aiwei Liu, Leyi Pan, Yijian Lu, Jingjing Li, Xuming Hu, Xi Zhang, Lijie Wen, Irwin King, Hui Xiong, Philip S. Yu

This paper conducts a comprehensive survey of the current state of text watermarking technology, covering four main aspects: (1) an overview and comparison of different text watermarking techniques; (2) evaluation methods for text watermarking algorithms, including their detectability, impact on text or LLM quality, robustness under target or untargeted attacks; (3) potential application scenarios for text watermarking technology; (4) current challenges and future directions for text watermarking.

Dialogue Generation Survey

Prompt Me Up: Unleashing the Power of Alignments for Multimodal Entity and Relation Extraction

1 code implementation25 Oct 2023 Xuming Hu, Junzhe Chen, Aiwei Liu, Shiao Meng, Lijie Wen, Philip S. Yu

Additionally, our method is orthogonal to previous multimodal fusions, and using it on prior SOTA fusions further improves 5. 47% F1.

Relation Relation Extraction

RAPL: A Relation-Aware Prototype Learning Approach for Few-Shot Document-Level Relation Extraction

1 code implementation24 Oct 2023 Shiao Meng, Xuming Hu, Aiwei Liu, Shu'ang Li, Fukun Ma, Yawen Yang, Lijie Wen

However, existing works often struggle to obtain class prototypes with accurate relational semantics: 1) To build prototype for a target relation type, they aggregate the representations of all entity pairs holding that relation, while these entity pairs may also hold other relations, thus disturbing the prototype.

Document-level Relation Extraction Meta-Learning +1

A Semantic Invariant Robust Watermark for Large Language Models

2 code implementations10 Oct 2023 Aiwei Liu, Leyi Pan, Xuming Hu, Shiao Meng, Lijie Wen

In this work, we propose a semantic invariant watermarking method for LLMs that provides both attack robustness and security robustness.

An Unforgeable Publicly Verifiable Watermark for Large Language Models

3 code implementations30 Jul 2023 Aiwei Liu, Leyi Pan, Xuming Hu, Shu'ang Li, Lijie Wen, Irwin King, Philip S. Yu

Experiments demonstrate that our algorithm attains high detection accuracy and computational efficiency through neural networks.

Computational Efficiency

Exploring the Compositional Generalization in Context Dependent Text-to-SQL Parsing

no code implementations29 May 2023 Aiwei Liu, Wei Liu, Xuming Hu, Shuang Li, Fukun Ma, Yawen Yang, Lijie Wen

Based on these observations, we propose a method named \texttt{p-align} to improve the compositional generalization of Text-to-SQL models.

SQL Parsing Text to SQL +1

GDA: Generative Data Augmentation Techniques for Relation Extraction Tasks

no code implementations26 May 2023 Xuming Hu, Aiwei Liu, Zeqi Tan, Xin Zhang, Chenwei Zhang, Irwin King, Philip S. Yu

These techniques neither preserve the semantic consistency of the original sentences when rule-based augmentations are adopted, nor preserve the syntax structure of sentences when expressing relations using seq2seq models, resulting in less diverse augmentations.

Data Augmentation Relation +1

Entity-to-Text based Data Augmentation for various Named Entity Recognition Tasks

no code implementations19 Oct 2022 Xuming Hu, Yong Jiang, Aiwei Liu, Zhongqiang Huang, Pengjun Xie, Fei Huang, Lijie Wen, Philip S. Yu

Data augmentation techniques have been used to alleviate the problem of scarce labeled data in various NER tasks (flat, nested, and discontinuous NER tasks).

Data Augmentation Diversity +4

Semantic Enhanced Text-to-SQL Parsing via Iteratively Learning Schema Linking Graph

1 code implementation8 Aug 2022 Aiwei Liu, Xuming Hu, Li Lin, Lijie Wen

First, we extract a schema linking graph from PLMs through a probing procedure in an unsupervised manner.

Graph Learning SQL Parsing +2

A Multi-level Supervised Contrastive Learning Framework for Low-Resource Natural Language Inference

no code implementations31 May 2022 Shu'ang Li, Xuming Hu, Li Lin, Aiwei Liu, Lijie Wen, Philip S. Yu

Natural Language Inference (NLI) is a growingly essential task in natural language understanding, which requires inferring the relationship between the sentence pairs (premise and hypothesis).

Contrastive Learning Data Augmentation +5

Cannot find the paper you are looking for? You can Submit a new open access paper.