Search Results for author: Lihai Nie

Found 5 papers, 1 papers with code

Benchmarking Poisoning Attacks against Retrieval-Augmented Generation

no code implementations24 May 2025 Baolei Zhang, Haoran Xin, Jiatong Li, Dongzhe Zhang, Minghong Fang, Zhuqing Liu, Lihai Nie, Zheli Liu

Our benchmark covers 5 standard question answering (QA) datasets and 10 expanded variants, along with 13 poisoning attack methods and 7 defense mechanisms, representing a broad spectrum of existing techniques.

Benchmarking Question Answering +3

When Safety Detectors Aren't Enough: A Stealthy and Effective Jailbreak Attack on LLMs via Steganographic Techniques

no code implementations22 May 2025 Jianing Geng, Biao Yi, Zekun Fei, Tongxi Wu, Lihai Nie, Zheli Liu

Jailbreak attacks pose a serious threat to large language models (LLMs) by bypassing built-in safety mechanisms and leading to harmful outputs.

Benchmarking

CTRAP: Embedding Collapse Trap to Safeguard Large Language Models from Harmful Fine-Tuning

no code implementations22 May 2025 Biao Yi, Tiansheng Huang, Baolei Zhang, Tong Li, Lihai Nie, Zheli Liu, Li Shen

Fine-tuning-as-a-service, while commercially successful for Large Language Model (LLM) providers, exposes models to harmful fine-tuning attacks.

Language Modeling Language Modelling +2

Practical Poisoning Attacks against Retrieval-Augmented Generation

no code implementations4 Apr 2025 Baolei Zhang, Yuxi Chen, Minghong Fang, Zhuqing Liu, Lihai Nie, Tong Li, Zheli Liu

Large language models (LLMs) have demonstrated impressive natural language processing abilities but face challenges such as hallucination and outdated knowledge.

Hallucination RAG +2

Your Semantic-Independent Watermark is Fragile: A Semantic Perturbation Attack against EaaS Watermark

1 code implementation14 Nov 2024 Zekun Fei, Biao Yi, Jianing Geng, Ruiqi He, Lihai Nie, Zheli Liu

Embedding-as-a-Service (EaaS) has emerged as a successful business pattern but faces significant challenges related to various forms of copyright infringement, particularly, the API misuse and model extraction attacks.

Model extraction

Cannot find the paper you are looking for? You can Submit a new open access paper.