Search Results for author: Zekun Fei

Found 2 papers, 1 papers with code

When Safety Detectors Aren't Enough: A Stealthy and Effective Jailbreak Attack on LLMs via Steganographic Techniques

no code implementations22 May 2025 Jianing Geng, Biao Yi, Zekun Fei, Tongxi Wu, Lihai Nie, Zheli Liu

Jailbreak attacks pose a serious threat to large language models (LLMs) by bypassing built-in safety mechanisms and leading to harmful outputs.

Benchmarking

Your Semantic-Independent Watermark is Fragile: A Semantic Perturbation Attack against EaaS Watermark

1 code implementation14 Nov 2024 Zekun Fei, Biao Yi, Jianing Geng, Ruiqi He, Lihai Nie, Zheli Liu

Embedding-as-a-Service (EaaS) has emerged as a successful business pattern but faces significant challenges related to various forms of copyright infringement, particularly, the API misuse and model extraction attacks.

Model extraction

Cannot find the paper you are looking for? You can Submit a new open access paper.