Search Results for author: Yebo Cao

Found 1 papers, 0 papers with code

Hidden You Malicious Goal Into Benign Narratives: Jailbreak Large Language Models through Logic Chain Injection

no code implementations7 Apr 2024 Zhilong Wang, Yebo Cao, Peng Liu

This paper proposes a new type of jailbreak attacks which can deceive both the LLMs and human (i. e., security analyst).

Language Modelling

Cannot find the paper you are looking for? You can Submit a new open access paper.