Search Results for author: Jiaqi Huang

Found 1 papers, 0 papers with code

Subtoxic Questions: Dive Into Attitude Change of LLM's Response in Jailbreak Attempts

no code implementations12 Apr 2024 Tianyu Zhang, Zixuan Zhao, Jiaqi Huang, Jingyu Hua, Sheng Zhong

As Large Language Models (LLMs) of Prompt Jailbreaking are getting more and more attention, it is of great significance to raise a generalized research paradigm to evaluate attack strengths and a basic model to conduct subtler experiments.

Cannot find the paper you are looking for? You can Submit a new open access paper.