Search Results for author: Lizhi Lin

Found 2 papers, 0 papers with code

Against The Achilles' Heel: A Survey on Red Teaming for Generative Models

no code implementations31 Mar 2024 Lizhi Lin, Honglin Mu, Zenan Zhai, Minghan Wang, Yuxia Wang, Renxi Wang, Junjie Gao, Yixuan Zhang, Wanxiang Che, Timothy Baldwin, Xudong Han, Haonan Li

Generative models are rapidly gaining popularity and being integrated into everyday applications, raising concerns over their safety issues as various vulnerabilities are exposed.

A Chinese Dataset for Evaluating the Safeguards in Large Language Models

no code implementations19 Feb 2024 Yuxia Wang, Zenan Zhai, Haonan Li, Xudong Han, Lizhi Lin, Zhenxuan Zhang, Jingru Zhao, Preslav Nakov, Timothy Baldwin

Previous studies have proposed comprehensive taxonomies of the risks posed by LLMs, as well as corresponding prompts that can be used to examine the safety mechanisms of LLMs.

Cannot find the paper you are looking for? You can Submit a new open access paper.