1 code implementation • 27 Mar 2024 • Keyan Guo, Ayush Utkarsh, Wenbo Ding, Isabelle Ondracek, Ziming Zhao, Guo Freeman, Nishant Vishwamitra, Hongxin Hu
Online user-generated content games (UGCGs) are increasingly popular among children and adolescents for social interaction and more creative online entertainment.
no code implementations • 7 Jan 2024 • Keyan Guo, Alexander Hu, Jaden Mu, Ziheng Shi, Ziming Zhao, Nishant Vishwamitra, Hongxin Hu
Our study reveals that a meticulously crafted reasoning prompt can effectively capture the context of hate speech by fully utilizing the knowledge base in LLMs, significantly outperforming existing techniques.
1 code implementation • 22 Dec 2023 • Nishant Vishwamitra, Keyan Guo, Farhan Tajwar Romit, Isabelle Ondracek, Long Cheng, Ziming Zhao, Hongxin Hu
HATEGUARD further achieves prompt-based zero-shot detection by automatically generating and updating detection prompts with new derogatory terms and targets in new wave samples to effectively address new waves of online hate.
no code implementations • 5 Sep 2023 • Yuze Liu, Ziming Zhao, Tiehua Zhang, Kang Wang, Xin Chen, Xiaowei Huang, Jun Yin, Zhishu Shen
Sleep stage classification is crucial for detecting patients' health conditions.
no code implementations • 1 Dec 2022 • Ziqi Yang, Lijin Wang, Da Yang, Jie Wan, Ziming Zhao, Ee-Chien Chang, Fan Zhang, Kui Ren
Besides, our further experiments show that PURIFIER is also effective in defending adversarial model inversion attacks and attribute inference attacks.
1 code implementation • 8 Jun 2022 • Jun Yan, Huilin Yin, Xiaoyang Deng, Ziming Zhao, Wancheng Ge, Hao Zhang, Gerhard Rigoll
Since adversarial vulnerability can be regarded as a high-frequency phenomenon, it is essential to regulate the adversarially-trained neural network models in the frequency domain.
no code implementations • 10 Mar 2022 • Junjie Shen, Ningfei Wang, Ziwen Wan, Yunpeng Luo, Takami Sato, Zhisheng Hu, Xinyang Zhang, Shengjian Guo, Zhenyu Zhong, Kang Li, Ziming Zhao, Chunming Qiao, Qi Alfred Chen
In this paper, we perform the first systematization of knowledge of such growing semantic AD AI security research space.
no code implementations • 22 Dec 2021 • Nishant Vishwamitra, Hongxin Hu, Ziming Zhao, Long Cheng, Feng Luo
We then introduce a new type of multimodal adversarial attacks called decoupling attack in MUROAN that aims to compromise multimodal models by decoupling their fused modalities.
1 code implementation • 23 Feb 2016 • Sailik Sengupta, Satya Gautam Vadlamudi, Subbarao Kambhampati, Marthony Taguinod, Adam Doupé, Ziming Zhao, Gail-Joon Ahn
We also address the issue of prioritizing vulnerabilities that when fixed, improves the security of the MTD system.