Search Results for author: Ziming Zhao

Found 9 papers, 4 papers with code

Moderating Illicit Online Image Promotion for Unsafe User-Generated Content Games Using Large Vision-Language Models

1 code implementation27 Mar 2024 Keyan Guo, Ayush Utkarsh, Wenbo Ding, Isabelle Ondracek, Ziming Zhao, Guo Freeman, Nishant Vishwamitra, Hongxin Hu

Online user-generated content games (UGCGs) are increasingly popular among children and adolescents for social interaction and more creative online entertainment.

Domain Adaptation

An Investigation of Large Language Models for Real-World Hate Speech Detection

no code implementations7 Jan 2024 Keyan Guo, Alexander Hu, Jaden Mu, Ziheng Shi, Ziming Zhao, Nishant Vishwamitra, Hongxin Hu

Our study reveals that a meticulously crafted reasoning prompt can effectively capture the context of hate speech by fully utilizing the knowledge base in LLMs, significantly outperforming existing techniques.

Hate Speech Detection

Moderating New Waves of Online Hate with Chain-of-Thought Reasoning in Large Language Models

1 code implementation22 Dec 2023 Nishant Vishwamitra, Keyan Guo, Farhan Tajwar Romit, Isabelle Ondracek, Long Cheng, Ziming Zhao, Hongxin Hu

HATEGUARD further achieves prompt-based zero-shot detection by automatically generating and updating detection prompts with new derogatory terms and targets in new wave samples to effectively address new waves of online hate.

Purifier: Defending Data Inference Attacks via Transforming Confidence Scores

no code implementations1 Dec 2022 Ziqi Yang, Lijin Wang, Da Yang, Jie Wan, Ziming Zhao, Ee-Chien Chang, Fan Zhang, Kui Ren

Besides, our further experiments show that PURIFIER is also effective in defending adversarial model inversion attacks and attribute inference attacks.

Attribute Inference Attack +1

Wavelet Regularization Benefits Adversarial Training

1 code implementation8 Jun 2022 Jun Yan, Huilin Yin, Xiaoyang Deng, Ziming Zhao, Wancheng Ge, Hao Zhang, Gerhard Rigoll

Since adversarial vulnerability can be regarded as a high-frequency phenomenon, it is essential to regulate the adversarially-trained neural network models in the frequency domain.

Adversarial Robustness

Understanding and Measuring Robustness of Multimodal Learning

no code implementations22 Dec 2021 Nishant Vishwamitra, Hongxin Hu, Ziming Zhao, Long Cheng, Feng Luo

We then introduce a new type of multimodal adversarial attacks called decoupling attack in MUROAN that aims to compromise multimodal models by decoupling their fused modalities.

Adversarial Robustness

Moving Target Defense for Web Applications using Bayesian Stackelberg Games

1 code implementation23 Feb 2016 Sailik Sengupta, Satya Gautam Vadlamudi, Subbarao Kambhampati, Marthony Taguinod, Adam Doupé, Ziming Zhao, Gail-Joon Ahn

We also address the issue of prioritizing vulnerabilities that when fixed, improves the security of the MTD system.

Cannot find the paper you are looking for? You can Submit a new open access paper.