1 code implementation • 26 Sep 2024 • Kaike Zhang, Qi Cao, Yunfan Wu, Fei Sun, HuaWei Shen, Xueqi Cheng
Leveraging these insights, we introduce the Vulnerability-aware Adversarial Training (VAT), designed to defend against poisoning attacks in recommender systems.
no code implementations • 20 Aug 2024 • Yunfan Wu, Qi Cao, Shuchang Tao, Kaike Zhang, Fei Sun, HuaWei Shen
Recent studies have demonstrated the vulnerability of recommender systems to data poisoning attacks, where adversaries inject carefully crafted fake user interactions into the training data of recommenders to promote target items.
no code implementations • 31 Jan 2024 • Kaike Zhang, Qi Cao, Yunfan Wu, Fei Sun, HuaWei Shen, Xueqi Cheng
Traditional defense strategies predominantly depend on predefined assumptions or rules extracted from specific known attacks, limiting their generalizability to unknown attack types.
no code implementations • 5 Sep 2023 • Kaike Zhang, Qi Cao, Fei Sun, Yunfan Wu, Shuchang Tao, HuaWei Shen, Xueqi Cheng
With the rapid growth of information, recommender systems have become integral for providing personalized suggestions and overcoming information overload.
no code implementations • 25 May 2023 • Shuchang Tao, Qi Cao, HuaWei Shen, Yunfan Wu, Bingbing Xu, Xueqi Cheng
To address these limitations, we analyze the causalities in graph adversarial attacks and conclude that causal features are key to achieve graph adversarial robustness, owing to their determinedness for labels and invariance across attacks.
1 code implementation • 9 May 2023 • YuanHao Liu, Qi Cao, HuaWei Shen, Yunfan Wu, Shuchang Tao, Xueqi Cheng
In this paper, we propose a new criterion for popularity debiasing, i. e., in an unbiased recommender system, both popular and unpopular items should receive Interactions Proportional to the number of users who Like it, namely IPL criterion.
1 code implementation • 16 Feb 2023 • Shuchang Tao, HuaWei Shen, Qi Cao, Yunfan Wu, Liang Hou, Xueqi Cheng
In this paper, we propose and formulate graph adversarial immunization, i. e., vaccinating part of graph structure to improve certifiable robustness of graph against any admissible adversarial attack.
1 code implementation • 3 Aug 2022 • Shuchang Tao, Qi Cao, HuaWei Shen, Yunfan Wu, Liang Hou, Fei Sun, Xueqi Cheng
In this paper, we first propose and define camouflage as distribution similarity between ego networks of injected nodes and normal nodes.
1 code implementation • 30 Aug 2021 • Shuchang Tao, Qi Cao, HuaWei Shen, JunJie Huang, Yunfan Wu, Xueqi Cheng
In this paper, we focus on an extremely limited scenario of single node injection evasion attack, i. e., the attacker is only allowed to inject one single node during the test phase to hurt GNN's performance.
1 code implementation • 2 Aug 2021 • Shi Qiu, Yunfan Wu, Saeed Anwar, Chongyi Li
Object detection in three-dimensional (3D) space attracts much interest from academia and industry since it is an essential task in AI-driven applications such as robotics, autonomous driving, and augmented reality.
1 code implementation • 12 Jul 2021 • Yunfan Wu, Qi Cao, HuaWei Shen, Shuchang Tao, Xueqi Cheng
INMO generates the inductive embeddings for users (items) by characterizing their interactions with some template items (template users), instead of employing an embedding lookup table.