1 code implementation • 14 Oct 2024 • Jintang Li, Ruofan Wu, Yuchang Zhu, Huizhe Zhang, Xinzhou Jin, Guibin Zhang, Zulun Zhu, Zibin Zheng, Liang Chen
Graph autoencoders (GAEs) are self-supervised learning models that can learn meaningful representations of graph-structured data by reconstructing the input graph from a low-dimensional latent space.
1 code implementation • 19 Jul 2024 • Xinzhou Jin, Jintang Li, Liang Chen, Chenyun Yu, Yuanzhen Xie, Tao Xie, Chengxiang Zhuo, Zang Li, Zibin Zheng
Surprisingly, we find that L2CL, using only one-hop contrastive learning paradigm, is able to capture intrinsic semantic structures and improve the quality of node representation, leading to a simple yet effective architecture.
1 code implementation • 20 Jun 2024 • Yunfei Liu, Jintang Li, Yuehe Chen, Ruofan Wu, Ericbk Wang, Jing Zhou, Sheng Tian, Shuheng Shen, Xing Fu, Changhua Meng, Weiqiang Wang, Liang Chen
Another promising line of research involves the adoption of modularity maximization, a popular and effective measure for community detection, as the guiding principle for clustering tasks.
1 code implementation • 19 Jun 2024 • Yuchang Zhu, Jintang Li, Yatao Bian, Zibin Zheng, Liang Chen
Accordingly, we propose a graph fairness framework based on invariant learning, namely FairINV, which enables the training of fair GNNs to accommodate various sensitive attributes within a single training session.
1 code implementation • 3 Jun 2024 • Jintang Li, Ruofan Wu, Xinzhou Jin, Boqun Ma, Liang Chen, Zibin Zheng
Recently, state space models (SSMs), which are framed as discretized representations of an underlying continuous-time linear dynamical system, have garnered substantial attention and achieved breakthrough advancements in independent sequence modeling.
1 code implementation • 11 May 2024 • Yuchang Zhu, Jintang Li, Zibin Zheng, Liang Chen
In particular, the objective of group fairness is to ensure that the decisions made by GNNs are independent of the sensitive attribute.
1 code implementation • 26 Mar 2024 • Huizhe Zhang, Jintang Li, Liang Chen, Zibin Zheng
However, the costs behind outstanding performances of GTs are higher energy consumption and computational overhead.
1 code implementation • 5 Dec 2023 • Wangbin Sun, Jintang Li, Liang Chen, Bingzhe Wu, Yatao Bian, Zibin Zheng
Graph contrastive learning (GCL) has emerged as a representative paradigm in graph self-supervised learning, where negative samples are commonly regarded as the key to preventing model collapse and producing distinguishable representations.
1 code implementation • 29 Nov 2023 • Yuchang Zhu, Jintang Li, Liang Chen, Zibin Zheng
Experiments on several benchmark datasets demonstrate that FairGKD, which does not require access to demographic information, significantly improves the fairness of GNNs by a large margin while maintaining their utility.
no code implementations • 28 Nov 2023 • Jintang Li, Jiawang Dan, Ruofan Wu, Jing Zhou, Sheng Tian, Yunfei Liu, Baokun Wang, Changhua Meng, Weiqiang Wang, Yuchang Zhu, Liang Chen, Zibin Zheng
Over the past few years, graph neural networks (GNNs) have become powerful and practical tools for learning on (static) graph-structure data.
no code implementations • 18 Oct 2023 • Jintang Li, Zheng Wei, Jiawang Dan, Jing Zhou, Yuchang Zhu, Ruofan Wu, Baokun Wang, Zhang Zhen, Changhua Meng, Hong Jin, Zibin Zheng, Liang Chen
Through in-depth investigations on several real-world heterogeneous graphs exhibiting varying levels of heterophily, we have observed that heterogeneous graph neural networks (HGNNs), which inherit many mechanisms from GNNs designed for homogeneous graphs, fail to generalize to heterogeneous graphs with heterophily or low level of homophily.
1 code implementation • 13 Aug 2023 • Jie Liao, Jintang Li, Liang Chen, Bingzhe Wu, Yatao Bian, Zibin Zheng
In the pursuit of promoting the expressiveness of GNNs for tail nodes, we explore how the deficiency of structural information deteriorates the performance of tail nodes and propose a general Structural Augmentation based taIL nOde Representation learning framework, dubbed as SAILOR, which can jointly learn to augment the graph structure and extract more informative representations for tail nodes.
1 code implementation • 3 Jun 2023 • Jintang Li, Wangbin Sun, Ruofan Wu, Yuchang Zhu, Liang Chen, Zibin Zheng
Oversmoothing is a common phenomenon observed in graph neural networks (GNNs), in which an increase in the network depth leads to a deterioration in their performance.
1 code implementation • 30 May 2023 • Jintang Li, Huizhe Zhang, Ruofan Wu, Zulun Zhu, Baokun Wang, Changhua Meng, Zibin Zheng, Liang Chen
While contrastive self-supervised learning has become the de-facto learning paradigm for graph neural networks, the pursuit of higher task accuracy requires a larger hidden dimensionality to learn informative and discriminative full-precision representations, raising concerns about computation, memory footprint, and energy consumption burden (largely overlooked) for real-world applications.
1 code implementation • 23 May 2023 • Sheng Tian, Jihai Dong, Jintang Li, Wenlong Zhao, Xiaolong Xu, Baokun Wang, Bowen Song, Changhua Meng, Tianyi Zhang, Liang Chen
Anomaly detection aims to distinguish abnormal instances that deviate significantly from the majority of benign ones.
1 code implementation • 18 May 2023 • Jintang Li, Sheng Tian, Ruofan Wu, Liang Zhu, Welong Zhao, Changhua Meng, Liang Chen, Zibin Zheng, Hongzhi Yin
We approach the problem by our proposed STEP, a self-supervised temporal pruning framework that learns to remove potentially redundant edges from input dynamic graphs.
1 code implementation • 20 Nov 2022 • Jintang Li, Jiaying Peng, Liang Chen, Zibin Zheng, TingTing Liang, Qing Ling
In this work, we seek to address these challenges and propose Spectral Adversarial Training (SAT), a simple yet effective adversarial training approach for GNNs.
no code implementations • 9 Nov 2022 • Zishan Gu, Jintang Li, Liang Chen
Graph Neural Networks (GNNs) as deep learning models working on graph-structure data have achieved advanced performance in many works.
1 code implementation • 15 Aug 2022 • Jintang Li, Zhouxin Yu, Zulun Zhu, Liang Chen, Qi Yu, Zibin Zheng, Sheng Tian, Ruofan Wu, Changhua Meng
We explore a new direction in that we can capture the evolving dynamics of temporal graphs with spiking neural networks (SNNs) instead of RNNs.
no code implementations • 20 May 2022 • Bingzhe Wu, Jintang Li, Junchi Yu, Yatao Bian, Hengtong Zhang, Chaochao Chen, Chengbin Hou, Guoji Fu, Liang Chen, Tingyang Xu, Yu Rong, Xiaolin Zheng, Junzhou Huang, Ran He, Baoyuan Wu, Guangyu Sun, Peng Cui, Zibin Zheng, Zhe Liu, Peilin Zhao
Deep graph learning has achieved remarkable progresses in both business and scientific areas ranging from finance and e-commerce, to drug and advanced material discovery.
2 code implementations • 20 May 2022 • Jintang Li, Ruofan Wu, Wangbin Sun, Liang Chen, Sheng Tian, Liang Zhu, Changhua Meng, Zibin Zheng, Weiqiang Wang
The last years have witnessed the emergence of a promising self-supervised learning strategy, referred to as masked autoencoding.
1 code implementation • 5 May 2022 • Zulun Zhu, Jiaying Peng, Jintang Li, Liang Chen, Qi Yu, Siqiang Luo
Graph Convolutional Networks (GCNs) achieve an impressive performance due to the remarkable representation ability in learning the graph information.
1 code implementation • 20 Apr 2022 • Jintang Li, Jie Liao, Ruofan Wu, Liang Chen, Zibin Zheng, Jiawang Dan, Changhua Meng, Weiqiang Wang
To mitigate such a threat, considerable research efforts have been devoted to increasing the robustness of GCNs against adversarial attacks.
no code implementations • 15 Feb 2022 • Jintang Li, Bingzhe Wu, Chengbin Hou, Guoji Fu, Yatao Bian, Liang Chen, Junzhou Huang, Zibin Zheng
Despite the progress, applying DGL to real-world applications faces a series of reliability threats including inherent noise, distribution shift, and adversarial attacks.
no code implementations • 17 Jan 2022 • Liang Chen, Qibiao Peng, Jintang Li, Yang Liu, Jiawei Chen, Yong Li, Zibin Zheng
To address such a challenge, we set the trigger as a single node, and the backdoor is activated when the trigger node is connected to the target node.
1 code implementation • 13 Aug 2021 • Liang Chen, Jintang Li, Qibiao Peng, Yang Liu, Zibin Zheng, Carl Yang
In this work, we theoretically and empirically demonstrate that structural adversarial examples can be attributed to the non-robust aggregation scheme (i. e., the weighted mean) of GCNs.
1 code implementation • 16 Feb 2021 • Jintang Li, Kun Xu, Liang Chen, Zibin Zheng, Xiao Liu
Graph Neural Networks (GNNs) have recently shown to be powerful tools for representing and analyzing graph data.
1 code implementation • 8 Sep 2020 • Jintang Li, Tao Xie, Liang Chen, Fenfang Xie, Xiangnan He, Zibin Zheng
Currently, most works on attacking GNNs are mainly using gradient information to guide the attack and achieve outstanding performance.
2 code implementations • 10 Mar 2020 • Liang Chen, Jintang Li, Jiaying Peng, Tao Xie, Zengxu Cao, Kun Xu, Xiangnan He, Zibin Zheng, Bingzhe Wu
To bridge this gap, we investigate and summarize the existing works on graph adversarial learning tasks systemically.