Search Results for author: Daokun Zhang

Found 11 papers, 6 papers with code

Combating Confirmation Bias: A Unified Pseudo-Labeling Framework for Entity Alignment

no code implementations5 Jul 2023 Qijie Ding, Jie Yin, Daokun Zhang, Junbin Gao

Entity alignment (EA) aims at identifying equivalent entity pairs across different knowledge graphs (KGs) that refer to the same real-world identity.

Entity Alignment Knowledge Graphs +1

Towards Complex Dynamic Physics System Simulation with Graph Neural ODEs

no code implementations21 May 2023 Guangsi Shi, Daokun Zhang, Ming Jin, Shirui Pan, Philip S. Yu

To better comprehend the complex physical laws, this paper proposes a novel learning based simulation model- Graph Networks with Spatial-Temporal neural Ordinary Equations (GNSTODE)- that characterizes the varying spatial and temporal dependencies in particle systems using a united end-to-end framework.

Beyond Smoothing: Unsupervised Graph Representation Learning with Edge Heterophily Discriminating

1 code implementation25 Nov 2022 Yixin Liu, Yizhen Zheng, Daokun Zhang, Vincent CS Lee, Shirui Pan

Node representations are learned through contrasting the dual-channel encodings obtained from the discriminated homophilic and heterophilic edges.

Graph Representation Learning

Conflict-Aware Pseudo Labeling via Optimal Transport for Entity Alignment

1 code implementation5 Sep 2022 Qijie Ding, Daokun Zhang, Jie Yin

The key idea is to iteratively pseudo-label alignment pairs empowered with conflict-aware optimal transport (OT) modeling to boost the precision of entity alignment.

Entity Alignment Entity Embeddings +2

Link Prediction with Contextualized Self-Supervision

no code implementations25 Jan 2022 Daokun Zhang, Jie Yin, Philip S. Yu

To generate informative node embeddings for link prediction, structural context prediction is leveraged as a self-supervised learning task to boost the link prediction performance.

Attribute Inductive Link Prediction +1

Towards Unsupervised Deep Graph Structure Learning

1 code implementation17 Jan 2022 Yixin Liu, Yu Zheng, Daokun Zhang, Hongxu Chen, Hao Peng, Shirui Pan

To solve the unsupervised GSL problem, we propose a novel StrUcture Bootstrapping contrastive LearnIng fraMEwork (SUBLIME for abbreviation) with the aid of self-supervised contrastive learning.

Contrastive Learning Graph structure learning

Search Efficient Binary Network Embedding

1 code implementation14 Jan 2019 Daokun Zhang, Jie Yin, Xingquan Zhu, Chengqi Zhang

In this paper, we propose a search efficient binary network embedding algorithm called BinaryNE to learn a binary code for each node, by simultaneously modeling node context relations and node attribute relations through a three-layer neural network.

Attribute Network Embedding +2

Attributed Network Embedding via Subspace Discovery

1 code implementation14 Jan 2019 Daokun Zhang, Jie Yin, Xingquan Zhu, Chengqi Zhang

In this paper, we propose a unified framework for attributed network embedding-attri2vec-that learns node embeddings by discovering a latent node attribute subspace via a network structure guided transformation performed on the original attribute space.

Attribute Clustering +4

SINE: Scalable Incomplete Network Embedding

2 code implementations16 Oct 2018 Daokun Zhang, Jie Yin, Xingquan Zhu, Chengqi Zhang

In this paper, we propose a Scalable Incomplete Network Embedding (SINE) algorithm for learning node representations from incomplete graphs.

Social and Information Networks

MetaGraph2Vec: Complex Semantic Path Augmented Heterogeneous Network Embedding

no code implementations7 Mar 2018 Daokun Zhang, Jie Yin, Xingquan Zhu, Chengqi Zhang

Network embedding in heterogeneous information networks (HINs) is a challenging task, due to complications of different node types and rich relationships between nodes.

Social and Information Networks

Network Representation Learning: A Survey

no code implementations4 Dec 2017 Daokun Zhang, Jie Yin, Xingquan Zhu, Chengqi Zhang

Network representation learning has been recently proposed as a new learning paradigm to embed network vertices into a low-dimensional vector space, by preserving network topology structure, vertex content, and other side information.

Representation Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.