no code implementations • 26 Oct 2023 • Donald Loveland, Rajmonda Caceres
However, the factors which govern gradient-based editing are understudied, obscuring why edges are chosen and if edits are grounded in an edge's importance.
no code implementations • 8 Jun 2023 • Donald Loveland, Jiong Zhu, Mark Heimann, Benjamin Fish, Michael T. Schaub, Danai Koutra
We ground the practical implications of this work through granular analysis on five real-world datasets with varying global homophily levels, demonstrating that (a) GNNs can fail to generalize to test nodes that deviate from the global homophily of a graph, and (b) high local homophily does not necessarily confer high performance for a node.
no code implementations • 10 Jul 2022 • Donald Loveland, Jiong Zhu, Mark Heimann, Ben Fish, Michael T. Schaub, Danai Koutra
We study the task of node classification for graph neural networks (GNNs) and establish a connection between group fairness, as measured by statistical parity and equal opportunity, and local assortativity, i. e., the tendency of linked nodes to have similar attributes.
no code implementations • 4 Jun 2022 • Ioannis Tsaknakis, Bhavya Kailkhura, Sijia Liu, Donald Loveland, James Diffenderfer, Anna Maria Hiszpanski, Mingyi Hong
Existing knowledge integration approaches are limited to using differentiable knowledge source to be compatible with first-order DL training paradigm.
1 code implementation • 10 Jan 2022 • Donald Loveland, Jiayi Pan, Aaresh Farrokh Bhathena, Yiyang Lu
While edge deletion is a common method used to promote fairness in GNNs, it fails to consider when data is inherently missing fair connections.
no code implementations • 25 Jun 2021 • Donald Loveland, Shusen Liu, Bhavya Kailkhura, Anna Hiszpanski, Yong Han
Graph neural network (GNN) explanations have largely been facilitated through post-hoc introspection.
1 code implementation • 14 Jun 2021 • Jiong Zhu, Junchen Jin, Donald Loveland, Michael T. Schaub, Danai Koutra
We bridge two research directions on graph neural networks (GNNs), by formalizing the relation between heterophily of node labels (i. e., connected nodes tend to have dissimilar labels) and the robustness of GNNs to adversarial attacks.
no code implementations • 16 Jul 2020 • Shusen Liu, Bhavya Kailkhura, Jize Zhang, Anna M. Hiszpanski, Emily Robertson, Donald Loveland, T. Yong-Jin Han
The scientific community has been increasingly interested in harnessing the power of deep learning to solve various domain challenges.
no code implementations • 30 Jun 2020 • Shusen Liu, Bhavya Kailkhura, Jize Zhang, Anna M. Hiszpanski, Emily Robertson, Donald Loveland, T. Yong-Jin Han
The scientific community has been increasingly interested in harnessing the power of deep learning to solve various domain challenges.
no code implementations • 6 Jul 2019 • Shusen Liu, Bhavya Kailkhura, Donald Loveland, Yong Han
In this work, we propose an introspection technique for deep neural networks that relies on a generative model to instigate salient editing of the input image for model interpretation.