1 code implementation • 28 Sep 2020 • Peiyuan Liao, Han Zhao, Keyulu Xu, Tommi Jaakkola, Geoffrey Gordon, Stefanie Jegelka, Ruslan Salakhutdinov
While the advent of Graph Neural Networks (GNNs) has greatly improved node and graph representation learning in many applications, the neighborhood aggregation scheme exposes additional vulnerabilities to adversaries seeking to extract node-level information about sensitive attributes.
2 code implementations • NeurIPS 2019 • Renato Negrinho, Darshan Patil, Nghia Le, Daniel Ferreira, Matthew Gormley, Geoffrey Gordon
We release an implementation of our language with this paper.
2 code implementations • ICML 2018 • Ahmed Hefny, Zita Marinho, Wen Sun, Siddhartha Srinivasa, Geoffrey Gordon
Predictive state policy networks consist of a recursive filter, which keeps track of a belief about the state of the environment, and a reactive policy that directly maps beliefs to actions, to maximize the cumulative reward.
no code implementations • 6 Dec 2017 • Borui Wang, Geoffrey Gordon
Learning general latent-variable probabilistic graphical models is a key theoretical challenge in machine learning and artificial intelligence.
no code implementations • NeurIPS 2017 • Carlton Downey, Ahmed Hefny, Boyue Li, Byron Boots, Geoffrey Gordon
We present a new model, Predictive State Recurrent Neural Networks (PSRNNs), for filtering and prediction in dynamical systems.
no code implementations • 14 Feb 2017 • Carlton Downey, Ahmed Hefny, Geoffrey Gordon
Unfortunately it is not obvious how to apply apply an EM style algorithm in the context of PSRs as the Log Likelihood is not well defined for all PSRs.
no code implementations • NeurIPS 2015 • Ahmed Hefny, Carlton Downey, Geoffrey Gordon
To address this problem, we present a new view of dynamical system learning: we show how to learn dynamical systems by solving a sequence of ordinary supervised learning problems, thereby allowing users to incorporate prior knowledge via standard techniques such as L1 regularization.
no code implementations • 26 Sep 2013 • Byron Boots, Geoffrey Gordon, Arthur Gretton
The essence is to represent the state as a nonparametric conditional embedding operator in a Reproducing Kernel Hilbert Space (RKHS) and leverage recent work in kernel methods to estimate, predict, and update the representation.