no code implementations • 21 Feb 2024 • Heli Ben-Hamu, Omri Puny, Itai Gat, Brian Karrer, Uriel Singer, Yaron Lipman
Taming the generation outcome of state of the art Diffusion and Flow-Matching (FM) models without having to re-train a task-specific model unlocks a powerful tool for solving inverse problems, conditional generation, and controlled generation in general.
no code implementations • 14 Dec 2023 • Lior Yariv, Omri Puny, Natalia Neverova, Oran Gafni, Yaron Lipman
Current diffusion or flow-based generative models for 3D shapes divide to two: distilling pre-trained 2D image diffusion models, and training directly on 3D shapes.
no code implementations • 22 Feb 2023 • Omri Puny, Derek Lim, Bobak T. Kiani, Haggai Maron, Yaron Lipman
This paper introduces an alternative expressive power hierarchy based on the ability of GNNs to calculate equivariant polynomials of a certain degree.
no code implementations • ICLR 2022 • Omri Puny, Matan Atzmon, Heli Ben-Hamu, Ishan Misra, Aditya Grover, Edward J. Smith, Yaron Lipman
For example, Euclidean motion invariant/equivariant graph or point cloud neural networks.
3 code implementations • 14 Jun 2020 • Omri Puny, Heli Ben-Hamu, Yaron Lipman
This paper advocates incorporating a Low-Rank Global Attention (LRGA) module, a computation and memory efficient variant of the dot-product attention (Vaswani et al., 2017), to Graph Neural Networks (GNNs) for improving their generalization power.
Ranked #1 on Link Property Prediction on ogbl-ddi