Search Results for author: Gal Mishne

Found 30 papers, 11 papers with code

Diffusion Nets

no code implementations25 Jun 2015 Gal Mishne, Uri Shaham, Alexander Cloninger, Israel Cohen

In this paper, we propose a manifold learning algorithm based on deep learning to create an encoder, which maps a high-dimensional dataset and its low-dimensional embedding, and a decoder, which takes the embedded data back to the high-dimensional space.

Outlier Detection

Hierarchical Coupled Geometry Analysis for Neuronal Structure and Activity Pattern Discovery

no code implementations6 Nov 2015 Gal Mishne, Ronen Talmon, Ron Meir, Jackie Schiller, Uri Dubin, Ronald R. Coifman

In the wake of recent advances in experimental methods in neuroscience, the ability to record in-vivo neuronal activity from awake animals has become feasible.

The Geometry of Nodal Sets and Outlier Detection

no code implementations5 Jun 2017 Xiuyuan Cheng, Gal Mishne, Stefan Steinerberger

Let $(M, g)$ be a compact manifold and let $-\Delta \phi_k = \lambda_k \phi_k$ be the sequence of Laplacian eigenfunctions.

Outlier Detection

Data-Driven Tree Transforms and Metrics

1 code implementation18 Aug 2017 Gal Mishne, Ronen Talmon, Israel Cohen, Ronald R. Coifman, Yuval Kluger

Often the data is such that the observations do not reside on a regular grid, and the given order of the features is arbitrary and does not convey a notion of locality.

Clustering

Randomized Near Neighbor Graphs, Giant Components, and Applications in Data Science

3 code implementations13 Nov 2017 George C. Linderman, Gal Mishne, Yuval Kluger, Stefan Steinerberger

If we pick $n$ random points uniformly in $[0, 1]^d$ and connect each point to its $k-$nearest neighbors, then it is well known that there exists a giant connected component with high probability.

Co-manifold learning with missing data

no code implementations16 Oct 2018 Gal Mishne, Eric C. Chi, Ronald R. Coifman

We propose utilizing this coupled structure to perform co-manifold learning: uncovering the underlying geometry of both the rows and the columns of a given matrix, where we focus on a missing data setting.

Clustering Data Visualization +1

Spectral Embedding Norm: Looking Deep into the Spectrum of the Graph Laplacian

1 code implementation25 Oct 2018 Xiuyuan Cheng, Gal Mishne

The extraction of clusters from a dataset which includes multiple clusters and a significant background component is a non-trivial task of practical importance.

Anomaly Detection Clustering +1

Multi-way Graph Signal Processing on Tensors: Integrative analysis of irregular geometries

no code implementations30 Jun 2020 Jay S. Stanley III, Eric C. Chi, Gal Mishne

Graph signal processing (GSP) is an important methodology for studying data residing on irregular structures.

Kernel-based parameter estimation of dynamical systems with unknown observation functions

no code implementations9 Sep 2020 Ofir Lindenbaum, Amir Sagiv, Gal Mishne, Ronen Talmon

A low-dimensional dynamical system is observed in an experiment as a high-dimensional signal; for example, a video of a chaotic pendulums system.

Provable Robustness by Geometric Regularization of ReLU Networks

no code implementations1 Jan 2021 Chester Holtz, Changhao Shi, Gal Mishne

Recent work has demonstrated that neural networks are vulnerable to small, adversarial perturbations of their input.

Online Adversarial Purification based on Self-supervised Learning

no code implementations ICLR 2021 Changhao Shi, Chester Holtz, Gal Mishne

Deep neural networks are known to be vulnerable to adversarial examples, where a perturbation in the input space leads to an amplified shift in the latent network representation.

Representation Learning Self-Supervised Learning

Online Adversarial Purification based on Self-Supervision

no code implementations23 Jan 2021 Changhao Shi, Chester Holtz, Gal Mishne

To the best of our knowledge, our paper is the first that generalizes the idea of using self-supervised signals to perform online test-time purification.

Representation Learning

LDLE: Low Distortion Local Eigenmaps

1 code implementation ICLR Workshop GTRL 2021 Dhruv Kohli, Alexander Cloninger, Gal Mishne

We present Low Distortion Local Eigenmaps (LDLE), a manifold learning technique which constructs a set of low distortion local views of a dataset in lower dimension and registers them to obtain a global embedding.

Learning Sample Reweighting for Adversarial Robustness

no code implementations29 Sep 2021 Chester Holtz, Tsui-Wei Weng, Gal Mishne

There has been great interest in enhancing the robustness of neural network classifiers to defend against adversarial perturbations through adversarial training, while balancing the trade-off between robust accuracy and standard accuracy.

Adversarial Robustness Bilevel Optimization

Learning Disentangled Behavior Embeddings

1 code implementation NeurIPS 2021 Changhao Shi, Sivan Schwartz, Shahar Levy, Shay Achvat, Maisan Abboud, Amir Ghanayim, Jackie Schiller, Gal Mishne

To understand the relationship between behavior and neural activity, experiments in neuroscience often include an animal performing a repeated behavior such as a motor task.

Data Processing of Functional Optical Microscopy for Neuroscience

no code implementations10 Jan 2022 Hadas Benisty, Alexander Song, Gal Mishne, Adam S. Charles

Functional optical imaging in neuroscience is rapidly growing with the development of new optical systems and fluorescence indicators.

Evaluating Disentanglement in Generative Models Without Knowledge of Latent Factors

no code implementations4 Oct 2022 Chester Holtz, Gal Mishne, Alexander Cloninger

Probabilistic generative models provide a flexible and systematic framework for learning the underlying geometry of data.

Disentanglement Fairness +2

Learning Sample Reweighting for Accuracy and Adversarial Robustness

no code implementations20 Oct 2022 Chester Holtz, Tsui-Wei Weng, Gal Mishne

There has been great interest in enhancing the robustness of neural network classifiers to defend against adversarial perturbations through adversarial training, while balancing the trade-off between robust accuracy and standard accuracy.

Adversarial Robustness Bilevel Optimization

Implicit Graphon Neural Representation

1 code implementation7 Nov 2022 Xinyue Xia, Gal Mishne, Yusu Wang

We also show that our model is suitable for graph representation learning and graph generation.

Graph Generation Graph Representation Learning

DiSC: Differential Spectral Clustering of Features

1 code implementation10 Nov 2022 Ram Dyuthi Sristi, Gal Mishne, Ariel Jaffe

Selecting subsets of features that differentiate between two conditions is a key task in a broad range of scientific domains.

Clustering Stochastic Block Model

SiBBlInGS: Similarity-driven Building-Block Inference using Graphs across States

1 code implementation7 Jun 2023 Noga Mudrik, Gal Mishne, Adam S. Charles

Time series data across scientific domains are often collected under distinct states (e. g., tasks), wherein latent processes (e. g., biological factors) create complex inter- and intra-state variability.

Dictionary Learning Time Series

Graph Laplacian Learning with Exponential Family Noise

no code implementations14 Jun 2023 Changhao Shi, Gal Mishne

A common challenge in applying graph machine learning methods is that the underlying graph of a system is often unknown.

Semi-Supervised Laplacian Learning on Stiefel Manifolds

no code implementations31 Jul 2023 Chester Holtz, PengWen Chen, Alexander Cloninger, Chung-Kuan Cheng, Gal Mishne

Motivated by the need to address the degeneracy of canonical Laplace learning algorithms in low label rates, we propose to reformulate graph-based semi-supervised learning as a nonconvex generalization of a \emph{Trust-Region Subproblem} (TRS).

Contextual Feature Selection with Conditional Stochastic Gates

no code implementations21 Dec 2023 Ram Dyuthi Sristi, Ofir Lindenbaum, Maria Lavzin, Jackie Schiller, Gal Mishne, Hadas Benisty

We study the problem of contextual feature selection, where the goal is to learn a predictive function while identifying subsets of informative features conditioned on specific contexts.

feature selection

Learning Cartesian Product Graphs with Laplacian Constraints

no code implementations12 Feb 2024 Changhao Shi, Gal Mishne

We establish statistical consistency for the penalized maximum likelihood estimation (MLE) of a Cartesian product Laplacian, and propose an efficient algorithm to solve the problem.

Graph Learning Imputation

Deep and shallow data science for multi-scale optical neuroscience

no code implementations13 Feb 2024 Gal Mishne, Adam Charles

Optical imaging of the brain has expanded dramatically in the past two decades.

Comparing Graph Transformers via Positional Encodings

no code implementations22 Feb 2024 Mitchell Black, Zhengchao Wan, Gal Mishne, Amir Nayyeri, Yusu Wang

The distinguishing power of graph transformers is closely tied to the choice of positional encoding: features used to augment the base transformer with information about the graph.

Navigate

Cannot find the paper you are looking for? You can Submit a new open access paper.