Search Results for author: Guruprasad Raghavan

Found 8 papers, 2 papers with code

Engineering flexible machine learning systems by traversing functionally-invariant paths

1 code implementation30 Apr 2022 Guruprasad Raghavan, Bahey Tharwat, Surya Narayanan Hari, Dhruvil Satani, Matt Thomson

We conceptualize the weight space of a neural network as a curved Riemannian manifold equipped with a metric tensor whose spectrum defines low rank subspaces in weight space that accommodate network adaptation without loss of prior knowledge.

Adversarial Robustness Continual Learning +4

Traversing Geodesics to Grow Biological Structures

no code implementations NeurIPS Workshop AI4Scien 2021 Pranav Bhamidipati, Guruprasad Raghavan, Matt Thomson

Biological tissues reliably grow into precise, functional structures from simple starting states during development.

Total Energy

Solving hybrid machine learning tasks by traversing weight space geodesics

no code implementations5 Jun 2021 Guruprasad Raghavan, Matt Thomson

Broadly, we introduce a geometric framework that unifies a range of machine learning objectives and that can be applied to multiple classes of neural network architectures.

BIG-bench Machine Learning

Sparsifying networks by traversing Geodesics

no code implementations NeurIPS Workshop DL-IG 2020 Guruprasad Raghavan, Matt Thomson

The geometry of weight spaces and functional manifolds of neural networks play an important role towards 'understanding' the intricacies of ML.

Architecture Agnostic Neural Networks

no code implementations5 Nov 2020 Sabera Talukder, Guruprasad Raghavan, Yisong Yue

Within this sparse, binary paradigm we sample many binary architectures to create families of architecture agnostic neural networks not trained via backpropagation.

Self-organization of multi-layer spiking neural networks

no code implementations12 Jun 2020 Guruprasad Raghavan, Cong Lin, Matt Thomson

Inspired by this strategy, we attempt to efficiently self-organize large neural networks with an arbitrary number of layers into a wide variety of architectures.

Geometric algorithms for predicting resilience and recovering damage in neural networks

no code implementations23 May 2020 Guruprasad Raghavan, Jiayi Li, Matt Thomson

Biological neural networks have evolved to maintain performance despite significant circuit damage.

Neural networks grown and self-organized by noise

2 code implementations NeurIPS 2019 Guruprasad Raghavan, Matt Thomson

The algorithm is adaptable to a wide-range of input-layer geometries, robust to malfunctioning units in the first layer, and so can be used to successfully grow and self-organize pooling architectures of different pool-sizes and shapes.

Cannot find the paper you are looking for? You can Submit a new open access paper.