Search Results for author: Siddhartha Mishra

Found 36 papers, 16 papers with code

SmallToLarge (S2L): Scalable Data Selection for Fine-tuning Large Language Models by Summarizing Training Trajectories of Small Models

no code implementations12 Mar 2024 Yu Yang, Siddhartha Mishra, Jeffrey N Chiang, Baharan Mirzasoleiman

In clinical text summarization on the MIMIC-III dataset (Johnson et al., 2016), S2L again outperforms training on the full dataset using only 50% of the data.

Math Text Summarization

A universal approximation theorem for nonlinear resistive networks

no code implementations22 Dec 2023 Benjamin Scellier, Siddhartha Mishra

Resistor networks have recently had a surge of interest as substrates for energy-efficient self-learning machines.

Self-Learning

An operator preconditioning perspective on training in physics-informed machine learning

no code implementations9 Oct 2023 Tim De Ryck, Florent Bonnet, Siddhartha Mishra, Emmanuel de Bézenac

In this paper, we investigate the behavior of gradient descent algorithms in physics-informed machine learning methods like PINNs, which minimize residuals connected to partial differential equations (PDEs).

Physics-informed machine learning

How does over-squashing affect the power of GNNs?

no code implementations6 Jun 2023 Francesco Di Giovanni, T. Konstantin Rusch, Michael M. Bronstein, Andreea Deac, Marc Lackenby, Siddhartha Mishra, Petar Veličković

In this paper, we provide a rigorous analysis to determine which function classes of node features can be learned by an MPNN of a given capacity.

A Structured Matrix Method for Nonequispaced Neural Operators

1 code implementation31 May 2023 Levi Lingsch, Mike Michelis, Emmanuel de Bezenac, Sirani M. Perera, Robert K. Katzschmann, Siddhartha Mishra

The computational efficiency of many neural operators, widely used for learning solutions of PDEs, relies on the fast Fourier transform (FFT) for performing spectral computations.

Computational Efficiency

A Survey on Oversmoothing in Graph Neural Networks

no code implementations20 Mar 2023 T. Konstantin Rusch, Michael M. Bronstein, Siddhartha Mishra

Node features of graph neural networks (GNNs) tend to become more similar with the increase of the network depth.

Graph Learning

Multi-Scale Message Passing Neural PDE Solvers

no code implementations7 Feb 2023 Léonard Equer, T. Konstantin Rusch, Siddhartha Mishra

We propose a novel multi-scale message passing neural network algorithm for learning the solutions of time-dependent PDEs.

Convolutional Neural Operators for robust and accurate learning of PDEs

1 code implementation NeurIPS 2023 Bogdan Raonić, Roberto Molinaro, Tim De Ryck, Tobias Rohner, Francesca Bartolucci, Rima Alaifari, Siddhartha Mishra, Emmanuel de Bézenac

Although very successfully used in conventional machine learning, convolution based neural network architectures -- believed to be inconsistent in function space -- have been largely ignored in the context of learning solution operators of PDEs.

Operator learning PDE Surrogate Modeling

Neural Inverse Operators for Solving PDE Inverse Problems

no code implementations26 Jan 2023 Roberto Molinaro, Yunan Yang, Björn Engquist, Siddhartha Mishra

A large class of inverse problems for PDEs are only well-defined as mappings from operators to functions.

Operator learning

wPINNs: Weak Physics informed neural networks for approximating entropy solutions of hyperbolic conservation laws

no code implementations18 Jul 2022 Tim De Ryck, Siddhartha Mishra, Roberto Molinaro

Physics informed neural networks (PINNs) require regularity of solutions of the underlying PDE to guarantee accurate approximation.

Error analysis for deep neural network approximations of parametric hyperbolic conservation laws

no code implementations15 Jul 2022 Tim De Ryck, Siddhartha Mishra

We derive rigorous bounds on the error resulting from the approximation of the solution of parametric hyperbolic scalar conservation laws with ReLU neural networks.

Agnostic Physics-Driven Deep Learning

no code implementations30 May 2022 Benjamin Scellier, Siddhartha Mishra, Yoshua Bengio, Yann Ollivier

This work establishes that a physical system can perform statistical learning without gradient computations, via an Agnostic Equilibrium Propagation (Aeqprop) procedure that combines energy minimization, homeostatic control, and nudging towards the correct response.

Variable-Input Deep Operator Networks

no code implementations23 May 2022 Michael Prasthofer, Tim De Ryck, Siddhartha Mishra

Existing architectures for operator learning require that the number and locations of sensors (where the input functions are evaluated) remain the same across all training and test samples, significantly restricting the range of their applicability.

Operator learning

Generic bounds on the approximation error for physics-informed (and) operator learning

no code implementations23 May 2022 Tim De Ryck, Siddhartha Mishra

We propose a very general framework for deriving rigorous bounds on the approximation error for physics-informed neural networks (PINNs) and operator learning architectures such as DeepONets and FNOs as well as for physics-informed operator learning.

Operator learning

Error estimates for physics informed neural networks approximating the Navier-Stokes equations

no code implementations17 Mar 2022 Tim De Ryck, Ameya D. Jagtap, Siddhartha Mishra

We prove rigorous bounds on the errors resulting from the approximation of the incompressible Navier-Stokes equations with (extended) physics informed neural networks.

Graph-Coupled Oscillator Networks

1 code implementation4 Feb 2022 T. Konstantin Rusch, Benjamin P. Chamberlain, James Rowbottom, Siddhartha Mishra, Michael M. Bronstein

This demonstrates that the proposed framework mitigates the oversmoothing problem.

Error analysis for physics informed neural networks (PINNs) approximating Kolmogorov PDEs

no code implementations28 Jun 2021 Tim De Ryck, Siddhartha Mishra

Moreover, we prove that the size of the PINNs and the number of training samples only grow polynomially with the underlying dimension, enabling PINNs to overcome the curse of dimensionality in this context.

On the approximation of functions by tanh neural networks

no code implementations18 Apr 2021 Tim De Ryck, Samuel Lanthaler, Siddhartha Mishra

We derive bounds on the error, in high-order Sobolev norms, incurred in the approximation of Sobolev-regular as well as analytic functions by neural networks with the hyperbolic tangent activation function.

UnICORNN: A recurrent model for learning very long time dependencies

1 code implementation9 Mar 2021 T. Konstantin Rusch, Siddhartha Mishra

The design of recurrent neural networks (RNNs) to accurately process sequential inputs with long-time dependencies is very challenging on account of the exploding and vanishing gradient problem.

Sentiment Analysis Sequential Image Classification +2

Iterative Surrogate Model Optimization (ISMO): An active learning algorithm for PDE constrained optimization with deep neural networks

4 code implementations13 Aug 2020 Kjetil O. Lye, Siddhartha Mishra, Deep Ray, Praveen Chandrasekhar

We present a novel active learning algorithm, termed as iterative surrogate model optimization (ISMO), for robust and efficient numerical approximation of PDE constrained optimization problems.

Active Learning Model Optimization

Estimates on the generalization error of Physics Informed Neural Networks (PINNs) for approximating PDEs

no code implementations29 Jun 2020 Siddhartha Mishra, Roberto Molinaro

Physics informed neural networks (PINNs) have recently been widely used for robust and accurate approximation of PDEs.

Estimates on the generalization error of Physics Informed Neural Networks (PINNs) for approximating a class of inverse problems for PDEs

no code implementations29 Jun 2020 Siddhartha Mishra, Roberto Molinaro

Physics informed neural networks (PINNs) have recently been very successfully applied for efficiently approximating inverse problems for PDEs.

Enhancing accuracy of deep learning algorithms by training with low-discrepancy sequences

1 code implementation26 May 2020 Siddhartha Mishra, T. Konstantin Rusch

We propose a deep supervised learning algorithm based on low-discrepancy sequences as the training set.

On the approximation of rough functions with deep neural networks

no code implementations13 Dec 2019 Tim De Ryck, Siddhartha Mishra, Deep Ray

Deep neural networks and the ENO procedure are both efficient frameworks for approximating rough functions.

Data Compression

A Multi-level procedure for enhancing accuracy of machine learning algorithms

1 code implementation20 Sep 2019 Kjetil O. Lye, Siddhartha Mishra, Roberto Molinaro

We propose a multi-level method to increase the accuracy of machine learning algorithms for approximating observables in scientific computing, particularly those that arise in systems modeled by differential equations.

BIG-bench Machine Learning Uncertainty Quantification

Statistical solutions of hyperbolic systems of conservation laws: numerical approximation

2 code implementations6 Jun 2019 Ulrik Skre Fjordholm, Kjetil Lye, Siddhartha Mishra, Franziska Weber

Statistical solutions are time-parameterized probability measures on spaces of integrable functions, that have been proposed recently as a framework for global solutions and uncertainty quantification for multi-dimensional hyperbolic system of conservation laws.

Numerical Analysis Fluid Dynamics 35L65, 65M08, 65C05, 65C30

Deep learning observables in computational fluid dynamics

3 code implementations7 Mar 2019 Kjetil O. Lye, Siddhartha Mishra, Deep Ray

Under the assumption that the underlying neural networks generalize well, we prove that the deep learning MC and QMC algorithms are guaranteed to be faster than the baseline (quasi-) Monte Carlo methods.

Efficient Neural Network Uncertainty Quantification

Numerical approximation of statistical solutions of scalar conservation laws

1 code implementation30 Oct 2017 Ulrik Skre Fjordholm, Kjetil Lye, Siddhartha Mishra

We propose efficient numerical algorithms for approximating statistical solutions of scalar conservation laws.

Numerical Analysis 35L65, 65M08, 65C05, 65C30

Cannot find the paper you are looking for? You can Submit a new open access paper.