Search Results for author: Rupesh Kumar Srivastava

Found 17 papers, 13 papers with code

Recurrent Highway Networks

5 code implementations ICML 2017 Julian Georg Zilly, Rupesh Kumar Srivastava, Jan Koutník, Jürgen Schmidhuber

We introduce a novel theoretical analysis of recurrent networks based on Gersgorin's circle theorem that illuminates several modeling and optimization issues and improves our understanding of the LSTM cell.

Language Modelling

EvoTorch: Scalable Evolutionary Computation in Python

1 code implementation24 Feb 2023 Nihat Engin Toklu, Timothy Atkinson, Vojtěch Micka, Paweł Liskowski, Rupesh Kumar Srivastava

Evolutionary computation is an important component within various fields such as artificial intelligence research, reinforcement learning, robotics, industrial automation and/or optimization, engineering design, etc.

reinforcement-learning Reinforcement Learning (RL)

Training Very Deep Networks

3 code implementations NeurIPS 2015 Rupesh Kumar Srivastava, Klaus Greff, Jürgen Schmidhuber

Theoretical and empirical evidence indicates that the depth of neural networks is crucial for their success.

Image Classification

Bayesian Flow Networks

1 code implementation14 Aug 2023 Alex Graves, Rupesh Kumar Srivastava, Timothy Atkinson, Faustino Gomez

Notably, the network inputs for discrete data lie on the probability simplex, and are therefore natively differentiable, paving the way for gradient-based sample guidance and few-step generation in discrete domains such as language modelling.

Bayesian Inference Data Compression +2

Highway Networks

3 code implementations3 May 2015 Rupesh Kumar Srivastava, Klaus Greff, Jürgen Schmidhuber

There is plenty of theoretical and empirical evidence that depth of neural networks is a crucial ingredient for their success.

Language Modelling

LSTM: A Search Space Odyssey

15 code implementations13 Mar 2015 Klaus Greff, Rupesh Kumar Srivastava, Jan Koutník, Bas R. Steunebrink, Jürgen Schmidhuber

Several variants of the Long Short-Term Memory (LSTM) architecture for recurrent neural networks have been proposed since its inception in 1995.

Handwriting Recognition Music Modeling +1

Training Agents using Upside-Down Reinforcement Learning

7 code implementations5 Dec 2019 Rupesh Kumar Srivastava, Pranav Shyam, Filipe Mutz, Wojciech Jaśkowski, Jürgen Schmidhuber

Many of its general principles are outlined in a companion report; the goal of this paper is to develop a practical learning algorithm and show that this conceptually simple perspective on agent training can produce a range of rewarding behaviors for multiple episodic environments.

reinforcement-learning Reinforcement Learning (RL)

ClipUp: A Simple and Powerful Optimizer for Distribution-based Policy Evolution

1 code implementation5 Aug 2020 Nihat Engin Toklu, Paweł Liskowski, Rupesh Kumar Srivastava

In these algorithms, gradients of the total reward with respect to the policy parameters are estimated using a population of solutions drawn from a search distribution, and then used for policy optimization with stochastic gradient ascent.

Continuous Control Humanoid Control

Binding via Reconstruction Clustering

1 code implementation19 Nov 2015 Klaus Greff, Rupesh Kumar Srivastava, Jürgen Schmidhuber

Disentangled distributed representations of data are desirable for machine learning, since they are more expressive and can generalize from fewer examples.

Clustering Denoising +1

Reward-Weighted Regression Converges to a Global Optimum

1 code implementation19 Jul 2021 Miroslav Štrupl, Francesco Faccio, Dylan R. Ashley, Rupesh Kumar Srivastava, Jürgen Schmidhuber

Reward-Weighted Regression (RWR) belongs to a family of widely known iterative Reinforcement Learning algorithms based on the Expectation-Maximization framework.

regression Reinforcement Learning (RL)

Multimeasurement Generative Models

1 code implementation ICLR 2022 Saeed Saremi, Rupesh Kumar Srivastava

We formally map the problem of sampling from an unknown distribution with a density in $\mathbb{R}^d$ to the problem of learning and sampling a smoother density in $\mathbb{R}^{Md}$ obtained by convolution with a fixed factorial kernel: the new density is referred to as M-density and the kernel as multimeasurement noise model (MNM).

Denoising

Upside-Down Reinforcement Learning Can Diverge in Stochastic Environments With Episodic Resets

1 code implementation13 May 2022 Miroslav Štrupl, Francesco Faccio, Dylan R. Ashley, Jürgen Schmidhuber, Rupesh Kumar Srivastava

Upside-Down Reinforcement Learning (UDRL) is an approach for solving RL problems that does not require value functions and uses only supervised learning, where the targets for given inputs in a dataset do not change over time.

reinforcement-learning Reinforcement Learning (RL)

ContextVP: Fully Context-Aware Video Prediction

no code implementations ECCV 2018 Wonmin Byeon, Qin Wang, Rupesh Kumar Srivastava, Petros Koumoutsakos

Video prediction models based on convolutional networks, recurrent networks, and their combinations often result in blurry predictions.

Video Prediction

Understanding Locally Competitive Networks

no code implementations5 Oct 2014 Rupesh Kumar Srivastava, Jonathan Masci, Faustino Gomez, Jürgen Schmidhuber

Recently proposed neural network activation functions such as rectified linear, maxout, and local winner-take-all have allowed for faster and more effective training of deep neural architectures on large and complex datasets.

Retrieval

Universal Smoothed Score Functions for Generative Modeling

no code implementations21 Mar 2023 Saeed Saremi, Rupesh Kumar Srivastava, Francis Bach

We consider the problem of generative modeling based on smoothing an unknown density of interest in $\mathbb{R}^d$ using factorial kernels with $M$ independent Gaussian channels with equal noise levels introduced by Saremi and Srivastava (2022).

Cannot find the paper you are looking for? You can Submit a new open access paper.