Search Results for author: Rahul Vashisht

Found 4 papers, 3 papers with code

On the Learning Dynamics of Attention Networks

1 code implementation25 Jul 2023 Rahul Vashisht, Harish G. Ramaswamy

Attention models are typically learned by optimizing one of three standard loss functions that are variously called -- soft attention, hard attention, and latent variable marginal likelihood (LVML) attention.

Hard Attention

On the Interpretability of Attention Networks

1 code implementation30 Dec 2022 Lakshmi Narayan Pandey, Rahul Vashisht, Harish G. Ramaswamy

In trained models with an attention mechanism, the outputs of an intermediate module that encodes the segment of input responsible for the output is often used as a way to peek into the `reasoning` of the network.

Image Captioning

Using noise resilience for ranking generalization of deep neural networks

1 code implementation16 Dec 2020 Depen Morwani, Rahul Vashisht, Harish G. Ramaswamy

Recent papers have shown that sufficiently overparameterized neural networks can perfectly fit even random labels.

Position

Structural Health Monitoring of Cantilever Beam, a Case Study -- Using Bayesian Neural Network AND Deep Learning

no code implementations17 Aug 2019 Rahul Vashisht, H. Viji, T. Sundararajan, D. Mohankumar, S. Sumitra

Deep learning architectures like CNN (Convolutional neural network) and LSTM(Long Short Term Memory) are good candidates for representation learning from high dimensional data.

BIG-bench Machine Learning Cantilever Beam +1

Cannot find the paper you are looking for? You can Submit a new open access paper.