Search Results for author: Shiv Shankar

Found 22 papers, 5 papers with code

Surprisingly Easy Hard-Attention for Sequence to Sequence Learning

1 code implementation EMNLP 2018 Shiv Shankar, Siddhant Garg, Sunita Sarawagi

In this paper we show that a simple beam approximation of the joint distribution between attention and output is an easy, accurate, and efficient attention mechanism for sequence to sequence learning.

Hard Attention Image Captioning +2

Optimizing for the Future in Non-Stationary MDPs

1 code implementation ICML 2020 Yash Chandak, Georgios Theocharous, Shiv Shankar, Martha White, Sridhar Mahadevan, Philip S. Thomas

Most reinforcement learning methods are based upon the key assumption that the transition dynamics and reward functions are fixed, that is, the underlying Markov decision process is stationary.

Differential Equation Units: Learning Functional Forms of Activation Functions from Data

1 code implementation6 Sep 2019 MohamadAli Torkamani, Shiv Shankar, Amirmohammad Rooshenas, Phillip Wallis

Most deep neural networks use simple, fixed activation functions, such as sigmoids or rectified linear units, regardless of domain or network structure.

Off-Policy Evaluation for Action-Dependent Non-Stationary Environments

1 code implementation24 Jan 2023 Yash Chandak, Shiv Shankar, Nathaniel D. Bastian, Bruno Castro da Silva, Emma Brunskil, Philip S. Thomas

Methods for sequential decision-making are often built upon a foundational assumption that the underlying decision process is stationary.

counterfactual Counterfactual Reasoning +2

Labeled Memory Networks for Online Model Adaptation

no code implementations5 Jul 2017 Shiv Shankar, Sunita Sarawagi

In this paper, we establish their potential in online adapting a batch trained neural network to domain-relevant labeled data at deployment time.

Few-Shot Learning

Learning Compact Neural Networks Using Ordinary Differential Equations as Activation Functions

no code implementations19 May 2019 MohamadAli Torkamani, Phillip Wallis, Shiv Shankar, Amirmohammad Rooshenas

Most deep neural networks use simple, fixed activation functions, such as sigmoids or rectified linear units, regardless of domain or network structure.

Untapped Potential of Data Augmentation: A Domain Generalization Viewpoint

no code implementations9 Jul 2020 Vihari Piratla, Shiv Shankar

It is believed that by processing augmented inputs in tandem with the original ones, the model learns a more robust set of features which are shared between the original and augmented counterparts.

Data Augmentation Domain Generalization

Bosonic Random Walk Networks for Graph Learning

no code implementations31 Dec 2020 Shiv Shankar, Don Towsley

The development of Graph Neural Networks (GNNs) has led to great progress in machine learning on graph-structured data.

BIG-bench Machine Learning Graph Learning +1

Three-quarter Sibling Regression for Denoising Observational Data

no code implementations31 Dec 2020 Shiv Shankar, Daniel Sheldon, Tao Sun, John Pickering, Thomas G. Dietterich

However, it will remove intrinsic variability if the variables are dependent, and therefore does not apply to many situations, including modeling of species counts that are controlled by common causes.

Denoising regression

High-Confidence Off-Policy (or Counterfactual) Variance Estimation

no code implementations25 Jan 2021 Yash Chandak, Shiv Shankar, Philip S. Thomas

Many sequential decision-making systems leverage data collected using prior policies to propose a new policy.

counterfactual Decision Making +1

Sibling Regression for Generalized Linear Models

no code implementations3 Jul 2021 Shiv Shankar, Daniel Sheldon

Field observations form the basis of many scientific studies, especially in ecological and social sciences.

regression

Adversarial Stein Training for Graph Energy Models

no code implementations30 Aug 2021 Shiv Shankar

Learning distributions over graph-structured data is a challenging task with many applications in biology and chemistry.

Graph Generation

Progressive Fusion for Multimodal Integration

no code implementations1 Sep 2022 Shiv Shankar, Laure Thompson, Madalina Fiterau

In this work, we present an iterative representation refinement approach, called Progressive Fusion, which mitigates the issues with late fusion representations.

Time Series Time Series Prediction

Privacy Aware Experiments without Cookies

no code implementations3 Nov 2022 Shiv Shankar, Ritwik Sinha, Saayan Mitra, Viswanathan Swaminathan, Sridhar Mahadevan, Moumita Sinha

We propose a two-stage experimental design, where the two brands only need to agree on high-level aggregate parameters of the experiment to test the alternate experiences.

Experimental Design valid

Implicit Training of Energy Model for Structure Prediction

no code implementations21 Nov 2022 Shiv Shankar, Vihari Piratla

Most deep learning research has focused on developing new model and training procedures.

Optimization using Parallel Gradient Evaluations on Multiple Parameters

no code implementations6 Feb 2023 Yash Chandak, Shiv Shankar, Venkata Gandikota, Philip S. Thomas, Arya Mazumdar

We propose a first-order method for convex optimization, where instead of being restricted to the gradient from a single parameter, gradients from multiple parameters can be used during each step of gradient descent.

Adaptive Instrument Design for Indirect Experiments

no code implementations5 Dec 2023 Yash Chandak, Shiv Shankar, Vasilis Syrgkanis, Emma Brunskill

Indirect experiments provide a valuable framework for estimating treatment effects in situations where conducting randomized control trials (RCTs) is impractical or unethical.

A/B testing under Interference with Partial Network Information

no code implementations16 Apr 2024 Shiv Shankar, Ritwik Sinha, Yash Chandak, Saayan Mitra, Madalina Fiterau

A/B tests are often required to be conducted on subjects that might have social connections.

Cannot find the paper you are looking for? You can Submit a new open access paper.