Search Results for author: Vijay Subramanian

Found 9 papers, 2 papers with code

Backward and Forward Inference in Interacting Independent-Cascade Processes: A Scalable and Convergent Message-Passing Approach

no code implementations29 Oct 2023 Nouman Khan, Kangle Mu, Mehrdad Moharrami, Vijay Subramanian

We study the problems of estimating the past and future evolutions of two diffusion processes that spread concurrently on a network.

Bayesian Learning of Optimal Policies in Markov Decision Processes with Countably Infinite State-Space

no code implementations NeurIPS 2023 Saghar Adler, Vijay Subramanian

Models of many real-life applications, such as queuing models of communication networks or computing systems, have a countably infinite state-space.

Thompson Sampling

Rarest-First with Probabilistic-Mode-Suppression

no code implementations1 Nov 2022 Nouman Khan, Mehrdad Moharrami, Vijay Subramanian

In this work, we propose a tunable piece-selection policy that minimizes this (undesirable) requisite by combining the (work-conserving but not stabilizing) rarest-first protocol with only an appropriate share of the (non-work conserving and stabilizing) mode-suppression protocol.

OpenGridGym: An Open-Source AI-Friendly Toolkit for Distribution Market Simulation

1 code implementation6 Mar 2022 Rayan El Helou, Kiyeob Lee, Dongqi Wu, Le Xie, Srinivas Shakkottai, Vijay Subramanian

This paper presents OpenGridGym, an open-source Python-based package that allows for seamless integration of distribution market simulation with state-of-the-art artificial intelligence (AI) decision-making algorithms.

Decision Making

Learning a Discrete Set of Optimal Allocation Rules in a Queueing System with Unknown Service Rate

no code implementations4 Feb 2022 Saghar Adler, Mehrdad Moharrami, Vijay Subramanian

Our goal is to design a dispatching policy that maximizes the long-term average reward for the dispatcher based on observing only the arrival times and the state of the system at each arrival that reflects a realistic sampling of such systems.

Blocking

Decentralized Cooperative Reinforcement Learning with Hierarchical Information Structure

no code implementations1 Nov 2021 Hsu Kao, Chen-Yu Wei, Vijay Subramanian

For the bandit setting, we propose a hierarchical bandit algorithm that achieves a near-optimal gap-independent regret of $\widetilde{\mathcal{O}}(\sqrt{ABT})$ and a near-optimal gap-dependent regret of $\mathcal{O}(\log(T))$, where $A$ and $B$ are the numbers of actions of the leader and the follower, respectively, and $T$ is the number of steps.

Multi-agent Reinforcement Learning Multi-Armed Bandits +3

Common Information based Approximate State Representations in Multi-Agent Reinforcement Learning

no code implementations25 Oct 2021 Hsu Kao, Vijay Subramanian

Due to information asymmetry, finding optimal policies for Decentralized Partially Observable Markov Decision Processes (Dec-POMDPs) is hard with the complexity growing doubly exponentially in the horizon length.

Multi-agent Reinforcement Learning reinforcement-learning +1

Empirical Policy Evaluation with Supergraphs

no code implementations18 Feb 2020 Daniel Vial, Vijay Subramanian

We devise and analyze algorithms for the empirical policy evaluation problem in reinforcement learning.

reinforcement-learning Reinforcement Learning +1

On the role of clustering in Personalized PageRank estimation

1 code implementation4 Jun 2017 Daniel Vial, Vijay Subramanian

We then show that the common underlying graph can be leveraged to efficiently and jointly estimate PPR for many pairs, rather than treating each pair separately using the primitive algorithm.

Social and Information Networks

Cannot find the paper you are looking for? You can Submit a new open access paper.