no code implementations • 6 Sep 2023 • Joshua R. Waite, Jiale Feng, Riley Tavassoli, Laura Harris, Sin Yong Tan, Subhadeep Chakraborty, Soumik Sarkar
The increasing concern surrounding gun violence in the United States has led to a focus on developing systems to improve public safety.
1 code implementation • 6 Dec 2021 • Zhanhong Jiang, Xian Yeow Lee, Sin Yong Tan, Kai Liang Tan, Aditya Balu, Young M. Lee, Chinmay Hegde, Soumik Sarkar
We propose a novel policy gradient method for multi-agent reinforcement learning, which leverages two different variance-reduction techniques and does not require large batches over iterations.
Multi-agent Reinforcement Learning Policy Gradient Methods +3
no code implementations • 9 Apr 2021 • Sin Yong Tan, Homagni Saha, Margarite Jacoby, Gregor P. Henze, Soumik Sarkar
Clustering is an unsupervised learning technique that is useful when working with a large volume of unlabeled data.
1 code implementation • 2 Mar 2021 • Yasaman Esfandiari, Sin Yong Tan, Zhanhong Jiang, Aditya Balu, Ethan Herron, Chinmay Hegde, Soumik Sarkar
Inspired by ideas from continual learning, we propose Cross-Gradient Aggregation (CGA), a novel decentralized learning algorithm where (i) each agent aggregates cross-gradient information, i. e., derivatives of its model with respect to its neighbors' datasets, and (ii) updates its model using a projected gradient based on quadratic programming (QP).
no code implementations • 21 Oct 2020 • Aditya Balu, Zhanhong Jiang, Sin Yong Tan, Chinmay Hedge, Young M Lee, Soumik Sarkar
In this context, we propose and analyze a novel decentralized deep learning algorithm where the agents interact over a fixed communication topology (without a central server).
2 code implementations • 11 Aug 2020 • Tryambak Gangopadhyay, Sin Yong Tan, Zhanhong Jiang, Rui Meng, Soumik Sarkar
Accurate interpretation of such prediction outcomes from a machine learning model that explicitly captures temporal correlations can significantly benefit the domain experts.
no code implementations • 15 Oct 2019 • Zhanhong Jiang, Aditya Balu, Sin Yong Tan, Young M. Lee, Chinmay Hegde, Soumik Sarkar
In this paper, we investigate the popular deep learning optimization routine, Adam, from the perspective of statistical moments.