no code implementations • ICML 2020 • Vidyashankar Sivakumar, Zhiwei Steven Wu, Arindam Banerjee
Bandit learning algorithms typically involve the balance of exploration and exploitation.
no code implementations • NeurIPS 2019 • Arindam Banerjee, Qilong Gu, Vidyashankar Sivakumar, Zhiwei Steven Wu
We also discuss stochastic process based forms of J-L, RIP, and sketching, to illustrate the generality of the results.
no code implementations • ICML 2017 • Vidyashankar Sivakumar, Arindam Banerjee
In this work we consider the problem of linear quantile regression in high dimensions where the number of predictor variables is much higher than the number of samples available for parameter estimation.
no code implementations • 17 Jun 2016 • Nicholas Johnson, Vidyashankar Sivakumar, Arindam Banerjee
The goal in such a problem is to minimize the (pseudo) regret which is the difference between the total expected loss of the algorithm and the total expected loss of the best fixed vector in hindsight.
no code implementations • NeurIPS 2015 • Vidyashankar Sivakumar, Arindam Banerjee, Pradeep K. Ravikumar
In contrast, for the sub-exponential setting, we show that the sample complexity and the estimation error will depend on the exponential width of the corresponding sets, and the analysis holds for any norm.
no code implementations • NeurIPS 2014 • Arindam Banerjee, Sheng Chen, Farideh Fazayeli, Vidyashankar Sivakumar
Analysis of non-asymptotic estimation error and structured statistical recovery based on norm regularized regression, such as Lasso, needs to consider four aspects: the norm, the loss function, the design matrix, and the noise model.
no code implementations • 1 Sep 2014 • Andre R. Goncalves, Puja Das, Soumyadeep Chatterjee, Vidyashankar Sivakumar, Fernando J. Von Zuben, Arindam Banerjee
We illustrate the effectiveness of the proposed model on a variety of synthetic and benchmark datasets for regression and classification.