no code implementations • 7 Jan 2025 • Aman Gupta, Shao Tang, Qingquan Song, Sirou Zhu, Jiwoo Hong, Ankan Saha, Viral Gupta, Noah Lee, Eunki Kim, Jason Zhu, Natesh Pillai, S. Sathiya Keerthi
In this paper, we argue that, for DAAs the reward (function) shape matters.
no code implementations • 12 Aug 2021 • Aman Gupta, Rohan Ramanath, Jun Shi, Anika Ramachandran, Sirou Zhou, Mingzhou Zhou, S. Sathiya Keerthi
Over-parameterized deep networks trained using gradient-based optimizers are a popular choice for solving classification and ranking problems.
1 code implementation • 9 Mar 2021 • Rohan Ramanath, S. Sathiya Keerthi, Yao Pan, Konstantin Salomatin, Kinjal Basu
We develop an intuition guided by theoretical and empirical analysis to show that when these instances follow certain structures, a large majority of the projections lie on vertices of the polytopes.
1 code implementation • 3 Mar 2020 • Saurav Manchanda, Khoa Doan, Pranjul Yadav, S. Sathiya Keerthi
This paper addresses the classic problem of regression, which involves the inductive learning of a map, $y=f(x, z)$, $z$ denoting noise, $f:\mathbb{R}^n\times \mathbb{R}^k \rightarrow \mathbb{R}^m$.
no code implementations • 7 Feb 2020 • Saurav Manchanda, Pranjul Yadav, Khoa Doan, S. Sathiya Keerthi
We present an experimental analysis on the historical logs of a major display advertising platform (https://www. criteo. com/).
1 code implementation • 30 May 2019 • Karan Aggarwal, Matthieu Kirchmeyer, Pranjul Yadav, S. Sathiya Keerthi, Patrick Gallinari
Such a real world situation is best represented using an implicit model in which an extra noise vector, $z$ is included with $x$ as input.
no code implementations • 22 Mar 2018 • Jennifer Ortiz, Magdalena Balazinska, Johannes Gehrke, S. Sathiya Keerthi
In the database field, query optimization remains a difficult problem.
no code implementations • 1 Feb 2018 • Chien-Chih Wang, Kent Loong Tan, Chun-Ting Chen, Yu-Hsiang Lin, S. Sathiya Keerthi, Dhruv Mahajan, S. Sundararajan, Chih-Jen Lin
First, to reduce the communication cost, we propose a diagonalization method such that an approximate Newton direction can be obtained without communication between machines.
no code implementations • 15 Nov 2017 • Dhruv Mahajan, Vivek Gupta, S. Sathiya Keerthi, Sellamanickam Sundararajan, Shravan Narayanamurthy, Rahul Kidambi
We also demonstrate their usefulness in making design choices such as the number of classifiers in the ensemble and the size of a subset of data used for training that is needed to achieve a certain value of generalization error.
no code implementations • ICML 2017 • Si Si, huan zhang, S. Sathiya Keerthi, Dhruv Mahajan, Inderjit S. Dhillon, Cho-Jui Hsieh
In this paper, we study the gradient boosted decision trees (GBDT) when the output space is high dimensional and sparse.
no code implementations • 22 Apr 2017 • Michał Dereziński, Dhruv Mahajan, S. Sathiya Keerthi, S. V. N. Vishwanathan, Markus Weimer
We propose Batch-Expansion Training (BET), a framework for running a batch optimizer on a gradually expanding dataset.
no code implementations • 6 Nov 2015 • S. Sathiya Keerthi, Tobias Schnabel, Rajiv Khanna
In a recent paper, Levy and Goldberg pointed out an interesting connection between prediction-based word embedding models and count models based on pointwise mutual information.
no code implementations • 18 May 2014 • Dhruv Mahajan, S. Sathiya Keerthi, S. Sundararajan
This paper concerns the distributed training of nonlinear kernel machines on Map-Reduce.
no code implementations • 18 May 2014 • Dhruv Mahajan, S. Sathiya Keerthi, S. Sundararajan
In this paper we design a distributed algorithm for $l_1$ regularization that is much better suited for such systems than existing algorithms.
no code implementations • 10 Nov 2013 • Vinod Nair, Rahul Kidambi, Sundararajan Sellamanickam, S. Sathiya Keerthi, Johannes Gehrke, Vijay Narayanan
We consider the problem of quantitatively evaluating missing value imputation algorithms.
no code implementations • 9 Nov 2013 • Rahul Kidambi, Vinod Nair, Sundararajan Sellamanickam, S. Sathiya Keerthi
In this paper we propose a structured output approach for missing value imputation that also incorporates domain constraints.
no code implementations • 4 Nov 2013 • Dhruv Mahajan, S. Sathiya Keerthi, S. Sundararajan, Leon Bottou
The method has strong convergence properties.
no code implementations • 31 Oct 2013 • Dhruv Mahajan, Nikunj Agrawal, S. Sathiya Keerthi, S. Sundararajan, Leon Bottou
In this paper we give a novel approach to the distributed training of linear classifiers (involving smooth losses and L2 regularization) that is designed to reduce the total communication costs.