2 code implementations • 30 Mar 2023 • Nicholas Meisburger, Vihan Lakshman, Benito Geordie, Joshua Engels, David Torres Ramos, Pratik Pranav, Benjamin Coleman, Benjamin Meisburger, Shubh Gupta, Yashwanth Adunukota, Tharun Medini, Anshumali Shrivastava
Efficient large-scale neural network training and inference on commodity CPU hardware is of immense practical significance in democratizing deep learning (DL) capabilities.
Ranked #2 on Node Classification on Yelp-Fraud
no code implementations • 29 Jan 2022 • Minghao Yan, Nicholas Meisburger, Tharun Medini, Anshumali Shrivastava
We show that with reduced communication, due to sparsity, we can train close to a billion parameter model on simple 4-16 core CPU nodes connected by basic low bandwidth interconnect.
no code implementations • 17 Mar 2021 • Gaurav Gupta, Tharun Medini, Anshumali Shrivastava, Alexander J Smola
Neural models have transformed the fundamental information retrieval problem of mapping a query to a giant set of items.
no code implementations • 1 Jan 2021 • Shabnam Daghaghi, Tharun Medini, Beidi Chen, Mengnan Zhao, Anshumali Shrivastava
Softmax classifiers with a very large number of classes naturally occur in many applications such as natural language processing and information retrieval.
no code implementations • 31 Dec 2020 • Shabnam Daghaghi, Tharun Medini, Nicholas Meisburger, Beidi Chen, Mengnan Zhao, Anshumali Shrivastava
Unfortunately, due to the dynamically updated parameters and data samples, there is no sampling scheme that is provably adaptive and samples the negative classes efficiently.
no code implementations • ICLR 2021 • Tharun Medini, Beidi Chen, Anshumali Shrivastava
The label vectors are random, sparse, and near-orthogonal by design, while the query vectors are learned and sparse.
1 code implementation • NeurIPS 2019 • Tharun Medini, Qixuan Huang, Yiqiu Wang, Vijai Mohan, Anshumali Shrivastava
Our largest model has 6. 4 billion parameters and trains in less than 35 hours on a single p3. 16x machine.
1 code implementation • 10 Oct 2019 • Gaurav Gupta, Minghao Yan, Benjamin Coleman, Bryce Kille, R. A. Leo Elworth, Tharun Medini, Todd Treangen, Anshumali Shrivastava
Interestingly, it is a count-min sketch type arrangement of a membership testing utility (Bloom Filter in our case).
1 code implementation • 7 Oct 2019 • Gaurav Gupta, Benjamin Coleman, Tharun Medini, Vijai Mohan, Anshumali Shrivastava
A simple array of Bloom Filters can achieve that.
no code implementations • 10 Sep 2019 • Shabnam Daghaghi, Tharun Medini, Anshumali Shrivastava
Zero-Shot Learning (ZSL) is a classification task where we do not have even a single training labeled example from a set of unseen classes.
3 code implementations • 7 Mar 2019 • Beidi Chen, Tharun Medini, James Farwell, Sameh Gobriel, Charlie Tai, Anshumali Shrivastava
On the same CPU hardware, SLIDE is over 10x faster than TF.
no code implementations • 9 Oct 2018 • Qixuan Huang, Yiqiu Wang, Tharun Medini, Anshumali Shrivastava
With MACH we can train ODP dataset with 100, 000 classes and 400, 000 features on a single Titan X GPU, with the classification accuracy of 19. 28%, which is the best-reported accuracy on this dataset.
no code implementations • 27 Sep 2018 • Tharun Medini, Anshumali Shrivastava
Imitation Learning is the task of mimicking the behavior of an expert player in a Reinforcement Learning(RL) Environment to enhance the training of a fresh agent (called novice) beginning from scratch.