Search Results for author: Tharun Medini

Found 13 papers, 5 papers with code

Distributed SLIDE: Enabling Training Large Neural Networks on Low Bandwidth and Simple CPU-Clusters via Model Parallelism and Sparsity

no code implementations29 Jan 2022 Minghao Yan, Nicholas Meisburger, Tharun Medini, Anshumali Shrivastava

We show that with reduced communication, due to sparsity, we can train close to a billion parameter model on simple 4-16 core CPU nodes connected by basic low bandwidth interconnect.

Cloud Computing

IRLI: Iterative Re-partitioning for Learning to Index

no code implementations17 Mar 2021 Gaurav Gupta, Tharun Medini, Anshumali Shrivastava, Alexander J Smola

Neural models have transformed the fundamental information retrieval problem of mapping a query to a giant set of items.

Information Retrieval Multi-Label Classification +1

A Truly Constant-time Distribution-aware Negative Sampling

no code implementations1 Jan 2021 Shabnam Daghaghi, Tharun Medini, Beidi Chen, Mengnan Zhao, Anshumali Shrivastava

Softmax classifiers with a very large number of classes naturally occur in many applications such as natural language processing and information retrieval.

Information Retrieval Retrieval

A Tale of Two Efficient and Informative Negative Sampling Distributions

no code implementations31 Dec 2020 Shabnam Daghaghi, Tharun Medini, Nicholas Meisburger, Beidi Chen, Mengnan Zhao, Anshumali Shrivastava

Unfortunately, due to the dynamically updated parameters and data samples, there is no sampling scheme that is provably adaptive and samples the negative classes efficiently.

Information Retrieval Retrieval +1

SOLAR: Sparse Orthogonal Learned and Random Embeddings

no code implementations ICLR 2021 Tharun Medini, Beidi Chen, Anshumali Shrivastava

The label vectors are random, sparse, and near-orthogonal by design, while the query vectors are learned and sparse.

Multi-Label Classification

SDM-Net: A Simple and Effective Model for Generalized Zero-Shot Learning

no code implementations10 Sep 2019 Shabnam Daghaghi, Tharun Medini, Anshumali Shrivastava

Zero-Shot Learning (ZSL) is a classification task where we do not have even a single training labeled example from a set of unseen classes.

Descriptive General Classification +3

Extreme Classification in Log Memory

no code implementations9 Oct 2018 Qixuan Huang, Yiqiu Wang, Tharun Medini, Anshumali Shrivastava

With MACH we can train ODP dataset with 100, 000 classes and 400, 000 features on a single Titan X GPU, with the classification accuracy of 19. 28%, which is the best-reported accuracy on this dataset.

Classification General Classification

Mimicking actions is a good strategy for beginners: Fast Reinforcement Learning with Expert Action Sequences

no code implementations27 Sep 2018 Tharun Medini, Anshumali Shrivastava

Imitation Learning is the task of mimicking the behavior of an expert player in a Reinforcement Learning(RL) Environment to enhance the training of a fresh agent (called novice) beginning from scratch.

Atari Games Imitation Learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.