Search Results for author: Eiman Ebrahimi

Found 4 papers, 3 papers with code

Large Graph Convolutional Network Training with GPU-Oriented Data Communication Architecture

1 code implementation4 Mar 2021 Seung Won Min, Kun Wu, Sitao Huang, Mert Hidayetoğlu, JinJun Xiong, Eiman Ebrahimi, Deming Chen, Wen-mei Hwu

In this work, we propose a novel GPU-oriented data communication approach for GCN training, where GPU threads directly access sparse features in host memory through zero-copy accesses without much CPU help.

Recommendation Systems

PyTorch-Direct: Enabling GPU Centric Data Access for Very Large Graph Neural Network Training with Irregular Accesses

1 code implementation20 Jan 2021 Seung Won Min, Kun Wu, Sitao Huang, Mert Hidayetoğlu, JinJun Xiong, Eiman Ebrahimi, Deming Chen, Wen-mei Hwu

While this process accounts for a significant portion of the training time, we find existing GNN implementations using popular deep neural network (DNN) libraries such as PyTorch are limited to a CPU-centric approach for the entire data preparation step.

At-Scale Sparse Deep Neural Network Inference with Efficient GPU Implementation

1 code implementation28 Jul 2020 Mert Hidayetoglu, Carl Pearson, Vikram Sharma Mailthody, Eiman Ebrahimi, JinJun Xiong, Rakesh Nagi, Wen-mei Hwu

This paper presents GPU performance optimization and scaling results for inference models of the Sparse Deep Neural Network Challenge 2020.

Optimizing Multi-GPU Parallelization Strategies for Deep Learning Training

no code implementations30 Jul 2019 Saptadeep Pal, Eiman Ebrahimi, Arslan Zulfiqar, Yaosheng Fu, Victor Zhang, Szymon Migacz, David Nellans, Puneet Gupta

This work explores hybrid parallelization, where each data parallel worker is comprised of more than one device, across which the model dataflow graph (DFG) is split using MP.

Cannot find the paper you are looking for? You can Submit a new open access paper.