Search Results for author: WenGuang Chen

Found 9 papers, 3 papers with code

A Comprehensive Survey on Distributed Training of Graph Neural Networks

no code implementations10 Nov 2022 Haiyang Lin, Mingyu Yan, Xiaochun Ye, Dongrui Fan, Shirui Pan, WenGuang Chen, Yuan Xie

This situation poses a considerable challenge for newcomers, hindering their ability to grasp a comprehensive understanding of the workflows, computational patterns, communication strategies, and optimization techniques employed in distributed GNN training.

Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss

no code implementations20 Jul 2022 Daning Cheng, WenGuang Chen

Based on the model's resilience to computational noise, model quantization is important for compressing models and improving computing speed.

Quantization

Quantization in Layer's Input is Matter

no code implementations10 Feb 2022 Daning Cheng, WenGuang Chen

In this paper, we will show that the quantization in layer's input is more important than parameters' quantization for loss function.

Quantization

GraphTheta: A Distributed Graph Neural Network Learning System With Flexible Training Strategy

1 code implementation21 Apr 2021 Yongchao Liu, Houyi Li, Guowei Zhang, Xintan Zeng, Yongyong Li, Bin Huang, Peng Zhang, Zhao Li, Xiaowei Zhu, Changhua He, WenGuang Chen

Herein, we present GraphTheta, the first distributed and scalable graph learning system built upon vertex-centric distributed graph processing with neural network operators implemented as user-defined functions.

Graph Learning

AIPerf: Automated machine learning as an AI-HPC benchmark

1 code implementation17 Aug 2020 Zhixiang Ren, Yongheng Liu, Tianhui Shi, Lei Xie, Yue Zhou, Jidong Zhai, Youhui Zhang, Yunquan Zhang, WenGuang Chen

The de facto HPC benchmark LINPACK can not reflect AI computing power and I/O performance without representative workload.

AutoML Benchmarking +1

Bridging the Gap Between Neural Networks and Neuromorphic Hardware with A Neural Network Compiler

no code implementations15 Nov 2017 Yu Ji, Youhui Zhang, WenGuang Chen, Yuan Xie

Different from developing neural networks (NNs) for general-purpose processors, the development for NN chips usually faces with some hardware-specific restrictions, such as limited precision of network signals and parameters, constrained computation scale, and limited types of non-linear functions.

SaberLDA: Sparsity-Aware Learning of Topic Models on GPUs

no code implementations8 Oct 2016 Kaiwei Li, Jianfei Chen, WenGuang Chen, Jun Zhu

Latent Dirichlet Allocation (LDA) is a popular tool for analyzing discrete count data such as text and images.

Topic Models

WarpLDA: a Cache Efficient O(1) Algorithm for Latent Dirichlet Allocation

no code implementations29 Oct 2015 Jianfei Chen, Kaiwei Li, Jun Zhu, WenGuang Chen

We then develop WarpLDA, an LDA sampler which achieves both the best O(1) time complexity per token and the best O(K) scope of random access.

Cannot find the paper you are looking for? You can Submit a new open access paper.