Search Results for author: Liu Ke

Found 4 papers, 0 papers with code

Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference

no code implementations10 Mar 2023 Haiyang Huang, Newsha Ardalani, Anna Sun, Liu Ke, Hsien-Hsin S. Lee, Anjali Sridhar, Shruti Bhosale, Carole-Jean Wu, Benjamin Lee

We propose three optimization techniques to mitigate sources of inefficiencies, namely (1) Dynamic gating, (2) Expert Buffering, and (3) Expert load balancing.

Language Modelling Machine Translation

Data Leakage via Access Patterns of Sparse Features in Deep Learning-based Recommendation Systems

no code implementations12 Dec 2022 Hanieh Hashemi, Wenjie Xiong, Liu Ke, Kiwan Maeng, Murali Annavaram, G. Edward Suh, Hsien-Hsin S. Lee

This paper explores the private information that may be learned by tracking a recommendation model's sparse feature access patterns.

Recommendation Systems

Neural Network-Inspired Analog-to-Digital Conversion to Achieve Super-Resolution with Low-Precision RRAM Devices

no code implementations28 Nov 2019 Weidong Cao, Liu Ke, Ayan Chakrabarti, Xuan Zhang

Recent works propose neural network- (NN-) inspired analog-to-digital converters (NNADCs) and demonstrate their great potentials in many emerging applications.

Quantization Robust Design +1

AxTrain: Hardware-Oriented Neural Network Training for Approximate Inference

no code implementations21 May 2018 Xin He, Liu Ke, Wenyan Lu, Guihai Yan, Xuan Zhang

The intrinsic error tolerance of neural network (NN) makes approximate computing a promising technique to improve the energy efficiency of NN inference.

Cannot find the paper you are looking for? You can Submit a new open access paper.