no code implementations • 9 Jan 2024 • Qinyi Luo, Penghan Wang, Wei zhang, Fan Lai, Jiachen Mao, Xiaohan Wei, Jun Song, Wei-Yu Tsai, Shuai Yang, Yuxi Hu, Xuehai Qian
Huge embedding tables in modern Deep Learning Recommender Models (DLRM) require prohibitively large memory during training and inference.
no code implementations • 20 Sep 2020 • Jiachen Mao, Mehdi Jafari, Audun Botterud
This paper introduces a mathematical formulation of energy storage systems into a generation capacity expansion framework to evaluate the role of energy storage in the decarbonization of distributed power systems.
no code implementations • 13 Sep 2019 • Qing Yang, Jiachen Mao, Zuoguan Wang, Hai Li
In addition to conventional compression techniques, e. g., weight pruning and quantization, removing unimportant activations can reduce the amount of data communication and the computation cost.
no code implementations • 7 Jan 2019 • Linghao Song, Jiachen Mao, Youwei Zhuo, Xuehai Qian, Hai Li, Yiran Chen
In this paper, inspired by recent work in machine learning systems, we propose a solution HyPar to determine layer-wise parallelism for deep neural network training with an array of DNN accelerators.