Search Results for author: Peiqi Yin

Found 3 papers, 1 papers with code

SPT: Fine-Tuning Transformer-based Language Models Efficiently with Sparsification

1 code implementation16 Dec 2023 Yuntao Gui, Xiao Yan, Peiqi Yin, Han Yang, James Cheng

Thus, we design the sparse MHA module, which computes and stores only large attention weights to reduce memory consumption, and the routed FFN module, which dynamically activates a subset of model parameters for each token to reduce computation cost.

Quantization

DGI: Easy and Efficient Inference for GNNs

no code implementations28 Nov 2022 Peiqi Yin, Xiao Yan, Jinjing Zhou, Qiang Fu, Zhenkun Cai, James Cheng, Bo Tang, Minjie Wang

In this paper, we develop Deep Graph Inference (DGI) -- a system for easy and efficient GNN model inference, which automatically translates the training code of a GNN model for layer-wise execution.

Cannot find the paper you are looking for? You can Submit a new open access paper.