Search Results for author: Rongqian Zhao

Found 1 papers, 0 papers with code

LocMoE: A Low-overhead MoE for Large Language Model Training

no code implementations25 Jan 2024 Jing Li, Zhijie Sun, Xuan He, Li Zeng, Yi Lin, Entong Li, Binfan Zheng, Rongqian Zhao, Xin Chen

However, the performance of MoE is limited by load imbalance and high latency of All-To-All communication, along with relatively redundant computation owing to large expert capacity.

Language Modelling Large Language Model

Cannot find the paper you are looking for? You can Submit a new open access paper.