no code implementations • 24 Jun 2024 • Deyuan Liu, Zhanyue Qin, Hairu Wang, Zhao Yang, Zecheng Wang, Fangying Rong, Qingbin Liu, Yanchao Hao, Xi Chen, Cunhang Fan, Zhao Lv, Zhiying Tu, Dianhui Chu, Bo Li, Dianbo Sui
While large language models (LLMs) excel in many domains, their complexity and scale challenge deployment in resource-limited environments.
no code implementations • 28 Mar 2024 • Deyuan Liu, Zecheng Wang, Bingning Wang, WeiPeng Chen, Chunshan Li, Zhiying Tu, Dianhui Chu, Bo Li, Dianbo Sui
The rapid proliferation of large language models (LLMs) such as GPT-4 and Gemini underscores the intense demand for resources during their training processes, posing significant challenges due to substantial computational and environmental costs.
1 code implementation • 1 Oct 2023 • Zecheng Wang, Che Wang, Zixuan Dong, Keith Ross
Recently, it has been shown that for offline deep reinforcement learning (DRL), pre-training Decision Transformer with a large language corpus can improve downstream performance (Reid et al., 2022).
1 code implementation • 8 Nov 2022 • Yik-Cheung Tam, Jiacheng Xu, Jiakai Zou, Zecheng Wang, Tinglong Liao, Shuhan Yuan
Knowledge cluster classification is boosted from 0. 7924 to 0. 9333 in Recall@1.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +4
1 code implementation • 6 Nov 2022 • Zecheng Wang, Yik-Cheung Tam
SUREALM employs an embedding retriever to search for training sentences in a data store that share similar word history during sequence generation.