1 code implementation • 9 Oct 2024 • Hao Jiang, Qi Liu, Rui Li, Shengyu Ye, Shijin Wang
In this work, we propose a new conversational framework that comprehensively integrates these information sources, collect data to train our models and evaluate their performance.
1 code implementation • 25 Sep 2024 • Yifei Liu, Jicheng Wen, Yang Wang, Shengyu Ye, Li Lyna Zhang, Ting Cao, Cheng Li, Mao Yang
Due to the redundancy in LLM weights, recent research has focused on pushing weight-only quantization to extremely low-bit (even down to 2 bits).