no code implementations • 22 Jan 2024 • Xunyu Zhu, Jian Li, Yong liu, Can Ma, Weiping Wang
This work addresses the challenge of democratizing advanced Large Language Models (LLMs) by compressing their mathematical reasoning capabilities into sub-billion parameter Small Language Models (SLMs) without compromising performance.
no code implementations • 15 Aug 2023 • Xunyu Zhu, Jian Li, Yong liu, Can Ma, Weiping Wang
As these challenges become increasingly pertinent, the field of model compression has emerged as a pivotal research area to alleviate these limitations.
no code implementations • 6 Apr 2023 • Xunyu Zhu, Jian Li, Yong liu, Weiping Wang
Neural Architectures Search (NAS) becomes more and more popular over these years.
1 code implementation • 11 Feb 2023 • Xunyu Zhu, Jian Li, Yong liu, Weiping Wang
It can effectively alleviate the unfair competition between operations during the search phase of DARTS by offsetting the inherent unfair advantage of the skip connection over other operations.
no code implementations • 11 Feb 2023 • Xunyu Zhu, Jian Li, Yong liu, Weiping Wang
Differentiable Architecture Search (DARTS) is a simple yet efficient Neural Architecture Search (NAS) method.