Search Results for author: Xunyu Zhu

Found 5 papers, 1 papers with code

Distilling Mathematical Reasoning Capabilities into Small Language Models

no code implementations22 Jan 2024 Xunyu Zhu, Jian Li, Yong liu, Can Ma, Weiping Wang

This work addresses the challenge of democratizing advanced Large Language Models (LLMs) by compressing their mathematical reasoning capabilities into sub-billion parameter Small Language Models (SLMs) without compromising performance.

Mathematical Reasoning

A Survey on Model Compression for Large Language Models

no code implementations15 Aug 2023 Xunyu Zhu, Jian Li, Yong liu, Can Ma, Weiping Wang

As these challenges become increasingly pertinent, the field of model compression has emerged as a pivotal research area to alleviate these limitations.

Benchmarking Knowledge Distillation +2

Robust Neural Architecture Search

no code implementations6 Apr 2023 Xunyu Zhu, Jian Li, Yong liu, Weiping Wang

Neural Architectures Search (NAS) becomes more and more popular over these years.

Image Classification Neural Architecture Search

Operation-level Progressive Differentiable Architecture Search

1 code implementation11 Feb 2023 Xunyu Zhu, Jian Li, Yong liu, Weiping Wang

It can effectively alleviate the unfair competition between operations during the search phase of DARTS by offsetting the inherent unfair advantage of the skip connection over other operations.

Neural Architecture Search

Improving Differentiable Architecture Search via Self-Distillation

no code implementations11 Feb 2023 Xunyu Zhu, Jian Li, Yong liu, Weiping Wang

Differentiable Architecture Search (DARTS) is a simple yet efficient Neural Architecture Search (NAS) method.

Neural Architecture Search

Cannot find the paper you are looking for? You can Submit a new open access paper.