Search Results for author: Ruikang Liu

Found 5 papers, 4 papers with code

A Simple Linear Patch Revives Layer-Pruned Large Language Models

no code implementations30 May 2025 Xinrui Chen, Haoli Bai, Tao Yuan, Ruikang Liu, Kang Zhao, Xianzhi Yu, Lu Hou, Tian Guan, Yonghong He, Chun Yuan

With only 5K samples, the retained performance of LinearPatch can be further boosted to 95. 16% within 30 minutes on a single computing card.

Knowledge Distillation Question Answering

Quantization Hurts Reasoning? An Empirical Study on Quantized Reasoning Models

1 code implementation7 Apr 2025 Ruikang Liu, Yuxuan Sun, Manyi Zhang, Haoli Bai, Xianzhi Yu, Tiezheng Yu, Chun Yuan, Lu Hou

In addition, strategically scaling the model sizes or reasoning steps can effectively enhance the performance.

Math Quantization

FlatQuant: Flatness Matters for LLM Quantization

1 code implementation12 Oct 2024 Yuxuan Sun, Ruikang Liu, Haoli Bai, Han Bao, Kang Zhao, Yuening Li, Jiaxin Hu, Xianzhi Yu, Lu Hou, Chun Yuan, Xin Jiang, Wulong Liu, Jun Yao

In this paper, we propose FlatQuant (Fast and Learnable Affine Transformation), a new post-training quantization approach to enhance flatness of weights and activations.

Quantization

IntactKV: Improving Large Language Model Quantization by Keeping Pivot Tokens Intact

2 code implementations2 Mar 2024 Ruikang Liu, Haoli Bai, Haokun Lin, Yuening Li, Han Gao, Zhengzhuo Xu, Lu Hou, Jun Yao, Chun Yuan

Such outliers are found to allocate most of the attention scores on initial tokens of input, termed as pivot tokens, which are crucial to the performance of quantized LLMs.

Language Modeling Language Modelling +2

Learning Imbalanced Data with Vision Transformers

1 code implementation CVPR 2023 Zhengzhuo Xu, Ruikang Liu, Shuo Yang, Zenghao Chai, Chun Yuan

In this paper, we systematically investigate the ViTs' performance in LTR and propose LiVT to train ViTs from scratch only with LT data.

Long-tail Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.