Search Results for author: Gaochen Dong

Found 2 papers, 0 papers with code

Blockwise Compression of Transformer-based Models without Retraining

no code implementations4 Apr 2023 Gaochen Dong, Wei Chen

This method mitigates data distribution deviation caused by quantization, eliminating the requirement for retraining.

Quantization

Block-wise Bit-Compression of Transformer-based Models

no code implementations16 Mar 2023 Gaochen Dong, Wei Chen

With the popularity of the recent Transformer-based models represented by BERT, GPT-3 and ChatGPT, there has been state-of-the-art performance in a range of natural language processing tasks.

Cannot find the paper you are looking for? You can Submit a new open access paper.