Search Results for author: Guoqing Jiang

Found 2 papers, 0 papers with code

SKDBERT: Compressing BERT via Stochastic Knowledge Distillation

no code implementations26 Nov 2022 Zixiang Ding, Guoqing Jiang, Shuai Zhang, Lin Guo, Wei Lin

In this paper, we propose Stochastic Knowledge Distillation (SKD) to obtain compact BERT-style language model dubbed SKDBERT.

Knowledge Distillation Language Modelling

Understanding Why Neural Networks Generalize Well Through GSNR of Parameters

no code implementations ICLR 2020 Jinlong Liu, Guoqing Jiang, Yunzhi Bai, Ting Chen, Huayan Wang

As deep neural networks (DNNs) achieve tremendous success across many application domains, researchers tried to explore in many aspects on why they generalize well.

Cannot find the paper you are looking for? You can Submit a new open access paper.