no code implementations • NAACL 2022 • Marzieh Tahaei, Ella Charlaix, Vahid Nia, Ali Ghodsi, Mehdi Rezagholizadeh
We push the limits of state-of-the-art Transformer-based pre-trained language model compression using Kronecker decomposition.
Knowledge Distillation Language Modelling +1