Transformers

Chinese Pre-trained Unbalanced Transformer

Introduced by Shao et al. in CPT: A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation

CPT, or Chinese Pre-trained Unbalanced Transformer, is a pre-trained unbalanced Transformer for Chinese natural language understanding (NLU) and natural language generation (NLG) tasks. CPT consists of three parts: a shared encoder, an understanding decoder, and a generation decoder. Two specific decoders with a shared encoder are pre-trained with masked language modeling (MLM) and denoising auto-encoding (DAE) tasks, respectively. With the partially shared architecture and multi-task pre-training, CPT can (1) learn specific knowledge of both NLU or NLG tasks with two decoders and (2) be fine-tuned flexibly that fully exploits the potential of the model. Two specific decoders with a shared encoder are pre-trained with masked language modeling (MLM) and denoising auto-encoding (DAE) tasks, respectively. With the partially shared architecture and multi-task pre-training, CPT can (1) learn specific knowledge of both NLU or NLG tasks with two decoders and (2) be fine-tuned flexibly that fully exploits the potential of the model.

Source: CPT: A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation

Papers


Paper Code Results Date Stars

Tasks


Task Papers Share
Denoising 1 25.00%
Language Modelling 1 25.00%
Natural Language Understanding 1 25.00%
Text Generation 1 25.00%

Components


Component Type
🤖 No Components Found You can add them if they exist; e.g. Mask R-CNN uses RoIAlign

Categories