CRISS, or Cross-lingual Retrievial for Iterative Self-Supervised Training (CRISS), is a self-supervised learning method for multilingual sequence generation. CRISS is developed based on the finding that the encoder outputs of multilingual denoising autoencoder can be used as language agnostic representation to retrieve parallel sentence pairs, and training the model on these retrieved sentence pairs can further improve its sentence retrieval and translation capabilities in an iterative manner. Using only unlabeled data from many different languages, CRISS iteratively mines for parallel sentences across languages, trains a new better multilingual model using these mined sentence pairs, mines again for better parallel sentences, and repeats.
Source: Cross-lingual Retrieval for Iterative Self-Supervised TrainingPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Machine Translation | 3 | 16.67% |
Translation | 3 | 16.67% |
Cross-Lingual Transfer | 2 | 11.11% |
Decoder | 2 | 11.11% |
NMT | 2 | 11.11% |
Abstractive Text Summarization | 1 | 5.56% |
XLM-R | 1 | 5.56% |
Zero-Shot Cross-Lingual Transfer | 1 | 5.56% |
Retrieval | 1 | 5.56% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |