Self-Supervised Learning

CRISS, or Cross-lingual Retrievial for Iterative Self-Supervised Training (CRISS), is a self-supervised learning method for multilingual sequence generation. CRISS is developed based on the finding that the encoder outputs of multilingual denoising autoencoder can be used as language agnostic representation to retrieve parallel sentence pairs, and training the model on these retrieved sentence pairs can further improve its sentence retrieval and translation capabilities in an iterative manner. Using only unlabeled data from many different languages, CRISS iteratively mines for parallel sentences across languages, trains a new better multilingual model using these mined sentence pairs, mines again for better parallel sentences, and repeats.

Source: Cross-lingual Retrieval for Iterative Self-Supervised Training

Papers


Paper Code Results Date Stars

Tasks


Task Papers Share
Machine Translation 3 18.75%
Translation 3 18.75%
Cross-Lingual Transfer 2 12.50%
NMT 2 12.50%
Abstractive Text Summarization 1 6.25%
XLM-R 1 6.25%
Zero-Shot Cross-Lingual Transfer 1 6.25%
Retrieval 1 6.25%
Sentence 1 6.25%

Components


Component Type
🤖 No Components Found You can add them if they exist; e.g. Mask R-CNN uses RoIAlign

Categories