fastText embeddings exploit subword information to construct word embeddings. Representations are learnt of character $n$-grams, and words represented as the sum of the $n$-gram vectors. This extends the word2vec type models with subword information. This helps the embeddings understand suffixes and prefixes. Once a word is represented using character $n$-grams, a skipgram model is trained to learn the embeddings.
Source: Enriching Word Vectors with Subword InformationPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Text Classification | 33 | 7.86% |
General Classification | 28 | 6.67% |
Sentence | 24 | 5.71% |
Sentiment Analysis | 22 | 5.24% |
Classification | 16 | 3.81% |
Named Entity Recognition (NER) | 15 | 3.57% |
Language Modelling | 12 | 2.86% |
Word Similarity | 11 | 2.62% |
Clustering | 7 | 1.67% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |