fastText embeddings exploit subword information to construct word embeddings. Representations are learnt of character $n$-grams, and words represented as the sum of the $n$-gram vectors. This extends the word2vec type models with subword information. This helps the embeddings understand suffixes and prefixes. Once a word is represented using character $n$-grams, a skipgram model is trained to learn the embeddings.
Source: Enriching Word Vectors with Subword InformationPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Text Classification | 30 | 10.14% |
General Classification | 28 | 9.46% |
Sentiment Analysis | 19 | 6.42% |
Named Entity Recognition | 13 | 4.39% |
Word Similarity | 11 | 3.72% |
Language Modelling | 10 | 3.38% |
Machine Learning | 7 | 2.36% |
Question Answering | 7 | 2.36% |
Machine Translation | 6 | 2.03% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |