SMITH, or Siamese Multi-depth Transformer-based Hierarchical Encoder, is a Transformer-based model for document representation learning and matching. It contains several design choices to adapt self-attention models for long text inputs. For the model pre-training, a masked sentence block language modeling task is used in addition to the original masked word language model task used in BERT, to capture sentence block relations within a document. Given a sequence of sentence block representation, the document level Transformers learn the contextual representation for each sentence block and the final document representation.
Source: Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document MatchingPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Clustering | 1 | 11.11% |
Information Retrieval | 1 | 11.11% |
Language Modelling | 1 | 11.11% |
Natural Language Understanding | 1 | 11.11% |
Playing the Game of 2048 | 1 | 11.11% |
Question Answering | 1 | 11.11% |
Retrieval | 1 | 11.11% |
Sentence | 1 | 11.11% |
Text Matching | 1 | 11.11% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |