Transformers

Siamese Multi-depth Transformer-based Hierarchical Encoder

Introduced by Yang et al. in Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching

SMITH, or Siamese Multi-depth Transformer-based Hierarchical Encoder, is a Transformer-based model for document representation learning and matching. It contains several design choices to adapt self-attention models for long text inputs. For the model pre-training, a masked sentence block language modeling task is used in addition to the original masked word language model task used in BERT, to capture sentence block relations within a document. Given a sequence of sentence block representation, the document level Transformers learn the contextual representation for each sentence block and the final document representation.

Source: Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching

Papers


Paper Code Results Date Stars

Tasks


Task Papers Share
Clustering 1 11.11%
Information Retrieval 1 11.11%
Language Modelling 1 11.11%
Natural Language Understanding 1 11.11%
Playing the Game of 2048 1 11.11%
Question Answering 1 11.11%
Retrieval 1 11.11%
Sentence 1 11.11%
Text Matching 1 11.11%

Components


Component Type
🤖 No Components Found You can add them if they exist; e.g. Mask R-CNN uses RoIAlign

Categories