Multi-Stage Document Ranking with BERT

31 Oct 2019 Rodrigo Nogueira Wei Yang Kyunghyun Cho Jimmy Lin

The advent of deep neural networks pre-trained via language modeling tasks has spurred a number of successful applications in natural language processing. This work explores one such popular model, BERT, in the context of document ranking... (read more)

PDF Abstract

Datasets


Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper


METHOD TYPE
Weight Decay
Regularization
Residual Connection
Skip Connections
Adam
Stochastic Optimization
Layer Normalization
Normalization
Softmax
Output Functions
Scaled Dot-Product Attention
Attention Mechanisms
Dropout
Regularization
GELU
Activation Functions
Multi-Head Attention
Attention Modules
Attention Dropout
Regularization
WordPiece
Subword Segmentation
Linear Warmup With Linear Decay
Learning Rate Schedules
Dense Connections
Feedforward Networks
BERT
Language Models