Retriever-Augmented Generation, or RAG, is a type of language generation model that combines pre-trained parametric and non-parametric memory for language generation. Specifically, the parametric memory is a pre-trained seq2seq model and the non-parametric memory is a dense vector index of Wikipedia, accessed with a pre-trained neural retriever. For query $x$, Maximum Inner Product Search (MIPS) is used to find the top-K documents $z_{i}$. For final prediction $y$, we treat $z$ as a latent variable and marginalize over seq2seq predictions given different documents.
Source: Retrieval-Augmented Generation for Knowledge-Intensive NLP TasksPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
RAG | 495 | 30.84% |
Retrieval | 353 | 21.99% |
Question Answering | 118 | 7.35% |
Large Language Model | 49 | 3.05% |
Language Modeling | 48 | 2.99% |
Language Modelling | 48 | 2.99% |
Information Retrieval | 33 | 2.06% |
Knowledge Graphs | 24 | 1.50% |
Benchmarking | 22 | 1.37% |