Stop Thinking, Just Do!

Sungsoo Kim's Blog

Retrieval-Augmented Language Model Pre-Training

tagsTags

21 November 2023


Article Source


Retrieval-Augmented Language Model Pre-Training

Abstract

REALM is just the BERT based language model (LM) augmented with a retriever. At pretrain time, MLM task is used. Pretraining is unsupervised where salient spans (entities, dates) are masked. Finetuning is done for Open QA task: predict (start, end) spans given question and retrieved docs z. Retriever is supported by knowledge corpus and MIPS. The paper showed great results for OpenQA task. It formed a basis for many future studies: RAG, RETRO, ATLAS.

In this I will talk about REALM training details and also its evaluation on open question answering datasets.


comments powered by Disqus