BERT large model is a pretrained model on English language which uses a Masked Language Modeling (MLM for short). It has 24 encoder layers. The model was first introduced in this paper and subsequently released in this repository.
Pre-requisite: BERT model, BERT for text summarization
Table of content:
What is BERT Large?A world before BERT: Word2Vec and GloVe
2.1 Elimination of Polysemy: ELMo and ULMFiT
2.2 And now BERTA look under BERT Large's architecture
3.1 Text pre-processing
3.2 Pre-tr...
Published on October 18, 2021 05:45