Research Article

Named Entity Recognition in Chinese Medical Literature Using Pretraining Models

Table 1

Parameters, pretraining tasks, and corpora of pretraining models.

Pretraining modelLHAPretraining taskPretraining corpora

BERT [18]1276812Masked Language Model, NSPBooks Corpus, Wikipedia
BERT-WWM [20]1276812Whole Word Masking, NSPWikipedia
BERT-WWM-EXT [20]1276812Whole Word Masking, NSPGeneral data (Baike, News, and QA), Wikipedia
ERNIE [21]1276812Phrase-level and entity-level masking, NSPChinese Wikipedia, Baidu Baike, News, and Tieba
ERNIE-tiny [22]3102412Phrase-level and entity-level masking, NSPChinese Wikipedia, Baidu Baike, News, and Tieba
RoBERTa [23]1276812Dynamic maskingBooks Corpus, Wikipedia