Research Article

Training Method and Device of Chemical Industry Chinese Language Model Based on Knowledge Distillation

Table 2

Distillation performance with BERT base.

ModelLayersHiddenAcc (%)F1 (%)

BERT (teacher)676894.1392.52
DistillBILSTM330091.4590.21
BERT PKD376892.8790.66
DistillBERT [36]376891.7789.63
BERT-of-Theseus376893.4391.14
BERT-EMD [37]376893.7791.34
BiLSTM-KD320093.1391.07