中文 预训练 ELECTRA 模型: 基于对抗学习 pretrain Chinese Model
中文 预训练 ELECTREA 模型: 基于对抗学习 pretrain Chinese Model
code Repost from google official code: https://github.com/google-research/electra
具体使用说明:参考 官方链接
electra-tiny google-drive
electra-tiny baidu-pan code:rs99
electra-tiny
metric | value |
---|---|
disc_accuracy | 0.95093095 |
disc_auc | 0.9762006 |
disc_loss | 0.14071295 |
disc_precision | 0.8018275 |
disc_recall | 0.6088053 |
loss | 9.516352 |
masked_lm_accuracy | 0.46732807 |
masked_lm_loss | 2.8209455 |
sampled_masked_lm_accuracy | 0.3504382 |
The model are trained on CLUE 10G Chinese Corpus with 1M-steps
注:only use pretrained electra-tiny with layer-wise learning rate decay without any distilaltion、data-augmentation. learning rate is set to 1e-4 for each task and run 10-epochs. (According to official results, the results may have large variance)
AFQMC | TNEWS | IFLYTEK | CMNLI | WSC | CSL | |
---|---|---|---|---|---|---|
Metrics | Acc | Acc | Acc | Acc | Acc | Acc |
ELECTRA-tiny | 70.319 | 54.280 | 53.538 | 73.745 | 64.336 | 78.700 |
Roberta-tiny | 69.904 | 54.150 | 56.808 | 74.037 | 64.336 | 74.133 |
注: