|
|
|
@ -57,6 +57,7 @@ large: BERT-NEZHA(a Chinese pretrained language model developed by Huawei, which
|
|
|
|
|
Functional Relative Posetional Encoding as an effective positional encoding scheme).
|
|
|
|
|
'''
|
|
|
|
|
if cfg.bert_network == 'base':
|
|
|
|
|
cfg.batch_size = 64
|
|
|
|
|
bert_net_cfg = BertConfig(
|
|
|
|
|
seq_length=128,
|
|
|
|
|
vocab_size=21128,
|
|
|
|
@ -75,6 +76,7 @@ if cfg.bert_network == 'base':
|
|
|
|
|
compute_type=mstype.float16
|
|
|
|
|
)
|
|
|
|
|
if cfg.bert_network == 'nezha':
|
|
|
|
|
cfg.batch_size = 96
|
|
|
|
|
bert_net_cfg = BertConfig(
|
|
|
|
|
seq_length=128,
|
|
|
|
|
vocab_size=21128,
|
|
|
|
@ -93,6 +95,7 @@ if cfg.bert_network == 'nezha':
|
|
|
|
|
compute_type=mstype.float16
|
|
|
|
|
)
|
|
|
|
|
if cfg.bert_network == 'large':
|
|
|
|
|
cfg.batch_size = 24
|
|
|
|
|
bert_net_cfg = BertConfig(
|
|
|
|
|
seq_length=512,
|
|
|
|
|
vocab_size=30522,
|
|
|
|
|