!6592 bugfix bert config

Merge pull request !6592 from yoonlee666/master
pull/6592/MERGE
mindspore-ci-bot 5 years ago committed by Gitee
commit 10add158f8

@ -57,6 +57,7 @@ large: BERT-NEZHA(a Chinese pretrained language model developed by Huawei, which
Functional Relative Posetional Encoding as an effective positional encoding scheme). Functional Relative Posetional Encoding as an effective positional encoding scheme).
''' '''
if cfg.bert_network == 'base': if cfg.bert_network == 'base':
cfg.batch_size = 64
bert_net_cfg = BertConfig( bert_net_cfg = BertConfig(
seq_length=128, seq_length=128,
vocab_size=21128, vocab_size=21128,
@ -75,6 +76,7 @@ if cfg.bert_network == 'base':
compute_type=mstype.float16 compute_type=mstype.float16
) )
if cfg.bert_network == 'nezha': if cfg.bert_network == 'nezha':
cfg.batch_size = 96
bert_net_cfg = BertConfig( bert_net_cfg = BertConfig(
seq_length=128, seq_length=128,
vocab_size=21128, vocab_size=21128,
@ -93,6 +95,7 @@ if cfg.bert_network == 'nezha':
compute_type=mstype.float16 compute_type=mstype.float16
) )
if cfg.bert_network == 'large': if cfg.bert_network == 'large':
cfg.batch_size = 24
bert_net_cfg = BertConfig( bert_net_cfg = BertConfig(
seq_length=512, seq_length=512,
vocab_size=30522, vocab_size=30522,

Loading…
Cancel
Save