mtasic85 commited on
Commit
afa8f6e
·
1 Parent(s): bee417a

global_batch_size: 256; micro_batch_size: 2

Browse files
Files changed (1) hide show
  1. scripts/pretrain-core-model.yaml +2 -2
scripts/pretrain-core-model.yaml CHANGED
@@ -67,8 +67,8 @@ train:
67
  global_batch_size: 256
68
 
69
  # Number of samples per data-parallel rank (type: int, default: 4)
70
- micro_batch_size: 4
71
- # micro_batch_size: 2
72
  # micro_batch_size: 1
73
 
74
  # Number of iterations with learning rate warmup active (type: int, default: 2000)
 
67
  global_batch_size: 256
68
 
69
  # Number of samples per data-parallel rank (type: int, default: 4)
70
+ # micro_batch_size: 4
71
+ micro_batch_size: 2
72
  # micro_batch_size: 1
73
 
74
  # Number of iterations with learning rate warmup active (type: int, default: 2000)