Update scheduler configs
Browse files
README.md
CHANGED
|
@@ -254,8 +254,18 @@ gradient_checkpointing: false
|
|
| 254 |
# stop training after this many evaluation losses have increased in a row
|
| 255 |
# https://huggingface.co/transformers/v4.2.2/_modules/transformers/trainer_callback.html#EarlyStoppingCallback
|
| 256 |
early_stopping_patience: 3
|
| 257 |
-
|
| 258 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 259 |
# specify optimizer
|
| 260 |
optimizer:
|
| 261 |
# specify weight decay
|
|
|
|
| 254 |
# stop training after this many evaluation losses have increased in a row
|
| 255 |
# https://huggingface.co/transformers/v4.2.2/_modules/transformers/trainer_callback.html#EarlyStoppingCallback
|
| 256 |
early_stopping_patience: 3
|
| 257 |
+
|
| 258 |
+
# specify a scheduler and kwargs to use with the optimizer
|
| 259 |
+
lr_scheduler: # 'one_cycle' | 'log_sweep' | empty for cosine
|
| 260 |
+
lr_scheduler_kwargs:
|
| 261 |
+
|
| 262 |
+
# for one_cycle optim
|
| 263 |
+
lr_div_factor: # learning rate div factor
|
| 264 |
+
|
| 265 |
+
# for log_sweep optim
|
| 266 |
+
log_sweep_min_lr:
|
| 267 |
+
log_sweep_max_lr:
|
| 268 |
+
|
| 269 |
# specify optimizer
|
| 270 |
optimizer:
|
| 271 |
# specify weight decay
|