Spaces:
Sleeping
Sleeping
Delete configs/callbacks/model_checkpoint.yaml with huggingface_hub
Browse files
configs/callbacks/model_checkpoint.yaml
DELETED
@@ -1,19 +0,0 @@
|
|
1 |
-
# https://pytorch-lightning.readthedocs.io/en/latest/api/pytorch_lightning.callbacks.ModelCheckpoint.html
|
2 |
-
|
3 |
-
# Save the model periodically by monitoring a quantity.
|
4 |
-
# Look at the above link for more detailed information.
|
5 |
-
model_checkpoint:
|
6 |
-
_target_: lightning.pytorch.callbacks.ModelCheckpoint
|
7 |
-
dirpath: ${paths.output_dir} # directory to save the model file
|
8 |
-
filename: "checkpoints/epoch_{epoch:03d}" # checkpoint filename
|
9 |
-
monitor: ${eval:'"val/loss" if ${data.train_val_test_split}[1] else "train/loss"'} # name of the logged metric which determines when model is improving
|
10 |
-
verbose: False # verbosity mode
|
11 |
-
save_last: True # additionally always save an exact copy of the last checkpoint to a file last.ckpt
|
12 |
-
save_top_k: 1 # save k best models (determined by above metric)
|
13 |
-
mode: "min" # "max" means higher metric value is better, can be also "min"
|
14 |
-
auto_insert_metric_name: False # when True, the checkpoints filenames will contain the metric name
|
15 |
-
save_weights_only: False # if True, then only the model’s weights will be saved
|
16 |
-
every_n_train_steps: null # number of training steps between checkpoints
|
17 |
-
train_time_interval: null # checkpoints are monitored at the specified time interval
|
18 |
-
every_n_epochs: null # number of epochs between checkpoints
|
19 |
-
save_on_train_epoch_end: null # whether to run checkpointing at the end of the training epoch or the end of validation
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|