"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/trainer_state.json +483 -6
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/trainer_state.json +1590 -6
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630226071.8984504/events.out.tfevents.1630226071.cc93b136ebf5.1086.99 +3 -0
- model-bin/finetune/base/log/1630226502.2753584/events.out.tfevents.1630226502.cc93b136ebf5.1086.101 +3 -0
- model-bin/finetune/base/log/1630226942.7046182/events.out.tfevents.1630226942.cc93b136ebf5.1086.103 +3 -0
- model-bin/finetune/base/log/1630227371.5174663/events.out.tfevents.1630227371.cc93b136ebf5.1086.105 +3 -0
- model-bin/finetune/base/log/1630227801.075265/events.out.tfevents.1630227801.cc93b136ebf5.1086.107 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630226071.cc93b136ebf5.1086.98 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630226502.cc93b136ebf5.1086.100 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630226942.cc93b136ebf5.1086.102 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630227371.cc93b136ebf5.1086.104 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630227801.cc93b136ebf5.1086.106 +3 -0
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:425263039b2e17958ef29df62ab9ab8878277ad6c4ead7176bda8210b2ed3fe6
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b2376bae8ab9408163ddcedcab213b016840a625804da3cbf1526010b8abc141
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80390ad562fb737219eaef200a69cb1cb5c9e3d1cb1722a1fdddeff6f6402b7d
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d760abfe85cbbe513ba97c70e9515e7e485b75c78f8b50035843503942c2409
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfefb4c24229027a9d0175af6026b93ceb96966446b698af9ac66780b0583d09
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
-
"best_metric": 0.
|
3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -284139,11 +284139,488 @@
|
|
284139 |
"eval_steps_per_second": 0.668,
|
284140 |
"eval_wer": 0.17565191913272782,
|
284141 |
"step": 145228
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
284142 |
}
|
284143 |
],
|
284144 |
-
"max_steps":
|
284145 |
"num_train_epochs": 5000,
|
284146 |
-
"total_flos": 4.
|
284147 |
"trial_name": null,
|
284148 |
"trial_params": null
|
284149 |
}
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.16716790141568252,
|
3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
|
4 |
+
"epoch": 1164.0,
|
5 |
+
"global_step": 145602,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
284139 |
"eval_steps_per_second": 0.668,
|
284140 |
"eval_wer": 0.17565191913272782,
|
284141 |
"step": 145228
|
284142 |
+
},
|
284143 |
+
{
|
284144 |
+
"epoch": 1171.02,
|
284145 |
+
"learning_rate": 7.670872374798062e-06,
|
284146 |
+
"loss": 0.3162,
|
284147 |
+
"step": 145230
|
284148 |
+
},
|
284149 |
+
{
|
284150 |
+
"epoch": 1171.06,
|
284151 |
+
"learning_rate": 7.670791599353798e-06,
|
284152 |
+
"loss": 0.268,
|
284153 |
+
"step": 145235
|
284154 |
+
},
|
284155 |
+
{
|
284156 |
+
"epoch": 1171.1,
|
284157 |
+
"learning_rate": 7.670710823909532e-06,
|
284158 |
+
"loss": 0.3148,
|
284159 |
+
"step": 145240
|
284160 |
+
},
|
284161 |
+
{
|
284162 |
+
"epoch": 1171.14,
|
284163 |
+
"learning_rate": 7.670630048465268e-06,
|
284164 |
+
"loss": 0.3431,
|
284165 |
+
"step": 145245
|
284166 |
+
},
|
284167 |
+
{
|
284168 |
+
"epoch": 1171.18,
|
284169 |
+
"learning_rate": 7.670549273021002e-06,
|
284170 |
+
"loss": 0.5562,
|
284171 |
+
"step": 145250
|
284172 |
+
},
|
284173 |
+
{
|
284174 |
+
"epoch": 1171.22,
|
284175 |
+
"learning_rate": 7.670468497576738e-06,
|
284176 |
+
"loss": 0.8706,
|
284177 |
+
"step": 145255
|
284178 |
+
},
|
284179 |
+
{
|
284180 |
+
"epoch": 1171.26,
|
284181 |
+
"learning_rate": 7.670387722132472e-06,
|
284182 |
+
"loss": 0.283,
|
284183 |
+
"step": 145260
|
284184 |
+
},
|
284185 |
+
{
|
284186 |
+
"epoch": 1171.3,
|
284187 |
+
"learning_rate": 7.670306946688208e-06,
|
284188 |
+
"loss": 0.3143,
|
284189 |
+
"step": 145265
|
284190 |
+
},
|
284191 |
+
{
|
284192 |
+
"epoch": 1171.34,
|
284193 |
+
"learning_rate": 7.670226171243944e-06,
|
284194 |
+
"loss": 0.2854,
|
284195 |
+
"step": 145270
|
284196 |
+
},
|
284197 |
+
{
|
284198 |
+
"epoch": 1171.38,
|
284199 |
+
"learning_rate": 7.670145395799678e-06,
|
284200 |
+
"loss": 0.5725,
|
284201 |
+
"step": 145275
|
284202 |
+
},
|
284203 |
+
{
|
284204 |
+
"epoch": 1171.42,
|
284205 |
+
"learning_rate": 7.670064620355414e-06,
|
284206 |
+
"loss": 0.9977,
|
284207 |
+
"step": 145280
|
284208 |
+
},
|
284209 |
+
{
|
284210 |
+
"epoch": 1171.46,
|
284211 |
+
"learning_rate": 7.669983844911148e-06,
|
284212 |
+
"loss": 0.2628,
|
284213 |
+
"step": 145285
|
284214 |
+
},
|
284215 |
+
{
|
284216 |
+
"epoch": 1171.5,
|
284217 |
+
"learning_rate": 7.669903069466884e-06,
|
284218 |
+
"loss": 0.2448,
|
284219 |
+
"step": 145290
|
284220 |
+
},
|
284221 |
+
{
|
284222 |
+
"epoch": 1171.54,
|
284223 |
+
"learning_rate": 7.669822294022618e-06,
|
284224 |
+
"loss": 0.3178,
|
284225 |
+
"step": 145295
|
284226 |
+
},
|
284227 |
+
{
|
284228 |
+
"epoch": 1171.58,
|
284229 |
+
"learning_rate": 7.669741518578354e-06,
|
284230 |
+
"loss": 0.6503,
|
284231 |
+
"step": 145300
|
284232 |
+
},
|
284233 |
+
{
|
284234 |
+
"epoch": 1171.62,
|
284235 |
+
"learning_rate": 7.669660743134088e-06,
|
284236 |
+
"loss": 0.9847,
|
284237 |
+
"step": 145305
|
284238 |
+
},
|
284239 |
+
{
|
284240 |
+
"epoch": 1171.66,
|
284241 |
+
"learning_rate": 7.669579967689824e-06,
|
284242 |
+
"loss": 0.2887,
|
284243 |
+
"step": 145310
|
284244 |
+
},
|
284245 |
+
{
|
284246 |
+
"epoch": 1171.7,
|
284247 |
+
"learning_rate": 7.669499192245558e-06,
|
284248 |
+
"loss": 0.2481,
|
284249 |
+
"step": 145315
|
284250 |
+
},
|
284251 |
+
{
|
284252 |
+
"epoch": 1171.74,
|
284253 |
+
"learning_rate": 7.669418416801294e-06,
|
284254 |
+
"loss": 0.2989,
|
284255 |
+
"step": 145320
|
284256 |
+
},
|
284257 |
+
{
|
284258 |
+
"epoch": 1171.78,
|
284259 |
+
"learning_rate": 7.669337641357028e-06,
|
284260 |
+
"loss": 0.6053,
|
284261 |
+
"step": 145325
|
284262 |
+
},
|
284263 |
+
{
|
284264 |
+
"epoch": 1171.82,
|
284265 |
+
"learning_rate": 7.669256865912764e-06,
|
284266 |
+
"loss": 0.9806,
|
284267 |
+
"step": 145330
|
284268 |
+
},
|
284269 |
+
{
|
284270 |
+
"epoch": 1171.86,
|
284271 |
+
"learning_rate": 7.669176090468498e-06,
|
284272 |
+
"loss": 0.3248,
|
284273 |
+
"step": 145335
|
284274 |
+
},
|
284275 |
+
{
|
284276 |
+
"epoch": 1171.9,
|
284277 |
+
"learning_rate": 7.669095315024234e-06,
|
284278 |
+
"loss": 0.3173,
|
284279 |
+
"step": 145340
|
284280 |
+
},
|
284281 |
+
{
|
284282 |
+
"epoch": 1171.94,
|
284283 |
+
"learning_rate": 7.66901453957997e-06,
|
284284 |
+
"loss": 0.3376,
|
284285 |
+
"step": 145345
|
284286 |
+
},
|
284287 |
+
{
|
284288 |
+
"epoch": 1171.98,
|
284289 |
+
"learning_rate": 7.668933764135704e-06,
|
284290 |
+
"loss": 0.6947,
|
284291 |
+
"step": 145350
|
284292 |
+
},
|
284293 |
+
{
|
284294 |
+
"epoch": 1172.0,
|
284295 |
+
"eval_loss": 0.47769203782081604,
|
284296 |
+
"eval_runtime": 39.6454,
|
284297 |
+
"eval_samples_per_second": 21.112,
|
284298 |
+
"eval_steps_per_second": 0.681,
|
284299 |
+
"eval_wer": 0.18728573557907288,
|
284300 |
+
"step": 145352
|
284301 |
+
},
|
284302 |
+
{
|
284303 |
+
"epoch": 1162.02,
|
284304 |
+
"learning_rate": 7.66885298869144e-06,
|
284305 |
+
"loss": 0.3284,
|
284306 |
+
"step": 145355
|
284307 |
+
},
|
284308 |
+
{
|
284309 |
+
"epoch": 1162.06,
|
284310 |
+
"learning_rate": 7.668772213247174e-06,
|
284311 |
+
"loss": 0.2665,
|
284312 |
+
"step": 145360
|
284313 |
+
},
|
284314 |
+
{
|
284315 |
+
"epoch": 1162.1,
|
284316 |
+
"learning_rate": 7.66869143780291e-06,
|
284317 |
+
"loss": 0.266,
|
284318 |
+
"step": 145365
|
284319 |
+
},
|
284320 |
+
{
|
284321 |
+
"epoch": 1162.14,
|
284322 |
+
"learning_rate": 7.668610662358644e-06,
|
284323 |
+
"loss": 0.3346,
|
284324 |
+
"step": 145370
|
284325 |
+
},
|
284326 |
+
{
|
284327 |
+
"epoch": 1162.18,
|
284328 |
+
"learning_rate": 7.66852988691438e-06,
|
284329 |
+
"loss": 0.6964,
|
284330 |
+
"step": 145375
|
284331 |
+
},
|
284332 |
+
{
|
284333 |
+
"epoch": 1162.22,
|
284334 |
+
"learning_rate": 7.668449111470114e-06,
|
284335 |
+
"loss": 0.8126,
|
284336 |
+
"step": 145380
|
284337 |
+
},
|
284338 |
+
{
|
284339 |
+
"epoch": 1162.26,
|
284340 |
+
"learning_rate": 7.66836833602585e-06,
|
284341 |
+
"loss": 0.2484,
|
284342 |
+
"step": 145385
|
284343 |
+
},
|
284344 |
+
{
|
284345 |
+
"epoch": 1162.3,
|
284346 |
+
"learning_rate": 7.668287560581584e-06,
|
284347 |
+
"loss": 0.2748,
|
284348 |
+
"step": 145390
|
284349 |
+
},
|
284350 |
+
{
|
284351 |
+
"epoch": 1162.34,
|
284352 |
+
"learning_rate": 7.66820678513732e-06,
|
284353 |
+
"loss": 0.3347,
|
284354 |
+
"step": 145395
|
284355 |
+
},
|
284356 |
+
{
|
284357 |
+
"epoch": 1162.38,
|
284358 |
+
"learning_rate": 7.668126009693053e-06,
|
284359 |
+
"loss": 0.7334,
|
284360 |
+
"step": 145400
|
284361 |
+
},
|
284362 |
+
{
|
284363 |
+
"epoch": 1162.42,
|
284364 |
+
"learning_rate": 7.66804523424879e-06,
|
284365 |
+
"loss": 0.7622,
|
284366 |
+
"step": 145405
|
284367 |
+
},
|
284368 |
+
{
|
284369 |
+
"epoch": 1162.46,
|
284370 |
+
"learning_rate": 7.667964458804525e-06,
|
284371 |
+
"loss": 0.2724,
|
284372 |
+
"step": 145410
|
284373 |
+
},
|
284374 |
+
{
|
284375 |
+
"epoch": 1162.5,
|
284376 |
+
"learning_rate": 7.66788368336026e-06,
|
284377 |
+
"loss": 0.2608,
|
284378 |
+
"step": 145415
|
284379 |
+
},
|
284380 |
+
{
|
284381 |
+
"epoch": 1162.54,
|
284382 |
+
"learning_rate": 7.667802907915995e-06,
|
284383 |
+
"loss": 0.3223,
|
284384 |
+
"step": 145420
|
284385 |
+
},
|
284386 |
+
{
|
284387 |
+
"epoch": 1162.58,
|
284388 |
+
"learning_rate": 7.66772213247173e-06,
|
284389 |
+
"loss": 0.6718,
|
284390 |
+
"step": 145425
|
284391 |
+
},
|
284392 |
+
{
|
284393 |
+
"epoch": 1162.62,
|
284394 |
+
"learning_rate": 7.667641357027465e-06,
|
284395 |
+
"loss": 0.8369,
|
284396 |
+
"step": 145430
|
284397 |
+
},
|
284398 |
+
{
|
284399 |
+
"epoch": 1162.66,
|
284400 |
+
"learning_rate": 7.6675605815832e-06,
|
284401 |
+
"loss": 0.2615,
|
284402 |
+
"step": 145435
|
284403 |
+
},
|
284404 |
+
{
|
284405 |
+
"epoch": 1162.7,
|
284406 |
+
"learning_rate": 7.667479806138935e-06,
|
284407 |
+
"loss": 0.2988,
|
284408 |
+
"step": 145440
|
284409 |
+
},
|
284410 |
+
{
|
284411 |
+
"epoch": 1162.74,
|
284412 |
+
"learning_rate": 7.66739903069467e-06,
|
284413 |
+
"loss": 0.3295,
|
284414 |
+
"step": 145445
|
284415 |
+
},
|
284416 |
+
{
|
284417 |
+
"epoch": 1162.78,
|
284418 |
+
"learning_rate": 7.667318255250405e-06,
|
284419 |
+
"loss": 0.7126,
|
284420 |
+
"step": 145450
|
284421 |
+
},
|
284422 |
+
{
|
284423 |
+
"epoch": 1162.82,
|
284424 |
+
"learning_rate": 7.66723747980614e-06,
|
284425 |
+
"loss": 0.8154,
|
284426 |
+
"step": 145455
|
284427 |
+
},
|
284428 |
+
{
|
284429 |
+
"epoch": 1162.86,
|
284430 |
+
"learning_rate": 7.667156704361875e-06,
|
284431 |
+
"loss": 0.2407,
|
284432 |
+
"step": 145460
|
284433 |
+
},
|
284434 |
+
{
|
284435 |
+
"epoch": 1162.9,
|
284436 |
+
"learning_rate": 7.66707592891761e-06,
|
284437 |
+
"loss": 0.2759,
|
284438 |
+
"step": 145465
|
284439 |
+
},
|
284440 |
+
{
|
284441 |
+
"epoch": 1162.94,
|
284442 |
+
"learning_rate": 7.666995153473345e-06,
|
284443 |
+
"loss": 0.2949,
|
284444 |
+
"step": 145470
|
284445 |
+
},
|
284446 |
+
{
|
284447 |
+
"epoch": 1162.98,
|
284448 |
+
"learning_rate": 7.66691437802908e-06,
|
284449 |
+
"loss": 0.7172,
|
284450 |
+
"step": 145475
|
284451 |
+
},
|
284452 |
+
{
|
284453 |
+
"epoch": 1163.0,
|
284454 |
+
"eval_loss": 0.3563687801361084,
|
284455 |
+
"eval_runtime": 41.8928,
|
284456 |
+
"eval_samples_per_second": 19.98,
|
284457 |
+
"eval_steps_per_second": 0.645,
|
284458 |
+
"eval_wer": 0.17516339869281045,
|
284459 |
+
"step": 145477
|
284460 |
+
},
|
284461 |
+
{
|
284462 |
+
"epoch": 1163.02,
|
284463 |
+
"learning_rate": 7.666833602584815e-06,
|
284464 |
+
"loss": 0.3148,
|
284465 |
+
"step": 145480
|
284466 |
+
},
|
284467 |
+
{
|
284468 |
+
"epoch": 1163.06,
|
284469 |
+
"learning_rate": 7.66675282714055e-06,
|
284470 |
+
"loss": 0.2606,
|
284471 |
+
"step": 145485
|
284472 |
+
},
|
284473 |
+
{
|
284474 |
+
"epoch": 1163.1,
|
284475 |
+
"learning_rate": 7.666672051696285e-06,
|
284476 |
+
"loss": 0.3123,
|
284477 |
+
"step": 145490
|
284478 |
+
},
|
284479 |
+
{
|
284480 |
+
"epoch": 1163.14,
|
284481 |
+
"learning_rate": 7.66659127625202e-06,
|
284482 |
+
"loss": 0.3515,
|
284483 |
+
"step": 145495
|
284484 |
+
},
|
284485 |
+
{
|
284486 |
+
"epoch": 1163.18,
|
284487 |
+
"learning_rate": 7.666510500807755e-06,
|
284488 |
+
"loss": 0.7378,
|
284489 |
+
"step": 145500
|
284490 |
+
},
|
284491 |
+
{
|
284492 |
+
"epoch": 1163.22,
|
284493 |
+
"learning_rate": 7.66642972536349e-06,
|
284494 |
+
"loss": 0.8376,
|
284495 |
+
"step": 145505
|
284496 |
+
},
|
284497 |
+
{
|
284498 |
+
"epoch": 1163.26,
|
284499 |
+
"learning_rate": 7.666348949919225e-06,
|
284500 |
+
"loss": 0.3011,
|
284501 |
+
"step": 145510
|
284502 |
+
},
|
284503 |
+
{
|
284504 |
+
"epoch": 1163.3,
|
284505 |
+
"learning_rate": 7.66626817447496e-06,
|
284506 |
+
"loss": 0.3181,
|
284507 |
+
"step": 145515
|
284508 |
+
},
|
284509 |
+
{
|
284510 |
+
"epoch": 1163.34,
|
284511 |
+
"learning_rate": 7.666187399030695e-06,
|
284512 |
+
"loss": 0.4141,
|
284513 |
+
"step": 145520
|
284514 |
+
},
|
284515 |
+
{
|
284516 |
+
"epoch": 1163.38,
|
284517 |
+
"learning_rate": 7.66610662358643e-06,
|
284518 |
+
"loss": 0.7503,
|
284519 |
+
"step": 145525
|
284520 |
+
},
|
284521 |
+
{
|
284522 |
+
"epoch": 1163.42,
|
284523 |
+
"learning_rate": 7.666025848142165e-06,
|
284524 |
+
"loss": 0.8573,
|
284525 |
+
"step": 145530
|
284526 |
+
},
|
284527 |
+
{
|
284528 |
+
"epoch": 1163.46,
|
284529 |
+
"learning_rate": 7.6659450726979e-06,
|
284530 |
+
"loss": 0.283,
|
284531 |
+
"step": 145535
|
284532 |
+
},
|
284533 |
+
{
|
284534 |
+
"epoch": 1163.5,
|
284535 |
+
"learning_rate": 7.665864297253635e-06,
|
284536 |
+
"loss": 0.2771,
|
284537 |
+
"step": 145540
|
284538 |
+
},
|
284539 |
+
{
|
284540 |
+
"epoch": 1163.54,
|
284541 |
+
"learning_rate": 7.66578352180937e-06,
|
284542 |
+
"loss": 0.5021,
|
284543 |
+
"step": 145545
|
284544 |
+
},
|
284545 |
+
{
|
284546 |
+
"epoch": 1163.58,
|
284547 |
+
"learning_rate": 7.665702746365107e-06,
|
284548 |
+
"loss": 0.7615,
|
284549 |
+
"step": 145550
|
284550 |
+
},
|
284551 |
+
{
|
284552 |
+
"epoch": 1163.62,
|
284553 |
+
"learning_rate": 7.66562197092084e-06,
|
284554 |
+
"loss": 0.8149,
|
284555 |
+
"step": 145555
|
284556 |
+
},
|
284557 |
+
{
|
284558 |
+
"epoch": 1163.66,
|
284559 |
+
"learning_rate": 7.665541195476576e-06,
|
284560 |
+
"loss": 0.2333,
|
284561 |
+
"step": 145560
|
284562 |
+
},
|
284563 |
+
{
|
284564 |
+
"epoch": 1163.7,
|
284565 |
+
"learning_rate": 7.66546042003231e-06,
|
284566 |
+
"loss": 0.293,
|
284567 |
+
"step": 145565
|
284568 |
+
},
|
284569 |
+
{
|
284570 |
+
"epoch": 1163.74,
|
284571 |
+
"learning_rate": 7.665379644588046e-06,
|
284572 |
+
"loss": 0.3133,
|
284573 |
+
"step": 145570
|
284574 |
+
},
|
284575 |
+
{
|
284576 |
+
"epoch": 1163.78,
|
284577 |
+
"learning_rate": 7.66529886914378e-06,
|
284578 |
+
"loss": 0.693,
|
284579 |
+
"step": 145575
|
284580 |
+
},
|
284581 |
+
{
|
284582 |
+
"epoch": 1163.82,
|
284583 |
+
"learning_rate": 7.665218093699516e-06,
|
284584 |
+
"loss": 0.9545,
|
284585 |
+
"step": 145580
|
284586 |
+
},
|
284587 |
+
{
|
284588 |
+
"epoch": 1163.86,
|
284589 |
+
"learning_rate": 7.66513731825525e-06,
|
284590 |
+
"loss": 0.3008,
|
284591 |
+
"step": 145585
|
284592 |
+
},
|
284593 |
+
{
|
284594 |
+
"epoch": 1163.9,
|
284595 |
+
"learning_rate": 7.665056542810986e-06,
|
284596 |
+
"loss": 0.3013,
|
284597 |
+
"step": 145590
|
284598 |
+
},
|
284599 |
+
{
|
284600 |
+
"epoch": 1163.94,
|
284601 |
+
"learning_rate": 7.66497576736672e-06,
|
284602 |
+
"loss": 0.538,
|
284603 |
+
"step": 145595
|
284604 |
+
},
|
284605 |
+
{
|
284606 |
+
"epoch": 1163.98,
|
284607 |
+
"learning_rate": 7.664894991922456e-06,
|
284608 |
+
"loss": 0.7369,
|
284609 |
+
"step": 145600
|
284610 |
+
},
|
284611 |
+
{
|
284612 |
+
"epoch": 1164.0,
|
284613 |
+
"eval_loss": 0.3466392159461975,
|
284614 |
+
"eval_runtime": 39.3229,
|
284615 |
+
"eval_samples_per_second": 21.285,
|
284616 |
+
"eval_steps_per_second": 0.687,
|
284617 |
+
"eval_wer": 0.16716790141568252,
|
284618 |
+
"step": 145602
|
284619 |
}
|
284620 |
],
|
284621 |
+
"max_steps": 625000,
|
284622 |
"num_train_epochs": 5000,
|
284623 |
+
"total_flos": 4.097272814922444e+20,
|
284624 |
"trial_name": null,
|
284625 |
"trial_params": null
|
284626 |
}
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c3e8284682b82934af7c2cb7a47d0aad1afbbab74b9a20b0bdf99e1f1b28b90
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7400c5774686da0c05e6115d2e42b7e425eb92bed75c7039ed458db7dc841243
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:791f31e71e6884d8a8929ef220193ccb17e2cfec3b0e299f7e8687cba4af3ce4
|
3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28a0b76f68ffd1c2d412704c5679e6f1d5de953fb95b5cacc4cb5d077d0d04b0
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7dff24b5bd65b8a22b77667e8a8418682532690368c32e2fbb3b1785624a9976
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
-
"best_metric": 0.
|
3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -283350,11 +283350,1595 @@
|
|
283350 |
"eval_steps_per_second": 0.679,
|
283351 |
"eval_wer": 0.16790141568253503,
|
283352 |
"step": 144607
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
283353 |
}
|
283354 |
],
|
283355 |
-
"max_steps":
|
283356 |
"num_train_epochs": 5000,
|
283357 |
-
"total_flos": 4.
|
283358 |
"trial_name": null,
|
283359 |
"trial_params": null
|
283360 |
}
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.16716790141568252,
|
3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
|
4 |
+
"epoch": 1166.0,
|
5 |
+
"global_step": 145852,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
283350 |
"eval_steps_per_second": 0.679,
|
283351 |
"eval_wer": 0.16790141568253503,
|
283352 |
"step": 144607
|
283353 |
+
},
|
283354 |
+
{
|
283355 |
+
"epoch": 1166.02,
|
283356 |
+
"learning_rate": 7.680888529886915e-06,
|
283357 |
+
"loss": 0.3216,
|
283358 |
+
"step": 144610
|
283359 |
+
},
|
283360 |
+
{
|
283361 |
+
"epoch": 1166.06,
|
283362 |
+
"learning_rate": 7.68080775444265e-06,
|
283363 |
+
"loss": 0.2238,
|
283364 |
+
"step": 144615
|
283365 |
+
},
|
283366 |
+
{
|
283367 |
+
"epoch": 1166.1,
|
283368 |
+
"learning_rate": 7.680726978998385e-06,
|
283369 |
+
"loss": 0.318,
|
283370 |
+
"step": 144620
|
283371 |
+
},
|
283372 |
+
{
|
283373 |
+
"epoch": 1166.14,
|
283374 |
+
"learning_rate": 7.680646203554121e-06,
|
283375 |
+
"loss": 0.357,
|
283376 |
+
"step": 144625
|
283377 |
+
},
|
283378 |
+
{
|
283379 |
+
"epoch": 1166.18,
|
283380 |
+
"learning_rate": 7.680565428109855e-06,
|
283381 |
+
"loss": 0.8028,
|
283382 |
+
"step": 144630
|
283383 |
+
},
|
283384 |
+
{
|
283385 |
+
"epoch": 1166.22,
|
283386 |
+
"learning_rate": 7.680484652665591e-06,
|
283387 |
+
"loss": 0.8776,
|
283388 |
+
"step": 144635
|
283389 |
+
},
|
283390 |
+
{
|
283391 |
+
"epoch": 1166.27,
|
283392 |
+
"learning_rate": 7.680403877221325e-06,
|
283393 |
+
"loss": 0.2889,
|
283394 |
+
"step": 144640
|
283395 |
+
},
|
283396 |
+
{
|
283397 |
+
"epoch": 1166.31,
|
283398 |
+
"learning_rate": 7.680323101777061e-06,
|
283399 |
+
"loss": 0.3034,
|
283400 |
+
"step": 144645
|
283401 |
+
},
|
283402 |
+
{
|
283403 |
+
"epoch": 1166.35,
|
283404 |
+
"learning_rate": 7.680242326332795e-06,
|
283405 |
+
"loss": 0.3402,
|
283406 |
+
"step": 144650
|
283407 |
+
},
|
283408 |
+
{
|
283409 |
+
"epoch": 1166.39,
|
283410 |
+
"learning_rate": 7.680161550888531e-06,
|
283411 |
+
"loss": 0.6687,
|
283412 |
+
"step": 144655
|
283413 |
+
},
|
283414 |
+
{
|
283415 |
+
"epoch": 1166.43,
|
283416 |
+
"learning_rate": 7.680080775444265e-06,
|
283417 |
+
"loss": 0.8415,
|
283418 |
+
"step": 144660
|
283419 |
+
},
|
283420 |
+
{
|
283421 |
+
"epoch": 1166.47,
|
283422 |
+
"learning_rate": 7.680000000000001e-06,
|
283423 |
+
"loss": 0.283,
|
283424 |
+
"step": 144665
|
283425 |
+
},
|
283426 |
+
{
|
283427 |
+
"epoch": 1166.51,
|
283428 |
+
"learning_rate": 7.679919224555735e-06,
|
283429 |
+
"loss": 0.2845,
|
283430 |
+
"step": 144670
|
283431 |
+
},
|
283432 |
+
{
|
283433 |
+
"epoch": 1166.55,
|
283434 |
+
"learning_rate": 7.679838449111471e-06,
|
283435 |
+
"loss": 0.3641,
|
283436 |
+
"step": 144675
|
283437 |
+
},
|
283438 |
+
{
|
283439 |
+
"epoch": 1166.59,
|
283440 |
+
"learning_rate": 7.679757673667205e-06,
|
283441 |
+
"loss": 0.6544,
|
283442 |
+
"step": 144680
|
283443 |
+
},
|
283444 |
+
{
|
283445 |
+
"epoch": 1166.63,
|
283446 |
+
"learning_rate": 7.679676898222941e-06,
|
283447 |
+
"loss": 0.8825,
|
283448 |
+
"step": 144685
|
283449 |
+
},
|
283450 |
+
{
|
283451 |
+
"epoch": 1166.67,
|
283452 |
+
"learning_rate": 7.679596122778677e-06,
|
283453 |
+
"loss": 0.308,
|
283454 |
+
"step": 144690
|
283455 |
+
},
|
283456 |
+
{
|
283457 |
+
"epoch": 1166.71,
|
283458 |
+
"learning_rate": 7.679515347334411e-06,
|
283459 |
+
"loss": 0.2938,
|
283460 |
+
"step": 144695
|
283461 |
+
},
|
283462 |
+
{
|
283463 |
+
"epoch": 1166.75,
|
283464 |
+
"learning_rate": 7.679434571890147e-06,
|
283465 |
+
"loss": 0.3421,
|
283466 |
+
"step": 144700
|
283467 |
+
},
|
283468 |
+
{
|
283469 |
+
"epoch": 1166.79,
|
283470 |
+
"learning_rate": 7.679353796445881e-06,
|
283471 |
+
"loss": 0.7238,
|
283472 |
+
"step": 144705
|
283473 |
+
},
|
283474 |
+
{
|
283475 |
+
"epoch": 1166.83,
|
283476 |
+
"learning_rate": 7.679273021001617e-06,
|
283477 |
+
"loss": 0.8722,
|
283478 |
+
"step": 144710
|
283479 |
+
},
|
283480 |
+
{
|
283481 |
+
"epoch": 1166.87,
|
283482 |
+
"learning_rate": 7.679192245557351e-06,
|
283483 |
+
"loss": 0.2458,
|
283484 |
+
"step": 144715
|
283485 |
+
},
|
283486 |
+
{
|
283487 |
+
"epoch": 1166.91,
|
283488 |
+
"learning_rate": 7.679111470113087e-06,
|
283489 |
+
"loss": 0.2701,
|
283490 |
+
"step": 144720
|
283491 |
+
},
|
283492 |
+
{
|
283493 |
+
"epoch": 1166.95,
|
283494 |
+
"learning_rate": 7.679030694668821e-06,
|
283495 |
+
"loss": 0.3769,
|
283496 |
+
"step": 144725
|
283497 |
+
},
|
283498 |
+
{
|
283499 |
+
"epoch": 1166.99,
|
283500 |
+
"learning_rate": 7.678949919224557e-06,
|
283501 |
+
"loss": 0.8216,
|
283502 |
+
"step": 144730
|
283503 |
+
},
|
283504 |
+
{
|
283505 |
+
"epoch": 1167.0,
|
283506 |
+
"eval_loss": 0.3796067237854004,
|
283507 |
+
"eval_runtime": 41.6728,
|
283508 |
+
"eval_samples_per_second": 20.085,
|
283509 |
+
"eval_steps_per_second": 0.648,
|
283510 |
+
"eval_wer": 0.18190386427898209,
|
283511 |
+
"step": 144731
|
283512 |
+
},
|
283513 |
+
{
|
283514 |
+
"epoch": 1167.03,
|
283515 |
+
"learning_rate": 7.67886914378029e-06,
|
283516 |
+
"loss": 0.2946,
|
283517 |
+
"step": 144735
|
283518 |
+
},
|
283519 |
+
{
|
283520 |
+
"epoch": 1167.07,
|
283521 |
+
"learning_rate": 7.678788368336027e-06,
|
283522 |
+
"loss": 0.2505,
|
283523 |
+
"step": 144740
|
283524 |
+
},
|
283525 |
+
{
|
283526 |
+
"epoch": 1167.11,
|
283527 |
+
"learning_rate": 7.678707592891763e-06,
|
283528 |
+
"loss": 0.2865,
|
283529 |
+
"step": 144745
|
283530 |
+
},
|
283531 |
+
{
|
283532 |
+
"epoch": 1167.15,
|
283533 |
+
"learning_rate": 7.678626817447497e-06,
|
283534 |
+
"loss": 0.3545,
|
283535 |
+
"step": 144750
|
283536 |
+
},
|
283537 |
+
{
|
283538 |
+
"epoch": 1167.19,
|
283539 |
+
"learning_rate": 7.678546042003232e-06,
|
283540 |
+
"loss": 0.9134,
|
283541 |
+
"step": 144755
|
283542 |
+
},
|
283543 |
+
{
|
283544 |
+
"epoch": 1167.23,
|
283545 |
+
"learning_rate": 7.678465266558967e-06,
|
283546 |
+
"loss": 0.655,
|
283547 |
+
"step": 144760
|
283548 |
+
},
|
283549 |
+
{
|
283550 |
+
"epoch": 1167.27,
|
283551 |
+
"learning_rate": 7.678384491114702e-06,
|
283552 |
+
"loss": 0.2705,
|
283553 |
+
"step": 144765
|
283554 |
+
},
|
283555 |
+
{
|
283556 |
+
"epoch": 1167.31,
|
283557 |
+
"learning_rate": 7.678303715670437e-06,
|
283558 |
+
"loss": 0.2562,
|
283559 |
+
"step": 144770
|
283560 |
+
},
|
283561 |
+
{
|
283562 |
+
"epoch": 1167.35,
|
283563 |
+
"learning_rate": 7.678222940226172e-06,
|
283564 |
+
"loss": 0.4007,
|
283565 |
+
"step": 144775
|
283566 |
+
},
|
283567 |
+
{
|
283568 |
+
"epoch": 1167.39,
|
283569 |
+
"learning_rate": 7.678142164781907e-06,
|
283570 |
+
"loss": 0.8377,
|
283571 |
+
"step": 144780
|
283572 |
+
},
|
283573 |
+
{
|
283574 |
+
"epoch": 1167.43,
|
283575 |
+
"learning_rate": 7.678061389337642e-06,
|
283576 |
+
"loss": 0.7035,
|
283577 |
+
"step": 144785
|
283578 |
+
},
|
283579 |
+
{
|
283580 |
+
"epoch": 1167.47,
|
283581 |
+
"learning_rate": 7.677980613893377e-06,
|
283582 |
+
"loss": 0.2795,
|
283583 |
+
"step": 144790
|
283584 |
+
},
|
283585 |
+
{
|
283586 |
+
"epoch": 1167.51,
|
283587 |
+
"learning_rate": 7.677899838449112e-06,
|
283588 |
+
"loss": 0.2829,
|
283589 |
+
"step": 144795
|
283590 |
+
},
|
283591 |
+
{
|
283592 |
+
"epoch": 1167.55,
|
283593 |
+
"learning_rate": 7.677819063004848e-06,
|
283594 |
+
"loss": 0.3805,
|
283595 |
+
"step": 144800
|
283596 |
+
},
|
283597 |
+
{
|
283598 |
+
"epoch": 1167.59,
|
283599 |
+
"learning_rate": 7.677738287560582e-06,
|
283600 |
+
"loss": 0.7594,
|
283601 |
+
"step": 144805
|
283602 |
+
},
|
283603 |
+
{
|
283604 |
+
"epoch": 1167.63,
|
283605 |
+
"learning_rate": 7.677657512116318e-06,
|
283606 |
+
"loss": 0.6312,
|
283607 |
+
"step": 144810
|
283608 |
+
},
|
283609 |
+
{
|
283610 |
+
"epoch": 1167.67,
|
283611 |
+
"learning_rate": 7.677576736672052e-06,
|
283612 |
+
"loss": 0.257,
|
283613 |
+
"step": 144815
|
283614 |
+
},
|
283615 |
+
{
|
283616 |
+
"epoch": 1167.71,
|
283617 |
+
"learning_rate": 7.677495961227788e-06,
|
283618 |
+
"loss": 0.2697,
|
283619 |
+
"step": 144820
|
283620 |
+
},
|
283621 |
+
{
|
283622 |
+
"epoch": 1167.76,
|
283623 |
+
"learning_rate": 7.677415185783522e-06,
|
283624 |
+
"loss": 0.3065,
|
283625 |
+
"step": 144825
|
283626 |
+
},
|
283627 |
+
{
|
283628 |
+
"epoch": 1167.8,
|
283629 |
+
"learning_rate": 7.677334410339258e-06,
|
283630 |
+
"loss": 0.8168,
|
283631 |
+
"step": 144830
|
283632 |
+
},
|
283633 |
+
{
|
283634 |
+
"epoch": 1167.84,
|
283635 |
+
"learning_rate": 7.677253634894992e-06,
|
283636 |
+
"loss": 0.6347,
|
283637 |
+
"step": 144835
|
283638 |
+
},
|
283639 |
+
{
|
283640 |
+
"epoch": 1167.88,
|
283641 |
+
"learning_rate": 7.677172859450728e-06,
|
283642 |
+
"loss": 0.2511,
|
283643 |
+
"step": 144840
|
283644 |
+
},
|
283645 |
+
{
|
283646 |
+
"epoch": 1167.92,
|
283647 |
+
"learning_rate": 7.677092084006462e-06,
|
283648 |
+
"loss": 0.2972,
|
283649 |
+
"step": 144845
|
283650 |
+
},
|
283651 |
+
{
|
283652 |
+
"epoch": 1167.96,
|
283653 |
+
"learning_rate": 7.677011308562198e-06,
|
283654 |
+
"loss": 0.4141,
|
283655 |
+
"step": 144850
|
283656 |
+
},
|
283657 |
+
{
|
283658 |
+
"epoch": 1168.0,
|
283659 |
+
"learning_rate": 7.676930533117932e-06,
|
283660 |
+
"loss": 1.1673,
|
283661 |
+
"step": 144855
|
283662 |
+
},
|
283663 |
+
{
|
283664 |
+
"epoch": 1168.0,
|
283665 |
+
"eval_loss": 0.34798669815063477,
|
283666 |
+
"eval_runtime": 41.3687,
|
283667 |
+
"eval_samples_per_second": 20.233,
|
283668 |
+
"eval_steps_per_second": 0.653,
|
283669 |
+
"eval_wer": 0.17150972252567184,
|
283670 |
+
"step": 144855
|
283671 |
+
},
|
283672 |
+
{
|
283673 |
+
"epoch": 1158.04,
|
283674 |
+
"learning_rate": 7.676849757673668e-06,
|
283675 |
+
"loss": 0.3284,
|
283676 |
+
"step": 144860
|
283677 |
+
},
|
283678 |
+
{
|
283679 |
+
"epoch": 1158.08,
|
283680 |
+
"learning_rate": 7.676768982229404e-06,
|
283681 |
+
"loss": 0.2587,
|
283682 |
+
"step": 144865
|
283683 |
+
},
|
283684 |
+
{
|
283685 |
+
"epoch": 1158.12,
|
283686 |
+
"learning_rate": 7.676688206785138e-06,
|
283687 |
+
"loss": 0.3396,
|
283688 |
+
"step": 144870
|
283689 |
+
},
|
283690 |
+
{
|
283691 |
+
"epoch": 1158.16,
|
283692 |
+
"learning_rate": 7.676607431340874e-06,
|
283693 |
+
"loss": 0.4414,
|
283694 |
+
"step": 144875
|
283695 |
+
},
|
283696 |
+
{
|
283697 |
+
"epoch": 1158.2,
|
283698 |
+
"learning_rate": 7.676526655896608e-06,
|
283699 |
+
"loss": 1.0351,
|
283700 |
+
"step": 144880
|
283701 |
+
},
|
283702 |
+
{
|
283703 |
+
"epoch": 1158.24,
|
283704 |
+
"learning_rate": 7.676445880452344e-06,
|
283705 |
+
"loss": 0.2941,
|
283706 |
+
"step": 144885
|
283707 |
+
},
|
283708 |
+
{
|
283709 |
+
"epoch": 1158.28,
|
283710 |
+
"learning_rate": 7.676365105008078e-06,
|
283711 |
+
"loss": 0.2294,
|
283712 |
+
"step": 144890
|
283713 |
+
},
|
283714 |
+
{
|
283715 |
+
"epoch": 1158.32,
|
283716 |
+
"learning_rate": 7.676284329563814e-06,
|
283717 |
+
"loss": 0.326,
|
283718 |
+
"step": 144895
|
283719 |
+
},
|
283720 |
+
{
|
283721 |
+
"epoch": 1158.36,
|
283722 |
+
"learning_rate": 7.676203554119548e-06,
|
283723 |
+
"loss": 0.4232,
|
283724 |
+
"step": 144900
|
283725 |
+
},
|
283726 |
+
{
|
283727 |
+
"epoch": 1158.4,
|
283728 |
+
"learning_rate": 7.676122778675284e-06,
|
283729 |
+
"loss": 1.1964,
|
283730 |
+
"step": 144905
|
283731 |
+
},
|
283732 |
+
{
|
283733 |
+
"epoch": 1158.44,
|
283734 |
+
"learning_rate": 7.676042003231018e-06,
|
283735 |
+
"loss": 0.2844,
|
283736 |
+
"step": 144910
|
283737 |
+
},
|
283738 |
+
{
|
283739 |
+
"epoch": 1158.48,
|
283740 |
+
"learning_rate": 7.675961227786754e-06,
|
283741 |
+
"loss": 0.2528,
|
283742 |
+
"step": 144915
|
283743 |
+
},
|
283744 |
+
{
|
283745 |
+
"epoch": 1158.52,
|
283746 |
+
"learning_rate": 7.67588045234249e-06,
|
283747 |
+
"loss": 0.313,
|
283748 |
+
"step": 144920
|
283749 |
+
},
|
283750 |
+
{
|
283751 |
+
"epoch": 1158.56,
|
283752 |
+
"learning_rate": 7.675799676898224e-06,
|
283753 |
+
"loss": 0.4891,
|
283754 |
+
"step": 144925
|
283755 |
+
},
|
283756 |
+
{
|
283757 |
+
"epoch": 1158.6,
|
283758 |
+
"learning_rate": 7.67571890145396e-06,
|
283759 |
+
"loss": 1.3052,
|
283760 |
+
"step": 144930
|
283761 |
+
},
|
283762 |
+
{
|
283763 |
+
"epoch": 1158.64,
|
283764 |
+
"learning_rate": 7.675638126009694e-06,
|
283765 |
+
"loss": 0.3103,
|
283766 |
+
"step": 144935
|
283767 |
+
},
|
283768 |
+
{
|
283769 |
+
"epoch": 1158.68,
|
283770 |
+
"learning_rate": 7.67555735056543e-06,
|
283771 |
+
"loss": 0.3308,
|
283772 |
+
"step": 144940
|
283773 |
+
},
|
283774 |
+
{
|
283775 |
+
"epoch": 1158.72,
|
283776 |
+
"learning_rate": 7.675476575121164e-06,
|
283777 |
+
"loss": 0.3203,
|
283778 |
+
"step": 144945
|
283779 |
+
},
|
283780 |
+
{
|
283781 |
+
"epoch": 1158.76,
|
283782 |
+
"learning_rate": 7.6753957996769e-06,
|
283783 |
+
"loss": 0.3745,
|
283784 |
+
"step": 144950
|
283785 |
+
},
|
283786 |
+
{
|
283787 |
+
"epoch": 1158.8,
|
283788 |
+
"learning_rate": 7.675315024232634e-06,
|
283789 |
+
"loss": 1.1487,
|
283790 |
+
"step": 144955
|
283791 |
+
},
|
283792 |
+
{
|
283793 |
+
"epoch": 1158.84,
|
283794 |
+
"learning_rate": 7.67523424878837e-06,
|
283795 |
+
"loss": 0.4522,
|
283796 |
+
"step": 144960
|
283797 |
+
},
|
283798 |
+
{
|
283799 |
+
"epoch": 1158.88,
|
283800 |
+
"learning_rate": 7.675153473344104e-06,
|
283801 |
+
"loss": 0.3011,
|
283802 |
+
"step": 144965
|
283803 |
+
},
|
283804 |
+
{
|
283805 |
+
"epoch": 1158.92,
|
283806 |
+
"learning_rate": 7.67507269789984e-06,
|
283807 |
+
"loss": 0.2663,
|
283808 |
+
"step": 144970
|
283809 |
+
},
|
283810 |
+
{
|
283811 |
+
"epoch": 1158.96,
|
283812 |
+
"learning_rate": 7.674991922455575e-06,
|
283813 |
+
"loss": 0.4301,
|
283814 |
+
"step": 144975
|
283815 |
+
},
|
283816 |
+
{
|
283817 |
+
"epoch": 1159.0,
|
283818 |
+
"learning_rate": 7.67491114701131e-06,
|
283819 |
+
"loss": 1.3414,
|
283820 |
+
"step": 144980
|
283821 |
+
},
|
283822 |
+
{
|
283823 |
+
"epoch": 1159.0,
|
283824 |
+
"eval_loss": 0.3729027807712555,
|
283825 |
+
"eval_runtime": 40.7509,
|
283826 |
+
"eval_samples_per_second": 20.539,
|
283827 |
+
"eval_steps_per_second": 0.663,
|
283828 |
+
"eval_wer": 0.17167350527549824,
|
283829 |
+
"step": 144980
|
283830 |
+
},
|
283831 |
+
{
|
283832 |
+
"epoch": 1169.04,
|
283833 |
+
"learning_rate": 7.674830371567045e-06,
|
283834 |
+
"loss": 0.3148,
|
283835 |
+
"step": 144985
|
283836 |
+
},
|
283837 |
+
{
|
283838 |
+
"epoch": 1169.08,
|
283839 |
+
"learning_rate": 7.67474959612278e-06,
|
283840 |
+
"loss": 0.2913,
|
283841 |
+
"step": 144990
|
283842 |
+
},
|
283843 |
+
{
|
283844 |
+
"epoch": 1169.12,
|
283845 |
+
"learning_rate": 7.674668820678515e-06,
|
283846 |
+
"loss": 0.2694,
|
283847 |
+
"step": 144995
|
283848 |
+
},
|
283849 |
+
{
|
283850 |
+
"epoch": 1169.16,
|
283851 |
+
"learning_rate": 7.67458804523425e-06,
|
283852 |
+
"loss": 0.478,
|
283853 |
+
"step": 145000
|
283854 |
+
},
|
283855 |
+
{
|
283856 |
+
"epoch": 1169.2,
|
283857 |
+
"learning_rate": 7.674507269789985e-06,
|
283858 |
+
"loss": 1.1648,
|
283859 |
+
"step": 145005
|
283860 |
+
},
|
283861 |
+
{
|
283862 |
+
"epoch": 1169.24,
|
283863 |
+
"learning_rate": 7.67442649434572e-06,
|
283864 |
+
"loss": 0.3068,
|
283865 |
+
"step": 145010
|
283866 |
+
},
|
283867 |
+
{
|
283868 |
+
"epoch": 1169.28,
|
283869 |
+
"learning_rate": 7.674345718901455e-06,
|
283870 |
+
"loss": 0.2923,
|
283871 |
+
"step": 145015
|
283872 |
+
},
|
283873 |
+
{
|
283874 |
+
"epoch": 1169.32,
|
283875 |
+
"learning_rate": 7.67426494345719e-06,
|
283876 |
+
"loss": 0.277,
|
283877 |
+
"step": 145020
|
283878 |
+
},
|
283879 |
+
{
|
283880 |
+
"epoch": 1169.36,
|
283881 |
+
"learning_rate": 7.674184168012925e-06,
|
283882 |
+
"loss": 0.5259,
|
283883 |
+
"step": 145025
|
283884 |
+
},
|
283885 |
+
{
|
283886 |
+
"epoch": 1169.4,
|
283887 |
+
"learning_rate": 7.67410339256866e-06,
|
283888 |
+
"loss": 1.3239,
|
283889 |
+
"step": 145030
|
283890 |
+
},
|
283891 |
+
{
|
283892 |
+
"epoch": 1169.44,
|
283893 |
+
"learning_rate": 7.674022617124395e-06,
|
283894 |
+
"loss": 0.2994,
|
283895 |
+
"step": 145035
|
283896 |
+
},
|
283897 |
+
{
|
283898 |
+
"epoch": 1169.48,
|
283899 |
+
"learning_rate": 7.673941841680131e-06,
|
283900 |
+
"loss": 0.2812,
|
283901 |
+
"step": 145040
|
283902 |
+
},
|
283903 |
+
{
|
283904 |
+
"epoch": 1169.52,
|
283905 |
+
"learning_rate": 7.673861066235865e-06,
|
283906 |
+
"loss": 0.2935,
|
283907 |
+
"step": 145045
|
283908 |
+
},
|
283909 |
+
{
|
283910 |
+
"epoch": 1169.56,
|
283911 |
+
"learning_rate": 7.673780290791601e-06,
|
283912 |
+
"loss": 0.4962,
|
283913 |
+
"step": 145050
|
283914 |
+
},
|
283915 |
+
{
|
283916 |
+
"epoch": 1169.6,
|
283917 |
+
"learning_rate": 7.673699515347335e-06,
|
283918 |
+
"loss": 1.2089,
|
283919 |
+
"step": 145055
|
283920 |
+
},
|
283921 |
+
{
|
283922 |
+
"epoch": 1169.64,
|
283923 |
+
"learning_rate": 7.673618739903071e-06,
|
283924 |
+
"loss": 0.2642,
|
283925 |
+
"step": 145060
|
283926 |
+
},
|
283927 |
+
{
|
283928 |
+
"epoch": 1169.68,
|
283929 |
+
"learning_rate": 7.673537964458805e-06,
|
283930 |
+
"loss": 0.2744,
|
283931 |
+
"step": 145065
|
283932 |
+
},
|
283933 |
+
{
|
283934 |
+
"epoch": 1169.72,
|
283935 |
+
"learning_rate": 7.673457189014541e-06,
|
283936 |
+
"loss": 0.278,
|
283937 |
+
"step": 145070
|
283938 |
+
},
|
283939 |
+
{
|
283940 |
+
"epoch": 1169.76,
|
283941 |
+
"learning_rate": 7.673376413570275e-06,
|
283942 |
+
"loss": 0.4382,
|
283943 |
+
"step": 145075
|
283944 |
+
},
|
283945 |
+
{
|
283946 |
+
"epoch": 1169.8,
|
283947 |
+
"learning_rate": 7.673295638126011e-06,
|
283948 |
+
"loss": 1.0989,
|
283949 |
+
"step": 145080
|
283950 |
+
},
|
283951 |
+
{
|
283952 |
+
"epoch": 1169.84,
|
283953 |
+
"learning_rate": 7.673214862681745e-06,
|
283954 |
+
"loss": 0.3134,
|
283955 |
+
"step": 145085
|
283956 |
+
},
|
283957 |
+
{
|
283958 |
+
"epoch": 1169.88,
|
283959 |
+
"learning_rate": 7.673134087237481e-06,
|
283960 |
+
"loss": 0.3014,
|
283961 |
+
"step": 145090
|
283962 |
+
},
|
283963 |
+
{
|
283964 |
+
"epoch": 1169.92,
|
283965 |
+
"learning_rate": 7.673053311793217e-06,
|
283966 |
+
"loss": 0.2862,
|
283967 |
+
"step": 145095
|
283968 |
+
},
|
283969 |
+
{
|
283970 |
+
"epoch": 1169.96,
|
283971 |
+
"learning_rate": 7.672972536348951e-06,
|
283972 |
+
"loss": 0.5695,
|
283973 |
+
"step": 145100
|
283974 |
+
},
|
283975 |
+
{
|
283976 |
+
"epoch": 1170.0,
|
283977 |
+
"eval_loss": 0.39969587326049805,
|
283978 |
+
"eval_runtime": 40.575,
|
283979 |
+
"eval_samples_per_second": 20.628,
|
283980 |
+
"eval_steps_per_second": 0.665,
|
283981 |
+
"eval_wer": 0.1783857926653191,
|
283982 |
+
"step": 145104
|
283983 |
+
},
|
283984 |
+
{
|
283985 |
+
"epoch": 1170.01,
|
283986 |
+
"learning_rate": 7.672891760904687e-06,
|
283987 |
+
"loss": 0.3477,
|
283988 |
+
"step": 145105
|
283989 |
+
},
|
283990 |
+
{
|
283991 |
+
"epoch": 1170.05,
|
283992 |
+
"learning_rate": 7.67281098546042e-06,
|
283993 |
+
"loss": 0.3213,
|
283994 |
+
"step": 145110
|
283995 |
+
},
|
283996 |
+
{
|
283997 |
+
"epoch": 1170.09,
|
283998 |
+
"learning_rate": 7.672730210016157e-06,
|
283999 |
+
"loss": 0.275,
|
284000 |
+
"step": 145115
|
284001 |
+
},
|
284002 |
+
{
|
284003 |
+
"epoch": 1170.13,
|
284004 |
+
"learning_rate": 7.67264943457189e-06,
|
284005 |
+
"loss": 0.3189,
|
284006 |
+
"step": 145120
|
284007 |
+
},
|
284008 |
+
{
|
284009 |
+
"epoch": 1170.17,
|
284010 |
+
"learning_rate": 7.672568659127627e-06,
|
284011 |
+
"loss": 0.5509,
|
284012 |
+
"step": 145125
|
284013 |
+
},
|
284014 |
+
{
|
284015 |
+
"epoch": 1170.21,
|
284016 |
+
"learning_rate": 7.67248788368336e-06,
|
284017 |
+
"loss": 1.0719,
|
284018 |
+
"step": 145130
|
284019 |
+
},
|
284020 |
+
{
|
284021 |
+
"epoch": 1170.25,
|
284022 |
+
"learning_rate": 7.672407108239097e-06,
|
284023 |
+
"loss": 0.2919,
|
284024 |
+
"step": 145135
|
284025 |
+
},
|
284026 |
+
{
|
284027 |
+
"epoch": 1170.29,
|
284028 |
+
"learning_rate": 7.67232633279483e-06,
|
284029 |
+
"loss": 0.2532,
|
284030 |
+
"step": 145140
|
284031 |
+
},
|
284032 |
+
{
|
284033 |
+
"epoch": 1170.33,
|
284034 |
+
"learning_rate": 7.672245557350567e-06,
|
284035 |
+
"loss": 0.3189,
|
284036 |
+
"step": 145145
|
284037 |
+
},
|
284038 |
+
{
|
284039 |
+
"epoch": 1170.37,
|
284040 |
+
"learning_rate": 7.672164781906302e-06,
|
284041 |
+
"loss": 0.4942,
|
284042 |
+
"step": 145150
|
284043 |
+
},
|
284044 |
+
{
|
284045 |
+
"epoch": 1170.41,
|
284046 |
+
"learning_rate": 7.672084006462037e-06,
|
284047 |
+
"loss": 1.1852,
|
284048 |
+
"step": 145155
|
284049 |
+
},
|
284050 |
+
{
|
284051 |
+
"epoch": 1170.45,
|
284052 |
+
"learning_rate": 7.672003231017772e-06,
|
284053 |
+
"loss": 0.3175,
|
284054 |
+
"step": 145160
|
284055 |
+
},
|
284056 |
+
{
|
284057 |
+
"epoch": 1170.49,
|
284058 |
+
"learning_rate": 7.671922455573507e-06,
|
284059 |
+
"loss": 0.3137,
|
284060 |
+
"step": 145165
|
284061 |
+
},
|
284062 |
+
{
|
284063 |
+
"epoch": 1170.53,
|
284064 |
+
"learning_rate": 7.671841680129242e-06,
|
284065 |
+
"loss": 0.2959,
|
284066 |
+
"step": 145170
|
284067 |
+
},
|
284068 |
+
{
|
284069 |
+
"epoch": 1170.57,
|
284070 |
+
"learning_rate": 7.671760904684977e-06,
|
284071 |
+
"loss": 0.4471,
|
284072 |
+
"step": 145175
|
284073 |
+
},
|
284074 |
+
{
|
284075 |
+
"epoch": 1170.61,
|
284076 |
+
"learning_rate": 7.671680129240712e-06,
|
284077 |
+
"loss": 1.0092,
|
284078 |
+
"step": 145180
|
284079 |
+
},
|
284080 |
+
{
|
284081 |
+
"epoch": 1170.65,
|
284082 |
+
"learning_rate": 7.671599353796446e-06,
|
284083 |
+
"loss": 0.3617,
|
284084 |
+
"step": 145185
|
284085 |
+
},
|
284086 |
+
{
|
284087 |
+
"epoch": 1170.69,
|
284088 |
+
"learning_rate": 7.671518578352182e-06,
|
284089 |
+
"loss": 0.3141,
|
284090 |
+
"step": 145190
|
284091 |
+
},
|
284092 |
+
{
|
284093 |
+
"epoch": 1170.73,
|
284094 |
+
"learning_rate": 7.671437802907916e-06,
|
284095 |
+
"loss": 0.2982,
|
284096 |
+
"step": 145195
|
284097 |
+
},
|
284098 |
+
{
|
284099 |
+
"epoch": 1170.77,
|
284100 |
+
"learning_rate": 7.671357027463652e-06,
|
284101 |
+
"loss": 0.4728,
|
284102 |
+
"step": 145200
|
284103 |
+
},
|
284104 |
+
{
|
284105 |
+
"epoch": 1170.81,
|
284106 |
+
"learning_rate": 7.671276252019386e-06,
|
284107 |
+
"loss": 1.086,
|
284108 |
+
"step": 145205
|
284109 |
+
},
|
284110 |
+
{
|
284111 |
+
"epoch": 1170.85,
|
284112 |
+
"learning_rate": 7.671195476575122e-06,
|
284113 |
+
"loss": 0.3326,
|
284114 |
+
"step": 145210
|
284115 |
+
},
|
284116 |
+
{
|
284117 |
+
"epoch": 1170.89,
|
284118 |
+
"learning_rate": 7.671114701130858e-06,
|
284119 |
+
"loss": 0.2725,
|
284120 |
+
"step": 145215
|
284121 |
+
},
|
284122 |
+
{
|
284123 |
+
"epoch": 1170.93,
|
284124 |
+
"learning_rate": 7.671033925686592e-06,
|
284125 |
+
"loss": 0.3219,
|
284126 |
+
"step": 145220
|
284127 |
+
},
|
284128 |
+
{
|
284129 |
+
"epoch": 1170.97,
|
284130 |
+
"learning_rate": 7.670953150242328e-06,
|
284131 |
+
"loss": 0.5993,
|
284132 |
+
"step": 145225
|
284133 |
+
},
|
284134 |
+
{
|
284135 |
+
"epoch": 1171.0,
|
284136 |
+
"eval_loss": 0.3494146466255188,
|
284137 |
+
"eval_runtime": 40.4075,
|
284138 |
+
"eval_samples_per_second": 20.714,
|
284139 |
+
"eval_steps_per_second": 0.668,
|
284140 |
+
"eval_wer": 0.17565191913272782,
|
284141 |
+
"step": 145228
|
284142 |
+
},
|
284143 |
+
{
|
284144 |
+
"epoch": 1171.02,
|
284145 |
+
"learning_rate": 7.670872374798062e-06,
|
284146 |
+
"loss": 0.3162,
|
284147 |
+
"step": 145230
|
284148 |
+
},
|
284149 |
+
{
|
284150 |
+
"epoch": 1171.06,
|
284151 |
+
"learning_rate": 7.670791599353798e-06,
|
284152 |
+
"loss": 0.268,
|
284153 |
+
"step": 145235
|
284154 |
+
},
|
284155 |
+
{
|
284156 |
+
"epoch": 1171.1,
|
284157 |
+
"learning_rate": 7.670710823909532e-06,
|
284158 |
+
"loss": 0.3148,
|
284159 |
+
"step": 145240
|
284160 |
+
},
|
284161 |
+
{
|
284162 |
+
"epoch": 1171.14,
|
284163 |
+
"learning_rate": 7.670630048465268e-06,
|
284164 |
+
"loss": 0.3431,
|
284165 |
+
"step": 145245
|
284166 |
+
},
|
284167 |
+
{
|
284168 |
+
"epoch": 1171.18,
|
284169 |
+
"learning_rate": 7.670549273021002e-06,
|
284170 |
+
"loss": 0.5562,
|
284171 |
+
"step": 145250
|
284172 |
+
},
|
284173 |
+
{
|
284174 |
+
"epoch": 1171.22,
|
284175 |
+
"learning_rate": 7.670468497576738e-06,
|
284176 |
+
"loss": 0.8706,
|
284177 |
+
"step": 145255
|
284178 |
+
},
|
284179 |
+
{
|
284180 |
+
"epoch": 1171.26,
|
284181 |
+
"learning_rate": 7.670387722132472e-06,
|
284182 |
+
"loss": 0.283,
|
284183 |
+
"step": 145260
|
284184 |
+
},
|
284185 |
+
{
|
284186 |
+
"epoch": 1171.3,
|
284187 |
+
"learning_rate": 7.670306946688208e-06,
|
284188 |
+
"loss": 0.3143,
|
284189 |
+
"step": 145265
|
284190 |
+
},
|
284191 |
+
{
|
284192 |
+
"epoch": 1171.34,
|
284193 |
+
"learning_rate": 7.670226171243944e-06,
|
284194 |
+
"loss": 0.2854,
|
284195 |
+
"step": 145270
|
284196 |
+
},
|
284197 |
+
{
|
284198 |
+
"epoch": 1171.38,
|
284199 |
+
"learning_rate": 7.670145395799678e-06,
|
284200 |
+
"loss": 0.5725,
|
284201 |
+
"step": 145275
|
284202 |
+
},
|
284203 |
+
{
|
284204 |
+
"epoch": 1171.42,
|
284205 |
+
"learning_rate": 7.670064620355414e-06,
|
284206 |
+
"loss": 0.9977,
|
284207 |
+
"step": 145280
|
284208 |
+
},
|
284209 |
+
{
|
284210 |
+
"epoch": 1171.46,
|
284211 |
+
"learning_rate": 7.669983844911148e-06,
|
284212 |
+
"loss": 0.2628,
|
284213 |
+
"step": 145285
|
284214 |
+
},
|
284215 |
+
{
|
284216 |
+
"epoch": 1171.5,
|
284217 |
+
"learning_rate": 7.669903069466884e-06,
|
284218 |
+
"loss": 0.2448,
|
284219 |
+
"step": 145290
|
284220 |
+
},
|
284221 |
+
{
|
284222 |
+
"epoch": 1171.54,
|
284223 |
+
"learning_rate": 7.669822294022618e-06,
|
284224 |
+
"loss": 0.3178,
|
284225 |
+
"step": 145295
|
284226 |
+
},
|
284227 |
+
{
|
284228 |
+
"epoch": 1171.58,
|
284229 |
+
"learning_rate": 7.669741518578354e-06,
|
284230 |
+
"loss": 0.6503,
|
284231 |
+
"step": 145300
|
284232 |
+
},
|
284233 |
+
{
|
284234 |
+
"epoch": 1171.62,
|
284235 |
+
"learning_rate": 7.669660743134088e-06,
|
284236 |
+
"loss": 0.9847,
|
284237 |
+
"step": 145305
|
284238 |
+
},
|
284239 |
+
{
|
284240 |
+
"epoch": 1171.66,
|
284241 |
+
"learning_rate": 7.669579967689824e-06,
|
284242 |
+
"loss": 0.2887,
|
284243 |
+
"step": 145310
|
284244 |
+
},
|
284245 |
+
{
|
284246 |
+
"epoch": 1171.7,
|
284247 |
+
"learning_rate": 7.669499192245558e-06,
|
284248 |
+
"loss": 0.2481,
|
284249 |
+
"step": 145315
|
284250 |
+
},
|
284251 |
+
{
|
284252 |
+
"epoch": 1171.74,
|
284253 |
+
"learning_rate": 7.669418416801294e-06,
|
284254 |
+
"loss": 0.2989,
|
284255 |
+
"step": 145320
|
284256 |
+
},
|
284257 |
+
{
|
284258 |
+
"epoch": 1171.78,
|
284259 |
+
"learning_rate": 7.669337641357028e-06,
|
284260 |
+
"loss": 0.6053,
|
284261 |
+
"step": 145325
|
284262 |
+
},
|
284263 |
+
{
|
284264 |
+
"epoch": 1171.82,
|
284265 |
+
"learning_rate": 7.669256865912764e-06,
|
284266 |
+
"loss": 0.9806,
|
284267 |
+
"step": 145330
|
284268 |
+
},
|
284269 |
+
{
|
284270 |
+
"epoch": 1171.86,
|
284271 |
+
"learning_rate": 7.669176090468498e-06,
|
284272 |
+
"loss": 0.3248,
|
284273 |
+
"step": 145335
|
284274 |
+
},
|
284275 |
+
{
|
284276 |
+
"epoch": 1171.9,
|
284277 |
+
"learning_rate": 7.669095315024234e-06,
|
284278 |
+
"loss": 0.3173,
|
284279 |
+
"step": 145340
|
284280 |
+
},
|
284281 |
+
{
|
284282 |
+
"epoch": 1171.94,
|
284283 |
+
"learning_rate": 7.66901453957997e-06,
|
284284 |
+
"loss": 0.3376,
|
284285 |
+
"step": 145345
|
284286 |
+
},
|
284287 |
+
{
|
284288 |
+
"epoch": 1171.98,
|
284289 |
+
"learning_rate": 7.668933764135704e-06,
|
284290 |
+
"loss": 0.6947,
|
284291 |
+
"step": 145350
|
284292 |
+
},
|
284293 |
+
{
|
284294 |
+
"epoch": 1172.0,
|
284295 |
+
"eval_loss": 0.47769203782081604,
|
284296 |
+
"eval_runtime": 39.6454,
|
284297 |
+
"eval_samples_per_second": 21.112,
|
284298 |
+
"eval_steps_per_second": 0.681,
|
284299 |
+
"eval_wer": 0.18728573557907288,
|
284300 |
+
"step": 145352
|
284301 |
+
},
|
284302 |
+
{
|
284303 |
+
"epoch": 1162.02,
|
284304 |
+
"learning_rate": 7.66885298869144e-06,
|
284305 |
+
"loss": 0.3284,
|
284306 |
+
"step": 145355
|
284307 |
+
},
|
284308 |
+
{
|
284309 |
+
"epoch": 1162.06,
|
284310 |
+
"learning_rate": 7.668772213247174e-06,
|
284311 |
+
"loss": 0.2665,
|
284312 |
+
"step": 145360
|
284313 |
+
},
|
284314 |
+
{
|
284315 |
+
"epoch": 1162.1,
|
284316 |
+
"learning_rate": 7.66869143780291e-06,
|
284317 |
+
"loss": 0.266,
|
284318 |
+
"step": 145365
|
284319 |
+
},
|
284320 |
+
{
|
284321 |
+
"epoch": 1162.14,
|
284322 |
+
"learning_rate": 7.668610662358644e-06,
|
284323 |
+
"loss": 0.3346,
|
284324 |
+
"step": 145370
|
284325 |
+
},
|
284326 |
+
{
|
284327 |
+
"epoch": 1162.18,
|
284328 |
+
"learning_rate": 7.66852988691438e-06,
|
284329 |
+
"loss": 0.6964,
|
284330 |
+
"step": 145375
|
284331 |
+
},
|
284332 |
+
{
|
284333 |
+
"epoch": 1162.22,
|
284334 |
+
"learning_rate": 7.668449111470114e-06,
|
284335 |
+
"loss": 0.8126,
|
284336 |
+
"step": 145380
|
284337 |
+
},
|
284338 |
+
{
|
284339 |
+
"epoch": 1162.26,
|
284340 |
+
"learning_rate": 7.66836833602585e-06,
|
284341 |
+
"loss": 0.2484,
|
284342 |
+
"step": 145385
|
284343 |
+
},
|
284344 |
+
{
|
284345 |
+
"epoch": 1162.3,
|
284346 |
+
"learning_rate": 7.668287560581584e-06,
|
284347 |
+
"loss": 0.2748,
|
284348 |
+
"step": 145390
|
284349 |
+
},
|
284350 |
+
{
|
284351 |
+
"epoch": 1162.34,
|
284352 |
+
"learning_rate": 7.66820678513732e-06,
|
284353 |
+
"loss": 0.3347,
|
284354 |
+
"step": 145395
|
284355 |
+
},
|
284356 |
+
{
|
284357 |
+
"epoch": 1162.38,
|
284358 |
+
"learning_rate": 7.668126009693053e-06,
|
284359 |
+
"loss": 0.7334,
|
284360 |
+
"step": 145400
|
284361 |
+
},
|
284362 |
+
{
|
284363 |
+
"epoch": 1162.42,
|
284364 |
+
"learning_rate": 7.66804523424879e-06,
|
284365 |
+
"loss": 0.7622,
|
284366 |
+
"step": 145405
|
284367 |
+
},
|
284368 |
+
{
|
284369 |
+
"epoch": 1162.46,
|
284370 |
+
"learning_rate": 7.667964458804525e-06,
|
284371 |
+
"loss": 0.2724,
|
284372 |
+
"step": 145410
|
284373 |
+
},
|
284374 |
+
{
|
284375 |
+
"epoch": 1162.5,
|
284376 |
+
"learning_rate": 7.66788368336026e-06,
|
284377 |
+
"loss": 0.2608,
|
284378 |
+
"step": 145415
|
284379 |
+
},
|
284380 |
+
{
|
284381 |
+
"epoch": 1162.54,
|
284382 |
+
"learning_rate": 7.667802907915995e-06,
|
284383 |
+
"loss": 0.3223,
|
284384 |
+
"step": 145420
|
284385 |
+
},
|
284386 |
+
{
|
284387 |
+
"epoch": 1162.58,
|
284388 |
+
"learning_rate": 7.66772213247173e-06,
|
284389 |
+
"loss": 0.6718,
|
284390 |
+
"step": 145425
|
284391 |
+
},
|
284392 |
+
{
|
284393 |
+
"epoch": 1162.62,
|
284394 |
+
"learning_rate": 7.667641357027465e-06,
|
284395 |
+
"loss": 0.8369,
|
284396 |
+
"step": 145430
|
284397 |
+
},
|
284398 |
+
{
|
284399 |
+
"epoch": 1162.66,
|
284400 |
+
"learning_rate": 7.6675605815832e-06,
|
284401 |
+
"loss": 0.2615,
|
284402 |
+
"step": 145435
|
284403 |
+
},
|
284404 |
+
{
|
284405 |
+
"epoch": 1162.7,
|
284406 |
+
"learning_rate": 7.667479806138935e-06,
|
284407 |
+
"loss": 0.2988,
|
284408 |
+
"step": 145440
|
284409 |
+
},
|
284410 |
+
{
|
284411 |
+
"epoch": 1162.74,
|
284412 |
+
"learning_rate": 7.66739903069467e-06,
|
284413 |
+
"loss": 0.3295,
|
284414 |
+
"step": 145445
|
284415 |
+
},
|
284416 |
+
{
|
284417 |
+
"epoch": 1162.78,
|
284418 |
+
"learning_rate": 7.667318255250405e-06,
|
284419 |
+
"loss": 0.7126,
|
284420 |
+
"step": 145450
|
284421 |
+
},
|
284422 |
+
{
|
284423 |
+
"epoch": 1162.82,
|
284424 |
+
"learning_rate": 7.66723747980614e-06,
|
284425 |
+
"loss": 0.8154,
|
284426 |
+
"step": 145455
|
284427 |
+
},
|
284428 |
+
{
|
284429 |
+
"epoch": 1162.86,
|
284430 |
+
"learning_rate": 7.667156704361875e-06,
|
284431 |
+
"loss": 0.2407,
|
284432 |
+
"step": 145460
|
284433 |
+
},
|
284434 |
+
{
|
284435 |
+
"epoch": 1162.9,
|
284436 |
+
"learning_rate": 7.66707592891761e-06,
|
284437 |
+
"loss": 0.2759,
|
284438 |
+
"step": 145465
|
284439 |
+
},
|
284440 |
+
{
|
284441 |
+
"epoch": 1162.94,
|
284442 |
+
"learning_rate": 7.666995153473345e-06,
|
284443 |
+
"loss": 0.2949,
|
284444 |
+
"step": 145470
|
284445 |
+
},
|
284446 |
+
{
|
284447 |
+
"epoch": 1162.98,
|
284448 |
+
"learning_rate": 7.66691437802908e-06,
|
284449 |
+
"loss": 0.7172,
|
284450 |
+
"step": 145475
|
284451 |
+
},
|
284452 |
+
{
|
284453 |
+
"epoch": 1163.0,
|
284454 |
+
"eval_loss": 0.3563687801361084,
|
284455 |
+
"eval_runtime": 41.8928,
|
284456 |
+
"eval_samples_per_second": 19.98,
|
284457 |
+
"eval_steps_per_second": 0.645,
|
284458 |
+
"eval_wer": 0.17516339869281045,
|
284459 |
+
"step": 145477
|
284460 |
+
},
|
284461 |
+
{
|
284462 |
+
"epoch": 1163.02,
|
284463 |
+
"learning_rate": 7.666833602584815e-06,
|
284464 |
+
"loss": 0.3148,
|
284465 |
+
"step": 145480
|
284466 |
+
},
|
284467 |
+
{
|
284468 |
+
"epoch": 1163.06,
|
284469 |
+
"learning_rate": 7.66675282714055e-06,
|
284470 |
+
"loss": 0.2606,
|
284471 |
+
"step": 145485
|
284472 |
+
},
|
284473 |
+
{
|
284474 |
+
"epoch": 1163.1,
|
284475 |
+
"learning_rate": 7.666672051696285e-06,
|
284476 |
+
"loss": 0.3123,
|
284477 |
+
"step": 145490
|
284478 |
+
},
|
284479 |
+
{
|
284480 |
+
"epoch": 1163.14,
|
284481 |
+
"learning_rate": 7.66659127625202e-06,
|
284482 |
+
"loss": 0.3515,
|
284483 |
+
"step": 145495
|
284484 |
+
},
|
284485 |
+
{
|
284486 |
+
"epoch": 1163.18,
|
284487 |
+
"learning_rate": 7.666510500807755e-06,
|
284488 |
+
"loss": 0.7378,
|
284489 |
+
"step": 145500
|
284490 |
+
},
|
284491 |
+
{
|
284492 |
+
"epoch": 1163.22,
|
284493 |
+
"learning_rate": 7.66642972536349e-06,
|
284494 |
+
"loss": 0.8376,
|
284495 |
+
"step": 145505
|
284496 |
+
},
|
284497 |
+
{
|
284498 |
+
"epoch": 1163.26,
|
284499 |
+
"learning_rate": 7.666348949919225e-06,
|
284500 |
+
"loss": 0.3011,
|
284501 |
+
"step": 145510
|
284502 |
+
},
|
284503 |
+
{
|
284504 |
+
"epoch": 1163.3,
|
284505 |
+
"learning_rate": 7.66626817447496e-06,
|
284506 |
+
"loss": 0.3181,
|
284507 |
+
"step": 145515
|
284508 |
+
},
|
284509 |
+
{
|
284510 |
+
"epoch": 1163.34,
|
284511 |
+
"learning_rate": 7.666187399030695e-06,
|
284512 |
+
"loss": 0.4141,
|
284513 |
+
"step": 145520
|
284514 |
+
},
|
284515 |
+
{
|
284516 |
+
"epoch": 1163.38,
|
284517 |
+
"learning_rate": 7.66610662358643e-06,
|
284518 |
+
"loss": 0.7503,
|
284519 |
+
"step": 145525
|
284520 |
+
},
|
284521 |
+
{
|
284522 |
+
"epoch": 1163.42,
|
284523 |
+
"learning_rate": 7.666025848142165e-06,
|
284524 |
+
"loss": 0.8573,
|
284525 |
+
"step": 145530
|
284526 |
+
},
|
284527 |
+
{
|
284528 |
+
"epoch": 1163.46,
|
284529 |
+
"learning_rate": 7.6659450726979e-06,
|
284530 |
+
"loss": 0.283,
|
284531 |
+
"step": 145535
|
284532 |
+
},
|
284533 |
+
{
|
284534 |
+
"epoch": 1163.5,
|
284535 |
+
"learning_rate": 7.665864297253635e-06,
|
284536 |
+
"loss": 0.2771,
|
284537 |
+
"step": 145540
|
284538 |
+
},
|
284539 |
+
{
|
284540 |
+
"epoch": 1163.54,
|
284541 |
+
"learning_rate": 7.66578352180937e-06,
|
284542 |
+
"loss": 0.5021,
|
284543 |
+
"step": 145545
|
284544 |
+
},
|
284545 |
+
{
|
284546 |
+
"epoch": 1163.58,
|
284547 |
+
"learning_rate": 7.665702746365107e-06,
|
284548 |
+
"loss": 0.7615,
|
284549 |
+
"step": 145550
|
284550 |
+
},
|
284551 |
+
{
|
284552 |
+
"epoch": 1163.62,
|
284553 |
+
"learning_rate": 7.66562197092084e-06,
|
284554 |
+
"loss": 0.8149,
|
284555 |
+
"step": 145555
|
284556 |
+
},
|
284557 |
+
{
|
284558 |
+
"epoch": 1163.66,
|
284559 |
+
"learning_rate": 7.665541195476576e-06,
|
284560 |
+
"loss": 0.2333,
|
284561 |
+
"step": 145560
|
284562 |
+
},
|
284563 |
+
{
|
284564 |
+
"epoch": 1163.7,
|
284565 |
+
"learning_rate": 7.66546042003231e-06,
|
284566 |
+
"loss": 0.293,
|
284567 |
+
"step": 145565
|
284568 |
+
},
|
284569 |
+
{
|
284570 |
+
"epoch": 1163.74,
|
284571 |
+
"learning_rate": 7.665379644588046e-06,
|
284572 |
+
"loss": 0.3133,
|
284573 |
+
"step": 145570
|
284574 |
+
},
|
284575 |
+
{
|
284576 |
+
"epoch": 1163.78,
|
284577 |
+
"learning_rate": 7.66529886914378e-06,
|
284578 |
+
"loss": 0.693,
|
284579 |
+
"step": 145575
|
284580 |
+
},
|
284581 |
+
{
|
284582 |
+
"epoch": 1163.82,
|
284583 |
+
"learning_rate": 7.665218093699516e-06,
|
284584 |
+
"loss": 0.9545,
|
284585 |
+
"step": 145580
|
284586 |
+
},
|
284587 |
+
{
|
284588 |
+
"epoch": 1163.86,
|
284589 |
+
"learning_rate": 7.66513731825525e-06,
|
284590 |
+
"loss": 0.3008,
|
284591 |
+
"step": 145585
|
284592 |
+
},
|
284593 |
+
{
|
284594 |
+
"epoch": 1163.9,
|
284595 |
+
"learning_rate": 7.665056542810986e-06,
|
284596 |
+
"loss": 0.3013,
|
284597 |
+
"step": 145590
|
284598 |
+
},
|
284599 |
+
{
|
284600 |
+
"epoch": 1163.94,
|
284601 |
+
"learning_rate": 7.66497576736672e-06,
|
284602 |
+
"loss": 0.538,
|
284603 |
+
"step": 145595
|
284604 |
+
},
|
284605 |
+
{
|
284606 |
+
"epoch": 1163.98,
|
284607 |
+
"learning_rate": 7.664894991922456e-06,
|
284608 |
+
"loss": 0.7369,
|
284609 |
+
"step": 145600
|
284610 |
+
},
|
284611 |
+
{
|
284612 |
+
"epoch": 1164.0,
|
284613 |
+
"eval_loss": 0.3466392159461975,
|
284614 |
+
"eval_runtime": 39.3229,
|
284615 |
+
"eval_samples_per_second": 21.285,
|
284616 |
+
"eval_steps_per_second": 0.687,
|
284617 |
+
"eval_wer": 0.16716790141568252,
|
284618 |
+
"step": 145602
|
284619 |
+
},
|
284620 |
+
{
|
284621 |
+
"epoch": 1164.02,
|
284622 |
+
"learning_rate": 7.66481421647819e-06,
|
284623 |
+
"loss": 0.3273,
|
284624 |
+
"step": 145605
|
284625 |
+
},
|
284626 |
+
{
|
284627 |
+
"epoch": 1164.06,
|
284628 |
+
"learning_rate": 7.664733441033926e-06,
|
284629 |
+
"loss": 0.2867,
|
284630 |
+
"step": 145610
|
284631 |
+
},
|
284632 |
+
{
|
284633 |
+
"epoch": 1164.1,
|
284634 |
+
"learning_rate": 7.664652665589662e-06,
|
284635 |
+
"loss": 0.3066,
|
284636 |
+
"step": 145615
|
284637 |
+
},
|
284638 |
+
{
|
284639 |
+
"epoch": 1164.14,
|
284640 |
+
"learning_rate": 7.664571890145396e-06,
|
284641 |
+
"loss": 0.3405,
|
284642 |
+
"step": 145620
|
284643 |
+
},
|
284644 |
+
{
|
284645 |
+
"epoch": 1164.18,
|
284646 |
+
"learning_rate": 7.664491114701132e-06,
|
284647 |
+
"loss": 0.7475,
|
284648 |
+
"step": 145625
|
284649 |
+
},
|
284650 |
+
{
|
284651 |
+
"epoch": 1164.22,
|
284652 |
+
"learning_rate": 7.664410339256866e-06,
|
284653 |
+
"loss": 0.9464,
|
284654 |
+
"step": 145630
|
284655 |
+
},
|
284656 |
+
{
|
284657 |
+
"epoch": 1164.26,
|
284658 |
+
"learning_rate": 7.664329563812602e-06,
|
284659 |
+
"loss": 0.3219,
|
284660 |
+
"step": 145635
|
284661 |
+
},
|
284662 |
+
{
|
284663 |
+
"epoch": 1164.3,
|
284664 |
+
"learning_rate": 7.664248788368336e-06,
|
284665 |
+
"loss": 0.3095,
|
284666 |
+
"step": 145640
|
284667 |
+
},
|
284668 |
+
{
|
284669 |
+
"epoch": 1164.34,
|
284670 |
+
"learning_rate": 7.664168012924072e-06,
|
284671 |
+
"loss": 0.3189,
|
284672 |
+
"step": 145645
|
284673 |
+
},
|
284674 |
+
{
|
284675 |
+
"epoch": 1164.38,
|
284676 |
+
"learning_rate": 7.664087237479806e-06,
|
284677 |
+
"loss": 0.7207,
|
284678 |
+
"step": 145650
|
284679 |
+
},
|
284680 |
+
{
|
284681 |
+
"epoch": 1164.42,
|
284682 |
+
"learning_rate": 7.664006462035542e-06,
|
284683 |
+
"loss": 0.8142,
|
284684 |
+
"step": 145655
|
284685 |
+
},
|
284686 |
+
{
|
284687 |
+
"epoch": 1164.46,
|
284688 |
+
"learning_rate": 7.663925686591276e-06,
|
284689 |
+
"loss": 0.2481,
|
284690 |
+
"step": 145660
|
284691 |
+
},
|
284692 |
+
{
|
284693 |
+
"epoch": 1164.5,
|
284694 |
+
"learning_rate": 7.663844911147012e-06,
|
284695 |
+
"loss": 0.2636,
|
284696 |
+
"step": 145665
|
284697 |
+
},
|
284698 |
+
{
|
284699 |
+
"epoch": 1164.54,
|
284700 |
+
"learning_rate": 7.663764135702746e-06,
|
284701 |
+
"loss": 0.3821,
|
284702 |
+
"step": 145670
|
284703 |
+
},
|
284704 |
+
{
|
284705 |
+
"epoch": 1164.58,
|
284706 |
+
"learning_rate": 7.663683360258482e-06,
|
284707 |
+
"loss": 0.7099,
|
284708 |
+
"step": 145675
|
284709 |
+
},
|
284710 |
+
{
|
284711 |
+
"epoch": 1164.62,
|
284712 |
+
"learning_rate": 7.663602584814218e-06,
|
284713 |
+
"loss": 0.7693,
|
284714 |
+
"step": 145680
|
284715 |
+
},
|
284716 |
+
{
|
284717 |
+
"epoch": 1164.66,
|
284718 |
+
"learning_rate": 7.663521809369952e-06,
|
284719 |
+
"loss": 0.3184,
|
284720 |
+
"step": 145685
|
284721 |
+
},
|
284722 |
+
{
|
284723 |
+
"epoch": 1164.7,
|
284724 |
+
"learning_rate": 7.663441033925688e-06,
|
284725 |
+
"loss": 0.2638,
|
284726 |
+
"step": 145690
|
284727 |
+
},
|
284728 |
+
{
|
284729 |
+
"epoch": 1164.74,
|
284730 |
+
"learning_rate": 7.663360258481422e-06,
|
284731 |
+
"loss": 0.4138,
|
284732 |
+
"step": 145695
|
284733 |
+
},
|
284734 |
+
{
|
284735 |
+
"epoch": 1164.78,
|
284736 |
+
"learning_rate": 7.663279483037158e-06,
|
284737 |
+
"loss": 0.7006,
|
284738 |
+
"step": 145700
|
284739 |
+
},
|
284740 |
+
{
|
284741 |
+
"epoch": 1164.82,
|
284742 |
+
"learning_rate": 7.663198707592892e-06,
|
284743 |
+
"loss": 0.9005,
|
284744 |
+
"step": 145705
|
284745 |
+
},
|
284746 |
+
{
|
284747 |
+
"epoch": 1164.86,
|
284748 |
+
"learning_rate": 7.663117932148628e-06,
|
284749 |
+
"loss": 0.2489,
|
284750 |
+
"step": 145710
|
284751 |
+
},
|
284752 |
+
{
|
284753 |
+
"epoch": 1164.9,
|
284754 |
+
"learning_rate": 7.663037156704362e-06,
|
284755 |
+
"loss": 0.2708,
|
284756 |
+
"step": 145715
|
284757 |
+
},
|
284758 |
+
{
|
284759 |
+
"epoch": 1164.94,
|
284760 |
+
"learning_rate": 7.662956381260098e-06,
|
284761 |
+
"loss": 0.3742,
|
284762 |
+
"step": 145720
|
284763 |
+
},
|
284764 |
+
{
|
284765 |
+
"epoch": 1164.98,
|
284766 |
+
"learning_rate": 7.662875605815832e-06,
|
284767 |
+
"loss": 0.7858,
|
284768 |
+
"step": 145725
|
284769 |
+
},
|
284770 |
+
{
|
284771 |
+
"epoch": 1165.0,
|
284772 |
+
"eval_loss": 0.38368138670921326,
|
284773 |
+
"eval_runtime": 40.6993,
|
284774 |
+
"eval_samples_per_second": 20.565,
|
284775 |
+
"eval_steps_per_second": 0.663,
|
284776 |
+
"eval_wer": 0.1798013485101138,
|
284777 |
+
"step": 145727
|
284778 |
+
},
|
284779 |
+
{
|
284780 |
+
"epoch": 1165.02,
|
284781 |
+
"learning_rate": 7.662794830371568e-06,
|
284782 |
+
"loss": 0.4956,
|
284783 |
+
"step": 145730
|
284784 |
+
},
|
284785 |
+
{
|
284786 |
+
"epoch": 1165.06,
|
284787 |
+
"learning_rate": 7.662714054927302e-06,
|
284788 |
+
"loss": 0.2444,
|
284789 |
+
"step": 145735
|
284790 |
+
},
|
284791 |
+
{
|
284792 |
+
"epoch": 1165.1,
|
284793 |
+
"learning_rate": 7.662633279483038e-06,
|
284794 |
+
"loss": 0.2835,
|
284795 |
+
"step": 145740
|
284796 |
+
},
|
284797 |
+
{
|
284798 |
+
"epoch": 1165.14,
|
284799 |
+
"learning_rate": 7.662552504038772e-06,
|
284800 |
+
"loss": 0.4172,
|
284801 |
+
"step": 145745
|
284802 |
+
},
|
284803 |
+
{
|
284804 |
+
"epoch": 1165.18,
|
284805 |
+
"learning_rate": 7.662471728594508e-06,
|
284806 |
+
"loss": 0.656,
|
284807 |
+
"step": 145750
|
284808 |
+
},
|
284809 |
+
{
|
284810 |
+
"epoch": 1165.22,
|
284811 |
+
"learning_rate": 7.662390953150244e-06,
|
284812 |
+
"loss": 0.8101,
|
284813 |
+
"step": 145755
|
284814 |
+
},
|
284815 |
+
{
|
284816 |
+
"epoch": 1165.26,
|
284817 |
+
"learning_rate": 7.662310177705978e-06,
|
284818 |
+
"loss": 0.2991,
|
284819 |
+
"step": 145760
|
284820 |
+
},
|
284821 |
+
{
|
284822 |
+
"epoch": 1165.3,
|
284823 |
+
"learning_rate": 7.662229402261714e-06,
|
284824 |
+
"loss": 0.2851,
|
284825 |
+
"step": 145765
|
284826 |
+
},
|
284827 |
+
{
|
284828 |
+
"epoch": 1165.34,
|
284829 |
+
"learning_rate": 7.662148626817448e-06,
|
284830 |
+
"loss": 0.3243,
|
284831 |
+
"step": 145770
|
284832 |
+
},
|
284833 |
+
{
|
284834 |
+
"epoch": 1165.38,
|
284835 |
+
"learning_rate": 7.662067851373183e-06,
|
284836 |
+
"loss": 0.7754,
|
284837 |
+
"step": 145775
|
284838 |
+
},
|
284839 |
+
{
|
284840 |
+
"epoch": 1165.42,
|
284841 |
+
"learning_rate": 7.661987075928918e-06,
|
284842 |
+
"loss": 0.9141,
|
284843 |
+
"step": 145780
|
284844 |
+
},
|
284845 |
+
{
|
284846 |
+
"epoch": 1165.46,
|
284847 |
+
"learning_rate": 7.661906300484653e-06,
|
284848 |
+
"loss": 0.2886,
|
284849 |
+
"step": 145785
|
284850 |
+
},
|
284851 |
+
{
|
284852 |
+
"epoch": 1165.5,
|
284853 |
+
"learning_rate": 7.661825525040388e-06,
|
284854 |
+
"loss": 0.305,
|
284855 |
+
"step": 145790
|
284856 |
+
},
|
284857 |
+
{
|
284858 |
+
"epoch": 1165.54,
|
284859 |
+
"learning_rate": 7.661744749596123e-06,
|
284860 |
+
"loss": 0.5382,
|
284861 |
+
"step": 145795
|
284862 |
+
},
|
284863 |
+
{
|
284864 |
+
"epoch": 1165.58,
|
284865 |
+
"learning_rate": 7.661663974151858e-06,
|
284866 |
+
"loss": 0.7651,
|
284867 |
+
"step": 145800
|
284868 |
+
},
|
284869 |
+
{
|
284870 |
+
"epoch": 1165.62,
|
284871 |
+
"learning_rate": 7.661583198707593e-06,
|
284872 |
+
"loss": 0.8388,
|
284873 |
+
"step": 145805
|
284874 |
+
},
|
284875 |
+
{
|
284876 |
+
"epoch": 1165.66,
|
284877 |
+
"learning_rate": 7.661502423263328e-06,
|
284878 |
+
"loss": 0.2597,
|
284879 |
+
"step": 145810
|
284880 |
+
},
|
284881 |
+
{
|
284882 |
+
"epoch": 1165.7,
|
284883 |
+
"learning_rate": 7.661421647819063e-06,
|
284884 |
+
"loss": 0.3229,
|
284885 |
+
"step": 145815
|
284886 |
+
},
|
284887 |
+
{
|
284888 |
+
"epoch": 1165.74,
|
284889 |
+
"learning_rate": 7.6613408723748e-06,
|
284890 |
+
"loss": 0.376,
|
284891 |
+
"step": 145820
|
284892 |
+
},
|
284893 |
+
{
|
284894 |
+
"epoch": 1165.78,
|
284895 |
+
"learning_rate": 7.661260096930533e-06,
|
284896 |
+
"loss": 0.6114,
|
284897 |
+
"step": 145825
|
284898 |
+
},
|
284899 |
+
{
|
284900 |
+
"epoch": 1165.82,
|
284901 |
+
"learning_rate": 7.66117932148627e-06,
|
284902 |
+
"loss": 0.8276,
|
284903 |
+
"step": 145830
|
284904 |
+
},
|
284905 |
+
{
|
284906 |
+
"epoch": 1165.86,
|
284907 |
+
"learning_rate": 7.661098546042003e-06,
|
284908 |
+
"loss": 0.2959,
|
284909 |
+
"step": 145835
|
284910 |
+
},
|
284911 |
+
{
|
284912 |
+
"epoch": 1165.9,
|
284913 |
+
"learning_rate": 7.66101777059774e-06,
|
284914 |
+
"loss": 0.2697,
|
284915 |
+
"step": 145840
|
284916 |
+
},
|
284917 |
+
{
|
284918 |
+
"epoch": 1165.94,
|
284919 |
+
"learning_rate": 7.660936995153473e-06,
|
284920 |
+
"loss": 0.3293,
|
284921 |
+
"step": 145845
|
284922 |
+
},
|
284923 |
+
{
|
284924 |
+
"epoch": 1165.98,
|
284925 |
+
"learning_rate": 7.660856219709209e-06,
|
284926 |
+
"loss": 0.7433,
|
284927 |
+
"step": 145850
|
284928 |
+
},
|
284929 |
+
{
|
284930 |
+
"epoch": 1166.0,
|
284931 |
+
"eval_loss": 0.33688074350357056,
|
284932 |
+
"eval_runtime": 41.2402,
|
284933 |
+
"eval_samples_per_second": 20.296,
|
284934 |
+
"eval_steps_per_second": 0.655,
|
284935 |
+
"eval_wer": 0.17012599228024178,
|
284936 |
+
"step": 145852
|
284937 |
}
|
284938 |
],
|
284939 |
+
"max_steps": 625000,
|
284940 |
"num_train_epochs": 5000,
|
284941 |
+
"total_flos": 4.104335476145737e+20,
|
284942 |
"trial_name": null,
|
284943 |
"trial_params": null
|
284944 |
}
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630226071.8984504/events.out.tfevents.1630226071.cc93b136ebf5.1086.99
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:950fda5fb6891bedbe19f04d898e4337085688846519f0cbf7e32d1f53a30bd7
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630226502.2753584/events.out.tfevents.1630226502.cc93b136ebf5.1086.101
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81e8a348434d43107b4bd23687b9036740d2656d44a85f95e0d48e8300c8960c
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630226942.7046182/events.out.tfevents.1630226942.cc93b136ebf5.1086.103
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65f84bc91cf2166792868abf376269402376f96ba7f36b17d1de5093ece4e578
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630227371.5174663/events.out.tfevents.1630227371.cc93b136ebf5.1086.105
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8138ddf57ae87fdc1b2496a5e815e82551c7bcac688069fc777167f16722bdba
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630227801.075265/events.out.tfevents.1630227801.cc93b136ebf5.1086.107
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b0c2ab0f952d96d77872c519ec99fddd0963e1be7c80ec37907c5e6271933548
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630226071.cc93b136ebf5.1086.98
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:831aa23a57f57c612b30af2c93fce20093b1b6a90cdf699d02c0d9b4538e0398
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630226502.cc93b136ebf5.1086.100
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5adf3892020b461c62a7cb3d13d02a2fe11fd02bb7189c650d98eefd2a55cf6e
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630226942.cc93b136ebf5.1086.102
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:04f87e0868a5a3a52b05bfbd849b69ed1c47f4407e5bc528eef400e93aac58c5
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630227371.cc93b136ebf5.1086.104
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1fd1780dc42c525f366f6d19f3c687da0f1f65bce9aa18df4267de2cfeec52a0
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630227801.cc93b136ebf5.1086.106
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14440456e8452b5d272fd86d57e8bd92f9707975bda22c4536cf19dadc451de5
|
3 |
+
size 8622
|