rakhman-llm commited on
Commit
fd6fa6c
·
verified ·
1 Parent(s): 23f738b

Training in progress, step 10500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5eb8bf06dd538fcb262c2fed2f9e68d7952360b525571db7ca0f1430e447d9ed
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:013308b4dc1251389723706bb70a4c12dd3e0f1c0451dc06722fb7fff47c38dc
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:942c2875abedaddb5c9436b198254369aa4f7b28a3b4b68fd6fcf41053e028fd
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:580eafdf89a0de8a5bf71611b2d46376bc396d69d739d11c02cf0fa1c01e5d26
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a7ea66d1531ee601a7eaa0403a8186669a6c59db91f4cad349d74bc0115c72f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b375d24b4f51731ac4974850b0a2bf3cfbc153b3c9a53e800d669342a8ff2a30
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7631547026bec9cd7a6ea58b5f8fb2fb117b688cae951965f0f9ff628a1476de
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62b81d85aecb38c3a9a2c050795de755f75a43f89e29b6a5cc7c6ab514e2f67e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.08249519765377045,
3
- "best_model_checkpoint": "./fine-tuned/checkpoint-10000",
4
- "epoch": 1.6,
5
  "eval_steps": 500,
6
- "global_step": 10000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1567,6 +1567,84 @@
1567
  "eval_samples_per_second": 17.101,
1568
  "eval_steps_per_second": 2.138,
1569
  "step": 10000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1570
  }
1571
  ],
1572
  "logging_steps": 50,
@@ -1586,7 +1664,7 @@
1586
  "attributes": {}
1587
  }
1588
  },
1589
- "total_flos": 4.87166312448e+16,
1590
  "train_batch_size": 8,
1591
  "trial_name": null,
1592
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.08225961029529572,
3
+ "best_model_checkpoint": "./fine-tuned/checkpoint-10500",
4
+ "epoch": 1.6800000000000002,
5
  "eval_steps": 500,
6
+ "global_step": 10500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1567
  "eval_samples_per_second": 17.101,
1568
  "eval_steps_per_second": 2.138,
1569
  "step": 10000
1570
+ },
1571
+ {
1572
+ "epoch": 1.608,
1573
+ "grad_norm": 5240.361328125,
1574
+ "learning_rate": 1.392e-05,
1575
+ "loss": 0.0546,
1576
+ "step": 10050
1577
+ },
1578
+ {
1579
+ "epoch": 1.616,
1580
+ "grad_norm": 7000.00927734375,
1581
+ "learning_rate": 1.384e-05,
1582
+ "loss": 0.0535,
1583
+ "step": 10100
1584
+ },
1585
+ {
1586
+ "epoch": 1.624,
1587
+ "grad_norm": 8141.75048828125,
1588
+ "learning_rate": 1.376e-05,
1589
+ "loss": 0.0555,
1590
+ "step": 10150
1591
+ },
1592
+ {
1593
+ "epoch": 1.6320000000000001,
1594
+ "grad_norm": 6566.3662109375,
1595
+ "learning_rate": 1.3680000000000001e-05,
1596
+ "loss": 0.0518,
1597
+ "step": 10200
1598
+ },
1599
+ {
1600
+ "epoch": 1.6400000000000001,
1601
+ "grad_norm": 7028.8935546875,
1602
+ "learning_rate": 1.36e-05,
1603
+ "loss": 0.0572,
1604
+ "step": 10250
1605
+ },
1606
+ {
1607
+ "epoch": 1.6480000000000001,
1608
+ "grad_norm": 13007.5703125,
1609
+ "learning_rate": 1.352e-05,
1610
+ "loss": 0.0567,
1611
+ "step": 10300
1612
+ },
1613
+ {
1614
+ "epoch": 1.6560000000000001,
1615
+ "grad_norm": 6286.06640625,
1616
+ "learning_rate": 1.344e-05,
1617
+ "loss": 0.0529,
1618
+ "step": 10350
1619
+ },
1620
+ {
1621
+ "epoch": 1.6640000000000001,
1622
+ "grad_norm": 6360.68408203125,
1623
+ "learning_rate": 1.336e-05,
1624
+ "loss": 0.054,
1625
+ "step": 10400
1626
+ },
1627
+ {
1628
+ "epoch": 1.6720000000000002,
1629
+ "grad_norm": 8098.84228515625,
1630
+ "learning_rate": 1.328e-05,
1631
+ "loss": 0.0592,
1632
+ "step": 10450
1633
+ },
1634
+ {
1635
+ "epoch": 1.6800000000000002,
1636
+ "grad_norm": 6886.65283203125,
1637
+ "learning_rate": 1.32e-05,
1638
+ "loss": 0.0524,
1639
+ "step": 10500
1640
+ },
1641
+ {
1642
+ "epoch": 1.6800000000000002,
1643
+ "eval_loss": 0.08225961029529572,
1644
+ "eval_runtime": 116.8647,
1645
+ "eval_samples_per_second": 17.114,
1646
+ "eval_steps_per_second": 2.139,
1647
+ "step": 10500
1648
  }
1649
  ],
1650
  "logging_steps": 50,
 
1664
  "attributes": {}
1665
  }
1666
  },
1667
+ "total_flos": 5.115246280704e+16,
1668
  "train_batch_size": 8,
1669
  "trial_name": null,
1670
  "trial_params": null