kodcode / trainer_state.json
sedrickkeh's picture
End of training
ae0a3b1 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.985163204747774,
"eval_steps": 500,
"global_step": 735,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006782534972445952,
"grad_norm": 7.135720261444181,
"learning_rate": 1.0810810810810812e-06,
"loss": 1.1988,
"step": 1
},
{
"epoch": 0.013565069944891903,
"grad_norm": 7.10061152713567,
"learning_rate": 2.1621621621621623e-06,
"loss": 1.1926,
"step": 2
},
{
"epoch": 0.020347604917337857,
"grad_norm": 7.012644871929241,
"learning_rate": 3.2432432432432437e-06,
"loss": 1.1873,
"step": 3
},
{
"epoch": 0.027130139889783807,
"grad_norm": 6.4062690637507425,
"learning_rate": 4.324324324324325e-06,
"loss": 1.1664,
"step": 4
},
{
"epoch": 0.03391267486222976,
"grad_norm": 4.819533337520507,
"learning_rate": 5.405405405405406e-06,
"loss": 1.1316,
"step": 5
},
{
"epoch": 0.040695209834675714,
"grad_norm": 3.0815069633752463,
"learning_rate": 6.486486486486487e-06,
"loss": 1.0897,
"step": 6
},
{
"epoch": 0.04747774480712166,
"grad_norm": 4.965671310290254,
"learning_rate": 7.567567567567569e-06,
"loss": 1.0541,
"step": 7
},
{
"epoch": 0.05426027977956761,
"grad_norm": 5.523597681999487,
"learning_rate": 8.64864864864865e-06,
"loss": 1.0812,
"step": 8
},
{
"epoch": 0.06104281475201356,
"grad_norm": 4.7758478538296245,
"learning_rate": 9.729729729729732e-06,
"loss": 1.0615,
"step": 9
},
{
"epoch": 0.06782534972445951,
"grad_norm": 4.37471758206583,
"learning_rate": 1.0810810810810812e-05,
"loss": 1.0034,
"step": 10
},
{
"epoch": 0.07460788469690546,
"grad_norm": 3.415263530374433,
"learning_rate": 1.1891891891891894e-05,
"loss": 0.9965,
"step": 11
},
{
"epoch": 0.08139041966935143,
"grad_norm": 2.1249671206986167,
"learning_rate": 1.2972972972972975e-05,
"loss": 0.9672,
"step": 12
},
{
"epoch": 0.08817295464179738,
"grad_norm": 2.804510150071328,
"learning_rate": 1.4054054054054055e-05,
"loss": 0.9248,
"step": 13
},
{
"epoch": 0.09495548961424333,
"grad_norm": 2.227059944094318,
"learning_rate": 1.5135135135135138e-05,
"loss": 0.9174,
"step": 14
},
{
"epoch": 0.10173802458668928,
"grad_norm": 1.9376154207127252,
"learning_rate": 1.6216216216216218e-05,
"loss": 0.9111,
"step": 15
},
{
"epoch": 0.10852055955913523,
"grad_norm": 1.5867898298795111,
"learning_rate": 1.72972972972973e-05,
"loss": 0.8742,
"step": 16
},
{
"epoch": 0.11530309453158118,
"grad_norm": 1.824405438277367,
"learning_rate": 1.8378378378378383e-05,
"loss": 0.8663,
"step": 17
},
{
"epoch": 0.12208562950402713,
"grad_norm": 1.2484939184792223,
"learning_rate": 1.9459459459459463e-05,
"loss": 0.8685,
"step": 18
},
{
"epoch": 0.1288681644764731,
"grad_norm": 1.6950608297204033,
"learning_rate": 2.054054054054054e-05,
"loss": 0.8643,
"step": 19
},
{
"epoch": 0.13565069944891903,
"grad_norm": 1.4257834623155268,
"learning_rate": 2.1621621621621624e-05,
"loss": 0.8589,
"step": 20
},
{
"epoch": 0.142433234421365,
"grad_norm": 1.2788529677027458,
"learning_rate": 2.2702702702702705e-05,
"loss": 0.832,
"step": 21
},
{
"epoch": 0.14921576939381093,
"grad_norm": 1.4653330199646428,
"learning_rate": 2.378378378378379e-05,
"loss": 0.8257,
"step": 22
},
{
"epoch": 0.1559983043662569,
"grad_norm": 1.1706491825282224,
"learning_rate": 2.4864864864864866e-05,
"loss": 0.8154,
"step": 23
},
{
"epoch": 0.16278083933870285,
"grad_norm": 1.575724350793329,
"learning_rate": 2.594594594594595e-05,
"loss": 0.83,
"step": 24
},
{
"epoch": 0.1695633743111488,
"grad_norm": 1.273387962120446,
"learning_rate": 2.702702702702703e-05,
"loss": 0.8327,
"step": 25
},
{
"epoch": 0.17634590928359475,
"grad_norm": 1.5867731702030288,
"learning_rate": 2.810810810810811e-05,
"loss": 0.8204,
"step": 26
},
{
"epoch": 0.1831284442560407,
"grad_norm": 0.8564117760497798,
"learning_rate": 2.918918918918919e-05,
"loss": 0.8056,
"step": 27
},
{
"epoch": 0.18991097922848665,
"grad_norm": 1.4715395766174262,
"learning_rate": 3.0270270270270275e-05,
"loss": 0.8055,
"step": 28
},
{
"epoch": 0.1966935142009326,
"grad_norm": 0.8375306490548892,
"learning_rate": 3.135135135135135e-05,
"loss": 0.7937,
"step": 29
},
{
"epoch": 0.20347604917337855,
"grad_norm": 1.620124215802785,
"learning_rate": 3.2432432432432436e-05,
"loss": 0.8001,
"step": 30
},
{
"epoch": 0.2102585841458245,
"grad_norm": 1.553317483743022,
"learning_rate": 3.351351351351351e-05,
"loss": 0.7876,
"step": 31
},
{
"epoch": 0.21704111911827045,
"grad_norm": 1.658909246335399,
"learning_rate": 3.45945945945946e-05,
"loss": 0.7891,
"step": 32
},
{
"epoch": 0.22382365409071642,
"grad_norm": 1.1987131034054377,
"learning_rate": 3.567567567567568e-05,
"loss": 0.7979,
"step": 33
},
{
"epoch": 0.23060618906316235,
"grad_norm": 2.2091757491329065,
"learning_rate": 3.6756756756756765e-05,
"loss": 0.7921,
"step": 34
},
{
"epoch": 0.23738872403560832,
"grad_norm": 2.173619464441479,
"learning_rate": 3.783783783783784e-05,
"loss": 0.7997,
"step": 35
},
{
"epoch": 0.24417125900805425,
"grad_norm": 1.3174649366078477,
"learning_rate": 3.8918918918918926e-05,
"loss": 0.768,
"step": 36
},
{
"epoch": 0.2509537939805002,
"grad_norm": 2.560545381997849,
"learning_rate": 4e-05,
"loss": 0.7938,
"step": 37
},
{
"epoch": 0.2577363289529462,
"grad_norm": 1.6460978741311298,
"learning_rate": 4.108108108108108e-05,
"loss": 0.7873,
"step": 38
},
{
"epoch": 0.2645188639253921,
"grad_norm": 2.1426591768978938,
"learning_rate": 4.2162162162162164e-05,
"loss": 0.781,
"step": 39
},
{
"epoch": 0.27130139889783805,
"grad_norm": 1.8160361307625077,
"learning_rate": 4.324324324324325e-05,
"loss": 0.7809,
"step": 40
},
{
"epoch": 0.278083933870284,
"grad_norm": 1.6910805765281172,
"learning_rate": 4.4324324324324325e-05,
"loss": 0.7744,
"step": 41
},
{
"epoch": 0.28486646884273,
"grad_norm": 1.9801882908067985,
"learning_rate": 4.540540540540541e-05,
"loss": 0.7732,
"step": 42
},
{
"epoch": 0.29164900381517594,
"grad_norm": 1.3912495139596244,
"learning_rate": 4.6486486486486486e-05,
"loss": 0.778,
"step": 43
},
{
"epoch": 0.29843153878762185,
"grad_norm": 1.9510858316270183,
"learning_rate": 4.756756756756758e-05,
"loss": 0.7628,
"step": 44
},
{
"epoch": 0.3052140737600678,
"grad_norm": 1.9755444838845442,
"learning_rate": 4.8648648648648654e-05,
"loss": 0.77,
"step": 45
},
{
"epoch": 0.3119966087325138,
"grad_norm": 1.7136629431095296,
"learning_rate": 4.972972972972973e-05,
"loss": 0.7776,
"step": 46
},
{
"epoch": 0.31877914370495974,
"grad_norm": 1.9262891096799022,
"learning_rate": 5.081081081081081e-05,
"loss": 0.7591,
"step": 47
},
{
"epoch": 0.3255616786774057,
"grad_norm": 1.8644045069770525,
"learning_rate": 5.18918918918919e-05,
"loss": 0.7564,
"step": 48
},
{
"epoch": 0.3323442136498516,
"grad_norm": 1.273771704881583,
"learning_rate": 5.2972972972972976e-05,
"loss": 0.7615,
"step": 49
},
{
"epoch": 0.3391267486222976,
"grad_norm": 2.4421063866094843,
"learning_rate": 5.405405405405406e-05,
"loss": 0.7803,
"step": 50
},
{
"epoch": 0.34590928359474354,
"grad_norm": 2.8264810278747197,
"learning_rate": 5.513513513513514e-05,
"loss": 0.7614,
"step": 51
},
{
"epoch": 0.3526918185671895,
"grad_norm": 2.2601831362375036,
"learning_rate": 5.621621621621622e-05,
"loss": 0.7727,
"step": 52
},
{
"epoch": 0.3594743535396354,
"grad_norm": 3.49645659835363,
"learning_rate": 5.7297297297297305e-05,
"loss": 0.7589,
"step": 53
},
{
"epoch": 0.3662568885120814,
"grad_norm": 2.8843912596082983,
"learning_rate": 5.837837837837838e-05,
"loss": 0.7561,
"step": 54
},
{
"epoch": 0.37303942348452734,
"grad_norm": 2.7371626553055965,
"learning_rate": 5.945945945945946e-05,
"loss": 0.7468,
"step": 55
},
{
"epoch": 0.3798219584569733,
"grad_norm": 2.4650676398708997,
"learning_rate": 6.054054054054055e-05,
"loss": 0.7501,
"step": 56
},
{
"epoch": 0.38660449342941927,
"grad_norm": 2.617962450345242,
"learning_rate": 6.162162162162163e-05,
"loss": 0.7504,
"step": 57
},
{
"epoch": 0.3933870284018652,
"grad_norm": 1.9045313378628999,
"learning_rate": 6.27027027027027e-05,
"loss": 0.7443,
"step": 58
},
{
"epoch": 0.40016956337431114,
"grad_norm": 3.214875277973139,
"learning_rate": 6.378378378378379e-05,
"loss": 0.756,
"step": 59
},
{
"epoch": 0.4069520983467571,
"grad_norm": 2.607985787054137,
"learning_rate": 6.486486486486487e-05,
"loss": 0.7575,
"step": 60
},
{
"epoch": 0.41373463331920307,
"grad_norm": 2.462297061139689,
"learning_rate": 6.594594594594596e-05,
"loss": 0.7395,
"step": 61
},
{
"epoch": 0.420517168291649,
"grad_norm": 1.7928802854683021,
"learning_rate": 6.702702702702703e-05,
"loss": 0.7443,
"step": 62
},
{
"epoch": 0.42729970326409494,
"grad_norm": 2.8782764835293015,
"learning_rate": 6.810810810810811e-05,
"loss": 0.7574,
"step": 63
},
{
"epoch": 0.4340822382365409,
"grad_norm": 2.107005343355103,
"learning_rate": 6.91891891891892e-05,
"loss": 0.7524,
"step": 64
},
{
"epoch": 0.44086477320898687,
"grad_norm": 2.400586779165952,
"learning_rate": 7.027027027027028e-05,
"loss": 0.7539,
"step": 65
},
{
"epoch": 0.44764730818143283,
"grad_norm": 1.8984648015553665,
"learning_rate": 7.135135135135136e-05,
"loss": 0.7661,
"step": 66
},
{
"epoch": 0.45442984315387874,
"grad_norm": 2.701931984195687,
"learning_rate": 7.243243243243243e-05,
"loss": 0.7423,
"step": 67
},
{
"epoch": 0.4612123781263247,
"grad_norm": 3.1216702832116807,
"learning_rate": 7.351351351351353e-05,
"loss": 0.7558,
"step": 68
},
{
"epoch": 0.46799491309877067,
"grad_norm": 1.7791622339381934,
"learning_rate": 7.45945945945946e-05,
"loss": 0.7338,
"step": 69
},
{
"epoch": 0.47477744807121663,
"grad_norm": 2.283223713003048,
"learning_rate": 7.567567567567568e-05,
"loss": 0.7512,
"step": 70
},
{
"epoch": 0.48155998304366254,
"grad_norm": 2.518755689574504,
"learning_rate": 7.675675675675675e-05,
"loss": 0.7371,
"step": 71
},
{
"epoch": 0.4883425180161085,
"grad_norm": 1.99481044416433,
"learning_rate": 7.783783783783785e-05,
"loss": 0.7478,
"step": 72
},
{
"epoch": 0.49512505298855447,
"grad_norm": 2.7463149548347183,
"learning_rate": 7.891891891891892e-05,
"loss": 0.7365,
"step": 73
},
{
"epoch": 0.5019075879610004,
"grad_norm": 2.138101333819343,
"learning_rate": 8e-05,
"loss": 0.7254,
"step": 74
},
{
"epoch": 0.5086901229334464,
"grad_norm": 2.640544503377941,
"learning_rate": 7.999954822103665e-05,
"loss": 0.7417,
"step": 75
},
{
"epoch": 0.5154726579058924,
"grad_norm": 2.2402410683129688,
"learning_rate": 7.999819289435179e-05,
"loss": 0.7354,
"step": 76
},
{
"epoch": 0.5222551928783383,
"grad_norm": 2.49341837988937,
"learning_rate": 7.999593405056084e-05,
"loss": 0.738,
"step": 77
},
{
"epoch": 0.5290377278507842,
"grad_norm": 2.3287753265703253,
"learning_rate": 7.999277174068872e-05,
"loss": 0.7247,
"step": 78
},
{
"epoch": 0.5358202628232301,
"grad_norm": 2.2524400032207716,
"learning_rate": 7.998870603616864e-05,
"loss": 0.7368,
"step": 79
},
{
"epoch": 0.5426027977956761,
"grad_norm": 2.686780411560504,
"learning_rate": 7.998373702884062e-05,
"loss": 0.7315,
"step": 80
},
{
"epoch": 0.5493853327681221,
"grad_norm": 1.7256862069861159,
"learning_rate": 7.997786483094931e-05,
"loss": 0.7492,
"step": 81
},
{
"epoch": 0.556167867740568,
"grad_norm": 2.6274112743201523,
"learning_rate": 7.997108957514146e-05,
"loss": 0.7229,
"step": 82
},
{
"epoch": 0.562950402713014,
"grad_norm": 2.0731276463095596,
"learning_rate": 7.9963411414463e-05,
"loss": 0.7375,
"step": 83
},
{
"epoch": 0.56973293768546,
"grad_norm": 2.6126411934480123,
"learning_rate": 7.99548305223555e-05,
"loss": 0.7284,
"step": 84
},
{
"epoch": 0.5765154726579059,
"grad_norm": 1.4633346753767393,
"learning_rate": 7.994534709265226e-05,
"loss": 0.7248,
"step": 85
},
{
"epoch": 0.5832980076303519,
"grad_norm": 3.4984880607635307,
"learning_rate": 7.993496133957401e-05,
"loss": 0.7517,
"step": 86
},
{
"epoch": 0.5900805426027977,
"grad_norm": 2.4422440886664147,
"learning_rate": 7.992367349772398e-05,
"loss": 0.7378,
"step": 87
},
{
"epoch": 0.5968630775752437,
"grad_norm": 1.8295430254229907,
"learning_rate": 7.991148382208265e-05,
"loss": 0.7241,
"step": 88
},
{
"epoch": 0.6036456125476897,
"grad_norm": 2.9619629036982755,
"learning_rate": 7.989839258800196e-05,
"loss": 0.7422,
"step": 89
},
{
"epoch": 0.6104281475201356,
"grad_norm": 1.9664968756669656,
"learning_rate": 7.988440009119911e-05,
"loss": 0.751,
"step": 90
},
{
"epoch": 0.6172106824925816,
"grad_norm": 1.6194583900993504,
"learning_rate": 7.986950664774992e-05,
"loss": 0.7346,
"step": 91
},
{
"epoch": 0.6239932174650276,
"grad_norm": 2.8148837821592796,
"learning_rate": 7.985371259408157e-05,
"loss": 0.7413,
"step": 92
},
{
"epoch": 0.6307757524374735,
"grad_norm": 1.486533554565982,
"learning_rate": 7.983701828696515e-05,
"loss": 0.7296,
"step": 93
},
{
"epoch": 0.6375582874099195,
"grad_norm": 3.229587748852655,
"learning_rate": 7.98194241035075e-05,
"loss": 0.7483,
"step": 94
},
{
"epoch": 0.6443408223823655,
"grad_norm": 2.518619876391423,
"learning_rate": 7.980093044114269e-05,
"loss": 0.7524,
"step": 95
},
{
"epoch": 0.6511233573548114,
"grad_norm": 2.6606747542222915,
"learning_rate": 7.978153771762311e-05,
"loss": 0.7531,
"step": 96
},
{
"epoch": 0.6579058923272573,
"grad_norm": 1.7367257209193254,
"learning_rate": 7.976124637101e-05,
"loss": 0.736,
"step": 97
},
{
"epoch": 0.6646884272997032,
"grad_norm": 1.8350871328487093,
"learning_rate": 7.974005685966354e-05,
"loss": 0.7248,
"step": 98
},
{
"epoch": 0.6714709622721492,
"grad_norm": 1.5007282529505472,
"learning_rate": 7.971796966223248e-05,
"loss": 0.7498,
"step": 99
},
{
"epoch": 0.6782534972445952,
"grad_norm": 1.4248528231943807,
"learning_rate": 7.969498527764341e-05,
"loss": 0.7208,
"step": 100
},
{
"epoch": 0.6850360322170411,
"grad_norm": 2.217134309648403,
"learning_rate": 7.967110422508936e-05,
"loss": 0.7295,
"step": 101
},
{
"epoch": 0.6918185671894871,
"grad_norm": 2.2429893368943277,
"learning_rate": 7.964632704401823e-05,
"loss": 0.7429,
"step": 102
},
{
"epoch": 0.698601102161933,
"grad_norm": 1.317977635899577,
"learning_rate": 7.962065429412046e-05,
"loss": 0.7311,
"step": 103
},
{
"epoch": 0.705383637134379,
"grad_norm": 3.1540990631312718,
"learning_rate": 7.959408655531646e-05,
"loss": 0.7331,
"step": 104
},
{
"epoch": 0.712166172106825,
"grad_norm": 2.0946781410468382,
"learning_rate": 7.956662442774351e-05,
"loss": 0.7357,
"step": 105
},
{
"epoch": 0.7189487070792708,
"grad_norm": 3.6514929315006834,
"learning_rate": 7.953826853174218e-05,
"loss": 0.724,
"step": 106
},
{
"epoch": 0.7257312420517168,
"grad_norm": 2.743797555116305,
"learning_rate": 7.950901950784236e-05,
"loss": 0.7337,
"step": 107
},
{
"epoch": 0.7325137770241628,
"grad_norm": 2.6025703631899364,
"learning_rate": 7.947887801674872e-05,
"loss": 0.7231,
"step": 108
},
{
"epoch": 0.7392963119966087,
"grad_norm": 2.33134578145788,
"learning_rate": 7.944784473932583e-05,
"loss": 0.7268,
"step": 109
},
{
"epoch": 0.7460788469690547,
"grad_norm": 2.369166475084867,
"learning_rate": 7.941592037658279e-05,
"loss": 0.7175,
"step": 110
},
{
"epoch": 0.7528613819415007,
"grad_norm": 1.5983852946337034,
"learning_rate": 7.93831056496574e-05,
"loss": 0.7215,
"step": 111
},
{
"epoch": 0.7596439169139466,
"grad_norm": 2.3446201220972696,
"learning_rate": 7.934940129979979e-05,
"loss": 0.7314,
"step": 112
},
{
"epoch": 0.7664264518863926,
"grad_norm": 1.872428876067427,
"learning_rate": 7.931480808835577e-05,
"loss": 0.7287,
"step": 113
},
{
"epoch": 0.7732089868588385,
"grad_norm": 1.982878649441548,
"learning_rate": 7.927932679674964e-05,
"loss": 0.7193,
"step": 114
},
{
"epoch": 0.7799915218312844,
"grad_norm": 1.5101882188042894,
"learning_rate": 7.924295822646643e-05,
"loss": 0.7217,
"step": 115
},
{
"epoch": 0.7867740568037304,
"grad_norm": 1.5820041393952022,
"learning_rate": 7.92057031990339e-05,
"loss": 0.7199,
"step": 116
},
{
"epoch": 0.7935565917761763,
"grad_norm": 1.5630207257272397,
"learning_rate": 7.91675625560039e-05,
"loss": 0.7141,
"step": 117
},
{
"epoch": 0.8003391267486223,
"grad_norm": 2.457059696195913,
"learning_rate": 7.91285371589335e-05,
"loss": 0.7082,
"step": 118
},
{
"epoch": 0.8071216617210683,
"grad_norm": 1.6028005358543407,
"learning_rate": 7.908862788936532e-05,
"loss": 0.7149,
"step": 119
},
{
"epoch": 0.8139041966935142,
"grad_norm": 1.8116395322587184,
"learning_rate": 7.904783564880779e-05,
"loss": 0.7123,
"step": 120
},
{
"epoch": 0.8206867316659602,
"grad_norm": 2.0950388250383996,
"learning_rate": 7.900616135871474e-05,
"loss": 0.7354,
"step": 121
},
{
"epoch": 0.8274692666384061,
"grad_norm": 1.8624393412855933,
"learning_rate": 7.896360596046453e-05,
"loss": 0.7156,
"step": 122
},
{
"epoch": 0.8342518016108521,
"grad_norm": 1.4851216429547307,
"learning_rate": 7.892017041533886e-05,
"loss": 0.7114,
"step": 123
},
{
"epoch": 0.841034336583298,
"grad_norm": 2.9385975867591765,
"learning_rate": 7.887585570450098e-05,
"loss": 0.7127,
"step": 124
},
{
"epoch": 0.8478168715557439,
"grad_norm": 1.8479656443451273,
"learning_rate": 7.883066282897362e-05,
"loss": 0.7173,
"step": 125
},
{
"epoch": 0.8545994065281899,
"grad_norm": 2.894143467483064,
"learning_rate": 7.878459280961629e-05,
"loss": 0.7183,
"step": 126
},
{
"epoch": 0.8613819415006359,
"grad_norm": 2.167706950028001,
"learning_rate": 7.873764668710228e-05,
"loss": 0.7237,
"step": 127
},
{
"epoch": 0.8681644764730818,
"grad_norm": 2.646470616726372,
"learning_rate": 7.868982552189514e-05,
"loss": 0.7132,
"step": 128
},
{
"epoch": 0.8749470114455278,
"grad_norm": 2.242758826547572,
"learning_rate": 7.864113039422464e-05,
"loss": 0.7296,
"step": 129
},
{
"epoch": 0.8817295464179737,
"grad_norm": 2.493893447513006,
"learning_rate": 7.859156240406252e-05,
"loss": 0.7185,
"step": 130
},
{
"epoch": 0.8885120813904197,
"grad_norm": 1.721360057791857,
"learning_rate": 7.854112267109756e-05,
"loss": 0.7244,
"step": 131
},
{
"epoch": 0.8952946163628657,
"grad_norm": 2.6683450004265583,
"learning_rate": 7.848981233471024e-05,
"loss": 0.7149,
"step": 132
},
{
"epoch": 0.9020771513353115,
"grad_norm": 2.061224678516105,
"learning_rate": 7.843763255394711e-05,
"loss": 0.7118,
"step": 133
},
{
"epoch": 0.9088596863077575,
"grad_norm": 1.9936485418092913,
"learning_rate": 7.838458450749452e-05,
"loss": 0.7057,
"step": 134
},
{
"epoch": 0.9156422212802034,
"grad_norm": 1.5127328421814061,
"learning_rate": 7.833066939365206e-05,
"loss": 0.7043,
"step": 135
},
{
"epoch": 0.9224247562526494,
"grad_norm": 2.040325487116571,
"learning_rate": 7.827588843030543e-05,
"loss": 0.7087,
"step": 136
},
{
"epoch": 0.9292072912250954,
"grad_norm": 1.6268257121266954,
"learning_rate": 7.822024285489896e-05,
"loss": 0.7105,
"step": 137
},
{
"epoch": 0.9359898261975413,
"grad_norm": 2.3150904956674263,
"learning_rate": 7.81637339244077e-05,
"loss": 0.7139,
"step": 138
},
{
"epoch": 0.9427723611699873,
"grad_norm": 1.841068832751663,
"learning_rate": 7.810636291530893e-05,
"loss": 0.7052,
"step": 139
},
{
"epoch": 0.9495548961424333,
"grad_norm": 1.9983783549939655,
"learning_rate": 7.804813112355339e-05,
"loss": 0.7115,
"step": 140
},
{
"epoch": 0.9563374311148792,
"grad_norm": 1.9754428939529725,
"learning_rate": 7.798903986453603e-05,
"loss": 0.7142,
"step": 141
},
{
"epoch": 0.9631199660873251,
"grad_norm": 1.735226023700741,
"learning_rate": 7.792909047306623e-05,
"loss": 0.7205,
"step": 142
},
{
"epoch": 0.969902501059771,
"grad_norm": 2.184380091482261,
"learning_rate": 7.786828430333769e-05,
"loss": 0.7107,
"step": 143
},
{
"epoch": 0.976685036032217,
"grad_norm": 1.8757020725047386,
"learning_rate": 7.78066227288978e-05,
"loss": 0.7198,
"step": 144
},
{
"epoch": 0.983467571004663,
"grad_norm": 1.0079051823953802,
"learning_rate": 7.77441071426167e-05,
"loss": 0.714,
"step": 145
},
{
"epoch": 0.9902501059771089,
"grad_norm": 1.5288723958450625,
"learning_rate": 7.768073895665573e-05,
"loss": 0.7087,
"step": 146
},
{
"epoch": 0.9970326409495549,
"grad_norm": 2.3013059467419996,
"learning_rate": 7.761651960243554e-05,
"loss": 0.7366,
"step": 147
},
{
"epoch": 1.0038151759220009,
"grad_norm": 2.269978916078414,
"learning_rate": 7.755145053060378e-05,
"loss": 1.1363,
"step": 148
},
{
"epoch": 1.0105977108944468,
"grad_norm": 1.6332374526993498,
"learning_rate": 7.748553321100238e-05,
"loss": 0.697,
"step": 149
},
{
"epoch": 1.0173802458668928,
"grad_norm": 2.0144072844842054,
"learning_rate": 7.741876913263422e-05,
"loss": 0.7169,
"step": 150
},
{
"epoch": 1.0241627808393388,
"grad_norm": 1.9963737399863932,
"learning_rate": 7.735115980362964e-05,
"loss": 0.6942,
"step": 151
},
{
"epoch": 1.0309453158117847,
"grad_norm": 2.010956402670031,
"learning_rate": 7.728270675121224e-05,
"loss": 0.7108,
"step": 152
},
{
"epoch": 1.0377278507842307,
"grad_norm": 1.1256640161683924,
"learning_rate": 7.721341152166448e-05,
"loss": 0.706,
"step": 153
},
{
"epoch": 1.0445103857566767,
"grad_norm": 2.7360045649430287,
"learning_rate": 7.714327568029272e-05,
"loss": 0.7035,
"step": 154
},
{
"epoch": 1.0512929207291226,
"grad_norm": 1.8273909265523915,
"learning_rate": 7.707230081139184e-05,
"loss": 0.7064,
"step": 155
},
{
"epoch": 1.0580754557015684,
"grad_norm": 2.155560932773446,
"learning_rate": 7.700048851820946e-05,
"loss": 0.6991,
"step": 156
},
{
"epoch": 1.0648579906740143,
"grad_norm": 2.1669039610273493,
"learning_rate": 7.692784042290976e-05,
"loss": 0.7099,
"step": 157
},
{
"epoch": 1.0716405256464603,
"grad_norm": 1.503612527062841,
"learning_rate": 7.685435816653681e-05,
"loss": 0.6973,
"step": 158
},
{
"epoch": 1.0784230606189062,
"grad_norm": 1.937309301259521,
"learning_rate": 7.678004340897747e-05,
"loss": 0.7049,
"step": 159
},
{
"epoch": 1.0852055955913522,
"grad_norm": 1.3969091003754739,
"learning_rate": 7.670489782892396e-05,
"loss": 0.6886,
"step": 160
},
{
"epoch": 1.0919881305637982,
"grad_norm": 1.9964168232948958,
"learning_rate": 7.662892312383592e-05,
"loss": 0.6883,
"step": 161
},
{
"epoch": 1.0987706655362441,
"grad_norm": 1.7569755198383514,
"learning_rate": 7.655212100990195e-05,
"loss": 0.6987,
"step": 162
},
{
"epoch": 1.10555320050869,
"grad_norm": 1.4447704185089303,
"learning_rate": 7.647449322200108e-05,
"loss": 0.6906,
"step": 163
},
{
"epoch": 1.112335735481136,
"grad_norm": 1.2497557944098203,
"learning_rate": 7.639604151366339e-05,
"loss": 0.7009,
"step": 164
},
{
"epoch": 1.119118270453582,
"grad_norm": 2.3116148247793267,
"learning_rate": 7.631676765703042e-05,
"loss": 0.6895,
"step": 165
},
{
"epoch": 1.125900805426028,
"grad_norm": 1.330812583383646,
"learning_rate": 7.623667344281522e-05,
"loss": 0.7004,
"step": 166
},
{
"epoch": 1.132683340398474,
"grad_norm": 1.8610464265540336,
"learning_rate": 7.615576068026187e-05,
"loss": 0.7077,
"step": 167
},
{
"epoch": 1.13946587537092,
"grad_norm": 1.7667897999541395,
"learning_rate": 7.607403119710453e-05,
"loss": 0.6913,
"step": 168
},
{
"epoch": 1.1462484103433659,
"grad_norm": 1.208336959762333,
"learning_rate": 7.599148683952628e-05,
"loss": 0.6839,
"step": 169
},
{
"epoch": 1.1530309453158119,
"grad_norm": 1.9033319484816924,
"learning_rate": 7.590812947211733e-05,
"loss": 0.6898,
"step": 170
},
{
"epoch": 1.1598134802882578,
"grad_norm": 1.213830580906014,
"learning_rate": 7.582396097783294e-05,
"loss": 0.6866,
"step": 171
},
{
"epoch": 1.1665960152607038,
"grad_norm": 2.309488724064315,
"learning_rate": 7.573898325795086e-05,
"loss": 0.697,
"step": 172
},
{
"epoch": 1.1733785502331497,
"grad_norm": 1.3337697012970882,
"learning_rate": 7.565319823202838e-05,
"loss": 0.7083,
"step": 173
},
{
"epoch": 1.1801610852055955,
"grad_norm": 1.644453886623139,
"learning_rate": 7.556660783785904e-05,
"loss": 0.6905,
"step": 174
},
{
"epoch": 1.1869436201780414,
"grad_norm": 1.3203501723312958,
"learning_rate": 7.547921403142874e-05,
"loss": 0.7048,
"step": 175
},
{
"epoch": 1.1937261551504874,
"grad_norm": 1.4545383798679439,
"learning_rate": 7.539101878687165e-05,
"loss": 0.6865,
"step": 176
},
{
"epoch": 1.2005086901229334,
"grad_norm": 2.1879709481996987,
"learning_rate": 7.53020240964256e-05,
"loss": 0.6964,
"step": 177
},
{
"epoch": 1.2072912250953793,
"grad_norm": 1.8322469465167428,
"learning_rate": 7.5212231970387e-05,
"loss": 0.6921,
"step": 178
},
{
"epoch": 1.2140737600678253,
"grad_norm": 1.2841938775963582,
"learning_rate": 7.512164443706555e-05,
"loss": 0.6816,
"step": 179
},
{
"epoch": 1.2208562950402713,
"grad_norm": 3.1832942767833003,
"learning_rate": 7.503026354273834e-05,
"loss": 0.7107,
"step": 180
},
{
"epoch": 1.2276388300127172,
"grad_norm": 2.223671181736338,
"learning_rate": 7.493809135160367e-05,
"loss": 0.7068,
"step": 181
},
{
"epoch": 1.2344213649851632,
"grad_norm": 2.5365710286141727,
"learning_rate": 7.484512994573438e-05,
"loss": 0.6997,
"step": 182
},
{
"epoch": 1.2412038999576092,
"grad_norm": 1.7033409104571895,
"learning_rate": 7.475138142503083e-05,
"loss": 0.6973,
"step": 183
},
{
"epoch": 1.2479864349300551,
"grad_norm": 2.666703064771071,
"learning_rate": 7.465684790717354e-05,
"loss": 0.6904,
"step": 184
},
{
"epoch": 1.254768969902501,
"grad_norm": 1.6449288965298707,
"learning_rate": 7.45615315275752e-05,
"loss": 0.7008,
"step": 185
},
{
"epoch": 1.261551504874947,
"grad_norm": 2.670948821208354,
"learning_rate": 7.446543443933258e-05,
"loss": 0.7187,
"step": 186
},
{
"epoch": 1.268334039847393,
"grad_norm": 2.052845388192151,
"learning_rate": 7.436855881317784e-05,
"loss": 0.7154,
"step": 187
},
{
"epoch": 1.275116574819839,
"grad_norm": 2.3638246999276054,
"learning_rate": 7.427090683742947e-05,
"loss": 0.6942,
"step": 188
},
{
"epoch": 1.281899109792285,
"grad_norm": 2.367916777331452,
"learning_rate": 7.417248071794288e-05,
"loss": 0.6978,
"step": 189
},
{
"epoch": 1.288681644764731,
"grad_norm": 1.919070978827847,
"learning_rate": 7.40732826780606e-05,
"loss": 0.6778,
"step": 190
},
{
"epoch": 1.2954641797371766,
"grad_norm": 1.5807447282945328,
"learning_rate": 7.397331495856199e-05,
"loss": 0.6878,
"step": 191
},
{
"epoch": 1.3022467147096228,
"grad_norm": 1.5725379589793347,
"learning_rate": 7.38725798176127e-05,
"loss": 0.6942,
"step": 192
},
{
"epoch": 1.3090292496820686,
"grad_norm": 1.5854647888888829,
"learning_rate": 7.37710795307136e-05,
"loss": 0.6942,
"step": 193
},
{
"epoch": 1.3158117846545148,
"grad_norm": 0.9840396505987244,
"learning_rate": 7.36688163906494e-05,
"loss": 0.6822,
"step": 194
},
{
"epoch": 1.3225943196269605,
"grad_norm": 1.9293846538238884,
"learning_rate": 7.356579270743689e-05,
"loss": 0.707,
"step": 195
},
{
"epoch": 1.3293768545994065,
"grad_norm": 1.297177627346237,
"learning_rate": 7.346201080827272e-05,
"loss": 0.6966,
"step": 196
},
{
"epoch": 1.3361593895718524,
"grad_norm": 2.3709366791443456,
"learning_rate": 7.335747303748079e-05,
"loss": 0.7001,
"step": 197
},
{
"epoch": 1.3429419245442984,
"grad_norm": 1.7251163867182135,
"learning_rate": 7.325218175645942e-05,
"loss": 0.694,
"step": 198
},
{
"epoch": 1.3497244595167444,
"grad_norm": 2.088301626825016,
"learning_rate": 7.314613934362788e-05,
"loss": 0.6879,
"step": 199
},
{
"epoch": 1.3565069944891903,
"grad_norm": 1.9529188107557567,
"learning_rate": 7.303934819437277e-05,
"loss": 0.6944,
"step": 200
},
{
"epoch": 1.3632895294616363,
"grad_norm": 1.6962527949805855,
"learning_rate": 7.293181072099377e-05,
"loss": 0.6917,
"step": 201
},
{
"epoch": 1.3700720644340822,
"grad_norm": 1.6169230360470628,
"learning_rate": 7.282352935264934e-05,
"loss": 0.684,
"step": 202
},
{
"epoch": 1.3768545994065282,
"grad_norm": 1.485732940861678,
"learning_rate": 7.271450653530167e-05,
"loss": 0.6862,
"step": 203
},
{
"epoch": 1.3836371343789742,
"grad_norm": 1.6704659351121771,
"learning_rate": 7.260474473166154e-05,
"loss": 0.6932,
"step": 204
},
{
"epoch": 1.3904196693514201,
"grad_norm": 1.5799837137385815,
"learning_rate": 7.249424642113266e-05,
"loss": 0.697,
"step": 205
},
{
"epoch": 1.397202204323866,
"grad_norm": 1.2977311485642662,
"learning_rate": 7.238301409975561e-05,
"loss": 0.6919,
"step": 206
},
{
"epoch": 1.403984739296312,
"grad_norm": 1.451392232428334,
"learning_rate": 7.227105028015156e-05,
"loss": 0.699,
"step": 207
},
{
"epoch": 1.410767274268758,
"grad_norm": 1.2941406989928823,
"learning_rate": 7.21583574914654e-05,
"loss": 0.6708,
"step": 208
},
{
"epoch": 1.417549809241204,
"grad_norm": 1.6593482637225496,
"learning_rate": 7.204493827930869e-05,
"loss": 0.6871,
"step": 209
},
{
"epoch": 1.4243323442136497,
"grad_norm": 1.528014842634458,
"learning_rate": 7.193079520570217e-05,
"loss": 0.6777,
"step": 210
},
{
"epoch": 1.431114879186096,
"grad_norm": 1.2836973493312132,
"learning_rate": 7.18159308490178e-05,
"loss": 0.6741,
"step": 211
},
{
"epoch": 1.4378974141585417,
"grad_norm": 1.2069873065957852,
"learning_rate": 7.170034780392055e-05,
"loss": 0.6852,
"step": 212
},
{
"epoch": 1.4446799491309876,
"grad_norm": 1.6652484448685945,
"learning_rate": 7.158404868130988e-05,
"loss": 0.6711,
"step": 213
},
{
"epoch": 1.4514624841034336,
"grad_norm": 1.2576599603912495,
"learning_rate": 7.14670361082606e-05,
"loss": 0.6849,
"step": 214
},
{
"epoch": 1.4582450190758796,
"grad_norm": 1.5405832213240351,
"learning_rate": 7.134931272796368e-05,
"loss": 0.691,
"step": 215
},
{
"epoch": 1.4650275540483255,
"grad_norm": 1.684185898740359,
"learning_rate": 7.123088119966647e-05,
"loss": 0.6868,
"step": 216
},
{
"epoch": 1.4718100890207715,
"grad_norm": 1.0827904024479826,
"learning_rate": 7.11117441986126e-05,
"loss": 0.6878,
"step": 217
},
{
"epoch": 1.4785926239932174,
"grad_norm": 1.5453825081747208,
"learning_rate": 7.099190441598161e-05,
"loss": 0.6838,
"step": 218
},
{
"epoch": 1.4853751589656634,
"grad_norm": 1.0691158300615973,
"learning_rate": 7.087136455882816e-05,
"loss": 0.6738,
"step": 219
},
{
"epoch": 1.4921576939381094,
"grad_norm": 2.077881844593604,
"learning_rate": 7.07501273500208e-05,
"loss": 0.6912,
"step": 220
},
{
"epoch": 1.4989402289105553,
"grad_norm": 0.9909091986869899,
"learning_rate": 7.06281955281806e-05,
"loss": 0.6875,
"step": 221
},
{
"epoch": 1.5057227638830013,
"grad_norm": 1.826572971343745,
"learning_rate": 7.050557184761915e-05,
"loss": 0.6855,
"step": 222
},
{
"epoch": 1.5125052988554473,
"grad_norm": 1.279113591847188,
"learning_rate": 7.038225907827639e-05,
"loss": 0.6961,
"step": 223
},
{
"epoch": 1.5192878338278932,
"grad_norm": 1.229774591769691,
"learning_rate": 7.02582600056581e-05,
"loss": 0.6802,
"step": 224
},
{
"epoch": 1.526070368800339,
"grad_norm": 1.3203007597562126,
"learning_rate": 7.013357743077289e-05,
"loss": 0.6965,
"step": 225
},
{
"epoch": 1.5328529037727852,
"grad_norm": 1.2862135502175833,
"learning_rate": 7.000821417006898e-05,
"loss": 0.6815,
"step": 226
},
{
"epoch": 1.539635438745231,
"grad_norm": 1.0628503441369677,
"learning_rate": 6.988217305537058e-05,
"loss": 0.6768,
"step": 227
},
{
"epoch": 1.546417973717677,
"grad_norm": 1.3535468021784238,
"learning_rate": 6.97554569338139e-05,
"loss": 0.6959,
"step": 228
},
{
"epoch": 1.5532005086901228,
"grad_norm": 1.1587176301112725,
"learning_rate": 6.962806866778284e-05,
"loss": 0.6821,
"step": 229
},
{
"epoch": 1.559983043662569,
"grad_norm": 1.6895118703555423,
"learning_rate": 6.950001113484432e-05,
"loss": 0.6832,
"step": 230
},
{
"epoch": 1.5667655786350148,
"grad_norm": 1.3111534840956083,
"learning_rate": 6.937128722768333e-05,
"loss": 0.6747,
"step": 231
},
{
"epoch": 1.573548113607461,
"grad_norm": 1.2243130191258123,
"learning_rate": 6.924189985403753e-05,
"loss": 0.69,
"step": 232
},
{
"epoch": 1.5803306485799067,
"grad_norm": 1.4287675639720405,
"learning_rate": 6.91118519366316e-05,
"loss": 0.6801,
"step": 233
},
{
"epoch": 1.5871131835523526,
"grad_norm": 0.9267438225948489,
"learning_rate": 6.898114641311122e-05,
"loss": 0.6845,
"step": 234
},
{
"epoch": 1.5938957185247986,
"grad_norm": 1.3352877429277719,
"learning_rate": 6.884978623597665e-05,
"loss": 0.6993,
"step": 235
},
{
"epoch": 1.6006782534972446,
"grad_norm": 1.334254924822497,
"learning_rate": 6.871777437251617e-05,
"loss": 0.6817,
"step": 236
},
{
"epoch": 1.6074607884696905,
"grad_norm": 1.2977369297336303,
"learning_rate": 6.858511380473887e-05,
"loss": 0.6856,
"step": 237
},
{
"epoch": 1.6142433234421365,
"grad_norm": 1.1900674515806826,
"learning_rate": 6.845180752930749e-05,
"loss": 0.6947,
"step": 238
},
{
"epoch": 1.6210258584145825,
"grad_norm": 2.7317108581757057,
"learning_rate": 6.831785855747054e-05,
"loss": 0.6985,
"step": 239
},
{
"epoch": 1.6278083933870284,
"grad_norm": 1.2267690972200642,
"learning_rate": 6.81832699149944e-05,
"loss": 0.6773,
"step": 240
},
{
"epoch": 1.6345909283594744,
"grad_norm": 3.8760882746190255,
"learning_rate": 6.804804464209498e-05,
"loss": 0.6895,
"step": 241
},
{
"epoch": 1.6413734633319204,
"grad_norm": 3.5268150192285157,
"learning_rate": 6.791218579336891e-05,
"loss": 0.6994,
"step": 242
},
{
"epoch": 1.6481559983043663,
"grad_norm": 1.7590143767536321,
"learning_rate": 6.777569643772472e-05,
"loss": 0.683,
"step": 243
},
{
"epoch": 1.654938533276812,
"grad_norm": 1.7753106738494688,
"learning_rate": 6.763857965831337e-05,
"loss": 0.6811,
"step": 244
},
{
"epoch": 1.6617210682492582,
"grad_norm": 1.7854517795067268,
"learning_rate": 6.750083855245869e-05,
"loss": 0.6818,
"step": 245
},
{
"epoch": 1.668503603221704,
"grad_norm": 1.2937361964461016,
"learning_rate": 6.736247623158738e-05,
"loss": 0.6825,
"step": 246
},
{
"epoch": 1.6752861381941502,
"grad_norm": 1.2452039091600213,
"learning_rate": 6.722349582115872e-05,
"loss": 0.6779,
"step": 247
},
{
"epoch": 1.682068673166596,
"grad_norm": 1.6654925410640917,
"learning_rate": 6.708390046059402e-05,
"loss": 0.6897,
"step": 248
},
{
"epoch": 1.688851208139042,
"grad_norm": 1.1695940537064082,
"learning_rate": 6.694369330320563e-05,
"loss": 0.6894,
"step": 249
},
{
"epoch": 1.6956337431114878,
"grad_norm": 1.790433077118171,
"learning_rate": 6.680287751612576e-05,
"loss": 0.6965,
"step": 250
},
{
"epoch": 1.702416278083934,
"grad_norm": 1.3940506254821923,
"learning_rate": 6.666145628023495e-05,
"loss": 0.6883,
"step": 251
},
{
"epoch": 1.7091988130563798,
"grad_norm": 1.5788853280407766,
"learning_rate": 6.651943279009015e-05,
"loss": 0.674,
"step": 252
},
{
"epoch": 1.7159813480288257,
"grad_norm": 1.135684283004857,
"learning_rate": 6.637681025385261e-05,
"loss": 0.6944,
"step": 253
},
{
"epoch": 1.7227638830012717,
"grad_norm": 1.4839512076861858,
"learning_rate": 6.62335918932154e-05,
"loss": 0.6752,
"step": 254
},
{
"epoch": 1.7295464179737177,
"grad_norm": 1.3150087814531235,
"learning_rate": 6.608978094333068e-05,
"loss": 0.6826,
"step": 255
},
{
"epoch": 1.7363289529461636,
"grad_norm": 1.0720984172971046,
"learning_rate": 6.594538065273652e-05,
"loss": 0.683,
"step": 256
},
{
"epoch": 1.7431114879186096,
"grad_norm": 1.46543732407952,
"learning_rate": 6.58003942832836e-05,
"loss": 0.6796,
"step": 257
},
{
"epoch": 1.7498940228910556,
"grad_norm": 0.9292343800735161,
"learning_rate": 6.56548251100615e-05,
"loss": 0.6686,
"step": 258
},
{
"epoch": 1.7566765578635015,
"grad_norm": 1.5664491391186621,
"learning_rate": 6.550867642132474e-05,
"loss": 0.6912,
"step": 259
},
{
"epoch": 1.7634590928359475,
"grad_norm": 0.9821764593726072,
"learning_rate": 6.536195151841847e-05,
"loss": 0.6702,
"step": 260
},
{
"epoch": 1.7702416278083932,
"grad_norm": 1.4298526249936745,
"learning_rate": 6.521465371570393e-05,
"loss": 0.6871,
"step": 261
},
{
"epoch": 1.7770241627808394,
"grad_norm": 1.2452682220878895,
"learning_rate": 6.506678634048353e-05,
"loss": 0.6677,
"step": 262
},
{
"epoch": 1.7838066977532852,
"grad_norm": 1.4920349233263195,
"learning_rate": 6.491835273292574e-05,
"loss": 0.6809,
"step": 263
},
{
"epoch": 1.7905892327257313,
"grad_norm": 1.212833170547237,
"learning_rate": 6.476935624598966e-05,
"loss": 0.6809,
"step": 264
},
{
"epoch": 1.797371767698177,
"grad_norm": 1.250429726318394,
"learning_rate": 6.461980024534918e-05,
"loss": 0.674,
"step": 265
},
{
"epoch": 1.8041543026706233,
"grad_norm": 1.0681520589021245,
"learning_rate": 6.446968810931707e-05,
"loss": 0.6767,
"step": 266
},
{
"epoch": 1.810936837643069,
"grad_norm": 1.470191760394408,
"learning_rate": 6.431902322876855e-05,
"loss": 0.6737,
"step": 267
},
{
"epoch": 1.8177193726155152,
"grad_norm": 1.1749343585137977,
"learning_rate": 6.416780900706484e-05,
"loss": 0.683,
"step": 268
},
{
"epoch": 1.824501907587961,
"grad_norm": 1.1727141312870375,
"learning_rate": 6.401604885997614e-05,
"loss": 0.6808,
"step": 269
},
{
"epoch": 1.831284442560407,
"grad_norm": 1.3495790960778822,
"learning_rate": 6.386374621560455e-05,
"loss": 0.6847,
"step": 270
},
{
"epoch": 1.8380669775328529,
"grad_norm": 1.075708125802829,
"learning_rate": 6.37109045143066e-05,
"loss": 0.676,
"step": 271
},
{
"epoch": 1.8448495125052988,
"grad_norm": 1.1555696102312405,
"learning_rate": 6.355752720861559e-05,
"loss": 0.6661,
"step": 272
},
{
"epoch": 1.8516320474777448,
"grad_norm": 1.383594134484124,
"learning_rate": 6.340361776316349e-05,
"loss": 0.6784,
"step": 273
},
{
"epoch": 1.8584145824501908,
"grad_norm": 1.063837234136454,
"learning_rate": 6.324917965460279e-05,
"loss": 0.6781,
"step": 274
},
{
"epoch": 1.8651971174226367,
"grad_norm": 1.173609581411256,
"learning_rate": 6.309421637152794e-05,
"loss": 0.6682,
"step": 275
},
{
"epoch": 1.8719796523950827,
"grad_norm": 1.0658200919149257,
"learning_rate": 6.29387314143965e-05,
"loss": 0.6878,
"step": 276
},
{
"epoch": 1.8787621873675286,
"grad_norm": 1.7400399121218273,
"learning_rate": 6.278272829545011e-05,
"loss": 0.6706,
"step": 277
},
{
"epoch": 1.8855447223399746,
"grad_norm": 0.7229472871914281,
"learning_rate": 6.26262105386351e-05,
"loss": 0.6847,
"step": 278
},
{
"epoch": 1.8923272573124206,
"grad_norm": 1.0742091230131698,
"learning_rate": 6.246918167952304e-05,
"loss": 0.6756,
"step": 279
},
{
"epoch": 1.8991097922848663,
"grad_norm": 2.3846707594542456,
"learning_rate": 6.231164526523063e-05,
"loss": 0.6776,
"step": 280
},
{
"epoch": 1.9058923272573125,
"grad_norm": 1.181568933400366,
"learning_rate": 6.21536048543398e-05,
"loss": 0.686,
"step": 281
},
{
"epoch": 1.9126748622297582,
"grad_norm": 3.197777220254654,
"learning_rate": 6.199506401681718e-05,
"loss": 0.7007,
"step": 282
},
{
"epoch": 1.9194573972022044,
"grad_norm": 2.538862401423035,
"learning_rate": 6.183602633393352e-05,
"loss": 0.7031,
"step": 283
},
{
"epoch": 1.9262399321746502,
"grad_norm": 1.8353914421208335,
"learning_rate": 6.167649539818284e-05,
"loss": 0.6663,
"step": 284
},
{
"epoch": 1.9330224671470964,
"grad_norm": 1.8245299177531569,
"learning_rate": 6.151647481320114e-05,
"loss": 0.7011,
"step": 285
},
{
"epoch": 1.939805002119542,
"grad_norm": 2.284062594732862,
"learning_rate": 6.135596819368512e-05,
"loss": 0.701,
"step": 286
},
{
"epoch": 1.9465875370919883,
"grad_norm": 1.5939451814656143,
"learning_rate": 6.119497916531053e-05,
"loss": 0.6898,
"step": 287
},
{
"epoch": 1.953370072064434,
"grad_norm": 2.382128196264195,
"learning_rate": 6.103351136465014e-05,
"loss": 0.689,
"step": 288
},
{
"epoch": 1.96015260703688,
"grad_norm": 2.255809021386828,
"learning_rate": 6.0871568439091743e-05,
"loss": 0.689,
"step": 289
},
{
"epoch": 1.966935142009326,
"grad_norm": 1.192524289628154,
"learning_rate": 6.070915404675571e-05,
"loss": 0.6709,
"step": 290
},
{
"epoch": 1.973717676981772,
"grad_norm": 1.2397061535202811,
"learning_rate": 6.0546271856412306e-05,
"loss": 0.687,
"step": 291
},
{
"epoch": 1.9805002119542179,
"grad_norm": 0.9812283662135952,
"learning_rate": 6.03829255473989e-05,
"loss": 0.677,
"step": 292
},
{
"epoch": 1.9872827469266638,
"grad_norm": 1.410310313066498,
"learning_rate": 6.0219118809536794e-05,
"loss": 0.6882,
"step": 293
},
{
"epoch": 1.9940652818991098,
"grad_norm": 0.7865238012992424,
"learning_rate": 6.0054855343047914e-05,
"loss": 0.6769,
"step": 294
},
{
"epoch": 2.0008478168715556,
"grad_norm": 1.9901508305957583,
"learning_rate": 5.989013885847117e-05,
"loss": 1.0693,
"step": 295
},
{
"epoch": 2.0076303518440017,
"grad_norm": 1.3663813100431552,
"learning_rate": 5.972497307657869e-05,
"loss": 0.6673,
"step": 296
},
{
"epoch": 2.0144128868164475,
"grad_norm": 1.1381480345523374,
"learning_rate": 5.955936172829179e-05,
"loss": 0.6676,
"step": 297
},
{
"epoch": 2.0211954217888937,
"grad_norm": 1.496250854193103,
"learning_rate": 5.939330855459661e-05,
"loss": 0.6551,
"step": 298
},
{
"epoch": 2.0279779567613394,
"grad_norm": 0.9716458462062207,
"learning_rate": 5.922681730645968e-05,
"loss": 0.6723,
"step": 299
},
{
"epoch": 2.0347604917337856,
"grad_norm": 1.501595368602735,
"learning_rate": 5.905989174474319e-05,
"loss": 0.6509,
"step": 300
},
{
"epoch": 2.0415430267062313,
"grad_norm": 0.849974962385986,
"learning_rate": 5.889253564011999e-05,
"loss": 0.6577,
"step": 301
},
{
"epoch": 2.0483255616786775,
"grad_norm": 1.418902591957157,
"learning_rate": 5.872475277298847e-05,
"loss": 0.6665,
"step": 302
},
{
"epoch": 2.0551080966511233,
"grad_norm": 0.8446815446152741,
"learning_rate": 5.855654693338711e-05,
"loss": 0.6556,
"step": 303
},
{
"epoch": 2.0618906316235694,
"grad_norm": 1.3364133345670484,
"learning_rate": 5.838792192090889e-05,
"loss": 0.6604,
"step": 304
},
{
"epoch": 2.068673166596015,
"grad_norm": 0.8691215260428371,
"learning_rate": 5.821888154461549e-05,
"loss": 0.6593,
"step": 305
},
{
"epoch": 2.0754557015684614,
"grad_norm": 1.2207875769882877,
"learning_rate": 5.8049429622951194e-05,
"loss": 0.6637,
"step": 306
},
{
"epoch": 2.082238236540907,
"grad_norm": 0.7775512997267359,
"learning_rate": 5.7879569983656694e-05,
"loss": 0.6433,
"step": 307
},
{
"epoch": 2.0890207715133533,
"grad_norm": 1.2237758983601073,
"learning_rate": 5.770930646368257e-05,
"loss": 0.6681,
"step": 308
},
{
"epoch": 2.095803306485799,
"grad_norm": 0.6935518063946065,
"learning_rate": 5.7538642909102654e-05,
"loss": 0.6571,
"step": 309
},
{
"epoch": 2.1025858414582452,
"grad_norm": 1.0509844665628039,
"learning_rate": 5.736758317502714e-05,
"loss": 0.6462,
"step": 310
},
{
"epoch": 2.109368376430691,
"grad_norm": 0.6501336655556583,
"learning_rate": 5.719613112551546e-05,
"loss": 0.642,
"step": 311
},
{
"epoch": 2.1161509114031367,
"grad_norm": 0.791518345366507,
"learning_rate": 5.702429063348912e-05,
"loss": 0.6523,
"step": 312
},
{
"epoch": 2.122933446375583,
"grad_norm": 0.7216839703720191,
"learning_rate": 5.685206558064407e-05,
"loss": 0.6485,
"step": 313
},
{
"epoch": 2.1297159813480286,
"grad_norm": 1.3490427513390442,
"learning_rate": 5.667945985736308e-05,
"loss": 0.6622,
"step": 314
},
{
"epoch": 2.136498516320475,
"grad_norm": 1.036563799894598,
"learning_rate": 5.6506477362627926e-05,
"loss": 0.6598,
"step": 315
},
{
"epoch": 2.1432810512929206,
"grad_norm": 0.8799949127384937,
"learning_rate": 5.6333122003931186e-05,
"loss": 0.6631,
"step": 316
},
{
"epoch": 2.1500635862653668,
"grad_norm": 0.9027804053204243,
"learning_rate": 5.615939769718809e-05,
"loss": 0.6543,
"step": 317
},
{
"epoch": 2.1568461212378125,
"grad_norm": 1.08723379469273,
"learning_rate": 5.5985308366647985e-05,
"loss": 0.6453,
"step": 318
},
{
"epoch": 2.1636286562102587,
"grad_norm": 1.198788894798635,
"learning_rate": 5.5810857944805744e-05,
"loss": 0.6466,
"step": 319
},
{
"epoch": 2.1704111911827044,
"grad_norm": 1.2287340016478474,
"learning_rate": 5.5636050372312896e-05,
"loss": 0.6776,
"step": 320
},
{
"epoch": 2.1771937261551506,
"grad_norm": 0.676307375412045,
"learning_rate": 5.546088959788862e-05,
"loss": 0.6517,
"step": 321
},
{
"epoch": 2.1839762611275964,
"grad_norm": 0.8023647253320252,
"learning_rate": 5.528537957823061e-05,
"loss": 0.6565,
"step": 322
},
{
"epoch": 2.1907587961000425,
"grad_norm": 1.4391382554036933,
"learning_rate": 5.510952427792559e-05,
"loss": 0.6587,
"step": 323
},
{
"epoch": 2.1975413310724883,
"grad_norm": 0.7633580340639281,
"learning_rate": 5.493332766935981e-05,
"loss": 0.6433,
"step": 324
},
{
"epoch": 2.2043238660449345,
"grad_norm": 1.0303353819004128,
"learning_rate": 5.4756793732629335e-05,
"loss": 0.6578,
"step": 325
},
{
"epoch": 2.21110640101738,
"grad_norm": 1.3526690779961053,
"learning_rate": 5.45799264554501e-05,
"loss": 0.6499,
"step": 326
},
{
"epoch": 2.2178889359898264,
"grad_norm": 0.5228459526464206,
"learning_rate": 5.440272983306789e-05,
"loss": 0.6394,
"step": 327
},
{
"epoch": 2.224671470962272,
"grad_norm": 1.4252351086818837,
"learning_rate": 5.4225207868167994e-05,
"loss": 0.6554,
"step": 328
},
{
"epoch": 2.231454005934718,
"grad_norm": 0.6950731562608328,
"learning_rate": 5.404736457078489e-05,
"loss": 0.6482,
"step": 329
},
{
"epoch": 2.238236540907164,
"grad_norm": 0.7644651875362564,
"learning_rate": 5.38692039582116e-05,
"loss": 0.6503,
"step": 330
},
{
"epoch": 2.24501907587961,
"grad_norm": 0.9023571857971306,
"learning_rate": 5.3690730054908985e-05,
"loss": 0.6468,
"step": 331
},
{
"epoch": 2.251801610852056,
"grad_norm": 0.9107802448508358,
"learning_rate": 5.3511946892414775e-05,
"loss": 0.6518,
"step": 332
},
{
"epoch": 2.2585841458245017,
"grad_norm": 1.1127754837574242,
"learning_rate": 5.333285850925256e-05,
"loss": 0.6565,
"step": 333
},
{
"epoch": 2.265366680796948,
"grad_norm": 0.8282253306938299,
"learning_rate": 5.315346895084056e-05,
"loss": 0.6564,
"step": 334
},
{
"epoch": 2.2721492157693937,
"grad_norm": 1.14130477616649,
"learning_rate": 5.297378226940019e-05,
"loss": 0.6516,
"step": 335
},
{
"epoch": 2.27893175074184,
"grad_norm": 1.0103971093744055,
"learning_rate": 5.279380252386461e-05,
"loss": 0.6626,
"step": 336
},
{
"epoch": 2.2857142857142856,
"grad_norm": 0.9536394369627204,
"learning_rate": 5.2613533779786945e-05,
"loss": 0.6663,
"step": 337
},
{
"epoch": 2.2924968206867318,
"grad_norm": 0.942528066117481,
"learning_rate": 5.243298010924852e-05,
"loss": 0.6408,
"step": 338
},
{
"epoch": 2.2992793556591775,
"grad_norm": 1.2117876187133245,
"learning_rate": 5.225214559076683e-05,
"loss": 0.6435,
"step": 339
},
{
"epoch": 2.3060618906316237,
"grad_norm": 0.8633776831955887,
"learning_rate": 5.207103430920345e-05,
"loss": 0.6395,
"step": 340
},
{
"epoch": 2.3128444256040694,
"grad_norm": 0.727531307349217,
"learning_rate": 5.1889650355671725e-05,
"loss": 0.654,
"step": 341
},
{
"epoch": 2.3196269605765156,
"grad_norm": 0.669365998637684,
"learning_rate": 5.1707997827444394e-05,
"loss": 0.6546,
"step": 342
},
{
"epoch": 2.3264094955489614,
"grad_norm": 0.8312703348077676,
"learning_rate": 5.152608082786098e-05,
"loss": 0.6646,
"step": 343
},
{
"epoch": 2.3331920305214076,
"grad_norm": 1.0187489492222126,
"learning_rate": 5.1343903466235174e-05,
"loss": 0.648,
"step": 344
},
{
"epoch": 2.3399745654938533,
"grad_norm": 1.034867027111884,
"learning_rate": 5.116146985776194e-05,
"loss": 0.6593,
"step": 345
},
{
"epoch": 2.3467571004662995,
"grad_norm": 1.1577447007405788,
"learning_rate": 5.0978784123424617e-05,
"loss": 0.6516,
"step": 346
},
{
"epoch": 2.3535396354387452,
"grad_norm": 1.0011140671455747,
"learning_rate": 5.0795850389901784e-05,
"loss": 0.6638,
"step": 347
},
{
"epoch": 2.360322170411191,
"grad_norm": 0.9415943625257757,
"learning_rate": 5.061267278947408e-05,
"loss": 0.6457,
"step": 348
},
{
"epoch": 2.367104705383637,
"grad_norm": 0.6919138650974885,
"learning_rate": 5.042925545993079e-05,
"loss": 0.6584,
"step": 349
},
{
"epoch": 2.373887240356083,
"grad_norm": 0.7266901542517143,
"learning_rate": 5.02456025444765e-05,
"loss": 0.6348,
"step": 350
},
{
"epoch": 2.380669775328529,
"grad_norm": 0.6635241827908179,
"learning_rate": 5.0061718191637394e-05,
"loss": 0.6574,
"step": 351
},
{
"epoch": 2.387452310300975,
"grad_norm": 0.5653141179051638,
"learning_rate": 4.987760655516757e-05,
"loss": 0.6502,
"step": 352
},
{
"epoch": 2.394234845273421,
"grad_norm": 0.7007173012785425,
"learning_rate": 4.9693271793955255e-05,
"loss": 0.6468,
"step": 353
},
{
"epoch": 2.4010173802458668,
"grad_norm": 0.5821543619857267,
"learning_rate": 4.95087180719288e-05,
"loss": 0.6486,
"step": 354
},
{
"epoch": 2.407799915218313,
"grad_norm": 0.4265984966323706,
"learning_rate": 4.9323949557962684e-05,
"loss": 0.6477,
"step": 355
},
{
"epoch": 2.4145824501907587,
"grad_norm": 0.6454773425594047,
"learning_rate": 4.913897042578327e-05,
"loss": 0.6479,
"step": 356
},
{
"epoch": 2.421364985163205,
"grad_norm": 0.615481868155692,
"learning_rate": 4.8953784853874624e-05,
"loss": 0.6605,
"step": 357
},
{
"epoch": 2.4281475201356506,
"grad_norm": 0.4923342474878763,
"learning_rate": 4.8768397025383996e-05,
"loss": 0.6414,
"step": 358
},
{
"epoch": 2.434930055108097,
"grad_norm": 0.4236872576222672,
"learning_rate": 4.858281112802745e-05,
"loss": 0.6526,
"step": 359
},
{
"epoch": 2.4417125900805425,
"grad_norm": 0.4376255004390753,
"learning_rate": 4.83970313539952e-05,
"loss": 0.6503,
"step": 360
},
{
"epoch": 2.4484951250529887,
"grad_norm": 0.7271305316785697,
"learning_rate": 4.821106189985693e-05,
"loss": 0.6447,
"step": 361
},
{
"epoch": 2.4552776600254345,
"grad_norm": 1.001429290369445,
"learning_rate": 4.8024906966467e-05,
"loss": 0.6566,
"step": 362
},
{
"epoch": 2.4620601949978806,
"grad_norm": 1.168304354121221,
"learning_rate": 4.783857075886956e-05,
"loss": 0.6614,
"step": 363
},
{
"epoch": 2.4688427299703264,
"grad_norm": 0.8021141929873027,
"learning_rate": 4.7652057486203525e-05,
"loss": 0.6446,
"step": 364
},
{
"epoch": 2.4756252649427726,
"grad_norm": 0.7075159940006046,
"learning_rate": 4.746537136160757e-05,
"loss": 0.6563,
"step": 365
},
{
"epoch": 2.4824077999152183,
"grad_norm": 0.7183111983007646,
"learning_rate": 4.727851660212487e-05,
"loss": 0.6502,
"step": 366
},
{
"epoch": 2.489190334887664,
"grad_norm": 0.755055422137936,
"learning_rate": 4.709149742860792e-05,
"loss": 0.6424,
"step": 367
},
{
"epoch": 2.4959728698601102,
"grad_norm": 0.7253802127810576,
"learning_rate": 4.690431806562311e-05,
"loss": 0.6496,
"step": 368
},
{
"epoch": 2.5027554048325564,
"grad_norm": 0.5125324963207099,
"learning_rate": 4.6716982741355386e-05,
"loss": 0.6533,
"step": 369
},
{
"epoch": 2.509537939805002,
"grad_norm": 0.5597998347086172,
"learning_rate": 4.652949568751267e-05,
"loss": 0.6402,
"step": 370
},
{
"epoch": 2.516320474777448,
"grad_norm": 0.7015467668995325,
"learning_rate": 4.63418611392303e-05,
"loss": 0.6482,
"step": 371
},
{
"epoch": 2.523103009749894,
"grad_norm": 0.511973221694181,
"learning_rate": 4.615408333497538e-05,
"loss": 0.6522,
"step": 372
},
{
"epoch": 2.52988554472234,
"grad_norm": 0.4401280437938751,
"learning_rate": 4.5966166516450985e-05,
"loss": 0.6573,
"step": 373
},
{
"epoch": 2.536668079694786,
"grad_norm": 0.614469648497891,
"learning_rate": 4.577811492850039e-05,
"loss": 0.6598,
"step": 374
},
{
"epoch": 2.5434506146672318,
"grad_norm": 0.7173941689992674,
"learning_rate": 4.558993281901116e-05,
"loss": 0.6465,
"step": 375
},
{
"epoch": 2.550233149639678,
"grad_norm": 0.6690409730578648,
"learning_rate": 4.540162443881922e-05,
"loss": 0.6488,
"step": 376
},
{
"epoch": 2.5570156846121237,
"grad_norm": 0.5848162533642309,
"learning_rate": 4.5213194041612824e-05,
"loss": 0.648,
"step": 377
},
{
"epoch": 2.56379821958457,
"grad_norm": 0.549981546670745,
"learning_rate": 4.5024645883836426e-05,
"loss": 0.6579,
"step": 378
},
{
"epoch": 2.5705807545570156,
"grad_norm": 0.556140263499365,
"learning_rate": 4.4835984224594586e-05,
"loss": 0.6621,
"step": 379
},
{
"epoch": 2.577363289529462,
"grad_norm": 0.5195224064491745,
"learning_rate": 4.464721332555577e-05,
"loss": 0.6512,
"step": 380
},
{
"epoch": 2.5841458245019076,
"grad_norm": 0.38259318048535124,
"learning_rate": 4.445833745085602e-05,
"loss": 0.6492,
"step": 381
},
{
"epoch": 2.5909283594743533,
"grad_norm": 0.3534037726943067,
"learning_rate": 4.4269360867002675e-05,
"loss": 0.6589,
"step": 382
},
{
"epoch": 2.5977108944467995,
"grad_norm": 0.315508715276319,
"learning_rate": 4.408028784277799e-05,
"loss": 0.6493,
"step": 383
},
{
"epoch": 2.6044934294192457,
"grad_norm": 0.3298656413250479,
"learning_rate": 4.389112264914273e-05,
"loss": 0.6589,
"step": 384
},
{
"epoch": 2.6112759643916914,
"grad_norm": 0.3016047037119661,
"learning_rate": 4.370186955913962e-05,
"loss": 0.6561,
"step": 385
},
{
"epoch": 2.618058499364137,
"grad_norm": 0.34845242594318765,
"learning_rate": 4.351253284779692e-05,
"loss": 0.6552,
"step": 386
},
{
"epoch": 2.6248410343365833,
"grad_norm": 0.38736846664169255,
"learning_rate": 4.332311679203177e-05,
"loss": 0.6364,
"step": 387
},
{
"epoch": 2.6316235693090295,
"grad_norm": 0.3951534285172612,
"learning_rate": 4.313362567055367e-05,
"loss": 0.6502,
"step": 388
},
{
"epoch": 2.6384061042814753,
"grad_norm": 0.4657313097001591,
"learning_rate": 4.294406376376771e-05,
"loss": 0.6561,
"step": 389
},
{
"epoch": 2.645188639253921,
"grad_norm": 0.4131171546046392,
"learning_rate": 4.2754435353677985e-05,
"loss": 0.6567,
"step": 390
},
{
"epoch": 2.651971174226367,
"grad_norm": 0.5305329556333895,
"learning_rate": 4.2564744723790835e-05,
"loss": 0.6568,
"step": 391
},
{
"epoch": 2.658753709198813,
"grad_norm": 0.7002749870897172,
"learning_rate": 4.237499615901805e-05,
"loss": 0.6552,
"step": 392
},
{
"epoch": 2.665536244171259,
"grad_norm": 0.7839360975679,
"learning_rate": 4.218519394558013e-05,
"loss": 0.6333,
"step": 393
},
{
"epoch": 2.672318779143705,
"grad_norm": 0.7621551938800204,
"learning_rate": 4.199534237090943e-05,
"loss": 0.6501,
"step": 394
},
{
"epoch": 2.679101314116151,
"grad_norm": 0.6652202397867919,
"learning_rate": 4.1805445723553346e-05,
"loss": 0.648,
"step": 395
},
{
"epoch": 2.685883849088597,
"grad_norm": 0.4811941000709009,
"learning_rate": 4.1615508293077394e-05,
"loss": 0.6544,
"step": 396
},
{
"epoch": 2.692666384061043,
"grad_norm": 0.3387872066498026,
"learning_rate": 4.142553436996834e-05,
"loss": 0.6558,
"step": 397
},
{
"epoch": 2.6994489190334887,
"grad_norm": 0.28441864482912343,
"learning_rate": 4.12355282455373e-05,
"loss": 0.6354,
"step": 398
},
{
"epoch": 2.706231454005935,
"grad_norm": 0.285244776358457,
"learning_rate": 4.1045494211822756e-05,
"loss": 0.6493,
"step": 399
},
{
"epoch": 2.7130139889783806,
"grad_norm": 0.3573803066995761,
"learning_rate": 4.085543656149365e-05,
"loss": 0.6397,
"step": 400
},
{
"epoch": 2.7197965239508264,
"grad_norm": 0.4278142282008419,
"learning_rate": 4.0665359587752394e-05,
"loss": 0.6537,
"step": 401
},
{
"epoch": 2.7265790589232726,
"grad_norm": 0.49944968397215816,
"learning_rate": 4.04752675842379e-05,
"loss": 0.656,
"step": 402
},
{
"epoch": 2.7333615938957188,
"grad_norm": 0.4642912024112797,
"learning_rate": 4.028516484492857e-05,
"loss": 0.6424,
"step": 403
},
{
"epoch": 2.7401441288681645,
"grad_norm": 0.40098112979297407,
"learning_rate": 4.009505566404535e-05,
"loss": 0.6377,
"step": 404
},
{
"epoch": 2.7469266638406102,
"grad_norm": 0.3708879898016676,
"learning_rate": 3.990494433595466e-05,
"loss": 0.6518,
"step": 405
},
{
"epoch": 2.7537091988130564,
"grad_norm": 0.40008123571174364,
"learning_rate": 3.9714835155071435e-05,
"loss": 0.6536,
"step": 406
},
{
"epoch": 2.760491733785502,
"grad_norm": 0.29605943254539613,
"learning_rate": 3.952473241576211e-05,
"loss": 0.6583,
"step": 407
},
{
"epoch": 2.7672742687579484,
"grad_norm": 0.3406525853896961,
"learning_rate": 3.933464041224761e-05,
"loss": 0.6545,
"step": 408
},
{
"epoch": 2.774056803730394,
"grad_norm": 0.4722405867001987,
"learning_rate": 3.914456343850637e-05,
"loss": 0.6513,
"step": 409
},
{
"epoch": 2.7808393387028403,
"grad_norm": 0.5263658831108335,
"learning_rate": 3.895450578817727e-05,
"loss": 0.6548,
"step": 410
},
{
"epoch": 2.787621873675286,
"grad_norm": 0.5203524120767699,
"learning_rate": 3.8764471754462714e-05,
"loss": 0.654,
"step": 411
},
{
"epoch": 2.794404408647732,
"grad_norm": 0.3902260359882798,
"learning_rate": 3.857446563003167e-05,
"loss": 0.651,
"step": 412
},
{
"epoch": 2.801186943620178,
"grad_norm": 0.3931366323184314,
"learning_rate": 3.838449170692262e-05,
"loss": 0.6418,
"step": 413
},
{
"epoch": 2.807969478592624,
"grad_norm": 0.4496897165040769,
"learning_rate": 3.819455427644666e-05,
"loss": 0.6566,
"step": 414
},
{
"epoch": 2.81475201356507,
"grad_norm": 0.4843275111384109,
"learning_rate": 3.800465762909057e-05,
"loss": 0.6494,
"step": 415
},
{
"epoch": 2.821534548537516,
"grad_norm": 0.4284479617853188,
"learning_rate": 3.781480605441989e-05,
"loss": 0.6504,
"step": 416
},
{
"epoch": 2.828317083509962,
"grad_norm": 0.3505752717784012,
"learning_rate": 3.7625003840981976e-05,
"loss": 0.6515,
"step": 417
},
{
"epoch": 2.835099618482408,
"grad_norm": 0.2739600638831855,
"learning_rate": 3.743525527620918e-05,
"loss": 0.6482,
"step": 418
},
{
"epoch": 2.8418821534548537,
"grad_norm": 0.2821715826512274,
"learning_rate": 3.724556464632203e-05,
"loss": 0.6558,
"step": 419
},
{
"epoch": 2.8486646884272995,
"grad_norm": 0.3001235929581994,
"learning_rate": 3.7055936236232296e-05,
"loss": 0.6283,
"step": 420
},
{
"epoch": 2.8554472233997457,
"grad_norm": 0.34084577684820244,
"learning_rate": 3.686637432944634e-05,
"loss": 0.6524,
"step": 421
},
{
"epoch": 2.862229758372192,
"grad_norm": 0.31354392380727714,
"learning_rate": 3.6676883207968226e-05,
"loss": 0.6407,
"step": 422
},
{
"epoch": 2.8690122933446376,
"grad_norm": 0.29938503730372684,
"learning_rate": 3.648746715220309e-05,
"loss": 0.6431,
"step": 423
},
{
"epoch": 2.8757948283170833,
"grad_norm": 0.292588496703024,
"learning_rate": 3.6298130440860394e-05,
"loss": 0.6457,
"step": 424
},
{
"epoch": 2.8825773632895295,
"grad_norm": 0.34211158982130174,
"learning_rate": 3.6108877350857296e-05,
"loss": 0.6413,
"step": 425
},
{
"epoch": 2.8893598982619753,
"grad_norm": 0.28453013315503256,
"learning_rate": 3.5919712157222014e-05,
"loss": 0.6493,
"step": 426
},
{
"epoch": 2.8961424332344214,
"grad_norm": 0.25000620848294364,
"learning_rate": 3.573063913299733e-05,
"loss": 0.6471,
"step": 427
},
{
"epoch": 2.902924968206867,
"grad_norm": 0.32984361513817445,
"learning_rate": 3.554166254914399e-05,
"loss": 0.6525,
"step": 428
},
{
"epoch": 2.9097075031793134,
"grad_norm": 0.34282163750459993,
"learning_rate": 3.535278667444423e-05,
"loss": 0.6435,
"step": 429
},
{
"epoch": 2.916490038151759,
"grad_norm": 0.3376932266610861,
"learning_rate": 3.5164015775405414e-05,
"loss": 0.6537,
"step": 430
},
{
"epoch": 2.9232725731242053,
"grad_norm": 0.2736761009463742,
"learning_rate": 3.4975354116163594e-05,
"loss": 0.6528,
"step": 431
},
{
"epoch": 2.930055108096651,
"grad_norm": 0.28401811879767763,
"learning_rate": 3.478680595838719e-05,
"loss": 0.6552,
"step": 432
},
{
"epoch": 2.9368376430690972,
"grad_norm": 0.21207447278197997,
"learning_rate": 3.4598375561180783e-05,
"loss": 0.6477,
"step": 433
},
{
"epoch": 2.943620178041543,
"grad_norm": 0.3273666184303206,
"learning_rate": 3.441006718098885e-05,
"loss": 0.6415,
"step": 434
},
{
"epoch": 2.9504027130139887,
"grad_norm": 0.3922599048141651,
"learning_rate": 3.422188507149962e-05,
"loss": 0.66,
"step": 435
},
{
"epoch": 2.957185247986435,
"grad_norm": 0.33853238933941526,
"learning_rate": 3.403383348354902e-05,
"loss": 0.6392,
"step": 436
},
{
"epoch": 2.963967782958881,
"grad_norm": 0.2904409798403406,
"learning_rate": 3.3845916665024626e-05,
"loss": 0.6571,
"step": 437
},
{
"epoch": 2.970750317931327,
"grad_norm": 0.25650378002806623,
"learning_rate": 3.36581388607697e-05,
"loss": 0.658,
"step": 438
},
{
"epoch": 2.9775328529037726,
"grad_norm": 0.30516122417008495,
"learning_rate": 3.347050431248735e-05,
"loss": 0.6445,
"step": 439
},
{
"epoch": 2.9843153878762188,
"grad_norm": 0.35193441095679195,
"learning_rate": 3.3283017258644634e-05,
"loss": 0.6514,
"step": 440
},
{
"epoch": 2.991097922848665,
"grad_norm": 0.26810523993687996,
"learning_rate": 3.30956819343769e-05,
"loss": 0.6479,
"step": 441
},
{
"epoch": 2.9978804578211107,
"grad_norm": 0.40136478388949387,
"learning_rate": 3.290850257139209e-05,
"loss": 0.7446,
"step": 442
},
{
"epoch": 3.0046629927935564,
"grad_norm": 0.6219962321725231,
"learning_rate": 3.2721483397875135e-05,
"loss": 0.9011,
"step": 443
},
{
"epoch": 3.0114455277660026,
"grad_norm": 0.6301524386828496,
"learning_rate": 3.253462863839243e-05,
"loss": 0.6098,
"step": 444
},
{
"epoch": 3.0182280627384483,
"grad_norm": 0.5440271202067795,
"learning_rate": 3.2347942513796475e-05,
"loss": 0.6367,
"step": 445
},
{
"epoch": 3.0250105977108945,
"grad_norm": 0.5386656826635444,
"learning_rate": 3.216142924113046e-05,
"loss": 0.6285,
"step": 446
},
{
"epoch": 3.0317931326833403,
"grad_norm": 0.4074410048009017,
"learning_rate": 3.1975093033533003e-05,
"loss": 0.6332,
"step": 447
},
{
"epoch": 3.0385756676557865,
"grad_norm": 0.25928533106144735,
"learning_rate": 3.1788938100143086e-05,
"loss": 0.6246,
"step": 448
},
{
"epoch": 3.045358202628232,
"grad_norm": 0.445439070033727,
"learning_rate": 3.160296864600482e-05,
"loss": 0.6193,
"step": 449
},
{
"epoch": 3.0521407376006784,
"grad_norm": 0.46469489323619156,
"learning_rate": 3.1417188871972554e-05,
"loss": 0.6264,
"step": 450
},
{
"epoch": 3.058923272573124,
"grad_norm": 0.3272979337949089,
"learning_rate": 3.123160297461601e-05,
"loss": 0.631,
"step": 451
},
{
"epoch": 3.0657058075455703,
"grad_norm": 0.3178424540888323,
"learning_rate": 3.104621514612538e-05,
"loss": 0.6344,
"step": 452
},
{
"epoch": 3.072488342518016,
"grad_norm": 0.2754042730785633,
"learning_rate": 3.086102957421672e-05,
"loss": 0.6129,
"step": 453
},
{
"epoch": 3.0792708774904622,
"grad_norm": 0.35442538475012997,
"learning_rate": 3.0676050442037336e-05,
"loss": 0.6269,
"step": 454
},
{
"epoch": 3.086053412462908,
"grad_norm": 0.2797121425291907,
"learning_rate": 3.0491281928071217e-05,
"loss": 0.6221,
"step": 455
},
{
"epoch": 3.092835947435354,
"grad_norm": 0.23765468818900978,
"learning_rate": 3.030672820604476e-05,
"loss": 0.6256,
"step": 456
},
{
"epoch": 3.0996184824078,
"grad_norm": 0.26572260092618777,
"learning_rate": 3.012239344483244e-05,
"loss": 0.6177,
"step": 457
},
{
"epoch": 3.1064010173802457,
"grad_norm": 0.33390446968478465,
"learning_rate": 2.993828180836262e-05,
"loss": 0.6113,
"step": 458
},
{
"epoch": 3.113183552352692,
"grad_norm": 0.2666734723327277,
"learning_rate": 2.9754397455523497e-05,
"loss": 0.6297,
"step": 459
},
{
"epoch": 3.1199660873251376,
"grad_norm": 0.29084625313932944,
"learning_rate": 2.957074454006921e-05,
"loss": 0.6196,
"step": 460
},
{
"epoch": 3.1267486222975838,
"grad_norm": 0.19046989657100505,
"learning_rate": 2.9387327210525942e-05,
"loss": 0.6188,
"step": 461
},
{
"epoch": 3.1335311572700295,
"grad_norm": 0.338123523854364,
"learning_rate": 2.9204149610098223e-05,
"loss": 0.6281,
"step": 462
},
{
"epoch": 3.1403136922424757,
"grad_norm": 0.22618376500160342,
"learning_rate": 2.902121587657539e-05,
"loss": 0.6157,
"step": 463
},
{
"epoch": 3.1470962272149214,
"grad_norm": 0.35386494541824426,
"learning_rate": 2.8838530142238076e-05,
"loss": 0.626,
"step": 464
},
{
"epoch": 3.1538787621873676,
"grad_norm": 0.2840090224796156,
"learning_rate": 2.865609653376484e-05,
"loss": 0.6199,
"step": 465
},
{
"epoch": 3.1606612971598134,
"grad_norm": 0.2835278774397419,
"learning_rate": 2.8473919172139034e-05,
"loss": 0.635,
"step": 466
},
{
"epoch": 3.1674438321322596,
"grad_norm": 0.2932289544670752,
"learning_rate": 2.8292002172555616e-05,
"loss": 0.6181,
"step": 467
},
{
"epoch": 3.1742263671047053,
"grad_norm": 0.21388735557547445,
"learning_rate": 2.8110349644328275e-05,
"loss": 0.6304,
"step": 468
},
{
"epoch": 3.1810089020771515,
"grad_norm": 0.25484830423076227,
"learning_rate": 2.7928965690796562e-05,
"loss": 0.6351,
"step": 469
},
{
"epoch": 3.187791437049597,
"grad_norm": 0.24843127490388153,
"learning_rate": 2.7747854409233187e-05,
"loss": 0.6301,
"step": 470
},
{
"epoch": 3.1945739720220434,
"grad_norm": 0.2451683485321557,
"learning_rate": 2.7567019890751493e-05,
"loss": 0.6207,
"step": 471
},
{
"epoch": 3.201356506994489,
"grad_norm": 0.3164960123899796,
"learning_rate": 2.7386466220213065e-05,
"loss": 0.6213,
"step": 472
},
{
"epoch": 3.2081390419669353,
"grad_norm": 0.2697409403652239,
"learning_rate": 2.7206197476135403e-05,
"loss": 0.6212,
"step": 473
},
{
"epoch": 3.214921576939381,
"grad_norm": 0.19669860705803766,
"learning_rate": 2.7026217730599814e-05,
"loss": 0.6284,
"step": 474
},
{
"epoch": 3.2217041119118273,
"grad_norm": 0.23674235246224137,
"learning_rate": 2.6846531049159454e-05,
"loss": 0.627,
"step": 475
},
{
"epoch": 3.228486646884273,
"grad_norm": 0.18982523269613027,
"learning_rate": 2.6667141490747454e-05,
"loss": 0.6236,
"step": 476
},
{
"epoch": 3.2352691818567187,
"grad_norm": 0.20626545109123062,
"learning_rate": 2.648805310758524e-05,
"loss": 0.6242,
"step": 477
},
{
"epoch": 3.242051716829165,
"grad_norm": 0.25273617130117126,
"learning_rate": 2.6309269945091025e-05,
"loss": 0.6135,
"step": 478
},
{
"epoch": 3.2488342518016107,
"grad_norm": 0.19396420036714845,
"learning_rate": 2.6130796041788404e-05,
"loss": 0.6401,
"step": 479
},
{
"epoch": 3.255616786774057,
"grad_norm": 0.26449583344475813,
"learning_rate": 2.5952635429215117e-05,
"loss": 0.6303,
"step": 480
},
{
"epoch": 3.2623993217465026,
"grad_norm": 0.2880771386228117,
"learning_rate": 2.5774792131832012e-05,
"loss": 0.6213,
"step": 481
},
{
"epoch": 3.269181856718949,
"grad_norm": 0.19301524096425413,
"learning_rate": 2.559727016693212e-05,
"loss": 0.6368,
"step": 482
},
{
"epoch": 3.2759643916913945,
"grad_norm": 0.31002080783316865,
"learning_rate": 2.54200735445499e-05,
"loss": 0.641,
"step": 483
},
{
"epoch": 3.2827469266638407,
"grad_norm": 0.24500950182263187,
"learning_rate": 2.5243206267370685e-05,
"loss": 0.6204,
"step": 484
},
{
"epoch": 3.2895294616362865,
"grad_norm": 0.25623143489960465,
"learning_rate": 2.506667233064021e-05,
"loss": 0.6207,
"step": 485
},
{
"epoch": 3.2963119966087326,
"grad_norm": 0.2918652621084026,
"learning_rate": 2.489047572207442e-05,
"loss": 0.6371,
"step": 486
},
{
"epoch": 3.3030945315811784,
"grad_norm": 0.23337516028968874,
"learning_rate": 2.4714620421769394e-05,
"loss": 0.6285,
"step": 487
},
{
"epoch": 3.3098770665536246,
"grad_norm": 0.23619697217842073,
"learning_rate": 2.4539110402111383e-05,
"loss": 0.631,
"step": 488
},
{
"epoch": 3.3166596015260703,
"grad_norm": 0.2448512796334409,
"learning_rate": 2.436394962768712e-05,
"loss": 0.6276,
"step": 489
},
{
"epoch": 3.3234421364985165,
"grad_norm": 0.1983649853769521,
"learning_rate": 2.4189142055194276e-05,
"loss": 0.6244,
"step": 490
},
{
"epoch": 3.3302246714709622,
"grad_norm": 0.21190165986074258,
"learning_rate": 2.4014691633352038e-05,
"loss": 0.6101,
"step": 491
},
{
"epoch": 3.337007206443408,
"grad_norm": 0.25804827992388285,
"learning_rate": 2.3840602302811923e-05,
"loss": 0.6135,
"step": 492
},
{
"epoch": 3.343789741415854,
"grad_norm": 0.18856304220278483,
"learning_rate": 2.3666877996068824e-05,
"loss": 0.6288,
"step": 493
},
{
"epoch": 3.3505722763883004,
"grad_norm": 0.3301373974007395,
"learning_rate": 2.3493522637372087e-05,
"loss": 0.6326,
"step": 494
},
{
"epoch": 3.357354811360746,
"grad_norm": 0.30315622915732504,
"learning_rate": 2.332054014263692e-05,
"loss": 0.6259,
"step": 495
},
{
"epoch": 3.364137346333192,
"grad_norm": 0.22646576230809926,
"learning_rate": 2.3147934419355935e-05,
"loss": 0.6245,
"step": 496
},
{
"epoch": 3.370919881305638,
"grad_norm": 0.30590058038547674,
"learning_rate": 2.2975709366510887e-05,
"loss": 0.6316,
"step": 497
},
{
"epoch": 3.3777024162780838,
"grad_norm": 0.26486089548607467,
"learning_rate": 2.2803868874484536e-05,
"loss": 0.6383,
"step": 498
},
{
"epoch": 3.38448495125053,
"grad_norm": 0.209806274707419,
"learning_rate": 2.2632416824972886e-05,
"loss": 0.614,
"step": 499
},
{
"epoch": 3.3912674862229757,
"grad_norm": 0.3076979703172019,
"learning_rate": 2.2461357090897352e-05,
"loss": 0.6317,
"step": 500
},
{
"epoch": 3.398050021195422,
"grad_norm": 0.27851643544183735,
"learning_rate": 2.2290693536317444e-05,
"loss": 0.6226,
"step": 501
},
{
"epoch": 3.4048325561678676,
"grad_norm": 0.26720054263757376,
"learning_rate": 2.212043001634332e-05,
"loss": 0.6306,
"step": 502
},
{
"epoch": 3.411615091140314,
"grad_norm": 0.3264075780455984,
"learning_rate": 2.1950570377048813e-05,
"loss": 0.6144,
"step": 503
},
{
"epoch": 3.4183976261127595,
"grad_norm": 0.24476904503391836,
"learning_rate": 2.178111845538453e-05,
"loss": 0.6317,
"step": 504
},
{
"epoch": 3.4251801610852057,
"grad_norm": 0.3445192467182261,
"learning_rate": 2.1612078079091125e-05,
"loss": 0.6227,
"step": 505
},
{
"epoch": 3.4319626960576515,
"grad_norm": 0.2248390312896755,
"learning_rate": 2.144345306661291e-05,
"loss": 0.6179,
"step": 506
},
{
"epoch": 3.4387452310300977,
"grad_norm": 0.24580333907022087,
"learning_rate": 2.1275247227011536e-05,
"loss": 0.6218,
"step": 507
},
{
"epoch": 3.4455277660025434,
"grad_norm": 0.2333810028522329,
"learning_rate": 2.1107464359880022e-05,
"loss": 0.6313,
"step": 508
},
{
"epoch": 3.4523103009749896,
"grad_norm": 0.20038699343598237,
"learning_rate": 2.0940108255256823e-05,
"loss": 0.6309,
"step": 509
},
{
"epoch": 3.4590928359474353,
"grad_norm": 0.19291801763604194,
"learning_rate": 2.0773182693540325e-05,
"loss": 0.6076,
"step": 510
},
{
"epoch": 3.465875370919881,
"grad_norm": 0.21138321490252843,
"learning_rate": 2.0606691445403398e-05,
"loss": 0.6238,
"step": 511
},
{
"epoch": 3.4726579058923273,
"grad_norm": 0.2081459000808386,
"learning_rate": 2.0440638271708226e-05,
"loss": 0.6135,
"step": 512
},
{
"epoch": 3.4794404408647734,
"grad_norm": 0.21359018594893273,
"learning_rate": 2.0275026923421315e-05,
"loss": 0.6347,
"step": 513
},
{
"epoch": 3.486222975837219,
"grad_norm": 0.262459399100271,
"learning_rate": 2.0109861141528853e-05,
"loss": 0.6289,
"step": 514
},
{
"epoch": 3.493005510809665,
"grad_norm": 0.24248389999156902,
"learning_rate": 1.9945144656952103e-05,
"loss": 0.6128,
"step": 515
},
{
"epoch": 3.499788045782111,
"grad_norm": 0.22920015918246003,
"learning_rate": 1.9780881190463222e-05,
"loss": 0.634,
"step": 516
},
{
"epoch": 3.506570580754557,
"grad_norm": 0.2824421315404272,
"learning_rate": 1.9617074452601113e-05,
"loss": 0.6322,
"step": 517
},
{
"epoch": 3.513353115727003,
"grad_norm": 0.19432350726010572,
"learning_rate": 1.9453728143587697e-05,
"loss": 0.6231,
"step": 518
},
{
"epoch": 3.520135650699449,
"grad_norm": 0.2846932920722668,
"learning_rate": 1.9290845953244307e-05,
"loss": 0.625,
"step": 519
},
{
"epoch": 3.526918185671895,
"grad_norm": 0.16416357169798437,
"learning_rate": 1.9128431560908263e-05,
"loss": 0.6191,
"step": 520
},
{
"epoch": 3.5337007206443407,
"grad_norm": 0.2747654462569946,
"learning_rate": 1.896648863534988e-05,
"loss": 0.622,
"step": 521
},
{
"epoch": 3.540483255616787,
"grad_norm": 0.15903407751400953,
"learning_rate": 1.8805020834689487e-05,
"loss": 0.6194,
"step": 522
},
{
"epoch": 3.5472657905892326,
"grad_norm": 0.20365263598678185,
"learning_rate": 1.864403180631489e-05,
"loss": 0.6255,
"step": 523
},
{
"epoch": 3.554048325561679,
"grad_norm": 0.16301528679144703,
"learning_rate": 1.8483525186798875e-05,
"loss": 0.6253,
"step": 524
},
{
"epoch": 3.5608308605341246,
"grad_norm": 0.23098070191792286,
"learning_rate": 1.832350460181717e-05,
"loss": 0.62,
"step": 525
},
{
"epoch": 3.5676133955065703,
"grad_norm": 0.16673057537934008,
"learning_rate": 1.8163973666066473e-05,
"loss": 0.629,
"step": 526
},
{
"epoch": 3.5743959304790165,
"grad_norm": 0.19378211117923547,
"learning_rate": 1.8004935983182835e-05,
"loss": 0.6344,
"step": 527
},
{
"epoch": 3.5811784654514627,
"grad_norm": 0.17643236928174555,
"learning_rate": 1.784639514566021e-05,
"loss": 0.6284,
"step": 528
},
{
"epoch": 3.5879610004239084,
"grad_norm": 0.19805537148789543,
"learning_rate": 1.7688354734769378e-05,
"loss": 0.6245,
"step": 529
},
{
"epoch": 3.594743535396354,
"grad_norm": 0.17444243321183842,
"learning_rate": 1.753081832047697e-05,
"loss": 0.6241,
"step": 530
},
{
"epoch": 3.6015260703688003,
"grad_norm": 0.22629631276174578,
"learning_rate": 1.7373789461364904e-05,
"loss": 0.6296,
"step": 531
},
{
"epoch": 3.6083086053412465,
"grad_norm": 0.2322655718660276,
"learning_rate": 1.7217271704549913e-05,
"loss": 0.6335,
"step": 532
},
{
"epoch": 3.6150911403136923,
"grad_norm": 0.20701067571130669,
"learning_rate": 1.7061268585603507e-05,
"loss": 0.6307,
"step": 533
},
{
"epoch": 3.621873675286138,
"grad_norm": 0.2206959741176191,
"learning_rate": 1.6905783628472074e-05,
"loss": 0.6347,
"step": 534
},
{
"epoch": 3.628656210258584,
"grad_norm": 0.15861912044210094,
"learning_rate": 1.6750820345397217e-05,
"loss": 0.6265,
"step": 535
},
{
"epoch": 3.63543874523103,
"grad_norm": 0.2001185331683702,
"learning_rate": 1.659638223683653e-05,
"loss": 0.6287,
"step": 536
},
{
"epoch": 3.642221280203476,
"grad_norm": 0.1746493458635906,
"learning_rate": 1.644247279138442e-05,
"loss": 0.6361,
"step": 537
},
{
"epoch": 3.649003815175922,
"grad_norm": 0.17095450468813203,
"learning_rate": 1.628909548569339e-05,
"loss": 0.6251,
"step": 538
},
{
"epoch": 3.655786350148368,
"grad_norm": 0.15588769920603499,
"learning_rate": 1.6136253784395455e-05,
"loss": 0.629,
"step": 539
},
{
"epoch": 3.662568885120814,
"grad_norm": 0.14759859028229927,
"learning_rate": 1.5983951140023864e-05,
"loss": 0.6174,
"step": 540
},
{
"epoch": 3.66935142009326,
"grad_norm": 0.16662793387428904,
"learning_rate": 1.583219099293516e-05,
"loss": 0.6192,
"step": 541
},
{
"epoch": 3.6761339550657057,
"grad_norm": 0.16325631134586066,
"learning_rate": 1.568097677123146e-05,
"loss": 0.6193,
"step": 542
},
{
"epoch": 3.682916490038152,
"grad_norm": 0.1590144709587067,
"learning_rate": 1.5530311890682946e-05,
"loss": 0.6126,
"step": 543
},
{
"epoch": 3.6896990250105977,
"grad_norm": 0.16971629601446409,
"learning_rate": 1.5380199754650838e-05,
"loss": 0.6204,
"step": 544
},
{
"epoch": 3.6964815599830434,
"grad_norm": 0.17653934494907608,
"learning_rate": 1.5230643754010355e-05,
"loss": 0.634,
"step": 545
},
{
"epoch": 3.7032640949554896,
"grad_norm": 0.1602687042337523,
"learning_rate": 1.508164726707427e-05,
"loss": 0.6132,
"step": 546
},
{
"epoch": 3.7100466299279358,
"grad_norm": 0.1853250131338786,
"learning_rate": 1.4933213659516485e-05,
"loss": 0.6263,
"step": 547
},
{
"epoch": 3.7168291649003815,
"grad_norm": 0.16166905381502983,
"learning_rate": 1.4785346284296078e-05,
"loss": 0.6158,
"step": 548
},
{
"epoch": 3.7236116998728273,
"grad_norm": 0.19092582748564702,
"learning_rate": 1.4638048481581537e-05,
"loss": 0.6235,
"step": 549
},
{
"epoch": 3.7303942348452734,
"grad_norm": 0.14148685680959427,
"learning_rate": 1.4491323578675265e-05,
"loss": 0.6137,
"step": 550
},
{
"epoch": 3.7371767698177196,
"grad_norm": 0.18984864298487686,
"learning_rate": 1.4345174889938514e-05,
"loss": 0.6329,
"step": 551
},
{
"epoch": 3.7439593047901654,
"grad_norm": 0.14327461316917306,
"learning_rate": 1.4199605716716414e-05,
"loss": 0.6272,
"step": 552
},
{
"epoch": 3.750741839762611,
"grad_norm": 0.18666689301215467,
"learning_rate": 1.4054619347263487e-05,
"loss": 0.6363,
"step": 553
},
{
"epoch": 3.7575243747350573,
"grad_norm": 0.14946115608698116,
"learning_rate": 1.3910219056669335e-05,
"loss": 0.6218,
"step": 554
},
{
"epoch": 3.764306909707503,
"grad_norm": 0.17777432794193568,
"learning_rate": 1.3766408106784601e-05,
"loss": 0.6278,
"step": 555
},
{
"epoch": 3.771089444679949,
"grad_norm": 0.16174854488646243,
"learning_rate": 1.3623189746147398e-05,
"loss": 0.6227,
"step": 556
},
{
"epoch": 3.777871979652395,
"grad_norm": 0.15621831808343503,
"learning_rate": 1.3480567209909863e-05,
"loss": 0.6296,
"step": 557
},
{
"epoch": 3.784654514624841,
"grad_norm": 0.17960424688042897,
"learning_rate": 1.333854371976505e-05,
"loss": 0.6175,
"step": 558
},
{
"epoch": 3.791437049597287,
"grad_norm": 0.161165540194335,
"learning_rate": 1.319712248387424e-05,
"loss": 0.6259,
"step": 559
},
{
"epoch": 3.798219584569733,
"grad_norm": 0.18195946037024635,
"learning_rate": 1.3056306696794377e-05,
"loss": 0.6259,
"step": 560
},
{
"epoch": 3.805002119542179,
"grad_norm": 0.15299855099029444,
"learning_rate": 1.2916099539405997e-05,
"loss": 0.623,
"step": 561
},
{
"epoch": 3.811784654514625,
"grad_norm": 0.19532294279356358,
"learning_rate": 1.2776504178841291e-05,
"loss": 0.629,
"step": 562
},
{
"epoch": 3.8185671894870707,
"grad_norm": 0.1505952319707712,
"learning_rate": 1.2637523768412633e-05,
"loss": 0.6239,
"step": 563
},
{
"epoch": 3.8253497244595165,
"grad_norm": 0.15785614162676143,
"learning_rate": 1.249916144754133e-05,
"loss": 0.6213,
"step": 564
},
{
"epoch": 3.8321322594319627,
"grad_norm": 0.17665393013957048,
"learning_rate": 1.2361420341686641e-05,
"loss": 0.6205,
"step": 565
},
{
"epoch": 3.838914794404409,
"grad_norm": 0.16420702224755918,
"learning_rate": 1.22243035622753e-05,
"loss": 0.6221,
"step": 566
},
{
"epoch": 3.8456973293768546,
"grad_norm": 0.17607498450226258,
"learning_rate": 1.2087814206631095e-05,
"loss": 0.6212,
"step": 567
},
{
"epoch": 3.8524798643493003,
"grad_norm": 0.19592268500165855,
"learning_rate": 1.1951955357905033e-05,
"loss": 0.6199,
"step": 568
},
{
"epoch": 3.8592623993217465,
"grad_norm": 0.1402346539755977,
"learning_rate": 1.1816730085005603e-05,
"loss": 0.6198,
"step": 569
},
{
"epoch": 3.8660449342941927,
"grad_norm": 0.18099131317071607,
"learning_rate": 1.1682141442529473e-05,
"loss": 0.6256,
"step": 570
},
{
"epoch": 3.8728274692666385,
"grad_norm": 0.19545074620930133,
"learning_rate": 1.1548192470692516e-05,
"loss": 0.6157,
"step": 571
},
{
"epoch": 3.879610004239084,
"grad_norm": 0.15514995332077897,
"learning_rate": 1.1414886195261135e-05,
"loss": 0.6262,
"step": 572
},
{
"epoch": 3.8863925392115304,
"grad_norm": 0.1534732088259275,
"learning_rate": 1.128222562748384e-05,
"loss": 0.6218,
"step": 573
},
{
"epoch": 3.893175074183976,
"grad_norm": 0.16016994591208966,
"learning_rate": 1.1150213764023357e-05,
"loss": 0.6264,
"step": 574
},
{
"epoch": 3.8999576091564223,
"grad_norm": 0.1393505701166259,
"learning_rate": 1.1018853586888794e-05,
"loss": 0.6217,
"step": 575
},
{
"epoch": 3.906740144128868,
"grad_norm": 0.13613614708020336,
"learning_rate": 1.0888148063368411e-05,
"loss": 0.6271,
"step": 576
},
{
"epoch": 3.9135226791013142,
"grad_norm": 0.13989845925554537,
"learning_rate": 1.075810014596248e-05,
"loss": 0.6246,
"step": 577
},
{
"epoch": 3.92030521407376,
"grad_norm": 0.14295575937313854,
"learning_rate": 1.0628712772316682e-05,
"loss": 0.6186,
"step": 578
},
{
"epoch": 3.927087749046206,
"grad_norm": 0.14578310176151307,
"learning_rate": 1.04999888651557e-05,
"loss": 0.6237,
"step": 579
},
{
"epoch": 3.933870284018652,
"grad_norm": 0.1418828430912816,
"learning_rate": 1.0371931332217175e-05,
"loss": 0.6247,
"step": 580
},
{
"epoch": 3.940652818991098,
"grad_norm": 0.1325595548114853,
"learning_rate": 1.0244543066186102e-05,
"loss": 0.623,
"step": 581
},
{
"epoch": 3.947435353963544,
"grad_norm": 0.1458164336431073,
"learning_rate": 1.0117826944629425e-05,
"loss": 0.6311,
"step": 582
},
{
"epoch": 3.9542178889359896,
"grad_norm": 0.1369850394553834,
"learning_rate": 9.991785829931024e-06,
"loss": 0.6352,
"step": 583
},
{
"epoch": 3.9610004239084358,
"grad_norm": 0.11861951638152013,
"learning_rate": 9.866422569227133e-06,
"loss": 0.616,
"step": 584
},
{
"epoch": 3.967782958880882,
"grad_norm": 0.12198529312439654,
"learning_rate": 9.74173999434192e-06,
"loss": 0.6267,
"step": 585
},
{
"epoch": 3.9745654938533277,
"grad_norm": 0.13835745132234695,
"learning_rate": 9.61774092172362e-06,
"loss": 0.6228,
"step": 586
},
{
"epoch": 3.9813480288257734,
"grad_norm": 0.14469581512913757,
"learning_rate": 9.494428152380868e-06,
"loss": 0.6249,
"step": 587
},
{
"epoch": 3.9881305637982196,
"grad_norm": 0.14538533970008014,
"learning_rate": 9.371804471819401e-06,
"loss": 0.6327,
"step": 588
},
{
"epoch": 3.994913098770666,
"grad_norm": 0.12842111782896357,
"learning_rate": 9.249872649979203e-06,
"loss": 0.6303,
"step": 589
},
{
"epoch": 4.001695633743111,
"grad_norm": 0.26900932028906,
"learning_rate": 9.128635441171854e-06,
"loss": 0.947,
"step": 590
},
{
"epoch": 4.008478168715557,
"grad_norm": 0.1546620990829906,
"learning_rate": 9.008095584018406e-06,
"loss": 0.6138,
"step": 591
},
{
"epoch": 4.0152607036880035,
"grad_norm": 0.15420921751444336,
"learning_rate": 8.888255801387417e-06,
"loss": 0.6084,
"step": 592
},
{
"epoch": 4.02204323866045,
"grad_norm": 0.15369650428751042,
"learning_rate": 8.76911880033354e-06,
"loss": 0.6196,
"step": 593
},
{
"epoch": 4.028825773632895,
"grad_norm": 0.1491321631121837,
"learning_rate": 8.65068727203633e-06,
"loss": 0.6193,
"step": 594
},
{
"epoch": 4.035608308605341,
"grad_norm": 0.14296040181465852,
"learning_rate": 8.53296389173941e-06,
"loss": 0.5971,
"step": 595
},
{
"epoch": 4.042390843577787,
"grad_norm": 0.17332216986843843,
"learning_rate": 8.415951318690134e-06,
"loss": 0.6085,
"step": 596
},
{
"epoch": 4.0491733785502335,
"grad_norm": 0.1392839623319017,
"learning_rate": 8.299652196079462e-06,
"loss": 0.6001,
"step": 597
},
{
"epoch": 4.055955913522679,
"grad_norm": 0.13152438443257536,
"learning_rate": 8.184069150982217e-06,
"loss": 0.6105,
"step": 598
},
{
"epoch": 4.062738448495125,
"grad_norm": 0.1534567228015376,
"learning_rate": 8.069204794297843e-06,
"loss": 0.6055,
"step": 599
},
{
"epoch": 4.069520983467571,
"grad_norm": 0.13360654544275674,
"learning_rate": 7.955061720691315e-06,
"loss": 0.6032,
"step": 600
},
{
"epoch": 4.076303518440017,
"grad_norm": 0.12240252460009761,
"learning_rate": 7.841642508534617e-06,
"loss": 0.6123,
"step": 601
},
{
"epoch": 4.083086053412463,
"grad_norm": 0.13532197093706563,
"learning_rate": 7.728949719848464e-06,
"loss": 0.6153,
"step": 602
},
{
"epoch": 4.089868588384909,
"grad_norm": 0.14193336659405498,
"learning_rate": 7.6169859002443954e-06,
"loss": 0.6168,
"step": 603
},
{
"epoch": 4.096651123357355,
"grad_norm": 0.12702143209359149,
"learning_rate": 7.505753578867354e-06,
"loss": 0.6151,
"step": 604
},
{
"epoch": 4.1034336583298,
"grad_norm": 0.13369676075093645,
"learning_rate": 7.395255268338459e-06,
"loss": 0.6166,
"step": 605
},
{
"epoch": 4.1102161933022465,
"grad_norm": 0.1334402641156527,
"learning_rate": 7.285493464698343e-06,
"loss": 0.6103,
"step": 606
},
{
"epoch": 4.116998728274693,
"grad_norm": 0.11322677082673437,
"learning_rate": 7.176470647350675e-06,
"loss": 0.6213,
"step": 607
},
{
"epoch": 4.123781263247139,
"grad_norm": 0.1301924506080975,
"learning_rate": 7.068189279006237e-06,
"loss": 0.6099,
"step": 608
},
{
"epoch": 4.130563798219584,
"grad_norm": 0.13846251217535666,
"learning_rate": 6.960651805627248e-06,
"loss": 0.5961,
"step": 609
},
{
"epoch": 4.13734633319203,
"grad_norm": 0.11426017628960868,
"learning_rate": 6.853860656372125e-06,
"loss": 0.6001,
"step": 610
},
{
"epoch": 4.144128868164477,
"grad_norm": 0.11173851423251971,
"learning_rate": 6.747818243540587e-06,
"loss": 0.6068,
"step": 611
},
{
"epoch": 4.150911403136923,
"grad_norm": 0.13703419453889504,
"learning_rate": 6.642526962519218e-06,
"loss": 0.6083,
"step": 612
},
{
"epoch": 4.157693938109368,
"grad_norm": 0.12103577637533855,
"learning_rate": 6.537989191727292e-06,
"loss": 0.6026,
"step": 613
},
{
"epoch": 4.164476473081814,
"grad_norm": 0.12350144156787667,
"learning_rate": 6.4342072925631125e-06,
"loss": 0.6019,
"step": 614
},
{
"epoch": 4.17125900805426,
"grad_norm": 0.11640263763641474,
"learning_rate": 6.331183609350602e-06,
"loss": 0.5954,
"step": 615
},
{
"epoch": 4.178041543026707,
"grad_norm": 0.10545213840827801,
"learning_rate": 6.228920469286408e-06,
"loss": 0.6117,
"step": 616
},
{
"epoch": 4.184824077999152,
"grad_norm": 0.11495651066805103,
"learning_rate": 6.127420182387314e-06,
"loss": 0.5946,
"step": 617
},
{
"epoch": 4.191606612971598,
"grad_norm": 0.11620572636315642,
"learning_rate": 6.026685041438018e-06,
"loss": 0.6034,
"step": 618
},
{
"epoch": 4.198389147944044,
"grad_norm": 0.10866699027698076,
"learning_rate": 5.926717321939417e-06,
"loss": 0.6048,
"step": 619
},
{
"epoch": 4.2051716829164905,
"grad_norm": 0.1245636539367363,
"learning_rate": 5.827519282057128e-06,
"loss": 0.6037,
"step": 620
},
{
"epoch": 4.211954217888936,
"grad_norm": 0.10644196388055245,
"learning_rate": 5.729093162570545e-06,
"loss": 0.607,
"step": 621
},
{
"epoch": 4.218736752861382,
"grad_norm": 0.10894266432537414,
"learning_rate": 5.631441186822168e-06,
"loss": 0.6135,
"step": 622
},
{
"epoch": 4.225519287833828,
"grad_norm": 0.12336549859525484,
"learning_rate": 5.534565560667426e-06,
"loss": 0.5981,
"step": 623
},
{
"epoch": 4.232301822806273,
"grad_norm": 0.10948607920606619,
"learning_rate": 5.4384684724248096e-06,
"loss": 0.6118,
"step": 624
},
{
"epoch": 4.23908435777872,
"grad_norm": 0.10785742485131251,
"learning_rate": 5.343152092826476e-06,
"loss": 0.6184,
"step": 625
},
{
"epoch": 4.245866892751166,
"grad_norm": 0.1346812509747976,
"learning_rate": 5.2486185749691665e-06,
"loss": 0.6058,
"step": 626
},
{
"epoch": 4.252649427723612,
"grad_norm": 0.10711676011900324,
"learning_rate": 5.154870054265635e-06,
"loss": 0.6071,
"step": 627
},
{
"epoch": 4.259431962696057,
"grad_norm": 0.09686690911531301,
"learning_rate": 5.061908648396338e-06,
"loss": 0.6019,
"step": 628
},
{
"epoch": 4.2662144976685035,
"grad_norm": 0.11286945799474804,
"learning_rate": 4.969736457261669e-06,
"loss": 0.604,
"step": 629
},
{
"epoch": 4.27299703264095,
"grad_norm": 0.12539528001284939,
"learning_rate": 4.878355562934465e-06,
"loss": 0.6103,
"step": 630
},
{
"epoch": 4.279779567613396,
"grad_norm": 0.11846976963374484,
"learning_rate": 4.787768029613014e-06,
"loss": 0.6102,
"step": 631
},
{
"epoch": 4.286562102585841,
"grad_norm": 0.11455949347213798,
"learning_rate": 4.697975903574419e-06,
"loss": 0.6044,
"step": 632
},
{
"epoch": 4.293344637558287,
"grad_norm": 0.11423245618436771,
"learning_rate": 4.608981213128347e-06,
"loss": 0.6104,
"step": 633
},
{
"epoch": 4.3001271725307335,
"grad_norm": 0.11763066458751498,
"learning_rate": 4.520785968571266e-06,
"loss": 0.6103,
"step": 634
},
{
"epoch": 4.30690970750318,
"grad_norm": 0.12407775475836062,
"learning_rate": 4.433392162140968e-06,
"loss": 0.6127,
"step": 635
},
{
"epoch": 4.313692242475625,
"grad_norm": 0.1177975105922243,
"learning_rate": 4.3468017679716245e-06,
"loss": 0.6258,
"step": 636
},
{
"epoch": 4.320474777448071,
"grad_norm": 0.11062581278413729,
"learning_rate": 4.261016742049155e-06,
"loss": 0.617,
"step": 637
},
{
"epoch": 4.327257312420517,
"grad_norm": 0.11328064594157465,
"learning_rate": 4.176039022167064e-06,
"loss": 0.6086,
"step": 638
},
{
"epoch": 4.3340398473929636,
"grad_norm": 0.12286256944458376,
"learning_rate": 4.0918705278826685e-06,
"loss": 0.6108,
"step": 639
},
{
"epoch": 4.340822382365409,
"grad_norm": 0.11726282791596607,
"learning_rate": 4.008513160473726e-06,
"loss": 0.6122,
"step": 640
},
{
"epoch": 4.347604917337855,
"grad_norm": 0.13243402516995725,
"learning_rate": 3.925968802895477e-06,
"loss": 0.6059,
"step": 641
},
{
"epoch": 4.354387452310301,
"grad_norm": 0.11732415982857679,
"learning_rate": 3.8442393197381456e-06,
"loss": 0.6159,
"step": 642
},
{
"epoch": 4.3611699872827465,
"grad_norm": 0.11432499142874704,
"learning_rate": 3.7633265571847787e-06,
"loss": 0.6053,
"step": 643
},
{
"epoch": 4.367952522255193,
"grad_norm": 0.12438661313712396,
"learning_rate": 3.683232342969594e-06,
"loss": 0.6185,
"step": 644
},
{
"epoch": 4.374735057227639,
"grad_norm": 0.12003558611356997,
"learning_rate": 3.603958486336625e-06,
"loss": 0.6311,
"step": 645
},
{
"epoch": 4.381517592200085,
"grad_norm": 0.10916503150131288,
"learning_rate": 3.5255067779989173e-06,
"loss": 0.6158,
"step": 646
},
{
"epoch": 4.38830012717253,
"grad_norm": 0.11467284610857227,
"learning_rate": 3.4478789900980545e-06,
"loss": 0.6176,
"step": 647
},
{
"epoch": 4.395082662144977,
"grad_norm": 0.11307569011740914,
"learning_rate": 3.371076876164101e-06,
"loss": 0.614,
"step": 648
},
{
"epoch": 4.401865197117423,
"grad_norm": 0.11385086584182105,
"learning_rate": 3.295102171076039e-06,
"loss": 0.6098,
"step": 649
},
{
"epoch": 4.408647732089869,
"grad_norm": 0.13858694628446455,
"learning_rate": 3.2199565910225305e-06,
"loss": 0.5976,
"step": 650
},
{
"epoch": 4.415430267062314,
"grad_norm": 0.12042166145746055,
"learning_rate": 3.145641833463198e-06,
"loss": 0.6164,
"step": 651
},
{
"epoch": 4.42221280203476,
"grad_norm": 0.11519596457203823,
"learning_rate": 3.0721595770902435e-06,
"loss": 0.6156,
"step": 652
},
{
"epoch": 4.428995337007207,
"grad_norm": 0.09905226368345378,
"learning_rate": 2.9995114817905493e-06,
"loss": 0.6001,
"step": 653
},
{
"epoch": 4.435777871979653,
"grad_norm": 0.11732702729180129,
"learning_rate": 2.927699188608171e-06,
"loss": 0.607,
"step": 654
},
{
"epoch": 4.442560406952098,
"grad_norm": 0.1179451846081634,
"learning_rate": 2.856724319707289e-06,
"loss": 0.6169,
"step": 655
},
{
"epoch": 4.449342941924544,
"grad_norm": 0.10442323225441018,
"learning_rate": 2.7865884783355234e-06,
"loss": 0.6179,
"step": 656
},
{
"epoch": 4.4561254768969905,
"grad_norm": 0.09435389572080079,
"learning_rate": 2.717293248787769e-06,
"loss": 0.6201,
"step": 657
},
{
"epoch": 4.462908011869436,
"grad_norm": 0.11295536763791658,
"learning_rate": 2.648840196370368e-06,
"loss": 0.6019,
"step": 658
},
{
"epoch": 4.469690546841882,
"grad_norm": 0.11594274422462209,
"learning_rate": 2.5812308673657825e-06,
"loss": 0.5977,
"step": 659
},
{
"epoch": 4.476473081814328,
"grad_norm": 0.10182885233809727,
"learning_rate": 2.5144667889976316e-06,
"loss": 0.6017,
"step": 660
},
{
"epoch": 4.483255616786774,
"grad_norm": 0.10348633393081054,
"learning_rate": 2.4485494693962197e-06,
"loss": 0.6086,
"step": 661
},
{
"epoch": 4.49003815175922,
"grad_norm": 0.10025894496329021,
"learning_rate": 2.3834803975644772e-06,
"loss": 0.6182,
"step": 662
},
{
"epoch": 4.496820686731666,
"grad_norm": 0.10187054293463914,
"learning_rate": 2.3192610433442784e-06,
"loss": 0.6074,
"step": 663
},
{
"epoch": 4.503603221704112,
"grad_norm": 0.09715916393818302,
"learning_rate": 2.2558928573833016e-06,
"loss": 0.6076,
"step": 664
},
{
"epoch": 4.510385756676558,
"grad_norm": 0.09756303138919038,
"learning_rate": 2.1933772711021995e-06,
"loss": 0.6019,
"step": 665
},
{
"epoch": 4.5171682916490035,
"grad_norm": 0.09525032911781027,
"learning_rate": 2.1317156966623243e-06,
"loss": 0.5995,
"step": 666
},
{
"epoch": 4.52395082662145,
"grad_norm": 0.09487242097788966,
"learning_rate": 2.0709095269337755e-06,
"loss": 0.6023,
"step": 667
},
{
"epoch": 4.530733361593896,
"grad_norm": 0.11476206706719008,
"learning_rate": 2.0109601354639706e-06,
"loss": 0.6143,
"step": 668
},
{
"epoch": 4.537515896566342,
"grad_norm": 0.09904862194550892,
"learning_rate": 1.9518688764466096e-06,
"loss": 0.6143,
"step": 669
},
{
"epoch": 4.544298431538787,
"grad_norm": 0.09278301353904947,
"learning_rate": 1.8936370846910845e-06,
"loss": 0.6092,
"step": 670
},
{
"epoch": 4.5510809665112335,
"grad_norm": 0.10688691601881584,
"learning_rate": 1.8362660755923079e-06,
"loss": 0.6186,
"step": 671
},
{
"epoch": 4.55786350148368,
"grad_norm": 0.09697916608330605,
"learning_rate": 1.7797571451010442e-06,
"loss": 0.5983,
"step": 672
},
{
"epoch": 4.564646036456125,
"grad_norm": 0.10471706981067595,
"learning_rate": 1.7241115696945821e-06,
"loss": 0.6132,
"step": 673
},
{
"epoch": 4.571428571428571,
"grad_norm": 0.09406802929392503,
"learning_rate": 1.669330606347952e-06,
"loss": 0.5962,
"step": 674
},
{
"epoch": 4.578211106401017,
"grad_norm": 0.0954037644840885,
"learning_rate": 1.6154154925054878e-06,
"loss": 0.6141,
"step": 675
},
{
"epoch": 4.5849936413734635,
"grad_norm": 0.10094861062769968,
"learning_rate": 1.5623674460529014e-06,
"loss": 0.6024,
"step": 676
},
{
"epoch": 4.59177617634591,
"grad_norm": 0.09573172404295381,
"learning_rate": 1.510187665289773e-06,
"loss": 0.6086,
"step": 677
},
{
"epoch": 4.598558711318355,
"grad_norm": 0.09995771347446304,
"learning_rate": 1.458877328902455e-06,
"loss": 0.6052,
"step": 678
},
{
"epoch": 4.605341246290801,
"grad_norm": 0.09656263063599439,
"learning_rate": 1.4084375959374864e-06,
"loss": 0.6072,
"step": 679
},
{
"epoch": 4.612123781263247,
"grad_norm": 0.09822686472765482,
"learning_rate": 1.3588696057753725e-06,
"loss": 0.6089,
"step": 680
},
{
"epoch": 4.618906316235693,
"grad_norm": 0.09676277658643694,
"learning_rate": 1.3101744781048774e-06,
"loss": 0.6016,
"step": 681
},
{
"epoch": 4.625688851208139,
"grad_norm": 0.0933744691144583,
"learning_rate": 1.262353312897715e-06,
"loss": 0.6075,
"step": 682
},
{
"epoch": 4.632471386180585,
"grad_norm": 0.0996442423520143,
"learning_rate": 1.2154071903837107e-06,
"loss": 0.6157,
"step": 683
},
{
"epoch": 4.639253921153031,
"grad_norm": 0.10109190980066803,
"learning_rate": 1.1693371710263857e-06,
"loss": 0.6049,
"step": 684
},
{
"epoch": 4.6460364561254766,
"grad_norm": 0.08703236124988406,
"learning_rate": 1.1241442954990301e-06,
"loss": 0.6048,
"step": 685
},
{
"epoch": 4.652818991097923,
"grad_norm": 0.09280509544326945,
"learning_rate": 1.0798295846611562e-06,
"loss": 0.5989,
"step": 686
},
{
"epoch": 4.659601526070369,
"grad_norm": 0.09785736700280491,
"learning_rate": 1.0363940395354777e-06,
"loss": 0.6098,
"step": 687
},
{
"epoch": 4.666384061042815,
"grad_norm": 0.09508110174195833,
"learning_rate": 9.938386412852652e-07,
"loss": 0.6149,
"step": 688
},
{
"epoch": 4.67316659601526,
"grad_norm": 0.09065950330259104,
"learning_rate": 9.52164351192213e-07,
"loss": 0.6088,
"step": 689
},
{
"epoch": 4.679949130987707,
"grad_norm": 0.09551867620930947,
"learning_rate": 9.113721106346918e-07,
"loss": 0.5989,
"step": 690
},
{
"epoch": 4.686731665960153,
"grad_norm": 0.09832290856026324,
"learning_rate": 8.714628410665082e-07,
"loss": 0.6172,
"step": 691
},
{
"epoch": 4.693514200932599,
"grad_norm": 0.09371892802996978,
"learning_rate": 8.324374439960947e-07,
"loss": 0.6067,
"step": 692
},
{
"epoch": 4.700296735905044,
"grad_norm": 0.09014551117205812,
"learning_rate": 7.942968009661123e-07,
"loss": 0.6142,
"step": 693
},
{
"epoch": 4.7070792708774905,
"grad_norm": 0.09724706571681141,
"learning_rate": 7.570417735335733e-07,
"loss": 0.614,
"step": 694
},
{
"epoch": 4.713861805849937,
"grad_norm": 0.08494611369686998,
"learning_rate": 7.206732032503638e-07,
"loss": 0.5969,
"step": 695
},
{
"epoch": 4.720644340822382,
"grad_norm": 0.0927927564675014,
"learning_rate": 6.851919116442274e-07,
"loss": 0.6122,
"step": 696
},
{
"epoch": 4.727426875794828,
"grad_norm": 0.09273724936533222,
"learning_rate": 6.505987002002245e-07,
"loss": 0.6092,
"step": 697
},
{
"epoch": 4.734209410767274,
"grad_norm": 0.08796818616185165,
"learning_rate": 6.168943503426139e-07,
"loss": 0.6038,
"step": 698
},
{
"epoch": 4.7409919457397205,
"grad_norm": 0.08686636687902749,
"learning_rate": 5.840796234172085e-07,
"loss": 0.5943,
"step": 699
},
{
"epoch": 4.747774480712166,
"grad_norm": 0.09236449250012803,
"learning_rate": 5.521552606741765e-07,
"loss": 0.6066,
"step": 700
},
{
"epoch": 4.754557015684612,
"grad_norm": 0.0926749240950783,
"learning_rate": 5.211219832512893e-07,
"loss": 0.6114,
"step": 701
},
{
"epoch": 4.761339550657058,
"grad_norm": 0.10499405025384424,
"learning_rate": 4.909804921576466e-07,
"loss": 0.6185,
"step": 702
},
{
"epoch": 4.768122085629504,
"grad_norm": 0.08848832101586154,
"learning_rate": 4.6173146825782224e-07,
"loss": 0.6067,
"step": 703
},
{
"epoch": 4.77490462060195,
"grad_norm": 0.09319395321366955,
"learning_rate": 4.3337557225650695e-07,
"loss": 0.6047,
"step": 704
},
{
"epoch": 4.781687155574396,
"grad_norm": 0.0915059840709739,
"learning_rate": 4.0591344468355666e-07,
"loss": 0.6118,
"step": 705
},
{
"epoch": 4.788469690546842,
"grad_norm": 0.10729723170818359,
"learning_rate": 3.793457058795591e-07,
"loss": 0.6189,
"step": 706
},
{
"epoch": 4.795252225519288,
"grad_norm": 0.09285587112058538,
"learning_rate": 3.5367295598178307e-07,
"loss": 0.6085,
"step": 707
},
{
"epoch": 4.8020347604917335,
"grad_norm": 0.08196537866623092,
"learning_rate": 3.2889577491064693e-07,
"loss": 0.6045,
"step": 708
},
{
"epoch": 4.80881729546418,
"grad_norm": 0.08720837958758754,
"learning_rate": 3.050147223566091e-07,
"loss": 0.6122,
"step": 709
},
{
"epoch": 4.815599830436626,
"grad_norm": 0.09158067797742139,
"learning_rate": 2.8203033776752487e-07,
"loss": 0.6033,
"step": 710
},
{
"epoch": 4.822382365409071,
"grad_norm": 0.09043295389558649,
"learning_rate": 2.599431403364694e-07,
"loss": 0.6094,
"step": 711
},
{
"epoch": 4.829164900381517,
"grad_norm": 0.09180388224113892,
"learning_rate": 2.3875362899000054e-07,
"loss": 0.604,
"step": 712
},
{
"epoch": 4.8359474353539635,
"grad_norm": 0.08651605692378189,
"learning_rate": 2.1846228237689226e-07,
"loss": 0.6087,
"step": 713
},
{
"epoch": 4.84272997032641,
"grad_norm": 0.08511470109771488,
"learning_rate": 1.9906955885732104e-07,
"loss": 0.6091,
"step": 714
},
{
"epoch": 4.849512505298856,
"grad_norm": 0.08454739481192287,
"learning_rate": 1.8057589649251862e-07,
"loss": 0.6146,
"step": 715
},
{
"epoch": 4.856295040271301,
"grad_norm": 0.08724704005095565,
"learning_rate": 1.62981713034851e-07,
"loss": 0.6042,
"step": 716
},
{
"epoch": 4.863077575243747,
"grad_norm": 0.08396574948424247,
"learning_rate": 1.462874059184305e-07,
"loss": 0.6086,
"step": 717
},
{
"epoch": 4.869860110216194,
"grad_norm": 0.09082130018441244,
"learning_rate": 1.3049335225009175e-07,
"loss": 0.6295,
"step": 718
},
{
"epoch": 4.876642645188639,
"grad_norm": 0.08722605095693269,
"learning_rate": 1.1559990880089189e-07,
"loss": 0.6085,
"step": 719
},
{
"epoch": 4.883425180161085,
"grad_norm": 0.0878159462681937,
"learning_rate": 1.0160741199805036e-07,
"loss": 0.5998,
"step": 720
},
{
"epoch": 4.890207715133531,
"grad_norm": 0.0878736816907958,
"learning_rate": 8.851617791735933e-08,
"loss": 0.6145,
"step": 721
},
{
"epoch": 4.896990250105977,
"grad_norm": 0.08847166155449576,
"learning_rate": 7.632650227602511e-08,
"loss": 0.604,
"step": 722
},
{
"epoch": 4.903772785078423,
"grad_norm": 0.08650107449834574,
"learning_rate": 6.503866042599338e-08,
"loss": 0.6187,
"step": 723
},
{
"epoch": 4.910555320050869,
"grad_norm": 0.09963600285093287,
"learning_rate": 5.4652907347745307e-08,
"loss": 0.6166,
"step": 724
},
{
"epoch": 4.917337855023315,
"grad_norm": 0.08660963081265105,
"learning_rate": 4.516947764451107e-08,
"loss": 0.6156,
"step": 725
},
{
"epoch": 4.924120389995761,
"grad_norm": 0.09162932265931903,
"learning_rate": 3.6588585537002955e-08,
"loss": 0.6085,
"step": 726
},
{
"epoch": 4.930902924968207,
"grad_norm": 0.08533741638096395,
"learning_rate": 2.8910424858543673e-08,
"loss": 0.6026,
"step": 727
},
{
"epoch": 4.937685459940653,
"grad_norm": 0.08431534489847897,
"learning_rate": 2.213516905070101e-08,
"loss": 0.6012,
"step": 728
},
{
"epoch": 4.944467994913099,
"grad_norm": 0.08633616914365788,
"learning_rate": 1.6262971159384243e-08,
"loss": 0.6146,
"step": 729
},
{
"epoch": 4.951250529885545,
"grad_norm": 0.0931259062674402,
"learning_rate": 1.1293963831366939e-08,
"loss": 0.6116,
"step": 730
},
{
"epoch": 4.9580330648579904,
"grad_norm": 0.08678440176862204,
"learning_rate": 7.228259311293784e-09,
"loss": 0.602,
"step": 731
},
{
"epoch": 4.964815599830437,
"grad_norm": 0.0942499520961841,
"learning_rate": 4.065949439158168e-09,
"loss": 0.6099,
"step": 732
},
{
"epoch": 4.971598134802883,
"grad_norm": 0.07935089647201646,
"learning_rate": 1.8071056482105166e-09,
"loss": 0.615,
"step": 733
},
{
"epoch": 4.978380669775328,
"grad_norm": 0.08485220331271215,
"learning_rate": 4.517789633551317e-10,
"loss": 0.6104,
"step": 734
},
{
"epoch": 4.985163204747774,
"grad_norm": 0.08704984136629952,
"learning_rate": 0.0,
"loss": 0.6111,
"step": 735
},
{
"epoch": 4.985163204747774,
"step": 735,
"total_flos": 1.8913738317545603e+19,
"train_loss": 0.673498132682982,
"train_runtime": 162938.9767,
"train_samples_per_second": 2.316,
"train_steps_per_second": 0.005
}
],
"logging_steps": 1.0,
"max_steps": 735,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.8913738317545603e+19,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}