diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,17647 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 2515, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0019880715705765406, + "grad_norm": 5.719694352988623, + "learning_rate": 3.174603174603175e-07, + "loss": 1.0761, + "step": 1 + }, + { + "epoch": 0.003976143141153081, + "grad_norm": 5.753446360073084, + "learning_rate": 6.34920634920635e-07, + "loss": 1.0759, + "step": 2 + }, + { + "epoch": 0.005964214711729622, + "grad_norm": 5.6362180717832855, + "learning_rate": 9.523809523809525e-07, + "loss": 1.0634, + "step": 3 + }, + { + "epoch": 0.007952286282306162, + "grad_norm": 5.735606400955685, + "learning_rate": 1.26984126984127e-06, + "loss": 1.079, + "step": 4 + }, + { + "epoch": 0.009940357852882704, + "grad_norm": 5.575817061249966, + "learning_rate": 1.5873015873015873e-06, + "loss": 1.0746, + "step": 5 + }, + { + "epoch": 0.011928429423459244, + "grad_norm": 5.260360155484042, + "learning_rate": 1.904761904761905e-06, + "loss": 1.0615, + "step": 6 + }, + { + "epoch": 0.013916500994035786, + "grad_norm": 4.183791741925817, + "learning_rate": 2.222222222222222e-06, + "loss": 1.023, + "step": 7 + }, + { + "epoch": 0.015904572564612324, + "grad_norm": 3.890604251583998, + "learning_rate": 2.53968253968254e-06, + "loss": 1.0225, + "step": 8 + }, + { + "epoch": 0.017892644135188866, + "grad_norm": 2.4418688472172203, + "learning_rate": 2.8571428571428573e-06, + "loss": 0.9791, + "step": 9 + }, + { + "epoch": 0.019880715705765408, + "grad_norm": 2.3267529009813304, + "learning_rate": 3.1746031746031746e-06, + "loss": 0.9752, + "step": 10 + }, + { + "epoch": 0.02186878727634195, + "grad_norm": 2.249898884427966, + "learning_rate": 3.492063492063492e-06, + "loss": 0.9821, + "step": 11 + }, + { + "epoch": 0.02385685884691849, + "grad_norm": 3.177129872043783, + "learning_rate": 3.80952380952381e-06, + "loss": 0.9412, + "step": 12 + }, + { + "epoch": 0.02584493041749503, + "grad_norm": 3.418098934813388, + "learning_rate": 4.126984126984127e-06, + "loss": 0.9534, + "step": 13 + }, + { + "epoch": 0.027833001988071572, + "grad_norm": 3.170560383167798, + "learning_rate": 4.444444444444444e-06, + "loss": 0.9408, + "step": 14 + }, + { + "epoch": 0.02982107355864811, + "grad_norm": 2.824593276204021, + "learning_rate": 4.761904761904762e-06, + "loss": 0.9202, + "step": 15 + }, + { + "epoch": 0.03180914512922465, + "grad_norm": 1.980942339361093, + "learning_rate": 5.07936507936508e-06, + "loss": 0.8779, + "step": 16 + }, + { + "epoch": 0.033797216699801194, + "grad_norm": 2.0329912524472835, + "learning_rate": 5.396825396825397e-06, + "loss": 0.8759, + "step": 17 + }, + { + "epoch": 0.03578528827037773, + "grad_norm": 1.7783267821238207, + "learning_rate": 5.7142857142857145e-06, + "loss": 0.8651, + "step": 18 + }, + { + "epoch": 0.03777335984095427, + "grad_norm": 1.2741893396855972, + "learning_rate": 6.031746031746032e-06, + "loss": 0.8495, + "step": 19 + }, + { + "epoch": 0.039761431411530816, + "grad_norm": 0.9557871843897173, + "learning_rate": 6.349206349206349e-06, + "loss": 0.8366, + "step": 20 + }, + { + "epoch": 0.041749502982107355, + "grad_norm": 1.0086159115273798, + "learning_rate": 6.666666666666667e-06, + "loss": 0.8291, + "step": 21 + }, + { + "epoch": 0.0437375745526839, + "grad_norm": 1.1612726770284232, + "learning_rate": 6.984126984126984e-06, + "loss": 0.8187, + "step": 22 + }, + { + "epoch": 0.04572564612326044, + "grad_norm": 1.0570750814322025, + "learning_rate": 7.301587301587301e-06, + "loss": 0.8107, + "step": 23 + }, + { + "epoch": 0.04771371769383698, + "grad_norm": 0.7751796252916353, + "learning_rate": 7.61904761904762e-06, + "loss": 0.7934, + "step": 24 + }, + { + "epoch": 0.04970178926441352, + "grad_norm": 0.7945616282295676, + "learning_rate": 7.936507936507936e-06, + "loss": 0.7779, + "step": 25 + }, + { + "epoch": 0.05168986083499006, + "grad_norm": 0.858160028742513, + "learning_rate": 8.253968253968254e-06, + "loss": 0.7796, + "step": 26 + }, + { + "epoch": 0.0536779324055666, + "grad_norm": 0.7579216923981651, + "learning_rate": 8.571428571428571e-06, + "loss": 0.7735, + "step": 27 + }, + { + "epoch": 0.055666003976143144, + "grad_norm": 0.6924307407288729, + "learning_rate": 8.888888888888888e-06, + "loss": 0.766, + "step": 28 + }, + { + "epoch": 0.05765407554671968, + "grad_norm": 0.6685114725284176, + "learning_rate": 9.206349206349207e-06, + "loss": 0.7672, + "step": 29 + }, + { + "epoch": 0.05964214711729622, + "grad_norm": 0.6913397934010306, + "learning_rate": 9.523809523809525e-06, + "loss": 0.7494, + "step": 30 + }, + { + "epoch": 0.061630218687872766, + "grad_norm": 0.6134605758081872, + "learning_rate": 9.841269841269842e-06, + "loss": 0.7552, + "step": 31 + }, + { + "epoch": 0.0636182902584493, + "grad_norm": 0.543250098390486, + "learning_rate": 1.015873015873016e-05, + "loss": 0.7505, + "step": 32 + }, + { + "epoch": 0.06560636182902585, + "grad_norm": 0.5141193730679742, + "learning_rate": 1.0476190476190477e-05, + "loss": 0.7446, + "step": 33 + }, + { + "epoch": 0.06759443339960239, + "grad_norm": 0.5115646872354824, + "learning_rate": 1.0793650793650794e-05, + "loss": 0.745, + "step": 34 + }, + { + "epoch": 0.06958250497017893, + "grad_norm": 0.5623419612031815, + "learning_rate": 1.1111111111111113e-05, + "loss": 0.7342, + "step": 35 + }, + { + "epoch": 0.07157057654075547, + "grad_norm": 0.5502558641447046, + "learning_rate": 1.1428571428571429e-05, + "loss": 0.7322, + "step": 36 + }, + { + "epoch": 0.073558648111332, + "grad_norm": 0.5359887902606768, + "learning_rate": 1.1746031746031748e-05, + "loss": 0.73, + "step": 37 + }, + { + "epoch": 0.07554671968190854, + "grad_norm": 0.44251623117375427, + "learning_rate": 1.2063492063492064e-05, + "loss": 0.7282, + "step": 38 + }, + { + "epoch": 0.0775347912524851, + "grad_norm": 0.4507677964518194, + "learning_rate": 1.2380952380952383e-05, + "loss": 0.7244, + "step": 39 + }, + { + "epoch": 0.07952286282306163, + "grad_norm": 0.4512848693711087, + "learning_rate": 1.2698412698412699e-05, + "loss": 0.7278, + "step": 40 + }, + { + "epoch": 0.08151093439363817, + "grad_norm": 0.41505660856918547, + "learning_rate": 1.3015873015873018e-05, + "loss": 0.7254, + "step": 41 + }, + { + "epoch": 0.08349900596421471, + "grad_norm": 0.48813041957576425, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.7194, + "step": 42 + }, + { + "epoch": 0.08548707753479125, + "grad_norm": 0.3877838036386138, + "learning_rate": 1.3650793650793652e-05, + "loss": 0.7176, + "step": 43 + }, + { + "epoch": 0.0874751491053678, + "grad_norm": 0.4181924462251027, + "learning_rate": 1.3968253968253968e-05, + "loss": 0.7156, + "step": 44 + }, + { + "epoch": 0.08946322067594434, + "grad_norm": 0.3939861879487352, + "learning_rate": 1.4285714285714287e-05, + "loss": 0.7056, + "step": 45 + }, + { + "epoch": 0.09145129224652088, + "grad_norm": 0.32819440379311476, + "learning_rate": 1.4603174603174603e-05, + "loss": 0.7053, + "step": 46 + }, + { + "epoch": 0.09343936381709742, + "grad_norm": 0.4345116128871954, + "learning_rate": 1.4920634920634922e-05, + "loss": 0.7097, + "step": 47 + }, + { + "epoch": 0.09542743538767395, + "grad_norm": 0.357575137500084, + "learning_rate": 1.523809523809524e-05, + "loss": 0.7002, + "step": 48 + }, + { + "epoch": 0.09741550695825049, + "grad_norm": 0.36960029460147453, + "learning_rate": 1.555555555555556e-05, + "loss": 0.6974, + "step": 49 + }, + { + "epoch": 0.09940357852882704, + "grad_norm": 0.3665609922915068, + "learning_rate": 1.5873015873015872e-05, + "loss": 0.6999, + "step": 50 + }, + { + "epoch": 0.10139165009940358, + "grad_norm": 0.3080867101340467, + "learning_rate": 1.6190476190476193e-05, + "loss": 0.7007, + "step": 51 + }, + { + "epoch": 0.10337972166998012, + "grad_norm": 0.29433385790793876, + "learning_rate": 1.6507936507936507e-05, + "loss": 0.6958, + "step": 52 + }, + { + "epoch": 0.10536779324055666, + "grad_norm": 0.2932180630106322, + "learning_rate": 1.6825396825396828e-05, + "loss": 0.6963, + "step": 53 + }, + { + "epoch": 0.1073558648111332, + "grad_norm": 0.390572352111665, + "learning_rate": 1.7142857142857142e-05, + "loss": 0.692, + "step": 54 + }, + { + "epoch": 0.10934393638170974, + "grad_norm": 0.46019460280581603, + "learning_rate": 1.7460317460317463e-05, + "loss": 0.6899, + "step": 55 + }, + { + "epoch": 0.11133200795228629, + "grad_norm": 0.4456425633626217, + "learning_rate": 1.7777777777777777e-05, + "loss": 0.6934, + "step": 56 + }, + { + "epoch": 0.11332007952286283, + "grad_norm": 0.5779834684485384, + "learning_rate": 1.8095238095238097e-05, + "loss": 0.7005, + "step": 57 + }, + { + "epoch": 0.11530815109343936, + "grad_norm": 0.43345848661699465, + "learning_rate": 1.8412698412698415e-05, + "loss": 0.688, + "step": 58 + }, + { + "epoch": 0.1172962226640159, + "grad_norm": 0.4702542027909562, + "learning_rate": 1.8730158730158732e-05, + "loss": 0.6838, + "step": 59 + }, + { + "epoch": 0.11928429423459244, + "grad_norm": 0.4532557455295433, + "learning_rate": 1.904761904761905e-05, + "loss": 0.6866, + "step": 60 + }, + { + "epoch": 0.12127236580516898, + "grad_norm": 0.6378809717045135, + "learning_rate": 1.9365079365079367e-05, + "loss": 0.6854, + "step": 61 + }, + { + "epoch": 0.12326043737574553, + "grad_norm": 0.8186502076427445, + "learning_rate": 1.9682539682539684e-05, + "loss": 0.6831, + "step": 62 + }, + { + "epoch": 0.12524850894632206, + "grad_norm": 0.9490391751989258, + "learning_rate": 2e-05, + "loss": 0.6798, + "step": 63 + }, + { + "epoch": 0.1272365805168986, + "grad_norm": 0.8318620223308042, + "learning_rate": 2.031746031746032e-05, + "loss": 0.6842, + "step": 64 + }, + { + "epoch": 0.12922465208747516, + "grad_norm": 0.6266730729415922, + "learning_rate": 2.063492063492064e-05, + "loss": 0.676, + "step": 65 + }, + { + "epoch": 0.1312127236580517, + "grad_norm": 0.4366217658351886, + "learning_rate": 2.0952380952380954e-05, + "loss": 0.6792, + "step": 66 + }, + { + "epoch": 0.13320079522862824, + "grad_norm": 0.4533555372067667, + "learning_rate": 2.126984126984127e-05, + "loss": 0.6797, + "step": 67 + }, + { + "epoch": 0.13518886679920478, + "grad_norm": 0.6802950259274722, + "learning_rate": 2.158730158730159e-05, + "loss": 0.6773, + "step": 68 + }, + { + "epoch": 0.13717693836978131, + "grad_norm": 1.1241036512188962, + "learning_rate": 2.190476190476191e-05, + "loss": 0.6783, + "step": 69 + }, + { + "epoch": 0.13916500994035785, + "grad_norm": 1.125722397572469, + "learning_rate": 2.2222222222222227e-05, + "loss": 0.6715, + "step": 70 + }, + { + "epoch": 0.1411530815109344, + "grad_norm": 0.8427474910536563, + "learning_rate": 2.253968253968254e-05, + "loss": 0.671, + "step": 71 + }, + { + "epoch": 0.14314115308151093, + "grad_norm": 1.0095817810923369, + "learning_rate": 2.2857142857142858e-05, + "loss": 0.673, + "step": 72 + }, + { + "epoch": 0.14512922465208747, + "grad_norm": 1.2860444414014913, + "learning_rate": 2.317460317460318e-05, + "loss": 0.6781, + "step": 73 + }, + { + "epoch": 0.147117296222664, + "grad_norm": 0.88467179821398, + "learning_rate": 2.3492063492063496e-05, + "loss": 0.6701, + "step": 74 + }, + { + "epoch": 0.14910536779324055, + "grad_norm": 1.1930360448025386, + "learning_rate": 2.380952380952381e-05, + "loss": 0.6714, + "step": 75 + }, + { + "epoch": 0.15109343936381708, + "grad_norm": 1.2462987345720242, + "learning_rate": 2.4126984126984128e-05, + "loss": 0.6674, + "step": 76 + }, + { + "epoch": 0.15308151093439365, + "grad_norm": 1.0610109095972218, + "learning_rate": 2.444444444444445e-05, + "loss": 0.6631, + "step": 77 + }, + { + "epoch": 0.1550695825049702, + "grad_norm": 1.13945144581175, + "learning_rate": 2.4761904761904766e-05, + "loss": 0.6686, + "step": 78 + }, + { + "epoch": 0.15705765407554673, + "grad_norm": 0.8784096423124421, + "learning_rate": 2.507936507936508e-05, + "loss": 0.666, + "step": 79 + }, + { + "epoch": 0.15904572564612326, + "grad_norm": 1.7453043654997522, + "learning_rate": 2.5396825396825397e-05, + "loss": 0.6682, + "step": 80 + }, + { + "epoch": 0.1610337972166998, + "grad_norm": 0.6565327581012248, + "learning_rate": 2.5714285714285718e-05, + "loss": 0.6619, + "step": 81 + }, + { + "epoch": 0.16302186878727634, + "grad_norm": 1.457355963249708, + "learning_rate": 2.6031746031746035e-05, + "loss": 0.6618, + "step": 82 + }, + { + "epoch": 0.16500994035785288, + "grad_norm": 1.3748440778156223, + "learning_rate": 2.6349206349206353e-05, + "loss": 0.662, + "step": 83 + }, + { + "epoch": 0.16699801192842942, + "grad_norm": 0.6576930469096256, + "learning_rate": 2.6666666666666667e-05, + "loss": 0.6645, + "step": 84 + }, + { + "epoch": 0.16898608349900596, + "grad_norm": 1.216101814201686, + "learning_rate": 2.6984126984126987e-05, + "loss": 0.6632, + "step": 85 + }, + { + "epoch": 0.1709741550695825, + "grad_norm": 1.6651254013048873, + "learning_rate": 2.7301587301587305e-05, + "loss": 0.6608, + "step": 86 + }, + { + "epoch": 0.17296222664015903, + "grad_norm": 0.736668310308581, + "learning_rate": 2.7619047619047622e-05, + "loss": 0.6614, + "step": 87 + }, + { + "epoch": 0.1749502982107356, + "grad_norm": 2.657224735711684, + "learning_rate": 2.7936507936507936e-05, + "loss": 0.6699, + "step": 88 + }, + { + "epoch": 0.17693836978131214, + "grad_norm": 1.8281737180328776, + "learning_rate": 2.8253968253968257e-05, + "loss": 0.6629, + "step": 89 + }, + { + "epoch": 0.17892644135188868, + "grad_norm": 2.458633290210547, + "learning_rate": 2.8571428571428574e-05, + "loss": 0.6602, + "step": 90 + }, + { + "epoch": 0.18091451292246521, + "grad_norm": 2.712443155137568, + "learning_rate": 2.888888888888889e-05, + "loss": 0.6749, + "step": 91 + }, + { + "epoch": 0.18290258449304175, + "grad_norm": 0.8261764251352429, + "learning_rate": 2.9206349206349206e-05, + "loss": 0.6612, + "step": 92 + }, + { + "epoch": 0.1848906560636183, + "grad_norm": 2.054547558417178, + "learning_rate": 2.9523809523809526e-05, + "loss": 0.667, + "step": 93 + }, + { + "epoch": 0.18687872763419483, + "grad_norm": 1.236783488406299, + "learning_rate": 2.9841269841269844e-05, + "loss": 0.6624, + "step": 94 + }, + { + "epoch": 0.18886679920477137, + "grad_norm": 2.162948847530382, + "learning_rate": 3.015873015873016e-05, + "loss": 0.6663, + "step": 95 + }, + { + "epoch": 0.1908548707753479, + "grad_norm": 1.930702254563974, + "learning_rate": 3.047619047619048e-05, + "loss": 0.6827, + "step": 96 + }, + { + "epoch": 0.19284294234592445, + "grad_norm": 1.609723044776149, + "learning_rate": 3.0793650793650796e-05, + "loss": 0.669, + "step": 97 + }, + { + "epoch": 0.19483101391650098, + "grad_norm": 1.5177019425537601, + "learning_rate": 3.111111111111112e-05, + "loss": 0.6689, + "step": 98 + }, + { + "epoch": 0.19681908548707752, + "grad_norm": 1.5351591826271913, + "learning_rate": 3.142857142857143e-05, + "loss": 0.6664, + "step": 99 + }, + { + "epoch": 0.1988071570576541, + "grad_norm": 1.0566494701757996, + "learning_rate": 3.1746031746031745e-05, + "loss": 0.6527, + "step": 100 + }, + { + "epoch": 0.20079522862823063, + "grad_norm": 1.4731697755973054, + "learning_rate": 3.2063492063492065e-05, + "loss": 0.6605, + "step": 101 + }, + { + "epoch": 0.20278330019880716, + "grad_norm": 1.1128607790678517, + "learning_rate": 3.2380952380952386e-05, + "loss": 0.6611, + "step": 102 + }, + { + "epoch": 0.2047713717693837, + "grad_norm": 1.478193180756222, + "learning_rate": 3.26984126984127e-05, + "loss": 0.6566, + "step": 103 + }, + { + "epoch": 0.20675944333996024, + "grad_norm": 1.3087065995108067, + "learning_rate": 3.3015873015873014e-05, + "loss": 0.66, + "step": 104 + }, + { + "epoch": 0.20874751491053678, + "grad_norm": 1.4156409974895987, + "learning_rate": 3.3333333333333335e-05, + "loss": 0.6638, + "step": 105 + }, + { + "epoch": 0.21073558648111332, + "grad_norm": 1.30062464042937, + "learning_rate": 3.3650793650793656e-05, + "loss": 0.6514, + "step": 106 + }, + { + "epoch": 0.21272365805168986, + "grad_norm": 1.4113886514836769, + "learning_rate": 3.396825396825397e-05, + "loss": 0.6569, + "step": 107 + }, + { + "epoch": 0.2147117296222664, + "grad_norm": 1.0156596270150422, + "learning_rate": 3.4285714285714284e-05, + "loss": 0.6504, + "step": 108 + }, + { + "epoch": 0.21669980119284293, + "grad_norm": 1.5825915741647385, + "learning_rate": 3.4603174603174604e-05, + "loss": 0.6533, + "step": 109 + }, + { + "epoch": 0.21868787276341947, + "grad_norm": 1.1296868604446741, + "learning_rate": 3.4920634920634925e-05, + "loss": 0.6534, + "step": 110 + }, + { + "epoch": 0.220675944333996, + "grad_norm": 1.3923998966758822, + "learning_rate": 3.523809523809524e-05, + "loss": 0.65, + "step": 111 + }, + { + "epoch": 0.22266401590457258, + "grad_norm": 1.2640733985651982, + "learning_rate": 3.555555555555555e-05, + "loss": 0.6519, + "step": 112 + }, + { + "epoch": 0.22465208747514911, + "grad_norm": 1.4802049067534304, + "learning_rate": 3.5873015873015874e-05, + "loss": 0.6508, + "step": 113 + }, + { + "epoch": 0.22664015904572565, + "grad_norm": 1.5035055412914218, + "learning_rate": 3.6190476190476195e-05, + "loss": 0.6473, + "step": 114 + }, + { + "epoch": 0.2286282306163022, + "grad_norm": 1.1937240148726864, + "learning_rate": 3.650793650793651e-05, + "loss": 0.6512, + "step": 115 + }, + { + "epoch": 0.23061630218687873, + "grad_norm": 1.2457856561461578, + "learning_rate": 3.682539682539683e-05, + "loss": 0.6447, + "step": 116 + }, + { + "epoch": 0.23260437375745527, + "grad_norm": 1.3273129343443546, + "learning_rate": 3.714285714285715e-05, + "loss": 0.6464, + "step": 117 + }, + { + "epoch": 0.2345924453280318, + "grad_norm": 1.3281865305863692, + "learning_rate": 3.7460317460317464e-05, + "loss": 0.6477, + "step": 118 + }, + { + "epoch": 0.23658051689860835, + "grad_norm": 1.0056815583069727, + "learning_rate": 3.777777777777778e-05, + "loss": 0.645, + "step": 119 + }, + { + "epoch": 0.23856858846918488, + "grad_norm": 1.382121816169979, + "learning_rate": 3.80952380952381e-05, + "loss": 0.6546, + "step": 120 + }, + { + "epoch": 0.24055666003976142, + "grad_norm": 1.3100694949960303, + "learning_rate": 3.841269841269842e-05, + "loss": 0.6488, + "step": 121 + }, + { + "epoch": 0.24254473161033796, + "grad_norm": 1.4947029004278205, + "learning_rate": 3.8730158730158734e-05, + "loss": 0.6477, + "step": 122 + }, + { + "epoch": 0.24453280318091453, + "grad_norm": 1.4644178209776242, + "learning_rate": 3.904761904761905e-05, + "loss": 0.652, + "step": 123 + }, + { + "epoch": 0.24652087475149106, + "grad_norm": 1.3769771435448332, + "learning_rate": 3.936507936507937e-05, + "loss": 0.6489, + "step": 124 + }, + { + "epoch": 0.2485089463220676, + "grad_norm": 1.0653896387340898, + "learning_rate": 3.968253968253969e-05, + "loss": 0.6459, + "step": 125 + }, + { + "epoch": 0.2504970178926441, + "grad_norm": 1.4120675480758071, + "learning_rate": 4e-05, + "loss": 0.6521, + "step": 126 + }, + { + "epoch": 0.2524850894632207, + "grad_norm": 1.2331141631565885, + "learning_rate": 4.031746031746032e-05, + "loss": 0.6412, + "step": 127 + }, + { + "epoch": 0.2544731610337972, + "grad_norm": 1.4510580599851584, + "learning_rate": 4.063492063492064e-05, + "loss": 0.6424, + "step": 128 + }, + { + "epoch": 0.25646123260437376, + "grad_norm": 1.1641813315566232, + "learning_rate": 4.095238095238095e-05, + "loss": 0.6456, + "step": 129 + }, + { + "epoch": 0.2584493041749503, + "grad_norm": 1.504580390843022, + "learning_rate": 4.126984126984128e-05, + "loss": 0.6408, + "step": 130 + }, + { + "epoch": 0.26043737574552683, + "grad_norm": 1.2257434898909374, + "learning_rate": 4.1587301587301594e-05, + "loss": 0.6429, + "step": 131 + }, + { + "epoch": 0.2624254473161034, + "grad_norm": 0.9912463823975768, + "learning_rate": 4.190476190476191e-05, + "loss": 0.6431, + "step": 132 + }, + { + "epoch": 0.2644135188866799, + "grad_norm": 1.5247568836492176, + "learning_rate": 4.222222222222223e-05, + "loss": 0.6555, + "step": 133 + }, + { + "epoch": 0.2664015904572565, + "grad_norm": 1.5773163450396874, + "learning_rate": 4.253968253968254e-05, + "loss": 0.6545, + "step": 134 + }, + { + "epoch": 0.268389662027833, + "grad_norm": 1.291828731349488, + "learning_rate": 4.2857142857142856e-05, + "loss": 0.6429, + "step": 135 + }, + { + "epoch": 0.27037773359840955, + "grad_norm": 1.0985522555843004, + "learning_rate": 4.317460317460318e-05, + "loss": 0.6408, + "step": 136 + }, + { + "epoch": 0.27236580516898606, + "grad_norm": 1.5828560475019822, + "learning_rate": 4.349206349206349e-05, + "loss": 0.6594, + "step": 137 + }, + { + "epoch": 0.27435387673956263, + "grad_norm": 1.3407587991159484, + "learning_rate": 4.380952380952382e-05, + "loss": 0.6382, + "step": 138 + }, + { + "epoch": 0.27634194831013914, + "grad_norm": 1.1882787897778562, + "learning_rate": 4.412698412698413e-05, + "loss": 0.6435, + "step": 139 + }, + { + "epoch": 0.2783300198807157, + "grad_norm": 1.4743146172658406, + "learning_rate": 4.444444444444445e-05, + "loss": 0.6426, + "step": 140 + }, + { + "epoch": 0.2803180914512923, + "grad_norm": 1.3198867552344011, + "learning_rate": 4.476190476190477e-05, + "loss": 0.6418, + "step": 141 + }, + { + "epoch": 0.2823061630218688, + "grad_norm": 1.087940001986759, + "learning_rate": 4.507936507936508e-05, + "loss": 0.6438, + "step": 142 + }, + { + "epoch": 0.28429423459244535, + "grad_norm": 1.7460817891906895, + "learning_rate": 4.53968253968254e-05, + "loss": 0.6438, + "step": 143 + }, + { + "epoch": 0.28628230616302186, + "grad_norm": 1.072098243868636, + "learning_rate": 4.5714285714285716e-05, + "loss": 0.6449, + "step": 144 + }, + { + "epoch": 0.2882703777335984, + "grad_norm": 2.1835116598411504, + "learning_rate": 4.603174603174603e-05, + "loss": 0.6489, + "step": 145 + }, + { + "epoch": 0.29025844930417494, + "grad_norm": 1.879705553582977, + "learning_rate": 4.634920634920636e-05, + "loss": 0.6514, + "step": 146 + }, + { + "epoch": 0.2922465208747515, + "grad_norm": 1.4090093524508818, + "learning_rate": 4.666666666666667e-05, + "loss": 0.6409, + "step": 147 + }, + { + "epoch": 0.294234592445328, + "grad_norm": 1.8436589830840873, + "learning_rate": 4.698412698412699e-05, + "loss": 0.6442, + "step": 148 + }, + { + "epoch": 0.2962226640159046, + "grad_norm": 1.0212379094945134, + "learning_rate": 4.7301587301587306e-05, + "loss": 0.6402, + "step": 149 + }, + { + "epoch": 0.2982107355864811, + "grad_norm": 1.7700790575852108, + "learning_rate": 4.761904761904762e-05, + "loss": 0.6415, + "step": 150 + }, + { + "epoch": 0.30019880715705766, + "grad_norm": 1.6067327390517923, + "learning_rate": 4.793650793650794e-05, + "loss": 0.6438, + "step": 151 + }, + { + "epoch": 0.30218687872763417, + "grad_norm": 1.1795271072398732, + "learning_rate": 4.8253968253968255e-05, + "loss": 0.635, + "step": 152 + }, + { + "epoch": 0.30417495029821073, + "grad_norm": 1.5562700844547241, + "learning_rate": 4.857142857142857e-05, + "loss": 0.6395, + "step": 153 + }, + { + "epoch": 0.3061630218687873, + "grad_norm": 1.1852659442882558, + "learning_rate": 4.88888888888889e-05, + "loss": 0.6353, + "step": 154 + }, + { + "epoch": 0.3081510934393638, + "grad_norm": 1.303291076586374, + "learning_rate": 4.920634920634921e-05, + "loss": 0.639, + "step": 155 + }, + { + "epoch": 0.3101391650099404, + "grad_norm": 1.3279480936113928, + "learning_rate": 4.952380952380953e-05, + "loss": 0.6408, + "step": 156 + }, + { + "epoch": 0.3121272365805169, + "grad_norm": 1.2480990916236967, + "learning_rate": 4.9841269841269845e-05, + "loss": 0.6395, + "step": 157 + }, + { + "epoch": 0.31411530815109345, + "grad_norm": 1.7232506056868664, + "learning_rate": 5.015873015873016e-05, + "loss": 0.6429, + "step": 158 + }, + { + "epoch": 0.31610337972166996, + "grad_norm": 1.1165626843230316, + "learning_rate": 5.047619047619048e-05, + "loss": 0.6377, + "step": 159 + }, + { + "epoch": 0.31809145129224653, + "grad_norm": 1.365984266306236, + "learning_rate": 5.0793650793650794e-05, + "loss": 0.6385, + "step": 160 + }, + { + "epoch": 0.32007952286282304, + "grad_norm": 1.161379924883256, + "learning_rate": 5.111111111111111e-05, + "loss": 0.6342, + "step": 161 + }, + { + "epoch": 0.3220675944333996, + "grad_norm": 1.5501649766832735, + "learning_rate": 5.1428571428571436e-05, + "loss": 0.6397, + "step": 162 + }, + { + "epoch": 0.3240556660039761, + "grad_norm": 1.236059822703464, + "learning_rate": 5.174603174603175e-05, + "loss": 0.6345, + "step": 163 + }, + { + "epoch": 0.3260437375745527, + "grad_norm": 1.1632007138023943, + "learning_rate": 5.206349206349207e-05, + "loss": 0.6331, + "step": 164 + }, + { + "epoch": 0.32803180914512925, + "grad_norm": 1.8957350418517902, + "learning_rate": 5.2380952380952384e-05, + "loss": 0.6424, + "step": 165 + }, + { + "epoch": 0.33001988071570576, + "grad_norm": 1.2762565748269343, + "learning_rate": 5.2698412698412705e-05, + "loss": 0.6388, + "step": 166 + }, + { + "epoch": 0.3320079522862823, + "grad_norm": 0.9926513996495598, + "learning_rate": 5.301587301587302e-05, + "loss": 0.6334, + "step": 167 + }, + { + "epoch": 0.33399602385685884, + "grad_norm": 1.5072734890729735, + "learning_rate": 5.333333333333333e-05, + "loss": 0.6402, + "step": 168 + }, + { + "epoch": 0.3359840954274354, + "grad_norm": 1.6848000191372772, + "learning_rate": 5.3650793650793654e-05, + "loss": 0.6314, + "step": 169 + }, + { + "epoch": 0.3379721669980119, + "grad_norm": 1.4016229963298041, + "learning_rate": 5.3968253968253975e-05, + "loss": 0.6278, + "step": 170 + }, + { + "epoch": 0.3399602385685885, + "grad_norm": 1.0923456202776507, + "learning_rate": 5.4285714285714295e-05, + "loss": 0.6297, + "step": 171 + }, + { + "epoch": 0.341948310139165, + "grad_norm": 2.5232602422771997, + "learning_rate": 5.460317460317461e-05, + "loss": 0.6382, + "step": 172 + }, + { + "epoch": 0.34393638170974156, + "grad_norm": 1.3834050979851695, + "learning_rate": 5.4920634920634923e-05, + "loss": 0.634, + "step": 173 + }, + { + "epoch": 0.34592445328031807, + "grad_norm": 3.0435619097934263, + "learning_rate": 5.5238095238095244e-05, + "loss": 0.6523, + "step": 174 + }, + { + "epoch": 0.34791252485089463, + "grad_norm": 2.9421056263010574, + "learning_rate": 5.555555555555556e-05, + "loss": 0.6548, + "step": 175 + }, + { + "epoch": 0.3499005964214712, + "grad_norm": 1.4643678503113071, + "learning_rate": 5.587301587301587e-05, + "loss": 0.6332, + "step": 176 + }, + { + "epoch": 0.3518886679920477, + "grad_norm": 1.9166228640874332, + "learning_rate": 5.619047619047619e-05, + "loss": 0.6409, + "step": 177 + }, + { + "epoch": 0.3538767395626243, + "grad_norm": 1.4000017299232366, + "learning_rate": 5.6507936507936514e-05, + "loss": 0.638, + "step": 178 + }, + { + "epoch": 0.3558648111332008, + "grad_norm": 1.6183232051146397, + "learning_rate": 5.6825396825396835e-05, + "loss": 0.6336, + "step": 179 + }, + { + "epoch": 0.35785288270377735, + "grad_norm": 1.5556653610365612, + "learning_rate": 5.714285714285715e-05, + "loss": 0.6382, + "step": 180 + }, + { + "epoch": 0.35984095427435386, + "grad_norm": 1.2516421633034058, + "learning_rate": 5.746031746031746e-05, + "loss": 0.6289, + "step": 181 + }, + { + "epoch": 0.36182902584493043, + "grad_norm": 1.3620554525101845, + "learning_rate": 5.777777777777778e-05, + "loss": 0.637, + "step": 182 + }, + { + "epoch": 0.36381709741550694, + "grad_norm": 0.8661709785769057, + "learning_rate": 5.80952380952381e-05, + "loss": 0.6379, + "step": 183 + }, + { + "epoch": 0.3658051689860835, + "grad_norm": 1.4436211568633606, + "learning_rate": 5.841269841269841e-05, + "loss": 0.6393, + "step": 184 + }, + { + "epoch": 0.36779324055666, + "grad_norm": 1.12234286687745, + "learning_rate": 5.873015873015873e-05, + "loss": 0.6326, + "step": 185 + }, + { + "epoch": 0.3697813121272366, + "grad_norm": 1.5736829330194766, + "learning_rate": 5.904761904761905e-05, + "loss": 0.6331, + "step": 186 + }, + { + "epoch": 0.3717693836978131, + "grad_norm": 1.3310376751788484, + "learning_rate": 5.9365079365079374e-05, + "loss": 0.6398, + "step": 187 + }, + { + "epoch": 0.37375745526838966, + "grad_norm": 1.2257869128697072, + "learning_rate": 5.968253968253969e-05, + "loss": 0.6335, + "step": 188 + }, + { + "epoch": 0.3757455268389662, + "grad_norm": 1.6139532041614482, + "learning_rate": 6.000000000000001e-05, + "loss": 0.6392, + "step": 189 + }, + { + "epoch": 0.37773359840954274, + "grad_norm": 1.3350592444809761, + "learning_rate": 6.031746031746032e-05, + "loss": 0.6382, + "step": 190 + }, + { + "epoch": 0.3797216699801193, + "grad_norm": 1.2377247673288536, + "learning_rate": 6.0634920634920636e-05, + "loss": 0.6246, + "step": 191 + }, + { + "epoch": 0.3817097415506958, + "grad_norm": 1.49455356135162, + "learning_rate": 6.095238095238096e-05, + "loss": 0.6318, + "step": 192 + }, + { + "epoch": 0.3836978131212724, + "grad_norm": 1.596684995637217, + "learning_rate": 6.126984126984128e-05, + "loss": 0.6287, + "step": 193 + }, + { + "epoch": 0.3856858846918489, + "grad_norm": 0.6665298437010337, + "learning_rate": 6.158730158730159e-05, + "loss": 0.6263, + "step": 194 + }, + { + "epoch": 0.38767395626242546, + "grad_norm": 1.756691203631161, + "learning_rate": 6.19047619047619e-05, + "loss": 0.6267, + "step": 195 + }, + { + "epoch": 0.38966202783300197, + "grad_norm": 1.0331188464436891, + "learning_rate": 6.222222222222223e-05, + "loss": 0.6212, + "step": 196 + }, + { + "epoch": 0.39165009940357853, + "grad_norm": 1.122470045661076, + "learning_rate": 6.253968253968255e-05, + "loss": 0.6284, + "step": 197 + }, + { + "epoch": 0.39363817097415504, + "grad_norm": 1.7664037452706398, + "learning_rate": 6.285714285714286e-05, + "loss": 0.642, + "step": 198 + }, + { + "epoch": 0.3956262425447316, + "grad_norm": 1.486929540611025, + "learning_rate": 6.317460317460318e-05, + "loss": 0.6246, + "step": 199 + }, + { + "epoch": 0.3976143141153082, + "grad_norm": 1.2175342983856243, + "learning_rate": 6.349206349206349e-05, + "loss": 0.6313, + "step": 200 + }, + { + "epoch": 0.3996023856858847, + "grad_norm": 1.4121892732145778, + "learning_rate": 6.380952380952382e-05, + "loss": 0.6329, + "step": 201 + }, + { + "epoch": 0.40159045725646125, + "grad_norm": 1.1436798208988694, + "learning_rate": 6.412698412698413e-05, + "loss": 0.6233, + "step": 202 + }, + { + "epoch": 0.40357852882703776, + "grad_norm": 1.8368150848847797, + "learning_rate": 6.444444444444446e-05, + "loss": 0.6224, + "step": 203 + }, + { + "epoch": 0.40556660039761433, + "grad_norm": 1.065714896345243, + "learning_rate": 6.476190476190477e-05, + "loss": 0.6279, + "step": 204 + }, + { + "epoch": 0.40755467196819084, + "grad_norm": 1.4516811834829857, + "learning_rate": 6.507936507936509e-05, + "loss": 0.635, + "step": 205 + }, + { + "epoch": 0.4095427435387674, + "grad_norm": 1.4806307032968837, + "learning_rate": 6.53968253968254e-05, + "loss": 0.6236, + "step": 206 + }, + { + "epoch": 0.4115308151093439, + "grad_norm": 1.8498374502351982, + "learning_rate": 6.571428571428571e-05, + "loss": 0.6285, + "step": 207 + }, + { + "epoch": 0.4135188866799205, + "grad_norm": 0.8403035362769382, + "learning_rate": 6.603174603174603e-05, + "loss": 0.6231, + "step": 208 + }, + { + "epoch": 0.415506958250497, + "grad_norm": 2.091792915751743, + "learning_rate": 6.634920634920636e-05, + "loss": 0.6309, + "step": 209 + }, + { + "epoch": 0.41749502982107356, + "grad_norm": 1.26117745196078, + "learning_rate": 6.666666666666667e-05, + "loss": 0.6325, + "step": 210 + }, + { + "epoch": 0.4194831013916501, + "grad_norm": 1.783033140481242, + "learning_rate": 6.6984126984127e-05, + "loss": 0.6309, + "step": 211 + }, + { + "epoch": 0.42147117296222664, + "grad_norm": 1.8515301788648455, + "learning_rate": 6.730158730158731e-05, + "loss": 0.6382, + "step": 212 + }, + { + "epoch": 0.4234592445328032, + "grad_norm": 1.1778109141989737, + "learning_rate": 6.761904761904763e-05, + "loss": 0.6289, + "step": 213 + }, + { + "epoch": 0.4254473161033797, + "grad_norm": 2.54816769327095, + "learning_rate": 6.793650793650794e-05, + "loss": 0.6371, + "step": 214 + }, + { + "epoch": 0.4274353876739563, + "grad_norm": 2.048704735899834, + "learning_rate": 6.825396825396825e-05, + "loss": 0.6398, + "step": 215 + }, + { + "epoch": 0.4294234592445328, + "grad_norm": 2.1019781611448853, + "learning_rate": 6.857142857142857e-05, + "loss": 0.6295, + "step": 216 + }, + { + "epoch": 0.43141153081510936, + "grad_norm": 1.7281786345970265, + "learning_rate": 6.88888888888889e-05, + "loss": 0.6275, + "step": 217 + }, + { + "epoch": 0.43339960238568587, + "grad_norm": 2.14951909400337, + "learning_rate": 6.920634920634921e-05, + "loss": 0.6331, + "step": 218 + }, + { + "epoch": 0.43538767395626243, + "grad_norm": 1.7356401726987714, + "learning_rate": 6.952380952380954e-05, + "loss": 0.6333, + "step": 219 + }, + { + "epoch": 0.43737574552683894, + "grad_norm": 1.9690563790102926, + "learning_rate": 6.984126984126985e-05, + "loss": 0.6267, + "step": 220 + }, + { + "epoch": 0.4393638170974155, + "grad_norm": 1.3726181777240785, + "learning_rate": 7.015873015873016e-05, + "loss": 0.6292, + "step": 221 + }, + { + "epoch": 0.441351888667992, + "grad_norm": 1.8939096029386184, + "learning_rate": 7.047619047619048e-05, + "loss": 0.6188, + "step": 222 + }, + { + "epoch": 0.4433399602385686, + "grad_norm": 1.2762772953205688, + "learning_rate": 7.079365079365079e-05, + "loss": 0.6273, + "step": 223 + }, + { + "epoch": 0.44532803180914515, + "grad_norm": 1.776596462339463, + "learning_rate": 7.11111111111111e-05, + "loss": 0.6239, + "step": 224 + }, + { + "epoch": 0.44731610337972166, + "grad_norm": 1.3274942963495349, + "learning_rate": 7.142857142857143e-05, + "loss": 0.6241, + "step": 225 + }, + { + "epoch": 0.44930417495029823, + "grad_norm": 1.8591216328666749, + "learning_rate": 7.174603174603175e-05, + "loss": 0.6249, + "step": 226 + }, + { + "epoch": 0.45129224652087474, + "grad_norm": 1.632104839191781, + "learning_rate": 7.206349206349208e-05, + "loss": 0.622, + "step": 227 + }, + { + "epoch": 0.4532803180914513, + "grad_norm": 1.5321880702968802, + "learning_rate": 7.238095238095239e-05, + "loss": 0.624, + "step": 228 + }, + { + "epoch": 0.4552683896620278, + "grad_norm": 1.3222149905940208, + "learning_rate": 7.26984126984127e-05, + "loss": 0.6176, + "step": 229 + }, + { + "epoch": 0.4572564612326044, + "grad_norm": 1.313836019162892, + "learning_rate": 7.301587301587302e-05, + "loss": 0.6143, + "step": 230 + }, + { + "epoch": 0.4592445328031809, + "grad_norm": 1.131038423221744, + "learning_rate": 7.333333333333333e-05, + "loss": 0.6188, + "step": 231 + }, + { + "epoch": 0.46123260437375746, + "grad_norm": 1.283678851750674, + "learning_rate": 7.365079365079366e-05, + "loss": 0.619, + "step": 232 + }, + { + "epoch": 0.46322067594433397, + "grad_norm": 1.6515171578755334, + "learning_rate": 7.396825396825397e-05, + "loss": 0.6262, + "step": 233 + }, + { + "epoch": 0.46520874751491054, + "grad_norm": 1.2107665827950345, + "learning_rate": 7.42857142857143e-05, + "loss": 0.6134, + "step": 234 + }, + { + "epoch": 0.4671968190854871, + "grad_norm": 1.3081158148843528, + "learning_rate": 7.460317460317461e-05, + "loss": 0.6207, + "step": 235 + }, + { + "epoch": 0.4691848906560636, + "grad_norm": 1.1721076028263735, + "learning_rate": 7.492063492063493e-05, + "loss": 0.625, + "step": 236 + }, + { + "epoch": 0.4711729622266402, + "grad_norm": 1.5047073563346702, + "learning_rate": 7.523809523809524e-05, + "loss": 0.62, + "step": 237 + }, + { + "epoch": 0.4731610337972167, + "grad_norm": 0.8406971392950173, + "learning_rate": 7.555555555555556e-05, + "loss": 0.621, + "step": 238 + }, + { + "epoch": 0.47514910536779326, + "grad_norm": 1.2339916132908735, + "learning_rate": 7.587301587301587e-05, + "loss": 0.6271, + "step": 239 + }, + { + "epoch": 0.47713717693836977, + "grad_norm": 1.3926448624868792, + "learning_rate": 7.61904761904762e-05, + "loss": 0.6218, + "step": 240 + }, + { + "epoch": 0.47912524850894633, + "grad_norm": 2.204044181851112, + "learning_rate": 7.650793650793651e-05, + "loss": 0.6193, + "step": 241 + }, + { + "epoch": 0.48111332007952284, + "grad_norm": 0.8533613284104712, + "learning_rate": 7.682539682539684e-05, + "loss": 0.6136, + "step": 242 + }, + { + "epoch": 0.4831013916500994, + "grad_norm": 1.9892818324819723, + "learning_rate": 7.714285714285715e-05, + "loss": 0.6392, + "step": 243 + }, + { + "epoch": 0.4850894632206759, + "grad_norm": 1.3210748896263227, + "learning_rate": 7.746031746031747e-05, + "loss": 0.6274, + "step": 244 + }, + { + "epoch": 0.4870775347912525, + "grad_norm": 1.257534879967076, + "learning_rate": 7.777777777777778e-05, + "loss": 0.6267, + "step": 245 + }, + { + "epoch": 0.48906560636182905, + "grad_norm": 1.5681584428773365, + "learning_rate": 7.80952380952381e-05, + "loss": 0.6334, + "step": 246 + }, + { + "epoch": 0.49105367793240556, + "grad_norm": 1.217282156325022, + "learning_rate": 7.841269841269841e-05, + "loss": 0.6188, + "step": 247 + }, + { + "epoch": 0.49304174950298213, + "grad_norm": 1.668210588102185, + "learning_rate": 7.873015873015874e-05, + "loss": 0.6289, + "step": 248 + }, + { + "epoch": 0.49502982107355864, + "grad_norm": 1.1732055301577373, + "learning_rate": 7.904761904761905e-05, + "loss": 0.6277, + "step": 249 + }, + { + "epoch": 0.4970178926441352, + "grad_norm": 1.1391292798020574, + "learning_rate": 7.936507936507938e-05, + "loss": 0.6252, + "step": 250 + }, + { + "epoch": 0.4990059642147117, + "grad_norm": 1.3420755566613047, + "learning_rate": 7.968253968253969e-05, + "loss": 0.6229, + "step": 251 + }, + { + "epoch": 0.5009940357852882, + "grad_norm": 1.7212519000230158, + "learning_rate": 8e-05, + "loss": 0.6283, + "step": 252 + }, + { + "epoch": 0.5029821073558648, + "grad_norm": 0.8460706799764476, + "learning_rate": 7.999996145566446e-05, + "loss": 0.6277, + "step": 253 + }, + { + "epoch": 0.5049701789264414, + "grad_norm": 1.2028891530381527, + "learning_rate": 7.999984582273212e-05, + "loss": 0.6298, + "step": 254 + }, + { + "epoch": 0.5069582504970179, + "grad_norm": 1.6321071690462996, + "learning_rate": 7.999965310142581e-05, + "loss": 0.6218, + "step": 255 + }, + { + "epoch": 0.5089463220675944, + "grad_norm": 0.8020880986907298, + "learning_rate": 7.999938329211695e-05, + "loss": 0.6145, + "step": 256 + }, + { + "epoch": 0.510934393638171, + "grad_norm": 1.5123244136663403, + "learning_rate": 7.999903639532553e-05, + "loss": 0.6233, + "step": 257 + }, + { + "epoch": 0.5129224652087475, + "grad_norm": 0.8619895667367529, + "learning_rate": 7.99986124117201e-05, + "loss": 0.6259, + "step": 258 + }, + { + "epoch": 0.5149105367793241, + "grad_norm": 1.674504615597631, + "learning_rate": 7.999811134211777e-05, + "loss": 0.623, + "step": 259 + }, + { + "epoch": 0.5168986083499006, + "grad_norm": 1.244705549383792, + "learning_rate": 7.99975331874842e-05, + "loss": 0.6259, + "step": 260 + }, + { + "epoch": 0.5188866799204771, + "grad_norm": 1.4164036526291948, + "learning_rate": 7.999687794893362e-05, + "loss": 0.6254, + "step": 261 + }, + { + "epoch": 0.5208747514910537, + "grad_norm": 1.1420458314329858, + "learning_rate": 7.999614562772881e-05, + "loss": 0.6224, + "step": 262 + }, + { + "epoch": 0.5228628230616302, + "grad_norm": 0.9971122950245115, + "learning_rate": 7.999533622528113e-05, + "loss": 0.6167, + "step": 263 + }, + { + "epoch": 0.5248508946322068, + "grad_norm": 1.53841350114554, + "learning_rate": 7.999444974315046e-05, + "loss": 0.6203, + "step": 264 + }, + { + "epoch": 0.5268389662027833, + "grad_norm": 1.355905799554643, + "learning_rate": 7.999348618304524e-05, + "loss": 0.6209, + "step": 265 + }, + { + "epoch": 0.5288270377733598, + "grad_norm": 1.1382955214819985, + "learning_rate": 7.999244554682249e-05, + "loss": 0.6137, + "step": 266 + }, + { + "epoch": 0.5308151093439364, + "grad_norm": 1.2022228391943435, + "learning_rate": 7.99913278364877e-05, + "loss": 0.6145, + "step": 267 + }, + { + "epoch": 0.532803180914513, + "grad_norm": 1.086139055972047, + "learning_rate": 7.999013305419497e-05, + "loss": 0.6145, + "step": 268 + }, + { + "epoch": 0.5347912524850894, + "grad_norm": 1.322274294817971, + "learning_rate": 7.998886120224688e-05, + "loss": 0.6234, + "step": 269 + }, + { + "epoch": 0.536779324055666, + "grad_norm": 0.9234044414783897, + "learning_rate": 7.998751228309457e-05, + "loss": 0.6144, + "step": 270 + }, + { + "epoch": 0.5387673956262425, + "grad_norm": 0.9326761013754535, + "learning_rate": 7.998608629933773e-05, + "loss": 0.61, + "step": 271 + }, + { + "epoch": 0.5407554671968191, + "grad_norm": 1.5279987086559388, + "learning_rate": 7.99845832537245e-05, + "loss": 0.6174, + "step": 272 + }, + { + "epoch": 0.5427435387673957, + "grad_norm": 0.9928341222876796, + "learning_rate": 7.99830031491516e-05, + "loss": 0.616, + "step": 273 + }, + { + "epoch": 0.5447316103379721, + "grad_norm": 1.3570459611370076, + "learning_rate": 7.998134598866422e-05, + "loss": 0.6187, + "step": 274 + }, + { + "epoch": 0.5467196819085487, + "grad_norm": 1.4557589984647394, + "learning_rate": 7.997961177545607e-05, + "loss": 0.6135, + "step": 275 + }, + { + "epoch": 0.5487077534791253, + "grad_norm": 0.7773435304804408, + "learning_rate": 7.997780051286937e-05, + "loss": 0.612, + "step": 276 + }, + { + "epoch": 0.5506958250497018, + "grad_norm": 1.223588643454511, + "learning_rate": 7.997591220439479e-05, + "loss": 0.6148, + "step": 277 + }, + { + "epoch": 0.5526838966202783, + "grad_norm": 1.4724994222951164, + "learning_rate": 7.997394685367153e-05, + "loss": 0.6149, + "step": 278 + }, + { + "epoch": 0.5546719681908548, + "grad_norm": 0.8204200684143781, + "learning_rate": 7.997190446448724e-05, + "loss": 0.6115, + "step": 279 + }, + { + "epoch": 0.5566600397614314, + "grad_norm": 1.3559612721067071, + "learning_rate": 7.996978504077805e-05, + "loss": 0.6089, + "step": 280 + }, + { + "epoch": 0.558648111332008, + "grad_norm": 1.0310273576587536, + "learning_rate": 7.996758858662855e-05, + "loss": 0.6174, + "step": 281 + }, + { + "epoch": 0.5606361829025845, + "grad_norm": 0.9946930215525382, + "learning_rate": 7.996531510627176e-05, + "loss": 0.6153, + "step": 282 + }, + { + "epoch": 0.562624254473161, + "grad_norm": 1.1224346378529806, + "learning_rate": 7.996296460408921e-05, + "loss": 0.6119, + "step": 283 + }, + { + "epoch": 0.5646123260437376, + "grad_norm": 1.3765896724427593, + "learning_rate": 7.996053708461078e-05, + "loss": 0.6156, + "step": 284 + }, + { + "epoch": 0.5666003976143141, + "grad_norm": 1.0089569688803084, + "learning_rate": 7.995803255251488e-05, + "loss": 0.6088, + "step": 285 + }, + { + "epoch": 0.5685884691848907, + "grad_norm": 0.950972323050223, + "learning_rate": 7.995545101262825e-05, + "loss": 0.6104, + "step": 286 + }, + { + "epoch": 0.5705765407554672, + "grad_norm": 1.4332474212744974, + "learning_rate": 7.995279246992608e-05, + "loss": 0.6166, + "step": 287 + }, + { + "epoch": 0.5725646123260437, + "grad_norm": 1.1557864927220203, + "learning_rate": 7.995005692953196e-05, + "loss": 0.607, + "step": 288 + }, + { + "epoch": 0.5745526838966203, + "grad_norm": 0.8094455060578747, + "learning_rate": 7.994724439671787e-05, + "loss": 0.6079, + "step": 289 + }, + { + "epoch": 0.5765407554671969, + "grad_norm": 1.058813447760782, + "learning_rate": 7.994435487690418e-05, + "loss": 0.6159, + "step": 290 + }, + { + "epoch": 0.5785288270377733, + "grad_norm": 1.7842490215031124, + "learning_rate": 7.994138837565962e-05, + "loss": 0.6064, + "step": 291 + }, + { + "epoch": 0.5805168986083499, + "grad_norm": 0.7937318931864666, + "learning_rate": 7.993834489870127e-05, + "loss": 0.61, + "step": 292 + }, + { + "epoch": 0.5825049701789264, + "grad_norm": 2.0448392448161266, + "learning_rate": 7.993522445189458e-05, + "loss": 0.6187, + "step": 293 + }, + { + "epoch": 0.584493041749503, + "grad_norm": 1.1616482588221761, + "learning_rate": 7.993202704125332e-05, + "loss": 0.6168, + "step": 294 + }, + { + "epoch": 0.5864811133200796, + "grad_norm": 1.7634514220529318, + "learning_rate": 7.992875267293958e-05, + "loss": 0.6202, + "step": 295 + }, + { + "epoch": 0.588469184890656, + "grad_norm": 1.3557395854086942, + "learning_rate": 7.99254013532638e-05, + "loss": 0.6287, + "step": 296 + }, + { + "epoch": 0.5904572564612326, + "grad_norm": 1.221924990086436, + "learning_rate": 7.99219730886847e-05, + "loss": 0.6177, + "step": 297 + }, + { + "epoch": 0.5924453280318092, + "grad_norm": 1.1570566780155553, + "learning_rate": 7.991846788580928e-05, + "loss": 0.6202, + "step": 298 + }, + { + "epoch": 0.5944333996023857, + "grad_norm": 0.8682851210033183, + "learning_rate": 7.991488575139283e-05, + "loss": 0.6108, + "step": 299 + }, + { + "epoch": 0.5964214711729622, + "grad_norm": 0.9892322185926827, + "learning_rate": 7.99112266923389e-05, + "loss": 0.6093, + "step": 300 + }, + { + "epoch": 0.5984095427435387, + "grad_norm": 1.3225369418198032, + "learning_rate": 7.990749071569928e-05, + "loss": 0.6116, + "step": 301 + }, + { + "epoch": 0.6003976143141153, + "grad_norm": 0.7090274559838418, + "learning_rate": 7.990367782867403e-05, + "loss": 0.6068, + "step": 302 + }, + { + "epoch": 0.6023856858846919, + "grad_norm": 1.1894865112831852, + "learning_rate": 7.989978803861138e-05, + "loss": 0.6106, + "step": 303 + }, + { + "epoch": 0.6043737574552683, + "grad_norm": 0.6833264633772178, + "learning_rate": 7.98958213530078e-05, + "loss": 0.6058, + "step": 304 + }, + { + "epoch": 0.6063618290258449, + "grad_norm": 0.9084651046263741, + "learning_rate": 7.989177777950798e-05, + "loss": 0.6014, + "step": 305 + }, + { + "epoch": 0.6083499005964215, + "grad_norm": 0.9853511757871085, + "learning_rate": 7.988765732590476e-05, + "loss": 0.6043, + "step": 306 + }, + { + "epoch": 0.610337972166998, + "grad_norm": 1.1489434663121463, + "learning_rate": 7.988346000013912e-05, + "loss": 0.6042, + "step": 307 + }, + { + "epoch": 0.6123260437375746, + "grad_norm": 0.9477747519287059, + "learning_rate": 7.987918581030024e-05, + "loss": 0.6039, + "step": 308 + }, + { + "epoch": 0.614314115308151, + "grad_norm": 1.1393103642754758, + "learning_rate": 7.987483476462539e-05, + "loss": 0.6028, + "step": 309 + }, + { + "epoch": 0.6163021868787276, + "grad_norm": 1.2313579785241444, + "learning_rate": 7.98704068715e-05, + "loss": 0.6103, + "step": 310 + }, + { + "epoch": 0.6182902584493042, + "grad_norm": 0.9717710635949429, + "learning_rate": 7.986590213945756e-05, + "loss": 0.606, + "step": 311 + }, + { + "epoch": 0.6202783300198808, + "grad_norm": 1.0672532374525363, + "learning_rate": 7.98613205771797e-05, + "loss": 0.6003, + "step": 312 + }, + { + "epoch": 0.6222664015904572, + "grad_norm": 1.3583742092426443, + "learning_rate": 7.985666219349604e-05, + "loss": 0.6022, + "step": 313 + }, + { + "epoch": 0.6242544731610338, + "grad_norm": 0.5837502834461586, + "learning_rate": 7.985192699738432e-05, + "loss": 0.5966, + "step": 314 + }, + { + "epoch": 0.6262425447316103, + "grad_norm": 0.6446776148368711, + "learning_rate": 7.984711499797029e-05, + "loss": 0.6038, + "step": 315 + }, + { + "epoch": 0.6282306163021869, + "grad_norm": 1.2386506977675524, + "learning_rate": 7.98422262045277e-05, + "loss": 0.6077, + "step": 316 + }, + { + "epoch": 0.6302186878727635, + "grad_norm": 1.1751514354670538, + "learning_rate": 7.983726062647833e-05, + "loss": 0.6042, + "step": 317 + }, + { + "epoch": 0.6322067594433399, + "grad_norm": 0.8754529089309102, + "learning_rate": 7.983221827339192e-05, + "loss": 0.6005, + "step": 318 + }, + { + "epoch": 0.6341948310139165, + "grad_norm": 1.070805762546688, + "learning_rate": 7.982709915498618e-05, + "loss": 0.603, + "step": 319 + }, + { + "epoch": 0.6361829025844931, + "grad_norm": 1.2753097622493266, + "learning_rate": 7.982190328112674e-05, + "loss": 0.601, + "step": 320 + }, + { + "epoch": 0.6381709741550696, + "grad_norm": 0.755670446629981, + "learning_rate": 7.981663066182721e-05, + "loss": 0.6011, + "step": 321 + }, + { + "epoch": 0.6401590457256461, + "grad_norm": 0.7148388951079416, + "learning_rate": 7.981128130724907e-05, + "loss": 0.6, + "step": 322 + }, + { + "epoch": 0.6421471172962226, + "grad_norm": 0.9122439016384183, + "learning_rate": 7.980585522770166e-05, + "loss": 0.6017, + "step": 323 + }, + { + "epoch": 0.6441351888667992, + "grad_norm": 1.3055890086405564, + "learning_rate": 7.980035243364221e-05, + "loss": 0.6086, + "step": 324 + }, + { + "epoch": 0.6461232604373758, + "grad_norm": 0.8349565637932869, + "learning_rate": 7.97947729356758e-05, + "loss": 0.5988, + "step": 325 + }, + { + "epoch": 0.6481113320079522, + "grad_norm": 0.9251569550381776, + "learning_rate": 7.978911674455538e-05, + "loss": 0.6086, + "step": 326 + }, + { + "epoch": 0.6500994035785288, + "grad_norm": 0.9646584755316521, + "learning_rate": 7.978338387118158e-05, + "loss": 0.602, + "step": 327 + }, + { + "epoch": 0.6520874751491054, + "grad_norm": 0.941633551491786, + "learning_rate": 7.977757432660295e-05, + "loss": 0.6055, + "step": 328 + }, + { + "epoch": 0.6540755467196819, + "grad_norm": 0.9205186480063723, + "learning_rate": 7.97716881220157e-05, + "loss": 0.6039, + "step": 329 + }, + { + "epoch": 0.6560636182902585, + "grad_norm": 0.9330713864830623, + "learning_rate": 7.976572526876386e-05, + "loss": 0.6007, + "step": 330 + }, + { + "epoch": 0.658051689860835, + "grad_norm": 1.259958088941969, + "learning_rate": 7.975968577833911e-05, + "loss": 0.6045, + "step": 331 + }, + { + "epoch": 0.6600397614314115, + "grad_norm": 0.8298038712817966, + "learning_rate": 7.975356966238087e-05, + "loss": 0.5998, + "step": 332 + }, + { + "epoch": 0.6620278330019881, + "grad_norm": 0.7181692734835541, + "learning_rate": 7.974737693267621e-05, + "loss": 0.6029, + "step": 333 + }, + { + "epoch": 0.6640159045725647, + "grad_norm": 0.7747035845063713, + "learning_rate": 7.97411076011599e-05, + "loss": 0.595, + "step": 334 + }, + { + "epoch": 0.6660039761431411, + "grad_norm": 0.8763620703275192, + "learning_rate": 7.973476167991426e-05, + "loss": 0.5947, + "step": 335 + }, + { + "epoch": 0.6679920477137177, + "grad_norm": 1.0767794678895612, + "learning_rate": 7.972833918116926e-05, + "loss": 0.596, + "step": 336 + }, + { + "epoch": 0.6699801192842942, + "grad_norm": 1.2632858490991365, + "learning_rate": 7.972184011730246e-05, + "loss": 0.5967, + "step": 337 + }, + { + "epoch": 0.6719681908548708, + "grad_norm": 0.8583135331342279, + "learning_rate": 7.971526450083897e-05, + "loss": 0.5942, + "step": 338 + }, + { + "epoch": 0.6739562624254473, + "grad_norm": 1.080095604857901, + "learning_rate": 7.970861234445142e-05, + "loss": 0.5962, + "step": 339 + }, + { + "epoch": 0.6759443339960238, + "grad_norm": 1.2462305949675103, + "learning_rate": 7.970188366095996e-05, + "loss": 0.5994, + "step": 340 + }, + { + "epoch": 0.6779324055666004, + "grad_norm": 0.8353002045769494, + "learning_rate": 7.969507846333221e-05, + "loss": 0.5925, + "step": 341 + }, + { + "epoch": 0.679920477137177, + "grad_norm": 0.6862279336154131, + "learning_rate": 7.968819676468327e-05, + "loss": 0.5942, + "step": 342 + }, + { + "epoch": 0.6819085487077535, + "grad_norm": 0.7767309327289988, + "learning_rate": 7.968123857827568e-05, + "loss": 0.5961, + "step": 343 + }, + { + "epoch": 0.68389662027833, + "grad_norm": 0.8656265135412583, + "learning_rate": 7.967420391751935e-05, + "loss": 0.6002, + "step": 344 + }, + { + "epoch": 0.6858846918489065, + "grad_norm": 0.8852817636204119, + "learning_rate": 7.96670927959716e-05, + "loss": 0.5973, + "step": 345 + }, + { + "epoch": 0.6878727634194831, + "grad_norm": 0.8201129197174135, + "learning_rate": 7.965990522733713e-05, + "loss": 0.5877, + "step": 346 + }, + { + "epoch": 0.6898608349900597, + "grad_norm": 0.7216882280944779, + "learning_rate": 7.96526412254679e-05, + "loss": 0.5933, + "step": 347 + }, + { + "epoch": 0.6918489065606361, + "grad_norm": 0.9822298632247115, + "learning_rate": 7.964530080436324e-05, + "loss": 0.5922, + "step": 348 + }, + { + "epoch": 0.6938369781312127, + "grad_norm": 1.2671518884810031, + "learning_rate": 7.963788397816974e-05, + "loss": 0.5913, + "step": 349 + }, + { + "epoch": 0.6958250497017893, + "grad_norm": 0.7838716714178371, + "learning_rate": 7.96303907611812e-05, + "loss": 0.591, + "step": 350 + }, + { + "epoch": 0.6978131212723658, + "grad_norm": 1.1891668128997623, + "learning_rate": 7.962282116783873e-05, + "loss": 0.5956, + "step": 351 + }, + { + "epoch": 0.6998011928429424, + "grad_norm": 1.1010146966411596, + "learning_rate": 7.96151752127305e-05, + "loss": 0.5896, + "step": 352 + }, + { + "epoch": 0.7017892644135189, + "grad_norm": 1.188668203426625, + "learning_rate": 7.9607452910592e-05, + "loss": 0.5934, + "step": 353 + }, + { + "epoch": 0.7037773359840954, + "grad_norm": 0.8731139591987406, + "learning_rate": 7.959965427630574e-05, + "loss": 0.5908, + "step": 354 + }, + { + "epoch": 0.705765407554672, + "grad_norm": 1.0694314172043198, + "learning_rate": 7.959177932490138e-05, + "loss": 0.5886, + "step": 355 + }, + { + "epoch": 0.7077534791252486, + "grad_norm": 1.2085859343148335, + "learning_rate": 7.958382807155566e-05, + "loss": 0.5894, + "step": 356 + }, + { + "epoch": 0.709741550695825, + "grad_norm": 0.8270753340299087, + "learning_rate": 7.957580053159237e-05, + "loss": 0.5891, + "step": 357 + }, + { + "epoch": 0.7117296222664016, + "grad_norm": 0.7137872070663243, + "learning_rate": 7.956769672048232e-05, + "loss": 0.5834, + "step": 358 + }, + { + "epoch": 0.7137176938369781, + "grad_norm": 0.8113240095237567, + "learning_rate": 7.955951665384333e-05, + "loss": 0.5882, + "step": 359 + }, + { + "epoch": 0.7157057654075547, + "grad_norm": 1.1469104720759737, + "learning_rate": 7.955126034744012e-05, + "loss": 0.5896, + "step": 360 + }, + { + "epoch": 0.7176938369781312, + "grad_norm": 1.180373367476227, + "learning_rate": 7.954292781718443e-05, + "loss": 0.5928, + "step": 361 + }, + { + "epoch": 0.7196819085487077, + "grad_norm": 0.9444939465878108, + "learning_rate": 7.953451907913482e-05, + "loss": 0.5918, + "step": 362 + }, + { + "epoch": 0.7216699801192843, + "grad_norm": 1.0006658840657565, + "learning_rate": 7.952603414949674e-05, + "loss": 0.5952, + "step": 363 + }, + { + "epoch": 0.7236580516898609, + "grad_norm": 1.2325093196009407, + "learning_rate": 7.951747304462255e-05, + "loss": 0.5951, + "step": 364 + }, + { + "epoch": 0.7256461232604374, + "grad_norm": 0.7749758827049822, + "learning_rate": 7.950883578101128e-05, + "loss": 0.589, + "step": 365 + }, + { + "epoch": 0.7276341948310139, + "grad_norm": 0.6979516318617547, + "learning_rate": 7.950012237530886e-05, + "loss": 0.5833, + "step": 366 + }, + { + "epoch": 0.7296222664015904, + "grad_norm": 0.7160911526302164, + "learning_rate": 7.949133284430788e-05, + "loss": 0.5896, + "step": 367 + }, + { + "epoch": 0.731610337972167, + "grad_norm": 0.9105506423506521, + "learning_rate": 7.948246720494769e-05, + "loss": 0.5923, + "step": 368 + }, + { + "epoch": 0.7335984095427436, + "grad_norm": 1.0461010915532079, + "learning_rate": 7.947352547431429e-05, + "loss": 0.5873, + "step": 369 + }, + { + "epoch": 0.73558648111332, + "grad_norm": 0.8558435340216544, + "learning_rate": 7.946450766964035e-05, + "loss": 0.5931, + "step": 370 + }, + { + "epoch": 0.7375745526838966, + "grad_norm": 0.7994654876766595, + "learning_rate": 7.945541380830512e-05, + "loss": 0.5923, + "step": 371 + }, + { + "epoch": 0.7395626242544732, + "grad_norm": 0.893131431334677, + "learning_rate": 7.944624390783444e-05, + "loss": 0.599, + "step": 372 + }, + { + "epoch": 0.7415506958250497, + "grad_norm": 1.2442016855861993, + "learning_rate": 7.94369979859007e-05, + "loss": 0.5952, + "step": 373 + }, + { + "epoch": 0.7435387673956262, + "grad_norm": 0.9079647599598836, + "learning_rate": 7.94276760603228e-05, + "loss": 0.5892, + "step": 374 + }, + { + "epoch": 0.7455268389662028, + "grad_norm": 0.9812110515231576, + "learning_rate": 7.941827814906611e-05, + "loss": 0.5898, + "step": 375 + }, + { + "epoch": 0.7475149105367793, + "grad_norm": 0.9969278983559571, + "learning_rate": 7.940880427024243e-05, + "loss": 0.587, + "step": 376 + }, + { + "epoch": 0.7495029821073559, + "grad_norm": 0.8529726891532479, + "learning_rate": 7.939925444211e-05, + "loss": 0.5843, + "step": 377 + }, + { + "epoch": 0.7514910536779325, + "grad_norm": 0.761433775409296, + "learning_rate": 7.93896286830734e-05, + "loss": 0.5914, + "step": 378 + }, + { + "epoch": 0.7534791252485089, + "grad_norm": 0.8206538735583792, + "learning_rate": 7.937992701168353e-05, + "loss": 0.5867, + "step": 379 + }, + { + "epoch": 0.7554671968190855, + "grad_norm": 0.9583546824707148, + "learning_rate": 7.937014944663766e-05, + "loss": 0.5882, + "step": 380 + }, + { + "epoch": 0.757455268389662, + "grad_norm": 1.2370646283419247, + "learning_rate": 7.936029600677924e-05, + "loss": 0.5888, + "step": 381 + }, + { + "epoch": 0.7594433399602386, + "grad_norm": 0.8519053228451124, + "learning_rate": 7.9350366711098e-05, + "loss": 0.5886, + "step": 382 + }, + { + "epoch": 0.7614314115308151, + "grad_norm": 0.7277634202579358, + "learning_rate": 7.934036157872983e-05, + "loss": 0.5889, + "step": 383 + }, + { + "epoch": 0.7634194831013916, + "grad_norm": 0.8707040598292405, + "learning_rate": 7.933028062895683e-05, + "loss": 0.5871, + "step": 384 + }, + { + "epoch": 0.7654075546719682, + "grad_norm": 1.0438936247793003, + "learning_rate": 7.932012388120714e-05, + "loss": 0.5905, + "step": 385 + }, + { + "epoch": 0.7673956262425448, + "grad_norm": 1.0450358135630395, + "learning_rate": 7.9309891355055e-05, + "loss": 0.5896, + "step": 386 + }, + { + "epoch": 0.7693836978131213, + "grad_norm": 1.2549615714436533, + "learning_rate": 7.929958307022074e-05, + "loss": 0.5844, + "step": 387 + }, + { + "epoch": 0.7713717693836978, + "grad_norm": 0.6622790149096027, + "learning_rate": 7.928919904657065e-05, + "loss": 0.5848, + "step": 388 + }, + { + "epoch": 0.7733598409542743, + "grad_norm": 0.6076325225296729, + "learning_rate": 7.9278739304117e-05, + "loss": 0.585, + "step": 389 + }, + { + "epoch": 0.7753479125248509, + "grad_norm": 0.8778111359677369, + "learning_rate": 7.926820386301795e-05, + "loss": 0.5915, + "step": 390 + }, + { + "epoch": 0.7773359840954275, + "grad_norm": 1.092006397904628, + "learning_rate": 7.925759274357763e-05, + "loss": 0.5897, + "step": 391 + }, + { + "epoch": 0.7793240556660039, + "grad_norm": 1.1071022336112284, + "learning_rate": 7.924690596624592e-05, + "loss": 0.5849, + "step": 392 + }, + { + "epoch": 0.7813121272365805, + "grad_norm": 0.8813657744060249, + "learning_rate": 7.923614355161859e-05, + "loss": 0.5844, + "step": 393 + }, + { + "epoch": 0.7833001988071571, + "grad_norm": 1.144804916630537, + "learning_rate": 7.922530552043712e-05, + "loss": 0.5888, + "step": 394 + }, + { + "epoch": 0.7852882703777336, + "grad_norm": 0.9647130509140422, + "learning_rate": 7.921439189358877e-05, + "loss": 0.5871, + "step": 395 + }, + { + "epoch": 0.7872763419483101, + "grad_norm": 0.8100748237359459, + "learning_rate": 7.920340269210643e-05, + "loss": 0.5895, + "step": 396 + }, + { + "epoch": 0.7892644135188867, + "grad_norm": 0.8424797920827156, + "learning_rate": 7.919233793716872e-05, + "loss": 0.5812, + "step": 397 + }, + { + "epoch": 0.7912524850894632, + "grad_norm": 0.981537746558217, + "learning_rate": 7.918119765009979e-05, + "loss": 0.5852, + "step": 398 + }, + { + "epoch": 0.7932405566600398, + "grad_norm": 1.1186079042253614, + "learning_rate": 7.916998185236939e-05, + "loss": 0.5844, + "step": 399 + }, + { + "epoch": 0.7952286282306164, + "grad_norm": 0.8679154615113002, + "learning_rate": 7.915869056559279e-05, + "loss": 0.5862, + "step": 400 + }, + { + "epoch": 0.7972166998011928, + "grad_norm": 0.7686569644536589, + "learning_rate": 7.914732381153079e-05, + "loss": 0.5849, + "step": 401 + }, + { + "epoch": 0.7992047713717694, + "grad_norm": 0.796490817813745, + "learning_rate": 7.913588161208952e-05, + "loss": 0.5858, + "step": 402 + }, + { + "epoch": 0.8011928429423459, + "grad_norm": 0.8035281403452298, + "learning_rate": 7.912436398932063e-05, + "loss": 0.5847, + "step": 403 + }, + { + "epoch": 0.8031809145129225, + "grad_norm": 1.0308204022548164, + "learning_rate": 7.911277096542106e-05, + "loss": 0.583, + "step": 404 + }, + { + "epoch": 0.805168986083499, + "grad_norm": 1.3305335250593018, + "learning_rate": 7.910110256273307e-05, + "loss": 0.5864, + "step": 405 + }, + { + "epoch": 0.8071570576540755, + "grad_norm": 0.6852998373300851, + "learning_rate": 7.908935880374423e-05, + "loss": 0.5804, + "step": 406 + }, + { + "epoch": 0.8091451292246521, + "grad_norm": 0.6572110197570675, + "learning_rate": 7.907753971108728e-05, + "loss": 0.58, + "step": 407 + }, + { + "epoch": 0.8111332007952287, + "grad_norm": 0.8879964008698213, + "learning_rate": 7.906564530754018e-05, + "loss": 0.5834, + "step": 408 + }, + { + "epoch": 0.8131212723658051, + "grad_norm": 1.0942411047383964, + "learning_rate": 7.905367561602603e-05, + "loss": 0.5824, + "step": 409 + }, + { + "epoch": 0.8151093439363817, + "grad_norm": 0.9734482127356657, + "learning_rate": 7.904163065961302e-05, + "loss": 0.5836, + "step": 410 + }, + { + "epoch": 0.8170974155069582, + "grad_norm": 0.9526293022396148, + "learning_rate": 7.90295104615144e-05, + "loss": 0.5873, + "step": 411 + }, + { + "epoch": 0.8190854870775348, + "grad_norm": 0.9111513732433741, + "learning_rate": 7.90173150450884e-05, + "loss": 0.5911, + "step": 412 + }, + { + "epoch": 0.8210735586481114, + "grad_norm": 1.000262045401118, + "learning_rate": 7.900504443383825e-05, + "loss": 0.5877, + "step": 413 + }, + { + "epoch": 0.8230616302186878, + "grad_norm": 1.0840607430740403, + "learning_rate": 7.899269865141206e-05, + "loss": 0.581, + "step": 414 + }, + { + "epoch": 0.8250497017892644, + "grad_norm": 1.0280039612546272, + "learning_rate": 7.898027772160284e-05, + "loss": 0.5774, + "step": 415 + }, + { + "epoch": 0.827037773359841, + "grad_norm": 1.052852112873632, + "learning_rate": 7.896778166834841e-05, + "loss": 0.5797, + "step": 416 + }, + { + "epoch": 0.8290258449304175, + "grad_norm": 0.9992150239594432, + "learning_rate": 7.895521051573139e-05, + "loss": 0.5839, + "step": 417 + }, + { + "epoch": 0.831013916500994, + "grad_norm": 0.8503482948129996, + "learning_rate": 7.894256428797909e-05, + "loss": 0.5784, + "step": 418 + }, + { + "epoch": 0.8330019880715706, + "grad_norm": 0.7307905930470503, + "learning_rate": 7.892984300946354e-05, + "loss": 0.5892, + "step": 419 + }, + { + "epoch": 0.8349900596421471, + "grad_norm": 0.7304703844232084, + "learning_rate": 7.891704670470142e-05, + "loss": 0.582, + "step": 420 + }, + { + "epoch": 0.8369781312127237, + "grad_norm": 0.8191311030518267, + "learning_rate": 7.890417539835395e-05, + "loss": 0.5793, + "step": 421 + }, + { + "epoch": 0.8389662027833003, + "grad_norm": 0.9761301233522803, + "learning_rate": 7.889122911522695e-05, + "loss": 0.5811, + "step": 422 + }, + { + "epoch": 0.8409542743538767, + "grad_norm": 1.1735275864088888, + "learning_rate": 7.887820788027072e-05, + "loss": 0.5835, + "step": 423 + }, + { + "epoch": 0.8429423459244533, + "grad_norm": 0.6856347281014243, + "learning_rate": 7.886511171858e-05, + "loss": 0.5856, + "step": 424 + }, + { + "epoch": 0.8449304174950298, + "grad_norm": 0.49809432528479436, + "learning_rate": 7.885194065539392e-05, + "loss": 0.5871, + "step": 425 + }, + { + "epoch": 0.8469184890656064, + "grad_norm": 0.6980729511674448, + "learning_rate": 7.883869471609598e-05, + "loss": 0.5854, + "step": 426 + }, + { + "epoch": 0.8489065606361829, + "grad_norm": 0.8088615313801084, + "learning_rate": 7.882537392621397e-05, + "loss": 0.5788, + "step": 427 + }, + { + "epoch": 0.8508946322067594, + "grad_norm": 0.6430145546564471, + "learning_rate": 7.881197831141994e-05, + "loss": 0.5834, + "step": 428 + }, + { + "epoch": 0.852882703777336, + "grad_norm": 0.5992622107260112, + "learning_rate": 7.879850789753016e-05, + "loss": 0.5819, + "step": 429 + }, + { + "epoch": 0.8548707753479126, + "grad_norm": 0.6800607847366943, + "learning_rate": 7.878496271050502e-05, + "loss": 0.5828, + "step": 430 + }, + { + "epoch": 0.856858846918489, + "grad_norm": 0.7499052373458801, + "learning_rate": 7.877134277644905e-05, + "loss": 0.5807, + "step": 431 + }, + { + "epoch": 0.8588469184890656, + "grad_norm": 0.9057195016780537, + "learning_rate": 7.875764812161079e-05, + "loss": 0.5694, + "step": 432 + }, + { + "epoch": 0.8608349900596421, + "grad_norm": 1.0520429016398172, + "learning_rate": 7.874387877238283e-05, + "loss": 0.5793, + "step": 433 + }, + { + "epoch": 0.8628230616302187, + "grad_norm": 1.0753300240675914, + "learning_rate": 7.873003475530169e-05, + "loss": 0.5835, + "step": 434 + }, + { + "epoch": 0.8648111332007953, + "grad_norm": 1.1933706500073453, + "learning_rate": 7.871611609704777e-05, + "loss": 0.5795, + "step": 435 + }, + { + "epoch": 0.8667992047713717, + "grad_norm": 0.9800602567078205, + "learning_rate": 7.870212282444536e-05, + "loss": 0.5777, + "step": 436 + }, + { + "epoch": 0.8687872763419483, + "grad_norm": 0.985646291001119, + "learning_rate": 7.868805496446254e-05, + "loss": 0.5825, + "step": 437 + }, + { + "epoch": 0.8707753479125249, + "grad_norm": 0.9173028079338591, + "learning_rate": 7.86739125442111e-05, + "loss": 0.5826, + "step": 438 + }, + { + "epoch": 0.8727634194831014, + "grad_norm": 0.8343252949496813, + "learning_rate": 7.865969559094657e-05, + "loss": 0.5765, + "step": 439 + }, + { + "epoch": 0.8747514910536779, + "grad_norm": 0.9011659813788296, + "learning_rate": 7.864540413206809e-05, + "loss": 0.5807, + "step": 440 + }, + { + "epoch": 0.8767395626242545, + "grad_norm": 1.1481938792997213, + "learning_rate": 7.863103819511841e-05, + "loss": 0.5756, + "step": 441 + }, + { + "epoch": 0.878727634194831, + "grad_norm": 0.8692225637264639, + "learning_rate": 7.861659780778379e-05, + "loss": 0.5797, + "step": 442 + }, + { + "epoch": 0.8807157057654076, + "grad_norm": 0.7792876237320553, + "learning_rate": 7.860208299789399e-05, + "loss": 0.5789, + "step": 443 + }, + { + "epoch": 0.882703777335984, + "grad_norm": 0.9806726470334519, + "learning_rate": 7.85874937934222e-05, + "loss": 0.572, + "step": 444 + }, + { + "epoch": 0.8846918489065606, + "grad_norm": 1.1128801316838084, + "learning_rate": 7.857283022248499e-05, + "loss": 0.5771, + "step": 445 + }, + { + "epoch": 0.8866799204771372, + "grad_norm": 0.7356169677900156, + "learning_rate": 7.855809231334222e-05, + "loss": 0.5706, + "step": 446 + }, + { + "epoch": 0.8886679920477137, + "grad_norm": 0.5716244438538673, + "learning_rate": 7.854328009439704e-05, + "loss": 0.5788, + "step": 447 + }, + { + "epoch": 0.8906560636182903, + "grad_norm": 0.7756825473319496, + "learning_rate": 7.852839359419583e-05, + "loss": 0.5726, + "step": 448 + }, + { + "epoch": 0.8926441351888668, + "grad_norm": 0.9597859189328001, + "learning_rate": 7.851343284142806e-05, + "loss": 0.5793, + "step": 449 + }, + { + "epoch": 0.8946322067594433, + "grad_norm": 0.9840514136250063, + "learning_rate": 7.849839786492638e-05, + "loss": 0.5816, + "step": 450 + }, + { + "epoch": 0.8966202783300199, + "grad_norm": 1.2387344205054815, + "learning_rate": 7.848328869366644e-05, + "loss": 0.5777, + "step": 451 + }, + { + "epoch": 0.8986083499005965, + "grad_norm": 0.9408504441388281, + "learning_rate": 7.846810535676687e-05, + "loss": 0.5686, + "step": 452 + }, + { + "epoch": 0.9005964214711729, + "grad_norm": 0.8661605445202164, + "learning_rate": 7.845284788348928e-05, + "loss": 0.5756, + "step": 453 + }, + { + "epoch": 0.9025844930417495, + "grad_norm": 0.8836259932861911, + "learning_rate": 7.843751630323811e-05, + "loss": 0.5754, + "step": 454 + }, + { + "epoch": 0.904572564612326, + "grad_norm": 0.7484011259578548, + "learning_rate": 7.842211064556065e-05, + "loss": 0.5747, + "step": 455 + }, + { + "epoch": 0.9065606361829026, + "grad_norm": 0.5916896881610786, + "learning_rate": 7.840663094014692e-05, + "loss": 0.5753, + "step": 456 + }, + { + "epoch": 0.9085487077534792, + "grad_norm": 0.6446171878161961, + "learning_rate": 7.839107721682971e-05, + "loss": 0.5758, + "step": 457 + }, + { + "epoch": 0.9105367793240556, + "grad_norm": 0.7331317560231023, + "learning_rate": 7.837544950558437e-05, + "loss": 0.5767, + "step": 458 + }, + { + "epoch": 0.9125248508946322, + "grad_norm": 0.7169698033515578, + "learning_rate": 7.835974783652892e-05, + "loss": 0.5774, + "step": 459 + }, + { + "epoch": 0.9145129224652088, + "grad_norm": 0.6975707545239594, + "learning_rate": 7.834397223992387e-05, + "loss": 0.5742, + "step": 460 + }, + { + "epoch": 0.9165009940357853, + "grad_norm": 0.7873223472670946, + "learning_rate": 7.83281227461722e-05, + "loss": 0.5752, + "step": 461 + }, + { + "epoch": 0.9184890656063618, + "grad_norm": 0.8677658380277969, + "learning_rate": 7.831219938581934e-05, + "loss": 0.5799, + "step": 462 + }, + { + "epoch": 0.9204771371769384, + "grad_norm": 0.8564280021238664, + "learning_rate": 7.829620218955303e-05, + "loss": 0.5764, + "step": 463 + }, + { + "epoch": 0.9224652087475149, + "grad_norm": 0.8770540530088526, + "learning_rate": 7.828013118820338e-05, + "loss": 0.5825, + "step": 464 + }, + { + "epoch": 0.9244532803180915, + "grad_norm": 0.8775401382866653, + "learning_rate": 7.826398641274266e-05, + "loss": 0.5743, + "step": 465 + }, + { + "epoch": 0.9264413518886679, + "grad_norm": 0.8396168474705734, + "learning_rate": 7.824776789428536e-05, + "loss": 0.5798, + "step": 466 + }, + { + "epoch": 0.9284294234592445, + "grad_norm": 0.6451882476229347, + "learning_rate": 7.823147566408808e-05, + "loss": 0.5739, + "step": 467 + }, + { + "epoch": 0.9304174950298211, + "grad_norm": 0.5473862955414479, + "learning_rate": 7.821510975354948e-05, + "loss": 0.572, + "step": 468 + }, + { + "epoch": 0.9324055666003976, + "grad_norm": 0.7334994329755088, + "learning_rate": 7.819867019421023e-05, + "loss": 0.5782, + "step": 469 + }, + { + "epoch": 0.9343936381709742, + "grad_norm": 0.8493149045492733, + "learning_rate": 7.818215701775289e-05, + "loss": 0.5691, + "step": 470 + }, + { + "epoch": 0.9363817097415507, + "grad_norm": 0.8916093585824147, + "learning_rate": 7.816557025600196e-05, + "loss": 0.5737, + "step": 471 + }, + { + "epoch": 0.9383697813121272, + "grad_norm": 1.0524874978245338, + "learning_rate": 7.814890994092371e-05, + "loss": 0.5783, + "step": 472 + }, + { + "epoch": 0.9403578528827038, + "grad_norm": 1.1448842494252496, + "learning_rate": 7.813217610462622e-05, + "loss": 0.5814, + "step": 473 + }, + { + "epoch": 0.9423459244532804, + "grad_norm": 0.8875324561425191, + "learning_rate": 7.811536877935916e-05, + "loss": 0.5692, + "step": 474 + }, + { + "epoch": 0.9443339960238568, + "grad_norm": 0.7862741860973999, + "learning_rate": 7.809848799751392e-05, + "loss": 0.579, + "step": 475 + }, + { + "epoch": 0.9463220675944334, + "grad_norm": 0.7404744964609349, + "learning_rate": 7.808153379162344e-05, + "loss": 0.5724, + "step": 476 + }, + { + "epoch": 0.94831013916501, + "grad_norm": 0.6210045203434688, + "learning_rate": 7.806450619436212e-05, + "loss": 0.5778, + "step": 477 + }, + { + "epoch": 0.9502982107355865, + "grad_norm": 0.5869709000454426, + "learning_rate": 7.804740523854584e-05, + "loss": 0.5707, + "step": 478 + }, + { + "epoch": 0.952286282306163, + "grad_norm": 0.690983672156381, + "learning_rate": 7.803023095713185e-05, + "loss": 0.5673, + "step": 479 + }, + { + "epoch": 0.9542743538767395, + "grad_norm": 0.7769763991531466, + "learning_rate": 7.801298338321873e-05, + "loss": 0.5716, + "step": 480 + }, + { + "epoch": 0.9562624254473161, + "grad_norm": 0.7773730859463129, + "learning_rate": 7.799566255004627e-05, + "loss": 0.5769, + "step": 481 + }, + { + "epoch": 0.9582504970178927, + "grad_norm": 0.6345198639510748, + "learning_rate": 7.797826849099549e-05, + "loss": 0.5707, + "step": 482 + }, + { + "epoch": 0.9602385685884692, + "grad_norm": 0.6588045885859571, + "learning_rate": 7.796080123958849e-05, + "loss": 0.5717, + "step": 483 + }, + { + "epoch": 0.9622266401590457, + "grad_norm": 0.8039607210702434, + "learning_rate": 7.794326082948847e-05, + "loss": 0.5779, + "step": 484 + }, + { + "epoch": 0.9642147117296223, + "grad_norm": 0.7648474063784596, + "learning_rate": 7.79256472944996e-05, + "loss": 0.5747, + "step": 485 + }, + { + "epoch": 0.9662027833001988, + "grad_norm": 0.9891199051114163, + "learning_rate": 7.790796066856695e-05, + "loss": 0.5761, + "step": 486 + }, + { + "epoch": 0.9681908548707754, + "grad_norm": 1.4830791374433934, + "learning_rate": 7.789020098577654e-05, + "loss": 0.578, + "step": 487 + }, + { + "epoch": 0.9701789264413518, + "grad_norm": 0.4001636007612993, + "learning_rate": 7.787236828035508e-05, + "loss": 0.565, + "step": 488 + }, + { + "epoch": 0.9721669980119284, + "grad_norm": 1.391894017796059, + "learning_rate": 7.785446258667006e-05, + "loss": 0.5728, + "step": 489 + }, + { + "epoch": 0.974155069582505, + "grad_norm": 0.8474494313233529, + "learning_rate": 7.783648393922968e-05, + "loss": 0.5755, + "step": 490 + }, + { + "epoch": 0.9761431411530815, + "grad_norm": 0.9368946226000615, + "learning_rate": 7.781843237268265e-05, + "loss": 0.5772, + "step": 491 + }, + { + "epoch": 0.9781312127236581, + "grad_norm": 1.0296054222965858, + "learning_rate": 7.780030792181826e-05, + "loss": 0.5754, + "step": 492 + }, + { + "epoch": 0.9801192842942346, + "grad_norm": 0.8078527279993658, + "learning_rate": 7.778211062156625e-05, + "loss": 0.5734, + "step": 493 + }, + { + "epoch": 0.9821073558648111, + "grad_norm": 0.8032865048136967, + "learning_rate": 7.776384050699677e-05, + "loss": 0.5722, + "step": 494 + }, + { + "epoch": 0.9840954274353877, + "grad_norm": 0.7330125115532979, + "learning_rate": 7.77454976133203e-05, + "loss": 0.5727, + "step": 495 + }, + { + "epoch": 0.9860834990059643, + "grad_norm": 0.5748768568400435, + "learning_rate": 7.772708197588755e-05, + "loss": 0.5814, + "step": 496 + }, + { + "epoch": 0.9880715705765407, + "grad_norm": 0.5663216889025084, + "learning_rate": 7.770859363018947e-05, + "loss": 0.575, + "step": 497 + }, + { + "epoch": 0.9900596421471173, + "grad_norm": 0.5130324016109733, + "learning_rate": 7.76900326118571e-05, + "loss": 0.5724, + "step": 498 + }, + { + "epoch": 0.9920477137176938, + "grad_norm": 0.6118536140979886, + "learning_rate": 7.767139895666154e-05, + "loss": 0.5741, + "step": 499 + }, + { + "epoch": 0.9940357852882704, + "grad_norm": 0.7903621371505978, + "learning_rate": 7.765269270051389e-05, + "loss": 0.5699, + "step": 500 + }, + { + "epoch": 0.9960238568588469, + "grad_norm": 0.9431939811737229, + "learning_rate": 7.763391387946515e-05, + "loss": 0.5752, + "step": 501 + }, + { + "epoch": 0.9980119284294234, + "grad_norm": 1.060145724961572, + "learning_rate": 7.76150625297062e-05, + "loss": 0.5766, + "step": 502 + }, + { + "epoch": 1.0, + "grad_norm": 0.949801921673401, + "learning_rate": 7.759613868756766e-05, + "loss": 0.5741, + "step": 503 + }, + { + "epoch": 1.0019880715705765, + "grad_norm": 0.9518658745713702, + "learning_rate": 7.757714238951987e-05, + "loss": 0.5507, + "step": 504 + }, + { + "epoch": 1.0039761431411531, + "grad_norm": 0.9224616680549119, + "learning_rate": 7.755807367217283e-05, + "loss": 0.5472, + "step": 505 + }, + { + "epoch": 1.0059642147117296, + "grad_norm": 0.9467697886335441, + "learning_rate": 7.75389325722761e-05, + "loss": 0.5397, + "step": 506 + }, + { + "epoch": 1.0079522862823063, + "grad_norm": 1.0032787528038696, + "learning_rate": 7.751971912671872e-05, + "loss": 0.5485, + "step": 507 + }, + { + "epoch": 1.0099403578528827, + "grad_norm": 1.0617180843326368, + "learning_rate": 7.750043337252915e-05, + "loss": 0.5486, + "step": 508 + }, + { + "epoch": 1.0119284294234592, + "grad_norm": 0.8345796460493181, + "learning_rate": 7.748107534687522e-05, + "loss": 0.5412, + "step": 509 + }, + { + "epoch": 1.0139165009940359, + "grad_norm": 0.5834246870142797, + "learning_rate": 7.746164508706406e-05, + "loss": 0.5452, + "step": 510 + }, + { + "epoch": 1.0159045725646123, + "grad_norm": 0.6308771152369433, + "learning_rate": 7.744214263054197e-05, + "loss": 0.5402, + "step": 511 + }, + { + "epoch": 1.0178926441351888, + "grad_norm": 0.8463052045590627, + "learning_rate": 7.742256801489446e-05, + "loss": 0.5479, + "step": 512 + }, + { + "epoch": 1.0198807157057654, + "grad_norm": 0.7841972760187628, + "learning_rate": 7.740292127784599e-05, + "loss": 0.5373, + "step": 513 + }, + { + "epoch": 1.021868787276342, + "grad_norm": 0.7132205161493853, + "learning_rate": 7.738320245726013e-05, + "loss": 0.5393, + "step": 514 + }, + { + "epoch": 1.0238568588469186, + "grad_norm": 0.7241903378381845, + "learning_rate": 7.736341159113932e-05, + "loss": 0.5468, + "step": 515 + }, + { + "epoch": 1.025844930417495, + "grad_norm": 0.6745690903906133, + "learning_rate": 7.734354871762481e-05, + "loss": 0.5443, + "step": 516 + }, + { + "epoch": 1.0278330019880715, + "grad_norm": 0.6972843853390581, + "learning_rate": 7.73236138749967e-05, + "loss": 0.544, + "step": 517 + }, + { + "epoch": 1.0298210735586482, + "grad_norm": 0.7989595482697651, + "learning_rate": 7.730360710167373e-05, + "loss": 0.5392, + "step": 518 + }, + { + "epoch": 1.0318091451292246, + "grad_norm": 0.9023867022544568, + "learning_rate": 7.728352843621332e-05, + "loss": 0.5423, + "step": 519 + }, + { + "epoch": 1.0337972166998013, + "grad_norm": 0.924903270385559, + "learning_rate": 7.726337791731139e-05, + "loss": 0.5406, + "step": 520 + }, + { + "epoch": 1.0357852882703777, + "grad_norm": 0.7157014642375783, + "learning_rate": 7.724315558380234e-05, + "loss": 0.5505, + "step": 521 + }, + { + "epoch": 1.0377733598409542, + "grad_norm": 0.5321338664276257, + "learning_rate": 7.722286147465902e-05, + "loss": 0.5407, + "step": 522 + }, + { + "epoch": 1.0397614314115309, + "grad_norm": 0.62645163866993, + "learning_rate": 7.720249562899257e-05, + "loss": 0.5454, + "step": 523 + }, + { + "epoch": 1.0417495029821073, + "grad_norm": 0.7501732694460348, + "learning_rate": 7.71820580860524e-05, + "loss": 0.5454, + "step": 524 + }, + { + "epoch": 1.0437375745526838, + "grad_norm": 0.9392351876663441, + "learning_rate": 7.716154888522606e-05, + "loss": 0.548, + "step": 525 + }, + { + "epoch": 1.0457256461232605, + "grad_norm": 1.1654586518995889, + "learning_rate": 7.714096806603925e-05, + "loss": 0.5436, + "step": 526 + }, + { + "epoch": 1.047713717693837, + "grad_norm": 0.7430639246487083, + "learning_rate": 7.712031566815564e-05, + "loss": 0.5435, + "step": 527 + }, + { + "epoch": 1.0497017892644136, + "grad_norm": 0.507332577185216, + "learning_rate": 7.709959173137691e-05, + "loss": 0.5459, + "step": 528 + }, + { + "epoch": 1.05168986083499, + "grad_norm": 0.47161280639381553, + "learning_rate": 7.707879629564258e-05, + "loss": 0.537, + "step": 529 + }, + { + "epoch": 1.0536779324055665, + "grad_norm": 0.49032557450749087, + "learning_rate": 7.705792940102992e-05, + "loss": 0.5354, + "step": 530 + }, + { + "epoch": 1.0556660039761432, + "grad_norm": 0.5367736352839478, + "learning_rate": 7.7036991087754e-05, + "loss": 0.5424, + "step": 531 + }, + { + "epoch": 1.0576540755467196, + "grad_norm": 0.5197894411370604, + "learning_rate": 7.701598139616748e-05, + "loss": 0.5452, + "step": 532 + }, + { + "epoch": 1.0596421471172963, + "grad_norm": 0.4882004191720187, + "learning_rate": 7.699490036676057e-05, + "loss": 0.5438, + "step": 533 + }, + { + "epoch": 1.0616302186878728, + "grad_norm": 0.5138919640591607, + "learning_rate": 7.697374804016102e-05, + "loss": 0.5419, + "step": 534 + }, + { + "epoch": 1.0636182902584492, + "grad_norm": 0.5587132412348711, + "learning_rate": 7.695252445713392e-05, + "loss": 0.5436, + "step": 535 + }, + { + "epoch": 1.065606361829026, + "grad_norm": 0.6000844383193432, + "learning_rate": 7.693122965858173e-05, + "loss": 0.5398, + "step": 536 + }, + { + "epoch": 1.0675944333996024, + "grad_norm": 0.7480777517729358, + "learning_rate": 7.690986368554414e-05, + "loss": 0.5411, + "step": 537 + }, + { + "epoch": 1.069582504970179, + "grad_norm": 0.9742584925928898, + "learning_rate": 7.688842657919799e-05, + "loss": 0.5398, + "step": 538 + }, + { + "epoch": 1.0715705765407555, + "grad_norm": 1.1824603415781014, + "learning_rate": 7.686691838085726e-05, + "loss": 0.5472, + "step": 539 + }, + { + "epoch": 1.073558648111332, + "grad_norm": 0.7779661582778407, + "learning_rate": 7.68453391319729e-05, + "loss": 0.5398, + "step": 540 + }, + { + "epoch": 1.0755467196819086, + "grad_norm": 0.660572845944423, + "learning_rate": 7.682368887413279e-05, + "loss": 0.546, + "step": 541 + }, + { + "epoch": 1.077534791252485, + "grad_norm": 0.647498839725535, + "learning_rate": 7.68019676490617e-05, + "loss": 0.5399, + "step": 542 + }, + { + "epoch": 1.0795228628230615, + "grad_norm": 0.5616817097968095, + "learning_rate": 7.678017549862111e-05, + "loss": 0.5414, + "step": 543 + }, + { + "epoch": 1.0815109343936382, + "grad_norm": 0.5077349122774221, + "learning_rate": 7.675831246480923e-05, + "loss": 0.537, + "step": 544 + }, + { + "epoch": 1.0834990059642147, + "grad_norm": 0.6460547464386392, + "learning_rate": 7.673637858976086e-05, + "loss": 0.5374, + "step": 545 + }, + { + "epoch": 1.0854870775347913, + "grad_norm": 0.7566165062457662, + "learning_rate": 7.671437391574735e-05, + "loss": 0.5454, + "step": 546 + }, + { + "epoch": 1.0874751491053678, + "grad_norm": 0.8765229126878328, + "learning_rate": 7.669229848517645e-05, + "loss": 0.5362, + "step": 547 + }, + { + "epoch": 1.0894632206759443, + "grad_norm": 0.9759995942905503, + "learning_rate": 7.667015234059231e-05, + "loss": 0.5489, + "step": 548 + }, + { + "epoch": 1.091451292246521, + "grad_norm": 1.0219449344079592, + "learning_rate": 7.664793552467537e-05, + "loss": 0.5501, + "step": 549 + }, + { + "epoch": 1.0934393638170974, + "grad_norm": 1.049500455165683, + "learning_rate": 7.662564808024221e-05, + "loss": 0.541, + "step": 550 + }, + { + "epoch": 1.0954274353876738, + "grad_norm": 1.040451711661253, + "learning_rate": 7.660329005024562e-05, + "loss": 0.5466, + "step": 551 + }, + { + "epoch": 1.0974155069582505, + "grad_norm": 0.9805330653702803, + "learning_rate": 7.658086147777433e-05, + "loss": 0.5447, + "step": 552 + }, + { + "epoch": 1.099403578528827, + "grad_norm": 0.7595556727657715, + "learning_rate": 7.655836240605306e-05, + "loss": 0.5425, + "step": 553 + }, + { + "epoch": 1.1013916500994037, + "grad_norm": 0.5629886257792873, + "learning_rate": 7.653579287844242e-05, + "loss": 0.5404, + "step": 554 + }, + { + "epoch": 1.10337972166998, + "grad_norm": 0.6066304118171499, + "learning_rate": 7.651315293843879e-05, + "loss": 0.5383, + "step": 555 + }, + { + "epoch": 1.1053677932405566, + "grad_norm": 0.7504168630502673, + "learning_rate": 7.649044262967421e-05, + "loss": 0.5401, + "step": 556 + }, + { + "epoch": 1.1073558648111332, + "grad_norm": 0.8672288992174086, + "learning_rate": 7.646766199591639e-05, + "loss": 0.5408, + "step": 557 + }, + { + "epoch": 1.1093439363817097, + "grad_norm": 0.7124336695406922, + "learning_rate": 7.644481108106853e-05, + "loss": 0.5379, + "step": 558 + }, + { + "epoch": 1.1113320079522864, + "grad_norm": 0.4827659980230668, + "learning_rate": 7.642188992916933e-05, + "loss": 0.5406, + "step": 559 + }, + { + "epoch": 1.1133200795228628, + "grad_norm": 0.4284195546204533, + "learning_rate": 7.639889858439279e-05, + "loss": 0.5408, + "step": 560 + }, + { + "epoch": 1.1153081510934393, + "grad_norm": 0.5213339666965665, + "learning_rate": 7.637583709104822e-05, + "loss": 0.5404, + "step": 561 + }, + { + "epoch": 1.117296222664016, + "grad_norm": 0.4100543308356987, + "learning_rate": 7.635270549358012e-05, + "loss": 0.5398, + "step": 562 + }, + { + "epoch": 1.1192842942345924, + "grad_norm": 0.4894280047781026, + "learning_rate": 7.63295038365681e-05, + "loss": 0.5339, + "step": 563 + }, + { + "epoch": 1.121272365805169, + "grad_norm": 0.6660381872451082, + "learning_rate": 7.630623216472677e-05, + "loss": 0.5382, + "step": 564 + }, + { + "epoch": 1.1232604373757455, + "grad_norm": 0.7384209893024911, + "learning_rate": 7.62828905229057e-05, + "loss": 0.5462, + "step": 565 + }, + { + "epoch": 1.125248508946322, + "grad_norm": 0.8439530140670745, + "learning_rate": 7.62594789560893e-05, + "loss": 0.54, + "step": 566 + }, + { + "epoch": 1.1272365805168987, + "grad_norm": 0.9437603370716825, + "learning_rate": 7.623599750939669e-05, + "loss": 0.5428, + "step": 567 + }, + { + "epoch": 1.1292246520874751, + "grad_norm": 0.9838775457498312, + "learning_rate": 7.621244622808176e-05, + "loss": 0.5417, + "step": 568 + }, + { + "epoch": 1.1312127236580518, + "grad_norm": 0.9050007349320468, + "learning_rate": 7.61888251575329e-05, + "loss": 0.5331, + "step": 569 + }, + { + "epoch": 1.1332007952286283, + "grad_norm": 0.6124525965371876, + "learning_rate": 7.616513434327306e-05, + "loss": 0.5443, + "step": 570 + }, + { + "epoch": 1.1351888667992047, + "grad_norm": 0.4152960954671294, + "learning_rate": 7.614137383095956e-05, + "loss": 0.5399, + "step": 571 + }, + { + "epoch": 1.1371769383697814, + "grad_norm": 0.5708141529701758, + "learning_rate": 7.611754366638405e-05, + "loss": 0.5429, + "step": 572 + }, + { + "epoch": 1.1391650099403579, + "grad_norm": 0.5984148199496552, + "learning_rate": 7.609364389547244e-05, + "loss": 0.5391, + "step": 573 + }, + { + "epoch": 1.1411530815109343, + "grad_norm": 0.4907009934797348, + "learning_rate": 7.606967456428477e-05, + "loss": 0.5387, + "step": 574 + }, + { + "epoch": 1.143141153081511, + "grad_norm": 0.4557259695946618, + "learning_rate": 7.60456357190151e-05, + "loss": 0.5362, + "step": 575 + }, + { + "epoch": 1.1451292246520874, + "grad_norm": 0.6032487760061391, + "learning_rate": 7.602152740599156e-05, + "loss": 0.5386, + "step": 576 + }, + { + "epoch": 1.147117296222664, + "grad_norm": 0.9574111490642987, + "learning_rate": 7.599734967167604e-05, + "loss": 0.5415, + "step": 577 + }, + { + "epoch": 1.1491053677932406, + "grad_norm": 1.2259900507384356, + "learning_rate": 7.59731025626643e-05, + "loss": 0.5439, + "step": 578 + }, + { + "epoch": 1.151093439363817, + "grad_norm": 0.5769700449630275, + "learning_rate": 7.594878612568577e-05, + "loss": 0.5366, + "step": 579 + }, + { + "epoch": 1.1530815109343937, + "grad_norm": 0.5807398397185227, + "learning_rate": 7.59244004076035e-05, + "loss": 0.5361, + "step": 580 + }, + { + "epoch": 1.1550695825049702, + "grad_norm": 0.9087659346410144, + "learning_rate": 7.589994545541406e-05, + "loss": 0.5465, + "step": 581 + }, + { + "epoch": 1.1570576540755466, + "grad_norm": 1.0866888064181934, + "learning_rate": 7.587542131624741e-05, + "loss": 0.5393, + "step": 582 + }, + { + "epoch": 1.1590457256461233, + "grad_norm": 0.8604614363303615, + "learning_rate": 7.585082803736692e-05, + "loss": 0.5394, + "step": 583 + }, + { + "epoch": 1.1610337972166997, + "grad_norm": 0.767814178995087, + "learning_rate": 7.582616566616917e-05, + "loss": 0.5358, + "step": 584 + }, + { + "epoch": 1.1630218687872764, + "grad_norm": 0.7506490371579211, + "learning_rate": 7.580143425018387e-05, + "loss": 0.5445, + "step": 585 + }, + { + "epoch": 1.1650099403578529, + "grad_norm": 0.743704401144008, + "learning_rate": 7.577663383707384e-05, + "loss": 0.5381, + "step": 586 + }, + { + "epoch": 1.1669980119284293, + "grad_norm": 0.586644447631041, + "learning_rate": 7.575176447463483e-05, + "loss": 0.5378, + "step": 587 + }, + { + "epoch": 1.168986083499006, + "grad_norm": 0.5352393673773019, + "learning_rate": 7.572682621079551e-05, + "loss": 0.5368, + "step": 588 + }, + { + "epoch": 1.1709741550695825, + "grad_norm": 0.5344762767308605, + "learning_rate": 7.570181909361733e-05, + "loss": 0.5394, + "step": 589 + }, + { + "epoch": 1.1729622266401591, + "grad_norm": 0.5395614621565161, + "learning_rate": 7.567674317129441e-05, + "loss": 0.5387, + "step": 590 + }, + { + "epoch": 1.1749502982107356, + "grad_norm": 0.575796280722566, + "learning_rate": 7.56515984921535e-05, + "loss": 0.5354, + "step": 591 + }, + { + "epoch": 1.176938369781312, + "grad_norm": 0.5915049989827494, + "learning_rate": 7.562638510465382e-05, + "loss": 0.5425, + "step": 592 + }, + { + "epoch": 1.1789264413518887, + "grad_norm": 0.5412929415802115, + "learning_rate": 7.560110305738708e-05, + "loss": 0.5414, + "step": 593 + }, + { + "epoch": 1.1809145129224652, + "grad_norm": 0.48254434256782, + "learning_rate": 7.557575239907721e-05, + "loss": 0.5372, + "step": 594 + }, + { + "epoch": 1.1829025844930419, + "grad_norm": 0.5181528971463945, + "learning_rate": 7.555033317858048e-05, + "loss": 0.5397, + "step": 595 + }, + { + "epoch": 1.1848906560636183, + "grad_norm": 0.5263184354853958, + "learning_rate": 7.552484544488519e-05, + "loss": 0.5391, + "step": 596 + }, + { + "epoch": 1.1868787276341948, + "grad_norm": 0.4115690316834255, + "learning_rate": 7.549928924711175e-05, + "loss": 0.5317, + "step": 597 + }, + { + "epoch": 1.1888667992047715, + "grad_norm": 0.43537299584180583, + "learning_rate": 7.547366463451249e-05, + "loss": 0.5344, + "step": 598 + }, + { + "epoch": 1.190854870775348, + "grad_norm": 0.4995055021742822, + "learning_rate": 7.544797165647162e-05, + "loss": 0.5344, + "step": 599 + }, + { + "epoch": 1.1928429423459244, + "grad_norm": 0.5090345933385654, + "learning_rate": 7.542221036250503e-05, + "loss": 0.5306, + "step": 600 + }, + { + "epoch": 1.194831013916501, + "grad_norm": 1.1457859586008614, + "learning_rate": 7.539638080226035e-05, + "loss": 0.5395, + "step": 601 + }, + { + "epoch": 1.1968190854870775, + "grad_norm": 0.7643479173221415, + "learning_rate": 7.537048302551672e-05, + "loss": 0.5452, + "step": 602 + }, + { + "epoch": 1.1988071570576542, + "grad_norm": 2.7209885317474662, + "learning_rate": 7.534451708218481e-05, + "loss": 0.5903, + "step": 603 + }, + { + "epoch": 1.2007952286282306, + "grad_norm": 15.140365769874222, + "learning_rate": 7.531848302230657e-05, + "loss": 0.9038, + "step": 604 + }, + { + "epoch": 1.202783300198807, + "grad_norm": 1.0923219501841632, + "learning_rate": 7.529238089605531e-05, + "loss": 0.6039, + "step": 605 + }, + { + "epoch": 1.2047713717693838, + "grad_norm": 1.6894455257731635, + "learning_rate": 7.526621075373549e-05, + "loss": 0.5814, + "step": 606 + }, + { + "epoch": 1.2067594433399602, + "grad_norm": 0.9739200902278735, + "learning_rate": 7.523997264578262e-05, + "loss": 0.5805, + "step": 607 + }, + { + "epoch": 1.2087475149105367, + "grad_norm": 1.4388276351964948, + "learning_rate": 7.521366662276324e-05, + "loss": 0.5814, + "step": 608 + }, + { + "epoch": 1.2107355864811133, + "grad_norm": 0.9122412624447593, + "learning_rate": 7.518729273537474e-05, + "loss": 0.5831, + "step": 609 + }, + { + "epoch": 1.2127236580516898, + "grad_norm": 1.11468588542839, + "learning_rate": 7.516085103444535e-05, + "loss": 0.5831, + "step": 610 + }, + { + "epoch": 1.2147117296222665, + "grad_norm": 0.7927291919940254, + "learning_rate": 7.513434157093395e-05, + "loss": 0.5744, + "step": 611 + }, + { + "epoch": 1.216699801192843, + "grad_norm": 0.9231028210170642, + "learning_rate": 7.510776439593e-05, + "loss": 0.5649, + "step": 612 + }, + { + "epoch": 1.2186878727634194, + "grad_norm": 0.7854830072173685, + "learning_rate": 7.508111956065351e-05, + "loss": 0.5706, + "step": 613 + }, + { + "epoch": 1.220675944333996, + "grad_norm": 0.7486027346054723, + "learning_rate": 7.505440711645482e-05, + "loss": 0.5633, + "step": 614 + }, + { + "epoch": 1.2226640159045725, + "grad_norm": 0.6883033522680998, + "learning_rate": 7.502762711481464e-05, + "loss": 0.5653, + "step": 615 + }, + { + "epoch": 1.2246520874751492, + "grad_norm": 0.7389094070760283, + "learning_rate": 7.50007796073438e-05, + "loss": 0.5611, + "step": 616 + }, + { + "epoch": 1.2266401590457257, + "grad_norm": 0.8318535060569524, + "learning_rate": 7.497386464578329e-05, + "loss": 0.5619, + "step": 617 + }, + { + "epoch": 1.228628230616302, + "grad_norm": 3.438501053226825, + "learning_rate": 7.494688228200407e-05, + "loss": 0.5655, + "step": 618 + }, + { + "epoch": 1.2306163021868788, + "grad_norm": 1.3177086303181726, + "learning_rate": 7.4919832568007e-05, + "loss": 0.5725, + "step": 619 + }, + { + "epoch": 1.2326043737574552, + "grad_norm": 1.275711618972336, + "learning_rate": 7.489271555592274e-05, + "loss": 0.5619, + "step": 620 + }, + { + "epoch": 1.234592445328032, + "grad_norm": 0.816816158692742, + "learning_rate": 7.486553129801167e-05, + "loss": 0.5532, + "step": 621 + }, + { + "epoch": 1.2365805168986084, + "grad_norm": 1.0789526717800226, + "learning_rate": 7.483827984666372e-05, + "loss": 0.5636, + "step": 622 + }, + { + "epoch": 1.2385685884691848, + "grad_norm": 0.931570699652363, + "learning_rate": 7.481096125439836e-05, + "loss": 0.5586, + "step": 623 + }, + { + "epoch": 1.2405566600397615, + "grad_norm": 1.0353236647558606, + "learning_rate": 7.478357557386444e-05, + "loss": 0.5576, + "step": 624 + }, + { + "epoch": 1.242544731610338, + "grad_norm": 0.7121241024787056, + "learning_rate": 7.47561228578401e-05, + "loss": 0.5527, + "step": 625 + }, + { + "epoch": 1.2445328031809146, + "grad_norm": 0.9354960795229111, + "learning_rate": 7.472860315923269e-05, + "loss": 0.5563, + "step": 626 + }, + { + "epoch": 1.246520874751491, + "grad_norm": 0.6810399481449622, + "learning_rate": 7.47010165310786e-05, + "loss": 0.5484, + "step": 627 + }, + { + "epoch": 1.2485089463220675, + "grad_norm": 0.564339776191756, + "learning_rate": 7.467336302654327e-05, + "loss": 0.5464, + "step": 628 + }, + { + "epoch": 1.250497017892644, + "grad_norm": 0.6329153920308194, + "learning_rate": 7.4645642698921e-05, + "loss": 0.5541, + "step": 629 + }, + { + "epoch": 1.2524850894632207, + "grad_norm": 0.5058821553568681, + "learning_rate": 7.461785560163485e-05, + "loss": 0.546, + "step": 630 + }, + { + "epoch": 1.2544731610337971, + "grad_norm": 0.4873097467195202, + "learning_rate": 7.459000178823661e-05, + "loss": 0.5478, + "step": 631 + }, + { + "epoch": 1.2564612326043738, + "grad_norm": 0.36686120495912883, + "learning_rate": 7.456208131240658e-05, + "loss": 0.5501, + "step": 632 + }, + { + "epoch": 1.2584493041749503, + "grad_norm": 0.4802564106494995, + "learning_rate": 7.45340942279536e-05, + "loss": 0.5447, + "step": 633 + }, + { + "epoch": 1.2604373757455267, + "grad_norm": 0.5250721782990612, + "learning_rate": 7.450604058881481e-05, + "loss": 0.5447, + "step": 634 + }, + { + "epoch": 1.2624254473161034, + "grad_norm": 0.4944252793040313, + "learning_rate": 7.44779204490557e-05, + "loss": 0.5517, + "step": 635 + }, + { + "epoch": 1.2644135188866799, + "grad_norm": 0.719034766972675, + "learning_rate": 7.444973386286985e-05, + "loss": 0.5453, + "step": 636 + }, + { + "epoch": 1.2664015904572565, + "grad_norm": 0.41783092273177835, + "learning_rate": 7.442148088457894e-05, + "loss": 0.5445, + "step": 637 + }, + { + "epoch": 1.268389662027833, + "grad_norm": 0.553215774804116, + "learning_rate": 7.439316156863255e-05, + "loss": 0.5416, + "step": 638 + }, + { + "epoch": 1.2703777335984094, + "grad_norm": 0.5408797993525253, + "learning_rate": 7.436477596960817e-05, + "loss": 0.5497, + "step": 639 + }, + { + "epoch": 1.2723658051689861, + "grad_norm": 0.4548242204966512, + "learning_rate": 7.4336324142211e-05, + "loss": 0.5432, + "step": 640 + }, + { + "epoch": 1.2743538767395626, + "grad_norm": 0.39363033210078713, + "learning_rate": 7.430780614127387e-05, + "loss": 0.5427, + "step": 641 + }, + { + "epoch": 1.2763419483101393, + "grad_norm": 0.589540690704029, + "learning_rate": 7.427922202175715e-05, + "loss": 0.5427, + "step": 642 + }, + { + "epoch": 1.2783300198807157, + "grad_norm": 0.7312646026221609, + "learning_rate": 7.425057183874865e-05, + "loss": 0.5459, + "step": 643 + }, + { + "epoch": 1.2803180914512922, + "grad_norm": 0.8432430495180779, + "learning_rate": 7.422185564746347e-05, + "loss": 0.5516, + "step": 644 + }, + { + "epoch": 1.2823061630218688, + "grad_norm": 0.7907499027073033, + "learning_rate": 7.419307350324392e-05, + "loss": 0.5477, + "step": 645 + }, + { + "epoch": 1.2842942345924453, + "grad_norm": 0.7675777704728362, + "learning_rate": 7.416422546155947e-05, + "loss": 0.5477, + "step": 646 + }, + { + "epoch": 1.286282306163022, + "grad_norm": 0.7818952066954505, + "learning_rate": 7.41353115780065e-05, + "loss": 0.5445, + "step": 647 + }, + { + "epoch": 1.2882703777335984, + "grad_norm": 0.6321072667200449, + "learning_rate": 7.410633190830839e-05, + "loss": 0.5475, + "step": 648 + }, + { + "epoch": 1.2902584493041749, + "grad_norm": 0.5418717449116404, + "learning_rate": 7.407728650831519e-05, + "loss": 0.5381, + "step": 649 + }, + { + "epoch": 1.2922465208747516, + "grad_norm": 0.47733226736370155, + "learning_rate": 7.404817543400372e-05, + "loss": 0.5358, + "step": 650 + }, + { + "epoch": 1.294234592445328, + "grad_norm": 0.4087774811685782, + "learning_rate": 7.401899874147733e-05, + "loss": 0.5423, + "step": 651 + }, + { + "epoch": 1.2962226640159047, + "grad_norm": 0.34492073630409487, + "learning_rate": 7.39897564869658e-05, + "loss": 0.5377, + "step": 652 + }, + { + "epoch": 1.2982107355864811, + "grad_norm": 0.41364092660903046, + "learning_rate": 7.396044872682533e-05, + "loss": 0.5409, + "step": 653 + }, + { + "epoch": 1.3001988071570576, + "grad_norm": 0.49300751266085974, + "learning_rate": 7.393107551753829e-05, + "loss": 0.5418, + "step": 654 + }, + { + "epoch": 1.302186878727634, + "grad_norm": 0.5115889514144143, + "learning_rate": 7.390163691571325e-05, + "loss": 0.541, + "step": 655 + }, + { + "epoch": 1.3041749502982107, + "grad_norm": 0.4520722977521369, + "learning_rate": 7.387213297808477e-05, + "loss": 0.5364, + "step": 656 + }, + { + "epoch": 1.3061630218687874, + "grad_norm": 0.45730722103763394, + "learning_rate": 7.384256376151333e-05, + "loss": 0.5433, + "step": 657 + }, + { + "epoch": 1.3081510934393639, + "grad_norm": 0.35991381976270714, + "learning_rate": 7.381292932298524e-05, + "loss": 0.5393, + "step": 658 + }, + { + "epoch": 1.3101391650099403, + "grad_norm": 0.3715007427958041, + "learning_rate": 7.378322971961245e-05, + "loss": 0.5386, + "step": 659 + }, + { + "epoch": 1.3121272365805168, + "grad_norm": 0.5035889463103559, + "learning_rate": 7.375346500863256e-05, + "loss": 0.5355, + "step": 660 + }, + { + "epoch": 1.3141153081510935, + "grad_norm": 0.6845352231571457, + "learning_rate": 7.372363524740862e-05, + "loss": 0.5434, + "step": 661 + }, + { + "epoch": 1.31610337972167, + "grad_norm": 0.7310717499538159, + "learning_rate": 7.369374049342904e-05, + "loss": 0.531, + "step": 662 + }, + { + "epoch": 1.3180914512922466, + "grad_norm": 0.5741347090908184, + "learning_rate": 7.366378080430748e-05, + "loss": 0.5424, + "step": 663 + }, + { + "epoch": 1.320079522862823, + "grad_norm": 0.4353235586521284, + "learning_rate": 7.36337562377828e-05, + "loss": 0.5359, + "step": 664 + }, + { + "epoch": 1.3220675944333995, + "grad_norm": 0.33576960536141043, + "learning_rate": 7.360366685171878e-05, + "loss": 0.535, + "step": 665 + }, + { + "epoch": 1.3240556660039762, + "grad_norm": 0.28581021576787774, + "learning_rate": 7.357351270410424e-05, + "loss": 0.5337, + "step": 666 + }, + { + "epoch": 1.3260437375745526, + "grad_norm": 0.34667350546819015, + "learning_rate": 7.354329385305274e-05, + "loss": 0.5344, + "step": 667 + }, + { + "epoch": 1.3280318091451293, + "grad_norm": 0.38791896920198515, + "learning_rate": 7.351301035680256e-05, + "loss": 0.5336, + "step": 668 + }, + { + "epoch": 1.3300198807157058, + "grad_norm": 0.3445729798343968, + "learning_rate": 7.348266227371657e-05, + "loss": 0.5304, + "step": 669 + }, + { + "epoch": 1.3320079522862822, + "grad_norm": 0.36399698860400687, + "learning_rate": 7.34522496622821e-05, + "loss": 0.5377, + "step": 670 + }, + { + "epoch": 1.333996023856859, + "grad_norm": 0.4179520009930122, + "learning_rate": 7.342177258111085e-05, + "loss": 0.5345, + "step": 671 + }, + { + "epoch": 1.3359840954274353, + "grad_norm": 0.5056757926940009, + "learning_rate": 7.339123108893873e-05, + "loss": 0.5315, + "step": 672 + }, + { + "epoch": 1.337972166998012, + "grad_norm": 0.6568326228633579, + "learning_rate": 7.336062524462586e-05, + "loss": 0.5336, + "step": 673 + }, + { + "epoch": 1.3399602385685885, + "grad_norm": 0.8051581651906589, + "learning_rate": 7.332995510715633e-05, + "loss": 0.5321, + "step": 674 + }, + { + "epoch": 1.341948310139165, + "grad_norm": 1.0061718045293344, + "learning_rate": 7.329922073563812e-05, + "loss": 0.5389, + "step": 675 + }, + { + "epoch": 1.3439363817097416, + "grad_norm": 1.224198955636859, + "learning_rate": 7.326842218930303e-05, + "loss": 0.5323, + "step": 676 + }, + { + "epoch": 1.345924453280318, + "grad_norm": 0.6806294285463886, + "learning_rate": 7.323755952750654e-05, + "loss": 0.5358, + "step": 677 + }, + { + "epoch": 1.3479125248508947, + "grad_norm": 0.5216238484730512, + "learning_rate": 7.32066328097277e-05, + "loss": 0.535, + "step": 678 + }, + { + "epoch": 1.3499005964214712, + "grad_norm": 0.6531151290656164, + "learning_rate": 7.317564209556902e-05, + "loss": 0.5379, + "step": 679 + }, + { + "epoch": 1.3518886679920477, + "grad_norm": 0.9240874377962419, + "learning_rate": 7.314458744475627e-05, + "loss": 0.5361, + "step": 680 + }, + { + "epoch": 1.3538767395626243, + "grad_norm": 1.0912021119009678, + "learning_rate": 7.311346891713852e-05, + "loss": 0.5324, + "step": 681 + }, + { + "epoch": 1.3558648111332008, + "grad_norm": 0.7376919034069003, + "learning_rate": 7.308228657268793e-05, + "loss": 0.5342, + "step": 682 + }, + { + "epoch": 1.3578528827037775, + "grad_norm": 0.5350891144993892, + "learning_rate": 7.305104047149961e-05, + "loss": 0.5268, + "step": 683 + }, + { + "epoch": 1.359840954274354, + "grad_norm": 0.5603454960089224, + "learning_rate": 7.30197306737916e-05, + "loss": 0.5347, + "step": 684 + }, + { + "epoch": 1.3618290258449304, + "grad_norm": 0.6375682349348152, + "learning_rate": 7.298835723990465e-05, + "loss": 0.5256, + "step": 685 + }, + { + "epoch": 1.3638170974155068, + "grad_norm": 0.5979905914047293, + "learning_rate": 7.295692023030217e-05, + "loss": 0.5351, + "step": 686 + }, + { + "epoch": 1.3658051689860835, + "grad_norm": 0.6240089079981558, + "learning_rate": 7.29254197055701e-05, + "loss": 0.5297, + "step": 687 + }, + { + "epoch": 1.36779324055666, + "grad_norm": 0.8050947571296337, + "learning_rate": 7.289385572641678e-05, + "loss": 0.5335, + "step": 688 + }, + { + "epoch": 1.3697813121272366, + "grad_norm": 0.8059108298679306, + "learning_rate": 7.286222835367283e-05, + "loss": 0.5354, + "step": 689 + }, + { + "epoch": 1.371769383697813, + "grad_norm": 0.6260685868632353, + "learning_rate": 7.283053764829106e-05, + "loss": 0.5382, + "step": 690 + }, + { + "epoch": 1.3737574552683895, + "grad_norm": 0.48277003276738417, + "learning_rate": 7.279878367134633e-05, + "loss": 0.5366, + "step": 691 + }, + { + "epoch": 1.3757455268389662, + "grad_norm": 0.6008163142885341, + "learning_rate": 7.276696648403542e-05, + "loss": 0.5331, + "step": 692 + }, + { + "epoch": 1.3777335984095427, + "grad_norm": 0.6809704933082544, + "learning_rate": 7.273508614767698e-05, + "loss": 0.5314, + "step": 693 + }, + { + "epoch": 1.3797216699801194, + "grad_norm": 0.5791983308971101, + "learning_rate": 7.270314272371129e-05, + "loss": 0.5384, + "step": 694 + }, + { + "epoch": 1.3817097415506958, + "grad_norm": 0.46173801711286394, + "learning_rate": 7.26711362737003e-05, + "loss": 0.5402, + "step": 695 + }, + { + "epoch": 1.3836978131212723, + "grad_norm": 0.3789583200923854, + "learning_rate": 7.263906685932734e-05, + "loss": 0.5308, + "step": 696 + }, + { + "epoch": 1.385685884691849, + "grad_norm": 0.35662981556630236, + "learning_rate": 7.260693454239712e-05, + "loss": 0.533, + "step": 697 + }, + { + "epoch": 1.3876739562624254, + "grad_norm": 0.4313635452570124, + "learning_rate": 7.257473938483561e-05, + "loss": 0.5361, + "step": 698 + }, + { + "epoch": 1.389662027833002, + "grad_norm": 0.42693701726569916, + "learning_rate": 7.254248144868985e-05, + "loss": 0.5269, + "step": 699 + }, + { + "epoch": 1.3916500994035785, + "grad_norm": 0.46448401272739637, + "learning_rate": 7.251016079612786e-05, + "loss": 0.5281, + "step": 700 + }, + { + "epoch": 1.393638170974155, + "grad_norm": 0.45165956462569934, + "learning_rate": 7.247777748943854e-05, + "loss": 0.5302, + "step": 701 + }, + { + "epoch": 1.3956262425447317, + "grad_norm": 0.5051001158050499, + "learning_rate": 7.244533159103157e-05, + "loss": 0.5319, + "step": 702 + }, + { + "epoch": 1.3976143141153081, + "grad_norm": 0.5235203957856401, + "learning_rate": 7.24128231634372e-05, + "loss": 0.5293, + "step": 703 + }, + { + "epoch": 1.3996023856858848, + "grad_norm": 0.4566107243597437, + "learning_rate": 7.238025226930624e-05, + "loss": 0.5358, + "step": 704 + }, + { + "epoch": 1.4015904572564613, + "grad_norm": 0.4625460874004843, + "learning_rate": 7.234761897140986e-05, + "loss": 0.5304, + "step": 705 + }, + { + "epoch": 1.4035785288270377, + "grad_norm": 0.41859360558238123, + "learning_rate": 7.231492333263949e-05, + "loss": 0.5307, + "step": 706 + }, + { + "epoch": 1.4055666003976144, + "grad_norm": 0.48973969900503483, + "learning_rate": 7.228216541600671e-05, + "loss": 0.5331, + "step": 707 + }, + { + "epoch": 1.4075546719681908, + "grad_norm": 0.7812920436986611, + "learning_rate": 7.224934528464314e-05, + "loss": 0.5372, + "step": 708 + }, + { + "epoch": 1.4095427435387675, + "grad_norm": 0.7104262912732389, + "learning_rate": 7.221646300180027e-05, + "loss": 0.5318, + "step": 709 + }, + { + "epoch": 1.411530815109344, + "grad_norm": 0.7596596767184741, + "learning_rate": 7.218351863084942e-05, + "loss": 0.5328, + "step": 710 + }, + { + "epoch": 1.4135188866799204, + "grad_norm": 0.7856583093617283, + "learning_rate": 7.21505122352815e-05, + "loss": 0.5361, + "step": 711 + }, + { + "epoch": 1.4155069582504969, + "grad_norm": 0.7573687326885941, + "learning_rate": 7.2117443878707e-05, + "loss": 0.5369, + "step": 712 + }, + { + "epoch": 1.4174950298210736, + "grad_norm": 0.7884477181941297, + "learning_rate": 7.20843136248558e-05, + "loss": 0.5377, + "step": 713 + }, + { + "epoch": 1.4194831013916502, + "grad_norm": 0.7951212883224608, + "learning_rate": 7.205112153757713e-05, + "loss": 0.5294, + "step": 714 + }, + { + "epoch": 1.4214711729622267, + "grad_norm": 0.7295618655254292, + "learning_rate": 7.201786768083929e-05, + "loss": 0.5387, + "step": 715 + }, + { + "epoch": 1.4234592445328031, + "grad_norm": 0.6204183333887281, + "learning_rate": 7.198455211872966e-05, + "loss": 0.5328, + "step": 716 + }, + { + "epoch": 1.4254473161033796, + "grad_norm": 0.5131018383315537, + "learning_rate": 7.19511749154546e-05, + "loss": 0.5324, + "step": 717 + }, + { + "epoch": 1.4274353876739563, + "grad_norm": 0.6107281423459198, + "learning_rate": 7.191773613533919e-05, + "loss": 0.5376, + "step": 718 + }, + { + "epoch": 1.4294234592445327, + "grad_norm": 0.9103425902361078, + "learning_rate": 7.188423584282719e-05, + "loss": 0.5445, + "step": 719 + }, + { + "epoch": 1.4314115308151094, + "grad_norm": 1.145266589470004, + "learning_rate": 7.185067410248096e-05, + "loss": 0.5295, + "step": 720 + }, + { + "epoch": 1.4333996023856859, + "grad_norm": 0.6522122472416725, + "learning_rate": 7.181705097898122e-05, + "loss": 0.5314, + "step": 721 + }, + { + "epoch": 1.4353876739562623, + "grad_norm": 0.4808333705940563, + "learning_rate": 7.178336653712704e-05, + "loss": 0.5283, + "step": 722 + }, + { + "epoch": 1.437375745526839, + "grad_norm": 1.4207277257992248, + "learning_rate": 7.174962084183563e-05, + "loss": 0.5458, + "step": 723 + }, + { + "epoch": 1.4393638170974155, + "grad_norm": 0.5015133463112434, + "learning_rate": 7.171581395814225e-05, + "loss": 0.5323, + "step": 724 + }, + { + "epoch": 1.4413518886679921, + "grad_norm": 0.7617435821622607, + "learning_rate": 7.168194595120011e-05, + "loss": 0.5367, + "step": 725 + }, + { + "epoch": 1.4433399602385686, + "grad_norm": 0.9226878333434035, + "learning_rate": 7.16480168862802e-05, + "loss": 0.5336, + "step": 726 + }, + { + "epoch": 1.445328031809145, + "grad_norm": 1.0981884524716032, + "learning_rate": 7.161402682877117e-05, + "loss": 0.5382, + "step": 727 + }, + { + "epoch": 1.4473161033797217, + "grad_norm": 0.7264677197882256, + "learning_rate": 7.157997584417927e-05, + "loss": 0.5349, + "step": 728 + }, + { + "epoch": 1.4493041749502982, + "grad_norm": 0.7884303052990086, + "learning_rate": 7.154586399812808e-05, + "loss": 0.5363, + "step": 729 + }, + { + "epoch": 1.4512922465208749, + "grad_norm": 0.7470918521247032, + "learning_rate": 7.151169135635853e-05, + "loss": 0.5374, + "step": 730 + }, + { + "epoch": 1.4532803180914513, + "grad_norm": 0.6053092802449975, + "learning_rate": 7.147745798472871e-05, + "loss": 0.5401, + "step": 731 + }, + { + "epoch": 1.4552683896620278, + "grad_norm": 0.7160809940534572, + "learning_rate": 7.144316394921378e-05, + "loss": 0.5428, + "step": 732 + }, + { + "epoch": 1.4572564612326044, + "grad_norm": 0.46588637989423676, + "learning_rate": 7.140880931590575e-05, + "loss": 0.5409, + "step": 733 + }, + { + "epoch": 1.459244532803181, + "grad_norm": 0.42391000958772374, + "learning_rate": 7.137439415101345e-05, + "loss": 0.535, + "step": 734 + }, + { + "epoch": 1.4612326043737576, + "grad_norm": 0.4462128319583528, + "learning_rate": 7.133991852086235e-05, + "loss": 0.539, + "step": 735 + }, + { + "epoch": 1.463220675944334, + "grad_norm": 0.3737682098835748, + "learning_rate": 7.13053824918945e-05, + "loss": 0.5374, + "step": 736 + }, + { + "epoch": 1.4652087475149105, + "grad_norm": 0.4379515087175158, + "learning_rate": 7.12707861306683e-05, + "loss": 0.5395, + "step": 737 + }, + { + "epoch": 1.4671968190854872, + "grad_norm": 0.49246889187865955, + "learning_rate": 7.12361295038584e-05, + "loss": 0.5338, + "step": 738 + }, + { + "epoch": 1.4691848906560636, + "grad_norm": 1.0681705732859723, + "learning_rate": 7.120141267825567e-05, + "loss": 0.5564, + "step": 739 + }, + { + "epoch": 1.4711729622266403, + "grad_norm": 0.6301239024315092, + "learning_rate": 7.116663572076697e-05, + "loss": 0.5423, + "step": 740 + }, + { + "epoch": 1.4731610337972167, + "grad_norm": 0.5869716649862752, + "learning_rate": 7.113179869841502e-05, + "loss": 0.5423, + "step": 741 + }, + { + "epoch": 1.4751491053677932, + "grad_norm": 0.6262438855252204, + "learning_rate": 7.109690167833829e-05, + "loss": 0.5517, + "step": 742 + }, + { + "epoch": 1.4771371769383697, + "grad_norm": 0.5891280338185073, + "learning_rate": 7.106194472779093e-05, + "loss": 0.5433, + "step": 743 + }, + { + "epoch": 1.4791252485089463, + "grad_norm": 0.6468565422433893, + "learning_rate": 7.102692791414257e-05, + "loss": 0.549, + "step": 744 + }, + { + "epoch": 1.4811133200795228, + "grad_norm": 0.5481646916057995, + "learning_rate": 7.099185130487818e-05, + "loss": 0.5348, + "step": 745 + }, + { + "epoch": 1.4831013916500995, + "grad_norm": 0.4199924161739955, + "learning_rate": 7.0956714967598e-05, + "loss": 0.54, + "step": 746 + }, + { + "epoch": 1.485089463220676, + "grad_norm": 0.48686262677138054, + "learning_rate": 7.092151897001734e-05, + "loss": 0.5411, + "step": 747 + }, + { + "epoch": 1.4870775347912524, + "grad_norm": 0.4745863525083435, + "learning_rate": 7.088626337996657e-05, + "loss": 0.5342, + "step": 748 + }, + { + "epoch": 1.489065606361829, + "grad_norm": 0.42679968740429375, + "learning_rate": 7.085094826539083e-05, + "loss": 0.5346, + "step": 749 + }, + { + "epoch": 1.4910536779324055, + "grad_norm": 0.4972102291339057, + "learning_rate": 7.081557369434999e-05, + "loss": 0.5368, + "step": 750 + }, + { + "epoch": 1.4930417495029822, + "grad_norm": 0.5632447683848434, + "learning_rate": 7.078013973501853e-05, + "loss": 0.5338, + "step": 751 + }, + { + "epoch": 1.4950298210735586, + "grad_norm": 0.6050601239636209, + "learning_rate": 7.074464645568535e-05, + "loss": 0.5313, + "step": 752 + }, + { + "epoch": 1.497017892644135, + "grad_norm": 0.7162448792755521, + "learning_rate": 7.070909392475372e-05, + "loss": 0.531, + "step": 753 + }, + { + "epoch": 1.4990059642147118, + "grad_norm": 0.6731491376009496, + "learning_rate": 7.067348221074106e-05, + "loss": 0.5261, + "step": 754 + }, + { + "epoch": 1.5009940357852882, + "grad_norm": 0.8022863170941446, + "learning_rate": 7.063781138227888e-05, + "loss": 0.5447, + "step": 755 + }, + { + "epoch": 1.502982107355865, + "grad_norm": 0.5914280370903948, + "learning_rate": 7.060208150811257e-05, + "loss": 0.5379, + "step": 756 + }, + { + "epoch": 1.5049701789264414, + "grad_norm": 0.41190704446363857, + "learning_rate": 7.056629265710137e-05, + "loss": 0.5396, + "step": 757 + }, + { + "epoch": 1.5069582504970178, + "grad_norm": 0.401420962458254, + "learning_rate": 7.053044489821812e-05, + "loss": 0.5419, + "step": 758 + }, + { + "epoch": 1.5089463220675943, + "grad_norm": 0.5144759429269341, + "learning_rate": 7.049453830054927e-05, + "loss": 0.5388, + "step": 759 + }, + { + "epoch": 1.510934393638171, + "grad_norm": 0.5723949652263456, + "learning_rate": 7.045857293329457e-05, + "loss": 0.542, + "step": 760 + }, + { + "epoch": 1.5129224652087476, + "grad_norm": 1.1438855059147868, + "learning_rate": 7.042254886576712e-05, + "loss": 0.5351, + "step": 761 + }, + { + "epoch": 1.514910536779324, + "grad_norm": 0.4874743210763008, + "learning_rate": 7.038646616739309e-05, + "loss": 0.5408, + "step": 762 + }, + { + "epoch": 1.5168986083499005, + "grad_norm": 0.4142150568688663, + "learning_rate": 7.035032490771165e-05, + "loss": 0.5379, + "step": 763 + }, + { + "epoch": 1.518886679920477, + "grad_norm": 0.36121114882100824, + "learning_rate": 7.031412515637485e-05, + "loss": 0.5419, + "step": 764 + }, + { + "epoch": 1.5208747514910537, + "grad_norm": 0.35876241898828937, + "learning_rate": 7.027786698314746e-05, + "loss": 0.5457, + "step": 765 + }, + { + "epoch": 1.5228628230616303, + "grad_norm": 0.3641469911154147, + "learning_rate": 7.024155045790684e-05, + "loss": 0.5404, + "step": 766 + }, + { + "epoch": 1.5248508946322068, + "grad_norm": 0.33445543544945516, + "learning_rate": 7.020517565064282e-05, + "loss": 0.5333, + "step": 767 + }, + { + "epoch": 1.5268389662027833, + "grad_norm": 0.34325800594144645, + "learning_rate": 7.016874263145752e-05, + "loss": 0.5341, + "step": 768 + }, + { + "epoch": 1.5288270377733597, + "grad_norm": 0.4686449206678608, + "learning_rate": 7.013225147056526e-05, + "loss": 0.5351, + "step": 769 + }, + { + "epoch": 1.5308151093439364, + "grad_norm": 0.41485737972871756, + "learning_rate": 7.009570223829244e-05, + "loss": 0.5318, + "step": 770 + }, + { + "epoch": 1.532803180914513, + "grad_norm": 0.5248903855009908, + "learning_rate": 7.005909500507735e-05, + "loss": 0.5368, + "step": 771 + }, + { + "epoch": 1.5347912524850895, + "grad_norm": 0.47323182987265955, + "learning_rate": 7.002242984147005e-05, + "loss": 0.537, + "step": 772 + }, + { + "epoch": 1.536779324055666, + "grad_norm": 0.541703280414164, + "learning_rate": 6.998570681813226e-05, + "loss": 0.536, + "step": 773 + }, + { + "epoch": 1.5387673956262424, + "grad_norm": 0.5737224485492203, + "learning_rate": 6.994892600583722e-05, + "loss": 0.5372, + "step": 774 + }, + { + "epoch": 1.540755467196819, + "grad_norm": 0.4745838189750734, + "learning_rate": 6.991208747546951e-05, + "loss": 0.5403, + "step": 775 + }, + { + "epoch": 1.5427435387673958, + "grad_norm": 0.4354551274230896, + "learning_rate": 6.987519129802499e-05, + "loss": 0.5395, + "step": 776 + }, + { + "epoch": 1.5447316103379722, + "grad_norm": 0.4475851156104487, + "learning_rate": 6.983823754461056e-05, + "loss": 0.5293, + "step": 777 + }, + { + "epoch": 1.5467196819085487, + "grad_norm": 0.35033211689343957, + "learning_rate": 6.980122628644414e-05, + "loss": 0.5367, + "step": 778 + }, + { + "epoch": 1.5487077534791251, + "grad_norm": 0.32233645639354963, + "learning_rate": 6.976415759485443e-05, + "loss": 0.5327, + "step": 779 + }, + { + "epoch": 1.5506958250497018, + "grad_norm": 0.29258727348920655, + "learning_rate": 6.972703154128084e-05, + "loss": 0.529, + "step": 780 + }, + { + "epoch": 1.5526838966202783, + "grad_norm": 0.2799251045164188, + "learning_rate": 6.968984819727334e-05, + "loss": 0.5307, + "step": 781 + }, + { + "epoch": 1.554671968190855, + "grad_norm": 0.25309699205970554, + "learning_rate": 6.965260763449227e-05, + "loss": 0.5282, + "step": 782 + }, + { + "epoch": 1.5566600397614314, + "grad_norm": 0.32992663486582197, + "learning_rate": 6.961530992470828e-05, + "loss": 0.5383, + "step": 783 + }, + { + "epoch": 1.5586481113320079, + "grad_norm": 0.41878571772654466, + "learning_rate": 6.957795513980214e-05, + "loss": 0.533, + "step": 784 + }, + { + "epoch": 1.5606361829025845, + "grad_norm": 0.42375355349911037, + "learning_rate": 6.954054335176461e-05, + "loss": 0.5292, + "step": 785 + }, + { + "epoch": 1.562624254473161, + "grad_norm": 0.37729083805071806, + "learning_rate": 6.950307463269633e-05, + "loss": 0.5337, + "step": 786 + }, + { + "epoch": 1.5646123260437377, + "grad_norm": 0.5172787811154693, + "learning_rate": 6.946554905480764e-05, + "loss": 0.5311, + "step": 787 + }, + { + "epoch": 1.5666003976143141, + "grad_norm": 0.6430291497025111, + "learning_rate": 6.942796669041845e-05, + "loss": 0.5266, + "step": 788 + }, + { + "epoch": 1.5685884691848906, + "grad_norm": 0.7341682662010934, + "learning_rate": 6.939032761195814e-05, + "loss": 0.5319, + "step": 789 + }, + { + "epoch": 1.570576540755467, + "grad_norm": 0.7733283890973897, + "learning_rate": 6.935263189196536e-05, + "loss": 0.525, + "step": 790 + }, + { + "epoch": 1.5725646123260437, + "grad_norm": 0.7479417608111679, + "learning_rate": 6.931487960308795e-05, + "loss": 0.5311, + "step": 791 + }, + { + "epoch": 1.5745526838966204, + "grad_norm": 0.6757799604868806, + "learning_rate": 6.927707081808274e-05, + "loss": 0.5302, + "step": 792 + }, + { + "epoch": 1.5765407554671969, + "grad_norm": 0.7227696163173898, + "learning_rate": 6.923920560981547e-05, + "loss": 0.5289, + "step": 793 + }, + { + "epoch": 1.5785288270377733, + "grad_norm": 0.824107106148003, + "learning_rate": 6.920128405126057e-05, + "loss": 0.5256, + "step": 794 + }, + { + "epoch": 1.5805168986083498, + "grad_norm": 0.8628110973646821, + "learning_rate": 6.916330621550116e-05, + "loss": 0.5295, + "step": 795 + }, + { + "epoch": 1.5825049701789264, + "grad_norm": 0.7654075564561619, + "learning_rate": 6.912527217572872e-05, + "loss": 0.5295, + "step": 796 + }, + { + "epoch": 1.5844930417495031, + "grad_norm": 0.6046199689394902, + "learning_rate": 6.90871820052431e-05, + "loss": 0.5244, + "step": 797 + }, + { + "epoch": 1.5864811133200796, + "grad_norm": 0.4598376110935637, + "learning_rate": 6.90490357774523e-05, + "loss": 0.5242, + "step": 798 + }, + { + "epoch": 1.588469184890656, + "grad_norm": 0.383063601508151, + "learning_rate": 6.90108335658724e-05, + "loss": 0.528, + "step": 799 + }, + { + "epoch": 1.5904572564612325, + "grad_norm": 0.3312610747130047, + "learning_rate": 6.897257544412733e-05, + "loss": 0.5233, + "step": 800 + }, + { + "epoch": 1.5924453280318092, + "grad_norm": 0.37975991863998054, + "learning_rate": 6.893426148594878e-05, + "loss": 0.5258, + "step": 801 + }, + { + "epoch": 1.5944333996023858, + "grad_norm": 0.518783862803526, + "learning_rate": 6.889589176517605e-05, + "loss": 0.5279, + "step": 802 + }, + { + "epoch": 1.5964214711729623, + "grad_norm": 0.620415152981916, + "learning_rate": 6.885746635575594e-05, + "loss": 0.5277, + "step": 803 + }, + { + "epoch": 1.5984095427435387, + "grad_norm": 0.5372244798306283, + "learning_rate": 6.88189853317425e-05, + "loss": 0.528, + "step": 804 + }, + { + "epoch": 1.6003976143141152, + "grad_norm": 0.35429448129413066, + "learning_rate": 6.878044876729703e-05, + "loss": 0.5311, + "step": 805 + }, + { + "epoch": 1.6023856858846919, + "grad_norm": 0.471670336402686, + "learning_rate": 6.874185673668783e-05, + "loss": 0.5295, + "step": 806 + }, + { + "epoch": 1.6043737574552683, + "grad_norm": 0.29450520073608794, + "learning_rate": 6.870320931429013e-05, + "loss": 0.5227, + "step": 807 + }, + { + "epoch": 1.606361829025845, + "grad_norm": 0.3606020368465704, + "learning_rate": 6.866450657458587e-05, + "loss": 0.5269, + "step": 808 + }, + { + "epoch": 1.6083499005964215, + "grad_norm": 0.4286070793201051, + "learning_rate": 6.862574859216363e-05, + "loss": 0.5255, + "step": 809 + }, + { + "epoch": 1.610337972166998, + "grad_norm": 0.42927764814865915, + "learning_rate": 6.858693544171845e-05, + "loss": 0.5279, + "step": 810 + }, + { + "epoch": 1.6123260437375746, + "grad_norm": 0.4410104674784239, + "learning_rate": 6.854806719805166e-05, + "loss": 0.5265, + "step": 811 + }, + { + "epoch": 1.614314115308151, + "grad_norm": 0.531026735276354, + "learning_rate": 6.850914393607082e-05, + "loss": 0.5309, + "step": 812 + }, + { + "epoch": 1.6163021868787277, + "grad_norm": 0.6716814454381097, + "learning_rate": 6.847016573078947e-05, + "loss": 0.5366, + "step": 813 + }, + { + "epoch": 1.6182902584493042, + "grad_norm": 0.7934744929903685, + "learning_rate": 6.843113265732708e-05, + "loss": 0.5192, + "step": 814 + }, + { + "epoch": 1.6202783300198806, + "grad_norm": 0.9038605001312126, + "learning_rate": 6.839204479090883e-05, + "loss": 0.5283, + "step": 815 + }, + { + "epoch": 1.622266401590457, + "grad_norm": 0.9290349986214314, + "learning_rate": 6.835290220686553e-05, + "loss": 0.529, + "step": 816 + }, + { + "epoch": 1.6242544731610338, + "grad_norm": 0.835195207409939, + "learning_rate": 6.831370498063341e-05, + "loss": 0.5316, + "step": 817 + }, + { + "epoch": 1.6262425447316105, + "grad_norm": 0.6921457969116479, + "learning_rate": 6.827445318775401e-05, + "loss": 0.5268, + "step": 818 + }, + { + "epoch": 1.628230616302187, + "grad_norm": 0.4237508185833249, + "learning_rate": 6.823514690387407e-05, + "loss": 0.5195, + "step": 819 + }, + { + "epoch": 1.6302186878727634, + "grad_norm": 0.3942746067868676, + "learning_rate": 6.81957862047453e-05, + "loss": 0.5276, + "step": 820 + }, + { + "epoch": 1.6322067594433398, + "grad_norm": 0.5755721778151188, + "learning_rate": 6.815637116622432e-05, + "loss": 0.5344, + "step": 821 + }, + { + "epoch": 1.6341948310139165, + "grad_norm": 0.6478367776398117, + "learning_rate": 6.811690186427244e-05, + "loss": 0.5237, + "step": 822 + }, + { + "epoch": 1.6361829025844932, + "grad_norm": 0.5020337919490125, + "learning_rate": 6.807737837495556e-05, + "loss": 0.5285, + "step": 823 + }, + { + "epoch": 1.6381709741550696, + "grad_norm": 0.36435989435714544, + "learning_rate": 6.803780077444401e-05, + "loss": 0.5259, + "step": 824 + }, + { + "epoch": 1.640159045725646, + "grad_norm": 0.3882654380737712, + "learning_rate": 6.799816913901242e-05, + "loss": 0.5271, + "step": 825 + }, + { + "epoch": 1.6421471172962225, + "grad_norm": 0.49404535476174083, + "learning_rate": 6.795848354503951e-05, + "loss": 0.5247, + "step": 826 + }, + { + "epoch": 1.6441351888667992, + "grad_norm": 0.5112014431648076, + "learning_rate": 6.791874406900806e-05, + "loss": 0.5265, + "step": 827 + }, + { + "epoch": 1.646123260437376, + "grad_norm": 0.45808472140668066, + "learning_rate": 6.787895078750465e-05, + "loss": 0.525, + "step": 828 + }, + { + "epoch": 1.6481113320079523, + "grad_norm": 1.2914831242671405, + "learning_rate": 6.783910377721954e-05, + "loss": 0.52, + "step": 829 + }, + { + "epoch": 1.6500994035785288, + "grad_norm": 0.3384619854495341, + "learning_rate": 6.779920311494657e-05, + "loss": 0.5272, + "step": 830 + }, + { + "epoch": 1.6520874751491053, + "grad_norm": 0.5083556366105404, + "learning_rate": 6.775924887758296e-05, + "loss": 0.5195, + "step": 831 + }, + { + "epoch": 1.654075546719682, + "grad_norm": 0.6386003447389207, + "learning_rate": 6.771924114212917e-05, + "loss": 0.5243, + "step": 832 + }, + { + "epoch": 1.6560636182902586, + "grad_norm": 0.5423661412327074, + "learning_rate": 6.767917998568882e-05, + "loss": 0.5214, + "step": 833 + }, + { + "epoch": 1.658051689860835, + "grad_norm": 0.46973576508674253, + "learning_rate": 6.763906548546841e-05, + "loss": 0.53, + "step": 834 + }, + { + "epoch": 1.6600397614314115, + "grad_norm": 0.40006367573081403, + "learning_rate": 6.75988977187773e-05, + "loss": 0.5302, + "step": 835 + }, + { + "epoch": 1.662027833001988, + "grad_norm": 0.4694950009603063, + "learning_rate": 6.755867676302747e-05, + "loss": 0.5212, + "step": 836 + }, + { + "epoch": 1.6640159045725647, + "grad_norm": 0.5151875858962307, + "learning_rate": 6.751840269573342e-05, + "loss": 0.5286, + "step": 837 + }, + { + "epoch": 1.666003976143141, + "grad_norm": 1.2198920217976148, + "learning_rate": 6.747807559451201e-05, + "loss": 0.5271, + "step": 838 + }, + { + "epoch": 1.6679920477137178, + "grad_norm": 0.3313618013972345, + "learning_rate": 6.74376955370823e-05, + "loss": 0.5336, + "step": 839 + }, + { + "epoch": 1.6699801192842942, + "grad_norm": 0.4987390488162863, + "learning_rate": 6.739726260126545e-05, + "loss": 0.5317, + "step": 840 + }, + { + "epoch": 1.6719681908548707, + "grad_norm": 0.7074939333172882, + "learning_rate": 6.735677686498443e-05, + "loss": 0.537, + "step": 841 + }, + { + "epoch": 1.6739562624254472, + "grad_norm": 0.6747029464356521, + "learning_rate": 6.731623840626408e-05, + "loss": 0.5372, + "step": 842 + }, + { + "epoch": 1.6759443339960238, + "grad_norm": 0.6466193766227213, + "learning_rate": 6.727564730323077e-05, + "loss": 0.5296, + "step": 843 + }, + { + "epoch": 1.6779324055666005, + "grad_norm": 0.5584777552614117, + "learning_rate": 6.723500363411238e-05, + "loss": 0.5252, + "step": 844 + }, + { + "epoch": 1.679920477137177, + "grad_norm": 0.453649217028817, + "learning_rate": 6.719430747723805e-05, + "loss": 0.5288, + "step": 845 + }, + { + "epoch": 1.6819085487077534, + "grad_norm": 0.48188541292199083, + "learning_rate": 6.715355891103811e-05, + "loss": 0.5264, + "step": 846 + }, + { + "epoch": 1.6838966202783299, + "grad_norm": 0.4254994544193942, + "learning_rate": 6.711275801404387e-05, + "loss": 0.529, + "step": 847 + }, + { + "epoch": 1.6858846918489065, + "grad_norm": 0.380870644142891, + "learning_rate": 6.707190486488751e-05, + "loss": 0.5276, + "step": 848 + }, + { + "epoch": 1.6878727634194832, + "grad_norm": 0.4177074544613175, + "learning_rate": 6.70309995423019e-05, + "loss": 0.5318, + "step": 849 + }, + { + "epoch": 1.6898608349900597, + "grad_norm": 0.395911119103237, + "learning_rate": 6.699004212512046e-05, + "loss": 0.526, + "step": 850 + }, + { + "epoch": 1.6918489065606361, + "grad_norm": 0.45543127339921896, + "learning_rate": 6.694903269227703e-05, + "loss": 0.5248, + "step": 851 + }, + { + "epoch": 1.6938369781312126, + "grad_norm": 0.44150579657087613, + "learning_rate": 6.690797132280565e-05, + "loss": 0.5243, + "step": 852 + }, + { + "epoch": 1.6958250497017893, + "grad_norm": 0.3842977414436349, + "learning_rate": 6.686685809584051e-05, + "loss": 0.5172, + "step": 853 + }, + { + "epoch": 1.697813121272366, + "grad_norm": 0.3267241356619446, + "learning_rate": 6.682569309061567e-05, + "loss": 0.5359, + "step": 854 + }, + { + "epoch": 1.6998011928429424, + "grad_norm": 0.31023857384364334, + "learning_rate": 6.678447638646507e-05, + "loss": 0.5203, + "step": 855 + }, + { + "epoch": 1.7017892644135189, + "grad_norm": 0.4582609159970721, + "learning_rate": 6.67432080628222e-05, + "loss": 0.5246, + "step": 856 + }, + { + "epoch": 1.7037773359840953, + "grad_norm": 0.4371367746577152, + "learning_rate": 6.670188819922006e-05, + "loss": 0.5256, + "step": 857 + }, + { + "epoch": 1.705765407554672, + "grad_norm": 0.3601588642012915, + "learning_rate": 6.666051687529102e-05, + "loss": 0.527, + "step": 858 + }, + { + "epoch": 1.7077534791252487, + "grad_norm": 0.3598932666256967, + "learning_rate": 6.661909417076656e-05, + "loss": 0.522, + "step": 859 + }, + { + "epoch": 1.7097415506958251, + "grad_norm": 0.3814306793726626, + "learning_rate": 6.65776201654772e-05, + "loss": 0.5263, + "step": 860 + }, + { + "epoch": 1.7117296222664016, + "grad_norm": 0.5155709892788785, + "learning_rate": 6.653609493935239e-05, + "loss": 0.5277, + "step": 861 + }, + { + "epoch": 1.713717693836978, + "grad_norm": 0.592297522606716, + "learning_rate": 6.649451857242018e-05, + "loss": 0.5302, + "step": 862 + }, + { + "epoch": 1.7157057654075547, + "grad_norm": 0.62317820887808, + "learning_rate": 6.64528911448073e-05, + "loss": 0.5314, + "step": 863 + }, + { + "epoch": 1.7176938369781312, + "grad_norm": 0.5293921255569864, + "learning_rate": 6.641121273673877e-05, + "loss": 0.5341, + "step": 864 + }, + { + "epoch": 1.7196819085487078, + "grad_norm": 0.3877701813428404, + "learning_rate": 6.636948342853797e-05, + "loss": 0.5298, + "step": 865 + }, + { + "epoch": 1.7216699801192843, + "grad_norm": 0.3207068664780065, + "learning_rate": 6.632770330062628e-05, + "loss": 0.5185, + "step": 866 + }, + { + "epoch": 1.7236580516898607, + "grad_norm": 0.3000147535618842, + "learning_rate": 6.62858724335231e-05, + "loss": 0.5279, + "step": 867 + }, + { + "epoch": 1.7256461232604374, + "grad_norm": 0.27110479588599784, + "learning_rate": 6.624399090784555e-05, + "loss": 0.5166, + "step": 868 + }, + { + "epoch": 1.7276341948310139, + "grad_norm": 0.2534706020052216, + "learning_rate": 6.62020588043084e-05, + "loss": 0.5299, + "step": 869 + }, + { + "epoch": 1.7296222664015906, + "grad_norm": 0.2651273278605272, + "learning_rate": 6.616007620372396e-05, + "loss": 0.5213, + "step": 870 + }, + { + "epoch": 1.731610337972167, + "grad_norm": 0.3008823267991322, + "learning_rate": 6.611804318700174e-05, + "loss": 0.5231, + "step": 871 + }, + { + "epoch": 1.7335984095427435, + "grad_norm": 0.36886074204926794, + "learning_rate": 6.60759598351485e-05, + "loss": 0.5184, + "step": 872 + }, + { + "epoch": 1.73558648111332, + "grad_norm": 0.4053040242003246, + "learning_rate": 6.603382622926801e-05, + "loss": 0.5272, + "step": 873 + }, + { + "epoch": 1.7375745526838966, + "grad_norm": 0.35686447014595274, + "learning_rate": 6.599164245056081e-05, + "loss": 0.5087, + "step": 874 + }, + { + "epoch": 1.7395626242544733, + "grad_norm": 0.3429941917807466, + "learning_rate": 6.594940858032422e-05, + "loss": 0.5195, + "step": 875 + }, + { + "epoch": 1.7415506958250497, + "grad_norm": 0.34576654170390286, + "learning_rate": 6.590712469995204e-05, + "loss": 0.5175, + "step": 876 + }, + { + "epoch": 1.7435387673956262, + "grad_norm": 0.3310782562648362, + "learning_rate": 6.586479089093451e-05, + "loss": 0.5133, + "step": 877 + }, + { + "epoch": 1.7455268389662026, + "grad_norm": 0.4318684757964282, + "learning_rate": 6.582240723485801e-05, + "loss": 0.5184, + "step": 878 + }, + { + "epoch": 1.7475149105367793, + "grad_norm": 0.6198244221196105, + "learning_rate": 6.577997381340509e-05, + "loss": 0.5278, + "step": 879 + }, + { + "epoch": 1.749502982107356, + "grad_norm": 0.707683440817649, + "learning_rate": 6.573749070835409e-05, + "loss": 0.5237, + "step": 880 + }, + { + "epoch": 1.7514910536779325, + "grad_norm": 0.676898274391904, + "learning_rate": 6.56949580015792e-05, + "loss": 0.5203, + "step": 881 + }, + { + "epoch": 1.753479125248509, + "grad_norm": 0.6183840795701814, + "learning_rate": 6.565237577505016e-05, + "loss": 0.5108, + "step": 882 + }, + { + "epoch": 1.7554671968190854, + "grad_norm": 0.5793914378061121, + "learning_rate": 6.560974411083214e-05, + "loss": 0.5205, + "step": 883 + }, + { + "epoch": 1.757455268389662, + "grad_norm": 0.5772648421166539, + "learning_rate": 6.556706309108562e-05, + "loss": 0.5206, + "step": 884 + }, + { + "epoch": 1.7594433399602387, + "grad_norm": 0.6223903009611668, + "learning_rate": 6.552433279806617e-05, + "loss": 0.5275, + "step": 885 + }, + { + "epoch": 1.7614314115308152, + "grad_norm": 0.694303234381892, + "learning_rate": 6.548155331412431e-05, + "loss": 0.5181, + "step": 886 + }, + { + "epoch": 1.7634194831013916, + "grad_norm": 0.7367498764846339, + "learning_rate": 6.54387247217054e-05, + "loss": 0.5212, + "step": 887 + }, + { + "epoch": 1.765407554671968, + "grad_norm": 0.6573127727786128, + "learning_rate": 6.539584710334941e-05, + "loss": 0.5165, + "step": 888 + }, + { + "epoch": 1.7673956262425448, + "grad_norm": 0.5423711722715193, + "learning_rate": 6.535292054169083e-05, + "loss": 0.5255, + "step": 889 + }, + { + "epoch": 1.7693836978131214, + "grad_norm": 0.37296323316331076, + "learning_rate": 6.530994511945841e-05, + "loss": 0.5175, + "step": 890 + }, + { + "epoch": 1.771371769383698, + "grad_norm": 0.390630809690755, + "learning_rate": 6.526692091947513e-05, + "loss": 0.5231, + "step": 891 + }, + { + "epoch": 1.7733598409542743, + "grad_norm": 0.628946177002013, + "learning_rate": 6.522384802465792e-05, + "loss": 0.516, + "step": 892 + }, + { + "epoch": 1.7753479125248508, + "grad_norm": 0.7131786851847514, + "learning_rate": 6.518072651801763e-05, + "loss": 0.5139, + "step": 893 + }, + { + "epoch": 1.7773359840954275, + "grad_norm": 0.677048790263275, + "learning_rate": 6.513755648265874e-05, + "loss": 0.5085, + "step": 894 + }, + { + "epoch": 1.779324055666004, + "grad_norm": 0.5545892812569629, + "learning_rate": 6.509433800177925e-05, + "loss": 0.5157, + "step": 895 + }, + { + "epoch": 1.7813121272365806, + "grad_norm": 0.4221871273105676, + "learning_rate": 6.505107115867053e-05, + "loss": 0.5129, + "step": 896 + }, + { + "epoch": 1.783300198807157, + "grad_norm": 0.33895277193033474, + "learning_rate": 6.500775603671721e-05, + "loss": 0.5269, + "step": 897 + }, + { + "epoch": 1.7852882703777335, + "grad_norm": 0.4106684204987088, + "learning_rate": 6.496439271939689e-05, + "loss": 0.5251, + "step": 898 + }, + { + "epoch": 1.78727634194831, + "grad_norm": 0.4813385967804267, + "learning_rate": 6.492098129028008e-05, + "loss": 0.5208, + "step": 899 + }, + { + "epoch": 1.7892644135188867, + "grad_norm": 0.5095331170047483, + "learning_rate": 6.487752183303002e-05, + "loss": 0.5175, + "step": 900 + }, + { + "epoch": 1.7912524850894633, + "grad_norm": 0.5785645590013878, + "learning_rate": 6.483401443140251e-05, + "loss": 0.5182, + "step": 901 + }, + { + "epoch": 1.7932405566600398, + "grad_norm": 0.547793961103489, + "learning_rate": 6.479045916924574e-05, + "loss": 0.5165, + "step": 902 + }, + { + "epoch": 1.7952286282306162, + "grad_norm": 0.41600517135228604, + "learning_rate": 6.474685613050014e-05, + "loss": 0.5049, + "step": 903 + }, + { + "epoch": 1.7972166998011927, + "grad_norm": 0.3299862272982457, + "learning_rate": 6.470320539919823e-05, + "loss": 0.5191, + "step": 904 + }, + { + "epoch": 1.7992047713717694, + "grad_norm": 0.32073772048850396, + "learning_rate": 6.465950705946441e-05, + "loss": 0.5117, + "step": 905 + }, + { + "epoch": 1.801192842942346, + "grad_norm": 0.35624437991584945, + "learning_rate": 6.461576119551488e-05, + "loss": 0.5152, + "step": 906 + }, + { + "epoch": 1.8031809145129225, + "grad_norm": 0.3359272702550711, + "learning_rate": 6.457196789165737e-05, + "loss": 0.5208, + "step": 907 + }, + { + "epoch": 1.805168986083499, + "grad_norm": 0.3001896290540342, + "learning_rate": 6.452812723229111e-05, + "loss": 0.5155, + "step": 908 + }, + { + "epoch": 1.8071570576540754, + "grad_norm": 0.43244608682808966, + "learning_rate": 6.448423930190653e-05, + "loss": 0.5168, + "step": 909 + }, + { + "epoch": 1.809145129224652, + "grad_norm": 0.33914062705350506, + "learning_rate": 6.444030418508517e-05, + "loss": 0.5215, + "step": 910 + }, + { + "epoch": 1.8111332007952288, + "grad_norm": 0.3793585723969184, + "learning_rate": 6.439632196649955e-05, + "loss": 0.5235, + "step": 911 + }, + { + "epoch": 1.8131212723658052, + "grad_norm": 0.36209092348601185, + "learning_rate": 6.435229273091295e-05, + "loss": 0.5067, + "step": 912 + }, + { + "epoch": 1.8151093439363817, + "grad_norm": 0.39315793371620283, + "learning_rate": 6.430821656317921e-05, + "loss": 0.5202, + "step": 913 + }, + { + "epoch": 1.8170974155069581, + "grad_norm": 0.3870294576875787, + "learning_rate": 6.42640935482427e-05, + "loss": 0.525, + "step": 914 + }, + { + "epoch": 1.8190854870775348, + "grad_norm": 0.3740502813906515, + "learning_rate": 6.421992377113801e-05, + "loss": 0.5189, + "step": 915 + }, + { + "epoch": 1.8210735586481115, + "grad_norm": 0.36305841904786135, + "learning_rate": 6.417570731698989e-05, + "loss": 0.5191, + "step": 916 + }, + { + "epoch": 1.823061630218688, + "grad_norm": 0.37807892969813695, + "learning_rate": 6.413144427101303e-05, + "loss": 0.5144, + "step": 917 + }, + { + "epoch": 1.8250497017892644, + "grad_norm": 0.3458228510072188, + "learning_rate": 6.40871347185119e-05, + "loss": 0.5189, + "step": 918 + }, + { + "epoch": 1.8270377733598409, + "grad_norm": 0.38758188828148943, + "learning_rate": 6.404277874488063e-05, + "loss": 0.5108, + "step": 919 + }, + { + "epoch": 1.8290258449304175, + "grad_norm": 0.41572979496537815, + "learning_rate": 6.399837643560279e-05, + "loss": 0.5136, + "step": 920 + }, + { + "epoch": 1.831013916500994, + "grad_norm": 0.42898037039016224, + "learning_rate": 6.395392787625126e-05, + "loss": 0.5235, + "step": 921 + }, + { + "epoch": 1.8330019880715707, + "grad_norm": 0.4978473830264871, + "learning_rate": 6.390943315248804e-05, + "loss": 0.5175, + "step": 922 + }, + { + "epoch": 1.8349900596421471, + "grad_norm": 0.6629837956376023, + "learning_rate": 6.386489235006412e-05, + "loss": 0.5178, + "step": 923 + }, + { + "epoch": 1.8369781312127236, + "grad_norm": 0.5000042265219343, + "learning_rate": 6.382030555481928e-05, + "loss": 0.5158, + "step": 924 + }, + { + "epoch": 1.8389662027833003, + "grad_norm": 0.4279576070326289, + "learning_rate": 6.377567285268192e-05, + "loss": 0.5223, + "step": 925 + }, + { + "epoch": 1.8409542743538767, + "grad_norm": 0.3459736856516694, + "learning_rate": 6.373099432966895e-05, + "loss": 0.5189, + "step": 926 + }, + { + "epoch": 1.8429423459244534, + "grad_norm": 0.2961388042649445, + "learning_rate": 6.368627007188556e-05, + "loss": 0.5159, + "step": 927 + }, + { + "epoch": 1.8449304174950298, + "grad_norm": 1.158821404766412, + "learning_rate": 6.364150016552511e-05, + "loss": 0.5118, + "step": 928 + }, + { + "epoch": 1.8469184890656063, + "grad_norm": 0.3313243590186058, + "learning_rate": 6.359668469686891e-05, + "loss": 0.5133, + "step": 929 + }, + { + "epoch": 1.8489065606361827, + "grad_norm": 0.3442419248332627, + "learning_rate": 6.355182375228606e-05, + "loss": 0.5188, + "step": 930 + }, + { + "epoch": 1.8508946322067594, + "grad_norm": 0.46405524867421283, + "learning_rate": 6.350691741823334e-05, + "loss": 0.5234, + "step": 931 + }, + { + "epoch": 1.852882703777336, + "grad_norm": 0.5552544859564328, + "learning_rate": 6.346196578125499e-05, + "loss": 0.5283, + "step": 932 + }, + { + "epoch": 1.8548707753479126, + "grad_norm": 0.6149770086477317, + "learning_rate": 6.341696892798256e-05, + "loss": 0.5213, + "step": 933 + }, + { + "epoch": 1.856858846918489, + "grad_norm": 0.6066033014098224, + "learning_rate": 6.337192694513474e-05, + "loss": 0.5258, + "step": 934 + }, + { + "epoch": 1.8588469184890655, + "grad_norm": 0.5534949185853664, + "learning_rate": 6.33268399195172e-05, + "loss": 0.5217, + "step": 935 + }, + { + "epoch": 1.8608349900596421, + "grad_norm": 0.46419762093019756, + "learning_rate": 6.328170793802239e-05, + "loss": 0.5184, + "step": 936 + }, + { + "epoch": 1.8628230616302188, + "grad_norm": 0.40115737747047064, + "learning_rate": 6.323653108762946e-05, + "loss": 0.5136, + "step": 937 + }, + { + "epoch": 1.8648111332007953, + "grad_norm": 0.4968544590324144, + "learning_rate": 6.319130945540397e-05, + "loss": 0.5173, + "step": 938 + }, + { + "epoch": 1.8667992047713717, + "grad_norm": 0.8922179521301115, + "learning_rate": 6.31460431284978e-05, + "loss": 0.5387, + "step": 939 + }, + { + "epoch": 1.8687872763419482, + "grad_norm": 0.8274425993048475, + "learning_rate": 6.310073219414897e-05, + "loss": 0.5171, + "step": 940 + }, + { + "epoch": 1.8707753479125249, + "grad_norm": 0.9470368749962408, + "learning_rate": 6.305537673968152e-05, + "loss": 0.5233, + "step": 941 + }, + { + "epoch": 1.8727634194831015, + "grad_norm": 0.9116078804524538, + "learning_rate": 6.300997685250518e-05, + "loss": 0.519, + "step": 942 + }, + { + "epoch": 1.874751491053678, + "grad_norm": 0.7644404414742432, + "learning_rate": 6.296453262011543e-05, + "loss": 0.522, + "step": 943 + }, + { + "epoch": 1.8767395626242545, + "grad_norm": 0.6794360427532787, + "learning_rate": 6.291904413009313e-05, + "loss": 0.5068, + "step": 944 + }, + { + "epoch": 1.878727634194831, + "grad_norm": 6.258038066728078, + "learning_rate": 6.287351147010445e-05, + "loss": 0.5197, + "step": 945 + }, + { + "epoch": 1.8807157057654076, + "grad_norm": 1.0714907278173418, + "learning_rate": 6.282793472790073e-05, + "loss": 0.5129, + "step": 946 + }, + { + "epoch": 1.882703777335984, + "grad_norm": 1.6860408777693041, + "learning_rate": 6.27823139913182e-05, + "loss": 0.5158, + "step": 947 + }, + { + "epoch": 1.8846918489065607, + "grad_norm": 0.9157025608605575, + "learning_rate": 6.273664934827792e-05, + "loss": 0.5254, + "step": 948 + }, + { + "epoch": 1.8866799204771372, + "grad_norm": 2.3654517437481464, + "learning_rate": 6.269094088678555e-05, + "loss": 0.5227, + "step": 949 + }, + { + "epoch": 1.8886679920477136, + "grad_norm": 1.941802659426266, + "learning_rate": 6.264518869493123e-05, + "loss": 0.5288, + "step": 950 + }, + { + "epoch": 1.8906560636182903, + "grad_norm": 1.3723093581829342, + "learning_rate": 6.259939286088932e-05, + "loss": 0.5272, + "step": 951 + }, + { + "epoch": 1.8926441351888668, + "grad_norm": 1.1729092180787042, + "learning_rate": 6.255355347291833e-05, + "loss": 0.5269, + "step": 952 + }, + { + "epoch": 1.8946322067594434, + "grad_norm": 1.5119822561319634, + "learning_rate": 6.25076706193607e-05, + "loss": 0.5199, + "step": 953 + }, + { + "epoch": 1.89662027833002, + "grad_norm": 1.0227380539137576, + "learning_rate": 6.246174438864264e-05, + "loss": 0.5208, + "step": 954 + }, + { + "epoch": 1.8986083499005963, + "grad_norm": 1.6110735368657032, + "learning_rate": 6.241577486927396e-05, + "loss": 0.5267, + "step": 955 + }, + { + "epoch": 1.9005964214711728, + "grad_norm": 1.3696020584560245, + "learning_rate": 6.236976214984787e-05, + "loss": 0.5307, + "step": 956 + }, + { + "epoch": 1.9025844930417495, + "grad_norm": 1.1164032790705993, + "learning_rate": 6.232370631904085e-05, + "loss": 0.5235, + "step": 957 + }, + { + "epoch": 1.9045725646123262, + "grad_norm": 0.9403848803470333, + "learning_rate": 6.227760746561248e-05, + "loss": 0.5126, + "step": 958 + }, + { + "epoch": 1.9065606361829026, + "grad_norm": 1.1097967355382794, + "learning_rate": 6.223146567840526e-05, + "loss": 0.5213, + "step": 959 + }, + { + "epoch": 1.908548707753479, + "grad_norm": 0.7036299666285428, + "learning_rate": 6.218528104634439e-05, + "loss": 0.5165, + "step": 960 + }, + { + "epoch": 1.9105367793240555, + "grad_norm": 1.1611817296430542, + "learning_rate": 6.213905365843767e-05, + "loss": 0.5172, + "step": 961 + }, + { + "epoch": 1.9125248508946322, + "grad_norm": 0.6447298644520358, + "learning_rate": 6.209278360377532e-05, + "loss": 0.5239, + "step": 962 + }, + { + "epoch": 1.9145129224652089, + "grad_norm": 0.961035900760647, + "learning_rate": 6.204647097152976e-05, + "loss": 0.5158, + "step": 963 + }, + { + "epoch": 1.9165009940357853, + "grad_norm": 0.5611928909928571, + "learning_rate": 6.200011585095546e-05, + "loss": 0.5207, + "step": 964 + }, + { + "epoch": 1.9184890656063618, + "grad_norm": 0.8349582131183718, + "learning_rate": 6.195371833138878e-05, + "loss": 0.517, + "step": 965 + }, + { + "epoch": 1.9204771371769382, + "grad_norm": 0.4906951073133178, + "learning_rate": 6.190727850224781e-05, + "loss": 0.5205, + "step": 966 + }, + { + "epoch": 1.922465208747515, + "grad_norm": 0.5781023232362354, + "learning_rate": 6.186079645303217e-05, + "loss": 0.5109, + "step": 967 + }, + { + "epoch": 1.9244532803180916, + "grad_norm": 0.4523458594191586, + "learning_rate": 6.181427227332284e-05, + "loss": 0.5163, + "step": 968 + }, + { + "epoch": 1.926441351888668, + "grad_norm": 0.5895393535936876, + "learning_rate": 6.1767706052782e-05, + "loss": 0.5101, + "step": 969 + }, + { + "epoch": 1.9284294234592445, + "grad_norm": 0.4494991711450939, + "learning_rate": 6.172109788115286e-05, + "loss": 0.5134, + "step": 970 + }, + { + "epoch": 1.930417495029821, + "grad_norm": 0.49354196995819166, + "learning_rate": 6.167444784825945e-05, + "loss": 0.5209, + "step": 971 + }, + { + "epoch": 1.9324055666003976, + "grad_norm": 0.734680878386189, + "learning_rate": 6.162775604400653e-05, + "loss": 0.538, + "step": 972 + }, + { + "epoch": 1.9343936381709743, + "grad_norm": 0.4126375739890944, + "learning_rate": 6.158102255837928e-05, + "loss": 0.5089, + "step": 973 + }, + { + "epoch": 1.9363817097415508, + "grad_norm": 2.5240012713966222, + "learning_rate": 6.153424748144331e-05, + "loss": 0.5484, + "step": 974 + }, + { + "epoch": 1.9383697813121272, + "grad_norm": 0.6639225231429577, + "learning_rate": 6.148743090334429e-05, + "loss": 0.5243, + "step": 975 + }, + { + "epoch": 1.9403578528827037, + "grad_norm": 0.7918093768800656, + "learning_rate": 6.144057291430794e-05, + "loss": 0.5137, + "step": 976 + }, + { + "epoch": 1.9423459244532804, + "grad_norm": 1.4149190907951867, + "learning_rate": 6.139367360463976e-05, + "loss": 0.5224, + "step": 977 + }, + { + "epoch": 1.9443339960238568, + "grad_norm": 0.46582696350343256, + "learning_rate": 6.134673306472487e-05, + "loss": 0.5182, + "step": 978 + }, + { + "epoch": 1.9463220675944335, + "grad_norm": 0.9635660986087046, + "learning_rate": 6.129975138502788e-05, + "loss": 0.528, + "step": 979 + }, + { + "epoch": 1.94831013916501, + "grad_norm": 1.1033730040373586, + "learning_rate": 6.125272865609267e-05, + "loss": 0.5189, + "step": 980 + }, + { + "epoch": 1.9502982107355864, + "grad_norm": 0.6710161371168709, + "learning_rate": 6.120566496854225e-05, + "loss": 0.5233, + "step": 981 + }, + { + "epoch": 1.9522862823061629, + "grad_norm": 0.5342304493013306, + "learning_rate": 6.11585604130785e-05, + "loss": 0.5188, + "step": 982 + }, + { + "epoch": 1.9542743538767395, + "grad_norm": 0.7351872609836071, + "learning_rate": 6.111141508048217e-05, + "loss": 0.5083, + "step": 983 + }, + { + "epoch": 1.9562624254473162, + "grad_norm": 0.6600541489771485, + "learning_rate": 6.106422906161248e-05, + "loss": 0.5234, + "step": 984 + }, + { + "epoch": 1.9582504970178927, + "grad_norm": 0.5571169506680288, + "learning_rate": 6.101700244740715e-05, + "loss": 0.5258, + "step": 985 + }, + { + "epoch": 1.9602385685884691, + "grad_norm": 0.7165724959209299, + "learning_rate": 6.0969735328882086e-05, + "loss": 0.5238, + "step": 986 + }, + { + "epoch": 1.9622266401590456, + "grad_norm": 0.643330366441739, + "learning_rate": 6.092242779713131e-05, + "loss": 0.5156, + "step": 987 + }, + { + "epoch": 1.9642147117296223, + "grad_norm": 0.476323219464556, + "learning_rate": 6.087507994332663e-05, + "loss": 0.5165, + "step": 988 + }, + { + "epoch": 1.966202783300199, + "grad_norm": 0.41439350527321883, + "learning_rate": 6.0827691858717674e-05, + "loss": 0.521, + "step": 989 + }, + { + "epoch": 1.9681908548707754, + "grad_norm": 0.5424398593597871, + "learning_rate": 6.078026363463153e-05, + "loss": 0.5212, + "step": 990 + }, + { + "epoch": 1.9701789264413518, + "grad_norm": 0.4603082917256751, + "learning_rate": 6.073279536247268e-05, + "loss": 0.5138, + "step": 991 + }, + { + "epoch": 1.9721669980119283, + "grad_norm": 0.5001963424623456, + "learning_rate": 6.068528713372276e-05, + "loss": 0.5247, + "step": 992 + }, + { + "epoch": 1.974155069582505, + "grad_norm": 0.4730380781447341, + "learning_rate": 6.063773903994043e-05, + "loss": 0.5141, + "step": 993 + }, + { + "epoch": 1.9761431411530817, + "grad_norm": 0.43352243194755535, + "learning_rate": 6.0590151172761186e-05, + "loss": 0.5205, + "step": 994 + }, + { + "epoch": 1.978131212723658, + "grad_norm": 0.48285340824373285, + "learning_rate": 6.054252362389715e-05, + "loss": 0.5118, + "step": 995 + }, + { + "epoch": 1.9801192842942346, + "grad_norm": 0.40877908110477573, + "learning_rate": 6.049485648513696e-05, + "loss": 0.5225, + "step": 996 + }, + { + "epoch": 1.982107355864811, + "grad_norm": 0.3780132592901253, + "learning_rate": 6.0447149848345476e-05, + "loss": 0.5206, + "step": 997 + }, + { + "epoch": 1.9840954274353877, + "grad_norm": 0.35453842643953837, + "learning_rate": 6.0399403805463767e-05, + "loss": 0.5126, + "step": 998 + }, + { + "epoch": 1.9860834990059644, + "grad_norm": 0.3032849729752592, + "learning_rate": 6.03516184485088e-05, + "loss": 0.5185, + "step": 999 + }, + { + "epoch": 1.9880715705765408, + "grad_norm": 0.49672445891973527, + "learning_rate": 6.030379386957331e-05, + "loss": 0.5257, + "step": 1000 + }, + { + "epoch": 1.9900596421471173, + "grad_norm": 0.34605524527043097, + "learning_rate": 6.025593016082564e-05, + "loss": 0.5149, + "step": 1001 + }, + { + "epoch": 1.9920477137176937, + "grad_norm": 0.34364685471639445, + "learning_rate": 6.020802741450951e-05, + "loss": 0.5118, + "step": 1002 + }, + { + "epoch": 1.9940357852882704, + "grad_norm": 0.3179870833998065, + "learning_rate": 6.016008572294392e-05, + "loss": 0.515, + "step": 1003 + }, + { + "epoch": 1.9960238568588469, + "grad_norm": 0.3567363057139461, + "learning_rate": 6.01121051785229e-05, + "loss": 0.5226, + "step": 1004 + }, + { + "epoch": 1.9980119284294235, + "grad_norm": 0.29232456854575584, + "learning_rate": 6.006408587371535e-05, + "loss": 0.5145, + "step": 1005 + }, + { + "epoch": 2.0, + "grad_norm": 0.3614805776515675, + "learning_rate": 6.001602790106488e-05, + "loss": 0.5081, + "step": 1006 + }, + { + "epoch": 2.0019880715705765, + "grad_norm": 0.4265486606705852, + "learning_rate": 5.996793135318962e-05, + "loss": 0.4912, + "step": 1007 + }, + { + "epoch": 2.003976143141153, + "grad_norm": 0.41944427653109145, + "learning_rate": 5.991979632278206e-05, + "loss": 0.4852, + "step": 1008 + }, + { + "epoch": 2.00596421471173, + "grad_norm": 0.46970990607031543, + "learning_rate": 5.987162290260883e-05, + "loss": 0.4798, + "step": 1009 + }, + { + "epoch": 2.0079522862823063, + "grad_norm": 0.3494875624015793, + "learning_rate": 5.9823411185510546e-05, + "loss": 0.48, + "step": 1010 + }, + { + "epoch": 2.0099403578528827, + "grad_norm": 0.33653022283326256, + "learning_rate": 5.9775161264401636e-05, + "loss": 0.4841, + "step": 1011 + }, + { + "epoch": 2.011928429423459, + "grad_norm": 0.3886220441073982, + "learning_rate": 5.972687323227016e-05, + "loss": 0.4829, + "step": 1012 + }, + { + "epoch": 2.0139165009940356, + "grad_norm": 0.35168024551500693, + "learning_rate": 5.967854718217764e-05, + "loss": 0.4802, + "step": 1013 + }, + { + "epoch": 2.0159045725646125, + "grad_norm": 0.29225835757850427, + "learning_rate": 5.963018320725884e-05, + "loss": 0.4809, + "step": 1014 + }, + { + "epoch": 2.017892644135189, + "grad_norm": 0.2910719808738326, + "learning_rate": 5.9581781400721605e-05, + "loss": 0.4722, + "step": 1015 + }, + { + "epoch": 2.0198807157057654, + "grad_norm": 0.2961140641840679, + "learning_rate": 5.9533341855846745e-05, + "loss": 0.4749, + "step": 1016 + }, + { + "epoch": 2.021868787276342, + "grad_norm": 0.2297321304146483, + "learning_rate": 5.948486466598773e-05, + "loss": 0.479, + "step": 1017 + }, + { + "epoch": 2.0238568588469183, + "grad_norm": 0.276146779314003, + "learning_rate": 5.943634992457065e-05, + "loss": 0.4816, + "step": 1018 + }, + { + "epoch": 2.0258449304174952, + "grad_norm": 0.3894209253927516, + "learning_rate": 5.938779772509388e-05, + "loss": 0.4845, + "step": 1019 + }, + { + "epoch": 2.0278330019880717, + "grad_norm": 0.3399631549933991, + "learning_rate": 5.933920816112807e-05, + "loss": 0.4718, + "step": 1020 + }, + { + "epoch": 2.029821073558648, + "grad_norm": 0.34376727093560305, + "learning_rate": 5.929058132631584e-05, + "loss": 0.4817, + "step": 1021 + }, + { + "epoch": 2.0318091451292246, + "grad_norm": 0.39737386982055917, + "learning_rate": 5.924191731437162e-05, + "loss": 0.4827, + "step": 1022 + }, + { + "epoch": 2.033797216699801, + "grad_norm": 0.2551201885904131, + "learning_rate": 5.9193216219081546e-05, + "loss": 0.4813, + "step": 1023 + }, + { + "epoch": 2.0357852882703775, + "grad_norm": 0.25758071434624624, + "learning_rate": 5.914447813430315e-05, + "loss": 0.475, + "step": 1024 + }, + { + "epoch": 2.0377733598409544, + "grad_norm": 0.2655696299091699, + "learning_rate": 5.909570315396529e-05, + "loss": 0.4821, + "step": 1025 + }, + { + "epoch": 2.039761431411531, + "grad_norm": 0.3183828091106384, + "learning_rate": 5.904689137206796e-05, + "loss": 0.4801, + "step": 1026 + }, + { + "epoch": 2.0417495029821073, + "grad_norm": 0.3018621992324004, + "learning_rate": 5.8998042882682024e-05, + "loss": 0.4754, + "step": 1027 + }, + { + "epoch": 2.043737574552684, + "grad_norm": 0.23706986324039556, + "learning_rate": 5.8949157779949094e-05, + "loss": 0.4699, + "step": 1028 + }, + { + "epoch": 2.0457256461232602, + "grad_norm": 0.21273568862734663, + "learning_rate": 5.8900236158081373e-05, + "loss": 0.476, + "step": 1029 + }, + { + "epoch": 2.047713717693837, + "grad_norm": 0.31270371437555916, + "learning_rate": 5.8851278111361444e-05, + "loss": 0.4799, + "step": 1030 + }, + { + "epoch": 2.0497017892644136, + "grad_norm": 0.39052899581994754, + "learning_rate": 5.8802283734142074e-05, + "loss": 0.4818, + "step": 1031 + }, + { + "epoch": 2.05168986083499, + "grad_norm": 0.3025793126930875, + "learning_rate": 5.875325312084603e-05, + "loss": 0.467, + "step": 1032 + }, + { + "epoch": 2.0536779324055665, + "grad_norm": 0.22993823523500936, + "learning_rate": 5.870418636596594e-05, + "loss": 0.4769, + "step": 1033 + }, + { + "epoch": 2.055666003976143, + "grad_norm": 0.3656561691652938, + "learning_rate": 5.865508356406408e-05, + "loss": 0.4777, + "step": 1034 + }, + { + "epoch": 2.05765407554672, + "grad_norm": 0.4134849810413476, + "learning_rate": 5.86059448097722e-05, + "loss": 0.4775, + "step": 1035 + }, + { + "epoch": 2.0596421471172963, + "grad_norm": 0.29706263006774, + "learning_rate": 5.8556770197791334e-05, + "loss": 0.4759, + "step": 1036 + }, + { + "epoch": 2.0616302186878728, + "grad_norm": 0.4075218161502847, + "learning_rate": 5.850755982289159e-05, + "loss": 0.4794, + "step": 1037 + }, + { + "epoch": 2.0636182902584492, + "grad_norm": 0.33462269323112476, + "learning_rate": 5.845831377991206e-05, + "loss": 0.4812, + "step": 1038 + }, + { + "epoch": 2.0656063618290257, + "grad_norm": 0.4806307657616057, + "learning_rate": 5.8409032163760536e-05, + "loss": 0.4781, + "step": 1039 + }, + { + "epoch": 2.0675944333996026, + "grad_norm": 0.5577446854340833, + "learning_rate": 5.835971506941338e-05, + "loss": 0.4889, + "step": 1040 + }, + { + "epoch": 2.069582504970179, + "grad_norm": 0.5725622742873799, + "learning_rate": 5.83103625919153e-05, + "loss": 0.4829, + "step": 1041 + }, + { + "epoch": 2.0715705765407555, + "grad_norm": 0.45898725687359726, + "learning_rate": 5.826097482637924e-05, + "loss": 0.4723, + "step": 1042 + }, + { + "epoch": 2.073558648111332, + "grad_norm": 0.2939925335093428, + "learning_rate": 5.8211551867986126e-05, + "loss": 0.4766, + "step": 1043 + }, + { + "epoch": 2.0755467196819084, + "grad_norm": 0.3276561950082936, + "learning_rate": 5.8162093811984714e-05, + "loss": 0.481, + "step": 1044 + }, + { + "epoch": 2.0775347912524853, + "grad_norm": 0.36792672163421464, + "learning_rate": 5.8112600753691404e-05, + "loss": 0.4844, + "step": 1045 + }, + { + "epoch": 2.0795228628230618, + "grad_norm": 0.41267081081703366, + "learning_rate": 5.806307278849003e-05, + "loss": 0.4824, + "step": 1046 + }, + { + "epoch": 2.081510934393638, + "grad_norm": 0.28869818706409717, + "learning_rate": 5.8013510011831726e-05, + "loss": 0.4808, + "step": 1047 + }, + { + "epoch": 2.0834990059642147, + "grad_norm": 0.22647781008653403, + "learning_rate": 5.7963912519234716e-05, + "loss": 0.4671, + "step": 1048 + }, + { + "epoch": 2.085487077534791, + "grad_norm": 0.2643026256357433, + "learning_rate": 5.791428040628413e-05, + "loss": 0.4763, + "step": 1049 + }, + { + "epoch": 2.0874751491053676, + "grad_norm": 0.3184898788545838, + "learning_rate": 5.7864613768631764e-05, + "loss": 0.4704, + "step": 1050 + }, + { + "epoch": 2.0894632206759445, + "grad_norm": 0.3180159256253107, + "learning_rate": 5.781491270199604e-05, + "loss": 0.4766, + "step": 1051 + }, + { + "epoch": 2.091451292246521, + "grad_norm": 0.2508605320554848, + "learning_rate": 5.776517730216166e-05, + "loss": 0.4752, + "step": 1052 + }, + { + "epoch": 2.0934393638170974, + "grad_norm": 0.23712239688416356, + "learning_rate": 5.7715407664979545e-05, + "loss": 0.4774, + "step": 1053 + }, + { + "epoch": 2.095427435387674, + "grad_norm": 0.2527292992526699, + "learning_rate": 5.7665603886366556e-05, + "loss": 0.4758, + "step": 1054 + }, + { + "epoch": 2.0974155069582503, + "grad_norm": 0.27237694330213974, + "learning_rate": 5.761576606230538e-05, + "loss": 0.474, + "step": 1055 + }, + { + "epoch": 2.099403578528827, + "grad_norm": 0.3069196097337193, + "learning_rate": 5.75658942888443e-05, + "loss": 0.4788, + "step": 1056 + }, + { + "epoch": 2.1013916500994037, + "grad_norm": 0.24640187280522557, + "learning_rate": 5.751598866209704e-05, + "loss": 0.4766, + "step": 1057 + }, + { + "epoch": 2.10337972166998, + "grad_norm": 0.3103053687041832, + "learning_rate": 5.746604927824257e-05, + "loss": 0.4706, + "step": 1058 + }, + { + "epoch": 2.1053677932405566, + "grad_norm": 0.40603031427539615, + "learning_rate": 5.741607623352488e-05, + "loss": 0.4736, + "step": 1059 + }, + { + "epoch": 2.107355864811133, + "grad_norm": 0.4925667009481897, + "learning_rate": 5.736606962425288e-05, + "loss": 0.4812, + "step": 1060 + }, + { + "epoch": 2.10934393638171, + "grad_norm": 0.5061455538281211, + "learning_rate": 5.731602954680016e-05, + "loss": 0.4767, + "step": 1061 + }, + { + "epoch": 2.1113320079522864, + "grad_norm": 0.3629127454241806, + "learning_rate": 5.726595609760477e-05, + "loss": 0.4811, + "step": 1062 + }, + { + "epoch": 2.113320079522863, + "grad_norm": 0.26290247683922574, + "learning_rate": 5.721584937316912e-05, + "loss": 0.4827, + "step": 1063 + }, + { + "epoch": 2.1153081510934393, + "grad_norm": 0.2265400929052351, + "learning_rate": 5.716570947005972e-05, + "loss": 0.4734, + "step": 1064 + }, + { + "epoch": 2.1172962226640157, + "grad_norm": 0.28429428323409667, + "learning_rate": 5.7115536484907034e-05, + "loss": 0.4671, + "step": 1065 + }, + { + "epoch": 2.1192842942345926, + "grad_norm": 0.3535690679025093, + "learning_rate": 5.706533051440529e-05, + "loss": 0.4877, + "step": 1066 + }, + { + "epoch": 2.121272365805169, + "grad_norm": 0.4987602517860749, + "learning_rate": 5.701509165531226e-05, + "loss": 0.4828, + "step": 1067 + }, + { + "epoch": 2.1232604373757455, + "grad_norm": 0.44471007932552276, + "learning_rate": 5.696482000444913e-05, + "loss": 0.479, + "step": 1068 + }, + { + "epoch": 2.125248508946322, + "grad_norm": 0.7333534347504769, + "learning_rate": 5.691451565870026e-05, + "loss": 0.4833, + "step": 1069 + }, + { + "epoch": 2.1272365805168985, + "grad_norm": 0.41727447435676135, + "learning_rate": 5.686417871501304e-05, + "loss": 0.4772, + "step": 1070 + }, + { + "epoch": 2.1292246520874754, + "grad_norm": 0.4401221787273442, + "learning_rate": 5.681380927039767e-05, + "loss": 0.4746, + "step": 1071 + }, + { + "epoch": 2.131212723658052, + "grad_norm": 0.5208507936208158, + "learning_rate": 5.676340742192697e-05, + "loss": 0.4704, + "step": 1072 + }, + { + "epoch": 2.1332007952286283, + "grad_norm": 0.4273310201857551, + "learning_rate": 5.671297326673625e-05, + "loss": 0.4735, + "step": 1073 + }, + { + "epoch": 2.1351888667992047, + "grad_norm": 0.32801220821324406, + "learning_rate": 5.666250690202304e-05, + "loss": 0.4749, + "step": 1074 + }, + { + "epoch": 2.137176938369781, + "grad_norm": 0.3278924275873546, + "learning_rate": 5.661200842504698e-05, + "loss": 0.4756, + "step": 1075 + }, + { + "epoch": 2.139165009940358, + "grad_norm": 0.34059455870639094, + "learning_rate": 5.6561477933129605e-05, + "loss": 0.484, + "step": 1076 + }, + { + "epoch": 2.1411530815109345, + "grad_norm": 0.3902473929759777, + "learning_rate": 5.651091552365407e-05, + "loss": 0.4747, + "step": 1077 + }, + { + "epoch": 2.143141153081511, + "grad_norm": 0.45074247141393353, + "learning_rate": 5.646032129406514e-05, + "loss": 0.4738, + "step": 1078 + }, + { + "epoch": 2.1451292246520874, + "grad_norm": 0.5573194267206498, + "learning_rate": 5.640969534186886e-05, + "loss": 0.4848, + "step": 1079 + }, + { + "epoch": 2.147117296222664, + "grad_norm": 0.5620699864501703, + "learning_rate": 5.635903776463242e-05, + "loss": 0.4831, + "step": 1080 + }, + { + "epoch": 2.1491053677932404, + "grad_norm": 0.3560378056402117, + "learning_rate": 5.630834865998391e-05, + "loss": 0.4806, + "step": 1081 + }, + { + "epoch": 2.1510934393638173, + "grad_norm": 0.259702279452807, + "learning_rate": 5.625762812561226e-05, + "loss": 0.473, + "step": 1082 + }, + { + "epoch": 2.1530815109343937, + "grad_norm": 0.37619445513774913, + "learning_rate": 5.620687625926693e-05, + "loss": 0.4815, + "step": 1083 + }, + { + "epoch": 2.15506958250497, + "grad_norm": 0.46465125114676165, + "learning_rate": 5.6156093158757776e-05, + "loss": 0.4686, + "step": 1084 + }, + { + "epoch": 2.1570576540755466, + "grad_norm": 0.49932017268218537, + "learning_rate": 5.6105278921954834e-05, + "loss": 0.4756, + "step": 1085 + }, + { + "epoch": 2.159045725646123, + "grad_norm": 0.4587350037975631, + "learning_rate": 5.605443364678813e-05, + "loss": 0.4805, + "step": 1086 + }, + { + "epoch": 2.1610337972167, + "grad_norm": 0.2846640559400187, + "learning_rate": 5.600355743124756e-05, + "loss": 0.4666, + "step": 1087 + }, + { + "epoch": 2.1630218687872764, + "grad_norm": 0.2747078862140328, + "learning_rate": 5.595265037338262e-05, + "loss": 0.4741, + "step": 1088 + }, + { + "epoch": 2.165009940357853, + "grad_norm": 0.34597139882807243, + "learning_rate": 5.590171257130225e-05, + "loss": 0.4731, + "step": 1089 + }, + { + "epoch": 2.1669980119284293, + "grad_norm": 0.32077150910072605, + "learning_rate": 5.58507441231746e-05, + "loss": 0.4782, + "step": 1090 + }, + { + "epoch": 2.168986083499006, + "grad_norm": 0.325683453143863, + "learning_rate": 5.579974512722695e-05, + "loss": 0.4778, + "step": 1091 + }, + { + "epoch": 2.1709741550695827, + "grad_norm": 0.23366416594281797, + "learning_rate": 5.574871568174542e-05, + "loss": 0.4801, + "step": 1092 + }, + { + "epoch": 2.172962226640159, + "grad_norm": 0.2615690824569999, + "learning_rate": 5.569765588507481e-05, + "loss": 0.478, + "step": 1093 + }, + { + "epoch": 2.1749502982107356, + "grad_norm": 0.24438298458929783, + "learning_rate": 5.5646565835618414e-05, + "loss": 0.4732, + "step": 1094 + }, + { + "epoch": 2.176938369781312, + "grad_norm": 0.25924952804351686, + "learning_rate": 5.5595445631837826e-05, + "loss": 0.485, + "step": 1095 + }, + { + "epoch": 2.1789264413518885, + "grad_norm": 0.26261769213553204, + "learning_rate": 5.554429537225277e-05, + "loss": 0.4675, + "step": 1096 + }, + { + "epoch": 2.1809145129224654, + "grad_norm": 0.25706757072349795, + "learning_rate": 5.549311515544088e-05, + "loss": 0.4755, + "step": 1097 + }, + { + "epoch": 2.182902584493042, + "grad_norm": 0.263310508037158, + "learning_rate": 5.544190508003755e-05, + "loss": 0.479, + "step": 1098 + }, + { + "epoch": 2.1848906560636183, + "grad_norm": 0.282958409875485, + "learning_rate": 5.5390665244735644e-05, + "loss": 0.477, + "step": 1099 + }, + { + "epoch": 2.1868787276341948, + "grad_norm": 0.32353132229505105, + "learning_rate": 5.533939574828548e-05, + "loss": 0.4725, + "step": 1100 + }, + { + "epoch": 2.1888667992047712, + "grad_norm": 0.27942916331621664, + "learning_rate": 5.528809668949448e-05, + "loss": 0.4801, + "step": 1101 + }, + { + "epoch": 2.1908548707753477, + "grad_norm": 0.26257454184150586, + "learning_rate": 5.523676816722705e-05, + "loss": 0.4749, + "step": 1102 + }, + { + "epoch": 2.1928429423459246, + "grad_norm": 0.3092973711478264, + "learning_rate": 5.518541028040437e-05, + "loss": 0.4705, + "step": 1103 + }, + { + "epoch": 2.194831013916501, + "grad_norm": 0.289875960480621, + "learning_rate": 5.5134023128004236e-05, + "loss": 0.4698, + "step": 1104 + }, + { + "epoch": 2.1968190854870775, + "grad_norm": 0.3272973896667466, + "learning_rate": 5.508260680906081e-05, + "loss": 0.476, + "step": 1105 + }, + { + "epoch": 2.198807157057654, + "grad_norm": 0.3043857410112243, + "learning_rate": 5.50311614226645e-05, + "loss": 0.4781, + "step": 1106 + }, + { + "epoch": 2.2007952286282304, + "grad_norm": 0.25728137972199694, + "learning_rate": 5.497968706796173e-05, + "loss": 0.4788, + "step": 1107 + }, + { + "epoch": 2.2027833001988073, + "grad_norm": 0.2496025926087033, + "learning_rate": 5.49281838441547e-05, + "loss": 0.4689, + "step": 1108 + }, + { + "epoch": 2.2047713717693838, + "grad_norm": 0.2552013584123278, + "learning_rate": 5.4876651850501325e-05, + "loss": 0.4721, + "step": 1109 + }, + { + "epoch": 2.20675944333996, + "grad_norm": 0.2508265413893804, + "learning_rate": 5.482509118631491e-05, + "loss": 0.47, + "step": 1110 + }, + { + "epoch": 2.2087475149105367, + "grad_norm": 0.31405764647416207, + "learning_rate": 5.4773501950964055e-05, + "loss": 0.4795, + "step": 1111 + }, + { + "epoch": 2.210735586481113, + "grad_norm": 0.28939539182682655, + "learning_rate": 5.472188424387236e-05, + "loss": 0.4703, + "step": 1112 + }, + { + "epoch": 2.21272365805169, + "grad_norm": 0.21249749698185788, + "learning_rate": 5.467023816451836e-05, + "loss": 0.4804, + "step": 1113 + }, + { + "epoch": 2.2147117296222665, + "grad_norm": 0.3614011385798107, + "learning_rate": 5.461856381243524e-05, + "loss": 0.4747, + "step": 1114 + }, + { + "epoch": 2.216699801192843, + "grad_norm": 0.21423784260449702, + "learning_rate": 5.456686128721069e-05, + "loss": 0.4725, + "step": 1115 + }, + { + "epoch": 2.2186878727634194, + "grad_norm": 0.3235629875050971, + "learning_rate": 5.451513068848668e-05, + "loss": 0.4772, + "step": 1116 + }, + { + "epoch": 2.220675944333996, + "grad_norm": 0.25439106188409116, + "learning_rate": 5.4463372115959266e-05, + "loss": 0.4726, + "step": 1117 + }, + { + "epoch": 2.2226640159045727, + "grad_norm": 0.2590887452156136, + "learning_rate": 5.4411585669378464e-05, + "loss": 0.4732, + "step": 1118 + }, + { + "epoch": 2.224652087475149, + "grad_norm": 0.26919093398382715, + "learning_rate": 5.4359771448547966e-05, + "loss": 0.4728, + "step": 1119 + }, + { + "epoch": 2.2266401590457257, + "grad_norm": 0.36365985427405484, + "learning_rate": 5.430792955332503e-05, + "loss": 0.4732, + "step": 1120 + }, + { + "epoch": 2.228628230616302, + "grad_norm": 0.3915941810323402, + "learning_rate": 5.425606008362019e-05, + "loss": 0.4781, + "step": 1121 + }, + { + "epoch": 2.2306163021868786, + "grad_norm": 0.2610839854362163, + "learning_rate": 5.420416313939719e-05, + "loss": 0.4831, + "step": 1122 + }, + { + "epoch": 2.2326043737574555, + "grad_norm": 0.4388412247539103, + "learning_rate": 5.415223882067269e-05, + "loss": 0.4973, + "step": 1123 + }, + { + "epoch": 2.234592445328032, + "grad_norm": 0.2698215651291218, + "learning_rate": 5.410028722751609e-05, + "loss": 0.476, + "step": 1124 + }, + { + "epoch": 2.2365805168986084, + "grad_norm": 0.26389275172674886, + "learning_rate": 5.404830846004938e-05, + "loss": 0.4703, + "step": 1125 + }, + { + "epoch": 2.238568588469185, + "grad_norm": 0.3120918809111035, + "learning_rate": 5.3996302618446906e-05, + "loss": 0.4769, + "step": 1126 + }, + { + "epoch": 2.2405566600397613, + "grad_norm": 0.32108060374276975, + "learning_rate": 5.394426980293522e-05, + "loss": 0.4669, + "step": 1127 + }, + { + "epoch": 2.242544731610338, + "grad_norm": 0.36707667301171554, + "learning_rate": 5.389221011379281e-05, + "loss": 0.4715, + "step": 1128 + }, + { + "epoch": 2.2445328031809146, + "grad_norm": 0.3816486880160509, + "learning_rate": 5.384012365135002e-05, + "loss": 0.4767, + "step": 1129 + }, + { + "epoch": 2.246520874751491, + "grad_norm": 0.46296188269694993, + "learning_rate": 5.3788010515988695e-05, + "loss": 0.4702, + "step": 1130 + }, + { + "epoch": 2.2485089463220675, + "grad_norm": 0.4896953557986629, + "learning_rate": 5.37358708081422e-05, + "loss": 0.4685, + "step": 1131 + }, + { + "epoch": 2.250497017892644, + "grad_norm": 0.43869657749170266, + "learning_rate": 5.368370462829503e-05, + "loss": 0.465, + "step": 1132 + }, + { + "epoch": 2.252485089463221, + "grad_norm": 0.5371500014095143, + "learning_rate": 5.363151207698273e-05, + "loss": 0.4706, + "step": 1133 + }, + { + "epoch": 2.2544731610337974, + "grad_norm": 0.5863174393137796, + "learning_rate": 5.357929325479164e-05, + "loss": 0.4746, + "step": 1134 + }, + { + "epoch": 2.256461232604374, + "grad_norm": 0.39887424277932715, + "learning_rate": 5.3527048262358774e-05, + "loss": 0.4712, + "step": 1135 + }, + { + "epoch": 2.2584493041749503, + "grad_norm": 0.30753503519660036, + "learning_rate": 5.347477720037156e-05, + "loss": 0.4745, + "step": 1136 + }, + { + "epoch": 2.2604373757455267, + "grad_norm": 0.45617523478874633, + "learning_rate": 5.342248016956765e-05, + "loss": 0.4696, + "step": 1137 + }, + { + "epoch": 2.2624254473161036, + "grad_norm": 0.5535883206642056, + "learning_rate": 5.3370157270734764e-05, + "loss": 0.4705, + "step": 1138 + }, + { + "epoch": 2.26441351888668, + "grad_norm": 0.4652282323599418, + "learning_rate": 5.331780860471046e-05, + "loss": 0.4743, + "step": 1139 + }, + { + "epoch": 2.2664015904572565, + "grad_norm": 0.36379549274018436, + "learning_rate": 5.326543427238199e-05, + "loss": 0.4726, + "step": 1140 + }, + { + "epoch": 2.268389662027833, + "grad_norm": 0.3171847241802109, + "learning_rate": 5.321303437468603e-05, + "loss": 0.4655, + "step": 1141 + }, + { + "epoch": 2.2703777335984094, + "grad_norm": 0.2881646521576697, + "learning_rate": 5.3160609012608545e-05, + "loss": 0.4633, + "step": 1142 + }, + { + "epoch": 2.272365805168986, + "grad_norm": 0.29641268848872576, + "learning_rate": 5.310815828718457e-05, + "loss": 0.4708, + "step": 1143 + }, + { + "epoch": 2.274353876739563, + "grad_norm": 0.40275523394654794, + "learning_rate": 5.305568229949802e-05, + "loss": 0.4673, + "step": 1144 + }, + { + "epoch": 2.2763419483101393, + "grad_norm": 0.35883228801414907, + "learning_rate": 5.3003181150681506e-05, + "loss": 0.4712, + "step": 1145 + }, + { + "epoch": 2.2783300198807157, + "grad_norm": 0.2687510267313907, + "learning_rate": 5.295065494191612e-05, + "loss": 0.468, + "step": 1146 + }, + { + "epoch": 2.280318091451292, + "grad_norm": 0.2892946413120926, + "learning_rate": 5.289810377443127e-05, + "loss": 0.473, + "step": 1147 + }, + { + "epoch": 2.2823061630218686, + "grad_norm": 0.2943293689076449, + "learning_rate": 5.284552774950441e-05, + "loss": 0.4711, + "step": 1148 + }, + { + "epoch": 2.2842942345924455, + "grad_norm": 0.2434140998108816, + "learning_rate": 5.279292696846096e-05, + "loss": 0.4687, + "step": 1149 + }, + { + "epoch": 2.286282306163022, + "grad_norm": 0.22090774927042578, + "learning_rate": 5.274030153267403e-05, + "loss": 0.4756, + "step": 1150 + }, + { + "epoch": 2.2882703777335984, + "grad_norm": 0.2475754664083401, + "learning_rate": 5.268765154356425e-05, + "loss": 0.4748, + "step": 1151 + }, + { + "epoch": 2.290258449304175, + "grad_norm": 0.24994849706871775, + "learning_rate": 5.263497710259953e-05, + "loss": 0.4737, + "step": 1152 + }, + { + "epoch": 2.2922465208747513, + "grad_norm": 0.2215857652666882, + "learning_rate": 5.2582278311294975e-05, + "loss": 0.4639, + "step": 1153 + }, + { + "epoch": 2.294234592445328, + "grad_norm": 0.2170348912725335, + "learning_rate": 5.252955527121255e-05, + "loss": 0.4733, + "step": 1154 + }, + { + "epoch": 2.2962226640159047, + "grad_norm": 0.21834748510730956, + "learning_rate": 5.2476808083961e-05, + "loss": 0.4635, + "step": 1155 + }, + { + "epoch": 2.298210735586481, + "grad_norm": 0.2351001402329863, + "learning_rate": 5.242403685119559e-05, + "loss": 0.4729, + "step": 1156 + }, + { + "epoch": 2.3001988071570576, + "grad_norm": 0.5861647894402029, + "learning_rate": 5.2371241674617924e-05, + "loss": 0.4885, + "step": 1157 + }, + { + "epoch": 2.302186878727634, + "grad_norm": 0.26021493801320544, + "learning_rate": 5.231842265597573e-05, + "loss": 0.4698, + "step": 1158 + }, + { + "epoch": 2.3041749502982105, + "grad_norm": 0.33674055818667714, + "learning_rate": 5.2265579897062724e-05, + "loss": 0.4695, + "step": 1159 + }, + { + "epoch": 2.3061630218687874, + "grad_norm": 0.30732795040623506, + "learning_rate": 5.2212713499718374e-05, + "loss": 0.4682, + "step": 1160 + }, + { + "epoch": 2.308151093439364, + "grad_norm": 0.2351863418623793, + "learning_rate": 5.2159823565827656e-05, + "loss": 0.4692, + "step": 1161 + }, + { + "epoch": 2.3101391650099403, + "grad_norm": 0.24745733032782763, + "learning_rate": 5.210691019732096e-05, + "loss": 0.4676, + "step": 1162 + }, + { + "epoch": 2.3121272365805168, + "grad_norm": 0.21317022863905552, + "learning_rate": 5.2053973496173814e-05, + "loss": 0.4708, + "step": 1163 + }, + { + "epoch": 2.3141153081510932, + "grad_norm": 0.27443341498457285, + "learning_rate": 5.2001013564406704e-05, + "loss": 0.4704, + "step": 1164 + }, + { + "epoch": 2.31610337972167, + "grad_norm": 0.2621084832032726, + "learning_rate": 5.194803050408491e-05, + "loss": 0.4673, + "step": 1165 + }, + { + "epoch": 2.3180914512922466, + "grad_norm": 0.23873064722197226, + "learning_rate": 5.1895024417318275e-05, + "loss": 0.4702, + "step": 1166 + }, + { + "epoch": 2.320079522862823, + "grad_norm": 0.2860158902919176, + "learning_rate": 5.184199540626102e-05, + "loss": 0.4741, + "step": 1167 + }, + { + "epoch": 2.3220675944333995, + "grad_norm": 0.3120161852959375, + "learning_rate": 5.178894357311155e-05, + "loss": 0.4651, + "step": 1168 + }, + { + "epoch": 2.324055666003976, + "grad_norm": 0.2686323464939365, + "learning_rate": 5.173586902011224e-05, + "loss": 0.4694, + "step": 1169 + }, + { + "epoch": 2.326043737574553, + "grad_norm": 0.25928705925335244, + "learning_rate": 5.168277184954924e-05, + "loss": 0.4615, + "step": 1170 + }, + { + "epoch": 2.3280318091451293, + "grad_norm": 0.2401126856878729, + "learning_rate": 5.1629652163752326e-05, + "loss": 0.4675, + "step": 1171 + }, + { + "epoch": 2.3300198807157058, + "grad_norm": 0.23622257410914968, + "learning_rate": 5.157651006509467e-05, + "loss": 0.4738, + "step": 1172 + }, + { + "epoch": 2.332007952286282, + "grad_norm": 0.2398042204246071, + "learning_rate": 5.152334565599257e-05, + "loss": 0.4685, + "step": 1173 + }, + { + "epoch": 2.3339960238568587, + "grad_norm": 0.3099714198955679, + "learning_rate": 5.147015903890541e-05, + "loss": 0.4734, + "step": 1174 + }, + { + "epoch": 2.3359840954274356, + "grad_norm": 0.2410518103522495, + "learning_rate": 5.1416950316335305e-05, + "loss": 0.4692, + "step": 1175 + }, + { + "epoch": 2.337972166998012, + "grad_norm": 0.22715511588645854, + "learning_rate": 5.1363719590827005e-05, + "loss": 0.4723, + "step": 1176 + }, + { + "epoch": 2.3399602385685885, + "grad_norm": 0.23992255270613838, + "learning_rate": 5.131046696496766e-05, + "loss": 0.4682, + "step": 1177 + }, + { + "epoch": 2.341948310139165, + "grad_norm": 0.2148629396505402, + "learning_rate": 5.1257192541386617e-05, + "loss": 0.4715, + "step": 1178 + }, + { + "epoch": 2.3439363817097414, + "grad_norm": 0.21680845387800704, + "learning_rate": 5.120389642275525e-05, + "loss": 0.4668, + "step": 1179 + }, + { + "epoch": 2.3459244532803183, + "grad_norm": 0.24226307278763232, + "learning_rate": 5.115057871178672e-05, + "loss": 0.4618, + "step": 1180 + }, + { + "epoch": 2.3479125248508947, + "grad_norm": 0.2861167812722164, + "learning_rate": 5.1097239511235825e-05, + "loss": 0.4723, + "step": 1181 + }, + { + "epoch": 2.349900596421471, + "grad_norm": 0.27609528840749975, + "learning_rate": 5.104387892389877e-05, + "loss": 0.467, + "step": 1182 + }, + { + "epoch": 2.3518886679920477, + "grad_norm": 0.38245826721422277, + "learning_rate": 5.099049705261294e-05, + "loss": 0.4755, + "step": 1183 + }, + { + "epoch": 2.353876739562624, + "grad_norm": 0.29912531232454764, + "learning_rate": 5.093709400025682e-05, + "loss": 0.4721, + "step": 1184 + }, + { + "epoch": 2.355864811133201, + "grad_norm": 0.2502048153204002, + "learning_rate": 5.088366986974965e-05, + "loss": 0.4667, + "step": 1185 + }, + { + "epoch": 2.3578528827037775, + "grad_norm": 0.2203348693194436, + "learning_rate": 5.0830224764051304e-05, + "loss": 0.4663, + "step": 1186 + }, + { + "epoch": 2.359840954274354, + "grad_norm": 0.3196618924057666, + "learning_rate": 5.0776758786162087e-05, + "loss": 0.4702, + "step": 1187 + }, + { + "epoch": 2.3618290258449304, + "grad_norm": 0.2756931543786987, + "learning_rate": 5.072327203912254e-05, + "loss": 0.4605, + "step": 1188 + }, + { + "epoch": 2.363817097415507, + "grad_norm": 0.20491551253798782, + "learning_rate": 5.066976462601321e-05, + "loss": 0.4665, + "step": 1189 + }, + { + "epoch": 2.3658051689860837, + "grad_norm": 0.2784356458213145, + "learning_rate": 5.0616236649954486e-05, + "loss": 0.4707, + "step": 1190 + }, + { + "epoch": 2.36779324055666, + "grad_norm": 0.3183004798915327, + "learning_rate": 5.056268821410638e-05, + "loss": 0.4727, + "step": 1191 + }, + { + "epoch": 2.3697813121272366, + "grad_norm": 0.42961542435749583, + "learning_rate": 5.0509119421668315e-05, + "loss": 0.4751, + "step": 1192 + }, + { + "epoch": 2.371769383697813, + "grad_norm": 0.3334202159622228, + "learning_rate": 5.0455530375879e-05, + "loss": 0.4743, + "step": 1193 + }, + { + "epoch": 2.3737574552683895, + "grad_norm": 0.41604702703185387, + "learning_rate": 5.040192118001614e-05, + "loss": 0.4687, + "step": 1194 + }, + { + "epoch": 2.3757455268389664, + "grad_norm": 0.4822424655743456, + "learning_rate": 5.034829193739625e-05, + "loss": 0.4729, + "step": 1195 + }, + { + "epoch": 2.377733598409543, + "grad_norm": 0.3387106786694204, + "learning_rate": 5.029464275137453e-05, + "loss": 0.4665, + "step": 1196 + }, + { + "epoch": 2.3797216699801194, + "grad_norm": 0.3595173653277191, + "learning_rate": 5.0240973725344584e-05, + "loss": 0.4794, + "step": 1197 + }, + { + "epoch": 2.381709741550696, + "grad_norm": 0.46424663733541854, + "learning_rate": 5.0187284962738254e-05, + "loss": 0.4742, + "step": 1198 + }, + { + "epoch": 2.3836978131212723, + "grad_norm": 0.3198119518435221, + "learning_rate": 5.0133576567025436e-05, + "loss": 0.479, + "step": 1199 + }, + { + "epoch": 2.3856858846918487, + "grad_norm": 0.32264968680235223, + "learning_rate": 5.007984864171384e-05, + "loss": 0.4755, + "step": 1200 + }, + { + "epoch": 2.3876739562624256, + "grad_norm": 0.2680481041894317, + "learning_rate": 5.002610129034883e-05, + "loss": 0.4682, + "step": 1201 + }, + { + "epoch": 2.389662027833002, + "grad_norm": 0.24662752780204666, + "learning_rate": 4.99723346165132e-05, + "loss": 0.4733, + "step": 1202 + }, + { + "epoch": 2.3916500994035785, + "grad_norm": 0.2585957373284863, + "learning_rate": 4.991854872382699e-05, + "loss": 0.4779, + "step": 1203 + }, + { + "epoch": 2.393638170974155, + "grad_norm": 0.23387709556355635, + "learning_rate": 4.9864743715947275e-05, + "loss": 0.4764, + "step": 1204 + }, + { + "epoch": 2.3956262425447314, + "grad_norm": 0.2909207043279319, + "learning_rate": 4.9810919696567966e-05, + "loss": 0.4763, + "step": 1205 + }, + { + "epoch": 2.3976143141153083, + "grad_norm": 0.24287573819888061, + "learning_rate": 4.9757076769419616e-05, + "loss": 0.474, + "step": 1206 + }, + { + "epoch": 2.399602385685885, + "grad_norm": 0.26216025207203086, + "learning_rate": 4.970321503826922e-05, + "loss": 0.4804, + "step": 1207 + }, + { + "epoch": 2.4015904572564613, + "grad_norm": 0.2677040788546583, + "learning_rate": 4.964933460692001e-05, + "loss": 0.4783, + "step": 1208 + }, + { + "epoch": 2.4035785288270377, + "grad_norm": 0.2552777608049991, + "learning_rate": 4.959543557921126e-05, + "loss": 0.473, + "step": 1209 + }, + { + "epoch": 2.405566600397614, + "grad_norm": 0.22672195489552985, + "learning_rate": 4.9541518059018073e-05, + "loss": 0.4702, + "step": 1210 + }, + { + "epoch": 2.4075546719681906, + "grad_norm": 0.2515384275613973, + "learning_rate": 4.94875821502512e-05, + "loss": 0.466, + "step": 1211 + }, + { + "epoch": 2.4095427435387675, + "grad_norm": 0.24081500343214746, + "learning_rate": 4.9433627956856835e-05, + "loss": 0.4731, + "step": 1212 + }, + { + "epoch": 2.411530815109344, + "grad_norm": 0.23961419355896613, + "learning_rate": 4.9379655582816396e-05, + "loss": 0.4644, + "step": 1213 + }, + { + "epoch": 2.4135188866799204, + "grad_norm": 0.2378021223509304, + "learning_rate": 4.9325665132146354e-05, + "loss": 0.4659, + "step": 1214 + }, + { + "epoch": 2.415506958250497, + "grad_norm": 0.2803020254737336, + "learning_rate": 4.927165670889801e-05, + "loss": 0.4731, + "step": 1215 + }, + { + "epoch": 2.4174950298210733, + "grad_norm": 0.31478058063257236, + "learning_rate": 4.9217630417157306e-05, + "loss": 0.4691, + "step": 1216 + }, + { + "epoch": 2.4194831013916502, + "grad_norm": 0.27040398082481854, + "learning_rate": 4.9163586361044615e-05, + "loss": 0.4739, + "step": 1217 + }, + { + "epoch": 2.4214711729622267, + "grad_norm": 0.22978169619172967, + "learning_rate": 4.910952464471455e-05, + "loss": 0.4738, + "step": 1218 + }, + { + "epoch": 2.423459244532803, + "grad_norm": 0.3003478653396795, + "learning_rate": 4.905544537235575e-05, + "loss": 0.4768, + "step": 1219 + }, + { + "epoch": 2.4254473161033796, + "grad_norm": 0.3230073628448877, + "learning_rate": 4.9001348648190715e-05, + "loss": 0.4724, + "step": 1220 + }, + { + "epoch": 2.427435387673956, + "grad_norm": 0.38849544363848615, + "learning_rate": 4.894723457647554e-05, + "loss": 0.4813, + "step": 1221 + }, + { + "epoch": 2.429423459244533, + "grad_norm": 0.355162411969774, + "learning_rate": 4.889310326149979e-05, + "loss": 0.4696, + "step": 1222 + }, + { + "epoch": 2.4314115308151094, + "grad_norm": 0.3414594723704106, + "learning_rate": 4.883895480758621e-05, + "loss": 0.4706, + "step": 1223 + }, + { + "epoch": 2.433399602385686, + "grad_norm": 0.30815325615464595, + "learning_rate": 4.8784789319090644e-05, + "loss": 0.4667, + "step": 1224 + }, + { + "epoch": 2.4353876739562623, + "grad_norm": 0.341796797546769, + "learning_rate": 4.8730606900401715e-05, + "loss": 0.4756, + "step": 1225 + }, + { + "epoch": 2.4373757455268388, + "grad_norm": 0.47306285227016065, + "learning_rate": 4.8676407655940694e-05, + "loss": 0.4726, + "step": 1226 + }, + { + "epoch": 2.4393638170974157, + "grad_norm": 0.3497635037514457, + "learning_rate": 4.862219169016127e-05, + "loss": 0.4709, + "step": 1227 + }, + { + "epoch": 2.441351888667992, + "grad_norm": 0.4504529577189983, + "learning_rate": 4.856795910754937e-05, + "loss": 0.4654, + "step": 1228 + }, + { + "epoch": 2.4433399602385686, + "grad_norm": 0.34689558387001274, + "learning_rate": 4.851371001262292e-05, + "loss": 0.4717, + "step": 1229 + }, + { + "epoch": 2.445328031809145, + "grad_norm": 0.2872733269653433, + "learning_rate": 4.84594445099317e-05, + "loss": 0.4733, + "step": 1230 + }, + { + "epoch": 2.4473161033797215, + "grad_norm": 0.27243647138976984, + "learning_rate": 4.840516270405712e-05, + "loss": 0.4742, + "step": 1231 + }, + { + "epoch": 2.4493041749502984, + "grad_norm": 0.30305149495329026, + "learning_rate": 4.8350864699611924e-05, + "loss": 0.4626, + "step": 1232 + }, + { + "epoch": 2.451292246520875, + "grad_norm": 0.27978195741025513, + "learning_rate": 4.8296550601240186e-05, + "loss": 0.4753, + "step": 1233 + }, + { + "epoch": 2.4532803180914513, + "grad_norm": 0.29267098699687766, + "learning_rate": 4.8242220513616945e-05, + "loss": 0.471, + "step": 1234 + }, + { + "epoch": 2.4552683896620278, + "grad_norm": 0.23437662644156787, + "learning_rate": 4.8187874541448046e-05, + "loss": 0.4688, + "step": 1235 + }, + { + "epoch": 2.457256461232604, + "grad_norm": 0.23103521775278718, + "learning_rate": 4.8133512789469945e-05, + "loss": 0.4823, + "step": 1236 + }, + { + "epoch": 2.459244532803181, + "grad_norm": 0.20865348483141843, + "learning_rate": 4.807913536244955e-05, + "loss": 0.4743, + "step": 1237 + }, + { + "epoch": 2.4612326043737576, + "grad_norm": 0.3100863032015534, + "learning_rate": 4.8024742365183946e-05, + "loss": 0.4706, + "step": 1238 + }, + { + "epoch": 2.463220675944334, + "grad_norm": 0.27950183594554934, + "learning_rate": 4.797033390250021e-05, + "loss": 0.4634, + "step": 1239 + }, + { + "epoch": 2.4652087475149105, + "grad_norm": 0.25304117732962556, + "learning_rate": 4.791591007925526e-05, + "loss": 0.4638, + "step": 1240 + }, + { + "epoch": 2.467196819085487, + "grad_norm": 0.28945569502035434, + "learning_rate": 4.7861471000335604e-05, + "loss": 0.4744, + "step": 1241 + }, + { + "epoch": 2.469184890656064, + "grad_norm": 0.23301689545756774, + "learning_rate": 4.780701677065714e-05, + "loss": 0.4689, + "step": 1242 + }, + { + "epoch": 2.4711729622266403, + "grad_norm": 0.2884587227277944, + "learning_rate": 4.775254749516497e-05, + "loss": 0.4713, + "step": 1243 + }, + { + "epoch": 2.4731610337972167, + "grad_norm": 0.31338204568964245, + "learning_rate": 4.76980632788332e-05, + "loss": 0.4797, + "step": 1244 + }, + { + "epoch": 2.475149105367793, + "grad_norm": 0.29265797464452964, + "learning_rate": 4.764356422666473e-05, + "loss": 0.4726, + "step": 1245 + }, + { + "epoch": 2.4771371769383697, + "grad_norm": 0.23282823698357283, + "learning_rate": 4.758905044369105e-05, + "loss": 0.4671, + "step": 1246 + }, + { + "epoch": 2.4791252485089466, + "grad_norm": 0.2929501814614166, + "learning_rate": 4.753452203497202e-05, + "loss": 0.473, + "step": 1247 + }, + { + "epoch": 2.481113320079523, + "grad_norm": 0.297087580600837, + "learning_rate": 4.747997910559573e-05, + "loss": 0.4698, + "step": 1248 + }, + { + "epoch": 2.4831013916500995, + "grad_norm": 0.83213556842206, + "learning_rate": 4.7425421760678195e-05, + "loss": 0.4798, + "step": 1249 + }, + { + "epoch": 2.485089463220676, + "grad_norm": 0.4143963209466649, + "learning_rate": 4.737085010536328e-05, + "loss": 0.47, + "step": 1250 + }, + { + "epoch": 2.4870775347912524, + "grad_norm": 0.3331788869276754, + "learning_rate": 4.731626424482238e-05, + "loss": 0.473, + "step": 1251 + }, + { + "epoch": 2.4890656063618293, + "grad_norm": 0.29929653590699296, + "learning_rate": 4.726166428425427e-05, + "loss": 0.4771, + "step": 1252 + }, + { + "epoch": 2.4910536779324057, + "grad_norm": 0.34503964283270694, + "learning_rate": 4.720705032888492e-05, + "loss": 0.4672, + "step": 1253 + }, + { + "epoch": 2.493041749502982, + "grad_norm": 0.3732218590332561, + "learning_rate": 4.715242248396728e-05, + "loss": 0.471, + "step": 1254 + }, + { + "epoch": 2.4950298210735586, + "grad_norm": 0.3583811530564549, + "learning_rate": 4.709778085478101e-05, + "loss": 0.4761, + "step": 1255 + }, + { + "epoch": 2.497017892644135, + "grad_norm": 0.3230101452445875, + "learning_rate": 4.70431255466324e-05, + "loss": 0.4614, + "step": 1256 + }, + { + "epoch": 2.4990059642147116, + "grad_norm": 0.28794588330894255, + "learning_rate": 4.698845666485408e-05, + "loss": 0.4699, + "step": 1257 + }, + { + "epoch": 2.500994035785288, + "grad_norm": 0.3780763497100984, + "learning_rate": 4.693377431480482e-05, + "loss": 0.4694, + "step": 1258 + }, + { + "epoch": 2.502982107355865, + "grad_norm": 0.45136607030451903, + "learning_rate": 4.6879078601869374e-05, + "loss": 0.4747, + "step": 1259 + }, + { + "epoch": 2.5049701789264414, + "grad_norm": 0.24953986065483028, + "learning_rate": 4.682436963145823e-05, + "loss": 0.4685, + "step": 1260 + }, + { + "epoch": 2.506958250497018, + "grad_norm": 0.336007300317526, + "learning_rate": 4.6769647509007443e-05, + "loss": 0.4738, + "step": 1261 + }, + { + "epoch": 2.5089463220675943, + "grad_norm": 0.33628369876467346, + "learning_rate": 4.671491233997841e-05, + "loss": 0.468, + "step": 1262 + }, + { + "epoch": 2.5109343936381707, + "grad_norm": 0.2711421941817858, + "learning_rate": 4.666016422985764e-05, + "loss": 0.4701, + "step": 1263 + }, + { + "epoch": 2.5129224652087476, + "grad_norm": 0.34640170723170177, + "learning_rate": 4.660540328415663e-05, + "loss": 0.4595, + "step": 1264 + }, + { + "epoch": 2.514910536779324, + "grad_norm": 0.3054801483553342, + "learning_rate": 4.6550629608411596e-05, + "loss": 0.4558, + "step": 1265 + }, + { + "epoch": 2.5168986083499005, + "grad_norm": 0.2278535685525667, + "learning_rate": 4.6495843308183275e-05, + "loss": 0.4648, + "step": 1266 + }, + { + "epoch": 2.518886679920477, + "grad_norm": 0.2947939908046167, + "learning_rate": 4.644104448905675e-05, + "loss": 0.4698, + "step": 1267 + }, + { + "epoch": 2.5208747514910534, + "grad_norm": 0.43545344724955837, + "learning_rate": 4.638623325664123e-05, + "loss": 0.4786, + "step": 1268 + }, + { + "epoch": 2.5228628230616303, + "grad_norm": 0.45373043954413755, + "learning_rate": 4.633140971656983e-05, + "loss": 0.4713, + "step": 1269 + }, + { + "epoch": 2.524850894632207, + "grad_norm": 0.33146377812850386, + "learning_rate": 4.6276573974499394e-05, + "loss": 0.4684, + "step": 1270 + }, + { + "epoch": 2.5268389662027833, + "grad_norm": 0.2680817520803437, + "learning_rate": 4.622172613611029e-05, + "loss": 0.4691, + "step": 1271 + }, + { + "epoch": 2.5288270377733597, + "grad_norm": 0.29840968057455064, + "learning_rate": 4.6166866307106195e-05, + "loss": 0.4681, + "step": 1272 + }, + { + "epoch": 2.530815109343936, + "grad_norm": 0.2928890490184993, + "learning_rate": 4.611199459321389e-05, + "loss": 0.4788, + "step": 1273 + }, + { + "epoch": 2.532803180914513, + "grad_norm": 0.24379232071112766, + "learning_rate": 4.605711110018307e-05, + "loss": 0.4738, + "step": 1274 + }, + { + "epoch": 2.5347912524850895, + "grad_norm": 0.22108967349738823, + "learning_rate": 4.600221593378611e-05, + "loss": 0.4688, + "step": 1275 + }, + { + "epoch": 2.536779324055666, + "grad_norm": 0.26959273399137357, + "learning_rate": 4.594730919981791e-05, + "loss": 0.4573, + "step": 1276 + }, + { + "epoch": 2.5387673956262424, + "grad_norm": 0.2989159213403466, + "learning_rate": 4.589239100409563e-05, + "loss": 0.465, + "step": 1277 + }, + { + "epoch": 2.540755467196819, + "grad_norm": 0.21391977504533247, + "learning_rate": 4.583746145245854e-05, + "loss": 0.467, + "step": 1278 + }, + { + "epoch": 2.542743538767396, + "grad_norm": 0.29971289779725213, + "learning_rate": 4.578252065076782e-05, + "loss": 0.4614, + "step": 1279 + }, + { + "epoch": 2.5447316103379722, + "grad_norm": 0.36235889419318323, + "learning_rate": 4.5727568704906285e-05, + "loss": 0.4709, + "step": 1280 + }, + { + "epoch": 2.5467196819085487, + "grad_norm": 0.3184243494913832, + "learning_rate": 4.5672605720778235e-05, + "loss": 0.4702, + "step": 1281 + }, + { + "epoch": 2.548707753479125, + "grad_norm": 0.3367075112146258, + "learning_rate": 4.561763180430928e-05, + "loss": 0.4681, + "step": 1282 + }, + { + "epoch": 2.5506958250497016, + "grad_norm": 0.38423599258480656, + "learning_rate": 4.5562647061446065e-05, + "loss": 0.4658, + "step": 1283 + }, + { + "epoch": 2.5526838966202785, + "grad_norm": 0.30156605639178163, + "learning_rate": 4.5507651598156104e-05, + "loss": 0.4715, + "step": 1284 + }, + { + "epoch": 2.554671968190855, + "grad_norm": 0.2539054825100618, + "learning_rate": 4.545264552042758e-05, + "loss": 0.4718, + "step": 1285 + }, + { + "epoch": 2.5566600397614314, + "grad_norm": 0.2832379512707468, + "learning_rate": 4.539762893426913e-05, + "loss": 0.4705, + "step": 1286 + }, + { + "epoch": 2.558648111332008, + "grad_norm": 0.2729857124864947, + "learning_rate": 4.534260194570964e-05, + "loss": 0.4629, + "step": 1287 + }, + { + "epoch": 2.5606361829025843, + "grad_norm": 0.9159869395713054, + "learning_rate": 4.528756466079804e-05, + "loss": 0.4611, + "step": 1288 + }, + { + "epoch": 2.5626242544731612, + "grad_norm": 0.2947830909553326, + "learning_rate": 4.5232517185603124e-05, + "loss": 0.4625, + "step": 1289 + }, + { + "epoch": 2.5646123260437377, + "grad_norm": 0.30617876929071663, + "learning_rate": 4.5177459626213294e-05, + "loss": 0.4575, + "step": 1290 + }, + { + "epoch": 2.566600397614314, + "grad_norm": 0.3906173771205208, + "learning_rate": 4.51223920887364e-05, + "loss": 0.4697, + "step": 1291 + }, + { + "epoch": 2.5685884691848906, + "grad_norm": 0.5107505465341972, + "learning_rate": 4.506731467929955e-05, + "loss": 0.4843, + "step": 1292 + }, + { + "epoch": 2.570576540755467, + "grad_norm": 0.26067879748684747, + "learning_rate": 4.501222750404883e-05, + "loss": 0.4687, + "step": 1293 + }, + { + "epoch": 2.572564612326044, + "grad_norm": 0.27991711122868407, + "learning_rate": 4.495713066914917e-05, + "loss": 0.483, + "step": 1294 + }, + { + "epoch": 2.5745526838966204, + "grad_norm": 0.3234266564676175, + "learning_rate": 4.4902024280784114e-05, + "loss": 0.4634, + "step": 1295 + }, + { + "epoch": 2.576540755467197, + "grad_norm": 0.29382217694112184, + "learning_rate": 4.4846908445155625e-05, + "loss": 0.4622, + "step": 1296 + }, + { + "epoch": 2.5785288270377733, + "grad_norm": 0.4240599462547126, + "learning_rate": 4.4791783268483864e-05, + "loss": 0.4707, + "step": 1297 + }, + { + "epoch": 2.5805168986083498, + "grad_norm": 0.29831663140195225, + "learning_rate": 4.473664885700699e-05, + "loss": 0.473, + "step": 1298 + }, + { + "epoch": 2.5825049701789267, + "grad_norm": 0.2867259906466539, + "learning_rate": 4.468150531698098e-05, + "loss": 0.474, + "step": 1299 + }, + { + "epoch": 2.584493041749503, + "grad_norm": 0.2380064860884348, + "learning_rate": 4.462635275467937e-05, + "loss": 0.4718, + "step": 1300 + }, + { + "epoch": 2.5864811133200796, + "grad_norm": 0.2817938743929372, + "learning_rate": 4.4571191276393125e-05, + "loss": 0.4721, + "step": 1301 + }, + { + "epoch": 2.588469184890656, + "grad_norm": 0.2792752528880275, + "learning_rate": 4.4516020988430346e-05, + "loss": 0.4677, + "step": 1302 + }, + { + "epoch": 2.5904572564612325, + "grad_norm": 0.27683589472201287, + "learning_rate": 4.4460841997116166e-05, + "loss": 0.4701, + "step": 1303 + }, + { + "epoch": 2.5924453280318094, + "grad_norm": 0.31437221453408365, + "learning_rate": 4.440565440879244e-05, + "loss": 0.4684, + "step": 1304 + }, + { + "epoch": 2.594433399602386, + "grad_norm": 0.30714606345078876, + "learning_rate": 4.435045832981763e-05, + "loss": 0.4642, + "step": 1305 + }, + { + "epoch": 2.5964214711729623, + "grad_norm": 0.2974009434374631, + "learning_rate": 4.429525386656653e-05, + "loss": 0.4744, + "step": 1306 + }, + { + "epoch": 2.5984095427435387, + "grad_norm": 0.3586097986600065, + "learning_rate": 4.424004112543012e-05, + "loss": 0.468, + "step": 1307 + }, + { + "epoch": 2.600397614314115, + "grad_norm": 0.43266237162260085, + "learning_rate": 4.4184820212815306e-05, + "loss": 0.4655, + "step": 1308 + }, + { + "epoch": 2.602385685884692, + "grad_norm": 0.3732894233625949, + "learning_rate": 4.4129591235144776e-05, + "loss": 0.4682, + "step": 1309 + }, + { + "epoch": 2.604373757455268, + "grad_norm": 0.42145804477336907, + "learning_rate": 4.407435429885674e-05, + "loss": 0.4598, + "step": 1310 + }, + { + "epoch": 2.606361829025845, + "grad_norm": 0.3143273970433766, + "learning_rate": 4.401910951040473e-05, + "loss": 0.4707, + "step": 1311 + }, + { + "epoch": 2.6083499005964215, + "grad_norm": 0.2402074218552476, + "learning_rate": 4.396385697625745e-05, + "loss": 0.4649, + "step": 1312 + }, + { + "epoch": 2.610337972166998, + "grad_norm": 0.3650833856208346, + "learning_rate": 4.39085968028985e-05, + "loss": 0.4705, + "step": 1313 + }, + { + "epoch": 2.612326043737575, + "grad_norm": 0.3102695268556827, + "learning_rate": 4.385332909682623e-05, + "loss": 0.4581, + "step": 1314 + }, + { + "epoch": 2.614314115308151, + "grad_norm": 0.2274936765796469, + "learning_rate": 4.3798053964553464e-05, + "loss": 0.4707, + "step": 1315 + }, + { + "epoch": 2.6163021868787277, + "grad_norm": 0.21562809171671227, + "learning_rate": 4.374277151260737e-05, + "loss": 0.4702, + "step": 1316 + }, + { + "epoch": 2.618290258449304, + "grad_norm": 0.2554204822921317, + "learning_rate": 4.3687481847529233e-05, + "loss": 0.4614, + "step": 1317 + }, + { + "epoch": 2.6202783300198806, + "grad_norm": 0.29349056607983115, + "learning_rate": 4.3632185075874214e-05, + "loss": 0.4704, + "step": 1318 + }, + { + "epoch": 2.622266401590457, + "grad_norm": 0.20690342951276153, + "learning_rate": 4.357688130421117e-05, + "loss": 0.477, + "step": 1319 + }, + { + "epoch": 2.6242544731610336, + "grad_norm": 0.24912650600700706, + "learning_rate": 4.3521570639122474e-05, + "loss": 0.4676, + "step": 1320 + }, + { + "epoch": 2.6262425447316105, + "grad_norm": 0.31060871647011534, + "learning_rate": 4.346625318720376e-05, + "loss": 0.4731, + "step": 1321 + }, + { + "epoch": 2.628230616302187, + "grad_norm": 0.3175262763718928, + "learning_rate": 4.341092905506375e-05, + "loss": 0.4647, + "step": 1322 + }, + { + "epoch": 2.6302186878727634, + "grad_norm": 0.22988315191632902, + "learning_rate": 4.3355598349324034e-05, + "loss": 0.4652, + "step": 1323 + }, + { + "epoch": 2.63220675944334, + "grad_norm": 0.29964998822607253, + "learning_rate": 4.330026117661888e-05, + "loss": 0.4597, + "step": 1324 + }, + { + "epoch": 2.6341948310139163, + "grad_norm": 0.2862821590914649, + "learning_rate": 4.324491764359502e-05, + "loss": 0.4705, + "step": 1325 + }, + { + "epoch": 2.636182902584493, + "grad_norm": 0.22238282029132211, + "learning_rate": 4.318956785691144e-05, + "loss": 0.4743, + "step": 1326 + }, + { + "epoch": 2.6381709741550696, + "grad_norm": 0.20835203612719516, + "learning_rate": 4.313421192323917e-05, + "loss": 0.4674, + "step": 1327 + }, + { + "epoch": 2.640159045725646, + "grad_norm": 0.21296138418439783, + "learning_rate": 4.30788499492611e-05, + "loss": 0.4669, + "step": 1328 + }, + { + "epoch": 2.6421471172962225, + "grad_norm": 0.6063626470982033, + "learning_rate": 4.3023482041671754e-05, + "loss": 0.4646, + "step": 1329 + }, + { + "epoch": 2.644135188866799, + "grad_norm": 0.22456488467008662, + "learning_rate": 4.296810830717709e-05, + "loss": 0.4576, + "step": 1330 + }, + { + "epoch": 2.646123260437376, + "grad_norm": 0.21173368053034997, + "learning_rate": 4.29127288524943e-05, + "loss": 0.4627, + "step": 1331 + }, + { + "epoch": 2.6481113320079523, + "grad_norm": 0.20369594046938577, + "learning_rate": 4.28573437843516e-05, + "loss": 0.4623, + "step": 1332 + }, + { + "epoch": 2.650099403578529, + "grad_norm": 0.23192364825311867, + "learning_rate": 4.280195320948801e-05, + "loss": 0.4562, + "step": 1333 + }, + { + "epoch": 2.6520874751491053, + "grad_norm": 0.20702404270208882, + "learning_rate": 4.274655723465319e-05, + "loss": 0.4587, + "step": 1334 + }, + { + "epoch": 2.6540755467196817, + "grad_norm": 0.2043349186973809, + "learning_rate": 4.26911559666072e-05, + "loss": 0.4686, + "step": 1335 + }, + { + "epoch": 2.6560636182902586, + "grad_norm": 0.2140823713672654, + "learning_rate": 4.263574951212027e-05, + "loss": 0.467, + "step": 1336 + }, + { + "epoch": 2.658051689860835, + "grad_norm": 0.21150791516481285, + "learning_rate": 4.2580337977972664e-05, + "loss": 0.4664, + "step": 1337 + }, + { + "epoch": 2.6600397614314115, + "grad_norm": 0.22487112272684076, + "learning_rate": 4.2524921470954406e-05, + "loss": 0.4693, + "step": 1338 + }, + { + "epoch": 2.662027833001988, + "grad_norm": 0.21793132020525952, + "learning_rate": 4.2469500097865134e-05, + "loss": 0.4729, + "step": 1339 + }, + { + "epoch": 2.6640159045725644, + "grad_norm": 0.2133804189419038, + "learning_rate": 4.2414073965513846e-05, + "loss": 0.4667, + "step": 1340 + }, + { + "epoch": 2.6660039761431413, + "grad_norm": 0.18948465569873307, + "learning_rate": 4.235864318071871e-05, + "loss": 0.4583, + "step": 1341 + }, + { + "epoch": 2.667992047713718, + "grad_norm": 0.19523854934494658, + "learning_rate": 4.230320785030686e-05, + "loss": 0.4624, + "step": 1342 + }, + { + "epoch": 2.6699801192842942, + "grad_norm": 1.6838215088450827, + "learning_rate": 4.2247768081114204e-05, + "loss": 0.486, + "step": 1343 + }, + { + "epoch": 2.6719681908548707, + "grad_norm": 0.22540813024187262, + "learning_rate": 4.219232397998518e-05, + "loss": 0.4574, + "step": 1344 + }, + { + "epoch": 2.673956262425447, + "grad_norm": 0.28028646989057787, + "learning_rate": 4.2136875653772614e-05, + "loss": 0.4695, + "step": 1345 + }, + { + "epoch": 2.675944333996024, + "grad_norm": 0.23378583807339337, + "learning_rate": 4.2081423209337426e-05, + "loss": 0.4677, + "step": 1346 + }, + { + "epoch": 2.6779324055666005, + "grad_norm": 0.23288832160542997, + "learning_rate": 4.202596675354851e-05, + "loss": 0.4641, + "step": 1347 + }, + { + "epoch": 2.679920477137177, + "grad_norm": 0.34485144522745426, + "learning_rate": 4.1970506393282483e-05, + "loss": 0.465, + "step": 1348 + }, + { + "epoch": 2.6819085487077534, + "grad_norm": 0.22517376125305072, + "learning_rate": 4.191504223542348e-05, + "loss": 0.4698, + "step": 1349 + }, + { + "epoch": 2.68389662027833, + "grad_norm": 0.2601189519472589, + "learning_rate": 4.185957438686294e-05, + "loss": 0.4748, + "step": 1350 + }, + { + "epoch": 2.6858846918489068, + "grad_norm": 0.3474035598879338, + "learning_rate": 4.180410295449944e-05, + "loss": 0.4668, + "step": 1351 + }, + { + "epoch": 2.6878727634194832, + "grad_norm": 0.24685855580960245, + "learning_rate": 4.174862804523847e-05, + "loss": 0.4577, + "step": 1352 + }, + { + "epoch": 2.6898608349900597, + "grad_norm": 0.21318028985851886, + "learning_rate": 4.1693149765992185e-05, + "loss": 0.4677, + "step": 1353 + }, + { + "epoch": 2.691848906560636, + "grad_norm": 0.25285941693159386, + "learning_rate": 4.1637668223679265e-05, + "loss": 0.4649, + "step": 1354 + }, + { + "epoch": 2.6938369781312126, + "grad_norm": 0.25170908983832746, + "learning_rate": 4.158218352522468e-05, + "loss": 0.4576, + "step": 1355 + }, + { + "epoch": 2.6958250497017895, + "grad_norm": 0.24902625836971093, + "learning_rate": 4.152669577755945e-05, + "loss": 0.4694, + "step": 1356 + }, + { + "epoch": 2.697813121272366, + "grad_norm": 0.26522375080926597, + "learning_rate": 4.147120508762051e-05, + "loss": 0.4685, + "step": 1357 + }, + { + "epoch": 2.6998011928429424, + "grad_norm": 0.21543694379170508, + "learning_rate": 4.141571156235044e-05, + "loss": 0.4684, + "step": 1358 + }, + { + "epoch": 2.701789264413519, + "grad_norm": 0.19796151498045514, + "learning_rate": 4.1360215308697294e-05, + "loss": 0.4586, + "step": 1359 + }, + { + "epoch": 2.7037773359840953, + "grad_norm": 0.2360071821488426, + "learning_rate": 4.130471643361438e-05, + "loss": 0.4622, + "step": 1360 + }, + { + "epoch": 2.705765407554672, + "grad_norm": 0.2401591059548515, + "learning_rate": 4.124921504406008e-05, + "loss": 0.4697, + "step": 1361 + }, + { + "epoch": 2.7077534791252487, + "grad_norm": 0.19632371914009855, + "learning_rate": 4.119371124699757e-05, + "loss": 0.4637, + "step": 1362 + }, + { + "epoch": 2.709741550695825, + "grad_norm": 0.24309468602335702, + "learning_rate": 4.1138205149394723e-05, + "loss": 0.4702, + "step": 1363 + }, + { + "epoch": 2.7117296222664016, + "grad_norm": 0.22446170452361522, + "learning_rate": 4.1082696858223826e-05, + "loss": 0.4625, + "step": 1364 + }, + { + "epoch": 2.713717693836978, + "grad_norm": 0.2258465756554639, + "learning_rate": 4.1027186480461375e-05, + "loss": 0.4579, + "step": 1365 + }, + { + "epoch": 2.715705765407555, + "grad_norm": 0.2644422390314895, + "learning_rate": 4.09716741230879e-05, + "loss": 0.4709, + "step": 1366 + }, + { + "epoch": 2.717693836978131, + "grad_norm": 0.2103180179626641, + "learning_rate": 4.0916159893087756e-05, + "loss": 0.4731, + "step": 1367 + }, + { + "epoch": 2.719681908548708, + "grad_norm": 0.27361880683699913, + "learning_rate": 4.0860643897448894e-05, + "loss": 0.4678, + "step": 1368 + }, + { + "epoch": 2.7216699801192843, + "grad_norm": 0.28798144219878985, + "learning_rate": 4.080512624316267e-05, + "loss": 0.4726, + "step": 1369 + }, + { + "epoch": 2.7236580516898607, + "grad_norm": 0.2264593548600379, + "learning_rate": 4.074960703722365e-05, + "loss": 0.4662, + "step": 1370 + }, + { + "epoch": 2.7256461232604376, + "grad_norm": 0.2242660243057448, + "learning_rate": 4.069408638662937e-05, + "loss": 0.4644, + "step": 1371 + }, + { + "epoch": 2.7276341948310137, + "grad_norm": 0.2621409442094008, + "learning_rate": 4.0638564398380163e-05, + "loss": 0.4655, + "step": 1372 + }, + { + "epoch": 2.7296222664015906, + "grad_norm": 0.2595547616945105, + "learning_rate": 4.0583041179478936e-05, + "loss": 0.4638, + "step": 1373 + }, + { + "epoch": 2.731610337972167, + "grad_norm": 0.2274598504625181, + "learning_rate": 4.052751683693095e-05, + "loss": 0.4633, + "step": 1374 + }, + { + "epoch": 2.7335984095427435, + "grad_norm": 0.23999707275096588, + "learning_rate": 4.047199147774368e-05, + "loss": 0.4631, + "step": 1375 + }, + { + "epoch": 2.73558648111332, + "grad_norm": 0.22767172991322615, + "learning_rate": 4.041646520892651e-05, + "loss": 0.4665, + "step": 1376 + }, + { + "epoch": 2.7375745526838964, + "grad_norm": 0.27587767902251026, + "learning_rate": 4.0360938137490604e-05, + "loss": 0.473, + "step": 1377 + }, + { + "epoch": 2.7395626242544733, + "grad_norm": 0.21764221482701487, + "learning_rate": 4.030541037044866e-05, + "loss": 0.4566, + "step": 1378 + }, + { + "epoch": 2.7415506958250497, + "grad_norm": 0.2666834855042023, + "learning_rate": 4.0249882014814735e-05, + "loss": 0.4624, + "step": 1379 + }, + { + "epoch": 2.743538767395626, + "grad_norm": 0.2364553717028977, + "learning_rate": 4.019435317760399e-05, + "loss": 0.4616, + "step": 1380 + }, + { + "epoch": 2.7455268389662026, + "grad_norm": 0.19496138813490002, + "learning_rate": 4.013882396583255e-05, + "loss": 0.4587, + "step": 1381 + }, + { + "epoch": 2.747514910536779, + "grad_norm": 0.20073634106913804, + "learning_rate": 4.008329448651722e-05, + "loss": 0.46, + "step": 1382 + }, + { + "epoch": 2.749502982107356, + "grad_norm": 0.2095676619154563, + "learning_rate": 4.0027764846675375e-05, + "loss": 0.4615, + "step": 1383 + }, + { + "epoch": 2.7514910536779325, + "grad_norm": 0.20024726668039622, + "learning_rate": 3.997223515332464e-05, + "loss": 0.4627, + "step": 1384 + }, + { + "epoch": 2.753479125248509, + "grad_norm": 0.18961125621675926, + "learning_rate": 3.991670551348279e-05, + "loss": 0.4655, + "step": 1385 + }, + { + "epoch": 2.7554671968190854, + "grad_norm": 0.1941167507355979, + "learning_rate": 3.9861176034167464e-05, + "loss": 0.4662, + "step": 1386 + }, + { + "epoch": 2.757455268389662, + "grad_norm": 0.23045374898538842, + "learning_rate": 3.980564682239603e-05, + "loss": 0.4686, + "step": 1387 + }, + { + "epoch": 2.7594433399602387, + "grad_norm": 0.3605198895596153, + "learning_rate": 3.9750117985185265e-05, + "loss": 0.4621, + "step": 1388 + }, + { + "epoch": 2.761431411530815, + "grad_norm": 0.29058441579436517, + "learning_rate": 3.969458962955134e-05, + "loss": 0.4699, + "step": 1389 + }, + { + "epoch": 2.7634194831013916, + "grad_norm": 0.2299369126959643, + "learning_rate": 3.96390618625094e-05, + "loss": 0.4644, + "step": 1390 + }, + { + "epoch": 2.765407554671968, + "grad_norm": 0.19284704992279508, + "learning_rate": 3.9583534791073496e-05, + "loss": 0.474, + "step": 1391 + }, + { + "epoch": 2.7673956262425445, + "grad_norm": 0.2294396001810075, + "learning_rate": 3.952800852225633e-05, + "loss": 0.4692, + "step": 1392 + }, + { + "epoch": 2.7693836978131214, + "grad_norm": 0.21459748499096268, + "learning_rate": 3.947248316306906e-05, + "loss": 0.4549, + "step": 1393 + }, + { + "epoch": 2.771371769383698, + "grad_norm": 0.21514167913930246, + "learning_rate": 3.9416958820521084e-05, + "loss": 0.4676, + "step": 1394 + }, + { + "epoch": 2.7733598409542743, + "grad_norm": 0.21325784114659638, + "learning_rate": 3.936143560161985e-05, + "loss": 0.4642, + "step": 1395 + }, + { + "epoch": 2.775347912524851, + "grad_norm": 0.24281523036646144, + "learning_rate": 3.930591361337065e-05, + "loss": 0.4628, + "step": 1396 + }, + { + "epoch": 2.7773359840954273, + "grad_norm": 0.265175164446194, + "learning_rate": 3.925039296277635e-05, + "loss": 0.4579, + "step": 1397 + }, + { + "epoch": 2.779324055666004, + "grad_norm": 0.25122618281563747, + "learning_rate": 3.919487375683733e-05, + "loss": 0.4571, + "step": 1398 + }, + { + "epoch": 2.7813121272365806, + "grad_norm": 0.21129482504489955, + "learning_rate": 3.913935610255112e-05, + "loss": 0.4486, + "step": 1399 + }, + { + "epoch": 2.783300198807157, + "grad_norm": 0.3147948552145016, + "learning_rate": 3.908384010691226e-05, + "loss": 0.4579, + "step": 1400 + }, + { + "epoch": 2.7852882703777335, + "grad_norm": 0.2455473863371928, + "learning_rate": 3.902832587691211e-05, + "loss": 0.4512, + "step": 1401 + }, + { + "epoch": 2.78727634194831, + "grad_norm": 0.2735323746731553, + "learning_rate": 3.897281351953864e-05, + "loss": 0.4549, + "step": 1402 + }, + { + "epoch": 2.789264413518887, + "grad_norm": 0.28524076892759626, + "learning_rate": 3.891730314177619e-05, + "loss": 0.4635, + "step": 1403 + }, + { + "epoch": 2.7912524850894633, + "grad_norm": 0.2566773378554369, + "learning_rate": 3.886179485060528e-05, + "loss": 0.4602, + "step": 1404 + }, + { + "epoch": 2.79324055666004, + "grad_norm": 0.2357393748079123, + "learning_rate": 3.8806288753002446e-05, + "loss": 0.4738, + "step": 1405 + }, + { + "epoch": 2.7952286282306162, + "grad_norm": 0.19262960449941974, + "learning_rate": 3.875078495593993e-05, + "loss": 0.4552, + "step": 1406 + }, + { + "epoch": 2.7972166998011927, + "grad_norm": 0.2201663667450906, + "learning_rate": 3.869528356638562e-05, + "loss": 0.4611, + "step": 1407 + }, + { + "epoch": 2.7992047713717696, + "grad_norm": 0.27795764096966974, + "learning_rate": 3.863978469130272e-05, + "loss": 0.4572, + "step": 1408 + }, + { + "epoch": 2.801192842942346, + "grad_norm": 0.2335116245887707, + "learning_rate": 3.8584288437649574e-05, + "loss": 0.4694, + "step": 1409 + }, + { + "epoch": 2.8031809145129225, + "grad_norm": 0.21856549040110454, + "learning_rate": 3.8528794912379506e-05, + "loss": 0.4577, + "step": 1410 + }, + { + "epoch": 2.805168986083499, + "grad_norm": 0.25356968098822286, + "learning_rate": 3.8473304222440556e-05, + "loss": 0.4636, + "step": 1411 + }, + { + "epoch": 2.8071570576540754, + "grad_norm": 0.21605953966990488, + "learning_rate": 3.841781647477533e-05, + "loss": 0.4671, + "step": 1412 + }, + { + "epoch": 2.8091451292246523, + "grad_norm": 0.25322156738563867, + "learning_rate": 3.836233177632075e-05, + "loss": 0.4584, + "step": 1413 + }, + { + "epoch": 2.8111332007952288, + "grad_norm": 0.2229043276554981, + "learning_rate": 3.8306850234007835e-05, + "loss": 0.46, + "step": 1414 + }, + { + "epoch": 2.8131212723658052, + "grad_norm": 0.22038124684629573, + "learning_rate": 3.825137195476154e-05, + "loss": 0.4614, + "step": 1415 + }, + { + "epoch": 2.8151093439363817, + "grad_norm": 0.25286192519399997, + "learning_rate": 3.819589704550056e-05, + "loss": 0.4535, + "step": 1416 + }, + { + "epoch": 2.817097415506958, + "grad_norm": 0.25163637021592183, + "learning_rate": 3.814042561313708e-05, + "loss": 0.4622, + "step": 1417 + }, + { + "epoch": 2.819085487077535, + "grad_norm": 0.21455983513210608, + "learning_rate": 3.808495776457654e-05, + "loss": 0.456, + "step": 1418 + }, + { + "epoch": 2.8210735586481115, + "grad_norm": 0.2410203088285688, + "learning_rate": 3.802949360671753e-05, + "loss": 0.4593, + "step": 1419 + }, + { + "epoch": 2.823061630218688, + "grad_norm": 0.22666498127120266, + "learning_rate": 3.7974033246451496e-05, + "loss": 0.4674, + "step": 1420 + }, + { + "epoch": 2.8250497017892644, + "grad_norm": 0.2678467891624614, + "learning_rate": 3.791857679066259e-05, + "loss": 0.4648, + "step": 1421 + }, + { + "epoch": 2.827037773359841, + "grad_norm": 0.24714758794770142, + "learning_rate": 3.786312434622741e-05, + "loss": 0.4563, + "step": 1422 + }, + { + "epoch": 2.8290258449304178, + "grad_norm": 0.3159535557855933, + "learning_rate": 3.780767602001484e-05, + "loss": 0.4691, + "step": 1423 + }, + { + "epoch": 2.8310139165009938, + "grad_norm": 0.24286637586695814, + "learning_rate": 3.77522319188858e-05, + "loss": 0.4564, + "step": 1424 + }, + { + "epoch": 2.8330019880715707, + "grad_norm": 0.2686674277466147, + "learning_rate": 3.7696792149693144e-05, + "loss": 0.46, + "step": 1425 + }, + { + "epoch": 2.834990059642147, + "grad_norm": 0.2813965137999818, + "learning_rate": 3.764135681928131e-05, + "loss": 0.4676, + "step": 1426 + }, + { + "epoch": 2.8369781312127236, + "grad_norm": 0.31227772048480595, + "learning_rate": 3.758592603448617e-05, + "loss": 0.4614, + "step": 1427 + }, + { + "epoch": 2.8389662027833005, + "grad_norm": 0.27050892530225235, + "learning_rate": 3.753049990213487e-05, + "loss": 0.4595, + "step": 1428 + }, + { + "epoch": 2.8409542743538765, + "grad_norm": 0.3253972996223399, + "learning_rate": 3.74750785290456e-05, + "loss": 0.4541, + "step": 1429 + }, + { + "epoch": 2.8429423459244534, + "grad_norm": 0.8290885706508856, + "learning_rate": 3.7419662022027356e-05, + "loss": 0.4626, + "step": 1430 + }, + { + "epoch": 2.84493041749503, + "grad_norm": 0.3278453956679973, + "learning_rate": 3.736425048787976e-05, + "loss": 0.4595, + "step": 1431 + }, + { + "epoch": 2.8469184890656063, + "grad_norm": 0.3380893038877377, + "learning_rate": 3.730884403339282e-05, + "loss": 0.458, + "step": 1432 + }, + { + "epoch": 2.8489065606361827, + "grad_norm": 0.2208597079738313, + "learning_rate": 3.725344276534681e-05, + "loss": 0.4588, + "step": 1433 + }, + { + "epoch": 2.850894632206759, + "grad_norm": 0.2229228221920045, + "learning_rate": 3.7198046790512e-05, + "loss": 0.472, + "step": 1434 + }, + { + "epoch": 2.852882703777336, + "grad_norm": 0.34642900940428056, + "learning_rate": 3.714265621564842e-05, + "loss": 0.4594, + "step": 1435 + }, + { + "epoch": 2.8548707753479126, + "grad_norm": 0.26258153715517213, + "learning_rate": 3.708727114750572e-05, + "loss": 0.4698, + "step": 1436 + }, + { + "epoch": 2.856858846918489, + "grad_norm": 0.21133696660065732, + "learning_rate": 3.7031891692822924e-05, + "loss": 0.47, + "step": 1437 + }, + { + "epoch": 2.8588469184890655, + "grad_norm": 0.2805733965805607, + "learning_rate": 3.697651795832826e-05, + "loss": 0.4618, + "step": 1438 + }, + { + "epoch": 2.860834990059642, + "grad_norm": 0.28771730407413304, + "learning_rate": 3.692115005073891e-05, + "loss": 0.4622, + "step": 1439 + }, + { + "epoch": 2.862823061630219, + "grad_norm": 0.257745593969288, + "learning_rate": 3.686578807676085e-05, + "loss": 0.4601, + "step": 1440 + }, + { + "epoch": 2.8648111332007953, + "grad_norm": 0.29359923657472803, + "learning_rate": 3.681043214308856e-05, + "loss": 0.4672, + "step": 1441 + }, + { + "epoch": 2.8667992047713717, + "grad_norm": 0.193949825815246, + "learning_rate": 3.6755082356404976e-05, + "loss": 0.4547, + "step": 1442 + }, + { + "epoch": 2.868787276341948, + "grad_norm": 0.26037388451793225, + "learning_rate": 3.6699738823381125e-05, + "loss": 0.4653, + "step": 1443 + }, + { + "epoch": 2.8707753479125246, + "grad_norm": 0.337949377251768, + "learning_rate": 3.664440165067598e-05, + "loss": 0.4588, + "step": 1444 + }, + { + "epoch": 2.8727634194831015, + "grad_norm": 0.9950831510061162, + "learning_rate": 3.6589070944936264e-05, + "loss": 0.4635, + "step": 1445 + }, + { + "epoch": 2.874751491053678, + "grad_norm": 0.2688681333914108, + "learning_rate": 3.653374681279625e-05, + "loss": 0.4609, + "step": 1446 + }, + { + "epoch": 2.8767395626242545, + "grad_norm": 0.26931338222240986, + "learning_rate": 3.647842936087754e-05, + "loss": 0.4537, + "step": 1447 + }, + { + "epoch": 2.878727634194831, + "grad_norm": 0.2657472291582106, + "learning_rate": 3.6423118695788834e-05, + "loss": 0.4639, + "step": 1448 + }, + { + "epoch": 2.8807157057654074, + "grad_norm": 0.30395367100041154, + "learning_rate": 3.6367814924125806e-05, + "loss": 0.4579, + "step": 1449 + }, + { + "epoch": 2.8827037773359843, + "grad_norm": 0.26228486695021913, + "learning_rate": 3.631251815247077e-05, + "loss": 0.4622, + "step": 1450 + }, + { + "epoch": 2.8846918489065607, + "grad_norm": 0.25042552063623835, + "learning_rate": 3.625722848739263e-05, + "loss": 0.4578, + "step": 1451 + }, + { + "epoch": 2.886679920477137, + "grad_norm": 0.37353549732274993, + "learning_rate": 3.620194603544655e-05, + "loss": 0.4583, + "step": 1452 + }, + { + "epoch": 2.8886679920477136, + "grad_norm": 0.2702383408957765, + "learning_rate": 3.6146670903173785e-05, + "loss": 0.4605, + "step": 1453 + }, + { + "epoch": 2.89065606361829, + "grad_norm": 0.20949177632500207, + "learning_rate": 3.6091403197101504e-05, + "loss": 0.4594, + "step": 1454 + }, + { + "epoch": 2.892644135188867, + "grad_norm": 0.20249165415901857, + "learning_rate": 3.6036143023742553e-05, + "loss": 0.4586, + "step": 1455 + }, + { + "epoch": 2.8946322067594434, + "grad_norm": 0.2013107734031283, + "learning_rate": 3.5980890489595275e-05, + "loss": 0.4582, + "step": 1456 + }, + { + "epoch": 2.89662027833002, + "grad_norm": 0.17825803453477496, + "learning_rate": 3.592564570114327e-05, + "loss": 0.4667, + "step": 1457 + }, + { + "epoch": 2.8986083499005963, + "grad_norm": 0.2043117194799363, + "learning_rate": 3.587040876485524e-05, + "loss": 0.4523, + "step": 1458 + }, + { + "epoch": 2.900596421471173, + "grad_norm": 0.1889988045208227, + "learning_rate": 3.5815179787184694e-05, + "loss": 0.4471, + "step": 1459 + }, + { + "epoch": 2.9025844930417497, + "grad_norm": 0.17301877144085692, + "learning_rate": 3.575995887456989e-05, + "loss": 0.4616, + "step": 1460 + }, + { + "epoch": 2.904572564612326, + "grad_norm": 0.1858563184396588, + "learning_rate": 3.570474613343348e-05, + "loss": 0.4538, + "step": 1461 + }, + { + "epoch": 2.9065606361829026, + "grad_norm": 0.17475933284775572, + "learning_rate": 3.5649541670182385e-05, + "loss": 0.4644, + "step": 1462 + }, + { + "epoch": 2.908548707753479, + "grad_norm": 0.19020230936864846, + "learning_rate": 3.559434559120757e-05, + "loss": 0.4662, + "step": 1463 + }, + { + "epoch": 2.9105367793240555, + "grad_norm": 0.19660105128297872, + "learning_rate": 3.553915800288385e-05, + "loss": 0.4679, + "step": 1464 + }, + { + "epoch": 2.9125248508946324, + "grad_norm": 0.2046735778964775, + "learning_rate": 3.548397901156966e-05, + "loss": 0.4664, + "step": 1465 + }, + { + "epoch": 2.914512922465209, + "grad_norm": 0.21619552346200047, + "learning_rate": 3.542880872360689e-05, + "loss": 0.4566, + "step": 1466 + }, + { + "epoch": 2.9165009940357853, + "grad_norm": 0.19360108054912628, + "learning_rate": 3.537364724532065e-05, + "loss": 0.4655, + "step": 1467 + }, + { + "epoch": 2.918489065606362, + "grad_norm": 0.18241404196740055, + "learning_rate": 3.5318494683019025e-05, + "loss": 0.4662, + "step": 1468 + }, + { + "epoch": 2.9204771371769382, + "grad_norm": 0.21577796418109538, + "learning_rate": 3.5263351142993e-05, + "loss": 0.4599, + "step": 1469 + }, + { + "epoch": 2.922465208747515, + "grad_norm": 0.2087684977429425, + "learning_rate": 3.520821673151614e-05, + "loss": 0.465, + "step": 1470 + }, + { + "epoch": 2.9244532803180916, + "grad_norm": 0.6884222775945866, + "learning_rate": 3.515309155484438e-05, + "loss": 0.4607, + "step": 1471 + }, + { + "epoch": 2.926441351888668, + "grad_norm": 0.224933874693449, + "learning_rate": 3.509797571921589e-05, + "loss": 0.4556, + "step": 1472 + }, + { + "epoch": 2.9284294234592445, + "grad_norm": 0.23471362139521423, + "learning_rate": 3.5042869330850844e-05, + "loss": 0.4643, + "step": 1473 + }, + { + "epoch": 2.930417495029821, + "grad_norm": 0.233852872228427, + "learning_rate": 3.498777249595118e-05, + "loss": 0.452, + "step": 1474 + }, + { + "epoch": 2.932405566600398, + "grad_norm": 0.2210891714850947, + "learning_rate": 3.493268532070046e-05, + "loss": 0.4678, + "step": 1475 + }, + { + "epoch": 2.9343936381709743, + "grad_norm": 0.3535398087630285, + "learning_rate": 3.487760791126361e-05, + "loss": 0.469, + "step": 1476 + }, + { + "epoch": 2.9363817097415508, + "grad_norm": 0.25585773718055604, + "learning_rate": 3.482254037378671e-05, + "loss": 0.4532, + "step": 1477 + }, + { + "epoch": 2.9383697813121272, + "grad_norm": 0.23694109296102414, + "learning_rate": 3.476748281439688e-05, + "loss": 0.4564, + "step": 1478 + }, + { + "epoch": 2.9403578528827037, + "grad_norm": 0.19910372458518003, + "learning_rate": 3.471243533920197e-05, + "loss": 0.4621, + "step": 1479 + }, + { + "epoch": 2.9423459244532806, + "grad_norm": 0.2237982620944968, + "learning_rate": 3.465739805429038e-05, + "loss": 0.4668, + "step": 1480 + }, + { + "epoch": 2.9443339960238566, + "grad_norm": 0.21397024941277665, + "learning_rate": 3.460237106573089e-05, + "loss": 0.4585, + "step": 1481 + }, + { + "epoch": 2.9463220675944335, + "grad_norm": 0.21613520324167362, + "learning_rate": 3.454735447957244e-05, + "loss": 0.4643, + "step": 1482 + }, + { + "epoch": 2.94831013916501, + "grad_norm": 0.20222193441002365, + "learning_rate": 3.44923484018439e-05, + "loss": 0.4553, + "step": 1483 + }, + { + "epoch": 2.9502982107355864, + "grad_norm": 0.20180090085759034, + "learning_rate": 3.443735293855395e-05, + "loss": 0.4608, + "step": 1484 + }, + { + "epoch": 2.952286282306163, + "grad_norm": 0.1888317542842441, + "learning_rate": 3.438236819569074e-05, + "loss": 0.4578, + "step": 1485 + }, + { + "epoch": 2.9542743538767393, + "grad_norm": 0.1921413976782572, + "learning_rate": 3.432739427922176e-05, + "loss": 0.4525, + "step": 1486 + }, + { + "epoch": 2.956262425447316, + "grad_norm": 0.18167928678331619, + "learning_rate": 3.427243129509372e-05, + "loss": 0.4519, + "step": 1487 + }, + { + "epoch": 2.9582504970178927, + "grad_norm": 0.18458189230588334, + "learning_rate": 3.421747934923219e-05, + "loss": 0.4683, + "step": 1488 + }, + { + "epoch": 2.960238568588469, + "grad_norm": 0.2082868127458548, + "learning_rate": 3.4162538547541465e-05, + "loss": 0.4552, + "step": 1489 + }, + { + "epoch": 2.9622266401590456, + "grad_norm": 0.18584151995338427, + "learning_rate": 3.410760899590439e-05, + "loss": 0.4631, + "step": 1490 + }, + { + "epoch": 2.964214711729622, + "grad_norm": 0.2383682645766151, + "learning_rate": 3.4052690800182104e-05, + "loss": 0.4689, + "step": 1491 + }, + { + "epoch": 2.966202783300199, + "grad_norm": 0.2404793594300668, + "learning_rate": 3.39977840662139e-05, + "loss": 0.467, + "step": 1492 + }, + { + "epoch": 2.9681908548707754, + "grad_norm": 0.20090767861090764, + "learning_rate": 3.394288889981695e-05, + "loss": 0.4533, + "step": 1493 + }, + { + "epoch": 2.970178926441352, + "grad_norm": 0.22162509945461517, + "learning_rate": 3.38880054067861e-05, + "loss": 0.4591, + "step": 1494 + }, + { + "epoch": 2.9721669980119283, + "grad_norm": 0.18800071463657672, + "learning_rate": 3.38331336928938e-05, + "loss": 0.4561, + "step": 1495 + }, + { + "epoch": 2.9741550695825048, + "grad_norm": 0.206244471673574, + "learning_rate": 3.377827386388971e-05, + "loss": 0.4654, + "step": 1496 + }, + { + "epoch": 2.9761431411530817, + "grad_norm": 0.22880555976938174, + "learning_rate": 3.372342602550062e-05, + "loss": 0.4558, + "step": 1497 + }, + { + "epoch": 2.978131212723658, + "grad_norm": 0.17949305996147424, + "learning_rate": 3.3668590283430185e-05, + "loss": 0.4627, + "step": 1498 + }, + { + "epoch": 2.9801192842942346, + "grad_norm": 0.18517810607930787, + "learning_rate": 3.3613766743358785e-05, + "loss": 0.4597, + "step": 1499 + }, + { + "epoch": 2.982107355864811, + "grad_norm": 0.23291876613271462, + "learning_rate": 3.3558955510943256e-05, + "loss": 0.4669, + "step": 1500 + }, + { + "epoch": 2.9840954274353875, + "grad_norm": 0.1820730482432967, + "learning_rate": 3.350415669181673e-05, + "loss": 0.4555, + "step": 1501 + }, + { + "epoch": 2.9860834990059644, + "grad_norm": 0.19413753257771896, + "learning_rate": 3.3449370391588424e-05, + "loss": 0.4633, + "step": 1502 + }, + { + "epoch": 2.988071570576541, + "grad_norm": 0.19346750713312172, + "learning_rate": 3.339459671584337e-05, + "loss": 0.4555, + "step": 1503 + }, + { + "epoch": 2.9900596421471173, + "grad_norm": 0.18401415157508247, + "learning_rate": 3.3339835770142364e-05, + "loss": 0.4621, + "step": 1504 + }, + { + "epoch": 2.9920477137176937, + "grad_norm": 0.19007332079668393, + "learning_rate": 3.32850876600216e-05, + "loss": 0.4604, + "step": 1505 + }, + { + "epoch": 2.99403578528827, + "grad_norm": 0.20364951200142725, + "learning_rate": 3.323035249099256e-05, + "loss": 0.4643, + "step": 1506 + }, + { + "epoch": 2.996023856858847, + "grad_norm": 0.17759550097595844, + "learning_rate": 3.3175630368541776e-05, + "loss": 0.4589, + "step": 1507 + }, + { + "epoch": 2.9980119284294235, + "grad_norm": 0.19573811055059043, + "learning_rate": 3.312092139813064e-05, + "loss": 0.461, + "step": 1508 + }, + { + "epoch": 3.0, + "grad_norm": 0.18051472969551505, + "learning_rate": 3.3066225685195194e-05, + "loss": 0.4438, + "step": 1509 + }, + { + "epoch": 3.0019880715705765, + "grad_norm": 0.21866713244429514, + "learning_rate": 3.301154333514594e-05, + "loss": 0.4275, + "step": 1510 + }, + { + "epoch": 3.003976143141153, + "grad_norm": 0.19247081598906532, + "learning_rate": 3.295687445336762e-05, + "loss": 0.4225, + "step": 1511 + }, + { + "epoch": 3.00596421471173, + "grad_norm": 0.2309539724817374, + "learning_rate": 3.290221914521899e-05, + "loss": 0.4219, + "step": 1512 + }, + { + "epoch": 3.0079522862823063, + "grad_norm": 0.21933537357774005, + "learning_rate": 3.284757751603273e-05, + "loss": 0.4293, + "step": 1513 + }, + { + "epoch": 3.0099403578528827, + "grad_norm": 0.23026407562644793, + "learning_rate": 3.279294967111508e-05, + "loss": 0.4192, + "step": 1514 + }, + { + "epoch": 3.011928429423459, + "grad_norm": 0.23983833817576242, + "learning_rate": 3.273833571574574e-05, + "loss": 0.4207, + "step": 1515 + }, + { + "epoch": 3.0139165009940356, + "grad_norm": 0.18930871849206557, + "learning_rate": 3.268373575517764e-05, + "loss": 0.4326, + "step": 1516 + }, + { + "epoch": 3.0159045725646125, + "grad_norm": 0.1911013501133307, + "learning_rate": 3.262914989463673e-05, + "loss": 0.429, + "step": 1517 + }, + { + "epoch": 3.017892644135189, + "grad_norm": 0.20511791924121267, + "learning_rate": 3.257457823932181e-05, + "loss": 0.426, + "step": 1518 + }, + { + "epoch": 3.0198807157057654, + "grad_norm": 0.20338555749017675, + "learning_rate": 3.252002089440429e-05, + "loss": 0.4288, + "step": 1519 + }, + { + "epoch": 3.021868787276342, + "grad_norm": 0.2266632220816163, + "learning_rate": 3.2465477965028e-05, + "loss": 0.4095, + "step": 1520 + }, + { + "epoch": 3.0238568588469183, + "grad_norm": 0.1895880120844912, + "learning_rate": 3.241094955630895e-05, + "loss": 0.4261, + "step": 1521 + }, + { + "epoch": 3.0258449304174952, + "grad_norm": 0.23483445478957796, + "learning_rate": 3.235643577333527e-05, + "loss": 0.4193, + "step": 1522 + }, + { + "epoch": 3.0278330019880717, + "grad_norm": 0.3732412917899254, + "learning_rate": 3.230193672116681e-05, + "loss": 0.4239, + "step": 1523 + }, + { + "epoch": 3.029821073558648, + "grad_norm": 0.21687047596976053, + "learning_rate": 3.224745250483504e-05, + "loss": 0.4278, + "step": 1524 + }, + { + "epoch": 3.0318091451292246, + "grad_norm": 0.2039436298095718, + "learning_rate": 3.219298322934287e-05, + "loss": 0.4184, + "step": 1525 + }, + { + "epoch": 3.033797216699801, + "grad_norm": 0.19594156067182716, + "learning_rate": 3.213852899966441e-05, + "loss": 0.4259, + "step": 1526 + }, + { + "epoch": 3.0357852882703775, + "grad_norm": 0.20179469518562662, + "learning_rate": 3.2084089920744746e-05, + "loss": 0.4309, + "step": 1527 + }, + { + "epoch": 3.0377733598409544, + "grad_norm": 0.19670377201579528, + "learning_rate": 3.2029666097499804e-05, + "loss": 0.416, + "step": 1528 + }, + { + "epoch": 3.039761431411531, + "grad_norm": 0.17957230536692131, + "learning_rate": 3.197525763481608e-05, + "loss": 0.422, + "step": 1529 + }, + { + "epoch": 3.0417495029821073, + "grad_norm": 0.15981443945964646, + "learning_rate": 3.1920864637550454e-05, + "loss": 0.4205, + "step": 1530 + }, + { + "epoch": 3.043737574552684, + "grad_norm": 0.20291061700552804, + "learning_rate": 3.186648721053005e-05, + "loss": 0.4269, + "step": 1531 + }, + { + "epoch": 3.0457256461232602, + "grad_norm": 0.17478282257132408, + "learning_rate": 3.1812125458551974e-05, + "loss": 0.4132, + "step": 1532 + }, + { + "epoch": 3.047713717693837, + "grad_norm": 0.16948145356552408, + "learning_rate": 3.175777948638307e-05, + "loss": 0.4289, + "step": 1533 + }, + { + "epoch": 3.0497017892644136, + "grad_norm": 0.16854257729121044, + "learning_rate": 3.170344939875982e-05, + "loss": 0.4282, + "step": 1534 + }, + { + "epoch": 3.05168986083499, + "grad_norm": 0.16732134458370454, + "learning_rate": 3.164913530038808e-05, + "loss": 0.4177, + "step": 1535 + }, + { + "epoch": 3.0536779324055665, + "grad_norm": 0.17964783746191185, + "learning_rate": 3.15948372959429e-05, + "loss": 0.4156, + "step": 1536 + }, + { + "epoch": 3.055666003976143, + "grad_norm": 0.17083246059369028, + "learning_rate": 3.1540555490068304e-05, + "loss": 0.4212, + "step": 1537 + }, + { + "epoch": 3.05765407554672, + "grad_norm": 0.17480502617064078, + "learning_rate": 3.14862899873771e-05, + "loss": 0.4201, + "step": 1538 + }, + { + "epoch": 3.0596421471172963, + "grad_norm": 0.15811181049785739, + "learning_rate": 3.143204089245063e-05, + "loss": 0.4236, + "step": 1539 + }, + { + "epoch": 3.0616302186878728, + "grad_norm": 0.20090880336661895, + "learning_rate": 3.137780830983873e-05, + "loss": 0.4164, + "step": 1540 + }, + { + "epoch": 3.0636182902584492, + "grad_norm": 0.2116311394564602, + "learning_rate": 3.132359234405931e-05, + "loss": 0.4285, + "step": 1541 + }, + { + "epoch": 3.0656063618290257, + "grad_norm": 0.17291665249112143, + "learning_rate": 3.12693930995983e-05, + "loss": 0.4227, + "step": 1542 + }, + { + "epoch": 3.0675944333996026, + "grad_norm": 0.20723641513115693, + "learning_rate": 3.121521068090937e-05, + "loss": 0.4265, + "step": 1543 + }, + { + "epoch": 3.069582504970179, + "grad_norm": 0.1839141741250624, + "learning_rate": 3.11610451924138e-05, + "loss": 0.4234, + "step": 1544 + }, + { + "epoch": 3.0715705765407555, + "grad_norm": 0.19373127551015817, + "learning_rate": 3.110689673850023e-05, + "loss": 0.4191, + "step": 1545 + }, + { + "epoch": 3.073558648111332, + "grad_norm": 0.1660825761301819, + "learning_rate": 3.105276542352447e-05, + "loss": 0.4267, + "step": 1546 + }, + { + "epoch": 3.0755467196819084, + "grad_norm": 0.1648308561329296, + "learning_rate": 3.0998651351809305e-05, + "loss": 0.4217, + "step": 1547 + }, + { + "epoch": 3.0775347912524853, + "grad_norm": 0.18042851709340152, + "learning_rate": 3.094455462764425e-05, + "loss": 0.4211, + "step": 1548 + }, + { + "epoch": 3.0795228628230618, + "grad_norm": 0.18388627104468944, + "learning_rate": 3.089047535528545e-05, + "loss": 0.4187, + "step": 1549 + }, + { + "epoch": 3.081510934393638, + "grad_norm": 0.18338332550558414, + "learning_rate": 3.08364136389554e-05, + "loss": 0.4222, + "step": 1550 + }, + { + "epoch": 3.0834990059642147, + "grad_norm": 0.5457393482733713, + "learning_rate": 3.07823695828427e-05, + "loss": 0.4392, + "step": 1551 + }, + { + "epoch": 3.085487077534791, + "grad_norm": 0.16946702499846994, + "learning_rate": 3.0728343291102e-05, + "loss": 0.425, + "step": 1552 + }, + { + "epoch": 3.0874751491053676, + "grad_norm": 0.21982162026792132, + "learning_rate": 3.067433486785366e-05, + "loss": 0.4252, + "step": 1553 + }, + { + "epoch": 3.0894632206759445, + "grad_norm": 0.20012161267571643, + "learning_rate": 3.062034441718362e-05, + "loss": 0.4271, + "step": 1554 + }, + { + "epoch": 3.091451292246521, + "grad_norm": 0.17533381870496842, + "learning_rate": 3.056637204314318e-05, + "loss": 0.4179, + "step": 1555 + }, + { + "epoch": 3.0934393638170974, + "grad_norm": 0.3163113330290082, + "learning_rate": 3.0512417849748805e-05, + "loss": 0.4221, + "step": 1556 + }, + { + "epoch": 3.095427435387674, + "grad_norm": 0.1649768820807243, + "learning_rate": 3.0458481940981933e-05, + "loss": 0.432, + "step": 1557 + }, + { + "epoch": 3.0974155069582503, + "grad_norm": 0.15470699826266648, + "learning_rate": 3.0404564420788745e-05, + "loss": 0.4308, + "step": 1558 + }, + { + "epoch": 3.099403578528827, + "grad_norm": 0.1702220476604976, + "learning_rate": 3.035066539308e-05, + "loss": 0.427, + "step": 1559 + }, + { + "epoch": 3.1013916500994037, + "grad_norm": 0.18688827814809164, + "learning_rate": 3.029678496173079e-05, + "loss": 0.4289, + "step": 1560 + }, + { + "epoch": 3.10337972166998, + "grad_norm": 0.14448275902724284, + "learning_rate": 3.024292323058039e-05, + "loss": 0.4204, + "step": 1561 + }, + { + "epoch": 3.1053677932405566, + "grad_norm": 0.1825200775041438, + "learning_rate": 3.0189080303432044e-05, + "loss": 0.4264, + "step": 1562 + }, + { + "epoch": 3.107355864811133, + "grad_norm": 0.15196658182073003, + "learning_rate": 3.013525628405274e-05, + "loss": 0.4248, + "step": 1563 + }, + { + "epoch": 3.10934393638171, + "grad_norm": 0.23580394328013937, + "learning_rate": 3.008145127617302e-05, + "loss": 0.4274, + "step": 1564 + }, + { + "epoch": 3.1113320079522864, + "grad_norm": 0.17167581205154928, + "learning_rate": 3.0027665383486803e-05, + "loss": 0.419, + "step": 1565 + }, + { + "epoch": 3.113320079522863, + "grad_norm": 0.17616777841111306, + "learning_rate": 2.997389870965118e-05, + "loss": 0.4213, + "step": 1566 + }, + { + "epoch": 3.1153081510934393, + "grad_norm": 0.17527288324883256, + "learning_rate": 2.9920151358286166e-05, + "loss": 0.425, + "step": 1567 + }, + { + "epoch": 3.1172962226640157, + "grad_norm": 0.17865409252565498, + "learning_rate": 2.986642343297458e-05, + "loss": 0.4245, + "step": 1568 + }, + { + "epoch": 3.1192842942345926, + "grad_norm": 0.17682031138461274, + "learning_rate": 2.9812715037261756e-05, + "loss": 0.4291, + "step": 1569 + }, + { + "epoch": 3.121272365805169, + "grad_norm": 0.16631730027106617, + "learning_rate": 2.9759026274655433e-05, + "loss": 0.4301, + "step": 1570 + }, + { + "epoch": 3.1232604373757455, + "grad_norm": 0.3566716216311293, + "learning_rate": 2.970535724862548e-05, + "loss": 0.4324, + "step": 1571 + }, + { + "epoch": 3.125248508946322, + "grad_norm": 0.16423762796865618, + "learning_rate": 2.9651708062603763e-05, + "loss": 0.4223, + "step": 1572 + }, + { + "epoch": 3.1272365805168985, + "grad_norm": 0.15581656406502245, + "learning_rate": 2.959807881998389e-05, + "loss": 0.4152, + "step": 1573 + }, + { + "epoch": 3.1292246520874754, + "grad_norm": 0.17130907591898645, + "learning_rate": 2.9544469624121e-05, + "loss": 0.4206, + "step": 1574 + }, + { + "epoch": 3.131212723658052, + "grad_norm": 0.16199415820717575, + "learning_rate": 2.9490880578331685e-05, + "loss": 0.4222, + "step": 1575 + }, + { + "epoch": 3.1332007952286283, + "grad_norm": 0.16445570164173293, + "learning_rate": 2.9437311785893635e-05, + "loss": 0.419, + "step": 1576 + }, + { + "epoch": 3.1351888667992047, + "grad_norm": 0.16037847818725806, + "learning_rate": 2.938376335004553e-05, + "loss": 0.4228, + "step": 1577 + }, + { + "epoch": 3.137176938369781, + "grad_norm": 0.16786978324050583, + "learning_rate": 2.9330235373986803e-05, + "loss": 0.4247, + "step": 1578 + }, + { + "epoch": 3.139165009940358, + "grad_norm": 0.18278838840353107, + "learning_rate": 2.9276727960877473e-05, + "loss": 0.4315, + "step": 1579 + }, + { + "epoch": 3.1411530815109345, + "grad_norm": 0.1651561323450754, + "learning_rate": 2.9223241213837923e-05, + "loss": 0.4092, + "step": 1580 + }, + { + "epoch": 3.143141153081511, + "grad_norm": 0.16966254220617266, + "learning_rate": 2.916977523594871e-05, + "loss": 0.4181, + "step": 1581 + }, + { + "epoch": 3.1451292246520874, + "grad_norm": 0.17505463062318927, + "learning_rate": 2.9116330130250374e-05, + "loss": 0.4266, + "step": 1582 + }, + { + "epoch": 3.147117296222664, + "grad_norm": 0.16115989680942328, + "learning_rate": 2.906290599974318e-05, + "loss": 0.4319, + "step": 1583 + }, + { + "epoch": 3.1491053677932404, + "grad_norm": 0.18383791657309656, + "learning_rate": 2.9009502947387056e-05, + "loss": 0.4194, + "step": 1584 + }, + { + "epoch": 3.1510934393638173, + "grad_norm": 0.14571585197850928, + "learning_rate": 2.8956121076101244e-05, + "loss": 0.4231, + "step": 1585 + }, + { + "epoch": 3.1530815109343937, + "grad_norm": 0.17282383882108132, + "learning_rate": 2.890276048876419e-05, + "loss": 0.4198, + "step": 1586 + }, + { + "epoch": 3.15506958250497, + "grad_norm": 0.16960750212497086, + "learning_rate": 2.884942128821329e-05, + "loss": 0.4203, + "step": 1587 + }, + { + "epoch": 3.1570576540755466, + "grad_norm": 0.14433740341835166, + "learning_rate": 2.8796103577244766e-05, + "loss": 0.4194, + "step": 1588 + }, + { + "epoch": 3.159045725646123, + "grad_norm": 0.18244597958945752, + "learning_rate": 2.8742807458613393e-05, + "loss": 0.4205, + "step": 1589 + }, + { + "epoch": 3.1610337972167, + "grad_norm": 0.2348787312272368, + "learning_rate": 2.8689533035032355e-05, + "loss": 0.4257, + "step": 1590 + }, + { + "epoch": 3.1630218687872764, + "grad_norm": 0.1731480343868321, + "learning_rate": 2.8636280409173012e-05, + "loss": 0.4278, + "step": 1591 + }, + { + "epoch": 3.165009940357853, + "grad_norm": 0.15285486539504065, + "learning_rate": 2.8583049683664698e-05, + "loss": 0.4245, + "step": 1592 + }, + { + "epoch": 3.1669980119284293, + "grad_norm": 1.1577676169956606, + "learning_rate": 2.8529840961094596e-05, + "loss": 0.4473, + "step": 1593 + }, + { + "epoch": 3.168986083499006, + "grad_norm": 0.15314522438830153, + "learning_rate": 2.8476654344007438e-05, + "loss": 0.417, + "step": 1594 + }, + { + "epoch": 3.1709741550695827, + "grad_norm": 0.1854340474178689, + "learning_rate": 2.8423489934905348e-05, + "loss": 0.4199, + "step": 1595 + }, + { + "epoch": 3.172962226640159, + "grad_norm": 0.2189527140355646, + "learning_rate": 2.837034783624768e-05, + "loss": 0.4232, + "step": 1596 + }, + { + "epoch": 3.1749502982107356, + "grad_norm": 0.16477213116978373, + "learning_rate": 2.8317228150450775e-05, + "loss": 0.4232, + "step": 1597 + }, + { + "epoch": 3.176938369781312, + "grad_norm": 0.18766127242746655, + "learning_rate": 2.8264130979887786e-05, + "loss": 0.4189, + "step": 1598 + }, + { + "epoch": 3.1789264413518885, + "grad_norm": 0.16962240140631765, + "learning_rate": 2.8211056426888466e-05, + "loss": 0.4303, + "step": 1599 + }, + { + "epoch": 3.1809145129224654, + "grad_norm": 0.23562016505996533, + "learning_rate": 2.8158004593738994e-05, + "loss": 0.4198, + "step": 1600 + }, + { + "epoch": 3.182902584493042, + "grad_norm": 0.16522091902575367, + "learning_rate": 2.810497558268172e-05, + "loss": 0.4236, + "step": 1601 + }, + { + "epoch": 3.1848906560636183, + "grad_norm": 0.20700531289678575, + "learning_rate": 2.8051969495915087e-05, + "loss": 0.4253, + "step": 1602 + }, + { + "epoch": 3.1868787276341948, + "grad_norm": 0.15198735404558872, + "learning_rate": 2.7998986435593306e-05, + "loss": 0.4235, + "step": 1603 + }, + { + "epoch": 3.1888667992047712, + "grad_norm": 0.16179795866633792, + "learning_rate": 2.7946026503826203e-05, + "loss": 0.4223, + "step": 1604 + }, + { + "epoch": 3.1908548707753477, + "grad_norm": 0.16649993865726812, + "learning_rate": 2.7893089802679055e-05, + "loss": 0.4257, + "step": 1605 + }, + { + "epoch": 3.1928429423459246, + "grad_norm": 0.15903132930020156, + "learning_rate": 2.7840176434172357e-05, + "loss": 0.4263, + "step": 1606 + }, + { + "epoch": 3.194831013916501, + "grad_norm": 0.16054210668205585, + "learning_rate": 2.778728650028164e-05, + "loss": 0.4182, + "step": 1607 + }, + { + "epoch": 3.1968190854870775, + "grad_norm": 0.15793274003584565, + "learning_rate": 2.773442010293728e-05, + "loss": 0.4196, + "step": 1608 + }, + { + "epoch": 3.198807157057654, + "grad_norm": 0.16707398647771807, + "learning_rate": 2.7681577344024274e-05, + "loss": 0.4291, + "step": 1609 + }, + { + "epoch": 3.2007952286282304, + "grad_norm": 0.1615709751294597, + "learning_rate": 2.7628758325382086e-05, + "loss": 0.4196, + "step": 1610 + }, + { + "epoch": 3.2027833001988073, + "grad_norm": 0.18962014963322602, + "learning_rate": 2.757596314880441e-05, + "loss": 0.4249, + "step": 1611 + }, + { + "epoch": 3.2047713717693838, + "grad_norm": 0.2498774125522842, + "learning_rate": 2.7523191916039e-05, + "loss": 0.4244, + "step": 1612 + }, + { + "epoch": 3.20675944333996, + "grad_norm": 0.33194601474779295, + "learning_rate": 2.7470444728787457e-05, + "loss": 0.4236, + "step": 1613 + }, + { + "epoch": 3.2087475149105367, + "grad_norm": 0.14356716664383448, + "learning_rate": 2.7417721688705038e-05, + "loss": 0.412, + "step": 1614 + }, + { + "epoch": 3.210735586481113, + "grad_norm": 0.1605633240701923, + "learning_rate": 2.7365022897400474e-05, + "loss": 0.424, + "step": 1615 + }, + { + "epoch": 3.21272365805169, + "grad_norm": 0.15677114515341178, + "learning_rate": 2.7312348456435766e-05, + "loss": 0.4158, + "step": 1616 + }, + { + "epoch": 3.2147117296222665, + "grad_norm": 0.14935248284076827, + "learning_rate": 2.7259698467325977e-05, + "loss": 0.433, + "step": 1617 + }, + { + "epoch": 3.216699801192843, + "grad_norm": 0.1451275646279355, + "learning_rate": 2.720707303153904e-05, + "loss": 0.42, + "step": 1618 + }, + { + "epoch": 3.2186878727634194, + "grad_norm": 0.14493464992235816, + "learning_rate": 2.7154472250495597e-05, + "loss": 0.4203, + "step": 1619 + }, + { + "epoch": 3.220675944333996, + "grad_norm": 0.14380332893928113, + "learning_rate": 2.7101896225568744e-05, + "loss": 0.4295, + "step": 1620 + }, + { + "epoch": 3.2226640159045727, + "grad_norm": 0.14091520617084688, + "learning_rate": 2.704934505808389e-05, + "loss": 0.4178, + "step": 1621 + }, + { + "epoch": 3.224652087475149, + "grad_norm": 0.14513943134763385, + "learning_rate": 2.6996818849318508e-05, + "loss": 0.4183, + "step": 1622 + }, + { + "epoch": 3.2266401590457257, + "grad_norm": 0.14432815766342816, + "learning_rate": 2.6944317700501992e-05, + "loss": 0.4166, + "step": 1623 + }, + { + "epoch": 3.228628230616302, + "grad_norm": 0.16739958916754763, + "learning_rate": 2.6891841712815448e-05, + "loss": 0.4298, + "step": 1624 + }, + { + "epoch": 3.2306163021868786, + "grad_norm": 0.17632490104186138, + "learning_rate": 2.6839390987391475e-05, + "loss": 0.4151, + "step": 1625 + }, + { + "epoch": 3.2326043737574555, + "grad_norm": 0.1701061576690492, + "learning_rate": 2.678696562531398e-05, + "loss": 0.4186, + "step": 1626 + }, + { + "epoch": 3.234592445328032, + "grad_norm": 0.14923451020759781, + "learning_rate": 2.6734565727618016e-05, + "loss": 0.4226, + "step": 1627 + }, + { + "epoch": 3.2365805168986084, + "grad_norm": 0.1402483490014172, + "learning_rate": 2.6682191395289537e-05, + "loss": 0.4277, + "step": 1628 + }, + { + "epoch": 3.238568588469185, + "grad_norm": 0.21580075521102834, + "learning_rate": 2.6629842729265246e-05, + "loss": 0.4096, + "step": 1629 + }, + { + "epoch": 3.2405566600397613, + "grad_norm": 0.24086136258461305, + "learning_rate": 2.6577519830432365e-05, + "loss": 0.4206, + "step": 1630 + }, + { + "epoch": 3.242544731610338, + "grad_norm": 0.17137383827651592, + "learning_rate": 2.6525222799628455e-05, + "loss": 0.4153, + "step": 1631 + }, + { + "epoch": 3.2445328031809146, + "grad_norm": 0.27384042950995446, + "learning_rate": 2.647295173764123e-05, + "loss": 0.4216, + "step": 1632 + }, + { + "epoch": 3.246520874751491, + "grad_norm": 0.23612530448432742, + "learning_rate": 2.6420706745208368e-05, + "loss": 0.4158, + "step": 1633 + }, + { + "epoch": 3.2485089463220675, + "grad_norm": 0.19382473812414455, + "learning_rate": 2.6368487923017286e-05, + "loss": 0.4225, + "step": 1634 + }, + { + "epoch": 3.250497017892644, + "grad_norm": 0.2945482875629765, + "learning_rate": 2.631629537170498e-05, + "loss": 0.4223, + "step": 1635 + }, + { + "epoch": 3.252485089463221, + "grad_norm": 0.1557550330207868, + "learning_rate": 2.6264129191857805e-05, + "loss": 0.4145, + "step": 1636 + }, + { + "epoch": 3.2544731610337974, + "grad_norm": 0.2396368267625787, + "learning_rate": 2.62119894840113e-05, + "loss": 0.416, + "step": 1637 + }, + { + "epoch": 3.256461232604374, + "grad_norm": 0.17272951015488022, + "learning_rate": 2.6159876348649994e-05, + "loss": 0.4314, + "step": 1638 + }, + { + "epoch": 3.2584493041749503, + "grad_norm": 0.1765073891960342, + "learning_rate": 2.6107789886207195e-05, + "loss": 0.416, + "step": 1639 + }, + { + "epoch": 3.2604373757455267, + "grad_norm": 0.19928549933644896, + "learning_rate": 2.605573019706479e-05, + "loss": 0.4262, + "step": 1640 + }, + { + "epoch": 3.2624254473161036, + "grad_norm": 0.16788212491095592, + "learning_rate": 2.60036973815531e-05, + "loss": 0.4174, + "step": 1641 + }, + { + "epoch": 3.26441351888668, + "grad_norm": 0.2014485731227408, + "learning_rate": 2.5951691539950635e-05, + "loss": 0.4223, + "step": 1642 + }, + { + "epoch": 3.2664015904572565, + "grad_norm": 0.19786684925638573, + "learning_rate": 2.5899712772483926e-05, + "loss": 0.42, + "step": 1643 + }, + { + "epoch": 3.268389662027833, + "grad_norm": 0.17771485372367102, + "learning_rate": 2.5847761179327334e-05, + "loss": 0.417, + "step": 1644 + }, + { + "epoch": 3.2703777335984094, + "grad_norm": 0.18677854197656066, + "learning_rate": 2.5795836860602808e-05, + "loss": 0.4267, + "step": 1645 + }, + { + "epoch": 3.272365805168986, + "grad_norm": 0.15790603246827162, + "learning_rate": 2.5743939916379805e-05, + "loss": 0.4179, + "step": 1646 + }, + { + "epoch": 3.274353876739563, + "grad_norm": 0.1768610231628309, + "learning_rate": 2.569207044667498e-05, + "loss": 0.4275, + "step": 1647 + }, + { + "epoch": 3.2763419483101393, + "grad_norm": 0.2092056932711368, + "learning_rate": 2.564022855145204e-05, + "loss": 0.4223, + "step": 1648 + }, + { + "epoch": 3.2783300198807157, + "grad_norm": 0.1784649030856477, + "learning_rate": 2.558841433062155e-05, + "loss": 0.4217, + "step": 1649 + }, + { + "epoch": 3.280318091451292, + "grad_norm": 0.16695975013592268, + "learning_rate": 2.5536627884040748e-05, + "loss": 0.4258, + "step": 1650 + }, + { + "epoch": 3.2823061630218686, + "grad_norm": 0.16257746343716772, + "learning_rate": 2.548486931151334e-05, + "loss": 0.4146, + "step": 1651 + }, + { + "epoch": 3.2842942345924455, + "grad_norm": 0.16035482107017218, + "learning_rate": 2.5433138712789324e-05, + "loss": 0.4263, + "step": 1652 + }, + { + "epoch": 3.286282306163022, + "grad_norm": 0.1668332300916889, + "learning_rate": 2.5381436187564776e-05, + "loss": 0.4271, + "step": 1653 + }, + { + "epoch": 3.2882703777335984, + "grad_norm": 0.18765944464929687, + "learning_rate": 2.5329761835481647e-05, + "loss": 0.4305, + "step": 1654 + }, + { + "epoch": 3.290258449304175, + "grad_norm": 0.22260684778448311, + "learning_rate": 2.5278115756127646e-05, + "loss": 0.4158, + "step": 1655 + }, + { + "epoch": 3.2922465208747513, + "grad_norm": 0.175348581154186, + "learning_rate": 2.5226498049035958e-05, + "loss": 0.4187, + "step": 1656 + }, + { + "epoch": 3.294234592445328, + "grad_norm": 0.137384946554788, + "learning_rate": 2.5174908813685094e-05, + "loss": 0.4262, + "step": 1657 + }, + { + "epoch": 3.2962226640159047, + "grad_norm": 0.2289530209064395, + "learning_rate": 2.5123348149498684e-05, + "loss": 0.4176, + "step": 1658 + }, + { + "epoch": 3.298210735586481, + "grad_norm": 0.24077183815627887, + "learning_rate": 2.5071816155845304e-05, + "loss": 0.4232, + "step": 1659 + }, + { + "epoch": 3.3001988071570576, + "grad_norm": 0.1515863899105361, + "learning_rate": 2.5020312932038285e-05, + "loss": 0.4219, + "step": 1660 + }, + { + "epoch": 3.302186878727634, + "grad_norm": 0.22154638035278512, + "learning_rate": 2.4968838577335507e-05, + "loss": 0.4218, + "step": 1661 + }, + { + "epoch": 3.3041749502982105, + "grad_norm": 0.2787216556594714, + "learning_rate": 2.491739319093919e-05, + "loss": 0.4189, + "step": 1662 + }, + { + "epoch": 3.3061630218687874, + "grad_norm": 0.15341434669320073, + "learning_rate": 2.486597687199577e-05, + "loss": 0.4101, + "step": 1663 + }, + { + "epoch": 3.308151093439364, + "grad_norm": 0.18926114022594837, + "learning_rate": 2.481458971959563e-05, + "loss": 0.4279, + "step": 1664 + }, + { + "epoch": 3.3101391650099403, + "grad_norm": 0.15294495380705728, + "learning_rate": 2.4763231832772966e-05, + "loss": 0.4194, + "step": 1665 + }, + { + "epoch": 3.3121272365805168, + "grad_norm": 0.15660817811423164, + "learning_rate": 2.4711903310505525e-05, + "loss": 0.421, + "step": 1666 + }, + { + "epoch": 3.3141153081510932, + "grad_norm": 0.162337406875044, + "learning_rate": 2.4660604251714526e-05, + "loss": 0.4233, + "step": 1667 + }, + { + "epoch": 3.31610337972167, + "grad_norm": 0.14327890979444038, + "learning_rate": 2.460933475526436e-05, + "loss": 0.4118, + "step": 1668 + }, + { + "epoch": 3.3180914512922466, + "grad_norm": 0.14965507258044433, + "learning_rate": 2.455809491996247e-05, + "loss": 0.4228, + "step": 1669 + }, + { + "epoch": 3.320079522862823, + "grad_norm": 0.14560822410161192, + "learning_rate": 2.4506884844559124e-05, + "loss": 0.4148, + "step": 1670 + }, + { + "epoch": 3.3220675944333995, + "grad_norm": 0.1426400454941638, + "learning_rate": 2.445570462774723e-05, + "loss": 0.4297, + "step": 1671 + }, + { + "epoch": 3.324055666003976, + "grad_norm": 0.1569048324686495, + "learning_rate": 2.4404554368162177e-05, + "loss": 0.4142, + "step": 1672 + }, + { + "epoch": 3.326043737574553, + "grad_norm": 0.1643392865084774, + "learning_rate": 2.4353434164381593e-05, + "loss": 0.4206, + "step": 1673 + }, + { + "epoch": 3.3280318091451293, + "grad_norm": 0.1498429617187632, + "learning_rate": 2.4302344114925203e-05, + "loss": 0.4137, + "step": 1674 + }, + { + "epoch": 3.3300198807157058, + "grad_norm": 0.13721526264782666, + "learning_rate": 2.425128431825459e-05, + "loss": 0.4212, + "step": 1675 + }, + { + "epoch": 3.332007952286282, + "grad_norm": 0.1600371084528462, + "learning_rate": 2.4200254872773056e-05, + "loss": 0.4193, + "step": 1676 + }, + { + "epoch": 3.3339960238568587, + "grad_norm": 0.15972808411829598, + "learning_rate": 2.4149255876825413e-05, + "loss": 0.4204, + "step": 1677 + }, + { + "epoch": 3.3359840954274356, + "grad_norm": 0.19172287503702923, + "learning_rate": 2.409828742869777e-05, + "loss": 0.4233, + "step": 1678 + }, + { + "epoch": 3.337972166998012, + "grad_norm": 0.17154708460144463, + "learning_rate": 2.4047349626617383e-05, + "loss": 0.4269, + "step": 1679 + }, + { + "epoch": 3.3399602385685885, + "grad_norm": 0.14615085908668787, + "learning_rate": 2.3996442568752434e-05, + "loss": 0.4281, + "step": 1680 + }, + { + "epoch": 3.341948310139165, + "grad_norm": 0.16245718655783098, + "learning_rate": 2.3945566353211872e-05, + "loss": 0.4192, + "step": 1681 + }, + { + "epoch": 3.3439363817097414, + "grad_norm": 0.13505783886664183, + "learning_rate": 2.389472107804518e-05, + "loss": 0.4072, + "step": 1682 + }, + { + "epoch": 3.3459244532803183, + "grad_norm": 0.14338107591716415, + "learning_rate": 2.3843906841242234e-05, + "loss": 0.4165, + "step": 1683 + }, + { + "epoch": 3.3479125248508947, + "grad_norm": 0.13551779503546382, + "learning_rate": 2.3793123740733074e-05, + "loss": 0.419, + "step": 1684 + }, + { + "epoch": 3.349900596421471, + "grad_norm": 0.18878082132312443, + "learning_rate": 2.374237187438775e-05, + "loss": 0.4262, + "step": 1685 + }, + { + "epoch": 3.3518886679920477, + "grad_norm": 0.19720983891195784, + "learning_rate": 2.3691651340016106e-05, + "loss": 0.413, + "step": 1686 + }, + { + "epoch": 3.353876739562624, + "grad_norm": 0.15864169989169807, + "learning_rate": 2.364096223536761e-05, + "loss": 0.4127, + "step": 1687 + }, + { + "epoch": 3.355864811133201, + "grad_norm": 0.1965166357303236, + "learning_rate": 2.359030465813115e-05, + "loss": 0.4207, + "step": 1688 + }, + { + "epoch": 3.3578528827037775, + "grad_norm": 0.38613530552613956, + "learning_rate": 2.3539678705934854e-05, + "loss": 0.4317, + "step": 1689 + }, + { + "epoch": 3.359840954274354, + "grad_norm": 0.18727427096552424, + "learning_rate": 2.348908447634593e-05, + "loss": 0.417, + "step": 1690 + }, + { + "epoch": 3.3618290258449304, + "grad_norm": 0.17011695468121943, + "learning_rate": 2.3438522066870405e-05, + "loss": 0.4196, + "step": 1691 + }, + { + "epoch": 3.363817097415507, + "grad_norm": 0.18341050658376595, + "learning_rate": 2.338799157495302e-05, + "loss": 0.4202, + "step": 1692 + }, + { + "epoch": 3.3658051689860837, + "grad_norm": 0.1775335051454464, + "learning_rate": 2.333749309797696e-05, + "loss": 0.4177, + "step": 1693 + }, + { + "epoch": 3.36779324055666, + "grad_norm": 0.15602473128210675, + "learning_rate": 2.328702673326376e-05, + "loss": 0.4246, + "step": 1694 + }, + { + "epoch": 3.3697813121272366, + "grad_norm": 0.16113387149141137, + "learning_rate": 2.3236592578073056e-05, + "loss": 0.4221, + "step": 1695 + }, + { + "epoch": 3.371769383697813, + "grad_norm": 0.14998557774497948, + "learning_rate": 2.3186190729602364e-05, + "loss": 0.4279, + "step": 1696 + }, + { + "epoch": 3.3737574552683895, + "grad_norm": 0.15917274374720317, + "learning_rate": 2.3135821284986983e-05, + "loss": 0.4232, + "step": 1697 + }, + { + "epoch": 3.3757455268389664, + "grad_norm": 0.17516252336995275, + "learning_rate": 2.3085484341299743e-05, + "loss": 0.4235, + "step": 1698 + }, + { + "epoch": 3.377733598409543, + "grad_norm": 0.1515880856245295, + "learning_rate": 2.3035179995550876e-05, + "loss": 0.4208, + "step": 1699 + }, + { + "epoch": 3.3797216699801194, + "grad_norm": 0.19601830982433405, + "learning_rate": 2.2984908344687745e-05, + "loss": 0.4236, + "step": 1700 + }, + { + "epoch": 3.381709741550696, + "grad_norm": 0.18377008019381375, + "learning_rate": 2.2934669485594716e-05, + "loss": 0.4211, + "step": 1701 + }, + { + "epoch": 3.3836978131212723, + "grad_norm": 0.16561826986439118, + "learning_rate": 2.288446351509297e-05, + "loss": 0.4262, + "step": 1702 + }, + { + "epoch": 3.3856858846918487, + "grad_norm": 0.21353955552866088, + "learning_rate": 2.2834290529940284e-05, + "loss": 0.4242, + "step": 1703 + }, + { + "epoch": 3.3876739562624256, + "grad_norm": 0.14923813005518957, + "learning_rate": 2.2784150626830898e-05, + "loss": 0.4217, + "step": 1704 + }, + { + "epoch": 3.389662027833002, + "grad_norm": 0.24069279516554903, + "learning_rate": 2.2734043902395245e-05, + "loss": 0.4176, + "step": 1705 + }, + { + "epoch": 3.3916500994035785, + "grad_norm": 0.1702158291378274, + "learning_rate": 2.2683970453199865e-05, + "loss": 0.4231, + "step": 1706 + }, + { + "epoch": 3.393638170974155, + "grad_norm": 0.17350287526667996, + "learning_rate": 2.2633930375747118e-05, + "loss": 0.4183, + "step": 1707 + }, + { + "epoch": 3.3956262425447314, + "grad_norm": 0.23437538422451198, + "learning_rate": 2.2583923766475125e-05, + "loss": 0.4178, + "step": 1708 + }, + { + "epoch": 3.3976143141153083, + "grad_norm": 0.14198550996726472, + "learning_rate": 2.2533950721757447e-05, + "loss": 0.414, + "step": 1709 + }, + { + "epoch": 3.399602385685885, + "grad_norm": 0.2102399025316218, + "learning_rate": 2.2484011337902968e-05, + "loss": 0.424, + "step": 1710 + }, + { + "epoch": 3.4015904572564613, + "grad_norm": 0.1576665890073455, + "learning_rate": 2.2434105711155706e-05, + "loss": 0.4262, + "step": 1711 + }, + { + "epoch": 3.4035785288270377, + "grad_norm": 0.2942487214647327, + "learning_rate": 2.2384233937694626e-05, + "loss": 0.4138, + "step": 1712 + }, + { + "epoch": 3.405566600397614, + "grad_norm": 0.1734635276860347, + "learning_rate": 2.2334396113633464e-05, + "loss": 0.4212, + "step": 1713 + }, + { + "epoch": 3.4075546719681906, + "grad_norm": 0.20395877825842865, + "learning_rate": 2.2284592335020475e-05, + "loss": 0.4275, + "step": 1714 + }, + { + "epoch": 3.4095427435387675, + "grad_norm": 0.18126084824037503, + "learning_rate": 2.223482269783834e-05, + "loss": 0.4182, + "step": 1715 + }, + { + "epoch": 3.411530815109344, + "grad_norm": 0.15914404766557488, + "learning_rate": 2.2185087298003968e-05, + "loss": 0.4118, + "step": 1716 + }, + { + "epoch": 3.4135188866799204, + "grad_norm": 0.1590304746258962, + "learning_rate": 2.213538623136824e-05, + "loss": 0.4172, + "step": 1717 + }, + { + "epoch": 3.415506958250497, + "grad_norm": 0.17825298578092894, + "learning_rate": 2.2085719593715886e-05, + "loss": 0.4132, + "step": 1718 + }, + { + "epoch": 3.4174950298210733, + "grad_norm": 0.16302076276162225, + "learning_rate": 2.203608748076528e-05, + "loss": 0.4169, + "step": 1719 + }, + { + "epoch": 3.4194831013916502, + "grad_norm": 0.17318949422352425, + "learning_rate": 2.198648998816827e-05, + "loss": 0.4187, + "step": 1720 + }, + { + "epoch": 3.4214711729622267, + "grad_norm": 0.15012555647563922, + "learning_rate": 2.193692721150998e-05, + "loss": 0.4226, + "step": 1721 + }, + { + "epoch": 3.423459244532803, + "grad_norm": 0.18815489284348785, + "learning_rate": 2.1887399246308623e-05, + "loss": 0.413, + "step": 1722 + }, + { + "epoch": 3.4254473161033796, + "grad_norm": 0.13978612071322793, + "learning_rate": 2.1837906188015303e-05, + "loss": 0.4227, + "step": 1723 + }, + { + "epoch": 3.427435387673956, + "grad_norm": 0.19168527604634417, + "learning_rate": 2.1788448132013874e-05, + "loss": 0.4249, + "step": 1724 + }, + { + "epoch": 3.429423459244533, + "grad_norm": 0.1535627672588957, + "learning_rate": 2.173902517362076e-05, + "loss": 0.4247, + "step": 1725 + }, + { + "epoch": 3.4314115308151094, + "grad_norm": 0.13874613052414875, + "learning_rate": 2.1689637408084703e-05, + "loss": 0.418, + "step": 1726 + }, + { + "epoch": 3.433399602385686, + "grad_norm": 0.13587604761472546, + "learning_rate": 2.1640284930586635e-05, + "loss": 0.417, + "step": 1727 + }, + { + "epoch": 3.4353876739562623, + "grad_norm": 0.16274376236534926, + "learning_rate": 2.1590967836239464e-05, + "loss": 0.4091, + "step": 1728 + }, + { + "epoch": 3.4373757455268388, + "grad_norm": 0.1389554882025859, + "learning_rate": 2.1541686220087938e-05, + "loss": 0.425, + "step": 1729 + }, + { + "epoch": 3.4393638170974157, + "grad_norm": 0.15567485102640824, + "learning_rate": 2.1492440177108423e-05, + "loss": 0.4211, + "step": 1730 + }, + { + "epoch": 3.441351888667992, + "grad_norm": 0.14753218577823196, + "learning_rate": 2.144322980220869e-05, + "loss": 0.4223, + "step": 1731 + }, + { + "epoch": 3.4433399602385686, + "grad_norm": 0.1458705887737164, + "learning_rate": 2.1394055190227814e-05, + "loss": 0.4319, + "step": 1732 + }, + { + "epoch": 3.445328031809145, + "grad_norm": 0.1397058371801398, + "learning_rate": 2.1344916435935915e-05, + "loss": 0.4113, + "step": 1733 + }, + { + "epoch": 3.4473161033797215, + "grad_norm": 0.14684655321915865, + "learning_rate": 2.129581363403406e-05, + "loss": 0.4237, + "step": 1734 + }, + { + "epoch": 3.4493041749502984, + "grad_norm": 0.17704552353843994, + "learning_rate": 2.124674687915398e-05, + "loss": 0.4226, + "step": 1735 + }, + { + "epoch": 3.451292246520875, + "grad_norm": 0.14030210704985885, + "learning_rate": 2.1197716265857932e-05, + "loss": 0.4188, + "step": 1736 + }, + { + "epoch": 3.4532803180914513, + "grad_norm": 0.17008441652493428, + "learning_rate": 2.1148721888638556e-05, + "loss": 0.4168, + "step": 1737 + }, + { + "epoch": 3.4552683896620278, + "grad_norm": 0.1836350292563087, + "learning_rate": 2.1099763841918626e-05, + "loss": 0.413, + "step": 1738 + }, + { + "epoch": 3.457256461232604, + "grad_norm": 0.4358843485324826, + "learning_rate": 2.105084222005093e-05, + "loss": 0.4398, + "step": 1739 + }, + { + "epoch": 3.459244532803181, + "grad_norm": 0.19076079127469725, + "learning_rate": 2.1001957117318002e-05, + "loss": 0.4156, + "step": 1740 + }, + { + "epoch": 3.4612326043737576, + "grad_norm": 0.15919930414992362, + "learning_rate": 2.095310862793206e-05, + "loss": 0.4207, + "step": 1741 + }, + { + "epoch": 3.463220675944334, + "grad_norm": 0.20473446973024545, + "learning_rate": 2.0904296846034703e-05, + "loss": 0.4174, + "step": 1742 + }, + { + "epoch": 3.4652087475149105, + "grad_norm": 0.15137987512270168, + "learning_rate": 2.085552186569686e-05, + "loss": 0.4117, + "step": 1743 + }, + { + "epoch": 3.467196819085487, + "grad_norm": 0.18074067785558987, + "learning_rate": 2.0806783780918467e-05, + "loss": 0.4092, + "step": 1744 + }, + { + "epoch": 3.469184890656064, + "grad_norm": 0.13226943416324957, + "learning_rate": 2.0758082685628377e-05, + "loss": 0.4146, + "step": 1745 + }, + { + "epoch": 3.4711729622266403, + "grad_norm": 0.16993781310376027, + "learning_rate": 2.0709418673684166e-05, + "loss": 0.4224, + "step": 1746 + }, + { + "epoch": 3.4731610337972167, + "grad_norm": 0.15112977680745895, + "learning_rate": 2.066079183887193e-05, + "loss": 0.4038, + "step": 1747 + }, + { + "epoch": 3.475149105367793, + "grad_norm": 0.15100626586795593, + "learning_rate": 2.0612202274906137e-05, + "loss": 0.4081, + "step": 1748 + }, + { + "epoch": 3.4771371769383697, + "grad_norm": 1.9680321094660187, + "learning_rate": 2.0563650075429376e-05, + "loss": 0.4662, + "step": 1749 + }, + { + "epoch": 3.4791252485089466, + "grad_norm": 0.19286841860615947, + "learning_rate": 2.0515135334012283e-05, + "loss": 0.4157, + "step": 1750 + }, + { + "epoch": 3.481113320079523, + "grad_norm": 0.14764009814938028, + "learning_rate": 2.046665814415326e-05, + "loss": 0.4188, + "step": 1751 + }, + { + "epoch": 3.4831013916500995, + "grad_norm": 0.18319314928246067, + "learning_rate": 2.0418218599278388e-05, + "loss": 0.4331, + "step": 1752 + }, + { + "epoch": 3.485089463220676, + "grad_norm": 0.14820529438687846, + "learning_rate": 2.036981679274117e-05, + "loss": 0.4261, + "step": 1753 + }, + { + "epoch": 3.4870775347912524, + "grad_norm": 0.1497754362666427, + "learning_rate": 2.0321452817822363e-05, + "loss": 0.414, + "step": 1754 + }, + { + "epoch": 3.4890656063618293, + "grad_norm": 0.15919162663516262, + "learning_rate": 2.027312676772984e-05, + "loss": 0.4297, + "step": 1755 + }, + { + "epoch": 3.4910536779324057, + "grad_norm": 0.1758929970525375, + "learning_rate": 2.0224838735598374e-05, + "loss": 0.4187, + "step": 1756 + }, + { + "epoch": 3.493041749502982, + "grad_norm": 0.1452678866846859, + "learning_rate": 2.0176588814489477e-05, + "loss": 0.4241, + "step": 1757 + }, + { + "epoch": 3.4950298210735586, + "grad_norm": 0.1514360850159552, + "learning_rate": 2.0128377097391193e-05, + "loss": 0.4169, + "step": 1758 + }, + { + "epoch": 3.497017892644135, + "grad_norm": 0.145255134633476, + "learning_rate": 2.0080203677217955e-05, + "loss": 0.4127, + "step": 1759 + }, + { + "epoch": 3.4990059642147116, + "grad_norm": 0.1548974716711143, + "learning_rate": 2.0032068646810378e-05, + "loss": 0.4199, + "step": 1760 + }, + { + "epoch": 3.500994035785288, + "grad_norm": 0.1506452726505713, + "learning_rate": 1.9983972098935122e-05, + "loss": 0.4188, + "step": 1761 + }, + { + "epoch": 3.502982107355865, + "grad_norm": 0.14658095527849782, + "learning_rate": 1.9935914126284655e-05, + "loss": 0.414, + "step": 1762 + }, + { + "epoch": 3.5049701789264414, + "grad_norm": 0.14632438106244663, + "learning_rate": 1.9887894821477106e-05, + "loss": 0.4258, + "step": 1763 + }, + { + "epoch": 3.506958250497018, + "grad_norm": 0.14429915729010442, + "learning_rate": 1.983991427705608e-05, + "loss": 0.4144, + "step": 1764 + }, + { + "epoch": 3.5089463220675943, + "grad_norm": 0.1429848169311897, + "learning_rate": 1.979197258549049e-05, + "loss": 0.4127, + "step": 1765 + }, + { + "epoch": 3.5109343936381707, + "grad_norm": 0.1676451471694739, + "learning_rate": 1.9744069839174384e-05, + "loss": 0.4153, + "step": 1766 + }, + { + "epoch": 3.5129224652087476, + "grad_norm": 0.14066248382152913, + "learning_rate": 1.969620613042671e-05, + "loss": 0.4185, + "step": 1767 + }, + { + "epoch": 3.514910536779324, + "grad_norm": 0.16764366472616893, + "learning_rate": 1.9648381551491224e-05, + "loss": 0.4141, + "step": 1768 + }, + { + "epoch": 3.5168986083499005, + "grad_norm": 0.14003669923848128, + "learning_rate": 1.960059619453624e-05, + "loss": 0.4213, + "step": 1769 + }, + { + "epoch": 3.518886679920477, + "grad_norm": 0.14366167229047724, + "learning_rate": 1.955285015165453e-05, + "loss": 0.4206, + "step": 1770 + }, + { + "epoch": 3.5208747514910534, + "grad_norm": 0.15242720064957377, + "learning_rate": 1.9505143514863056e-05, + "loss": 0.4009, + "step": 1771 + }, + { + "epoch": 3.5228628230616303, + "grad_norm": 0.1785700156717405, + "learning_rate": 1.945747637610285e-05, + "loss": 0.4224, + "step": 1772 + }, + { + "epoch": 3.524850894632207, + "grad_norm": 0.1529136898284363, + "learning_rate": 1.9409848827238814e-05, + "loss": 0.4189, + "step": 1773 + }, + { + "epoch": 3.5268389662027833, + "grad_norm": 0.15141172921996104, + "learning_rate": 1.936226096005957e-05, + "loss": 0.4175, + "step": 1774 + }, + { + "epoch": 3.5288270377733597, + "grad_norm": 0.1440969504372814, + "learning_rate": 1.9314712866277264e-05, + "loss": 0.4084, + "step": 1775 + }, + { + "epoch": 3.530815109343936, + "grad_norm": 0.14316810507788563, + "learning_rate": 1.9267204637527345e-05, + "loss": 0.4218, + "step": 1776 + }, + { + "epoch": 3.532803180914513, + "grad_norm": 0.14564934266921534, + "learning_rate": 1.921973636536848e-05, + "loss": 0.4051, + "step": 1777 + }, + { + "epoch": 3.5347912524850895, + "grad_norm": 0.25175630532587917, + "learning_rate": 1.9172308141282336e-05, + "loss": 0.4207, + "step": 1778 + }, + { + "epoch": 3.536779324055666, + "grad_norm": 0.14020457933643266, + "learning_rate": 1.9124920056673374e-05, + "loss": 0.4193, + "step": 1779 + }, + { + "epoch": 3.5387673956262424, + "grad_norm": 0.15928874940395493, + "learning_rate": 1.907757220286871e-05, + "loss": 0.4132, + "step": 1780 + }, + { + "epoch": 3.540755467196819, + "grad_norm": 0.15522223883560438, + "learning_rate": 1.9030264671117914e-05, + "loss": 0.4199, + "step": 1781 + }, + { + "epoch": 3.542743538767396, + "grad_norm": 0.24239762715881097, + "learning_rate": 1.8982997552592863e-05, + "loss": 0.4183, + "step": 1782 + }, + { + "epoch": 3.5447316103379722, + "grad_norm": 0.1660873246133442, + "learning_rate": 1.893577093838753e-05, + "loss": 0.4242, + "step": 1783 + }, + { + "epoch": 3.5467196819085487, + "grad_norm": 0.16085140583809687, + "learning_rate": 1.8888584919517867e-05, + "loss": 0.4177, + "step": 1784 + }, + { + "epoch": 3.548707753479125, + "grad_norm": 0.16069903654575152, + "learning_rate": 1.8841439586921515e-05, + "loss": 0.4182, + "step": 1785 + }, + { + "epoch": 3.5506958250497016, + "grad_norm": 0.15442226396451475, + "learning_rate": 1.8794335031457763e-05, + "loss": 0.4168, + "step": 1786 + }, + { + "epoch": 3.5526838966202785, + "grad_norm": 0.15050396138925257, + "learning_rate": 1.874727134390733e-05, + "loss": 0.4196, + "step": 1787 + }, + { + "epoch": 3.554671968190855, + "grad_norm": 0.15316322830236687, + "learning_rate": 1.8700248614972122e-05, + "loss": 0.4261, + "step": 1788 + }, + { + "epoch": 3.5566600397614314, + "grad_norm": 0.15262385894358507, + "learning_rate": 1.865326693527514e-05, + "loss": 0.4236, + "step": 1789 + }, + { + "epoch": 3.558648111332008, + "grad_norm": 0.13953426409619946, + "learning_rate": 1.8606326395360254e-05, + "loss": 0.4278, + "step": 1790 + }, + { + "epoch": 3.5606361829025843, + "grad_norm": 0.1453249229248661, + "learning_rate": 1.8559427085692068e-05, + "loss": 0.4151, + "step": 1791 + }, + { + "epoch": 3.5626242544731612, + "grad_norm": 0.13768057292816951, + "learning_rate": 1.8512569096655717e-05, + "loss": 0.4112, + "step": 1792 + }, + { + "epoch": 3.5646123260437377, + "grad_norm": 0.15733429051521475, + "learning_rate": 1.8465752518556715e-05, + "loss": 0.4106, + "step": 1793 + }, + { + "epoch": 3.566600397614314, + "grad_norm": 0.15307708019233576, + "learning_rate": 1.8418977441620737e-05, + "loss": 0.4197, + "step": 1794 + }, + { + "epoch": 3.5685884691848906, + "grad_norm": 0.14136017029964, + "learning_rate": 1.8372243955993484e-05, + "loss": 0.4229, + "step": 1795 + }, + { + "epoch": 3.570576540755467, + "grad_norm": 0.14308762935160746, + "learning_rate": 1.8325552151740553e-05, + "loss": 0.4156, + "step": 1796 + }, + { + "epoch": 3.572564612326044, + "grad_norm": 0.14090812672312653, + "learning_rate": 1.827890211884715e-05, + "loss": 0.4139, + "step": 1797 + }, + { + "epoch": 3.5745526838966204, + "grad_norm": 0.1450916042416356, + "learning_rate": 1.8232293947218004e-05, + "loss": 0.4116, + "step": 1798 + }, + { + "epoch": 3.576540755467197, + "grad_norm": 0.14762435550768865, + "learning_rate": 1.8185727726677167e-05, + "loss": 0.4312, + "step": 1799 + }, + { + "epoch": 3.5785288270377733, + "grad_norm": 0.15042990698055161, + "learning_rate": 1.8139203546967835e-05, + "loss": 0.421, + "step": 1800 + }, + { + "epoch": 3.5805168986083498, + "grad_norm": 0.14061082501100122, + "learning_rate": 1.8092721497752194e-05, + "loss": 0.4236, + "step": 1801 + }, + { + "epoch": 3.5825049701789267, + "grad_norm": 0.14111997641843832, + "learning_rate": 1.8046281668611242e-05, + "loss": 0.4196, + "step": 1802 + }, + { + "epoch": 3.584493041749503, + "grad_norm": 0.16047052141744386, + "learning_rate": 1.799988414904457e-05, + "loss": 0.413, + "step": 1803 + }, + { + "epoch": 3.5864811133200796, + "grad_norm": 0.14324152977090976, + "learning_rate": 1.795352902847025e-05, + "loss": 0.4214, + "step": 1804 + }, + { + "epoch": 3.588469184890656, + "grad_norm": 0.1637034156499913, + "learning_rate": 1.790721639622468e-05, + "loss": 0.42, + "step": 1805 + }, + { + "epoch": 3.5904572564612325, + "grad_norm": 0.14535956431472496, + "learning_rate": 1.786094634156233e-05, + "loss": 0.4217, + "step": 1806 + }, + { + "epoch": 3.5924453280318094, + "grad_norm": 0.15532983320228866, + "learning_rate": 1.781471895365562e-05, + "loss": 0.4142, + "step": 1807 + }, + { + "epoch": 3.594433399602386, + "grad_norm": 0.1648501719516284, + "learning_rate": 1.7768534321594755e-05, + "loss": 0.422, + "step": 1808 + }, + { + "epoch": 3.5964214711729623, + "grad_norm": 0.13739363913925556, + "learning_rate": 1.7722392534387526e-05, + "loss": 0.4195, + "step": 1809 + }, + { + "epoch": 3.5984095427435387, + "grad_norm": 0.1684875509378481, + "learning_rate": 1.7676293680959178e-05, + "loss": 0.4154, + "step": 1810 + }, + { + "epoch": 3.600397614314115, + "grad_norm": 0.15049792505533202, + "learning_rate": 1.763023785015216e-05, + "loss": 0.4283, + "step": 1811 + }, + { + "epoch": 3.602385685884692, + "grad_norm": 0.14341126106705412, + "learning_rate": 1.7584225130726066e-05, + "loss": 0.4148, + "step": 1812 + }, + { + "epoch": 3.604373757455268, + "grad_norm": 0.141627160897845, + "learning_rate": 1.7538255611357358e-05, + "loss": 0.4267, + "step": 1813 + }, + { + "epoch": 3.606361829025845, + "grad_norm": 0.1574393061219025, + "learning_rate": 1.7492329380639298e-05, + "loss": 0.4216, + "step": 1814 + }, + { + "epoch": 3.6083499005964215, + "grad_norm": 0.13802470610512332, + "learning_rate": 1.7446446527081673e-05, + "loss": 0.4178, + "step": 1815 + }, + { + "epoch": 3.610337972166998, + "grad_norm": 0.15359173777670673, + "learning_rate": 1.7400607139110687e-05, + "loss": 0.4098, + "step": 1816 + }, + { + "epoch": 3.612326043737575, + "grad_norm": 0.13739662485676052, + "learning_rate": 1.735481130506878e-05, + "loss": 0.4125, + "step": 1817 + }, + { + "epoch": 3.614314115308151, + "grad_norm": 0.1671537840554596, + "learning_rate": 1.730905911321445e-05, + "loss": 0.4167, + "step": 1818 + }, + { + "epoch": 3.6163021868787277, + "grad_norm": 0.15557775501383705, + "learning_rate": 1.7263350651722103e-05, + "loss": 0.4165, + "step": 1819 + }, + { + "epoch": 3.618290258449304, + "grad_norm": 0.14372860836642484, + "learning_rate": 1.7217686008681825e-05, + "loss": 0.4157, + "step": 1820 + }, + { + "epoch": 3.6202783300198806, + "grad_norm": 0.22155274518208556, + "learning_rate": 1.7172065272099297e-05, + "loss": 0.421, + "step": 1821 + }, + { + "epoch": 3.622266401590457, + "grad_norm": 0.1305430733649789, + "learning_rate": 1.7126488529895548e-05, + "loss": 0.4057, + "step": 1822 + }, + { + "epoch": 3.6242544731610336, + "grad_norm": 0.15254680452691538, + "learning_rate": 1.7080955869906874e-05, + "loss": 0.4159, + "step": 1823 + }, + { + "epoch": 3.6262425447316105, + "grad_norm": 0.14049400385188401, + "learning_rate": 1.703546737988457e-05, + "loss": 0.4142, + "step": 1824 + }, + { + "epoch": 3.628230616302187, + "grad_norm": 0.13644463312609267, + "learning_rate": 1.6990023147494814e-05, + "loss": 0.4226, + "step": 1825 + }, + { + "epoch": 3.6302186878727634, + "grad_norm": 0.17220133648201097, + "learning_rate": 1.694462326031849e-05, + "loss": 0.4138, + "step": 1826 + }, + { + "epoch": 3.63220675944334, + "grad_norm": 0.14493215532450543, + "learning_rate": 1.6899267805851028e-05, + "loss": 0.4186, + "step": 1827 + }, + { + "epoch": 3.6341948310139163, + "grad_norm": 0.20593834230752084, + "learning_rate": 1.685395687150223e-05, + "loss": 0.4155, + "step": 1828 + }, + { + "epoch": 3.636182902584493, + "grad_norm": 0.14491486894512906, + "learning_rate": 1.6808690544596062e-05, + "loss": 0.4125, + "step": 1829 + }, + { + "epoch": 3.6381709741550696, + "grad_norm": 0.1523534114676863, + "learning_rate": 1.6763468912370547e-05, + "loss": 0.4185, + "step": 1830 + }, + { + "epoch": 3.640159045725646, + "grad_norm": 0.14504280147138135, + "learning_rate": 1.6718292061977604e-05, + "loss": 0.4283, + "step": 1831 + }, + { + "epoch": 3.6421471172962225, + "grad_norm": 0.15299410652292605, + "learning_rate": 1.667316008048281e-05, + "loss": 0.4146, + "step": 1832 + }, + { + "epoch": 3.644135188866799, + "grad_norm": 0.13635691345377363, + "learning_rate": 1.6628073054865262e-05, + "loss": 0.407, + "step": 1833 + }, + { + "epoch": 3.646123260437376, + "grad_norm": 0.14474262221855225, + "learning_rate": 1.658303107201744e-05, + "loss": 0.4165, + "step": 1834 + }, + { + "epoch": 3.6481113320079523, + "grad_norm": 0.13242000912309584, + "learning_rate": 1.653803421874501e-05, + "loss": 0.4144, + "step": 1835 + }, + { + "epoch": 3.650099403578529, + "grad_norm": 0.1417432474902151, + "learning_rate": 1.6493082581766664e-05, + "loss": 0.412, + "step": 1836 + }, + { + "epoch": 3.6520874751491053, + "grad_norm": 0.13649931195067325, + "learning_rate": 1.6448176247713963e-05, + "loss": 0.4202, + "step": 1837 + }, + { + "epoch": 3.6540755467196817, + "grad_norm": 0.13140071457176428, + "learning_rate": 1.6403315303131117e-05, + "loss": 0.4071, + "step": 1838 + }, + { + "epoch": 3.6560636182902586, + "grad_norm": 0.12785529079057562, + "learning_rate": 1.6358499834474888e-05, + "loss": 0.4238, + "step": 1839 + }, + { + "epoch": 3.658051689860835, + "grad_norm": 0.1328523406546393, + "learning_rate": 1.6313729928114436e-05, + "loss": 0.4177, + "step": 1840 + }, + { + "epoch": 3.6600397614314115, + "grad_norm": 0.12763913723943507, + "learning_rate": 1.6269005670331056e-05, + "loss": 0.4191, + "step": 1841 + }, + { + "epoch": 3.662027833001988, + "grad_norm": 0.1258915652562026, + "learning_rate": 1.6224327147318088e-05, + "loss": 0.4154, + "step": 1842 + }, + { + "epoch": 3.6640159045725644, + "grad_norm": 0.1419369252184392, + "learning_rate": 1.6179694445180734e-05, + "loss": 0.4046, + "step": 1843 + }, + { + "epoch": 3.6660039761431413, + "grad_norm": 0.11679097361725023, + "learning_rate": 1.613510764993589e-05, + "loss": 0.4105, + "step": 1844 + }, + { + "epoch": 3.667992047713718, + "grad_norm": 0.12616122435207, + "learning_rate": 1.6090566847511963e-05, + "loss": 0.424, + "step": 1845 + }, + { + "epoch": 3.6699801192842942, + "grad_norm": 0.14828367320020833, + "learning_rate": 1.604607212374876e-05, + "loss": 0.4201, + "step": 1846 + }, + { + "epoch": 3.6719681908548707, + "grad_norm": 0.13104162926966564, + "learning_rate": 1.600162356439723e-05, + "loss": 0.4128, + "step": 1847 + }, + { + "epoch": 3.673956262425447, + "grad_norm": 0.1476910246192276, + "learning_rate": 1.5957221255119372e-05, + "loss": 0.4102, + "step": 1848 + }, + { + "epoch": 3.675944333996024, + "grad_norm": 0.1368512969102732, + "learning_rate": 1.5912865281488102e-05, + "loss": 0.4191, + "step": 1849 + }, + { + "epoch": 3.6779324055666005, + "grad_norm": 0.1281089423568805, + "learning_rate": 1.586855572898698e-05, + "loss": 0.4076, + "step": 1850 + }, + { + "epoch": 3.679920477137177, + "grad_norm": 0.1395949823972034, + "learning_rate": 1.582429268301011e-05, + "loss": 0.4134, + "step": 1851 + }, + { + "epoch": 3.6819085487077534, + "grad_norm": 0.13951337217782803, + "learning_rate": 1.5780076228861993e-05, + "loss": 0.4168, + "step": 1852 + }, + { + "epoch": 3.68389662027833, + "grad_norm": 0.12829014970739838, + "learning_rate": 1.5735906451757307e-05, + "loss": 0.4198, + "step": 1853 + }, + { + "epoch": 3.6858846918489068, + "grad_norm": 0.13774909426196918, + "learning_rate": 1.5691783436820794e-05, + "loss": 0.4192, + "step": 1854 + }, + { + "epoch": 3.6878727634194832, + "grad_norm": 0.28908814343759603, + "learning_rate": 1.564770726908708e-05, + "loss": 0.4079, + "step": 1855 + }, + { + "epoch": 3.6898608349900597, + "grad_norm": 0.14430553920756006, + "learning_rate": 1.5603678033500466e-05, + "loss": 0.4152, + "step": 1856 + }, + { + "epoch": 3.691848906560636, + "grad_norm": 0.16044176639548022, + "learning_rate": 1.5559695814914833e-05, + "loss": 0.4172, + "step": 1857 + }, + { + "epoch": 3.6938369781312126, + "grad_norm": 0.14847470198565352, + "learning_rate": 1.5515760698093485e-05, + "loss": 0.4052, + "step": 1858 + }, + { + "epoch": 3.6958250497017895, + "grad_norm": 0.13066672984376299, + "learning_rate": 1.54718727677089e-05, + "loss": 0.4125, + "step": 1859 + }, + { + "epoch": 3.697813121272366, + "grad_norm": 0.16208398204896493, + "learning_rate": 1.542803210834263e-05, + "loss": 0.4109, + "step": 1860 + }, + { + "epoch": 3.6998011928429424, + "grad_norm": 0.14308392969629824, + "learning_rate": 1.5384238804485132e-05, + "loss": 0.4171, + "step": 1861 + }, + { + "epoch": 3.701789264413519, + "grad_norm": 0.13472552616176814, + "learning_rate": 1.5340492940535594e-05, + "loss": 0.4188, + "step": 1862 + }, + { + "epoch": 3.7037773359840953, + "grad_norm": 0.16245696982606775, + "learning_rate": 1.529679460080178e-05, + "loss": 0.4159, + "step": 1863 + }, + { + "epoch": 3.705765407554672, + "grad_norm": 0.12641490699261004, + "learning_rate": 1.5253143869499875e-05, + "loss": 0.4142, + "step": 1864 + }, + { + "epoch": 3.7077534791252487, + "grad_norm": 0.1407468125478082, + "learning_rate": 1.5209540830754282e-05, + "loss": 0.4081, + "step": 1865 + }, + { + "epoch": 3.709741550695825, + "grad_norm": 0.12063476957620818, + "learning_rate": 1.5165985568597497e-05, + "loss": 0.4199, + "step": 1866 + }, + { + "epoch": 3.7117296222664016, + "grad_norm": 0.1508589666989841, + "learning_rate": 1.5122478166969985e-05, + "loss": 0.4204, + "step": 1867 + }, + { + "epoch": 3.713717693836978, + "grad_norm": 0.27787074894019453, + "learning_rate": 1.5079018709719928e-05, + "loss": 0.4181, + "step": 1868 + }, + { + "epoch": 3.715705765407555, + "grad_norm": 0.12122504335335331, + "learning_rate": 1.503560728060312e-05, + "loss": 0.4049, + "step": 1869 + }, + { + "epoch": 3.717693836978131, + "grad_norm": 0.1285881332318176, + "learning_rate": 1.4992243963282796e-05, + "loss": 0.4121, + "step": 1870 + }, + { + "epoch": 3.719681908548708, + "grad_norm": 0.13056341767813032, + "learning_rate": 1.4948928841329466e-05, + "loss": 0.4129, + "step": 1871 + }, + { + "epoch": 3.7216699801192843, + "grad_norm": 0.1176335456819128, + "learning_rate": 1.4905661998220761e-05, + "loss": 0.4229, + "step": 1872 + }, + { + "epoch": 3.7236580516898607, + "grad_norm": 0.127572659098762, + "learning_rate": 1.4862443517341283e-05, + "loss": 0.4085, + "step": 1873 + }, + { + "epoch": 3.7256461232604376, + "grad_norm": 0.12448568682751915, + "learning_rate": 1.481927348198238e-05, + "loss": 0.4159, + "step": 1874 + }, + { + "epoch": 3.7276341948310137, + "grad_norm": 0.12127432985554247, + "learning_rate": 1.477615197534208e-05, + "loss": 0.4147, + "step": 1875 + }, + { + "epoch": 3.7296222664015906, + "grad_norm": 0.13778721644970837, + "learning_rate": 1.473307908052489e-05, + "loss": 0.4244, + "step": 1876 + }, + { + "epoch": 3.731610337972167, + "grad_norm": 0.12777566872269489, + "learning_rate": 1.4690054880541604e-05, + "loss": 0.4125, + "step": 1877 + }, + { + "epoch": 3.7335984095427435, + "grad_norm": 0.12730243157884172, + "learning_rate": 1.4647079458309189e-05, + "loss": 0.4035, + "step": 1878 + }, + { + "epoch": 3.73558648111332, + "grad_norm": 0.12401266431341217, + "learning_rate": 1.460415289665059e-05, + "loss": 0.4158, + "step": 1879 + }, + { + "epoch": 3.7375745526838964, + "grad_norm": 0.13340393869019873, + "learning_rate": 1.4561275278294602e-05, + "loss": 0.4122, + "step": 1880 + }, + { + "epoch": 3.7395626242544733, + "grad_norm": 0.12525013952678124, + "learning_rate": 1.4518446685875689e-05, + "loss": 0.4173, + "step": 1881 + }, + { + "epoch": 3.7415506958250497, + "grad_norm": 0.12718806203575092, + "learning_rate": 1.4475667201933851e-05, + "loss": 0.402, + "step": 1882 + }, + { + "epoch": 3.743538767395626, + "grad_norm": 0.12373186687056051, + "learning_rate": 1.4432936908914381e-05, + "loss": 0.4142, + "step": 1883 + }, + { + "epoch": 3.7455268389662026, + "grad_norm": 0.12193879482225009, + "learning_rate": 1.4390255889167856e-05, + "loss": 0.4101, + "step": 1884 + }, + { + "epoch": 3.747514910536779, + "grad_norm": 0.13219048436791064, + "learning_rate": 1.4347624224949845e-05, + "loss": 0.4219, + "step": 1885 + }, + { + "epoch": 3.749502982107356, + "grad_norm": 0.1215258991560494, + "learning_rate": 1.4305041998420808e-05, + "loss": 0.4133, + "step": 1886 + }, + { + "epoch": 3.7514910536779325, + "grad_norm": 0.13401458332037963, + "learning_rate": 1.4262509291645921e-05, + "loss": 0.413, + "step": 1887 + }, + { + "epoch": 3.753479125248509, + "grad_norm": 0.12356450775309526, + "learning_rate": 1.4220026186594932e-05, + "loss": 0.418, + "step": 1888 + }, + { + "epoch": 3.7554671968190854, + "grad_norm": 0.12577960300004168, + "learning_rate": 1.417759276514199e-05, + "loss": 0.4025, + "step": 1889 + }, + { + "epoch": 3.757455268389662, + "grad_norm": 0.1317678943066002, + "learning_rate": 1.4135209109065517e-05, + "loss": 0.409, + "step": 1890 + }, + { + "epoch": 3.7594433399602387, + "grad_norm": 0.11834469035412429, + "learning_rate": 1.4092875300047974e-05, + "loss": 0.4055, + "step": 1891 + }, + { + "epoch": 3.761431411530815, + "grad_norm": 0.12895864349124414, + "learning_rate": 1.4050591419675792e-05, + "loss": 0.4108, + "step": 1892 + }, + { + "epoch": 3.7634194831013916, + "grad_norm": 0.13329492236504692, + "learning_rate": 1.4008357549439202e-05, + "loss": 0.4257, + "step": 1893 + }, + { + "epoch": 3.765407554671968, + "grad_norm": 0.129072823887272, + "learning_rate": 1.3966173770732007e-05, + "loss": 0.4209, + "step": 1894 + }, + { + "epoch": 3.7673956262425445, + "grad_norm": 0.12921413700028228, + "learning_rate": 1.3924040164851497e-05, + "loss": 0.4117, + "step": 1895 + }, + { + "epoch": 3.7693836978131214, + "grad_norm": 0.13042817761575895, + "learning_rate": 1.3881956812998269e-05, + "loss": 0.4198, + "step": 1896 + }, + { + "epoch": 3.771371769383698, + "grad_norm": 0.12792815199593058, + "learning_rate": 1.3839923796276055e-05, + "loss": 0.413, + "step": 1897 + }, + { + "epoch": 3.7733598409542743, + "grad_norm": 0.11728201250249243, + "learning_rate": 1.37979411956916e-05, + "loss": 0.4155, + "step": 1898 + }, + { + "epoch": 3.775347912524851, + "grad_norm": 0.12476125551103852, + "learning_rate": 1.3756009092154482e-05, + "loss": 0.412, + "step": 1899 + }, + { + "epoch": 3.7773359840954273, + "grad_norm": 0.21481338746143638, + "learning_rate": 1.3714127566476933e-05, + "loss": 0.4131, + "step": 1900 + }, + { + "epoch": 3.779324055666004, + "grad_norm": 0.1470134918262448, + "learning_rate": 1.3672296699373729e-05, + "loss": 0.4161, + "step": 1901 + }, + { + "epoch": 3.7813121272365806, + "grad_norm": 0.12788644729106496, + "learning_rate": 1.3630516571462047e-05, + "loss": 0.4207, + "step": 1902 + }, + { + "epoch": 3.783300198807157, + "grad_norm": 0.1301078309175956, + "learning_rate": 1.3588787263261235e-05, + "loss": 0.4132, + "step": 1903 + }, + { + "epoch": 3.7852882703777335, + "grad_norm": 0.12211189542063601, + "learning_rate": 1.3547108855192717e-05, + "loss": 0.4143, + "step": 1904 + }, + { + "epoch": 3.78727634194831, + "grad_norm": 0.11690508523129137, + "learning_rate": 1.3505481427579822e-05, + "loss": 0.4108, + "step": 1905 + }, + { + "epoch": 3.789264413518887, + "grad_norm": 0.11402347692962254, + "learning_rate": 1.3463905060647627e-05, + "loss": 0.4115, + "step": 1906 + }, + { + "epoch": 3.7912524850894633, + "grad_norm": 0.16898178577525383, + "learning_rate": 1.3422379834522795e-05, + "loss": 0.4155, + "step": 1907 + }, + { + "epoch": 3.79324055666004, + "grad_norm": 0.12007712230669085, + "learning_rate": 1.3380905829233464e-05, + "loss": 0.4155, + "step": 1908 + }, + { + "epoch": 3.7952286282306162, + "grad_norm": 0.12168887034461463, + "learning_rate": 1.3339483124709003e-05, + "loss": 0.4164, + "step": 1909 + }, + { + "epoch": 3.7972166998011927, + "grad_norm": 0.1304638038783227, + "learning_rate": 1.3298111800779938e-05, + "loss": 0.4141, + "step": 1910 + }, + { + "epoch": 3.7992047713717696, + "grad_norm": 0.12286293611002731, + "learning_rate": 1.325679193717781e-05, + "loss": 0.4266, + "step": 1911 + }, + { + "epoch": 3.801192842942346, + "grad_norm": 0.11873178024386737, + "learning_rate": 1.321552361353494e-05, + "loss": 0.4094, + "step": 1912 + }, + { + "epoch": 3.8031809145129225, + "grad_norm": 0.11875300394160675, + "learning_rate": 1.3174306909384327e-05, + "loss": 0.405, + "step": 1913 + }, + { + "epoch": 3.805168986083499, + "grad_norm": 0.1305765918885014, + "learning_rate": 1.3133141904159504e-05, + "loss": 0.4167, + "step": 1914 + }, + { + "epoch": 3.8071570576540754, + "grad_norm": 0.1155018724842256, + "learning_rate": 1.3092028677194355e-05, + "loss": 0.4017, + "step": 1915 + }, + { + "epoch": 3.8091451292246523, + "grad_norm": 0.12128244400314282, + "learning_rate": 1.3050967307722981e-05, + "loss": 0.4207, + "step": 1916 + }, + { + "epoch": 3.8111332007952288, + "grad_norm": 0.126233600797627, + "learning_rate": 1.3009957874879557e-05, + "loss": 0.4243, + "step": 1917 + }, + { + "epoch": 3.8131212723658052, + "grad_norm": 0.11307120413410887, + "learning_rate": 1.2969000457698125e-05, + "loss": 0.41, + "step": 1918 + }, + { + "epoch": 3.8151093439363817, + "grad_norm": 0.13789982064344947, + "learning_rate": 1.2928095135112501e-05, + "loss": 0.4197, + "step": 1919 + }, + { + "epoch": 3.817097415506958, + "grad_norm": 0.12235858354606367, + "learning_rate": 1.288724198595614e-05, + "loss": 0.4176, + "step": 1920 + }, + { + "epoch": 3.819085487077535, + "grad_norm": 0.12135903666566722, + "learning_rate": 1.28464410889619e-05, + "loss": 0.4165, + "step": 1921 + }, + { + "epoch": 3.8210735586481115, + "grad_norm": 0.12771649763190396, + "learning_rate": 1.2805692522761955e-05, + "loss": 0.4116, + "step": 1922 + }, + { + "epoch": 3.823061630218688, + "grad_norm": 0.13045665895650455, + "learning_rate": 1.2764996365887624e-05, + "loss": 0.4207, + "step": 1923 + }, + { + "epoch": 3.8250497017892644, + "grad_norm": 0.12744409710040272, + "learning_rate": 1.2724352696769233e-05, + "loss": 0.4114, + "step": 1924 + }, + { + "epoch": 3.827037773359841, + "grad_norm": 0.13256940553885835, + "learning_rate": 1.268376159373593e-05, + "loss": 0.4115, + "step": 1925 + }, + { + "epoch": 3.8290258449304178, + "grad_norm": 0.1196732645649503, + "learning_rate": 1.264322313501559e-05, + "loss": 0.4182, + "step": 1926 + }, + { + "epoch": 3.8310139165009938, + "grad_norm": 0.12898140736842106, + "learning_rate": 1.2602737398734583e-05, + "loss": 0.4217, + "step": 1927 + }, + { + "epoch": 3.8330019880715707, + "grad_norm": 0.12259254324513066, + "learning_rate": 1.2562304462917698e-05, + "loss": 0.41, + "step": 1928 + }, + { + "epoch": 3.834990059642147, + "grad_norm": 0.14314796016876316, + "learning_rate": 1.2521924405487998e-05, + "loss": 0.4175, + "step": 1929 + }, + { + "epoch": 3.8369781312127236, + "grad_norm": 0.11735250306416119, + "learning_rate": 1.2481597304266585e-05, + "loss": 0.4116, + "step": 1930 + }, + { + "epoch": 3.8389662027833005, + "grad_norm": 0.1365058855329985, + "learning_rate": 1.2441323236972536e-05, + "loss": 0.4057, + "step": 1931 + }, + { + "epoch": 3.8409542743538765, + "grad_norm": 0.13339748891673117, + "learning_rate": 1.2401102281222705e-05, + "loss": 0.4229, + "step": 1932 + }, + { + "epoch": 3.8429423459244534, + "grad_norm": 0.11594781861973515, + "learning_rate": 1.2360934514531589e-05, + "loss": 0.4095, + "step": 1933 + }, + { + "epoch": 3.84493041749503, + "grad_norm": 0.13220912477680097, + "learning_rate": 1.2320820014311186e-05, + "loss": 0.4115, + "step": 1934 + }, + { + "epoch": 3.8469184890656063, + "grad_norm": 0.1313681631522006, + "learning_rate": 1.2280758857870842e-05, + "loss": 0.4192, + "step": 1935 + }, + { + "epoch": 3.8489065606361827, + "grad_norm": 0.11599884335856218, + "learning_rate": 1.2240751122417058e-05, + "loss": 0.4138, + "step": 1936 + }, + { + "epoch": 3.850894632206759, + "grad_norm": 0.12653486094497657, + "learning_rate": 1.2200796885053445e-05, + "loss": 0.4017, + "step": 1937 + }, + { + "epoch": 3.852882703777336, + "grad_norm": 0.1130518233332399, + "learning_rate": 1.2160896222780471e-05, + "loss": 0.4145, + "step": 1938 + }, + { + "epoch": 3.8548707753479126, + "grad_norm": 0.13866914719348072, + "learning_rate": 1.2121049212495359e-05, + "loss": 0.4099, + "step": 1939 + }, + { + "epoch": 3.856858846918489, + "grad_norm": 0.12066796601207383, + "learning_rate": 1.2081255930991942e-05, + "loss": 0.406, + "step": 1940 + }, + { + "epoch": 3.8588469184890655, + "grad_norm": 0.11330237818106631, + "learning_rate": 1.2041516454960492e-05, + "loss": 0.4187, + "step": 1941 + }, + { + "epoch": 3.860834990059642, + "grad_norm": 0.12538978699781372, + "learning_rate": 1.2001830860987598e-05, + "loss": 0.4091, + "step": 1942 + }, + { + "epoch": 3.862823061630219, + "grad_norm": 0.12467804244722842, + "learning_rate": 1.1962199225555996e-05, + "loss": 0.4107, + "step": 1943 + }, + { + "epoch": 3.8648111332007953, + "grad_norm": 0.12249180781767287, + "learning_rate": 1.1922621625044464e-05, + "loss": 0.4074, + "step": 1944 + }, + { + "epoch": 3.8667992047713717, + "grad_norm": 0.1306550240810388, + "learning_rate": 1.1883098135727571e-05, + "loss": 0.4207, + "step": 1945 + }, + { + "epoch": 3.868787276341948, + "grad_norm": 0.11933610478296443, + "learning_rate": 1.1843628833775687e-05, + "loss": 0.4115, + "step": 1946 + }, + { + "epoch": 3.8707753479125246, + "grad_norm": 0.1278251014710002, + "learning_rate": 1.1804213795254697e-05, + "loss": 0.4114, + "step": 1947 + }, + { + "epoch": 3.8727634194831015, + "grad_norm": 0.12541159535916258, + "learning_rate": 1.1764853096125935e-05, + "loss": 0.4188, + "step": 1948 + }, + { + "epoch": 3.874751491053678, + "grad_norm": 0.12727972434053444, + "learning_rate": 1.1725546812245994e-05, + "loss": 0.4101, + "step": 1949 + }, + { + "epoch": 3.8767395626242545, + "grad_norm": 0.21485568213880235, + "learning_rate": 1.1686295019366605e-05, + "loss": 0.4184, + "step": 1950 + }, + { + "epoch": 3.878727634194831, + "grad_norm": 0.12396718977956772, + "learning_rate": 1.1647097793134478e-05, + "loss": 0.4035, + "step": 1951 + }, + { + "epoch": 3.8807157057654074, + "grad_norm": 0.19187344793079394, + "learning_rate": 1.160795520909117e-05, + "loss": 0.4141, + "step": 1952 + }, + { + "epoch": 3.8827037773359843, + "grad_norm": 0.13972243163789239, + "learning_rate": 1.1568867342672939e-05, + "loss": 0.418, + "step": 1953 + }, + { + "epoch": 3.8846918489065607, + "grad_norm": 0.12246725141298273, + "learning_rate": 1.152983426921053e-05, + "loss": 0.4087, + "step": 1954 + }, + { + "epoch": 3.886679920477137, + "grad_norm": 0.1441970347739633, + "learning_rate": 1.149085606392919e-05, + "loss": 0.4107, + "step": 1955 + }, + { + "epoch": 3.8886679920477136, + "grad_norm": 0.11793533898722837, + "learning_rate": 1.1451932801948345e-05, + "loss": 0.4094, + "step": 1956 + }, + { + "epoch": 3.89065606361829, + "grad_norm": 0.11984657088654536, + "learning_rate": 1.1413064558281564e-05, + "loss": 0.4022, + "step": 1957 + }, + { + "epoch": 3.892644135188867, + "grad_norm": 0.13747510019212833, + "learning_rate": 1.1374251407836372e-05, + "loss": 0.4165, + "step": 1958 + }, + { + "epoch": 3.8946322067594434, + "grad_norm": 0.12283933628622347, + "learning_rate": 1.1335493425414135e-05, + "loss": 0.4251, + "step": 1959 + }, + { + "epoch": 3.89662027833002, + "grad_norm": 0.12767675489119157, + "learning_rate": 1.1296790685709875e-05, + "loss": 0.4068, + "step": 1960 + }, + { + "epoch": 3.8986083499005963, + "grad_norm": 0.11737134954171792, + "learning_rate": 1.1258143263312183e-05, + "loss": 0.414, + "step": 1961 + }, + { + "epoch": 3.900596421471173, + "grad_norm": 0.12017326193658819, + "learning_rate": 1.1219551232702993e-05, + "loss": 0.4087, + "step": 1962 + }, + { + "epoch": 3.9025844930417497, + "grad_norm": 0.1181570800422918, + "learning_rate": 1.1181014668257508e-05, + "loss": 0.4136, + "step": 1963 + }, + { + "epoch": 3.904572564612326, + "grad_norm": 0.11786433898774867, + "learning_rate": 1.1142533644244074e-05, + "loss": 0.4163, + "step": 1964 + }, + { + "epoch": 3.9065606361829026, + "grad_norm": 0.11951359580180812, + "learning_rate": 1.1104108234823946e-05, + "loss": 0.4167, + "step": 1965 + }, + { + "epoch": 3.908548707753479, + "grad_norm": 0.11430020006073105, + "learning_rate": 1.1065738514051225e-05, + "loss": 0.4108, + "step": 1966 + }, + { + "epoch": 3.9105367793240555, + "grad_norm": 0.12399755011503341, + "learning_rate": 1.1027424555872676e-05, + "loss": 0.4152, + "step": 1967 + }, + { + "epoch": 3.9125248508946324, + "grad_norm": 0.10379447658508856, + "learning_rate": 1.0989166434127604e-05, + "loss": 0.4191, + "step": 1968 + }, + { + "epoch": 3.914512922465209, + "grad_norm": 0.11841545442440568, + "learning_rate": 1.0950964222547706e-05, + "loss": 0.4074, + "step": 1969 + }, + { + "epoch": 3.9165009940357853, + "grad_norm": 0.12174811592361247, + "learning_rate": 1.0912817994756928e-05, + "loss": 0.4071, + "step": 1970 + }, + { + "epoch": 3.918489065606362, + "grad_norm": 0.11738800063725485, + "learning_rate": 1.0874727824271307e-05, + "loss": 0.4026, + "step": 1971 + }, + { + "epoch": 3.9204771371769382, + "grad_norm": 0.12683667073776073, + "learning_rate": 1.083669378449885e-05, + "loss": 0.4149, + "step": 1972 + }, + { + "epoch": 3.922465208747515, + "grad_norm": 0.23578844123586262, + "learning_rate": 1.0798715948739424e-05, + "loss": 0.4243, + "step": 1973 + }, + { + "epoch": 3.9244532803180916, + "grad_norm": 0.12470565801211261, + "learning_rate": 1.0760794390184542e-05, + "loss": 0.4084, + "step": 1974 + }, + { + "epoch": 3.926441351888668, + "grad_norm": 0.14135849783994955, + "learning_rate": 1.0722929181917264e-05, + "loss": 0.404, + "step": 1975 + }, + { + "epoch": 3.9284294234592445, + "grad_norm": 0.12266354579516658, + "learning_rate": 1.0685120396912057e-05, + "loss": 0.4111, + "step": 1976 + }, + { + "epoch": 3.930417495029821, + "grad_norm": 0.13413047404837405, + "learning_rate": 1.0647368108034644e-05, + "loss": 0.4123, + "step": 1977 + }, + { + "epoch": 3.932405566600398, + "grad_norm": 0.1253547852041438, + "learning_rate": 1.0609672388041864e-05, + "loss": 0.4062, + "step": 1978 + }, + { + "epoch": 3.9343936381709743, + "grad_norm": 0.11410535557037213, + "learning_rate": 1.0572033309581569e-05, + "loss": 0.4091, + "step": 1979 + }, + { + "epoch": 3.9363817097415508, + "grad_norm": 0.14930707729055762, + "learning_rate": 1.0534450945192382e-05, + "loss": 0.4194, + "step": 1980 + }, + { + "epoch": 3.9383697813121272, + "grad_norm": 0.11785277834457099, + "learning_rate": 1.0496925367303677e-05, + "loss": 0.4083, + "step": 1981 + }, + { + "epoch": 3.9403578528827037, + "grad_norm": 0.11621703078477161, + "learning_rate": 1.0459456648235395e-05, + "loss": 0.4265, + "step": 1982 + }, + { + "epoch": 3.9423459244532806, + "grad_norm": 0.12441075712798819, + "learning_rate": 1.0422044860197867e-05, + "loss": 0.4236, + "step": 1983 + }, + { + "epoch": 3.9443339960238566, + "grad_norm": 0.11732262831105307, + "learning_rate": 1.0384690075291725e-05, + "loss": 0.413, + "step": 1984 + }, + { + "epoch": 3.9463220675944335, + "grad_norm": 0.11971044861609616, + "learning_rate": 1.0347392365507734e-05, + "loss": 0.4064, + "step": 1985 + }, + { + "epoch": 3.94831013916501, + "grad_norm": 0.12454131199916985, + "learning_rate": 1.0310151802726667e-05, + "loss": 0.4024, + "step": 1986 + }, + { + "epoch": 3.9502982107355864, + "grad_norm": 0.1243323308512183, + "learning_rate": 1.0272968458719158e-05, + "loss": 0.4114, + "step": 1987 + }, + { + "epoch": 3.952286282306163, + "grad_norm": 0.10706316079552054, + "learning_rate": 1.0235842405145587e-05, + "loss": 0.4023, + "step": 1988 + }, + { + "epoch": 3.9542743538767393, + "grad_norm": 0.11034771625495471, + "learning_rate": 1.0198773713555878e-05, + "loss": 0.4021, + "step": 1989 + }, + { + "epoch": 3.956262425447316, + "grad_norm": 0.1114590814911575, + "learning_rate": 1.0161762455389445e-05, + "loss": 0.4144, + "step": 1990 + }, + { + "epoch": 3.9582504970178927, + "grad_norm": 0.11480269527129816, + "learning_rate": 1.0124808701975021e-05, + "loss": 0.4019, + "step": 1991 + }, + { + "epoch": 3.960238568588469, + "grad_norm": 0.10951780763117114, + "learning_rate": 1.0087912524530492e-05, + "loss": 0.4105, + "step": 1992 + }, + { + "epoch": 3.9622266401590456, + "grad_norm": 0.10730065026334104, + "learning_rate": 1.005107399416279e-05, + "loss": 0.4119, + "step": 1993 + }, + { + "epoch": 3.964214711729622, + "grad_norm": 0.1157812587471413, + "learning_rate": 1.0014293181867743e-05, + "loss": 0.4048, + "step": 1994 + }, + { + "epoch": 3.966202783300199, + "grad_norm": 0.10634464575554442, + "learning_rate": 9.977570158529958e-06, + "loss": 0.4126, + "step": 1995 + }, + { + "epoch": 3.9681908548707754, + "grad_norm": 0.14226769514993798, + "learning_rate": 9.940904994922658e-06, + "loss": 0.4027, + "step": 1996 + }, + { + "epoch": 3.970178926441352, + "grad_norm": 0.10876356513781965, + "learning_rate": 9.904297761707569e-06, + "loss": 0.4158, + "step": 1997 + }, + { + "epoch": 3.9721669980119283, + "grad_norm": 0.10968453251673606, + "learning_rate": 9.867748529434737e-06, + "loss": 0.4025, + "step": 1998 + }, + { + "epoch": 3.9741550695825048, + "grad_norm": 0.1106822867587708, + "learning_rate": 9.831257368542486e-06, + "loss": 0.4126, + "step": 1999 + }, + { + "epoch": 3.9761431411530817, + "grad_norm": 0.1458577023566835, + "learning_rate": 9.794824349357186e-06, + "loss": 0.416, + "step": 2000 + }, + { + "epoch": 3.978131212723658, + "grad_norm": 0.11447590869393695, + "learning_rate": 9.75844954209316e-06, + "loss": 0.4115, + "step": 2001 + }, + { + "epoch": 3.9801192842942346, + "grad_norm": 0.11780031913072884, + "learning_rate": 9.722133016852547e-06, + "loss": 0.4034, + "step": 2002 + }, + { + "epoch": 3.982107355864811, + "grad_norm": 0.11200150629333681, + "learning_rate": 9.685874843625158e-06, + "loss": 0.4071, + "step": 2003 + }, + { + "epoch": 3.9840954274353875, + "grad_norm": 0.11362356694063454, + "learning_rate": 9.649675092288366e-06, + "loss": 0.4086, + "step": 2004 + }, + { + "epoch": 3.9860834990059644, + "grad_norm": 0.11178076261068912, + "learning_rate": 9.613533832606925e-06, + "loss": 0.415, + "step": 2005 + }, + { + "epoch": 3.988071570576541, + "grad_norm": 0.1303616732073129, + "learning_rate": 9.5774511342329e-06, + "loss": 0.4162, + "step": 2006 + }, + { + "epoch": 3.9900596421471173, + "grad_norm": 0.11590806959347276, + "learning_rate": 9.54142706670543e-06, + "loss": 0.4104, + "step": 2007 + }, + { + "epoch": 3.9920477137176937, + "grad_norm": 0.11347675690643563, + "learning_rate": 9.505461699450747e-06, + "loss": 0.4154, + "step": 2008 + }, + { + "epoch": 3.99403578528827, + "grad_norm": 0.10616599308683848, + "learning_rate": 9.469555101781886e-06, + "loss": 0.4126, + "step": 2009 + }, + { + "epoch": 3.996023856858847, + "grad_norm": 0.11591746562675516, + "learning_rate": 9.433707342898647e-06, + "loss": 0.4065, + "step": 2010 + }, + { + "epoch": 3.9980119284294235, + "grad_norm": 0.1169087220375494, + "learning_rate": 9.397918491887443e-06, + "loss": 0.4106, + "step": 2011 + }, + { + "epoch": 4.0, + "grad_norm": 0.11577369538103958, + "learning_rate": 9.362188617721135e-06, + "loss": 0.3946, + "step": 2012 + }, + { + "epoch": 4.001988071570577, + "grad_norm": 0.16242295220761954, + "learning_rate": 9.326517789258944e-06, + "loss": 0.3935, + "step": 2013 + }, + { + "epoch": 4.003976143141153, + "grad_norm": 0.12270671213229685, + "learning_rate": 9.290906075246288e-06, + "loss": 0.3796, + "step": 2014 + }, + { + "epoch": 4.00596421471173, + "grad_norm": 0.13856681126979406, + "learning_rate": 9.255353544314669e-06, + "loss": 0.3887, + "step": 2015 + }, + { + "epoch": 4.007952286282306, + "grad_norm": 0.14376438682763779, + "learning_rate": 9.219860264981486e-06, + "loss": 0.3893, + "step": 2016 + }, + { + "epoch": 4.009940357852883, + "grad_norm": 0.12730114055456757, + "learning_rate": 9.184426305650022e-06, + "loss": 0.3939, + "step": 2017 + }, + { + "epoch": 4.01192842942346, + "grad_norm": 0.13390495368426292, + "learning_rate": 9.149051734609183e-06, + "loss": 0.3882, + "step": 2018 + }, + { + "epoch": 4.013916500994036, + "grad_norm": 0.15428620649225006, + "learning_rate": 9.11373662003343e-06, + "loss": 0.4004, + "step": 2019 + }, + { + "epoch": 4.0159045725646125, + "grad_norm": 0.12561990727486522, + "learning_rate": 9.078481029982655e-06, + "loss": 0.3893, + "step": 2020 + }, + { + "epoch": 4.0178926441351885, + "grad_norm": 0.14373181736551474, + "learning_rate": 9.043285032402016e-06, + "loss": 0.3801, + "step": 2021 + }, + { + "epoch": 4.019880715705765, + "grad_norm": 0.12268500210699507, + "learning_rate": 9.008148695121832e-06, + "loss": 0.3992, + "step": 2022 + }, + { + "epoch": 4.021868787276342, + "grad_norm": 0.13904765421921883, + "learning_rate": 8.973072085857439e-06, + "loss": 0.3813, + "step": 2023 + }, + { + "epoch": 4.023856858846918, + "grad_norm": 0.11969320329871334, + "learning_rate": 8.938055272209083e-06, + "loss": 0.3939, + "step": 2024 + }, + { + "epoch": 4.025844930417495, + "grad_norm": 0.1274287401386006, + "learning_rate": 8.903098321661718e-06, + "loss": 0.3919, + "step": 2025 + }, + { + "epoch": 4.027833001988071, + "grad_norm": 0.12563635221967606, + "learning_rate": 8.868201301584997e-06, + "loss": 0.3895, + "step": 2026 + }, + { + "epoch": 4.029821073558648, + "grad_norm": 0.13763721797276127, + "learning_rate": 8.833364279233038e-06, + "loss": 0.388, + "step": 2027 + }, + { + "epoch": 4.031809145129225, + "grad_norm": 0.12122680857919836, + "learning_rate": 8.798587321744327e-06, + "loss": 0.3951, + "step": 2028 + }, + { + "epoch": 4.033797216699801, + "grad_norm": 0.12255329580322401, + "learning_rate": 8.763870496141606e-06, + "loss": 0.3765, + "step": 2029 + }, + { + "epoch": 4.035785288270378, + "grad_norm": 0.1335435726236594, + "learning_rate": 8.729213869331725e-06, + "loss": 0.3896, + "step": 2030 + }, + { + "epoch": 4.037773359840954, + "grad_norm": 0.11542654602014098, + "learning_rate": 8.69461750810551e-06, + "loss": 0.3823, + "step": 2031 + }, + { + "epoch": 4.039761431411531, + "grad_norm": 0.11640952234172514, + "learning_rate": 8.660081479137653e-06, + "loss": 0.3857, + "step": 2032 + }, + { + "epoch": 4.041749502982108, + "grad_norm": 0.11853107645718568, + "learning_rate": 8.625605848986577e-06, + "loss": 0.3783, + "step": 2033 + }, + { + "epoch": 4.043737574552684, + "grad_norm": 0.10880512742851284, + "learning_rate": 8.591190684094263e-06, + "loss": 0.3917, + "step": 2034 + }, + { + "epoch": 4.045725646123261, + "grad_norm": 0.12050622630816221, + "learning_rate": 8.556836050786228e-06, + "loss": 0.3812, + "step": 2035 + }, + { + "epoch": 4.047713717693837, + "grad_norm": 0.11073118657630984, + "learning_rate": 8.522542015271291e-06, + "loss": 0.3863, + "step": 2036 + }, + { + "epoch": 4.049701789264414, + "grad_norm": 0.11162845876453863, + "learning_rate": 8.488308643641483e-06, + "loss": 0.3854, + "step": 2037 + }, + { + "epoch": 4.0516898608349905, + "grad_norm": 0.10817652025454551, + "learning_rate": 8.45413600187194e-06, + "loss": 0.3773, + "step": 2038 + }, + { + "epoch": 4.0536779324055665, + "grad_norm": 0.1213300740692339, + "learning_rate": 8.420024155820745e-06, + "loss": 0.3873, + "step": 2039 + }, + { + "epoch": 4.055666003976143, + "grad_norm": 0.11152699715766383, + "learning_rate": 8.385973171228823e-06, + "loss": 0.3858, + "step": 2040 + }, + { + "epoch": 4.057654075546719, + "grad_norm": 0.11647996747982567, + "learning_rate": 8.351983113719812e-06, + "loss": 0.388, + "step": 2041 + }, + { + "epoch": 4.059642147117296, + "grad_norm": 0.11879327665095078, + "learning_rate": 8.318054048799903e-06, + "loss": 0.3853, + "step": 2042 + }, + { + "epoch": 4.061630218687872, + "grad_norm": 0.1070875071355291, + "learning_rate": 8.284186041857754e-06, + "loss": 0.3845, + "step": 2043 + }, + { + "epoch": 4.063618290258449, + "grad_norm": 0.10630385206510362, + "learning_rate": 8.250379158164383e-06, + "loss": 0.3938, + "step": 2044 + }, + { + "epoch": 4.065606361829026, + "grad_norm": 0.11597291042447759, + "learning_rate": 8.216633462872972e-06, + "loss": 0.3907, + "step": 2045 + }, + { + "epoch": 4.067594433399602, + "grad_norm": 0.7364567886782528, + "learning_rate": 8.182949021018784e-06, + "loss": 0.4164, + "step": 2046 + }, + { + "epoch": 4.069582504970179, + "grad_norm": 0.11061595816141373, + "learning_rate": 8.149325897519049e-06, + "loss": 0.3897, + "step": 2047 + }, + { + "epoch": 4.071570576540755, + "grad_norm": 0.11232959746711056, + "learning_rate": 8.115764157172812e-06, + "loss": 0.3988, + "step": 2048 + }, + { + "epoch": 4.073558648111332, + "grad_norm": 0.1016202934512297, + "learning_rate": 8.082263864660822e-06, + "loss": 0.3803, + "step": 2049 + }, + { + "epoch": 4.075546719681909, + "grad_norm": 0.11318725010275028, + "learning_rate": 8.048825084545413e-06, + "loss": 0.4008, + "step": 2050 + }, + { + "epoch": 4.077534791252485, + "grad_norm": 0.11326433626229489, + "learning_rate": 8.015447881270337e-06, + "loss": 0.4066, + "step": 2051 + }, + { + "epoch": 4.079522862823062, + "grad_norm": 0.10820812710830408, + "learning_rate": 7.982132319160722e-06, + "loss": 0.3814, + "step": 2052 + }, + { + "epoch": 4.081510934393638, + "grad_norm": 0.10664925736228041, + "learning_rate": 7.94887846242288e-06, + "loss": 0.3872, + "step": 2053 + }, + { + "epoch": 4.083499005964215, + "grad_norm": 0.10160843998696288, + "learning_rate": 7.915686375144193e-06, + "loss": 0.3936, + "step": 2054 + }, + { + "epoch": 4.085487077534792, + "grad_norm": 0.11288434735669675, + "learning_rate": 7.88255612129301e-06, + "loss": 0.3796, + "step": 2055 + }, + { + "epoch": 4.087475149105368, + "grad_norm": 0.1122952920063878, + "learning_rate": 7.849487764718514e-06, + "loss": 0.3817, + "step": 2056 + }, + { + "epoch": 4.0894632206759445, + "grad_norm": 0.11093903242708326, + "learning_rate": 7.816481369150594e-06, + "loss": 0.3888, + "step": 2057 + }, + { + "epoch": 4.0914512922465205, + "grad_norm": 0.10739815028668943, + "learning_rate": 7.783536998199732e-06, + "loss": 0.3961, + "step": 2058 + }, + { + "epoch": 4.093439363817097, + "grad_norm": 0.12452803775102535, + "learning_rate": 7.750654715356884e-06, + "loss": 0.3955, + "step": 2059 + }, + { + "epoch": 4.095427435387674, + "grad_norm": 0.11411657187569366, + "learning_rate": 7.717834583993298e-06, + "loss": 0.3933, + "step": 2060 + }, + { + "epoch": 4.09741550695825, + "grad_norm": 0.11299300329262432, + "learning_rate": 7.685076667360523e-06, + "loss": 0.3897, + "step": 2061 + }, + { + "epoch": 4.099403578528827, + "grad_norm": 0.11173542019812274, + "learning_rate": 7.652381028590152e-06, + "loss": 0.3873, + "step": 2062 + }, + { + "epoch": 4.101391650099403, + "grad_norm": 0.1174313343930306, + "learning_rate": 7.619747730693765e-06, + "loss": 0.3823, + "step": 2063 + }, + { + "epoch": 4.10337972166998, + "grad_norm": 0.11445309254950768, + "learning_rate": 7.587176836562799e-06, + "loss": 0.3882, + "step": 2064 + }, + { + "epoch": 4.105367793240557, + "grad_norm": 0.10898913985158198, + "learning_rate": 7.5546684089684395e-06, + "loss": 0.404, + "step": 2065 + }, + { + "epoch": 4.107355864811133, + "grad_norm": 0.12297485161873528, + "learning_rate": 7.522222510561463e-06, + "loss": 0.3901, + "step": 2066 + }, + { + "epoch": 4.10934393638171, + "grad_norm": 0.1250730138667789, + "learning_rate": 7.489839203872153e-06, + "loss": 0.3811, + "step": 2067 + }, + { + "epoch": 4.111332007952286, + "grad_norm": 0.11397541482738091, + "learning_rate": 7.45751855131017e-06, + "loss": 0.3944, + "step": 2068 + }, + { + "epoch": 4.113320079522863, + "grad_norm": 0.10709840073577899, + "learning_rate": 7.4252606151643894e-06, + "loss": 0.389, + "step": 2069 + }, + { + "epoch": 4.11530815109344, + "grad_norm": 0.11044500826907738, + "learning_rate": 7.393065457602877e-06, + "loss": 0.3918, + "step": 2070 + }, + { + "epoch": 4.117296222664016, + "grad_norm": 0.10317075798494911, + "learning_rate": 7.360933140672668e-06, + "loss": 0.3885, + "step": 2071 + }, + { + "epoch": 4.119284294234593, + "grad_norm": 0.10854131372035143, + "learning_rate": 7.328863726299707e-06, + "loss": 0.3919, + "step": 2072 + }, + { + "epoch": 4.121272365805169, + "grad_norm": 0.1030453432518548, + "learning_rate": 7.296857276288704e-06, + "loss": 0.3806, + "step": 2073 + }, + { + "epoch": 4.1232604373757455, + "grad_norm": 0.10895214824409698, + "learning_rate": 7.264913852323027e-06, + "loss": 0.3809, + "step": 2074 + }, + { + "epoch": 4.1252485089463224, + "grad_norm": 0.10521539908738563, + "learning_rate": 7.233033515964582e-06, + "loss": 0.3913, + "step": 2075 + }, + { + "epoch": 4.1272365805168985, + "grad_norm": 0.1046089740427435, + "learning_rate": 7.20121632865368e-06, + "loss": 0.3894, + "step": 2076 + }, + { + "epoch": 4.129224652087475, + "grad_norm": 0.11238242975442705, + "learning_rate": 7.169462351708958e-06, + "loss": 0.3765, + "step": 2077 + }, + { + "epoch": 4.131212723658051, + "grad_norm": 0.10931635695014968, + "learning_rate": 7.137771646327176e-06, + "loss": 0.3851, + "step": 2078 + }, + { + "epoch": 4.133200795228628, + "grad_norm": 0.1073786462890413, + "learning_rate": 7.106144273583227e-06, + "loss": 0.3819, + "step": 2079 + }, + { + "epoch": 4.135188866799205, + "grad_norm": 0.10454150184031392, + "learning_rate": 7.074580294429902e-06, + "loss": 0.3885, + "step": 2080 + }, + { + "epoch": 4.137176938369781, + "grad_norm": 0.11587471164722234, + "learning_rate": 7.043079769697833e-06, + "loss": 0.3968, + "step": 2081 + }, + { + "epoch": 4.139165009940358, + "grad_norm": 0.1049751719664736, + "learning_rate": 7.0116427600953565e-06, + "loss": 0.3887, + "step": 2082 + }, + { + "epoch": 4.141153081510934, + "grad_norm": 0.10651452982037905, + "learning_rate": 6.9802693262084106e-06, + "loss": 0.3897, + "step": 2083 + }, + { + "epoch": 4.143141153081511, + "grad_norm": 0.11240955066490231, + "learning_rate": 6.948959528500396e-06, + "loss": 0.3865, + "step": 2084 + }, + { + "epoch": 4.145129224652088, + "grad_norm": 0.11296598058113064, + "learning_rate": 6.91771342731209e-06, + "loss": 0.387, + "step": 2085 + }, + { + "epoch": 4.147117296222664, + "grad_norm": 0.09714300484621313, + "learning_rate": 6.886531082861503e-06, + "loss": 0.4024, + "step": 2086 + }, + { + "epoch": 4.149105367793241, + "grad_norm": 0.11486164583751605, + "learning_rate": 6.855412555243743e-06, + "loss": 0.3884, + "step": 2087 + }, + { + "epoch": 4.151093439363817, + "grad_norm": 0.10858190643310542, + "learning_rate": 6.824357904430998e-06, + "loss": 0.3943, + "step": 2088 + }, + { + "epoch": 4.153081510934394, + "grad_norm": 0.10169998242634819, + "learning_rate": 6.7933671902722955e-06, + "loss": 0.3973, + "step": 2089 + }, + { + "epoch": 4.155069582504971, + "grad_norm": 0.11225341199262792, + "learning_rate": 6.762440472493459e-06, + "loss": 0.3884, + "step": 2090 + }, + { + "epoch": 4.157057654075547, + "grad_norm": 0.10809234552066299, + "learning_rate": 6.731577810696981e-06, + "loss": 0.383, + "step": 2091 + }, + { + "epoch": 4.1590457256461235, + "grad_norm": 0.1162247680241023, + "learning_rate": 6.700779264361896e-06, + "loss": 0.3948, + "step": 2092 + }, + { + "epoch": 4.1610337972166995, + "grad_norm": 0.10331984098873934, + "learning_rate": 6.670044892843686e-06, + "loss": 0.3904, + "step": 2093 + }, + { + "epoch": 4.163021868787276, + "grad_norm": 0.11263645547156585, + "learning_rate": 6.6393747553741416e-06, + "loss": 0.3935, + "step": 2094 + }, + { + "epoch": 4.165009940357853, + "grad_norm": 0.11731311349840723, + "learning_rate": 6.608768911061281e-06, + "loss": 0.3939, + "step": 2095 + }, + { + "epoch": 4.166998011928429, + "grad_norm": 0.10993203224686413, + "learning_rate": 6.5782274188891696e-06, + "loss": 0.3956, + "step": 2096 + }, + { + "epoch": 4.168986083499006, + "grad_norm": 0.1517967414260752, + "learning_rate": 6.547750337717906e-06, + "loss": 0.3921, + "step": 2097 + }, + { + "epoch": 4.170974155069582, + "grad_norm": 0.10352035003831739, + "learning_rate": 6.517337726283437e-06, + "loss": 0.3901, + "step": 2098 + }, + { + "epoch": 4.172962226640159, + "grad_norm": 0.1019881318386552, + "learning_rate": 6.486989643197446e-06, + "loss": 0.374, + "step": 2099 + }, + { + "epoch": 4.174950298210735, + "grad_norm": 0.1058930276911533, + "learning_rate": 6.456706146947271e-06, + "loss": 0.3952, + "step": 2100 + }, + { + "epoch": 4.176938369781312, + "grad_norm": 0.10884888055550279, + "learning_rate": 6.426487295895772e-06, + "loss": 0.3755, + "step": 2101 + }, + { + "epoch": 4.178926441351889, + "grad_norm": 0.10366779903871269, + "learning_rate": 6.396333148281231e-06, + "loss": 0.387, + "step": 2102 + }, + { + "epoch": 4.180914512922465, + "grad_norm": 0.1047566988829119, + "learning_rate": 6.366243762217226e-06, + "loss": 0.3964, + "step": 2103 + }, + { + "epoch": 4.182902584493042, + "grad_norm": 0.10172953838653928, + "learning_rate": 6.336219195692521e-06, + "loss": 0.3911, + "step": 2104 + }, + { + "epoch": 4.184890656063618, + "grad_norm": 0.10614996378516255, + "learning_rate": 6.306259506570968e-06, + "loss": 0.3826, + "step": 2105 + }, + { + "epoch": 4.186878727634195, + "grad_norm": 0.11907843686153102, + "learning_rate": 6.276364752591386e-06, + "loss": 0.3804, + "step": 2106 + }, + { + "epoch": 4.188866799204772, + "grad_norm": 0.11026202432242263, + "learning_rate": 6.2465349913674435e-06, + "loss": 0.3828, + "step": 2107 + }, + { + "epoch": 4.190854870775348, + "grad_norm": 0.10861909520932465, + "learning_rate": 6.216770280387554e-06, + "loss": 0.3825, + "step": 2108 + }, + { + "epoch": 4.192842942345925, + "grad_norm": 0.11130931558291521, + "learning_rate": 6.18707067701477e-06, + "loss": 0.3889, + "step": 2109 + }, + { + "epoch": 4.194831013916501, + "grad_norm": 0.10489299799335888, + "learning_rate": 6.157436238486671e-06, + "loss": 0.3868, + "step": 2110 + }, + { + "epoch": 4.1968190854870775, + "grad_norm": 0.11564230886287205, + "learning_rate": 6.127867021915235e-06, + "loss": 0.3829, + "step": 2111 + }, + { + "epoch": 4.198807157057654, + "grad_norm": 0.10738774680568806, + "learning_rate": 6.098363084286765e-06, + "loss": 0.3902, + "step": 2112 + }, + { + "epoch": 4.20079522862823, + "grad_norm": 0.10870599960288968, + "learning_rate": 6.068924482461715e-06, + "loss": 0.389, + "step": 2113 + }, + { + "epoch": 4.202783300198807, + "grad_norm": 0.1160575426010257, + "learning_rate": 6.0395512731746865e-06, + "loss": 0.3827, + "step": 2114 + }, + { + "epoch": 4.204771371769383, + "grad_norm": 0.11306539716214972, + "learning_rate": 6.0102435130342085e-06, + "loss": 0.3926, + "step": 2115 + }, + { + "epoch": 4.20675944333996, + "grad_norm": 0.11006470554733454, + "learning_rate": 5.981001258522683e-06, + "loss": 0.3984, + "step": 2116 + }, + { + "epoch": 4.208747514910537, + "grad_norm": 0.1099261910301456, + "learning_rate": 5.951824565996282e-06, + "loss": 0.3809, + "step": 2117 + }, + { + "epoch": 4.210735586481113, + "grad_norm": 0.11092749383381181, + "learning_rate": 5.922713491684811e-06, + "loss": 0.3862, + "step": 2118 + }, + { + "epoch": 4.21272365805169, + "grad_norm": 0.09827057623284234, + "learning_rate": 5.893668091691625e-06, + "loss": 0.3935, + "step": 2119 + }, + { + "epoch": 4.214711729622266, + "grad_norm": 0.1087451514423072, + "learning_rate": 5.864688421993499e-06, + "loss": 0.3906, + "step": 2120 + }, + { + "epoch": 4.216699801192843, + "grad_norm": 0.10709955492019746, + "learning_rate": 5.835774538440557e-06, + "loss": 0.3878, + "step": 2121 + }, + { + "epoch": 4.21868787276342, + "grad_norm": 0.10181961138592265, + "learning_rate": 5.806926496756084e-06, + "loss": 0.3812, + "step": 2122 + }, + { + "epoch": 4.220675944333996, + "grad_norm": 0.1062337392337739, + "learning_rate": 5.778144352536545e-06, + "loss": 0.3836, + "step": 2123 + }, + { + "epoch": 4.222664015904573, + "grad_norm": 0.10197581797668581, + "learning_rate": 5.74942816125136e-06, + "loss": 0.3824, + "step": 2124 + }, + { + "epoch": 4.224652087475149, + "grad_norm": 0.10219575099997118, + "learning_rate": 5.720777978242851e-06, + "loss": 0.3863, + "step": 2125 + }, + { + "epoch": 4.226640159045726, + "grad_norm": 0.1040267273692468, + "learning_rate": 5.692193858726134e-06, + "loss": 0.3874, + "step": 2126 + }, + { + "epoch": 4.2286282306163026, + "grad_norm": 0.11269177444316061, + "learning_rate": 5.663675857789006e-06, + "loss": 0.3865, + "step": 2127 + }, + { + "epoch": 4.230616302186879, + "grad_norm": 0.09995425296412798, + "learning_rate": 5.635224030391832e-06, + "loss": 0.3933, + "step": 2128 + }, + { + "epoch": 4.2326043737574555, + "grad_norm": 0.10689562263264148, + "learning_rate": 5.606838431367454e-06, + "loss": 0.3921, + "step": 2129 + }, + { + "epoch": 4.2345924453280315, + "grad_norm": 0.11122546471139345, + "learning_rate": 5.578519115421084e-06, + "loss": 0.3888, + "step": 2130 + }, + { + "epoch": 4.236580516898608, + "grad_norm": 0.10344950837778637, + "learning_rate": 5.550266137130154e-06, + "loss": 0.3947, + "step": 2131 + }, + { + "epoch": 4.238568588469185, + "grad_norm": 0.10216325643865894, + "learning_rate": 5.5220795509443085e-06, + "loss": 0.3901, + "step": 2132 + }, + { + "epoch": 4.240556660039761, + "grad_norm": 0.10694524647608616, + "learning_rate": 5.493959411185197e-06, + "loss": 0.3942, + "step": 2133 + }, + { + "epoch": 4.242544731610338, + "grad_norm": 0.10609283465242796, + "learning_rate": 5.465905772046425e-06, + "loss": 0.3793, + "step": 2134 + }, + { + "epoch": 4.244532803180914, + "grad_norm": 0.1079949840251626, + "learning_rate": 5.437918687593433e-06, + "loss": 0.4008, + "step": 2135 + }, + { + "epoch": 4.246520874751491, + "grad_norm": 0.11738376878238611, + "learning_rate": 5.40999821176341e-06, + "loss": 0.3848, + "step": 2136 + }, + { + "epoch": 4.248508946322068, + "grad_norm": 0.11082300608152638, + "learning_rate": 5.3821443983651525e-06, + "loss": 0.3987, + "step": 2137 + }, + { + "epoch": 4.250497017892644, + "grad_norm": 0.10101109241687987, + "learning_rate": 5.3543573010790095e-06, + "loss": 0.394, + "step": 2138 + }, + { + "epoch": 4.252485089463221, + "grad_norm": 0.11468964097949289, + "learning_rate": 5.326636973456741e-06, + "loss": 0.4014, + "step": 2139 + }, + { + "epoch": 4.254473161033797, + "grad_norm": 0.10452528767883948, + "learning_rate": 5.298983468921402e-06, + "loss": 0.383, + "step": 2140 + }, + { + "epoch": 4.256461232604374, + "grad_norm": 0.10689689020767205, + "learning_rate": 5.2713968407673225e-06, + "loss": 0.3879, + "step": 2141 + }, + { + "epoch": 4.258449304174951, + "grad_norm": 0.10520846283587974, + "learning_rate": 5.243877142159899e-06, + "loss": 0.3919, + "step": 2142 + }, + { + "epoch": 4.260437375745527, + "grad_norm": 0.11093695400298927, + "learning_rate": 5.21642442613556e-06, + "loss": 0.378, + "step": 2143 + }, + { + "epoch": 4.262425447316104, + "grad_norm": 0.11329605962872928, + "learning_rate": 5.1890387456016425e-06, + "loss": 0.3902, + "step": 2144 + }, + { + "epoch": 4.26441351888668, + "grad_norm": 0.10901614052465748, + "learning_rate": 5.161720153336287e-06, + "loss": 0.3883, + "step": 2145 + }, + { + "epoch": 4.2664015904572565, + "grad_norm": 0.11033984670188425, + "learning_rate": 5.134468701988345e-06, + "loss": 0.3848, + "step": 2146 + }, + { + "epoch": 4.2683896620278325, + "grad_norm": 0.1112581347434242, + "learning_rate": 5.107284444077265e-06, + "loss": 0.3777, + "step": 2147 + }, + { + "epoch": 4.270377733598409, + "grad_norm": 0.11604790354287506, + "learning_rate": 5.080167431993018e-06, + "loss": 0.3936, + "step": 2148 + }, + { + "epoch": 4.272365805168986, + "grad_norm": 0.10130942223708193, + "learning_rate": 5.053117717995939e-06, + "loss": 0.3916, + "step": 2149 + }, + { + "epoch": 4.274353876739562, + "grad_norm": 0.10942647069946232, + "learning_rate": 5.026135354216717e-06, + "loss": 0.3885, + "step": 2150 + }, + { + "epoch": 4.276341948310139, + "grad_norm": 0.11395657047861649, + "learning_rate": 4.999220392656203e-06, + "loss": 0.3945, + "step": 2151 + }, + { + "epoch": 4.278330019880716, + "grad_norm": 0.10700487248942017, + "learning_rate": 4.972372885185368e-06, + "loss": 0.389, + "step": 2152 + }, + { + "epoch": 4.280318091451292, + "grad_norm": 0.10391399259588788, + "learning_rate": 4.945592883545179e-06, + "loss": 0.3934, + "step": 2153 + }, + { + "epoch": 4.282306163021869, + "grad_norm": 0.10037438756173153, + "learning_rate": 4.918880439346496e-06, + "loss": 0.381, + "step": 2154 + }, + { + "epoch": 4.284294234592445, + "grad_norm": 0.10597748791604826, + "learning_rate": 4.892235604069999e-06, + "loss": 0.3916, + "step": 2155 + }, + { + "epoch": 4.286282306163022, + "grad_norm": 0.10257742187290152, + "learning_rate": 4.86565842906606e-06, + "loss": 0.3873, + "step": 2156 + }, + { + "epoch": 4.288270377733598, + "grad_norm": 0.10673572419919322, + "learning_rate": 4.8391489655546495e-06, + "loss": 0.3734, + "step": 2157 + }, + { + "epoch": 4.290258449304175, + "grad_norm": 0.11498553199702016, + "learning_rate": 4.81270726462526e-06, + "loss": 0.3931, + "step": 2158 + }, + { + "epoch": 4.292246520874752, + "grad_norm": 0.10798065238474126, + "learning_rate": 4.786333377236773e-06, + "loss": 0.4006, + "step": 2159 + }, + { + "epoch": 4.294234592445328, + "grad_norm": 0.09816552037676422, + "learning_rate": 4.760027354217394e-06, + "loss": 0.3847, + "step": 2160 + }, + { + "epoch": 4.296222664015905, + "grad_norm": 0.10237023642683052, + "learning_rate": 4.733789246264526e-06, + "loss": 0.3985, + "step": 2161 + }, + { + "epoch": 4.298210735586481, + "grad_norm": 0.09693280463118617, + "learning_rate": 4.707619103944696e-06, + "loss": 0.3851, + "step": 2162 + }, + { + "epoch": 4.300198807157058, + "grad_norm": 0.10605120716456504, + "learning_rate": 4.681516977693435e-06, + "loss": 0.3883, + "step": 2163 + }, + { + "epoch": 4.3021868787276345, + "grad_norm": 0.09960557557894663, + "learning_rate": 4.6554829178152035e-06, + "loss": 0.3921, + "step": 2164 + }, + { + "epoch": 4.3041749502982105, + "grad_norm": 0.09498386093083336, + "learning_rate": 4.629516974483279e-06, + "loss": 0.3931, + "step": 2165 + }, + { + "epoch": 4.306163021868787, + "grad_norm": 0.09991291721444792, + "learning_rate": 4.603619197739662e-06, + "loss": 0.3997, + "step": 2166 + }, + { + "epoch": 4.308151093439363, + "grad_norm": 0.10582505708208029, + "learning_rate": 4.577789637494978e-06, + "loss": 0.3879, + "step": 2167 + }, + { + "epoch": 4.31013916500994, + "grad_norm": 0.1114427277686471, + "learning_rate": 4.552028343528396e-06, + "loss": 0.3773, + "step": 2168 + }, + { + "epoch": 4.312127236580517, + "grad_norm": 0.09989286001429934, + "learning_rate": 4.526335365487509e-06, + "loss": 0.3837, + "step": 2169 + }, + { + "epoch": 4.314115308151093, + "grad_norm": 0.10342275152788724, + "learning_rate": 4.500710752888253e-06, + "loss": 0.3883, + "step": 2170 + }, + { + "epoch": 4.31610337972167, + "grad_norm": 0.10316817695567211, + "learning_rate": 4.47515455511482e-06, + "loss": 0.4026, + "step": 2171 + }, + { + "epoch": 4.318091451292246, + "grad_norm": 0.09876129834768971, + "learning_rate": 4.449666821419536e-06, + "loss": 0.3831, + "step": 2172 + }, + { + "epoch": 4.320079522862823, + "grad_norm": 0.11477440375009836, + "learning_rate": 4.424247600922793e-06, + "loss": 0.3872, + "step": 2173 + }, + { + "epoch": 4.3220675944334, + "grad_norm": 0.10099717624479602, + "learning_rate": 4.398896942612934e-06, + "loss": 0.3947, + "step": 2174 + }, + { + "epoch": 4.324055666003976, + "grad_norm": 0.10121838635908194, + "learning_rate": 4.373614895346178e-06, + "loss": 0.3931, + "step": 2175 + }, + { + "epoch": 4.326043737574553, + "grad_norm": 0.10716998309395787, + "learning_rate": 4.348401507846509e-06, + "loss": 0.3837, + "step": 2176 + }, + { + "epoch": 4.328031809145129, + "grad_norm": 0.0976333146806361, + "learning_rate": 4.323256828705593e-06, + "loss": 0.3882, + "step": 2177 + }, + { + "epoch": 4.330019880715706, + "grad_norm": 0.10697024916266006, + "learning_rate": 4.2981809063826715e-06, + "loss": 0.3816, + "step": 2178 + }, + { + "epoch": 4.332007952286283, + "grad_norm": 0.10052526954739248, + "learning_rate": 4.273173789204488e-06, + "loss": 0.3811, + "step": 2179 + }, + { + "epoch": 4.333996023856859, + "grad_norm": 0.09928114782318379, + "learning_rate": 4.248235525365174e-06, + "loss": 0.379, + "step": 2180 + }, + { + "epoch": 4.335984095427436, + "grad_norm": 0.0999330585185468, + "learning_rate": 4.2233661629261746e-06, + "loss": 0.4004, + "step": 2181 + }, + { + "epoch": 4.337972166998012, + "grad_norm": 0.1093719529042732, + "learning_rate": 4.198565749816142e-06, + "loss": 0.4014, + "step": 2182 + }, + { + "epoch": 4.3399602385685885, + "grad_norm": 0.0990449385264406, + "learning_rate": 4.173834333830842e-06, + "loss": 0.3851, + "step": 2183 + }, + { + "epoch": 4.341948310139165, + "grad_norm": 0.09842163804658173, + "learning_rate": 4.149171962633083e-06, + "loss": 0.3771, + "step": 2184 + }, + { + "epoch": 4.343936381709741, + "grad_norm": 0.098524628734478, + "learning_rate": 4.124578683752596e-06, + "loss": 0.3851, + "step": 2185 + }, + { + "epoch": 4.345924453280318, + "grad_norm": 0.09531813272536584, + "learning_rate": 4.100054544585961e-06, + "loss": 0.3912, + "step": 2186 + }, + { + "epoch": 4.347912524850894, + "grad_norm": 0.09791263411124397, + "learning_rate": 4.075599592396509e-06, + "loss": 0.3826, + "step": 2187 + }, + { + "epoch": 4.349900596421471, + "grad_norm": 0.09273942541729, + "learning_rate": 4.051213874314237e-06, + "loss": 0.3951, + "step": 2188 + }, + { + "epoch": 4.351888667992048, + "grad_norm": 0.1050206316055863, + "learning_rate": 4.026897437335708e-06, + "loss": 0.3808, + "step": 2189 + }, + { + "epoch": 4.353876739562624, + "grad_norm": 0.10040959376895207, + "learning_rate": 4.002650328323969e-06, + "loss": 0.385, + "step": 2190 + }, + { + "epoch": 4.355864811133201, + "grad_norm": 0.1016395750899665, + "learning_rate": 3.978472594008454e-06, + "loss": 0.3944, + "step": 2191 + }, + { + "epoch": 4.357852882703777, + "grad_norm": 0.0993656184763306, + "learning_rate": 3.954364280984906e-06, + "loss": 0.3904, + "step": 2192 + }, + { + "epoch": 4.359840954274354, + "grad_norm": 0.0943218423219894, + "learning_rate": 3.930325435715245e-06, + "loss": 0.3936, + "step": 2193 + }, + { + "epoch": 4.361829025844931, + "grad_norm": 0.09681840686767862, + "learning_rate": 3.906356104527565e-06, + "loss": 0.3947, + "step": 2194 + }, + { + "epoch": 4.363817097415507, + "grad_norm": 0.09709721598226503, + "learning_rate": 3.882456333615952e-06, + "loss": 0.3878, + "step": 2195 + }, + { + "epoch": 4.365805168986084, + "grad_norm": 0.09913382152372414, + "learning_rate": 3.858626169040447e-06, + "loss": 0.3843, + "step": 2196 + }, + { + "epoch": 4.36779324055666, + "grad_norm": 0.09980848225328527, + "learning_rate": 3.834865656726945e-06, + "loss": 0.3909, + "step": 2197 + }, + { + "epoch": 4.369781312127237, + "grad_norm": 0.09420624149064247, + "learning_rate": 3.8111748424671e-06, + "loss": 0.3942, + "step": 2198 + }, + { + "epoch": 4.3717693836978135, + "grad_norm": 0.09687823459593438, + "learning_rate": 3.7875537719182486e-06, + "loss": 0.3869, + "step": 2199 + }, + { + "epoch": 4.3737574552683895, + "grad_norm": 0.10595045404187366, + "learning_rate": 3.7640024906033134e-06, + "loss": 0.3891, + "step": 2200 + }, + { + "epoch": 4.3757455268389664, + "grad_norm": 0.09612367938332525, + "learning_rate": 3.7405210439107254e-06, + "loss": 0.379, + "step": 2201 + }, + { + "epoch": 4.3777335984095425, + "grad_norm": 0.1023095190589734, + "learning_rate": 3.7171094770943026e-06, + "loss": 0.3875, + "step": 2202 + }, + { + "epoch": 4.379721669980119, + "grad_norm": 0.099151488292962, + "learning_rate": 3.69376783527323e-06, + "loss": 0.3898, + "step": 2203 + }, + { + "epoch": 4.381709741550695, + "grad_norm": 0.09958329292525986, + "learning_rate": 3.6704961634319046e-06, + "loss": 0.3851, + "step": 2204 + }, + { + "epoch": 4.383697813121272, + "grad_norm": 0.09555701123264873, + "learning_rate": 3.6472945064198827e-06, + "loss": 0.3889, + "step": 2205 + }, + { + "epoch": 4.385685884691849, + "grad_norm": 0.09980480198930511, + "learning_rate": 3.6241629089517873e-06, + "loss": 0.3804, + "step": 2206 + }, + { + "epoch": 4.387673956262425, + "grad_norm": 0.09835469378518892, + "learning_rate": 3.6011014156072242e-06, + "loss": 0.3924, + "step": 2207 + }, + { + "epoch": 4.389662027833002, + "grad_norm": 0.10440527313630382, + "learning_rate": 3.5781100708306827e-06, + "loss": 0.3905, + "step": 2208 + }, + { + "epoch": 4.391650099403579, + "grad_norm": 0.10830643367489604, + "learning_rate": 3.5551889189314738e-06, + "loss": 0.396, + "step": 2209 + }, + { + "epoch": 4.393638170974155, + "grad_norm": 0.10721642274121215, + "learning_rate": 3.532338004083631e-06, + "loss": 0.3916, + "step": 2210 + }, + { + "epoch": 4.395626242544732, + "grad_norm": 0.09987362704278649, + "learning_rate": 3.5095573703258023e-06, + "loss": 0.3914, + "step": 2211 + }, + { + "epoch": 4.397614314115308, + "grad_norm": 0.10841817698714662, + "learning_rate": 3.4868470615612247e-06, + "loss": 0.3831, + "step": 2212 + }, + { + "epoch": 4.399602385685885, + "grad_norm": 0.10340887834099978, + "learning_rate": 3.464207121557581e-06, + "loss": 0.4053, + "step": 2213 + }, + { + "epoch": 4.401590457256461, + "grad_norm": 0.09365502481903334, + "learning_rate": 3.4416375939469427e-06, + "loss": 0.3876, + "step": 2214 + }, + { + "epoch": 4.403578528827038, + "grad_norm": 0.0999030246429339, + "learning_rate": 3.4191385222256845e-06, + "loss": 0.3802, + "step": 2215 + }, + { + "epoch": 4.405566600397615, + "grad_norm": 0.1012356919744598, + "learning_rate": 3.396709949754393e-06, + "loss": 0.3889, + "step": 2216 + }, + { + "epoch": 4.407554671968191, + "grad_norm": 0.10343516451567464, + "learning_rate": 3.37435191975779e-06, + "loss": 0.3996, + "step": 2217 + }, + { + "epoch": 4.4095427435387675, + "grad_norm": 0.596392899901363, + "learning_rate": 3.3520644753246432e-06, + "loss": 0.4044, + "step": 2218 + }, + { + "epoch": 4.4115308151093435, + "grad_norm": 0.09379139186711735, + "learning_rate": 3.3298476594076923e-06, + "loss": 0.3791, + "step": 2219 + }, + { + "epoch": 4.41351888667992, + "grad_norm": 0.10187480860768795, + "learning_rate": 3.307701514823558e-06, + "loss": 0.3834, + "step": 2220 + }, + { + "epoch": 4.415506958250497, + "grad_norm": 0.10571085909393454, + "learning_rate": 3.28562608425266e-06, + "loss": 0.39, + "step": 2221 + }, + { + "epoch": 4.417495029821073, + "grad_norm": 0.10014337973432315, + "learning_rate": 3.263621410239144e-06, + "loss": 0.3913, + "step": 2222 + }, + { + "epoch": 4.41948310139165, + "grad_norm": 0.1092625763710919, + "learning_rate": 3.241687535190776e-06, + "loss": 0.4004, + "step": 2223 + }, + { + "epoch": 4.421471172962226, + "grad_norm": 0.09879099002909277, + "learning_rate": 3.2198245013788943e-06, + "loss": 0.393, + "step": 2224 + }, + { + "epoch": 4.423459244532803, + "grad_norm": 0.36421280599677913, + "learning_rate": 3.198032350938309e-06, + "loss": 0.3872, + "step": 2225 + }, + { + "epoch": 4.42544731610338, + "grad_norm": 0.10043741715437189, + "learning_rate": 3.1763111258672084e-06, + "loss": 0.3811, + "step": 2226 + }, + { + "epoch": 4.427435387673956, + "grad_norm": 0.10483503269426613, + "learning_rate": 3.1546608680271105e-06, + "loss": 0.3925, + "step": 2227 + }, + { + "epoch": 4.429423459244533, + "grad_norm": 0.09693474001090108, + "learning_rate": 3.1330816191427505e-06, + "loss": 0.3928, + "step": 2228 + }, + { + "epoch": 4.431411530815109, + "grad_norm": 0.09737601508794358, + "learning_rate": 3.111573420802021e-06, + "loss": 0.3834, + "step": 2229 + }, + { + "epoch": 4.433399602385686, + "grad_norm": 0.1010936127284044, + "learning_rate": 3.0901363144558804e-06, + "loss": 0.3856, + "step": 2230 + }, + { + "epoch": 4.435387673956263, + "grad_norm": 0.1048068301583792, + "learning_rate": 3.0687703414182814e-06, + "loss": 0.3873, + "step": 2231 + }, + { + "epoch": 4.437375745526839, + "grad_norm": 0.09218979311028978, + "learning_rate": 3.0474755428660853e-06, + "loss": 0.3869, + "step": 2232 + }, + { + "epoch": 4.439363817097416, + "grad_norm": 0.09891825070474362, + "learning_rate": 3.0262519598389883e-06, + "loss": 0.392, + "step": 2233 + }, + { + "epoch": 4.441351888667992, + "grad_norm": 0.09824471049330011, + "learning_rate": 3.005099633239428e-06, + "loss": 0.3878, + "step": 2234 + }, + { + "epoch": 4.443339960238569, + "grad_norm": 0.10111959954987662, + "learning_rate": 2.9840186038325326e-06, + "loss": 0.3925, + "step": 2235 + }, + { + "epoch": 4.4453280318091455, + "grad_norm": 0.09410522943055849, + "learning_rate": 2.963008912246008e-06, + "loss": 0.391, + "step": 2236 + }, + { + "epoch": 4.4473161033797215, + "grad_norm": 0.1025495604201037, + "learning_rate": 2.9420705989700882e-06, + "loss": 0.3953, + "step": 2237 + }, + { + "epoch": 4.449304174950298, + "grad_norm": 0.09987461410476814, + "learning_rate": 2.9212037043574402e-06, + "loss": 0.384, + "step": 2238 + }, + { + "epoch": 4.451292246520874, + "grad_norm": 0.0926074391620245, + "learning_rate": 2.9004082686230916e-06, + "loss": 0.3895, + "step": 2239 + }, + { + "epoch": 4.453280318091451, + "grad_norm": 0.09482564866145016, + "learning_rate": 2.879684331844361e-06, + "loss": 0.3762, + "step": 2240 + }, + { + "epoch": 4.455268389662028, + "grad_norm": 0.13865518028486007, + "learning_rate": 2.859031933960763e-06, + "loss": 0.3906, + "step": 2241 + }, + { + "epoch": 4.457256461232604, + "grad_norm": 0.09264050462975766, + "learning_rate": 2.8384511147739437e-06, + "loss": 0.4027, + "step": 2242 + }, + { + "epoch": 4.459244532803181, + "grad_norm": 0.10005832449719292, + "learning_rate": 2.817941913947606e-06, + "loss": 0.3805, + "step": 2243 + }, + { + "epoch": 4.461232604373757, + "grad_norm": 0.09821994260465117, + "learning_rate": 2.797504371007427e-06, + "loss": 0.3937, + "step": 2244 + }, + { + "epoch": 4.463220675944334, + "grad_norm": 0.09631788391400439, + "learning_rate": 2.777138525340979e-06, + "loss": 0.3827, + "step": 2245 + }, + { + "epoch": 4.465208747514911, + "grad_norm": 0.09039255877447051, + "learning_rate": 2.756844416197666e-06, + "loss": 0.3825, + "step": 2246 + }, + { + "epoch": 4.467196819085487, + "grad_norm": 0.09465359257625011, + "learning_rate": 2.7366220826886294e-06, + "loss": 0.3809, + "step": 2247 + }, + { + "epoch": 4.469184890656064, + "grad_norm": 0.09254465134150146, + "learning_rate": 2.716471563786689e-06, + "loss": 0.3979, + "step": 2248 + }, + { + "epoch": 4.47117296222664, + "grad_norm": 0.09612944852010466, + "learning_rate": 2.6963928983262745e-06, + "loss": 0.3848, + "step": 2249 + }, + { + "epoch": 4.473161033797217, + "grad_norm": 0.105893288150653, + "learning_rate": 2.6763861250033118e-06, + "loss": 0.3916, + "step": 2250 + }, + { + "epoch": 4.475149105367794, + "grad_norm": 0.09300696972063668, + "learning_rate": 2.6564512823751987e-06, + "loss": 0.3994, + "step": 2251 + }, + { + "epoch": 4.47713717693837, + "grad_norm": 0.0930437296362057, + "learning_rate": 2.636588408860701e-06, + "loss": 0.3853, + "step": 2252 + }, + { + "epoch": 4.4791252485089466, + "grad_norm": 0.09325528669971057, + "learning_rate": 2.616797542739873e-06, + "loss": 0.3775, + "step": 2253 + }, + { + "epoch": 4.481113320079523, + "grad_norm": 0.09465774340633075, + "learning_rate": 2.597078722154014e-06, + "loss": 0.387, + "step": 2254 + }, + { + "epoch": 4.4831013916500995, + "grad_norm": 0.13768709356858644, + "learning_rate": 2.5774319851055564e-06, + "loss": 0.383, + "step": 2255 + }, + { + "epoch": 4.485089463220676, + "grad_norm": 0.09645823543283955, + "learning_rate": 2.5578573694580254e-06, + "loss": 0.3807, + "step": 2256 + }, + { + "epoch": 4.487077534791252, + "grad_norm": 0.09758883625704344, + "learning_rate": 2.5383549129359473e-06, + "loss": 0.3853, + "step": 2257 + }, + { + "epoch": 4.489065606361829, + "grad_norm": 0.09776875136116465, + "learning_rate": 2.51892465312479e-06, + "loss": 0.3795, + "step": 2258 + }, + { + "epoch": 4.491053677932405, + "grad_norm": 0.09531997029739311, + "learning_rate": 2.4995666274708664e-06, + "loss": 0.3859, + "step": 2259 + }, + { + "epoch": 4.493041749502982, + "grad_norm": 0.09613584567556226, + "learning_rate": 2.4802808732812935e-06, + "loss": 0.3963, + "step": 2260 + }, + { + "epoch": 4.495029821073558, + "grad_norm": 0.10310822349287414, + "learning_rate": 2.4610674277239046e-06, + "loss": 0.3944, + "step": 2261 + }, + { + "epoch": 4.497017892644135, + "grad_norm": 0.09144966609722156, + "learning_rate": 2.4419263278271643e-06, + "loss": 0.3956, + "step": 2262 + }, + { + "epoch": 4.499005964214712, + "grad_norm": 0.09492031385454708, + "learning_rate": 2.4228576104801295e-06, + "loss": 0.3877, + "step": 2263 + }, + { + "epoch": 4.500994035785288, + "grad_norm": 0.09327970875824858, + "learning_rate": 2.4038613124323496e-06, + "loss": 0.4014, + "step": 2264 + }, + { + "epoch": 4.502982107355865, + "grad_norm": 0.09558401559285043, + "learning_rate": 2.3849374702938065e-06, + "loss": 0.397, + "step": 2265 + }, + { + "epoch": 4.504970178926442, + "grad_norm": 0.09747152770568815, + "learning_rate": 2.366086120534852e-06, + "loss": 0.3801, + "step": 2266 + }, + { + "epoch": 4.506958250497018, + "grad_norm": 0.09409003277053672, + "learning_rate": 2.3473072994861167e-06, + "loss": 0.3935, + "step": 2267 + }, + { + "epoch": 4.508946322067595, + "grad_norm": 0.09512871106275056, + "learning_rate": 2.3286010433384656e-06, + "loss": 0.3903, + "step": 2268 + }, + { + "epoch": 4.510934393638171, + "grad_norm": 0.0937640062202609, + "learning_rate": 2.3099673881429086e-06, + "loss": 0.3839, + "step": 2269 + }, + { + "epoch": 4.512922465208748, + "grad_norm": 0.09330775039486519, + "learning_rate": 2.2914063698105338e-06, + "loss": 0.3978, + "step": 2270 + }, + { + "epoch": 4.514910536779324, + "grad_norm": 0.09712706610467936, + "learning_rate": 2.2729180241124517e-06, + "loss": 0.3844, + "step": 2271 + }, + { + "epoch": 4.5168986083499005, + "grad_norm": 0.09403036480214051, + "learning_rate": 2.254502386679711e-06, + "loss": 0.3943, + "step": 2272 + }, + { + "epoch": 4.518886679920477, + "grad_norm": 0.10067963044125913, + "learning_rate": 2.2361594930032338e-06, + "loss": 0.4011, + "step": 2273 + }, + { + "epoch": 4.5208747514910534, + "grad_norm": 0.09169835381585156, + "learning_rate": 2.2178893784337594e-06, + "loss": 0.3937, + "step": 2274 + }, + { + "epoch": 4.52286282306163, + "grad_norm": 0.08850262851562063, + "learning_rate": 2.1996920781817544e-06, + "loss": 0.3802, + "step": 2275 + }, + { + "epoch": 4.524850894632207, + "grad_norm": 0.093917345011886, + "learning_rate": 2.1815676273173604e-06, + "loss": 0.3927, + "step": 2276 + }, + { + "epoch": 4.526838966202783, + "grad_norm": 0.09501270473791533, + "learning_rate": 2.1635160607703256e-06, + "loss": 0.3828, + "step": 2277 + }, + { + "epoch": 4.52882703777336, + "grad_norm": 0.09135115060354339, + "learning_rate": 2.145537413329932e-06, + "loss": 0.3835, + "step": 2278 + }, + { + "epoch": 4.530815109343936, + "grad_norm": 0.08990517715951528, + "learning_rate": 2.1276317196449315e-06, + "loss": 0.3955, + "step": 2279 + }, + { + "epoch": 4.532803180914513, + "grad_norm": 0.30857334159915706, + "learning_rate": 2.109799014223475e-06, + "loss": 0.3999, + "step": 2280 + }, + { + "epoch": 4.534791252485089, + "grad_norm": 0.09985728888122387, + "learning_rate": 2.092039331433049e-06, + "loss": 0.388, + "step": 2281 + }, + { + "epoch": 4.536779324055666, + "grad_norm": 0.0938150353363061, + "learning_rate": 2.0743527055004176e-06, + "loss": 0.3816, + "step": 2282 + }, + { + "epoch": 4.538767395626243, + "grad_norm": 0.09823097175682717, + "learning_rate": 2.0567391705115416e-06, + "loss": 0.3965, + "step": 2283 + }, + { + "epoch": 4.540755467196819, + "grad_norm": 0.086397808420212, + "learning_rate": 2.0391987604115204e-06, + "loss": 0.3896, + "step": 2284 + }, + { + "epoch": 4.542743538767396, + "grad_norm": 0.09653654555914719, + "learning_rate": 2.021731509004523e-06, + "loss": 0.4017, + "step": 2285 + }, + { + "epoch": 4.544731610337972, + "grad_norm": 0.09598593518878275, + "learning_rate": 2.0043374499537373e-06, + "loss": 0.3805, + "step": 2286 + }, + { + "epoch": 4.546719681908549, + "grad_norm": 0.09535832154199035, + "learning_rate": 1.987016616781281e-06, + "loss": 0.389, + "step": 2287 + }, + { + "epoch": 4.548707753479126, + "grad_norm": 0.09030336197873803, + "learning_rate": 1.969769042868155e-06, + "loss": 0.3782, + "step": 2288 + }, + { + "epoch": 4.550695825049702, + "grad_norm": 0.0900881624984173, + "learning_rate": 1.9525947614541694e-06, + "loss": 0.3825, + "step": 2289 + }, + { + "epoch": 4.5526838966202785, + "grad_norm": 0.0932574491444472, + "learning_rate": 1.9354938056378935e-06, + "loss": 0.3779, + "step": 2290 + }, + { + "epoch": 4.5546719681908545, + "grad_norm": 0.09410751821503942, + "learning_rate": 1.918466208376573e-06, + "loss": 0.3943, + "step": 2291 + }, + { + "epoch": 4.556660039761431, + "grad_norm": 0.09098872687261256, + "learning_rate": 1.9015120024860811e-06, + "loss": 0.3785, + "step": 2292 + }, + { + "epoch": 4.558648111332008, + "grad_norm": 0.08926273695926924, + "learning_rate": 1.8846312206408468e-06, + "loss": 0.3871, + "step": 2293 + }, + { + "epoch": 4.560636182902584, + "grad_norm": 0.08923403092076015, + "learning_rate": 1.8678238953737926e-06, + "loss": 0.392, + "step": 2294 + }, + { + "epoch": 4.562624254473161, + "grad_norm": 0.08970320411457725, + "learning_rate": 1.8510900590762838e-06, + "loss": 0.3769, + "step": 2295 + }, + { + "epoch": 4.564612326043737, + "grad_norm": 0.0928633771471017, + "learning_rate": 1.8344297439980475e-06, + "loss": 0.4015, + "step": 2296 + }, + { + "epoch": 4.566600397614314, + "grad_norm": 0.09530337273078132, + "learning_rate": 1.81784298224712e-06, + "loss": 0.393, + "step": 2297 + }, + { + "epoch": 4.568588469184891, + "grad_norm": 0.09252969398882711, + "learning_rate": 1.8013298057897932e-06, + "loss": 0.3896, + "step": 2298 + }, + { + "epoch": 4.570576540755467, + "grad_norm": 0.09219676491217316, + "learning_rate": 1.7848902464505303e-06, + "loss": 0.3884, + "step": 2299 + }, + { + "epoch": 4.572564612326044, + "grad_norm": 0.09242046744560646, + "learning_rate": 1.7685243359119298e-06, + "loss": 0.3792, + "step": 2300 + }, + { + "epoch": 4.57455268389662, + "grad_norm": 0.08935728010398887, + "learning_rate": 1.7522321057146508e-06, + "loss": 0.3864, + "step": 2301 + }, + { + "epoch": 4.576540755467197, + "grad_norm": 0.0931849223946967, + "learning_rate": 1.7360135872573502e-06, + "loss": 0.3801, + "step": 2302 + }, + { + "epoch": 4.578528827037774, + "grad_norm": 0.09107346999398934, + "learning_rate": 1.71986881179663e-06, + "loss": 0.3974, + "step": 2303 + }, + { + "epoch": 4.58051689860835, + "grad_norm": 0.09210780909523932, + "learning_rate": 1.70379781044697e-06, + "loss": 0.3834, + "step": 2304 + }, + { + "epoch": 4.582504970178927, + "grad_norm": 0.0884785884728689, + "learning_rate": 1.6878006141806747e-06, + "loss": 0.397, + "step": 2305 + }, + { + "epoch": 4.584493041749503, + "grad_norm": 0.09331749408939408, + "learning_rate": 1.6718772538278117e-06, + "loss": 0.3915, + "step": 2306 + }, + { + "epoch": 4.58648111332008, + "grad_norm": 0.09223920546338787, + "learning_rate": 1.6560277600761487e-06, + "loss": 0.3816, + "step": 2307 + }, + { + "epoch": 4.588469184890656, + "grad_norm": 0.09681340642252376, + "learning_rate": 1.6402521634710877e-06, + "loss": 0.3785, + "step": 2308 + }, + { + "epoch": 4.5904572564612325, + "grad_norm": 0.09629143650589808, + "learning_rate": 1.6245504944156332e-06, + "loss": 0.3916, + "step": 2309 + }, + { + "epoch": 4.592445328031809, + "grad_norm": 0.09627909480747703, + "learning_rate": 1.608922783170299e-06, + "loss": 0.3833, + "step": 2310 + }, + { + "epoch": 4.594433399602385, + "grad_norm": 0.08841171190126924, + "learning_rate": 1.5933690598530737e-06, + "loss": 0.3752, + "step": 2311 + }, + { + "epoch": 4.596421471172962, + "grad_norm": 0.09050689164750318, + "learning_rate": 1.577889354439357e-06, + "loss": 0.3826, + "step": 2312 + }, + { + "epoch": 4.598409542743539, + "grad_norm": 0.08629910372587375, + "learning_rate": 1.5624836967618939e-06, + "loss": 0.3769, + "step": 2313 + }, + { + "epoch": 4.600397614314115, + "grad_norm": 0.09722776788968594, + "learning_rate": 1.547152116510726e-06, + "loss": 0.3923, + "step": 2314 + }, + { + "epoch": 4.602385685884692, + "grad_norm": 0.09237377126717078, + "learning_rate": 1.5318946432331338e-06, + "loss": 0.3878, + "step": 2315 + }, + { + "epoch": 4.604373757455268, + "grad_norm": 0.08858713383649396, + "learning_rate": 1.5167113063335736e-06, + "loss": 0.3945, + "step": 2316 + }, + { + "epoch": 4.606361829025845, + "grad_norm": 0.10149378277319007, + "learning_rate": 1.5016021350736253e-06, + "loss": 0.394, + "step": 2317 + }, + { + "epoch": 4.608349900596421, + "grad_norm": 0.09370047111339964, + "learning_rate": 1.4865671585719477e-06, + "loss": 0.3791, + "step": 2318 + }, + { + "epoch": 4.610337972166998, + "grad_norm": 0.10062037398190488, + "learning_rate": 1.4716064058041846e-06, + "loss": 0.3834, + "step": 2319 + }, + { + "epoch": 4.612326043737575, + "grad_norm": 0.08806754588403382, + "learning_rate": 1.4567199056029613e-06, + "loss": 0.3972, + "step": 2320 + }, + { + "epoch": 4.614314115308151, + "grad_norm": 0.08442134067610135, + "learning_rate": 1.4419076866577864e-06, + "loss": 0.3748, + "step": 2321 + }, + { + "epoch": 4.616302186878728, + "grad_norm": 0.08796926416731948, + "learning_rate": 1.4271697775150162e-06, + "loss": 0.3834, + "step": 2322 + }, + { + "epoch": 4.618290258449305, + "grad_norm": 0.0910818428968802, + "learning_rate": 1.4125062065777973e-06, + "loss": 0.3827, + "step": 2323 + }, + { + "epoch": 4.620278330019881, + "grad_norm": 0.09065119233786946, + "learning_rate": 1.397917002106013e-06, + "loss": 0.3881, + "step": 2324 + }, + { + "epoch": 4.6222664015904575, + "grad_norm": 0.09104282484132259, + "learning_rate": 1.3834021922162167e-06, + "loss": 0.4014, + "step": 2325 + }, + { + "epoch": 4.6242544731610336, + "grad_norm": 0.09429251418317382, + "learning_rate": 1.3689618048815966e-06, + "loss": 0.3883, + "step": 2326 + }, + { + "epoch": 4.6262425447316105, + "grad_norm": 0.09749806535855796, + "learning_rate": 1.3545958679319093e-06, + "loss": 0.3879, + "step": 2327 + }, + { + "epoch": 4.6282306163021865, + "grad_norm": 0.09057348522000637, + "learning_rate": 1.3403044090534301e-06, + "loss": 0.3911, + "step": 2328 + }, + { + "epoch": 4.630218687872763, + "grad_norm": 0.08893250766924893, + "learning_rate": 1.3260874557889002e-06, + "loss": 0.4004, + "step": 2329 + }, + { + "epoch": 4.63220675944334, + "grad_norm": 0.08812327485932064, + "learning_rate": 1.3119450355374651e-06, + "loss": 0.3871, + "step": 2330 + }, + { + "epoch": 4.634194831013916, + "grad_norm": 0.08984430375875535, + "learning_rate": 1.2978771755546382e-06, + "loss": 0.389, + "step": 2331 + }, + { + "epoch": 4.636182902584493, + "grad_norm": 0.11238909857144332, + "learning_rate": 1.2838839029522344e-06, + "loss": 0.3882, + "step": 2332 + }, + { + "epoch": 4.63817097415507, + "grad_norm": 0.08542870167083381, + "learning_rate": 1.2699652446983213e-06, + "loss": 0.3998, + "step": 2333 + }, + { + "epoch": 4.640159045725646, + "grad_norm": 0.09072455538699621, + "learning_rate": 1.2561212276171753e-06, + "loss": 0.3906, + "step": 2334 + }, + { + "epoch": 4.642147117296223, + "grad_norm": 0.09077176680951171, + "learning_rate": 1.242351878389214e-06, + "loss": 0.399, + "step": 2335 + }, + { + "epoch": 4.644135188866799, + "grad_norm": 0.09087377061392632, + "learning_rate": 1.228657223550962e-06, + "loss": 0.3775, + "step": 2336 + }, + { + "epoch": 4.646123260437376, + "grad_norm": 0.08493200594804953, + "learning_rate": 1.2150372894949869e-06, + "loss": 0.3984, + "step": 2337 + }, + { + "epoch": 4.648111332007952, + "grad_norm": 0.08817667865233096, + "learning_rate": 1.2014921024698523e-06, + "loss": 0.3922, + "step": 2338 + }, + { + "epoch": 4.650099403578529, + "grad_norm": 0.08927232121946307, + "learning_rate": 1.1880216885800676e-06, + "loss": 0.3971, + "step": 2339 + }, + { + "epoch": 4.652087475149106, + "grad_norm": 0.08613839924431113, + "learning_rate": 1.1746260737860449e-06, + "loss": 0.3983, + "step": 2340 + }, + { + "epoch": 4.654075546719682, + "grad_norm": 0.0918930802800742, + "learning_rate": 1.1613052839040395e-06, + "loss": 0.3911, + "step": 2341 + }, + { + "epoch": 4.656063618290259, + "grad_norm": 0.08905521353173235, + "learning_rate": 1.148059344606094e-06, + "loss": 0.3769, + "step": 2342 + }, + { + "epoch": 4.658051689860835, + "grad_norm": 0.08840020767025152, + "learning_rate": 1.1348882814200103e-06, + "loss": 0.3869, + "step": 2343 + }, + { + "epoch": 4.6600397614314115, + "grad_norm": 0.09020713218255047, + "learning_rate": 1.1217921197292836e-06, + "loss": 0.3808, + "step": 2344 + }, + { + "epoch": 4.662027833001988, + "grad_norm": 0.08584989452622944, + "learning_rate": 1.1087708847730494e-06, + "loss": 0.387, + "step": 2345 + }, + { + "epoch": 4.664015904572564, + "grad_norm": 0.08782265773096049, + "learning_rate": 1.095824601646056e-06, + "loss": 0.3805, + "step": 2346 + }, + { + "epoch": 4.666003976143141, + "grad_norm": 0.0856685513359666, + "learning_rate": 1.0829532952985987e-06, + "loss": 0.3816, + "step": 2347 + }, + { + "epoch": 4.667992047713717, + "grad_norm": 0.09328575113682944, + "learning_rate": 1.0701569905364706e-06, + "loss": 0.3886, + "step": 2348 + }, + { + "epoch": 4.669980119284294, + "grad_norm": 0.09334769831602094, + "learning_rate": 1.0574357120209266e-06, + "loss": 0.388, + "step": 2349 + }, + { + "epoch": 4.671968190854871, + "grad_norm": 0.09183604939307251, + "learning_rate": 1.0447894842686269e-06, + "loss": 0.3854, + "step": 2350 + }, + { + "epoch": 4.673956262425447, + "grad_norm": 0.08849811025639472, + "learning_rate": 1.0322183316515955e-06, + "loss": 0.3928, + "step": 2351 + }, + { + "epoch": 4.675944333996024, + "grad_norm": 0.0861408438176817, + "learning_rate": 1.0197222783971728e-06, + "loss": 0.382, + "step": 2352 + }, + { + "epoch": 4.6779324055666, + "grad_norm": 0.09179102020068529, + "learning_rate": 1.0073013485879523e-06, + "loss": 0.3924, + "step": 2353 + }, + { + "epoch": 4.679920477137177, + "grad_norm": 0.08985169454264477, + "learning_rate": 9.949555661617638e-07, + "loss": 0.3911, + "step": 2354 + }, + { + "epoch": 4.681908548707754, + "grad_norm": 0.08712350247200025, + "learning_rate": 9.826849549116103e-07, + "loss": 0.3851, + "step": 2355 + }, + { + "epoch": 4.68389662027833, + "grad_norm": 0.08603921934702671, + "learning_rate": 9.704895384856106e-07, + "loss": 0.39, + "step": 2356 + }, + { + "epoch": 4.685884691848907, + "grad_norm": 0.08904593449849948, + "learning_rate": 9.583693403869864e-07, + "loss": 0.3821, + "step": 2357 + }, + { + "epoch": 4.687872763419483, + "grad_norm": 0.09273629581237089, + "learning_rate": 9.463243839739778e-07, + "loss": 0.382, + "step": 2358 + }, + { + "epoch": 4.68986083499006, + "grad_norm": 0.0926055652891825, + "learning_rate": 9.343546924598291e-07, + "loss": 0.3652, + "step": 2359 + }, + { + "epoch": 4.691848906560637, + "grad_norm": 0.08902119464039318, + "learning_rate": 9.224602889127321e-07, + "loss": 0.3839, + "step": 2360 + }, + { + "epoch": 4.693836978131213, + "grad_norm": 0.09461665383225337, + "learning_rate": 9.106411962557816e-07, + "loss": 0.3815, + "step": 2361 + }, + { + "epoch": 4.6958250497017895, + "grad_norm": 0.08978190990006801, + "learning_rate": 8.988974372669302e-07, + "loss": 0.3893, + "step": 2362 + }, + { + "epoch": 4.6978131212723655, + "grad_norm": 0.08480912747580577, + "learning_rate": 8.872290345789492e-07, + "loss": 0.3763, + "step": 2363 + }, + { + "epoch": 4.699801192842942, + "grad_norm": 0.08888694711926962, + "learning_rate": 8.756360106793793e-07, + "loss": 0.3827, + "step": 2364 + }, + { + "epoch": 4.701789264413518, + "grad_norm": 0.08768867258285498, + "learning_rate": 8.641183879104864e-07, + "loss": 0.3948, + "step": 2365 + }, + { + "epoch": 4.703777335984095, + "grad_norm": 0.0862141848794459, + "learning_rate": 8.526761884692259e-07, + "loss": 0.3924, + "step": 2366 + }, + { + "epoch": 4.705765407554672, + "grad_norm": 0.0892669179569011, + "learning_rate": 8.41309434407207e-07, + "loss": 0.389, + "step": 2367 + }, + { + "epoch": 4.707753479125248, + "grad_norm": 0.09566520976823947, + "learning_rate": 8.300181476306179e-07, + "loss": 0.3919, + "step": 2368 + }, + { + "epoch": 4.709741550695825, + "grad_norm": 0.08459337487960555, + "learning_rate": 8.188023499002206e-07, + "loss": 0.3838, + "step": 2369 + }, + { + "epoch": 4.711729622266402, + "grad_norm": 0.08665109130706619, + "learning_rate": 8.076620628312892e-07, + "loss": 0.3822, + "step": 2370 + }, + { + "epoch": 4.713717693836978, + "grad_norm": 0.09024375966314252, + "learning_rate": 7.965973078935696e-07, + "loss": 0.3813, + "step": 2371 + }, + { + "epoch": 4.715705765407555, + "grad_norm": 0.09285780165532485, + "learning_rate": 7.856081064112441e-07, + "loss": 0.3912, + "step": 2372 + }, + { + "epoch": 4.717693836978131, + "grad_norm": 0.0943616237445835, + "learning_rate": 7.746944795628874e-07, + "loss": 0.3862, + "step": 2373 + }, + { + "epoch": 4.719681908548708, + "grad_norm": 0.08785063174101017, + "learning_rate": 7.638564483814215e-07, + "loss": 0.3947, + "step": 2374 + }, + { + "epoch": 4.721669980119284, + "grad_norm": 0.09095838243764175, + "learning_rate": 7.530940337540848e-07, + "loss": 0.3871, + "step": 2375 + }, + { + "epoch": 4.723658051689861, + "grad_norm": 0.09266236841123172, + "learning_rate": 7.424072564223795e-07, + "loss": 0.3792, + "step": 2376 + }, + { + "epoch": 4.725646123260438, + "grad_norm": 0.08856421790524949, + "learning_rate": 7.317961369820481e-07, + "loss": 0.3984, + "step": 2377 + }, + { + "epoch": 4.727634194831014, + "grad_norm": 0.09046046436239621, + "learning_rate": 7.212606958830126e-07, + "loss": 0.388, + "step": 2378 + }, + { + "epoch": 4.729622266401591, + "grad_norm": 0.08810710235320644, + "learning_rate": 7.108009534293514e-07, + "loss": 0.3926, + "step": 2379 + }, + { + "epoch": 4.7316103379721675, + "grad_norm": 0.08776085645147964, + "learning_rate": 7.00416929779264e-07, + "loss": 0.3835, + "step": 2380 + }, + { + "epoch": 4.7335984095427435, + "grad_norm": 0.09018333553170645, + "learning_rate": 6.901086449450045e-07, + "loss": 0.3926, + "step": 2381 + }, + { + "epoch": 4.73558648111332, + "grad_norm": 0.08976241821965435, + "learning_rate": 6.798761187928815e-07, + "loss": 0.3824, + "step": 2382 + }, + { + "epoch": 4.737574552683896, + "grad_norm": 0.08687984662867637, + "learning_rate": 6.697193710431827e-07, + "loss": 0.3894, + "step": 2383 + }, + { + "epoch": 4.739562624254473, + "grad_norm": 0.08513429550866758, + "learning_rate": 6.596384212701656e-07, + "loss": 0.3889, + "step": 2384 + }, + { + "epoch": 4.741550695825049, + "grad_norm": 0.09283958366235047, + "learning_rate": 6.496332889020096e-07, + "loss": 0.3917, + "step": 2385 + }, + { + "epoch": 4.743538767395626, + "grad_norm": 0.08978086123224327, + "learning_rate": 6.397039932207705e-07, + "loss": 0.3874, + "step": 2386 + }, + { + "epoch": 4.745526838966203, + "grad_norm": 0.09028630878201256, + "learning_rate": 6.298505533623544e-07, + "loss": 0.3979, + "step": 2387 + }, + { + "epoch": 4.747514910536779, + "grad_norm": 0.08511974572102049, + "learning_rate": 6.200729883164736e-07, + "loss": 0.3868, + "step": 2388 + }, + { + "epoch": 4.749502982107356, + "grad_norm": 0.09014999836299889, + "learning_rate": 6.10371316926619e-07, + "loss": 0.3832, + "step": 2389 + }, + { + "epoch": 4.751491053677933, + "grad_norm": 0.08562371944420073, + "learning_rate": 6.007455578900079e-07, + "loss": 0.3835, + "step": 2390 + }, + { + "epoch": 4.753479125248509, + "grad_norm": 0.08437078505374138, + "learning_rate": 5.911957297575743e-07, + "loss": 0.3995, + "step": 2391 + }, + { + "epoch": 4.755467196819086, + "grad_norm": 0.08509023606898676, + "learning_rate": 5.817218509339029e-07, + "loss": 0.382, + "step": 2392 + }, + { + "epoch": 4.757455268389662, + "grad_norm": 0.0862939742400007, + "learning_rate": 5.723239396772107e-07, + "loss": 0.3751, + "step": 2393 + }, + { + "epoch": 4.759443339960239, + "grad_norm": 0.08427683497565448, + "learning_rate": 5.630020140993076e-07, + "loss": 0.3906, + "step": 2394 + }, + { + "epoch": 4.761431411530815, + "grad_norm": 0.08659925642974421, + "learning_rate": 5.537560921655694e-07, + "loss": 0.3739, + "step": 2395 + }, + { + "epoch": 4.763419483101392, + "grad_norm": 0.08684889514283894, + "learning_rate": 5.445861916948891e-07, + "loss": 0.3883, + "step": 2396 + }, + { + "epoch": 4.7654075546719685, + "grad_norm": 0.08586227981413348, + "learning_rate": 5.354923303596593e-07, + "loss": 0.3882, + "step": 2397 + }, + { + "epoch": 4.7673956262425445, + "grad_norm": 0.08933801071830935, + "learning_rate": 5.26474525685714e-07, + "loss": 0.3921, + "step": 2398 + }, + { + "epoch": 4.769383697813121, + "grad_norm": 0.08393331580601653, + "learning_rate": 5.175327950523201e-07, + "loss": 0.3823, + "step": 2399 + }, + { + "epoch": 4.7713717693836974, + "grad_norm": 0.08312263829480795, + "learning_rate": 5.086671556921286e-07, + "loss": 0.3914, + "step": 2400 + }, + { + "epoch": 4.773359840954274, + "grad_norm": 0.08583116775976779, + "learning_rate": 4.998776246911563e-07, + "loss": 0.3879, + "step": 2401 + }, + { + "epoch": 4.775347912524851, + "grad_norm": 0.08780350297111986, + "learning_rate": 4.911642189887289e-07, + "loss": 0.3822, + "step": 2402 + }, + { + "epoch": 4.777335984095427, + "grad_norm": 0.08554313989570071, + "learning_rate": 4.82526955377467e-07, + "loss": 0.3767, + "step": 2403 + }, + { + "epoch": 4.779324055666004, + "grad_norm": 0.08765985124847109, + "learning_rate": 4.7396585050325074e-07, + "loss": 0.3847, + "step": 2404 + }, + { + "epoch": 4.78131212723658, + "grad_norm": 0.08780793267342793, + "learning_rate": 4.654809208651889e-07, + "loss": 0.3844, + "step": 2405 + }, + { + "epoch": 4.783300198807157, + "grad_norm": 0.08604341225517562, + "learning_rate": 4.5707218281557883e-07, + "loss": 0.393, + "step": 2406 + }, + { + "epoch": 4.785288270377734, + "grad_norm": 0.08308372988525528, + "learning_rate": 4.487396525598797e-07, + "loss": 0.3913, + "step": 2407 + }, + { + "epoch": 4.78727634194831, + "grad_norm": 0.08542634110771802, + "learning_rate": 4.4048334615668156e-07, + "loss": 0.4001, + "step": 2408 + }, + { + "epoch": 4.789264413518887, + "grad_norm": 0.08706213681011674, + "learning_rate": 4.323032795176785e-07, + "loss": 0.382, + "step": 2409 + }, + { + "epoch": 4.791252485089463, + "grad_norm": 0.0859838895047276, + "learning_rate": 4.241994684076378e-07, + "loss": 0.3961, + "step": 2410 + }, + { + "epoch": 4.79324055666004, + "grad_norm": 0.08596617254705141, + "learning_rate": 4.1617192844435105e-07, + "loss": 0.3814, + "step": 2411 + }, + { + "epoch": 4.795228628230617, + "grad_norm": 0.09187239171244084, + "learning_rate": 4.0822067509863397e-07, + "loss": 0.3964, + "step": 2412 + }, + { + "epoch": 4.797216699801193, + "grad_norm": 0.09151958557517886, + "learning_rate": 4.0034572369427315e-07, + "loss": 0.3825, + "step": 2413 + }, + { + "epoch": 4.79920477137177, + "grad_norm": 0.08886486817433736, + "learning_rate": 3.9254708940800855e-07, + "loss": 0.3905, + "step": 2414 + }, + { + "epoch": 4.801192842942346, + "grad_norm": 0.08799982211453036, + "learning_rate": 3.848247872694977e-07, + "loss": 0.3881, + "step": 2415 + }, + { + "epoch": 4.8031809145129225, + "grad_norm": 0.0902921148546499, + "learning_rate": 3.7717883216128904e-07, + "loss": 0.3855, + "step": 2416 + }, + { + "epoch": 4.805168986083499, + "grad_norm": 0.08950875599619121, + "learning_rate": 3.6960923881879994e-07, + "loss": 0.3997, + "step": 2417 + }, + { + "epoch": 4.807157057654075, + "grad_norm": 0.08405639164843014, + "learning_rate": 3.6211602183027216e-07, + "loss": 0.3827, + "step": 2418 + }, + { + "epoch": 4.809145129224652, + "grad_norm": 0.08385315321218918, + "learning_rate": 3.5469919563676735e-07, + "loss": 0.3853, + "step": 2419 + }, + { + "epoch": 4.811133200795228, + "grad_norm": 0.08505559287027027, + "learning_rate": 3.4735877453210944e-07, + "loss": 0.3839, + "step": 2420 + }, + { + "epoch": 4.813121272365805, + "grad_norm": 0.0877458325572006, + "learning_rate": 3.400947726628845e-07, + "loss": 0.3889, + "step": 2421 + }, + { + "epoch": 4.815109343936381, + "grad_norm": 0.08379392354212882, + "learning_rate": 3.329072040284009e-07, + "loss": 0.3813, + "step": 2422 + }, + { + "epoch": 4.817097415506958, + "grad_norm": 0.08663727574036406, + "learning_rate": 3.257960824806583e-07, + "loss": 0.3724, + "step": 2423 + }, + { + "epoch": 4.819085487077535, + "grad_norm": 0.08972288136722377, + "learning_rate": 3.187614217243296e-07, + "loss": 0.3904, + "step": 2424 + }, + { + "epoch": 4.821073558648111, + "grad_norm": 0.08914702376767991, + "learning_rate": 3.1180323531673e-07, + "loss": 0.3902, + "step": 2425 + }, + { + "epoch": 4.823061630218688, + "grad_norm": 0.08226333159404381, + "learning_rate": 3.049215366677949e-07, + "loss": 0.3919, + "step": 2426 + }, + { + "epoch": 4.825049701789265, + "grad_norm": 0.08530573163134422, + "learning_rate": 2.9811633904004876e-07, + "loss": 0.3813, + "step": 2427 + }, + { + "epoch": 4.827037773359841, + "grad_norm": 0.08624288085838655, + "learning_rate": 2.913876555485873e-07, + "loss": 0.3863, + "step": 2428 + }, + { + "epoch": 4.829025844930418, + "grad_norm": 0.08814860299045851, + "learning_rate": 2.8473549916103295e-07, + "loss": 0.3923, + "step": 2429 + }, + { + "epoch": 4.831013916500994, + "grad_norm": 0.0834424995172425, + "learning_rate": 2.7815988269753957e-07, + "loss": 0.3816, + "step": 2430 + }, + { + "epoch": 4.833001988071571, + "grad_norm": 0.0875951945007766, + "learning_rate": 2.7166081883074793e-07, + "loss": 0.3831, + "step": 2431 + }, + { + "epoch": 4.834990059642147, + "grad_norm": 0.08573228313795739, + "learning_rate": 2.6523832008575447e-07, + "loss": 0.3797, + "step": 2432 + }, + { + "epoch": 4.836978131212724, + "grad_norm": 0.08471682476679154, + "learning_rate": 2.5889239884011594e-07, + "loss": 0.386, + "step": 2433 + }, + { + "epoch": 4.8389662027833005, + "grad_norm": 0.08533182226317534, + "learning_rate": 2.526230673237873e-07, + "loss": 0.3911, + "step": 2434 + }, + { + "epoch": 4.8409542743538765, + "grad_norm": 0.08729608897177847, + "learning_rate": 2.46430337619139e-07, + "loss": 0.3933, + "step": 2435 + }, + { + "epoch": 4.842942345924453, + "grad_norm": 0.08464771353563329, + "learning_rate": 2.4031422166090447e-07, + "loss": 0.3877, + "step": 2436 + }, + { + "epoch": 4.84493041749503, + "grad_norm": 0.08529987456863733, + "learning_rate": 2.3427473123615263e-07, + "loss": 0.3872, + "step": 2437 + }, + { + "epoch": 4.846918489065606, + "grad_norm": 0.08251198483736942, + "learning_rate": 2.2831187798430633e-07, + "loss": 0.3851, + "step": 2438 + }, + { + "epoch": 4.848906560636183, + "grad_norm": 0.08697926770346306, + "learning_rate": 2.2242567339706645e-07, + "loss": 0.3891, + "step": 2439 + }, + { + "epoch": 4.850894632206759, + "grad_norm": 0.08389709541673776, + "learning_rate": 2.1661612881842542e-07, + "loss": 0.3974, + "step": 2440 + }, + { + "epoch": 4.852882703777336, + "grad_norm": 0.08176299785081795, + "learning_rate": 2.108832554446405e-07, + "loss": 0.3865, + "step": 2441 + }, + { + "epoch": 4.854870775347912, + "grad_norm": 0.0859453841790713, + "learning_rate": 2.0522706432419382e-07, + "loss": 0.3894, + "step": 2442 + }, + { + "epoch": 4.856858846918489, + "grad_norm": 0.09171848286661113, + "learning_rate": 1.9964756635780125e-07, + "loss": 0.388, + "step": 2443 + }, + { + "epoch": 4.858846918489066, + "grad_norm": 0.08974957511489433, + "learning_rate": 1.9414477229835472e-07, + "loss": 0.3798, + "step": 2444 + }, + { + "epoch": 4.860834990059642, + "grad_norm": 0.08912058067697651, + "learning_rate": 1.887186927509399e-07, + "loss": 0.3844, + "step": 2445 + }, + { + "epoch": 4.862823061630219, + "grad_norm": 0.08776737055560599, + "learning_rate": 1.8336933817278746e-07, + "loss": 0.3827, + "step": 2446 + }, + { + "epoch": 4.864811133200796, + "grad_norm": 0.08734208552490895, + "learning_rate": 1.780967188732552e-07, + "loss": 0.3988, + "step": 2447 + }, + { + "epoch": 4.866799204771372, + "grad_norm": 0.08473457345009698, + "learning_rate": 1.7290084501383254e-07, + "loss": 0.39, + "step": 2448 + }, + { + "epoch": 4.868787276341949, + "grad_norm": 0.08742974713015055, + "learning_rate": 1.6778172660809167e-07, + "loss": 0.3841, + "step": 2449 + }, + { + "epoch": 4.870775347912525, + "grad_norm": 0.08420030904113016, + "learning_rate": 1.627393735216787e-07, + "loss": 0.3871, + "step": 2450 + }, + { + "epoch": 4.8727634194831015, + "grad_norm": 0.0862672824125834, + "learning_rate": 1.5777379547230908e-07, + "loss": 0.3875, + "step": 2451 + }, + { + "epoch": 4.8747514910536776, + "grad_norm": 0.08692886838458265, + "learning_rate": 1.5288500202972346e-07, + "loss": 0.3726, + "step": 2452 + }, + { + "epoch": 4.8767395626242545, + "grad_norm": 0.08351633165920348, + "learning_rate": 1.4807300261568735e-07, + "loss": 0.3844, + "step": 2453 + }, + { + "epoch": 4.878727634194831, + "grad_norm": 0.08207322266636974, + "learning_rate": 1.4333780650396922e-07, + "loss": 0.3836, + "step": 2454 + }, + { + "epoch": 4.880715705765407, + "grad_norm": 0.08636151753971841, + "learning_rate": 1.3867942282031365e-07, + "loss": 0.3814, + "step": 2455 + }, + { + "epoch": 4.882703777335984, + "grad_norm": 0.08372475407996112, + "learning_rate": 1.3409786054244144e-07, + "loss": 0.3864, + "step": 2456 + }, + { + "epoch": 4.88469184890656, + "grad_norm": 0.08656783681493549, + "learning_rate": 1.295931285000096e-07, + "loss": 0.3936, + "step": 2457 + }, + { + "epoch": 4.886679920477137, + "grad_norm": 0.0865359135074539, + "learning_rate": 1.2516523537462023e-07, + "loss": 0.3854, + "step": 2458 + }, + { + "epoch": 4.888667992047714, + "grad_norm": 0.08370382158458198, + "learning_rate": 1.208141896997761e-07, + "loss": 0.3874, + "step": 2459 + }, + { + "epoch": 4.89065606361829, + "grad_norm": 0.08898094046346416, + "learning_rate": 1.1653999986089404e-07, + "loss": 0.3816, + "step": 2460 + }, + { + "epoch": 4.892644135188867, + "grad_norm": 0.08630171969853209, + "learning_rate": 1.1234267409525601e-07, + "loss": 0.3923, + "step": 2461 + }, + { + "epoch": 4.894632206759443, + "grad_norm": 0.0822988617920718, + "learning_rate": 1.0822222049202247e-07, + "loss": 0.3764, + "step": 2462 + }, + { + "epoch": 4.89662027833002, + "grad_norm": 0.08320129070038762, + "learning_rate": 1.0417864699220126e-07, + "loss": 0.3761, + "step": 2463 + }, + { + "epoch": 4.898608349900597, + "grad_norm": 0.0842454627956477, + "learning_rate": 1.0021196138863432e-07, + "loss": 0.3837, + "step": 2464 + }, + { + "epoch": 4.900596421471173, + "grad_norm": 0.08427608839565273, + "learning_rate": 9.632217132598431e-08, + "loss": 0.3962, + "step": 2465 + }, + { + "epoch": 4.90258449304175, + "grad_norm": 0.08425550035154374, + "learning_rate": 9.250928430072138e-08, + "loss": 0.3826, + "step": 2466 + }, + { + "epoch": 4.904572564612326, + "grad_norm": 0.0870394245800322, + "learning_rate": 8.877330766110526e-08, + "loss": 0.3899, + "step": 2467 + }, + { + "epoch": 4.906560636182903, + "grad_norm": 0.08676508699928365, + "learning_rate": 8.51142486071721e-08, + "loss": 0.3924, + "step": 2468 + }, + { + "epoch": 4.9085487077534795, + "grad_norm": 0.08646978372665866, + "learning_rate": 8.153211419072104e-08, + "loss": 0.3844, + "step": 2469 + }, + { + "epoch": 4.9105367793240555, + "grad_norm": 0.08355087037223606, + "learning_rate": 7.802691131530093e-08, + "loss": 0.3936, + "step": 2470 + }, + { + "epoch": 4.912524850894632, + "grad_norm": 0.0825099428809249, + "learning_rate": 7.459864673619698e-08, + "loss": 0.3844, + "step": 2471 + }, + { + "epoch": 4.914512922465208, + "grad_norm": 0.0874220543050281, + "learning_rate": 7.124732706042636e-08, + "loss": 0.3839, + "step": 2472 + }, + { + "epoch": 4.916500994035785, + "grad_norm": 0.08504844158831366, + "learning_rate": 6.797295874669818e-08, + "loss": 0.3775, + "step": 2473 + }, + { + "epoch": 4.918489065606362, + "grad_norm": 0.08380430885759332, + "learning_rate": 6.47755481054313e-08, + "loss": 0.3864, + "step": 2474 + }, + { + "epoch": 4.920477137176938, + "grad_norm": 0.08542485101299789, + "learning_rate": 6.165510129873653e-08, + "loss": 0.3831, + "step": 2475 + }, + { + "epoch": 4.922465208747515, + "grad_norm": 0.08407973206991785, + "learning_rate": 5.8611624340385585e-08, + "loss": 0.3918, + "step": 2476 + }, + { + "epoch": 4.924453280318091, + "grad_norm": 0.08941652662215563, + "learning_rate": 5.564512309581993e-08, + "loss": 0.3857, + "step": 2477 + }, + { + "epoch": 4.926441351888668, + "grad_norm": 0.08476705011477939, + "learning_rate": 5.2755603282128585e-08, + "loss": 0.3996, + "step": 2478 + }, + { + "epoch": 4.928429423459244, + "grad_norm": 0.08424740037805077, + "learning_rate": 4.994307046804814e-08, + "loss": 0.4005, + "step": 2479 + }, + { + "epoch": 4.930417495029821, + "grad_norm": 0.08089692105587475, + "learning_rate": 4.720753007393164e-08, + "loss": 0.3856, + "step": 2480 + }, + { + "epoch": 4.932405566600398, + "grad_norm": 0.0828129028116285, + "learning_rate": 4.454898737176194e-08, + "loss": 0.3896, + "step": 2481 + }, + { + "epoch": 4.934393638170974, + "grad_norm": 0.08581140480052221, + "learning_rate": 4.1967447485129486e-08, + "loss": 0.3877, + "step": 2482 + }, + { + "epoch": 4.936381709741551, + "grad_norm": 0.0841501559497828, + "learning_rate": 3.946291538921454e-08, + "loss": 0.3932, + "step": 2483 + }, + { + "epoch": 4.938369781312128, + "grad_norm": 0.084945831920411, + "learning_rate": 3.703539591080052e-08, + "loss": 0.3893, + "step": 2484 + }, + { + "epoch": 4.940357852882704, + "grad_norm": 0.08602665475516638, + "learning_rate": 3.468489372823847e-08, + "loss": 0.3857, + "step": 2485 + }, + { + "epoch": 4.942345924453281, + "grad_norm": 0.0843351047913454, + "learning_rate": 3.2411413371460364e-08, + "loss": 0.3944, + "step": 2486 + }, + { + "epoch": 4.944333996023857, + "grad_norm": 0.09458575400648904, + "learning_rate": 3.02149592219525e-08, + "loss": 0.371, + "step": 2487 + }, + { + "epoch": 4.9463220675944335, + "grad_norm": 0.09060041810386756, + "learning_rate": 2.8095535512759898e-08, + "loss": 0.3823, + "step": 2488 + }, + { + "epoch": 4.9483101391650095, + "grad_norm": 0.08243819150909268, + "learning_rate": 2.605314632846856e-08, + "loss": 0.3857, + "step": 2489 + }, + { + "epoch": 4.950298210735586, + "grad_norm": 0.08673906431689675, + "learning_rate": 2.4087795605209906e-08, + "loss": 0.3857, + "step": 2490 + }, + { + "epoch": 4.952286282306163, + "grad_norm": 0.18285784263478053, + "learning_rate": 2.2199487130638575e-08, + "loss": 0.3808, + "step": 2491 + }, + { + "epoch": 4.954274353876739, + "grad_norm": 0.08505699269511605, + "learning_rate": 2.038822454393241e-08, + "loss": 0.3919, + "step": 2492 + }, + { + "epoch": 4.956262425447316, + "grad_norm": 0.08105600682700584, + "learning_rate": 1.8654011335788037e-08, + "loss": 0.388, + "step": 2493 + }, + { + "epoch": 4.958250497017893, + "grad_norm": 0.08527795126915254, + "learning_rate": 1.699685084840752e-08, + "loss": 0.384, + "step": 2494 + }, + { + "epoch": 4.960238568588469, + "grad_norm": 0.08653563160420075, + "learning_rate": 1.5416746275507267e-08, + "loss": 0.3962, + "step": 2495 + }, + { + "epoch": 4.962226640159046, + "grad_norm": 0.08979740263573663, + "learning_rate": 1.3913700662282482e-08, + "loss": 0.3914, + "step": 2496 + }, + { + "epoch": 4.964214711729622, + "grad_norm": 0.08523863622899382, + "learning_rate": 1.2487716905429382e-08, + "loss": 0.3787, + "step": 2497 + }, + { + "epoch": 4.966202783300199, + "grad_norm": 0.08508629391165508, + "learning_rate": 1.1138797753131868e-08, + "loss": 0.3885, + "step": 2498 + }, + { + "epoch": 4.968190854870775, + "grad_norm": 0.08532887882999854, + "learning_rate": 9.866945805048211e-09, + "loss": 0.3769, + "step": 2499 + }, + { + "epoch": 4.970178926441352, + "grad_norm": 0.08378156608125406, + "learning_rate": 8.672163512311039e-09, + "loss": 0.3835, + "step": 2500 + }, + { + "epoch": 4.972166998011929, + "grad_norm": 0.08684442546437508, + "learning_rate": 7.554453177522902e-09, + "loss": 0.389, + "step": 2501 + }, + { + "epoch": 4.974155069582505, + "grad_norm": 0.0813736466741287, + "learning_rate": 6.513816954760721e-09, + "loss": 0.3744, + "step": 2502 + }, + { + "epoch": 4.976143141153082, + "grad_norm": 0.08714909210347142, + "learning_rate": 5.5502568495491295e-09, + "loss": 0.3766, + "step": 2503 + }, + { + "epoch": 4.9781312127236585, + "grad_norm": 0.08760221313835241, + "learning_rate": 4.663774718878245e-09, + "loss": 0.3885, + "step": 2504 + }, + { + "epoch": 4.980119284294235, + "grad_norm": 0.08699109801070153, + "learning_rate": 3.854372271194784e-09, + "loss": 0.3888, + "step": 2505 + }, + { + "epoch": 4.9821073558648115, + "grad_norm": 0.08377737770104218, + "learning_rate": 3.1220510663931836e-09, + "loss": 0.3841, + "step": 2506 + }, + { + "epoch": 4.9840954274353875, + "grad_norm": 0.0860928281202727, + "learning_rate": 2.4668125158111567e-09, + "loss": 0.3874, + "step": 2507 + }, + { + "epoch": 4.986083499005964, + "grad_norm": 0.08466195525302847, + "learning_rate": 1.8886578822385758e-09, + "loss": 0.3968, + "step": 2508 + }, + { + "epoch": 4.98807157057654, + "grad_norm": 0.08539612831774646, + "learning_rate": 1.3875882799041506e-09, + "loss": 0.389, + "step": 2509 + }, + { + "epoch": 4.990059642147117, + "grad_norm": 0.0871537888144173, + "learning_rate": 9.636046744754268e-10, + "loss": 0.3887, + "step": 2510 + }, + { + "epoch": 4.992047713717694, + "grad_norm": 0.09005357737681634, + "learning_rate": 6.167078830587869e-10, + "loss": 0.3828, + "step": 2511 + }, + { + "epoch": 4.99403578528827, + "grad_norm": 0.08342821477503674, + "learning_rate": 3.4689857420833193e-10, + "loss": 0.3765, + "step": 2512 + }, + { + "epoch": 4.996023856858847, + "grad_norm": 0.08551466138197758, + "learning_rate": 1.5417726789923594e-10, + "loss": 0.399, + "step": 2513 + }, + { + "epoch": 4.998011928429423, + "grad_norm": 0.08572188617520322, + "learning_rate": 3.8544335545509515e-11, + "loss": 0.3885, + "step": 2514 + }, + { + "epoch": 5.0, + "grad_norm": 0.08448061468786694, + "learning_rate": 0.0, + "loss": 0.3759, + "step": 2515 + }, + { + "epoch": 5.0, + "step": 2515, + "total_flos": 4.219199777931264e+16, + "train_loss": 0.48938420348565576, + "train_runtime": 39399.5763, + "train_samples_per_second": 32.661, + "train_steps_per_second": 0.064 + } + ], + "logging_steps": 1, + "max_steps": 2515, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 4.219199777931264e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}