diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,104333 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.9998993322371732, + "eval_steps": 500, + "global_step": 29800, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00013422368376900105, + "grad_norm": 42.767356872558594, + "learning_rate": 2.2371364653243848e-07, + "loss": 7.8985, + "step": 2 + }, + { + "epoch": 0.0002684473675380021, + "grad_norm": 44.26445388793945, + "learning_rate": 4.4742729306487696e-07, + "loss": 6.497, + "step": 4 + }, + { + "epoch": 0.00040267105130700314, + "grad_norm": 40.350433349609375, + "learning_rate": 6.711409395973154e-07, + "loss": 6.4155, + "step": 6 + }, + { + "epoch": 0.0005368947350760042, + "grad_norm": 40.29481887817383, + "learning_rate": 8.948545861297539e-07, + "loss": 6.4425, + "step": 8 + }, + { + "epoch": 0.0006711184188450052, + "grad_norm": 34.15610122680664, + "learning_rate": 1.1185682326621925e-06, + "loss": 6.1634, + "step": 10 + }, + { + "epoch": 0.0008053421026140063, + "grad_norm": 35.55307388305664, + "learning_rate": 1.3422818791946309e-06, + "loss": 5.9724, + "step": 12 + }, + { + "epoch": 0.0009395657863830073, + "grad_norm": 34.54287338256836, + "learning_rate": 1.5659955257270695e-06, + "loss": 5.5537, + "step": 14 + }, + { + "epoch": 0.0010737894701520084, + "grad_norm": 31.5373592376709, + "learning_rate": 1.7897091722595078e-06, + "loss": 4.9947, + "step": 16 + }, + { + "epoch": 0.0012080131539210094, + "grad_norm": 29.093385696411133, + "learning_rate": 2.013422818791946e-06, + "loss": 4.7559, + "step": 18 + }, + { + "epoch": 0.0013422368376900104, + "grad_norm": 17.960554122924805, + "learning_rate": 2.237136465324385e-06, + "loss": 3.9112, + "step": 20 + }, + { + "epoch": 0.0014764605214590114, + "grad_norm": 15.002157211303711, + "learning_rate": 2.4608501118568234e-06, + "loss": 3.6361, + "step": 22 + }, + { + "epoch": 0.0016106842052280126, + "grad_norm": 12.236384391784668, + "learning_rate": 2.6845637583892617e-06, + "loss": 3.4589, + "step": 24 + }, + { + "epoch": 0.0017449078889970136, + "grad_norm": 8.323893547058105, + "learning_rate": 2.9082774049217e-06, + "loss": 3.1198, + "step": 26 + }, + { + "epoch": 0.0018791315727660146, + "grad_norm": 7.51531982421875, + "learning_rate": 3.131991051454139e-06, + "loss": 3.0164, + "step": 28 + }, + { + "epoch": 0.0020133552565350155, + "grad_norm": 5.39280891418457, + "learning_rate": 3.3557046979865773e-06, + "loss": 2.8598, + "step": 30 + }, + { + "epoch": 0.0021475789403040168, + "grad_norm": 4.7940239906311035, + "learning_rate": 3.5794183445190157e-06, + "loss": 2.7762, + "step": 32 + }, + { + "epoch": 0.0022818026240730175, + "grad_norm": 4.117138385772705, + "learning_rate": 3.803131991051454e-06, + "loss": 2.6936, + "step": 34 + }, + { + "epoch": 0.0024160263078420187, + "grad_norm": 3.4565351009368896, + "learning_rate": 4.026845637583892e-06, + "loss": 2.5653, + "step": 36 + }, + { + "epoch": 0.00255024999161102, + "grad_norm": 5.539157390594482, + "learning_rate": 4.250559284116332e-06, + "loss": 2.592, + "step": 38 + }, + { + "epoch": 0.0026844736753800207, + "grad_norm": 3.1694529056549072, + "learning_rate": 4.47427293064877e-06, + "loss": 2.5212, + "step": 40 + }, + { + "epoch": 0.002818697359149022, + "grad_norm": 4.043001651763916, + "learning_rate": 4.697986577181209e-06, + "loss": 2.4646, + "step": 42 + }, + { + "epoch": 0.0029529210429180227, + "grad_norm": 3.426333427429199, + "learning_rate": 4.921700223713647e-06, + "loss": 2.4251, + "step": 44 + }, + { + "epoch": 0.003087144726687024, + "grad_norm": 2.9743473529815674, + "learning_rate": 5.1454138702460855e-06, + "loss": 2.3507, + "step": 46 + }, + { + "epoch": 0.003221368410456025, + "grad_norm": 2.7875661849975586, + "learning_rate": 5.3691275167785235e-06, + "loss": 2.3166, + "step": 48 + }, + { + "epoch": 0.003355592094225026, + "grad_norm": 5.016522407531738, + "learning_rate": 5.592841163310962e-06, + "loss": 2.2963, + "step": 50 + }, + { + "epoch": 0.003489815777994027, + "grad_norm": 3.7552225589752197, + "learning_rate": 5.8165548098434e-06, + "loss": 2.2321, + "step": 52 + }, + { + "epoch": 0.003624039461763028, + "grad_norm": 2.941863775253296, + "learning_rate": 6.04026845637584e-06, + "loss": 2.2089, + "step": 54 + }, + { + "epoch": 0.003758263145532029, + "grad_norm": 2.9160988330841064, + "learning_rate": 6.263982102908278e-06, + "loss": 2.1362, + "step": 56 + }, + { + "epoch": 0.0038924868293010303, + "grad_norm": 2.726508378982544, + "learning_rate": 6.487695749440716e-06, + "loss": 2.149, + "step": 58 + }, + { + "epoch": 0.004026710513070031, + "grad_norm": 2.86901593208313, + "learning_rate": 6.7114093959731546e-06, + "loss": 2.0731, + "step": 60 + }, + { + "epoch": 0.004160934196839032, + "grad_norm": 2.479349374771118, + "learning_rate": 6.935123042505594e-06, + "loss": 2.0715, + "step": 62 + }, + { + "epoch": 0.0042951578806080335, + "grad_norm": 3.0036346912384033, + "learning_rate": 7.158836689038031e-06, + "loss": 2.0446, + "step": 64 + }, + { + "epoch": 0.004429381564377034, + "grad_norm": 2.9378981590270996, + "learning_rate": 7.382550335570471e-06, + "loss": 1.9705, + "step": 66 + }, + { + "epoch": 0.004563605248146035, + "grad_norm": 2.7089834213256836, + "learning_rate": 7.606263982102908e-06, + "loss": 1.9511, + "step": 68 + }, + { + "epoch": 0.004697828931915037, + "grad_norm": 2.9577481746673584, + "learning_rate": 7.829977628635348e-06, + "loss": 1.916, + "step": 70 + }, + { + "epoch": 0.0048320526156840375, + "grad_norm": 3.2720839977264404, + "learning_rate": 8.053691275167785e-06, + "loss": 1.838, + "step": 72 + }, + { + "epoch": 0.004966276299453038, + "grad_norm": 3.753499984741211, + "learning_rate": 8.277404921700224e-06, + "loss": 1.7891, + "step": 74 + }, + { + "epoch": 0.00510049998322204, + "grad_norm": 3.103024482727051, + "learning_rate": 8.501118568232664e-06, + "loss": 1.7397, + "step": 76 + }, + { + "epoch": 0.005234723666991041, + "grad_norm": 3.117305040359497, + "learning_rate": 8.724832214765101e-06, + "loss": 1.7192, + "step": 78 + }, + { + "epoch": 0.0053689473507600415, + "grad_norm": 3.049574375152588, + "learning_rate": 8.94854586129754e-06, + "loss": 1.649, + "step": 80 + }, + { + "epoch": 0.005503171034529042, + "grad_norm": 2.904085159301758, + "learning_rate": 9.172259507829977e-06, + "loss": 1.5617, + "step": 82 + }, + { + "epoch": 0.005637394718298044, + "grad_norm": 2.859712839126587, + "learning_rate": 9.395973154362418e-06, + "loss": 1.5642, + "step": 84 + }, + { + "epoch": 0.005771618402067045, + "grad_norm": 2.6117401123046875, + "learning_rate": 9.619686800894855e-06, + "loss": 1.529, + "step": 86 + }, + { + "epoch": 0.005905842085836045, + "grad_norm": 2.8871121406555176, + "learning_rate": 9.843400447427293e-06, + "loss": 1.4863, + "step": 88 + }, + { + "epoch": 0.006040065769605047, + "grad_norm": 2.6381306648254395, + "learning_rate": 1.006711409395973e-05, + "loss": 1.4858, + "step": 90 + }, + { + "epoch": 0.006174289453374048, + "grad_norm": 2.521726131439209, + "learning_rate": 1.0290827740492171e-05, + "loss": 1.443, + "step": 92 + }, + { + "epoch": 0.006308513137143049, + "grad_norm": 2.3711681365966797, + "learning_rate": 1.051454138702461e-05, + "loss": 1.4161, + "step": 94 + }, + { + "epoch": 0.00644273682091205, + "grad_norm": 2.5755703449249268, + "learning_rate": 1.0738255033557047e-05, + "loss": 1.461, + "step": 96 + }, + { + "epoch": 0.006576960504681051, + "grad_norm": 2.566465139389038, + "learning_rate": 1.0961968680089486e-05, + "loss": 1.3901, + "step": 98 + }, + { + "epoch": 0.006711184188450052, + "grad_norm": 2.506310224533081, + "learning_rate": 1.1185682326621925e-05, + "loss": 1.3596, + "step": 100 + }, + { + "epoch": 0.0068454078722190535, + "grad_norm": 2.2585983276367188, + "learning_rate": 1.1409395973154363e-05, + "loss": 1.3593, + "step": 102 + }, + { + "epoch": 0.006979631555988054, + "grad_norm": 2.6175777912139893, + "learning_rate": 1.16331096196868e-05, + "loss": 1.3012, + "step": 104 + }, + { + "epoch": 0.007113855239757055, + "grad_norm": 2.2929115295410156, + "learning_rate": 1.185682326621924e-05, + "loss": 1.3368, + "step": 106 + }, + { + "epoch": 0.007248078923526056, + "grad_norm": 3.431950092315674, + "learning_rate": 1.208053691275168e-05, + "loss": 1.2798, + "step": 108 + }, + { + "epoch": 0.007382302607295057, + "grad_norm": 2.7987656593322754, + "learning_rate": 1.2304250559284117e-05, + "loss": 1.2695, + "step": 110 + }, + { + "epoch": 0.007516526291064058, + "grad_norm": 2.331099033355713, + "learning_rate": 1.2527964205816556e-05, + "loss": 1.2599, + "step": 112 + }, + { + "epoch": 0.007650749974833059, + "grad_norm": 2.243494749069214, + "learning_rate": 1.2751677852348994e-05, + "loss": 1.3625, + "step": 114 + }, + { + "epoch": 0.007784973658602061, + "grad_norm": 2.421570301055908, + "learning_rate": 1.2975391498881432e-05, + "loss": 1.3264, + "step": 116 + }, + { + "epoch": 0.007919197342371061, + "grad_norm": 2.063924789428711, + "learning_rate": 1.319910514541387e-05, + "loss": 1.2795, + "step": 118 + }, + { + "epoch": 0.008053421026140062, + "grad_norm": 2.33018159866333, + "learning_rate": 1.3422818791946309e-05, + "loss": 1.2582, + "step": 120 + }, + { + "epoch": 0.008187644709909063, + "grad_norm": 2.2538163661956787, + "learning_rate": 1.3646532438478748e-05, + "loss": 1.2508, + "step": 122 + }, + { + "epoch": 0.008321868393678064, + "grad_norm": 2.2358736991882324, + "learning_rate": 1.3870246085011188e-05, + "loss": 1.1971, + "step": 124 + }, + { + "epoch": 0.008456092077447066, + "grad_norm": 2.3848955631256104, + "learning_rate": 1.4093959731543624e-05, + "loss": 1.2728, + "step": 126 + }, + { + "epoch": 0.008590315761216067, + "grad_norm": 2.260687828063965, + "learning_rate": 1.4317673378076063e-05, + "loss": 1.2465, + "step": 128 + }, + { + "epoch": 0.008724539444985068, + "grad_norm": 2.33004093170166, + "learning_rate": 1.4541387024608501e-05, + "loss": 1.1897, + "step": 130 + }, + { + "epoch": 0.008858763128754069, + "grad_norm": 2.219352960586548, + "learning_rate": 1.4765100671140942e-05, + "loss": 1.2068, + "step": 132 + }, + { + "epoch": 0.00899298681252307, + "grad_norm": 2.1695876121520996, + "learning_rate": 1.4988814317673377e-05, + "loss": 1.1861, + "step": 134 + }, + { + "epoch": 0.00912721049629207, + "grad_norm": 2.1830992698669434, + "learning_rate": 1.5212527964205816e-05, + "loss": 1.2202, + "step": 136 + }, + { + "epoch": 0.009261434180061073, + "grad_norm": 2.1508290767669678, + "learning_rate": 1.5436241610738255e-05, + "loss": 1.152, + "step": 138 + }, + { + "epoch": 0.009395657863830073, + "grad_norm": 2.224714517593384, + "learning_rate": 1.5659955257270695e-05, + "loss": 1.174, + "step": 140 + }, + { + "epoch": 0.009529881547599074, + "grad_norm": 2.093791961669922, + "learning_rate": 1.5883668903803133e-05, + "loss": 1.1588, + "step": 142 + }, + { + "epoch": 0.009664105231368075, + "grad_norm": 1.9894508123397827, + "learning_rate": 1.610738255033557e-05, + "loss": 1.1574, + "step": 144 + }, + { + "epoch": 0.009798328915137076, + "grad_norm": 2.047668933868408, + "learning_rate": 1.633109619686801e-05, + "loss": 1.1913, + "step": 146 + }, + { + "epoch": 0.009932552598906077, + "grad_norm": 2.0811944007873535, + "learning_rate": 1.6554809843400447e-05, + "loss": 1.1582, + "step": 148 + }, + { + "epoch": 0.010066776282675077, + "grad_norm": 2.1576690673828125, + "learning_rate": 1.6778523489932888e-05, + "loss": 1.1453, + "step": 150 + }, + { + "epoch": 0.01020099996644408, + "grad_norm": 2.1650378704071045, + "learning_rate": 1.7002237136465328e-05, + "loss": 1.1698, + "step": 152 + }, + { + "epoch": 0.01033522365021308, + "grad_norm": 2.089860439300537, + "learning_rate": 1.7225950782997762e-05, + "loss": 1.111, + "step": 154 + }, + { + "epoch": 0.010469447333982081, + "grad_norm": 2.0436043739318848, + "learning_rate": 1.7449664429530202e-05, + "loss": 1.1292, + "step": 156 + }, + { + "epoch": 0.010603671017751082, + "grad_norm": 2.8769757747650146, + "learning_rate": 1.767337807606264e-05, + "loss": 1.1203, + "step": 158 + }, + { + "epoch": 0.010737894701520083, + "grad_norm": 2.236624240875244, + "learning_rate": 1.789709172259508e-05, + "loss": 1.1019, + "step": 160 + }, + { + "epoch": 0.010872118385289084, + "grad_norm": 2.022063970565796, + "learning_rate": 1.8120805369127517e-05, + "loss": 1.1133, + "step": 162 + }, + { + "epoch": 0.011006342069058084, + "grad_norm": 1.8289309740066528, + "learning_rate": 1.8344519015659954e-05, + "loss": 1.1115, + "step": 164 + }, + { + "epoch": 0.011140565752827087, + "grad_norm": 2.115225076675415, + "learning_rate": 1.8568232662192395e-05, + "loss": 1.1549, + "step": 166 + }, + { + "epoch": 0.011274789436596088, + "grad_norm": 2.106482982635498, + "learning_rate": 1.8791946308724835e-05, + "loss": 1.1644, + "step": 168 + }, + { + "epoch": 0.011409013120365089, + "grad_norm": 2.041001319885254, + "learning_rate": 1.9015659955257272e-05, + "loss": 1.0633, + "step": 170 + }, + { + "epoch": 0.01154323680413409, + "grad_norm": 2.135566473007202, + "learning_rate": 1.923937360178971e-05, + "loss": 1.141, + "step": 172 + }, + { + "epoch": 0.01167746048790309, + "grad_norm": 2.074963092803955, + "learning_rate": 1.946308724832215e-05, + "loss": 1.1373, + "step": 174 + }, + { + "epoch": 0.01181168417167209, + "grad_norm": 5.725609302520752, + "learning_rate": 1.9686800894854587e-05, + "loss": 1.033, + "step": 176 + }, + { + "epoch": 0.011945907855441093, + "grad_norm": 2.2156283855438232, + "learning_rate": 1.9910514541387027e-05, + "loss": 1.0946, + "step": 178 + }, + { + "epoch": 0.012080131539210094, + "grad_norm": 1.979102373123169, + "learning_rate": 2.013422818791946e-05, + "loss": 1.1088, + "step": 180 + }, + { + "epoch": 0.012214355222979095, + "grad_norm": 1.9270720481872559, + "learning_rate": 2.03579418344519e-05, + "loss": 1.116, + "step": 182 + }, + { + "epoch": 0.012348578906748096, + "grad_norm": 2.0316152572631836, + "learning_rate": 2.0581655480984342e-05, + "loss": 1.0833, + "step": 184 + }, + { + "epoch": 0.012482802590517096, + "grad_norm": 2.2409253120422363, + "learning_rate": 2.080536912751678e-05, + "loss": 1.1118, + "step": 186 + }, + { + "epoch": 0.012617026274286097, + "grad_norm": 2.080007791519165, + "learning_rate": 2.102908277404922e-05, + "loss": 1.1066, + "step": 188 + }, + { + "epoch": 0.012751249958055098, + "grad_norm": 1.9611796140670776, + "learning_rate": 2.1252796420581657e-05, + "loss": 1.1154, + "step": 190 + }, + { + "epoch": 0.0128854736418241, + "grad_norm": 1.939950704574585, + "learning_rate": 2.1476510067114094e-05, + "loss": 1.0574, + "step": 192 + }, + { + "epoch": 0.013019697325593101, + "grad_norm": 2.051105499267578, + "learning_rate": 2.1700223713646534e-05, + "loss": 1.1122, + "step": 194 + }, + { + "epoch": 0.013153921009362102, + "grad_norm": 2.430405855178833, + "learning_rate": 2.192393736017897e-05, + "loss": 1.1435, + "step": 196 + }, + { + "epoch": 0.013288144693131103, + "grad_norm": 2.2696692943573, + "learning_rate": 2.2147651006711412e-05, + "loss": 1.126, + "step": 198 + }, + { + "epoch": 0.013422368376900104, + "grad_norm": 1.884081482887268, + "learning_rate": 2.237136465324385e-05, + "loss": 1.0558, + "step": 200 + }, + { + "epoch": 0.013556592060669104, + "grad_norm": 1.9265978336334229, + "learning_rate": 2.2595078299776286e-05, + "loss": 1.1367, + "step": 202 + }, + { + "epoch": 0.013690815744438107, + "grad_norm": 2.098081111907959, + "learning_rate": 2.2818791946308727e-05, + "loss": 1.0358, + "step": 204 + }, + { + "epoch": 0.013825039428207108, + "grad_norm": 2.078112840652466, + "learning_rate": 2.3042505592841164e-05, + "loss": 1.0662, + "step": 206 + }, + { + "epoch": 0.013959263111976108, + "grad_norm": 1.9842445850372314, + "learning_rate": 2.32662192393736e-05, + "loss": 1.0274, + "step": 208 + }, + { + "epoch": 0.01409348679574511, + "grad_norm": 1.8199636936187744, + "learning_rate": 2.348993288590604e-05, + "loss": 0.9903, + "step": 210 + }, + { + "epoch": 0.01422771047951411, + "grad_norm": 1.8890475034713745, + "learning_rate": 2.371364653243848e-05, + "loss": 1.0526, + "step": 212 + }, + { + "epoch": 0.01436193416328311, + "grad_norm": 1.9008369445800781, + "learning_rate": 2.393736017897092e-05, + "loss": 1.0305, + "step": 214 + }, + { + "epoch": 0.014496157847052112, + "grad_norm": 1.9247174263000488, + "learning_rate": 2.416107382550336e-05, + "loss": 1.0382, + "step": 216 + }, + { + "epoch": 0.014630381530821114, + "grad_norm": 1.8395044803619385, + "learning_rate": 2.4384787472035793e-05, + "loss": 0.9981, + "step": 218 + }, + { + "epoch": 0.014764605214590115, + "grad_norm": 1.9873464107513428, + "learning_rate": 2.4608501118568234e-05, + "loss": 1.0409, + "step": 220 + }, + { + "epoch": 0.014898828898359116, + "grad_norm": 2.246591806411743, + "learning_rate": 2.4832214765100674e-05, + "loss": 1.0352, + "step": 222 + }, + { + "epoch": 0.015033052582128116, + "grad_norm": 1.8422189950942993, + "learning_rate": 2.505592841163311e-05, + "loss": 1.016, + "step": 224 + }, + { + "epoch": 0.015167276265897117, + "grad_norm": 2.0310373306274414, + "learning_rate": 2.527964205816555e-05, + "loss": 1.0845, + "step": 226 + }, + { + "epoch": 0.015301499949666118, + "grad_norm": 1.8543283939361572, + "learning_rate": 2.550335570469799e-05, + "loss": 1.077, + "step": 228 + }, + { + "epoch": 0.01543572363343512, + "grad_norm": 1.8772192001342773, + "learning_rate": 2.5727069351230426e-05, + "loss": 1.0509, + "step": 230 + }, + { + "epoch": 0.015569947317204121, + "grad_norm": 1.918837547302246, + "learning_rate": 2.5950782997762863e-05, + "loss": 1.0099, + "step": 232 + }, + { + "epoch": 0.01570417100097312, + "grad_norm": 2.334092140197754, + "learning_rate": 2.6174496644295304e-05, + "loss": 1.0434, + "step": 234 + }, + { + "epoch": 0.015838394684742123, + "grad_norm": 2.450557231903076, + "learning_rate": 2.639821029082774e-05, + "loss": 1.0655, + "step": 236 + }, + { + "epoch": 0.015972618368511125, + "grad_norm": 1.8649179935455322, + "learning_rate": 2.662192393736018e-05, + "loss": 1.0242, + "step": 238 + }, + { + "epoch": 0.016106842052280124, + "grad_norm": 1.8493071794509888, + "learning_rate": 2.6845637583892618e-05, + "loss": 1.0186, + "step": 240 + }, + { + "epoch": 0.016241065736049127, + "grad_norm": 1.9221071004867554, + "learning_rate": 2.7069351230425055e-05, + "loss": 1.0608, + "step": 242 + }, + { + "epoch": 0.016375289419818126, + "grad_norm": 1.8620386123657227, + "learning_rate": 2.7293064876957496e-05, + "loss": 0.9906, + "step": 244 + }, + { + "epoch": 0.01650951310358713, + "grad_norm": 2.163280963897705, + "learning_rate": 2.7516778523489933e-05, + "loss": 0.9914, + "step": 246 + }, + { + "epoch": 0.016643736787356127, + "grad_norm": 9.815702438354492, + "learning_rate": 2.7740492170022377e-05, + "loss": 1.0617, + "step": 248 + }, + { + "epoch": 0.01677796047112513, + "grad_norm": 1.879356026649475, + "learning_rate": 2.796420581655481e-05, + "loss": 0.9977, + "step": 250 + }, + { + "epoch": 0.016912184154894132, + "grad_norm": 1.9691245555877686, + "learning_rate": 2.8187919463087248e-05, + "loss": 1.0689, + "step": 252 + }, + { + "epoch": 0.01704640783866313, + "grad_norm": 1.7776098251342773, + "learning_rate": 2.8411633109619688e-05, + "loss": 0.9844, + "step": 254 + }, + { + "epoch": 0.017180631522432134, + "grad_norm": 2.298013925552368, + "learning_rate": 2.8635346756152125e-05, + "loss": 1.0274, + "step": 256 + }, + { + "epoch": 0.017314855206201133, + "grad_norm": 1.880511999130249, + "learning_rate": 2.885906040268457e-05, + "loss": 1.0339, + "step": 258 + }, + { + "epoch": 0.017449078889970136, + "grad_norm": 2.170077323913574, + "learning_rate": 2.9082774049217003e-05, + "loss": 0.9712, + "step": 260 + }, + { + "epoch": 0.017583302573739135, + "grad_norm": 1.9864907264709473, + "learning_rate": 2.930648769574944e-05, + "loss": 1.0496, + "step": 262 + }, + { + "epoch": 0.017717526257508137, + "grad_norm": 1.7236459255218506, + "learning_rate": 2.9530201342281884e-05, + "loss": 0.9276, + "step": 264 + }, + { + "epoch": 0.01785174994127714, + "grad_norm": 1.8995026350021362, + "learning_rate": 2.9753914988814318e-05, + "loss": 0.9976, + "step": 266 + }, + { + "epoch": 0.01798597362504614, + "grad_norm": 1.8944072723388672, + "learning_rate": 2.9977628635346755e-05, + "loss": 0.9927, + "step": 268 + }, + { + "epoch": 0.01812019730881514, + "grad_norm": 1.9850423336029053, + "learning_rate": 3.02013422818792e-05, + "loss": 1.0039, + "step": 270 + }, + { + "epoch": 0.01825442099258414, + "grad_norm": 2.0685665607452393, + "learning_rate": 3.0425055928411632e-05, + "loss": 1.0208, + "step": 272 + }, + { + "epoch": 0.018388644676353143, + "grad_norm": 1.7906346321105957, + "learning_rate": 3.0648769574944076e-05, + "loss": 1.0011, + "step": 274 + }, + { + "epoch": 0.018522868360122145, + "grad_norm": 2.530409097671509, + "learning_rate": 3.087248322147651e-05, + "loss": 1.0667, + "step": 276 + }, + { + "epoch": 0.018657092043891144, + "grad_norm": 1.7903863191604614, + "learning_rate": 3.109619686800895e-05, + "loss": 0.9743, + "step": 278 + }, + { + "epoch": 0.018791315727660147, + "grad_norm": 2.157287120819092, + "learning_rate": 3.131991051454139e-05, + "loss": 1.036, + "step": 280 + }, + { + "epoch": 0.018925539411429146, + "grad_norm": 1.8101394176483154, + "learning_rate": 3.1543624161073825e-05, + "loss": 1.0131, + "step": 282 + }, + { + "epoch": 0.01905976309519815, + "grad_norm": 1.8443541526794434, + "learning_rate": 3.1767337807606265e-05, + "loss": 0.9968, + "step": 284 + }, + { + "epoch": 0.019193986778967147, + "grad_norm": 3.132554531097412, + "learning_rate": 3.1991051454138706e-05, + "loss": 0.9758, + "step": 286 + }, + { + "epoch": 0.01932821046273615, + "grad_norm": 1.8236656188964844, + "learning_rate": 3.221476510067114e-05, + "loss": 0.9747, + "step": 288 + }, + { + "epoch": 0.019462434146505152, + "grad_norm": 2.611768960952759, + "learning_rate": 3.243847874720358e-05, + "loss": 0.9804, + "step": 290 + }, + { + "epoch": 0.01959665783027415, + "grad_norm": 19.7513427734375, + "learning_rate": 3.266219239373602e-05, + "loss": 1.0925, + "step": 292 + }, + { + "epoch": 0.019730881514043154, + "grad_norm": 2.0081746578216553, + "learning_rate": 3.288590604026846e-05, + "loss": 1.0048, + "step": 294 + }, + { + "epoch": 0.019865105197812153, + "grad_norm": 2.01193904876709, + "learning_rate": 3.3109619686800894e-05, + "loss": 0.9301, + "step": 296 + }, + { + "epoch": 0.019999328881581156, + "grad_norm": 1.8901691436767578, + "learning_rate": 3.3333333333333335e-05, + "loss": 0.9385, + "step": 298 + }, + { + "epoch": 0.020133552565350155, + "grad_norm": 2.4309654235839844, + "learning_rate": 3.3557046979865775e-05, + "loss": 1.0148, + "step": 300 + }, + { + "epoch": 0.020267776249119157, + "grad_norm": 1.8135578632354736, + "learning_rate": 3.378076062639821e-05, + "loss": 0.9746, + "step": 302 + }, + { + "epoch": 0.02040199993288816, + "grad_norm": 2.1611545085906982, + "learning_rate": 3.4004474272930656e-05, + "loss": 0.9685, + "step": 304 + }, + { + "epoch": 0.02053622361665716, + "grad_norm": 1.7547329664230347, + "learning_rate": 3.422818791946309e-05, + "loss": 1.0273, + "step": 306 + }, + { + "epoch": 0.02067044730042616, + "grad_norm": 1.6789021492004395, + "learning_rate": 3.4451901565995524e-05, + "loss": 0.962, + "step": 308 + }, + { + "epoch": 0.02080467098419516, + "grad_norm": 1.7096121311187744, + "learning_rate": 3.4675615212527964e-05, + "loss": 1.0389, + "step": 310 + }, + { + "epoch": 0.020938894667964163, + "grad_norm": 2.264564037322998, + "learning_rate": 3.4899328859060405e-05, + "loss": 1.0188, + "step": 312 + }, + { + "epoch": 0.021073118351733162, + "grad_norm": 2.4550116062164307, + "learning_rate": 3.512304250559284e-05, + "loss": 0.9969, + "step": 314 + }, + { + "epoch": 0.021207342035502164, + "grad_norm": 1.9647586345672607, + "learning_rate": 3.534675615212528e-05, + "loss": 0.934, + "step": 316 + }, + { + "epoch": 0.021341565719271167, + "grad_norm": 1.7615418434143066, + "learning_rate": 3.557046979865772e-05, + "loss": 0.9689, + "step": 318 + }, + { + "epoch": 0.021475789403040166, + "grad_norm": 1.8934327363967896, + "learning_rate": 3.579418344519016e-05, + "loss": 1.0117, + "step": 320 + }, + { + "epoch": 0.02161001308680917, + "grad_norm": 1.7119038105010986, + "learning_rate": 3.6017897091722594e-05, + "loss": 0.9061, + "step": 322 + }, + { + "epoch": 0.021744236770578167, + "grad_norm": 1.767134666442871, + "learning_rate": 3.6241610738255034e-05, + "loss": 0.9391, + "step": 324 + }, + { + "epoch": 0.02187846045434717, + "grad_norm": 2.058410406112671, + "learning_rate": 3.6465324384787475e-05, + "loss": 0.9918, + "step": 326 + }, + { + "epoch": 0.02201268413811617, + "grad_norm": 1.6547894477844238, + "learning_rate": 3.668903803131991e-05, + "loss": 1.0361, + "step": 328 + }, + { + "epoch": 0.02214690782188517, + "grad_norm": 1.8306688070297241, + "learning_rate": 3.6912751677852356e-05, + "loss": 0.9505, + "step": 330 + }, + { + "epoch": 0.022281131505654174, + "grad_norm": 1.8601231575012207, + "learning_rate": 3.713646532438479e-05, + "loss": 0.9661, + "step": 332 + }, + { + "epoch": 0.022415355189423173, + "grad_norm": 1.877942442893982, + "learning_rate": 3.736017897091722e-05, + "loss": 0.9401, + "step": 334 + }, + { + "epoch": 0.022549578873192175, + "grad_norm": 1.8389990329742432, + "learning_rate": 3.758389261744967e-05, + "loss": 0.9276, + "step": 336 + }, + { + "epoch": 0.022683802556961175, + "grad_norm": 1.6361632347106934, + "learning_rate": 3.7807606263982104e-05, + "loss": 0.9313, + "step": 338 + }, + { + "epoch": 0.022818026240730177, + "grad_norm": 1.736832857131958, + "learning_rate": 3.8031319910514545e-05, + "loss": 1.0363, + "step": 340 + }, + { + "epoch": 0.02295224992449918, + "grad_norm": 1.808683156967163, + "learning_rate": 3.8255033557046985e-05, + "loss": 0.9628, + "step": 342 + }, + { + "epoch": 0.02308647360826818, + "grad_norm": 1.7691764831542969, + "learning_rate": 3.847874720357942e-05, + "loss": 0.9395, + "step": 344 + }, + { + "epoch": 0.02322069729203718, + "grad_norm": 1.6702338457107544, + "learning_rate": 3.870246085011186e-05, + "loss": 0.9588, + "step": 346 + }, + { + "epoch": 0.02335492097580618, + "grad_norm": 1.7142435312271118, + "learning_rate": 3.89261744966443e-05, + "loss": 0.903, + "step": 348 + }, + { + "epoch": 0.023489144659575183, + "grad_norm": 1.6802270412445068, + "learning_rate": 3.914988814317674e-05, + "loss": 0.9571, + "step": 350 + }, + { + "epoch": 0.02362336834334418, + "grad_norm": 1.9458545446395874, + "learning_rate": 3.9373601789709174e-05, + "loss": 0.9587, + "step": 352 + }, + { + "epoch": 0.023757592027113184, + "grad_norm": 1.8980857133865356, + "learning_rate": 3.959731543624161e-05, + "loss": 0.9452, + "step": 354 + }, + { + "epoch": 0.023891815710882187, + "grad_norm": 1.8086082935333252, + "learning_rate": 3.9821029082774055e-05, + "loss": 0.8951, + "step": 356 + }, + { + "epoch": 0.024026039394651186, + "grad_norm": 1.7780733108520508, + "learning_rate": 4.004474272930649e-05, + "loss": 0.8955, + "step": 358 + }, + { + "epoch": 0.02416026307842019, + "grad_norm": 2.0118095874786377, + "learning_rate": 4.026845637583892e-05, + "loss": 0.9621, + "step": 360 + }, + { + "epoch": 0.024294486762189187, + "grad_norm": 1.6194727420806885, + "learning_rate": 4.049217002237137e-05, + "loss": 0.8891, + "step": 362 + }, + { + "epoch": 0.02442871044595819, + "grad_norm": 1.8351916074752808, + "learning_rate": 4.07158836689038e-05, + "loss": 0.9886, + "step": 364 + }, + { + "epoch": 0.02456293412972719, + "grad_norm": 1.8265271186828613, + "learning_rate": 4.0939597315436244e-05, + "loss": 1.0141, + "step": 366 + }, + { + "epoch": 0.02469715781349619, + "grad_norm": 1.84307861328125, + "learning_rate": 4.1163310961968684e-05, + "loss": 0.987, + "step": 368 + }, + { + "epoch": 0.024831381497265194, + "grad_norm": 2.075024127960205, + "learning_rate": 4.138702460850112e-05, + "loss": 0.9706, + "step": 370 + }, + { + "epoch": 0.024965605181034193, + "grad_norm": 1.7007570266723633, + "learning_rate": 4.161073825503356e-05, + "loss": 0.9534, + "step": 372 + }, + { + "epoch": 0.025099828864803195, + "grad_norm": 1.9438354969024658, + "learning_rate": 4.1834451901566e-05, + "loss": 0.9997, + "step": 374 + }, + { + "epoch": 0.025234052548572194, + "grad_norm": 1.5149259567260742, + "learning_rate": 4.205816554809844e-05, + "loss": 0.9208, + "step": 376 + }, + { + "epoch": 0.025368276232341197, + "grad_norm": 1.6345700025558472, + "learning_rate": 4.228187919463087e-05, + "loss": 0.9345, + "step": 378 + }, + { + "epoch": 0.025502499916110196, + "grad_norm": 2.3215291500091553, + "learning_rate": 4.2505592841163314e-05, + "loss": 0.9775, + "step": 380 + }, + { + "epoch": 0.0256367235998792, + "grad_norm": 2.234036684036255, + "learning_rate": 4.2729306487695754e-05, + "loss": 0.9709, + "step": 382 + }, + { + "epoch": 0.0257709472836482, + "grad_norm": 1.8222061395645142, + "learning_rate": 4.295302013422819e-05, + "loss": 0.9838, + "step": 384 + }, + { + "epoch": 0.0259051709674172, + "grad_norm": 1.5880529880523682, + "learning_rate": 4.317673378076063e-05, + "loss": 0.941, + "step": 386 + }, + { + "epoch": 0.026039394651186203, + "grad_norm": 1.7884408235549927, + "learning_rate": 4.340044742729307e-05, + "loss": 1.025, + "step": 388 + }, + { + "epoch": 0.0261736183349552, + "grad_norm": 1.5966376066207886, + "learning_rate": 4.36241610738255e-05, + "loss": 0.9362, + "step": 390 + }, + { + "epoch": 0.026307842018724204, + "grad_norm": 1.804855227470398, + "learning_rate": 4.384787472035794e-05, + "loss": 0.9968, + "step": 392 + }, + { + "epoch": 0.026442065702493203, + "grad_norm": 1.6474556922912598, + "learning_rate": 4.4071588366890384e-05, + "loss": 0.9353, + "step": 394 + }, + { + "epoch": 0.026576289386262206, + "grad_norm": 2.1431021690368652, + "learning_rate": 4.4295302013422824e-05, + "loss": 0.9527, + "step": 396 + }, + { + "epoch": 0.026710513070031208, + "grad_norm": 2.439883232116699, + "learning_rate": 4.451901565995526e-05, + "loss": 0.9578, + "step": 398 + }, + { + "epoch": 0.026844736753800207, + "grad_norm": 2.5266048908233643, + "learning_rate": 4.47427293064877e-05, + "loss": 0.9169, + "step": 400 + }, + { + "epoch": 0.02697896043756921, + "grad_norm": 7.857470512390137, + "learning_rate": 4.496644295302014e-05, + "loss": 0.9875, + "step": 402 + }, + { + "epoch": 0.02711318412133821, + "grad_norm": 1.719964861869812, + "learning_rate": 4.519015659955257e-05, + "loss": 0.9324, + "step": 404 + }, + { + "epoch": 0.02724740780510721, + "grad_norm": 1.7738561630249023, + "learning_rate": 4.541387024608501e-05, + "loss": 0.9648, + "step": 406 + }, + { + "epoch": 0.027381631488876214, + "grad_norm": 4.7296953201293945, + "learning_rate": 4.5637583892617453e-05, + "loss": 0.9465, + "step": 408 + }, + { + "epoch": 0.027515855172645213, + "grad_norm": 1.8566882610321045, + "learning_rate": 4.586129753914989e-05, + "loss": 0.9163, + "step": 410 + }, + { + "epoch": 0.027650078856414215, + "grad_norm": 1.6954864263534546, + "learning_rate": 4.608501118568233e-05, + "loss": 0.9287, + "step": 412 + }, + { + "epoch": 0.027784302540183214, + "grad_norm": 1.734711766242981, + "learning_rate": 4.630872483221477e-05, + "loss": 0.9313, + "step": 414 + }, + { + "epoch": 0.027918526223952217, + "grad_norm": 1.7496058940887451, + "learning_rate": 4.65324384787472e-05, + "loss": 0.926, + "step": 416 + }, + { + "epoch": 0.028052749907721216, + "grad_norm": 1.6309984922409058, + "learning_rate": 4.675615212527964e-05, + "loss": 0.8993, + "step": 418 + }, + { + "epoch": 0.02818697359149022, + "grad_norm": 3.022975444793701, + "learning_rate": 4.697986577181208e-05, + "loss": 0.8894, + "step": 420 + }, + { + "epoch": 0.02832119727525922, + "grad_norm": 1.860723614692688, + "learning_rate": 4.720357941834452e-05, + "loss": 0.9366, + "step": 422 + }, + { + "epoch": 0.02845542095902822, + "grad_norm": 1.670973777770996, + "learning_rate": 4.742729306487696e-05, + "loss": 0.9438, + "step": 424 + }, + { + "epoch": 0.028589644642797223, + "grad_norm": 1.8910622596740723, + "learning_rate": 4.76510067114094e-05, + "loss": 0.9366, + "step": 426 + }, + { + "epoch": 0.02872386832656622, + "grad_norm": 1.6946468353271484, + "learning_rate": 4.787472035794184e-05, + "loss": 0.9455, + "step": 428 + }, + { + "epoch": 0.028858092010335224, + "grad_norm": 1.9580165147781372, + "learning_rate": 4.809843400447427e-05, + "loss": 0.9467, + "step": 430 + }, + { + "epoch": 0.028992315694104223, + "grad_norm": 1.765116810798645, + "learning_rate": 4.832214765100672e-05, + "loss": 0.9342, + "step": 432 + }, + { + "epoch": 0.029126539377873226, + "grad_norm": 2.1234090328216553, + "learning_rate": 4.854586129753915e-05, + "loss": 0.9607, + "step": 434 + }, + { + "epoch": 0.029260763061642228, + "grad_norm": 2.8784382343292236, + "learning_rate": 4.8769574944071586e-05, + "loss": 0.9569, + "step": 436 + }, + { + "epoch": 0.029394986745411227, + "grad_norm": 2.9103879928588867, + "learning_rate": 4.8993288590604034e-05, + "loss": 0.8813, + "step": 438 + }, + { + "epoch": 0.02952921042918023, + "grad_norm": 1.9417873620986938, + "learning_rate": 4.921700223713647e-05, + "loss": 0.9726, + "step": 440 + }, + { + "epoch": 0.02966343411294923, + "grad_norm": 1.5041873455047607, + "learning_rate": 4.944071588366891e-05, + "loss": 0.9298, + "step": 442 + }, + { + "epoch": 0.02979765779671823, + "grad_norm": 2.0654478073120117, + "learning_rate": 4.966442953020135e-05, + "loss": 1.0488, + "step": 444 + }, + { + "epoch": 0.02993188148048723, + "grad_norm": 1.5653998851776123, + "learning_rate": 4.988814317673378e-05, + "loss": 0.9493, + "step": 446 + }, + { + "epoch": 0.030066105164256233, + "grad_norm": 2.1645994186401367, + "learning_rate": 5.011185682326622e-05, + "loss": 0.9166, + "step": 448 + }, + { + "epoch": 0.030200328848025235, + "grad_norm": 1.866364598274231, + "learning_rate": 5.033557046979866e-05, + "loss": 0.9697, + "step": 450 + }, + { + "epoch": 0.030334552531794234, + "grad_norm": 1.411466360092163, + "learning_rate": 5.05592841163311e-05, + "loss": 0.92, + "step": 452 + }, + { + "epoch": 0.030468776215563237, + "grad_norm": 1.7987719774246216, + "learning_rate": 5.078299776286354e-05, + "loss": 0.995, + "step": 454 + }, + { + "epoch": 0.030602999899332236, + "grad_norm": 2.065213680267334, + "learning_rate": 5.100671140939598e-05, + "loss": 0.9905, + "step": 456 + }, + { + "epoch": 0.03073722358310124, + "grad_norm": 3.715909004211426, + "learning_rate": 5.123042505592841e-05, + "loss": 0.9419, + "step": 458 + }, + { + "epoch": 0.03087144726687024, + "grad_norm": 1.5822199583053589, + "learning_rate": 5.145413870246085e-05, + "loss": 0.9306, + "step": 460 + }, + { + "epoch": 0.03100567095063924, + "grad_norm": 1.9820096492767334, + "learning_rate": 5.167785234899329e-05, + "loss": 1.0166, + "step": 462 + }, + { + "epoch": 0.031139894634408243, + "grad_norm": 7.59510612487793, + "learning_rate": 5.1901565995525726e-05, + "loss": 0.9553, + "step": 464 + }, + { + "epoch": 0.031274118318177245, + "grad_norm": 1.6523329019546509, + "learning_rate": 5.212527964205817e-05, + "loss": 0.9754, + "step": 466 + }, + { + "epoch": 0.03140834200194624, + "grad_norm": 1.549682378768921, + "learning_rate": 5.234899328859061e-05, + "loss": 0.8897, + "step": 468 + }, + { + "epoch": 0.03154256568571524, + "grad_norm": 2.2443296909332275, + "learning_rate": 5.257270693512305e-05, + "loss": 0.9679, + "step": 470 + }, + { + "epoch": 0.031676789369484246, + "grad_norm": 3.2198832035064697, + "learning_rate": 5.279642058165548e-05, + "loss": 0.9503, + "step": 472 + }, + { + "epoch": 0.03181101305325325, + "grad_norm": 1.7302112579345703, + "learning_rate": 5.302013422818792e-05, + "loss": 0.9569, + "step": 474 + }, + { + "epoch": 0.03194523673702225, + "grad_norm": 1.6648476123809814, + "learning_rate": 5.324384787472036e-05, + "loss": 0.9226, + "step": 476 + }, + { + "epoch": 0.032079460420791246, + "grad_norm": 1.5022692680358887, + "learning_rate": 5.3467561521252796e-05, + "loss": 0.8694, + "step": 478 + }, + { + "epoch": 0.03221368410456025, + "grad_norm": 1.6610078811645508, + "learning_rate": 5.3691275167785237e-05, + "loss": 0.9539, + "step": 480 + }, + { + "epoch": 0.03234790778832925, + "grad_norm": 1.7546595335006714, + "learning_rate": 5.391498881431768e-05, + "loss": 0.8589, + "step": 482 + }, + { + "epoch": 0.032482131472098254, + "grad_norm": 3.204794406890869, + "learning_rate": 5.413870246085011e-05, + "loss": 0.885, + "step": 484 + }, + { + "epoch": 0.03261635515586725, + "grad_norm": 1.8887150287628174, + "learning_rate": 5.436241610738255e-05, + "loss": 0.9423, + "step": 486 + }, + { + "epoch": 0.03275057883963625, + "grad_norm": 1.6688555479049683, + "learning_rate": 5.458612975391499e-05, + "loss": 0.9337, + "step": 488 + }, + { + "epoch": 0.032884802523405254, + "grad_norm": 1.4868736267089844, + "learning_rate": 5.4809843400447426e-05, + "loss": 0.8962, + "step": 490 + }, + { + "epoch": 0.03301902620717426, + "grad_norm": 2.3861632347106934, + "learning_rate": 5.5033557046979866e-05, + "loss": 0.9746, + "step": 492 + }, + { + "epoch": 0.03315324989094326, + "grad_norm": 1.8033841848373413, + "learning_rate": 5.5257270693512306e-05, + "loss": 0.9252, + "step": 494 + }, + { + "epoch": 0.033287473574712255, + "grad_norm": 1.5896399021148682, + "learning_rate": 5.5480984340044754e-05, + "loss": 0.8963, + "step": 496 + }, + { + "epoch": 0.03342169725848126, + "grad_norm": 1.4402724504470825, + "learning_rate": 5.570469798657718e-05, + "loss": 0.9249, + "step": 498 + }, + { + "epoch": 0.03355592094225026, + "grad_norm": 1.7836982011795044, + "learning_rate": 5.592841163310962e-05, + "loss": 0.8972, + "step": 500 + }, + { + "epoch": 0.03369014462601926, + "grad_norm": 1.5181845426559448, + "learning_rate": 5.615212527964206e-05, + "loss": 0.8425, + "step": 502 + }, + { + "epoch": 0.033824368309788265, + "grad_norm": 1.7089219093322754, + "learning_rate": 5.6375838926174495e-05, + "loss": 0.9004, + "step": 504 + }, + { + "epoch": 0.03395859199355726, + "grad_norm": 2.010953903198242, + "learning_rate": 5.6599552572706936e-05, + "loss": 0.9736, + "step": 506 + }, + { + "epoch": 0.03409281567732626, + "grad_norm": 2.0464272499084473, + "learning_rate": 5.6823266219239376e-05, + "loss": 0.9807, + "step": 508 + }, + { + "epoch": 0.034227039361095266, + "grad_norm": 1.6039035320281982, + "learning_rate": 5.704697986577181e-05, + "loss": 0.8955, + "step": 510 + }, + { + "epoch": 0.03436126304486427, + "grad_norm": 5.811178207397461, + "learning_rate": 5.727069351230425e-05, + "loss": 0.9033, + "step": 512 + }, + { + "epoch": 0.03449548672863327, + "grad_norm": 3.845050096511841, + "learning_rate": 5.749440715883669e-05, + "loss": 0.9547, + "step": 514 + }, + { + "epoch": 0.034629710412402266, + "grad_norm": 1.4917224645614624, + "learning_rate": 5.771812080536914e-05, + "loss": 0.9389, + "step": 516 + }, + { + "epoch": 0.03476393409617127, + "grad_norm": 1.537062644958496, + "learning_rate": 5.7941834451901565e-05, + "loss": 0.8869, + "step": 518 + }, + { + "epoch": 0.03489815777994027, + "grad_norm": 1.544792890548706, + "learning_rate": 5.8165548098434006e-05, + "loss": 0.9185, + "step": 520 + }, + { + "epoch": 0.035032381463709274, + "grad_norm": 1.5967966318130493, + "learning_rate": 5.838926174496645e-05, + "loss": 0.9002, + "step": 522 + }, + { + "epoch": 0.03516660514747827, + "grad_norm": 1.4114470481872559, + "learning_rate": 5.861297539149888e-05, + "loss": 0.8574, + "step": 524 + }, + { + "epoch": 0.03530082883124727, + "grad_norm": 1.5922441482543945, + "learning_rate": 5.883668903803132e-05, + "loss": 0.9034, + "step": 526 + }, + { + "epoch": 0.035435052515016274, + "grad_norm": 2.062743663787842, + "learning_rate": 5.906040268456377e-05, + "loss": 0.9114, + "step": 528 + }, + { + "epoch": 0.03556927619878528, + "grad_norm": 1.4497499465942383, + "learning_rate": 5.9284116331096195e-05, + "loss": 0.8768, + "step": 530 + }, + { + "epoch": 0.03570349988255428, + "grad_norm": 2.300522565841675, + "learning_rate": 5.9507829977628635e-05, + "loss": 0.9472, + "step": 532 + }, + { + "epoch": 0.035837723566323275, + "grad_norm": 2.232165575027466, + "learning_rate": 5.973154362416108e-05, + "loss": 0.8641, + "step": 534 + }, + { + "epoch": 0.03597194725009228, + "grad_norm": 4.513120174407959, + "learning_rate": 5.995525727069351e-05, + "loss": 0.8452, + "step": 536 + }, + { + "epoch": 0.03610617093386128, + "grad_norm": 1.7872861623764038, + "learning_rate": 6.017897091722595e-05, + "loss": 1.0023, + "step": 538 + }, + { + "epoch": 0.03624039461763028, + "grad_norm": 3.58272647857666, + "learning_rate": 6.04026845637584e-05, + "loss": 0.9371, + "step": 540 + }, + { + "epoch": 0.036374618301399285, + "grad_norm": 1.8548665046691895, + "learning_rate": 6.062639821029084e-05, + "loss": 0.9036, + "step": 542 + }, + { + "epoch": 0.03650884198516828, + "grad_norm": 1.9858945608139038, + "learning_rate": 6.0850111856823265e-05, + "loss": 0.9069, + "step": 544 + }, + { + "epoch": 0.03664306566893728, + "grad_norm": 1.9066975116729736, + "learning_rate": 6.107382550335571e-05, + "loss": 0.9916, + "step": 546 + }, + { + "epoch": 0.036777289352706286, + "grad_norm": 1.6789088249206543, + "learning_rate": 6.129753914988815e-05, + "loss": 0.9138, + "step": 548 + }, + { + "epoch": 0.03691151303647529, + "grad_norm": 2.7989633083343506, + "learning_rate": 6.152125279642058e-05, + "loss": 0.895, + "step": 550 + }, + { + "epoch": 0.03704573672024429, + "grad_norm": 2.9937233924865723, + "learning_rate": 6.174496644295302e-05, + "loss": 0.9045, + "step": 552 + }, + { + "epoch": 0.037179960404013286, + "grad_norm": 2.912078380584717, + "learning_rate": 6.196868008948546e-05, + "loss": 0.9123, + "step": 554 + }, + { + "epoch": 0.03731418408778229, + "grad_norm": 2.75984525680542, + "learning_rate": 6.21923937360179e-05, + "loss": 0.9155, + "step": 556 + }, + { + "epoch": 0.03744840777155129, + "grad_norm": 2.853611707687378, + "learning_rate": 6.241610738255034e-05, + "loss": 0.9572, + "step": 558 + }, + { + "epoch": 0.037582631455320294, + "grad_norm": 2.731768846511841, + "learning_rate": 6.263982102908278e-05, + "loss": 0.8789, + "step": 560 + }, + { + "epoch": 0.03771685513908929, + "grad_norm": 1.396867036819458, + "learning_rate": 6.286353467561522e-05, + "loss": 0.9149, + "step": 562 + }, + { + "epoch": 0.03785107882285829, + "grad_norm": 3.020078659057617, + "learning_rate": 6.308724832214765e-05, + "loss": 0.9413, + "step": 564 + }, + { + "epoch": 0.037985302506627294, + "grad_norm": 9.90017318725586, + "learning_rate": 6.331096196868009e-05, + "loss": 0.8593, + "step": 566 + }, + { + "epoch": 0.0381195261903963, + "grad_norm": 3.2886898517608643, + "learning_rate": 6.353467561521253e-05, + "loss": 0.9688, + "step": 568 + }, + { + "epoch": 0.0382537498741653, + "grad_norm": 2.8568029403686523, + "learning_rate": 6.375838926174497e-05, + "loss": 0.8769, + "step": 570 + }, + { + "epoch": 0.038387973557934295, + "grad_norm": 1.5326567888259888, + "learning_rate": 6.398210290827741e-05, + "loss": 0.8189, + "step": 572 + }, + { + "epoch": 0.0385221972417033, + "grad_norm": 9.17284107208252, + "learning_rate": 6.420581655480985e-05, + "loss": 0.9056, + "step": 574 + }, + { + "epoch": 0.0386564209254723, + "grad_norm": 1.7160284519195557, + "learning_rate": 6.442953020134228e-05, + "loss": 0.8903, + "step": 576 + }, + { + "epoch": 0.0387906446092413, + "grad_norm": 1.760678768157959, + "learning_rate": 6.465324384787472e-05, + "loss": 0.9099, + "step": 578 + }, + { + "epoch": 0.038924868293010305, + "grad_norm": 3.0596044063568115, + "learning_rate": 6.487695749440716e-05, + "loss": 0.8456, + "step": 580 + }, + { + "epoch": 0.0390590919767793, + "grad_norm": 1.4947329759597778, + "learning_rate": 6.51006711409396e-05, + "loss": 0.9065, + "step": 582 + }, + { + "epoch": 0.0391933156605483, + "grad_norm": 2.0749616622924805, + "learning_rate": 6.532438478747204e-05, + "loss": 0.9234, + "step": 584 + }, + { + "epoch": 0.039327539344317305, + "grad_norm": 1.5716087818145752, + "learning_rate": 6.554809843400448e-05, + "loss": 0.9152, + "step": 586 + }, + { + "epoch": 0.03946176302808631, + "grad_norm": 2.856884717941284, + "learning_rate": 6.577181208053692e-05, + "loss": 0.8517, + "step": 588 + }, + { + "epoch": 0.039595986711855304, + "grad_norm": 1.5679426193237305, + "learning_rate": 6.599552572706935e-05, + "loss": 0.8785, + "step": 590 + }, + { + "epoch": 0.039730210395624306, + "grad_norm": 1.7654825448989868, + "learning_rate": 6.621923937360179e-05, + "loss": 0.8845, + "step": 592 + }, + { + "epoch": 0.03986443407939331, + "grad_norm": 2.7273428440093994, + "learning_rate": 6.644295302013423e-05, + "loss": 0.9262, + "step": 594 + }, + { + "epoch": 0.03999865776316231, + "grad_norm": 1.99534273147583, + "learning_rate": 6.666666666666667e-05, + "loss": 0.8871, + "step": 596 + }, + { + "epoch": 0.040132881446931314, + "grad_norm": 2.444655656814575, + "learning_rate": 6.689038031319911e-05, + "loss": 0.9251, + "step": 598 + }, + { + "epoch": 0.04026710513070031, + "grad_norm": 3.578488826751709, + "learning_rate": 6.711409395973155e-05, + "loss": 0.8834, + "step": 600 + }, + { + "epoch": 0.04040132881446931, + "grad_norm": 6.7625837326049805, + "learning_rate": 6.733780760626398e-05, + "loss": 0.9143, + "step": 602 + }, + { + "epoch": 0.040535552498238314, + "grad_norm": 1.28508722782135, + "learning_rate": 6.756152125279642e-05, + "loss": 0.8798, + "step": 604 + }, + { + "epoch": 0.04066977618200732, + "grad_norm": 1.5711547136306763, + "learning_rate": 6.778523489932886e-05, + "loss": 0.9331, + "step": 606 + }, + { + "epoch": 0.04080399986577632, + "grad_norm": 1.9061214923858643, + "learning_rate": 6.800894854586131e-05, + "loss": 0.918, + "step": 608 + }, + { + "epoch": 0.040938223549545315, + "grad_norm": 1.4584358930587769, + "learning_rate": 6.823266219239374e-05, + "loss": 0.8318, + "step": 610 + }, + { + "epoch": 0.04107244723331432, + "grad_norm": 1.6867270469665527, + "learning_rate": 6.845637583892618e-05, + "loss": 0.9338, + "step": 612 + }, + { + "epoch": 0.04120667091708332, + "grad_norm": 3.235154151916504, + "learning_rate": 6.868008948545862e-05, + "loss": 0.9091, + "step": 614 + }, + { + "epoch": 0.04134089460085232, + "grad_norm": 1.7500040531158447, + "learning_rate": 6.890380313199105e-05, + "loss": 0.9071, + "step": 616 + }, + { + "epoch": 0.041475118284621325, + "grad_norm": 1.5308209657669067, + "learning_rate": 6.912751677852349e-05, + "loss": 0.8818, + "step": 618 + }, + { + "epoch": 0.04160934196839032, + "grad_norm": 1.5142631530761719, + "learning_rate": 6.935123042505593e-05, + "loss": 0.98, + "step": 620 + }, + { + "epoch": 0.04174356565215932, + "grad_norm": 1.6299635171890259, + "learning_rate": 6.957494407158837e-05, + "loss": 0.8267, + "step": 622 + }, + { + "epoch": 0.041877789335928325, + "grad_norm": 1.7747793197631836, + "learning_rate": 6.979865771812081e-05, + "loss": 0.8581, + "step": 624 + }, + { + "epoch": 0.04201201301969733, + "grad_norm": 1.6121944189071655, + "learning_rate": 7.002237136465325e-05, + "loss": 0.8964, + "step": 626 + }, + { + "epoch": 0.042146236703466324, + "grad_norm": 1.6563115119934082, + "learning_rate": 7.024608501118568e-05, + "loss": 0.9141, + "step": 628 + }, + { + "epoch": 0.042280460387235326, + "grad_norm": 1.6622352600097656, + "learning_rate": 7.046979865771812e-05, + "loss": 0.8935, + "step": 630 + }, + { + "epoch": 0.04241468407100433, + "grad_norm": 1.409108281135559, + "learning_rate": 7.069351230425056e-05, + "loss": 0.9248, + "step": 632 + }, + { + "epoch": 0.04254890775477333, + "grad_norm": 1.6550589799880981, + "learning_rate": 7.091722595078301e-05, + "loss": 0.997, + "step": 634 + }, + { + "epoch": 0.042683131438542334, + "grad_norm": 1.4970465898513794, + "learning_rate": 7.114093959731544e-05, + "loss": 0.844, + "step": 636 + }, + { + "epoch": 0.04281735512231133, + "grad_norm": 2.5428895950317383, + "learning_rate": 7.136465324384788e-05, + "loss": 0.8903, + "step": 638 + }, + { + "epoch": 0.04295157880608033, + "grad_norm": 1.9868848323822021, + "learning_rate": 7.158836689038032e-05, + "loss": 0.8593, + "step": 640 + }, + { + "epoch": 0.043085802489849334, + "grad_norm": 1.4770828485488892, + "learning_rate": 7.181208053691275e-05, + "loss": 0.8941, + "step": 642 + }, + { + "epoch": 0.04322002617361834, + "grad_norm": 1.4597129821777344, + "learning_rate": 7.203579418344519e-05, + "loss": 0.8786, + "step": 644 + }, + { + "epoch": 0.04335424985738734, + "grad_norm": 2.5897531509399414, + "learning_rate": 7.225950782997764e-05, + "loss": 0.9089, + "step": 646 + }, + { + "epoch": 0.043488473541156335, + "grad_norm": 3.4969544410705566, + "learning_rate": 7.248322147651007e-05, + "loss": 0.8873, + "step": 648 + }, + { + "epoch": 0.04362269722492534, + "grad_norm": 2.3789241313934326, + "learning_rate": 7.270693512304251e-05, + "loss": 0.8342, + "step": 650 + }, + { + "epoch": 0.04375692090869434, + "grad_norm": 2.3509631156921387, + "learning_rate": 7.293064876957495e-05, + "loss": 0.9211, + "step": 652 + }, + { + "epoch": 0.04389114459246334, + "grad_norm": 2.083207845687866, + "learning_rate": 7.315436241610739e-05, + "loss": 0.9198, + "step": 654 + }, + { + "epoch": 0.04402536827623234, + "grad_norm": 3.4255850315093994, + "learning_rate": 7.337807606263982e-05, + "loss": 0.9134, + "step": 656 + }, + { + "epoch": 0.04415959196000134, + "grad_norm": 3.6358864307403564, + "learning_rate": 7.360178970917227e-05, + "loss": 0.8847, + "step": 658 + }, + { + "epoch": 0.04429381564377034, + "grad_norm": 2.201115131378174, + "learning_rate": 7.382550335570471e-05, + "loss": 0.9123, + "step": 660 + }, + { + "epoch": 0.044428039327539345, + "grad_norm": 1.5794076919555664, + "learning_rate": 7.404921700223714e-05, + "loss": 0.8586, + "step": 662 + }, + { + "epoch": 0.04456226301130835, + "grad_norm": 4.102825164794922, + "learning_rate": 7.427293064876958e-05, + "loss": 0.9089, + "step": 664 + }, + { + "epoch": 0.04469648669507734, + "grad_norm": 1.4240411520004272, + "learning_rate": 7.449664429530202e-05, + "loss": 0.87, + "step": 666 + }, + { + "epoch": 0.044830710378846346, + "grad_norm": 1.6875033378601074, + "learning_rate": 7.472035794183445e-05, + "loss": 0.8644, + "step": 668 + }, + { + "epoch": 0.04496493406261535, + "grad_norm": 1.5658398866653442, + "learning_rate": 7.494407158836689e-05, + "loss": 0.9216, + "step": 670 + }, + { + "epoch": 0.04509915774638435, + "grad_norm": 1.396859884262085, + "learning_rate": 7.516778523489934e-05, + "loss": 0.8531, + "step": 672 + }, + { + "epoch": 0.045233381430153353, + "grad_norm": 1.5729717016220093, + "learning_rate": 7.539149888143177e-05, + "loss": 0.9405, + "step": 674 + }, + { + "epoch": 0.04536760511392235, + "grad_norm": 2.3833537101745605, + "learning_rate": 7.561521252796421e-05, + "loss": 0.9004, + "step": 676 + }, + { + "epoch": 0.04550182879769135, + "grad_norm": 2.166736364364624, + "learning_rate": 7.583892617449665e-05, + "loss": 0.9107, + "step": 678 + }, + { + "epoch": 0.045636052481460354, + "grad_norm": 1.4044255018234253, + "learning_rate": 7.606263982102909e-05, + "loss": 0.8947, + "step": 680 + }, + { + "epoch": 0.04577027616522936, + "grad_norm": 1.9272996187210083, + "learning_rate": 7.628635346756152e-05, + "loss": 0.9089, + "step": 682 + }, + { + "epoch": 0.04590449984899836, + "grad_norm": 3.3325276374816895, + "learning_rate": 7.651006711409397e-05, + "loss": 0.8839, + "step": 684 + }, + { + "epoch": 0.046038723532767355, + "grad_norm": 2.3696067333221436, + "learning_rate": 7.673378076062641e-05, + "loss": 0.8734, + "step": 686 + }, + { + "epoch": 0.04617294721653636, + "grad_norm": 1.3236571550369263, + "learning_rate": 7.695749440715884e-05, + "loss": 0.8133, + "step": 688 + }, + { + "epoch": 0.04630717090030536, + "grad_norm": 1.7845690250396729, + "learning_rate": 7.718120805369128e-05, + "loss": 0.8593, + "step": 690 + }, + { + "epoch": 0.04644139458407436, + "grad_norm": 1.9854339361190796, + "learning_rate": 7.740492170022372e-05, + "loss": 0.8604, + "step": 692 + }, + { + "epoch": 0.04657561826784336, + "grad_norm": 2.004243850708008, + "learning_rate": 7.762863534675615e-05, + "loss": 0.9429, + "step": 694 + }, + { + "epoch": 0.04670984195161236, + "grad_norm": 1.6457959413528442, + "learning_rate": 7.78523489932886e-05, + "loss": 0.9183, + "step": 696 + }, + { + "epoch": 0.04684406563538136, + "grad_norm": 1.4421991109848022, + "learning_rate": 7.807606263982104e-05, + "loss": 0.8882, + "step": 698 + }, + { + "epoch": 0.046978289319150365, + "grad_norm": 1.4433611631393433, + "learning_rate": 7.829977628635348e-05, + "loss": 0.8521, + "step": 700 + }, + { + "epoch": 0.04711251300291937, + "grad_norm": 1.7360423803329468, + "learning_rate": 7.852348993288591e-05, + "loss": 0.8972, + "step": 702 + }, + { + "epoch": 0.04724673668668836, + "grad_norm": 2.0711636543273926, + "learning_rate": 7.874720357941835e-05, + "loss": 0.896, + "step": 704 + }, + { + "epoch": 0.047380960370457366, + "grad_norm": 1.478452444076538, + "learning_rate": 7.897091722595079e-05, + "loss": 0.8851, + "step": 706 + }, + { + "epoch": 0.04751518405422637, + "grad_norm": 1.7454901933670044, + "learning_rate": 7.919463087248322e-05, + "loss": 0.9579, + "step": 708 + }, + { + "epoch": 0.04764940773799537, + "grad_norm": 2.5465352535247803, + "learning_rate": 7.941834451901567e-05, + "loss": 0.8532, + "step": 710 + }, + { + "epoch": 0.04778363142176437, + "grad_norm": 1.8807421922683716, + "learning_rate": 7.964205816554811e-05, + "loss": 0.8988, + "step": 712 + }, + { + "epoch": 0.04791785510553337, + "grad_norm": 1.62044095993042, + "learning_rate": 7.986577181208054e-05, + "loss": 0.8757, + "step": 714 + }, + { + "epoch": 0.04805207878930237, + "grad_norm": 1.3398281335830688, + "learning_rate": 8.008948545861298e-05, + "loss": 0.838, + "step": 716 + }, + { + "epoch": 0.048186302473071374, + "grad_norm": 1.6650171279907227, + "learning_rate": 8.031319910514542e-05, + "loss": 0.9152, + "step": 718 + }, + { + "epoch": 0.04832052615684038, + "grad_norm": 1.874007225036621, + "learning_rate": 8.053691275167784e-05, + "loss": 0.9129, + "step": 720 + }, + { + "epoch": 0.04845474984060937, + "grad_norm": 1.5779155492782593, + "learning_rate": 8.07606263982103e-05, + "loss": 0.8926, + "step": 722 + }, + { + "epoch": 0.048588973524378375, + "grad_norm": 1.4982414245605469, + "learning_rate": 8.098434004474274e-05, + "loss": 0.9389, + "step": 724 + }, + { + "epoch": 0.04872319720814738, + "grad_norm": 2.013594388961792, + "learning_rate": 8.120805369127518e-05, + "loss": 0.8542, + "step": 726 + }, + { + "epoch": 0.04885742089191638, + "grad_norm": 1.6982237100601196, + "learning_rate": 8.14317673378076e-05, + "loss": 0.9011, + "step": 728 + }, + { + "epoch": 0.04899164457568538, + "grad_norm": 3.0389931201934814, + "learning_rate": 8.165548098434005e-05, + "loss": 0.8657, + "step": 730 + }, + { + "epoch": 0.04912586825945438, + "grad_norm": 2.0701937675476074, + "learning_rate": 8.187919463087249e-05, + "loss": 0.8703, + "step": 732 + }, + { + "epoch": 0.04926009194322338, + "grad_norm": 1.7981239557266235, + "learning_rate": 8.210290827740493e-05, + "loss": 0.9151, + "step": 734 + }, + { + "epoch": 0.04939431562699238, + "grad_norm": 2.2304553985595703, + "learning_rate": 8.232662192393737e-05, + "loss": 0.8721, + "step": 736 + }, + { + "epoch": 0.049528539310761385, + "grad_norm": 2.8598451614379883, + "learning_rate": 8.255033557046981e-05, + "loss": 0.925, + "step": 738 + }, + { + "epoch": 0.04966276299453039, + "grad_norm": 1.5226657390594482, + "learning_rate": 8.277404921700224e-05, + "loss": 0.8633, + "step": 740 + }, + { + "epoch": 0.04979698667829938, + "grad_norm": 1.9706357717514038, + "learning_rate": 8.299776286353468e-05, + "loss": 0.8499, + "step": 742 + }, + { + "epoch": 0.049931210362068386, + "grad_norm": 1.6518230438232422, + "learning_rate": 8.322147651006712e-05, + "loss": 0.7827, + "step": 744 + }, + { + "epoch": 0.05006543404583739, + "grad_norm": 1.5772185325622559, + "learning_rate": 8.344519015659956e-05, + "loss": 0.946, + "step": 746 + }, + { + "epoch": 0.05019965772960639, + "grad_norm": 2.105870246887207, + "learning_rate": 8.3668903803132e-05, + "loss": 0.849, + "step": 748 + }, + { + "epoch": 0.05033388141337539, + "grad_norm": 1.616303563117981, + "learning_rate": 8.389261744966444e-05, + "loss": 0.8467, + "step": 750 + }, + { + "epoch": 0.05046810509714439, + "grad_norm": 2.292529582977295, + "learning_rate": 8.411633109619688e-05, + "loss": 0.9027, + "step": 752 + }, + { + "epoch": 0.05060232878091339, + "grad_norm": 1.5480836629867554, + "learning_rate": 8.43400447427293e-05, + "loss": 0.8651, + "step": 754 + }, + { + "epoch": 0.050736552464682394, + "grad_norm": 1.5585349798202515, + "learning_rate": 8.456375838926175e-05, + "loss": 0.8554, + "step": 756 + }, + { + "epoch": 0.050870776148451397, + "grad_norm": 1.6302483081817627, + "learning_rate": 8.478747203579419e-05, + "loss": 0.8242, + "step": 758 + }, + { + "epoch": 0.05100499983222039, + "grad_norm": 1.8211963176727295, + "learning_rate": 8.501118568232663e-05, + "loss": 0.9038, + "step": 760 + }, + { + "epoch": 0.051139223515989395, + "grad_norm": 1.9375659227371216, + "learning_rate": 8.523489932885907e-05, + "loss": 0.8742, + "step": 762 + }, + { + "epoch": 0.0512734471997584, + "grad_norm": 1.5411709547042847, + "learning_rate": 8.545861297539151e-05, + "loss": 0.8768, + "step": 764 + }, + { + "epoch": 0.0514076708835274, + "grad_norm": 1.6576482057571411, + "learning_rate": 8.568232662192394e-05, + "loss": 0.8962, + "step": 766 + }, + { + "epoch": 0.0515418945672964, + "grad_norm": 1.7040925025939941, + "learning_rate": 8.590604026845638e-05, + "loss": 0.8104, + "step": 768 + }, + { + "epoch": 0.0516761182510654, + "grad_norm": 1.7019544839859009, + "learning_rate": 8.612975391498882e-05, + "loss": 0.885, + "step": 770 + }, + { + "epoch": 0.0518103419348344, + "grad_norm": 1.4850620031356812, + "learning_rate": 8.635346756152126e-05, + "loss": 0.9345, + "step": 772 + }, + { + "epoch": 0.0519445656186034, + "grad_norm": 2.2853121757507324, + "learning_rate": 8.65771812080537e-05, + "loss": 0.9004, + "step": 774 + }, + { + "epoch": 0.052078789302372405, + "grad_norm": 1.718302845954895, + "learning_rate": 8.680089485458614e-05, + "loss": 0.8885, + "step": 776 + }, + { + "epoch": 0.05221301298614141, + "grad_norm": 2.4001340866088867, + "learning_rate": 8.702460850111858e-05, + "loss": 0.883, + "step": 778 + }, + { + "epoch": 0.0523472366699104, + "grad_norm": 1.4307639598846436, + "learning_rate": 8.7248322147651e-05, + "loss": 0.8925, + "step": 780 + }, + { + "epoch": 0.052481460353679406, + "grad_norm": 1.3757809400558472, + "learning_rate": 8.747203579418345e-05, + "loss": 0.8874, + "step": 782 + }, + { + "epoch": 0.05261568403744841, + "grad_norm": 1.4396053552627563, + "learning_rate": 8.769574944071589e-05, + "loss": 0.9032, + "step": 784 + }, + { + "epoch": 0.05274990772121741, + "grad_norm": 1.6641589403152466, + "learning_rate": 8.791946308724833e-05, + "loss": 0.9667, + "step": 786 + }, + { + "epoch": 0.052884131404986406, + "grad_norm": 1.66237211227417, + "learning_rate": 8.814317673378077e-05, + "loss": 0.9228, + "step": 788 + }, + { + "epoch": 0.05301835508875541, + "grad_norm": 1.4634531736373901, + "learning_rate": 8.836689038031321e-05, + "loss": 0.9049, + "step": 790 + }, + { + "epoch": 0.05315257877252441, + "grad_norm": 2.0408074855804443, + "learning_rate": 8.859060402684565e-05, + "loss": 0.9163, + "step": 792 + }, + { + "epoch": 0.053286802456293414, + "grad_norm": 1.4953192472457886, + "learning_rate": 8.881431767337808e-05, + "loss": 0.922, + "step": 794 + }, + { + "epoch": 0.053421026140062416, + "grad_norm": 2.698669672012329, + "learning_rate": 8.903803131991052e-05, + "loss": 0.8641, + "step": 796 + }, + { + "epoch": 0.05355524982383141, + "grad_norm": 1.5639328956604004, + "learning_rate": 8.926174496644296e-05, + "loss": 0.8754, + "step": 798 + }, + { + "epoch": 0.053689473507600415, + "grad_norm": 1.4404573440551758, + "learning_rate": 8.94854586129754e-05, + "loss": 0.8646, + "step": 800 + }, + { + "epoch": 0.05382369719136942, + "grad_norm": 2.0549895763397217, + "learning_rate": 8.970917225950784e-05, + "loss": 0.8184, + "step": 802 + }, + { + "epoch": 0.05395792087513842, + "grad_norm": 2.4482076168060303, + "learning_rate": 8.993288590604028e-05, + "loss": 0.8865, + "step": 804 + }, + { + "epoch": 0.05409214455890742, + "grad_norm": 2.617295980453491, + "learning_rate": 9.01565995525727e-05, + "loss": 0.9126, + "step": 806 + }, + { + "epoch": 0.05422636824267642, + "grad_norm": 1.6565312147140503, + "learning_rate": 9.038031319910515e-05, + "loss": 0.8668, + "step": 808 + }, + { + "epoch": 0.05436059192644542, + "grad_norm": 1.6836875677108765, + "learning_rate": 9.060402684563759e-05, + "loss": 0.8412, + "step": 810 + }, + { + "epoch": 0.05449481561021442, + "grad_norm": 3.3620588779449463, + "learning_rate": 9.082774049217003e-05, + "loss": 0.8967, + "step": 812 + }, + { + "epoch": 0.054629039293983425, + "grad_norm": 1.7039563655853271, + "learning_rate": 9.105145413870247e-05, + "loss": 0.8374, + "step": 814 + }, + { + "epoch": 0.05476326297775243, + "grad_norm": 1.5863765478134155, + "learning_rate": 9.127516778523491e-05, + "loss": 0.821, + "step": 816 + }, + { + "epoch": 0.05489748666152142, + "grad_norm": 1.436874270439148, + "learning_rate": 9.149888143176735e-05, + "loss": 0.9001, + "step": 818 + }, + { + "epoch": 0.055031710345290426, + "grad_norm": 1.9585341215133667, + "learning_rate": 9.172259507829977e-05, + "loss": 0.8977, + "step": 820 + }, + { + "epoch": 0.05516593402905943, + "grad_norm": 1.5336610078811646, + "learning_rate": 9.194630872483221e-05, + "loss": 0.8992, + "step": 822 + }, + { + "epoch": 0.05530015771282843, + "grad_norm": 2.7829861640930176, + "learning_rate": 9.217002237136466e-05, + "loss": 0.922, + "step": 824 + }, + { + "epoch": 0.055434381396597426, + "grad_norm": 1.4213629961013794, + "learning_rate": 9.23937360178971e-05, + "loss": 0.9008, + "step": 826 + }, + { + "epoch": 0.05556860508036643, + "grad_norm": 1.4592857360839844, + "learning_rate": 9.261744966442954e-05, + "loss": 0.884, + "step": 828 + }, + { + "epoch": 0.05570282876413543, + "grad_norm": 1.7585809230804443, + "learning_rate": 9.284116331096198e-05, + "loss": 0.8807, + "step": 830 + }, + { + "epoch": 0.055837052447904434, + "grad_norm": 1.8087443113327026, + "learning_rate": 9.30648769574944e-05, + "loss": 0.867, + "step": 832 + }, + { + "epoch": 0.055971276131673436, + "grad_norm": 2.120878219604492, + "learning_rate": 9.328859060402684e-05, + "loss": 0.8508, + "step": 834 + }, + { + "epoch": 0.05610549981544243, + "grad_norm": 3.5708112716674805, + "learning_rate": 9.351230425055928e-05, + "loss": 0.9109, + "step": 836 + }, + { + "epoch": 0.056239723499211434, + "grad_norm": 1.4622738361358643, + "learning_rate": 9.373601789709174e-05, + "loss": 0.8472, + "step": 838 + }, + { + "epoch": 0.05637394718298044, + "grad_norm": 1.4327478408813477, + "learning_rate": 9.395973154362417e-05, + "loss": 0.8546, + "step": 840 + }, + { + "epoch": 0.05650817086674944, + "grad_norm": 1.8220202922821045, + "learning_rate": 9.41834451901566e-05, + "loss": 0.9217, + "step": 842 + }, + { + "epoch": 0.05664239455051844, + "grad_norm": 3.180845260620117, + "learning_rate": 9.440715883668905e-05, + "loss": 0.9166, + "step": 844 + }, + { + "epoch": 0.05677661823428744, + "grad_norm": 1.602735161781311, + "learning_rate": 9.463087248322147e-05, + "loss": 0.9221, + "step": 846 + }, + { + "epoch": 0.05691084191805644, + "grad_norm": 1.3909780979156494, + "learning_rate": 9.485458612975391e-05, + "loss": 0.8352, + "step": 848 + }, + { + "epoch": 0.05704506560182544, + "grad_norm": 1.4142225980758667, + "learning_rate": 9.507829977628635e-05, + "loss": 0.8591, + "step": 850 + }, + { + "epoch": 0.057179289285594445, + "grad_norm": 3.468468427658081, + "learning_rate": 9.53020134228188e-05, + "loss": 0.8948, + "step": 852 + }, + { + "epoch": 0.05731351296936345, + "grad_norm": 1.839051604270935, + "learning_rate": 9.552572706935124e-05, + "loss": 0.8648, + "step": 854 + }, + { + "epoch": 0.05744773665313244, + "grad_norm": 1.7877297401428223, + "learning_rate": 9.574944071588368e-05, + "loss": 0.8961, + "step": 856 + }, + { + "epoch": 0.057581960336901446, + "grad_norm": 1.9843209981918335, + "learning_rate": 9.59731543624161e-05, + "loss": 0.8792, + "step": 858 + }, + { + "epoch": 0.05771618402067045, + "grad_norm": 1.921899437904358, + "learning_rate": 9.619686800894854e-05, + "loss": 0.8588, + "step": 860 + }, + { + "epoch": 0.05785040770443945, + "grad_norm": 2.093482255935669, + "learning_rate": 9.642058165548098e-05, + "loss": 0.9604, + "step": 862 + }, + { + "epoch": 0.057984631388208446, + "grad_norm": 1.8832120895385742, + "learning_rate": 9.664429530201344e-05, + "loss": 0.908, + "step": 864 + }, + { + "epoch": 0.05811885507197745, + "grad_norm": 1.2885879278182983, + "learning_rate": 9.686800894854587e-05, + "loss": 0.8453, + "step": 866 + }, + { + "epoch": 0.05825307875574645, + "grad_norm": 1.905735969543457, + "learning_rate": 9.70917225950783e-05, + "loss": 0.9275, + "step": 868 + }, + { + "epoch": 0.058387302439515454, + "grad_norm": 2.0157928466796875, + "learning_rate": 9.731543624161075e-05, + "loss": 0.8819, + "step": 870 + }, + { + "epoch": 0.058521526123284456, + "grad_norm": 1.3370037078857422, + "learning_rate": 9.753914988814317e-05, + "loss": 0.8153, + "step": 872 + }, + { + "epoch": 0.05865574980705345, + "grad_norm": 1.4842098951339722, + "learning_rate": 9.776286353467561e-05, + "loss": 0.8796, + "step": 874 + }, + { + "epoch": 0.058789973490822454, + "grad_norm": 1.4972621202468872, + "learning_rate": 9.798657718120807e-05, + "loss": 0.8647, + "step": 876 + }, + { + "epoch": 0.05892419717459146, + "grad_norm": 1.740038275718689, + "learning_rate": 9.82102908277405e-05, + "loss": 0.8925, + "step": 878 + }, + { + "epoch": 0.05905842085836046, + "grad_norm": 1.52963387966156, + "learning_rate": 9.843400447427293e-05, + "loss": 0.9363, + "step": 880 + }, + { + "epoch": 0.05919264454212946, + "grad_norm": 2.290513038635254, + "learning_rate": 9.865771812080538e-05, + "loss": 0.8879, + "step": 882 + }, + { + "epoch": 0.05932686822589846, + "grad_norm": 1.6631953716278076, + "learning_rate": 9.888143176733782e-05, + "loss": 0.8762, + "step": 884 + }, + { + "epoch": 0.05946109190966746, + "grad_norm": 1.472398042678833, + "learning_rate": 9.910514541387024e-05, + "loss": 0.8953, + "step": 886 + }, + { + "epoch": 0.05959531559343646, + "grad_norm": 1.6663011312484741, + "learning_rate": 9.93288590604027e-05, + "loss": 0.8678, + "step": 888 + }, + { + "epoch": 0.059729539277205465, + "grad_norm": 2.3357253074645996, + "learning_rate": 9.955257270693514e-05, + "loss": 0.898, + "step": 890 + }, + { + "epoch": 0.05986376296097446, + "grad_norm": 1.9164422750473022, + "learning_rate": 9.977628635346756e-05, + "loss": 0.8616, + "step": 892 + }, + { + "epoch": 0.05999798664474346, + "grad_norm": 1.9118332862854004, + "learning_rate": 0.0001, + "loss": 0.9123, + "step": 894 + }, + { + "epoch": 0.060132210328512466, + "grad_norm": 1.4247066974639893, + "learning_rate": 9.999999881879917e-05, + "loss": 0.8485, + "step": 896 + }, + { + "epoch": 0.06026643401228147, + "grad_norm": 1.6532388925552368, + "learning_rate": 9.99999952751967e-05, + "loss": 0.8958, + "step": 898 + }, + { + "epoch": 0.06040065769605047, + "grad_norm": 1.411152958869934, + "learning_rate": 9.999998936919278e-05, + "loss": 0.8478, + "step": 900 + }, + { + "epoch": 0.060534881379819466, + "grad_norm": 1.296415090560913, + "learning_rate": 9.999998110078769e-05, + "loss": 0.8662, + "step": 902 + }, + { + "epoch": 0.06066910506358847, + "grad_norm": 1.436594009399414, + "learning_rate": 9.99999704699818e-05, + "loss": 0.8919, + "step": 904 + }, + { + "epoch": 0.06080332874735747, + "grad_norm": 1.9760971069335938, + "learning_rate": 9.999995747677564e-05, + "loss": 0.801, + "step": 906 + }, + { + "epoch": 0.060937552431126474, + "grad_norm": 2.1906111240386963, + "learning_rate": 9.99999421211698e-05, + "loss": 0.85, + "step": 908 + }, + { + "epoch": 0.061071776114895476, + "grad_norm": 1.5004881620407104, + "learning_rate": 9.999992440316502e-05, + "loss": 0.8712, + "step": 910 + }, + { + "epoch": 0.06120599979866447, + "grad_norm": 1.4594885110855103, + "learning_rate": 9.999990432276214e-05, + "loss": 0.8529, + "step": 912 + }, + { + "epoch": 0.061340223482433474, + "grad_norm": 1.8031158447265625, + "learning_rate": 9.999988187996208e-05, + "loss": 0.885, + "step": 914 + }, + { + "epoch": 0.06147444716620248, + "grad_norm": 1.571834921836853, + "learning_rate": 9.999985707476594e-05, + "loss": 0.8803, + "step": 916 + }, + { + "epoch": 0.06160867084997148, + "grad_norm": 1.6122395992279053, + "learning_rate": 9.999982990717487e-05, + "loss": 0.882, + "step": 918 + }, + { + "epoch": 0.06174289453374048, + "grad_norm": 2.581127643585205, + "learning_rate": 9.999980037719016e-05, + "loss": 0.8602, + "step": 920 + }, + { + "epoch": 0.06187711821750948, + "grad_norm": 1.770944595336914, + "learning_rate": 9.99997684848132e-05, + "loss": 0.9097, + "step": 922 + }, + { + "epoch": 0.06201134190127848, + "grad_norm": 1.7628514766693115, + "learning_rate": 9.99997342300455e-05, + "loss": 0.8812, + "step": 924 + }, + { + "epoch": 0.06214556558504748, + "grad_norm": 1.4451676607131958, + "learning_rate": 9.999969761288868e-05, + "loss": 0.8952, + "step": 926 + }, + { + "epoch": 0.062279789268816485, + "grad_norm": 2.1536195278167725, + "learning_rate": 9.999965863334445e-05, + "loss": 0.8244, + "step": 928 + }, + { + "epoch": 0.06241401295258548, + "grad_norm": 1.5444164276123047, + "learning_rate": 9.99996172914147e-05, + "loss": 0.9155, + "step": 930 + }, + { + "epoch": 0.06254823663635449, + "grad_norm": 1.5740525722503662, + "learning_rate": 9.999957358710132e-05, + "loss": 0.8623, + "step": 932 + }, + { + "epoch": 0.06268246032012349, + "grad_norm": 1.4494963884353638, + "learning_rate": 9.999952752040643e-05, + "loss": 0.9155, + "step": 934 + }, + { + "epoch": 0.06281668400389248, + "grad_norm": 1.6744060516357422, + "learning_rate": 9.999947909133219e-05, + "loss": 0.8704, + "step": 936 + }, + { + "epoch": 0.06295090768766148, + "grad_norm": 1.5943561792373657, + "learning_rate": 9.999942829988086e-05, + "loss": 0.8005, + "step": 938 + }, + { + "epoch": 0.06308513137143049, + "grad_norm": 1.6639518737792969, + "learning_rate": 9.999937514605486e-05, + "loss": 0.834, + "step": 940 + }, + { + "epoch": 0.06321935505519949, + "grad_norm": 1.3946268558502197, + "learning_rate": 9.999931962985674e-05, + "loss": 0.8614, + "step": 942 + }, + { + "epoch": 0.06335357873896849, + "grad_norm": 1.7072865962982178, + "learning_rate": 9.999926175128905e-05, + "loss": 0.8979, + "step": 944 + }, + { + "epoch": 0.0634878024227375, + "grad_norm": 2.105038642883301, + "learning_rate": 9.999920151035458e-05, + "loss": 0.9333, + "step": 946 + }, + { + "epoch": 0.0636220261065065, + "grad_norm": 1.498768925666809, + "learning_rate": 9.999913890705616e-05, + "loss": 0.8761, + "step": 948 + }, + { + "epoch": 0.0637562497902755, + "grad_norm": 1.3514175415039062, + "learning_rate": 9.999907394139674e-05, + "loss": 0.8673, + "step": 950 + }, + { + "epoch": 0.0638904734740445, + "grad_norm": 1.5056959390640259, + "learning_rate": 9.99990066133794e-05, + "loss": 0.89, + "step": 952 + }, + { + "epoch": 0.06402469715781349, + "grad_norm": 1.3694546222686768, + "learning_rate": 9.99989369230073e-05, + "loss": 0.8391, + "step": 954 + }, + { + "epoch": 0.06415892084158249, + "grad_norm": 1.7158812284469604, + "learning_rate": 9.999886487028376e-05, + "loss": 0.9251, + "step": 956 + }, + { + "epoch": 0.0642931445253515, + "grad_norm": 1.927911400794983, + "learning_rate": 9.999879045521218e-05, + "loss": 0.9081, + "step": 958 + }, + { + "epoch": 0.0644273682091205, + "grad_norm": 1.925579309463501, + "learning_rate": 9.999871367779606e-05, + "loss": 0.854, + "step": 960 + }, + { + "epoch": 0.0645615918928895, + "grad_norm": 1.2850353717803955, + "learning_rate": 9.999863453803904e-05, + "loss": 0.8755, + "step": 962 + }, + { + "epoch": 0.0646958155766585, + "grad_norm": 1.4252276420593262, + "learning_rate": 9.999855303594485e-05, + "loss": 0.8555, + "step": 964 + }, + { + "epoch": 0.0648300392604275, + "grad_norm": 3.0897445678710938, + "learning_rate": 9.999846917151737e-05, + "loss": 0.8602, + "step": 966 + }, + { + "epoch": 0.06496426294419651, + "grad_norm": 1.5266180038452148, + "learning_rate": 9.999838294476051e-05, + "loss": 0.8598, + "step": 968 + }, + { + "epoch": 0.06509848662796551, + "grad_norm": 2.7690606117248535, + "learning_rate": 9.99982943556784e-05, + "loss": 0.885, + "step": 970 + }, + { + "epoch": 0.0652327103117345, + "grad_norm": 1.3687183856964111, + "learning_rate": 9.999820340427517e-05, + "loss": 0.8545, + "step": 972 + }, + { + "epoch": 0.0653669339955035, + "grad_norm": 1.5934265851974487, + "learning_rate": 9.999811009055518e-05, + "loss": 0.8277, + "step": 974 + }, + { + "epoch": 0.0655011576792725, + "grad_norm": 2.1981096267700195, + "learning_rate": 9.999801441452278e-05, + "loss": 0.7975, + "step": 976 + }, + { + "epoch": 0.0656353813630415, + "grad_norm": 1.4350436925888062, + "learning_rate": 9.999791637618252e-05, + "loss": 0.8324, + "step": 978 + }, + { + "epoch": 0.06576960504681051, + "grad_norm": 2.4169490337371826, + "learning_rate": 9.999781597553903e-05, + "loss": 0.8526, + "step": 980 + }, + { + "epoch": 0.06590382873057951, + "grad_norm": 1.2815511226654053, + "learning_rate": 9.999771321259705e-05, + "loss": 0.8884, + "step": 982 + }, + { + "epoch": 0.06603805241434851, + "grad_norm": 1.4573643207550049, + "learning_rate": 9.999760808736145e-05, + "loss": 0.8733, + "step": 984 + }, + { + "epoch": 0.06617227609811752, + "grad_norm": 1.8162658214569092, + "learning_rate": 9.999750059983716e-05, + "loss": 0.8241, + "step": 986 + }, + { + "epoch": 0.06630649978188652, + "grad_norm": 1.3659766912460327, + "learning_rate": 9.999739075002931e-05, + "loss": 0.895, + "step": 988 + }, + { + "epoch": 0.06644072346565552, + "grad_norm": 1.7499064207077026, + "learning_rate": 9.999727853794305e-05, + "loss": 0.928, + "step": 990 + }, + { + "epoch": 0.06657494714942451, + "grad_norm": 1.3851163387298584, + "learning_rate": 9.999716396358369e-05, + "loss": 0.8947, + "step": 992 + }, + { + "epoch": 0.06670917083319351, + "grad_norm": 2.0054306983947754, + "learning_rate": 9.999704702695664e-05, + "loss": 0.884, + "step": 994 + }, + { + "epoch": 0.06684339451696251, + "grad_norm": 1.4638731479644775, + "learning_rate": 9.999692772806746e-05, + "loss": 0.9536, + "step": 996 + }, + { + "epoch": 0.06697761820073152, + "grad_norm": 1.4908126592636108, + "learning_rate": 9.999680606692174e-05, + "loss": 0.8523, + "step": 998 + }, + { + "epoch": 0.06711184188450052, + "grad_norm": 1.5436162948608398, + "learning_rate": 9.999668204352526e-05, + "loss": 0.8326, + "step": 1000 + }, + { + "epoch": 0.06724606556826952, + "grad_norm": 1.7729190587997437, + "learning_rate": 9.999655565788385e-05, + "loss": 0.859, + "step": 1002 + }, + { + "epoch": 0.06738028925203852, + "grad_norm": 2.2067604064941406, + "learning_rate": 9.99964269100035e-05, + "loss": 0.793, + "step": 1004 + }, + { + "epoch": 0.06751451293580753, + "grad_norm": 1.7928110361099243, + "learning_rate": 9.999629579989032e-05, + "loss": 0.8179, + "step": 1006 + }, + { + "epoch": 0.06764873661957653, + "grad_norm": 1.3039089441299438, + "learning_rate": 9.999616232755045e-05, + "loss": 0.8545, + "step": 1008 + }, + { + "epoch": 0.06778296030334552, + "grad_norm": 1.576532244682312, + "learning_rate": 9.999602649299022e-05, + "loss": 0.842, + "step": 1010 + }, + { + "epoch": 0.06791718398711452, + "grad_norm": 1.6710187196731567, + "learning_rate": 9.999588829621606e-05, + "loss": 0.911, + "step": 1012 + }, + { + "epoch": 0.06805140767088352, + "grad_norm": 1.279616117477417, + "learning_rate": 9.99957477372345e-05, + "loss": 0.8623, + "step": 1014 + }, + { + "epoch": 0.06818563135465253, + "grad_norm": 1.345814824104309, + "learning_rate": 9.999560481605217e-05, + "loss": 0.8028, + "step": 1016 + }, + { + "epoch": 0.06831985503842153, + "grad_norm": 1.5044865608215332, + "learning_rate": 9.999545953267582e-05, + "loss": 0.8385, + "step": 1018 + }, + { + "epoch": 0.06845407872219053, + "grad_norm": 1.6629892587661743, + "learning_rate": 9.999531188711232e-05, + "loss": 0.8897, + "step": 1020 + }, + { + "epoch": 0.06858830240595953, + "grad_norm": 1.2309874296188354, + "learning_rate": 9.999516187936864e-05, + "loss": 0.8042, + "step": 1022 + }, + { + "epoch": 0.06872252608972854, + "grad_norm": 1.2960044145584106, + "learning_rate": 9.999500950945188e-05, + "loss": 0.8592, + "step": 1024 + }, + { + "epoch": 0.06885674977349754, + "grad_norm": 3.2066714763641357, + "learning_rate": 9.999485477736923e-05, + "loss": 0.8411, + "step": 1026 + }, + { + "epoch": 0.06899097345726654, + "grad_norm": 1.5629687309265137, + "learning_rate": 9.999469768312799e-05, + "loss": 0.9141, + "step": 1028 + }, + { + "epoch": 0.06912519714103553, + "grad_norm": 1.96613609790802, + "learning_rate": 9.99945382267356e-05, + "loss": 0.9002, + "step": 1030 + }, + { + "epoch": 0.06925942082480453, + "grad_norm": 1.3429502248764038, + "learning_rate": 9.999437640819959e-05, + "loss": 0.9162, + "step": 1032 + }, + { + "epoch": 0.06939364450857353, + "grad_norm": 1.2782080173492432, + "learning_rate": 9.999421222752763e-05, + "loss": 0.8635, + "step": 1034 + }, + { + "epoch": 0.06952786819234254, + "grad_norm": 1.330491065979004, + "learning_rate": 9.999404568472742e-05, + "loss": 0.884, + "step": 1036 + }, + { + "epoch": 0.06966209187611154, + "grad_norm": 1.3575077056884766, + "learning_rate": 9.999387677980687e-05, + "loss": 0.8559, + "step": 1038 + }, + { + "epoch": 0.06979631555988054, + "grad_norm": 1.6015293598175049, + "learning_rate": 9.999370551277395e-05, + "loss": 0.8575, + "step": 1040 + }, + { + "epoch": 0.06993053924364954, + "grad_norm": 1.6036041975021362, + "learning_rate": 9.999353188363676e-05, + "loss": 0.8912, + "step": 1042 + }, + { + "epoch": 0.07006476292741855, + "grad_norm": 1.4285792112350464, + "learning_rate": 9.999335589240348e-05, + "loss": 0.8507, + "step": 1044 + }, + { + "epoch": 0.07019898661118755, + "grad_norm": 1.7658170461654663, + "learning_rate": 9.999317753908246e-05, + "loss": 0.869, + "step": 1046 + }, + { + "epoch": 0.07033321029495654, + "grad_norm": 1.4544686079025269, + "learning_rate": 9.999299682368211e-05, + "loss": 0.8702, + "step": 1048 + }, + { + "epoch": 0.07046743397872554, + "grad_norm": 1.428667426109314, + "learning_rate": 9.999281374621095e-05, + "loss": 0.8041, + "step": 1050 + }, + { + "epoch": 0.07060165766249454, + "grad_norm": 1.5885188579559326, + "learning_rate": 9.999262830667766e-05, + "loss": 0.8531, + "step": 1052 + }, + { + "epoch": 0.07073588134626355, + "grad_norm": 6.514537334442139, + "learning_rate": 9.999244050509098e-05, + "loss": 0.869, + "step": 1054 + }, + { + "epoch": 0.07087010503003255, + "grad_norm": 1.5298740863800049, + "learning_rate": 9.999225034145979e-05, + "loss": 0.8104, + "step": 1056 + }, + { + "epoch": 0.07100432871380155, + "grad_norm": 1.3994238376617432, + "learning_rate": 9.999205781579309e-05, + "loss": 0.8596, + "step": 1058 + }, + { + "epoch": 0.07113855239757055, + "grad_norm": 1.8185546398162842, + "learning_rate": 9.999186292809995e-05, + "loss": 0.8485, + "step": 1060 + }, + { + "epoch": 0.07127277608133956, + "grad_norm": 1.4394915103912354, + "learning_rate": 9.99916656783896e-05, + "loss": 0.8019, + "step": 1062 + }, + { + "epoch": 0.07140699976510856, + "grad_norm": 1.567618727684021, + "learning_rate": 9.999146606667135e-05, + "loss": 0.8233, + "step": 1064 + }, + { + "epoch": 0.07154122344887756, + "grad_norm": 1.8687235116958618, + "learning_rate": 9.999126409295463e-05, + "loss": 0.8539, + "step": 1066 + }, + { + "epoch": 0.07167544713264655, + "grad_norm": 1.3390531539916992, + "learning_rate": 9.999105975724898e-05, + "loss": 0.8283, + "step": 1068 + }, + { + "epoch": 0.07180967081641555, + "grad_norm": 2.7048981189727783, + "learning_rate": 9.999085305956406e-05, + "loss": 0.819, + "step": 1070 + }, + { + "epoch": 0.07194389450018455, + "grad_norm": 1.721372365951538, + "learning_rate": 9.999064399990964e-05, + "loss": 0.8514, + "step": 1072 + }, + { + "epoch": 0.07207811818395356, + "grad_norm": 1.280475378036499, + "learning_rate": 9.999043257829561e-05, + "loss": 0.7827, + "step": 1074 + }, + { + "epoch": 0.07221234186772256, + "grad_norm": 3.6173923015594482, + "learning_rate": 9.999021879473192e-05, + "loss": 0.7947, + "step": 1076 + }, + { + "epoch": 0.07234656555149156, + "grad_norm": 2.0217697620391846, + "learning_rate": 9.99900026492287e-05, + "loss": 0.8223, + "step": 1078 + }, + { + "epoch": 0.07248078923526056, + "grad_norm": 1.4395509958267212, + "learning_rate": 9.998978414179617e-05, + "loss": 0.8473, + "step": 1080 + }, + { + "epoch": 0.07261501291902957, + "grad_norm": 1.426866054534912, + "learning_rate": 9.998956327244462e-05, + "loss": 0.9186, + "step": 1082 + }, + { + "epoch": 0.07274923660279857, + "grad_norm": 1.339833378791809, + "learning_rate": 9.998934004118452e-05, + "loss": 0.8104, + "step": 1084 + }, + { + "epoch": 0.07288346028656756, + "grad_norm": 1.3739291429519653, + "learning_rate": 9.99891144480264e-05, + "loss": 0.808, + "step": 1086 + }, + { + "epoch": 0.07301768397033656, + "grad_norm": 1.93855881690979, + "learning_rate": 9.99888864929809e-05, + "loss": 0.8228, + "step": 1088 + }, + { + "epoch": 0.07315190765410556, + "grad_norm": 1.507676362991333, + "learning_rate": 9.998865617605883e-05, + "loss": 0.8457, + "step": 1090 + }, + { + "epoch": 0.07328613133787457, + "grad_norm": 1.841133952140808, + "learning_rate": 9.998842349727107e-05, + "loss": 0.8103, + "step": 1092 + }, + { + "epoch": 0.07342035502164357, + "grad_norm": 1.2533656358718872, + "learning_rate": 9.998818845662859e-05, + "loss": 0.8232, + "step": 1094 + }, + { + "epoch": 0.07355457870541257, + "grad_norm": 1.4022369384765625, + "learning_rate": 9.998795105414248e-05, + "loss": 0.8436, + "step": 1096 + }, + { + "epoch": 0.07368880238918157, + "grad_norm": 1.5202876329421997, + "learning_rate": 9.998771128982399e-05, + "loss": 0.8699, + "step": 1098 + }, + { + "epoch": 0.07382302607295058, + "grad_norm": 1.4362784624099731, + "learning_rate": 9.998746916368444e-05, + "loss": 0.8737, + "step": 1100 + }, + { + "epoch": 0.07395724975671958, + "grad_norm": 1.4843257665634155, + "learning_rate": 9.998722467573528e-05, + "loss": 0.8605, + "step": 1102 + }, + { + "epoch": 0.07409147344048858, + "grad_norm": 1.275189757347107, + "learning_rate": 9.998697782598804e-05, + "loss": 0.7974, + "step": 1104 + }, + { + "epoch": 0.07422569712425757, + "grad_norm": 1.3372191190719604, + "learning_rate": 9.998672861445439e-05, + "loss": 0.9027, + "step": 1106 + }, + { + "epoch": 0.07435992080802657, + "grad_norm": 1.7303627729415894, + "learning_rate": 9.998647704114612e-05, + "loss": 0.8333, + "step": 1108 + }, + { + "epoch": 0.07449414449179557, + "grad_norm": 1.6017955541610718, + "learning_rate": 9.998622310607508e-05, + "loss": 0.8252, + "step": 1110 + }, + { + "epoch": 0.07462836817556458, + "grad_norm": 1.5187143087387085, + "learning_rate": 9.998596680925331e-05, + "loss": 0.868, + "step": 1112 + }, + { + "epoch": 0.07476259185933358, + "grad_norm": 1.3681526184082031, + "learning_rate": 9.99857081506929e-05, + "loss": 0.8218, + "step": 1114 + }, + { + "epoch": 0.07489681554310258, + "grad_norm": 1.9026960134506226, + "learning_rate": 9.998544713040608e-05, + "loss": 0.7745, + "step": 1116 + }, + { + "epoch": 0.07503103922687158, + "grad_norm": 1.8776766061782837, + "learning_rate": 9.998518374840515e-05, + "loss": 0.8585, + "step": 1118 + }, + { + "epoch": 0.07516526291064059, + "grad_norm": 2.2423903942108154, + "learning_rate": 9.998491800470259e-05, + "loss": 0.8332, + "step": 1120 + }, + { + "epoch": 0.07529948659440959, + "grad_norm": 1.6247942447662354, + "learning_rate": 9.998464989931097e-05, + "loss": 0.8407, + "step": 1122 + }, + { + "epoch": 0.07543371027817858, + "grad_norm": 1.3288408517837524, + "learning_rate": 9.998437943224292e-05, + "loss": 0.865, + "step": 1124 + }, + { + "epoch": 0.07556793396194758, + "grad_norm": 1.5527228116989136, + "learning_rate": 9.998410660351121e-05, + "loss": 0.8223, + "step": 1126 + }, + { + "epoch": 0.07570215764571658, + "grad_norm": 1.4501368999481201, + "learning_rate": 9.998383141312877e-05, + "loss": 0.8355, + "step": 1128 + }, + { + "epoch": 0.07583638132948559, + "grad_norm": 1.5103774070739746, + "learning_rate": 9.99835538611086e-05, + "loss": 0.8287, + "step": 1130 + }, + { + "epoch": 0.07597060501325459, + "grad_norm": 1.4005253314971924, + "learning_rate": 9.998327394746378e-05, + "loss": 0.8736, + "step": 1132 + }, + { + "epoch": 0.07610482869702359, + "grad_norm": 2.2452094554901123, + "learning_rate": 9.998299167220755e-05, + "loss": 0.8813, + "step": 1134 + }, + { + "epoch": 0.0762390523807926, + "grad_norm": 1.320500373840332, + "learning_rate": 9.998270703535326e-05, + "loss": 0.7934, + "step": 1136 + }, + { + "epoch": 0.0763732760645616, + "grad_norm": 1.582572340965271, + "learning_rate": 9.998242003691434e-05, + "loss": 0.9164, + "step": 1138 + }, + { + "epoch": 0.0765074997483306, + "grad_norm": 1.6067897081375122, + "learning_rate": 9.998213067690436e-05, + "loss": 0.8988, + "step": 1140 + }, + { + "epoch": 0.07664172343209959, + "grad_norm": 1.3594622611999512, + "learning_rate": 9.998183895533701e-05, + "loss": 0.8876, + "step": 1142 + }, + { + "epoch": 0.07677594711586859, + "grad_norm": 1.7617809772491455, + "learning_rate": 9.998154487222602e-05, + "loss": 0.8808, + "step": 1144 + }, + { + "epoch": 0.07691017079963759, + "grad_norm": 1.5601156949996948, + "learning_rate": 9.998124842758535e-05, + "loss": 0.8613, + "step": 1146 + }, + { + "epoch": 0.0770443944834066, + "grad_norm": 1.5522006750106812, + "learning_rate": 9.998094962142897e-05, + "loss": 0.8229, + "step": 1148 + }, + { + "epoch": 0.0771786181671756, + "grad_norm": 1.2793421745300293, + "learning_rate": 9.9980648453771e-05, + "loss": 0.8181, + "step": 1150 + }, + { + "epoch": 0.0773128418509446, + "grad_norm": 1.309046745300293, + "learning_rate": 9.998034492462567e-05, + "loss": 0.8801, + "step": 1152 + }, + { + "epoch": 0.0774470655347136, + "grad_norm": 1.305838704109192, + "learning_rate": 9.998003903400732e-05, + "loss": 0.8232, + "step": 1154 + }, + { + "epoch": 0.0775812892184826, + "grad_norm": 1.352232575416565, + "learning_rate": 9.997973078193041e-05, + "loss": 0.9048, + "step": 1156 + }, + { + "epoch": 0.07771551290225161, + "grad_norm": 1.3795783519744873, + "learning_rate": 9.99794201684095e-05, + "loss": 0.8422, + "step": 1158 + }, + { + "epoch": 0.07784973658602061, + "grad_norm": 1.46577787399292, + "learning_rate": 9.997910719345928e-05, + "loss": 0.8454, + "step": 1160 + }, + { + "epoch": 0.0779839602697896, + "grad_norm": 1.3010011911392212, + "learning_rate": 9.997879185709453e-05, + "loss": 0.8205, + "step": 1162 + }, + { + "epoch": 0.0781181839535586, + "grad_norm": 1.2190300226211548, + "learning_rate": 9.997847415933012e-05, + "loss": 0.7828, + "step": 1164 + }, + { + "epoch": 0.0782524076373276, + "grad_norm": 1.3660764694213867, + "learning_rate": 9.997815410018111e-05, + "loss": 0.9083, + "step": 1166 + }, + { + "epoch": 0.0783866313210966, + "grad_norm": 1.1576415300369263, + "learning_rate": 9.997783167966258e-05, + "loss": 0.8664, + "step": 1168 + }, + { + "epoch": 0.07852085500486561, + "grad_norm": 3.223151445388794, + "learning_rate": 9.997750689778978e-05, + "loss": 0.8677, + "step": 1170 + }, + { + "epoch": 0.07865507868863461, + "grad_norm": 1.2480010986328125, + "learning_rate": 9.997717975457807e-05, + "loss": 0.8692, + "step": 1172 + }, + { + "epoch": 0.07878930237240361, + "grad_norm": 1.337230920791626, + "learning_rate": 9.997685025004288e-05, + "loss": 0.84, + "step": 1174 + }, + { + "epoch": 0.07892352605617262, + "grad_norm": 1.5925257205963135, + "learning_rate": 9.997651838419979e-05, + "loss": 0.828, + "step": 1176 + }, + { + "epoch": 0.07905774973994162, + "grad_norm": 1.4853055477142334, + "learning_rate": 9.997618415706448e-05, + "loss": 0.8157, + "step": 1178 + }, + { + "epoch": 0.07919197342371061, + "grad_norm": 1.3261574506759644, + "learning_rate": 9.997584756865274e-05, + "loss": 0.8429, + "step": 1180 + }, + { + "epoch": 0.07932619710747961, + "grad_norm": 1.4375333786010742, + "learning_rate": 9.997550861898049e-05, + "loss": 0.9062, + "step": 1182 + }, + { + "epoch": 0.07946042079124861, + "grad_norm": 1.9652042388916016, + "learning_rate": 9.997516730806372e-05, + "loss": 0.8545, + "step": 1184 + }, + { + "epoch": 0.07959464447501761, + "grad_norm": 1.3153257369995117, + "learning_rate": 9.997482363591857e-05, + "loss": 0.8545, + "step": 1186 + }, + { + "epoch": 0.07972886815878662, + "grad_norm": 1.7656164169311523, + "learning_rate": 9.997447760256126e-05, + "loss": 0.8648, + "step": 1188 + }, + { + "epoch": 0.07986309184255562, + "grad_norm": 1.261059284210205, + "learning_rate": 9.997412920800817e-05, + "loss": 0.7956, + "step": 1190 + }, + { + "epoch": 0.07999731552632462, + "grad_norm": 1.4606881141662598, + "learning_rate": 9.997377845227576e-05, + "loss": 0.8576, + "step": 1192 + }, + { + "epoch": 0.08013153921009362, + "grad_norm": 1.3973236083984375, + "learning_rate": 9.997342533538056e-05, + "loss": 0.8701, + "step": 1194 + }, + { + "epoch": 0.08026576289386263, + "grad_norm": 1.4433778524398804, + "learning_rate": 9.99730698573393e-05, + "loss": 0.8336, + "step": 1196 + }, + { + "epoch": 0.08039998657763163, + "grad_norm": 1.434862732887268, + "learning_rate": 9.997271201816873e-05, + "loss": 0.825, + "step": 1198 + }, + { + "epoch": 0.08053421026140062, + "grad_norm": 1.2260946035385132, + "learning_rate": 9.99723518178858e-05, + "loss": 0.7844, + "step": 1200 + }, + { + "epoch": 0.08066843394516962, + "grad_norm": 1.6296569108963013, + "learning_rate": 9.997198925650753e-05, + "loss": 0.8702, + "step": 1202 + }, + { + "epoch": 0.08080265762893862, + "grad_norm": 1.4024724960327148, + "learning_rate": 9.9971624334051e-05, + "loss": 0.8359, + "step": 1204 + }, + { + "epoch": 0.08093688131270763, + "grad_norm": 2.7965261936187744, + "learning_rate": 9.997125705053352e-05, + "loss": 0.853, + "step": 1206 + }, + { + "epoch": 0.08107110499647663, + "grad_norm": 1.1983357667922974, + "learning_rate": 9.997088740597237e-05, + "loss": 0.835, + "step": 1208 + }, + { + "epoch": 0.08120532868024563, + "grad_norm": 1.1802141666412354, + "learning_rate": 9.997051540038508e-05, + "loss": 0.8028, + "step": 1210 + }, + { + "epoch": 0.08133955236401463, + "grad_norm": 5.62499475479126, + "learning_rate": 9.997014103378921e-05, + "loss": 0.8374, + "step": 1212 + }, + { + "epoch": 0.08147377604778364, + "grad_norm": 1.3405348062515259, + "learning_rate": 9.996976430620241e-05, + "loss": 0.8063, + "step": 1214 + }, + { + "epoch": 0.08160799973155264, + "grad_norm": 1.4804438352584839, + "learning_rate": 9.996938521764254e-05, + "loss": 0.8195, + "step": 1216 + }, + { + "epoch": 0.08174222341532163, + "grad_norm": 1.225172758102417, + "learning_rate": 9.996900376812746e-05, + "loss": 0.8781, + "step": 1218 + }, + { + "epoch": 0.08187644709909063, + "grad_norm": 1.286395788192749, + "learning_rate": 9.996861995767522e-05, + "loss": 0.8469, + "step": 1220 + }, + { + "epoch": 0.08201067078285963, + "grad_norm": 1.6874159574508667, + "learning_rate": 9.996823378630393e-05, + "loss": 0.8572, + "step": 1222 + }, + { + "epoch": 0.08214489446662863, + "grad_norm": 1.407341718673706, + "learning_rate": 9.996784525403186e-05, + "loss": 0.7872, + "step": 1224 + }, + { + "epoch": 0.08227911815039764, + "grad_norm": 1.9326817989349365, + "learning_rate": 9.996745436087736e-05, + "loss": 0.8664, + "step": 1226 + }, + { + "epoch": 0.08241334183416664, + "grad_norm": 1.7948651313781738, + "learning_rate": 9.99670611068589e-05, + "loss": 0.8661, + "step": 1228 + }, + { + "epoch": 0.08254756551793564, + "grad_norm": 1.3383947610855103, + "learning_rate": 9.996666549199505e-05, + "loss": 0.8789, + "step": 1230 + }, + { + "epoch": 0.08268178920170464, + "grad_norm": 1.5959744453430176, + "learning_rate": 9.996626751630453e-05, + "loss": 0.8675, + "step": 1232 + }, + { + "epoch": 0.08281601288547365, + "grad_norm": 1.5789822340011597, + "learning_rate": 9.996586717980611e-05, + "loss": 0.8179, + "step": 1234 + }, + { + "epoch": 0.08295023656924265, + "grad_norm": 1.6039397716522217, + "learning_rate": 9.996546448251871e-05, + "loss": 0.8497, + "step": 1236 + }, + { + "epoch": 0.08308446025301164, + "grad_norm": 1.5541913509368896, + "learning_rate": 9.996505942446139e-05, + "loss": 0.8189, + "step": 1238 + }, + { + "epoch": 0.08321868393678064, + "grad_norm": 1.6467313766479492, + "learning_rate": 9.996465200565324e-05, + "loss": 0.908, + "step": 1240 + }, + { + "epoch": 0.08335290762054964, + "grad_norm": 1.3486804962158203, + "learning_rate": 9.996424222611356e-05, + "loss": 0.7749, + "step": 1242 + }, + { + "epoch": 0.08348713130431865, + "grad_norm": 1.6886804103851318, + "learning_rate": 9.996383008586165e-05, + "loss": 0.8229, + "step": 1244 + }, + { + "epoch": 0.08362135498808765, + "grad_norm": 1.831268310546875, + "learning_rate": 9.996341558491706e-05, + "loss": 0.9068, + "step": 1246 + }, + { + "epoch": 0.08375557867185665, + "grad_norm": 1.4950191974639893, + "learning_rate": 9.996299872329931e-05, + "loss": 0.8092, + "step": 1248 + }, + { + "epoch": 0.08388980235562565, + "grad_norm": 1.272981882095337, + "learning_rate": 9.996257950102811e-05, + "loss": 0.7781, + "step": 1250 + }, + { + "epoch": 0.08402402603939466, + "grad_norm": 1.3095083236694336, + "learning_rate": 9.996215791812328e-05, + "loss": 0.8071, + "step": 1252 + }, + { + "epoch": 0.08415824972316366, + "grad_norm": 1.740781545639038, + "learning_rate": 9.996173397460475e-05, + "loss": 0.8787, + "step": 1254 + }, + { + "epoch": 0.08429247340693265, + "grad_norm": 2.427889347076416, + "learning_rate": 9.996130767049252e-05, + "loss": 0.7607, + "step": 1256 + }, + { + "epoch": 0.08442669709070165, + "grad_norm": 1.7244151830673218, + "learning_rate": 9.996087900580675e-05, + "loss": 0.8009, + "step": 1258 + }, + { + "epoch": 0.08456092077447065, + "grad_norm": 2.351905107498169, + "learning_rate": 9.996044798056769e-05, + "loss": 0.8508, + "step": 1260 + }, + { + "epoch": 0.08469514445823965, + "grad_norm": 1.3471221923828125, + "learning_rate": 9.996001459479572e-05, + "loss": 0.8101, + "step": 1262 + }, + { + "epoch": 0.08482936814200866, + "grad_norm": 1.319439172744751, + "learning_rate": 9.995957884851129e-05, + "loss": 0.827, + "step": 1264 + }, + { + "epoch": 0.08496359182577766, + "grad_norm": 1.4922178983688354, + "learning_rate": 9.995914074173501e-05, + "loss": 0.7704, + "step": 1266 + }, + { + "epoch": 0.08509781550954666, + "grad_norm": 2.1654322147369385, + "learning_rate": 9.995870027448756e-05, + "loss": 0.8398, + "step": 1268 + }, + { + "epoch": 0.08523203919331566, + "grad_norm": 1.5658771991729736, + "learning_rate": 9.995825744678976e-05, + "loss": 0.8643, + "step": 1270 + }, + { + "epoch": 0.08536626287708467, + "grad_norm": 1.5930389165878296, + "learning_rate": 9.995781225866254e-05, + "loss": 0.8004, + "step": 1272 + }, + { + "epoch": 0.08550048656085366, + "grad_norm": 1.2441797256469727, + "learning_rate": 9.995736471012693e-05, + "loss": 0.8926, + "step": 1274 + }, + { + "epoch": 0.08563471024462266, + "grad_norm": 1.4061142206192017, + "learning_rate": 9.995691480120408e-05, + "loss": 0.7998, + "step": 1276 + }, + { + "epoch": 0.08576893392839166, + "grad_norm": 1.9022417068481445, + "learning_rate": 9.995646253191522e-05, + "loss": 0.8846, + "step": 1278 + }, + { + "epoch": 0.08590315761216066, + "grad_norm": 1.862823486328125, + "learning_rate": 9.995600790228176e-05, + "loss": 0.7921, + "step": 1280 + }, + { + "epoch": 0.08603738129592967, + "grad_norm": 1.2491618394851685, + "learning_rate": 9.995555091232516e-05, + "loss": 0.7846, + "step": 1282 + }, + { + "epoch": 0.08617160497969867, + "grad_norm": 1.292881727218628, + "learning_rate": 9.995509156206701e-05, + "loss": 0.8009, + "step": 1284 + }, + { + "epoch": 0.08630582866346767, + "grad_norm": 2.969083309173584, + "learning_rate": 9.995462985152902e-05, + "loss": 0.8369, + "step": 1286 + }, + { + "epoch": 0.08644005234723667, + "grad_norm": 1.9158148765563965, + "learning_rate": 9.995416578073299e-05, + "loss": 0.7613, + "step": 1288 + }, + { + "epoch": 0.08657427603100568, + "grad_norm": 1.4865663051605225, + "learning_rate": 9.995369934970085e-05, + "loss": 0.8059, + "step": 1290 + }, + { + "epoch": 0.08670849971477468, + "grad_norm": 1.2080954313278198, + "learning_rate": 9.995323055845466e-05, + "loss": 0.7653, + "step": 1292 + }, + { + "epoch": 0.08684272339854367, + "grad_norm": 1.413210391998291, + "learning_rate": 9.995275940701657e-05, + "loss": 0.8324, + "step": 1294 + }, + { + "epoch": 0.08697694708231267, + "grad_norm": 1.22736394405365, + "learning_rate": 9.995228589540881e-05, + "loss": 0.7899, + "step": 1296 + }, + { + "epoch": 0.08711117076608167, + "grad_norm": 1.8747268915176392, + "learning_rate": 9.995181002365376e-05, + "loss": 0.8409, + "step": 1298 + }, + { + "epoch": 0.08724539444985067, + "grad_norm": 1.453848123550415, + "learning_rate": 9.995133179177391e-05, + "loss": 0.7499, + "step": 1300 + }, + { + "epoch": 0.08737961813361968, + "grad_norm": 1.2807241678237915, + "learning_rate": 9.995085119979189e-05, + "loss": 0.7725, + "step": 1302 + }, + { + "epoch": 0.08751384181738868, + "grad_norm": 1.6054444313049316, + "learning_rate": 9.995036824773034e-05, + "loss": 0.7963, + "step": 1304 + }, + { + "epoch": 0.08764806550115768, + "grad_norm": 1.579632043838501, + "learning_rate": 9.994988293561213e-05, + "loss": 0.8583, + "step": 1306 + }, + { + "epoch": 0.08778228918492668, + "grad_norm": 1.4864579439163208, + "learning_rate": 9.994939526346016e-05, + "loss": 0.8355, + "step": 1308 + }, + { + "epoch": 0.08791651286869569, + "grad_norm": 1.6583774089813232, + "learning_rate": 9.99489052312975e-05, + "loss": 0.7561, + "step": 1310 + }, + { + "epoch": 0.08805073655246468, + "grad_norm": 2.079428195953369, + "learning_rate": 9.99484128391473e-05, + "loss": 0.7822, + "step": 1312 + }, + { + "epoch": 0.08818496023623368, + "grad_norm": 1.3199563026428223, + "learning_rate": 9.994791808703279e-05, + "loss": 0.7904, + "step": 1314 + }, + { + "epoch": 0.08831918392000268, + "grad_norm": 2.335811138153076, + "learning_rate": 9.994742097497737e-05, + "loss": 0.9113, + "step": 1316 + }, + { + "epoch": 0.08845340760377168, + "grad_norm": 1.2894052267074585, + "learning_rate": 9.994692150300453e-05, + "loss": 0.8324, + "step": 1318 + }, + { + "epoch": 0.08858763128754069, + "grad_norm": 1.2020514011383057, + "learning_rate": 9.994641967113787e-05, + "loss": 0.7979, + "step": 1320 + }, + { + "epoch": 0.08872185497130969, + "grad_norm": 1.4542573690414429, + "learning_rate": 9.994591547940109e-05, + "loss": 0.7408, + "step": 1322 + }, + { + "epoch": 0.08885607865507869, + "grad_norm": 1.7090483903884888, + "learning_rate": 9.994540892781802e-05, + "loss": 0.8605, + "step": 1324 + }, + { + "epoch": 0.0889903023388477, + "grad_norm": 1.3629621267318726, + "learning_rate": 9.994490001641258e-05, + "loss": 0.8248, + "step": 1326 + }, + { + "epoch": 0.0891245260226167, + "grad_norm": 1.5533065795898438, + "learning_rate": 9.994438874520885e-05, + "loss": 0.801, + "step": 1328 + }, + { + "epoch": 0.0892587497063857, + "grad_norm": 1.7974445819854736, + "learning_rate": 9.994387511423096e-05, + "loss": 0.8156, + "step": 1330 + }, + { + "epoch": 0.08939297339015469, + "grad_norm": 1.3404579162597656, + "learning_rate": 9.994335912350317e-05, + "loss": 0.8292, + "step": 1332 + }, + { + "epoch": 0.08952719707392369, + "grad_norm": 1.7092863321304321, + "learning_rate": 9.994284077304987e-05, + "loss": 0.8455, + "step": 1334 + }, + { + "epoch": 0.08966142075769269, + "grad_norm": 1.2564599514007568, + "learning_rate": 9.994232006289554e-05, + "loss": 0.8207, + "step": 1336 + }, + { + "epoch": 0.0897956444414617, + "grad_norm": 8.875020980834961, + "learning_rate": 9.994179699306483e-05, + "loss": 0.7694, + "step": 1338 + }, + { + "epoch": 0.0899298681252307, + "grad_norm": 1.4143531322479248, + "learning_rate": 9.99412715635824e-05, + "loss": 0.9383, + "step": 1340 + }, + { + "epoch": 0.0900640918089997, + "grad_norm": 1.8751009702682495, + "learning_rate": 9.994074377447309e-05, + "loss": 0.8285, + "step": 1342 + }, + { + "epoch": 0.0901983154927687, + "grad_norm": 1.322713851928711, + "learning_rate": 9.994021362576184e-05, + "loss": 0.7885, + "step": 1344 + }, + { + "epoch": 0.0903325391765377, + "grad_norm": 1.5340343713760376, + "learning_rate": 9.99396811174737e-05, + "loss": 0.8407, + "step": 1346 + }, + { + "epoch": 0.09046676286030671, + "grad_norm": 1.544365406036377, + "learning_rate": 9.993914624963383e-05, + "loss": 0.7966, + "step": 1348 + }, + { + "epoch": 0.0906009865440757, + "grad_norm": 1.9333579540252686, + "learning_rate": 9.99386090222675e-05, + "loss": 0.7956, + "step": 1350 + }, + { + "epoch": 0.0907352102278447, + "grad_norm": 1.710904836654663, + "learning_rate": 9.993806943540009e-05, + "loss": 0.8167, + "step": 1352 + }, + { + "epoch": 0.0908694339116137, + "grad_norm": 1.344147801399231, + "learning_rate": 9.993752748905712e-05, + "loss": 0.8044, + "step": 1354 + }, + { + "epoch": 0.0910036575953827, + "grad_norm": 1.5693535804748535, + "learning_rate": 9.993698318326416e-05, + "loss": 0.8063, + "step": 1356 + }, + { + "epoch": 0.0911378812791517, + "grad_norm": 1.494345784187317, + "learning_rate": 9.993643651804694e-05, + "loss": 0.8564, + "step": 1358 + }, + { + "epoch": 0.09127210496292071, + "grad_norm": 1.3548997640609741, + "learning_rate": 9.99358874934313e-05, + "loss": 0.8763, + "step": 1360 + }, + { + "epoch": 0.09140632864668971, + "grad_norm": 1.4821975231170654, + "learning_rate": 9.993533610944315e-05, + "loss": 0.8288, + "step": 1362 + }, + { + "epoch": 0.09154055233045871, + "grad_norm": 1.3371686935424805, + "learning_rate": 9.993478236610858e-05, + "loss": 0.8288, + "step": 1364 + }, + { + "epoch": 0.09167477601422772, + "grad_norm": 1.7221264839172363, + "learning_rate": 9.993422626345373e-05, + "loss": 0.8626, + "step": 1366 + }, + { + "epoch": 0.09180899969799672, + "grad_norm": 1.32262122631073, + "learning_rate": 9.993366780150488e-05, + "loss": 0.8873, + "step": 1368 + }, + { + "epoch": 0.0919432233817657, + "grad_norm": 1.3162240982055664, + "learning_rate": 9.993310698028842e-05, + "loss": 0.8074, + "step": 1370 + }, + { + "epoch": 0.09207744706553471, + "grad_norm": 1.3181543350219727, + "learning_rate": 9.993254379983084e-05, + "loss": 0.8264, + "step": 1372 + }, + { + "epoch": 0.09221167074930371, + "grad_norm": 1.367629885673523, + "learning_rate": 9.993197826015874e-05, + "loss": 0.8927, + "step": 1374 + }, + { + "epoch": 0.09234589443307271, + "grad_norm": 1.7153240442276, + "learning_rate": 9.993141036129887e-05, + "loss": 0.7883, + "step": 1376 + }, + { + "epoch": 0.09248011811684172, + "grad_norm": 1.9687062501907349, + "learning_rate": 9.993084010327804e-05, + "loss": 0.8382, + "step": 1378 + }, + { + "epoch": 0.09261434180061072, + "grad_norm": 1.3036432266235352, + "learning_rate": 9.993026748612322e-05, + "loss": 0.8184, + "step": 1380 + }, + { + "epoch": 0.09274856548437972, + "grad_norm": 1.311419129371643, + "learning_rate": 9.992969250986142e-05, + "loss": 0.863, + "step": 1382 + }, + { + "epoch": 0.09288278916814872, + "grad_norm": 1.3586655855178833, + "learning_rate": 9.992911517451985e-05, + "loss": 0.8568, + "step": 1384 + }, + { + "epoch": 0.09301701285191773, + "grad_norm": 1.15739107131958, + "learning_rate": 9.992853548012576e-05, + "loss": 0.8365, + "step": 1386 + }, + { + "epoch": 0.09315123653568672, + "grad_norm": 1.3566397428512573, + "learning_rate": 9.992795342670656e-05, + "loss": 0.7763, + "step": 1388 + }, + { + "epoch": 0.09328546021945572, + "grad_norm": 1.740161418914795, + "learning_rate": 9.992736901428971e-05, + "loss": 0.7798, + "step": 1390 + }, + { + "epoch": 0.09341968390322472, + "grad_norm": 6.3728485107421875, + "learning_rate": 9.992678224290288e-05, + "loss": 0.7612, + "step": 1392 + }, + { + "epoch": 0.09355390758699372, + "grad_norm": 1.6933761835098267, + "learning_rate": 9.992619311257376e-05, + "loss": 0.7963, + "step": 1394 + }, + { + "epoch": 0.09368813127076273, + "grad_norm": 1.3581308126449585, + "learning_rate": 9.992560162333019e-05, + "loss": 0.8041, + "step": 1396 + }, + { + "epoch": 0.09382235495453173, + "grad_norm": 1.2820748090744019, + "learning_rate": 9.992500777520011e-05, + "loss": 0.8395, + "step": 1398 + }, + { + "epoch": 0.09395657863830073, + "grad_norm": 1.3277995586395264, + "learning_rate": 9.99244115682116e-05, + "loss": 0.8477, + "step": 1400 + }, + { + "epoch": 0.09409080232206973, + "grad_norm": 1.349748134613037, + "learning_rate": 9.992381300239281e-05, + "loss": 0.8246, + "step": 1402 + }, + { + "epoch": 0.09422502600583874, + "grad_norm": 1.4513624906539917, + "learning_rate": 9.992321207777202e-05, + "loss": 0.8087, + "step": 1404 + }, + { + "epoch": 0.09435924968960774, + "grad_norm": 1.3624792098999023, + "learning_rate": 9.992260879437763e-05, + "loss": 0.7429, + "step": 1406 + }, + { + "epoch": 0.09449347337337673, + "grad_norm": 1.508120059967041, + "learning_rate": 9.992200315223815e-05, + "loss": 0.8151, + "step": 1408 + }, + { + "epoch": 0.09462769705714573, + "grad_norm": 1.3839117288589478, + "learning_rate": 9.992139515138219e-05, + "loss": 0.839, + "step": 1410 + }, + { + "epoch": 0.09476192074091473, + "grad_norm": 1.907886028289795, + "learning_rate": 9.992078479183847e-05, + "loss": 0.8181, + "step": 1412 + }, + { + "epoch": 0.09489614442468373, + "grad_norm": 1.682834506034851, + "learning_rate": 9.992017207363584e-05, + "loss": 0.8222, + "step": 1414 + }, + { + "epoch": 0.09503036810845274, + "grad_norm": 1.2854427099227905, + "learning_rate": 9.991955699680322e-05, + "loss": 0.7843, + "step": 1416 + }, + { + "epoch": 0.09516459179222174, + "grad_norm": 1.4885661602020264, + "learning_rate": 9.991893956136973e-05, + "loss": 0.8888, + "step": 1418 + }, + { + "epoch": 0.09529881547599074, + "grad_norm": 1.8255189657211304, + "learning_rate": 9.991831976736447e-05, + "loss": 0.8754, + "step": 1420 + }, + { + "epoch": 0.09543303915975974, + "grad_norm": 1.947938084602356, + "learning_rate": 9.99176976148168e-05, + "loss": 0.8436, + "step": 1422 + }, + { + "epoch": 0.09556726284352875, + "grad_norm": 1.3944926261901855, + "learning_rate": 9.991707310375604e-05, + "loss": 0.7863, + "step": 1424 + }, + { + "epoch": 0.09570148652729774, + "grad_norm": 2.084347724914551, + "learning_rate": 9.991644623421176e-05, + "loss": 0.7776, + "step": 1426 + }, + { + "epoch": 0.09583571021106674, + "grad_norm": 1.5374326705932617, + "learning_rate": 9.991581700621355e-05, + "loss": 0.82, + "step": 1428 + }, + { + "epoch": 0.09596993389483574, + "grad_norm": 1.429213523864746, + "learning_rate": 9.991518541979113e-05, + "loss": 0.8114, + "step": 1430 + }, + { + "epoch": 0.09610415757860474, + "grad_norm": 1.3940026760101318, + "learning_rate": 9.991455147497435e-05, + "loss": 0.8303, + "step": 1432 + }, + { + "epoch": 0.09623838126237375, + "grad_norm": 1.309457540512085, + "learning_rate": 9.991391517179318e-05, + "loss": 0.8646, + "step": 1434 + }, + { + "epoch": 0.09637260494614275, + "grad_norm": 1.4260421991348267, + "learning_rate": 9.991327651027765e-05, + "loss": 0.9089, + "step": 1436 + }, + { + "epoch": 0.09650682862991175, + "grad_norm": 1.6267179250717163, + "learning_rate": 9.991263549045797e-05, + "loss": 0.7937, + "step": 1438 + }, + { + "epoch": 0.09664105231368075, + "grad_norm": 1.6733784675598145, + "learning_rate": 9.991199211236442e-05, + "loss": 0.8057, + "step": 1440 + }, + { + "epoch": 0.09677527599744976, + "grad_norm": 1.1236393451690674, + "learning_rate": 9.991134637602737e-05, + "loss": 0.7769, + "step": 1442 + }, + { + "epoch": 0.09690949968121874, + "grad_norm": 1.145785927772522, + "learning_rate": 9.991069828147737e-05, + "loss": 0.7233, + "step": 1444 + }, + { + "epoch": 0.09704372336498775, + "grad_norm": 1.2535909414291382, + "learning_rate": 9.9910047828745e-05, + "loss": 0.8332, + "step": 1446 + }, + { + "epoch": 0.09717794704875675, + "grad_norm": 1.3010847568511963, + "learning_rate": 9.990939501786103e-05, + "loss": 0.761, + "step": 1448 + }, + { + "epoch": 0.09731217073252575, + "grad_norm": 2.345953941345215, + "learning_rate": 9.990873984885629e-05, + "loss": 0.8331, + "step": 1450 + }, + { + "epoch": 0.09744639441629475, + "grad_norm": 1.5649274587631226, + "learning_rate": 9.990808232176172e-05, + "loss": 0.8558, + "step": 1452 + }, + { + "epoch": 0.09758061810006376, + "grad_norm": 1.2294329404830933, + "learning_rate": 9.99074224366084e-05, + "loss": 0.7881, + "step": 1454 + }, + { + "epoch": 0.09771484178383276, + "grad_norm": 1.3104562759399414, + "learning_rate": 9.990676019342752e-05, + "loss": 0.798, + "step": 1456 + }, + { + "epoch": 0.09784906546760176, + "grad_norm": 1.2666356563568115, + "learning_rate": 9.990609559225036e-05, + "loss": 0.8305, + "step": 1458 + }, + { + "epoch": 0.09798328915137076, + "grad_norm": 1.2787644863128662, + "learning_rate": 9.990542863310831e-05, + "loss": 0.7108, + "step": 1460 + }, + { + "epoch": 0.09811751283513977, + "grad_norm": 1.3989942073822021, + "learning_rate": 9.990475931603289e-05, + "loss": 0.838, + "step": 1462 + }, + { + "epoch": 0.09825173651890876, + "grad_norm": 1.323578953742981, + "learning_rate": 9.990408764105575e-05, + "loss": 0.81, + "step": 1464 + }, + { + "epoch": 0.09838596020267776, + "grad_norm": 1.2336007356643677, + "learning_rate": 9.990341360820856e-05, + "loss": 0.8711, + "step": 1466 + }, + { + "epoch": 0.09852018388644676, + "grad_norm": 1.4644780158996582, + "learning_rate": 9.990273721752324e-05, + "loss": 0.8542, + "step": 1468 + }, + { + "epoch": 0.09865440757021576, + "grad_norm": 1.164896011352539, + "learning_rate": 9.99020584690317e-05, + "loss": 0.7895, + "step": 1470 + }, + { + "epoch": 0.09878863125398477, + "grad_norm": 2.3305513858795166, + "learning_rate": 9.990137736276604e-05, + "loss": 0.796, + "step": 1472 + }, + { + "epoch": 0.09892285493775377, + "grad_norm": 1.4515163898468018, + "learning_rate": 9.990069389875843e-05, + "loss": 0.7958, + "step": 1474 + }, + { + "epoch": 0.09905707862152277, + "grad_norm": 1.335404634475708, + "learning_rate": 9.990000807704114e-05, + "loss": 0.782, + "step": 1476 + }, + { + "epoch": 0.09919130230529177, + "grad_norm": 1.3235502243041992, + "learning_rate": 9.98993198976466e-05, + "loss": 0.8255, + "step": 1478 + }, + { + "epoch": 0.09932552598906078, + "grad_norm": 2.2413065433502197, + "learning_rate": 9.989862936060731e-05, + "loss": 0.8303, + "step": 1480 + }, + { + "epoch": 0.09945974967282976, + "grad_norm": 5.080524921417236, + "learning_rate": 9.989793646595591e-05, + "loss": 0.7998, + "step": 1482 + }, + { + "epoch": 0.09959397335659877, + "grad_norm": 1.2764524221420288, + "learning_rate": 9.989724121372514e-05, + "loss": 0.7724, + "step": 1484 + }, + { + "epoch": 0.09972819704036777, + "grad_norm": 1.6263426542282104, + "learning_rate": 9.989654360394782e-05, + "loss": 0.8485, + "step": 1486 + }, + { + "epoch": 0.09986242072413677, + "grad_norm": 1.8585193157196045, + "learning_rate": 9.989584363665696e-05, + "loss": 0.8305, + "step": 1488 + }, + { + "epoch": 0.09999664440790577, + "grad_norm": 3.082002878189087, + "learning_rate": 9.989514131188559e-05, + "loss": 0.787, + "step": 1490 + }, + { + "epoch": 0.10013086809167478, + "grad_norm": 1.289965271949768, + "learning_rate": 9.989443662966691e-05, + "loss": 0.7783, + "step": 1492 + }, + { + "epoch": 0.10026509177544378, + "grad_norm": 1.5314586162567139, + "learning_rate": 9.989372959003421e-05, + "loss": 0.8543, + "step": 1494 + }, + { + "epoch": 0.10039931545921278, + "grad_norm": 3.5703372955322266, + "learning_rate": 9.98930201930209e-05, + "loss": 0.8007, + "step": 1496 + }, + { + "epoch": 0.10053353914298178, + "grad_norm": 1.9307204484939575, + "learning_rate": 9.989230843866049e-05, + "loss": 0.824, + "step": 1498 + }, + { + "epoch": 0.10066776282675079, + "grad_norm": 2.095611810684204, + "learning_rate": 9.989159432698663e-05, + "loss": 0.7722, + "step": 1500 + }, + { + "epoch": 0.10080198651051978, + "grad_norm": 1.2669951915740967, + "learning_rate": 9.989087785803303e-05, + "loss": 0.8146, + "step": 1502 + }, + { + "epoch": 0.10093621019428878, + "grad_norm": 1.3628166913986206, + "learning_rate": 9.989015903183357e-05, + "loss": 0.7819, + "step": 1504 + }, + { + "epoch": 0.10107043387805778, + "grad_norm": 1.2423489093780518, + "learning_rate": 9.98894378484222e-05, + "loss": 0.8651, + "step": 1506 + }, + { + "epoch": 0.10120465756182678, + "grad_norm": 1.501184105873108, + "learning_rate": 9.988871430783298e-05, + "loss": 0.7944, + "step": 1508 + }, + { + "epoch": 0.10133888124559579, + "grad_norm": 1.2144012451171875, + "learning_rate": 9.988798841010012e-05, + "loss": 0.8069, + "step": 1510 + }, + { + "epoch": 0.10147310492936479, + "grad_norm": 2.4115240573883057, + "learning_rate": 9.98872601552579e-05, + "loss": 0.7664, + "step": 1512 + }, + { + "epoch": 0.10160732861313379, + "grad_norm": 2.0335206985473633, + "learning_rate": 9.988652954334076e-05, + "loss": 0.8062, + "step": 1514 + }, + { + "epoch": 0.10174155229690279, + "grad_norm": 1.4311667680740356, + "learning_rate": 9.988579657438317e-05, + "loss": 0.8448, + "step": 1516 + }, + { + "epoch": 0.1018757759806718, + "grad_norm": 3.5951297283172607, + "learning_rate": 9.988506124841981e-05, + "loss": 0.7939, + "step": 1518 + }, + { + "epoch": 0.10200999966444078, + "grad_norm": 4.994314670562744, + "learning_rate": 9.98843235654854e-05, + "loss": 0.7847, + "step": 1520 + }, + { + "epoch": 0.10214422334820979, + "grad_norm": 4.0133256912231445, + "learning_rate": 9.988358352561478e-05, + "loss": 0.7028, + "step": 1522 + }, + { + "epoch": 0.10227844703197879, + "grad_norm": 1.512060284614563, + "learning_rate": 9.988284112884294e-05, + "loss": 0.7783, + "step": 1524 + }, + { + "epoch": 0.10241267071574779, + "grad_norm": 1.9130018949508667, + "learning_rate": 9.988209637520494e-05, + "loss": 0.8547, + "step": 1526 + }, + { + "epoch": 0.1025468943995168, + "grad_norm": 1.427221417427063, + "learning_rate": 9.988134926473598e-05, + "loss": 0.8566, + "step": 1528 + }, + { + "epoch": 0.1026811180832858, + "grad_norm": 1.2133654356002808, + "learning_rate": 9.988059979747135e-05, + "loss": 0.7705, + "step": 1530 + }, + { + "epoch": 0.1028153417670548, + "grad_norm": 2.223356246948242, + "learning_rate": 9.987984797344648e-05, + "loss": 0.8361, + "step": 1532 + }, + { + "epoch": 0.1029495654508238, + "grad_norm": 1.1635984182357788, + "learning_rate": 9.987909379269686e-05, + "loss": 0.8032, + "step": 1534 + }, + { + "epoch": 0.1030837891345928, + "grad_norm": 1.4475617408752441, + "learning_rate": 9.987833725525815e-05, + "loss": 0.8738, + "step": 1536 + }, + { + "epoch": 0.1032180128183618, + "grad_norm": 1.554733157157898, + "learning_rate": 9.987757836116608e-05, + "loss": 0.7886, + "step": 1538 + }, + { + "epoch": 0.1033522365021308, + "grad_norm": 1.3741666078567505, + "learning_rate": 9.987681711045652e-05, + "loss": 0.8057, + "step": 1540 + }, + { + "epoch": 0.1034864601858998, + "grad_norm": 1.1904531717300415, + "learning_rate": 9.987605350316542e-05, + "loss": 0.8379, + "step": 1542 + }, + { + "epoch": 0.1036206838696688, + "grad_norm": 1.6170859336853027, + "learning_rate": 9.987528753932888e-05, + "loss": 0.8036, + "step": 1544 + }, + { + "epoch": 0.1037549075534378, + "grad_norm": 1.4818545579910278, + "learning_rate": 9.987451921898307e-05, + "loss": 0.7636, + "step": 1546 + }, + { + "epoch": 0.1038891312372068, + "grad_norm": 1.4961395263671875, + "learning_rate": 9.987374854216431e-05, + "loss": 0.8496, + "step": 1548 + }, + { + "epoch": 0.10402335492097581, + "grad_norm": 1.1231153011322021, + "learning_rate": 9.9872975508909e-05, + "loss": 0.7967, + "step": 1550 + }, + { + "epoch": 0.10415757860474481, + "grad_norm": 1.0882415771484375, + "learning_rate": 9.987220011925367e-05, + "loss": 0.6914, + "step": 1552 + }, + { + "epoch": 0.10429180228851381, + "grad_norm": 1.2804229259490967, + "learning_rate": 9.987142237323495e-05, + "loss": 0.7904, + "step": 1554 + }, + { + "epoch": 0.10442602597228282, + "grad_norm": 1.3405996561050415, + "learning_rate": 9.98706422708896e-05, + "loss": 0.8317, + "step": 1556 + }, + { + "epoch": 0.1045602496560518, + "grad_norm": 1.7399771213531494, + "learning_rate": 9.986985981225445e-05, + "loss": 0.7981, + "step": 1558 + }, + { + "epoch": 0.1046944733398208, + "grad_norm": 1.1339941024780273, + "learning_rate": 9.98690749973665e-05, + "loss": 0.771, + "step": 1560 + }, + { + "epoch": 0.10482869702358981, + "grad_norm": 1.4163483381271362, + "learning_rate": 9.986828782626282e-05, + "loss": 0.7263, + "step": 1562 + }, + { + "epoch": 0.10496292070735881, + "grad_norm": 1.2849148511886597, + "learning_rate": 9.986749829898061e-05, + "loss": 0.7603, + "step": 1564 + }, + { + "epoch": 0.10509714439112781, + "grad_norm": 1.2235387563705444, + "learning_rate": 9.986670641555715e-05, + "loss": 0.8495, + "step": 1566 + }, + { + "epoch": 0.10523136807489682, + "grad_norm": 1.3458445072174072, + "learning_rate": 9.986591217602988e-05, + "loss": 0.7594, + "step": 1568 + }, + { + "epoch": 0.10536559175866582, + "grad_norm": 1.1595379114151, + "learning_rate": 9.986511558043631e-05, + "loss": 0.7621, + "step": 1570 + }, + { + "epoch": 0.10549981544243482, + "grad_norm": 1.22896146774292, + "learning_rate": 9.98643166288141e-05, + "loss": 0.7538, + "step": 1572 + }, + { + "epoch": 0.10563403912620382, + "grad_norm": 1.3033969402313232, + "learning_rate": 9.986351532120097e-05, + "loss": 0.8397, + "step": 1574 + }, + { + "epoch": 0.10576826280997281, + "grad_norm": 1.3369084596633911, + "learning_rate": 9.98627116576348e-05, + "loss": 0.7519, + "step": 1576 + }, + { + "epoch": 0.10590248649374182, + "grad_norm": 1.3159362077713013, + "learning_rate": 9.986190563815355e-05, + "loss": 0.7714, + "step": 1578 + }, + { + "epoch": 0.10603671017751082, + "grad_norm": 1.2473654747009277, + "learning_rate": 9.986109726279531e-05, + "loss": 0.7457, + "step": 1580 + }, + { + "epoch": 0.10617093386127982, + "grad_norm": 1.3414995670318604, + "learning_rate": 9.986028653159826e-05, + "loss": 0.8176, + "step": 1582 + }, + { + "epoch": 0.10630515754504882, + "grad_norm": 3.310617446899414, + "learning_rate": 9.985947344460074e-05, + "loss": 0.8616, + "step": 1584 + }, + { + "epoch": 0.10643938122881783, + "grad_norm": 1.5509215593338013, + "learning_rate": 9.985865800184113e-05, + "loss": 0.8301, + "step": 1586 + }, + { + "epoch": 0.10657360491258683, + "grad_norm": 1.2767397165298462, + "learning_rate": 9.985784020335798e-05, + "loss": 0.8217, + "step": 1588 + }, + { + "epoch": 0.10670782859635583, + "grad_norm": 1.5381598472595215, + "learning_rate": 9.985702004918992e-05, + "loss": 0.83, + "step": 1590 + }, + { + "epoch": 0.10684205228012483, + "grad_norm": 6.233184337615967, + "learning_rate": 9.98561975393757e-05, + "loss": 0.8475, + "step": 1592 + }, + { + "epoch": 0.10697627596389384, + "grad_norm": 2.839657783508301, + "learning_rate": 9.985537267395418e-05, + "loss": 0.7681, + "step": 1594 + }, + { + "epoch": 0.10711049964766282, + "grad_norm": 1.9623489379882812, + "learning_rate": 9.985454545296434e-05, + "loss": 0.8005, + "step": 1596 + }, + { + "epoch": 0.10724472333143183, + "grad_norm": 1.5133533477783203, + "learning_rate": 9.985371587644526e-05, + "loss": 0.868, + "step": 1598 + }, + { + "epoch": 0.10737894701520083, + "grad_norm": 1.517056941986084, + "learning_rate": 9.985288394443615e-05, + "loss": 0.803, + "step": 1600 + }, + { + "epoch": 0.10751317069896983, + "grad_norm": 1.431883454322815, + "learning_rate": 9.98520496569763e-05, + "loss": 0.7725, + "step": 1602 + }, + { + "epoch": 0.10764739438273883, + "grad_norm": 2.0910532474517822, + "learning_rate": 9.985121301410511e-05, + "loss": 0.7608, + "step": 1604 + }, + { + "epoch": 0.10778161806650784, + "grad_norm": 1.8099147081375122, + "learning_rate": 9.985037401586217e-05, + "loss": 0.9247, + "step": 1606 + }, + { + "epoch": 0.10791584175027684, + "grad_norm": 1.9033899307250977, + "learning_rate": 9.984953266228707e-05, + "loss": 0.821, + "step": 1608 + }, + { + "epoch": 0.10805006543404584, + "grad_norm": 1.296483039855957, + "learning_rate": 9.984868895341957e-05, + "loss": 0.7958, + "step": 1610 + }, + { + "epoch": 0.10818428911781484, + "grad_norm": 1.6510436534881592, + "learning_rate": 9.984784288929953e-05, + "loss": 0.769, + "step": 1612 + }, + { + "epoch": 0.10831851280158383, + "grad_norm": 1.635801911354065, + "learning_rate": 9.984699446996697e-05, + "loss": 0.8483, + "step": 1614 + }, + { + "epoch": 0.10845273648535284, + "grad_norm": 1.598320484161377, + "learning_rate": 9.984614369546191e-05, + "loss": 0.8505, + "step": 1616 + }, + { + "epoch": 0.10858696016912184, + "grad_norm": 1.5453705787658691, + "learning_rate": 9.984529056582459e-05, + "loss": 0.8062, + "step": 1618 + }, + { + "epoch": 0.10872118385289084, + "grad_norm": 1.5651856660842896, + "learning_rate": 9.984443508109531e-05, + "loss": 0.7983, + "step": 1620 + }, + { + "epoch": 0.10885540753665984, + "grad_norm": 1.8090884685516357, + "learning_rate": 9.984357724131448e-05, + "loss": 0.8308, + "step": 1622 + }, + { + "epoch": 0.10898963122042885, + "grad_norm": 1.8616065979003906, + "learning_rate": 9.984271704652263e-05, + "loss": 0.7526, + "step": 1624 + }, + { + "epoch": 0.10912385490419785, + "grad_norm": 1.312053918838501, + "learning_rate": 9.984185449676044e-05, + "loss": 0.7644, + "step": 1626 + }, + { + "epoch": 0.10925807858796685, + "grad_norm": 1.403348445892334, + "learning_rate": 9.984098959206863e-05, + "loss": 0.7711, + "step": 1628 + }, + { + "epoch": 0.10939230227173585, + "grad_norm": 1.3508669137954712, + "learning_rate": 9.984012233248805e-05, + "loss": 0.7732, + "step": 1630 + }, + { + "epoch": 0.10952652595550486, + "grad_norm": 1.1792352199554443, + "learning_rate": 9.98392527180597e-05, + "loss": 0.7759, + "step": 1632 + }, + { + "epoch": 0.10966074963927384, + "grad_norm": 1.2566429376602173, + "learning_rate": 9.983838074882467e-05, + "loss": 0.8051, + "step": 1634 + }, + { + "epoch": 0.10979497332304285, + "grad_norm": 1.272409200668335, + "learning_rate": 9.983750642482414e-05, + "loss": 0.7872, + "step": 1636 + }, + { + "epoch": 0.10992919700681185, + "grad_norm": 1.3588621616363525, + "learning_rate": 9.983662974609945e-05, + "loss": 0.8099, + "step": 1638 + }, + { + "epoch": 0.11006342069058085, + "grad_norm": 1.3502743244171143, + "learning_rate": 9.9835750712692e-05, + "loss": 0.7576, + "step": 1640 + }, + { + "epoch": 0.11019764437434985, + "grad_norm": 1.369414210319519, + "learning_rate": 9.983486932464332e-05, + "loss": 0.7747, + "step": 1642 + }, + { + "epoch": 0.11033186805811886, + "grad_norm": 1.192598581314087, + "learning_rate": 9.983398558199506e-05, + "loss": 0.7855, + "step": 1644 + }, + { + "epoch": 0.11046609174188786, + "grad_norm": 1.53266441822052, + "learning_rate": 9.983309948478898e-05, + "loss": 0.8317, + "step": 1646 + }, + { + "epoch": 0.11060031542565686, + "grad_norm": 1.2673954963684082, + "learning_rate": 9.983221103306695e-05, + "loss": 0.7258, + "step": 1648 + }, + { + "epoch": 0.11073453910942586, + "grad_norm": 1.2609288692474365, + "learning_rate": 9.983132022687093e-05, + "loss": 0.7564, + "step": 1650 + }, + { + "epoch": 0.11086876279319485, + "grad_norm": 1.4011356830596924, + "learning_rate": 9.983042706624302e-05, + "loss": 0.8574, + "step": 1652 + }, + { + "epoch": 0.11100298647696386, + "grad_norm": 3.235201835632324, + "learning_rate": 9.982953155122542e-05, + "loss": 0.7768, + "step": 1654 + }, + { + "epoch": 0.11113721016073286, + "grad_norm": 1.3096420764923096, + "learning_rate": 9.982863368186044e-05, + "loss": 0.7925, + "step": 1656 + }, + { + "epoch": 0.11127143384450186, + "grad_norm": 1.4204448461532593, + "learning_rate": 9.98277334581905e-05, + "loss": 0.7952, + "step": 1658 + }, + { + "epoch": 0.11140565752827086, + "grad_norm": 1.3249876499176025, + "learning_rate": 9.982683088025813e-05, + "loss": 0.7743, + "step": 1660 + }, + { + "epoch": 0.11153988121203987, + "grad_norm": 1.5150166749954224, + "learning_rate": 9.982592594810599e-05, + "loss": 0.8212, + "step": 1662 + }, + { + "epoch": 0.11167410489580887, + "grad_norm": 1.3663066625595093, + "learning_rate": 9.982501866177682e-05, + "loss": 0.8411, + "step": 1664 + }, + { + "epoch": 0.11180832857957787, + "grad_norm": 1.3154817819595337, + "learning_rate": 9.98241090213135e-05, + "loss": 0.7834, + "step": 1666 + }, + { + "epoch": 0.11194255226334687, + "grad_norm": 1.3979624509811401, + "learning_rate": 9.982319702675901e-05, + "loss": 0.8053, + "step": 1668 + }, + { + "epoch": 0.11207677594711588, + "grad_norm": 1.3888903856277466, + "learning_rate": 9.982228267815643e-05, + "loss": 0.7757, + "step": 1670 + }, + { + "epoch": 0.11221099963088486, + "grad_norm": 1.4778214693069458, + "learning_rate": 9.982136597554896e-05, + "loss": 0.8161, + "step": 1672 + }, + { + "epoch": 0.11234522331465387, + "grad_norm": 1.4513918161392212, + "learning_rate": 9.982044691897991e-05, + "loss": 0.7305, + "step": 1674 + }, + { + "epoch": 0.11247944699842287, + "grad_norm": 1.3353772163391113, + "learning_rate": 9.981952550849273e-05, + "loss": 0.7575, + "step": 1676 + }, + { + "epoch": 0.11261367068219187, + "grad_norm": 1.2371762990951538, + "learning_rate": 9.981860174413092e-05, + "loss": 0.819, + "step": 1678 + }, + { + "epoch": 0.11274789436596087, + "grad_norm": 3.9118690490722656, + "learning_rate": 9.981767562593815e-05, + "loss": 0.6956, + "step": 1680 + }, + { + "epoch": 0.11288211804972988, + "grad_norm": 1.2148282527923584, + "learning_rate": 9.981674715395816e-05, + "loss": 0.7897, + "step": 1682 + }, + { + "epoch": 0.11301634173349888, + "grad_norm": 1.2970951795578003, + "learning_rate": 9.981581632823485e-05, + "loss": 0.8284, + "step": 1684 + }, + { + "epoch": 0.11315056541726788, + "grad_norm": 1.3304328918457031, + "learning_rate": 9.981488314881215e-05, + "loss": 0.8311, + "step": 1686 + }, + { + "epoch": 0.11328478910103688, + "grad_norm": 1.2172995805740356, + "learning_rate": 9.981394761573419e-05, + "loss": 0.7544, + "step": 1688 + }, + { + "epoch": 0.11341901278480587, + "grad_norm": 1.2176858186721802, + "learning_rate": 9.981300972904515e-05, + "loss": 0.8118, + "step": 1690 + }, + { + "epoch": 0.11355323646857488, + "grad_norm": 1.348036527633667, + "learning_rate": 9.981206948878937e-05, + "loss": 0.7689, + "step": 1692 + }, + { + "epoch": 0.11368746015234388, + "grad_norm": 1.1857105493545532, + "learning_rate": 9.981112689501126e-05, + "loss": 0.8124, + "step": 1694 + }, + { + "epoch": 0.11382168383611288, + "grad_norm": 1.2678489685058594, + "learning_rate": 9.981018194775533e-05, + "loss": 0.8076, + "step": 1696 + }, + { + "epoch": 0.11395590751988188, + "grad_norm": 1.3750667572021484, + "learning_rate": 9.980923464706627e-05, + "loss": 0.8819, + "step": 1698 + }, + { + "epoch": 0.11409013120365089, + "grad_norm": 1.183547854423523, + "learning_rate": 9.980828499298882e-05, + "loss": 0.8644, + "step": 1700 + }, + { + "epoch": 0.11422435488741989, + "grad_norm": 1.2928284406661987, + "learning_rate": 9.980733298556783e-05, + "loss": 0.8045, + "step": 1702 + }, + { + "epoch": 0.11435857857118889, + "grad_norm": 1.5439093112945557, + "learning_rate": 9.980637862484832e-05, + "loss": 0.8166, + "step": 1704 + }, + { + "epoch": 0.11449280225495789, + "grad_norm": 1.4209173917770386, + "learning_rate": 9.980542191087535e-05, + "loss": 0.8048, + "step": 1706 + }, + { + "epoch": 0.1146270259387269, + "grad_norm": 2.5467655658721924, + "learning_rate": 9.980446284369413e-05, + "loss": 0.7793, + "step": 1708 + }, + { + "epoch": 0.11476124962249588, + "grad_norm": 1.5948156118392944, + "learning_rate": 9.980350142334998e-05, + "loss": 0.797, + "step": 1710 + }, + { + "epoch": 0.11489547330626489, + "grad_norm": 1.2480472326278687, + "learning_rate": 9.980253764988832e-05, + "loss": 0.8435, + "step": 1712 + }, + { + "epoch": 0.11502969699003389, + "grad_norm": 1.4319919347763062, + "learning_rate": 9.980157152335467e-05, + "loss": 0.8062, + "step": 1714 + }, + { + "epoch": 0.11516392067380289, + "grad_norm": 1.3824477195739746, + "learning_rate": 9.980060304379472e-05, + "loss": 0.7959, + "step": 1716 + }, + { + "epoch": 0.1152981443575719, + "grad_norm": 1.350187063217163, + "learning_rate": 9.979963221125421e-05, + "loss": 0.73, + "step": 1718 + }, + { + "epoch": 0.1154323680413409, + "grad_norm": 1.605605125427246, + "learning_rate": 9.9798659025779e-05, + "loss": 0.7791, + "step": 1720 + }, + { + "epoch": 0.1155665917251099, + "grad_norm": 1.3328030109405518, + "learning_rate": 9.979768348741507e-05, + "loss": 0.7717, + "step": 1722 + }, + { + "epoch": 0.1157008154088789, + "grad_norm": 1.287878155708313, + "learning_rate": 9.979670559620851e-05, + "loss": 0.7704, + "step": 1724 + }, + { + "epoch": 0.1158350390926479, + "grad_norm": 1.2989848852157593, + "learning_rate": 9.979572535220555e-05, + "loss": 0.7726, + "step": 1726 + }, + { + "epoch": 0.11596926277641689, + "grad_norm": 1.447620153427124, + "learning_rate": 9.979474275545248e-05, + "loss": 0.7897, + "step": 1728 + }, + { + "epoch": 0.1161034864601859, + "grad_norm": 1.369988203048706, + "learning_rate": 9.979375780599573e-05, + "loss": 0.7513, + "step": 1730 + }, + { + "epoch": 0.1162377101439549, + "grad_norm": 1.8917629718780518, + "learning_rate": 9.979277050388183e-05, + "loss": 0.7262, + "step": 1732 + }, + { + "epoch": 0.1163719338277239, + "grad_norm": 1.0357433557510376, + "learning_rate": 9.979178084915745e-05, + "loss": 0.7571, + "step": 1734 + }, + { + "epoch": 0.1165061575114929, + "grad_norm": 1.284207820892334, + "learning_rate": 9.979078884186933e-05, + "loss": 0.7646, + "step": 1736 + }, + { + "epoch": 0.1166403811952619, + "grad_norm": 1.2012271881103516, + "learning_rate": 9.978979448206434e-05, + "loss": 0.8242, + "step": 1738 + }, + { + "epoch": 0.11677460487903091, + "grad_norm": 1.6155526638031006, + "learning_rate": 9.978879776978949e-05, + "loss": 0.8832, + "step": 1740 + }, + { + "epoch": 0.11690882856279991, + "grad_norm": 1.3117786645889282, + "learning_rate": 9.978779870509182e-05, + "loss": 0.7964, + "step": 1742 + }, + { + "epoch": 0.11704305224656891, + "grad_norm": 1.2522165775299072, + "learning_rate": 9.978679728801859e-05, + "loss": 0.8288, + "step": 1744 + }, + { + "epoch": 0.1171772759303379, + "grad_norm": 1.4759392738342285, + "learning_rate": 9.978579351861707e-05, + "loss": 0.815, + "step": 1746 + }, + { + "epoch": 0.1173114996141069, + "grad_norm": 1.2629553079605103, + "learning_rate": 9.978478739693473e-05, + "loss": 0.7777, + "step": 1748 + }, + { + "epoch": 0.1174457232978759, + "grad_norm": 1.3460296392440796, + "learning_rate": 9.978377892301906e-05, + "loss": 0.7317, + "step": 1750 + }, + { + "epoch": 0.11757994698164491, + "grad_norm": 1.3826546669006348, + "learning_rate": 9.978276809691776e-05, + "loss": 0.822, + "step": 1752 + }, + { + "epoch": 0.11771417066541391, + "grad_norm": 1.285803198814392, + "learning_rate": 9.978175491867854e-05, + "loss": 0.7972, + "step": 1754 + }, + { + "epoch": 0.11784839434918291, + "grad_norm": 1.267152190208435, + "learning_rate": 9.978073938834929e-05, + "loss": 0.7369, + "step": 1756 + }, + { + "epoch": 0.11798261803295192, + "grad_norm": 1.4843337535858154, + "learning_rate": 9.977972150597799e-05, + "loss": 0.8302, + "step": 1758 + }, + { + "epoch": 0.11811684171672092, + "grad_norm": 1.3544261455535889, + "learning_rate": 9.977870127161275e-05, + "loss": 0.7733, + "step": 1760 + }, + { + "epoch": 0.11825106540048992, + "grad_norm": 1.9223833084106445, + "learning_rate": 9.977767868530176e-05, + "loss": 0.8213, + "step": 1762 + }, + { + "epoch": 0.11838528908425892, + "grad_norm": 1.3144159317016602, + "learning_rate": 9.977665374709333e-05, + "loss": 0.8232, + "step": 1764 + }, + { + "epoch": 0.11851951276802791, + "grad_norm": 1.1810052394866943, + "learning_rate": 9.977562645703589e-05, + "loss": 0.7614, + "step": 1766 + }, + { + "epoch": 0.11865373645179692, + "grad_norm": 1.575556993484497, + "learning_rate": 9.977459681517798e-05, + "loss": 0.8387, + "step": 1768 + }, + { + "epoch": 0.11878796013556592, + "grad_norm": 1.113950252532959, + "learning_rate": 9.977356482156825e-05, + "loss": 0.7608, + "step": 1770 + }, + { + "epoch": 0.11892218381933492, + "grad_norm": 1.2446372509002686, + "learning_rate": 9.977253047625546e-05, + "loss": 0.7678, + "step": 1772 + }, + { + "epoch": 0.11905640750310392, + "grad_norm": 1.3604047298431396, + "learning_rate": 9.977149377928847e-05, + "loss": 0.7857, + "step": 1774 + }, + { + "epoch": 0.11919063118687293, + "grad_norm": 1.0783966779708862, + "learning_rate": 9.977045473071627e-05, + "loss": 0.7241, + "step": 1776 + }, + { + "epoch": 0.11932485487064193, + "grad_norm": 1.2782478332519531, + "learning_rate": 9.976941333058796e-05, + "loss": 0.8046, + "step": 1778 + }, + { + "epoch": 0.11945907855441093, + "grad_norm": 1.35762619972229, + "learning_rate": 9.976836957895275e-05, + "loss": 0.8035, + "step": 1780 + }, + { + "epoch": 0.11959330223817993, + "grad_norm": 1.316946268081665, + "learning_rate": 9.976732347585993e-05, + "loss": 0.8468, + "step": 1782 + }, + { + "epoch": 0.11972752592194892, + "grad_norm": 1.3204083442687988, + "learning_rate": 9.976627502135894e-05, + "loss": 0.8752, + "step": 1784 + }, + { + "epoch": 0.11986174960571792, + "grad_norm": 1.148432970046997, + "learning_rate": 9.976522421549932e-05, + "loss": 0.6699, + "step": 1786 + }, + { + "epoch": 0.11999597328948693, + "grad_norm": 1.3192583322525024, + "learning_rate": 9.97641710583307e-05, + "loss": 0.7993, + "step": 1788 + }, + { + "epoch": 0.12013019697325593, + "grad_norm": 1.2164489030838013, + "learning_rate": 9.976311554990287e-05, + "loss": 0.7252, + "step": 1790 + }, + { + "epoch": 0.12026442065702493, + "grad_norm": 1.1925063133239746, + "learning_rate": 9.976205769026568e-05, + "loss": 0.7562, + "step": 1792 + }, + { + "epoch": 0.12039864434079393, + "grad_norm": 1.2228387594223022, + "learning_rate": 9.976099747946912e-05, + "loss": 0.7322, + "step": 1794 + }, + { + "epoch": 0.12053286802456294, + "grad_norm": 1.2622096538543701, + "learning_rate": 9.975993491756328e-05, + "loss": 0.8457, + "step": 1796 + }, + { + "epoch": 0.12066709170833194, + "grad_norm": 1.1262799501419067, + "learning_rate": 9.975887000459835e-05, + "loss": 0.8479, + "step": 1798 + }, + { + "epoch": 0.12080131539210094, + "grad_norm": 34.44312286376953, + "learning_rate": 9.975780274062468e-05, + "loss": 0.765, + "step": 1800 + }, + { + "epoch": 0.12093553907586994, + "grad_norm": 1.2758334875106812, + "learning_rate": 9.975673312569267e-05, + "loss": 0.7277, + "step": 1802 + }, + { + "epoch": 0.12106976275963893, + "grad_norm": 1.367239236831665, + "learning_rate": 9.975566115985284e-05, + "loss": 0.7898, + "step": 1804 + }, + { + "epoch": 0.12120398644340794, + "grad_norm": 1.4623262882232666, + "learning_rate": 9.975458684315588e-05, + "loss": 0.7649, + "step": 1806 + }, + { + "epoch": 0.12133821012717694, + "grad_norm": 2.027376651763916, + "learning_rate": 9.975351017565253e-05, + "loss": 0.7479, + "step": 1808 + }, + { + "epoch": 0.12147243381094594, + "grad_norm": 1.2474169731140137, + "learning_rate": 9.975243115739366e-05, + "loss": 0.7568, + "step": 1810 + }, + { + "epoch": 0.12160665749471494, + "grad_norm": 1.9691351652145386, + "learning_rate": 9.975134978843026e-05, + "loss": 0.8102, + "step": 1812 + }, + { + "epoch": 0.12174088117848395, + "grad_norm": 1.175947666168213, + "learning_rate": 9.97502660688134e-05, + "loss": 0.7581, + "step": 1814 + }, + { + "epoch": 0.12187510486225295, + "grad_norm": 1.529521107673645, + "learning_rate": 9.97491799985943e-05, + "loss": 0.8436, + "step": 1816 + }, + { + "epoch": 0.12200932854602195, + "grad_norm": 1.532962441444397, + "learning_rate": 9.974809157782427e-05, + "loss": 0.7984, + "step": 1818 + }, + { + "epoch": 0.12214355222979095, + "grad_norm": 1.0973433256149292, + "learning_rate": 9.974700080655475e-05, + "loss": 0.7525, + "step": 1820 + }, + { + "epoch": 0.12227777591355994, + "grad_norm": 1.2945022583007812, + "learning_rate": 9.974590768483725e-05, + "loss": 0.767, + "step": 1822 + }, + { + "epoch": 0.12241199959732894, + "grad_norm": 1.1806293725967407, + "learning_rate": 9.974481221272345e-05, + "loss": 0.7899, + "step": 1824 + }, + { + "epoch": 0.12254622328109795, + "grad_norm": 1.2637357711791992, + "learning_rate": 9.974371439026508e-05, + "loss": 0.7686, + "step": 1826 + }, + { + "epoch": 0.12268044696486695, + "grad_norm": 1.3107693195343018, + "learning_rate": 9.974261421751403e-05, + "loss": 0.7571, + "step": 1828 + }, + { + "epoch": 0.12281467064863595, + "grad_norm": 1.2245479822158813, + "learning_rate": 9.974151169452226e-05, + "loss": 0.8113, + "step": 1830 + }, + { + "epoch": 0.12294889433240495, + "grad_norm": 1.2956514358520508, + "learning_rate": 9.974040682134189e-05, + "loss": 0.8173, + "step": 1832 + }, + { + "epoch": 0.12308311801617396, + "grad_norm": 1.2990756034851074, + "learning_rate": 9.97392995980251e-05, + "loss": 0.8011, + "step": 1834 + }, + { + "epoch": 0.12321734169994296, + "grad_norm": 1.2273094654083252, + "learning_rate": 9.973819002462421e-05, + "loss": 0.8189, + "step": 1836 + }, + { + "epoch": 0.12335156538371196, + "grad_norm": 1.1717866659164429, + "learning_rate": 9.973707810119165e-05, + "loss": 0.759, + "step": 1838 + }, + { + "epoch": 0.12348578906748096, + "grad_norm": 1.330411434173584, + "learning_rate": 9.973596382777995e-05, + "loss": 0.8096, + "step": 1840 + }, + { + "epoch": 0.12362001275124995, + "grad_norm": 1.3452109098434448, + "learning_rate": 9.973484720444178e-05, + "loss": 0.7882, + "step": 1842 + }, + { + "epoch": 0.12375423643501895, + "grad_norm": 1.2101787328720093, + "learning_rate": 9.973372823122985e-05, + "loss": 0.8313, + "step": 1844 + }, + { + "epoch": 0.12388846011878796, + "grad_norm": 1.41432523727417, + "learning_rate": 9.973260690819708e-05, + "loss": 0.7717, + "step": 1846 + }, + { + "epoch": 0.12402268380255696, + "grad_norm": 1.2841819524765015, + "learning_rate": 9.973148323539641e-05, + "loss": 0.7722, + "step": 1848 + }, + { + "epoch": 0.12415690748632596, + "grad_norm": 1.3827474117279053, + "learning_rate": 9.973035721288096e-05, + "loss": 0.7811, + "step": 1850 + }, + { + "epoch": 0.12429113117009497, + "grad_norm": 1.5363903045654297, + "learning_rate": 9.972922884070392e-05, + "loss": 0.752, + "step": 1852 + }, + { + "epoch": 0.12442535485386397, + "grad_norm": 1.2475894689559937, + "learning_rate": 9.972809811891861e-05, + "loss": 0.7954, + "step": 1854 + }, + { + "epoch": 0.12455957853763297, + "grad_norm": 1.1693847179412842, + "learning_rate": 9.972696504757846e-05, + "loss": 0.8019, + "step": 1856 + }, + { + "epoch": 0.12469380222140197, + "grad_norm": 1.9490900039672852, + "learning_rate": 9.972582962673698e-05, + "loss": 0.7501, + "step": 1858 + }, + { + "epoch": 0.12482802590517096, + "grad_norm": 2.058365821838379, + "learning_rate": 9.972469185644783e-05, + "loss": 0.7717, + "step": 1860 + }, + { + "epoch": 0.12496224958893996, + "grad_norm": 2.113640069961548, + "learning_rate": 9.972355173676478e-05, + "loss": 0.8192, + "step": 1862 + }, + { + "epoch": 0.12509647327270898, + "grad_norm": 1.2641154527664185, + "learning_rate": 9.972240926774168e-05, + "loss": 0.7613, + "step": 1864 + }, + { + "epoch": 0.12523069695647798, + "grad_norm": 1.1489163637161255, + "learning_rate": 9.972126444943252e-05, + "loss": 0.7567, + "step": 1866 + }, + { + "epoch": 0.12536492064024699, + "grad_norm": 1.123280644416809, + "learning_rate": 9.97201172818914e-05, + "loss": 0.7765, + "step": 1868 + }, + { + "epoch": 0.12549914432401596, + "grad_norm": 1.388689637184143, + "learning_rate": 9.97189677651725e-05, + "loss": 0.8062, + "step": 1870 + }, + { + "epoch": 0.12563336800778496, + "grad_norm": 1.3234788179397583, + "learning_rate": 9.971781589933012e-05, + "loss": 0.7392, + "step": 1872 + }, + { + "epoch": 0.12576759169155396, + "grad_norm": 1.1948199272155762, + "learning_rate": 9.971666168441872e-05, + "loss": 0.8034, + "step": 1874 + }, + { + "epoch": 0.12590181537532297, + "grad_norm": 2.0882232189178467, + "learning_rate": 9.971550512049281e-05, + "loss": 0.7683, + "step": 1876 + }, + { + "epoch": 0.12603603905909197, + "grad_norm": 1.1946005821228027, + "learning_rate": 9.971434620760707e-05, + "loss": 0.7405, + "step": 1878 + }, + { + "epoch": 0.12617026274286097, + "grad_norm": 1.5380767583847046, + "learning_rate": 9.97131849458162e-05, + "loss": 0.7712, + "step": 1880 + }, + { + "epoch": 0.12630448642662997, + "grad_norm": 1.0007857084274292, + "learning_rate": 9.971202133517512e-05, + "loss": 0.7146, + "step": 1882 + }, + { + "epoch": 0.12643871011039898, + "grad_norm": 1.260170578956604, + "learning_rate": 9.971085537573879e-05, + "loss": 0.8351, + "step": 1884 + }, + { + "epoch": 0.12657293379416798, + "grad_norm": 1.1478792428970337, + "learning_rate": 9.970968706756227e-05, + "loss": 0.7442, + "step": 1886 + }, + { + "epoch": 0.12670715747793698, + "grad_norm": 1.2533572912216187, + "learning_rate": 9.970851641070081e-05, + "loss": 0.8, + "step": 1888 + }, + { + "epoch": 0.12684138116170598, + "grad_norm": 1.5237817764282227, + "learning_rate": 9.970734340520969e-05, + "loss": 0.7082, + "step": 1890 + }, + { + "epoch": 0.126975604845475, + "grad_norm": 1.4419610500335693, + "learning_rate": 9.970616805114434e-05, + "loss": 0.801, + "step": 1892 + }, + { + "epoch": 0.127109828529244, + "grad_norm": 1.7004404067993164, + "learning_rate": 9.970499034856029e-05, + "loss": 0.8262, + "step": 1894 + }, + { + "epoch": 0.127244052213013, + "grad_norm": 1.3557778596878052, + "learning_rate": 9.970381029751319e-05, + "loss": 0.8065, + "step": 1896 + }, + { + "epoch": 0.127378275896782, + "grad_norm": 1.2651209831237793, + "learning_rate": 9.970262789805878e-05, + "loss": 0.7901, + "step": 1898 + }, + { + "epoch": 0.127512499580551, + "grad_norm": 1.9335846900939941, + "learning_rate": 9.970144315025296e-05, + "loss": 0.8032, + "step": 1900 + }, + { + "epoch": 0.12764672326432, + "grad_norm": 3.0118868350982666, + "learning_rate": 9.970025605415166e-05, + "loss": 0.7242, + "step": 1902 + }, + { + "epoch": 0.127780946948089, + "grad_norm": 1.3224263191223145, + "learning_rate": 9.9699066609811e-05, + "loss": 0.8135, + "step": 1904 + }, + { + "epoch": 0.127915170631858, + "grad_norm": 1.142478108406067, + "learning_rate": 9.969787481728718e-05, + "loss": 0.7557, + "step": 1906 + }, + { + "epoch": 0.12804939431562698, + "grad_norm": 1.0734022855758667, + "learning_rate": 9.969668067663652e-05, + "loss": 0.7483, + "step": 1908 + }, + { + "epoch": 0.12818361799939598, + "grad_norm": 1.2380741834640503, + "learning_rate": 9.969548418791539e-05, + "loss": 0.7703, + "step": 1910 + }, + { + "epoch": 0.12831784168316498, + "grad_norm": 1.385408639907837, + "learning_rate": 9.969428535118036e-05, + "loss": 0.7688, + "step": 1912 + }, + { + "epoch": 0.128452065366934, + "grad_norm": 1.2676340341567993, + "learning_rate": 9.969308416648807e-05, + "loss": 0.7792, + "step": 1914 + }, + { + "epoch": 0.128586289050703, + "grad_norm": 1.098507046699524, + "learning_rate": 9.969188063389528e-05, + "loss": 0.8204, + "step": 1916 + }, + { + "epoch": 0.128720512734472, + "grad_norm": 1.145999550819397, + "learning_rate": 9.969067475345884e-05, + "loss": 0.7539, + "step": 1918 + }, + { + "epoch": 0.128854736418241, + "grad_norm": 1.6709622144699097, + "learning_rate": 9.968946652523572e-05, + "loss": 0.7803, + "step": 1920 + }, + { + "epoch": 0.12898896010201, + "grad_norm": 1.552955985069275, + "learning_rate": 9.968825594928302e-05, + "loss": 0.8612, + "step": 1922 + }, + { + "epoch": 0.129123183785779, + "grad_norm": 1.707472324371338, + "learning_rate": 9.968704302565794e-05, + "loss": 0.78, + "step": 1924 + }, + { + "epoch": 0.129257407469548, + "grad_norm": 1.2199907302856445, + "learning_rate": 9.968582775441778e-05, + "loss": 0.7672, + "step": 1926 + }, + { + "epoch": 0.129391631153317, + "grad_norm": 1.230023741722107, + "learning_rate": 9.968461013561995e-05, + "loss": 0.7606, + "step": 1928 + }, + { + "epoch": 0.129525854837086, + "grad_norm": 1.389344334602356, + "learning_rate": 9.968339016932202e-05, + "loss": 0.7466, + "step": 1930 + }, + { + "epoch": 0.129660078520855, + "grad_norm": 2.679879665374756, + "learning_rate": 9.968216785558158e-05, + "loss": 0.7859, + "step": 1932 + }, + { + "epoch": 0.129794302204624, + "grad_norm": 1.282226324081421, + "learning_rate": 9.968094319445642e-05, + "loss": 0.7685, + "step": 1934 + }, + { + "epoch": 0.12992852588839301, + "grad_norm": 1.1735801696777344, + "learning_rate": 9.967971618600437e-05, + "loss": 0.7154, + "step": 1936 + }, + { + "epoch": 0.13006274957216202, + "grad_norm": 1.2819246053695679, + "learning_rate": 9.967848683028343e-05, + "loss": 0.8303, + "step": 1938 + }, + { + "epoch": 0.13019697325593102, + "grad_norm": 1.1663737297058105, + "learning_rate": 9.967725512735169e-05, + "loss": 0.6946, + "step": 1940 + }, + { + "epoch": 0.13033119693970002, + "grad_norm": 1.334855556488037, + "learning_rate": 9.96760210772673e-05, + "loss": 0.7953, + "step": 1942 + }, + { + "epoch": 0.130465420623469, + "grad_norm": 1.795897126197815, + "learning_rate": 9.967478468008861e-05, + "loss": 0.8466, + "step": 1944 + }, + { + "epoch": 0.130599644307238, + "grad_norm": 1.366719126701355, + "learning_rate": 9.967354593587403e-05, + "loss": 0.8178, + "step": 1946 + }, + { + "epoch": 0.130733867991007, + "grad_norm": 1.5758782625198364, + "learning_rate": 9.96723048446821e-05, + "loss": 0.7448, + "step": 1948 + }, + { + "epoch": 0.130868091674776, + "grad_norm": 6.708095073699951, + "learning_rate": 9.967106140657143e-05, + "loss": 0.8304, + "step": 1950 + }, + { + "epoch": 0.131002315358545, + "grad_norm": 1.4056740999221802, + "learning_rate": 9.966981562160077e-05, + "loss": 0.8205, + "step": 1952 + }, + { + "epoch": 0.131136539042314, + "grad_norm": 1.2714340686798096, + "learning_rate": 9.9668567489829e-05, + "loss": 0.8092, + "step": 1954 + }, + { + "epoch": 0.131270762726083, + "grad_norm": 1.224878191947937, + "learning_rate": 9.966731701131509e-05, + "loss": 0.7486, + "step": 1956 + }, + { + "epoch": 0.13140498640985201, + "grad_norm": 2.1951212882995605, + "learning_rate": 9.966606418611811e-05, + "loss": 0.7046, + "step": 1958 + }, + { + "epoch": 0.13153921009362102, + "grad_norm": 1.5360804796218872, + "learning_rate": 9.966480901429727e-05, + "loss": 0.7998, + "step": 1960 + }, + { + "epoch": 0.13167343377739002, + "grad_norm": 1.2967774868011475, + "learning_rate": 9.966355149591187e-05, + "loss": 0.8602, + "step": 1962 + }, + { + "epoch": 0.13180765746115902, + "grad_norm": 1.301439642906189, + "learning_rate": 9.96622916310213e-05, + "loss": 0.7479, + "step": 1964 + }, + { + "epoch": 0.13194188114492802, + "grad_norm": 1.2981282472610474, + "learning_rate": 9.966102941968512e-05, + "loss": 0.749, + "step": 1966 + }, + { + "epoch": 0.13207610482869703, + "grad_norm": 1.6208810806274414, + "learning_rate": 9.965976486196295e-05, + "loss": 0.724, + "step": 1968 + }, + { + "epoch": 0.13221032851246603, + "grad_norm": 1.2995665073394775, + "learning_rate": 9.965849795791455e-05, + "loss": 0.7397, + "step": 1970 + }, + { + "epoch": 0.13234455219623503, + "grad_norm": 1.1837626695632935, + "learning_rate": 9.965722870759977e-05, + "loss": 0.7759, + "step": 1972 + }, + { + "epoch": 0.13247877588000403, + "grad_norm": 1.2880407571792603, + "learning_rate": 9.965595711107858e-05, + "loss": 0.8343, + "step": 1974 + }, + { + "epoch": 0.13261299956377304, + "grad_norm": 1.2756503820419312, + "learning_rate": 9.965468316841106e-05, + "loss": 0.7975, + "step": 1976 + }, + { + "epoch": 0.13274722324754204, + "grad_norm": 1.225896954536438, + "learning_rate": 9.96534068796574e-05, + "loss": 0.7585, + "step": 1978 + }, + { + "epoch": 0.13288144693131104, + "grad_norm": 2.2436439990997314, + "learning_rate": 9.965212824487791e-05, + "loss": 0.7737, + "step": 1980 + }, + { + "epoch": 0.13301567061508002, + "grad_norm": 3.3554556369781494, + "learning_rate": 9.965084726413298e-05, + "loss": 0.7495, + "step": 1982 + }, + { + "epoch": 0.13314989429884902, + "grad_norm": 1.4639302492141724, + "learning_rate": 9.964956393748317e-05, + "loss": 0.7274, + "step": 1984 + }, + { + "epoch": 0.13328411798261802, + "grad_norm": 1.5518916845321655, + "learning_rate": 9.964827826498909e-05, + "loss": 0.7824, + "step": 1986 + }, + { + "epoch": 0.13341834166638702, + "grad_norm": 1.7486345767974854, + "learning_rate": 9.964699024671148e-05, + "loss": 0.8203, + "step": 1988 + }, + { + "epoch": 0.13355256535015603, + "grad_norm": 1.6883330345153809, + "learning_rate": 9.964569988271122e-05, + "loss": 0.7276, + "step": 1990 + }, + { + "epoch": 0.13368678903392503, + "grad_norm": 1.7586199045181274, + "learning_rate": 9.964440717304926e-05, + "loss": 0.748, + "step": 1992 + }, + { + "epoch": 0.13382101271769403, + "grad_norm": 1.22627854347229, + "learning_rate": 9.964311211778667e-05, + "loss": 0.7706, + "step": 1994 + }, + { + "epoch": 0.13395523640146303, + "grad_norm": 1.7036383152008057, + "learning_rate": 9.964181471698469e-05, + "loss": 0.7716, + "step": 1996 + }, + { + "epoch": 0.13408946008523204, + "grad_norm": 1.3230777978897095, + "learning_rate": 9.964051497070455e-05, + "loss": 0.7399, + "step": 1998 + }, + { + "epoch": 0.13422368376900104, + "grad_norm": 1.5753309726715088, + "learning_rate": 9.963921287900769e-05, + "loss": 0.8049, + "step": 2000 + }, + { + "epoch": 0.13435790745277004, + "grad_norm": 1.1064035892486572, + "learning_rate": 9.963790844195563e-05, + "loss": 0.7231, + "step": 2002 + }, + { + "epoch": 0.13449213113653904, + "grad_norm": 1.4460866451263428, + "learning_rate": 9.963660165961002e-05, + "loss": 0.7572, + "step": 2004 + }, + { + "epoch": 0.13462635482030805, + "grad_norm": 1.347019910812378, + "learning_rate": 9.963529253203259e-05, + "loss": 0.7903, + "step": 2006 + }, + { + "epoch": 0.13476057850407705, + "grad_norm": 2.20070481300354, + "learning_rate": 9.963398105928519e-05, + "loss": 0.7821, + "step": 2008 + }, + { + "epoch": 0.13489480218784605, + "grad_norm": 1.2994256019592285, + "learning_rate": 9.963266724142976e-05, + "loss": 0.7841, + "step": 2010 + }, + { + "epoch": 0.13502902587161505, + "grad_norm": 2.113593339920044, + "learning_rate": 9.963135107852844e-05, + "loss": 0.7278, + "step": 2012 + }, + { + "epoch": 0.13516324955538406, + "grad_norm": 1.2047135829925537, + "learning_rate": 9.963003257064336e-05, + "loss": 0.7626, + "step": 2014 + }, + { + "epoch": 0.13529747323915306, + "grad_norm": 1.3871992826461792, + "learning_rate": 9.962871171783684e-05, + "loss": 0.8016, + "step": 2016 + }, + { + "epoch": 0.13543169692292206, + "grad_norm": 1.2552212476730347, + "learning_rate": 9.962738852017126e-05, + "loss": 0.7771, + "step": 2018 + }, + { + "epoch": 0.13556592060669104, + "grad_norm": 3.0351154804229736, + "learning_rate": 9.962606297770917e-05, + "loss": 0.8385, + "step": 2020 + }, + { + "epoch": 0.13570014429046004, + "grad_norm": 1.8649026155471802, + "learning_rate": 9.962473509051319e-05, + "loss": 0.7947, + "step": 2022 + }, + { + "epoch": 0.13583436797422904, + "grad_norm": 1.1793270111083984, + "learning_rate": 9.962340485864608e-05, + "loss": 0.779, + "step": 2024 + }, + { + "epoch": 0.13596859165799804, + "grad_norm": 3.530226230621338, + "learning_rate": 9.962207228217066e-05, + "loss": 0.7629, + "step": 2026 + }, + { + "epoch": 0.13610281534176705, + "grad_norm": 1.3800299167633057, + "learning_rate": 9.962073736114989e-05, + "loss": 0.7415, + "step": 2028 + }, + { + "epoch": 0.13623703902553605, + "grad_norm": 1.15921950340271, + "learning_rate": 9.961940009564688e-05, + "loss": 0.7576, + "step": 2030 + }, + { + "epoch": 0.13637126270930505, + "grad_norm": 1.1775150299072266, + "learning_rate": 9.961806048572477e-05, + "loss": 0.7293, + "step": 2032 + }, + { + "epoch": 0.13650548639307405, + "grad_norm": 1.3775979280471802, + "learning_rate": 9.961671853144687e-05, + "loss": 0.7293, + "step": 2034 + }, + { + "epoch": 0.13663971007684306, + "grad_norm": 1.3519424200057983, + "learning_rate": 9.96153742328766e-05, + "loss": 0.8251, + "step": 2036 + }, + { + "epoch": 0.13677393376061206, + "grad_norm": 1.2024866342544556, + "learning_rate": 9.961402759007742e-05, + "loss": 0.7602, + "step": 2038 + }, + { + "epoch": 0.13690815744438106, + "grad_norm": 1.2619240283966064, + "learning_rate": 9.961267860311305e-05, + "loss": 0.7521, + "step": 2040 + }, + { + "epoch": 0.13704238112815006, + "grad_norm": 1.3698983192443848, + "learning_rate": 9.961132727204716e-05, + "loss": 0.8188, + "step": 2042 + }, + { + "epoch": 0.13717660481191907, + "grad_norm": 1.5280874967575073, + "learning_rate": 9.96099735969436e-05, + "loss": 0.8327, + "step": 2044 + }, + { + "epoch": 0.13731082849568807, + "grad_norm": 1.436668038368225, + "learning_rate": 9.960861757786634e-05, + "loss": 0.7705, + "step": 2046 + }, + { + "epoch": 0.13744505217945707, + "grad_norm": 1.3540871143341064, + "learning_rate": 9.960725921487947e-05, + "loss": 0.8292, + "step": 2048 + }, + { + "epoch": 0.13757927586322607, + "grad_norm": 1.2535234689712524, + "learning_rate": 9.960589850804713e-05, + "loss": 0.8032, + "step": 2050 + }, + { + "epoch": 0.13771349954699508, + "grad_norm": 1.1608707904815674, + "learning_rate": 9.960453545743365e-05, + "loss": 0.7469, + "step": 2052 + }, + { + "epoch": 0.13784772323076408, + "grad_norm": 1.1416162252426147, + "learning_rate": 9.96031700631034e-05, + "loss": 0.7772, + "step": 2054 + }, + { + "epoch": 0.13798194691453308, + "grad_norm": 1.2960506677627563, + "learning_rate": 9.96018023251209e-05, + "loss": 0.7593, + "step": 2056 + }, + { + "epoch": 0.13811617059830206, + "grad_norm": 1.4882822036743164, + "learning_rate": 9.96004322435508e-05, + "loss": 0.7428, + "step": 2058 + }, + { + "epoch": 0.13825039428207106, + "grad_norm": 1.3273555040359497, + "learning_rate": 9.959905981845781e-05, + "loss": 0.7656, + "step": 2060 + }, + { + "epoch": 0.13838461796584006, + "grad_norm": 1.2009254693984985, + "learning_rate": 9.959768504990675e-05, + "loss": 0.7915, + "step": 2062 + }, + { + "epoch": 0.13851884164960906, + "grad_norm": 1.2466890811920166, + "learning_rate": 9.959630793796262e-05, + "loss": 0.7866, + "step": 2064 + }, + { + "epoch": 0.13865306533337807, + "grad_norm": 1.2222994565963745, + "learning_rate": 9.959492848269047e-05, + "loss": 0.7919, + "step": 2066 + }, + { + "epoch": 0.13878728901714707, + "grad_norm": 1.122150182723999, + "learning_rate": 9.959354668415546e-05, + "loss": 0.7601, + "step": 2068 + }, + { + "epoch": 0.13892151270091607, + "grad_norm": 1.397875428199768, + "learning_rate": 9.95921625424229e-05, + "loss": 0.7757, + "step": 2070 + }, + { + "epoch": 0.13905573638468507, + "grad_norm": 1.2442747354507446, + "learning_rate": 9.959077605755818e-05, + "loss": 0.7413, + "step": 2072 + }, + { + "epoch": 0.13918996006845408, + "grad_norm": 1.3520575761795044, + "learning_rate": 9.95893872296268e-05, + "loss": 0.7089, + "step": 2074 + }, + { + "epoch": 0.13932418375222308, + "grad_norm": 1.3335973024368286, + "learning_rate": 9.958799605869438e-05, + "loss": 0.7301, + "step": 2076 + }, + { + "epoch": 0.13945840743599208, + "grad_norm": 1.2340714931488037, + "learning_rate": 9.958660254482667e-05, + "loss": 0.7388, + "step": 2078 + }, + { + "epoch": 0.13959263111976108, + "grad_norm": 1.120605707168579, + "learning_rate": 9.95852066880895e-05, + "loss": 0.8445, + "step": 2080 + }, + { + "epoch": 0.1397268548035301, + "grad_norm": 1.170307993888855, + "learning_rate": 9.95838084885488e-05, + "loss": 0.7694, + "step": 2082 + }, + { + "epoch": 0.1398610784872991, + "grad_norm": 1.257315754890442, + "learning_rate": 9.958240794627067e-05, + "loss": 0.7403, + "step": 2084 + }, + { + "epoch": 0.1399953021710681, + "grad_norm": 1.2532832622528076, + "learning_rate": 9.958100506132127e-05, + "loss": 0.8383, + "step": 2086 + }, + { + "epoch": 0.1401295258548371, + "grad_norm": 1.278404951095581, + "learning_rate": 9.957959983376686e-05, + "loss": 0.7382, + "step": 2088 + }, + { + "epoch": 0.1402637495386061, + "grad_norm": 1.2620716094970703, + "learning_rate": 9.957819226367385e-05, + "loss": 0.7767, + "step": 2090 + }, + { + "epoch": 0.1403979732223751, + "grad_norm": 1.158979058265686, + "learning_rate": 9.957678235110877e-05, + "loss": 0.7545, + "step": 2092 + }, + { + "epoch": 0.1405321969061441, + "grad_norm": 1.125051498413086, + "learning_rate": 9.957537009613819e-05, + "loss": 0.7712, + "step": 2094 + }, + { + "epoch": 0.14066642058991308, + "grad_norm": 1.1082146167755127, + "learning_rate": 9.957395549882887e-05, + "loss": 0.811, + "step": 2096 + }, + { + "epoch": 0.14080064427368208, + "grad_norm": 1.2813564538955688, + "learning_rate": 9.957253855924761e-05, + "loss": 0.7631, + "step": 2098 + }, + { + "epoch": 0.14093486795745108, + "grad_norm": 1.2051936388015747, + "learning_rate": 9.957111927746143e-05, + "loss": 0.7933, + "step": 2100 + }, + { + "epoch": 0.14106909164122008, + "grad_norm": 1.1182860136032104, + "learning_rate": 9.956969765353731e-05, + "loss": 0.7487, + "step": 2102 + }, + { + "epoch": 0.1412033153249891, + "grad_norm": 1.590226650238037, + "learning_rate": 9.956827368754246e-05, + "loss": 0.8099, + "step": 2104 + }, + { + "epoch": 0.1413375390087581, + "grad_norm": 1.163851261138916, + "learning_rate": 9.956684737954414e-05, + "loss": 0.7571, + "step": 2106 + }, + { + "epoch": 0.1414717626925271, + "grad_norm": 1.29503333568573, + "learning_rate": 9.956541872960976e-05, + "loss": 0.7603, + "step": 2108 + }, + { + "epoch": 0.1416059863762961, + "grad_norm": 1.2573994398117065, + "learning_rate": 9.956398773780682e-05, + "loss": 0.7937, + "step": 2110 + }, + { + "epoch": 0.1417402100600651, + "grad_norm": 1.3426892757415771, + "learning_rate": 9.95625544042029e-05, + "loss": 0.7615, + "step": 2112 + }, + { + "epoch": 0.1418744337438341, + "grad_norm": 1.237913727760315, + "learning_rate": 9.956111872886576e-05, + "loss": 0.7442, + "step": 2114 + }, + { + "epoch": 0.1420086574276031, + "grad_norm": 1.1821072101593018, + "learning_rate": 9.95596807118632e-05, + "loss": 0.6547, + "step": 2116 + }, + { + "epoch": 0.1421428811113721, + "grad_norm": 1.3989217281341553, + "learning_rate": 9.955824035326321e-05, + "loss": 0.738, + "step": 2118 + }, + { + "epoch": 0.1422771047951411, + "grad_norm": 1.27780282497406, + "learning_rate": 9.95567976531338e-05, + "loss": 0.8016, + "step": 2120 + }, + { + "epoch": 0.1424113284789101, + "grad_norm": 1.3274186849594116, + "learning_rate": 9.955535261154316e-05, + "loss": 0.7972, + "step": 2122 + }, + { + "epoch": 0.1425455521626791, + "grad_norm": 1.2454166412353516, + "learning_rate": 9.955390522855954e-05, + "loss": 0.7838, + "step": 2124 + }, + { + "epoch": 0.14267977584644811, + "grad_norm": 1.2846773862838745, + "learning_rate": 9.955245550425135e-05, + "loss": 0.7931, + "step": 2126 + }, + { + "epoch": 0.14281399953021712, + "grad_norm": 1.2487084865570068, + "learning_rate": 9.955100343868709e-05, + "loss": 0.7326, + "step": 2128 + }, + { + "epoch": 0.14294822321398612, + "grad_norm": 1.3661062717437744, + "learning_rate": 9.954954903193533e-05, + "loss": 0.7615, + "step": 2130 + }, + { + "epoch": 0.14308244689775512, + "grad_norm": 1.4582746028900146, + "learning_rate": 9.954809228406483e-05, + "loss": 0.7534, + "step": 2132 + }, + { + "epoch": 0.1432166705815241, + "grad_norm": 1.4687801599502563, + "learning_rate": 9.954663319514439e-05, + "loss": 0.8336, + "step": 2134 + }, + { + "epoch": 0.1433508942652931, + "grad_norm": 1.1722335815429688, + "learning_rate": 9.954517176524298e-05, + "loss": 0.7461, + "step": 2136 + }, + { + "epoch": 0.1434851179490621, + "grad_norm": 1.2503559589385986, + "learning_rate": 9.954370799442961e-05, + "loss": 0.8353, + "step": 2138 + }, + { + "epoch": 0.1436193416328311, + "grad_norm": 1.614410400390625, + "learning_rate": 9.954224188277347e-05, + "loss": 0.6964, + "step": 2140 + }, + { + "epoch": 0.1437535653166001, + "grad_norm": 1.238930344581604, + "learning_rate": 9.954077343034383e-05, + "loss": 0.7151, + "step": 2142 + }, + { + "epoch": 0.1438877890003691, + "grad_norm": 1.317571997642517, + "learning_rate": 9.953930263721003e-05, + "loss": 0.8119, + "step": 2144 + }, + { + "epoch": 0.1440220126841381, + "grad_norm": 1.0393397808074951, + "learning_rate": 9.953782950344164e-05, + "loss": 0.7762, + "step": 2146 + }, + { + "epoch": 0.14415623636790711, + "grad_norm": 1.179162621498108, + "learning_rate": 9.95363540291082e-05, + "loss": 0.8064, + "step": 2148 + }, + { + "epoch": 0.14429046005167612, + "grad_norm": 1.8216201066970825, + "learning_rate": 9.953487621427942e-05, + "loss": 0.7538, + "step": 2150 + }, + { + "epoch": 0.14442468373544512, + "grad_norm": 1.174189805984497, + "learning_rate": 9.953339605902517e-05, + "loss": 0.8036, + "step": 2152 + }, + { + "epoch": 0.14455890741921412, + "grad_norm": 1.2796666622161865, + "learning_rate": 9.953191356341535e-05, + "loss": 0.7963, + "step": 2154 + }, + { + "epoch": 0.14469313110298312, + "grad_norm": 1.1608229875564575, + "learning_rate": 9.953042872752003e-05, + "loss": 0.7536, + "step": 2156 + }, + { + "epoch": 0.14482735478675213, + "grad_norm": 1.2113354206085205, + "learning_rate": 9.952894155140932e-05, + "loss": 0.7396, + "step": 2158 + }, + { + "epoch": 0.14496157847052113, + "grad_norm": 1.3290759325027466, + "learning_rate": 9.952745203515354e-05, + "loss": 0.7895, + "step": 2160 + }, + { + "epoch": 0.14509580215429013, + "grad_norm": 1.383651614189148, + "learning_rate": 9.952596017882304e-05, + "loss": 0.7644, + "step": 2162 + }, + { + "epoch": 0.14523002583805913, + "grad_norm": 1.2829383611679077, + "learning_rate": 9.95244659824883e-05, + "loss": 0.7665, + "step": 2164 + }, + { + "epoch": 0.14536424952182814, + "grad_norm": 1.3651251792907715, + "learning_rate": 9.952296944621994e-05, + "loss": 0.7422, + "step": 2166 + }, + { + "epoch": 0.14549847320559714, + "grad_norm": 1.5312341451644897, + "learning_rate": 9.952147057008864e-05, + "loss": 0.7046, + "step": 2168 + }, + { + "epoch": 0.14563269688936614, + "grad_norm": 1.7080609798431396, + "learning_rate": 9.951996935416525e-05, + "loss": 0.7642, + "step": 2170 + }, + { + "epoch": 0.14576692057313512, + "grad_norm": 1.1974818706512451, + "learning_rate": 9.951846579852069e-05, + "loss": 0.7425, + "step": 2172 + }, + { + "epoch": 0.14590114425690412, + "grad_norm": 1.213015079498291, + "learning_rate": 9.951695990322598e-05, + "loss": 0.8013, + "step": 2174 + }, + { + "epoch": 0.14603536794067312, + "grad_norm": 1.1024574041366577, + "learning_rate": 9.95154516683523e-05, + "loss": 0.7126, + "step": 2176 + }, + { + "epoch": 0.14616959162444212, + "grad_norm": 1.175889492034912, + "learning_rate": 9.951394109397088e-05, + "loss": 0.7567, + "step": 2178 + }, + { + "epoch": 0.14630381530821113, + "grad_norm": 3.2251601219177246, + "learning_rate": 9.951242818015312e-05, + "loss": 0.7992, + "step": 2180 + }, + { + "epoch": 0.14643803899198013, + "grad_norm": 1.39309561252594, + "learning_rate": 9.951091292697048e-05, + "loss": 0.756, + "step": 2182 + }, + { + "epoch": 0.14657226267574913, + "grad_norm": 1.3651174306869507, + "learning_rate": 9.950939533449458e-05, + "loss": 0.8032, + "step": 2184 + }, + { + "epoch": 0.14670648635951813, + "grad_norm": 1.6942890882492065, + "learning_rate": 9.950787540279711e-05, + "loss": 0.7455, + "step": 2186 + }, + { + "epoch": 0.14684071004328714, + "grad_norm": 1.5253287553787231, + "learning_rate": 9.950635313194986e-05, + "loss": 0.7913, + "step": 2188 + }, + { + "epoch": 0.14697493372705614, + "grad_norm": 1.0081595182418823, + "learning_rate": 9.950482852202478e-05, + "loss": 0.71, + "step": 2190 + }, + { + "epoch": 0.14710915741082514, + "grad_norm": 1.2482656240463257, + "learning_rate": 9.95033015730939e-05, + "loss": 0.8345, + "step": 2192 + }, + { + "epoch": 0.14724338109459414, + "grad_norm": 1.436821460723877, + "learning_rate": 9.950177228522937e-05, + "loss": 0.783, + "step": 2194 + }, + { + "epoch": 0.14737760477836315, + "grad_norm": 1.3209800720214844, + "learning_rate": 9.950024065850343e-05, + "loss": 0.708, + "step": 2196 + }, + { + "epoch": 0.14751182846213215, + "grad_norm": 1.070220947265625, + "learning_rate": 9.949870669298846e-05, + "loss": 0.7609, + "step": 2198 + }, + { + "epoch": 0.14764605214590115, + "grad_norm": 1.7431410551071167, + "learning_rate": 9.949717038875695e-05, + "loss": 0.7533, + "step": 2200 + }, + { + "epoch": 0.14778027582967015, + "grad_norm": 1.4989782571792603, + "learning_rate": 9.949563174588146e-05, + "loss": 0.7089, + "step": 2202 + }, + { + "epoch": 0.14791449951343916, + "grad_norm": 1.8121305704116821, + "learning_rate": 9.949409076443468e-05, + "loss": 0.7706, + "step": 2204 + }, + { + "epoch": 0.14804872319720816, + "grad_norm": 1.2245267629623413, + "learning_rate": 9.949254744448946e-05, + "loss": 0.8267, + "step": 2206 + }, + { + "epoch": 0.14818294688097716, + "grad_norm": 2.5223753452301025, + "learning_rate": 9.94910017861187e-05, + "loss": 0.7736, + "step": 2208 + }, + { + "epoch": 0.14831717056474614, + "grad_norm": 1.1899269819259644, + "learning_rate": 9.948945378939542e-05, + "loss": 0.7145, + "step": 2210 + }, + { + "epoch": 0.14845139424851514, + "grad_norm": 1.1237188577651978, + "learning_rate": 9.948790345439276e-05, + "loss": 0.8445, + "step": 2212 + }, + { + "epoch": 0.14858561793228414, + "grad_norm": 1.4916682243347168, + "learning_rate": 9.948635078118398e-05, + "loss": 0.8115, + "step": 2214 + }, + { + "epoch": 0.14871984161605314, + "grad_norm": 1.5387001037597656, + "learning_rate": 9.948479576984242e-05, + "loss": 0.7568, + "step": 2216 + }, + { + "epoch": 0.14885406529982215, + "grad_norm": 1.1785844564437866, + "learning_rate": 9.948323842044159e-05, + "loss": 0.7708, + "step": 2218 + }, + { + "epoch": 0.14898828898359115, + "grad_norm": 1.6287435293197632, + "learning_rate": 9.948167873305503e-05, + "loss": 0.7399, + "step": 2220 + }, + { + "epoch": 0.14912251266736015, + "grad_norm": 1.3606981039047241, + "learning_rate": 9.948011670775647e-05, + "loss": 0.8457, + "step": 2222 + }, + { + "epoch": 0.14925673635112915, + "grad_norm": 1.2144297361373901, + "learning_rate": 9.94785523446197e-05, + "loss": 0.7864, + "step": 2224 + }, + { + "epoch": 0.14939096003489816, + "grad_norm": 1.1448192596435547, + "learning_rate": 9.947698564371859e-05, + "loss": 0.751, + "step": 2226 + }, + { + "epoch": 0.14952518371866716, + "grad_norm": 1.1770398616790771, + "learning_rate": 9.947541660512722e-05, + "loss": 0.7395, + "step": 2228 + }, + { + "epoch": 0.14965940740243616, + "grad_norm": 1.417425274848938, + "learning_rate": 9.947384522891972e-05, + "loss": 0.803, + "step": 2230 + }, + { + "epoch": 0.14979363108620516, + "grad_norm": 1.1650129556655884, + "learning_rate": 9.947227151517032e-05, + "loss": 0.7982, + "step": 2232 + }, + { + "epoch": 0.14992785476997417, + "grad_norm": 1.4188095331192017, + "learning_rate": 9.947069546395336e-05, + "loss": 0.7795, + "step": 2234 + }, + { + "epoch": 0.15006207845374317, + "grad_norm": 1.1999107599258423, + "learning_rate": 9.946911707534331e-05, + "loss": 0.7865, + "step": 2236 + }, + { + "epoch": 0.15019630213751217, + "grad_norm": 1.2209099531173706, + "learning_rate": 9.946753634941477e-05, + "loss": 0.817, + "step": 2238 + }, + { + "epoch": 0.15033052582128117, + "grad_norm": 1.3145641088485718, + "learning_rate": 9.94659532862424e-05, + "loss": 0.7132, + "step": 2240 + }, + { + "epoch": 0.15046474950505018, + "grad_norm": 1.1500871181488037, + "learning_rate": 9.9464367885901e-05, + "loss": 0.7218, + "step": 2242 + }, + { + "epoch": 0.15059897318881918, + "grad_norm": 1.2443076372146606, + "learning_rate": 9.946278014846551e-05, + "loss": 0.8128, + "step": 2244 + }, + { + "epoch": 0.15073319687258815, + "grad_norm": 1.2562689781188965, + "learning_rate": 9.946119007401091e-05, + "loss": 0.7358, + "step": 2246 + }, + { + "epoch": 0.15086742055635716, + "grad_norm": 2.3361754417419434, + "learning_rate": 9.945959766261235e-05, + "loss": 0.7404, + "step": 2248 + }, + { + "epoch": 0.15100164424012616, + "grad_norm": 1.2442772388458252, + "learning_rate": 9.945800291434504e-05, + "loss": 0.7834, + "step": 2250 + }, + { + "epoch": 0.15113586792389516, + "grad_norm": 1.1339282989501953, + "learning_rate": 9.945640582928437e-05, + "loss": 0.6752, + "step": 2252 + }, + { + "epoch": 0.15127009160766416, + "grad_norm": 1.4449889659881592, + "learning_rate": 9.945480640750577e-05, + "loss": 0.7898, + "step": 2254 + }, + { + "epoch": 0.15140431529143317, + "grad_norm": 1.2559311389923096, + "learning_rate": 9.945320464908481e-05, + "loss": 0.7696, + "step": 2256 + }, + { + "epoch": 0.15153853897520217, + "grad_norm": 1.77239191532135, + "learning_rate": 9.945160055409717e-05, + "loss": 0.8126, + "step": 2258 + }, + { + "epoch": 0.15167276265897117, + "grad_norm": 1.2011572122573853, + "learning_rate": 9.944999412261866e-05, + "loss": 0.7859, + "step": 2260 + }, + { + "epoch": 0.15180698634274017, + "grad_norm": 1.8600788116455078, + "learning_rate": 9.944838535472515e-05, + "loss": 0.7774, + "step": 2262 + }, + { + "epoch": 0.15194121002650918, + "grad_norm": 1.2087199687957764, + "learning_rate": 9.944677425049268e-05, + "loss": 0.7841, + "step": 2264 + }, + { + "epoch": 0.15207543371027818, + "grad_norm": 1.272459864616394, + "learning_rate": 9.944516080999735e-05, + "loss": 0.7719, + "step": 2266 + }, + { + "epoch": 0.15220965739404718, + "grad_norm": 1.721838116645813, + "learning_rate": 9.944354503331541e-05, + "loss": 0.7264, + "step": 2268 + }, + { + "epoch": 0.15234388107781618, + "grad_norm": 1.2229008674621582, + "learning_rate": 9.94419269205232e-05, + "loss": 0.7513, + "step": 2270 + }, + { + "epoch": 0.1524781047615852, + "grad_norm": 1.1715528964996338, + "learning_rate": 9.944030647169715e-05, + "loss": 0.7562, + "step": 2272 + }, + { + "epoch": 0.1526123284453542, + "grad_norm": 1.1329717636108398, + "learning_rate": 9.943868368691386e-05, + "loss": 0.702, + "step": 2274 + }, + { + "epoch": 0.1527465521291232, + "grad_norm": 1.1848491430282593, + "learning_rate": 9.943705856624996e-05, + "loss": 0.7484, + "step": 2276 + }, + { + "epoch": 0.1528807758128922, + "grad_norm": 1.3173327445983887, + "learning_rate": 9.943543110978227e-05, + "loss": 0.7915, + "step": 2278 + }, + { + "epoch": 0.1530149994966612, + "grad_norm": 1.2425185441970825, + "learning_rate": 9.943380131758768e-05, + "loss": 0.7463, + "step": 2280 + }, + { + "epoch": 0.1531492231804302, + "grad_norm": 1.7477085590362549, + "learning_rate": 9.943216918974317e-05, + "loss": 0.7949, + "step": 2282 + }, + { + "epoch": 0.15328344686419917, + "grad_norm": 1.1816725730895996, + "learning_rate": 9.943053472632587e-05, + "loss": 0.7867, + "step": 2284 + }, + { + "epoch": 0.15341767054796818, + "grad_norm": 1.3466622829437256, + "learning_rate": 9.942889792741302e-05, + "loss": 0.7513, + "step": 2286 + }, + { + "epoch": 0.15355189423173718, + "grad_norm": 1.5214899778366089, + "learning_rate": 9.942725879308192e-05, + "loss": 0.7092, + "step": 2288 + }, + { + "epoch": 0.15368611791550618, + "grad_norm": 1.2291330099105835, + "learning_rate": 9.942561732341005e-05, + "loss": 0.7821, + "step": 2290 + }, + { + "epoch": 0.15382034159927518, + "grad_norm": 1.209675908088684, + "learning_rate": 9.942397351847494e-05, + "loss": 0.6945, + "step": 2292 + }, + { + "epoch": 0.1539545652830442, + "grad_norm": 1.1875338554382324, + "learning_rate": 9.942232737835428e-05, + "loss": 0.8003, + "step": 2294 + }, + { + "epoch": 0.1540887889668132, + "grad_norm": 1.1507275104522705, + "learning_rate": 9.942067890312584e-05, + "loss": 0.8093, + "step": 2296 + }, + { + "epoch": 0.1542230126505822, + "grad_norm": 1.1249173879623413, + "learning_rate": 9.94190280928675e-05, + "loss": 0.7864, + "step": 2298 + }, + { + "epoch": 0.1543572363343512, + "grad_norm": 1.3070746660232544, + "learning_rate": 9.941737494765725e-05, + "loss": 0.6958, + "step": 2300 + }, + { + "epoch": 0.1544914600181202, + "grad_norm": 1.1317988634109497, + "learning_rate": 9.941571946757321e-05, + "loss": 0.7691, + "step": 2302 + }, + { + "epoch": 0.1546256837018892, + "grad_norm": 1.2521330118179321, + "learning_rate": 9.941406165269362e-05, + "loss": 0.7484, + "step": 2304 + }, + { + "epoch": 0.1547599073856582, + "grad_norm": 1.2915416955947876, + "learning_rate": 9.941240150309676e-05, + "loss": 0.829, + "step": 2306 + }, + { + "epoch": 0.1548941310694272, + "grad_norm": 0.9967454671859741, + "learning_rate": 9.94107390188611e-05, + "loss": 0.6904, + "step": 2308 + }, + { + "epoch": 0.1550283547531962, + "grad_norm": 1.4336018562316895, + "learning_rate": 9.94090742000652e-05, + "loss": 0.7557, + "step": 2310 + }, + { + "epoch": 0.1551625784369652, + "grad_norm": 1.0555182695388794, + "learning_rate": 9.940740704678768e-05, + "loss": 0.7726, + "step": 2312 + }, + { + "epoch": 0.1552968021207342, + "grad_norm": 1.298625111579895, + "learning_rate": 9.940573755910735e-05, + "loss": 0.7815, + "step": 2314 + }, + { + "epoch": 0.15543102580450321, + "grad_norm": 1.1737966537475586, + "learning_rate": 9.940406573710306e-05, + "loss": 0.7837, + "step": 2316 + }, + { + "epoch": 0.15556524948827222, + "grad_norm": 2.8313827514648438, + "learning_rate": 9.940239158085382e-05, + "loss": 0.8697, + "step": 2318 + }, + { + "epoch": 0.15569947317204122, + "grad_norm": 1.422286033630371, + "learning_rate": 9.940071509043872e-05, + "loss": 0.7542, + "step": 2320 + }, + { + "epoch": 0.1558336968558102, + "grad_norm": 1.3354989290237427, + "learning_rate": 9.939903626593698e-05, + "loss": 0.8018, + "step": 2322 + }, + { + "epoch": 0.1559679205395792, + "grad_norm": 1.183380126953125, + "learning_rate": 9.939735510742792e-05, + "loss": 0.777, + "step": 2324 + }, + { + "epoch": 0.1561021442233482, + "grad_norm": 1.3062756061553955, + "learning_rate": 9.939567161499095e-05, + "loss": 0.7899, + "step": 2326 + }, + { + "epoch": 0.1562363679071172, + "grad_norm": 1.3748116493225098, + "learning_rate": 9.939398578870563e-05, + "loss": 0.7707, + "step": 2328 + }, + { + "epoch": 0.1563705915908862, + "grad_norm": 1.2097219228744507, + "learning_rate": 9.939229762865164e-05, + "loss": 0.7341, + "step": 2330 + }, + { + "epoch": 0.1565048152746552, + "grad_norm": 1.2697770595550537, + "learning_rate": 9.939060713490868e-05, + "loss": 0.6468, + "step": 2332 + }, + { + "epoch": 0.1566390389584242, + "grad_norm": 1.3282665014266968, + "learning_rate": 9.938891430755666e-05, + "loss": 0.7525, + "step": 2334 + }, + { + "epoch": 0.1567732626421932, + "grad_norm": 1.0913504362106323, + "learning_rate": 9.938721914667557e-05, + "loss": 0.7398, + "step": 2336 + }, + { + "epoch": 0.15690748632596221, + "grad_norm": 1.1177167892456055, + "learning_rate": 9.938552165234548e-05, + "loss": 0.7344, + "step": 2338 + }, + { + "epoch": 0.15704171000973122, + "grad_norm": 1.3176368474960327, + "learning_rate": 9.93838218246466e-05, + "loss": 0.7472, + "step": 2340 + }, + { + "epoch": 0.15717593369350022, + "grad_norm": 1.211304783821106, + "learning_rate": 9.938211966365926e-05, + "loss": 0.7897, + "step": 2342 + }, + { + "epoch": 0.15731015737726922, + "grad_norm": 1.1732478141784668, + "learning_rate": 9.938041516946389e-05, + "loss": 0.7204, + "step": 2344 + }, + { + "epoch": 0.15744438106103822, + "grad_norm": 1.1366077661514282, + "learning_rate": 9.937870834214097e-05, + "loss": 0.6967, + "step": 2346 + }, + { + "epoch": 0.15757860474480723, + "grad_norm": 5.17569637298584, + "learning_rate": 9.93769991817712e-05, + "loss": 0.7289, + "step": 2348 + }, + { + "epoch": 0.15771282842857623, + "grad_norm": 1.351745843887329, + "learning_rate": 9.93752876884353e-05, + "loss": 0.7601, + "step": 2350 + }, + { + "epoch": 0.15784705211234523, + "grad_norm": 1.280282974243164, + "learning_rate": 9.937357386221416e-05, + "loss": 0.7786, + "step": 2352 + }, + { + "epoch": 0.15798127579611423, + "grad_norm": 1.163209080696106, + "learning_rate": 9.937185770318874e-05, + "loss": 0.748, + "step": 2354 + }, + { + "epoch": 0.15811549947988324, + "grad_norm": 1.4269531965255737, + "learning_rate": 9.937013921144014e-05, + "loss": 0.7422, + "step": 2356 + }, + { + "epoch": 0.15824972316365224, + "grad_norm": 1.350040078163147, + "learning_rate": 9.936841838704953e-05, + "loss": 0.7644, + "step": 2358 + }, + { + "epoch": 0.15838394684742121, + "grad_norm": 1.2202732563018799, + "learning_rate": 9.936669523009823e-05, + "loss": 0.7566, + "step": 2360 + }, + { + "epoch": 0.15851817053119022, + "grad_norm": 1.2352739572525024, + "learning_rate": 9.936496974066767e-05, + "loss": 0.7225, + "step": 2362 + }, + { + "epoch": 0.15865239421495922, + "grad_norm": 1.2050681114196777, + "learning_rate": 9.936324191883935e-05, + "loss": 0.7364, + "step": 2364 + }, + { + "epoch": 0.15878661789872822, + "grad_norm": 1.1253108978271484, + "learning_rate": 9.936151176469493e-05, + "loss": 0.7951, + "step": 2366 + }, + { + "epoch": 0.15892084158249722, + "grad_norm": 1.2299094200134277, + "learning_rate": 9.935977927831612e-05, + "loss": 0.7207, + "step": 2368 + }, + { + "epoch": 0.15905506526626623, + "grad_norm": 1.3539718389511108, + "learning_rate": 9.935804445978483e-05, + "loss": 0.8654, + "step": 2370 + }, + { + "epoch": 0.15918928895003523, + "grad_norm": 1.3129518032073975, + "learning_rate": 9.935630730918297e-05, + "loss": 0.7026, + "step": 2372 + }, + { + "epoch": 0.15932351263380423, + "grad_norm": 1.2837015390396118, + "learning_rate": 9.935456782659267e-05, + "loss": 0.8147, + "step": 2374 + }, + { + "epoch": 0.15945773631757323, + "grad_norm": 1.6276519298553467, + "learning_rate": 9.935282601209608e-05, + "loss": 0.7906, + "step": 2376 + }, + { + "epoch": 0.15959196000134224, + "grad_norm": 1.1305961608886719, + "learning_rate": 9.93510818657755e-05, + "loss": 0.7622, + "step": 2378 + }, + { + "epoch": 0.15972618368511124, + "grad_norm": 1.7124518156051636, + "learning_rate": 9.934933538771336e-05, + "loss": 0.7523, + "step": 2380 + }, + { + "epoch": 0.15986040736888024, + "grad_norm": 1.5002055168151855, + "learning_rate": 9.934758657799218e-05, + "loss": 0.8082, + "step": 2382 + }, + { + "epoch": 0.15999463105264924, + "grad_norm": 1.216407060623169, + "learning_rate": 9.934583543669453e-05, + "loss": 0.7598, + "step": 2384 + }, + { + "epoch": 0.16012885473641825, + "grad_norm": 1.3850566148757935, + "learning_rate": 9.934408196390322e-05, + "loss": 0.7452, + "step": 2386 + }, + { + "epoch": 0.16026307842018725, + "grad_norm": 1.1947745084762573, + "learning_rate": 9.934232615970107e-05, + "loss": 0.7384, + "step": 2388 + }, + { + "epoch": 0.16039730210395625, + "grad_norm": 1.2187963724136353, + "learning_rate": 9.934056802417101e-05, + "loss": 0.7995, + "step": 2390 + }, + { + "epoch": 0.16053152578772525, + "grad_norm": 1.179424524307251, + "learning_rate": 9.933880755739616e-05, + "loss": 0.8486, + "step": 2392 + }, + { + "epoch": 0.16066574947149426, + "grad_norm": 1.32096266746521, + "learning_rate": 9.933704475945966e-05, + "loss": 0.7991, + "step": 2394 + }, + { + "epoch": 0.16079997315526326, + "grad_norm": 1.4193588495254517, + "learning_rate": 9.933527963044483e-05, + "loss": 0.7433, + "step": 2396 + }, + { + "epoch": 0.16093419683903223, + "grad_norm": 1.4676076173782349, + "learning_rate": 9.933351217043504e-05, + "loss": 0.7406, + "step": 2398 + }, + { + "epoch": 0.16106842052280124, + "grad_norm": 1.3093665838241577, + "learning_rate": 9.933174237951381e-05, + "loss": 0.7232, + "step": 2400 + }, + { + "epoch": 0.16120264420657024, + "grad_norm": 1.3477256298065186, + "learning_rate": 9.932997025776475e-05, + "loss": 0.7955, + "step": 2402 + }, + { + "epoch": 0.16133686789033924, + "grad_norm": 1.4739902019500732, + "learning_rate": 9.932819580527162e-05, + "loss": 0.7765, + "step": 2404 + }, + { + "epoch": 0.16147109157410824, + "grad_norm": 0.9575392007827759, + "learning_rate": 9.932641902211821e-05, + "loss": 0.7153, + "step": 2406 + }, + { + "epoch": 0.16160531525787725, + "grad_norm": 1.135473370552063, + "learning_rate": 9.932463990838851e-05, + "loss": 0.7088, + "step": 2408 + }, + { + "epoch": 0.16173953894164625, + "grad_norm": 1.268178939819336, + "learning_rate": 9.932285846416658e-05, + "loss": 0.7715, + "step": 2410 + }, + { + "epoch": 0.16187376262541525, + "grad_norm": 1.320458173751831, + "learning_rate": 9.932107468953656e-05, + "loss": 0.7896, + "step": 2412 + }, + { + "epoch": 0.16200798630918425, + "grad_norm": 1.2818015813827515, + "learning_rate": 9.931928858458276e-05, + "loss": 0.7705, + "step": 2414 + }, + { + "epoch": 0.16214220999295326, + "grad_norm": 1.1835918426513672, + "learning_rate": 9.931750014938954e-05, + "loss": 0.7115, + "step": 2416 + }, + { + "epoch": 0.16227643367672226, + "grad_norm": 1.1886675357818604, + "learning_rate": 9.931570938404144e-05, + "loss": 0.7545, + "step": 2418 + }, + { + "epoch": 0.16241065736049126, + "grad_norm": 1.196901559829712, + "learning_rate": 9.931391628862304e-05, + "loss": 0.7196, + "step": 2420 + }, + { + "epoch": 0.16254488104426026, + "grad_norm": 1.235708475112915, + "learning_rate": 9.931212086321905e-05, + "loss": 0.8074, + "step": 2422 + }, + { + "epoch": 0.16267910472802927, + "grad_norm": 1.2353520393371582, + "learning_rate": 9.931032310791432e-05, + "loss": 0.7673, + "step": 2424 + }, + { + "epoch": 0.16281332841179827, + "grad_norm": 1.1860045194625854, + "learning_rate": 9.930852302279381e-05, + "loss": 0.7585, + "step": 2426 + }, + { + "epoch": 0.16294755209556727, + "grad_norm": 1.1550395488739014, + "learning_rate": 9.930672060794253e-05, + "loss": 0.7325, + "step": 2428 + }, + { + "epoch": 0.16308177577933627, + "grad_norm": 1.867236614227295, + "learning_rate": 9.930491586344565e-05, + "loss": 0.7485, + "step": 2430 + }, + { + "epoch": 0.16321599946310528, + "grad_norm": 1.3778027296066284, + "learning_rate": 9.930310878938846e-05, + "loss": 0.7444, + "step": 2432 + }, + { + "epoch": 0.16335022314687428, + "grad_norm": 1.1479414701461792, + "learning_rate": 9.930129938585633e-05, + "loss": 0.781, + "step": 2434 + }, + { + "epoch": 0.16348444683064325, + "grad_norm": 1.8815853595733643, + "learning_rate": 9.929948765293477e-05, + "loss": 0.6867, + "step": 2436 + }, + { + "epoch": 0.16361867051441226, + "grad_norm": 1.6327052116394043, + "learning_rate": 9.929767359070934e-05, + "loss": 0.6813, + "step": 2438 + }, + { + "epoch": 0.16375289419818126, + "grad_norm": 1.1803863048553467, + "learning_rate": 9.929585719926577e-05, + "loss": 0.7515, + "step": 2440 + }, + { + "epoch": 0.16388711788195026, + "grad_norm": 1.522848129272461, + "learning_rate": 9.92940384786899e-05, + "loss": 0.7298, + "step": 2442 + }, + { + "epoch": 0.16402134156571926, + "grad_norm": 1.3202654123306274, + "learning_rate": 9.929221742906763e-05, + "loss": 0.8428, + "step": 2444 + }, + { + "epoch": 0.16415556524948827, + "grad_norm": 1.2601298093795776, + "learning_rate": 9.929039405048501e-05, + "loss": 0.7443, + "step": 2446 + }, + { + "epoch": 0.16428978893325727, + "grad_norm": 1.2158600091934204, + "learning_rate": 9.928856834302823e-05, + "loss": 0.7908, + "step": 2448 + }, + { + "epoch": 0.16442401261702627, + "grad_norm": 1.2021377086639404, + "learning_rate": 9.928674030678348e-05, + "loss": 0.7545, + "step": 2450 + }, + { + "epoch": 0.16455823630079527, + "grad_norm": 1.3095088005065918, + "learning_rate": 9.928490994183719e-05, + "loss": 0.7767, + "step": 2452 + }, + { + "epoch": 0.16469245998456428, + "grad_norm": 1.6881237030029297, + "learning_rate": 9.928307724827581e-05, + "loss": 0.739, + "step": 2454 + }, + { + "epoch": 0.16482668366833328, + "grad_norm": 1.1968653202056885, + "learning_rate": 9.928124222618594e-05, + "loss": 0.6853, + "step": 2456 + }, + { + "epoch": 0.16496090735210228, + "grad_norm": 2.4138987064361572, + "learning_rate": 9.92794048756543e-05, + "loss": 0.7295, + "step": 2458 + }, + { + "epoch": 0.16509513103587128, + "grad_norm": 1.1000226736068726, + "learning_rate": 9.927756519676765e-05, + "loss": 0.7345, + "step": 2460 + }, + { + "epoch": 0.1652293547196403, + "grad_norm": 1.4773651361465454, + "learning_rate": 9.927572318961299e-05, + "loss": 0.7501, + "step": 2462 + }, + { + "epoch": 0.1653635784034093, + "grad_norm": 1.304039478302002, + "learning_rate": 9.927387885427726e-05, + "loss": 0.6835, + "step": 2464 + }, + { + "epoch": 0.1654978020871783, + "grad_norm": 1.3185865879058838, + "learning_rate": 9.927203219084766e-05, + "loss": 0.7706, + "step": 2466 + }, + { + "epoch": 0.1656320257709473, + "grad_norm": 1.181311845779419, + "learning_rate": 9.927018319941145e-05, + "loss": 0.7814, + "step": 2468 + }, + { + "epoch": 0.1657662494547163, + "grad_norm": 2.8532204627990723, + "learning_rate": 9.926833188005595e-05, + "loss": 0.7775, + "step": 2470 + }, + { + "epoch": 0.1659004731384853, + "grad_norm": 1.2215230464935303, + "learning_rate": 9.926647823286865e-05, + "loss": 0.7126, + "step": 2472 + }, + { + "epoch": 0.16603469682225427, + "grad_norm": 1.2203389406204224, + "learning_rate": 9.926462225793714e-05, + "loss": 0.7673, + "step": 2474 + }, + { + "epoch": 0.16616892050602328, + "grad_norm": 1.7585606575012207, + "learning_rate": 9.926276395534911e-05, + "loss": 0.7177, + "step": 2476 + }, + { + "epoch": 0.16630314418979228, + "grad_norm": 1.2644014358520508, + "learning_rate": 9.926090332519234e-05, + "loss": 0.8059, + "step": 2478 + }, + { + "epoch": 0.16643736787356128, + "grad_norm": 1.0948673486709595, + "learning_rate": 9.925904036755475e-05, + "loss": 0.8598, + "step": 2480 + }, + { + "epoch": 0.16657159155733028, + "grad_norm": 1.1442246437072754, + "learning_rate": 9.92571750825244e-05, + "loss": 0.7273, + "step": 2482 + }, + { + "epoch": 0.1667058152410993, + "grad_norm": 1.0907037258148193, + "learning_rate": 9.925530747018935e-05, + "loss": 0.7578, + "step": 2484 + }, + { + "epoch": 0.1668400389248683, + "grad_norm": 1.1021397113800049, + "learning_rate": 9.92534375306379e-05, + "loss": 0.6874, + "step": 2486 + }, + { + "epoch": 0.1669742626086373, + "grad_norm": 1.3376507759094238, + "learning_rate": 9.925156526395836e-05, + "loss": 0.7817, + "step": 2488 + }, + { + "epoch": 0.1671084862924063, + "grad_norm": 1.2467944622039795, + "learning_rate": 9.924969067023922e-05, + "loss": 0.7608, + "step": 2490 + }, + { + "epoch": 0.1672427099761753, + "grad_norm": 1.2369590997695923, + "learning_rate": 9.924781374956905e-05, + "loss": 0.7391, + "step": 2492 + }, + { + "epoch": 0.1673769336599443, + "grad_norm": 1.2221477031707764, + "learning_rate": 9.924593450203652e-05, + "loss": 0.7096, + "step": 2494 + }, + { + "epoch": 0.1675111573437133, + "grad_norm": 1.2566503286361694, + "learning_rate": 9.924405292773041e-05, + "loss": 0.7927, + "step": 2496 + }, + { + "epoch": 0.1676453810274823, + "grad_norm": 1.7431105375289917, + "learning_rate": 9.924216902673964e-05, + "loss": 0.7695, + "step": 2498 + }, + { + "epoch": 0.1677796047112513, + "grad_norm": 1.2656195163726807, + "learning_rate": 9.924028279915323e-05, + "loss": 0.7307, + "step": 2500 + }, + { + "epoch": 0.1679138283950203, + "grad_norm": 1.2934205532073975, + "learning_rate": 9.923839424506026e-05, + "loss": 0.7277, + "step": 2502 + }, + { + "epoch": 0.1680480520787893, + "grad_norm": 1.0522221326828003, + "learning_rate": 9.923650336454999e-05, + "loss": 0.6516, + "step": 2504 + }, + { + "epoch": 0.16818227576255831, + "grad_norm": 1.1378988027572632, + "learning_rate": 9.923461015771177e-05, + "loss": 0.749, + "step": 2506 + }, + { + "epoch": 0.16831649944632732, + "grad_norm": 1.086676001548767, + "learning_rate": 9.923271462463502e-05, + "loss": 0.7114, + "step": 2508 + }, + { + "epoch": 0.16845072313009632, + "grad_norm": 1.1618993282318115, + "learning_rate": 9.923081676540933e-05, + "loss": 0.7368, + "step": 2510 + }, + { + "epoch": 0.1685849468138653, + "grad_norm": 1.5692874193191528, + "learning_rate": 9.922891658012435e-05, + "loss": 0.7363, + "step": 2512 + }, + { + "epoch": 0.1687191704976343, + "grad_norm": 1.5985716581344604, + "learning_rate": 9.922701406886987e-05, + "loss": 0.7275, + "step": 2514 + }, + { + "epoch": 0.1688533941814033, + "grad_norm": 1.7021191120147705, + "learning_rate": 9.922510923173576e-05, + "loss": 0.7594, + "step": 2516 + }, + { + "epoch": 0.1689876178651723, + "grad_norm": 1.0292367935180664, + "learning_rate": 9.922320206881205e-05, + "loss": 0.7647, + "step": 2518 + }, + { + "epoch": 0.1691218415489413, + "grad_norm": 1.0268237590789795, + "learning_rate": 9.922129258018883e-05, + "loss": 0.7616, + "step": 2520 + }, + { + "epoch": 0.1692560652327103, + "grad_norm": 1.1957460641860962, + "learning_rate": 9.921938076595632e-05, + "loss": 0.7571, + "step": 2522 + }, + { + "epoch": 0.1693902889164793, + "grad_norm": 1.319554328918457, + "learning_rate": 9.921746662620488e-05, + "loss": 0.7842, + "step": 2524 + }, + { + "epoch": 0.1695245126002483, + "grad_norm": 1.523512601852417, + "learning_rate": 9.92155501610249e-05, + "loss": 0.7729, + "step": 2526 + }, + { + "epoch": 0.16965873628401731, + "grad_norm": 1.1168030500411987, + "learning_rate": 9.921363137050696e-05, + "loss": 0.7467, + "step": 2528 + }, + { + "epoch": 0.16979295996778632, + "grad_norm": 1.394824504852295, + "learning_rate": 9.921171025474171e-05, + "loss": 0.7121, + "step": 2530 + }, + { + "epoch": 0.16992718365155532, + "grad_norm": 1.3063008785247803, + "learning_rate": 9.920978681381991e-05, + "loss": 0.7918, + "step": 2532 + }, + { + "epoch": 0.17006140733532432, + "grad_norm": 1.3497661352157593, + "learning_rate": 9.920786104783247e-05, + "loss": 0.7768, + "step": 2534 + }, + { + "epoch": 0.17019563101909332, + "grad_norm": 1.41562819480896, + "learning_rate": 9.920593295687035e-05, + "loss": 0.7616, + "step": 2536 + }, + { + "epoch": 0.17032985470286233, + "grad_norm": 1.2921581268310547, + "learning_rate": 9.920400254102466e-05, + "loss": 0.7841, + "step": 2538 + }, + { + "epoch": 0.17046407838663133, + "grad_norm": 1.0917983055114746, + "learning_rate": 9.92020698003866e-05, + "loss": 0.7187, + "step": 2540 + }, + { + "epoch": 0.17059830207040033, + "grad_norm": 1.3144159317016602, + "learning_rate": 9.92001347350475e-05, + "loss": 0.7462, + "step": 2542 + }, + { + "epoch": 0.17073252575416933, + "grad_norm": 1.0488479137420654, + "learning_rate": 9.919819734509878e-05, + "loss": 0.704, + "step": 2544 + }, + { + "epoch": 0.17086674943793834, + "grad_norm": 1.1551920175552368, + "learning_rate": 9.919625763063197e-05, + "loss": 0.7127, + "step": 2546 + }, + { + "epoch": 0.1710009731217073, + "grad_norm": 1.1880587339401245, + "learning_rate": 9.919431559173874e-05, + "loss": 0.8004, + "step": 2548 + }, + { + "epoch": 0.17113519680547631, + "grad_norm": 1.3170595169067383, + "learning_rate": 9.919237122851084e-05, + "loss": 0.7678, + "step": 2550 + }, + { + "epoch": 0.17126942048924532, + "grad_norm": 1.301263451576233, + "learning_rate": 9.91904245410401e-05, + "loss": 0.7392, + "step": 2552 + }, + { + "epoch": 0.17140364417301432, + "grad_norm": 1.723443627357483, + "learning_rate": 9.918847552941856e-05, + "loss": 0.7954, + "step": 2554 + }, + { + "epoch": 0.17153786785678332, + "grad_norm": 1.2289232015609741, + "learning_rate": 9.918652419373827e-05, + "loss": 0.78, + "step": 2556 + }, + { + "epoch": 0.17167209154055232, + "grad_norm": 1.0448615550994873, + "learning_rate": 9.918457053409142e-05, + "loss": 0.7678, + "step": 2558 + }, + { + "epoch": 0.17180631522432133, + "grad_norm": 1.0582432746887207, + "learning_rate": 9.918261455057036e-05, + "loss": 0.6728, + "step": 2560 + }, + { + "epoch": 0.17194053890809033, + "grad_norm": 1.1932666301727295, + "learning_rate": 9.918065624326745e-05, + "loss": 0.7933, + "step": 2562 + }, + { + "epoch": 0.17207476259185933, + "grad_norm": 1.028738021850586, + "learning_rate": 9.917869561227524e-05, + "loss": 0.685, + "step": 2564 + }, + { + "epoch": 0.17220898627562833, + "grad_norm": 1.0490139722824097, + "learning_rate": 9.917673265768639e-05, + "loss": 0.687, + "step": 2566 + }, + { + "epoch": 0.17234320995939734, + "grad_norm": 1.2028472423553467, + "learning_rate": 9.917476737959361e-05, + "loss": 0.7658, + "step": 2568 + }, + { + "epoch": 0.17247743364316634, + "grad_norm": 2.2009756565093994, + "learning_rate": 9.917279977808976e-05, + "loss": 0.7275, + "step": 2570 + }, + { + "epoch": 0.17261165732693534, + "grad_norm": 1.2236677408218384, + "learning_rate": 9.917082985326782e-05, + "loss": 0.7573, + "step": 2572 + }, + { + "epoch": 0.17274588101070434, + "grad_norm": 1.2253838777542114, + "learning_rate": 9.916885760522087e-05, + "loss": 0.7352, + "step": 2574 + }, + { + "epoch": 0.17288010469447335, + "grad_norm": 1.1593698263168335, + "learning_rate": 9.916688303404208e-05, + "loss": 0.7484, + "step": 2576 + }, + { + "epoch": 0.17301432837824235, + "grad_norm": 1.2147655487060547, + "learning_rate": 9.916490613982474e-05, + "loss": 0.7174, + "step": 2578 + }, + { + "epoch": 0.17314855206201135, + "grad_norm": 1.321434736251831, + "learning_rate": 9.916292692266228e-05, + "loss": 0.7979, + "step": 2580 + }, + { + "epoch": 0.17328277574578035, + "grad_norm": 1.3140137195587158, + "learning_rate": 9.91609453826482e-05, + "loss": 0.7674, + "step": 2582 + }, + { + "epoch": 0.17341699942954936, + "grad_norm": 1.2361663579940796, + "learning_rate": 9.915896151987609e-05, + "loss": 0.704, + "step": 2584 + }, + { + "epoch": 0.17355122311331833, + "grad_norm": 1.2271445989608765, + "learning_rate": 9.915697533443976e-05, + "loss": 0.7282, + "step": 2586 + }, + { + "epoch": 0.17368544679708733, + "grad_norm": 1.3761489391326904, + "learning_rate": 9.915498682643297e-05, + "loss": 0.7634, + "step": 2588 + }, + { + "epoch": 0.17381967048085634, + "grad_norm": 1.174750566482544, + "learning_rate": 9.915299599594973e-05, + "loss": 0.7094, + "step": 2590 + }, + { + "epoch": 0.17395389416462534, + "grad_norm": 1.1923400163650513, + "learning_rate": 9.915100284308407e-05, + "loss": 0.7698, + "step": 2592 + }, + { + "epoch": 0.17408811784839434, + "grad_norm": 1.2211850881576538, + "learning_rate": 9.91490073679302e-05, + "loss": 0.7688, + "step": 2594 + }, + { + "epoch": 0.17422234153216334, + "grad_norm": 1.0700055360794067, + "learning_rate": 9.914700957058235e-05, + "loss": 0.6985, + "step": 2596 + }, + { + "epoch": 0.17435656521593235, + "grad_norm": 1.142652153968811, + "learning_rate": 9.914500945113496e-05, + "loss": 0.7337, + "step": 2598 + }, + { + "epoch": 0.17449078889970135, + "grad_norm": 1.1199309825897217, + "learning_rate": 9.91430070096825e-05, + "loss": 0.7095, + "step": 2600 + }, + { + "epoch": 0.17462501258347035, + "grad_norm": 1.0610016584396362, + "learning_rate": 9.914100224631962e-05, + "loss": 0.7307, + "step": 2602 + }, + { + "epoch": 0.17475923626723935, + "grad_norm": 1.1697499752044678, + "learning_rate": 9.9138995161141e-05, + "loss": 0.7198, + "step": 2604 + }, + { + "epoch": 0.17489345995100836, + "grad_norm": 1.088790774345398, + "learning_rate": 9.913698575424148e-05, + "loss": 0.6844, + "step": 2606 + }, + { + "epoch": 0.17502768363477736, + "grad_norm": 1.5284569263458252, + "learning_rate": 9.913497402571602e-05, + "loss": 0.7349, + "step": 2608 + }, + { + "epoch": 0.17516190731854636, + "grad_norm": 1.1123801469802856, + "learning_rate": 9.913295997565966e-05, + "loss": 0.6767, + "step": 2610 + }, + { + "epoch": 0.17529613100231536, + "grad_norm": 1.2049285173416138, + "learning_rate": 9.913094360416755e-05, + "loss": 0.7509, + "step": 2612 + }, + { + "epoch": 0.17543035468608437, + "grad_norm": 1.16732656955719, + "learning_rate": 9.912892491133496e-05, + "loss": 0.7729, + "step": 2614 + }, + { + "epoch": 0.17556457836985337, + "grad_norm": 1.1883516311645508, + "learning_rate": 9.912690389725727e-05, + "loss": 0.7081, + "step": 2616 + }, + { + "epoch": 0.17569880205362237, + "grad_norm": 1.1990382671356201, + "learning_rate": 9.912488056202998e-05, + "loss": 0.7833, + "step": 2618 + }, + { + "epoch": 0.17583302573739137, + "grad_norm": 1.6179962158203125, + "learning_rate": 9.91228549057487e-05, + "loss": 0.7501, + "step": 2620 + }, + { + "epoch": 0.17596724942116038, + "grad_norm": 1.2496074438095093, + "learning_rate": 9.91208269285091e-05, + "loss": 0.8079, + "step": 2622 + }, + { + "epoch": 0.17610147310492935, + "grad_norm": 2.569533109664917, + "learning_rate": 9.911879663040704e-05, + "loss": 0.7947, + "step": 2624 + }, + { + "epoch": 0.17623569678869835, + "grad_norm": 1.2028528451919556, + "learning_rate": 9.911676401153842e-05, + "loss": 0.7544, + "step": 2626 + }, + { + "epoch": 0.17636992047246736, + "grad_norm": 1.208221435546875, + "learning_rate": 9.911472907199928e-05, + "loss": 0.7546, + "step": 2628 + }, + { + "epoch": 0.17650414415623636, + "grad_norm": 1.1322695016860962, + "learning_rate": 9.911269181188575e-05, + "loss": 0.7467, + "step": 2630 + }, + { + "epoch": 0.17663836784000536, + "grad_norm": 1.445098638534546, + "learning_rate": 9.911065223129414e-05, + "loss": 0.7239, + "step": 2632 + }, + { + "epoch": 0.17677259152377436, + "grad_norm": 1.2913486957550049, + "learning_rate": 9.910861033032079e-05, + "loss": 0.7957, + "step": 2634 + }, + { + "epoch": 0.17690681520754337, + "grad_norm": 1.304382085800171, + "learning_rate": 9.910656610906214e-05, + "loss": 0.7491, + "step": 2636 + }, + { + "epoch": 0.17704103889131237, + "grad_norm": 1.1376643180847168, + "learning_rate": 9.910451956761482e-05, + "loss": 0.7934, + "step": 2638 + }, + { + "epoch": 0.17717526257508137, + "grad_norm": 1.0980056524276733, + "learning_rate": 9.910247070607552e-05, + "loss": 0.7526, + "step": 2640 + }, + { + "epoch": 0.17730948625885037, + "grad_norm": 1.2822057008743286, + "learning_rate": 9.910041952454103e-05, + "loss": 0.7221, + "step": 2642 + }, + { + "epoch": 0.17744370994261938, + "grad_norm": 1.14082932472229, + "learning_rate": 9.909836602310824e-05, + "loss": 0.7206, + "step": 2644 + }, + { + "epoch": 0.17757793362638838, + "grad_norm": 1.1832576990127563, + "learning_rate": 9.909631020187424e-05, + "loss": 0.7794, + "step": 2646 + }, + { + "epoch": 0.17771215731015738, + "grad_norm": 1.0556637048721313, + "learning_rate": 9.909425206093613e-05, + "loss": 0.7347, + "step": 2648 + }, + { + "epoch": 0.17784638099392638, + "grad_norm": 1.192689299583435, + "learning_rate": 9.909219160039112e-05, + "loss": 0.7317, + "step": 2650 + }, + { + "epoch": 0.1779806046776954, + "grad_norm": 1.3593887090682983, + "learning_rate": 9.90901288203366e-05, + "loss": 0.7381, + "step": 2652 + }, + { + "epoch": 0.1781148283614644, + "grad_norm": 1.3047209978103638, + "learning_rate": 9.908806372087002e-05, + "loss": 0.7594, + "step": 2654 + }, + { + "epoch": 0.1782490520452334, + "grad_norm": 1.2017946243286133, + "learning_rate": 9.9085996302089e-05, + "loss": 0.881, + "step": 2656 + }, + { + "epoch": 0.1783832757290024, + "grad_norm": 1.2056033611297607, + "learning_rate": 9.908392656409113e-05, + "loss": 0.6891, + "step": 2658 + }, + { + "epoch": 0.1785174994127714, + "grad_norm": 1.1817090511322021, + "learning_rate": 9.908185450697428e-05, + "loss": 0.7388, + "step": 2660 + }, + { + "epoch": 0.17865172309654037, + "grad_norm": 1.1967507600784302, + "learning_rate": 9.907978013083629e-05, + "loss": 0.7806, + "step": 2662 + }, + { + "epoch": 0.17878594678030937, + "grad_norm": 1.319650411605835, + "learning_rate": 9.907770343577522e-05, + "loss": 0.7853, + "step": 2664 + }, + { + "epoch": 0.17892017046407838, + "grad_norm": 2.1139121055603027, + "learning_rate": 9.907562442188916e-05, + "loss": 0.7972, + "step": 2666 + }, + { + "epoch": 0.17905439414784738, + "grad_norm": 1.6605976819992065, + "learning_rate": 9.907354308927635e-05, + "loss": 0.7991, + "step": 2668 + }, + { + "epoch": 0.17918861783161638, + "grad_norm": 1.182556390762329, + "learning_rate": 9.907145943803513e-05, + "loss": 0.7995, + "step": 2670 + }, + { + "epoch": 0.17932284151538538, + "grad_norm": 1.0889657735824585, + "learning_rate": 9.906937346826395e-05, + "loss": 0.7726, + "step": 2672 + }, + { + "epoch": 0.1794570651991544, + "grad_norm": 1.0999469757080078, + "learning_rate": 9.906728518006136e-05, + "loss": 0.7414, + "step": 2674 + }, + { + "epoch": 0.1795912888829234, + "grad_norm": 1.4655449390411377, + "learning_rate": 9.906519457352605e-05, + "loss": 0.7299, + "step": 2676 + }, + { + "epoch": 0.1797255125666924, + "grad_norm": 1.482847809791565, + "learning_rate": 9.906310164875676e-05, + "loss": 0.7254, + "step": 2678 + }, + { + "epoch": 0.1798597362504614, + "grad_norm": 1.191411018371582, + "learning_rate": 9.90610064058524e-05, + "loss": 0.6878, + "step": 2680 + }, + { + "epoch": 0.1799939599342304, + "grad_norm": 1.2766669988632202, + "learning_rate": 9.905890884491195e-05, + "loss": 0.7835, + "step": 2682 + }, + { + "epoch": 0.1801281836179994, + "grad_norm": 1.205926537513733, + "learning_rate": 9.905680896603455e-05, + "loss": 0.6985, + "step": 2684 + }, + { + "epoch": 0.1802624073017684, + "grad_norm": 1.0893765687942505, + "learning_rate": 9.905470676931938e-05, + "loss": 0.7839, + "step": 2686 + }, + { + "epoch": 0.1803966309855374, + "grad_norm": 42.618675231933594, + "learning_rate": 9.905260225486577e-05, + "loss": 0.7174, + "step": 2688 + }, + { + "epoch": 0.1805308546693064, + "grad_norm": 1.4459474086761475, + "learning_rate": 9.905049542277318e-05, + "loss": 0.7727, + "step": 2690 + }, + { + "epoch": 0.1806650783530754, + "grad_norm": 1.1787508726119995, + "learning_rate": 9.904838627314112e-05, + "loss": 0.7852, + "step": 2692 + }, + { + "epoch": 0.1807993020368444, + "grad_norm": 1.2559529542922974, + "learning_rate": 9.904627480606926e-05, + "loss": 0.78, + "step": 2694 + }, + { + "epoch": 0.18093352572061341, + "grad_norm": 1.2696396112442017, + "learning_rate": 9.904416102165736e-05, + "loss": 0.6922, + "step": 2696 + }, + { + "epoch": 0.18106774940438242, + "grad_norm": 1.4660208225250244, + "learning_rate": 9.90420449200053e-05, + "loss": 0.7632, + "step": 2698 + }, + { + "epoch": 0.1812019730881514, + "grad_norm": 1.0987337827682495, + "learning_rate": 9.903992650121306e-05, + "loss": 0.6834, + "step": 2700 + }, + { + "epoch": 0.1813361967719204, + "grad_norm": 1.1610790491104126, + "learning_rate": 9.903780576538071e-05, + "loss": 0.7703, + "step": 2702 + }, + { + "epoch": 0.1814704204556894, + "grad_norm": 1.284844160079956, + "learning_rate": 9.903568271260847e-05, + "loss": 0.8427, + "step": 2704 + }, + { + "epoch": 0.1816046441394584, + "grad_norm": 1.110213041305542, + "learning_rate": 9.903355734299664e-05, + "loss": 0.6959, + "step": 2706 + }, + { + "epoch": 0.1817388678232274, + "grad_norm": 1.186299443244934, + "learning_rate": 9.903142965664566e-05, + "loss": 0.7515, + "step": 2708 + }, + { + "epoch": 0.1818730915069964, + "grad_norm": 1.2254539728164673, + "learning_rate": 9.902929965365603e-05, + "loss": 0.7767, + "step": 2710 + }, + { + "epoch": 0.1820073151907654, + "grad_norm": 1.0632877349853516, + "learning_rate": 9.902716733412842e-05, + "loss": 0.7166, + "step": 2712 + }, + { + "epoch": 0.1821415388745344, + "grad_norm": 1.3026443719863892, + "learning_rate": 9.902503269816356e-05, + "loss": 0.7088, + "step": 2714 + }, + { + "epoch": 0.1822757625583034, + "grad_norm": 1.2889468669891357, + "learning_rate": 9.90228957458623e-05, + "loss": 0.7798, + "step": 2716 + }, + { + "epoch": 0.1824099862420724, + "grad_norm": 1.083335041999817, + "learning_rate": 9.902075647732563e-05, + "loss": 0.7136, + "step": 2718 + }, + { + "epoch": 0.18254420992584142, + "grad_norm": 1.224172830581665, + "learning_rate": 9.90186148926546e-05, + "loss": 0.7451, + "step": 2720 + }, + { + "epoch": 0.18267843360961042, + "grad_norm": 1.1782457828521729, + "learning_rate": 9.90164709919504e-05, + "loss": 0.8336, + "step": 2722 + }, + { + "epoch": 0.18281265729337942, + "grad_norm": 1.1387137174606323, + "learning_rate": 9.901432477531433e-05, + "loss": 0.7322, + "step": 2724 + }, + { + "epoch": 0.18294688097714842, + "grad_norm": 1.1830766201019287, + "learning_rate": 9.901217624284782e-05, + "loss": 0.8164, + "step": 2726 + }, + { + "epoch": 0.18308110466091743, + "grad_norm": 1.3433092832565308, + "learning_rate": 9.901002539465234e-05, + "loss": 0.8382, + "step": 2728 + }, + { + "epoch": 0.18321532834468643, + "grad_norm": 1.133381962776184, + "learning_rate": 9.900787223082955e-05, + "loss": 0.779, + "step": 2730 + }, + { + "epoch": 0.18334955202845543, + "grad_norm": 1.3183567523956299, + "learning_rate": 9.900571675148116e-05, + "loss": 0.701, + "step": 2732 + }, + { + "epoch": 0.18348377571222443, + "grad_norm": 1.3738702535629272, + "learning_rate": 9.9003558956709e-05, + "loss": 0.7616, + "step": 2734 + }, + { + "epoch": 0.18361799939599344, + "grad_norm": 1.0567593574523926, + "learning_rate": 9.900139884661507e-05, + "loss": 0.7768, + "step": 2736 + }, + { + "epoch": 0.1837522230797624, + "grad_norm": 1.253952145576477, + "learning_rate": 9.899923642130139e-05, + "loss": 0.7426, + "step": 2738 + }, + { + "epoch": 0.1838864467635314, + "grad_norm": 1.1834818124771118, + "learning_rate": 9.899707168087013e-05, + "loss": 0.7258, + "step": 2740 + }, + { + "epoch": 0.18402067044730042, + "grad_norm": 1.1730256080627441, + "learning_rate": 9.89949046254236e-05, + "loss": 0.7649, + "step": 2742 + }, + { + "epoch": 0.18415489413106942, + "grad_norm": 1.2688709497451782, + "learning_rate": 9.899273525506417e-05, + "loss": 0.7731, + "step": 2744 + }, + { + "epoch": 0.18428911781483842, + "grad_norm": 0.988409698009491, + "learning_rate": 9.899056356989434e-05, + "loss": 0.6975, + "step": 2746 + }, + { + "epoch": 0.18442334149860742, + "grad_norm": 1.1317416429519653, + "learning_rate": 9.89883895700167e-05, + "loss": 0.6988, + "step": 2748 + }, + { + "epoch": 0.18455756518237643, + "grad_norm": 1.400775671005249, + "learning_rate": 9.8986213255534e-05, + "loss": 0.8486, + "step": 2750 + }, + { + "epoch": 0.18469178886614543, + "grad_norm": 2.1011314392089844, + "learning_rate": 9.898403462654904e-05, + "loss": 0.8074, + "step": 2752 + }, + { + "epoch": 0.18482601254991443, + "grad_norm": 1.2800923585891724, + "learning_rate": 9.898185368316477e-05, + "loss": 0.7832, + "step": 2754 + }, + { + "epoch": 0.18496023623368343, + "grad_norm": 7.7373270988464355, + "learning_rate": 9.897967042548424e-05, + "loss": 0.7684, + "step": 2756 + }, + { + "epoch": 0.18509445991745244, + "grad_norm": 1.1353752613067627, + "learning_rate": 9.897748485361059e-05, + "loss": 0.7304, + "step": 2758 + }, + { + "epoch": 0.18522868360122144, + "grad_norm": 2.234422445297241, + "learning_rate": 9.89752969676471e-05, + "loss": 0.7119, + "step": 2760 + }, + { + "epoch": 0.18536290728499044, + "grad_norm": 1.1421473026275635, + "learning_rate": 9.897310676769712e-05, + "loss": 0.8281, + "step": 2762 + }, + { + "epoch": 0.18549713096875944, + "grad_norm": 1.2406892776489258, + "learning_rate": 9.897091425386415e-05, + "loss": 0.7532, + "step": 2764 + }, + { + "epoch": 0.18563135465252845, + "grad_norm": 1.4110692739486694, + "learning_rate": 9.896871942625179e-05, + "loss": 0.7723, + "step": 2766 + }, + { + "epoch": 0.18576557833629745, + "grad_norm": 1.292081356048584, + "learning_rate": 9.896652228496372e-05, + "loss": 0.7827, + "step": 2768 + }, + { + "epoch": 0.18589980202006645, + "grad_norm": 1.1562920808792114, + "learning_rate": 9.896432283010376e-05, + "loss": 0.7404, + "step": 2770 + }, + { + "epoch": 0.18603402570383545, + "grad_norm": 1.7341324090957642, + "learning_rate": 9.896212106177583e-05, + "loss": 0.7552, + "step": 2772 + }, + { + "epoch": 0.18616824938760446, + "grad_norm": 1.6327306032180786, + "learning_rate": 9.895991698008397e-05, + "loss": 0.7682, + "step": 2774 + }, + { + "epoch": 0.18630247307137343, + "grad_norm": 1.3916778564453125, + "learning_rate": 9.89577105851323e-05, + "loss": 0.7868, + "step": 2776 + }, + { + "epoch": 0.18643669675514243, + "grad_norm": 1.4612019062042236, + "learning_rate": 9.895550187702506e-05, + "loss": 0.7867, + "step": 2778 + }, + { + "epoch": 0.18657092043891144, + "grad_norm": 1.1375740766525269, + "learning_rate": 9.895329085586667e-05, + "loss": 0.7608, + "step": 2780 + }, + { + "epoch": 0.18670514412268044, + "grad_norm": 1.4708541631698608, + "learning_rate": 9.895107752176152e-05, + "loss": 0.8019, + "step": 2782 + }, + { + "epoch": 0.18683936780644944, + "grad_norm": 1.0651980638504028, + "learning_rate": 9.894886187481421e-05, + "loss": 0.7935, + "step": 2784 + }, + { + "epoch": 0.18697359149021844, + "grad_norm": 1.7364146709442139, + "learning_rate": 9.894664391512943e-05, + "loss": 0.7939, + "step": 2786 + }, + { + "epoch": 0.18710781517398745, + "grad_norm": 1.8116856813430786, + "learning_rate": 9.894442364281197e-05, + "loss": 0.7369, + "step": 2788 + }, + { + "epoch": 0.18724203885775645, + "grad_norm": 1.2194713354110718, + "learning_rate": 9.894220105796676e-05, + "loss": 0.8202, + "step": 2790 + }, + { + "epoch": 0.18737626254152545, + "grad_norm": 1.329644799232483, + "learning_rate": 9.893997616069878e-05, + "loss": 0.7858, + "step": 2792 + }, + { + "epoch": 0.18751048622529445, + "grad_norm": 1.1791480779647827, + "learning_rate": 9.893774895111317e-05, + "loss": 0.7183, + "step": 2794 + }, + { + "epoch": 0.18764470990906346, + "grad_norm": 1.3106368780136108, + "learning_rate": 9.893551942931514e-05, + "loss": 0.7934, + "step": 2796 + }, + { + "epoch": 0.18777893359283246, + "grad_norm": 1.014682412147522, + "learning_rate": 9.893328759541003e-05, + "loss": 0.794, + "step": 2798 + }, + { + "epoch": 0.18791315727660146, + "grad_norm": 1.1085726022720337, + "learning_rate": 9.893105344950333e-05, + "loss": 0.6797, + "step": 2800 + }, + { + "epoch": 0.18804738096037046, + "grad_norm": 1.0696598291397095, + "learning_rate": 9.892881699170058e-05, + "loss": 0.6568, + "step": 2802 + }, + { + "epoch": 0.18818160464413947, + "grad_norm": 1.4964417219161987, + "learning_rate": 9.892657822210742e-05, + "loss": 0.7017, + "step": 2804 + }, + { + "epoch": 0.18831582832790847, + "grad_norm": 1.2987319231033325, + "learning_rate": 9.892433714082966e-05, + "loss": 0.7513, + "step": 2806 + }, + { + "epoch": 0.18845005201167747, + "grad_norm": 1.3627650737762451, + "learning_rate": 9.892209374797318e-05, + "loss": 0.7948, + "step": 2808 + }, + { + "epoch": 0.18858427569544647, + "grad_norm": 1.2367647886276245, + "learning_rate": 9.891984804364395e-05, + "loss": 0.7252, + "step": 2810 + }, + { + "epoch": 0.18871849937921548, + "grad_norm": 1.2189453840255737, + "learning_rate": 9.891760002794812e-05, + "loss": 0.7563, + "step": 2812 + }, + { + "epoch": 0.18885272306298445, + "grad_norm": 1.0545687675476074, + "learning_rate": 9.891534970099188e-05, + "loss": 0.7729, + "step": 2814 + }, + { + "epoch": 0.18898694674675345, + "grad_norm": 1.0687874555587769, + "learning_rate": 9.891309706288154e-05, + "loss": 0.7083, + "step": 2816 + }, + { + "epoch": 0.18912117043052246, + "grad_norm": 1.707055687904358, + "learning_rate": 9.891084211372356e-05, + "loss": 0.8, + "step": 2818 + }, + { + "epoch": 0.18925539411429146, + "grad_norm": 1.1517010927200317, + "learning_rate": 9.890858485362447e-05, + "loss": 0.7723, + "step": 2820 + }, + { + "epoch": 0.18938961779806046, + "grad_norm": 1.7331767082214355, + "learning_rate": 9.89063252826909e-05, + "loss": 0.7459, + "step": 2822 + }, + { + "epoch": 0.18952384148182946, + "grad_norm": 1.2639738321304321, + "learning_rate": 9.890406340102964e-05, + "loss": 0.7863, + "step": 2824 + }, + { + "epoch": 0.18965806516559847, + "grad_norm": 1.186524748802185, + "learning_rate": 9.890179920874756e-05, + "loss": 0.7206, + "step": 2826 + }, + { + "epoch": 0.18979228884936747, + "grad_norm": 1.2795987129211426, + "learning_rate": 9.889953270595162e-05, + "loss": 0.751, + "step": 2828 + }, + { + "epoch": 0.18992651253313647, + "grad_norm": 1.4070851802825928, + "learning_rate": 9.889726389274892e-05, + "loss": 0.7062, + "step": 2830 + }, + { + "epoch": 0.19006073621690547, + "grad_norm": 2.162493944168091, + "learning_rate": 9.889499276924666e-05, + "loss": 0.7934, + "step": 2832 + }, + { + "epoch": 0.19019495990067448, + "grad_norm": 1.2679576873779297, + "learning_rate": 9.889271933555213e-05, + "loss": 0.793, + "step": 2834 + }, + { + "epoch": 0.19032918358444348, + "grad_norm": 1.405962586402893, + "learning_rate": 9.889044359177277e-05, + "loss": 0.7238, + "step": 2836 + }, + { + "epoch": 0.19046340726821248, + "grad_norm": 1.2353532314300537, + "learning_rate": 9.888816553801608e-05, + "loss": 0.7139, + "step": 2838 + }, + { + "epoch": 0.19059763095198148, + "grad_norm": 1.1028170585632324, + "learning_rate": 9.888588517438968e-05, + "loss": 0.7244, + "step": 2840 + }, + { + "epoch": 0.1907318546357505, + "grad_norm": 1.1188403367996216, + "learning_rate": 9.888360250100137e-05, + "loss": 0.7216, + "step": 2842 + }, + { + "epoch": 0.1908660783195195, + "grad_norm": 1.157393455505371, + "learning_rate": 9.888131751795895e-05, + "loss": 0.7636, + "step": 2844 + }, + { + "epoch": 0.1910003020032885, + "grad_norm": 1.4509886503219604, + "learning_rate": 9.88790302253704e-05, + "loss": 0.7643, + "step": 2846 + }, + { + "epoch": 0.1911345256870575, + "grad_norm": 1.1343168020248413, + "learning_rate": 9.887674062334377e-05, + "loss": 0.8027, + "step": 2848 + }, + { + "epoch": 0.19126874937082647, + "grad_norm": 1.3045190572738647, + "learning_rate": 9.88744487119873e-05, + "loss": 0.7703, + "step": 2850 + }, + { + "epoch": 0.19140297305459547, + "grad_norm": 1.1510552167892456, + "learning_rate": 9.88721544914092e-05, + "loss": 0.7222, + "step": 2852 + }, + { + "epoch": 0.19153719673836447, + "grad_norm": 1.2157753705978394, + "learning_rate": 9.886985796171792e-05, + "loss": 0.7933, + "step": 2854 + }, + { + "epoch": 0.19167142042213348, + "grad_norm": 1.9779458045959473, + "learning_rate": 9.886755912302194e-05, + "loss": 0.68, + "step": 2856 + }, + { + "epoch": 0.19180564410590248, + "grad_norm": 1.150827169418335, + "learning_rate": 9.886525797542989e-05, + "loss": 0.761, + "step": 2858 + }, + { + "epoch": 0.19193986778967148, + "grad_norm": 1.3666388988494873, + "learning_rate": 9.88629545190505e-05, + "loss": 0.784, + "step": 2860 + }, + { + "epoch": 0.19207409147344048, + "grad_norm": 1.2915219068527222, + "learning_rate": 9.88606487539926e-05, + "loss": 0.7966, + "step": 2862 + }, + { + "epoch": 0.1922083151572095, + "grad_norm": 1.4001399278640747, + "learning_rate": 9.88583406803651e-05, + "loss": 0.7708, + "step": 2864 + }, + { + "epoch": 0.1923425388409785, + "grad_norm": 1.7778263092041016, + "learning_rate": 9.885603029827707e-05, + "loss": 0.7163, + "step": 2866 + }, + { + "epoch": 0.1924767625247475, + "grad_norm": 1.3420580625534058, + "learning_rate": 9.885371760783772e-05, + "loss": 0.7487, + "step": 2868 + }, + { + "epoch": 0.1926109862085165, + "grad_norm": 1.06627357006073, + "learning_rate": 9.885140260915625e-05, + "loss": 0.7426, + "step": 2870 + }, + { + "epoch": 0.1927452098922855, + "grad_norm": 1.1110459566116333, + "learning_rate": 9.884908530234208e-05, + "loss": 0.7587, + "step": 2872 + }, + { + "epoch": 0.1928794335760545, + "grad_norm": 0.9999328851699829, + "learning_rate": 9.884676568750469e-05, + "loss": 0.7246, + "step": 2874 + }, + { + "epoch": 0.1930136572598235, + "grad_norm": 1.1353189945220947, + "learning_rate": 9.884444376475367e-05, + "loss": 0.7546, + "step": 2876 + }, + { + "epoch": 0.1931478809435925, + "grad_norm": 1.070284128189087, + "learning_rate": 9.884211953419873e-05, + "loss": 0.6763, + "step": 2878 + }, + { + "epoch": 0.1932821046273615, + "grad_norm": 1.1456369161605835, + "learning_rate": 9.883979299594969e-05, + "loss": 0.7325, + "step": 2880 + }, + { + "epoch": 0.1934163283111305, + "grad_norm": 1.1208562850952148, + "learning_rate": 9.883746415011646e-05, + "loss": 0.7168, + "step": 2882 + }, + { + "epoch": 0.1935505519948995, + "grad_norm": 1.173606276512146, + "learning_rate": 9.88351329968091e-05, + "loss": 0.7472, + "step": 2884 + }, + { + "epoch": 0.1936847756786685, + "grad_norm": 1.1535316705703735, + "learning_rate": 9.883279953613771e-05, + "loss": 0.7096, + "step": 2886 + }, + { + "epoch": 0.1938189993624375, + "grad_norm": 1.0993741750717163, + "learning_rate": 9.88304637682126e-05, + "loss": 0.7317, + "step": 2888 + }, + { + "epoch": 0.1939532230462065, + "grad_norm": 1.6121855974197388, + "learning_rate": 9.882812569314408e-05, + "loss": 0.6954, + "step": 2890 + }, + { + "epoch": 0.1940874467299755, + "grad_norm": 1.5288883447647095, + "learning_rate": 9.882578531104263e-05, + "loss": 0.772, + "step": 2892 + }, + { + "epoch": 0.1942216704137445, + "grad_norm": 0.9637719988822937, + "learning_rate": 9.882344262201884e-05, + "loss": 0.72, + "step": 2894 + }, + { + "epoch": 0.1943558940975135, + "grad_norm": 1.0743874311447144, + "learning_rate": 9.882109762618342e-05, + "loss": 0.725, + "step": 2896 + }, + { + "epoch": 0.1944901177812825, + "grad_norm": 1.1397347450256348, + "learning_rate": 9.88187503236471e-05, + "loss": 0.7995, + "step": 2898 + }, + { + "epoch": 0.1946243414650515, + "grad_norm": 1.1381994485855103, + "learning_rate": 9.881640071452085e-05, + "loss": 0.7373, + "step": 2900 + }, + { + "epoch": 0.1947585651488205, + "grad_norm": 2.827315330505371, + "learning_rate": 9.881404879891565e-05, + "loss": 0.7867, + "step": 2902 + }, + { + "epoch": 0.1948927888325895, + "grad_norm": 1.2178807258605957, + "learning_rate": 9.881169457694263e-05, + "loss": 0.7632, + "step": 2904 + }, + { + "epoch": 0.1950270125163585, + "grad_norm": 1.188071846961975, + "learning_rate": 9.880933804871304e-05, + "loss": 0.7936, + "step": 2906 + }, + { + "epoch": 0.1951612362001275, + "grad_norm": 1.5488351583480835, + "learning_rate": 9.88069792143382e-05, + "loss": 0.7501, + "step": 2908 + }, + { + "epoch": 0.19529545988389652, + "grad_norm": 1.2302963733673096, + "learning_rate": 9.880461807392956e-05, + "loss": 0.7269, + "step": 2910 + }, + { + "epoch": 0.19542968356766552, + "grad_norm": 1.212137222290039, + "learning_rate": 9.880225462759869e-05, + "loss": 0.7771, + "step": 2912 + }, + { + "epoch": 0.19556390725143452, + "grad_norm": 1.40646493434906, + "learning_rate": 9.879988887545726e-05, + "loss": 0.7424, + "step": 2914 + }, + { + "epoch": 0.19569813093520352, + "grad_norm": 1.6506620645523071, + "learning_rate": 9.879752081761704e-05, + "loss": 0.719, + "step": 2916 + }, + { + "epoch": 0.19583235461897253, + "grad_norm": 1.1950587034225464, + "learning_rate": 9.87951504541899e-05, + "loss": 0.7966, + "step": 2918 + }, + { + "epoch": 0.19596657830274153, + "grad_norm": 1.2168084383010864, + "learning_rate": 9.879277778528788e-05, + "loss": 0.7682, + "step": 2920 + }, + { + "epoch": 0.19610080198651053, + "grad_norm": 1.1859941482543945, + "learning_rate": 9.879040281102303e-05, + "loss": 0.7092, + "step": 2922 + }, + { + "epoch": 0.19623502567027953, + "grad_norm": 1.2239441871643066, + "learning_rate": 9.878802553150762e-05, + "loss": 0.7683, + "step": 2924 + }, + { + "epoch": 0.1963692493540485, + "grad_norm": 1.762573480606079, + "learning_rate": 9.87856459468539e-05, + "loss": 0.7422, + "step": 2926 + }, + { + "epoch": 0.1965034730378175, + "grad_norm": 1.6585900783538818, + "learning_rate": 9.878326405717438e-05, + "loss": 0.7766, + "step": 2928 + }, + { + "epoch": 0.1966376967215865, + "grad_norm": 1.1981463432312012, + "learning_rate": 9.878087986258156e-05, + "loss": 0.7542, + "step": 2930 + }, + { + "epoch": 0.19677192040535552, + "grad_norm": 1.0024964809417725, + "learning_rate": 9.877849336318807e-05, + "loss": 0.7245, + "step": 2932 + }, + { + "epoch": 0.19690614408912452, + "grad_norm": 1.1262283325195312, + "learning_rate": 9.877610455910668e-05, + "loss": 0.7791, + "step": 2934 + }, + { + "epoch": 0.19704036777289352, + "grad_norm": 1.1668745279312134, + "learning_rate": 9.877371345045029e-05, + "loss": 0.7428, + "step": 2936 + }, + { + "epoch": 0.19717459145666252, + "grad_norm": 1.3850470781326294, + "learning_rate": 9.877132003733182e-05, + "loss": 0.6995, + "step": 2938 + }, + { + "epoch": 0.19730881514043153, + "grad_norm": 1.1988468170166016, + "learning_rate": 9.876892431986442e-05, + "loss": 0.7714, + "step": 2940 + }, + { + "epoch": 0.19744303882420053, + "grad_norm": 1.1874380111694336, + "learning_rate": 9.876652629816122e-05, + "loss": 0.7281, + "step": 2942 + }, + { + "epoch": 0.19757726250796953, + "grad_norm": 1.0807650089263916, + "learning_rate": 9.876412597233555e-05, + "loss": 0.8305, + "step": 2944 + }, + { + "epoch": 0.19771148619173853, + "grad_norm": 1.122131586074829, + "learning_rate": 9.876172334250082e-05, + "loss": 0.7169, + "step": 2946 + }, + { + "epoch": 0.19784570987550754, + "grad_norm": 1.265494704246521, + "learning_rate": 9.875931840877055e-05, + "loss": 0.7753, + "step": 2948 + }, + { + "epoch": 0.19797993355927654, + "grad_norm": 1.19541597366333, + "learning_rate": 9.875691117125837e-05, + "loss": 0.7584, + "step": 2950 + }, + { + "epoch": 0.19811415724304554, + "grad_norm": 1.149675965309143, + "learning_rate": 9.875450163007801e-05, + "loss": 0.7827, + "step": 2952 + }, + { + "epoch": 0.19824838092681454, + "grad_norm": 1.1761040687561035, + "learning_rate": 9.875208978534331e-05, + "loss": 0.6823, + "step": 2954 + }, + { + "epoch": 0.19838260461058355, + "grad_norm": 1.0206571817398071, + "learning_rate": 9.874967563716826e-05, + "loss": 0.6873, + "step": 2956 + }, + { + "epoch": 0.19851682829435255, + "grad_norm": 1.573801040649414, + "learning_rate": 9.874725918566687e-05, + "loss": 0.7422, + "step": 2958 + }, + { + "epoch": 0.19865105197812155, + "grad_norm": 3.296018123626709, + "learning_rate": 9.874484043095336e-05, + "loss": 0.7217, + "step": 2960 + }, + { + "epoch": 0.19878527566189055, + "grad_norm": 2.1112844944000244, + "learning_rate": 9.874241937314199e-05, + "loss": 0.6715, + "step": 2962 + }, + { + "epoch": 0.19891949934565953, + "grad_norm": 1.3631985187530518, + "learning_rate": 9.873999601234715e-05, + "loss": 0.7548, + "step": 2964 + }, + { + "epoch": 0.19905372302942853, + "grad_norm": 1.7285358905792236, + "learning_rate": 9.873757034868333e-05, + "loss": 0.7102, + "step": 2966 + }, + { + "epoch": 0.19918794671319753, + "grad_norm": 1.4606434106826782, + "learning_rate": 9.873514238226515e-05, + "loss": 0.7486, + "step": 2968 + }, + { + "epoch": 0.19932217039696654, + "grad_norm": 1.2391377687454224, + "learning_rate": 9.873271211320735e-05, + "loss": 0.7302, + "step": 2970 + }, + { + "epoch": 0.19945639408073554, + "grad_norm": 1.065616250038147, + "learning_rate": 9.873027954162471e-05, + "loss": 0.7424, + "step": 2972 + }, + { + "epoch": 0.19959061776450454, + "grad_norm": 1.0928055047988892, + "learning_rate": 9.87278446676322e-05, + "loss": 0.7631, + "step": 2974 + }, + { + "epoch": 0.19972484144827354, + "grad_norm": 1.2849832773208618, + "learning_rate": 9.872540749134484e-05, + "loss": 0.8011, + "step": 2976 + }, + { + "epoch": 0.19985906513204255, + "grad_norm": 1.105771780014038, + "learning_rate": 9.872296801287779e-05, + "loss": 0.8108, + "step": 2978 + }, + { + "epoch": 0.19999328881581155, + "grad_norm": 1.8333889245986938, + "learning_rate": 9.872052623234632e-05, + "loss": 0.7429, + "step": 2980 + }, + { + "epoch": 0.20012751249958055, + "grad_norm": 1.1117559671401978, + "learning_rate": 9.871808214986578e-05, + "loss": 0.6958, + "step": 2982 + }, + { + "epoch": 0.20026173618334955, + "grad_norm": 1.1552224159240723, + "learning_rate": 9.871563576555165e-05, + "loss": 0.7331, + "step": 2984 + }, + { + "epoch": 0.20039595986711856, + "grad_norm": 1.2907450199127197, + "learning_rate": 9.871318707951953e-05, + "loss": 0.7443, + "step": 2986 + }, + { + "epoch": 0.20053018355088756, + "grad_norm": 1.1583023071289062, + "learning_rate": 9.871073609188513e-05, + "loss": 0.8166, + "step": 2988 + }, + { + "epoch": 0.20066440723465656, + "grad_norm": 1.2795884609222412, + "learning_rate": 9.87082828027642e-05, + "loss": 0.7008, + "step": 2990 + }, + { + "epoch": 0.20079863091842556, + "grad_norm": 1.3416017293930054, + "learning_rate": 9.870582721227273e-05, + "loss": 0.822, + "step": 2992 + }, + { + "epoch": 0.20093285460219457, + "grad_norm": 1.2741094827651978, + "learning_rate": 9.870336932052667e-05, + "loss": 0.7139, + "step": 2994 + }, + { + "epoch": 0.20106707828596357, + "grad_norm": 1.5517258644104004, + "learning_rate": 9.87009091276422e-05, + "loss": 0.7441, + "step": 2996 + }, + { + "epoch": 0.20120130196973257, + "grad_norm": 1.0434761047363281, + "learning_rate": 9.869844663373553e-05, + "loss": 0.7364, + "step": 2998 + }, + { + "epoch": 0.20133552565350157, + "grad_norm": 1.3809441328048706, + "learning_rate": 9.8695981838923e-05, + "loss": 0.7398, + "step": 3000 + }, + { + "epoch": 0.20146974933727055, + "grad_norm": 1.060219168663025, + "learning_rate": 9.869351474332111e-05, + "loss": 0.7235, + "step": 3002 + }, + { + "epoch": 0.20160397302103955, + "grad_norm": 1.6353628635406494, + "learning_rate": 9.869104534704641e-05, + "loss": 0.7376, + "step": 3004 + }, + { + "epoch": 0.20173819670480855, + "grad_norm": 1.215761661529541, + "learning_rate": 9.868857365021553e-05, + "loss": 0.8069, + "step": 3006 + }, + { + "epoch": 0.20187242038857756, + "grad_norm": 1.2343037128448486, + "learning_rate": 9.86860996529453e-05, + "loss": 0.72, + "step": 3008 + }, + { + "epoch": 0.20200664407234656, + "grad_norm": 1.1783651113510132, + "learning_rate": 9.868362335535262e-05, + "loss": 0.7734, + "step": 3010 + }, + { + "epoch": 0.20214086775611556, + "grad_norm": 1.0304306745529175, + "learning_rate": 9.868114475755445e-05, + "loss": 0.6904, + "step": 3012 + }, + { + "epoch": 0.20227509143988456, + "grad_norm": 1.1342130899429321, + "learning_rate": 9.867866385966791e-05, + "loss": 0.7364, + "step": 3014 + }, + { + "epoch": 0.20240931512365357, + "grad_norm": 1.1649483442306519, + "learning_rate": 9.867618066181023e-05, + "loss": 0.7879, + "step": 3016 + }, + { + "epoch": 0.20254353880742257, + "grad_norm": 1.2916783094406128, + "learning_rate": 9.867369516409874e-05, + "loss": 0.738, + "step": 3018 + }, + { + "epoch": 0.20267776249119157, + "grad_norm": 1.258411169052124, + "learning_rate": 9.867120736665087e-05, + "loss": 0.7294, + "step": 3020 + }, + { + "epoch": 0.20281198617496057, + "grad_norm": 1.307192087173462, + "learning_rate": 9.866871726958415e-05, + "loss": 0.8615, + "step": 3022 + }, + { + "epoch": 0.20294620985872958, + "grad_norm": 1.1338318586349487, + "learning_rate": 9.866622487301624e-05, + "loss": 0.6777, + "step": 3024 + }, + { + "epoch": 0.20308043354249858, + "grad_norm": 1.2602858543395996, + "learning_rate": 9.866373017706492e-05, + "loss": 0.7936, + "step": 3026 + }, + { + "epoch": 0.20321465722626758, + "grad_norm": 0.9843763113021851, + "learning_rate": 9.866123318184803e-05, + "loss": 0.6511, + "step": 3028 + }, + { + "epoch": 0.20334888091003658, + "grad_norm": 1.1347159147262573, + "learning_rate": 9.865873388748354e-05, + "loss": 0.7671, + "step": 3030 + }, + { + "epoch": 0.20348310459380559, + "grad_norm": 1.4263908863067627, + "learning_rate": 9.865623229408959e-05, + "loss": 0.7339, + "step": 3032 + }, + { + "epoch": 0.2036173282775746, + "grad_norm": 1.1282978057861328, + "learning_rate": 9.865372840178433e-05, + "loss": 0.7729, + "step": 3034 + }, + { + "epoch": 0.2037515519613436, + "grad_norm": 1.1372108459472656, + "learning_rate": 9.865122221068608e-05, + "loss": 0.7619, + "step": 3036 + }, + { + "epoch": 0.2038857756451126, + "grad_norm": 1.1885900497436523, + "learning_rate": 9.864871372091324e-05, + "loss": 0.7027, + "step": 3038 + }, + { + "epoch": 0.20401999932888157, + "grad_norm": 1.1478683948516846, + "learning_rate": 9.864620293258434e-05, + "loss": 0.7334, + "step": 3040 + }, + { + "epoch": 0.20415422301265057, + "grad_norm": 1.4754353761672974, + "learning_rate": 9.864368984581803e-05, + "loss": 0.6868, + "step": 3042 + }, + { + "epoch": 0.20428844669641957, + "grad_norm": 1.1996697187423706, + "learning_rate": 9.8641174460733e-05, + "loss": 0.7198, + "step": 3044 + }, + { + "epoch": 0.20442267038018858, + "grad_norm": 1.235029697418213, + "learning_rate": 9.863865677744814e-05, + "loss": 0.7288, + "step": 3046 + }, + { + "epoch": 0.20455689406395758, + "grad_norm": 1.098764419555664, + "learning_rate": 9.863613679608239e-05, + "loss": 0.6877, + "step": 3048 + }, + { + "epoch": 0.20469111774772658, + "grad_norm": 1.268288254737854, + "learning_rate": 9.863361451675481e-05, + "loss": 0.817, + "step": 3050 + }, + { + "epoch": 0.20482534143149558, + "grad_norm": 1.1356425285339355, + "learning_rate": 9.863108993958459e-05, + "loss": 0.7465, + "step": 3052 + }, + { + "epoch": 0.20495956511526459, + "grad_norm": 1.187319278717041, + "learning_rate": 9.862856306469099e-05, + "loss": 0.7622, + "step": 3054 + }, + { + "epoch": 0.2050937887990336, + "grad_norm": 1.3413240909576416, + "learning_rate": 9.86260338921934e-05, + "loss": 0.6907, + "step": 3056 + }, + { + "epoch": 0.2052280124828026, + "grad_norm": 1.3412761688232422, + "learning_rate": 9.862350242221135e-05, + "loss": 0.7947, + "step": 3058 + }, + { + "epoch": 0.2053622361665716, + "grad_norm": 1.2627809047698975, + "learning_rate": 9.862096865486441e-05, + "loss": 0.7168, + "step": 3060 + }, + { + "epoch": 0.2054964598503406, + "grad_norm": 1.1120837926864624, + "learning_rate": 9.861843259027233e-05, + "loss": 0.7121, + "step": 3062 + }, + { + "epoch": 0.2056306835341096, + "grad_norm": 1.2272433042526245, + "learning_rate": 9.861589422855488e-05, + "loss": 0.7741, + "step": 3064 + }, + { + "epoch": 0.2057649072178786, + "grad_norm": 2.011345863342285, + "learning_rate": 9.861335356983206e-05, + "loss": 0.7038, + "step": 3066 + }, + { + "epoch": 0.2058991309016476, + "grad_norm": 1.1610370874404907, + "learning_rate": 9.861081061422386e-05, + "loss": 0.7797, + "step": 3068 + }, + { + "epoch": 0.2060333545854166, + "grad_norm": 1.5199135541915894, + "learning_rate": 9.860826536185044e-05, + "loss": 0.7256, + "step": 3070 + }, + { + "epoch": 0.2061675782691856, + "grad_norm": 0.9823091626167297, + "learning_rate": 9.860571781283208e-05, + "loss": 0.7143, + "step": 3072 + }, + { + "epoch": 0.2063018019529546, + "grad_norm": 1.1092530488967896, + "learning_rate": 9.860316796728912e-05, + "loss": 0.7797, + "step": 3074 + }, + { + "epoch": 0.2064360256367236, + "grad_norm": 1.133023977279663, + "learning_rate": 9.860061582534205e-05, + "loss": 0.7105, + "step": 3076 + }, + { + "epoch": 0.2065702493204926, + "grad_norm": 1.3843505382537842, + "learning_rate": 9.859806138711148e-05, + "loss": 0.7875, + "step": 3078 + }, + { + "epoch": 0.2067044730042616, + "grad_norm": 1.0685838460922241, + "learning_rate": 9.859550465271804e-05, + "loss": 0.6607, + "step": 3080 + }, + { + "epoch": 0.2068386966880306, + "grad_norm": 1.1929254531860352, + "learning_rate": 9.859294562228258e-05, + "loss": 0.8085, + "step": 3082 + }, + { + "epoch": 0.2069729203717996, + "grad_norm": 2.1357104778289795, + "learning_rate": 9.859038429592599e-05, + "loss": 0.7222, + "step": 3084 + }, + { + "epoch": 0.2071071440555686, + "grad_norm": 1.1408915519714355, + "learning_rate": 9.858782067376928e-05, + "loss": 0.6628, + "step": 3086 + }, + { + "epoch": 0.2072413677393376, + "grad_norm": 1.2059365510940552, + "learning_rate": 9.85852547559336e-05, + "loss": 0.8336, + "step": 3088 + }, + { + "epoch": 0.2073755914231066, + "grad_norm": 1.0125360488891602, + "learning_rate": 9.858268654254017e-05, + "loss": 0.7075, + "step": 3090 + }, + { + "epoch": 0.2075098151068756, + "grad_norm": 1.3566681146621704, + "learning_rate": 9.858011603371033e-05, + "loss": 0.7074, + "step": 3092 + }, + { + "epoch": 0.2076440387906446, + "grad_norm": 1.080464482307434, + "learning_rate": 9.857754322956554e-05, + "loss": 0.696, + "step": 3094 + }, + { + "epoch": 0.2077782624744136, + "grad_norm": 1.0486165285110474, + "learning_rate": 9.857496813022735e-05, + "loss": 0.7453, + "step": 3096 + }, + { + "epoch": 0.2079124861581826, + "grad_norm": 1.0846431255340576, + "learning_rate": 9.857239073581743e-05, + "loss": 0.7727, + "step": 3098 + }, + { + "epoch": 0.20804670984195162, + "grad_norm": 1.1815550327301025, + "learning_rate": 9.856981104645757e-05, + "loss": 0.7271, + "step": 3100 + }, + { + "epoch": 0.20818093352572062, + "grad_norm": 1.200812578201294, + "learning_rate": 9.856722906226965e-05, + "loss": 0.7766, + "step": 3102 + }, + { + "epoch": 0.20831515720948962, + "grad_norm": 1.3870935440063477, + "learning_rate": 9.856464478337566e-05, + "loss": 0.7684, + "step": 3104 + }, + { + "epoch": 0.20844938089325862, + "grad_norm": 2.0267977714538574, + "learning_rate": 9.85620582098977e-05, + "loss": 0.727, + "step": 3106 + }, + { + "epoch": 0.20858360457702763, + "grad_norm": 1.2615206241607666, + "learning_rate": 9.855946934195799e-05, + "loss": 0.7919, + "step": 3108 + }, + { + "epoch": 0.20871782826079663, + "grad_norm": 1.1016186475753784, + "learning_rate": 9.855687817967882e-05, + "loss": 0.7548, + "step": 3110 + }, + { + "epoch": 0.20885205194456563, + "grad_norm": 0.9793462157249451, + "learning_rate": 9.855428472318267e-05, + "loss": 0.7518, + "step": 3112 + }, + { + "epoch": 0.20898627562833463, + "grad_norm": 1.139752745628357, + "learning_rate": 9.855168897259202e-05, + "loss": 0.7284, + "step": 3114 + }, + { + "epoch": 0.2091204993121036, + "grad_norm": 1.2097362279891968, + "learning_rate": 9.854909092802955e-05, + "loss": 0.7188, + "step": 3116 + }, + { + "epoch": 0.2092547229958726, + "grad_norm": 1.1110161542892456, + "learning_rate": 9.854649058961799e-05, + "loss": 0.7279, + "step": 3118 + }, + { + "epoch": 0.2093889466796416, + "grad_norm": 1.1635180711746216, + "learning_rate": 9.854388795748022e-05, + "loss": 0.7384, + "step": 3120 + }, + { + "epoch": 0.20952317036341062, + "grad_norm": 1.0395607948303223, + "learning_rate": 9.854128303173919e-05, + "loss": 0.6984, + "step": 3122 + }, + { + "epoch": 0.20965739404717962, + "grad_norm": 1.2108814716339111, + "learning_rate": 9.8538675812518e-05, + "loss": 0.7615, + "step": 3124 + }, + { + "epoch": 0.20979161773094862, + "grad_norm": 1.134364128112793, + "learning_rate": 9.853606629993983e-05, + "loss": 0.7576, + "step": 3126 + }, + { + "epoch": 0.20992584141471762, + "grad_norm": 1.1714712381362915, + "learning_rate": 9.853345449412796e-05, + "loss": 0.6579, + "step": 3128 + }, + { + "epoch": 0.21006006509848663, + "grad_norm": 1.1478958129882812, + "learning_rate": 9.853084039520581e-05, + "loss": 0.7723, + "step": 3130 + }, + { + "epoch": 0.21019428878225563, + "grad_norm": 1.2991362810134888, + "learning_rate": 9.852822400329688e-05, + "loss": 0.7406, + "step": 3132 + }, + { + "epoch": 0.21032851246602463, + "grad_norm": 1.1406856775283813, + "learning_rate": 9.852560531852479e-05, + "loss": 0.7334, + "step": 3134 + }, + { + "epoch": 0.21046273614979363, + "grad_norm": 1.0486959218978882, + "learning_rate": 9.852298434101328e-05, + "loss": 0.7012, + "step": 3136 + }, + { + "epoch": 0.21059695983356264, + "grad_norm": 1.4187973737716675, + "learning_rate": 9.852036107088617e-05, + "loss": 0.7311, + "step": 3138 + }, + { + "epoch": 0.21073118351733164, + "grad_norm": 1.5801842212677002, + "learning_rate": 9.851773550826742e-05, + "loss": 0.7044, + "step": 3140 + }, + { + "epoch": 0.21086540720110064, + "grad_norm": 1.2946140766143799, + "learning_rate": 9.851510765328105e-05, + "loss": 0.7428, + "step": 3142 + }, + { + "epoch": 0.21099963088486964, + "grad_norm": 1.1796568632125854, + "learning_rate": 9.851247750605126e-05, + "loss": 0.7211, + "step": 3144 + }, + { + "epoch": 0.21113385456863865, + "grad_norm": 1.0621049404144287, + "learning_rate": 9.85098450667023e-05, + "loss": 0.7527, + "step": 3146 + }, + { + "epoch": 0.21126807825240765, + "grad_norm": 1.4693660736083984, + "learning_rate": 9.850721033535854e-05, + "loss": 0.759, + "step": 3148 + }, + { + "epoch": 0.21140230193617665, + "grad_norm": 1.233821153640747, + "learning_rate": 9.85045733121445e-05, + "loss": 0.7018, + "step": 3150 + }, + { + "epoch": 0.21153652561994563, + "grad_norm": 1.652405023574829, + "learning_rate": 9.850193399718475e-05, + "loss": 0.7492, + "step": 3152 + }, + { + "epoch": 0.21167074930371463, + "grad_norm": 1.35104501247406, + "learning_rate": 9.849929239060398e-05, + "loss": 0.7357, + "step": 3154 + }, + { + "epoch": 0.21180497298748363, + "grad_norm": 1.1674814224243164, + "learning_rate": 9.849664849252701e-05, + "loss": 0.7184, + "step": 3156 + }, + { + "epoch": 0.21193919667125263, + "grad_norm": 1.2722527980804443, + "learning_rate": 9.849400230307877e-05, + "loss": 0.7412, + "step": 3158 + }, + { + "epoch": 0.21207342035502164, + "grad_norm": 1.0681673288345337, + "learning_rate": 9.849135382238428e-05, + "loss": 0.7271, + "step": 3160 + }, + { + "epoch": 0.21220764403879064, + "grad_norm": 1.156601071357727, + "learning_rate": 9.848870305056867e-05, + "loss": 0.7621, + "step": 3162 + }, + { + "epoch": 0.21234186772255964, + "grad_norm": 1.1709725856781006, + "learning_rate": 9.848604998775721e-05, + "loss": 0.7921, + "step": 3164 + }, + { + "epoch": 0.21247609140632864, + "grad_norm": 1.0551742315292358, + "learning_rate": 9.848339463407521e-05, + "loss": 0.6156, + "step": 3166 + }, + { + "epoch": 0.21261031509009765, + "grad_norm": 1.2272999286651611, + "learning_rate": 9.848073698964817e-05, + "loss": 0.6912, + "step": 3168 + }, + { + "epoch": 0.21274453877386665, + "grad_norm": 1.103464126586914, + "learning_rate": 9.847807705460163e-05, + "loss": 0.7523, + "step": 3170 + }, + { + "epoch": 0.21287876245763565, + "grad_norm": 1.215071439743042, + "learning_rate": 9.847541482906129e-05, + "loss": 0.7767, + "step": 3172 + }, + { + "epoch": 0.21301298614140465, + "grad_norm": 1.2491865158081055, + "learning_rate": 9.84727503131529e-05, + "loss": 0.7524, + "step": 3174 + }, + { + "epoch": 0.21314720982517366, + "grad_norm": 1.4180876016616821, + "learning_rate": 9.847008350700239e-05, + "loss": 0.767, + "step": 3176 + }, + { + "epoch": 0.21328143350894266, + "grad_norm": 1.2222609519958496, + "learning_rate": 9.846741441073574e-05, + "loss": 0.7615, + "step": 3178 + }, + { + "epoch": 0.21341565719271166, + "grad_norm": 1.1580231189727783, + "learning_rate": 9.846474302447907e-05, + "loss": 0.7834, + "step": 3180 + }, + { + "epoch": 0.21354988087648066, + "grad_norm": 1.3348591327667236, + "learning_rate": 9.846206934835859e-05, + "loss": 0.7133, + "step": 3182 + }, + { + "epoch": 0.21368410456024967, + "grad_norm": 1.0967212915420532, + "learning_rate": 9.845939338250063e-05, + "loss": 0.707, + "step": 3184 + }, + { + "epoch": 0.21381832824401867, + "grad_norm": 1.2718641757965088, + "learning_rate": 9.845671512703163e-05, + "loss": 0.7405, + "step": 3186 + }, + { + "epoch": 0.21395255192778767, + "grad_norm": 1.4138965606689453, + "learning_rate": 9.845403458207813e-05, + "loss": 0.7712, + "step": 3188 + }, + { + "epoch": 0.21408677561155665, + "grad_norm": 6.629738807678223, + "learning_rate": 9.845135174776676e-05, + "loss": 0.7772, + "step": 3190 + }, + { + "epoch": 0.21422099929532565, + "grad_norm": 1.3174983263015747, + "learning_rate": 9.844866662422432e-05, + "loss": 0.7585, + "step": 3192 + }, + { + "epoch": 0.21435522297909465, + "grad_norm": 1.3816879987716675, + "learning_rate": 9.844597921157764e-05, + "loss": 0.7098, + "step": 3194 + }, + { + "epoch": 0.21448944666286365, + "grad_norm": 1.203576683998108, + "learning_rate": 9.84432895099537e-05, + "loss": 0.8389, + "step": 3196 + }, + { + "epoch": 0.21462367034663266, + "grad_norm": 1.1303554773330688, + "learning_rate": 9.844059751947959e-05, + "loss": 0.7791, + "step": 3198 + }, + { + "epoch": 0.21475789403040166, + "grad_norm": 1.2763458490371704, + "learning_rate": 9.84379032402825e-05, + "loss": 0.6964, + "step": 3200 + }, + { + "epoch": 0.21489211771417066, + "grad_norm": 1.4167712926864624, + "learning_rate": 9.843520667248974e-05, + "loss": 0.6939, + "step": 3202 + }, + { + "epoch": 0.21502634139793966, + "grad_norm": 1.0749961137771606, + "learning_rate": 9.84325078162287e-05, + "loss": 0.7213, + "step": 3204 + }, + { + "epoch": 0.21516056508170867, + "grad_norm": 1.409028172492981, + "learning_rate": 9.84298066716269e-05, + "loss": 0.707, + "step": 3206 + }, + { + "epoch": 0.21529478876547767, + "grad_norm": 2.0518510341644287, + "learning_rate": 9.842710323881199e-05, + "loss": 0.7577, + "step": 3208 + }, + { + "epoch": 0.21542901244924667, + "grad_norm": 0.9805545806884766, + "learning_rate": 9.842439751791169e-05, + "loss": 0.7061, + "step": 3210 + }, + { + "epoch": 0.21556323613301567, + "grad_norm": 1.0923629999160767, + "learning_rate": 9.842168950905379e-05, + "loss": 0.6747, + "step": 3212 + }, + { + "epoch": 0.21569745981678468, + "grad_norm": 1.0803589820861816, + "learning_rate": 9.84189792123663e-05, + "loss": 0.6973, + "step": 3214 + }, + { + "epoch": 0.21583168350055368, + "grad_norm": 1.091895580291748, + "learning_rate": 9.841626662797725e-05, + "loss": 0.7432, + "step": 3216 + }, + { + "epoch": 0.21596590718432268, + "grad_norm": 1.1444942951202393, + "learning_rate": 9.841355175601481e-05, + "loss": 0.744, + "step": 3218 + }, + { + "epoch": 0.21610013086809168, + "grad_norm": 1.2099897861480713, + "learning_rate": 9.841083459660725e-05, + "loss": 0.7323, + "step": 3220 + }, + { + "epoch": 0.21623435455186069, + "grad_norm": 1.3110309839248657, + "learning_rate": 9.840811514988294e-05, + "loss": 0.7389, + "step": 3222 + }, + { + "epoch": 0.2163685782356297, + "grad_norm": 1.2931196689605713, + "learning_rate": 9.840539341597039e-05, + "loss": 0.7125, + "step": 3224 + }, + { + "epoch": 0.2165028019193987, + "grad_norm": 1.1830400228500366, + "learning_rate": 9.840266939499818e-05, + "loss": 0.7337, + "step": 3226 + }, + { + "epoch": 0.21663702560316767, + "grad_norm": 1.1818979978561401, + "learning_rate": 9.839994308709504e-05, + "loss": 0.6791, + "step": 3228 + }, + { + "epoch": 0.21677124928693667, + "grad_norm": 1.1020506620407104, + "learning_rate": 9.839721449238974e-05, + "loss": 0.6538, + "step": 3230 + }, + { + "epoch": 0.21690547297070567, + "grad_norm": 1.7511346340179443, + "learning_rate": 9.839448361101124e-05, + "loss": 0.7263, + "step": 3232 + }, + { + "epoch": 0.21703969665447467, + "grad_norm": 1.1966168880462646, + "learning_rate": 9.839175044308854e-05, + "loss": 0.7244, + "step": 3234 + }, + { + "epoch": 0.21717392033824368, + "grad_norm": 1.5103763341903687, + "learning_rate": 9.838901498875081e-05, + "loss": 0.7085, + "step": 3236 + }, + { + "epoch": 0.21730814402201268, + "grad_norm": 1.7880642414093018, + "learning_rate": 9.838627724812725e-05, + "loss": 0.7046, + "step": 3238 + }, + { + "epoch": 0.21744236770578168, + "grad_norm": 1.1055731773376465, + "learning_rate": 9.838353722134725e-05, + "loss": 0.7563, + "step": 3240 + }, + { + "epoch": 0.21757659138955068, + "grad_norm": 1.201663851737976, + "learning_rate": 9.838079490854027e-05, + "loss": 0.7018, + "step": 3242 + }, + { + "epoch": 0.21771081507331969, + "grad_norm": 1.2205414772033691, + "learning_rate": 9.837805030983585e-05, + "loss": 0.7647, + "step": 3244 + }, + { + "epoch": 0.2178450387570887, + "grad_norm": 1.239234447479248, + "learning_rate": 9.837530342536368e-05, + "loss": 0.7548, + "step": 3246 + }, + { + "epoch": 0.2179792624408577, + "grad_norm": 1.5525140762329102, + "learning_rate": 9.837255425525356e-05, + "loss": 0.7301, + "step": 3248 + }, + { + "epoch": 0.2181134861246267, + "grad_norm": 1.1218043565750122, + "learning_rate": 9.836980279963537e-05, + "loss": 0.7664, + "step": 3250 + }, + { + "epoch": 0.2182477098083957, + "grad_norm": 1.2930965423583984, + "learning_rate": 9.836704905863911e-05, + "loss": 0.715, + "step": 3252 + }, + { + "epoch": 0.2183819334921647, + "grad_norm": 1.274599552154541, + "learning_rate": 9.836429303239491e-05, + "loss": 0.7677, + "step": 3254 + }, + { + "epoch": 0.2185161571759337, + "grad_norm": 1.1840789318084717, + "learning_rate": 9.836153472103296e-05, + "loss": 0.7674, + "step": 3256 + }, + { + "epoch": 0.2186503808597027, + "grad_norm": 1.3341907262802124, + "learning_rate": 9.835877412468357e-05, + "loss": 0.7489, + "step": 3258 + }, + { + "epoch": 0.2187846045434717, + "grad_norm": 1.1839338541030884, + "learning_rate": 9.835601124347722e-05, + "loss": 0.6916, + "step": 3260 + }, + { + "epoch": 0.2189188282272407, + "grad_norm": 1.022170901298523, + "learning_rate": 9.835324607754442e-05, + "loss": 0.689, + "step": 3262 + }, + { + "epoch": 0.2190530519110097, + "grad_norm": 1.1673866510391235, + "learning_rate": 9.835047862701583e-05, + "loss": 0.8045, + "step": 3264 + }, + { + "epoch": 0.21918727559477869, + "grad_norm": 1.1706578731536865, + "learning_rate": 9.834770889202219e-05, + "loss": 0.7545, + "step": 3266 + }, + { + "epoch": 0.2193214992785477, + "grad_norm": 1.2601052522659302, + "learning_rate": 9.834493687269438e-05, + "loss": 0.7145, + "step": 3268 + }, + { + "epoch": 0.2194557229623167, + "grad_norm": 1.116899847984314, + "learning_rate": 9.834216256916337e-05, + "loss": 0.7458, + "step": 3270 + }, + { + "epoch": 0.2195899466460857, + "grad_norm": 1.189274787902832, + "learning_rate": 9.833938598156025e-05, + "loss": 0.7316, + "step": 3272 + }, + { + "epoch": 0.2197241703298547, + "grad_norm": 1.1440093517303467, + "learning_rate": 9.833660711001619e-05, + "loss": 0.7593, + "step": 3274 + }, + { + "epoch": 0.2198583940136237, + "grad_norm": 1.5465357303619385, + "learning_rate": 9.833382595466249e-05, + "loss": 0.6599, + "step": 3276 + }, + { + "epoch": 0.2199926176973927, + "grad_norm": 2.1189398765563965, + "learning_rate": 9.833104251563056e-05, + "loss": 0.6945, + "step": 3278 + }, + { + "epoch": 0.2201268413811617, + "grad_norm": 1.1823686361312866, + "learning_rate": 9.832825679305191e-05, + "loss": 0.7261, + "step": 3280 + }, + { + "epoch": 0.2202610650649307, + "grad_norm": 1.389985203742981, + "learning_rate": 9.832546878705817e-05, + "loss": 0.6992, + "step": 3282 + }, + { + "epoch": 0.2203952887486997, + "grad_norm": 1.2685588598251343, + "learning_rate": 9.832267849778106e-05, + "loss": 0.74, + "step": 3284 + }, + { + "epoch": 0.2205295124324687, + "grad_norm": 1.1789789199829102, + "learning_rate": 9.83198859253524e-05, + "loss": 0.7712, + "step": 3286 + }, + { + "epoch": 0.2206637361162377, + "grad_norm": 1.079740047454834, + "learning_rate": 9.831709106990414e-05, + "loss": 0.7319, + "step": 3288 + }, + { + "epoch": 0.22079795980000672, + "grad_norm": 1.25736665725708, + "learning_rate": 9.831429393156834e-05, + "loss": 0.7148, + "step": 3290 + }, + { + "epoch": 0.22093218348377572, + "grad_norm": 1.6037640571594238, + "learning_rate": 9.831149451047718e-05, + "loss": 0.7842, + "step": 3292 + }, + { + "epoch": 0.22106640716754472, + "grad_norm": 1.8668400049209595, + "learning_rate": 9.83086928067629e-05, + "loss": 0.7163, + "step": 3294 + }, + { + "epoch": 0.22120063085131372, + "grad_norm": 1.6986749172210693, + "learning_rate": 9.830588882055786e-05, + "loss": 0.7139, + "step": 3296 + }, + { + "epoch": 0.22133485453508273, + "grad_norm": 1.2642039060592651, + "learning_rate": 9.830308255199457e-05, + "loss": 0.747, + "step": 3298 + }, + { + "epoch": 0.22146907821885173, + "grad_norm": 1.3078261613845825, + "learning_rate": 9.830027400120561e-05, + "loss": 0.7421, + "step": 3300 + }, + { + "epoch": 0.22160330190262073, + "grad_norm": 1.1431974172592163, + "learning_rate": 9.82974631683237e-05, + "loss": 0.7438, + "step": 3302 + }, + { + "epoch": 0.2217375255863897, + "grad_norm": 1.0984482765197754, + "learning_rate": 9.829465005348162e-05, + "loss": 0.7345, + "step": 3304 + }, + { + "epoch": 0.2218717492701587, + "grad_norm": 1.1009687185287476, + "learning_rate": 9.829183465681229e-05, + "loss": 0.686, + "step": 3306 + }, + { + "epoch": 0.2220059729539277, + "grad_norm": 1.0883309841156006, + "learning_rate": 9.828901697844872e-05, + "loss": 0.6682, + "step": 3308 + }, + { + "epoch": 0.2221401966376967, + "grad_norm": 1.512415885925293, + "learning_rate": 9.828619701852407e-05, + "loss": 0.7897, + "step": 3310 + }, + { + "epoch": 0.22227442032146572, + "grad_norm": 1.1906039714813232, + "learning_rate": 9.828337477717157e-05, + "loss": 0.6654, + "step": 3312 + }, + { + "epoch": 0.22240864400523472, + "grad_norm": 1.738184928894043, + "learning_rate": 9.828055025452454e-05, + "loss": 0.8456, + "step": 3314 + }, + { + "epoch": 0.22254286768900372, + "grad_norm": 1.3221595287322998, + "learning_rate": 9.827772345071647e-05, + "loss": 0.697, + "step": 3316 + }, + { + "epoch": 0.22267709137277272, + "grad_norm": 1.672614336013794, + "learning_rate": 9.827489436588088e-05, + "loss": 0.7487, + "step": 3318 + }, + { + "epoch": 0.22281131505654173, + "grad_norm": 1.4702619314193726, + "learning_rate": 9.827206300015147e-05, + "loss": 0.7818, + "step": 3320 + }, + { + "epoch": 0.22294553874031073, + "grad_norm": 1.2230552434921265, + "learning_rate": 9.8269229353662e-05, + "loss": 0.7932, + "step": 3322 + }, + { + "epoch": 0.22307976242407973, + "grad_norm": 1.1639833450317383, + "learning_rate": 9.826639342654636e-05, + "loss": 0.7041, + "step": 3324 + }, + { + "epoch": 0.22321398610784873, + "grad_norm": 1.333631992340088, + "learning_rate": 9.826355521893855e-05, + "loss": 0.7125, + "step": 3326 + }, + { + "epoch": 0.22334820979161774, + "grad_norm": 1.2048940658569336, + "learning_rate": 9.826071473097265e-05, + "loss": 0.7808, + "step": 3328 + }, + { + "epoch": 0.22348243347538674, + "grad_norm": 1.131531834602356, + "learning_rate": 9.82578719627829e-05, + "loss": 0.7575, + "step": 3330 + }, + { + "epoch": 0.22361665715915574, + "grad_norm": 1.1711511611938477, + "learning_rate": 9.825502691450357e-05, + "loss": 0.6958, + "step": 3332 + }, + { + "epoch": 0.22375088084292474, + "grad_norm": 1.0046534538269043, + "learning_rate": 9.825217958626913e-05, + "loss": 0.7534, + "step": 3334 + }, + { + "epoch": 0.22388510452669375, + "grad_norm": 1.0132005214691162, + "learning_rate": 9.824932997821408e-05, + "loss": 0.6879, + "step": 3336 + }, + { + "epoch": 0.22401932821046275, + "grad_norm": 1.1994860172271729, + "learning_rate": 9.824647809047306e-05, + "loss": 0.7525, + "step": 3338 + }, + { + "epoch": 0.22415355189423175, + "grad_norm": 1.171962022781372, + "learning_rate": 9.824362392318082e-05, + "loss": 0.6529, + "step": 3340 + }, + { + "epoch": 0.22428777557800073, + "grad_norm": 1.1129786968231201, + "learning_rate": 9.824076747647223e-05, + "loss": 0.7097, + "step": 3342 + }, + { + "epoch": 0.22442199926176973, + "grad_norm": 1.3192722797393799, + "learning_rate": 9.823790875048224e-05, + "loss": 0.7779, + "step": 3344 + }, + { + "epoch": 0.22455622294553873, + "grad_norm": 1.1418710947036743, + "learning_rate": 9.823504774534591e-05, + "loss": 0.7458, + "step": 3346 + }, + { + "epoch": 0.22469044662930773, + "grad_norm": 1.180791974067688, + "learning_rate": 9.823218446119842e-05, + "loss": 0.7012, + "step": 3348 + }, + { + "epoch": 0.22482467031307674, + "grad_norm": 1.1937333345413208, + "learning_rate": 9.822931889817506e-05, + "loss": 0.7746, + "step": 3350 + }, + { + "epoch": 0.22495889399684574, + "grad_norm": 1.2063286304473877, + "learning_rate": 9.822645105641123e-05, + "loss": 0.7043, + "step": 3352 + }, + { + "epoch": 0.22509311768061474, + "grad_norm": 1.2942754030227661, + "learning_rate": 9.822358093604242e-05, + "loss": 0.7285, + "step": 3354 + }, + { + "epoch": 0.22522734136438374, + "grad_norm": 1.1679829359054565, + "learning_rate": 9.822070853720421e-05, + "loss": 0.683, + "step": 3356 + }, + { + "epoch": 0.22536156504815275, + "grad_norm": 1.1674364805221558, + "learning_rate": 9.821783386003239e-05, + "loss": 0.6875, + "step": 3358 + }, + { + "epoch": 0.22549578873192175, + "grad_norm": 1.1562505960464478, + "learning_rate": 9.821495690466272e-05, + "loss": 0.7452, + "step": 3360 + }, + { + "epoch": 0.22563001241569075, + "grad_norm": 1.2197908163070679, + "learning_rate": 9.821207767123113e-05, + "loss": 0.8058, + "step": 3362 + }, + { + "epoch": 0.22576423609945975, + "grad_norm": 1.1897050142288208, + "learning_rate": 9.820919615987368e-05, + "loss": 0.6991, + "step": 3364 + }, + { + "epoch": 0.22589845978322876, + "grad_norm": 1.2827985286712646, + "learning_rate": 9.820631237072652e-05, + "loss": 0.84, + "step": 3366 + }, + { + "epoch": 0.22603268346699776, + "grad_norm": 1.0616706609725952, + "learning_rate": 9.82034263039259e-05, + "loss": 0.6679, + "step": 3368 + }, + { + "epoch": 0.22616690715076676, + "grad_norm": 1.5039023160934448, + "learning_rate": 9.820053795960815e-05, + "loss": 0.7722, + "step": 3370 + }, + { + "epoch": 0.22630113083453576, + "grad_norm": 1.2132271528244019, + "learning_rate": 9.819764733790979e-05, + "loss": 0.696, + "step": 3372 + }, + { + "epoch": 0.22643535451830477, + "grad_norm": 1.3512502908706665, + "learning_rate": 9.819475443896736e-05, + "loss": 0.6855, + "step": 3374 + }, + { + "epoch": 0.22656957820207377, + "grad_norm": 1.2658389806747437, + "learning_rate": 9.819185926291754e-05, + "loss": 0.7526, + "step": 3376 + }, + { + "epoch": 0.22670380188584277, + "grad_norm": 1.086442232131958, + "learning_rate": 9.818896180989716e-05, + "loss": 0.7226, + "step": 3378 + }, + { + "epoch": 0.22683802556961175, + "grad_norm": 1.312869668006897, + "learning_rate": 9.818606208004309e-05, + "loss": 0.7276, + "step": 3380 + }, + { + "epoch": 0.22697224925338075, + "grad_norm": 1.1030393838882446, + "learning_rate": 9.818316007349232e-05, + "loss": 0.8007, + "step": 3382 + }, + { + "epoch": 0.22710647293714975, + "grad_norm": 1.296844244003296, + "learning_rate": 9.8180255790382e-05, + "loss": 0.7222, + "step": 3384 + }, + { + "epoch": 0.22724069662091875, + "grad_norm": 2.0905978679656982, + "learning_rate": 9.817734923084934e-05, + "loss": 0.7195, + "step": 3386 + }, + { + "epoch": 0.22737492030468776, + "grad_norm": 1.3652046918869019, + "learning_rate": 9.817444039503165e-05, + "loss": 0.7275, + "step": 3388 + }, + { + "epoch": 0.22750914398845676, + "grad_norm": 1.1128994226455688, + "learning_rate": 9.817152928306638e-05, + "loss": 0.7207, + "step": 3390 + }, + { + "epoch": 0.22764336767222576, + "grad_norm": 1.7706009149551392, + "learning_rate": 9.81686158950911e-05, + "loss": 0.7097, + "step": 3392 + }, + { + "epoch": 0.22777759135599476, + "grad_norm": 1.4544775485992432, + "learning_rate": 9.816570023124342e-05, + "loss": 0.834, + "step": 3394 + }, + { + "epoch": 0.22791181503976377, + "grad_norm": 1.928645133972168, + "learning_rate": 9.816278229166114e-05, + "loss": 0.7695, + "step": 3396 + }, + { + "epoch": 0.22804603872353277, + "grad_norm": 1.073394775390625, + "learning_rate": 9.815986207648208e-05, + "loss": 0.7945, + "step": 3398 + }, + { + "epoch": 0.22818026240730177, + "grad_norm": 1.1026980876922607, + "learning_rate": 9.815693958584424e-05, + "loss": 0.7195, + "step": 3400 + }, + { + "epoch": 0.22831448609107077, + "grad_norm": 1.298104166984558, + "learning_rate": 9.815401481988571e-05, + "loss": 0.7352, + "step": 3402 + }, + { + "epoch": 0.22844870977483978, + "grad_norm": 1.1682909727096558, + "learning_rate": 9.815108777874467e-05, + "loss": 0.7504, + "step": 3404 + }, + { + "epoch": 0.22858293345860878, + "grad_norm": 1.234245777130127, + "learning_rate": 9.814815846255942e-05, + "loss": 0.819, + "step": 3406 + }, + { + "epoch": 0.22871715714237778, + "grad_norm": 1.2632322311401367, + "learning_rate": 9.814522687146837e-05, + "loss": 0.6739, + "step": 3408 + }, + { + "epoch": 0.22885138082614678, + "grad_norm": 1.0744209289550781, + "learning_rate": 9.814229300560999e-05, + "loss": 0.7162, + "step": 3410 + }, + { + "epoch": 0.22898560450991579, + "grad_norm": 1.0212805271148682, + "learning_rate": 9.813935686512297e-05, + "loss": 0.7079, + "step": 3412 + }, + { + "epoch": 0.2291198281936848, + "grad_norm": 1.13587486743927, + "learning_rate": 9.813641845014599e-05, + "loss": 0.8045, + "step": 3414 + }, + { + "epoch": 0.2292540518774538, + "grad_norm": 1.1257480382919312, + "learning_rate": 9.813347776081789e-05, + "loss": 0.6868, + "step": 3416 + }, + { + "epoch": 0.22938827556122277, + "grad_norm": 1.2473704814910889, + "learning_rate": 9.813053479727761e-05, + "loss": 0.714, + "step": 3418 + }, + { + "epoch": 0.22952249924499177, + "grad_norm": 1.0271937847137451, + "learning_rate": 9.812758955966421e-05, + "loss": 0.6735, + "step": 3420 + }, + { + "epoch": 0.22965672292876077, + "grad_norm": 1.1803767681121826, + "learning_rate": 9.812464204811686e-05, + "loss": 0.7347, + "step": 3422 + }, + { + "epoch": 0.22979094661252977, + "grad_norm": 1.1167150735855103, + "learning_rate": 9.812169226277479e-05, + "loss": 0.6919, + "step": 3424 + }, + { + "epoch": 0.22992517029629878, + "grad_norm": 1.2198318243026733, + "learning_rate": 9.811874020377738e-05, + "loss": 0.7548, + "step": 3426 + }, + { + "epoch": 0.23005939398006778, + "grad_norm": 1.0730876922607422, + "learning_rate": 9.811578587126413e-05, + "loss": 0.6881, + "step": 3428 + }, + { + "epoch": 0.23019361766383678, + "grad_norm": 1.3007014989852905, + "learning_rate": 9.81128292653746e-05, + "loss": 0.6966, + "step": 3430 + }, + { + "epoch": 0.23032784134760578, + "grad_norm": 1.1373518705368042, + "learning_rate": 9.810987038624851e-05, + "loss": 0.7638, + "step": 3432 + }, + { + "epoch": 0.23046206503137479, + "grad_norm": 1.1992700099945068, + "learning_rate": 9.810690923402566e-05, + "loss": 0.7312, + "step": 3434 + }, + { + "epoch": 0.2305962887151438, + "grad_norm": 1.0444557666778564, + "learning_rate": 9.810394580884592e-05, + "loss": 0.7302, + "step": 3436 + }, + { + "epoch": 0.2307305123989128, + "grad_norm": 1.0445059537887573, + "learning_rate": 9.810098011084935e-05, + "loss": 0.7635, + "step": 3438 + }, + { + "epoch": 0.2308647360826818, + "grad_norm": 1.4251956939697266, + "learning_rate": 9.809801214017604e-05, + "loss": 0.7156, + "step": 3440 + }, + { + "epoch": 0.2309989597664508, + "grad_norm": 1.1660226583480835, + "learning_rate": 9.809504189696626e-05, + "loss": 0.7105, + "step": 3442 + }, + { + "epoch": 0.2311331834502198, + "grad_norm": 1.0929772853851318, + "learning_rate": 9.809206938136031e-05, + "loss": 0.738, + "step": 3444 + }, + { + "epoch": 0.2312674071339888, + "grad_norm": 1.3212532997131348, + "learning_rate": 9.808909459349865e-05, + "loss": 0.7501, + "step": 3446 + }, + { + "epoch": 0.2314016308177578, + "grad_norm": 1.182216763496399, + "learning_rate": 9.808611753352184e-05, + "loss": 0.737, + "step": 3448 + }, + { + "epoch": 0.2315358545015268, + "grad_norm": 1.378854513168335, + "learning_rate": 9.808313820157052e-05, + "loss": 0.7129, + "step": 3450 + }, + { + "epoch": 0.2316700781852958, + "grad_norm": 1.1713718175888062, + "learning_rate": 9.808015659778549e-05, + "loss": 0.7528, + "step": 3452 + }, + { + "epoch": 0.23180430186906478, + "grad_norm": 1.11350679397583, + "learning_rate": 9.80771727223076e-05, + "loss": 0.6979, + "step": 3454 + }, + { + "epoch": 0.23193852555283379, + "grad_norm": 1.063978910446167, + "learning_rate": 9.807418657527782e-05, + "loss": 0.7262, + "step": 3456 + }, + { + "epoch": 0.2320727492366028, + "grad_norm": 1.4309468269348145, + "learning_rate": 9.807119815683728e-05, + "loss": 0.6923, + "step": 3458 + }, + { + "epoch": 0.2322069729203718, + "grad_norm": 1.379732370376587, + "learning_rate": 9.806820746712716e-05, + "loss": 0.7342, + "step": 3460 + }, + { + "epoch": 0.2323411966041408, + "grad_norm": 1.1750215291976929, + "learning_rate": 9.806521450628875e-05, + "loss": 0.7704, + "step": 3462 + }, + { + "epoch": 0.2324754202879098, + "grad_norm": 1.1312384605407715, + "learning_rate": 9.806221927446347e-05, + "loss": 0.7967, + "step": 3464 + }, + { + "epoch": 0.2326096439716788, + "grad_norm": 1.3090623617172241, + "learning_rate": 9.805922177179283e-05, + "loss": 0.7725, + "step": 3466 + }, + { + "epoch": 0.2327438676554478, + "grad_norm": 1.1082390546798706, + "learning_rate": 9.805622199841848e-05, + "loss": 0.6999, + "step": 3468 + }, + { + "epoch": 0.2328780913392168, + "grad_norm": 1.0914182662963867, + "learning_rate": 9.805321995448214e-05, + "loss": 0.7262, + "step": 3470 + }, + { + "epoch": 0.2330123150229858, + "grad_norm": 1.1261818408966064, + "learning_rate": 9.805021564012564e-05, + "loss": 0.7699, + "step": 3472 + }, + { + "epoch": 0.2331465387067548, + "grad_norm": 1.0641417503356934, + "learning_rate": 9.804720905549094e-05, + "loss": 0.7033, + "step": 3474 + }, + { + "epoch": 0.2332807623905238, + "grad_norm": 1.4888118505477905, + "learning_rate": 9.80442002007201e-05, + "loss": 0.6867, + "step": 3476 + }, + { + "epoch": 0.2334149860742928, + "grad_norm": 1.3334269523620605, + "learning_rate": 9.804118907595527e-05, + "loss": 0.6697, + "step": 3478 + }, + { + "epoch": 0.23354920975806182, + "grad_norm": 1.0359936952590942, + "learning_rate": 9.803817568133872e-05, + "loss": 0.7034, + "step": 3480 + }, + { + "epoch": 0.23368343344183082, + "grad_norm": 1.256037712097168, + "learning_rate": 9.803516001701286e-05, + "loss": 0.7702, + "step": 3482 + }, + { + "epoch": 0.23381765712559982, + "grad_norm": 1.2835520505905151, + "learning_rate": 9.80321420831201e-05, + "loss": 0.7173, + "step": 3484 + }, + { + "epoch": 0.23395188080936882, + "grad_norm": 1.2579686641693115, + "learning_rate": 9.80291218798031e-05, + "loss": 0.7454, + "step": 3486 + }, + { + "epoch": 0.23408610449313783, + "grad_norm": 1.1662805080413818, + "learning_rate": 9.802609940720455e-05, + "loss": 0.7573, + "step": 3488 + }, + { + "epoch": 0.23422032817690683, + "grad_norm": 1.117235541343689, + "learning_rate": 9.802307466546723e-05, + "loss": 0.7579, + "step": 3490 + }, + { + "epoch": 0.2343545518606758, + "grad_norm": 1.1694008111953735, + "learning_rate": 9.802004765473407e-05, + "loss": 0.7195, + "step": 3492 + }, + { + "epoch": 0.2344887755444448, + "grad_norm": 1.1704126596450806, + "learning_rate": 9.801701837514808e-05, + "loss": 0.6869, + "step": 3494 + }, + { + "epoch": 0.2346229992282138, + "grad_norm": 1.1846356391906738, + "learning_rate": 9.801398682685238e-05, + "loss": 0.7383, + "step": 3496 + }, + { + "epoch": 0.2347572229119828, + "grad_norm": 1.1462559700012207, + "learning_rate": 9.801095300999024e-05, + "loss": 0.7334, + "step": 3498 + }, + { + "epoch": 0.2348914465957518, + "grad_norm": 1.1507761478424072, + "learning_rate": 9.800791692470497e-05, + "loss": 0.66, + "step": 3500 + }, + { + "epoch": 0.23502567027952082, + "grad_norm": 1.2698543071746826, + "learning_rate": 9.800487857114004e-05, + "loss": 0.7343, + "step": 3502 + }, + { + "epoch": 0.23515989396328982, + "grad_norm": 1.1707394123077393, + "learning_rate": 9.800183794943898e-05, + "loss": 0.7248, + "step": 3504 + }, + { + "epoch": 0.23529411764705882, + "grad_norm": 1.3977922201156616, + "learning_rate": 9.799879505974548e-05, + "loss": 0.7618, + "step": 3506 + }, + { + "epoch": 0.23542834133082782, + "grad_norm": 1.113440752029419, + "learning_rate": 9.799574990220328e-05, + "loss": 0.6986, + "step": 3508 + }, + { + "epoch": 0.23556256501459683, + "grad_norm": 0.9917773008346558, + "learning_rate": 9.79927024769563e-05, + "loss": 0.6308, + "step": 3510 + }, + { + "epoch": 0.23569678869836583, + "grad_norm": 1.1032965183258057, + "learning_rate": 9.798965278414849e-05, + "loss": 0.7304, + "step": 3512 + }, + { + "epoch": 0.23583101238213483, + "grad_norm": 1.1461377143859863, + "learning_rate": 9.798660082392396e-05, + "loss": 0.6918, + "step": 3514 + }, + { + "epoch": 0.23596523606590383, + "grad_norm": 1.1212458610534668, + "learning_rate": 9.798354659642691e-05, + "loss": 0.7687, + "step": 3516 + }, + { + "epoch": 0.23609945974967284, + "grad_norm": 1.6978212594985962, + "learning_rate": 9.798049010180161e-05, + "loss": 0.7388, + "step": 3518 + }, + { + "epoch": 0.23623368343344184, + "grad_norm": 1.3478388786315918, + "learning_rate": 9.797743134019253e-05, + "loss": 0.7947, + "step": 3520 + }, + { + "epoch": 0.23636790711721084, + "grad_norm": 1.1857460737228394, + "learning_rate": 9.797437031174414e-05, + "loss": 0.7414, + "step": 3522 + }, + { + "epoch": 0.23650213080097984, + "grad_norm": 1.2205296754837036, + "learning_rate": 9.797130701660111e-05, + "loss": 0.7317, + "step": 3524 + }, + { + "epoch": 0.23663635448474885, + "grad_norm": 1.3379762172698975, + "learning_rate": 9.796824145490815e-05, + "loss": 0.7165, + "step": 3526 + }, + { + "epoch": 0.23677057816851785, + "grad_norm": 1.1166516542434692, + "learning_rate": 9.79651736268101e-05, + "loss": 0.7088, + "step": 3528 + }, + { + "epoch": 0.23690480185228682, + "grad_norm": 1.2244601249694824, + "learning_rate": 9.796210353245192e-05, + "loss": 0.7266, + "step": 3530 + }, + { + "epoch": 0.23703902553605583, + "grad_norm": 1.0700269937515259, + "learning_rate": 9.795903117197867e-05, + "loss": 0.7143, + "step": 3532 + }, + { + "epoch": 0.23717324921982483, + "grad_norm": 1.1071746349334717, + "learning_rate": 9.795595654553548e-05, + "loss": 0.697, + "step": 3534 + }, + { + "epoch": 0.23730747290359383, + "grad_norm": 1.3859741687774658, + "learning_rate": 9.795287965326767e-05, + "loss": 0.7405, + "step": 3536 + }, + { + "epoch": 0.23744169658736283, + "grad_norm": 1.2678968906402588, + "learning_rate": 9.794980049532058e-05, + "loss": 0.706, + "step": 3538 + }, + { + "epoch": 0.23757592027113184, + "grad_norm": 1.073473572731018, + "learning_rate": 9.79467190718397e-05, + "loss": 0.7804, + "step": 3540 + }, + { + "epoch": 0.23771014395490084, + "grad_norm": 1.770054578781128, + "learning_rate": 9.794363538297065e-05, + "loss": 0.7759, + "step": 3542 + }, + { + "epoch": 0.23784436763866984, + "grad_norm": 1.256622552871704, + "learning_rate": 9.794054942885909e-05, + "loss": 0.7375, + "step": 3544 + }, + { + "epoch": 0.23797859132243884, + "grad_norm": 1.2577251195907593, + "learning_rate": 9.793746120965083e-05, + "loss": 0.7204, + "step": 3546 + }, + { + "epoch": 0.23811281500620785, + "grad_norm": 1.1353765726089478, + "learning_rate": 9.793437072549181e-05, + "loss": 0.6705, + "step": 3548 + }, + { + "epoch": 0.23824703868997685, + "grad_norm": 1.0381370782852173, + "learning_rate": 9.793127797652801e-05, + "loss": 0.6757, + "step": 3550 + }, + { + "epoch": 0.23838126237374585, + "grad_norm": 1.1946134567260742, + "learning_rate": 9.79281829629056e-05, + "loss": 0.7714, + "step": 3552 + }, + { + "epoch": 0.23851548605751485, + "grad_norm": 1.4038199186325073, + "learning_rate": 9.792508568477078e-05, + "loss": 0.723, + "step": 3554 + }, + { + "epoch": 0.23864970974128386, + "grad_norm": 1.2247889041900635, + "learning_rate": 9.792198614226992e-05, + "loss": 0.7129, + "step": 3556 + }, + { + "epoch": 0.23878393342505286, + "grad_norm": 1.3369977474212646, + "learning_rate": 9.791888433554943e-05, + "loss": 0.7371, + "step": 3558 + }, + { + "epoch": 0.23891815710882186, + "grad_norm": 1.1259257793426514, + "learning_rate": 9.79157802647559e-05, + "loss": 0.6986, + "step": 3560 + }, + { + "epoch": 0.23905238079259086, + "grad_norm": 1.241822600364685, + "learning_rate": 9.791267393003596e-05, + "loss": 0.7072, + "step": 3562 + }, + { + "epoch": 0.23918660447635987, + "grad_norm": 1.2874438762664795, + "learning_rate": 9.79095653315364e-05, + "loss": 0.7947, + "step": 3564 + }, + { + "epoch": 0.23932082816012887, + "grad_norm": 1.553829312324524, + "learning_rate": 9.790645446940408e-05, + "loss": 0.7108, + "step": 3566 + }, + { + "epoch": 0.23945505184389784, + "grad_norm": 1.1112264394760132, + "learning_rate": 9.7903341343786e-05, + "loss": 0.7198, + "step": 3568 + }, + { + "epoch": 0.23958927552766685, + "grad_norm": 1.469152808189392, + "learning_rate": 9.790022595482924e-05, + "loss": 0.7685, + "step": 3570 + }, + { + "epoch": 0.23972349921143585, + "grad_norm": 1.1694190502166748, + "learning_rate": 9.789710830268099e-05, + "loss": 0.7018, + "step": 3572 + }, + { + "epoch": 0.23985772289520485, + "grad_norm": 1.2232519388198853, + "learning_rate": 9.789398838748856e-05, + "loss": 0.7373, + "step": 3574 + }, + { + "epoch": 0.23999194657897385, + "grad_norm": 1.0650718212127686, + "learning_rate": 9.789086620939936e-05, + "loss": 0.7908, + "step": 3576 + }, + { + "epoch": 0.24012617026274286, + "grad_norm": 1.0550053119659424, + "learning_rate": 9.78877417685609e-05, + "loss": 0.7505, + "step": 3578 + }, + { + "epoch": 0.24026039394651186, + "grad_norm": 1.0913403034210205, + "learning_rate": 9.788461506512081e-05, + "loss": 0.6891, + "step": 3580 + }, + { + "epoch": 0.24039461763028086, + "grad_norm": 1.256256103515625, + "learning_rate": 9.788148609922682e-05, + "loss": 0.7299, + "step": 3582 + }, + { + "epoch": 0.24052884131404986, + "grad_norm": 1.370922327041626, + "learning_rate": 9.787835487102677e-05, + "loss": 0.78, + "step": 3584 + }, + { + "epoch": 0.24066306499781887, + "grad_norm": 1.1956902742385864, + "learning_rate": 9.78752213806686e-05, + "loss": 0.7117, + "step": 3586 + }, + { + "epoch": 0.24079728868158787, + "grad_norm": 1.1800501346588135, + "learning_rate": 9.787208562830036e-05, + "loss": 0.6698, + "step": 3588 + }, + { + "epoch": 0.24093151236535687, + "grad_norm": 1.1705611944198608, + "learning_rate": 9.786894761407021e-05, + "loss": 0.6684, + "step": 3590 + }, + { + "epoch": 0.24106573604912587, + "grad_norm": 1.2129614353179932, + "learning_rate": 9.786580733812643e-05, + "loss": 0.7731, + "step": 3592 + }, + { + "epoch": 0.24119995973289488, + "grad_norm": 1.265656590461731, + "learning_rate": 9.786266480061737e-05, + "loss": 0.7194, + "step": 3594 + }, + { + "epoch": 0.24133418341666388, + "grad_norm": 1.2038565874099731, + "learning_rate": 9.78595200016915e-05, + "loss": 0.7071, + "step": 3596 + }, + { + "epoch": 0.24146840710043288, + "grad_norm": 1.1142795085906982, + "learning_rate": 9.785637294149743e-05, + "loss": 0.7693, + "step": 3598 + }, + { + "epoch": 0.24160263078420188, + "grad_norm": 1.061290979385376, + "learning_rate": 9.785322362018385e-05, + "loss": 0.7344, + "step": 3600 + }, + { + "epoch": 0.24173685446797089, + "grad_norm": 1.1317564249038696, + "learning_rate": 9.785007203789955e-05, + "loss": 0.7455, + "step": 3602 + }, + { + "epoch": 0.2418710781517399, + "grad_norm": 1.1659226417541504, + "learning_rate": 9.784691819479343e-05, + "loss": 0.7227, + "step": 3604 + }, + { + "epoch": 0.24200530183550886, + "grad_norm": 1.1892750263214111, + "learning_rate": 9.784376209101454e-05, + "loss": 0.7811, + "step": 3606 + }, + { + "epoch": 0.24213952551927787, + "grad_norm": 1.2177784442901611, + "learning_rate": 9.784060372671195e-05, + "loss": 0.7072, + "step": 3608 + }, + { + "epoch": 0.24227374920304687, + "grad_norm": 1.257696270942688, + "learning_rate": 9.783744310203491e-05, + "loss": 0.7335, + "step": 3610 + }, + { + "epoch": 0.24240797288681587, + "grad_norm": 1.164030909538269, + "learning_rate": 9.783428021713274e-05, + "loss": 0.7008, + "step": 3612 + }, + { + "epoch": 0.24254219657058487, + "grad_norm": 1.138877034187317, + "learning_rate": 9.783111507215491e-05, + "loss": 0.7451, + "step": 3614 + }, + { + "epoch": 0.24267642025435388, + "grad_norm": 1.120921015739441, + "learning_rate": 9.782794766725094e-05, + "loss": 0.8045, + "step": 3616 + }, + { + "epoch": 0.24281064393812288, + "grad_norm": 1.1520529985427856, + "learning_rate": 9.78247780025705e-05, + "loss": 0.7086, + "step": 3618 + }, + { + "epoch": 0.24294486762189188, + "grad_norm": 1.222719669342041, + "learning_rate": 9.782160607826334e-05, + "loss": 0.7644, + "step": 3620 + }, + { + "epoch": 0.24307909130566088, + "grad_norm": 1.2890983819961548, + "learning_rate": 9.781843189447933e-05, + "loss": 0.7014, + "step": 3622 + }, + { + "epoch": 0.24321331498942989, + "grad_norm": 1.2486456632614136, + "learning_rate": 9.781525545136844e-05, + "loss": 0.7205, + "step": 3624 + }, + { + "epoch": 0.2433475386731989, + "grad_norm": 1.1125035285949707, + "learning_rate": 9.781207674908076e-05, + "loss": 0.7771, + "step": 3626 + }, + { + "epoch": 0.2434817623569679, + "grad_norm": 1.0747028589248657, + "learning_rate": 9.780889578776647e-05, + "loss": 0.664, + "step": 3628 + }, + { + "epoch": 0.2436159860407369, + "grad_norm": 1.0456690788269043, + "learning_rate": 9.780571256757587e-05, + "loss": 0.6862, + "step": 3630 + }, + { + "epoch": 0.2437502097245059, + "grad_norm": 1.097830057144165, + "learning_rate": 9.780252708865936e-05, + "loss": 0.692, + "step": 3632 + }, + { + "epoch": 0.2438844334082749, + "grad_norm": 1.0377219915390015, + "learning_rate": 9.779933935116742e-05, + "loss": 0.6649, + "step": 3634 + }, + { + "epoch": 0.2440186570920439, + "grad_norm": 1.1227837800979614, + "learning_rate": 9.779614935525073e-05, + "loss": 0.7859, + "step": 3636 + }, + { + "epoch": 0.2441528807758129, + "grad_norm": 1.0132102966308594, + "learning_rate": 9.779295710105993e-05, + "loss": 0.7911, + "step": 3638 + }, + { + "epoch": 0.2442871044595819, + "grad_norm": 1.0197548866271973, + "learning_rate": 9.77897625887459e-05, + "loss": 0.6874, + "step": 3640 + }, + { + "epoch": 0.2444213281433509, + "grad_norm": 1.0421596765518188, + "learning_rate": 9.778656581845958e-05, + "loss": 0.6859, + "step": 3642 + }, + { + "epoch": 0.24455555182711988, + "grad_norm": 1.015177845954895, + "learning_rate": 9.778336679035197e-05, + "loss": 0.7016, + "step": 3644 + }, + { + "epoch": 0.24468977551088889, + "grad_norm": 1.1596684455871582, + "learning_rate": 9.778016550457425e-05, + "loss": 0.6877, + "step": 3646 + }, + { + "epoch": 0.2448239991946579, + "grad_norm": 1.020958662033081, + "learning_rate": 9.777696196127766e-05, + "loss": 0.697, + "step": 3648 + }, + { + "epoch": 0.2449582228784269, + "grad_norm": 1.4052140712738037, + "learning_rate": 9.777375616061359e-05, + "loss": 0.7571, + "step": 3650 + }, + { + "epoch": 0.2450924465621959, + "grad_norm": 1.1902647018432617, + "learning_rate": 9.777054810273345e-05, + "loss": 0.7328, + "step": 3652 + }, + { + "epoch": 0.2452266702459649, + "grad_norm": 1.2707990407943726, + "learning_rate": 9.776733778778888e-05, + "loss": 0.7506, + "step": 3654 + }, + { + "epoch": 0.2453608939297339, + "grad_norm": 1.1604020595550537, + "learning_rate": 9.776412521593152e-05, + "loss": 0.6732, + "step": 3656 + }, + { + "epoch": 0.2454951176135029, + "grad_norm": 1.068184733390808, + "learning_rate": 9.776091038731317e-05, + "loss": 0.7242, + "step": 3658 + }, + { + "epoch": 0.2456293412972719, + "grad_norm": 1.2311961650848389, + "learning_rate": 9.775769330208571e-05, + "loss": 0.6708, + "step": 3660 + }, + { + "epoch": 0.2457635649810409, + "grad_norm": 1.0739384889602661, + "learning_rate": 9.775447396040116e-05, + "loss": 0.735, + "step": 3662 + }, + { + "epoch": 0.2458977886648099, + "grad_norm": 1.1574194431304932, + "learning_rate": 9.775125236241161e-05, + "loss": 0.7217, + "step": 3664 + }, + { + "epoch": 0.2460320123485789, + "grad_norm": 1.1732994318008423, + "learning_rate": 9.77480285082693e-05, + "loss": 0.7277, + "step": 3666 + }, + { + "epoch": 0.2461662360323479, + "grad_norm": 1.008643627166748, + "learning_rate": 9.774480239812653e-05, + "loss": 0.6997, + "step": 3668 + }, + { + "epoch": 0.24630045971611692, + "grad_norm": 1.0346351861953735, + "learning_rate": 9.774157403213573e-05, + "loss": 0.7228, + "step": 3670 + }, + { + "epoch": 0.24643468339988592, + "grad_norm": 1.1072207689285278, + "learning_rate": 9.773834341044944e-05, + "loss": 0.7401, + "step": 3672 + }, + { + "epoch": 0.24656890708365492, + "grad_norm": 1.3879812955856323, + "learning_rate": 9.77351105332203e-05, + "loss": 0.729, + "step": 3674 + }, + { + "epoch": 0.24670313076742392, + "grad_norm": 1.1985526084899902, + "learning_rate": 9.773187540060105e-05, + "loss": 0.8118, + "step": 3676 + }, + { + "epoch": 0.24683735445119293, + "grad_norm": 1.0382481813430786, + "learning_rate": 9.772863801274455e-05, + "loss": 0.7627, + "step": 3678 + }, + { + "epoch": 0.24697157813496193, + "grad_norm": 1.1838608980178833, + "learning_rate": 9.772539836980376e-05, + "loss": 0.7657, + "step": 3680 + }, + { + "epoch": 0.2471058018187309, + "grad_norm": 0.9234886169433594, + "learning_rate": 9.772215647193174e-05, + "loss": 0.6451, + "step": 3682 + }, + { + "epoch": 0.2472400255024999, + "grad_norm": 1.0064630508422852, + "learning_rate": 9.771891231928167e-05, + "loss": 0.6762, + "step": 3684 + }, + { + "epoch": 0.2473742491862689, + "grad_norm": 1.1998895406723022, + "learning_rate": 9.771566591200682e-05, + "loss": 0.659, + "step": 3686 + }, + { + "epoch": 0.2475084728700379, + "grad_norm": 1.4650298357009888, + "learning_rate": 9.77124172502606e-05, + "loss": 0.8039, + "step": 3688 + }, + { + "epoch": 0.2476426965538069, + "grad_norm": 1.1969459056854248, + "learning_rate": 9.77091663341965e-05, + "loss": 0.7041, + "step": 3690 + }, + { + "epoch": 0.24777692023757591, + "grad_norm": 1.0319349765777588, + "learning_rate": 9.770591316396807e-05, + "loss": 0.7602, + "step": 3692 + }, + { + "epoch": 0.24791114392134492, + "grad_norm": 1.1042534112930298, + "learning_rate": 9.770265773972906e-05, + "loss": 0.7217, + "step": 3694 + }, + { + "epoch": 0.24804536760511392, + "grad_norm": 0.9951987266540527, + "learning_rate": 9.769940006163329e-05, + "loss": 0.8003, + "step": 3696 + }, + { + "epoch": 0.24817959128888292, + "grad_norm": 1.2236394882202148, + "learning_rate": 9.769614012983465e-05, + "loss": 0.8061, + "step": 3698 + }, + { + "epoch": 0.24831381497265193, + "grad_norm": 1.1980949640274048, + "learning_rate": 9.769287794448721e-05, + "loss": 0.6583, + "step": 3700 + }, + { + "epoch": 0.24844803865642093, + "grad_norm": 1.1455384492874146, + "learning_rate": 9.768961350574503e-05, + "loss": 0.6776, + "step": 3702 + }, + { + "epoch": 0.24858226234018993, + "grad_norm": 0.9626195430755615, + "learning_rate": 9.768634681376243e-05, + "loss": 0.6675, + "step": 3704 + }, + { + "epoch": 0.24871648602395893, + "grad_norm": 1.0661108493804932, + "learning_rate": 9.768307786869369e-05, + "loss": 0.7096, + "step": 3706 + }, + { + "epoch": 0.24885070970772794, + "grad_norm": 1.1644963026046753, + "learning_rate": 9.767980667069328e-05, + "loss": 0.7599, + "step": 3708 + }, + { + "epoch": 0.24898493339149694, + "grad_norm": 1.3132227659225464, + "learning_rate": 9.767653321991578e-05, + "loss": 0.7179, + "step": 3710 + }, + { + "epoch": 0.24911915707526594, + "grad_norm": 1.076185703277588, + "learning_rate": 9.767325751651583e-05, + "loss": 0.7005, + "step": 3712 + }, + { + "epoch": 0.24925338075903494, + "grad_norm": 1.1602646112442017, + "learning_rate": 9.76699795606482e-05, + "loss": 0.681, + "step": 3714 + }, + { + "epoch": 0.24938760444280395, + "grad_norm": 1.0248345136642456, + "learning_rate": 9.766669935246778e-05, + "loss": 0.6841, + "step": 3716 + }, + { + "epoch": 0.24952182812657295, + "grad_norm": 1.0485111474990845, + "learning_rate": 9.766341689212956e-05, + "loss": 0.7047, + "step": 3718 + }, + { + "epoch": 0.24965605181034192, + "grad_norm": 1.3126776218414307, + "learning_rate": 9.76601321797886e-05, + "loss": 0.697, + "step": 3720 + }, + { + "epoch": 0.24979027549411092, + "grad_norm": 1.0859121084213257, + "learning_rate": 9.765684521560012e-05, + "loss": 0.7444, + "step": 3722 + }, + { + "epoch": 0.24992449917787993, + "grad_norm": 1.1804007291793823, + "learning_rate": 9.765355599971942e-05, + "loss": 0.6579, + "step": 3724 + }, + { + "epoch": 0.25005872286164893, + "grad_norm": 1.4063502550125122, + "learning_rate": 9.765026453230191e-05, + "loss": 0.6788, + "step": 3726 + }, + { + "epoch": 0.25019294654541796, + "grad_norm": 1.0992275476455688, + "learning_rate": 9.76469708135031e-05, + "loss": 0.6705, + "step": 3728 + }, + { + "epoch": 0.25032717022918693, + "grad_norm": 1.1454664468765259, + "learning_rate": 9.764367484347861e-05, + "loss": 0.6742, + "step": 3730 + }, + { + "epoch": 0.25046139391295597, + "grad_norm": 1.3076444864273071, + "learning_rate": 9.764037662238417e-05, + "loss": 0.7585, + "step": 3732 + }, + { + "epoch": 0.25059561759672494, + "grad_norm": 1.3327702283859253, + "learning_rate": 9.763707615037561e-05, + "loss": 0.7608, + "step": 3734 + }, + { + "epoch": 0.25072984128049397, + "grad_norm": 1.3393803834915161, + "learning_rate": 9.763377342760888e-05, + "loss": 0.7539, + "step": 3736 + }, + { + "epoch": 0.25086406496426294, + "grad_norm": 1.4489805698394775, + "learning_rate": 9.763046845424002e-05, + "loss": 0.8044, + "step": 3738 + }, + { + "epoch": 0.2509982886480319, + "grad_norm": 1.1607704162597656, + "learning_rate": 9.762716123042519e-05, + "loss": 0.6865, + "step": 3740 + }, + { + "epoch": 0.25113251233180095, + "grad_norm": 1.2858433723449707, + "learning_rate": 9.762385175632065e-05, + "loss": 0.7094, + "step": 3742 + }, + { + "epoch": 0.2512667360155699, + "grad_norm": 1.336488127708435, + "learning_rate": 9.762054003208276e-05, + "loss": 0.8006, + "step": 3744 + }, + { + "epoch": 0.25140095969933896, + "grad_norm": 1.2179168462753296, + "learning_rate": 9.761722605786799e-05, + "loss": 0.7414, + "step": 3746 + }, + { + "epoch": 0.25153518338310793, + "grad_norm": 1.1243705749511719, + "learning_rate": 9.761390983383294e-05, + "loss": 0.6437, + "step": 3748 + }, + { + "epoch": 0.25166940706687696, + "grad_norm": 1.0953590869903564, + "learning_rate": 9.761059136013426e-05, + "loss": 0.7216, + "step": 3750 + }, + { + "epoch": 0.25180363075064593, + "grad_norm": 1.3589489459991455, + "learning_rate": 9.760727063692878e-05, + "loss": 0.8207, + "step": 3752 + }, + { + "epoch": 0.25193785443441497, + "grad_norm": 1.2893167734146118, + "learning_rate": 9.760394766437335e-05, + "loss": 0.7309, + "step": 3754 + }, + { + "epoch": 0.25207207811818394, + "grad_norm": 1.1073765754699707, + "learning_rate": 9.760062244262502e-05, + "loss": 0.7182, + "step": 3756 + }, + { + "epoch": 0.25220630180195297, + "grad_norm": 0.994770884513855, + "learning_rate": 9.759729497184089e-05, + "loss": 0.7595, + "step": 3758 + }, + { + "epoch": 0.25234052548572194, + "grad_norm": 1.1640686988830566, + "learning_rate": 9.759396525217817e-05, + "loss": 0.7536, + "step": 3760 + }, + { + "epoch": 0.252474749169491, + "grad_norm": 1.36345636844635, + "learning_rate": 9.759063328379416e-05, + "loss": 0.7381, + "step": 3762 + }, + { + "epoch": 0.25260897285325995, + "grad_norm": 1.3841369152069092, + "learning_rate": 9.758729906684632e-05, + "loss": 0.614, + "step": 3764 + }, + { + "epoch": 0.252743196537029, + "grad_norm": 1.0968348979949951, + "learning_rate": 9.758396260149219e-05, + "loss": 0.6944, + "step": 3766 + }, + { + "epoch": 0.25287742022079795, + "grad_norm": 1.3005449771881104, + "learning_rate": 9.758062388788937e-05, + "loss": 0.7201, + "step": 3768 + }, + { + "epoch": 0.253011643904567, + "grad_norm": 1.2965426445007324, + "learning_rate": 9.757728292619566e-05, + "loss": 0.7133, + "step": 3770 + }, + { + "epoch": 0.25314586758833596, + "grad_norm": 1.0011999607086182, + "learning_rate": 9.757393971656888e-05, + "loss": 0.7186, + "step": 3772 + }, + { + "epoch": 0.253280091272105, + "grad_norm": 1.1350724697113037, + "learning_rate": 9.7570594259167e-05, + "loss": 0.7019, + "step": 3774 + }, + { + "epoch": 0.25341431495587396, + "grad_norm": 1.1533780097961426, + "learning_rate": 9.756724655414807e-05, + "loss": 0.706, + "step": 3776 + }, + { + "epoch": 0.25354853863964294, + "grad_norm": 1.064712405204773, + "learning_rate": 9.75638966016703e-05, + "loss": 0.7427, + "step": 3778 + }, + { + "epoch": 0.25368276232341197, + "grad_norm": 1.1156611442565918, + "learning_rate": 9.756054440189191e-05, + "loss": 0.6738, + "step": 3780 + }, + { + "epoch": 0.25381698600718094, + "grad_norm": 0.9972400665283203, + "learning_rate": 9.755718995497136e-05, + "loss": 0.6887, + "step": 3782 + }, + { + "epoch": 0.25395120969095, + "grad_norm": 1.187756896018982, + "learning_rate": 9.755383326106709e-05, + "loss": 0.7279, + "step": 3784 + }, + { + "epoch": 0.25408543337471895, + "grad_norm": 1.1233904361724854, + "learning_rate": 9.75504743203377e-05, + "loss": 0.7038, + "step": 3786 + }, + { + "epoch": 0.254219657058488, + "grad_norm": 1.1424480676651, + "learning_rate": 9.75471131329419e-05, + "loss": 0.6967, + "step": 3788 + }, + { + "epoch": 0.25435388074225695, + "grad_norm": 1.0352827310562134, + "learning_rate": 9.754374969903852e-05, + "loss": 0.6602, + "step": 3790 + }, + { + "epoch": 0.254488104426026, + "grad_norm": 1.3024020195007324, + "learning_rate": 9.754038401878645e-05, + "loss": 0.7572, + "step": 3792 + }, + { + "epoch": 0.25462232810979496, + "grad_norm": 1.093445062637329, + "learning_rate": 9.753701609234471e-05, + "loss": 0.751, + "step": 3794 + }, + { + "epoch": 0.254756551793564, + "grad_norm": 1.159218192100525, + "learning_rate": 9.753364591987244e-05, + "loss": 0.7295, + "step": 3796 + }, + { + "epoch": 0.25489077547733296, + "grad_norm": 1.127556562423706, + "learning_rate": 9.753027350152888e-05, + "loss": 0.7962, + "step": 3798 + }, + { + "epoch": 0.255024999161102, + "grad_norm": 1.3137211799621582, + "learning_rate": 9.752689883747335e-05, + "loss": 0.738, + "step": 3800 + }, + { + "epoch": 0.25515922284487097, + "grad_norm": 1.182447910308838, + "learning_rate": 9.752352192786531e-05, + "loss": 0.7343, + "step": 3802 + }, + { + "epoch": 0.25529344652864, + "grad_norm": 1.095441222190857, + "learning_rate": 9.752014277286432e-05, + "loss": 0.6628, + "step": 3804 + }, + { + "epoch": 0.255427670212409, + "grad_norm": 1.229159951210022, + "learning_rate": 9.751676137263002e-05, + "loss": 0.6822, + "step": 3806 + }, + { + "epoch": 0.255561893896178, + "grad_norm": 1.1966331005096436, + "learning_rate": 9.751337772732218e-05, + "loss": 0.7565, + "step": 3808 + }, + { + "epoch": 0.255696117579947, + "grad_norm": 1.3662197589874268, + "learning_rate": 9.750999183710068e-05, + "loss": 0.7229, + "step": 3810 + }, + { + "epoch": 0.255830341263716, + "grad_norm": 1.0455504655838013, + "learning_rate": 9.750660370212549e-05, + "loss": 0.6916, + "step": 3812 + }, + { + "epoch": 0.255964564947485, + "grad_norm": 1.1308659315109253, + "learning_rate": 9.75032133225567e-05, + "loss": 0.7304, + "step": 3814 + }, + { + "epoch": 0.25609878863125396, + "grad_norm": 1.267609715461731, + "learning_rate": 9.749982069855448e-05, + "loss": 0.6578, + "step": 3816 + }, + { + "epoch": 0.256233012315023, + "grad_norm": 1.3140232563018799, + "learning_rate": 9.749642583027914e-05, + "loss": 0.7273, + "step": 3818 + }, + { + "epoch": 0.25636723599879196, + "grad_norm": 1.0035659074783325, + "learning_rate": 9.749302871789107e-05, + "loss": 0.6648, + "step": 3820 + }, + { + "epoch": 0.256501459682561, + "grad_norm": 1.5064566135406494, + "learning_rate": 9.748962936155079e-05, + "loss": 0.8053, + "step": 3822 + }, + { + "epoch": 0.25663568336632997, + "grad_norm": 1.0891283750534058, + "learning_rate": 9.748622776141892e-05, + "loss": 0.7495, + "step": 3824 + }, + { + "epoch": 0.256769907050099, + "grad_norm": 1.0988272428512573, + "learning_rate": 9.748282391765615e-05, + "loss": 0.6811, + "step": 3826 + }, + { + "epoch": 0.256904130733868, + "grad_norm": 1.0769731998443604, + "learning_rate": 9.747941783042332e-05, + "loss": 0.7194, + "step": 3828 + }, + { + "epoch": 0.257038354417637, + "grad_norm": 1.1300938129425049, + "learning_rate": 9.747600949988136e-05, + "loss": 0.7761, + "step": 3830 + }, + { + "epoch": 0.257172578101406, + "grad_norm": 1.0362656116485596, + "learning_rate": 9.747259892619132e-05, + "loss": 0.7048, + "step": 3832 + }, + { + "epoch": 0.257306801785175, + "grad_norm": 1.1735504865646362, + "learning_rate": 9.746918610951433e-05, + "loss": 0.7157, + "step": 3834 + }, + { + "epoch": 0.257441025468944, + "grad_norm": 1.0931183099746704, + "learning_rate": 9.746577105001163e-05, + "loss": 0.6933, + "step": 3836 + }, + { + "epoch": 0.257575249152713, + "grad_norm": 1.193933367729187, + "learning_rate": 9.74623537478446e-05, + "loss": 0.7582, + "step": 3838 + }, + { + "epoch": 0.257709472836482, + "grad_norm": 1.1191264390945435, + "learning_rate": 9.745893420317469e-05, + "loss": 0.6414, + "step": 3840 + }, + { + "epoch": 0.257843696520251, + "grad_norm": 1.2166557312011719, + "learning_rate": 9.745551241616344e-05, + "loss": 0.7545, + "step": 3842 + }, + { + "epoch": 0.25797792020402, + "grad_norm": 1.238582730293274, + "learning_rate": 9.745208838697255e-05, + "loss": 0.7233, + "step": 3844 + }, + { + "epoch": 0.258112143887789, + "grad_norm": 1.1257667541503906, + "learning_rate": 9.744866211576381e-05, + "loss": 0.7227, + "step": 3846 + }, + { + "epoch": 0.258246367571558, + "grad_norm": 1.0283488035202026, + "learning_rate": 9.744523360269909e-05, + "loss": 0.6334, + "step": 3848 + }, + { + "epoch": 0.25838059125532703, + "grad_norm": 1.229082703590393, + "learning_rate": 9.744180284794035e-05, + "loss": 0.7463, + "step": 3850 + }, + { + "epoch": 0.258514814939096, + "grad_norm": 1.1621450185775757, + "learning_rate": 9.743836985164974e-05, + "loss": 0.7193, + "step": 3852 + }, + { + "epoch": 0.258649038622865, + "grad_norm": 1.0491615533828735, + "learning_rate": 9.743493461398942e-05, + "loss": 0.7096, + "step": 3854 + }, + { + "epoch": 0.258783262306634, + "grad_norm": 1.0686851739883423, + "learning_rate": 9.743149713512175e-05, + "loss": 0.7385, + "step": 3856 + }, + { + "epoch": 0.258917485990403, + "grad_norm": 1.1188812255859375, + "learning_rate": 9.742805741520908e-05, + "loss": 0.7903, + "step": 3858 + }, + { + "epoch": 0.259051709674172, + "grad_norm": 1.0035182237625122, + "learning_rate": 9.742461545441398e-05, + "loss": 0.7312, + "step": 3860 + }, + { + "epoch": 0.259185933357941, + "grad_norm": 1.1080131530761719, + "learning_rate": 9.742117125289904e-05, + "loss": 0.7621, + "step": 3862 + }, + { + "epoch": 0.25932015704171, + "grad_norm": 1.1767661571502686, + "learning_rate": 9.741772481082702e-05, + "loss": 0.6926, + "step": 3864 + }, + { + "epoch": 0.259454380725479, + "grad_norm": 1.1401045322418213, + "learning_rate": 9.741427612836074e-05, + "loss": 0.7, + "step": 3866 + }, + { + "epoch": 0.259588604409248, + "grad_norm": 1.1692506074905396, + "learning_rate": 9.741082520566314e-05, + "loss": 0.7517, + "step": 3868 + }, + { + "epoch": 0.259722828093017, + "grad_norm": 1.2039555311203003, + "learning_rate": 9.740737204289729e-05, + "loss": 0.7437, + "step": 3870 + }, + { + "epoch": 0.25985705177678603, + "grad_norm": 1.1465946435928345, + "learning_rate": 9.740391664022633e-05, + "loss": 0.7207, + "step": 3872 + }, + { + "epoch": 0.259991275460555, + "grad_norm": 1.2403197288513184, + "learning_rate": 9.740045899781352e-05, + "loss": 0.7601, + "step": 3874 + }, + { + "epoch": 0.26012549914432403, + "grad_norm": 1.3701558113098145, + "learning_rate": 9.739699911582225e-05, + "loss": 0.8242, + "step": 3876 + }, + { + "epoch": 0.260259722828093, + "grad_norm": 1.1910128593444824, + "learning_rate": 9.739353699441596e-05, + "loss": 0.6879, + "step": 3878 + }, + { + "epoch": 0.26039394651186204, + "grad_norm": 1.1024234294891357, + "learning_rate": 9.739007263375823e-05, + "loss": 0.7299, + "step": 3880 + }, + { + "epoch": 0.260528170195631, + "grad_norm": 1.1535654067993164, + "learning_rate": 9.738660603401277e-05, + "loss": 0.7267, + "step": 3882 + }, + { + "epoch": 0.26066239387940005, + "grad_norm": 1.0920493602752686, + "learning_rate": 9.738313719534337e-05, + "loss": 0.7854, + "step": 3884 + }, + { + "epoch": 0.260796617563169, + "grad_norm": 1.7549275159835815, + "learning_rate": 9.73796661179139e-05, + "loss": 0.682, + "step": 3886 + }, + { + "epoch": 0.260930841246938, + "grad_norm": 1.212325930595398, + "learning_rate": 9.737619280188837e-05, + "loss": 0.7691, + "step": 3888 + }, + { + "epoch": 0.261065064930707, + "grad_norm": 1.0737411975860596, + "learning_rate": 9.737271724743088e-05, + "loss": 0.652, + "step": 3890 + }, + { + "epoch": 0.261199288614476, + "grad_norm": 1.1670502424240112, + "learning_rate": 9.736923945470568e-05, + "loss": 0.7321, + "step": 3892 + }, + { + "epoch": 0.26133351229824503, + "grad_norm": 1.2263405323028564, + "learning_rate": 9.736575942387706e-05, + "loss": 0.7116, + "step": 3894 + }, + { + "epoch": 0.261467735982014, + "grad_norm": 1.1557703018188477, + "learning_rate": 9.736227715510944e-05, + "loss": 0.7032, + "step": 3896 + }, + { + "epoch": 0.26160195966578303, + "grad_norm": 1.3054964542388916, + "learning_rate": 9.735879264856736e-05, + "loss": 0.7081, + "step": 3898 + }, + { + "epoch": 0.261736183349552, + "grad_norm": 1.119599461555481, + "learning_rate": 9.735530590441545e-05, + "loss": 0.6495, + "step": 3900 + }, + { + "epoch": 0.26187040703332104, + "grad_norm": 1.9122085571289062, + "learning_rate": 9.735181692281846e-05, + "loss": 0.7446, + "step": 3902 + }, + { + "epoch": 0.26200463071709, + "grad_norm": 1.135762095451355, + "learning_rate": 9.734832570394124e-05, + "loss": 0.7072, + "step": 3904 + }, + { + "epoch": 0.26213885440085904, + "grad_norm": 1.079358696937561, + "learning_rate": 9.734483224794872e-05, + "loss": 0.764, + "step": 3906 + }, + { + "epoch": 0.262273078084628, + "grad_norm": 1.1305729150772095, + "learning_rate": 9.7341336555006e-05, + "loss": 0.7986, + "step": 3908 + }, + { + "epoch": 0.26240730176839705, + "grad_norm": 1.1853477954864502, + "learning_rate": 9.73378386252782e-05, + "loss": 0.6972, + "step": 3910 + }, + { + "epoch": 0.262541525452166, + "grad_norm": 1.19010329246521, + "learning_rate": 9.73343384589306e-05, + "loss": 0.6651, + "step": 3912 + }, + { + "epoch": 0.26267574913593505, + "grad_norm": 1.41134774684906, + "learning_rate": 9.733083605612863e-05, + "loss": 0.7109, + "step": 3914 + }, + { + "epoch": 0.26280997281970403, + "grad_norm": 1.4223922491073608, + "learning_rate": 9.732733141703769e-05, + "loss": 0.6819, + "step": 3916 + }, + { + "epoch": 0.26294419650347306, + "grad_norm": 1.5214942693710327, + "learning_rate": 9.732382454182343e-05, + "loss": 0.8066, + "step": 3918 + }, + { + "epoch": 0.26307842018724203, + "grad_norm": 1.2094883918762207, + "learning_rate": 9.73203154306515e-05, + "loss": 0.6308, + "step": 3920 + }, + { + "epoch": 0.26321264387101106, + "grad_norm": 1.1626900434494019, + "learning_rate": 9.731680408368772e-05, + "loss": 0.6826, + "step": 3922 + }, + { + "epoch": 0.26334686755478004, + "grad_norm": 1.4367871284484863, + "learning_rate": 9.7313290501098e-05, + "loss": 0.7448, + "step": 3924 + }, + { + "epoch": 0.263481091238549, + "grad_norm": 1.5802116394042969, + "learning_rate": 9.730977468304834e-05, + "loss": 0.7858, + "step": 3926 + }, + { + "epoch": 0.26361531492231804, + "grad_norm": 1.0798490047454834, + "learning_rate": 9.730625662970485e-05, + "loss": 0.7485, + "step": 3928 + }, + { + "epoch": 0.263749538606087, + "grad_norm": 1.8378727436065674, + "learning_rate": 9.730273634123377e-05, + "loss": 0.6474, + "step": 3930 + }, + { + "epoch": 0.26388376228985605, + "grad_norm": 1.8537728786468506, + "learning_rate": 9.72992138178014e-05, + "loss": 0.7069, + "step": 3932 + }, + { + "epoch": 0.264017985973625, + "grad_norm": 1.439657211303711, + "learning_rate": 9.72956890595742e-05, + "loss": 0.6436, + "step": 3934 + }, + { + "epoch": 0.26415220965739405, + "grad_norm": 1.0422866344451904, + "learning_rate": 9.729216206671868e-05, + "loss": 0.6997, + "step": 3936 + }, + { + "epoch": 0.26428643334116303, + "grad_norm": 1.206998586654663, + "learning_rate": 9.728863283940151e-05, + "loss": 0.7195, + "step": 3938 + }, + { + "epoch": 0.26442065702493206, + "grad_norm": 1.0577679872512817, + "learning_rate": 9.728510137778944e-05, + "loss": 0.669, + "step": 3940 + }, + { + "epoch": 0.26455488070870103, + "grad_norm": 1.1963422298431396, + "learning_rate": 9.728156768204928e-05, + "loss": 0.7314, + "step": 3942 + }, + { + "epoch": 0.26468910439247006, + "grad_norm": 1.545266032218933, + "learning_rate": 9.727803175234804e-05, + "loss": 0.7109, + "step": 3944 + }, + { + "epoch": 0.26482332807623904, + "grad_norm": 1.0092322826385498, + "learning_rate": 9.727449358885276e-05, + "loss": 0.753, + "step": 3946 + }, + { + "epoch": 0.26495755176000807, + "grad_norm": 1.293685793876648, + "learning_rate": 9.727095319173065e-05, + "loss": 0.7639, + "step": 3948 + }, + { + "epoch": 0.26509177544377704, + "grad_norm": 1.1440637111663818, + "learning_rate": 9.726741056114892e-05, + "loss": 0.8083, + "step": 3950 + }, + { + "epoch": 0.2652259991275461, + "grad_norm": 1.0889347791671753, + "learning_rate": 9.726386569727501e-05, + "loss": 0.6597, + "step": 3952 + }, + { + "epoch": 0.26536022281131505, + "grad_norm": 1.1356900930404663, + "learning_rate": 9.726031860027637e-05, + "loss": 0.7465, + "step": 3954 + }, + { + "epoch": 0.2654944464950841, + "grad_norm": 1.1200767755508423, + "learning_rate": 9.725676927032061e-05, + "loss": 0.686, + "step": 3956 + }, + { + "epoch": 0.26562867017885305, + "grad_norm": 1.100643515586853, + "learning_rate": 9.725321770757545e-05, + "loss": 0.7333, + "step": 3958 + }, + { + "epoch": 0.2657628938626221, + "grad_norm": 1.022718906402588, + "learning_rate": 9.724966391220865e-05, + "loss": 0.792, + "step": 3960 + }, + { + "epoch": 0.26589711754639106, + "grad_norm": 1.3356891870498657, + "learning_rate": 9.724610788438815e-05, + "loss": 0.709, + "step": 3962 + }, + { + "epoch": 0.26603134123016003, + "grad_norm": 1.2211649417877197, + "learning_rate": 9.724254962428196e-05, + "loss": 0.6901, + "step": 3964 + }, + { + "epoch": 0.26616556491392906, + "grad_norm": 1.1071388721466064, + "learning_rate": 9.72389891320582e-05, + "loss": 0.6893, + "step": 3966 + }, + { + "epoch": 0.26629978859769804, + "grad_norm": 1.1717225313186646, + "learning_rate": 9.723542640788509e-05, + "loss": 0.7193, + "step": 3968 + }, + { + "epoch": 0.26643401228146707, + "grad_norm": 1.0993868112564087, + "learning_rate": 9.723186145193097e-05, + "loss": 0.7268, + "step": 3970 + }, + { + "epoch": 0.26656823596523604, + "grad_norm": 1.0340200662612915, + "learning_rate": 9.722829426436427e-05, + "loss": 0.7291, + "step": 3972 + }, + { + "epoch": 0.2667024596490051, + "grad_norm": 0.938714325428009, + "learning_rate": 9.722472484535354e-05, + "loss": 0.716, + "step": 3974 + }, + { + "epoch": 0.26683668333277405, + "grad_norm": 1.1499898433685303, + "learning_rate": 9.722115319506743e-05, + "loss": 0.7458, + "step": 3976 + }, + { + "epoch": 0.2669709070165431, + "grad_norm": 1.1436411142349243, + "learning_rate": 9.721757931367468e-05, + "loss": 0.7299, + "step": 3978 + }, + { + "epoch": 0.26710513070031205, + "grad_norm": 1.2237372398376465, + "learning_rate": 9.721400320134415e-05, + "loss": 0.7948, + "step": 3980 + }, + { + "epoch": 0.2672393543840811, + "grad_norm": 1.1530967950820923, + "learning_rate": 9.721042485824483e-05, + "loss": 0.7577, + "step": 3982 + }, + { + "epoch": 0.26737357806785006, + "grad_norm": 1.2369847297668457, + "learning_rate": 9.720684428454576e-05, + "loss": 0.6774, + "step": 3984 + }, + { + "epoch": 0.2675078017516191, + "grad_norm": 1.1894441843032837, + "learning_rate": 9.720326148041612e-05, + "loss": 0.7308, + "step": 3986 + }, + { + "epoch": 0.26764202543538806, + "grad_norm": 1.0823352336883545, + "learning_rate": 9.719967644602521e-05, + "loss": 0.7019, + "step": 3988 + }, + { + "epoch": 0.2677762491191571, + "grad_norm": 0.9706602096557617, + "learning_rate": 9.71960891815424e-05, + "loss": 0.6352, + "step": 3990 + }, + { + "epoch": 0.26791047280292607, + "grad_norm": 1.227729082107544, + "learning_rate": 9.719249968713717e-05, + "loss": 0.7034, + "step": 3992 + }, + { + "epoch": 0.2680446964866951, + "grad_norm": 1.4342589378356934, + "learning_rate": 9.718890796297914e-05, + "loss": 0.7531, + "step": 3994 + }, + { + "epoch": 0.2681789201704641, + "grad_norm": 1.027266025543213, + "learning_rate": 9.7185314009238e-05, + "loss": 0.6469, + "step": 3996 + }, + { + "epoch": 0.2683131438542331, + "grad_norm": 1.0618791580200195, + "learning_rate": 9.718171782608356e-05, + "loss": 0.7386, + "step": 3998 + }, + { + "epoch": 0.2684473675380021, + "grad_norm": 1.1263371706008911, + "learning_rate": 9.717811941368574e-05, + "loss": 0.696, + "step": 4000 + }, + { + "epoch": 0.26858159122177105, + "grad_norm": 1.2227270603179932, + "learning_rate": 9.717451877221453e-05, + "loss": 0.7411, + "step": 4002 + }, + { + "epoch": 0.2687158149055401, + "grad_norm": 1.1450586318969727, + "learning_rate": 9.717091590184008e-05, + "loss": 0.7522, + "step": 4004 + }, + { + "epoch": 0.26885003858930906, + "grad_norm": 1.0368503332138062, + "learning_rate": 9.71673108027326e-05, + "loss": 0.718, + "step": 4006 + }, + { + "epoch": 0.2689842622730781, + "grad_norm": 1.1287481784820557, + "learning_rate": 9.716370347506247e-05, + "loss": 0.7373, + "step": 4008 + }, + { + "epoch": 0.26911848595684706, + "grad_norm": 1.2839189767837524, + "learning_rate": 9.716009391900006e-05, + "loss": 0.6749, + "step": 4010 + }, + { + "epoch": 0.2692527096406161, + "grad_norm": 1.0585741996765137, + "learning_rate": 9.715648213471597e-05, + "loss": 0.7737, + "step": 4012 + }, + { + "epoch": 0.26938693332438507, + "grad_norm": 0.9859800338745117, + "learning_rate": 9.715286812238082e-05, + "loss": 0.6506, + "step": 4014 + }, + { + "epoch": 0.2695211570081541, + "grad_norm": 1.1902844905853271, + "learning_rate": 9.714925188216537e-05, + "loss": 0.8032, + "step": 4016 + }, + { + "epoch": 0.2696553806919231, + "grad_norm": 2.5191569328308105, + "learning_rate": 9.714563341424048e-05, + "loss": 0.7063, + "step": 4018 + }, + { + "epoch": 0.2697896043756921, + "grad_norm": 1.1799323558807373, + "learning_rate": 9.714201271877713e-05, + "loss": 0.7384, + "step": 4020 + }, + { + "epoch": 0.2699238280594611, + "grad_norm": 1.308539867401123, + "learning_rate": 9.713838979594638e-05, + "loss": 0.6858, + "step": 4022 + }, + { + "epoch": 0.2700580517432301, + "grad_norm": 1.099339485168457, + "learning_rate": 9.71347646459194e-05, + "loss": 0.679, + "step": 4024 + }, + { + "epoch": 0.2701922754269991, + "grad_norm": 1.1468719244003296, + "learning_rate": 9.713113726886747e-05, + "loss": 0.6644, + "step": 4026 + }, + { + "epoch": 0.2703264991107681, + "grad_norm": 1.1587730646133423, + "learning_rate": 9.712750766496201e-05, + "loss": 0.6816, + "step": 4028 + }, + { + "epoch": 0.2704607227945371, + "grad_norm": 1.0334495306015015, + "learning_rate": 9.712387583437445e-05, + "loss": 0.7228, + "step": 4030 + }, + { + "epoch": 0.2705949464783061, + "grad_norm": 1.2018488645553589, + "learning_rate": 9.712024177727645e-05, + "loss": 0.7465, + "step": 4032 + }, + { + "epoch": 0.2707291701620751, + "grad_norm": 1.5213717222213745, + "learning_rate": 9.711660549383967e-05, + "loss": 0.7471, + "step": 4034 + }, + { + "epoch": 0.2708633938458441, + "grad_norm": 1.8163784742355347, + "learning_rate": 9.711296698423593e-05, + "loss": 0.7182, + "step": 4036 + }, + { + "epoch": 0.2709976175296131, + "grad_norm": 1.145598292350769, + "learning_rate": 9.710932624863715e-05, + "loss": 0.7038, + "step": 4038 + }, + { + "epoch": 0.2711318412133821, + "grad_norm": 1.237273097038269, + "learning_rate": 9.710568328721534e-05, + "loss": 0.7144, + "step": 4040 + }, + { + "epoch": 0.2712660648971511, + "grad_norm": 1.2341762781143188, + "learning_rate": 9.710203810014262e-05, + "loss": 0.778, + "step": 4042 + }, + { + "epoch": 0.2714002885809201, + "grad_norm": 1.0736732482910156, + "learning_rate": 9.709839068759123e-05, + "loss": 0.6782, + "step": 4044 + }, + { + "epoch": 0.2715345122646891, + "grad_norm": 1.2260222434997559, + "learning_rate": 9.70947410497335e-05, + "loss": 0.6977, + "step": 4046 + }, + { + "epoch": 0.2716687359484581, + "grad_norm": 1.1734833717346191, + "learning_rate": 9.709108918674185e-05, + "loss": 0.7427, + "step": 4048 + }, + { + "epoch": 0.2718029596322271, + "grad_norm": 1.1103187799453735, + "learning_rate": 9.708743509878884e-05, + "loss": 0.7538, + "step": 4050 + }, + { + "epoch": 0.2719371833159961, + "grad_norm": 1.0368175506591797, + "learning_rate": 9.70837787860471e-05, + "loss": 0.7464, + "step": 4052 + }, + { + "epoch": 0.2720714069997651, + "grad_norm": 1.0606738328933716, + "learning_rate": 9.708012024868942e-05, + "loss": 0.6899, + "step": 4054 + }, + { + "epoch": 0.2722056306835341, + "grad_norm": 1.254096269607544, + "learning_rate": 9.707645948688863e-05, + "loss": 0.7723, + "step": 4056 + }, + { + "epoch": 0.2723398543673031, + "grad_norm": 1.1519017219543457, + "learning_rate": 9.70727965008177e-05, + "loss": 0.6875, + "step": 4058 + }, + { + "epoch": 0.2724740780510721, + "grad_norm": 2.8999814987182617, + "learning_rate": 9.706913129064971e-05, + "loss": 0.7196, + "step": 4060 + }, + { + "epoch": 0.27260830173484113, + "grad_norm": 1.135886549949646, + "learning_rate": 9.706546385655781e-05, + "loss": 0.7366, + "step": 4062 + }, + { + "epoch": 0.2727425254186101, + "grad_norm": 1.0506887435913086, + "learning_rate": 9.706179419871531e-05, + "loss": 0.6983, + "step": 4064 + }, + { + "epoch": 0.27287674910237913, + "grad_norm": 1.1773637533187866, + "learning_rate": 9.705812231729557e-05, + "loss": 0.7328, + "step": 4066 + }, + { + "epoch": 0.2730109727861481, + "grad_norm": 2.1838226318359375, + "learning_rate": 9.705444821247208e-05, + "loss": 0.7691, + "step": 4068 + }, + { + "epoch": 0.27314519646991714, + "grad_norm": 1.373435139656067, + "learning_rate": 9.705077188441844e-05, + "loss": 0.6969, + "step": 4070 + }, + { + "epoch": 0.2732794201536861, + "grad_norm": 1.2124875783920288, + "learning_rate": 9.704709333330836e-05, + "loss": 0.711, + "step": 4072 + }, + { + "epoch": 0.27341364383745514, + "grad_norm": 1.3007575273513794, + "learning_rate": 9.704341255931562e-05, + "loss": 0.6932, + "step": 4074 + }, + { + "epoch": 0.2735478675212241, + "grad_norm": 1.0921796560287476, + "learning_rate": 9.703972956261416e-05, + "loss": 0.7336, + "step": 4076 + }, + { + "epoch": 0.2736820912049931, + "grad_norm": 1.2120072841644287, + "learning_rate": 9.703604434337797e-05, + "loss": 0.7159, + "step": 4078 + }, + { + "epoch": 0.2738163148887621, + "grad_norm": 1.47358238697052, + "learning_rate": 9.703235690178118e-05, + "loss": 0.7175, + "step": 4080 + }, + { + "epoch": 0.2739505385725311, + "grad_norm": 1.1698148250579834, + "learning_rate": 9.7028667237998e-05, + "loss": 0.6971, + "step": 4082 + }, + { + "epoch": 0.27408476225630013, + "grad_norm": 1.2034008502960205, + "learning_rate": 9.702497535220278e-05, + "loss": 0.769, + "step": 4084 + }, + { + "epoch": 0.2742189859400691, + "grad_norm": 1.183003544807434, + "learning_rate": 9.702128124456996e-05, + "loss": 0.7371, + "step": 4086 + }, + { + "epoch": 0.27435320962383813, + "grad_norm": 1.3367705345153809, + "learning_rate": 9.701758491527404e-05, + "loss": 0.8031, + "step": 4088 + }, + { + "epoch": 0.2744874333076071, + "grad_norm": 1.158776044845581, + "learning_rate": 9.701388636448969e-05, + "loss": 0.7452, + "step": 4090 + }, + { + "epoch": 0.27462165699137614, + "grad_norm": 1.1235275268554688, + "learning_rate": 9.701018559239167e-05, + "loss": 0.7204, + "step": 4092 + }, + { + "epoch": 0.2747558806751451, + "grad_norm": 1.0379655361175537, + "learning_rate": 9.700648259915481e-05, + "loss": 0.6754, + "step": 4094 + }, + { + "epoch": 0.27489010435891414, + "grad_norm": 1.151607632637024, + "learning_rate": 9.700277738495409e-05, + "loss": 0.7668, + "step": 4096 + }, + { + "epoch": 0.2750243280426831, + "grad_norm": 1.1385493278503418, + "learning_rate": 9.699906994996457e-05, + "loss": 0.7487, + "step": 4098 + }, + { + "epoch": 0.27515855172645215, + "grad_norm": 1.1888922452926636, + "learning_rate": 9.69953602943614e-05, + "loss": 0.7085, + "step": 4100 + }, + { + "epoch": 0.2752927754102211, + "grad_norm": 1.007498860359192, + "learning_rate": 9.699164841831989e-05, + "loss": 0.7052, + "step": 4102 + }, + { + "epoch": 0.27542699909399015, + "grad_norm": 1.2998079061508179, + "learning_rate": 9.698793432201538e-05, + "loss": 0.7781, + "step": 4104 + }, + { + "epoch": 0.27556122277775913, + "grad_norm": 1.1245256662368774, + "learning_rate": 9.698421800562338e-05, + "loss": 0.7311, + "step": 4106 + }, + { + "epoch": 0.27569544646152816, + "grad_norm": 1.0242562294006348, + "learning_rate": 9.698049946931947e-05, + "loss": 0.7246, + "step": 4108 + }, + { + "epoch": 0.27582967014529713, + "grad_norm": 1.1672991514205933, + "learning_rate": 9.697677871327933e-05, + "loss": 0.6993, + "step": 4110 + }, + { + "epoch": 0.27596389382906616, + "grad_norm": 1.0333305597305298, + "learning_rate": 9.697305573767879e-05, + "loss": 0.7259, + "step": 4112 + }, + { + "epoch": 0.27609811751283514, + "grad_norm": 1.0109241008758545, + "learning_rate": 9.696933054269372e-05, + "loss": 0.6522, + "step": 4114 + }, + { + "epoch": 0.2762323411966041, + "grad_norm": 1.0484569072723389, + "learning_rate": 9.696560312850015e-05, + "loss": 0.7308, + "step": 4116 + }, + { + "epoch": 0.27636656488037314, + "grad_norm": 1.2440683841705322, + "learning_rate": 9.69618734952742e-05, + "loss": 0.7118, + "step": 4118 + }, + { + "epoch": 0.2765007885641421, + "grad_norm": 1.4809832572937012, + "learning_rate": 9.695814164319204e-05, + "loss": 0.7568, + "step": 4120 + }, + { + "epoch": 0.27663501224791115, + "grad_norm": 1.076040267944336, + "learning_rate": 9.695440757243005e-05, + "loss": 0.6586, + "step": 4122 + }, + { + "epoch": 0.2767692359316801, + "grad_norm": 1.0714213848114014, + "learning_rate": 9.695067128316463e-05, + "loss": 0.6833, + "step": 4124 + }, + { + "epoch": 0.27690345961544915, + "grad_norm": 1.0056148767471313, + "learning_rate": 9.694693277557232e-05, + "loss": 0.681, + "step": 4126 + }, + { + "epoch": 0.27703768329921813, + "grad_norm": 1.216422438621521, + "learning_rate": 9.694319204982974e-05, + "loss": 0.7266, + "step": 4128 + }, + { + "epoch": 0.27717190698298716, + "grad_norm": 1.3905694484710693, + "learning_rate": 9.693944910611365e-05, + "loss": 0.6919, + "step": 4130 + }, + { + "epoch": 0.27730613066675613, + "grad_norm": 1.0062493085861206, + "learning_rate": 9.693570394460091e-05, + "loss": 0.6494, + "step": 4132 + }, + { + "epoch": 0.27744035435052516, + "grad_norm": 1.5734044313430786, + "learning_rate": 9.693195656546843e-05, + "loss": 0.685, + "step": 4134 + }, + { + "epoch": 0.27757457803429414, + "grad_norm": 1.1906888484954834, + "learning_rate": 9.692820696889331e-05, + "loss": 0.7182, + "step": 4136 + }, + { + "epoch": 0.27770880171806317, + "grad_norm": 1.1630241870880127, + "learning_rate": 9.692445515505268e-05, + "loss": 0.6852, + "step": 4138 + }, + { + "epoch": 0.27784302540183214, + "grad_norm": 1.123489499092102, + "learning_rate": 9.692070112412382e-05, + "loss": 0.6891, + "step": 4140 + }, + { + "epoch": 0.2779772490856012, + "grad_norm": 1.1171189546585083, + "learning_rate": 9.69169448762841e-05, + "loss": 0.7437, + "step": 4142 + }, + { + "epoch": 0.27811147276937015, + "grad_norm": 1.5453418493270874, + "learning_rate": 9.691318641171099e-05, + "loss": 0.7297, + "step": 4144 + }, + { + "epoch": 0.2782456964531392, + "grad_norm": 1.3269058465957642, + "learning_rate": 9.690942573058207e-05, + "loss": 0.6759, + "step": 4146 + }, + { + "epoch": 0.27837992013690815, + "grad_norm": 1.1371320486068726, + "learning_rate": 9.690566283307503e-05, + "loss": 0.707, + "step": 4148 + }, + { + "epoch": 0.2785141438206772, + "grad_norm": 1.9872580766677856, + "learning_rate": 9.690189771936766e-05, + "loss": 0.6828, + "step": 4150 + }, + { + "epoch": 0.27864836750444616, + "grad_norm": 1.037161111831665, + "learning_rate": 9.689813038963784e-05, + "loss": 0.6583, + "step": 4152 + }, + { + "epoch": 0.27878259118821513, + "grad_norm": 1.2190905809402466, + "learning_rate": 9.68943608440636e-05, + "loss": 0.7327, + "step": 4154 + }, + { + "epoch": 0.27891681487198416, + "grad_norm": 1.0231003761291504, + "learning_rate": 9.6890589082823e-05, + "loss": 0.714, + "step": 4156 + }, + { + "epoch": 0.27905103855575314, + "grad_norm": 1.1530766487121582, + "learning_rate": 9.68868151060943e-05, + "loss": 0.7366, + "step": 4158 + }, + { + "epoch": 0.27918526223952217, + "grad_norm": 1.200112223625183, + "learning_rate": 9.688303891405576e-05, + "loss": 0.7856, + "step": 4160 + }, + { + "epoch": 0.27931948592329114, + "grad_norm": 1.2123029232025146, + "learning_rate": 9.687926050688583e-05, + "loss": 0.748, + "step": 4162 + }, + { + "epoch": 0.2794537096070602, + "grad_norm": 1.2193679809570312, + "learning_rate": 9.687547988476303e-05, + "loss": 0.7167, + "step": 4164 + }, + { + "epoch": 0.27958793329082915, + "grad_norm": 1.103472113609314, + "learning_rate": 9.687169704786599e-05, + "loss": 0.683, + "step": 4166 + }, + { + "epoch": 0.2797221569745982, + "grad_norm": 1.1211434602737427, + "learning_rate": 9.686791199637342e-05, + "loss": 0.6931, + "step": 4168 + }, + { + "epoch": 0.27985638065836715, + "grad_norm": 1.15460205078125, + "learning_rate": 9.686412473046418e-05, + "loss": 0.7261, + "step": 4170 + }, + { + "epoch": 0.2799906043421362, + "grad_norm": 0.9894868731498718, + "learning_rate": 9.686033525031719e-05, + "loss": 0.7467, + "step": 4172 + }, + { + "epoch": 0.28012482802590516, + "grad_norm": 1.2631585597991943, + "learning_rate": 9.685654355611151e-05, + "loss": 0.7588, + "step": 4174 + }, + { + "epoch": 0.2802590517096742, + "grad_norm": 1.0129464864730835, + "learning_rate": 9.685274964802629e-05, + "loss": 0.75, + "step": 4176 + }, + { + "epoch": 0.28039327539344316, + "grad_norm": 1.4937212467193604, + "learning_rate": 9.684895352624077e-05, + "loss": 0.667, + "step": 4178 + }, + { + "epoch": 0.2805274990772122, + "grad_norm": 1.0413357019424438, + "learning_rate": 9.684515519093433e-05, + "loss": 0.6708, + "step": 4180 + }, + { + "epoch": 0.28066172276098117, + "grad_norm": 1.2190228700637817, + "learning_rate": 9.684135464228643e-05, + "loss": 0.7621, + "step": 4182 + }, + { + "epoch": 0.2807959464447502, + "grad_norm": 1.024556040763855, + "learning_rate": 9.683755188047663e-05, + "loss": 0.701, + "step": 4184 + }, + { + "epoch": 0.2809301701285192, + "grad_norm": 1.073548436164856, + "learning_rate": 9.68337469056846e-05, + "loss": 0.7019, + "step": 4186 + }, + { + "epoch": 0.2810643938122882, + "grad_norm": 1.0884594917297363, + "learning_rate": 9.682993971809012e-05, + "loss": 0.7868, + "step": 4188 + }, + { + "epoch": 0.2811986174960572, + "grad_norm": 1.0661922693252563, + "learning_rate": 9.682613031787308e-05, + "loss": 0.7261, + "step": 4190 + }, + { + "epoch": 0.28133284117982615, + "grad_norm": 1.2550932168960571, + "learning_rate": 9.682231870521347e-05, + "loss": 0.726, + "step": 4192 + }, + { + "epoch": 0.2814670648635952, + "grad_norm": 1.5789484977722168, + "learning_rate": 9.681850488029136e-05, + "loss": 0.6917, + "step": 4194 + }, + { + "epoch": 0.28160128854736416, + "grad_norm": 1.1192291975021362, + "learning_rate": 9.681468884328697e-05, + "loss": 0.6583, + "step": 4196 + }, + { + "epoch": 0.2817355122311332, + "grad_norm": 1.3032783269882202, + "learning_rate": 9.681087059438059e-05, + "loss": 0.6452, + "step": 4198 + }, + { + "epoch": 0.28186973591490216, + "grad_norm": 1.3085541725158691, + "learning_rate": 9.680705013375261e-05, + "loss": 0.7515, + "step": 4200 + }, + { + "epoch": 0.2820039595986712, + "grad_norm": 1.0685060024261475, + "learning_rate": 9.680322746158357e-05, + "loss": 0.7283, + "step": 4202 + }, + { + "epoch": 0.28213818328244017, + "grad_norm": 0.9785438179969788, + "learning_rate": 9.679940257805404e-05, + "loss": 0.7339, + "step": 4204 + }, + { + "epoch": 0.2822724069662092, + "grad_norm": 1.0365419387817383, + "learning_rate": 9.679557548334479e-05, + "loss": 0.6785, + "step": 4206 + }, + { + "epoch": 0.2824066306499782, + "grad_norm": 1.2138546705245972, + "learning_rate": 9.679174617763662e-05, + "loss": 0.7403, + "step": 4208 + }, + { + "epoch": 0.2825408543337472, + "grad_norm": 1.4012144804000854, + "learning_rate": 9.678791466111044e-05, + "loss": 0.8049, + "step": 4210 + }, + { + "epoch": 0.2826750780175162, + "grad_norm": 1.1018271446228027, + "learning_rate": 9.67840809339473e-05, + "loss": 0.7087, + "step": 4212 + }, + { + "epoch": 0.2828093017012852, + "grad_norm": 1.0963165760040283, + "learning_rate": 9.678024499632834e-05, + "loss": 0.7312, + "step": 4214 + }, + { + "epoch": 0.2829435253850542, + "grad_norm": 1.218801498413086, + "learning_rate": 9.677640684843478e-05, + "loss": 0.7484, + "step": 4216 + }, + { + "epoch": 0.2830777490688232, + "grad_norm": 1.1199160814285278, + "learning_rate": 9.6772566490448e-05, + "loss": 0.6566, + "step": 4218 + }, + { + "epoch": 0.2832119727525922, + "grad_norm": 1.5163363218307495, + "learning_rate": 9.676872392254941e-05, + "loss": 0.6935, + "step": 4220 + }, + { + "epoch": 0.2833461964363612, + "grad_norm": 1.098585844039917, + "learning_rate": 9.67648791449206e-05, + "loss": 0.7089, + "step": 4222 + }, + { + "epoch": 0.2834804201201302, + "grad_norm": 1.1444215774536133, + "learning_rate": 9.676103215774319e-05, + "loss": 0.6433, + "step": 4224 + }, + { + "epoch": 0.2836146438038992, + "grad_norm": 1.0866196155548096, + "learning_rate": 9.675718296119898e-05, + "loss": 0.7273, + "step": 4226 + }, + { + "epoch": 0.2837488674876682, + "grad_norm": 1.1252814531326294, + "learning_rate": 9.675333155546979e-05, + "loss": 0.6798, + "step": 4228 + }, + { + "epoch": 0.2838830911714372, + "grad_norm": 1.1624395847320557, + "learning_rate": 9.674947794073765e-05, + "loss": 0.7036, + "step": 4230 + }, + { + "epoch": 0.2840173148552062, + "grad_norm": 1.0577150583267212, + "learning_rate": 9.674562211718459e-05, + "loss": 0.62, + "step": 4232 + }, + { + "epoch": 0.2841515385389752, + "grad_norm": 1.045554280281067, + "learning_rate": 9.674176408499281e-05, + "loss": 0.6564, + "step": 4234 + }, + { + "epoch": 0.2842857622227442, + "grad_norm": 1.1526782512664795, + "learning_rate": 9.67379038443446e-05, + "loss": 0.7117, + "step": 4236 + }, + { + "epoch": 0.2844199859065132, + "grad_norm": 1.0404531955718994, + "learning_rate": 9.673404139542233e-05, + "loss": 0.7032, + "step": 4238 + }, + { + "epoch": 0.2845542095902822, + "grad_norm": 1.041201114654541, + "learning_rate": 9.67301767384085e-05, + "loss": 0.7019, + "step": 4240 + }, + { + "epoch": 0.2846884332740512, + "grad_norm": 1.1981208324432373, + "learning_rate": 9.672630987348573e-05, + "loss": 0.7029, + "step": 4242 + }, + { + "epoch": 0.2848226569578202, + "grad_norm": 1.1003867387771606, + "learning_rate": 9.672244080083668e-05, + "loss": 0.7255, + "step": 4244 + }, + { + "epoch": 0.2849568806415892, + "grad_norm": 2.205414056777954, + "learning_rate": 9.671856952064418e-05, + "loss": 0.7091, + "step": 4246 + }, + { + "epoch": 0.2850911043253582, + "grad_norm": 1.0728976726531982, + "learning_rate": 9.671469603309113e-05, + "loss": 0.617, + "step": 4248 + }, + { + "epoch": 0.2852253280091272, + "grad_norm": 1.2097114324569702, + "learning_rate": 9.671082033836057e-05, + "loss": 0.7732, + "step": 4250 + }, + { + "epoch": 0.28535955169289623, + "grad_norm": 1.173802375793457, + "learning_rate": 9.67069424366356e-05, + "loss": 0.7405, + "step": 4252 + }, + { + "epoch": 0.2854937753766652, + "grad_norm": 1.1673674583435059, + "learning_rate": 9.670306232809945e-05, + "loss": 0.7246, + "step": 4254 + }, + { + "epoch": 0.28562799906043423, + "grad_norm": 1.9087454080581665, + "learning_rate": 9.669918001293543e-05, + "loss": 0.7625, + "step": 4256 + }, + { + "epoch": 0.2857622227442032, + "grad_norm": 1.0080323219299316, + "learning_rate": 9.669529549132699e-05, + "loss": 0.6792, + "step": 4258 + }, + { + "epoch": 0.28589644642797224, + "grad_norm": 1.1245241165161133, + "learning_rate": 9.669140876345765e-05, + "loss": 0.7519, + "step": 4260 + }, + { + "epoch": 0.2860306701117412, + "grad_norm": 1.1891417503356934, + "learning_rate": 9.668751982951108e-05, + "loss": 0.7294, + "step": 4262 + }, + { + "epoch": 0.28616489379551024, + "grad_norm": 0.9736157655715942, + "learning_rate": 9.6683628689671e-05, + "loss": 0.7211, + "step": 4264 + }, + { + "epoch": 0.2862991174792792, + "grad_norm": 1.0116606950759888, + "learning_rate": 9.667973534412125e-05, + "loss": 0.6763, + "step": 4266 + }, + { + "epoch": 0.2864333411630482, + "grad_norm": 1.6707595586776733, + "learning_rate": 9.66758397930458e-05, + "loss": 0.7135, + "step": 4268 + }, + { + "epoch": 0.2865675648468172, + "grad_norm": 1.379185676574707, + "learning_rate": 9.667194203662873e-05, + "loss": 0.6941, + "step": 4270 + }, + { + "epoch": 0.2867017885305862, + "grad_norm": 2.073378086090088, + "learning_rate": 9.666804207505414e-05, + "loss": 0.7383, + "step": 4272 + }, + { + "epoch": 0.28683601221435523, + "grad_norm": 1.1643736362457275, + "learning_rate": 9.666413990850635e-05, + "loss": 0.6586, + "step": 4274 + }, + { + "epoch": 0.2869702358981242, + "grad_norm": 0.9920051693916321, + "learning_rate": 9.666023553716971e-05, + "loss": 0.6988, + "step": 4276 + }, + { + "epoch": 0.28710445958189323, + "grad_norm": 1.1142268180847168, + "learning_rate": 9.665632896122869e-05, + "loss": 0.6993, + "step": 4278 + }, + { + "epoch": 0.2872386832656622, + "grad_norm": 1.9538670778274536, + "learning_rate": 9.665242018086786e-05, + "loss": 0.7178, + "step": 4280 + }, + { + "epoch": 0.28737290694943124, + "grad_norm": 2.505720376968384, + "learning_rate": 9.664850919627193e-05, + "loss": 0.7012, + "step": 4282 + }, + { + "epoch": 0.2875071306332002, + "grad_norm": 0.9926694631576538, + "learning_rate": 9.664459600762568e-05, + "loss": 0.7295, + "step": 4284 + }, + { + "epoch": 0.28764135431696924, + "grad_norm": 0.9303231239318848, + "learning_rate": 9.664068061511397e-05, + "loss": 0.6918, + "step": 4286 + }, + { + "epoch": 0.2877755780007382, + "grad_norm": 1.1642636060714722, + "learning_rate": 9.663676301892182e-05, + "loss": 0.7, + "step": 4288 + }, + { + "epoch": 0.28790980168450725, + "grad_norm": 1.8136674165725708, + "learning_rate": 9.663284321923434e-05, + "loss": 0.71, + "step": 4290 + }, + { + "epoch": 0.2880440253682762, + "grad_norm": 2.031097888946533, + "learning_rate": 9.66289212162367e-05, + "loss": 0.785, + "step": 4292 + }, + { + "epoch": 0.28817824905204525, + "grad_norm": 1.2717703580856323, + "learning_rate": 9.662499701011424e-05, + "loss": 0.7158, + "step": 4294 + }, + { + "epoch": 0.28831247273581423, + "grad_norm": 1.3626692295074463, + "learning_rate": 9.662107060105234e-05, + "loss": 0.663, + "step": 4296 + }, + { + "epoch": 0.28844669641958326, + "grad_norm": 1.123256802558899, + "learning_rate": 9.661714198923654e-05, + "loss": 0.6952, + "step": 4298 + }, + { + "epoch": 0.28858092010335223, + "grad_norm": 1.127862811088562, + "learning_rate": 9.661321117485244e-05, + "loss": 0.7416, + "step": 4300 + }, + { + "epoch": 0.28871514378712126, + "grad_norm": 1.430205225944519, + "learning_rate": 9.66092781580858e-05, + "loss": 0.6815, + "step": 4302 + }, + { + "epoch": 0.28884936747089024, + "grad_norm": 1.0847445726394653, + "learning_rate": 9.66053429391224e-05, + "loss": 0.7995, + "step": 4304 + }, + { + "epoch": 0.2889835911546592, + "grad_norm": 1.3557614088058472, + "learning_rate": 9.660140551814817e-05, + "loss": 0.7073, + "step": 4306 + }, + { + "epoch": 0.28911781483842824, + "grad_norm": 1.0713691711425781, + "learning_rate": 9.65974658953492e-05, + "loss": 0.6449, + "step": 4308 + }, + { + "epoch": 0.2892520385221972, + "grad_norm": 1.2485902309417725, + "learning_rate": 9.659352407091159e-05, + "loss": 0.6684, + "step": 4310 + }, + { + "epoch": 0.28938626220596625, + "grad_norm": 1.1673352718353271, + "learning_rate": 9.658958004502158e-05, + "loss": 0.7202, + "step": 4312 + }, + { + "epoch": 0.2895204858897352, + "grad_norm": 0.8654835224151611, + "learning_rate": 9.658563381786554e-05, + "loss": 0.61, + "step": 4314 + }, + { + "epoch": 0.28965470957350425, + "grad_norm": 1.395545244216919, + "learning_rate": 9.65816853896299e-05, + "loss": 0.6299, + "step": 4316 + }, + { + "epoch": 0.28978893325727323, + "grad_norm": 1.2566602230072021, + "learning_rate": 9.657773476050123e-05, + "loss": 0.7165, + "step": 4318 + }, + { + "epoch": 0.28992315694104226, + "grad_norm": 1.102645754814148, + "learning_rate": 9.657378193066617e-05, + "loss": 0.7139, + "step": 4320 + }, + { + "epoch": 0.29005738062481123, + "grad_norm": 1.0785845518112183, + "learning_rate": 9.65698269003115e-05, + "loss": 0.6792, + "step": 4322 + }, + { + "epoch": 0.29019160430858026, + "grad_norm": 1.2975529432296753, + "learning_rate": 9.656586966962411e-05, + "loss": 0.6619, + "step": 4324 + }, + { + "epoch": 0.29032582799234924, + "grad_norm": 1.0212528705596924, + "learning_rate": 9.656191023879092e-05, + "loss": 0.7072, + "step": 4326 + }, + { + "epoch": 0.29046005167611827, + "grad_norm": 0.9956158995628357, + "learning_rate": 9.655794860799904e-05, + "loss": 0.7268, + "step": 4328 + }, + { + "epoch": 0.29059427535988724, + "grad_norm": 1.0832117795944214, + "learning_rate": 9.655398477743565e-05, + "loss": 0.7002, + "step": 4330 + }, + { + "epoch": 0.2907284990436563, + "grad_norm": 1.1135586500167847, + "learning_rate": 9.6550018747288e-05, + "loss": 0.714, + "step": 4332 + }, + { + "epoch": 0.29086272272742525, + "grad_norm": 1.170379400253296, + "learning_rate": 9.654605051774352e-05, + "loss": 0.6517, + "step": 4334 + }, + { + "epoch": 0.2909969464111943, + "grad_norm": 1.1077511310577393, + "learning_rate": 9.654208008898968e-05, + "loss": 0.7301, + "step": 4336 + }, + { + "epoch": 0.29113117009496325, + "grad_norm": 1.5347226858139038, + "learning_rate": 9.653810746121407e-05, + "loss": 0.7131, + "step": 4338 + }, + { + "epoch": 0.2912653937787323, + "grad_norm": 1.0661876201629639, + "learning_rate": 9.65341326346044e-05, + "loss": 0.7085, + "step": 4340 + }, + { + "epoch": 0.29139961746250126, + "grad_norm": 1.0869132280349731, + "learning_rate": 9.653015560934846e-05, + "loss": 0.7179, + "step": 4342 + }, + { + "epoch": 0.29153384114627023, + "grad_norm": 2.238389015197754, + "learning_rate": 9.652617638563417e-05, + "loss": 0.8169, + "step": 4344 + }, + { + "epoch": 0.29166806483003926, + "grad_norm": 1.0856190919876099, + "learning_rate": 9.652219496364954e-05, + "loss": 0.721, + "step": 4346 + }, + { + "epoch": 0.29180228851380824, + "grad_norm": 1.2137451171875, + "learning_rate": 9.651821134358268e-05, + "loss": 0.7266, + "step": 4348 + }, + { + "epoch": 0.29193651219757727, + "grad_norm": 1.2103468179702759, + "learning_rate": 9.651422552562181e-05, + "loss": 0.6986, + "step": 4350 + }, + { + "epoch": 0.29207073588134624, + "grad_norm": 1.219045877456665, + "learning_rate": 9.651023750995525e-05, + "loss": 0.7148, + "step": 4352 + }, + { + "epoch": 0.2922049595651153, + "grad_norm": 1.0039916038513184, + "learning_rate": 9.650624729677141e-05, + "loss": 0.7019, + "step": 4354 + }, + { + "epoch": 0.29233918324888425, + "grad_norm": 1.551803469657898, + "learning_rate": 9.650225488625886e-05, + "loss": 0.7499, + "step": 4356 + }, + { + "epoch": 0.2924734069326533, + "grad_norm": 1.1392501592636108, + "learning_rate": 9.649826027860619e-05, + "loss": 0.6889, + "step": 4358 + }, + { + "epoch": 0.29260763061642225, + "grad_norm": 1.0280663967132568, + "learning_rate": 9.649426347400217e-05, + "loss": 0.679, + "step": 4360 + }, + { + "epoch": 0.2927418543001913, + "grad_norm": 1.082579493522644, + "learning_rate": 9.649026447263561e-05, + "loss": 0.7187, + "step": 4362 + }, + { + "epoch": 0.29287607798396026, + "grad_norm": 0.9832870960235596, + "learning_rate": 9.648626327469549e-05, + "loss": 0.7013, + "step": 4364 + }, + { + "epoch": 0.2930103016677293, + "grad_norm": 1.9217787981033325, + "learning_rate": 9.648225988037083e-05, + "loss": 0.6945, + "step": 4366 + }, + { + "epoch": 0.29314452535149826, + "grad_norm": 1.3493105173110962, + "learning_rate": 9.64782542898508e-05, + "loss": 0.7189, + "step": 4368 + }, + { + "epoch": 0.2932787490352673, + "grad_norm": 0.9947900772094727, + "learning_rate": 9.647424650332467e-05, + "loss": 0.668, + "step": 4370 + }, + { + "epoch": 0.29341297271903627, + "grad_norm": 1.1923651695251465, + "learning_rate": 9.647023652098174e-05, + "loss": 0.7023, + "step": 4372 + }, + { + "epoch": 0.2935471964028053, + "grad_norm": 1.172889232635498, + "learning_rate": 9.646622434301154e-05, + "loss": 0.7749, + "step": 4374 + }, + { + "epoch": 0.2936814200865743, + "grad_norm": 1.2299495935440063, + "learning_rate": 9.64622099696036e-05, + "loss": 0.7138, + "step": 4376 + }, + { + "epoch": 0.2938156437703433, + "grad_norm": 1.1374621391296387, + "learning_rate": 9.645819340094762e-05, + "loss": 0.684, + "step": 4378 + }, + { + "epoch": 0.2939498674541123, + "grad_norm": 1.3243783712387085, + "learning_rate": 9.645417463723335e-05, + "loss": 0.7502, + "step": 4380 + }, + { + "epoch": 0.29408409113788125, + "grad_norm": 1.0015956163406372, + "learning_rate": 9.645015367865067e-05, + "loss": 0.7016, + "step": 4382 + }, + { + "epoch": 0.2942183148216503, + "grad_norm": 1.0454777479171753, + "learning_rate": 9.644613052538957e-05, + "loss": 0.7469, + "step": 4384 + }, + { + "epoch": 0.29435253850541926, + "grad_norm": 1.108790397644043, + "learning_rate": 9.644210517764014e-05, + "loss": 0.6956, + "step": 4386 + }, + { + "epoch": 0.2944867621891883, + "grad_norm": 1.1916991472244263, + "learning_rate": 9.643807763559258e-05, + "loss": 0.6804, + "step": 4388 + }, + { + "epoch": 0.29462098587295726, + "grad_norm": 1.013068437576294, + "learning_rate": 9.643404789943713e-05, + "loss": 0.699, + "step": 4390 + }, + { + "epoch": 0.2947552095567263, + "grad_norm": 0.9876919984817505, + "learning_rate": 9.643001596936427e-05, + "loss": 0.671, + "step": 4392 + }, + { + "epoch": 0.29488943324049527, + "grad_norm": 1.1245033740997314, + "learning_rate": 9.642598184556442e-05, + "loss": 0.7324, + "step": 4394 + }, + { + "epoch": 0.2950236569242643, + "grad_norm": 1.155880331993103, + "learning_rate": 9.642194552822823e-05, + "loss": 0.7868, + "step": 4396 + }, + { + "epoch": 0.2951578806080333, + "grad_norm": 1.117091417312622, + "learning_rate": 9.64179070175464e-05, + "loss": 0.7233, + "step": 4398 + }, + { + "epoch": 0.2952921042918023, + "grad_norm": 1.099109411239624, + "learning_rate": 9.641386631370976e-05, + "loss": 0.7429, + "step": 4400 + }, + { + "epoch": 0.2954263279755713, + "grad_norm": 1.0578813552856445, + "learning_rate": 9.640982341690918e-05, + "loss": 0.6906, + "step": 4402 + }, + { + "epoch": 0.2955605516593403, + "grad_norm": 1.064645528793335, + "learning_rate": 9.640577832733571e-05, + "loss": 0.7581, + "step": 4404 + }, + { + "epoch": 0.2956947753431093, + "grad_norm": 0.9522479772567749, + "learning_rate": 9.640173104518047e-05, + "loss": 0.692, + "step": 4406 + }, + { + "epoch": 0.2958289990268783, + "grad_norm": 1.1719841957092285, + "learning_rate": 9.63976815706347e-05, + "loss": 0.7418, + "step": 4408 + }, + { + "epoch": 0.2959632227106473, + "grad_norm": 1.7045280933380127, + "learning_rate": 9.639362990388969e-05, + "loss": 0.7086, + "step": 4410 + }, + { + "epoch": 0.2960974463944163, + "grad_norm": 1.1607673168182373, + "learning_rate": 9.63895760451369e-05, + "loss": 0.7272, + "step": 4412 + }, + { + "epoch": 0.2962316700781853, + "grad_norm": 1.157461404800415, + "learning_rate": 9.638551999456786e-05, + "loss": 0.729, + "step": 4414 + }, + { + "epoch": 0.2963658937619543, + "grad_norm": 1.1352097988128662, + "learning_rate": 9.638146175237421e-05, + "loss": 0.718, + "step": 4416 + }, + { + "epoch": 0.2965001174457233, + "grad_norm": 1.4608018398284912, + "learning_rate": 9.637740131874771e-05, + "loss": 0.7127, + "step": 4418 + }, + { + "epoch": 0.2966343411294923, + "grad_norm": 1.2369742393493652, + "learning_rate": 9.63733386938802e-05, + "loss": 0.7211, + "step": 4420 + }, + { + "epoch": 0.2967685648132613, + "grad_norm": 1.0688071250915527, + "learning_rate": 9.636927387796361e-05, + "loss": 0.6845, + "step": 4422 + }, + { + "epoch": 0.2969027884970303, + "grad_norm": 1.10688316822052, + "learning_rate": 9.636520687119002e-05, + "loss": 0.6995, + "step": 4424 + }, + { + "epoch": 0.2970370121807993, + "grad_norm": 0.9537732005119324, + "learning_rate": 9.636113767375158e-05, + "loss": 0.6519, + "step": 4426 + }, + { + "epoch": 0.2971712358645683, + "grad_norm": 1.0720096826553345, + "learning_rate": 9.635706628584054e-05, + "loss": 0.6166, + "step": 4428 + }, + { + "epoch": 0.2973054595483373, + "grad_norm": 1.4682979583740234, + "learning_rate": 9.63529927076493e-05, + "loss": 0.6336, + "step": 4430 + }, + { + "epoch": 0.2974396832321063, + "grad_norm": 1.0823240280151367, + "learning_rate": 9.634891693937026e-05, + "loss": 0.6681, + "step": 4432 + }, + { + "epoch": 0.2975739069158753, + "grad_norm": 1.2542070150375366, + "learning_rate": 9.634483898119608e-05, + "loss": 0.6513, + "step": 4434 + }, + { + "epoch": 0.2977081305996443, + "grad_norm": 2.113034248352051, + "learning_rate": 9.634075883331937e-05, + "loss": 0.7436, + "step": 4436 + }, + { + "epoch": 0.2978423542834133, + "grad_norm": 1.283736228942871, + "learning_rate": 9.633667649593294e-05, + "loss": 0.6824, + "step": 4438 + }, + { + "epoch": 0.2979765779671823, + "grad_norm": 1.3601338863372803, + "learning_rate": 9.633259196922966e-05, + "loss": 0.6553, + "step": 4440 + }, + { + "epoch": 0.29811080165095133, + "grad_norm": 1.5369936227798462, + "learning_rate": 9.632850525340251e-05, + "loss": 0.7418, + "step": 4442 + }, + { + "epoch": 0.2982450253347203, + "grad_norm": 1.0400851964950562, + "learning_rate": 9.63244163486446e-05, + "loss": 0.6815, + "step": 4444 + }, + { + "epoch": 0.29837924901848933, + "grad_norm": 1.124594807624817, + "learning_rate": 9.63203252551491e-05, + "loss": 0.6522, + "step": 4446 + }, + { + "epoch": 0.2985134727022583, + "grad_norm": 1.2869653701782227, + "learning_rate": 9.63162319731093e-05, + "loss": 0.798, + "step": 4448 + }, + { + "epoch": 0.29864769638602734, + "grad_norm": 1.1588083505630493, + "learning_rate": 9.631213650271864e-05, + "loss": 0.7211, + "step": 4450 + }, + { + "epoch": 0.2987819200697963, + "grad_norm": 0.9154360890388489, + "learning_rate": 9.630803884417061e-05, + "loss": 0.6761, + "step": 4452 + }, + { + "epoch": 0.29891614375356534, + "grad_norm": 0.9952070713043213, + "learning_rate": 9.630393899765878e-05, + "loss": 0.7719, + "step": 4454 + }, + { + "epoch": 0.2990503674373343, + "grad_norm": 1.147867202758789, + "learning_rate": 9.62998369633769e-05, + "loss": 0.7057, + "step": 4456 + }, + { + "epoch": 0.2991845911211033, + "grad_norm": 1.7422637939453125, + "learning_rate": 9.629573274151876e-05, + "loss": 0.6712, + "step": 4458 + }, + { + "epoch": 0.2993188148048723, + "grad_norm": 1.0809777975082397, + "learning_rate": 9.62916263322783e-05, + "loss": 0.7134, + "step": 4460 + }, + { + "epoch": 0.2994530384886413, + "grad_norm": 1.0853334665298462, + "learning_rate": 9.628751773584951e-05, + "loss": 0.6778, + "step": 4462 + }, + { + "epoch": 0.29958726217241033, + "grad_norm": 1.066697120666504, + "learning_rate": 9.628340695242652e-05, + "loss": 0.7024, + "step": 4464 + }, + { + "epoch": 0.2997214858561793, + "grad_norm": 1.0515681505203247, + "learning_rate": 9.627929398220358e-05, + "loss": 0.6421, + "step": 4466 + }, + { + "epoch": 0.29985570953994833, + "grad_norm": 1.065488338470459, + "learning_rate": 9.6275178825375e-05, + "loss": 0.6713, + "step": 4468 + }, + { + "epoch": 0.2999899332237173, + "grad_norm": 1.2193208932876587, + "learning_rate": 9.627106148213522e-05, + "loss": 0.6997, + "step": 4470 + }, + { + "epoch": 0.30012415690748634, + "grad_norm": 1.091118335723877, + "learning_rate": 9.626694195267876e-05, + "loss": 0.6888, + "step": 4472 + }, + { + "epoch": 0.3002583805912553, + "grad_norm": 1.1276484727859497, + "learning_rate": 9.626282023720028e-05, + "loss": 0.7017, + "step": 4474 + }, + { + "epoch": 0.30039260427502434, + "grad_norm": 1.1557211875915527, + "learning_rate": 9.625869633589453e-05, + "loss": 0.6885, + "step": 4476 + }, + { + "epoch": 0.3005268279587933, + "grad_norm": 1.1216962337493896, + "learning_rate": 9.625457024895632e-05, + "loss": 0.7308, + "step": 4478 + }, + { + "epoch": 0.30066105164256235, + "grad_norm": 1.0295031070709229, + "learning_rate": 9.625044197658063e-05, + "loss": 0.6237, + "step": 4480 + }, + { + "epoch": 0.3007952753263313, + "grad_norm": 1.1262531280517578, + "learning_rate": 9.624631151896251e-05, + "loss": 0.6779, + "step": 4482 + }, + { + "epoch": 0.30092949901010035, + "grad_norm": 1.1350386142730713, + "learning_rate": 9.62421788762971e-05, + "loss": 0.6781, + "step": 4484 + }, + { + "epoch": 0.30106372269386933, + "grad_norm": 1.081994652748108, + "learning_rate": 9.623804404877967e-05, + "loss": 0.7855, + "step": 4486 + }, + { + "epoch": 0.30119794637763836, + "grad_norm": 1.0471010208129883, + "learning_rate": 9.623390703660559e-05, + "loss": 0.758, + "step": 4488 + }, + { + "epoch": 0.30133217006140733, + "grad_norm": 1.1623107194900513, + "learning_rate": 9.62297678399703e-05, + "loss": 0.8105, + "step": 4490 + }, + { + "epoch": 0.3014663937451763, + "grad_norm": 0.9763181209564209, + "learning_rate": 9.62256264590694e-05, + "loss": 0.6587, + "step": 4492 + }, + { + "epoch": 0.30160061742894534, + "grad_norm": 1.1519794464111328, + "learning_rate": 9.622148289409855e-05, + "loss": 0.6951, + "step": 4494 + }, + { + "epoch": 0.3017348411127143, + "grad_norm": 1.0183149576187134, + "learning_rate": 9.621733714525353e-05, + "loss": 0.6649, + "step": 4496 + }, + { + "epoch": 0.30186906479648334, + "grad_norm": 1.2081207036972046, + "learning_rate": 9.621318921273021e-05, + "loss": 0.7239, + "step": 4498 + }, + { + "epoch": 0.3020032884802523, + "grad_norm": 1.100765585899353, + "learning_rate": 9.620903909672457e-05, + "loss": 0.6723, + "step": 4500 + }, + { + "epoch": 0.30213751216402135, + "grad_norm": 1.1360970735549927, + "learning_rate": 9.620488679743269e-05, + "loss": 0.6997, + "step": 4502 + }, + { + "epoch": 0.3022717358477903, + "grad_norm": 1.055786371231079, + "learning_rate": 9.620073231505078e-05, + "loss": 0.6099, + "step": 4504 + }, + { + "epoch": 0.30240595953155935, + "grad_norm": 1.0756354331970215, + "learning_rate": 9.61965756497751e-05, + "loss": 0.6755, + "step": 4506 + }, + { + "epoch": 0.30254018321532833, + "grad_norm": 1.2146291732788086, + "learning_rate": 9.619241680180209e-05, + "loss": 0.6447, + "step": 4508 + }, + { + "epoch": 0.30267440689909736, + "grad_norm": 1.0501223802566528, + "learning_rate": 9.61882557713282e-05, + "loss": 0.7336, + "step": 4510 + }, + { + "epoch": 0.30280863058286633, + "grad_norm": 1.0172021389007568, + "learning_rate": 9.618409255855006e-05, + "loss": 0.6122, + "step": 4512 + }, + { + "epoch": 0.30294285426663536, + "grad_norm": 1.4607539176940918, + "learning_rate": 9.617992716366435e-05, + "loss": 0.7468, + "step": 4514 + }, + { + "epoch": 0.30307707795040434, + "grad_norm": 1.0794657468795776, + "learning_rate": 9.61757595868679e-05, + "loss": 0.7288, + "step": 4516 + }, + { + "epoch": 0.30321130163417337, + "grad_norm": 1.0251612663269043, + "learning_rate": 9.617158982835761e-05, + "loss": 0.7091, + "step": 4518 + }, + { + "epoch": 0.30334552531794234, + "grad_norm": 0.9694096446037292, + "learning_rate": 9.61674178883305e-05, + "loss": 0.7073, + "step": 4520 + }, + { + "epoch": 0.3034797490017114, + "grad_norm": 1.239879846572876, + "learning_rate": 9.616324376698366e-05, + "loss": 0.6995, + "step": 4522 + }, + { + "epoch": 0.30361397268548035, + "grad_norm": 1.0317822694778442, + "learning_rate": 9.615906746451435e-05, + "loss": 0.6413, + "step": 4524 + }, + { + "epoch": 0.3037481963692494, + "grad_norm": 1.1398820877075195, + "learning_rate": 9.615488898111985e-05, + "loss": 0.6995, + "step": 4526 + }, + { + "epoch": 0.30388242005301835, + "grad_norm": 1.0622879266738892, + "learning_rate": 9.615070831699762e-05, + "loss": 0.6748, + "step": 4528 + }, + { + "epoch": 0.30401664373678733, + "grad_norm": 1.4225927591323853, + "learning_rate": 9.614652547234516e-05, + "loss": 0.7386, + "step": 4530 + }, + { + "epoch": 0.30415086742055636, + "grad_norm": 1.211207389831543, + "learning_rate": 9.614234044736012e-05, + "loss": 0.7205, + "step": 4532 + }, + { + "epoch": 0.30428509110432533, + "grad_norm": 1.1871228218078613, + "learning_rate": 9.613815324224023e-05, + "loss": 0.743, + "step": 4534 + }, + { + "epoch": 0.30441931478809436, + "grad_norm": 1.1464089155197144, + "learning_rate": 9.613396385718334e-05, + "loss": 0.7394, + "step": 4536 + }, + { + "epoch": 0.30455353847186334, + "grad_norm": 1.1008272171020508, + "learning_rate": 9.612977229238735e-05, + "loss": 0.7076, + "step": 4538 + }, + { + "epoch": 0.30468776215563237, + "grad_norm": 1.11635422706604, + "learning_rate": 9.612557854805036e-05, + "loss": 0.769, + "step": 4540 + }, + { + "epoch": 0.30482198583940134, + "grad_norm": 1.2722200155258179, + "learning_rate": 9.612138262437046e-05, + "loss": 0.7641, + "step": 4542 + }, + { + "epoch": 0.3049562095231704, + "grad_norm": 2.5558648109436035, + "learning_rate": 9.611718452154594e-05, + "loss": 0.7306, + "step": 4544 + }, + { + "epoch": 0.30509043320693935, + "grad_norm": 1.1128522157669067, + "learning_rate": 9.611298423977512e-05, + "loss": 0.7284, + "step": 4546 + }, + { + "epoch": 0.3052246568907084, + "grad_norm": 0.9909085035324097, + "learning_rate": 9.610878177925648e-05, + "loss": 0.7257, + "step": 4548 + }, + { + "epoch": 0.30535888057447735, + "grad_norm": 1.0871301889419556, + "learning_rate": 9.610457714018857e-05, + "loss": 0.7338, + "step": 4550 + }, + { + "epoch": 0.3054931042582464, + "grad_norm": 1.0989935398101807, + "learning_rate": 9.610037032277007e-05, + "loss": 0.6772, + "step": 4552 + }, + { + "epoch": 0.30562732794201536, + "grad_norm": 1.3562226295471191, + "learning_rate": 9.609616132719971e-05, + "loss": 0.7434, + "step": 4554 + }, + { + "epoch": 0.3057615516257844, + "grad_norm": 0.993280827999115, + "learning_rate": 9.609195015367636e-05, + "loss": 0.6481, + "step": 4556 + }, + { + "epoch": 0.30589577530955336, + "grad_norm": 1.5082858800888062, + "learning_rate": 9.608773680239902e-05, + "loss": 0.7755, + "step": 4558 + }, + { + "epoch": 0.3060299989933224, + "grad_norm": 1.336323857307434, + "learning_rate": 9.608352127356672e-05, + "loss": 0.6694, + "step": 4560 + }, + { + "epoch": 0.30616422267709137, + "grad_norm": 0.9614841938018799, + "learning_rate": 9.607930356737869e-05, + "loss": 0.6919, + "step": 4562 + }, + { + "epoch": 0.3062984463608604, + "grad_norm": 1.0784282684326172, + "learning_rate": 9.607508368403415e-05, + "loss": 0.6726, + "step": 4564 + }, + { + "epoch": 0.3064326700446294, + "grad_norm": 2.376894474029541, + "learning_rate": 9.607086162373253e-05, + "loss": 0.7332, + "step": 4566 + }, + { + "epoch": 0.30656689372839835, + "grad_norm": 1.072051763534546, + "learning_rate": 9.606663738667328e-05, + "loss": 0.7185, + "step": 4568 + }, + { + "epoch": 0.3067011174121674, + "grad_norm": 1.0158129930496216, + "learning_rate": 9.6062410973056e-05, + "loss": 0.6662, + "step": 4570 + }, + { + "epoch": 0.30683534109593635, + "grad_norm": 1.059302806854248, + "learning_rate": 9.605818238308038e-05, + "loss": 0.6646, + "step": 4572 + }, + { + "epoch": 0.3069695647797054, + "grad_norm": 1.0993070602416992, + "learning_rate": 9.605395161694621e-05, + "loss": 0.7113, + "step": 4574 + }, + { + "epoch": 0.30710378846347436, + "grad_norm": 1.1712669134140015, + "learning_rate": 9.60497186748534e-05, + "loss": 0.8254, + "step": 4576 + }, + { + "epoch": 0.3072380121472434, + "grad_norm": 1.1879925727844238, + "learning_rate": 9.604548355700194e-05, + "loss": 0.734, + "step": 4578 + }, + { + "epoch": 0.30737223583101236, + "grad_norm": 1.0698761940002441, + "learning_rate": 9.60412462635919e-05, + "loss": 0.6339, + "step": 4580 + }, + { + "epoch": 0.3075064595147814, + "grad_norm": 0.9980600476264954, + "learning_rate": 9.603700679482352e-05, + "loss": 0.6412, + "step": 4582 + }, + { + "epoch": 0.30764068319855037, + "grad_norm": 0.9460391402244568, + "learning_rate": 9.603276515089711e-05, + "loss": 0.6941, + "step": 4584 + }, + { + "epoch": 0.3077749068823194, + "grad_norm": 1.068634033203125, + "learning_rate": 9.602852133201305e-05, + "loss": 0.6952, + "step": 4586 + }, + { + "epoch": 0.3079091305660884, + "grad_norm": 1.2626107931137085, + "learning_rate": 9.602427533837188e-05, + "loss": 0.7765, + "step": 4588 + }, + { + "epoch": 0.3080433542498574, + "grad_norm": 1.057005763053894, + "learning_rate": 9.60200271701742e-05, + "loss": 0.7368, + "step": 4590 + }, + { + "epoch": 0.3081775779336264, + "grad_norm": 0.9701996445655823, + "learning_rate": 9.601577682762072e-05, + "loss": 0.6842, + "step": 4592 + }, + { + "epoch": 0.3083118016173954, + "grad_norm": 1.1067408323287964, + "learning_rate": 9.60115243109123e-05, + "loss": 0.7384, + "step": 4594 + }, + { + "epoch": 0.3084460253011644, + "grad_norm": 1.0144931077957153, + "learning_rate": 9.60072696202498e-05, + "loss": 0.727, + "step": 4596 + }, + { + "epoch": 0.3085802489849334, + "grad_norm": 1.1220800876617432, + "learning_rate": 9.60030127558343e-05, + "loss": 0.673, + "step": 4598 + }, + { + "epoch": 0.3087144726687024, + "grad_norm": 1.1260027885437012, + "learning_rate": 9.59987537178669e-05, + "loss": 0.6752, + "step": 4600 + }, + { + "epoch": 0.3088486963524714, + "grad_norm": 1.149194359779358, + "learning_rate": 9.599449250654884e-05, + "loss": 0.7066, + "step": 4602 + }, + { + "epoch": 0.3089829200362404, + "grad_norm": 0.9491012096405029, + "learning_rate": 9.599022912208145e-05, + "loss": 0.6873, + "step": 4604 + }, + { + "epoch": 0.30911714372000937, + "grad_norm": 1.353082537651062, + "learning_rate": 9.598596356466618e-05, + "loss": 0.6817, + "step": 4606 + }, + { + "epoch": 0.3092513674037784, + "grad_norm": 1.115274429321289, + "learning_rate": 9.598169583450455e-05, + "loss": 0.6553, + "step": 4608 + }, + { + "epoch": 0.3093855910875474, + "grad_norm": 1.2805801630020142, + "learning_rate": 9.597742593179822e-05, + "loss": 0.6605, + "step": 4610 + }, + { + "epoch": 0.3095198147713164, + "grad_norm": 1.1053426265716553, + "learning_rate": 9.597315385674893e-05, + "loss": 0.6783, + "step": 4612 + }, + { + "epoch": 0.3096540384550854, + "grad_norm": 0.9258078932762146, + "learning_rate": 9.596887960955849e-05, + "loss": 0.6579, + "step": 4614 + }, + { + "epoch": 0.3097882621388544, + "grad_norm": 1.1637978553771973, + "learning_rate": 9.596460319042891e-05, + "loss": 0.6987, + "step": 4616 + }, + { + "epoch": 0.3099224858226234, + "grad_norm": 1.219173789024353, + "learning_rate": 9.596032459956222e-05, + "loss": 0.7033, + "step": 4618 + }, + { + "epoch": 0.3100567095063924, + "grad_norm": 1.0405182838439941, + "learning_rate": 9.595604383716055e-05, + "loss": 0.7149, + "step": 4620 + }, + { + "epoch": 0.3101909331901614, + "grad_norm": 1.7751362323760986, + "learning_rate": 9.59517609034262e-05, + "loss": 0.7148, + "step": 4622 + }, + { + "epoch": 0.3103251568739304, + "grad_norm": 1.0881564617156982, + "learning_rate": 9.594747579856149e-05, + "loss": 0.7243, + "step": 4624 + }, + { + "epoch": 0.3104593805576994, + "grad_norm": 1.2595254182815552, + "learning_rate": 9.59431885227689e-05, + "loss": 0.685, + "step": 4626 + }, + { + "epoch": 0.3105936042414684, + "grad_norm": 1.3012375831604004, + "learning_rate": 9.5938899076251e-05, + "loss": 0.68, + "step": 4628 + }, + { + "epoch": 0.3107278279252374, + "grad_norm": 1.1929621696472168, + "learning_rate": 9.593460745921046e-05, + "loss": 0.7239, + "step": 4630 + }, + { + "epoch": 0.31086205160900643, + "grad_norm": 1.0787216424942017, + "learning_rate": 9.593031367185003e-05, + "loss": 0.6862, + "step": 4632 + }, + { + "epoch": 0.3109962752927754, + "grad_norm": 1.2381178140640259, + "learning_rate": 9.592601771437261e-05, + "loss": 0.6973, + "step": 4634 + }, + { + "epoch": 0.31113049897654443, + "grad_norm": 1.4702860116958618, + "learning_rate": 9.592171958698115e-05, + "loss": 0.752, + "step": 4636 + }, + { + "epoch": 0.3112647226603134, + "grad_norm": 1.1221622228622437, + "learning_rate": 9.591741928987876e-05, + "loss": 0.7334, + "step": 4638 + }, + { + "epoch": 0.31139894634408244, + "grad_norm": 1.3333992958068848, + "learning_rate": 9.591311682326859e-05, + "loss": 0.7177, + "step": 4640 + }, + { + "epoch": 0.3115331700278514, + "grad_norm": 0.9714601039886475, + "learning_rate": 9.590881218735394e-05, + "loss": 0.6296, + "step": 4642 + }, + { + "epoch": 0.3116673937116204, + "grad_norm": 1.1566107273101807, + "learning_rate": 9.590450538233817e-05, + "loss": 0.6886, + "step": 4644 + }, + { + "epoch": 0.3118016173953894, + "grad_norm": 1.1195974349975586, + "learning_rate": 9.590019640842482e-05, + "loss": 0.6605, + "step": 4646 + }, + { + "epoch": 0.3119358410791584, + "grad_norm": 1.0662175416946411, + "learning_rate": 9.589588526581741e-05, + "loss": 0.7264, + "step": 4648 + }, + { + "epoch": 0.3120700647629274, + "grad_norm": 1.082502007484436, + "learning_rate": 9.58915719547197e-05, + "loss": 0.7297, + "step": 4650 + }, + { + "epoch": 0.3122042884466964, + "grad_norm": 1.0617055892944336, + "learning_rate": 9.588725647533545e-05, + "loss": 0.7096, + "step": 4652 + }, + { + "epoch": 0.31233851213046543, + "grad_norm": 0.9861075282096863, + "learning_rate": 9.588293882786857e-05, + "loss": 0.6781, + "step": 4654 + }, + { + "epoch": 0.3124727358142344, + "grad_norm": 1.15412175655365, + "learning_rate": 9.587861901252305e-05, + "loss": 0.648, + "step": 4656 + }, + { + "epoch": 0.31260695949800343, + "grad_norm": 1.5281078815460205, + "learning_rate": 9.5874297029503e-05, + "loss": 0.7575, + "step": 4658 + }, + { + "epoch": 0.3127411831817724, + "grad_norm": 1.079062819480896, + "learning_rate": 9.586997287901262e-05, + "loss": 0.7238, + "step": 4660 + }, + { + "epoch": 0.31287540686554144, + "grad_norm": 1.133266806602478, + "learning_rate": 9.586564656125623e-05, + "loss": 0.7354, + "step": 4662 + }, + { + "epoch": 0.3130096305493104, + "grad_norm": 1.4366016387939453, + "learning_rate": 9.586131807643822e-05, + "loss": 0.6718, + "step": 4664 + }, + { + "epoch": 0.31314385423307944, + "grad_norm": 0.9520801305770874, + "learning_rate": 9.585698742476311e-05, + "loss": 0.6562, + "step": 4666 + }, + { + "epoch": 0.3132780779168484, + "grad_norm": 1.0416120290756226, + "learning_rate": 9.585265460643553e-05, + "loss": 0.7258, + "step": 4668 + }, + { + "epoch": 0.31341230160061745, + "grad_norm": 1.092458724975586, + "learning_rate": 9.584831962166017e-05, + "loss": 0.6946, + "step": 4670 + }, + { + "epoch": 0.3135465252843864, + "grad_norm": 1.0674761533737183, + "learning_rate": 9.584398247064188e-05, + "loss": 0.6967, + "step": 4672 + }, + { + "epoch": 0.31368074896815545, + "grad_norm": 1.082927942276001, + "learning_rate": 9.583964315358555e-05, + "loss": 0.6804, + "step": 4674 + }, + { + "epoch": 0.31381497265192443, + "grad_norm": 1.2659660577774048, + "learning_rate": 9.583530167069626e-05, + "loss": 0.7403, + "step": 4676 + }, + { + "epoch": 0.31394919633569346, + "grad_norm": 0.9640771150588989, + "learning_rate": 9.583095802217905e-05, + "loss": 0.6655, + "step": 4678 + }, + { + "epoch": 0.31408342001946243, + "grad_norm": 1.516160011291504, + "learning_rate": 9.582661220823922e-05, + "loss": 0.6648, + "step": 4680 + }, + { + "epoch": 0.3142176437032314, + "grad_norm": 1.02597177028656, + "learning_rate": 9.582226422908207e-05, + "loss": 0.7132, + "step": 4682 + }, + { + "epoch": 0.31435186738700044, + "grad_norm": 1.0739665031433105, + "learning_rate": 9.581791408491305e-05, + "loss": 0.6212, + "step": 4684 + }, + { + "epoch": 0.3144860910707694, + "grad_norm": 1.6776080131530762, + "learning_rate": 9.581356177593767e-05, + "loss": 0.7355, + "step": 4686 + }, + { + "epoch": 0.31462031475453844, + "grad_norm": 1.3557602167129517, + "learning_rate": 9.58092073023616e-05, + "loss": 0.7082, + "step": 4688 + }, + { + "epoch": 0.3147545384383074, + "grad_norm": 1.0693676471710205, + "learning_rate": 9.580485066439056e-05, + "loss": 0.6973, + "step": 4690 + }, + { + "epoch": 0.31488876212207645, + "grad_norm": 1.0692919492721558, + "learning_rate": 9.58004918622304e-05, + "loss": 0.6769, + "step": 4692 + }, + { + "epoch": 0.3150229858058454, + "grad_norm": 1.411782145500183, + "learning_rate": 9.579613089608705e-05, + "loss": 0.6878, + "step": 4694 + }, + { + "epoch": 0.31515720948961445, + "grad_norm": 1.0561375617980957, + "learning_rate": 9.579176776616658e-05, + "loss": 0.6816, + "step": 4696 + }, + { + "epoch": 0.31529143317338343, + "grad_norm": 1.1592587232589722, + "learning_rate": 9.578740247267514e-05, + "loss": 0.7211, + "step": 4698 + }, + { + "epoch": 0.31542565685715246, + "grad_norm": 0.9945803880691528, + "learning_rate": 9.578303501581895e-05, + "loss": 0.6684, + "step": 4700 + }, + { + "epoch": 0.31555988054092143, + "grad_norm": 1.0777732133865356, + "learning_rate": 9.57786653958044e-05, + "loss": 0.6935, + "step": 4702 + }, + { + "epoch": 0.31569410422469046, + "grad_norm": 0.9888481497764587, + "learning_rate": 9.577429361283792e-05, + "loss": 0.6799, + "step": 4704 + }, + { + "epoch": 0.31582832790845944, + "grad_norm": 1.0442769527435303, + "learning_rate": 9.576991966712607e-05, + "loss": 0.6754, + "step": 4706 + }, + { + "epoch": 0.31596255159222847, + "grad_norm": 1.0910941362380981, + "learning_rate": 9.576554355887554e-05, + "loss": 0.6806, + "step": 4708 + }, + { + "epoch": 0.31609677527599744, + "grad_norm": 1.2258208990097046, + "learning_rate": 9.576116528829306e-05, + "loss": 0.726, + "step": 4710 + }, + { + "epoch": 0.3162309989597665, + "grad_norm": 1.060171127319336, + "learning_rate": 9.575678485558551e-05, + "loss": 0.7486, + "step": 4712 + }, + { + "epoch": 0.31636522264353545, + "grad_norm": 1.1440684795379639, + "learning_rate": 9.575240226095984e-05, + "loss": 0.6969, + "step": 4714 + }, + { + "epoch": 0.3164994463273045, + "grad_norm": 1.1293154954910278, + "learning_rate": 9.574801750462315e-05, + "loss": 0.7284, + "step": 4716 + }, + { + "epoch": 0.31663367001107345, + "grad_norm": 1.2209240198135376, + "learning_rate": 9.574363058678257e-05, + "loss": 0.74, + "step": 4718 + }, + { + "epoch": 0.31676789369484243, + "grad_norm": 1.7266669273376465, + "learning_rate": 9.573924150764541e-05, + "loss": 0.7009, + "step": 4720 + }, + { + "epoch": 0.31690211737861146, + "grad_norm": 1.1648743152618408, + "learning_rate": 9.573485026741902e-05, + "loss": 0.6951, + "step": 4722 + }, + { + "epoch": 0.31703634106238043, + "grad_norm": 1.3447242975234985, + "learning_rate": 9.57304568663109e-05, + "loss": 0.6777, + "step": 4724 + }, + { + "epoch": 0.31717056474614946, + "grad_norm": 1.0942497253417969, + "learning_rate": 9.572606130452862e-05, + "loss": 0.6868, + "step": 4726 + }, + { + "epoch": 0.31730478842991844, + "grad_norm": 1.0130993127822876, + "learning_rate": 9.572166358227985e-05, + "loss": 0.6962, + "step": 4728 + }, + { + "epoch": 0.31743901211368747, + "grad_norm": 1.2938907146453857, + "learning_rate": 9.571726369977239e-05, + "loss": 0.7458, + "step": 4730 + }, + { + "epoch": 0.31757323579745644, + "grad_norm": 1.0879647731781006, + "learning_rate": 9.571286165721412e-05, + "loss": 0.7154, + "step": 4732 + }, + { + "epoch": 0.3177074594812255, + "grad_norm": 1.0983290672302246, + "learning_rate": 9.570845745481303e-05, + "loss": 0.717, + "step": 4734 + }, + { + "epoch": 0.31784168316499445, + "grad_norm": 0.8874186277389526, + "learning_rate": 9.570405109277719e-05, + "loss": 0.6164, + "step": 4736 + }, + { + "epoch": 0.3179759068487635, + "grad_norm": 1.0488560199737549, + "learning_rate": 9.569964257131484e-05, + "loss": 0.7797, + "step": 4738 + }, + { + "epoch": 0.31811013053253245, + "grad_norm": 1.1975808143615723, + "learning_rate": 9.56952318906342e-05, + "loss": 0.7539, + "step": 4740 + }, + { + "epoch": 0.3182443542163015, + "grad_norm": 1.1216825246810913, + "learning_rate": 9.569081905094375e-05, + "loss": 0.6643, + "step": 4742 + }, + { + "epoch": 0.31837857790007046, + "grad_norm": 1.1932557821273804, + "learning_rate": 9.568640405245192e-05, + "loss": 0.6596, + "step": 4744 + }, + { + "epoch": 0.3185128015838395, + "grad_norm": 1.1482906341552734, + "learning_rate": 9.568198689536734e-05, + "loss": 0.7082, + "step": 4746 + }, + { + "epoch": 0.31864702526760846, + "grad_norm": 1.2021764516830444, + "learning_rate": 9.567756757989872e-05, + "loss": 0.6848, + "step": 4748 + }, + { + "epoch": 0.3187812489513775, + "grad_norm": 2.35219144821167, + "learning_rate": 9.567314610625485e-05, + "loss": 0.6858, + "step": 4750 + }, + { + "epoch": 0.31891547263514647, + "grad_norm": 1.1158368587493896, + "learning_rate": 9.566872247464464e-05, + "loss": 0.6739, + "step": 4752 + }, + { + "epoch": 0.3190496963189155, + "grad_norm": 1.0408796072006226, + "learning_rate": 9.56642966852771e-05, + "loss": 0.7017, + "step": 4754 + }, + { + "epoch": 0.3191839200026845, + "grad_norm": 1.1878182888031006, + "learning_rate": 9.565986873836132e-05, + "loss": 0.7339, + "step": 4756 + }, + { + "epoch": 0.31931814368645345, + "grad_norm": 1.1423181295394897, + "learning_rate": 9.565543863410654e-05, + "loss": 0.6595, + "step": 4758 + }, + { + "epoch": 0.3194523673702225, + "grad_norm": 1.1221919059753418, + "learning_rate": 9.565100637272206e-05, + "loss": 0.6839, + "step": 4760 + }, + { + "epoch": 0.31958659105399145, + "grad_norm": 1.0802282094955444, + "learning_rate": 9.564657195441731e-05, + "loss": 0.6891, + "step": 4762 + }, + { + "epoch": 0.3197208147377605, + "grad_norm": 1.8912112712860107, + "learning_rate": 9.564213537940177e-05, + "loss": 0.7711, + "step": 4764 + }, + { + "epoch": 0.31985503842152946, + "grad_norm": 1.4157387018203735, + "learning_rate": 9.563769664788511e-05, + "loss": 0.6955, + "step": 4766 + }, + { + "epoch": 0.3199892621052985, + "grad_norm": 1.271424412727356, + "learning_rate": 9.563325576007701e-05, + "loss": 0.7484, + "step": 4768 + }, + { + "epoch": 0.32012348578906746, + "grad_norm": 1.1220924854278564, + "learning_rate": 9.562881271618732e-05, + "loss": 0.6604, + "step": 4770 + }, + { + "epoch": 0.3202577094728365, + "grad_norm": 1.0575189590454102, + "learning_rate": 9.562436751642593e-05, + "loss": 0.6773, + "step": 4772 + }, + { + "epoch": 0.32039193315660547, + "grad_norm": 1.2651679515838623, + "learning_rate": 9.561992016100293e-05, + "loss": 0.6567, + "step": 4774 + }, + { + "epoch": 0.3205261568403745, + "grad_norm": 1.2005423307418823, + "learning_rate": 9.561547065012839e-05, + "loss": 0.7591, + "step": 4776 + }, + { + "epoch": 0.3206603805241435, + "grad_norm": 1.3069040775299072, + "learning_rate": 9.561101898401255e-05, + "loss": 0.7619, + "step": 4778 + }, + { + "epoch": 0.3207946042079125, + "grad_norm": 1.2713435888290405, + "learning_rate": 9.560656516286577e-05, + "loss": 0.7043, + "step": 4780 + }, + { + "epoch": 0.3209288278916815, + "grad_norm": 1.2976057529449463, + "learning_rate": 9.560210918689847e-05, + "loss": 0.7565, + "step": 4782 + }, + { + "epoch": 0.3210630515754505, + "grad_norm": 1.0996086597442627, + "learning_rate": 9.559765105632117e-05, + "loss": 0.7015, + "step": 4784 + }, + { + "epoch": 0.3211972752592195, + "grad_norm": 1.0666567087173462, + "learning_rate": 9.559319077134453e-05, + "loss": 0.6563, + "step": 4786 + }, + { + "epoch": 0.3213314989429885, + "grad_norm": 1.154266357421875, + "learning_rate": 9.558872833217927e-05, + "loss": 0.6679, + "step": 4788 + }, + { + "epoch": 0.3214657226267575, + "grad_norm": 1.0351626873016357, + "learning_rate": 9.558426373903626e-05, + "loss": 0.6951, + "step": 4790 + }, + { + "epoch": 0.3215999463105265, + "grad_norm": 1.0043928623199463, + "learning_rate": 9.557979699212642e-05, + "loss": 0.6525, + "step": 4792 + }, + { + "epoch": 0.3217341699942955, + "grad_norm": 1.2971715927124023, + "learning_rate": 9.557532809166079e-05, + "loss": 0.7109, + "step": 4794 + }, + { + "epoch": 0.32186839367806447, + "grad_norm": 1.2481286525726318, + "learning_rate": 9.557085703785054e-05, + "loss": 0.7872, + "step": 4796 + }, + { + "epoch": 0.3220026173618335, + "grad_norm": 1.089882493019104, + "learning_rate": 9.55663838309069e-05, + "loss": 0.7012, + "step": 4798 + }, + { + "epoch": 0.3221368410456025, + "grad_norm": 1.3476847410202026, + "learning_rate": 9.556190847104123e-05, + "loss": 0.7281, + "step": 4800 + }, + { + "epoch": 0.3222710647293715, + "grad_norm": 1.4910799264907837, + "learning_rate": 9.555743095846497e-05, + "loss": 0.6774, + "step": 4802 + }, + { + "epoch": 0.3224052884131405, + "grad_norm": 1.1183050870895386, + "learning_rate": 9.555295129338969e-05, + "loss": 0.635, + "step": 4804 + }, + { + "epoch": 0.3225395120969095, + "grad_norm": 1.0580859184265137, + "learning_rate": 9.554846947602704e-05, + "loss": 0.7013, + "step": 4806 + }, + { + "epoch": 0.3226737357806785, + "grad_norm": 1.1748672723770142, + "learning_rate": 9.554398550658876e-05, + "loss": 0.6827, + "step": 4808 + }, + { + "epoch": 0.3228079594644475, + "grad_norm": 1.1253255605697632, + "learning_rate": 9.553949938528675e-05, + "loss": 0.6665, + "step": 4810 + }, + { + "epoch": 0.3229421831482165, + "grad_norm": 1.010771632194519, + "learning_rate": 9.553501111233292e-05, + "loss": 0.7021, + "step": 4812 + }, + { + "epoch": 0.3230764068319855, + "grad_norm": 1.1532608270645142, + "learning_rate": 9.553052068793937e-05, + "loss": 0.7278, + "step": 4814 + }, + { + "epoch": 0.3232106305157545, + "grad_norm": 1.0641151666641235, + "learning_rate": 9.552602811231824e-05, + "loss": 0.7161, + "step": 4816 + }, + { + "epoch": 0.3233448541995235, + "grad_norm": 1.0781819820404053, + "learning_rate": 9.552153338568181e-05, + "loss": 0.7942, + "step": 4818 + }, + { + "epoch": 0.3234790778832925, + "grad_norm": 1.099227786064148, + "learning_rate": 9.551703650824243e-05, + "loss": 0.7583, + "step": 4820 + }, + { + "epoch": 0.32361330156706153, + "grad_norm": 1.1302250623703003, + "learning_rate": 9.551253748021259e-05, + "loss": 0.7083, + "step": 4822 + }, + { + "epoch": 0.3237475252508305, + "grad_norm": 1.2145006656646729, + "learning_rate": 9.550803630180485e-05, + "loss": 0.7189, + "step": 4824 + }, + { + "epoch": 0.32388174893459953, + "grad_norm": 1.0137630701065063, + "learning_rate": 9.55035329732319e-05, + "loss": 0.6715, + "step": 4826 + }, + { + "epoch": 0.3240159726183685, + "grad_norm": 0.9891653060913086, + "learning_rate": 9.549902749470646e-05, + "loss": 0.6566, + "step": 4828 + }, + { + "epoch": 0.32415019630213754, + "grad_norm": 1.0337892770767212, + "learning_rate": 9.549451986644147e-05, + "loss": 0.6769, + "step": 4830 + }, + { + "epoch": 0.3242844199859065, + "grad_norm": 1.083341360092163, + "learning_rate": 9.549001008864987e-05, + "loss": 0.7068, + "step": 4832 + }, + { + "epoch": 0.3244186436696755, + "grad_norm": 1.1820749044418335, + "learning_rate": 9.548549816154473e-05, + "loss": 0.7483, + "step": 4834 + }, + { + "epoch": 0.3245528673534445, + "grad_norm": 0.8702498078346252, + "learning_rate": 9.548098408533926e-05, + "loss": 0.6265, + "step": 4836 + }, + { + "epoch": 0.3246870910372135, + "grad_norm": 1.1651930809020996, + "learning_rate": 9.547646786024673e-05, + "loss": 0.7112, + "step": 4838 + }, + { + "epoch": 0.3248213147209825, + "grad_norm": 1.0169198513031006, + "learning_rate": 9.547194948648051e-05, + "loss": 0.6766, + "step": 4840 + }, + { + "epoch": 0.3249555384047515, + "grad_norm": 1.4236356019973755, + "learning_rate": 9.546742896425409e-05, + "loss": 0.7752, + "step": 4842 + }, + { + "epoch": 0.32508976208852053, + "grad_norm": 1.0329314470291138, + "learning_rate": 9.546290629378107e-05, + "loss": 0.7007, + "step": 4844 + }, + { + "epoch": 0.3252239857722895, + "grad_norm": 0.963914155960083, + "learning_rate": 9.545838147527512e-05, + "loss": 0.6976, + "step": 4846 + }, + { + "epoch": 0.32535820945605853, + "grad_norm": 1.0919042825698853, + "learning_rate": 9.545385450895003e-05, + "loss": 0.7028, + "step": 4848 + }, + { + "epoch": 0.3254924331398275, + "grad_norm": 1.2006242275238037, + "learning_rate": 9.544932539501971e-05, + "loss": 0.6799, + "step": 4850 + }, + { + "epoch": 0.32562665682359654, + "grad_norm": 1.0634151697158813, + "learning_rate": 9.544479413369814e-05, + "loss": 0.7073, + "step": 4852 + }, + { + "epoch": 0.3257608805073655, + "grad_norm": 1.1870208978652954, + "learning_rate": 9.54402607251994e-05, + "loss": 0.7336, + "step": 4854 + }, + { + "epoch": 0.32589510419113454, + "grad_norm": 1.11781907081604, + "learning_rate": 9.543572516973769e-05, + "loss": 0.6152, + "step": 4856 + }, + { + "epoch": 0.3260293278749035, + "grad_norm": 1.5316169261932373, + "learning_rate": 9.543118746752733e-05, + "loss": 0.6583, + "step": 4858 + }, + { + "epoch": 0.32616355155867255, + "grad_norm": 1.018977165222168, + "learning_rate": 9.542664761878269e-05, + "loss": 0.691, + "step": 4860 + }, + { + "epoch": 0.3262977752424415, + "grad_norm": 1.1612788438796997, + "learning_rate": 9.542210562371828e-05, + "loss": 0.685, + "step": 4862 + }, + { + "epoch": 0.32643199892621055, + "grad_norm": 1.6843246221542358, + "learning_rate": 9.541756148254867e-05, + "loss": 0.6789, + "step": 4864 + }, + { + "epoch": 0.32656622260997953, + "grad_norm": 1.2154302597045898, + "learning_rate": 9.541301519548861e-05, + "loss": 0.7006, + "step": 4866 + }, + { + "epoch": 0.32670044629374856, + "grad_norm": 0.9375450015068054, + "learning_rate": 9.54084667627529e-05, + "loss": 0.6265, + "step": 4868 + }, + { + "epoch": 0.32683466997751753, + "grad_norm": 1.1680549383163452, + "learning_rate": 9.540391618455641e-05, + "loss": 0.7199, + "step": 4870 + }, + { + "epoch": 0.3269688936612865, + "grad_norm": 1.1197963953018188, + "learning_rate": 9.539936346111416e-05, + "loss": 0.7492, + "step": 4872 + }, + { + "epoch": 0.32710311734505554, + "grad_norm": 1.323502540588379, + "learning_rate": 9.539480859264128e-05, + "loss": 0.7417, + "step": 4874 + }, + { + "epoch": 0.3272373410288245, + "grad_norm": 1.6119531393051147, + "learning_rate": 9.539025157935292e-05, + "loss": 0.7183, + "step": 4876 + }, + { + "epoch": 0.32737156471259354, + "grad_norm": 1.082277774810791, + "learning_rate": 9.538569242146447e-05, + "loss": 0.7381, + "step": 4878 + }, + { + "epoch": 0.3275057883963625, + "grad_norm": 1.2153085470199585, + "learning_rate": 9.53811311191913e-05, + "loss": 0.6995, + "step": 4880 + }, + { + "epoch": 0.32764001208013155, + "grad_norm": 1.1135528087615967, + "learning_rate": 9.53765676727489e-05, + "loss": 0.6961, + "step": 4882 + }, + { + "epoch": 0.3277742357639005, + "grad_norm": 1.0991860628128052, + "learning_rate": 9.537200208235291e-05, + "loss": 0.723, + "step": 4884 + }, + { + "epoch": 0.32790845944766955, + "grad_norm": 1.1873701810836792, + "learning_rate": 9.536743434821904e-05, + "loss": 0.667, + "step": 4886 + }, + { + "epoch": 0.32804268313143853, + "grad_norm": 1.292967677116394, + "learning_rate": 9.536286447056311e-05, + "loss": 0.7801, + "step": 4888 + }, + { + "epoch": 0.32817690681520756, + "grad_norm": 1.1023390293121338, + "learning_rate": 9.535829244960104e-05, + "loss": 0.7174, + "step": 4890 + }, + { + "epoch": 0.32831113049897653, + "grad_norm": 1.0848222970962524, + "learning_rate": 9.535371828554884e-05, + "loss": 0.6969, + "step": 4892 + }, + { + "epoch": 0.32844535418274556, + "grad_norm": 1.024688482284546, + "learning_rate": 9.534914197862266e-05, + "loss": 0.6672, + "step": 4894 + }, + { + "epoch": 0.32857957786651454, + "grad_norm": 1.1134281158447266, + "learning_rate": 9.534456352903866e-05, + "loss": 0.6991, + "step": 4896 + }, + { + "epoch": 0.32871380155028357, + "grad_norm": 1.0506900548934937, + "learning_rate": 9.533998293701323e-05, + "loss": 0.6822, + "step": 4898 + }, + { + "epoch": 0.32884802523405254, + "grad_norm": 1.1279277801513672, + "learning_rate": 9.533540020276274e-05, + "loss": 0.719, + "step": 4900 + }, + { + "epoch": 0.3289822489178216, + "grad_norm": 1.0671149492263794, + "learning_rate": 9.533081532650375e-05, + "loss": 0.6608, + "step": 4902 + }, + { + "epoch": 0.32911647260159055, + "grad_norm": 1.2312167882919312, + "learning_rate": 9.532622830845287e-05, + "loss": 0.6822, + "step": 4904 + }, + { + "epoch": 0.3292506962853596, + "grad_norm": 1.6732779741287231, + "learning_rate": 9.532163914882685e-05, + "loss": 0.6851, + "step": 4906 + }, + { + "epoch": 0.32938491996912855, + "grad_norm": 1.053810477256775, + "learning_rate": 9.531704784784248e-05, + "loss": 0.644, + "step": 4908 + }, + { + "epoch": 0.32951914365289753, + "grad_norm": 0.9314039945602417, + "learning_rate": 9.531245440571672e-05, + "loss": 0.6335, + "step": 4910 + }, + { + "epoch": 0.32965336733666656, + "grad_norm": 1.0545871257781982, + "learning_rate": 9.53078588226666e-05, + "loss": 0.6814, + "step": 4912 + }, + { + "epoch": 0.32978759102043553, + "grad_norm": 1.5250245332717896, + "learning_rate": 9.530326109890924e-05, + "loss": 0.6458, + "step": 4914 + }, + { + "epoch": 0.32992181470420456, + "grad_norm": 1.0881139039993286, + "learning_rate": 9.529866123466187e-05, + "loss": 0.6858, + "step": 4916 + }, + { + "epoch": 0.33005603838797354, + "grad_norm": 1.2018465995788574, + "learning_rate": 9.529405923014183e-05, + "loss": 0.6176, + "step": 4918 + }, + { + "epoch": 0.33019026207174257, + "grad_norm": 1.7062848806381226, + "learning_rate": 9.528945508556656e-05, + "loss": 0.7139, + "step": 4920 + }, + { + "epoch": 0.33032448575551154, + "grad_norm": 0.940497100353241, + "learning_rate": 9.528484880115361e-05, + "loss": 0.6638, + "step": 4922 + }, + { + "epoch": 0.3304587094392806, + "grad_norm": 1.4594491720199585, + "learning_rate": 9.52802403771206e-05, + "loss": 0.8803, + "step": 4924 + }, + { + "epoch": 0.33059293312304955, + "grad_norm": 1.187603235244751, + "learning_rate": 9.527562981368525e-05, + "loss": 0.6924, + "step": 4926 + }, + { + "epoch": 0.3307271568068186, + "grad_norm": 1.187653660774231, + "learning_rate": 9.527101711106546e-05, + "loss": 0.6564, + "step": 4928 + }, + { + "epoch": 0.33086138049058755, + "grad_norm": 1.141374945640564, + "learning_rate": 9.52664022694791e-05, + "loss": 0.7499, + "step": 4930 + }, + { + "epoch": 0.3309956041743566, + "grad_norm": 1.122341275215149, + "learning_rate": 9.526178528914425e-05, + "loss": 0.6705, + "step": 4932 + }, + { + "epoch": 0.33112982785812556, + "grad_norm": 1.0938045978546143, + "learning_rate": 9.525716617027906e-05, + "loss": 0.6893, + "step": 4934 + }, + { + "epoch": 0.3312640515418946, + "grad_norm": 1.0191075801849365, + "learning_rate": 9.525254491310176e-05, + "loss": 0.6375, + "step": 4936 + }, + { + "epoch": 0.33139827522566356, + "grad_norm": 1.026029109954834, + "learning_rate": 9.524792151783069e-05, + "loss": 0.6443, + "step": 4938 + }, + { + "epoch": 0.3315324989094326, + "grad_norm": 1.1697235107421875, + "learning_rate": 9.524329598468431e-05, + "loss": 0.6307, + "step": 4940 + }, + { + "epoch": 0.33166672259320157, + "grad_norm": 1.4177303314208984, + "learning_rate": 9.523866831388116e-05, + "loss": 0.727, + "step": 4942 + }, + { + "epoch": 0.3318009462769706, + "grad_norm": 1.0223388671875, + "learning_rate": 9.52340385056399e-05, + "loss": 0.6492, + "step": 4944 + }, + { + "epoch": 0.3319351699607396, + "grad_norm": 1.0136746168136597, + "learning_rate": 9.522940656017926e-05, + "loss": 0.7031, + "step": 4946 + }, + { + "epoch": 0.33206939364450855, + "grad_norm": 1.5003118515014648, + "learning_rate": 9.52247724777181e-05, + "loss": 0.6505, + "step": 4948 + }, + { + "epoch": 0.3322036173282776, + "grad_norm": 1.07683265209198, + "learning_rate": 9.522013625847537e-05, + "loss": 0.7223, + "step": 4950 + }, + { + "epoch": 0.33233784101204655, + "grad_norm": 1.5895766019821167, + "learning_rate": 9.521549790267013e-05, + "loss": 0.6667, + "step": 4952 + }, + { + "epoch": 0.3324720646958156, + "grad_norm": 1.0356003046035767, + "learning_rate": 9.521085741052152e-05, + "loss": 0.6584, + "step": 4954 + }, + { + "epoch": 0.33260628837958456, + "grad_norm": 1.148694634437561, + "learning_rate": 9.52062147822488e-05, + "loss": 0.6871, + "step": 4956 + }, + { + "epoch": 0.3327405120633536, + "grad_norm": 1.2603338956832886, + "learning_rate": 9.520157001807133e-05, + "loss": 0.7459, + "step": 4958 + }, + { + "epoch": 0.33287473574712256, + "grad_norm": 1.1903748512268066, + "learning_rate": 9.519692311820856e-05, + "loss": 0.6372, + "step": 4960 + }, + { + "epoch": 0.3330089594308916, + "grad_norm": 0.9854759573936462, + "learning_rate": 9.519227408288006e-05, + "loss": 0.6916, + "step": 4962 + }, + { + "epoch": 0.33314318311466057, + "grad_norm": 1.0852856636047363, + "learning_rate": 9.518762291230546e-05, + "loss": 0.7023, + "step": 4964 + }, + { + "epoch": 0.3332774067984296, + "grad_norm": 0.9276344776153564, + "learning_rate": 9.518296960670455e-05, + "loss": 0.6828, + "step": 4966 + }, + { + "epoch": 0.3334116304821986, + "grad_norm": 1.0925185680389404, + "learning_rate": 9.517831416629716e-05, + "loss": 0.6997, + "step": 4968 + }, + { + "epoch": 0.3335458541659676, + "grad_norm": 1.4964567422866821, + "learning_rate": 9.517365659130326e-05, + "loss": 0.671, + "step": 4970 + }, + { + "epoch": 0.3336800778497366, + "grad_norm": 1.0770866870880127, + "learning_rate": 9.516899688194294e-05, + "loss": 0.6594, + "step": 4972 + }, + { + "epoch": 0.3338143015335056, + "grad_norm": 1.1555149555206299, + "learning_rate": 9.516433503843631e-05, + "loss": 0.7312, + "step": 4974 + }, + { + "epoch": 0.3339485252172746, + "grad_norm": 1.0918995141983032, + "learning_rate": 9.515967106100368e-05, + "loss": 0.7366, + "step": 4976 + }, + { + "epoch": 0.3340827489010436, + "grad_norm": 1.065820574760437, + "learning_rate": 9.515500494986541e-05, + "loss": 0.7003, + "step": 4978 + }, + { + "epoch": 0.3342169725848126, + "grad_norm": 1.0946167707443237, + "learning_rate": 9.515033670524192e-05, + "loss": 0.7005, + "step": 4980 + }, + { + "epoch": 0.3343511962685816, + "grad_norm": 1.5920342206954956, + "learning_rate": 9.514566632735382e-05, + "loss": 0.6515, + "step": 4982 + }, + { + "epoch": 0.3344854199523506, + "grad_norm": 1.3612632751464844, + "learning_rate": 9.514099381642175e-05, + "loss": 0.7155, + "step": 4984 + }, + { + "epoch": 0.33461964363611957, + "grad_norm": 1.3884814977645874, + "learning_rate": 9.51363191726665e-05, + "loss": 0.7438, + "step": 4986 + }, + { + "epoch": 0.3347538673198886, + "grad_norm": 0.9352781772613525, + "learning_rate": 9.513164239630891e-05, + "loss": 0.6303, + "step": 4988 + }, + { + "epoch": 0.3348880910036576, + "grad_norm": 1.0428996086120605, + "learning_rate": 9.512696348756997e-05, + "loss": 0.6896, + "step": 4990 + }, + { + "epoch": 0.3350223146874266, + "grad_norm": 1.1056615114212036, + "learning_rate": 9.512228244667076e-05, + "loss": 0.668, + "step": 4992 + }, + { + "epoch": 0.3351565383711956, + "grad_norm": 1.1554951667785645, + "learning_rate": 9.511759927383243e-05, + "loss": 0.6924, + "step": 4994 + }, + { + "epoch": 0.3352907620549646, + "grad_norm": 1.0461509227752686, + "learning_rate": 9.511291396927625e-05, + "loss": 0.6861, + "step": 4996 + }, + { + "epoch": 0.3354249857387336, + "grad_norm": 1.0818904638290405, + "learning_rate": 9.510822653322359e-05, + "loss": 0.6708, + "step": 4998 + }, + { + "epoch": 0.3355592094225026, + "grad_norm": 1.0191214084625244, + "learning_rate": 9.510353696589593e-05, + "loss": 0.6678, + "step": 5000 + }, + { + "epoch": 0.3356934331062716, + "grad_norm": 1.0327562093734741, + "learning_rate": 9.509884526751485e-05, + "loss": 0.6998, + "step": 5002 + }, + { + "epoch": 0.3358276567900406, + "grad_norm": 1.0621801614761353, + "learning_rate": 9.5094151438302e-05, + "loss": 0.7418, + "step": 5004 + }, + { + "epoch": 0.3359618804738096, + "grad_norm": 1.079667329788208, + "learning_rate": 9.508945547847916e-05, + "loss": 0.7226, + "step": 5006 + }, + { + "epoch": 0.3360961041575786, + "grad_norm": 1.0097723007202148, + "learning_rate": 9.508475738826823e-05, + "loss": 0.748, + "step": 5008 + }, + { + "epoch": 0.3362303278413476, + "grad_norm": 1.095018744468689, + "learning_rate": 9.508005716789117e-05, + "loss": 0.7191, + "step": 5010 + }, + { + "epoch": 0.33636455152511663, + "grad_norm": 1.254999041557312, + "learning_rate": 9.507535481757005e-05, + "loss": 0.7584, + "step": 5012 + }, + { + "epoch": 0.3364987752088856, + "grad_norm": 1.6734964847564697, + "learning_rate": 9.507065033752704e-05, + "loss": 0.7268, + "step": 5014 + }, + { + "epoch": 0.33663299889265463, + "grad_norm": 1.0575109720230103, + "learning_rate": 9.506594372798446e-05, + "loss": 0.7239, + "step": 5016 + }, + { + "epoch": 0.3367672225764236, + "grad_norm": 1.4216424226760864, + "learning_rate": 9.506123498916463e-05, + "loss": 0.6463, + "step": 5018 + }, + { + "epoch": 0.33690144626019264, + "grad_norm": 0.9909161329269409, + "learning_rate": 9.505652412129008e-05, + "loss": 0.6506, + "step": 5020 + }, + { + "epoch": 0.3370356699439616, + "grad_norm": 1.262615442276001, + "learning_rate": 9.505181112458335e-05, + "loss": 0.7405, + "step": 5022 + }, + { + "epoch": 0.3371698936277306, + "grad_norm": 1.5525568723678589, + "learning_rate": 9.504709599926712e-05, + "loss": 0.6885, + "step": 5024 + }, + { + "epoch": 0.3373041173114996, + "grad_norm": 1.0351464748382568, + "learning_rate": 9.504237874556421e-05, + "loss": 0.6916, + "step": 5026 + }, + { + "epoch": 0.3374383409952686, + "grad_norm": 1.0131417512893677, + "learning_rate": 9.503765936369746e-05, + "loss": 0.6773, + "step": 5028 + }, + { + "epoch": 0.3375725646790376, + "grad_norm": 1.277593731880188, + "learning_rate": 9.503293785388987e-05, + "loss": 0.7126, + "step": 5030 + }, + { + "epoch": 0.3377067883628066, + "grad_norm": 1.1509110927581787, + "learning_rate": 9.502821421636454e-05, + "loss": 0.7054, + "step": 5032 + }, + { + "epoch": 0.33784101204657563, + "grad_norm": 1.7441537380218506, + "learning_rate": 9.502348845134461e-05, + "loss": 0.68, + "step": 5034 + }, + { + "epoch": 0.3379752357303446, + "grad_norm": 1.1158937215805054, + "learning_rate": 9.501876055905339e-05, + "loss": 0.7326, + "step": 5036 + }, + { + "epoch": 0.33810945941411363, + "grad_norm": 1.407645583152771, + "learning_rate": 9.501403053971427e-05, + "loss": 0.6986, + "step": 5038 + }, + { + "epoch": 0.3382436830978826, + "grad_norm": 1.029934048652649, + "learning_rate": 9.500929839355071e-05, + "loss": 0.6293, + "step": 5040 + }, + { + "epoch": 0.33837790678165164, + "grad_norm": 1.1010067462921143, + "learning_rate": 9.500456412078631e-05, + "loss": 0.7056, + "step": 5042 + }, + { + "epoch": 0.3385121304654206, + "grad_norm": 1.2496752738952637, + "learning_rate": 9.499982772164475e-05, + "loss": 0.7108, + "step": 5044 + }, + { + "epoch": 0.33864635414918964, + "grad_norm": 1.099348783493042, + "learning_rate": 9.499508919634983e-05, + "loss": 0.6907, + "step": 5046 + }, + { + "epoch": 0.3387805778329586, + "grad_norm": 1.123132348060608, + "learning_rate": 9.499034854512542e-05, + "loss": 0.713, + "step": 5048 + }, + { + "epoch": 0.33891480151672765, + "grad_norm": 1.031817078590393, + "learning_rate": 9.49856057681955e-05, + "loss": 0.7342, + "step": 5050 + }, + { + "epoch": 0.3390490252004966, + "grad_norm": 1.2734098434448242, + "learning_rate": 9.498086086578418e-05, + "loss": 0.7487, + "step": 5052 + }, + { + "epoch": 0.33918324888426565, + "grad_norm": 1.0864171981811523, + "learning_rate": 9.497611383811564e-05, + "loss": 0.6994, + "step": 5054 + }, + { + "epoch": 0.33931747256803463, + "grad_norm": 1.0995585918426514, + "learning_rate": 9.497136468541415e-05, + "loss": 0.6424, + "step": 5056 + }, + { + "epoch": 0.33945169625180366, + "grad_norm": 1.089625358581543, + "learning_rate": 9.496661340790411e-05, + "loss": 0.7, + "step": 5058 + }, + { + "epoch": 0.33958591993557263, + "grad_norm": 1.0574417114257812, + "learning_rate": 9.496186000581002e-05, + "loss": 0.7276, + "step": 5060 + }, + { + "epoch": 0.3397201436193416, + "grad_norm": 1.0373708009719849, + "learning_rate": 9.495710447935646e-05, + "loss": 0.681, + "step": 5062 + }, + { + "epoch": 0.33985436730311064, + "grad_norm": 1.1043150424957275, + "learning_rate": 9.49523468287681e-05, + "loss": 0.7475, + "step": 5064 + }, + { + "epoch": 0.3399885909868796, + "grad_norm": 1.123861312866211, + "learning_rate": 9.494758705426978e-05, + "loss": 0.7078, + "step": 5066 + }, + { + "epoch": 0.34012281467064864, + "grad_norm": 1.1560646295547485, + "learning_rate": 9.494282515608632e-05, + "loss": 0.6853, + "step": 5068 + }, + { + "epoch": 0.3402570383544176, + "grad_norm": 1.0215603113174438, + "learning_rate": 9.493806113444277e-05, + "loss": 0.732, + "step": 5070 + }, + { + "epoch": 0.34039126203818665, + "grad_norm": 0.9919569492340088, + "learning_rate": 9.493329498956421e-05, + "loss": 0.6517, + "step": 5072 + }, + { + "epoch": 0.3405254857219556, + "grad_norm": 1.1932328939437866, + "learning_rate": 9.492852672167578e-05, + "loss": 0.6514, + "step": 5074 + }, + { + "epoch": 0.34065970940572465, + "grad_norm": 1.3314809799194336, + "learning_rate": 9.492375633100283e-05, + "loss": 0.6408, + "step": 5076 + }, + { + "epoch": 0.34079393308949363, + "grad_norm": 1.1061758995056152, + "learning_rate": 9.491898381777074e-05, + "loss": 0.6561, + "step": 5078 + }, + { + "epoch": 0.34092815677326266, + "grad_norm": 1.0593011379241943, + "learning_rate": 9.4914209182205e-05, + "loss": 0.7318, + "step": 5080 + }, + { + "epoch": 0.34106238045703163, + "grad_norm": 1.0847229957580566, + "learning_rate": 9.490943242453118e-05, + "loss": 0.693, + "step": 5082 + }, + { + "epoch": 0.34119660414080066, + "grad_norm": 1.0208863019943237, + "learning_rate": 9.4904653544975e-05, + "loss": 0.6298, + "step": 5084 + }, + { + "epoch": 0.34133082782456964, + "grad_norm": 1.1964168548583984, + "learning_rate": 9.489987254376222e-05, + "loss": 0.7066, + "step": 5086 + }, + { + "epoch": 0.34146505150833867, + "grad_norm": 1.3700635433197021, + "learning_rate": 9.489508942111878e-05, + "loss": 0.7003, + "step": 5088 + }, + { + "epoch": 0.34159927519210764, + "grad_norm": 1.1736687421798706, + "learning_rate": 9.489030417727063e-05, + "loss": 0.7113, + "step": 5090 + }, + { + "epoch": 0.3417334988758767, + "grad_norm": 1.2480403184890747, + "learning_rate": 9.488551681244388e-05, + "loss": 0.7044, + "step": 5092 + }, + { + "epoch": 0.34186772255964565, + "grad_norm": 0.9867148995399475, + "learning_rate": 9.488072732686474e-05, + "loss": 0.6922, + "step": 5094 + }, + { + "epoch": 0.3420019462434146, + "grad_norm": 0.9716460704803467, + "learning_rate": 9.487593572075948e-05, + "loss": 0.6429, + "step": 5096 + }, + { + "epoch": 0.34213616992718365, + "grad_norm": 1.0517593622207642, + "learning_rate": 9.48711419943545e-05, + "loss": 0.7262, + "step": 5098 + }, + { + "epoch": 0.34227039361095263, + "grad_norm": 1.0895615816116333, + "learning_rate": 9.486634614787631e-05, + "loss": 0.6671, + "step": 5100 + }, + { + "epoch": 0.34240461729472166, + "grad_norm": 1.1830127239227295, + "learning_rate": 9.486154818155146e-05, + "loss": 0.7024, + "step": 5102 + }, + { + "epoch": 0.34253884097849063, + "grad_norm": 0.9760610461235046, + "learning_rate": 9.485674809560669e-05, + "loss": 0.6577, + "step": 5104 + }, + { + "epoch": 0.34267306466225966, + "grad_norm": 1.0928126573562622, + "learning_rate": 9.485194589026878e-05, + "loss": 0.6874, + "step": 5106 + }, + { + "epoch": 0.34280728834602864, + "grad_norm": 1.0870157480239868, + "learning_rate": 9.484714156576464e-05, + "loss": 0.7024, + "step": 5108 + }, + { + "epoch": 0.34294151202979767, + "grad_norm": 1.1574848890304565, + "learning_rate": 9.484233512232123e-05, + "loss": 0.727, + "step": 5110 + }, + { + "epoch": 0.34307573571356664, + "grad_norm": 1.5628575086593628, + "learning_rate": 9.483752656016567e-05, + "loss": 0.7628, + "step": 5112 + }, + { + "epoch": 0.3432099593973357, + "grad_norm": 1.4135750532150269, + "learning_rate": 9.483271587952515e-05, + "loss": 0.6959, + "step": 5114 + }, + { + "epoch": 0.34334418308110465, + "grad_norm": 1.5119282007217407, + "learning_rate": 9.482790308062697e-05, + "loss": 0.7533, + "step": 5116 + }, + { + "epoch": 0.3434784067648737, + "grad_norm": 1.1340827941894531, + "learning_rate": 9.48230881636985e-05, + "loss": 0.7502, + "step": 5118 + }, + { + "epoch": 0.34361263044864265, + "grad_norm": 1.1876959800720215, + "learning_rate": 9.481827112896727e-05, + "loss": 0.7444, + "step": 5120 + }, + { + "epoch": 0.3437468541324117, + "grad_norm": 1.0453839302062988, + "learning_rate": 9.481345197666087e-05, + "loss": 0.6689, + "step": 5122 + }, + { + "epoch": 0.34388107781618066, + "grad_norm": 1.0681465864181519, + "learning_rate": 9.480863070700696e-05, + "loss": 0.6624, + "step": 5124 + }, + { + "epoch": 0.3440153014999497, + "grad_norm": 1.1562570333480835, + "learning_rate": 9.480380732023338e-05, + "loss": 0.7213, + "step": 5126 + }, + { + "epoch": 0.34414952518371866, + "grad_norm": 1.0983346700668335, + "learning_rate": 9.479898181656801e-05, + "loss": 0.6863, + "step": 5128 + }, + { + "epoch": 0.3442837488674877, + "grad_norm": 1.2765334844589233, + "learning_rate": 9.479415419623883e-05, + "loss": 0.7889, + "step": 5130 + }, + { + "epoch": 0.34441797255125667, + "grad_norm": 1.126746654510498, + "learning_rate": 9.478932445947395e-05, + "loss": 0.6746, + "step": 5132 + }, + { + "epoch": 0.34455219623502564, + "grad_norm": 1.1382261514663696, + "learning_rate": 9.478449260650158e-05, + "loss": 0.7237, + "step": 5134 + }, + { + "epoch": 0.3446864199187947, + "grad_norm": 1.141695261001587, + "learning_rate": 9.477965863754998e-05, + "loss": 0.6959, + "step": 5136 + }, + { + "epoch": 0.34482064360256365, + "grad_norm": 1.0475116968154907, + "learning_rate": 9.477482255284757e-05, + "loss": 0.7596, + "step": 5138 + }, + { + "epoch": 0.3449548672863327, + "grad_norm": 1.4221386909484863, + "learning_rate": 9.476998435262284e-05, + "loss": 0.6576, + "step": 5140 + }, + { + "epoch": 0.34508909097010165, + "grad_norm": 0.9887280464172363, + "learning_rate": 9.476514403710439e-05, + "loss": 0.7574, + "step": 5142 + }, + { + "epoch": 0.3452233146538707, + "grad_norm": 0.9547103047370911, + "learning_rate": 9.476030160652091e-05, + "loss": 0.62, + "step": 5144 + }, + { + "epoch": 0.34535753833763966, + "grad_norm": 1.191341757774353, + "learning_rate": 9.475545706110119e-05, + "loss": 0.7415, + "step": 5146 + }, + { + "epoch": 0.3454917620214087, + "grad_norm": 1.0935919284820557, + "learning_rate": 9.475061040107414e-05, + "loss": 0.7202, + "step": 5148 + }, + { + "epoch": 0.34562598570517766, + "grad_norm": 1.2453701496124268, + "learning_rate": 9.474576162666874e-05, + "loss": 0.7048, + "step": 5150 + }, + { + "epoch": 0.3457602093889467, + "grad_norm": 1.120084524154663, + "learning_rate": 9.474091073811409e-05, + "loss": 0.6792, + "step": 5152 + }, + { + "epoch": 0.34589443307271567, + "grad_norm": 1.1592801809310913, + "learning_rate": 9.47360577356394e-05, + "loss": 0.7244, + "step": 5154 + }, + { + "epoch": 0.3460286567564847, + "grad_norm": 1.1230835914611816, + "learning_rate": 9.473120261947395e-05, + "loss": 0.7103, + "step": 5156 + }, + { + "epoch": 0.3461628804402537, + "grad_norm": 0.9806675910949707, + "learning_rate": 9.472634538984712e-05, + "loss": 0.6911, + "step": 5158 + }, + { + "epoch": 0.3462971041240227, + "grad_norm": 1.457768201828003, + "learning_rate": 9.472148604698843e-05, + "loss": 0.7743, + "step": 5160 + }, + { + "epoch": 0.3464313278077917, + "grad_norm": 1.0284276008605957, + "learning_rate": 9.471662459112747e-05, + "loss": 0.7379, + "step": 5162 + }, + { + "epoch": 0.3465655514915607, + "grad_norm": 1.2664660215377808, + "learning_rate": 9.471176102249393e-05, + "loss": 0.587, + "step": 5164 + }, + { + "epoch": 0.3466997751753297, + "grad_norm": 1.3976407051086426, + "learning_rate": 9.47068953413176e-05, + "loss": 0.7207, + "step": 5166 + }, + { + "epoch": 0.3468339988590987, + "grad_norm": 1.140342354774475, + "learning_rate": 9.470202754782837e-05, + "loss": 0.6351, + "step": 5168 + }, + { + "epoch": 0.3469682225428677, + "grad_norm": 1.0120048522949219, + "learning_rate": 9.469715764225626e-05, + "loss": 0.6729, + "step": 5170 + }, + { + "epoch": 0.34710244622663666, + "grad_norm": 1.1098517179489136, + "learning_rate": 9.469228562483132e-05, + "loss": 0.6339, + "step": 5172 + }, + { + "epoch": 0.3472366699104057, + "grad_norm": 0.980413556098938, + "learning_rate": 9.468741149578379e-05, + "loss": 0.6917, + "step": 5174 + }, + { + "epoch": 0.34737089359417467, + "grad_norm": 1.0559087991714478, + "learning_rate": 9.468253525534393e-05, + "loss": 0.7193, + "step": 5176 + }, + { + "epoch": 0.3475051172779437, + "grad_norm": 1.0527195930480957, + "learning_rate": 9.467765690374214e-05, + "loss": 0.6606, + "step": 5178 + }, + { + "epoch": 0.3476393409617127, + "grad_norm": 1.15939462184906, + "learning_rate": 9.467277644120893e-05, + "loss": 0.7285, + "step": 5180 + }, + { + "epoch": 0.3477735646454817, + "grad_norm": 1.1253271102905273, + "learning_rate": 9.466789386797486e-05, + "loss": 0.7025, + "step": 5182 + }, + { + "epoch": 0.3479077883292507, + "grad_norm": 1.0979676246643066, + "learning_rate": 9.466300918427065e-05, + "loss": 0.7205, + "step": 5184 + }, + { + "epoch": 0.3480420120130197, + "grad_norm": 0.9581397771835327, + "learning_rate": 9.465812239032708e-05, + "loss": 0.6445, + "step": 5186 + }, + { + "epoch": 0.3481762356967887, + "grad_norm": 1.073517918586731, + "learning_rate": 9.465323348637505e-05, + "loss": 0.6466, + "step": 5188 + }, + { + "epoch": 0.3483104593805577, + "grad_norm": 1.308125376701355, + "learning_rate": 9.464834247264553e-05, + "loss": 0.6924, + "step": 5190 + }, + { + "epoch": 0.3484446830643267, + "grad_norm": 1.2113579511642456, + "learning_rate": 9.464344934936964e-05, + "loss": 0.6725, + "step": 5192 + }, + { + "epoch": 0.3485789067480957, + "grad_norm": 1.140137791633606, + "learning_rate": 9.463855411677856e-05, + "loss": 0.6954, + "step": 5194 + }, + { + "epoch": 0.3487131304318647, + "grad_norm": 1.337956190109253, + "learning_rate": 9.463365677510357e-05, + "loss": 0.7545, + "step": 5196 + }, + { + "epoch": 0.3488473541156337, + "grad_norm": 1.0394006967544556, + "learning_rate": 9.462875732457606e-05, + "loss": 0.6959, + "step": 5198 + }, + { + "epoch": 0.3489815777994027, + "grad_norm": 1.1676454544067383, + "learning_rate": 9.462385576542752e-05, + "loss": 0.6698, + "step": 5200 + }, + { + "epoch": 0.34911580148317173, + "grad_norm": 1.0103447437286377, + "learning_rate": 9.461895209788956e-05, + "loss": 0.6978, + "step": 5202 + }, + { + "epoch": 0.3492500251669407, + "grad_norm": 1.0410352945327759, + "learning_rate": 9.461404632219384e-05, + "loss": 0.6244, + "step": 5204 + }, + { + "epoch": 0.34938424885070973, + "grad_norm": 1.0641223192214966, + "learning_rate": 9.460913843857217e-05, + "loss": 0.7342, + "step": 5206 + }, + { + "epoch": 0.3495184725344787, + "grad_norm": 1.1432600021362305, + "learning_rate": 9.460422844725642e-05, + "loss": 0.6896, + "step": 5208 + }, + { + "epoch": 0.3496526962182477, + "grad_norm": 1.2629886865615845, + "learning_rate": 9.459931634847859e-05, + "loss": 0.7591, + "step": 5210 + }, + { + "epoch": 0.3497869199020167, + "grad_norm": 1.0171542167663574, + "learning_rate": 9.459440214247077e-05, + "loss": 0.6011, + "step": 5212 + }, + { + "epoch": 0.3499211435857857, + "grad_norm": 1.1639806032180786, + "learning_rate": 9.458948582946514e-05, + "loss": 0.6729, + "step": 5214 + }, + { + "epoch": 0.3500553672695547, + "grad_norm": 1.0054445266723633, + "learning_rate": 9.458456740969397e-05, + "loss": 0.644, + "step": 5216 + }, + { + "epoch": 0.3501895909533237, + "grad_norm": 0.9366871118545532, + "learning_rate": 9.457964688338967e-05, + "loss": 0.6568, + "step": 5218 + }, + { + "epoch": 0.3503238146370927, + "grad_norm": 1.5994693040847778, + "learning_rate": 9.457472425078473e-05, + "loss": 0.7491, + "step": 5220 + }, + { + "epoch": 0.3504580383208617, + "grad_norm": 1.1239392757415771, + "learning_rate": 9.456979951211172e-05, + "loss": 0.5976, + "step": 5222 + }, + { + "epoch": 0.35059226200463073, + "grad_norm": 1.031184196472168, + "learning_rate": 9.45648726676033e-05, + "loss": 0.6977, + "step": 5224 + }, + { + "epoch": 0.3507264856883997, + "grad_norm": 1.017400860786438, + "learning_rate": 9.455994371749231e-05, + "loss": 0.6405, + "step": 5226 + }, + { + "epoch": 0.35086070937216873, + "grad_norm": 1.146761417388916, + "learning_rate": 9.45550126620116e-05, + "loss": 0.7208, + "step": 5228 + }, + { + "epoch": 0.3509949330559377, + "grad_norm": 1.1883848905563354, + "learning_rate": 9.455007950139412e-05, + "loss": 0.6999, + "step": 5230 + }, + { + "epoch": 0.35112915673970674, + "grad_norm": 1.1775646209716797, + "learning_rate": 9.454514423587301e-05, + "loss": 0.6619, + "step": 5232 + }, + { + "epoch": 0.3512633804234757, + "grad_norm": 1.1438260078430176, + "learning_rate": 9.454020686568143e-05, + "loss": 0.727, + "step": 5234 + }, + { + "epoch": 0.35139760410724474, + "grad_norm": 1.1475664377212524, + "learning_rate": 9.453526739105267e-05, + "loss": 0.6989, + "step": 5236 + }, + { + "epoch": 0.3515318277910137, + "grad_norm": 1.0647696256637573, + "learning_rate": 9.45303258122201e-05, + "loss": 0.6298, + "step": 5238 + }, + { + "epoch": 0.35166605147478275, + "grad_norm": 2.985783338546753, + "learning_rate": 9.452538212941719e-05, + "loss": 0.6947, + "step": 5240 + }, + { + "epoch": 0.3518002751585517, + "grad_norm": 0.9413696527481079, + "learning_rate": 9.452043634287753e-05, + "loss": 0.714, + "step": 5242 + }, + { + "epoch": 0.35193449884232075, + "grad_norm": 1.4837936162948608, + "learning_rate": 9.45154884528348e-05, + "loss": 0.7204, + "step": 5244 + }, + { + "epoch": 0.35206872252608973, + "grad_norm": 1.350868821144104, + "learning_rate": 9.451053845952278e-05, + "loss": 0.6835, + "step": 5246 + }, + { + "epoch": 0.3522029462098587, + "grad_norm": 1.0339003801345825, + "learning_rate": 9.450558636317533e-05, + "loss": 0.6535, + "step": 5248 + }, + { + "epoch": 0.35233716989362773, + "grad_norm": 1.4247851371765137, + "learning_rate": 9.450063216402644e-05, + "loss": 0.7949, + "step": 5250 + }, + { + "epoch": 0.3524713935773967, + "grad_norm": 1.0391966104507446, + "learning_rate": 9.44956758623102e-05, + "loss": 0.6573, + "step": 5252 + }, + { + "epoch": 0.35260561726116574, + "grad_norm": 1.1045339107513428, + "learning_rate": 9.44907174582608e-05, + "loss": 0.7038, + "step": 5254 + }, + { + "epoch": 0.3527398409449347, + "grad_norm": 1.0392239093780518, + "learning_rate": 9.448575695211244e-05, + "loss": 0.7089, + "step": 5256 + }, + { + "epoch": 0.35287406462870374, + "grad_norm": 1.0227470397949219, + "learning_rate": 9.448079434409956e-05, + "loss": 0.6423, + "step": 5258 + }, + { + "epoch": 0.3530082883124727, + "grad_norm": 1.0412817001342773, + "learning_rate": 9.447582963445663e-05, + "loss": 0.6564, + "step": 5260 + }, + { + "epoch": 0.35314251199624175, + "grad_norm": 1.1213363409042358, + "learning_rate": 9.447086282341818e-05, + "loss": 0.6377, + "step": 5262 + }, + { + "epoch": 0.3532767356800107, + "grad_norm": 1.0754303932189941, + "learning_rate": 9.446589391121893e-05, + "loss": 0.7076, + "step": 5264 + }, + { + "epoch": 0.35341095936377975, + "grad_norm": 1.133129596710205, + "learning_rate": 9.446092289809361e-05, + "loss": 0.6513, + "step": 5266 + }, + { + "epoch": 0.35354518304754873, + "grad_norm": 1.0103533267974854, + "learning_rate": 9.445594978427714e-05, + "loss": 0.652, + "step": 5268 + }, + { + "epoch": 0.35367940673131776, + "grad_norm": 1.0910634994506836, + "learning_rate": 9.445097457000444e-05, + "loss": 0.7455, + "step": 5270 + }, + { + "epoch": 0.35381363041508673, + "grad_norm": 1.0608628988265991, + "learning_rate": 9.444599725551061e-05, + "loss": 0.7294, + "step": 5272 + }, + { + "epoch": 0.35394785409885576, + "grad_norm": 1.2099720239639282, + "learning_rate": 9.444101784103082e-05, + "loss": 0.6946, + "step": 5274 + }, + { + "epoch": 0.35408207778262474, + "grad_norm": 1.0302938222885132, + "learning_rate": 9.443603632680031e-05, + "loss": 0.6494, + "step": 5276 + }, + { + "epoch": 0.35421630146639377, + "grad_norm": 1.0468542575836182, + "learning_rate": 9.443105271305445e-05, + "loss": 0.6824, + "step": 5278 + }, + { + "epoch": 0.35435052515016274, + "grad_norm": 1.090032696723938, + "learning_rate": 9.442606700002874e-05, + "loss": 0.6674, + "step": 5280 + }, + { + "epoch": 0.3544847488339318, + "grad_norm": 1.011366844177246, + "learning_rate": 9.442107918795873e-05, + "loss": 0.7143, + "step": 5282 + }, + { + "epoch": 0.35461897251770075, + "grad_norm": 1.0054188966751099, + "learning_rate": 9.441608927708006e-05, + "loss": 0.746, + "step": 5284 + }, + { + "epoch": 0.3547531962014697, + "grad_norm": 1.2857420444488525, + "learning_rate": 9.441109726762852e-05, + "loss": 0.7379, + "step": 5286 + }, + { + "epoch": 0.35488741988523875, + "grad_norm": 1.0563364028930664, + "learning_rate": 9.440610315983998e-05, + "loss": 0.6357, + "step": 5288 + }, + { + "epoch": 0.3550216435690077, + "grad_norm": 1.1448321342468262, + "learning_rate": 9.440110695395037e-05, + "loss": 0.6954, + "step": 5290 + }, + { + "epoch": 0.35515586725277676, + "grad_norm": 1.1138883829116821, + "learning_rate": 9.439610865019577e-05, + "loss": 0.6362, + "step": 5292 + }, + { + "epoch": 0.35529009093654573, + "grad_norm": 1.0992587804794312, + "learning_rate": 9.439110824881232e-05, + "loss": 0.7334, + "step": 5294 + }, + { + "epoch": 0.35542431462031476, + "grad_norm": 1.2044397592544556, + "learning_rate": 9.438610575003632e-05, + "loss": 0.708, + "step": 5296 + }, + { + "epoch": 0.35555853830408374, + "grad_norm": 1.0396875143051147, + "learning_rate": 9.43811011541041e-05, + "loss": 0.6138, + "step": 5298 + }, + { + "epoch": 0.35569276198785277, + "grad_norm": 1.1773802042007446, + "learning_rate": 9.437609446125211e-05, + "loss": 0.7468, + "step": 5300 + }, + { + "epoch": 0.35582698567162174, + "grad_norm": 1.0798813104629517, + "learning_rate": 9.437108567171693e-05, + "loss": 0.7499, + "step": 5302 + }, + { + "epoch": 0.3559612093553908, + "grad_norm": 1.066114902496338, + "learning_rate": 9.436607478573522e-05, + "loss": 0.6672, + "step": 5304 + }, + { + "epoch": 0.35609543303915975, + "grad_norm": 1.2825912237167358, + "learning_rate": 9.436106180354369e-05, + "loss": 0.7209, + "step": 5306 + }, + { + "epoch": 0.3562296567229288, + "grad_norm": 1.0812764167785645, + "learning_rate": 9.435604672537924e-05, + "loss": 0.6479, + "step": 5308 + }, + { + "epoch": 0.35636388040669775, + "grad_norm": 1.2417733669281006, + "learning_rate": 9.43510295514788e-05, + "loss": 0.631, + "step": 5310 + }, + { + "epoch": 0.3564981040904668, + "grad_norm": 0.9054554104804993, + "learning_rate": 9.434601028207942e-05, + "loss": 0.5859, + "step": 5312 + }, + { + "epoch": 0.35663232777423576, + "grad_norm": 1.10965895652771, + "learning_rate": 9.434098891741827e-05, + "loss": 0.6654, + "step": 5314 + }, + { + "epoch": 0.3567665514580048, + "grad_norm": 1.175485372543335, + "learning_rate": 9.433596545773258e-05, + "loss": 0.7067, + "step": 5316 + }, + { + "epoch": 0.35690077514177376, + "grad_norm": 1.0619306564331055, + "learning_rate": 9.43309399032597e-05, + "loss": 0.7069, + "step": 5318 + }, + { + "epoch": 0.3570349988255428, + "grad_norm": 1.0990415811538696, + "learning_rate": 9.432591225423708e-05, + "loss": 0.7396, + "step": 5320 + }, + { + "epoch": 0.35716922250931177, + "grad_norm": 1.2514533996582031, + "learning_rate": 9.432088251090228e-05, + "loss": 0.6848, + "step": 5322 + }, + { + "epoch": 0.35730344619308074, + "grad_norm": 1.0155177116394043, + "learning_rate": 9.431585067349293e-05, + "loss": 0.6888, + "step": 5324 + }, + { + "epoch": 0.3574376698768498, + "grad_norm": 1.1702946424484253, + "learning_rate": 9.431081674224677e-05, + "loss": 0.6642, + "step": 5326 + }, + { + "epoch": 0.35757189356061875, + "grad_norm": 1.1770367622375488, + "learning_rate": 9.430578071740167e-05, + "loss": 0.6797, + "step": 5328 + }, + { + "epoch": 0.3577061172443878, + "grad_norm": 1.0894922018051147, + "learning_rate": 9.430074259919554e-05, + "loss": 0.6688, + "step": 5330 + }, + { + "epoch": 0.35784034092815675, + "grad_norm": 1.1621495485305786, + "learning_rate": 9.429570238786645e-05, + "loss": 0.6791, + "step": 5332 + }, + { + "epoch": 0.3579745646119258, + "grad_norm": 1.0528497695922852, + "learning_rate": 9.429066008365251e-05, + "loss": 0.7114, + "step": 5334 + }, + { + "epoch": 0.35810878829569476, + "grad_norm": 1.1805702447891235, + "learning_rate": 9.428561568679199e-05, + "loss": 0.7519, + "step": 5336 + }, + { + "epoch": 0.3582430119794638, + "grad_norm": 1.0912284851074219, + "learning_rate": 9.428056919752319e-05, + "loss": 0.7414, + "step": 5338 + }, + { + "epoch": 0.35837723566323276, + "grad_norm": 1.135192632675171, + "learning_rate": 9.42755206160846e-05, + "loss": 0.6375, + "step": 5340 + }, + { + "epoch": 0.3585114593470018, + "grad_norm": 1.0422782897949219, + "learning_rate": 9.427046994271471e-05, + "loss": 0.7249, + "step": 5342 + }, + { + "epoch": 0.35864568303077077, + "grad_norm": 1.1697744131088257, + "learning_rate": 9.426541717765216e-05, + "loss": 0.6435, + "step": 5344 + }, + { + "epoch": 0.3587799067145398, + "grad_norm": 1.6935707330703735, + "learning_rate": 9.426036232113571e-05, + "loss": 0.7125, + "step": 5346 + }, + { + "epoch": 0.3589141303983088, + "grad_norm": 1.2502233982086182, + "learning_rate": 9.425530537340417e-05, + "loss": 0.7024, + "step": 5348 + }, + { + "epoch": 0.3590483540820778, + "grad_norm": 0.9787608981132507, + "learning_rate": 9.425024633469647e-05, + "loss": 0.6551, + "step": 5350 + }, + { + "epoch": 0.3591825777658468, + "grad_norm": 1.5133332014083862, + "learning_rate": 9.424518520525165e-05, + "loss": 0.7225, + "step": 5352 + }, + { + "epoch": 0.3593168014496158, + "grad_norm": 1.0226023197174072, + "learning_rate": 9.424012198530882e-05, + "loss": 0.6953, + "step": 5354 + }, + { + "epoch": 0.3594510251333848, + "grad_norm": 1.0732944011688232, + "learning_rate": 9.423505667510724e-05, + "loss": 0.6278, + "step": 5356 + }, + { + "epoch": 0.3595852488171538, + "grad_norm": 1.168982744216919, + "learning_rate": 9.42299892748862e-05, + "loss": 0.6749, + "step": 5358 + }, + { + "epoch": 0.3597194725009228, + "grad_norm": 1.084213376045227, + "learning_rate": 9.422491978488515e-05, + "loss": 0.7054, + "step": 5360 + }, + { + "epoch": 0.35985369618469176, + "grad_norm": 0.9596426486968994, + "learning_rate": 9.42198482053436e-05, + "loss": 0.6319, + "step": 5362 + }, + { + "epoch": 0.3599879198684608, + "grad_norm": 1.0419789552688599, + "learning_rate": 9.421477453650118e-05, + "loss": 0.5885, + "step": 5364 + }, + { + "epoch": 0.36012214355222977, + "grad_norm": 1.0027025938034058, + "learning_rate": 9.420969877859761e-05, + "loss": 0.661, + "step": 5366 + }, + { + "epoch": 0.3602563672359988, + "grad_norm": 0.8679866194725037, + "learning_rate": 9.420462093187271e-05, + "loss": 0.6068, + "step": 5368 + }, + { + "epoch": 0.3603905909197678, + "grad_norm": 2.679222583770752, + "learning_rate": 9.419954099656638e-05, + "loss": 0.7255, + "step": 5370 + }, + { + "epoch": 0.3605248146035368, + "grad_norm": 1.037379264831543, + "learning_rate": 9.419445897291867e-05, + "loss": 0.7083, + "step": 5372 + }, + { + "epoch": 0.3606590382873058, + "grad_norm": 1.331444501876831, + "learning_rate": 9.418937486116968e-05, + "loss": 0.673, + "step": 5374 + }, + { + "epoch": 0.3607932619710748, + "grad_norm": 1.0762038230895996, + "learning_rate": 9.418428866155961e-05, + "loss": 0.7332, + "step": 5376 + }, + { + "epoch": 0.3609274856548438, + "grad_norm": 1.0162066221237183, + "learning_rate": 9.417920037432879e-05, + "loss": 0.5862, + "step": 5378 + }, + { + "epoch": 0.3610617093386128, + "grad_norm": 1.0344343185424805, + "learning_rate": 9.417410999971762e-05, + "loss": 0.6889, + "step": 5380 + }, + { + "epoch": 0.3611959330223818, + "grad_norm": 1.0645183324813843, + "learning_rate": 9.416901753796663e-05, + "loss": 0.6554, + "step": 5382 + }, + { + "epoch": 0.3613301567061508, + "grad_norm": 1.1452938318252563, + "learning_rate": 9.41639229893164e-05, + "loss": 0.7387, + "step": 5384 + }, + { + "epoch": 0.3614643803899198, + "grad_norm": 1.0661026239395142, + "learning_rate": 9.415882635400768e-05, + "loss": 0.7022, + "step": 5386 + }, + { + "epoch": 0.3615986040736888, + "grad_norm": 1.194131851196289, + "learning_rate": 9.415372763228123e-05, + "loss": 0.7464, + "step": 5388 + }, + { + "epoch": 0.3617328277574578, + "grad_norm": 1.0507525205612183, + "learning_rate": 9.414862682437797e-05, + "loss": 0.6617, + "step": 5390 + }, + { + "epoch": 0.36186705144122683, + "grad_norm": 1.1302318572998047, + "learning_rate": 9.414352393053891e-05, + "loss": 0.6155, + "step": 5392 + }, + { + "epoch": 0.3620012751249958, + "grad_norm": 1.166967511177063, + "learning_rate": 9.413841895100515e-05, + "loss": 0.6621, + "step": 5394 + }, + { + "epoch": 0.36213549880876483, + "grad_norm": 1.0846643447875977, + "learning_rate": 9.413331188601791e-05, + "loss": 0.7087, + "step": 5396 + }, + { + "epoch": 0.3622697224925338, + "grad_norm": 1.0447102785110474, + "learning_rate": 9.412820273581844e-05, + "loss": 0.6652, + "step": 5398 + }, + { + "epoch": 0.3624039461763028, + "grad_norm": 1.3294551372528076, + "learning_rate": 9.412309150064817e-05, + "loss": 0.651, + "step": 5400 + }, + { + "epoch": 0.3625381698600718, + "grad_norm": 1.0539995431900024, + "learning_rate": 9.411797818074861e-05, + "loss": 0.6983, + "step": 5402 + }, + { + "epoch": 0.3626723935438408, + "grad_norm": 1.0341007709503174, + "learning_rate": 9.411286277636131e-05, + "loss": 0.6941, + "step": 5404 + }, + { + "epoch": 0.3628066172276098, + "grad_norm": 1.1077064275741577, + "learning_rate": 9.410774528772802e-05, + "loss": 0.6889, + "step": 5406 + }, + { + "epoch": 0.3629408409113788, + "grad_norm": 1.3066242933273315, + "learning_rate": 9.410262571509046e-05, + "loss": 0.7075, + "step": 5408 + }, + { + "epoch": 0.3630750645951478, + "grad_norm": 1.058754563331604, + "learning_rate": 9.409750405869058e-05, + "loss": 0.6989, + "step": 5410 + }, + { + "epoch": 0.3632092882789168, + "grad_norm": 1.1940826177597046, + "learning_rate": 9.409238031877034e-05, + "loss": 0.7646, + "step": 5412 + }, + { + "epoch": 0.36334351196268583, + "grad_norm": 1.1361353397369385, + "learning_rate": 9.408725449557184e-05, + "loss": 0.7342, + "step": 5414 + }, + { + "epoch": 0.3634777356464548, + "grad_norm": 1.0520503520965576, + "learning_rate": 9.408212658933726e-05, + "loss": 0.6822, + "step": 5416 + }, + { + "epoch": 0.36361195933022383, + "grad_norm": 1.1402872800827026, + "learning_rate": 9.407699660030888e-05, + "loss": 0.7148, + "step": 5418 + }, + { + "epoch": 0.3637461830139928, + "grad_norm": 1.1878187656402588, + "learning_rate": 9.407186452872908e-05, + "loss": 0.7087, + "step": 5420 + }, + { + "epoch": 0.36388040669776184, + "grad_norm": 1.087986707687378, + "learning_rate": 9.406673037484035e-05, + "loss": 0.721, + "step": 5422 + }, + { + "epoch": 0.3640146303815308, + "grad_norm": 0.9601892232894897, + "learning_rate": 9.406159413888527e-05, + "loss": 0.6224, + "step": 5424 + }, + { + "epoch": 0.36414885406529984, + "grad_norm": 1.0169936418533325, + "learning_rate": 9.40564558211065e-05, + "loss": 0.6989, + "step": 5426 + }, + { + "epoch": 0.3642830777490688, + "grad_norm": 1.0120033025741577, + "learning_rate": 9.405131542174684e-05, + "loss": 0.73, + "step": 5428 + }, + { + "epoch": 0.36441730143283785, + "grad_norm": 1.2148478031158447, + "learning_rate": 9.404617294104911e-05, + "loss": 0.8287, + "step": 5430 + }, + { + "epoch": 0.3645515251166068, + "grad_norm": 1.4467118978500366, + "learning_rate": 9.404102837925637e-05, + "loss": 0.7313, + "step": 5432 + }, + { + "epoch": 0.36468574880037585, + "grad_norm": 1.2116284370422363, + "learning_rate": 9.403588173661162e-05, + "loss": 0.6943, + "step": 5434 + }, + { + "epoch": 0.3648199724841448, + "grad_norm": 1.098501443862915, + "learning_rate": 9.403073301335805e-05, + "loss": 0.7122, + "step": 5436 + }, + { + "epoch": 0.3649541961679138, + "grad_norm": 1.039089322090149, + "learning_rate": 9.402558220973892e-05, + "loss": 0.6847, + "step": 5438 + }, + { + "epoch": 0.36508841985168283, + "grad_norm": 1.4157084226608276, + "learning_rate": 9.402042932599762e-05, + "loss": 0.693, + "step": 5440 + }, + { + "epoch": 0.3652226435354518, + "grad_norm": 1.2092018127441406, + "learning_rate": 9.401527436237758e-05, + "loss": 0.7003, + "step": 5442 + }, + { + "epoch": 0.36535686721922084, + "grad_norm": 1.0189353227615356, + "learning_rate": 9.40101173191224e-05, + "loss": 0.609, + "step": 5444 + }, + { + "epoch": 0.3654910909029898, + "grad_norm": 1.053718090057373, + "learning_rate": 9.40049581964757e-05, + "loss": 0.6704, + "step": 5446 + }, + { + "epoch": 0.36562531458675884, + "grad_norm": 1.7575159072875977, + "learning_rate": 9.399979699468126e-05, + "loss": 0.6315, + "step": 5448 + }, + { + "epoch": 0.3657595382705278, + "grad_norm": 1.2005894184112549, + "learning_rate": 9.399463371398295e-05, + "loss": 0.7318, + "step": 5450 + }, + { + "epoch": 0.36589376195429685, + "grad_norm": 0.9507780075073242, + "learning_rate": 9.398946835462469e-05, + "loss": 0.6205, + "step": 5452 + }, + { + "epoch": 0.3660279856380658, + "grad_norm": 0.9422311186790466, + "learning_rate": 9.398430091685056e-05, + "loss": 0.6896, + "step": 5454 + }, + { + "epoch": 0.36616220932183485, + "grad_norm": 1.461273193359375, + "learning_rate": 9.397913140090471e-05, + "loss": 0.6858, + "step": 5456 + }, + { + "epoch": 0.3662964330056038, + "grad_norm": 1.0395759344100952, + "learning_rate": 9.397395980703137e-05, + "loss": 0.7179, + "step": 5458 + }, + { + "epoch": 0.36643065668937286, + "grad_norm": 0.9786043763160706, + "learning_rate": 9.39687861354749e-05, + "loss": 0.6653, + "step": 5460 + }, + { + "epoch": 0.36656488037314183, + "grad_norm": 1.1348570585250854, + "learning_rate": 9.396361038647976e-05, + "loss": 0.7807, + "step": 5462 + }, + { + "epoch": 0.36669910405691086, + "grad_norm": 1.1227689981460571, + "learning_rate": 9.395843256029047e-05, + "loss": 0.715, + "step": 5464 + }, + { + "epoch": 0.36683332774067984, + "grad_norm": 1.1113135814666748, + "learning_rate": 9.39532526571517e-05, + "loss": 0.7552, + "step": 5466 + }, + { + "epoch": 0.36696755142444887, + "grad_norm": 1.094340205192566, + "learning_rate": 9.394807067730814e-05, + "loss": 0.6482, + "step": 5468 + }, + { + "epoch": 0.36710177510821784, + "grad_norm": 1.2362637519836426, + "learning_rate": 9.394288662100467e-05, + "loss": 0.7001, + "step": 5470 + }, + { + "epoch": 0.3672359987919869, + "grad_norm": 0.9833071827888489, + "learning_rate": 9.393770048848622e-05, + "loss": 0.6472, + "step": 5472 + }, + { + "epoch": 0.36737022247575585, + "grad_norm": 1.1718744039535522, + "learning_rate": 9.393251227999784e-05, + "loss": 0.6698, + "step": 5474 + }, + { + "epoch": 0.3675044461595248, + "grad_norm": 1.0212267637252808, + "learning_rate": 9.392732199578462e-05, + "loss": 0.6415, + "step": 5476 + }, + { + "epoch": 0.36763866984329385, + "grad_norm": 1.1257481575012207, + "learning_rate": 9.392212963609183e-05, + "loss": 0.6981, + "step": 5478 + }, + { + "epoch": 0.3677728935270628, + "grad_norm": 1.2375941276550293, + "learning_rate": 9.391693520116477e-05, + "loss": 0.7004, + "step": 5480 + }, + { + "epoch": 0.36790711721083186, + "grad_norm": 1.1024177074432373, + "learning_rate": 9.391173869124889e-05, + "loss": 0.6756, + "step": 5482 + }, + { + "epoch": 0.36804134089460083, + "grad_norm": 1.1458148956298828, + "learning_rate": 9.390654010658971e-05, + "loss": 0.6584, + "step": 5484 + }, + { + "epoch": 0.36817556457836986, + "grad_norm": 1.274086594581604, + "learning_rate": 9.390133944743284e-05, + "loss": 0.6454, + "step": 5486 + }, + { + "epoch": 0.36830978826213884, + "grad_norm": 1.0998162031173706, + "learning_rate": 9.389613671402402e-05, + "loss": 0.6255, + "step": 5488 + }, + { + "epoch": 0.36844401194590787, + "grad_norm": 1.1039185523986816, + "learning_rate": 9.389093190660905e-05, + "loss": 0.7152, + "step": 5490 + }, + { + "epoch": 0.36857823562967684, + "grad_norm": 0.9752369523048401, + "learning_rate": 9.388572502543384e-05, + "loss": 0.6572, + "step": 5492 + }, + { + "epoch": 0.3687124593134459, + "grad_norm": 1.011757254600525, + "learning_rate": 9.388051607074445e-05, + "loss": 0.7229, + "step": 5494 + }, + { + "epoch": 0.36884668299721485, + "grad_norm": 1.072043776512146, + "learning_rate": 9.387530504278695e-05, + "loss": 0.6519, + "step": 5496 + }, + { + "epoch": 0.3689809066809839, + "grad_norm": 1.2205461263656616, + "learning_rate": 9.387009194180755e-05, + "loss": 0.6945, + "step": 5498 + }, + { + "epoch": 0.36911513036475285, + "grad_norm": 0.9959816336631775, + "learning_rate": 9.38648767680526e-05, + "loss": 0.6862, + "step": 5500 + }, + { + "epoch": 0.3692493540485219, + "grad_norm": 1.0489308834075928, + "learning_rate": 9.385965952176847e-05, + "loss": 0.6726, + "step": 5502 + }, + { + "epoch": 0.36938357773229086, + "grad_norm": 1.0031033754348755, + "learning_rate": 9.385444020320166e-05, + "loss": 0.6779, + "step": 5504 + }, + { + "epoch": 0.3695178014160599, + "grad_norm": 1.0508573055267334, + "learning_rate": 9.38492188125988e-05, + "loss": 0.6764, + "step": 5506 + }, + { + "epoch": 0.36965202509982886, + "grad_norm": 1.0428763628005981, + "learning_rate": 9.384399535020657e-05, + "loss": 0.6462, + "step": 5508 + }, + { + "epoch": 0.3697862487835979, + "grad_norm": 1.3005393743515015, + "learning_rate": 9.383876981627178e-05, + "loss": 0.7182, + "step": 5510 + }, + { + "epoch": 0.36992047246736687, + "grad_norm": 1.1237488985061646, + "learning_rate": 9.383354221104132e-05, + "loss": 0.7415, + "step": 5512 + }, + { + "epoch": 0.37005469615113584, + "grad_norm": 1.087746024131775, + "learning_rate": 9.382831253476219e-05, + "loss": 0.6635, + "step": 5514 + }, + { + "epoch": 0.3701889198349049, + "grad_norm": 1.189430832862854, + "learning_rate": 9.382308078768146e-05, + "loss": 0.6528, + "step": 5516 + }, + { + "epoch": 0.37032314351867385, + "grad_norm": 1.0363761186599731, + "learning_rate": 9.381784697004636e-05, + "loss": 0.6923, + "step": 5518 + }, + { + "epoch": 0.3704573672024429, + "grad_norm": 1.0994951725006104, + "learning_rate": 9.381261108210412e-05, + "loss": 0.6299, + "step": 5520 + }, + { + "epoch": 0.37059159088621185, + "grad_norm": 1.5011996030807495, + "learning_rate": 9.380737312410219e-05, + "loss": 0.591, + "step": 5522 + }, + { + "epoch": 0.3707258145699809, + "grad_norm": 1.8518561124801636, + "learning_rate": 9.380213309628803e-05, + "loss": 0.7117, + "step": 5524 + }, + { + "epoch": 0.37086003825374986, + "grad_norm": 1.2186483144760132, + "learning_rate": 9.379689099890921e-05, + "loss": 0.6424, + "step": 5526 + }, + { + "epoch": 0.3709942619375189, + "grad_norm": 1.0989634990692139, + "learning_rate": 9.37916468322134e-05, + "loss": 0.5955, + "step": 5528 + }, + { + "epoch": 0.37112848562128786, + "grad_norm": 1.1953972578048706, + "learning_rate": 9.378640059644839e-05, + "loss": 0.6873, + "step": 5530 + }, + { + "epoch": 0.3712627093050569, + "grad_norm": 1.0643166303634644, + "learning_rate": 9.378115229186207e-05, + "loss": 0.6677, + "step": 5532 + }, + { + "epoch": 0.37139693298882587, + "grad_norm": 1.1248139142990112, + "learning_rate": 9.377590191870236e-05, + "loss": 0.6331, + "step": 5534 + }, + { + "epoch": 0.3715311566725949, + "grad_norm": 1.00499427318573, + "learning_rate": 9.37706494772174e-05, + "loss": 0.6585, + "step": 5536 + }, + { + "epoch": 0.3716653803563639, + "grad_norm": 1.1205849647521973, + "learning_rate": 9.376539496765531e-05, + "loss": 0.6161, + "step": 5538 + }, + { + "epoch": 0.3717996040401329, + "grad_norm": 0.9915878772735596, + "learning_rate": 9.376013839026437e-05, + "loss": 0.6394, + "step": 5540 + }, + { + "epoch": 0.3719338277239019, + "grad_norm": 1.1141587495803833, + "learning_rate": 9.375487974529296e-05, + "loss": 0.6941, + "step": 5542 + }, + { + "epoch": 0.3720680514076709, + "grad_norm": 1.0914654731750488, + "learning_rate": 9.37496190329895e-05, + "loss": 0.6707, + "step": 5544 + }, + { + "epoch": 0.3722022750914399, + "grad_norm": 1.206699252128601, + "learning_rate": 9.374435625360259e-05, + "loss": 0.6699, + "step": 5546 + }, + { + "epoch": 0.3723364987752089, + "grad_norm": 0.985154390335083, + "learning_rate": 9.373909140738084e-05, + "loss": 0.6467, + "step": 5548 + }, + { + "epoch": 0.3724707224589779, + "grad_norm": 1.2127586603164673, + "learning_rate": 9.373382449457304e-05, + "loss": 0.7294, + "step": 5550 + }, + { + "epoch": 0.37260494614274686, + "grad_norm": 1.0873113870620728, + "learning_rate": 9.372855551542805e-05, + "loss": 0.7035, + "step": 5552 + }, + { + "epoch": 0.3727391698265159, + "grad_norm": 1.1156877279281616, + "learning_rate": 9.372328447019478e-05, + "loss": 0.7254, + "step": 5554 + }, + { + "epoch": 0.37287339351028487, + "grad_norm": 1.3185495138168335, + "learning_rate": 9.37180113591223e-05, + "loss": 0.7596, + "step": 5556 + }, + { + "epoch": 0.3730076171940539, + "grad_norm": 2.6542375087738037, + "learning_rate": 9.371273618245976e-05, + "loss": 0.6865, + "step": 5558 + }, + { + "epoch": 0.3731418408778229, + "grad_norm": 1.883700966835022, + "learning_rate": 9.370745894045639e-05, + "loss": 0.6406, + "step": 5560 + }, + { + "epoch": 0.3732760645615919, + "grad_norm": 1.2540524005889893, + "learning_rate": 9.370217963336152e-05, + "loss": 0.6706, + "step": 5562 + }, + { + "epoch": 0.3734102882453609, + "grad_norm": 1.111669898033142, + "learning_rate": 9.36968982614246e-05, + "loss": 0.7454, + "step": 5564 + }, + { + "epoch": 0.3735445119291299, + "grad_norm": 1.7078750133514404, + "learning_rate": 9.369161482489519e-05, + "loss": 0.6821, + "step": 5566 + }, + { + "epoch": 0.3736787356128989, + "grad_norm": 1.2800893783569336, + "learning_rate": 9.368632932402287e-05, + "loss": 0.7193, + "step": 5568 + }, + { + "epoch": 0.3738129592966679, + "grad_norm": 1.2646973133087158, + "learning_rate": 9.368104175905741e-05, + "loss": 0.7339, + "step": 5570 + }, + { + "epoch": 0.3739471829804369, + "grad_norm": 1.9984244108200073, + "learning_rate": 9.367575213024861e-05, + "loss": 0.7086, + "step": 5572 + }, + { + "epoch": 0.3740814066642059, + "grad_norm": 1.0493110418319702, + "learning_rate": 9.36704604378464e-05, + "loss": 0.7568, + "step": 5574 + }, + { + "epoch": 0.3742156303479749, + "grad_norm": 1.1971096992492676, + "learning_rate": 9.366516668210083e-05, + "loss": 0.7471, + "step": 5576 + }, + { + "epoch": 0.3743498540317439, + "grad_norm": 1.066906213760376, + "learning_rate": 9.365987086326198e-05, + "loss": 0.6573, + "step": 5578 + }, + { + "epoch": 0.3744840777155129, + "grad_norm": 1.0150924921035767, + "learning_rate": 9.365457298158009e-05, + "loss": 0.6744, + "step": 5580 + }, + { + "epoch": 0.3746183013992819, + "grad_norm": 1.3731441497802734, + "learning_rate": 9.364927303730549e-05, + "loss": 0.7055, + "step": 5582 + }, + { + "epoch": 0.3747525250830509, + "grad_norm": 2.7943010330200195, + "learning_rate": 9.364397103068854e-05, + "loss": 0.5967, + "step": 5584 + }, + { + "epoch": 0.37488674876681993, + "grad_norm": 1.1104239225387573, + "learning_rate": 9.36386669619798e-05, + "loss": 0.6855, + "step": 5586 + }, + { + "epoch": 0.3750209724505889, + "grad_norm": 1.2128210067749023, + "learning_rate": 9.363336083142986e-05, + "loss": 0.6556, + "step": 5588 + }, + { + "epoch": 0.3751551961343579, + "grad_norm": 1.0144624710083008, + "learning_rate": 9.36280526392894e-05, + "loss": 0.6796, + "step": 5590 + }, + { + "epoch": 0.3752894198181269, + "grad_norm": 1.0004570484161377, + "learning_rate": 9.362274238580926e-05, + "loss": 0.6382, + "step": 5592 + }, + { + "epoch": 0.3754236435018959, + "grad_norm": 1.1422725915908813, + "learning_rate": 9.361743007124032e-05, + "loss": 0.7098, + "step": 5594 + }, + { + "epoch": 0.3755578671856649, + "grad_norm": 1.0474408864974976, + "learning_rate": 9.36121156958336e-05, + "loss": 0.7866, + "step": 5596 + }, + { + "epoch": 0.3756920908694339, + "grad_norm": 1.7418471574783325, + "learning_rate": 9.360679925984013e-05, + "loss": 0.6917, + "step": 5598 + }, + { + "epoch": 0.3758263145532029, + "grad_norm": 1.1903367042541504, + "learning_rate": 9.360148076351117e-05, + "loss": 0.5844, + "step": 5600 + }, + { + "epoch": 0.3759605382369719, + "grad_norm": 1.1024620532989502, + "learning_rate": 9.359616020709798e-05, + "loss": 0.6292, + "step": 5602 + }, + { + "epoch": 0.3760947619207409, + "grad_norm": 1.4756131172180176, + "learning_rate": 9.359083759085195e-05, + "loss": 0.6778, + "step": 5604 + }, + { + "epoch": 0.3762289856045099, + "grad_norm": 1.1311850547790527, + "learning_rate": 9.358551291502456e-05, + "loss": 0.6545, + "step": 5606 + }, + { + "epoch": 0.37636320928827893, + "grad_norm": 1.1814398765563965, + "learning_rate": 9.358018617986739e-05, + "loss": 0.6401, + "step": 5608 + }, + { + "epoch": 0.3764974329720479, + "grad_norm": 1.0626928806304932, + "learning_rate": 9.357485738563212e-05, + "loss": 0.6659, + "step": 5610 + }, + { + "epoch": 0.37663165665581694, + "grad_norm": 1.2566537857055664, + "learning_rate": 9.356952653257051e-05, + "loss": 0.6408, + "step": 5612 + }, + { + "epoch": 0.3767658803395859, + "grad_norm": 1.0836503505706787, + "learning_rate": 9.356419362093449e-05, + "loss": 0.7051, + "step": 5614 + }, + { + "epoch": 0.37690010402335494, + "grad_norm": 1.1769944429397583, + "learning_rate": 9.355885865097595e-05, + "loss": 0.632, + "step": 5616 + }, + { + "epoch": 0.3770343277071239, + "grad_norm": 1.2091494798660278, + "learning_rate": 9.3553521622947e-05, + "loss": 0.741, + "step": 5618 + }, + { + "epoch": 0.37716855139089295, + "grad_norm": 1.059908390045166, + "learning_rate": 9.354818253709981e-05, + "loss": 0.7112, + "step": 5620 + }, + { + "epoch": 0.3773027750746619, + "grad_norm": 1.0945123434066772, + "learning_rate": 9.354284139368662e-05, + "loss": 0.7043, + "step": 5622 + }, + { + "epoch": 0.37743699875843095, + "grad_norm": 1.1156584024429321, + "learning_rate": 9.35374981929598e-05, + "loss": 0.7091, + "step": 5624 + }, + { + "epoch": 0.3775712224421999, + "grad_norm": 0.9157266616821289, + "learning_rate": 9.35321529351718e-05, + "loss": 0.6328, + "step": 5626 + }, + { + "epoch": 0.3777054461259689, + "grad_norm": 1.1768842935562134, + "learning_rate": 9.352680562057516e-05, + "loss": 0.6926, + "step": 5628 + }, + { + "epoch": 0.37783966980973793, + "grad_norm": 1.0927475690841675, + "learning_rate": 9.352145624942256e-05, + "loss": 0.6667, + "step": 5630 + }, + { + "epoch": 0.3779738934935069, + "grad_norm": 1.2773330211639404, + "learning_rate": 9.351610482196676e-05, + "loss": 0.656, + "step": 5632 + }, + { + "epoch": 0.37810811717727594, + "grad_norm": 1.2487109899520874, + "learning_rate": 9.351075133846054e-05, + "loss": 0.7294, + "step": 5634 + }, + { + "epoch": 0.3782423408610449, + "grad_norm": 0.9611471891403198, + "learning_rate": 9.35053957991569e-05, + "loss": 0.5967, + "step": 5636 + }, + { + "epoch": 0.37837656454481394, + "grad_norm": 0.9779294729232788, + "learning_rate": 9.350003820430885e-05, + "loss": 0.6705, + "step": 5638 + }, + { + "epoch": 0.3785107882285829, + "grad_norm": 1.0110894441604614, + "learning_rate": 9.349467855416953e-05, + "loss": 0.6288, + "step": 5640 + }, + { + "epoch": 0.37864501191235195, + "grad_norm": 1.2623742818832397, + "learning_rate": 9.348931684899219e-05, + "loss": 0.6304, + "step": 5642 + }, + { + "epoch": 0.3787792355961209, + "grad_norm": 1.0891246795654297, + "learning_rate": 9.348395308903014e-05, + "loss": 0.7221, + "step": 5644 + }, + { + "epoch": 0.37891345927988995, + "grad_norm": 2.084601402282715, + "learning_rate": 9.347858727453682e-05, + "loss": 0.7258, + "step": 5646 + }, + { + "epoch": 0.3790476829636589, + "grad_norm": 1.037649154663086, + "learning_rate": 9.347321940576575e-05, + "loss": 0.6064, + "step": 5648 + }, + { + "epoch": 0.37918190664742796, + "grad_norm": 1.1743371486663818, + "learning_rate": 9.346784948297054e-05, + "loss": 0.6693, + "step": 5650 + }, + { + "epoch": 0.37931613033119693, + "grad_norm": 1.0375677347183228, + "learning_rate": 9.346247750640491e-05, + "loss": 0.6843, + "step": 5652 + }, + { + "epoch": 0.37945035401496596, + "grad_norm": 1.1350818872451782, + "learning_rate": 9.34571034763227e-05, + "loss": 0.6466, + "step": 5654 + }, + { + "epoch": 0.37958457769873494, + "grad_norm": 1.1146163940429688, + "learning_rate": 9.34517273929778e-05, + "loss": 0.7157, + "step": 5656 + }, + { + "epoch": 0.37971880138250397, + "grad_norm": 1.2034099102020264, + "learning_rate": 9.34463492566242e-05, + "loss": 0.725, + "step": 5658 + }, + { + "epoch": 0.37985302506627294, + "grad_norm": 1.1982332468032837, + "learning_rate": 9.344096906751607e-05, + "loss": 0.659, + "step": 5660 + }, + { + "epoch": 0.379987248750042, + "grad_norm": 1.1427327394485474, + "learning_rate": 9.343558682590756e-05, + "loss": 0.7168, + "step": 5662 + }, + { + "epoch": 0.38012147243381095, + "grad_norm": 1.4069221019744873, + "learning_rate": 9.343020253205298e-05, + "loss": 0.6388, + "step": 5664 + }, + { + "epoch": 0.3802556961175799, + "grad_norm": 1.046435832977295, + "learning_rate": 9.342481618620673e-05, + "loss": 0.6666, + "step": 5666 + }, + { + "epoch": 0.38038991980134895, + "grad_norm": 1.6005109548568726, + "learning_rate": 9.341942778862331e-05, + "loss": 0.6731, + "step": 5668 + }, + { + "epoch": 0.3805241434851179, + "grad_norm": 1.3098866939544678, + "learning_rate": 9.341403733955732e-05, + "loss": 0.7199, + "step": 5670 + }, + { + "epoch": 0.38065836716888696, + "grad_norm": 1.012797474861145, + "learning_rate": 9.340864483926343e-05, + "loss": 0.7605, + "step": 5672 + }, + { + "epoch": 0.38079259085265593, + "grad_norm": 1.3212889432907104, + "learning_rate": 9.340325028799642e-05, + "loss": 0.6812, + "step": 5674 + }, + { + "epoch": 0.38092681453642496, + "grad_norm": 1.115822672843933, + "learning_rate": 9.339785368601119e-05, + "loss": 0.6059, + "step": 5676 + }, + { + "epoch": 0.38106103822019394, + "grad_norm": 0.9674661159515381, + "learning_rate": 9.339245503356271e-05, + "loss": 0.6786, + "step": 5678 + }, + { + "epoch": 0.38119526190396297, + "grad_norm": 1.1214262247085571, + "learning_rate": 9.338705433090607e-05, + "loss": 0.6805, + "step": 5680 + }, + { + "epoch": 0.38132948558773194, + "grad_norm": 1.024985909461975, + "learning_rate": 9.338165157829641e-05, + "loss": 0.653, + "step": 5682 + }, + { + "epoch": 0.381463709271501, + "grad_norm": 0.968446671962738, + "learning_rate": 9.337624677598903e-05, + "loss": 0.6053, + "step": 5684 + }, + { + "epoch": 0.38159793295526995, + "grad_norm": 1.6247999668121338, + "learning_rate": 9.337083992423927e-05, + "loss": 0.6536, + "step": 5686 + }, + { + "epoch": 0.381732156639039, + "grad_norm": 1.0137710571289062, + "learning_rate": 9.336543102330263e-05, + "loss": 0.6477, + "step": 5688 + }, + { + "epoch": 0.38186638032280795, + "grad_norm": 1.031615138053894, + "learning_rate": 9.336002007343464e-05, + "loss": 0.6506, + "step": 5690 + }, + { + "epoch": 0.382000604006577, + "grad_norm": 1.087575912475586, + "learning_rate": 9.335460707489097e-05, + "loss": 0.7036, + "step": 5692 + }, + { + "epoch": 0.38213482769034596, + "grad_norm": 0.9327123761177063, + "learning_rate": 9.334919202792736e-05, + "loss": 0.6221, + "step": 5694 + }, + { + "epoch": 0.382269051374115, + "grad_norm": 1.122715950012207, + "learning_rate": 9.334377493279968e-05, + "loss": 0.6935, + "step": 5696 + }, + { + "epoch": 0.38240327505788396, + "grad_norm": 1.0195330381393433, + "learning_rate": 9.333835578976385e-05, + "loss": 0.6045, + "step": 5698 + }, + { + "epoch": 0.38253749874165294, + "grad_norm": 1.1630983352661133, + "learning_rate": 9.333293459907595e-05, + "loss": 0.6921, + "step": 5700 + }, + { + "epoch": 0.38267172242542197, + "grad_norm": 1.014896273612976, + "learning_rate": 9.33275113609921e-05, + "loss": 0.6407, + "step": 5702 + }, + { + "epoch": 0.38280594610919094, + "grad_norm": 1.0407695770263672, + "learning_rate": 9.332208607576851e-05, + "loss": 0.6618, + "step": 5704 + }, + { + "epoch": 0.38294016979296, + "grad_norm": 1.3244556188583374, + "learning_rate": 9.331665874366156e-05, + "loss": 0.7152, + "step": 5706 + }, + { + "epoch": 0.38307439347672895, + "grad_norm": 1.077439785003662, + "learning_rate": 9.331122936492766e-05, + "loss": 0.6275, + "step": 5708 + }, + { + "epoch": 0.383208617160498, + "grad_norm": 1.2128117084503174, + "learning_rate": 9.330579793982335e-05, + "loss": 0.7595, + "step": 5710 + }, + { + "epoch": 0.38334284084426695, + "grad_norm": 1.2303240299224854, + "learning_rate": 9.330036446860524e-05, + "loss": 0.7631, + "step": 5712 + }, + { + "epoch": 0.383477064528036, + "grad_norm": 1.2520685195922852, + "learning_rate": 9.329492895153006e-05, + "loss": 0.6768, + "step": 5714 + }, + { + "epoch": 0.38361128821180496, + "grad_norm": 1.1363435983657837, + "learning_rate": 9.328949138885461e-05, + "loss": 0.6893, + "step": 5716 + }, + { + "epoch": 0.383745511895574, + "grad_norm": 1.4869840145111084, + "learning_rate": 9.328405178083584e-05, + "loss": 0.6519, + "step": 5718 + }, + { + "epoch": 0.38387973557934296, + "grad_norm": 1.1456031799316406, + "learning_rate": 9.327861012773071e-05, + "loss": 0.7197, + "step": 5720 + }, + { + "epoch": 0.384013959263112, + "grad_norm": 1.1507480144500732, + "learning_rate": 9.327316642979638e-05, + "loss": 0.7817, + "step": 5722 + }, + { + "epoch": 0.38414818294688097, + "grad_norm": 0.9522068500518799, + "learning_rate": 9.326772068729001e-05, + "loss": 0.6497, + "step": 5724 + }, + { + "epoch": 0.38428240663065, + "grad_norm": 1.0786789655685425, + "learning_rate": 9.326227290046892e-05, + "loss": 0.6879, + "step": 5726 + }, + { + "epoch": 0.384416630314419, + "grad_norm": 0.8930708765983582, + "learning_rate": 9.325682306959051e-05, + "loss": 0.6191, + "step": 5728 + }, + { + "epoch": 0.384550853998188, + "grad_norm": 1.2715744972229004, + "learning_rate": 9.325137119491227e-05, + "loss": 0.7194, + "step": 5730 + }, + { + "epoch": 0.384685077681957, + "grad_norm": 1.288090467453003, + "learning_rate": 9.324591727669181e-05, + "loss": 0.6743, + "step": 5732 + }, + { + "epoch": 0.384819301365726, + "grad_norm": 1.147087812423706, + "learning_rate": 9.324046131518678e-05, + "loss": 0.6997, + "step": 5734 + }, + { + "epoch": 0.384953525049495, + "grad_norm": 1.0619357824325562, + "learning_rate": 9.323500331065498e-05, + "loss": 0.6485, + "step": 5736 + }, + { + "epoch": 0.38508774873326396, + "grad_norm": 1.1486061811447144, + "learning_rate": 9.322954326335429e-05, + "loss": 0.7399, + "step": 5738 + }, + { + "epoch": 0.385221972417033, + "grad_norm": 0.9960038661956787, + "learning_rate": 9.322408117354271e-05, + "loss": 0.6371, + "step": 5740 + }, + { + "epoch": 0.38535619610080196, + "grad_norm": 1.1259183883666992, + "learning_rate": 9.321861704147827e-05, + "loss": 0.6672, + "step": 5742 + }, + { + "epoch": 0.385490419784571, + "grad_norm": 1.2630796432495117, + "learning_rate": 9.321315086741916e-05, + "loss": 0.6638, + "step": 5744 + }, + { + "epoch": 0.38562464346833997, + "grad_norm": 1.0901516675949097, + "learning_rate": 9.320768265162366e-05, + "loss": 0.6766, + "step": 5746 + }, + { + "epoch": 0.385758867152109, + "grad_norm": 1.1709264516830444, + "learning_rate": 9.320221239435012e-05, + "loss": 0.616, + "step": 5748 + }, + { + "epoch": 0.38589309083587797, + "grad_norm": 1.2259521484375, + "learning_rate": 9.319674009585699e-05, + "loss": 0.5973, + "step": 5750 + }, + { + "epoch": 0.386027314519647, + "grad_norm": 1.7706619501113892, + "learning_rate": 9.319126575640283e-05, + "loss": 0.6522, + "step": 5752 + }, + { + "epoch": 0.386161538203416, + "grad_norm": 1.1718233823776245, + "learning_rate": 9.318578937624629e-05, + "loss": 0.6479, + "step": 5754 + }, + { + "epoch": 0.386295761887185, + "grad_norm": 1.0942471027374268, + "learning_rate": 9.318031095564613e-05, + "loss": 0.6747, + "step": 5756 + }, + { + "epoch": 0.386429985570954, + "grad_norm": 1.0590591430664062, + "learning_rate": 9.31748304948612e-05, + "loss": 0.6361, + "step": 5758 + }, + { + "epoch": 0.386564209254723, + "grad_norm": 0.9840893745422363, + "learning_rate": 9.316934799415041e-05, + "loss": 0.67, + "step": 5760 + }, + { + "epoch": 0.386698432938492, + "grad_norm": 1.2189501523971558, + "learning_rate": 9.316386345377281e-05, + "loss": 0.6674, + "step": 5762 + }, + { + "epoch": 0.386832656622261, + "grad_norm": 1.0591330528259277, + "learning_rate": 9.315837687398756e-05, + "loss": 0.6833, + "step": 5764 + }, + { + "epoch": 0.38696688030603, + "grad_norm": 1.0182664394378662, + "learning_rate": 9.315288825505387e-05, + "loss": 0.6918, + "step": 5766 + }, + { + "epoch": 0.387101103989799, + "grad_norm": 1.3563718795776367, + "learning_rate": 9.314739759723105e-05, + "loss": 0.7408, + "step": 5768 + }, + { + "epoch": 0.387235327673568, + "grad_norm": 1.2516956329345703, + "learning_rate": 9.314190490077857e-05, + "loss": 0.6319, + "step": 5770 + }, + { + "epoch": 0.387369551357337, + "grad_norm": 1.058447241783142, + "learning_rate": 9.313641016595588e-05, + "loss": 0.6691, + "step": 5772 + }, + { + "epoch": 0.387503775041106, + "grad_norm": 1.2977604866027832, + "learning_rate": 9.313091339302267e-05, + "loss": 0.7333, + "step": 5774 + }, + { + "epoch": 0.387637998724875, + "grad_norm": 1.0655789375305176, + "learning_rate": 9.312541458223858e-05, + "loss": 0.7506, + "step": 5776 + }, + { + "epoch": 0.387772222408644, + "grad_norm": 1.0868114233016968, + "learning_rate": 9.311991373386349e-05, + "loss": 0.6913, + "step": 5778 + }, + { + "epoch": 0.387906446092413, + "grad_norm": 1.147434949874878, + "learning_rate": 9.311441084815724e-05, + "loss": 0.6807, + "step": 5780 + }, + { + "epoch": 0.388040669776182, + "grad_norm": 1.7061388492584229, + "learning_rate": 9.310890592537987e-05, + "loss": 0.6832, + "step": 5782 + }, + { + "epoch": 0.388174893459951, + "grad_norm": 1.2586777210235596, + "learning_rate": 9.310339896579145e-05, + "loss": 0.669, + "step": 5784 + }, + { + "epoch": 0.38830911714372, + "grad_norm": 0.9657377600669861, + "learning_rate": 9.30978899696522e-05, + "loss": 0.6274, + "step": 5786 + }, + { + "epoch": 0.388443340827489, + "grad_norm": 1.2519251108169556, + "learning_rate": 9.30923789372224e-05, + "loss": 0.6726, + "step": 5788 + }, + { + "epoch": 0.388577564511258, + "grad_norm": 1.0089620351791382, + "learning_rate": 9.308686586876243e-05, + "loss": 0.6184, + "step": 5790 + }, + { + "epoch": 0.388711788195027, + "grad_norm": 1.0273845195770264, + "learning_rate": 9.308135076453277e-05, + "loss": 0.6022, + "step": 5792 + }, + { + "epoch": 0.388846011878796, + "grad_norm": 1.6973581314086914, + "learning_rate": 9.307583362479402e-05, + "loss": 0.6645, + "step": 5794 + }, + { + "epoch": 0.388980235562565, + "grad_norm": 1.1413888931274414, + "learning_rate": 9.307031444980681e-05, + "loss": 0.6716, + "step": 5796 + }, + { + "epoch": 0.38911445924633403, + "grad_norm": 1.123063087463379, + "learning_rate": 9.306479323983195e-05, + "loss": 0.6733, + "step": 5798 + }, + { + "epoch": 0.389248682930103, + "grad_norm": 1.134331226348877, + "learning_rate": 9.305926999513029e-05, + "loss": 0.7427, + "step": 5800 + }, + { + "epoch": 0.38938290661387204, + "grad_norm": 1.0449963808059692, + "learning_rate": 9.30537447159628e-05, + "loss": 0.6834, + "step": 5802 + }, + { + "epoch": 0.389517130297641, + "grad_norm": 0.9834620952606201, + "learning_rate": 9.304821740259053e-05, + "loss": 0.6828, + "step": 5804 + }, + { + "epoch": 0.38965135398141004, + "grad_norm": 1.0525524616241455, + "learning_rate": 9.304268805527464e-05, + "loss": 0.7005, + "step": 5806 + }, + { + "epoch": 0.389785577665179, + "grad_norm": 1.0604172945022583, + "learning_rate": 9.303715667427639e-05, + "loss": 0.6736, + "step": 5808 + }, + { + "epoch": 0.38991980134894805, + "grad_norm": 1.0922068357467651, + "learning_rate": 9.30316232598571e-05, + "loss": 0.7082, + "step": 5810 + }, + { + "epoch": 0.390054025032717, + "grad_norm": 1.2460064888000488, + "learning_rate": 9.302608781227823e-05, + "loss": 0.7062, + "step": 5812 + }, + { + "epoch": 0.390188248716486, + "grad_norm": 1.2911078929901123, + "learning_rate": 9.302055033180133e-05, + "loss": 0.6592, + "step": 5814 + }, + { + "epoch": 0.390322472400255, + "grad_norm": 1.1606218814849854, + "learning_rate": 9.3015010818688e-05, + "loss": 0.6507, + "step": 5816 + }, + { + "epoch": 0.390456696084024, + "grad_norm": 1.1166902780532837, + "learning_rate": 9.300946927320001e-05, + "loss": 0.5937, + "step": 5818 + }, + { + "epoch": 0.39059091976779303, + "grad_norm": 1.0168946981430054, + "learning_rate": 9.300392569559917e-05, + "loss": 0.7135, + "step": 5820 + }, + { + "epoch": 0.390725143451562, + "grad_norm": 1.0619525909423828, + "learning_rate": 9.29983800861474e-05, + "loss": 0.7102, + "step": 5822 + }, + { + "epoch": 0.39085936713533104, + "grad_norm": 1.3438724279403687, + "learning_rate": 9.299283244510674e-05, + "loss": 0.7085, + "step": 5824 + }, + { + "epoch": 0.3909935908191, + "grad_norm": 1.0972028970718384, + "learning_rate": 9.298728277273927e-05, + "loss": 0.6942, + "step": 5826 + }, + { + "epoch": 0.39112781450286904, + "grad_norm": 1.483447551727295, + "learning_rate": 9.298173106930723e-05, + "loss": 0.5973, + "step": 5828 + }, + { + "epoch": 0.391262038186638, + "grad_norm": 1.2423871755599976, + "learning_rate": 9.297617733507291e-05, + "loss": 0.7694, + "step": 5830 + }, + { + "epoch": 0.39139626187040705, + "grad_norm": 0.929648756980896, + "learning_rate": 9.297062157029872e-05, + "loss": 0.6573, + "step": 5832 + }, + { + "epoch": 0.391530485554176, + "grad_norm": 1.2835553884506226, + "learning_rate": 9.296506377524716e-05, + "loss": 0.7118, + "step": 5834 + }, + { + "epoch": 0.39166470923794505, + "grad_norm": 1.2994301319122314, + "learning_rate": 9.295950395018084e-05, + "loss": 0.7424, + "step": 5836 + }, + { + "epoch": 0.391798932921714, + "grad_norm": 1.0983941555023193, + "learning_rate": 9.295394209536242e-05, + "loss": 0.758, + "step": 5838 + }, + { + "epoch": 0.39193315660548306, + "grad_norm": 1.120267391204834, + "learning_rate": 9.29483782110547e-05, + "loss": 0.7137, + "step": 5840 + }, + { + "epoch": 0.39206738028925203, + "grad_norm": 1.349510908126831, + "learning_rate": 9.294281229752057e-05, + "loss": 0.6632, + "step": 5842 + }, + { + "epoch": 0.39220160397302106, + "grad_norm": 1.0613240003585815, + "learning_rate": 9.2937244355023e-05, + "loss": 0.6693, + "step": 5844 + }, + { + "epoch": 0.39233582765679004, + "grad_norm": 1.2015703916549683, + "learning_rate": 9.293167438382506e-05, + "loss": 0.7193, + "step": 5846 + }, + { + "epoch": 0.39247005134055907, + "grad_norm": 1.14816415309906, + "learning_rate": 9.292610238418992e-05, + "loss": 0.6654, + "step": 5848 + }, + { + "epoch": 0.39260427502432804, + "grad_norm": 1.147283911705017, + "learning_rate": 9.292052835638088e-05, + "loss": 0.762, + "step": 5850 + }, + { + "epoch": 0.392738498708097, + "grad_norm": 1.0627939701080322, + "learning_rate": 9.291495230066125e-05, + "loss": 0.6812, + "step": 5852 + }, + { + "epoch": 0.39287272239186605, + "grad_norm": 1.2463438510894775, + "learning_rate": 9.290937421729454e-05, + "loss": 0.6162, + "step": 5854 + }, + { + "epoch": 0.393006946075635, + "grad_norm": 1.1472399234771729, + "learning_rate": 9.290379410654425e-05, + "loss": 0.6816, + "step": 5856 + }, + { + "epoch": 0.39314116975940405, + "grad_norm": 1.1549139022827148, + "learning_rate": 9.289821196867405e-05, + "loss": 0.7213, + "step": 5858 + }, + { + "epoch": 0.393275393443173, + "grad_norm": 1.0665532350540161, + "learning_rate": 9.289262780394772e-05, + "loss": 0.672, + "step": 5860 + }, + { + "epoch": 0.39340961712694206, + "grad_norm": 1.235317587852478, + "learning_rate": 9.288704161262904e-05, + "loss": 0.6619, + "step": 5862 + }, + { + "epoch": 0.39354384081071103, + "grad_norm": 1.0959912538528442, + "learning_rate": 9.2881453394982e-05, + "loss": 0.6623, + "step": 5864 + }, + { + "epoch": 0.39367806449448006, + "grad_norm": 1.1718889474868774, + "learning_rate": 9.28758631512706e-05, + "loss": 0.6712, + "step": 5866 + }, + { + "epoch": 0.39381228817824904, + "grad_norm": 1.1043980121612549, + "learning_rate": 9.287027088175898e-05, + "loss": 0.7072, + "step": 5868 + }, + { + "epoch": 0.39394651186201807, + "grad_norm": 1.0995796918869019, + "learning_rate": 9.286467658671136e-05, + "loss": 0.6571, + "step": 5870 + }, + { + "epoch": 0.39408073554578704, + "grad_norm": 1.475175380706787, + "learning_rate": 9.285908026639207e-05, + "loss": 0.7097, + "step": 5872 + }, + { + "epoch": 0.3942149592295561, + "grad_norm": 1.5103731155395508, + "learning_rate": 9.285348192106551e-05, + "loss": 0.7555, + "step": 5874 + }, + { + "epoch": 0.39434918291332505, + "grad_norm": 1.0760539770126343, + "learning_rate": 9.28478815509962e-05, + "loss": 0.6706, + "step": 5876 + }, + { + "epoch": 0.3944834065970941, + "grad_norm": 0.9949193596839905, + "learning_rate": 9.284227915644872e-05, + "loss": 0.7143, + "step": 5878 + }, + { + "epoch": 0.39461763028086305, + "grad_norm": 1.2008304595947266, + "learning_rate": 9.283667473768782e-05, + "loss": 0.6664, + "step": 5880 + }, + { + "epoch": 0.3947518539646321, + "grad_norm": 1.1227686405181885, + "learning_rate": 9.283106829497828e-05, + "loss": 0.7323, + "step": 5882 + }, + { + "epoch": 0.39488607764840106, + "grad_norm": 1.1749778985977173, + "learning_rate": 9.282545982858496e-05, + "loss": 0.6511, + "step": 5884 + }, + { + "epoch": 0.3950203013321701, + "grad_norm": 0.969157874584198, + "learning_rate": 9.28198493387729e-05, + "loss": 0.687, + "step": 5886 + }, + { + "epoch": 0.39515452501593906, + "grad_norm": 1.1056897640228271, + "learning_rate": 9.281423682580714e-05, + "loss": 0.7228, + "step": 5888 + }, + { + "epoch": 0.39528874869970804, + "grad_norm": 1.1171948909759521, + "learning_rate": 9.280862228995291e-05, + "loss": 0.7403, + "step": 5890 + }, + { + "epoch": 0.39542297238347707, + "grad_norm": 1.099923014640808, + "learning_rate": 9.280300573147542e-05, + "loss": 0.6672, + "step": 5892 + }, + { + "epoch": 0.39555719606724604, + "grad_norm": 1.0958002805709839, + "learning_rate": 9.27973871506401e-05, + "loss": 0.7194, + "step": 5894 + }, + { + "epoch": 0.39569141975101507, + "grad_norm": 1.0290558338165283, + "learning_rate": 9.27917665477124e-05, + "loss": 0.6969, + "step": 5896 + }, + { + "epoch": 0.39582564343478405, + "grad_norm": 1.0672677755355835, + "learning_rate": 9.278614392295786e-05, + "loss": 0.6818, + "step": 5898 + }, + { + "epoch": 0.3959598671185531, + "grad_norm": 1.061200737953186, + "learning_rate": 9.278051927664217e-05, + "loss": 0.6052, + "step": 5900 + }, + { + "epoch": 0.39609409080232205, + "grad_norm": 1.2648062705993652, + "learning_rate": 9.277489260903104e-05, + "loss": 0.6926, + "step": 5902 + }, + { + "epoch": 0.3962283144860911, + "grad_norm": 1.0376838445663452, + "learning_rate": 9.276926392039038e-05, + "loss": 0.6976, + "step": 5904 + }, + { + "epoch": 0.39636253816986006, + "grad_norm": 1.3055962324142456, + "learning_rate": 9.276363321098609e-05, + "loss": 0.6908, + "step": 5906 + }, + { + "epoch": 0.3964967618536291, + "grad_norm": 1.1608704328536987, + "learning_rate": 9.275800048108423e-05, + "loss": 0.6734, + "step": 5908 + }, + { + "epoch": 0.39663098553739806, + "grad_norm": 1.1547906398773193, + "learning_rate": 9.27523657309509e-05, + "loss": 0.7042, + "step": 5910 + }, + { + "epoch": 0.3967652092211671, + "grad_norm": 1.0295121669769287, + "learning_rate": 9.27467289608524e-05, + "loss": 0.6748, + "step": 5912 + }, + { + "epoch": 0.39689943290493607, + "grad_norm": 1.1790900230407715, + "learning_rate": 9.274109017105497e-05, + "loss": 0.6604, + "step": 5914 + }, + { + "epoch": 0.3970336565887051, + "grad_norm": 1.0177485942840576, + "learning_rate": 9.27354493618251e-05, + "loss": 0.6853, + "step": 5916 + }, + { + "epoch": 0.39716788027247407, + "grad_norm": 1.054178237915039, + "learning_rate": 9.272980653342929e-05, + "loss": 0.6572, + "step": 5918 + }, + { + "epoch": 0.3973021039562431, + "grad_norm": 0.9216815233230591, + "learning_rate": 9.272416168613414e-05, + "loss": 0.6137, + "step": 5920 + }, + { + "epoch": 0.3974363276400121, + "grad_norm": 1.038726806640625, + "learning_rate": 9.271851482020635e-05, + "loss": 0.6604, + "step": 5922 + }, + { + "epoch": 0.3975705513237811, + "grad_norm": 1.3206719160079956, + "learning_rate": 9.271286593591275e-05, + "loss": 0.7159, + "step": 5924 + }, + { + "epoch": 0.3977047750075501, + "grad_norm": 1.3097436428070068, + "learning_rate": 9.270721503352022e-05, + "loss": 0.7292, + "step": 5926 + }, + { + "epoch": 0.39783899869131906, + "grad_norm": 1.3265849351882935, + "learning_rate": 9.270156211329578e-05, + "loss": 0.6842, + "step": 5928 + }, + { + "epoch": 0.3979732223750881, + "grad_norm": 0.9673488736152649, + "learning_rate": 9.269590717550647e-05, + "loss": 0.6519, + "step": 5930 + }, + { + "epoch": 0.39810744605885706, + "grad_norm": 1.0370547771453857, + "learning_rate": 9.269025022041953e-05, + "loss": 0.6785, + "step": 5932 + }, + { + "epoch": 0.3982416697426261, + "grad_norm": 1.2383445501327515, + "learning_rate": 9.268459124830218e-05, + "loss": 0.6802, + "step": 5934 + }, + { + "epoch": 0.39837589342639507, + "grad_norm": 1.4812686443328857, + "learning_rate": 9.267893025942186e-05, + "loss": 0.6809, + "step": 5936 + }, + { + "epoch": 0.3985101171101641, + "grad_norm": 1.1149955987930298, + "learning_rate": 9.267326725404599e-05, + "loss": 0.712, + "step": 5938 + }, + { + "epoch": 0.39864434079393307, + "grad_norm": 1.1058670282363892, + "learning_rate": 9.266760223244218e-05, + "loss": 0.6745, + "step": 5940 + }, + { + "epoch": 0.3987785644777021, + "grad_norm": 1.172837257385254, + "learning_rate": 9.266193519487805e-05, + "loss": 0.698, + "step": 5942 + }, + { + "epoch": 0.3989127881614711, + "grad_norm": 1.115440845489502, + "learning_rate": 9.265626614162137e-05, + "loss": 0.693, + "step": 5944 + }, + { + "epoch": 0.3990470118452401, + "grad_norm": 1.037589430809021, + "learning_rate": 9.265059507294001e-05, + "loss": 0.7305, + "step": 5946 + }, + { + "epoch": 0.3991812355290091, + "grad_norm": 1.1750209331512451, + "learning_rate": 9.264492198910189e-05, + "loss": 0.7095, + "step": 5948 + }, + { + "epoch": 0.3993154592127781, + "grad_norm": 1.1001358032226562, + "learning_rate": 9.263924689037505e-05, + "loss": 0.6478, + "step": 5950 + }, + { + "epoch": 0.3994496828965471, + "grad_norm": 1.6089013814926147, + "learning_rate": 9.263356977702766e-05, + "loss": 0.7067, + "step": 5952 + }, + { + "epoch": 0.3995839065803161, + "grad_norm": 1.092583179473877, + "learning_rate": 9.262789064932794e-05, + "loss": 0.6607, + "step": 5954 + }, + { + "epoch": 0.3997181302640851, + "grad_norm": 1.0944645404815674, + "learning_rate": 9.262220950754419e-05, + "loss": 0.6722, + "step": 5956 + }, + { + "epoch": 0.3998523539478541, + "grad_norm": 1.061228632926941, + "learning_rate": 9.261652635194487e-05, + "loss": 0.6982, + "step": 5958 + }, + { + "epoch": 0.3999865776316231, + "grad_norm": 0.9771525263786316, + "learning_rate": 9.261084118279847e-05, + "loss": 0.6908, + "step": 5960 + }, + { + "epoch": 0.4001208013153921, + "grad_norm": 1.602399468421936, + "learning_rate": 9.260515400037362e-05, + "loss": 0.6764, + "step": 5962 + }, + { + "epoch": 0.4002550249991611, + "grad_norm": 1.0964266061782837, + "learning_rate": 9.259946480493902e-05, + "loss": 0.685, + "step": 5964 + }, + { + "epoch": 0.4003892486829301, + "grad_norm": 1.1376968622207642, + "learning_rate": 9.259377359676348e-05, + "loss": 0.7389, + "step": 5966 + }, + { + "epoch": 0.4005234723666991, + "grad_norm": 1.1643990278244019, + "learning_rate": 9.25880803761159e-05, + "loss": 0.5875, + "step": 5968 + }, + { + "epoch": 0.4006576960504681, + "grad_norm": 1.0562384128570557, + "learning_rate": 9.258238514326525e-05, + "loss": 0.605, + "step": 5970 + }, + { + "epoch": 0.4007919197342371, + "grad_norm": 1.1646043062210083, + "learning_rate": 9.257668789848067e-05, + "loss": 0.6567, + "step": 5972 + }, + { + "epoch": 0.4009261434180061, + "grad_norm": 1.1420818567276, + "learning_rate": 9.257098864203128e-05, + "loss": 0.6783, + "step": 5974 + }, + { + "epoch": 0.4010603671017751, + "grad_norm": 1.2688438892364502, + "learning_rate": 9.25652873741864e-05, + "loss": 0.6516, + "step": 5976 + }, + { + "epoch": 0.4011945907855441, + "grad_norm": 1.054469108581543, + "learning_rate": 9.255958409521538e-05, + "loss": 0.6125, + "step": 5978 + }, + { + "epoch": 0.4013288144693131, + "grad_norm": 1.1423333883285522, + "learning_rate": 9.25538788053877e-05, + "loss": 0.5963, + "step": 5980 + }, + { + "epoch": 0.4014630381530821, + "grad_norm": 1.0816683769226074, + "learning_rate": 9.254817150497295e-05, + "loss": 0.6697, + "step": 5982 + }, + { + "epoch": 0.4015972618368511, + "grad_norm": 0.975953221321106, + "learning_rate": 9.254246219424075e-05, + "loss": 0.6628, + "step": 5984 + }, + { + "epoch": 0.4017314855206201, + "grad_norm": 1.167656421661377, + "learning_rate": 9.253675087346087e-05, + "loss": 0.6356, + "step": 5986 + }, + { + "epoch": 0.40186570920438913, + "grad_norm": 1.33633553981781, + "learning_rate": 9.253103754290315e-05, + "loss": 0.6431, + "step": 5988 + }, + { + "epoch": 0.4019999328881581, + "grad_norm": 1.1282232999801636, + "learning_rate": 9.252532220283754e-05, + "loss": 0.6035, + "step": 5990 + }, + { + "epoch": 0.40213415657192714, + "grad_norm": 0.9276664853096008, + "learning_rate": 9.251960485353408e-05, + "loss": 0.6548, + "step": 5992 + }, + { + "epoch": 0.4022683802556961, + "grad_norm": 1.0622806549072266, + "learning_rate": 9.251388549526292e-05, + "loss": 0.6835, + "step": 5994 + }, + { + "epoch": 0.40240260393946514, + "grad_norm": 1.0541036128997803, + "learning_rate": 9.250816412829425e-05, + "loss": 0.7287, + "step": 5996 + }, + { + "epoch": 0.4025368276232341, + "grad_norm": 1.134677767753601, + "learning_rate": 9.250244075289843e-05, + "loss": 0.6533, + "step": 5998 + }, + { + "epoch": 0.40267105130700315, + "grad_norm": 1.189555048942566, + "learning_rate": 9.249671536934585e-05, + "loss": 0.6717, + "step": 6000 + }, + { + "epoch": 0.4028052749907721, + "grad_norm": 1.0003716945648193, + "learning_rate": 9.249098797790702e-05, + "loss": 0.6487, + "step": 6002 + }, + { + "epoch": 0.4029394986745411, + "grad_norm": 1.3909540176391602, + "learning_rate": 9.248525857885259e-05, + "loss": 0.6151, + "step": 6004 + }, + { + "epoch": 0.4030737223583101, + "grad_norm": 0.9759494066238403, + "learning_rate": 9.247952717245321e-05, + "loss": 0.7081, + "step": 6006 + }, + { + "epoch": 0.4032079460420791, + "grad_norm": 1.2941428422927856, + "learning_rate": 9.247379375897974e-05, + "loss": 0.7397, + "step": 6008 + }, + { + "epoch": 0.40334216972584813, + "grad_norm": 1.346400260925293, + "learning_rate": 9.2468058338703e-05, + "loss": 0.7726, + "step": 6010 + }, + { + "epoch": 0.4034763934096171, + "grad_norm": 1.0712957382202148, + "learning_rate": 9.246232091189402e-05, + "loss": 0.6759, + "step": 6012 + }, + { + "epoch": 0.40361061709338614, + "grad_norm": 1.012364387512207, + "learning_rate": 9.245658147882388e-05, + "loss": 0.6428, + "step": 6014 + }, + { + "epoch": 0.4037448407771551, + "grad_norm": 1.3964303731918335, + "learning_rate": 9.245084003976377e-05, + "loss": 0.7316, + "step": 6016 + }, + { + "epoch": 0.40387906446092414, + "grad_norm": 1.0563886165618896, + "learning_rate": 9.244509659498493e-05, + "loss": 0.7132, + "step": 6018 + }, + { + "epoch": 0.4040132881446931, + "grad_norm": 0.9809919595718384, + "learning_rate": 9.243935114475872e-05, + "loss": 0.656, + "step": 6020 + }, + { + "epoch": 0.40414751182846215, + "grad_norm": 1.0551362037658691, + "learning_rate": 9.243360368935666e-05, + "loss": 0.6481, + "step": 6022 + }, + { + "epoch": 0.4042817355122311, + "grad_norm": 1.0678483247756958, + "learning_rate": 9.242785422905025e-05, + "loss": 0.6739, + "step": 6024 + }, + { + "epoch": 0.40441595919600015, + "grad_norm": 0.9836398959159851, + "learning_rate": 9.242210276411115e-05, + "loss": 0.6616, + "step": 6026 + }, + { + "epoch": 0.4045501828797691, + "grad_norm": 1.125788927078247, + "learning_rate": 9.241634929481112e-05, + "loss": 0.6742, + "step": 6028 + }, + { + "epoch": 0.40468440656353816, + "grad_norm": 1.3970268964767456, + "learning_rate": 9.2410593821422e-05, + "loss": 0.638, + "step": 6030 + }, + { + "epoch": 0.40481863024730713, + "grad_norm": 1.178939938545227, + "learning_rate": 9.24048363442157e-05, + "loss": 0.6704, + "step": 6032 + }, + { + "epoch": 0.40495285393107616, + "grad_norm": 1.0103930234909058, + "learning_rate": 9.239907686346429e-05, + "loss": 0.643, + "step": 6034 + }, + { + "epoch": 0.40508707761484514, + "grad_norm": 1.0354652404785156, + "learning_rate": 9.239331537943987e-05, + "loss": 0.6443, + "step": 6036 + }, + { + "epoch": 0.40522130129861417, + "grad_norm": 1.1033591032028198, + "learning_rate": 9.238755189241466e-05, + "loss": 0.713, + "step": 6038 + }, + { + "epoch": 0.40535552498238314, + "grad_norm": 0.9593886733055115, + "learning_rate": 9.238178640266096e-05, + "loss": 0.638, + "step": 6040 + }, + { + "epoch": 0.4054897486661521, + "grad_norm": 0.9968031048774719, + "learning_rate": 9.23760189104512e-05, + "loss": 0.6368, + "step": 6042 + }, + { + "epoch": 0.40562397234992115, + "grad_norm": 1.142154574394226, + "learning_rate": 9.237024941605789e-05, + "loss": 0.6719, + "step": 6044 + }, + { + "epoch": 0.4057581960336901, + "grad_norm": 1.0583053827285767, + "learning_rate": 9.23644779197536e-05, + "loss": 0.6468, + "step": 6046 + }, + { + "epoch": 0.40589241971745915, + "grad_norm": 1.1052918434143066, + "learning_rate": 9.235870442181104e-05, + "loss": 0.6907, + "step": 6048 + }, + { + "epoch": 0.4060266434012281, + "grad_norm": 1.1614830493927002, + "learning_rate": 9.235292892250298e-05, + "loss": 0.7555, + "step": 6050 + }, + { + "epoch": 0.40616086708499716, + "grad_norm": 1.1473082304000854, + "learning_rate": 9.234715142210233e-05, + "loss": 0.6158, + "step": 6052 + }, + { + "epoch": 0.40629509076876613, + "grad_norm": 1.0020947456359863, + "learning_rate": 9.234137192088202e-05, + "loss": 0.6192, + "step": 6054 + }, + { + "epoch": 0.40642931445253516, + "grad_norm": 1.0485420227050781, + "learning_rate": 9.233559041911517e-05, + "loss": 0.6748, + "step": 6056 + }, + { + "epoch": 0.40656353813630414, + "grad_norm": 0.935448944568634, + "learning_rate": 9.232980691707491e-05, + "loss": 0.5965, + "step": 6058 + }, + { + "epoch": 0.40669776182007317, + "grad_norm": 1.0142656564712524, + "learning_rate": 9.232402141503452e-05, + "loss": 0.7221, + "step": 6060 + }, + { + "epoch": 0.40683198550384214, + "grad_norm": 1.1718939542770386, + "learning_rate": 9.231823391326734e-05, + "loss": 0.627, + "step": 6062 + }, + { + "epoch": 0.40696620918761117, + "grad_norm": 1.0399384498596191, + "learning_rate": 9.231244441204683e-05, + "loss": 0.6466, + "step": 6064 + }, + { + "epoch": 0.40710043287138015, + "grad_norm": 1.2102640867233276, + "learning_rate": 9.230665291164652e-05, + "loss": 0.6899, + "step": 6066 + }, + { + "epoch": 0.4072346565551492, + "grad_norm": 1.152836561203003, + "learning_rate": 9.230085941234006e-05, + "loss": 0.7211, + "step": 6068 + }, + { + "epoch": 0.40736888023891815, + "grad_norm": 1.1507648229599, + "learning_rate": 9.229506391440115e-05, + "loss": 0.7227, + "step": 6070 + }, + { + "epoch": 0.4075031039226872, + "grad_norm": 1.045613408088684, + "learning_rate": 9.228926641810367e-05, + "loss": 0.6501, + "step": 6072 + }, + { + "epoch": 0.40763732760645616, + "grad_norm": 1.1735528707504272, + "learning_rate": 9.22834669237215e-05, + "loss": 0.6691, + "step": 6074 + }, + { + "epoch": 0.4077715512902252, + "grad_norm": 1.136020541191101, + "learning_rate": 9.227766543152864e-05, + "loss": 0.6165, + "step": 6076 + }, + { + "epoch": 0.40790577497399416, + "grad_norm": 1.159848928451538, + "learning_rate": 9.227186194179925e-05, + "loss": 0.6709, + "step": 6078 + }, + { + "epoch": 0.40803999865776314, + "grad_norm": 1.3651721477508545, + "learning_rate": 9.22660564548075e-05, + "loss": 0.7813, + "step": 6080 + }, + { + "epoch": 0.40817422234153217, + "grad_norm": 1.0709141492843628, + "learning_rate": 9.22602489708277e-05, + "loss": 0.6063, + "step": 6082 + }, + { + "epoch": 0.40830844602530114, + "grad_norm": 1.0377559661865234, + "learning_rate": 9.225443949013424e-05, + "loss": 0.6476, + "step": 6084 + }, + { + "epoch": 0.40844266970907017, + "grad_norm": 1.231859803199768, + "learning_rate": 9.224862801300159e-05, + "loss": 0.7013, + "step": 6086 + }, + { + "epoch": 0.40857689339283915, + "grad_norm": 1.0269285440444946, + "learning_rate": 9.224281453970436e-05, + "loss": 0.6628, + "step": 6088 + }, + { + "epoch": 0.4087111170766082, + "grad_norm": 1.080043077468872, + "learning_rate": 9.22369990705172e-05, + "loss": 0.6303, + "step": 6090 + }, + { + "epoch": 0.40884534076037715, + "grad_norm": 1.0308558940887451, + "learning_rate": 9.223118160571489e-05, + "loss": 0.6574, + "step": 6092 + }, + { + "epoch": 0.4089795644441462, + "grad_norm": 1.0971088409423828, + "learning_rate": 9.22253621455723e-05, + "loss": 0.6433, + "step": 6094 + }, + { + "epoch": 0.40911378812791516, + "grad_norm": 0.9573982954025269, + "learning_rate": 9.221954069036438e-05, + "loss": 0.6165, + "step": 6096 + }, + { + "epoch": 0.4092480118116842, + "grad_norm": 1.0471969842910767, + "learning_rate": 9.221371724036619e-05, + "loss": 0.7094, + "step": 6098 + }, + { + "epoch": 0.40938223549545316, + "grad_norm": 1.1849780082702637, + "learning_rate": 9.220789179585286e-05, + "loss": 0.6669, + "step": 6100 + }, + { + "epoch": 0.4095164591792222, + "grad_norm": 1.0062451362609863, + "learning_rate": 9.220206435709963e-05, + "loss": 0.6624, + "step": 6102 + }, + { + "epoch": 0.40965068286299117, + "grad_norm": 1.4470155239105225, + "learning_rate": 9.219623492438185e-05, + "loss": 0.6079, + "step": 6104 + }, + { + "epoch": 0.4097849065467602, + "grad_norm": 1.346421241760254, + "learning_rate": 9.219040349797495e-05, + "loss": 0.7008, + "step": 6106 + }, + { + "epoch": 0.40991913023052917, + "grad_norm": 1.0121887922286987, + "learning_rate": 9.218457007815447e-05, + "loss": 0.6691, + "step": 6108 + }, + { + "epoch": 0.4100533539142982, + "grad_norm": 2.2194879055023193, + "learning_rate": 9.217873466519597e-05, + "loss": 0.7418, + "step": 6110 + }, + { + "epoch": 0.4101875775980672, + "grad_norm": 0.9041343331336975, + "learning_rate": 9.217289725937521e-05, + "loss": 0.615, + "step": 6112 + }, + { + "epoch": 0.4103218012818362, + "grad_norm": 1.1116787195205688, + "learning_rate": 9.2167057860968e-05, + "loss": 0.681, + "step": 6114 + }, + { + "epoch": 0.4104560249656052, + "grad_norm": 1.2395095825195312, + "learning_rate": 9.216121647025021e-05, + "loss": 0.7903, + "step": 6116 + }, + { + "epoch": 0.41059024864937416, + "grad_norm": 1.2876553535461426, + "learning_rate": 9.215537308749784e-05, + "loss": 0.6527, + "step": 6118 + }, + { + "epoch": 0.4107244723331432, + "grad_norm": 1.1910831928253174, + "learning_rate": 9.214952771298701e-05, + "loss": 0.6842, + "step": 6120 + }, + { + "epoch": 0.41085869601691216, + "grad_norm": 0.9551767706871033, + "learning_rate": 9.214368034699387e-05, + "loss": 0.6568, + "step": 6122 + }, + { + "epoch": 0.4109929197006812, + "grad_norm": 1.0212019681930542, + "learning_rate": 9.213783098979469e-05, + "loss": 0.7006, + "step": 6124 + }, + { + "epoch": 0.41112714338445017, + "grad_norm": 1.1502602100372314, + "learning_rate": 9.213197964166587e-05, + "loss": 0.6756, + "step": 6126 + }, + { + "epoch": 0.4112613670682192, + "grad_norm": 1.0294241905212402, + "learning_rate": 9.212612630288386e-05, + "loss": 0.6463, + "step": 6128 + }, + { + "epoch": 0.41139559075198817, + "grad_norm": 1.4980016946792603, + "learning_rate": 9.212027097372522e-05, + "loss": 0.7198, + "step": 6130 + }, + { + "epoch": 0.4115298144357572, + "grad_norm": 1.0276916027069092, + "learning_rate": 9.21144136544666e-05, + "loss": 0.6685, + "step": 6132 + }, + { + "epoch": 0.4116640381195262, + "grad_norm": 0.9319644570350647, + "learning_rate": 9.210855434538477e-05, + "loss": 0.6197, + "step": 6134 + }, + { + "epoch": 0.4117982618032952, + "grad_norm": 1.0256725549697876, + "learning_rate": 9.210269304675652e-05, + "loss": 0.6934, + "step": 6136 + }, + { + "epoch": 0.4119324854870642, + "grad_norm": 2.123508930206299, + "learning_rate": 9.209682975885882e-05, + "loss": 0.6992, + "step": 6138 + }, + { + "epoch": 0.4120667091708332, + "grad_norm": 1.2907676696777344, + "learning_rate": 9.209096448196872e-05, + "loss": 0.692, + "step": 6140 + }, + { + "epoch": 0.4122009328546022, + "grad_norm": 1.0666604042053223, + "learning_rate": 9.208509721636328e-05, + "loss": 0.6574, + "step": 6142 + }, + { + "epoch": 0.4123351565383712, + "grad_norm": 1.0116019248962402, + "learning_rate": 9.207922796231977e-05, + "loss": 0.6908, + "step": 6144 + }, + { + "epoch": 0.4124693802221402, + "grad_norm": 1.1583993434906006, + "learning_rate": 9.20733567201155e-05, + "loss": 0.6884, + "step": 6146 + }, + { + "epoch": 0.4126036039059092, + "grad_norm": 1.0761034488677979, + "learning_rate": 9.206748349002782e-05, + "loss": 0.663, + "step": 6148 + }, + { + "epoch": 0.4127378275896782, + "grad_norm": 1.0891854763031006, + "learning_rate": 9.20616082723343e-05, + "loss": 0.6682, + "step": 6150 + }, + { + "epoch": 0.4128720512734472, + "grad_norm": 1.028387427330017, + "learning_rate": 9.20557310673125e-05, + "loss": 0.7085, + "step": 6152 + }, + { + "epoch": 0.4130062749572162, + "grad_norm": 1.0554955005645752, + "learning_rate": 9.20498518752401e-05, + "loss": 0.586, + "step": 6154 + }, + { + "epoch": 0.4131404986409852, + "grad_norm": 0.9535325169563293, + "learning_rate": 9.204397069639486e-05, + "loss": 0.6625, + "step": 6156 + }, + { + "epoch": 0.4132747223247542, + "grad_norm": 1.0058443546295166, + "learning_rate": 9.203808753105471e-05, + "loss": 0.6522, + "step": 6158 + }, + { + "epoch": 0.4134089460085232, + "grad_norm": 1.4530055522918701, + "learning_rate": 9.203220237949758e-05, + "loss": 0.6723, + "step": 6160 + }, + { + "epoch": 0.4135431696922922, + "grad_norm": 0.9904844760894775, + "learning_rate": 9.202631524200153e-05, + "loss": 0.5549, + "step": 6162 + }, + { + "epoch": 0.4136773933760612, + "grad_norm": 1.3306118249893188, + "learning_rate": 9.202042611884475e-05, + "loss": 0.6845, + "step": 6164 + }, + { + "epoch": 0.4138116170598302, + "grad_norm": 0.8913776278495789, + "learning_rate": 9.201453501030546e-05, + "loss": 0.6191, + "step": 6166 + }, + { + "epoch": 0.4139458407435992, + "grad_norm": 1.0931206941604614, + "learning_rate": 9.200864191666199e-05, + "loss": 0.623, + "step": 6168 + }, + { + "epoch": 0.4140800644273682, + "grad_norm": 1.4609962701797485, + "learning_rate": 9.200274683819282e-05, + "loss": 0.653, + "step": 6170 + }, + { + "epoch": 0.4142142881111372, + "grad_norm": 1.0243514776229858, + "learning_rate": 9.199684977517645e-05, + "loss": 0.6721, + "step": 6172 + }, + { + "epoch": 0.4143485117949062, + "grad_norm": 1.3936306238174438, + "learning_rate": 9.199095072789149e-05, + "loss": 0.7181, + "step": 6174 + }, + { + "epoch": 0.4144827354786752, + "grad_norm": 1.211826205253601, + "learning_rate": 9.19850496966167e-05, + "loss": 0.6401, + "step": 6176 + }, + { + "epoch": 0.41461695916244423, + "grad_norm": 1.2502893209457397, + "learning_rate": 9.197914668163085e-05, + "loss": 0.6879, + "step": 6178 + }, + { + "epoch": 0.4147511828462132, + "grad_norm": 1.0616451501846313, + "learning_rate": 9.19732416832129e-05, + "loss": 0.6886, + "step": 6180 + }, + { + "epoch": 0.41488540652998224, + "grad_norm": 1.069578766822815, + "learning_rate": 9.19673347016418e-05, + "loss": 0.5796, + "step": 6182 + }, + { + "epoch": 0.4150196302137512, + "grad_norm": 1.216532588005066, + "learning_rate": 9.196142573719666e-05, + "loss": 0.734, + "step": 6184 + }, + { + "epoch": 0.41515385389752024, + "grad_norm": 1.1617928743362427, + "learning_rate": 9.195551479015667e-05, + "loss": 0.6501, + "step": 6186 + }, + { + "epoch": 0.4152880775812892, + "grad_norm": 1.0069537162780762, + "learning_rate": 9.19496018608011e-05, + "loss": 0.6785, + "step": 6188 + }, + { + "epoch": 0.41542230126505825, + "grad_norm": 1.260364294052124, + "learning_rate": 9.194368694940935e-05, + "loss": 0.6827, + "step": 6190 + }, + { + "epoch": 0.4155565249488272, + "grad_norm": 1.0657905340194702, + "learning_rate": 9.193777005626086e-05, + "loss": 0.6688, + "step": 6192 + }, + { + "epoch": 0.4156907486325962, + "grad_norm": 1.3277440071105957, + "learning_rate": 9.193185118163521e-05, + "loss": 0.6361, + "step": 6194 + }, + { + "epoch": 0.4158249723163652, + "grad_norm": 1.1172504425048828, + "learning_rate": 9.192593032581203e-05, + "loss": 0.6907, + "step": 6196 + }, + { + "epoch": 0.4159591960001342, + "grad_norm": 1.1119155883789062, + "learning_rate": 9.19200074890711e-05, + "loss": 0.6752, + "step": 6198 + }, + { + "epoch": 0.41609341968390323, + "grad_norm": 1.1108487844467163, + "learning_rate": 9.191408267169226e-05, + "loss": 0.7023, + "step": 6200 + }, + { + "epoch": 0.4162276433676722, + "grad_norm": 0.9239070415496826, + "learning_rate": 9.19081558739554e-05, + "loss": 0.6321, + "step": 6202 + }, + { + "epoch": 0.41636186705144124, + "grad_norm": 0.999114453792572, + "learning_rate": 9.190222709614061e-05, + "loss": 0.6482, + "step": 6204 + }, + { + "epoch": 0.4164960907352102, + "grad_norm": 1.2064570188522339, + "learning_rate": 9.189629633852799e-05, + "loss": 0.6746, + "step": 6206 + }, + { + "epoch": 0.41663031441897924, + "grad_norm": 1.2322992086410522, + "learning_rate": 9.189036360139773e-05, + "loss": 0.6422, + "step": 6208 + }, + { + "epoch": 0.4167645381027482, + "grad_norm": 1.1882107257843018, + "learning_rate": 9.188442888503018e-05, + "loss": 0.6634, + "step": 6210 + }, + { + "epoch": 0.41689876178651725, + "grad_norm": 1.0807157754898071, + "learning_rate": 9.187849218970572e-05, + "loss": 0.7188, + "step": 6212 + }, + { + "epoch": 0.4170329854702862, + "grad_norm": 1.0926183462142944, + "learning_rate": 9.187255351570487e-05, + "loss": 0.6573, + "step": 6214 + }, + { + "epoch": 0.41716720915405525, + "grad_norm": 1.1089380979537964, + "learning_rate": 9.18666128633082e-05, + "loss": 0.6824, + "step": 6216 + }, + { + "epoch": 0.4173014328378242, + "grad_norm": 1.1253862380981445, + "learning_rate": 9.186067023279639e-05, + "loss": 0.6919, + "step": 6218 + }, + { + "epoch": 0.41743565652159326, + "grad_norm": 0.9917927384376526, + "learning_rate": 9.185472562445022e-05, + "loss": 0.6455, + "step": 6220 + }, + { + "epoch": 0.41756988020536223, + "grad_norm": 1.724709153175354, + "learning_rate": 9.184877903855058e-05, + "loss": 0.7087, + "step": 6222 + }, + { + "epoch": 0.41770410388913126, + "grad_norm": 1.1316773891448975, + "learning_rate": 9.184283047537843e-05, + "loss": 0.6598, + "step": 6224 + }, + { + "epoch": 0.41783832757290024, + "grad_norm": 0.9227393865585327, + "learning_rate": 9.18368799352148e-05, + "loss": 0.6334, + "step": 6226 + }, + { + "epoch": 0.41797255125666927, + "grad_norm": 1.0776231288909912, + "learning_rate": 9.183092741834087e-05, + "loss": 0.693, + "step": 6228 + }, + { + "epoch": 0.41810677494043824, + "grad_norm": 1.0032908916473389, + "learning_rate": 9.182497292503789e-05, + "loss": 0.6563, + "step": 6230 + }, + { + "epoch": 0.4182409986242072, + "grad_norm": 1.2029250860214233, + "learning_rate": 9.181901645558717e-05, + "loss": 0.6592, + "step": 6232 + }, + { + "epoch": 0.41837522230797625, + "grad_norm": 1.0548990964889526, + "learning_rate": 9.181305801027015e-05, + "loss": 0.6138, + "step": 6234 + }, + { + "epoch": 0.4185094459917452, + "grad_norm": 0.9585376977920532, + "learning_rate": 9.180709758936839e-05, + "loss": 0.6225, + "step": 6236 + }, + { + "epoch": 0.41864366967551425, + "grad_norm": 1.2465183734893799, + "learning_rate": 9.180113519316345e-05, + "loss": 0.7468, + "step": 6238 + }, + { + "epoch": 0.4187778933592832, + "grad_norm": 1.2814899682998657, + "learning_rate": 9.179517082193709e-05, + "loss": 0.6708, + "step": 6240 + }, + { + "epoch": 0.41891211704305226, + "grad_norm": 1.0063660144805908, + "learning_rate": 9.178920447597108e-05, + "loss": 0.6204, + "step": 6242 + }, + { + "epoch": 0.41904634072682123, + "grad_norm": 1.1164515018463135, + "learning_rate": 9.178323615554733e-05, + "loss": 0.6057, + "step": 6244 + }, + { + "epoch": 0.41918056441059026, + "grad_norm": 0.9850273132324219, + "learning_rate": 9.177726586094785e-05, + "loss": 0.6054, + "step": 6246 + }, + { + "epoch": 0.41931478809435924, + "grad_norm": 1.115423321723938, + "learning_rate": 9.177129359245471e-05, + "loss": 0.6518, + "step": 6248 + }, + { + "epoch": 0.41944901177812827, + "grad_norm": 1.0860360860824585, + "learning_rate": 9.176531935035009e-05, + "loss": 0.6096, + "step": 6250 + }, + { + "epoch": 0.41958323546189724, + "grad_norm": 1.2603168487548828, + "learning_rate": 9.175934313491625e-05, + "loss": 0.7101, + "step": 6252 + }, + { + "epoch": 0.41971745914566627, + "grad_norm": 1.0043116807937622, + "learning_rate": 9.175336494643557e-05, + "loss": 0.7133, + "step": 6254 + }, + { + "epoch": 0.41985168282943525, + "grad_norm": 1.0733429193496704, + "learning_rate": 9.174738478519047e-05, + "loss": 0.6137, + "step": 6256 + }, + { + "epoch": 0.4199859065132043, + "grad_norm": 1.2858965396881104, + "learning_rate": 9.174140265146356e-05, + "loss": 0.6648, + "step": 6258 + }, + { + "epoch": 0.42012013019697325, + "grad_norm": 1.0367754697799683, + "learning_rate": 9.173541854553745e-05, + "loss": 0.6619, + "step": 6260 + }, + { + "epoch": 0.4202543538807423, + "grad_norm": 1.035481333732605, + "learning_rate": 9.172943246769489e-05, + "loss": 0.7092, + "step": 6262 + }, + { + "epoch": 0.42038857756451126, + "grad_norm": 0.9812489151954651, + "learning_rate": 9.17234444182187e-05, + "loss": 0.6362, + "step": 6264 + }, + { + "epoch": 0.42052280124828023, + "grad_norm": 1.0386441946029663, + "learning_rate": 9.17174543973918e-05, + "loss": 0.6318, + "step": 6266 + }, + { + "epoch": 0.42065702493204926, + "grad_norm": 1.134960651397705, + "learning_rate": 9.171146240549722e-05, + "loss": 0.7133, + "step": 6268 + }, + { + "epoch": 0.42079124861581824, + "grad_norm": 1.1966896057128906, + "learning_rate": 9.170546844281807e-05, + "loss": 0.7314, + "step": 6270 + }, + { + "epoch": 0.42092547229958727, + "grad_norm": 0.960020124912262, + "learning_rate": 9.169947250963753e-05, + "loss": 0.6676, + "step": 6272 + }, + { + "epoch": 0.42105969598335624, + "grad_norm": 1.9024641513824463, + "learning_rate": 9.169347460623892e-05, + "loss": 0.598, + "step": 6274 + }, + { + "epoch": 0.42119391966712527, + "grad_norm": 1.1077617406845093, + "learning_rate": 9.168747473290562e-05, + "loss": 0.7323, + "step": 6276 + }, + { + "epoch": 0.42132814335089425, + "grad_norm": 1.0826691389083862, + "learning_rate": 9.168147288992112e-05, + "loss": 0.7214, + "step": 6278 + }, + { + "epoch": 0.4214623670346633, + "grad_norm": 1.00456702709198, + "learning_rate": 9.167546907756898e-05, + "loss": 0.6772, + "step": 6280 + }, + { + "epoch": 0.42159659071843225, + "grad_norm": 1.0859038829803467, + "learning_rate": 9.166946329613288e-05, + "loss": 0.6402, + "step": 6282 + }, + { + "epoch": 0.4217308144022013, + "grad_norm": 1.00326669216156, + "learning_rate": 9.166345554589658e-05, + "loss": 0.5797, + "step": 6284 + }, + { + "epoch": 0.42186503808597026, + "grad_norm": 1.0553573369979858, + "learning_rate": 9.165744582714393e-05, + "loss": 0.6244, + "step": 6286 + }, + { + "epoch": 0.4219992617697393, + "grad_norm": 1.1167014837265015, + "learning_rate": 9.165143414015889e-05, + "loss": 0.7148, + "step": 6288 + }, + { + "epoch": 0.42213348545350826, + "grad_norm": 1.0160471200942993, + "learning_rate": 9.164542048522549e-05, + "loss": 0.6077, + "step": 6290 + }, + { + "epoch": 0.4222677091372773, + "grad_norm": 0.9857290983200073, + "learning_rate": 9.163940486262785e-05, + "loss": 0.6954, + "step": 6292 + }, + { + "epoch": 0.42240193282104627, + "grad_norm": 1.2125552892684937, + "learning_rate": 9.163338727265022e-05, + "loss": 0.6373, + "step": 6294 + }, + { + "epoch": 0.4225361565048153, + "grad_norm": 1.0877693891525269, + "learning_rate": 9.162736771557692e-05, + "loss": 0.7392, + "step": 6296 + }, + { + "epoch": 0.42267038018858427, + "grad_norm": 1.0825779438018799, + "learning_rate": 9.162134619169233e-05, + "loss": 0.6596, + "step": 6298 + }, + { + "epoch": 0.4228046038723533, + "grad_norm": 1.0345951318740845, + "learning_rate": 9.161532270128099e-05, + "loss": 0.6746, + "step": 6300 + }, + { + "epoch": 0.4229388275561223, + "grad_norm": 1.1396832466125488, + "learning_rate": 9.16092972446275e-05, + "loss": 0.7129, + "step": 6302 + }, + { + "epoch": 0.42307305123989125, + "grad_norm": 0.9923988580703735, + "learning_rate": 9.160326982201652e-05, + "loss": 0.6696, + "step": 6304 + }, + { + "epoch": 0.4232072749236603, + "grad_norm": 1.2008332014083862, + "learning_rate": 9.159724043373284e-05, + "loss": 0.6764, + "step": 6306 + }, + { + "epoch": 0.42334149860742926, + "grad_norm": 1.0264410972595215, + "learning_rate": 9.159120908006135e-05, + "loss": 0.6352, + "step": 6308 + }, + { + "epoch": 0.4234757222911983, + "grad_norm": 1.1287544965744019, + "learning_rate": 9.158517576128705e-05, + "loss": 0.642, + "step": 6310 + }, + { + "epoch": 0.42360994597496726, + "grad_norm": 1.1886465549468994, + "learning_rate": 9.157914047769493e-05, + "loss": 0.6993, + "step": 6312 + }, + { + "epoch": 0.4237441696587363, + "grad_norm": 1.4661725759506226, + "learning_rate": 9.15731032295702e-05, + "loss": 0.6129, + "step": 6314 + }, + { + "epoch": 0.42387839334250527, + "grad_norm": 1.1720086336135864, + "learning_rate": 9.15670640171981e-05, + "loss": 0.6844, + "step": 6316 + }, + { + "epoch": 0.4240126170262743, + "grad_norm": 2.242379903793335, + "learning_rate": 9.156102284086394e-05, + "loss": 0.7241, + "step": 6318 + }, + { + "epoch": 0.42414684071004327, + "grad_norm": 1.0734869241714478, + "learning_rate": 9.15549797008532e-05, + "loss": 0.6562, + "step": 6320 + }, + { + "epoch": 0.4242810643938123, + "grad_norm": 1.0534396171569824, + "learning_rate": 9.154893459745138e-05, + "loss": 0.6223, + "step": 6322 + }, + { + "epoch": 0.4244152880775813, + "grad_norm": 0.8479713201522827, + "learning_rate": 9.154288753094408e-05, + "loss": 0.5982, + "step": 6324 + }, + { + "epoch": 0.4245495117613503, + "grad_norm": 1.071876049041748, + "learning_rate": 9.153683850161706e-05, + "loss": 0.6977, + "step": 6326 + }, + { + "epoch": 0.4246837354451193, + "grad_norm": 1.4902291297912598, + "learning_rate": 9.15307875097561e-05, + "loss": 0.6598, + "step": 6328 + }, + { + "epoch": 0.4248179591288883, + "grad_norm": 1.0573935508728027, + "learning_rate": 9.152473455564708e-05, + "loss": 0.7148, + "step": 6330 + }, + { + "epoch": 0.4249521828126573, + "grad_norm": 1.0465713739395142, + "learning_rate": 9.151867963957601e-05, + "loss": 0.6745, + "step": 6332 + }, + { + "epoch": 0.4250864064964263, + "grad_norm": 1.1158103942871094, + "learning_rate": 9.151262276182898e-05, + "loss": 0.6817, + "step": 6334 + }, + { + "epoch": 0.4252206301801953, + "grad_norm": 1.0501621961593628, + "learning_rate": 9.150656392269215e-05, + "loss": 0.7074, + "step": 6336 + }, + { + "epoch": 0.4253548538639643, + "grad_norm": 1.1173028945922852, + "learning_rate": 9.15005031224518e-05, + "loss": 0.7018, + "step": 6338 + }, + { + "epoch": 0.4254890775477333, + "grad_norm": 1.1124805212020874, + "learning_rate": 9.149444036139427e-05, + "loss": 0.6565, + "step": 6340 + }, + { + "epoch": 0.42562330123150227, + "grad_norm": 1.1531590223312378, + "learning_rate": 9.148837563980606e-05, + "loss": 0.6702, + "step": 6342 + }, + { + "epoch": 0.4257575249152713, + "grad_norm": 1.3122427463531494, + "learning_rate": 9.148230895797366e-05, + "loss": 0.6935, + "step": 6344 + }, + { + "epoch": 0.4258917485990403, + "grad_norm": 1.274470329284668, + "learning_rate": 9.147624031618373e-05, + "loss": 0.6396, + "step": 6346 + }, + { + "epoch": 0.4260259722828093, + "grad_norm": 1.0647002458572388, + "learning_rate": 9.147016971472299e-05, + "loss": 0.6394, + "step": 6348 + }, + { + "epoch": 0.4261601959665783, + "grad_norm": 2.0520029067993164, + "learning_rate": 9.146409715387832e-05, + "loss": 0.6615, + "step": 6350 + }, + { + "epoch": 0.4262944196503473, + "grad_norm": 1.0095345973968506, + "learning_rate": 9.145802263393657e-05, + "loss": 0.5958, + "step": 6352 + }, + { + "epoch": 0.4264286433341163, + "grad_norm": 1.1976794004440308, + "learning_rate": 9.145194615518477e-05, + "loss": 0.6894, + "step": 6354 + }, + { + "epoch": 0.4265628670178853, + "grad_norm": 1.1440362930297852, + "learning_rate": 9.144586771791003e-05, + "loss": 0.6281, + "step": 6356 + }, + { + "epoch": 0.4266970907016543, + "grad_norm": 1.0716235637664795, + "learning_rate": 9.143978732239955e-05, + "loss": 0.7333, + "step": 6358 + }, + { + "epoch": 0.4268313143854233, + "grad_norm": 1.2851024866104126, + "learning_rate": 9.143370496894061e-05, + "loss": 0.7129, + "step": 6360 + }, + { + "epoch": 0.4269655380691923, + "grad_norm": 1.1375380754470825, + "learning_rate": 9.142762065782058e-05, + "loss": 0.658, + "step": 6362 + }, + { + "epoch": 0.4270997617529613, + "grad_norm": 0.9954057931900024, + "learning_rate": 9.142153438932693e-05, + "loss": 0.657, + "step": 6364 + }, + { + "epoch": 0.4272339854367303, + "grad_norm": 0.9236775636672974, + "learning_rate": 9.141544616374724e-05, + "loss": 0.6579, + "step": 6366 + }, + { + "epoch": 0.42736820912049933, + "grad_norm": 1.0957763195037842, + "learning_rate": 9.140935598136914e-05, + "loss": 0.7027, + "step": 6368 + }, + { + "epoch": 0.4275024328042683, + "grad_norm": 1.032928705215454, + "learning_rate": 9.140326384248042e-05, + "loss": 0.7045, + "step": 6370 + }, + { + "epoch": 0.42763665648803734, + "grad_norm": 1.0739259719848633, + "learning_rate": 9.139716974736889e-05, + "loss": 0.6771, + "step": 6372 + }, + { + "epoch": 0.4277708801718063, + "grad_norm": 1.215862512588501, + "learning_rate": 9.13910736963225e-05, + "loss": 0.6626, + "step": 6374 + }, + { + "epoch": 0.42790510385557534, + "grad_norm": 1.0815331935882568, + "learning_rate": 9.138497568962927e-05, + "loss": 0.6489, + "step": 6376 + }, + { + "epoch": 0.4280393275393443, + "grad_norm": 0.9747056365013123, + "learning_rate": 9.137887572757732e-05, + "loss": 0.6308, + "step": 6378 + }, + { + "epoch": 0.4281735512231133, + "grad_norm": 1.240158200263977, + "learning_rate": 9.137277381045486e-05, + "loss": 0.5727, + "step": 6380 + }, + { + "epoch": 0.4283077749068823, + "grad_norm": 1.1604822874069214, + "learning_rate": 9.136666993855018e-05, + "loss": 0.6793, + "step": 6382 + }, + { + "epoch": 0.4284419985906513, + "grad_norm": 1.0272921323776245, + "learning_rate": 9.13605641121517e-05, + "loss": 0.6959, + "step": 6384 + }, + { + "epoch": 0.4285762222744203, + "grad_norm": 1.115839958190918, + "learning_rate": 9.135445633154789e-05, + "loss": 0.684, + "step": 6386 + }, + { + "epoch": 0.4287104459581893, + "grad_norm": 1.3776438236236572, + "learning_rate": 9.134834659702736e-05, + "loss": 0.7394, + "step": 6388 + }, + { + "epoch": 0.42884466964195833, + "grad_norm": 1.1234407424926758, + "learning_rate": 9.134223490887875e-05, + "loss": 0.6542, + "step": 6390 + }, + { + "epoch": 0.4289788933257273, + "grad_norm": 1.107689619064331, + "learning_rate": 9.133612126739082e-05, + "loss": 0.6801, + "step": 6392 + }, + { + "epoch": 0.42911311700949634, + "grad_norm": 1.1047486066818237, + "learning_rate": 9.133000567285245e-05, + "loss": 0.6355, + "step": 6394 + }, + { + "epoch": 0.4292473406932653, + "grad_norm": 1.3371986150741577, + "learning_rate": 9.13238881255526e-05, + "loss": 0.7247, + "step": 6396 + }, + { + "epoch": 0.42938156437703434, + "grad_norm": 0.9681903719902039, + "learning_rate": 9.131776862578027e-05, + "loss": 0.6958, + "step": 6398 + }, + { + "epoch": 0.4295157880608033, + "grad_norm": 1.271085262298584, + "learning_rate": 9.131164717382466e-05, + "loss": 0.7858, + "step": 6400 + }, + { + "epoch": 0.42965001174457235, + "grad_norm": 1.1457453966140747, + "learning_rate": 9.130552376997492e-05, + "loss": 0.6198, + "step": 6402 + }, + { + "epoch": 0.4297842354283413, + "grad_norm": 1.0122690200805664, + "learning_rate": 9.129939841452042e-05, + "loss": 0.652, + "step": 6404 + }, + { + "epoch": 0.42991845911211035, + "grad_norm": 0.8656542301177979, + "learning_rate": 9.129327110775056e-05, + "loss": 0.5501, + "step": 6406 + }, + { + "epoch": 0.4300526827958793, + "grad_norm": 1.0972250699996948, + "learning_rate": 9.128714184995483e-05, + "loss": 0.62, + "step": 6408 + }, + { + "epoch": 0.43018690647964836, + "grad_norm": 1.1215505599975586, + "learning_rate": 9.128101064142285e-05, + "loss": 0.6519, + "step": 6410 + }, + { + "epoch": 0.43032113016341733, + "grad_norm": 1.1611055135726929, + "learning_rate": 9.127487748244427e-05, + "loss": 0.7007, + "step": 6412 + }, + { + "epoch": 0.43045535384718636, + "grad_norm": 1.1495709419250488, + "learning_rate": 9.126874237330891e-05, + "loss": 0.7069, + "step": 6414 + }, + { + "epoch": 0.43058957753095534, + "grad_norm": 1.1690107583999634, + "learning_rate": 9.126260531430662e-05, + "loss": 0.6409, + "step": 6416 + }, + { + "epoch": 0.4307238012147243, + "grad_norm": 1.081002116203308, + "learning_rate": 9.125646630572737e-05, + "loss": 0.6799, + "step": 6418 + }, + { + "epoch": 0.43085802489849334, + "grad_norm": 1.0002063512802124, + "learning_rate": 9.125032534786122e-05, + "loss": 0.6398, + "step": 6420 + }, + { + "epoch": 0.4309922485822623, + "grad_norm": 1.1421399116516113, + "learning_rate": 9.12441824409983e-05, + "loss": 0.6651, + "step": 6422 + }, + { + "epoch": 0.43112647226603135, + "grad_norm": 0.9874308109283447, + "learning_rate": 9.123803758542888e-05, + "loss": 0.5975, + "step": 6424 + }, + { + "epoch": 0.4312606959498003, + "grad_norm": 1.8552358150482178, + "learning_rate": 9.123189078144326e-05, + "loss": 0.6139, + "step": 6426 + }, + { + "epoch": 0.43139491963356935, + "grad_norm": 1.4092941284179688, + "learning_rate": 9.122574202933188e-05, + "loss": 0.7107, + "step": 6428 + }, + { + "epoch": 0.4315291433173383, + "grad_norm": 1.1505235433578491, + "learning_rate": 9.121959132938527e-05, + "loss": 0.6559, + "step": 6430 + }, + { + "epoch": 0.43166336700110736, + "grad_norm": 0.9634313583374023, + "learning_rate": 9.1213438681894e-05, + "loss": 0.6344, + "step": 6432 + }, + { + "epoch": 0.43179759068487633, + "grad_norm": 0.8595042824745178, + "learning_rate": 9.120728408714882e-05, + "loss": 0.6284, + "step": 6434 + }, + { + "epoch": 0.43193181436864536, + "grad_norm": 0.9590168595314026, + "learning_rate": 9.120112754544047e-05, + "loss": 0.654, + "step": 6436 + }, + { + "epoch": 0.43206603805241434, + "grad_norm": 0.9372844099998474, + "learning_rate": 9.119496905705989e-05, + "loss": 0.5882, + "step": 6438 + }, + { + "epoch": 0.43220026173618337, + "grad_norm": 1.245493769645691, + "learning_rate": 9.118880862229802e-05, + "loss": 0.7153, + "step": 6440 + }, + { + "epoch": 0.43233448541995234, + "grad_norm": 1.0889474153518677, + "learning_rate": 9.118264624144594e-05, + "loss": 0.6479, + "step": 6442 + }, + { + "epoch": 0.43246870910372137, + "grad_norm": 1.0484486818313599, + "learning_rate": 9.11764819147948e-05, + "loss": 0.6561, + "step": 6444 + }, + { + "epoch": 0.43260293278749035, + "grad_norm": 1.3504307270050049, + "learning_rate": 9.117031564263584e-05, + "loss": 0.7183, + "step": 6446 + }, + { + "epoch": 0.4327371564712594, + "grad_norm": 1.2466015815734863, + "learning_rate": 9.116414742526047e-05, + "loss": 0.6879, + "step": 6448 + }, + { + "epoch": 0.43287138015502835, + "grad_norm": 1.1774563789367676, + "learning_rate": 9.115797726296004e-05, + "loss": 0.6828, + "step": 6450 + }, + { + "epoch": 0.4330056038387974, + "grad_norm": 0.9712786674499512, + "learning_rate": 9.115180515602614e-05, + "loss": 0.6084, + "step": 6452 + }, + { + "epoch": 0.43313982752256636, + "grad_norm": 1.1016490459442139, + "learning_rate": 9.114563110475036e-05, + "loss": 0.6523, + "step": 6454 + }, + { + "epoch": 0.43327405120633533, + "grad_norm": 1.0148249864578247, + "learning_rate": 9.113945510942443e-05, + "loss": 0.6723, + "step": 6456 + }, + { + "epoch": 0.43340827489010436, + "grad_norm": 0.9622500538825989, + "learning_rate": 9.113327717034013e-05, + "loss": 0.6633, + "step": 6458 + }, + { + "epoch": 0.43354249857387334, + "grad_norm": 1.0464283227920532, + "learning_rate": 9.112709728778937e-05, + "loss": 0.6835, + "step": 6460 + }, + { + "epoch": 0.43367672225764237, + "grad_norm": 1.1165943145751953, + "learning_rate": 9.112091546206414e-05, + "loss": 0.5918, + "step": 6462 + }, + { + "epoch": 0.43381094594141134, + "grad_norm": 1.076734185218811, + "learning_rate": 9.111473169345652e-05, + "loss": 0.6961, + "step": 6464 + }, + { + "epoch": 0.43394516962518037, + "grad_norm": 1.0434982776641846, + "learning_rate": 9.110854598225867e-05, + "loss": 0.638, + "step": 6466 + }, + { + "epoch": 0.43407939330894935, + "grad_norm": 1.0204663276672363, + "learning_rate": 9.110235832876286e-05, + "loss": 0.6514, + "step": 6468 + }, + { + "epoch": 0.4342136169927184, + "grad_norm": 1.104552984237671, + "learning_rate": 9.109616873326144e-05, + "loss": 0.6957, + "step": 6470 + }, + { + "epoch": 0.43434784067648735, + "grad_norm": 1.1974577903747559, + "learning_rate": 9.108997719604687e-05, + "loss": 0.6785, + "step": 6472 + }, + { + "epoch": 0.4344820643602564, + "grad_norm": 0.9422672986984253, + "learning_rate": 9.108378371741167e-05, + "loss": 0.6053, + "step": 6474 + }, + { + "epoch": 0.43461628804402536, + "grad_norm": 1.0764330625534058, + "learning_rate": 9.107758829764848e-05, + "loss": 0.6145, + "step": 6476 + }, + { + "epoch": 0.4347505117277944, + "grad_norm": 0.8954029679298401, + "learning_rate": 9.107139093705001e-05, + "loss": 0.6287, + "step": 6478 + }, + { + "epoch": 0.43488473541156336, + "grad_norm": 1.232624888420105, + "learning_rate": 9.106519163590907e-05, + "loss": 0.7467, + "step": 6480 + }, + { + "epoch": 0.4350189590953324, + "grad_norm": 1.0775651931762695, + "learning_rate": 9.10589903945186e-05, + "loss": 0.6934, + "step": 6482 + }, + { + "epoch": 0.43515318277910137, + "grad_norm": 0.9535127282142639, + "learning_rate": 9.105278721317157e-05, + "loss": 0.6415, + "step": 6484 + }, + { + "epoch": 0.4352874064628704, + "grad_norm": 1.1277869939804077, + "learning_rate": 9.104658209216108e-05, + "loss": 0.7023, + "step": 6486 + }, + { + "epoch": 0.43542163014663937, + "grad_norm": 1.0773403644561768, + "learning_rate": 9.104037503178028e-05, + "loss": 0.6835, + "step": 6488 + }, + { + "epoch": 0.4355558538304084, + "grad_norm": 1.0601844787597656, + "learning_rate": 9.103416603232246e-05, + "loss": 0.6951, + "step": 6490 + }, + { + "epoch": 0.4356900775141774, + "grad_norm": 0.9843427538871765, + "learning_rate": 9.102795509408099e-05, + "loss": 0.6212, + "step": 6492 + }, + { + "epoch": 0.43582430119794635, + "grad_norm": 1.3423608541488647, + "learning_rate": 9.102174221734934e-05, + "loss": 0.7103, + "step": 6494 + }, + { + "epoch": 0.4359585248817154, + "grad_norm": 0.9906916618347168, + "learning_rate": 9.101552740242102e-05, + "loss": 0.6194, + "step": 6496 + }, + { + "epoch": 0.43609274856548436, + "grad_norm": 1.5656930208206177, + "learning_rate": 9.100931064958968e-05, + "loss": 0.6622, + "step": 6498 + }, + { + "epoch": 0.4362269722492534, + "grad_norm": 1.0899085998535156, + "learning_rate": 9.100309195914907e-05, + "loss": 0.6361, + "step": 6500 + }, + { + "epoch": 0.43636119593302236, + "grad_norm": 0.9742255806922913, + "learning_rate": 9.099687133139298e-05, + "loss": 0.5813, + "step": 6502 + }, + { + "epoch": 0.4364954196167914, + "grad_norm": 1.050657033920288, + "learning_rate": 9.099064876661533e-05, + "loss": 0.6577, + "step": 6504 + }, + { + "epoch": 0.43662964330056037, + "grad_norm": 1.137790560722351, + "learning_rate": 9.098442426511014e-05, + "loss": 0.6901, + "step": 6506 + }, + { + "epoch": 0.4367638669843294, + "grad_norm": 1.0315080881118774, + "learning_rate": 9.09781978271715e-05, + "loss": 0.6881, + "step": 6508 + }, + { + "epoch": 0.43689809066809837, + "grad_norm": 0.9549081325531006, + "learning_rate": 9.097196945309359e-05, + "loss": 0.6294, + "step": 6510 + }, + { + "epoch": 0.4370323143518674, + "grad_norm": 1.1911718845367432, + "learning_rate": 9.096573914317068e-05, + "loss": 0.6654, + "step": 6512 + }, + { + "epoch": 0.4371665380356364, + "grad_norm": 1.0279988050460815, + "learning_rate": 9.095950689769716e-05, + "loss": 0.6934, + "step": 6514 + }, + { + "epoch": 0.4373007617194054, + "grad_norm": 1.69241201877594, + "learning_rate": 9.095327271696749e-05, + "loss": 0.6478, + "step": 6516 + }, + { + "epoch": 0.4374349854031744, + "grad_norm": 1.1771697998046875, + "learning_rate": 9.094703660127622e-05, + "loss": 0.6668, + "step": 6518 + }, + { + "epoch": 0.4375692090869434, + "grad_norm": 1.045859932899475, + "learning_rate": 9.094079855091797e-05, + "loss": 0.7428, + "step": 6520 + }, + { + "epoch": 0.4377034327707124, + "grad_norm": 1.130373239517212, + "learning_rate": 9.09345585661875e-05, + "loss": 0.6426, + "step": 6522 + }, + { + "epoch": 0.4378376564544814, + "grad_norm": 1.051623821258545, + "learning_rate": 9.092831664737964e-05, + "loss": 0.7218, + "step": 6524 + }, + { + "epoch": 0.4379718801382504, + "grad_norm": 1.0282305479049683, + "learning_rate": 9.092207279478929e-05, + "loss": 0.6457, + "step": 6526 + }, + { + "epoch": 0.4381061038220194, + "grad_norm": 1.3251056671142578, + "learning_rate": 9.091582700871148e-05, + "loss": 0.6205, + "step": 6528 + }, + { + "epoch": 0.4382403275057884, + "grad_norm": 1.111969232559204, + "learning_rate": 9.090957928944129e-05, + "loss": 0.6603, + "step": 6530 + }, + { + "epoch": 0.43837455118955737, + "grad_norm": 1.0762337446212769, + "learning_rate": 9.090332963727393e-05, + "loss": 0.6861, + "step": 6532 + }, + { + "epoch": 0.4385087748733264, + "grad_norm": 1.1892808675765991, + "learning_rate": 9.089707805250468e-05, + "loss": 0.6627, + "step": 6534 + }, + { + "epoch": 0.4386429985570954, + "grad_norm": 0.946892499923706, + "learning_rate": 9.089082453542891e-05, + "loss": 0.6566, + "step": 6536 + }, + { + "epoch": 0.4387772222408644, + "grad_norm": 1.0272513628005981, + "learning_rate": 9.088456908634209e-05, + "loss": 0.6747, + "step": 6538 + }, + { + "epoch": 0.4389114459246334, + "grad_norm": 1.255175232887268, + "learning_rate": 9.087831170553978e-05, + "loss": 0.6437, + "step": 6540 + }, + { + "epoch": 0.4390456696084024, + "grad_norm": 0.9205337166786194, + "learning_rate": 9.087205239331762e-05, + "loss": 0.6389, + "step": 6542 + }, + { + "epoch": 0.4391798932921714, + "grad_norm": 1.233696699142456, + "learning_rate": 9.086579114997136e-05, + "loss": 0.6947, + "step": 6544 + }, + { + "epoch": 0.4393141169759404, + "grad_norm": 1.099592924118042, + "learning_rate": 9.085952797579682e-05, + "loss": 0.6951, + "step": 6546 + }, + { + "epoch": 0.4394483406597094, + "grad_norm": 1.2203367948532104, + "learning_rate": 9.085326287108995e-05, + "loss": 0.7214, + "step": 6548 + }, + { + "epoch": 0.4395825643434784, + "grad_norm": 1.1897811889648438, + "learning_rate": 9.084699583614673e-05, + "loss": 0.6923, + "step": 6550 + }, + { + "epoch": 0.4397167880272474, + "grad_norm": 1.0716934204101562, + "learning_rate": 9.084072687126327e-05, + "loss": 0.7315, + "step": 6552 + }, + { + "epoch": 0.4398510117110164, + "grad_norm": 1.0627871751785278, + "learning_rate": 9.083445597673578e-05, + "loss": 0.602, + "step": 6554 + }, + { + "epoch": 0.4399852353947854, + "grad_norm": 0.977236270904541, + "learning_rate": 9.082818315286055e-05, + "loss": 0.6165, + "step": 6556 + }, + { + "epoch": 0.44011945907855443, + "grad_norm": 0.9908711910247803, + "learning_rate": 9.082190839993395e-05, + "loss": 0.6062, + "step": 6558 + }, + { + "epoch": 0.4402536827623234, + "grad_norm": 1.441779613494873, + "learning_rate": 9.081563171825245e-05, + "loss": 0.6575, + "step": 6560 + }, + { + "epoch": 0.44038790644609244, + "grad_norm": 0.9981227517127991, + "learning_rate": 9.08093531081126e-05, + "loss": 0.686, + "step": 6562 + }, + { + "epoch": 0.4405221301298614, + "grad_norm": 1.0307862758636475, + "learning_rate": 9.080307256981109e-05, + "loss": 0.6502, + "step": 6564 + }, + { + "epoch": 0.44065635381363044, + "grad_norm": 1.0696779489517212, + "learning_rate": 9.079679010364461e-05, + "loss": 0.6779, + "step": 6566 + }, + { + "epoch": 0.4407905774973994, + "grad_norm": 0.9596073627471924, + "learning_rate": 9.079050570991004e-05, + "loss": 0.6073, + "step": 6568 + }, + { + "epoch": 0.4409248011811684, + "grad_norm": 1.1085256338119507, + "learning_rate": 9.078421938890426e-05, + "loss": 0.623, + "step": 6570 + }, + { + "epoch": 0.4410590248649374, + "grad_norm": 0.9325559139251709, + "learning_rate": 9.077793114092435e-05, + "loss": 0.6415, + "step": 6572 + }, + { + "epoch": 0.4411932485487064, + "grad_norm": 1.551746129989624, + "learning_rate": 9.077164096626736e-05, + "loss": 0.6947, + "step": 6574 + }, + { + "epoch": 0.4413274722324754, + "grad_norm": 1.1572011709213257, + "learning_rate": 9.07653488652305e-05, + "loss": 0.6643, + "step": 6576 + }, + { + "epoch": 0.4414616959162444, + "grad_norm": 1.0599066019058228, + "learning_rate": 9.075905483811109e-05, + "loss": 0.6904, + "step": 6578 + }, + { + "epoch": 0.44159591960001343, + "grad_norm": 1.0249269008636475, + "learning_rate": 9.075275888520647e-05, + "loss": 0.617, + "step": 6580 + }, + { + "epoch": 0.4417301432837824, + "grad_norm": 1.1034280061721802, + "learning_rate": 9.074646100681413e-05, + "loss": 0.6757, + "step": 6582 + }, + { + "epoch": 0.44186436696755144, + "grad_norm": 0.9876016974449158, + "learning_rate": 9.074016120323163e-05, + "loss": 0.632, + "step": 6584 + }, + { + "epoch": 0.4419985906513204, + "grad_norm": 1.3852652311325073, + "learning_rate": 9.073385947475664e-05, + "loss": 0.6871, + "step": 6586 + }, + { + "epoch": 0.44213281433508944, + "grad_norm": 1.137299656867981, + "learning_rate": 9.072755582168688e-05, + "loss": 0.6558, + "step": 6588 + }, + { + "epoch": 0.4422670380188584, + "grad_norm": 0.9819977283477783, + "learning_rate": 9.07212502443202e-05, + "loss": 0.6194, + "step": 6590 + }, + { + "epoch": 0.44240126170262745, + "grad_norm": 1.3733234405517578, + "learning_rate": 9.071494274295452e-05, + "loss": 0.6471, + "step": 6592 + }, + { + "epoch": 0.4425354853863964, + "grad_norm": 1.1099233627319336, + "learning_rate": 9.070863331788785e-05, + "loss": 0.679, + "step": 6594 + }, + { + "epoch": 0.44266970907016545, + "grad_norm": 1.0077110528945923, + "learning_rate": 9.07023219694183e-05, + "loss": 0.6723, + "step": 6596 + }, + { + "epoch": 0.4428039327539344, + "grad_norm": 0.9958982467651367, + "learning_rate": 9.06960086978441e-05, + "loss": 0.6577, + "step": 6598 + }, + { + "epoch": 0.44293815643770346, + "grad_norm": 1.4362531900405884, + "learning_rate": 9.068969350346349e-05, + "loss": 0.6941, + "step": 6600 + }, + { + "epoch": 0.44307238012147243, + "grad_norm": 2.014660358428955, + "learning_rate": 9.068337638657489e-05, + "loss": 0.6327, + "step": 6602 + }, + { + "epoch": 0.44320660380524146, + "grad_norm": 0.9608994126319885, + "learning_rate": 9.067705734747674e-05, + "loss": 0.6199, + "step": 6604 + }, + { + "epoch": 0.44334082748901044, + "grad_norm": 1.7143373489379883, + "learning_rate": 9.067073638646763e-05, + "loss": 0.7387, + "step": 6606 + }, + { + "epoch": 0.4434750511727794, + "grad_norm": 0.9786049723625183, + "learning_rate": 9.06644135038462e-05, + "loss": 0.6506, + "step": 6608 + }, + { + "epoch": 0.44360927485654844, + "grad_norm": 3.5571486949920654, + "learning_rate": 9.06580886999112e-05, + "loss": 0.6904, + "step": 6610 + }, + { + "epoch": 0.4437434985403174, + "grad_norm": 1.0379436016082764, + "learning_rate": 9.065176197496146e-05, + "loss": 0.6766, + "step": 6612 + }, + { + "epoch": 0.44387772222408645, + "grad_norm": 1.0329163074493408, + "learning_rate": 9.064543332929589e-05, + "loss": 0.6724, + "step": 6614 + }, + { + "epoch": 0.4440119459078554, + "grad_norm": 1.0833524465560913, + "learning_rate": 9.063910276321354e-05, + "loss": 0.6512, + "step": 6616 + }, + { + "epoch": 0.44414616959162445, + "grad_norm": 1.4965224266052246, + "learning_rate": 9.063277027701349e-05, + "loss": 0.6345, + "step": 6618 + }, + { + "epoch": 0.4442803932753934, + "grad_norm": 0.9639096856117249, + "learning_rate": 9.062643587099495e-05, + "loss": 0.6275, + "step": 6620 + }, + { + "epoch": 0.44441461695916246, + "grad_norm": 1.1125866174697876, + "learning_rate": 9.062009954545719e-05, + "loss": 0.5863, + "step": 6622 + }, + { + "epoch": 0.44454884064293143, + "grad_norm": 1.0979225635528564, + "learning_rate": 9.061376130069961e-05, + "loss": 0.696, + "step": 6624 + }, + { + "epoch": 0.44468306432670046, + "grad_norm": 0.9352512359619141, + "learning_rate": 9.060742113702168e-05, + "loss": 0.6487, + "step": 6626 + }, + { + "epoch": 0.44481728801046944, + "grad_norm": 1.1315749883651733, + "learning_rate": 9.060107905472294e-05, + "loss": 0.7242, + "step": 6628 + }, + { + "epoch": 0.44495151169423847, + "grad_norm": 1.142823338508606, + "learning_rate": 9.059473505410305e-05, + "loss": 0.6645, + "step": 6630 + }, + { + "epoch": 0.44508573537800744, + "grad_norm": 1.253380298614502, + "learning_rate": 9.058838913546178e-05, + "loss": 0.6615, + "step": 6632 + }, + { + "epoch": 0.44521995906177647, + "grad_norm": 1.1333796977996826, + "learning_rate": 9.058204129909891e-05, + "loss": 0.6809, + "step": 6634 + }, + { + "epoch": 0.44535418274554545, + "grad_norm": 1.2247251272201538, + "learning_rate": 9.05756915453144e-05, + "loss": 0.6851, + "step": 6636 + }, + { + "epoch": 0.4454884064293145, + "grad_norm": 1.5173280239105225, + "learning_rate": 9.056933987440825e-05, + "loss": 0.7467, + "step": 6638 + }, + { + "epoch": 0.44562263011308345, + "grad_norm": 1.110931158065796, + "learning_rate": 9.056298628668056e-05, + "loss": 0.722, + "step": 6640 + }, + { + "epoch": 0.4457568537968525, + "grad_norm": 1.1217375993728638, + "learning_rate": 9.055663078243156e-05, + "loss": 0.6467, + "step": 6642 + }, + { + "epoch": 0.44589107748062146, + "grad_norm": 1.1028882265090942, + "learning_rate": 9.055027336196146e-05, + "loss": 0.6725, + "step": 6644 + }, + { + "epoch": 0.44602530116439043, + "grad_norm": 1.1858761310577393, + "learning_rate": 9.054391402557072e-05, + "loss": 0.6381, + "step": 6646 + }, + { + "epoch": 0.44615952484815946, + "grad_norm": 1.0766549110412598, + "learning_rate": 9.053755277355976e-05, + "loss": 0.6382, + "step": 6648 + }, + { + "epoch": 0.44629374853192844, + "grad_norm": 1.1438759565353394, + "learning_rate": 9.053118960622915e-05, + "loss": 0.7142, + "step": 6650 + }, + { + "epoch": 0.44642797221569747, + "grad_norm": 1.0107367038726807, + "learning_rate": 9.052482452387953e-05, + "loss": 0.5961, + "step": 6652 + }, + { + "epoch": 0.44656219589946644, + "grad_norm": 0.9405455589294434, + "learning_rate": 9.051845752681163e-05, + "loss": 0.6392, + "step": 6654 + }, + { + "epoch": 0.44669641958323547, + "grad_norm": 1.139909029006958, + "learning_rate": 9.051208861532629e-05, + "loss": 0.6692, + "step": 6656 + }, + { + "epoch": 0.44683064326700445, + "grad_norm": 1.0859870910644531, + "learning_rate": 9.050571778972443e-05, + "loss": 0.7422, + "step": 6658 + }, + { + "epoch": 0.4469648669507735, + "grad_norm": 1.1717115640640259, + "learning_rate": 9.049934505030705e-05, + "loss": 0.599, + "step": 6660 + }, + { + "epoch": 0.44709909063454245, + "grad_norm": 1.094762921333313, + "learning_rate": 9.049297039737528e-05, + "loss": 0.6316, + "step": 6662 + }, + { + "epoch": 0.4472333143183115, + "grad_norm": 1.4241257905960083, + "learning_rate": 9.048659383123026e-05, + "loss": 0.6412, + "step": 6664 + }, + { + "epoch": 0.44736753800208046, + "grad_norm": 1.7000561952590942, + "learning_rate": 9.04802153521733e-05, + "loss": 0.7028, + "step": 6666 + }, + { + "epoch": 0.4475017616858495, + "grad_norm": 1.068803310394287, + "learning_rate": 9.047383496050576e-05, + "loss": 0.6925, + "step": 6668 + }, + { + "epoch": 0.44763598536961846, + "grad_norm": 0.993427574634552, + "learning_rate": 9.046745265652912e-05, + "loss": 0.69, + "step": 6670 + }, + { + "epoch": 0.4477702090533875, + "grad_norm": 1.00606369972229, + "learning_rate": 9.046106844054491e-05, + "loss": 0.6426, + "step": 6672 + }, + { + "epoch": 0.44790443273715647, + "grad_norm": 1.0299243927001953, + "learning_rate": 9.045468231285477e-05, + "loss": 0.6346, + "step": 6674 + }, + { + "epoch": 0.4480386564209255, + "grad_norm": 1.0653523206710815, + "learning_rate": 9.044829427376046e-05, + "loss": 0.7017, + "step": 6676 + }, + { + "epoch": 0.44817288010469447, + "grad_norm": 1.4495142698287964, + "learning_rate": 9.044190432356377e-05, + "loss": 0.6282, + "step": 6678 + }, + { + "epoch": 0.4483071037884635, + "grad_norm": 1.0920124053955078, + "learning_rate": 9.043551246256664e-05, + "loss": 0.6573, + "step": 6680 + }, + { + "epoch": 0.4484413274722325, + "grad_norm": 0.9434648752212524, + "learning_rate": 9.042911869107105e-05, + "loss": 0.6084, + "step": 6682 + }, + { + "epoch": 0.44857555115600145, + "grad_norm": 1.13992440700531, + "learning_rate": 9.04227230093791e-05, + "loss": 0.671, + "step": 6684 + }, + { + "epoch": 0.4487097748397705, + "grad_norm": 0.9534379243850708, + "learning_rate": 9.041632541779298e-05, + "loss": 0.6019, + "step": 6686 + }, + { + "epoch": 0.44884399852353946, + "grad_norm": 1.0360051393508911, + "learning_rate": 9.040992591661495e-05, + "loss": 0.6658, + "step": 6688 + }, + { + "epoch": 0.4489782222073085, + "grad_norm": 1.1044501066207886, + "learning_rate": 9.04035245061474e-05, + "loss": 0.6831, + "step": 6690 + }, + { + "epoch": 0.44911244589107746, + "grad_norm": 1.1218234300613403, + "learning_rate": 9.039712118669276e-05, + "loss": 0.7078, + "step": 6692 + }, + { + "epoch": 0.4492466695748465, + "grad_norm": 1.112831711769104, + "learning_rate": 9.039071595855357e-05, + "loss": 0.7103, + "step": 6694 + }, + { + "epoch": 0.44938089325861547, + "grad_norm": 1.0213844776153564, + "learning_rate": 9.038430882203249e-05, + "loss": 0.6505, + "step": 6696 + }, + { + "epoch": 0.4495151169423845, + "grad_norm": 1.2835336923599243, + "learning_rate": 9.037789977743223e-05, + "loss": 0.6708, + "step": 6698 + }, + { + "epoch": 0.44964934062615347, + "grad_norm": 1.0042214393615723, + "learning_rate": 9.03714888250556e-05, + "loss": 0.7034, + "step": 6700 + }, + { + "epoch": 0.4497835643099225, + "grad_norm": 1.216511845588684, + "learning_rate": 9.036507596520551e-05, + "loss": 0.6803, + "step": 6702 + }, + { + "epoch": 0.4499177879936915, + "grad_norm": 0.9730432033538818, + "learning_rate": 9.035866119818495e-05, + "loss": 0.625, + "step": 6704 + }, + { + "epoch": 0.4500520116774605, + "grad_norm": 1.2708172798156738, + "learning_rate": 9.035224452429703e-05, + "loss": 0.6832, + "step": 6706 + }, + { + "epoch": 0.4501862353612295, + "grad_norm": 1.033936858177185, + "learning_rate": 9.034582594384488e-05, + "loss": 0.6413, + "step": 6708 + }, + { + "epoch": 0.4503204590449985, + "grad_norm": 1.223681092262268, + "learning_rate": 9.033940545713182e-05, + "loss": 0.6432, + "step": 6710 + }, + { + "epoch": 0.4504546827287675, + "grad_norm": 1.0408893823623657, + "learning_rate": 9.033298306446115e-05, + "loss": 0.615, + "step": 6712 + }, + { + "epoch": 0.4505889064125365, + "grad_norm": 1.0709335803985596, + "learning_rate": 9.032655876613636e-05, + "loss": 0.6871, + "step": 6714 + }, + { + "epoch": 0.4507231300963055, + "grad_norm": 1.0904982089996338, + "learning_rate": 9.032013256246094e-05, + "loss": 0.6562, + "step": 6716 + }, + { + "epoch": 0.4508573537800745, + "grad_norm": 1.2430102825164795, + "learning_rate": 9.031370445373856e-05, + "loss": 0.6927, + "step": 6718 + }, + { + "epoch": 0.4509915774638435, + "grad_norm": 1.2163711786270142, + "learning_rate": 9.030727444027294e-05, + "loss": 0.6461, + "step": 6720 + }, + { + "epoch": 0.45112580114761247, + "grad_norm": 0.965525209903717, + "learning_rate": 9.030084252236783e-05, + "loss": 0.6482, + "step": 6722 + }, + { + "epoch": 0.4512600248313815, + "grad_norm": 1.178641438484192, + "learning_rate": 9.029440870032718e-05, + "loss": 0.6777, + "step": 6724 + }, + { + "epoch": 0.4513942485151505, + "grad_norm": 1.0191656351089478, + "learning_rate": 9.028797297445495e-05, + "loss": 0.5993, + "step": 6726 + }, + { + "epoch": 0.4515284721989195, + "grad_norm": 1.122101902961731, + "learning_rate": 9.028153534505522e-05, + "loss": 0.6529, + "step": 6728 + }, + { + "epoch": 0.4516626958826885, + "grad_norm": 1.0204761028289795, + "learning_rate": 9.027509581243214e-05, + "loss": 0.6585, + "step": 6730 + }, + { + "epoch": 0.4517969195664575, + "grad_norm": 1.0416470766067505, + "learning_rate": 9.026865437688998e-05, + "loss": 0.669, + "step": 6732 + }, + { + "epoch": 0.4519311432502265, + "grad_norm": 1.1773619651794434, + "learning_rate": 9.026221103873312e-05, + "loss": 0.6587, + "step": 6734 + }, + { + "epoch": 0.4520653669339955, + "grad_norm": 1.0783255100250244, + "learning_rate": 9.025576579826593e-05, + "loss": 0.647, + "step": 6736 + }, + { + "epoch": 0.4521995906177645, + "grad_norm": 1.2282097339630127, + "learning_rate": 9.024931865579296e-05, + "loss": 0.6124, + "step": 6738 + }, + { + "epoch": 0.4523338143015335, + "grad_norm": 0.9424877762794495, + "learning_rate": 9.024286961161885e-05, + "loss": 0.6388, + "step": 6740 + }, + { + "epoch": 0.4524680379853025, + "grad_norm": 1.1589195728302002, + "learning_rate": 9.023641866604829e-05, + "loss": 0.6858, + "step": 6742 + }, + { + "epoch": 0.4526022616690715, + "grad_norm": 1.2010003328323364, + "learning_rate": 9.022996581938605e-05, + "loss": 0.6983, + "step": 6744 + }, + { + "epoch": 0.4527364853528405, + "grad_norm": 0.9364185333251953, + "learning_rate": 9.022351107193704e-05, + "loss": 0.6288, + "step": 6746 + }, + { + "epoch": 0.45287070903660953, + "grad_norm": 0.9412966370582581, + "learning_rate": 9.021705442400623e-05, + "loss": 0.6178, + "step": 6748 + }, + { + "epoch": 0.4530049327203785, + "grad_norm": 1.013659119606018, + "learning_rate": 9.021059587589869e-05, + "loss": 0.6267, + "step": 6750 + }, + { + "epoch": 0.45313915640414754, + "grad_norm": 1.0910239219665527, + "learning_rate": 9.020413542791955e-05, + "loss": 0.6773, + "step": 6752 + }, + { + "epoch": 0.4532733800879165, + "grad_norm": 1.3612470626831055, + "learning_rate": 9.019767308037407e-05, + "loss": 0.5651, + "step": 6754 + }, + { + "epoch": 0.45340760377168554, + "grad_norm": 1.2523924112319946, + "learning_rate": 9.01912088335676e-05, + "loss": 0.673, + "step": 6756 + }, + { + "epoch": 0.4535418274554545, + "grad_norm": 1.0892006158828735, + "learning_rate": 9.018474268780553e-05, + "loss": 0.6062, + "step": 6758 + }, + { + "epoch": 0.4536760511392235, + "grad_norm": 1.0809118747711182, + "learning_rate": 9.017827464339338e-05, + "loss": 0.6266, + "step": 6760 + }, + { + "epoch": 0.4538102748229925, + "grad_norm": 1.1393630504608154, + "learning_rate": 9.017180470063679e-05, + "loss": 0.6653, + "step": 6762 + }, + { + "epoch": 0.4539444985067615, + "grad_norm": 1.0781728029251099, + "learning_rate": 9.01653328598414e-05, + "loss": 0.6775, + "step": 6764 + }, + { + "epoch": 0.4540787221905305, + "grad_norm": 1.149721622467041, + "learning_rate": 9.015885912131302e-05, + "loss": 0.7004, + "step": 6766 + }, + { + "epoch": 0.4542129458742995, + "grad_norm": 1.207176923751831, + "learning_rate": 9.015238348535751e-05, + "loss": 0.714, + "step": 6768 + }, + { + "epoch": 0.45434716955806853, + "grad_norm": 1.212417483329773, + "learning_rate": 9.014590595228086e-05, + "loss": 0.6742, + "step": 6770 + }, + { + "epoch": 0.4544813932418375, + "grad_norm": 1.082265019416809, + "learning_rate": 9.013942652238908e-05, + "loss": 0.6742, + "step": 6772 + }, + { + "epoch": 0.45461561692560654, + "grad_norm": 0.9447269439697266, + "learning_rate": 9.01329451959883e-05, + "loss": 0.6373, + "step": 6774 + }, + { + "epoch": 0.4547498406093755, + "grad_norm": 0.9641191959381104, + "learning_rate": 9.012646197338481e-05, + "loss": 0.654, + "step": 6776 + }, + { + "epoch": 0.45488406429314454, + "grad_norm": 0.9599052667617798, + "learning_rate": 9.011997685488489e-05, + "loss": 0.6235, + "step": 6778 + }, + { + "epoch": 0.4550182879769135, + "grad_norm": 1.0892101526260376, + "learning_rate": 9.011348984079496e-05, + "loss": 0.6639, + "step": 6780 + }, + { + "epoch": 0.45515251166068255, + "grad_norm": 1.0836642980575562, + "learning_rate": 9.010700093142151e-05, + "loss": 0.6323, + "step": 6782 + }, + { + "epoch": 0.4552867353444515, + "grad_norm": 1.3505182266235352, + "learning_rate": 9.010051012707114e-05, + "loss": 0.6974, + "step": 6784 + }, + { + "epoch": 0.45542095902822055, + "grad_norm": 0.9923969507217407, + "learning_rate": 9.009401742805052e-05, + "loss": 0.6549, + "step": 6786 + }, + { + "epoch": 0.4555551827119895, + "grad_norm": 0.9913681149482727, + "learning_rate": 9.008752283466641e-05, + "loss": 0.635, + "step": 6788 + }, + { + "epoch": 0.45568940639575856, + "grad_norm": 1.0151026248931885, + "learning_rate": 9.008102634722568e-05, + "loss": 0.615, + "step": 6790 + }, + { + "epoch": 0.45582363007952753, + "grad_norm": 1.1850115060806274, + "learning_rate": 9.007452796603526e-05, + "loss": 0.5593, + "step": 6792 + }, + { + "epoch": 0.45595785376329656, + "grad_norm": 1.1839693784713745, + "learning_rate": 9.006802769140221e-05, + "loss": 0.6833, + "step": 6794 + }, + { + "epoch": 0.45609207744706554, + "grad_norm": 1.0648049116134644, + "learning_rate": 9.006152552363363e-05, + "loss": 0.7479, + "step": 6796 + }, + { + "epoch": 0.4562263011308345, + "grad_norm": 1.0856976509094238, + "learning_rate": 9.005502146303676e-05, + "loss": 0.5972, + "step": 6798 + }, + { + "epoch": 0.45636052481460354, + "grad_norm": 1.185408353805542, + "learning_rate": 9.004851550991888e-05, + "loss": 0.6914, + "step": 6800 + }, + { + "epoch": 0.4564947484983725, + "grad_norm": 1.1428455114364624, + "learning_rate": 9.004200766458742e-05, + "loss": 0.6366, + "step": 6802 + }, + { + "epoch": 0.45662897218214155, + "grad_norm": 1.0512735843658447, + "learning_rate": 9.00354979273498e-05, + "loss": 0.6416, + "step": 6804 + }, + { + "epoch": 0.4567631958659105, + "grad_norm": 1.0457926988601685, + "learning_rate": 9.002898629851364e-05, + "loss": 0.6299, + "step": 6806 + }, + { + "epoch": 0.45689741954967955, + "grad_norm": 0.984696090221405, + "learning_rate": 9.00224727783866e-05, + "loss": 0.6399, + "step": 6808 + }, + { + "epoch": 0.4570316432334485, + "grad_norm": 1.4410641193389893, + "learning_rate": 9.001595736727642e-05, + "loss": 0.6988, + "step": 6810 + }, + { + "epoch": 0.45716586691721756, + "grad_norm": 1.2012965679168701, + "learning_rate": 9.000944006549095e-05, + "loss": 0.7485, + "step": 6812 + }, + { + "epoch": 0.45730009060098653, + "grad_norm": 1.048329472541809, + "learning_rate": 9.00029208733381e-05, + "loss": 0.6663, + "step": 6814 + }, + { + "epoch": 0.45743431428475556, + "grad_norm": 1.1386462450027466, + "learning_rate": 8.99963997911259e-05, + "loss": 0.7471, + "step": 6816 + }, + { + "epoch": 0.45756853796852454, + "grad_norm": 1.0403379201889038, + "learning_rate": 8.998987681916246e-05, + "loss": 0.628, + "step": 6818 + }, + { + "epoch": 0.45770276165229357, + "grad_norm": 1.0118883848190308, + "learning_rate": 8.998335195775599e-05, + "loss": 0.6844, + "step": 6820 + }, + { + "epoch": 0.45783698533606254, + "grad_norm": 1.1694881916046143, + "learning_rate": 8.997682520721476e-05, + "loss": 0.6405, + "step": 6822 + }, + { + "epoch": 0.45797120901983157, + "grad_norm": 1.1148027181625366, + "learning_rate": 8.997029656784715e-05, + "loss": 0.653, + "step": 6824 + }, + { + "epoch": 0.45810543270360055, + "grad_norm": 1.0107674598693848, + "learning_rate": 8.996376603996161e-05, + "loss": 0.6555, + "step": 6826 + }, + { + "epoch": 0.4582396563873696, + "grad_norm": 0.9955770969390869, + "learning_rate": 8.995723362386672e-05, + "loss": 0.6387, + "step": 6828 + }, + { + "epoch": 0.45837388007113855, + "grad_norm": 1.1577054262161255, + "learning_rate": 8.995069931987113e-05, + "loss": 0.6875, + "step": 6830 + }, + { + "epoch": 0.4585081037549076, + "grad_norm": 1.1350963115692139, + "learning_rate": 8.994416312828354e-05, + "loss": 0.6236, + "step": 6832 + }, + { + "epoch": 0.45864232743867656, + "grad_norm": 1.0730098485946655, + "learning_rate": 8.993762504941277e-05, + "loss": 0.7111, + "step": 6834 + }, + { + "epoch": 0.45877655112244553, + "grad_norm": 1.1010090112686157, + "learning_rate": 8.993108508356779e-05, + "loss": 0.6476, + "step": 6836 + }, + { + "epoch": 0.45891077480621456, + "grad_norm": 1.0718328952789307, + "learning_rate": 8.992454323105752e-05, + "loss": 0.7306, + "step": 6838 + }, + { + "epoch": 0.45904499848998354, + "grad_norm": 1.111706018447876, + "learning_rate": 8.991799949219112e-05, + "loss": 0.624, + "step": 6840 + }, + { + "epoch": 0.45917922217375257, + "grad_norm": 1.0027763843536377, + "learning_rate": 8.991145386727773e-05, + "loss": 0.6267, + "step": 6842 + }, + { + "epoch": 0.45931344585752154, + "grad_norm": 1.1393334865570068, + "learning_rate": 8.990490635662663e-05, + "loss": 0.6908, + "step": 6844 + }, + { + "epoch": 0.45944766954129057, + "grad_norm": 1.5427532196044922, + "learning_rate": 8.989835696054718e-05, + "loss": 0.7297, + "step": 6846 + }, + { + "epoch": 0.45958189322505955, + "grad_norm": 1.078510046005249, + "learning_rate": 8.989180567934881e-05, + "loss": 0.7104, + "step": 6848 + }, + { + "epoch": 0.4597161169088286, + "grad_norm": 1.0581449270248413, + "learning_rate": 8.988525251334106e-05, + "loss": 0.6171, + "step": 6850 + }, + { + "epoch": 0.45985034059259755, + "grad_norm": 1.0170172452926636, + "learning_rate": 8.987869746283358e-05, + "loss": 0.643, + "step": 6852 + }, + { + "epoch": 0.4599845642763666, + "grad_norm": 0.9322235584259033, + "learning_rate": 8.987214052813604e-05, + "loss": 0.6311, + "step": 6854 + }, + { + "epoch": 0.46011878796013556, + "grad_norm": 1.0942978858947754, + "learning_rate": 8.986558170955828e-05, + "loss": 0.6775, + "step": 6856 + }, + { + "epoch": 0.4602530116439046, + "grad_norm": 1.0409022569656372, + "learning_rate": 8.985902100741018e-05, + "loss": 0.5689, + "step": 6858 + }, + { + "epoch": 0.46038723532767356, + "grad_norm": 1.0800971984863281, + "learning_rate": 8.98524584220017e-05, + "loss": 0.69, + "step": 6860 + }, + { + "epoch": 0.4605214590114426, + "grad_norm": 1.049082636833191, + "learning_rate": 8.984589395364294e-05, + "loss": 0.649, + "step": 6862 + }, + { + "epoch": 0.46065568269521157, + "grad_norm": 1.1251814365386963, + "learning_rate": 8.983932760264405e-05, + "loss": 0.6719, + "step": 6864 + }, + { + "epoch": 0.4607899063789806, + "grad_norm": 1.1507809162139893, + "learning_rate": 8.983275936931526e-05, + "loss": 0.727, + "step": 6866 + }, + { + "epoch": 0.46092413006274957, + "grad_norm": 0.9597031474113464, + "learning_rate": 8.982618925396691e-05, + "loss": 0.63, + "step": 6868 + }, + { + "epoch": 0.46105835374651855, + "grad_norm": 1.1732813119888306, + "learning_rate": 8.981961725690943e-05, + "loss": 0.6308, + "step": 6870 + }, + { + "epoch": 0.4611925774302876, + "grad_norm": 1.0450263023376465, + "learning_rate": 8.981304337845337e-05, + "loss": 0.5938, + "step": 6872 + }, + { + "epoch": 0.46132680111405655, + "grad_norm": 1.109165906906128, + "learning_rate": 8.980646761890928e-05, + "loss": 0.6669, + "step": 6874 + }, + { + "epoch": 0.4614610247978256, + "grad_norm": 0.9738684296607971, + "learning_rate": 8.979988997858785e-05, + "loss": 0.6748, + "step": 6876 + }, + { + "epoch": 0.46159524848159456, + "grad_norm": 1.1553380489349365, + "learning_rate": 8.97933104577999e-05, + "loss": 0.6687, + "step": 6878 + }, + { + "epoch": 0.4617294721653636, + "grad_norm": 1.0648709535598755, + "learning_rate": 8.978672905685629e-05, + "loss": 0.6825, + "step": 6880 + }, + { + "epoch": 0.46186369584913256, + "grad_norm": 1.3166885375976562, + "learning_rate": 8.978014577606797e-05, + "loss": 0.6791, + "step": 6882 + }, + { + "epoch": 0.4619979195329016, + "grad_norm": 0.9989602565765381, + "learning_rate": 8.977356061574597e-05, + "loss": 0.6538, + "step": 6884 + }, + { + "epoch": 0.46213214321667057, + "grad_norm": 1.5037658214569092, + "learning_rate": 8.976697357620145e-05, + "loss": 0.6675, + "step": 6886 + }, + { + "epoch": 0.4622663669004396, + "grad_norm": 1.0083638429641724, + "learning_rate": 8.976038465774563e-05, + "loss": 0.5991, + "step": 6888 + }, + { + "epoch": 0.46240059058420857, + "grad_norm": 1.194140911102295, + "learning_rate": 8.975379386068981e-05, + "loss": 0.7116, + "step": 6890 + }, + { + "epoch": 0.4625348142679776, + "grad_norm": 1.4568372964859009, + "learning_rate": 8.974720118534541e-05, + "loss": 0.6345, + "step": 6892 + }, + { + "epoch": 0.4626690379517466, + "grad_norm": 1.069860577583313, + "learning_rate": 8.974060663202392e-05, + "loss": 0.7056, + "step": 6894 + }, + { + "epoch": 0.4628032616355156, + "grad_norm": 1.410301923751831, + "learning_rate": 8.97340102010369e-05, + "loss": 0.6908, + "step": 6896 + }, + { + "epoch": 0.4629374853192846, + "grad_norm": 1.1220239400863647, + "learning_rate": 8.972741189269605e-05, + "loss": 0.6735, + "step": 6898 + }, + { + "epoch": 0.4630717090030536, + "grad_norm": 2.3627586364746094, + "learning_rate": 8.972081170731307e-05, + "loss": 0.6223, + "step": 6900 + }, + { + "epoch": 0.4632059326868226, + "grad_norm": 1.1087870597839355, + "learning_rate": 8.971420964519988e-05, + "loss": 0.6791, + "step": 6902 + }, + { + "epoch": 0.4633401563705916, + "grad_norm": 1.1223286390304565, + "learning_rate": 8.970760570666839e-05, + "loss": 0.7227, + "step": 6904 + }, + { + "epoch": 0.4634743800543606, + "grad_norm": 0.9709183573722839, + "learning_rate": 8.970099989203058e-05, + "loss": 0.6467, + "step": 6906 + }, + { + "epoch": 0.46360860373812957, + "grad_norm": 1.0744274854660034, + "learning_rate": 8.96943922015986e-05, + "loss": 0.6976, + "step": 6908 + }, + { + "epoch": 0.4637428274218986, + "grad_norm": 1.0004558563232422, + "learning_rate": 8.968778263568465e-05, + "loss": 0.613, + "step": 6910 + }, + { + "epoch": 0.46387705110566757, + "grad_norm": 1.0825237035751343, + "learning_rate": 8.968117119460103e-05, + "loss": 0.7359, + "step": 6912 + }, + { + "epoch": 0.4640112747894366, + "grad_norm": 1.074562668800354, + "learning_rate": 8.967455787866007e-05, + "loss": 0.6747, + "step": 6914 + }, + { + "epoch": 0.4641454984732056, + "grad_norm": 1.0532876253128052, + "learning_rate": 8.966794268817427e-05, + "loss": 0.6445, + "step": 6916 + }, + { + "epoch": 0.4642797221569746, + "grad_norm": 1.5095680952072144, + "learning_rate": 8.96613256234562e-05, + "loss": 0.7003, + "step": 6918 + }, + { + "epoch": 0.4644139458407436, + "grad_norm": 2.5291311740875244, + "learning_rate": 8.965470668481848e-05, + "loss": 0.6705, + "step": 6920 + }, + { + "epoch": 0.4645481695245126, + "grad_norm": 1.0325822830200195, + "learning_rate": 8.964808587257386e-05, + "loss": 0.651, + "step": 6922 + }, + { + "epoch": 0.4646823932082816, + "grad_norm": 1.061231255531311, + "learning_rate": 8.964146318703512e-05, + "loss": 0.5798, + "step": 6924 + }, + { + "epoch": 0.4648166168920506, + "grad_norm": 1.285365104675293, + "learning_rate": 8.96348386285152e-05, + "loss": 0.7, + "step": 6926 + }, + { + "epoch": 0.4649508405758196, + "grad_norm": 1.1364279985427856, + "learning_rate": 8.962821219732711e-05, + "loss": 0.656, + "step": 6928 + }, + { + "epoch": 0.4650850642595886, + "grad_norm": 1.1411540508270264, + "learning_rate": 8.96215838937839e-05, + "loss": 0.6316, + "step": 6930 + }, + { + "epoch": 0.4652192879433576, + "grad_norm": 1.1594171524047852, + "learning_rate": 8.961495371819877e-05, + "loss": 0.723, + "step": 6932 + }, + { + "epoch": 0.4653535116271266, + "grad_norm": 1.0198676586151123, + "learning_rate": 8.960832167088498e-05, + "loss": 0.6404, + "step": 6934 + }, + { + "epoch": 0.4654877353108956, + "grad_norm": 0.9912198781967163, + "learning_rate": 8.960168775215588e-05, + "loss": 0.6453, + "step": 6936 + }, + { + "epoch": 0.46562195899466463, + "grad_norm": 1.1035021543502808, + "learning_rate": 8.95950519623249e-05, + "loss": 0.6325, + "step": 6938 + }, + { + "epoch": 0.4657561826784336, + "grad_norm": 0.9891259074211121, + "learning_rate": 8.95884143017056e-05, + "loss": 0.6341, + "step": 6940 + }, + { + "epoch": 0.46589040636220264, + "grad_norm": 0.9204497337341309, + "learning_rate": 8.958177477061154e-05, + "loss": 0.6293, + "step": 6942 + }, + { + "epoch": 0.4660246300459716, + "grad_norm": 1.120475172996521, + "learning_rate": 8.957513336935646e-05, + "loss": 0.6751, + "step": 6944 + }, + { + "epoch": 0.4661588537297406, + "grad_norm": 1.1103843450546265, + "learning_rate": 8.956849009825417e-05, + "loss": 0.6358, + "step": 6946 + }, + { + "epoch": 0.4662930774135096, + "grad_norm": 1.2845054864883423, + "learning_rate": 8.95618449576185e-05, + "loss": 0.7357, + "step": 6948 + }, + { + "epoch": 0.4664273010972786, + "grad_norm": 1.0931564569473267, + "learning_rate": 8.955519794776348e-05, + "loss": 0.6829, + "step": 6950 + }, + { + "epoch": 0.4665615247810476, + "grad_norm": 1.0615421533584595, + "learning_rate": 8.954854906900312e-05, + "loss": 0.6505, + "step": 6952 + }, + { + "epoch": 0.4666957484648166, + "grad_norm": 1.0076024532318115, + "learning_rate": 8.954189832165159e-05, + "loss": 0.6809, + "step": 6954 + }, + { + "epoch": 0.4668299721485856, + "grad_norm": 1.1049742698669434, + "learning_rate": 8.953524570602313e-05, + "loss": 0.7196, + "step": 6956 + }, + { + "epoch": 0.4669641958323546, + "grad_norm": 1.5184391736984253, + "learning_rate": 8.952859122243204e-05, + "loss": 0.7042, + "step": 6958 + }, + { + "epoch": 0.46709841951612363, + "grad_norm": 1.5310966968536377, + "learning_rate": 8.952193487119276e-05, + "loss": 0.6192, + "step": 6960 + }, + { + "epoch": 0.4672326431998926, + "grad_norm": 1.4648321866989136, + "learning_rate": 8.951527665261976e-05, + "loss": 0.6901, + "step": 6962 + }, + { + "epoch": 0.46736686688366164, + "grad_norm": 1.162032961845398, + "learning_rate": 8.950861656702764e-05, + "loss": 0.7611, + "step": 6964 + }, + { + "epoch": 0.4675010905674306, + "grad_norm": 1.1246246099472046, + "learning_rate": 8.950195461473109e-05, + "loss": 0.6745, + "step": 6966 + }, + { + "epoch": 0.46763531425119964, + "grad_norm": 1.2168980836868286, + "learning_rate": 8.949529079604485e-05, + "loss": 0.6573, + "step": 6968 + }, + { + "epoch": 0.4677695379349686, + "grad_norm": 1.2611074447631836, + "learning_rate": 8.94886251112838e-05, + "loss": 0.6753, + "step": 6970 + }, + { + "epoch": 0.46790376161873765, + "grad_norm": 1.0567899942398071, + "learning_rate": 8.948195756076285e-05, + "loss": 0.7266, + "step": 6972 + }, + { + "epoch": 0.4680379853025066, + "grad_norm": 1.4233819246292114, + "learning_rate": 8.947528814479704e-05, + "loss": 0.6134, + "step": 6974 + }, + { + "epoch": 0.46817220898627565, + "grad_norm": 1.2391204833984375, + "learning_rate": 8.94686168637015e-05, + "loss": 0.663, + "step": 6976 + }, + { + "epoch": 0.4683064326700446, + "grad_norm": 1.310097575187683, + "learning_rate": 8.946194371779142e-05, + "loss": 0.6546, + "step": 6978 + }, + { + "epoch": 0.46844065635381366, + "grad_norm": 1.034376621246338, + "learning_rate": 8.94552687073821e-05, + "loss": 0.6731, + "step": 6980 + }, + { + "epoch": 0.46857488003758263, + "grad_norm": 1.1510058641433716, + "learning_rate": 8.944859183278891e-05, + "loss": 0.6442, + "step": 6982 + }, + { + "epoch": 0.4687091037213516, + "grad_norm": 1.1591176986694336, + "learning_rate": 8.944191309432735e-05, + "loss": 0.6933, + "step": 6984 + }, + { + "epoch": 0.46884332740512064, + "grad_norm": 1.083055853843689, + "learning_rate": 8.943523249231293e-05, + "loss": 0.6641, + "step": 6986 + }, + { + "epoch": 0.4689775510888896, + "grad_norm": 1.4717953205108643, + "learning_rate": 8.942855002706134e-05, + "loss": 0.6953, + "step": 6988 + }, + { + "epoch": 0.46911177477265864, + "grad_norm": 1.2020087242126465, + "learning_rate": 8.942186569888829e-05, + "loss": 0.6102, + "step": 6990 + }, + { + "epoch": 0.4692459984564276, + "grad_norm": 1.069995403289795, + "learning_rate": 8.94151795081096e-05, + "loss": 0.6819, + "step": 6992 + }, + { + "epoch": 0.46938022214019665, + "grad_norm": 1.1206103563308716, + "learning_rate": 8.940849145504118e-05, + "loss": 0.683, + "step": 6994 + }, + { + "epoch": 0.4695144458239656, + "grad_norm": 1.803853988647461, + "learning_rate": 8.940180153999904e-05, + "loss": 0.6418, + "step": 6996 + }, + { + "epoch": 0.46964866950773465, + "grad_norm": 2.395535707473755, + "learning_rate": 8.939510976329927e-05, + "loss": 0.6804, + "step": 6998 + }, + { + "epoch": 0.4697828931915036, + "grad_norm": 1.123642921447754, + "learning_rate": 8.938841612525801e-05, + "loss": 0.6524, + "step": 7000 + }, + { + "epoch": 0.46991711687527266, + "grad_norm": 1.076267123222351, + "learning_rate": 8.938172062619155e-05, + "loss": 0.6429, + "step": 7002 + }, + { + "epoch": 0.47005134055904163, + "grad_norm": 1.2277798652648926, + "learning_rate": 8.937502326641622e-05, + "loss": 0.7775, + "step": 7004 + }, + { + "epoch": 0.47018556424281066, + "grad_norm": 0.8746130466461182, + "learning_rate": 8.936832404624848e-05, + "loss": 0.6474, + "step": 7006 + }, + { + "epoch": 0.47031978792657964, + "grad_norm": 1.0472476482391357, + "learning_rate": 8.936162296600486e-05, + "loss": 0.6987, + "step": 7008 + }, + { + "epoch": 0.47045401161034867, + "grad_norm": 1.4213472604751587, + "learning_rate": 8.935492002600194e-05, + "loss": 0.5801, + "step": 7010 + }, + { + "epoch": 0.47058823529411764, + "grad_norm": 1.0555846691131592, + "learning_rate": 8.934821522655642e-05, + "loss": 0.7152, + "step": 7012 + }, + { + "epoch": 0.47072245897788667, + "grad_norm": 1.1034077405929565, + "learning_rate": 8.934150856798514e-05, + "loss": 0.6107, + "step": 7014 + }, + { + "epoch": 0.47085668266165565, + "grad_norm": 1.109114408493042, + "learning_rate": 8.933480005060492e-05, + "loss": 0.7219, + "step": 7016 + }, + { + "epoch": 0.4709909063454247, + "grad_norm": 1.1220183372497559, + "learning_rate": 8.932808967473274e-05, + "loss": 0.641, + "step": 7018 + }, + { + "epoch": 0.47112513002919365, + "grad_norm": 0.983274519443512, + "learning_rate": 8.932137744068567e-05, + "loss": 0.604, + "step": 7020 + }, + { + "epoch": 0.4712593537129626, + "grad_norm": 1.0105772018432617, + "learning_rate": 8.931466334878085e-05, + "loss": 0.6437, + "step": 7022 + }, + { + "epoch": 0.47139357739673166, + "grad_norm": 1.0627435445785522, + "learning_rate": 8.930794739933547e-05, + "loss": 0.633, + "step": 7024 + }, + { + "epoch": 0.47152780108050063, + "grad_norm": 1.4478068351745605, + "learning_rate": 8.930122959266689e-05, + "loss": 0.6532, + "step": 7026 + }, + { + "epoch": 0.47166202476426966, + "grad_norm": 0.9602782130241394, + "learning_rate": 8.929450992909248e-05, + "loss": 0.7098, + "step": 7028 + }, + { + "epoch": 0.47179624844803864, + "grad_norm": 1.1455880403518677, + "learning_rate": 8.928778840892975e-05, + "loss": 0.6132, + "step": 7030 + }, + { + "epoch": 0.47193047213180767, + "grad_norm": 1.0074069499969482, + "learning_rate": 8.928106503249628e-05, + "loss": 0.6814, + "step": 7032 + }, + { + "epoch": 0.47206469581557664, + "grad_norm": 1.0082672834396362, + "learning_rate": 8.927433980010973e-05, + "loss": 0.5948, + "step": 7034 + }, + { + "epoch": 0.47219891949934567, + "grad_norm": 1.070309042930603, + "learning_rate": 8.926761271208785e-05, + "loss": 0.7057, + "step": 7036 + }, + { + "epoch": 0.47233314318311465, + "grad_norm": 1.1839375495910645, + "learning_rate": 8.926088376874849e-05, + "loss": 0.6744, + "step": 7038 + }, + { + "epoch": 0.4724673668668837, + "grad_norm": 1.0738030672073364, + "learning_rate": 8.925415297040957e-05, + "loss": 0.6774, + "step": 7040 + }, + { + "epoch": 0.47260159055065265, + "grad_norm": 1.1747937202453613, + "learning_rate": 8.924742031738911e-05, + "loss": 0.5994, + "step": 7042 + }, + { + "epoch": 0.4727358142344217, + "grad_norm": 0.9965662360191345, + "learning_rate": 8.924068581000521e-05, + "loss": 0.6201, + "step": 7044 + }, + { + "epoch": 0.47287003791819066, + "grad_norm": 1.0046024322509766, + "learning_rate": 8.923394944857609e-05, + "loss": 0.6126, + "step": 7046 + }, + { + "epoch": 0.4730042616019597, + "grad_norm": 0.9661778807640076, + "learning_rate": 8.922721123341999e-05, + "loss": 0.6749, + "step": 7048 + }, + { + "epoch": 0.47313848528572866, + "grad_norm": 1.0183511972427368, + "learning_rate": 8.922047116485532e-05, + "loss": 0.6765, + "step": 7050 + }, + { + "epoch": 0.4732727089694977, + "grad_norm": 0.9786390662193298, + "learning_rate": 8.921372924320048e-05, + "loss": 0.6522, + "step": 7052 + }, + { + "epoch": 0.47340693265326667, + "grad_norm": 1.198760747909546, + "learning_rate": 8.920698546877406e-05, + "loss": 0.6494, + "step": 7054 + }, + { + "epoch": 0.4735411563370357, + "grad_norm": 0.9723764061927795, + "learning_rate": 8.920023984189468e-05, + "loss": 0.6539, + "step": 7056 + }, + { + "epoch": 0.47367538002080467, + "grad_norm": 1.0999858379364014, + "learning_rate": 8.919349236288105e-05, + "loss": 0.6382, + "step": 7058 + }, + { + "epoch": 0.47380960370457365, + "grad_norm": 1.075578212738037, + "learning_rate": 8.918674303205197e-05, + "loss": 0.7138, + "step": 7060 + }, + { + "epoch": 0.4739438273883427, + "grad_norm": 0.970807671546936, + "learning_rate": 8.917999184972634e-05, + "loss": 0.654, + "step": 7062 + }, + { + "epoch": 0.47407805107211165, + "grad_norm": 1.1560735702514648, + "learning_rate": 8.917323881622314e-05, + "loss": 0.6951, + "step": 7064 + }, + { + "epoch": 0.4742122747558807, + "grad_norm": 1.1930887699127197, + "learning_rate": 8.916648393186143e-05, + "loss": 0.5947, + "step": 7066 + }, + { + "epoch": 0.47434649843964966, + "grad_norm": 1.014479637145996, + "learning_rate": 8.915972719696037e-05, + "loss": 0.6198, + "step": 7068 + }, + { + "epoch": 0.4744807221234187, + "grad_norm": 0.9746931791305542, + "learning_rate": 8.915296861183923e-05, + "loss": 0.7069, + "step": 7070 + }, + { + "epoch": 0.47461494580718766, + "grad_norm": 1.0965828895568848, + "learning_rate": 8.914620817681729e-05, + "loss": 0.6967, + "step": 7072 + }, + { + "epoch": 0.4747491694909567, + "grad_norm": 1.3043853044509888, + "learning_rate": 8.9139445892214e-05, + "loss": 0.6357, + "step": 7074 + }, + { + "epoch": 0.47488339317472567, + "grad_norm": 1.6396334171295166, + "learning_rate": 8.913268175834886e-05, + "loss": 0.7468, + "step": 7076 + }, + { + "epoch": 0.4750176168584947, + "grad_norm": 1.0765565633773804, + "learning_rate": 8.912591577554143e-05, + "loss": 0.6436, + "step": 7078 + }, + { + "epoch": 0.47515184054226367, + "grad_norm": 1.0507439374923706, + "learning_rate": 8.911914794411144e-05, + "loss": 0.6137, + "step": 7080 + }, + { + "epoch": 0.4752860642260327, + "grad_norm": 1.0619193315505981, + "learning_rate": 8.911237826437865e-05, + "loss": 0.6245, + "step": 7082 + }, + { + "epoch": 0.4754202879098017, + "grad_norm": 1.102828025817871, + "learning_rate": 8.910560673666289e-05, + "loss": 0.7056, + "step": 7084 + }, + { + "epoch": 0.4755545115935707, + "grad_norm": 1.1942352056503296, + "learning_rate": 8.909883336128408e-05, + "loss": 0.6584, + "step": 7086 + }, + { + "epoch": 0.4756887352773397, + "grad_norm": 0.9986782073974609, + "learning_rate": 8.909205813856232e-05, + "loss": 0.6168, + "step": 7088 + }, + { + "epoch": 0.4758229589611087, + "grad_norm": 0.9641457200050354, + "learning_rate": 8.908528106881765e-05, + "loss": 0.6498, + "step": 7090 + }, + { + "epoch": 0.4759571826448777, + "grad_norm": 1.2766104936599731, + "learning_rate": 8.907850215237032e-05, + "loss": 0.7591, + "step": 7092 + }, + { + "epoch": 0.4760914063286467, + "grad_norm": 0.9906812310218811, + "learning_rate": 8.907172138954061e-05, + "loss": 0.6064, + "step": 7094 + }, + { + "epoch": 0.4762256300124157, + "grad_norm": 1.2280668020248413, + "learning_rate": 8.90649387806489e-05, + "loss": 0.6616, + "step": 7096 + }, + { + "epoch": 0.47635985369618467, + "grad_norm": 0.9963435530662537, + "learning_rate": 8.905815432601566e-05, + "loss": 0.6256, + "step": 7098 + }, + { + "epoch": 0.4764940773799537, + "grad_norm": 1.073029637336731, + "learning_rate": 8.905136802596142e-05, + "loss": 0.624, + "step": 7100 + }, + { + "epoch": 0.47662830106372267, + "grad_norm": 0.966012716293335, + "learning_rate": 8.904457988080681e-05, + "loss": 0.651, + "step": 7102 + }, + { + "epoch": 0.4767625247474917, + "grad_norm": 1.126132607460022, + "learning_rate": 8.90377898908726e-05, + "loss": 0.6854, + "step": 7104 + }, + { + "epoch": 0.4768967484312607, + "grad_norm": 1.112329125404358, + "learning_rate": 8.903099805647959e-05, + "loss": 0.6875, + "step": 7106 + }, + { + "epoch": 0.4770309721150297, + "grad_norm": 0.9800264239311218, + "learning_rate": 8.902420437794865e-05, + "loss": 0.6274, + "step": 7108 + }, + { + "epoch": 0.4771651957987987, + "grad_norm": 1.270875096321106, + "learning_rate": 8.901740885560082e-05, + "loss": 0.6694, + "step": 7110 + }, + { + "epoch": 0.4772994194825677, + "grad_norm": 1.0601409673690796, + "learning_rate": 8.901061148975711e-05, + "loss": 0.6756, + "step": 7112 + }, + { + "epoch": 0.4774336431663367, + "grad_norm": 1.0308822393417358, + "learning_rate": 8.900381228073875e-05, + "loss": 0.6611, + "step": 7114 + }, + { + "epoch": 0.4775678668501057, + "grad_norm": 1.0516387224197388, + "learning_rate": 8.899701122886695e-05, + "loss": 0.6754, + "step": 7116 + }, + { + "epoch": 0.4777020905338747, + "grad_norm": 1.080004096031189, + "learning_rate": 8.899020833446304e-05, + "loss": 0.6428, + "step": 7118 + }, + { + "epoch": 0.4778363142176437, + "grad_norm": 1.009861707687378, + "learning_rate": 8.898340359784847e-05, + "loss": 0.5901, + "step": 7120 + }, + { + "epoch": 0.4779705379014127, + "grad_norm": 1.106646180152893, + "learning_rate": 8.897659701934474e-05, + "loss": 0.6663, + "step": 7122 + }, + { + "epoch": 0.4781047615851817, + "grad_norm": 1.251336693763733, + "learning_rate": 8.896978859927343e-05, + "loss": 0.6562, + "step": 7124 + }, + { + "epoch": 0.4782389852689507, + "grad_norm": 1.4340730905532837, + "learning_rate": 8.896297833795625e-05, + "loss": 0.5836, + "step": 7126 + }, + { + "epoch": 0.47837320895271973, + "grad_norm": 1.22294282913208, + "learning_rate": 8.895616623571497e-05, + "loss": 0.6578, + "step": 7128 + }, + { + "epoch": 0.4785074326364887, + "grad_norm": 1.2244223356246948, + "learning_rate": 8.894935229287142e-05, + "loss": 0.6836, + "step": 7130 + }, + { + "epoch": 0.47864165632025774, + "grad_norm": 0.978756844997406, + "learning_rate": 8.894253650974757e-05, + "loss": 0.6477, + "step": 7132 + }, + { + "epoch": 0.4787758800040267, + "grad_norm": 1.2327920198440552, + "learning_rate": 8.893571888666545e-05, + "loss": 0.6362, + "step": 7134 + }, + { + "epoch": 0.4789101036877957, + "grad_norm": 0.9988105893135071, + "learning_rate": 8.892889942394719e-05, + "loss": 0.6637, + "step": 7136 + }, + { + "epoch": 0.4790443273715647, + "grad_norm": 0.9424437880516052, + "learning_rate": 8.892207812191497e-05, + "loss": 0.5791, + "step": 7138 + }, + { + "epoch": 0.4791785510553337, + "grad_norm": 1.1618750095367432, + "learning_rate": 8.89152549808911e-05, + "loss": 0.6781, + "step": 7140 + }, + { + "epoch": 0.4793127747391027, + "grad_norm": 1.0431857109069824, + "learning_rate": 8.890843000119795e-05, + "loss": 0.6332, + "step": 7142 + }, + { + "epoch": 0.4794469984228717, + "grad_norm": 1.1647588014602661, + "learning_rate": 8.890160318315798e-05, + "loss": 0.7326, + "step": 7144 + }, + { + "epoch": 0.4795812221066407, + "grad_norm": 1.0543895959854126, + "learning_rate": 8.889477452709378e-05, + "loss": 0.6853, + "step": 7146 + }, + { + "epoch": 0.4797154457904097, + "grad_norm": 0.9098528027534485, + "learning_rate": 8.888794403332797e-05, + "loss": 0.5688, + "step": 7148 + }, + { + "epoch": 0.47984966947417873, + "grad_norm": 1.0702451467514038, + "learning_rate": 8.888111170218325e-05, + "loss": 0.6047, + "step": 7150 + }, + { + "epoch": 0.4799838931579477, + "grad_norm": 1.1213874816894531, + "learning_rate": 8.887427753398248e-05, + "loss": 0.6557, + "step": 7152 + }, + { + "epoch": 0.48011811684171674, + "grad_norm": 1.6079059839248657, + "learning_rate": 8.886744152904851e-05, + "loss": 0.6407, + "step": 7154 + }, + { + "epoch": 0.4802523405254857, + "grad_norm": 1.0884034633636475, + "learning_rate": 8.886060368770439e-05, + "loss": 0.6665, + "step": 7156 + }, + { + "epoch": 0.48038656420925474, + "grad_norm": 1.1436165571212769, + "learning_rate": 8.885376401027315e-05, + "loss": 0.7486, + "step": 7158 + }, + { + "epoch": 0.4805207878930237, + "grad_norm": 0.9711717963218689, + "learning_rate": 8.884692249707795e-05, + "loss": 0.6311, + "step": 7160 + }, + { + "epoch": 0.48065501157679275, + "grad_norm": 1.055991768836975, + "learning_rate": 8.884007914844208e-05, + "loss": 0.6775, + "step": 7162 + }, + { + "epoch": 0.4807892352605617, + "grad_norm": 0.982055127620697, + "learning_rate": 8.883323396468882e-05, + "loss": 0.5728, + "step": 7164 + }, + { + "epoch": 0.48092345894433075, + "grad_norm": 1.1967780590057373, + "learning_rate": 8.882638694614163e-05, + "loss": 0.6189, + "step": 7166 + }, + { + "epoch": 0.4810576826280997, + "grad_norm": 2.022672176361084, + "learning_rate": 8.8819538093124e-05, + "loss": 0.6913, + "step": 7168 + }, + { + "epoch": 0.48119190631186876, + "grad_norm": 1.0841823816299438, + "learning_rate": 8.881268740595954e-05, + "loss": 0.6773, + "step": 7170 + }, + { + "epoch": 0.48132612999563773, + "grad_norm": 1.0845587253570557, + "learning_rate": 8.880583488497192e-05, + "loss": 0.658, + "step": 7172 + }, + { + "epoch": 0.4814603536794067, + "grad_norm": 1.6183933019638062, + "learning_rate": 8.87989805304849e-05, + "loss": 0.5913, + "step": 7174 + }, + { + "epoch": 0.48159457736317574, + "grad_norm": 1.0903329849243164, + "learning_rate": 8.879212434282235e-05, + "loss": 0.6679, + "step": 7176 + }, + { + "epoch": 0.4817288010469447, + "grad_norm": 1.0819309949874878, + "learning_rate": 8.878526632230819e-05, + "loss": 0.6312, + "step": 7178 + }, + { + "epoch": 0.48186302473071374, + "grad_norm": 1.1857296228408813, + "learning_rate": 8.87784064692665e-05, + "loss": 0.6358, + "step": 7180 + }, + { + "epoch": 0.4819972484144827, + "grad_norm": 1.0356576442718506, + "learning_rate": 8.877154478402131e-05, + "loss": 0.6143, + "step": 7182 + }, + { + "epoch": 0.48213147209825175, + "grad_norm": 1.0137851238250732, + "learning_rate": 8.876468126689692e-05, + "loss": 0.7352, + "step": 7184 + }, + { + "epoch": 0.4822656957820207, + "grad_norm": 1.004191279411316, + "learning_rate": 8.875781591821754e-05, + "loss": 0.6567, + "step": 7186 + }, + { + "epoch": 0.48239991946578975, + "grad_norm": 1.1644058227539062, + "learning_rate": 8.875094873830758e-05, + "loss": 0.6588, + "step": 7188 + }, + { + "epoch": 0.4825341431495587, + "grad_norm": 1.0665473937988281, + "learning_rate": 8.87440797274915e-05, + "loss": 0.6818, + "step": 7190 + }, + { + "epoch": 0.48266836683332776, + "grad_norm": 1.214895248413086, + "learning_rate": 8.873720888609382e-05, + "loss": 0.6378, + "step": 7192 + }, + { + "epoch": 0.48280259051709673, + "grad_norm": 1.0585546493530273, + "learning_rate": 8.873033621443921e-05, + "loss": 0.6605, + "step": 7194 + }, + { + "epoch": 0.48293681420086576, + "grad_norm": 1.1378669738769531, + "learning_rate": 8.872346171285237e-05, + "loss": 0.6071, + "step": 7196 + }, + { + "epoch": 0.48307103788463474, + "grad_norm": 1.0823408365249634, + "learning_rate": 8.871658538165811e-05, + "loss": 0.661, + "step": 7198 + }, + { + "epoch": 0.48320526156840377, + "grad_norm": 1.0443452596664429, + "learning_rate": 8.870970722118132e-05, + "loss": 0.6439, + "step": 7200 + }, + { + "epoch": 0.48333948525217274, + "grad_norm": 1.0604746341705322, + "learning_rate": 8.870282723174699e-05, + "loss": 0.5891, + "step": 7202 + }, + { + "epoch": 0.48347370893594177, + "grad_norm": 1.1016902923583984, + "learning_rate": 8.869594541368017e-05, + "loss": 0.6408, + "step": 7204 + }, + { + "epoch": 0.48360793261971075, + "grad_norm": 1.0672671794891357, + "learning_rate": 8.868906176730602e-05, + "loss": 0.7246, + "step": 7206 + }, + { + "epoch": 0.4837421563034798, + "grad_norm": 1.0549395084381104, + "learning_rate": 8.868217629294979e-05, + "loss": 0.6581, + "step": 7208 + }, + { + "epoch": 0.48387637998724875, + "grad_norm": 1.0876435041427612, + "learning_rate": 8.86752889909368e-05, + "loss": 0.6263, + "step": 7210 + }, + { + "epoch": 0.4840106036710177, + "grad_norm": 0.9596108794212341, + "learning_rate": 8.866839986159244e-05, + "loss": 0.5984, + "step": 7212 + }, + { + "epoch": 0.48414482735478676, + "grad_norm": 2.628302574157715, + "learning_rate": 8.866150890524224e-05, + "loss": 0.6294, + "step": 7214 + }, + { + "epoch": 0.48427905103855573, + "grad_norm": 1.0163933038711548, + "learning_rate": 8.865461612221176e-05, + "loss": 0.6496, + "step": 7216 + }, + { + "epoch": 0.48441327472232476, + "grad_norm": 1.1279231309890747, + "learning_rate": 8.864772151282668e-05, + "loss": 0.6772, + "step": 7218 + }, + { + "epoch": 0.48454749840609374, + "grad_norm": 1.2656511068344116, + "learning_rate": 8.864082507741276e-05, + "loss": 0.6073, + "step": 7220 + }, + { + "epoch": 0.48468172208986277, + "grad_norm": 0.9335100650787354, + "learning_rate": 8.863392681629583e-05, + "loss": 0.5735, + "step": 7222 + }, + { + "epoch": 0.48481594577363174, + "grad_norm": 1.1231383085250854, + "learning_rate": 8.862702672980185e-05, + "loss": 0.6264, + "step": 7224 + }, + { + "epoch": 0.48495016945740077, + "grad_norm": 1.1034128665924072, + "learning_rate": 8.862012481825679e-05, + "loss": 0.5951, + "step": 7226 + }, + { + "epoch": 0.48508439314116975, + "grad_norm": 1.06645929813385, + "learning_rate": 8.861322108198678e-05, + "loss": 0.6167, + "step": 7228 + }, + { + "epoch": 0.4852186168249388, + "grad_norm": 1.1769659519195557, + "learning_rate": 8.860631552131801e-05, + "loss": 0.6778, + "step": 7230 + }, + { + "epoch": 0.48535284050870775, + "grad_norm": 1.1352919340133667, + "learning_rate": 8.859940813657675e-05, + "loss": 0.638, + "step": 7232 + }, + { + "epoch": 0.4854870641924768, + "grad_norm": 1.0001920461654663, + "learning_rate": 8.859249892808935e-05, + "loss": 0.6859, + "step": 7234 + }, + { + "epoch": 0.48562128787624576, + "grad_norm": 1.0392810106277466, + "learning_rate": 8.858558789618228e-05, + "loss": 0.6792, + "step": 7236 + }, + { + "epoch": 0.4857555115600148, + "grad_norm": 1.0945895910263062, + "learning_rate": 8.857867504118204e-05, + "loss": 0.6026, + "step": 7238 + }, + { + "epoch": 0.48588973524378376, + "grad_norm": 0.982492208480835, + "learning_rate": 8.857176036341526e-05, + "loss": 0.619, + "step": 7240 + }, + { + "epoch": 0.4860239589275528, + "grad_norm": 1.1728482246398926, + "learning_rate": 8.856484386320867e-05, + "loss": 0.6795, + "step": 7242 + }, + { + "epoch": 0.48615818261132177, + "grad_norm": 1.0668323040008545, + "learning_rate": 8.855792554088903e-05, + "loss": 0.61, + "step": 7244 + }, + { + "epoch": 0.4862924062950908, + "grad_norm": 1.027684211730957, + "learning_rate": 8.855100539678324e-05, + "loss": 0.6812, + "step": 7246 + }, + { + "epoch": 0.48642662997885977, + "grad_norm": 1.0103657245635986, + "learning_rate": 8.854408343121824e-05, + "loss": 0.6584, + "step": 7248 + }, + { + "epoch": 0.48656085366262875, + "grad_norm": 1.0728471279144287, + "learning_rate": 8.85371596445211e-05, + "loss": 0.725, + "step": 7250 + }, + { + "epoch": 0.4866950773463978, + "grad_norm": 1.1303855180740356, + "learning_rate": 8.853023403701894e-05, + "loss": 0.7265, + "step": 7252 + }, + { + "epoch": 0.48682930103016675, + "grad_norm": 0.9891423583030701, + "learning_rate": 8.852330660903899e-05, + "loss": 0.6167, + "step": 7254 + }, + { + "epoch": 0.4869635247139358, + "grad_norm": 1.1069655418395996, + "learning_rate": 8.851637736090857e-05, + "loss": 0.6709, + "step": 7256 + }, + { + "epoch": 0.48709774839770476, + "grad_norm": 1.0861634016036987, + "learning_rate": 8.850944629295503e-05, + "loss": 0.6759, + "step": 7258 + }, + { + "epoch": 0.4872319720814738, + "grad_norm": 1.3577293157577515, + "learning_rate": 8.850251340550591e-05, + "loss": 0.6849, + "step": 7260 + }, + { + "epoch": 0.48736619576524276, + "grad_norm": 1.1354912519454956, + "learning_rate": 8.849557869888872e-05, + "loss": 0.6488, + "step": 7262 + }, + { + "epoch": 0.4875004194490118, + "grad_norm": 1.0820255279541016, + "learning_rate": 8.848864217343114e-05, + "loss": 0.5862, + "step": 7264 + }, + { + "epoch": 0.48763464313278077, + "grad_norm": 0.9930956959724426, + "learning_rate": 8.848170382946091e-05, + "loss": 0.6174, + "step": 7266 + }, + { + "epoch": 0.4877688668165498, + "grad_norm": 1.0091296434402466, + "learning_rate": 8.847476366730584e-05, + "loss": 0.6347, + "step": 7268 + }, + { + "epoch": 0.48790309050031877, + "grad_norm": 0.9442898631095886, + "learning_rate": 8.846782168729384e-05, + "loss": 0.5684, + "step": 7270 + }, + { + "epoch": 0.4880373141840878, + "grad_norm": 1.0996882915496826, + "learning_rate": 8.846087788975292e-05, + "loss": 0.7144, + "step": 7272 + }, + { + "epoch": 0.4881715378678568, + "grad_norm": 1.119537591934204, + "learning_rate": 8.845393227501114e-05, + "loss": 0.6634, + "step": 7274 + }, + { + "epoch": 0.4883057615516258, + "grad_norm": 1.0286153554916382, + "learning_rate": 8.844698484339668e-05, + "loss": 0.6733, + "step": 7276 + }, + { + "epoch": 0.4884399852353948, + "grad_norm": 1.1285734176635742, + "learning_rate": 8.844003559523779e-05, + "loss": 0.6943, + "step": 7278 + }, + { + "epoch": 0.4885742089191638, + "grad_norm": 1.0654995441436768, + "learning_rate": 8.84330845308628e-05, + "loss": 0.6218, + "step": 7280 + }, + { + "epoch": 0.4887084326029328, + "grad_norm": 1.1435794830322266, + "learning_rate": 8.842613165060015e-05, + "loss": 0.6557, + "step": 7282 + }, + { + "epoch": 0.4888426562867018, + "grad_norm": 1.215554118156433, + "learning_rate": 8.841917695477835e-05, + "loss": 0.6423, + "step": 7284 + }, + { + "epoch": 0.4889768799704708, + "grad_norm": 1.0148378610610962, + "learning_rate": 8.841222044372597e-05, + "loss": 0.6413, + "step": 7286 + }, + { + "epoch": 0.48911110365423976, + "grad_norm": 0.9803034067153931, + "learning_rate": 8.840526211777172e-05, + "loss": 0.674, + "step": 7288 + }, + { + "epoch": 0.4892453273380088, + "grad_norm": 1.213284969329834, + "learning_rate": 8.839830197724435e-05, + "loss": 0.5822, + "step": 7290 + }, + { + "epoch": 0.48937955102177777, + "grad_norm": 1.2273319959640503, + "learning_rate": 8.839134002247272e-05, + "loss": 0.6526, + "step": 7292 + }, + { + "epoch": 0.4895137747055468, + "grad_norm": 0.9492382407188416, + "learning_rate": 8.838437625378575e-05, + "loss": 0.6438, + "step": 7294 + }, + { + "epoch": 0.4896479983893158, + "grad_norm": 1.0803555250167847, + "learning_rate": 8.83774106715125e-05, + "loss": 0.6604, + "step": 7296 + }, + { + "epoch": 0.4897822220730848, + "grad_norm": 1.1315208673477173, + "learning_rate": 8.837044327598206e-05, + "loss": 0.6355, + "step": 7298 + }, + { + "epoch": 0.4899164457568538, + "grad_norm": 1.0323400497436523, + "learning_rate": 8.836347406752363e-05, + "loss": 0.6169, + "step": 7300 + }, + { + "epoch": 0.4900506694406228, + "grad_norm": 1.054308295249939, + "learning_rate": 8.835650304646648e-05, + "loss": 0.6799, + "step": 7302 + }, + { + "epoch": 0.4901848931243918, + "grad_norm": 1.666773796081543, + "learning_rate": 8.834953021313999e-05, + "loss": 0.6705, + "step": 7304 + }, + { + "epoch": 0.4903191168081608, + "grad_norm": 1.113276720046997, + "learning_rate": 8.834255556787361e-05, + "loss": 0.7109, + "step": 7306 + }, + { + "epoch": 0.4904533404919298, + "grad_norm": 1.0723297595977783, + "learning_rate": 8.833557911099688e-05, + "loss": 0.743, + "step": 7308 + }, + { + "epoch": 0.4905875641756988, + "grad_norm": 0.982081413269043, + "learning_rate": 8.832860084283942e-05, + "loss": 0.6018, + "step": 7310 + }, + { + "epoch": 0.4907217878594678, + "grad_norm": 1.1384668350219727, + "learning_rate": 8.832162076373094e-05, + "loss": 0.649, + "step": 7312 + }, + { + "epoch": 0.4908560115432368, + "grad_norm": 1.1195060014724731, + "learning_rate": 8.831463887400122e-05, + "loss": 0.6211, + "step": 7314 + }, + { + "epoch": 0.4909902352270058, + "grad_norm": 1.0770231485366821, + "learning_rate": 8.830765517398017e-05, + "loss": 0.6519, + "step": 7316 + }, + { + "epoch": 0.49112445891077483, + "grad_norm": 1.090936541557312, + "learning_rate": 8.830066966399775e-05, + "loss": 0.7029, + "step": 7318 + }, + { + "epoch": 0.4912586825945438, + "grad_norm": 1.2828911542892456, + "learning_rate": 8.829368234438397e-05, + "loss": 0.5789, + "step": 7320 + }, + { + "epoch": 0.49139290627831284, + "grad_norm": 1.2686687707901, + "learning_rate": 8.828669321546902e-05, + "loss": 0.6519, + "step": 7322 + }, + { + "epoch": 0.4915271299620818, + "grad_norm": 1.336713194847107, + "learning_rate": 8.827970227758313e-05, + "loss": 0.713, + "step": 7324 + }, + { + "epoch": 0.4916613536458508, + "grad_norm": 1.0106137990951538, + "learning_rate": 8.827270953105655e-05, + "loss": 0.7171, + "step": 7326 + }, + { + "epoch": 0.4917955773296198, + "grad_norm": 1.1118881702423096, + "learning_rate": 8.826571497621972e-05, + "loss": 0.6424, + "step": 7328 + }, + { + "epoch": 0.4919298010133888, + "grad_norm": 0.9412743449211121, + "learning_rate": 8.825871861340308e-05, + "loss": 0.6053, + "step": 7330 + }, + { + "epoch": 0.4920640246971578, + "grad_norm": 1.0979979038238525, + "learning_rate": 8.825172044293725e-05, + "loss": 0.6517, + "step": 7332 + }, + { + "epoch": 0.4921982483809268, + "grad_norm": 1.1139260530471802, + "learning_rate": 8.824472046515283e-05, + "loss": 0.6765, + "step": 7334 + }, + { + "epoch": 0.4923324720646958, + "grad_norm": 1.2965672016143799, + "learning_rate": 8.823771868038058e-05, + "loss": 0.6566, + "step": 7336 + }, + { + "epoch": 0.4924666957484648, + "grad_norm": 1.1571414470672607, + "learning_rate": 8.823071508895131e-05, + "loss": 0.6023, + "step": 7338 + }, + { + "epoch": 0.49260091943223383, + "grad_norm": 1.5252083539962769, + "learning_rate": 8.822370969119592e-05, + "loss": 0.7048, + "step": 7340 + }, + { + "epoch": 0.4927351431160028, + "grad_norm": 1.0156404972076416, + "learning_rate": 8.821670248744542e-05, + "loss": 0.6371, + "step": 7342 + }, + { + "epoch": 0.49286936679977184, + "grad_norm": 1.6277072429656982, + "learning_rate": 8.820969347803088e-05, + "loss": 0.7344, + "step": 7344 + }, + { + "epoch": 0.4930035904835408, + "grad_norm": 1.3882527351379395, + "learning_rate": 8.820268266328345e-05, + "loss": 0.6687, + "step": 7346 + }, + { + "epoch": 0.49313781416730984, + "grad_norm": 1.1527955532073975, + "learning_rate": 8.81956700435344e-05, + "loss": 0.6568, + "step": 7348 + }, + { + "epoch": 0.4932720378510788, + "grad_norm": 1.0824893712997437, + "learning_rate": 8.818865561911504e-05, + "loss": 0.6134, + "step": 7350 + }, + { + "epoch": 0.49340626153484785, + "grad_norm": 1.1220204830169678, + "learning_rate": 8.81816393903568e-05, + "loss": 0.7174, + "step": 7352 + }, + { + "epoch": 0.4935404852186168, + "grad_norm": 1.1143301725387573, + "learning_rate": 8.817462135759117e-05, + "loss": 0.6575, + "step": 7354 + }, + { + "epoch": 0.49367470890238585, + "grad_norm": 1.1951769590377808, + "learning_rate": 8.816760152114976e-05, + "loss": 0.679, + "step": 7356 + }, + { + "epoch": 0.4938089325861548, + "grad_norm": 1.0683196783065796, + "learning_rate": 8.816057988136422e-05, + "loss": 0.6011, + "step": 7358 + }, + { + "epoch": 0.49394315626992386, + "grad_norm": 1.1153167486190796, + "learning_rate": 8.815355643856633e-05, + "loss": 0.6792, + "step": 7360 + }, + { + "epoch": 0.49407737995369283, + "grad_norm": 1.0507798194885254, + "learning_rate": 8.814653119308794e-05, + "loss": 0.6355, + "step": 7362 + }, + { + "epoch": 0.4942116036374618, + "grad_norm": 1.233003854751587, + "learning_rate": 8.813950414526093e-05, + "loss": 0.6855, + "step": 7364 + }, + { + "epoch": 0.49434582732123084, + "grad_norm": 0.9083017706871033, + "learning_rate": 8.813247529541737e-05, + "loss": 0.651, + "step": 7366 + }, + { + "epoch": 0.4944800510049998, + "grad_norm": 1.1303188800811768, + "learning_rate": 8.812544464388932e-05, + "loss": 0.7534, + "step": 7368 + }, + { + "epoch": 0.49461427468876884, + "grad_norm": 1.0329996347427368, + "learning_rate": 8.8118412191009e-05, + "loss": 0.5869, + "step": 7370 + }, + { + "epoch": 0.4947484983725378, + "grad_norm": 1.0308709144592285, + "learning_rate": 8.811137793710863e-05, + "loss": 0.6149, + "step": 7372 + }, + { + "epoch": 0.49488272205630685, + "grad_norm": 1.163661241531372, + "learning_rate": 8.810434188252063e-05, + "loss": 0.7506, + "step": 7374 + }, + { + "epoch": 0.4950169457400758, + "grad_norm": 1.409184217453003, + "learning_rate": 8.809730402757739e-05, + "loss": 0.5969, + "step": 7376 + }, + { + "epoch": 0.49515116942384485, + "grad_norm": 1.0647064447402954, + "learning_rate": 8.809026437261145e-05, + "loss": 0.6521, + "step": 7378 + }, + { + "epoch": 0.4952853931076138, + "grad_norm": 1.0129282474517822, + "learning_rate": 8.80832229179554e-05, + "loss": 0.6498, + "step": 7380 + }, + { + "epoch": 0.49541961679138286, + "grad_norm": 1.0250266790390015, + "learning_rate": 8.8076179663942e-05, + "loss": 0.6595, + "step": 7382 + }, + { + "epoch": 0.49555384047515183, + "grad_norm": 1.0052516460418701, + "learning_rate": 8.806913461090395e-05, + "loss": 0.6798, + "step": 7384 + }, + { + "epoch": 0.49568806415892086, + "grad_norm": 0.9906535744667053, + "learning_rate": 8.806208775917417e-05, + "loss": 0.6435, + "step": 7386 + }, + { + "epoch": 0.49582228784268984, + "grad_norm": 1.077197551727295, + "learning_rate": 8.805503910908557e-05, + "loss": 0.6533, + "step": 7388 + }, + { + "epoch": 0.49595651152645887, + "grad_norm": 1.6716821193695068, + "learning_rate": 8.804798866097121e-05, + "loss": 0.6529, + "step": 7390 + }, + { + "epoch": 0.49609073521022784, + "grad_norm": 0.9911879301071167, + "learning_rate": 8.80409364151642e-05, + "loss": 0.6959, + "step": 7392 + }, + { + "epoch": 0.49622495889399687, + "grad_norm": 1.0691368579864502, + "learning_rate": 8.803388237199776e-05, + "loss": 0.6926, + "step": 7394 + }, + { + "epoch": 0.49635918257776585, + "grad_norm": 0.9496676921844482, + "learning_rate": 8.802682653180516e-05, + "loss": 0.5831, + "step": 7396 + }, + { + "epoch": 0.4964934062615349, + "grad_norm": 1.2202308177947998, + "learning_rate": 8.801976889491979e-05, + "loss": 0.6378, + "step": 7398 + }, + { + "epoch": 0.49662762994530385, + "grad_norm": 1.1562025547027588, + "learning_rate": 8.80127094616751e-05, + "loss": 0.6617, + "step": 7400 + }, + { + "epoch": 0.4967618536290728, + "grad_norm": 0.94105064868927, + "learning_rate": 8.800564823240464e-05, + "loss": 0.6513, + "step": 7402 + }, + { + "epoch": 0.49689607731284186, + "grad_norm": 1.05088472366333, + "learning_rate": 8.799858520744201e-05, + "loss": 0.6262, + "step": 7404 + }, + { + "epoch": 0.49703030099661083, + "grad_norm": 1.0232298374176025, + "learning_rate": 8.799152038712099e-05, + "loss": 0.5954, + "step": 7406 + }, + { + "epoch": 0.49716452468037986, + "grad_norm": 1.0250871181488037, + "learning_rate": 8.798445377177531e-05, + "loss": 0.6446, + "step": 7408 + }, + { + "epoch": 0.49729874836414883, + "grad_norm": 0.9300850033760071, + "learning_rate": 8.79773853617389e-05, + "loss": 0.6087, + "step": 7410 + }, + { + "epoch": 0.49743297204791787, + "grad_norm": 1.0377005338668823, + "learning_rate": 8.797031515734571e-05, + "loss": 0.6592, + "step": 7412 + }, + { + "epoch": 0.49756719573168684, + "grad_norm": 1.0114991664886475, + "learning_rate": 8.796324315892978e-05, + "loss": 0.6064, + "step": 7414 + }, + { + "epoch": 0.49770141941545587, + "grad_norm": 0.962981641292572, + "learning_rate": 8.795616936682528e-05, + "loss": 0.6048, + "step": 7416 + }, + { + "epoch": 0.49783564309922484, + "grad_norm": 1.3811211585998535, + "learning_rate": 8.794909378136639e-05, + "loss": 0.6492, + "step": 7418 + }, + { + "epoch": 0.4979698667829939, + "grad_norm": 0.9624200463294983, + "learning_rate": 8.794201640288746e-05, + "loss": 0.6221, + "step": 7420 + }, + { + "epoch": 0.49810409046676285, + "grad_norm": 1.0491639375686646, + "learning_rate": 8.793493723172286e-05, + "loss": 0.6212, + "step": 7422 + }, + { + "epoch": 0.4982383141505319, + "grad_norm": 1.1057180166244507, + "learning_rate": 8.792785626820708e-05, + "loss": 0.6459, + "step": 7424 + }, + { + "epoch": 0.49837253783430085, + "grad_norm": 1.0114151239395142, + "learning_rate": 8.792077351267466e-05, + "loss": 0.6959, + "step": 7426 + }, + { + "epoch": 0.4985067615180699, + "grad_norm": 1.1589282751083374, + "learning_rate": 8.791368896546027e-05, + "loss": 0.6812, + "step": 7428 + }, + { + "epoch": 0.49864098520183886, + "grad_norm": 1.2169181108474731, + "learning_rate": 8.790660262689863e-05, + "loss": 0.7807, + "step": 7430 + }, + { + "epoch": 0.4987752088856079, + "grad_norm": 1.1884033679962158, + "learning_rate": 8.789951449732454e-05, + "loss": 0.6644, + "step": 7432 + }, + { + "epoch": 0.49890943256937687, + "grad_norm": 1.1186769008636475, + "learning_rate": 8.789242457707294e-05, + "loss": 0.6636, + "step": 7434 + }, + { + "epoch": 0.4990436562531459, + "grad_norm": 1.0410146713256836, + "learning_rate": 8.788533286647878e-05, + "loss": 0.587, + "step": 7436 + }, + { + "epoch": 0.49917787993691487, + "grad_norm": 1.07662832736969, + "learning_rate": 8.787823936587714e-05, + "loss": 0.6532, + "step": 7438 + }, + { + "epoch": 0.49931210362068384, + "grad_norm": 0.859747588634491, + "learning_rate": 8.787114407560317e-05, + "loss": 0.6396, + "step": 7440 + }, + { + "epoch": 0.4994463273044529, + "grad_norm": 1.036584496498108, + "learning_rate": 8.786404699599211e-05, + "loss": 0.7043, + "step": 7442 + }, + { + "epoch": 0.49958055098822185, + "grad_norm": 1.0626755952835083, + "learning_rate": 8.78569481273793e-05, + "loss": 0.6924, + "step": 7444 + }, + { + "epoch": 0.4997147746719909, + "grad_norm": 1.1623696088790894, + "learning_rate": 8.784984747010012e-05, + "loss": 0.6752, + "step": 7446 + }, + { + "epoch": 0.49984899835575985, + "grad_norm": 1.1692180633544922, + "learning_rate": 8.784274502449009e-05, + "loss": 0.6518, + "step": 7448 + }, + { + "epoch": 0.4999832220395289, + "grad_norm": 1.004929780960083, + "learning_rate": 8.783564079088477e-05, + "loss": 0.6102, + "step": 7450 + }, + { + "epoch": 0.5001174457232979, + "grad_norm": 1.0000412464141846, + "learning_rate": 8.782853476961981e-05, + "loss": 0.626, + "step": 7452 + }, + { + "epoch": 0.5002516694070669, + "grad_norm": 1.0467336177825928, + "learning_rate": 8.782142696103098e-05, + "loss": 0.675, + "step": 7454 + }, + { + "epoch": 0.5003858930908359, + "grad_norm": 1.0724281072616577, + "learning_rate": 8.781431736545409e-05, + "loss": 0.6264, + "step": 7456 + }, + { + "epoch": 0.5005201167746048, + "grad_norm": 0.9236459136009216, + "learning_rate": 8.780720598322507e-05, + "loss": 0.612, + "step": 7458 + }, + { + "epoch": 0.5006543404583739, + "grad_norm": 1.2468255758285522, + "learning_rate": 8.780009281467991e-05, + "loss": 0.669, + "step": 7460 + }, + { + "epoch": 0.5007885641421429, + "grad_norm": 1.2404422760009766, + "learning_rate": 8.77929778601547e-05, + "loss": 0.6444, + "step": 7462 + }, + { + "epoch": 0.5009227878259119, + "grad_norm": 0.9682061076164246, + "learning_rate": 8.778586111998561e-05, + "loss": 0.5992, + "step": 7464 + }, + { + "epoch": 0.5010570115096808, + "grad_norm": 1.0936394929885864, + "learning_rate": 8.777874259450887e-05, + "loss": 0.649, + "step": 7466 + }, + { + "epoch": 0.5011912351934499, + "grad_norm": 1.0420470237731934, + "learning_rate": 8.777162228406083e-05, + "loss": 0.6183, + "step": 7468 + }, + { + "epoch": 0.5013254588772189, + "grad_norm": 1.0366129875183105, + "learning_rate": 8.776450018897792e-05, + "loss": 0.6587, + "step": 7470 + }, + { + "epoch": 0.5014596825609879, + "grad_norm": 1.1497166156768799, + "learning_rate": 8.775737630959662e-05, + "loss": 0.6408, + "step": 7472 + }, + { + "epoch": 0.5015939062447569, + "grad_norm": 0.9807114601135254, + "learning_rate": 8.775025064625355e-05, + "loss": 0.6383, + "step": 7474 + }, + { + "epoch": 0.5017281299285259, + "grad_norm": 1.0735164880752563, + "learning_rate": 8.774312319928537e-05, + "loss": 0.6411, + "step": 7476 + }, + { + "epoch": 0.5018623536122949, + "grad_norm": 0.995513916015625, + "learning_rate": 8.773599396902886e-05, + "loss": 0.6344, + "step": 7478 + }, + { + "epoch": 0.5019965772960638, + "grad_norm": 0.891380786895752, + "learning_rate": 8.772886295582079e-05, + "loss": 0.5634, + "step": 7480 + }, + { + "epoch": 0.5021308009798329, + "grad_norm": 1.1450775861740112, + "learning_rate": 8.772173015999816e-05, + "loss": 0.5742, + "step": 7482 + }, + { + "epoch": 0.5022650246636019, + "grad_norm": 0.9486705660820007, + "learning_rate": 8.771459558189796e-05, + "loss": 0.6125, + "step": 7484 + }, + { + "epoch": 0.5023992483473709, + "grad_norm": 1.1187535524368286, + "learning_rate": 8.770745922185728e-05, + "loss": 0.64, + "step": 7486 + }, + { + "epoch": 0.5025334720311398, + "grad_norm": 1.4808590412139893, + "learning_rate": 8.770032108021331e-05, + "loss": 0.671, + "step": 7488 + }, + { + "epoch": 0.5026676957149089, + "grad_norm": 1.0867255926132202, + "learning_rate": 8.76931811573033e-05, + "loss": 0.6704, + "step": 7490 + }, + { + "epoch": 0.5028019193986779, + "grad_norm": 0.9657959342002869, + "learning_rate": 8.768603945346458e-05, + "loss": 0.6457, + "step": 7492 + }, + { + "epoch": 0.5029361430824469, + "grad_norm": 1.0877447128295898, + "learning_rate": 8.767889596903462e-05, + "loss": 0.6815, + "step": 7494 + }, + { + "epoch": 0.5030703667662159, + "grad_norm": 1.057449221611023, + "learning_rate": 8.767175070435092e-05, + "loss": 0.6401, + "step": 7496 + }, + { + "epoch": 0.5032045904499849, + "grad_norm": 5.432621002197266, + "learning_rate": 8.766460365975107e-05, + "loss": 0.5941, + "step": 7498 + }, + { + "epoch": 0.5033388141337539, + "grad_norm": 1.0450407266616821, + "learning_rate": 8.765745483557276e-05, + "loss": 0.671, + "step": 7500 + }, + { + "epoch": 0.503473037817523, + "grad_norm": 0.9949643015861511, + "learning_rate": 8.765030423215377e-05, + "loss": 0.6221, + "step": 7502 + }, + { + "epoch": 0.5036072615012919, + "grad_norm": 1.0133390426635742, + "learning_rate": 8.764315184983193e-05, + "loss": 0.6609, + "step": 7504 + }, + { + "epoch": 0.5037414851850609, + "grad_norm": 1.0307201147079468, + "learning_rate": 8.76359976889452e-05, + "loss": 0.6813, + "step": 7506 + }, + { + "epoch": 0.5038757088688299, + "grad_norm": 1.0315346717834473, + "learning_rate": 8.762884174983158e-05, + "loss": 0.6635, + "step": 7508 + }, + { + "epoch": 0.504009932552599, + "grad_norm": 1.23046875, + "learning_rate": 8.762168403282917e-05, + "loss": 0.7086, + "step": 7510 + }, + { + "epoch": 0.5041441562363679, + "grad_norm": 1.2309722900390625, + "learning_rate": 8.761452453827618e-05, + "loss": 0.6886, + "step": 7512 + }, + { + "epoch": 0.5042783799201369, + "grad_norm": 0.9779922962188721, + "learning_rate": 8.760736326651087e-05, + "loss": 0.6119, + "step": 7514 + }, + { + "epoch": 0.5044126036039059, + "grad_norm": 0.9598320126533508, + "learning_rate": 8.760020021787158e-05, + "loss": 0.62, + "step": 7516 + }, + { + "epoch": 0.5045468272876749, + "grad_norm": 2.2394909858703613, + "learning_rate": 8.759303539269679e-05, + "loss": 0.6471, + "step": 7518 + }, + { + "epoch": 0.5046810509714439, + "grad_norm": 1.0633375644683838, + "learning_rate": 8.7585868791325e-05, + "loss": 0.6494, + "step": 7520 + }, + { + "epoch": 0.5048152746552129, + "grad_norm": 1.1028856039047241, + "learning_rate": 8.757870041409481e-05, + "loss": 0.6536, + "step": 7522 + }, + { + "epoch": 0.504949498338982, + "grad_norm": 1.4106289148330688, + "learning_rate": 8.757153026134493e-05, + "loss": 0.6635, + "step": 7524 + }, + { + "epoch": 0.5050837220227509, + "grad_norm": 1.292608618736267, + "learning_rate": 8.756435833341412e-05, + "loss": 0.7536, + "step": 7526 + }, + { + "epoch": 0.5052179457065199, + "grad_norm": 1.009282112121582, + "learning_rate": 8.755718463064125e-05, + "loss": 0.601, + "step": 7528 + }, + { + "epoch": 0.5053521693902889, + "grad_norm": 0.946381688117981, + "learning_rate": 8.755000915336527e-05, + "loss": 0.6527, + "step": 7530 + }, + { + "epoch": 0.505486393074058, + "grad_norm": 1.0703035593032837, + "learning_rate": 8.754283190192516e-05, + "loss": 0.6132, + "step": 7532 + }, + { + "epoch": 0.5056206167578269, + "grad_norm": 1.0881116390228271, + "learning_rate": 8.75356528766601e-05, + "loss": 0.6117, + "step": 7534 + }, + { + "epoch": 0.5057548404415959, + "grad_norm": 1.224095344543457, + "learning_rate": 8.752847207790924e-05, + "loss": 0.6924, + "step": 7536 + }, + { + "epoch": 0.5058890641253649, + "grad_norm": 1.0547585487365723, + "learning_rate": 8.752128950601189e-05, + "loss": 0.6394, + "step": 7538 + }, + { + "epoch": 0.506023287809134, + "grad_norm": 1.133722186088562, + "learning_rate": 8.751410516130739e-05, + "loss": 0.5899, + "step": 7540 + }, + { + "epoch": 0.5061575114929029, + "grad_norm": 1.1165887117385864, + "learning_rate": 8.750691904413515e-05, + "loss": 0.7256, + "step": 7542 + }, + { + "epoch": 0.5062917351766719, + "grad_norm": 1.232200026512146, + "learning_rate": 8.749973115483478e-05, + "loss": 0.6446, + "step": 7544 + }, + { + "epoch": 0.506425958860441, + "grad_norm": 1.9479700326919556, + "learning_rate": 8.749254149374583e-05, + "loss": 0.6947, + "step": 7546 + }, + { + "epoch": 0.50656018254421, + "grad_norm": 0.9247474074363708, + "learning_rate": 8.748535006120804e-05, + "loss": 0.6398, + "step": 7548 + }, + { + "epoch": 0.5066944062279789, + "grad_norm": 1.1821279525756836, + "learning_rate": 8.747815685756115e-05, + "loss": 0.6105, + "step": 7550 + }, + { + "epoch": 0.5068286299117479, + "grad_norm": 1.039718508720398, + "learning_rate": 8.747096188314506e-05, + "loss": 0.6645, + "step": 7552 + }, + { + "epoch": 0.506962853595517, + "grad_norm": 1.0514421463012695, + "learning_rate": 8.74637651382997e-05, + "loss": 0.694, + "step": 7554 + }, + { + "epoch": 0.5070970772792859, + "grad_norm": 1.110547423362732, + "learning_rate": 8.745656662336511e-05, + "loss": 0.6288, + "step": 7556 + }, + { + "epoch": 0.5072313009630549, + "grad_norm": 1.2142102718353271, + "learning_rate": 8.744936633868139e-05, + "loss": 0.6727, + "step": 7558 + }, + { + "epoch": 0.5073655246468239, + "grad_norm": 0.9618511199951172, + "learning_rate": 8.744216428458878e-05, + "loss": 0.6716, + "step": 7560 + }, + { + "epoch": 0.507499748330593, + "grad_norm": 0.9959256052970886, + "learning_rate": 8.743496046142752e-05, + "loss": 0.6373, + "step": 7562 + }, + { + "epoch": 0.5076339720143619, + "grad_norm": 1.234066128730774, + "learning_rate": 8.742775486953799e-05, + "loss": 0.5901, + "step": 7564 + }, + { + "epoch": 0.5077681956981309, + "grad_norm": 1.105370044708252, + "learning_rate": 8.742054750926063e-05, + "loss": 0.6654, + "step": 7566 + }, + { + "epoch": 0.5079024193819, + "grad_norm": 1.0729789733886719, + "learning_rate": 8.7413338380936e-05, + "loss": 0.7325, + "step": 7568 + }, + { + "epoch": 0.508036643065669, + "grad_norm": 1.02970290184021, + "learning_rate": 8.740612748490468e-05, + "loss": 0.6863, + "step": 7570 + }, + { + "epoch": 0.5081708667494379, + "grad_norm": 1.0629090070724487, + "learning_rate": 8.739891482150741e-05, + "loss": 0.6224, + "step": 7572 + }, + { + "epoch": 0.5083050904332069, + "grad_norm": 1.0690653324127197, + "learning_rate": 8.739170039108495e-05, + "loss": 0.6694, + "step": 7574 + }, + { + "epoch": 0.508439314116976, + "grad_norm": 1.087717056274414, + "learning_rate": 8.738448419397818e-05, + "loss": 0.6956, + "step": 7576 + }, + { + "epoch": 0.508573537800745, + "grad_norm": 0.9817718863487244, + "learning_rate": 8.737726623052805e-05, + "loss": 0.6292, + "step": 7578 + }, + { + "epoch": 0.5087077614845139, + "grad_norm": 2.2005228996276855, + "learning_rate": 8.737004650107557e-05, + "loss": 0.6386, + "step": 7580 + }, + { + "epoch": 0.5088419851682829, + "grad_norm": 1.0479660034179688, + "learning_rate": 8.736282500596188e-05, + "loss": 0.6512, + "step": 7582 + }, + { + "epoch": 0.508976208852052, + "grad_norm": 1.2043403387069702, + "learning_rate": 8.735560174552821e-05, + "loss": 0.7167, + "step": 7584 + }, + { + "epoch": 0.509110432535821, + "grad_norm": 1.121976613998413, + "learning_rate": 8.73483767201158e-05, + "loss": 0.7294, + "step": 7586 + }, + { + "epoch": 0.5092446562195899, + "grad_norm": 1.1195271015167236, + "learning_rate": 8.734114993006602e-05, + "loss": 0.6299, + "step": 7588 + }, + { + "epoch": 0.509378879903359, + "grad_norm": 1.4202194213867188, + "learning_rate": 8.733392137572033e-05, + "loss": 0.6179, + "step": 7590 + }, + { + "epoch": 0.509513103587128, + "grad_norm": 1.0699143409729004, + "learning_rate": 8.732669105742028e-05, + "loss": 0.6533, + "step": 7592 + }, + { + "epoch": 0.5096473272708969, + "grad_norm": 1.2938518524169922, + "learning_rate": 8.731945897550748e-05, + "loss": 0.6644, + "step": 7594 + }, + { + "epoch": 0.5097815509546659, + "grad_norm": 1.3335658311843872, + "learning_rate": 8.731222513032362e-05, + "loss": 0.6138, + "step": 7596 + }, + { + "epoch": 0.509915774638435, + "grad_norm": 1.1592415571212769, + "learning_rate": 8.730498952221051e-05, + "loss": 0.6655, + "step": 7598 + }, + { + "epoch": 0.510049998322204, + "grad_norm": 1.2052128314971924, + "learning_rate": 8.729775215151e-05, + "loss": 0.6133, + "step": 7600 + }, + { + "epoch": 0.5101842220059729, + "grad_norm": 1.1826194524765015, + "learning_rate": 8.729051301856402e-05, + "loss": 0.5907, + "step": 7602 + }, + { + "epoch": 0.5103184456897419, + "grad_norm": 1.1111299991607666, + "learning_rate": 8.728327212371466e-05, + "loss": 0.7097, + "step": 7604 + }, + { + "epoch": 0.510452669373511, + "grad_norm": 1.5861774682998657, + "learning_rate": 8.727602946730397e-05, + "loss": 0.6082, + "step": 7606 + }, + { + "epoch": 0.51058689305728, + "grad_norm": 1.4667607545852661, + "learning_rate": 8.726878504967421e-05, + "loss": 0.667, + "step": 7608 + }, + { + "epoch": 0.5107211167410489, + "grad_norm": 0.9902428388595581, + "learning_rate": 8.726153887116766e-05, + "loss": 0.6366, + "step": 7610 + }, + { + "epoch": 0.510855340424818, + "grad_norm": 0.9585524201393127, + "learning_rate": 8.725429093212664e-05, + "loss": 0.6093, + "step": 7612 + }, + { + "epoch": 0.510989564108587, + "grad_norm": 1.0268088579177856, + "learning_rate": 8.724704123289365e-05, + "loss": 0.6577, + "step": 7614 + }, + { + "epoch": 0.511123787792356, + "grad_norm": 0.9739425778388977, + "learning_rate": 8.723978977381119e-05, + "loss": 0.6113, + "step": 7616 + }, + { + "epoch": 0.5112580114761249, + "grad_norm": 1.6853216886520386, + "learning_rate": 8.72325365552219e-05, + "loss": 0.6764, + "step": 7618 + }, + { + "epoch": 0.511392235159894, + "grad_norm": 1.0694867372512817, + "learning_rate": 8.722528157746847e-05, + "loss": 0.6613, + "step": 7620 + }, + { + "epoch": 0.511526458843663, + "grad_norm": 1.4094675779342651, + "learning_rate": 8.721802484089368e-05, + "loss": 0.6777, + "step": 7622 + }, + { + "epoch": 0.511660682527432, + "grad_norm": 1.1949818134307861, + "learning_rate": 8.721076634584042e-05, + "loss": 0.6498, + "step": 7624 + }, + { + "epoch": 0.5117949062112009, + "grad_norm": 1.0607755184173584, + "learning_rate": 8.72035060926516e-05, + "loss": 0.6183, + "step": 7626 + }, + { + "epoch": 0.51192912989497, + "grad_norm": 1.1353689432144165, + "learning_rate": 8.719624408167029e-05, + "loss": 0.5894, + "step": 7628 + }, + { + "epoch": 0.512063353578739, + "grad_norm": 0.9764440059661865, + "learning_rate": 8.71889803132396e-05, + "loss": 0.6442, + "step": 7630 + }, + { + "epoch": 0.5121975772625079, + "grad_norm": 1.0443223714828491, + "learning_rate": 8.718171478770269e-05, + "loss": 0.6744, + "step": 7632 + }, + { + "epoch": 0.512331800946277, + "grad_norm": 1.0360503196716309, + "learning_rate": 8.71744475054029e-05, + "loss": 0.6924, + "step": 7634 + }, + { + "epoch": 0.512466024630046, + "grad_norm": 1.087961196899414, + "learning_rate": 8.716717846668354e-05, + "loss": 0.6285, + "step": 7636 + }, + { + "epoch": 0.512600248313815, + "grad_norm": 1.0620614290237427, + "learning_rate": 8.715990767188811e-05, + "loss": 0.6147, + "step": 7638 + }, + { + "epoch": 0.5127344719975839, + "grad_norm": 1.312974214553833, + "learning_rate": 8.71526351213601e-05, + "loss": 0.6432, + "step": 7640 + }, + { + "epoch": 0.512868695681353, + "grad_norm": 1.0307996273040771, + "learning_rate": 8.714536081544314e-05, + "loss": 0.5833, + "step": 7642 + }, + { + "epoch": 0.513002919365122, + "grad_norm": 1.0482127666473389, + "learning_rate": 8.71380847544809e-05, + "loss": 0.6352, + "step": 7644 + }, + { + "epoch": 0.513137143048891, + "grad_norm": 1.0585205554962158, + "learning_rate": 8.713080693881723e-05, + "loss": 0.6624, + "step": 7646 + }, + { + "epoch": 0.5132713667326599, + "grad_norm": 1.017253041267395, + "learning_rate": 8.712352736879592e-05, + "loss": 0.7177, + "step": 7648 + }, + { + "epoch": 0.513405590416429, + "grad_norm": 0.938526451587677, + "learning_rate": 8.711624604476094e-05, + "loss": 0.6576, + "step": 7650 + }, + { + "epoch": 0.513539814100198, + "grad_norm": 1.0195128917694092, + "learning_rate": 8.710896296705634e-05, + "loss": 0.6401, + "step": 7652 + }, + { + "epoch": 0.513674037783967, + "grad_norm": 1.232199788093567, + "learning_rate": 8.710167813602619e-05, + "loss": 0.684, + "step": 7654 + }, + { + "epoch": 0.513808261467736, + "grad_norm": 1.0564061403274536, + "learning_rate": 8.70943915520147e-05, + "loss": 0.6007, + "step": 7656 + }, + { + "epoch": 0.513942485151505, + "grad_norm": 0.9898374676704407, + "learning_rate": 8.708710321536617e-05, + "loss": 0.5945, + "step": 7658 + }, + { + "epoch": 0.514076708835274, + "grad_norm": 1.0006496906280518, + "learning_rate": 8.707981312642494e-05, + "loss": 0.6414, + "step": 7660 + }, + { + "epoch": 0.514210932519043, + "grad_norm": 1.2971662282943726, + "learning_rate": 8.707252128553544e-05, + "loss": 0.6242, + "step": 7662 + }, + { + "epoch": 0.514345156202812, + "grad_norm": 1.0045400857925415, + "learning_rate": 8.706522769304224e-05, + "loss": 0.6339, + "step": 7664 + }, + { + "epoch": 0.514479379886581, + "grad_norm": 0.9474780559539795, + "learning_rate": 8.70579323492899e-05, + "loss": 0.6462, + "step": 7666 + }, + { + "epoch": 0.51461360357035, + "grad_norm": 1.4041409492492676, + "learning_rate": 8.705063525462312e-05, + "loss": 0.6008, + "step": 7668 + }, + { + "epoch": 0.5147478272541189, + "grad_norm": 0.9978337287902832, + "learning_rate": 8.704333640938669e-05, + "loss": 0.637, + "step": 7670 + }, + { + "epoch": 0.514882050937888, + "grad_norm": 1.0927438735961914, + "learning_rate": 8.703603581392546e-05, + "loss": 0.646, + "step": 7672 + }, + { + "epoch": 0.515016274621657, + "grad_norm": 1.1463407278060913, + "learning_rate": 8.702873346858434e-05, + "loss": 0.6386, + "step": 7674 + }, + { + "epoch": 0.515150498305426, + "grad_norm": 0.9773491024971008, + "learning_rate": 8.70214293737084e-05, + "loss": 0.6463, + "step": 7676 + }, + { + "epoch": 0.515284721989195, + "grad_norm": 1.1640729904174805, + "learning_rate": 8.701412352964274e-05, + "loss": 0.6623, + "step": 7678 + }, + { + "epoch": 0.515418945672964, + "grad_norm": 1.3043346405029297, + "learning_rate": 8.70068159367325e-05, + "loss": 0.6581, + "step": 7680 + }, + { + "epoch": 0.515553169356733, + "grad_norm": 1.5669822692871094, + "learning_rate": 8.699950659532298e-05, + "loss": 0.6526, + "step": 7682 + }, + { + "epoch": 0.515687393040502, + "grad_norm": 1.0059089660644531, + "learning_rate": 8.699219550575953e-05, + "loss": 0.6044, + "step": 7684 + }, + { + "epoch": 0.515821616724271, + "grad_norm": 1.0378648042678833, + "learning_rate": 8.698488266838759e-05, + "loss": 0.6286, + "step": 7686 + }, + { + "epoch": 0.51595584040804, + "grad_norm": 1.1167110204696655, + "learning_rate": 8.697756808355267e-05, + "loss": 0.634, + "step": 7688 + }, + { + "epoch": 0.516090064091809, + "grad_norm": 1.049477219581604, + "learning_rate": 8.697025175160039e-05, + "loss": 0.5723, + "step": 7690 + }, + { + "epoch": 0.516224287775578, + "grad_norm": 1.1996572017669678, + "learning_rate": 8.696293367287638e-05, + "loss": 0.7119, + "step": 7692 + }, + { + "epoch": 0.516358511459347, + "grad_norm": 1.0795083045959473, + "learning_rate": 8.695561384772646e-05, + "loss": 0.6275, + "step": 7694 + }, + { + "epoch": 0.516492735143116, + "grad_norm": 1.0214886665344238, + "learning_rate": 8.694829227649644e-05, + "loss": 0.6757, + "step": 7696 + }, + { + "epoch": 0.516626958826885, + "grad_norm": 1.1827250719070435, + "learning_rate": 8.694096895953227e-05, + "loss": 0.6685, + "step": 7698 + }, + { + "epoch": 0.5167611825106541, + "grad_norm": 1.0351839065551758, + "learning_rate": 8.693364389717998e-05, + "loss": 0.6215, + "step": 7700 + }, + { + "epoch": 0.516895406194423, + "grad_norm": 1.0110132694244385, + "learning_rate": 8.692631708978562e-05, + "loss": 0.6475, + "step": 7702 + }, + { + "epoch": 0.517029629878192, + "grad_norm": 1.1591826677322388, + "learning_rate": 8.69189885376954e-05, + "loss": 0.6171, + "step": 7704 + }, + { + "epoch": 0.517163853561961, + "grad_norm": 1.0220993757247925, + "learning_rate": 8.691165824125556e-05, + "loss": 0.6669, + "step": 7706 + }, + { + "epoch": 0.51729807724573, + "grad_norm": 1.1127986907958984, + "learning_rate": 8.690432620081246e-05, + "loss": 0.6426, + "step": 7708 + }, + { + "epoch": 0.517432300929499, + "grad_norm": 1.0023584365844727, + "learning_rate": 8.689699241671252e-05, + "loss": 0.5993, + "step": 7710 + }, + { + "epoch": 0.517566524613268, + "grad_norm": 1.3513555526733398, + "learning_rate": 8.688965688930224e-05, + "loss": 0.5884, + "step": 7712 + }, + { + "epoch": 0.517700748297037, + "grad_norm": 1.107893466949463, + "learning_rate": 8.68823196189282e-05, + "loss": 0.6851, + "step": 7714 + }, + { + "epoch": 0.517834971980806, + "grad_norm": 1.0814416408538818, + "learning_rate": 8.68749806059371e-05, + "loss": 0.7228, + "step": 7716 + }, + { + "epoch": 0.517969195664575, + "grad_norm": 1.3134574890136719, + "learning_rate": 8.686763985067568e-05, + "loss": 0.7092, + "step": 7718 + }, + { + "epoch": 0.518103419348344, + "grad_norm": 1.3009024858474731, + "learning_rate": 8.686029735349075e-05, + "loss": 0.6243, + "step": 7720 + }, + { + "epoch": 0.5182376430321131, + "grad_norm": 1.1668957471847534, + "learning_rate": 8.685295311472927e-05, + "loss": 0.6548, + "step": 7722 + }, + { + "epoch": 0.518371866715882, + "grad_norm": 1.0200467109680176, + "learning_rate": 8.684560713473822e-05, + "loss": 0.6426, + "step": 7724 + }, + { + "epoch": 0.518506090399651, + "grad_norm": 1.109863042831421, + "learning_rate": 8.68382594138647e-05, + "loss": 0.5925, + "step": 7726 + }, + { + "epoch": 0.51864031408342, + "grad_norm": 1.043623447418213, + "learning_rate": 8.683090995245584e-05, + "loss": 0.6926, + "step": 7728 + }, + { + "epoch": 0.5187745377671891, + "grad_norm": 1.005377173423767, + "learning_rate": 8.682355875085893e-05, + "loss": 0.5992, + "step": 7730 + }, + { + "epoch": 0.518908761450958, + "grad_norm": 1.0924162864685059, + "learning_rate": 8.681620580942127e-05, + "loss": 0.6099, + "step": 7732 + }, + { + "epoch": 0.519042985134727, + "grad_norm": 1.9004437923431396, + "learning_rate": 8.680885112849028e-05, + "loss": 0.6336, + "step": 7734 + }, + { + "epoch": 0.519177208818496, + "grad_norm": 1.0223441123962402, + "learning_rate": 8.680149470841346e-05, + "loss": 0.6434, + "step": 7736 + }, + { + "epoch": 0.5193114325022651, + "grad_norm": 1.0525568723678589, + "learning_rate": 8.679413654953837e-05, + "loss": 0.6352, + "step": 7738 + }, + { + "epoch": 0.519445656186034, + "grad_norm": 0.9689558744430542, + "learning_rate": 8.678677665221268e-05, + "loss": 0.5739, + "step": 7740 + }, + { + "epoch": 0.519579879869803, + "grad_norm": 0.9976610541343689, + "learning_rate": 8.677941501678415e-05, + "loss": 0.6204, + "step": 7742 + }, + { + "epoch": 0.5197141035535721, + "grad_norm": 1.1912943124771118, + "learning_rate": 8.677205164360059e-05, + "loss": 0.7589, + "step": 7744 + }, + { + "epoch": 0.519848327237341, + "grad_norm": 1.831667423248291, + "learning_rate": 8.676468653300987e-05, + "loss": 0.6202, + "step": 7746 + }, + { + "epoch": 0.51998255092111, + "grad_norm": 1.085209608078003, + "learning_rate": 8.675731968536002e-05, + "loss": 0.6968, + "step": 7748 + }, + { + "epoch": 0.520116774604879, + "grad_norm": 1.301666498184204, + "learning_rate": 8.674995110099911e-05, + "loss": 0.6577, + "step": 7750 + }, + { + "epoch": 0.5202509982886481, + "grad_norm": 0.9996120929718018, + "learning_rate": 8.674258078027525e-05, + "loss": 0.6601, + "step": 7752 + }, + { + "epoch": 0.520385221972417, + "grad_norm": 1.0176335573196411, + "learning_rate": 8.673520872353671e-05, + "loss": 0.6561, + "step": 7754 + }, + { + "epoch": 0.520519445656186, + "grad_norm": 1.3511710166931152, + "learning_rate": 8.672783493113179e-05, + "loss": 0.7215, + "step": 7756 + }, + { + "epoch": 0.520653669339955, + "grad_norm": 1.0890061855316162, + "learning_rate": 8.672045940340889e-05, + "loss": 0.6518, + "step": 7758 + }, + { + "epoch": 0.5207878930237241, + "grad_norm": 1.0245070457458496, + "learning_rate": 8.671308214071652e-05, + "loss": 0.6538, + "step": 7760 + }, + { + "epoch": 0.520922116707493, + "grad_norm": 1.0657215118408203, + "learning_rate": 8.670570314340318e-05, + "loss": 0.6401, + "step": 7762 + }, + { + "epoch": 0.521056340391262, + "grad_norm": 1.4132921695709229, + "learning_rate": 8.669832241181756e-05, + "loss": 0.67, + "step": 7764 + }, + { + "epoch": 0.5211905640750311, + "grad_norm": 1.603489875793457, + "learning_rate": 8.669093994630835e-05, + "loss": 0.7102, + "step": 7766 + }, + { + "epoch": 0.5213247877588001, + "grad_norm": 1.1183873414993286, + "learning_rate": 8.66835557472244e-05, + "loss": 0.6822, + "step": 7768 + }, + { + "epoch": 0.521459011442569, + "grad_norm": 1.2292097806930542, + "learning_rate": 8.667616981491458e-05, + "loss": 0.7148, + "step": 7770 + }, + { + "epoch": 0.521593235126338, + "grad_norm": 1.140467643737793, + "learning_rate": 8.666878214972783e-05, + "loss": 0.6574, + "step": 7772 + }, + { + "epoch": 0.5217274588101071, + "grad_norm": 1.070495367050171, + "learning_rate": 8.666139275201325e-05, + "loss": 0.6431, + "step": 7774 + }, + { + "epoch": 0.521861682493876, + "grad_norm": 1.0301347970962524, + "learning_rate": 8.665400162211995e-05, + "loss": 0.6536, + "step": 7776 + }, + { + "epoch": 0.521995906177645, + "grad_norm": 1.162405252456665, + "learning_rate": 8.664660876039715e-05, + "loss": 0.6408, + "step": 7778 + }, + { + "epoch": 0.522130129861414, + "grad_norm": 1.1394559144973755, + "learning_rate": 8.663921416719415e-05, + "loss": 0.7394, + "step": 7780 + }, + { + "epoch": 0.5222643535451831, + "grad_norm": 1.1406190395355225, + "learning_rate": 8.663181784286032e-05, + "loss": 0.6864, + "step": 7782 + }, + { + "epoch": 0.522398577228952, + "grad_norm": 0.9947435259819031, + "learning_rate": 8.662441978774514e-05, + "loss": 0.6074, + "step": 7784 + }, + { + "epoch": 0.522532800912721, + "grad_norm": 1.0343014001846313, + "learning_rate": 8.661702000219814e-05, + "loss": 0.6589, + "step": 7786 + }, + { + "epoch": 0.5226670245964901, + "grad_norm": 0.998464822769165, + "learning_rate": 8.660961848656895e-05, + "loss": 0.6141, + "step": 7788 + }, + { + "epoch": 0.5228012482802591, + "grad_norm": 1.8650562763214111, + "learning_rate": 8.660221524120727e-05, + "loss": 0.7038, + "step": 7790 + }, + { + "epoch": 0.522935471964028, + "grad_norm": 1.15292227268219, + "learning_rate": 8.65948102664629e-05, + "loss": 0.5998, + "step": 7792 + }, + { + "epoch": 0.523069695647797, + "grad_norm": 1.2766640186309814, + "learning_rate": 8.658740356268571e-05, + "loss": 0.6357, + "step": 7794 + }, + { + "epoch": 0.5232039193315661, + "grad_norm": 1.0240525007247925, + "learning_rate": 8.657999513022563e-05, + "loss": 0.6269, + "step": 7796 + }, + { + "epoch": 0.5233381430153351, + "grad_norm": 1.5196304321289062, + "learning_rate": 8.657258496943274e-05, + "loss": 0.6223, + "step": 7798 + }, + { + "epoch": 0.523472366699104, + "grad_norm": 1.0911345481872559, + "learning_rate": 8.65651730806571e-05, + "loss": 0.6661, + "step": 7800 + }, + { + "epoch": 0.523606590382873, + "grad_norm": 1.163367748260498, + "learning_rate": 8.655775946424895e-05, + "loss": 0.5699, + "step": 7802 + }, + { + "epoch": 0.5237408140666421, + "grad_norm": 1.1395347118377686, + "learning_rate": 8.655034412055856e-05, + "loss": 0.6649, + "step": 7804 + }, + { + "epoch": 0.5238750377504111, + "grad_norm": 1.1319963932037354, + "learning_rate": 8.654292704993627e-05, + "loss": 0.6253, + "step": 7806 + }, + { + "epoch": 0.52400926143418, + "grad_norm": 0.9931795597076416, + "learning_rate": 8.653550825273253e-05, + "loss": 0.6329, + "step": 7808 + }, + { + "epoch": 0.5241434851179491, + "grad_norm": 1.2011792659759521, + "learning_rate": 8.652808772929788e-05, + "loss": 0.7392, + "step": 7810 + }, + { + "epoch": 0.5242777088017181, + "grad_norm": 0.9433580040931702, + "learning_rate": 8.652066547998292e-05, + "loss": 0.6397, + "step": 7812 + }, + { + "epoch": 0.524411932485487, + "grad_norm": 1.039955973625183, + "learning_rate": 8.651324150513833e-05, + "loss": 0.6344, + "step": 7814 + }, + { + "epoch": 0.524546156169256, + "grad_norm": 1.0503945350646973, + "learning_rate": 8.650581580511487e-05, + "loss": 0.6448, + "step": 7816 + }, + { + "epoch": 0.5246803798530251, + "grad_norm": 1.0410972833633423, + "learning_rate": 8.64983883802634e-05, + "loss": 0.6067, + "step": 7818 + }, + { + "epoch": 0.5248146035367941, + "grad_norm": 1.023305892944336, + "learning_rate": 8.649095923093484e-05, + "loss": 0.5264, + "step": 7820 + }, + { + "epoch": 0.524948827220563, + "grad_norm": 1.0543869733810425, + "learning_rate": 8.648352835748024e-05, + "loss": 0.635, + "step": 7822 + }, + { + "epoch": 0.525083050904332, + "grad_norm": 1.1276301145553589, + "learning_rate": 8.647609576025064e-05, + "loss": 0.5973, + "step": 7824 + }, + { + "epoch": 0.5252172745881011, + "grad_norm": 1.0825186967849731, + "learning_rate": 8.646866143959725e-05, + "loss": 0.6765, + "step": 7826 + }, + { + "epoch": 0.5253514982718701, + "grad_norm": 0.993621826171875, + "learning_rate": 8.646122539587133e-05, + "loss": 0.5949, + "step": 7828 + }, + { + "epoch": 0.525485721955639, + "grad_norm": 1.1419330835342407, + "learning_rate": 8.64537876294242e-05, + "loss": 0.6804, + "step": 7830 + }, + { + "epoch": 0.5256199456394081, + "grad_norm": 1.189340591430664, + "learning_rate": 8.644634814060728e-05, + "loss": 0.5882, + "step": 7832 + }, + { + "epoch": 0.5257541693231771, + "grad_norm": 1.0714603662490845, + "learning_rate": 8.643890692977209e-05, + "loss": 0.6831, + "step": 7834 + }, + { + "epoch": 0.5258883930069461, + "grad_norm": 1.0567216873168945, + "learning_rate": 8.64314639972702e-05, + "loss": 0.644, + "step": 7836 + }, + { + "epoch": 0.526022616690715, + "grad_norm": 0.9821720719337463, + "learning_rate": 8.642401934345328e-05, + "loss": 0.6922, + "step": 7838 + }, + { + "epoch": 0.5261568403744841, + "grad_norm": 0.9419900178909302, + "learning_rate": 8.641657296867306e-05, + "loss": 0.6089, + "step": 7840 + }, + { + "epoch": 0.5262910640582531, + "grad_norm": 1.1534830331802368, + "learning_rate": 8.640912487328139e-05, + "loss": 0.6305, + "step": 7842 + }, + { + "epoch": 0.5264252877420221, + "grad_norm": 1.1612067222595215, + "learning_rate": 8.640167505763014e-05, + "loss": 0.6018, + "step": 7844 + }, + { + "epoch": 0.526559511425791, + "grad_norm": 0.9795418977737427, + "learning_rate": 8.639422352207136e-05, + "loss": 0.7098, + "step": 7846 + }, + { + "epoch": 0.5266937351095601, + "grad_norm": 1.0749777555465698, + "learning_rate": 8.638677026695707e-05, + "loss": 0.6147, + "step": 7848 + }, + { + "epoch": 0.5268279587933291, + "grad_norm": 1.1168949604034424, + "learning_rate": 8.637931529263943e-05, + "loss": 0.7036, + "step": 7850 + }, + { + "epoch": 0.526962182477098, + "grad_norm": 0.9598569869995117, + "learning_rate": 8.637185859947067e-05, + "loss": 0.5887, + "step": 7852 + }, + { + "epoch": 0.5270964061608671, + "grad_norm": 0.9963730573654175, + "learning_rate": 8.636440018780314e-05, + "loss": 0.6447, + "step": 7854 + }, + { + "epoch": 0.5272306298446361, + "grad_norm": 0.9982380270957947, + "learning_rate": 8.635694005798919e-05, + "loss": 0.6494, + "step": 7856 + }, + { + "epoch": 0.5273648535284051, + "grad_norm": 1.157914400100708, + "learning_rate": 8.634947821038132e-05, + "loss": 0.757, + "step": 7858 + }, + { + "epoch": 0.527499077212174, + "grad_norm": 1.0347570180892944, + "learning_rate": 8.634201464533207e-05, + "loss": 0.6752, + "step": 7860 + }, + { + "epoch": 0.5276333008959431, + "grad_norm": 1.0469486713409424, + "learning_rate": 8.633454936319411e-05, + "loss": 0.6475, + "step": 7862 + }, + { + "epoch": 0.5277675245797121, + "grad_norm": 1.2268787622451782, + "learning_rate": 8.632708236432015e-05, + "loss": 0.6541, + "step": 7864 + }, + { + "epoch": 0.5279017482634811, + "grad_norm": 1.1413471698760986, + "learning_rate": 8.631961364906296e-05, + "loss": 0.6259, + "step": 7866 + }, + { + "epoch": 0.52803597194725, + "grad_norm": 1.1279798746109009, + "learning_rate": 8.631214321777546e-05, + "loss": 0.7006, + "step": 7868 + }, + { + "epoch": 0.5281701956310191, + "grad_norm": 1.0315818786621094, + "learning_rate": 8.630467107081057e-05, + "loss": 0.6442, + "step": 7870 + }, + { + "epoch": 0.5283044193147881, + "grad_norm": 1.1192677021026611, + "learning_rate": 8.629719720852138e-05, + "loss": 0.6034, + "step": 7872 + }, + { + "epoch": 0.5284386429985571, + "grad_norm": 1.1972932815551758, + "learning_rate": 8.628972163126101e-05, + "loss": 0.6259, + "step": 7874 + }, + { + "epoch": 0.5285728666823261, + "grad_norm": 1.0784581899642944, + "learning_rate": 8.628224433938263e-05, + "loss": 0.622, + "step": 7876 + }, + { + "epoch": 0.5287070903660951, + "grad_norm": 1.0202914476394653, + "learning_rate": 8.627476533323957e-05, + "loss": 0.6532, + "step": 7878 + }, + { + "epoch": 0.5288413140498641, + "grad_norm": 1.1538416147232056, + "learning_rate": 8.626728461318518e-05, + "loss": 0.6699, + "step": 7880 + }, + { + "epoch": 0.5289755377336331, + "grad_norm": 1.0363773107528687, + "learning_rate": 8.62598021795729e-05, + "loss": 0.6393, + "step": 7882 + }, + { + "epoch": 0.5291097614174021, + "grad_norm": 1.0268653631210327, + "learning_rate": 8.625231803275627e-05, + "loss": 0.6782, + "step": 7884 + }, + { + "epoch": 0.5292439851011711, + "grad_norm": 0.8181347846984863, + "learning_rate": 8.62448321730889e-05, + "loss": 0.6044, + "step": 7886 + }, + { + "epoch": 0.5293782087849401, + "grad_norm": 1.0739057064056396, + "learning_rate": 8.623734460092449e-05, + "loss": 0.6223, + "step": 7888 + }, + { + "epoch": 0.529512432468709, + "grad_norm": 1.0722723007202148, + "learning_rate": 8.62298553166168e-05, + "loss": 0.6551, + "step": 7890 + }, + { + "epoch": 0.5296466561524781, + "grad_norm": 1.0251004695892334, + "learning_rate": 8.622236432051969e-05, + "loss": 0.6408, + "step": 7892 + }, + { + "epoch": 0.5297808798362471, + "grad_norm": 1.0956529378890991, + "learning_rate": 8.621487161298709e-05, + "loss": 0.6263, + "step": 7894 + }, + { + "epoch": 0.5299151035200161, + "grad_norm": 1.1363166570663452, + "learning_rate": 8.620737719437302e-05, + "loss": 0.6748, + "step": 7896 + }, + { + "epoch": 0.5300493272037851, + "grad_norm": 1.026837706565857, + "learning_rate": 8.619988106503157e-05, + "loss": 0.529, + "step": 7898 + }, + { + "epoch": 0.5301835508875541, + "grad_norm": 0.9920128583908081, + "learning_rate": 8.619238322531694e-05, + "loss": 0.64, + "step": 7900 + }, + { + "epoch": 0.5303177745713231, + "grad_norm": 1.0423779487609863, + "learning_rate": 8.618488367558335e-05, + "loss": 0.6391, + "step": 7902 + }, + { + "epoch": 0.5304519982550921, + "grad_norm": 1.183112382888794, + "learning_rate": 8.617738241618518e-05, + "loss": 0.6261, + "step": 7904 + }, + { + "epoch": 0.5305862219388611, + "grad_norm": 1.0162816047668457, + "learning_rate": 8.616987944747683e-05, + "loss": 0.6331, + "step": 7906 + }, + { + "epoch": 0.5307204456226301, + "grad_norm": 1.2763468027114868, + "learning_rate": 8.616237476981279e-05, + "loss": 0.6126, + "step": 7908 + }, + { + "epoch": 0.5308546693063991, + "grad_norm": 1.084007740020752, + "learning_rate": 8.615486838354765e-05, + "loss": 0.5957, + "step": 7910 + }, + { + "epoch": 0.5309888929901682, + "grad_norm": 1.0765655040740967, + "learning_rate": 8.61473602890361e-05, + "loss": 0.6723, + "step": 7912 + }, + { + "epoch": 0.5311231166739371, + "grad_norm": 1.0184767246246338, + "learning_rate": 8.613985048663282e-05, + "loss": 0.6472, + "step": 7914 + }, + { + "epoch": 0.5312573403577061, + "grad_norm": 3.213615655899048, + "learning_rate": 8.613233897669271e-05, + "loss": 0.5945, + "step": 7916 + }, + { + "epoch": 0.5313915640414751, + "grad_norm": 1.017124891281128, + "learning_rate": 8.612482575957062e-05, + "loss": 0.6631, + "step": 7918 + }, + { + "epoch": 0.5315257877252442, + "grad_norm": 1.061355710029602, + "learning_rate": 8.611731083562154e-05, + "loss": 0.5845, + "step": 7920 + }, + { + "epoch": 0.5316600114090131, + "grad_norm": 1.0305843353271484, + "learning_rate": 8.610979420520053e-05, + "loss": 0.6712, + "step": 7922 + }, + { + "epoch": 0.5317942350927821, + "grad_norm": 0.9948982000350952, + "learning_rate": 8.610227586866278e-05, + "loss": 0.5888, + "step": 7924 + }, + { + "epoch": 0.5319284587765511, + "grad_norm": 0.9479734301567078, + "learning_rate": 8.609475582636348e-05, + "loss": 0.5764, + "step": 7926 + }, + { + "epoch": 0.5320626824603201, + "grad_norm": 1.1093517541885376, + "learning_rate": 8.608723407865792e-05, + "loss": 0.6771, + "step": 7928 + }, + { + "epoch": 0.5321969061440891, + "grad_norm": 1.0552197694778442, + "learning_rate": 8.607971062590155e-05, + "loss": 0.6272, + "step": 7930 + }, + { + "epoch": 0.5323311298278581, + "grad_norm": 1.1534368991851807, + "learning_rate": 8.607218546844979e-05, + "loss": 0.6534, + "step": 7932 + }, + { + "epoch": 0.5324653535116272, + "grad_norm": 1.058334231376648, + "learning_rate": 8.606465860665819e-05, + "loss": 0.6327, + "step": 7934 + }, + { + "epoch": 0.5325995771953961, + "grad_norm": 1.1343920230865479, + "learning_rate": 8.605713004088238e-05, + "loss": 0.6907, + "step": 7936 + }, + { + "epoch": 0.5327338008791651, + "grad_norm": 1.3429607152938843, + "learning_rate": 8.60495997714781e-05, + "loss": 0.6221, + "step": 7938 + }, + { + "epoch": 0.5328680245629341, + "grad_norm": 1.3038303852081299, + "learning_rate": 8.60420677988011e-05, + "loss": 0.6281, + "step": 7940 + }, + { + "epoch": 0.5330022482467032, + "grad_norm": 1.0231696367263794, + "learning_rate": 8.60345341232073e-05, + "loss": 0.6278, + "step": 7942 + }, + { + "epoch": 0.5331364719304721, + "grad_norm": 1.0313324928283691, + "learning_rate": 8.60269987450526e-05, + "loss": 0.6253, + "step": 7944 + }, + { + "epoch": 0.5332706956142411, + "grad_norm": 1.0684877634048462, + "learning_rate": 8.601946166469305e-05, + "loss": 0.6228, + "step": 7946 + }, + { + "epoch": 0.5334049192980101, + "grad_norm": 1.0405035018920898, + "learning_rate": 8.601192288248478e-05, + "loss": 0.6356, + "step": 7948 + }, + { + "epoch": 0.5335391429817792, + "grad_norm": 1.089769721031189, + "learning_rate": 8.600438239878394e-05, + "loss": 0.6284, + "step": 7950 + }, + { + "epoch": 0.5336733666655481, + "grad_norm": 1.114942193031311, + "learning_rate": 8.599684021394686e-05, + "loss": 0.748, + "step": 7952 + }, + { + "epoch": 0.5338075903493171, + "grad_norm": 0.9413202404975891, + "learning_rate": 8.598929632832985e-05, + "loss": 0.6514, + "step": 7954 + }, + { + "epoch": 0.5339418140330862, + "grad_norm": 0.9680725336074829, + "learning_rate": 8.598175074228938e-05, + "loss": 0.5765, + "step": 7956 + }, + { + "epoch": 0.5340760377168552, + "grad_norm": 1.4321138858795166, + "learning_rate": 8.597420345618193e-05, + "loss": 0.6071, + "step": 7958 + }, + { + "epoch": 0.5342102614006241, + "grad_norm": 1.0866897106170654, + "learning_rate": 8.596665447036409e-05, + "loss": 0.6463, + "step": 7960 + }, + { + "epoch": 0.5343444850843931, + "grad_norm": 1.0365897417068481, + "learning_rate": 8.595910378519257e-05, + "loss": 0.6773, + "step": 7962 + }, + { + "epoch": 0.5344787087681622, + "grad_norm": 0.9632202982902527, + "learning_rate": 8.59515514010241e-05, + "loss": 0.6185, + "step": 7964 + }, + { + "epoch": 0.5346129324519311, + "grad_norm": 1.2617647647857666, + "learning_rate": 8.594399731821552e-05, + "loss": 0.6479, + "step": 7966 + }, + { + "epoch": 0.5347471561357001, + "grad_norm": 0.9706656336784363, + "learning_rate": 8.593644153712374e-05, + "loss": 0.6248, + "step": 7968 + }, + { + "epoch": 0.5348813798194691, + "grad_norm": 1.5335643291473389, + "learning_rate": 8.592888405810578e-05, + "loss": 0.6378, + "step": 7970 + }, + { + "epoch": 0.5350156035032382, + "grad_norm": 1.1157900094985962, + "learning_rate": 8.59213248815187e-05, + "loss": 0.6764, + "step": 7972 + }, + { + "epoch": 0.5351498271870071, + "grad_norm": 1.339160680770874, + "learning_rate": 8.591376400771964e-05, + "loss": 0.5938, + "step": 7974 + }, + { + "epoch": 0.5352840508707761, + "grad_norm": 1.19435715675354, + "learning_rate": 8.590620143706587e-05, + "loss": 0.7154, + "step": 7976 + }, + { + "epoch": 0.5354182745545452, + "grad_norm": 1.1539045572280884, + "learning_rate": 8.589863716991469e-05, + "loss": 0.7104, + "step": 7978 + }, + { + "epoch": 0.5355524982383142, + "grad_norm": 1.0074177980422974, + "learning_rate": 8.589107120662348e-05, + "loss": 0.6027, + "step": 7980 + }, + { + "epoch": 0.5356867219220831, + "grad_norm": 0.969353199005127, + "learning_rate": 8.588350354754973e-05, + "loss": 0.6428, + "step": 7982 + }, + { + "epoch": 0.5358209456058521, + "grad_norm": 1.1570994853973389, + "learning_rate": 8.587593419305101e-05, + "loss": 0.6547, + "step": 7984 + }, + { + "epoch": 0.5359551692896212, + "grad_norm": 0.9162207841873169, + "learning_rate": 8.586836314348494e-05, + "loss": 0.675, + "step": 7986 + }, + { + "epoch": 0.5360893929733902, + "grad_norm": 0.9713578820228577, + "learning_rate": 8.586079039920924e-05, + "loss": 0.6232, + "step": 7988 + }, + { + "epoch": 0.5362236166571591, + "grad_norm": 1.3750872611999512, + "learning_rate": 8.585321596058174e-05, + "loss": 0.7061, + "step": 7990 + }, + { + "epoch": 0.5363578403409281, + "grad_norm": 0.9677409529685974, + "learning_rate": 8.584563982796026e-05, + "loss": 0.6361, + "step": 7992 + }, + { + "epoch": 0.5364920640246972, + "grad_norm": 1.0356523990631104, + "learning_rate": 8.583806200170279e-05, + "loss": 0.6689, + "step": 7994 + }, + { + "epoch": 0.5366262877084662, + "grad_norm": 1.0907829999923706, + "learning_rate": 8.583048248216736e-05, + "loss": 0.6733, + "step": 7996 + }, + { + "epoch": 0.5367605113922351, + "grad_norm": 1.0503954887390137, + "learning_rate": 8.582290126971209e-05, + "loss": 0.5998, + "step": 7998 + }, + { + "epoch": 0.5368947350760042, + "grad_norm": 1.0227012634277344, + "learning_rate": 8.581531836469518e-05, + "loss": 0.6032, + "step": 8000 + }, + { + "epoch": 0.5370289587597732, + "grad_norm": 1.0412101745605469, + "learning_rate": 8.580773376747492e-05, + "loss": 0.6144, + "step": 8002 + }, + { + "epoch": 0.5371631824435421, + "grad_norm": 1.048636794090271, + "learning_rate": 8.580014747840964e-05, + "loss": 0.6312, + "step": 8004 + }, + { + "epoch": 0.5372974061273111, + "grad_norm": 1.1250464916229248, + "learning_rate": 8.579255949785779e-05, + "loss": 0.6273, + "step": 8006 + }, + { + "epoch": 0.5374316298110802, + "grad_norm": 1.1802406311035156, + "learning_rate": 8.578496982617788e-05, + "loss": 0.6626, + "step": 8008 + }, + { + "epoch": 0.5375658534948492, + "grad_norm": 0.9356011152267456, + "learning_rate": 8.577737846372853e-05, + "loss": 0.6837, + "step": 8010 + }, + { + "epoch": 0.5377000771786181, + "grad_norm": 1.0496151447296143, + "learning_rate": 8.576978541086838e-05, + "loss": 0.6203, + "step": 8012 + }, + { + "epoch": 0.5378343008623871, + "grad_norm": 1.1250532865524292, + "learning_rate": 8.576219066795622e-05, + "loss": 0.5904, + "step": 8014 + }, + { + "epoch": 0.5379685245461562, + "grad_norm": 0.9540520906448364, + "learning_rate": 8.575459423535089e-05, + "loss": 0.6346, + "step": 8016 + }, + { + "epoch": 0.5381027482299252, + "grad_norm": 1.5916898250579834, + "learning_rate": 8.574699611341125e-05, + "loss": 0.6193, + "step": 8018 + }, + { + "epoch": 0.5382369719136941, + "grad_norm": 0.9837737083435059, + "learning_rate": 8.573939630249638e-05, + "loss": 0.686, + "step": 8020 + }, + { + "epoch": 0.5383711955974632, + "grad_norm": 1.1416816711425781, + "learning_rate": 8.573179480296529e-05, + "loss": 0.7032, + "step": 8022 + }, + { + "epoch": 0.5385054192812322, + "grad_norm": 1.125101089477539, + "learning_rate": 8.572419161517716e-05, + "loss": 0.6531, + "step": 8024 + }, + { + "epoch": 0.5386396429650012, + "grad_norm": 1.0765750408172607, + "learning_rate": 8.571658673949124e-05, + "loss": 0.6607, + "step": 8026 + }, + { + "epoch": 0.5387738666487701, + "grad_norm": 0.9669967889785767, + "learning_rate": 8.570898017626681e-05, + "loss": 0.6227, + "step": 8028 + }, + { + "epoch": 0.5389080903325392, + "grad_norm": 1.1211695671081543, + "learning_rate": 8.570137192586329e-05, + "loss": 0.6798, + "step": 8030 + }, + { + "epoch": 0.5390423140163082, + "grad_norm": 0.96068274974823, + "learning_rate": 8.569376198864015e-05, + "loss": 0.6164, + "step": 8032 + }, + { + "epoch": 0.5391765377000772, + "grad_norm": 1.313602089881897, + "learning_rate": 8.568615036495696e-05, + "loss": 0.646, + "step": 8034 + }, + { + "epoch": 0.5393107613838461, + "grad_norm": 0.9256817698478699, + "learning_rate": 8.567853705517332e-05, + "loss": 0.5842, + "step": 8036 + }, + { + "epoch": 0.5394449850676152, + "grad_norm": 1.083078145980835, + "learning_rate": 8.567092205964897e-05, + "loss": 0.6093, + "step": 8038 + }, + { + "epoch": 0.5395792087513842, + "grad_norm": 1.7828437089920044, + "learning_rate": 8.566330537874369e-05, + "loss": 0.6242, + "step": 8040 + }, + { + "epoch": 0.5397134324351531, + "grad_norm": 1.140256404876709, + "learning_rate": 8.565568701281738e-05, + "loss": 0.5567, + "step": 8042 + }, + { + "epoch": 0.5398476561189222, + "grad_norm": 0.9843252301216125, + "learning_rate": 8.564806696222995e-05, + "loss": 0.5182, + "step": 8044 + }, + { + "epoch": 0.5399818798026912, + "grad_norm": 1.1507277488708496, + "learning_rate": 8.564044522734147e-05, + "loss": 0.63, + "step": 8046 + }, + { + "epoch": 0.5401161034864602, + "grad_norm": 1.0563549995422363, + "learning_rate": 8.563282180851203e-05, + "loss": 0.6801, + "step": 8048 + }, + { + "epoch": 0.5402503271702291, + "grad_norm": 1.0808017253875732, + "learning_rate": 8.562519670610183e-05, + "loss": 0.6887, + "step": 8050 + }, + { + "epoch": 0.5403845508539982, + "grad_norm": 1.0849084854125977, + "learning_rate": 8.561756992047114e-05, + "loss": 0.6625, + "step": 8052 + }, + { + "epoch": 0.5405187745377672, + "grad_norm": 0.9871025085449219, + "learning_rate": 8.560994145198031e-05, + "loss": 0.6241, + "step": 8054 + }, + { + "epoch": 0.5406529982215362, + "grad_norm": 1.465746521949768, + "learning_rate": 8.560231130098977e-05, + "loss": 0.6822, + "step": 8056 + }, + { + "epoch": 0.5407872219053051, + "grad_norm": 1.1308164596557617, + "learning_rate": 8.559467946786002e-05, + "loss": 0.6862, + "step": 8058 + }, + { + "epoch": 0.5409214455890742, + "grad_norm": 1.1542991399765015, + "learning_rate": 8.558704595295168e-05, + "loss": 0.6147, + "step": 8060 + }, + { + "epoch": 0.5410556692728432, + "grad_norm": 1.1769368648529053, + "learning_rate": 8.557941075662538e-05, + "loss": 0.6499, + "step": 8062 + }, + { + "epoch": 0.5411898929566122, + "grad_norm": 0.9702188372612, + "learning_rate": 8.55717738792419e-05, + "loss": 0.643, + "step": 8064 + }, + { + "epoch": 0.5413241166403812, + "grad_norm": 1.063549518585205, + "learning_rate": 8.556413532116204e-05, + "loss": 0.5977, + "step": 8066 + }, + { + "epoch": 0.5414583403241502, + "grad_norm": 1.269137978553772, + "learning_rate": 8.555649508274672e-05, + "loss": 0.6633, + "step": 8068 + }, + { + "epoch": 0.5415925640079192, + "grad_norm": 1.0866457223892212, + "learning_rate": 8.554885316435694e-05, + "loss": 0.6889, + "step": 8070 + }, + { + "epoch": 0.5417267876916882, + "grad_norm": 0.9888244867324829, + "learning_rate": 8.554120956635375e-05, + "loss": 0.6395, + "step": 8072 + }, + { + "epoch": 0.5418610113754572, + "grad_norm": 1.6463063955307007, + "learning_rate": 8.553356428909827e-05, + "loss": 0.647, + "step": 8074 + }, + { + "epoch": 0.5419952350592262, + "grad_norm": 1.0329785346984863, + "learning_rate": 8.552591733295179e-05, + "loss": 0.6518, + "step": 8076 + }, + { + "epoch": 0.5421294587429952, + "grad_norm": 1.084775447845459, + "learning_rate": 8.551826869827555e-05, + "loss": 0.714, + "step": 8078 + }, + { + "epoch": 0.5422636824267641, + "grad_norm": 1.1005643606185913, + "learning_rate": 8.551061838543097e-05, + "loss": 0.6846, + "step": 8080 + }, + { + "epoch": 0.5423979061105332, + "grad_norm": 1.1735957860946655, + "learning_rate": 8.550296639477948e-05, + "loss": 0.6822, + "step": 8082 + }, + { + "epoch": 0.5425321297943022, + "grad_norm": 1.0340960025787354, + "learning_rate": 8.549531272668266e-05, + "loss": 0.6334, + "step": 8084 + }, + { + "epoch": 0.5426663534780712, + "grad_norm": 1.2648706436157227, + "learning_rate": 8.54876573815021e-05, + "loss": 0.6676, + "step": 8086 + }, + { + "epoch": 0.5428005771618402, + "grad_norm": 1.3927974700927734, + "learning_rate": 8.548000035959953e-05, + "loss": 0.6405, + "step": 8088 + }, + { + "epoch": 0.5429348008456092, + "grad_norm": 1.0736503601074219, + "learning_rate": 8.547234166133671e-05, + "loss": 0.6455, + "step": 8090 + }, + { + "epoch": 0.5430690245293782, + "grad_norm": 1.0800048112869263, + "learning_rate": 8.546468128707548e-05, + "loss": 0.5956, + "step": 8092 + }, + { + "epoch": 0.5432032482131472, + "grad_norm": 1.027530312538147, + "learning_rate": 8.545701923717781e-05, + "loss": 0.7324, + "step": 8094 + }, + { + "epoch": 0.5433374718969162, + "grad_norm": 0.9228115081787109, + "learning_rate": 8.54493555120057e-05, + "loss": 0.5722, + "step": 8096 + }, + { + "epoch": 0.5434716955806852, + "grad_norm": 1.122521162033081, + "learning_rate": 8.544169011192125e-05, + "loss": 0.7119, + "step": 8098 + }, + { + "epoch": 0.5436059192644542, + "grad_norm": 1.0947986841201782, + "learning_rate": 8.543402303728665e-05, + "loss": 0.6815, + "step": 8100 + }, + { + "epoch": 0.5437401429482233, + "grad_norm": 1.00077486038208, + "learning_rate": 8.542635428846413e-05, + "loss": 0.6045, + "step": 8102 + }, + { + "epoch": 0.5438743666319922, + "grad_norm": 1.0414117574691772, + "learning_rate": 8.541868386581601e-05, + "loss": 0.6676, + "step": 8104 + }, + { + "epoch": 0.5440085903157612, + "grad_norm": 1.5793548822402954, + "learning_rate": 8.541101176970476e-05, + "loss": 0.7247, + "step": 8106 + }, + { + "epoch": 0.5441428139995302, + "grad_norm": 0.999428927898407, + "learning_rate": 8.540333800049282e-05, + "loss": 0.6897, + "step": 8108 + }, + { + "epoch": 0.5442770376832993, + "grad_norm": 1.000823974609375, + "learning_rate": 8.539566255854277e-05, + "loss": 0.6522, + "step": 8110 + }, + { + "epoch": 0.5444112613670682, + "grad_norm": 0.9637446999549866, + "learning_rate": 8.538798544421729e-05, + "loss": 0.6125, + "step": 8112 + }, + { + "epoch": 0.5445454850508372, + "grad_norm": 1.002723217010498, + "learning_rate": 8.538030665787906e-05, + "loss": 0.5744, + "step": 8114 + }, + { + "epoch": 0.5446797087346062, + "grad_norm": 1.1904152631759644, + "learning_rate": 8.537262619989093e-05, + "loss": 0.7247, + "step": 8116 + }, + { + "epoch": 0.5448139324183752, + "grad_norm": 1.1540770530700684, + "learning_rate": 8.536494407061575e-05, + "loss": 0.6186, + "step": 8118 + }, + { + "epoch": 0.5449481561021442, + "grad_norm": 0.9250359535217285, + "learning_rate": 8.535726027041652e-05, + "loss": 0.7193, + "step": 8120 + }, + { + "epoch": 0.5450823797859132, + "grad_norm": 0.9992945790290833, + "learning_rate": 8.534957479965629e-05, + "loss": 0.6211, + "step": 8122 + }, + { + "epoch": 0.5452166034696823, + "grad_norm": 0.9270564317703247, + "learning_rate": 8.534188765869813e-05, + "loss": 0.6254, + "step": 8124 + }, + { + "epoch": 0.5453508271534512, + "grad_norm": 1.0162103176116943, + "learning_rate": 8.533419884790528e-05, + "loss": 0.6142, + "step": 8126 + }, + { + "epoch": 0.5454850508372202, + "grad_norm": 1.0450429916381836, + "learning_rate": 8.532650836764102e-05, + "loss": 0.6941, + "step": 8128 + }, + { + "epoch": 0.5456192745209892, + "grad_norm": 1.0318299531936646, + "learning_rate": 8.531881621826871e-05, + "loss": 0.6611, + "step": 8130 + }, + { + "epoch": 0.5457534982047583, + "grad_norm": 1.0086703300476074, + "learning_rate": 8.531112240015178e-05, + "loss": 0.677, + "step": 8132 + }, + { + "epoch": 0.5458877218885272, + "grad_norm": 0.9467586278915405, + "learning_rate": 8.530342691365377e-05, + "loss": 0.6441, + "step": 8134 + }, + { + "epoch": 0.5460219455722962, + "grad_norm": 0.8377612233161926, + "learning_rate": 8.529572975913824e-05, + "loss": 0.565, + "step": 8136 + }, + { + "epoch": 0.5461561692560652, + "grad_norm": 0.8942846655845642, + "learning_rate": 8.528803093696889e-05, + "loss": 0.5783, + "step": 8138 + }, + { + "epoch": 0.5462903929398343, + "grad_norm": 1.1009197235107422, + "learning_rate": 8.528033044750947e-05, + "loss": 0.6599, + "step": 8140 + }, + { + "epoch": 0.5464246166236032, + "grad_norm": 1.0929917097091675, + "learning_rate": 8.527262829112382e-05, + "loss": 0.7262, + "step": 8142 + }, + { + "epoch": 0.5465588403073722, + "grad_norm": 1.159877896308899, + "learning_rate": 8.526492446817583e-05, + "loss": 0.6566, + "step": 8144 + }, + { + "epoch": 0.5466930639911413, + "grad_norm": 1.178993821144104, + "learning_rate": 8.52572189790295e-05, + "loss": 0.5907, + "step": 8146 + }, + { + "epoch": 0.5468272876749103, + "grad_norm": 1.305436372756958, + "learning_rate": 8.524951182404892e-05, + "loss": 0.593, + "step": 8148 + }, + { + "epoch": 0.5469615113586792, + "grad_norm": 1.1507524251937866, + "learning_rate": 8.52418030035982e-05, + "loss": 0.6921, + "step": 8150 + }, + { + "epoch": 0.5470957350424482, + "grad_norm": 1.0701565742492676, + "learning_rate": 8.523409251804163e-05, + "loss": 0.6252, + "step": 8152 + }, + { + "epoch": 0.5472299587262173, + "grad_norm": 1.0704950094223022, + "learning_rate": 8.522638036774345e-05, + "loss": 0.6535, + "step": 8154 + }, + { + "epoch": 0.5473641824099862, + "grad_norm": 0.9062939882278442, + "learning_rate": 8.521866655306805e-05, + "loss": 0.6828, + "step": 8156 + }, + { + "epoch": 0.5474984060937552, + "grad_norm": 1.1163350343704224, + "learning_rate": 8.521095107437993e-05, + "loss": 0.6318, + "step": 8158 + }, + { + "epoch": 0.5476326297775242, + "grad_norm": 0.9973238706588745, + "learning_rate": 8.52032339320436e-05, + "loss": 0.616, + "step": 8160 + }, + { + "epoch": 0.5477668534612933, + "grad_norm": 1.248987078666687, + "learning_rate": 8.51955151264237e-05, + "loss": 0.5971, + "step": 8162 + }, + { + "epoch": 0.5479010771450622, + "grad_norm": 1.2610836029052734, + "learning_rate": 8.518779465788491e-05, + "loss": 0.7045, + "step": 8164 + }, + { + "epoch": 0.5480353008288312, + "grad_norm": 0.9630738496780396, + "learning_rate": 8.518007252679202e-05, + "loss": 0.6211, + "step": 8166 + }, + { + "epoch": 0.5481695245126003, + "grad_norm": 1.0208569765090942, + "learning_rate": 8.517234873350987e-05, + "loss": 0.6194, + "step": 8168 + }, + { + "epoch": 0.5483037481963693, + "grad_norm": 1.011013388633728, + "learning_rate": 8.516462327840343e-05, + "loss": 0.5775, + "step": 8170 + }, + { + "epoch": 0.5484379718801382, + "grad_norm": 1.3043915033340454, + "learning_rate": 8.515689616183769e-05, + "loss": 0.642, + "step": 8172 + }, + { + "epoch": 0.5485721955639072, + "grad_norm": 1.1023441553115845, + "learning_rate": 8.51491673841777e-05, + "loss": 0.6545, + "step": 8174 + }, + { + "epoch": 0.5487064192476763, + "grad_norm": 1.0701701641082764, + "learning_rate": 8.51414369457887e-05, + "loss": 0.6236, + "step": 8176 + }, + { + "epoch": 0.5488406429314453, + "grad_norm": 1.1929540634155273, + "learning_rate": 8.513370484703591e-05, + "loss": 0.6672, + "step": 8178 + }, + { + "epoch": 0.5489748666152142, + "grad_norm": 1.1066863536834717, + "learning_rate": 8.512597108828464e-05, + "loss": 0.6638, + "step": 8180 + }, + { + "epoch": 0.5491090902989832, + "grad_norm": 1.0696840286254883, + "learning_rate": 8.511823566990031e-05, + "loss": 0.697, + "step": 8182 + }, + { + "epoch": 0.5492433139827523, + "grad_norm": 1.2313902378082275, + "learning_rate": 8.511049859224842e-05, + "loss": 0.7155, + "step": 8184 + }, + { + "epoch": 0.5493775376665213, + "grad_norm": 1.0531792640686035, + "learning_rate": 8.510275985569449e-05, + "loss": 0.6543, + "step": 8186 + }, + { + "epoch": 0.5495117613502902, + "grad_norm": 1.1346839666366577, + "learning_rate": 8.509501946060421e-05, + "loss": 0.6262, + "step": 8188 + }, + { + "epoch": 0.5496459850340593, + "grad_norm": 1.1444693803787231, + "learning_rate": 8.508727740734324e-05, + "loss": 0.64, + "step": 8190 + }, + { + "epoch": 0.5497802087178283, + "grad_norm": 1.1017465591430664, + "learning_rate": 8.507953369627743e-05, + "loss": 0.6515, + "step": 8192 + }, + { + "epoch": 0.5499144324015972, + "grad_norm": 1.0046477317810059, + "learning_rate": 8.507178832777263e-05, + "loss": 0.6283, + "step": 8194 + }, + { + "epoch": 0.5500486560853662, + "grad_norm": 1.0455659627914429, + "learning_rate": 8.506404130219479e-05, + "loss": 0.6674, + "step": 8196 + }, + { + "epoch": 0.5501828797691353, + "grad_norm": 1.0172463655471802, + "learning_rate": 8.505629261990995e-05, + "loss": 0.691, + "step": 8198 + }, + { + "epoch": 0.5503171034529043, + "grad_norm": 1.0176228284835815, + "learning_rate": 8.504854228128422e-05, + "loss": 0.6475, + "step": 8200 + }, + { + "epoch": 0.5504513271366732, + "grad_norm": 0.9945377707481384, + "learning_rate": 8.504079028668379e-05, + "loss": 0.6226, + "step": 8202 + }, + { + "epoch": 0.5505855508204422, + "grad_norm": 1.0326147079467773, + "learning_rate": 8.503303663647492e-05, + "loss": 0.5734, + "step": 8204 + }, + { + "epoch": 0.5507197745042113, + "grad_norm": 1.0489661693572998, + "learning_rate": 8.502528133102397e-05, + "loss": 0.5622, + "step": 8206 + }, + { + "epoch": 0.5508539981879803, + "grad_norm": 1.0645067691802979, + "learning_rate": 8.501752437069734e-05, + "loss": 0.6315, + "step": 8208 + }, + { + "epoch": 0.5509882218717492, + "grad_norm": 2.246250629425049, + "learning_rate": 8.500976575586155e-05, + "loss": 0.6108, + "step": 8210 + }, + { + "epoch": 0.5511224455555183, + "grad_norm": 0.9789879322052002, + "learning_rate": 8.500200548688318e-05, + "loss": 0.7113, + "step": 8212 + }, + { + "epoch": 0.5512566692392873, + "grad_norm": 1.8534557819366455, + "learning_rate": 8.499424356412886e-05, + "loss": 0.6391, + "step": 8214 + }, + { + "epoch": 0.5513908929230563, + "grad_norm": 1.0547062158584595, + "learning_rate": 8.498647998796534e-05, + "loss": 0.6031, + "step": 8216 + }, + { + "epoch": 0.5515251166068252, + "grad_norm": 1.0872576236724854, + "learning_rate": 8.497871475875946e-05, + "loss": 0.6412, + "step": 8218 + }, + { + "epoch": 0.5516593402905943, + "grad_norm": 1.004032850265503, + "learning_rate": 8.497094787687807e-05, + "loss": 0.6903, + "step": 8220 + }, + { + "epoch": 0.5517935639743633, + "grad_norm": 1.2047981023788452, + "learning_rate": 8.496317934268818e-05, + "loss": 0.6767, + "step": 8222 + }, + { + "epoch": 0.5519277876581323, + "grad_norm": 1.089982271194458, + "learning_rate": 8.49554091565568e-05, + "loss": 0.6252, + "step": 8224 + }, + { + "epoch": 0.5520620113419012, + "grad_norm": 1.1389589309692383, + "learning_rate": 8.49476373188511e-05, + "loss": 0.6627, + "step": 8226 + }, + { + "epoch": 0.5521962350256703, + "grad_norm": 1.0517256259918213, + "learning_rate": 8.493986382993823e-05, + "loss": 0.6275, + "step": 8228 + }, + { + "epoch": 0.5523304587094393, + "grad_norm": 0.9931901693344116, + "learning_rate": 8.49320886901855e-05, + "loss": 0.6115, + "step": 8230 + }, + { + "epoch": 0.5524646823932082, + "grad_norm": 0.9084005951881409, + "learning_rate": 8.492431189996029e-05, + "loss": 0.5321, + "step": 8232 + }, + { + "epoch": 0.5525989060769773, + "grad_norm": 1.0069876909255981, + "learning_rate": 8.491653345963002e-05, + "loss": 0.6006, + "step": 8234 + }, + { + "epoch": 0.5527331297607463, + "grad_norm": 1.0372812747955322, + "learning_rate": 8.490875336956219e-05, + "loss": 0.6397, + "step": 8236 + }, + { + "epoch": 0.5528673534445153, + "grad_norm": 1.2959024906158447, + "learning_rate": 8.49009716301244e-05, + "loss": 0.67, + "step": 8238 + }, + { + "epoch": 0.5530015771282842, + "grad_norm": 0.9682850241661072, + "learning_rate": 8.489318824168435e-05, + "loss": 0.6599, + "step": 8240 + }, + { + "epoch": 0.5531358008120533, + "grad_norm": 1.1090869903564453, + "learning_rate": 8.488540320460975e-05, + "loss": 0.6012, + "step": 8242 + }, + { + "epoch": 0.5532700244958223, + "grad_norm": 1.2157872915267944, + "learning_rate": 8.487761651926848e-05, + "loss": 0.6433, + "step": 8244 + }, + { + "epoch": 0.5534042481795913, + "grad_norm": 1.5864654779434204, + "learning_rate": 8.48698281860284e-05, + "loss": 0.6466, + "step": 8246 + }, + { + "epoch": 0.5535384718633602, + "grad_norm": 1.0488078594207764, + "learning_rate": 8.486203820525748e-05, + "loss": 0.5906, + "step": 8248 + }, + { + "epoch": 0.5536726955471293, + "grad_norm": 1.1140676736831665, + "learning_rate": 8.485424657732384e-05, + "loss": 0.6756, + "step": 8250 + }, + { + "epoch": 0.5538069192308983, + "grad_norm": 1.0485445261001587, + "learning_rate": 8.484645330259557e-05, + "loss": 0.6702, + "step": 8252 + }, + { + "epoch": 0.5539411429146673, + "grad_norm": 1.054240107536316, + "learning_rate": 8.483865838144092e-05, + "loss": 0.6003, + "step": 8254 + }, + { + "epoch": 0.5540753665984363, + "grad_norm": 1.060673475265503, + "learning_rate": 8.483086181422816e-05, + "loss": 0.604, + "step": 8256 + }, + { + "epoch": 0.5542095902822053, + "grad_norm": 1.134052038192749, + "learning_rate": 8.482306360132567e-05, + "loss": 0.6691, + "step": 8258 + }, + { + "epoch": 0.5543438139659743, + "grad_norm": 0.9479762315750122, + "learning_rate": 8.48152637431019e-05, + "loss": 0.6215, + "step": 8260 + }, + { + "epoch": 0.5544780376497433, + "grad_norm": 1.013800024986267, + "learning_rate": 8.480746223992539e-05, + "loss": 0.6914, + "step": 8262 + }, + { + "epoch": 0.5546122613335123, + "grad_norm": 1.0438390970230103, + "learning_rate": 8.479965909216471e-05, + "loss": 0.6848, + "step": 8264 + }, + { + "epoch": 0.5547464850172813, + "grad_norm": 1.0111439228057861, + "learning_rate": 8.479185430018858e-05, + "loss": 0.6174, + "step": 8266 + }, + { + "epoch": 0.5548807087010503, + "grad_norm": 1.811466097831726, + "learning_rate": 8.478404786436577e-05, + "loss": 0.6816, + "step": 8268 + }, + { + "epoch": 0.5550149323848192, + "grad_norm": 1.1030131578445435, + "learning_rate": 8.477623978506508e-05, + "loss": 0.6428, + "step": 8270 + }, + { + "epoch": 0.5551491560685883, + "grad_norm": 1.0027936697006226, + "learning_rate": 8.476843006265545e-05, + "loss": 0.5804, + "step": 8272 + }, + { + "epoch": 0.5552833797523573, + "grad_norm": 1.1092429161071777, + "learning_rate": 8.476061869750586e-05, + "loss": 0.698, + "step": 8274 + }, + { + "epoch": 0.5554176034361263, + "grad_norm": 1.0696680545806885, + "learning_rate": 8.475280568998541e-05, + "loss": 0.6074, + "step": 8276 + }, + { + "epoch": 0.5555518271198953, + "grad_norm": 2.5581133365631104, + "learning_rate": 8.474499104046322e-05, + "loss": 0.6166, + "step": 8278 + }, + { + "epoch": 0.5556860508036643, + "grad_norm": 1.0897672176361084, + "learning_rate": 8.473717474930851e-05, + "loss": 0.6882, + "step": 8280 + }, + { + "epoch": 0.5558202744874333, + "grad_norm": 1.124261736869812, + "learning_rate": 8.472935681689063e-05, + "loss": 0.5806, + "step": 8282 + }, + { + "epoch": 0.5559544981712023, + "grad_norm": 1.0583689212799072, + "learning_rate": 8.472153724357892e-05, + "loss": 0.7294, + "step": 8284 + }, + { + "epoch": 0.5560887218549713, + "grad_norm": 1.1011725664138794, + "learning_rate": 8.471371602974284e-05, + "loss": 0.6707, + "step": 8286 + }, + { + "epoch": 0.5562229455387403, + "grad_norm": 1.0401387214660645, + "learning_rate": 8.470589317575194e-05, + "loss": 0.5701, + "step": 8288 + }, + { + "epoch": 0.5563571692225093, + "grad_norm": 1.064677119255066, + "learning_rate": 8.469806868197585e-05, + "loss": 0.6673, + "step": 8290 + }, + { + "epoch": 0.5564913929062784, + "grad_norm": 1.0974767208099365, + "learning_rate": 8.469024254878424e-05, + "loss": 0.6954, + "step": 8292 + }, + { + "epoch": 0.5566256165900473, + "grad_norm": 1.0445501804351807, + "learning_rate": 8.468241477654688e-05, + "loss": 0.6022, + "step": 8294 + }, + { + "epoch": 0.5567598402738163, + "grad_norm": 1.1602227687835693, + "learning_rate": 8.467458536563362e-05, + "loss": 0.6687, + "step": 8296 + }, + { + "epoch": 0.5568940639575853, + "grad_norm": 0.9773215651512146, + "learning_rate": 8.46667543164144e-05, + "loss": 0.6445, + "step": 8298 + }, + { + "epoch": 0.5570282876413544, + "grad_norm": 1.013282299041748, + "learning_rate": 8.465892162925919e-05, + "loss": 0.6519, + "step": 8300 + }, + { + "epoch": 0.5571625113251233, + "grad_norm": 1.0231740474700928, + "learning_rate": 8.46510873045381e-05, + "loss": 0.643, + "step": 8302 + }, + { + "epoch": 0.5572967350088923, + "grad_norm": 1.308756709098816, + "learning_rate": 8.464325134262126e-05, + "loss": 0.6785, + "step": 8304 + }, + { + "epoch": 0.5574309586926613, + "grad_norm": 0.9831346273422241, + "learning_rate": 8.463541374387894e-05, + "loss": 0.7026, + "step": 8306 + }, + { + "epoch": 0.5575651823764303, + "grad_norm": 1.050607442855835, + "learning_rate": 8.462757450868139e-05, + "loss": 0.6293, + "step": 8308 + }, + { + "epoch": 0.5576994060601993, + "grad_norm": 1.0742460489273071, + "learning_rate": 8.461973363739906e-05, + "loss": 0.6521, + "step": 8310 + }, + { + "epoch": 0.5578336297439683, + "grad_norm": 1.148586392402649, + "learning_rate": 8.46118911304024e-05, + "loss": 0.6586, + "step": 8312 + }, + { + "epoch": 0.5579678534277374, + "grad_norm": 1.0171434879302979, + "learning_rate": 8.460404698806194e-05, + "loss": 0.6123, + "step": 8314 + }, + { + "epoch": 0.5581020771115063, + "grad_norm": 0.8852545022964478, + "learning_rate": 8.45962012107483e-05, + "loss": 0.5919, + "step": 8316 + }, + { + "epoch": 0.5582363007952753, + "grad_norm": 1.4615974426269531, + "learning_rate": 8.458835379883219e-05, + "loss": 0.6773, + "step": 8318 + }, + { + "epoch": 0.5583705244790443, + "grad_norm": 1.1346313953399658, + "learning_rate": 8.458050475268437e-05, + "loss": 0.6652, + "step": 8320 + }, + { + "epoch": 0.5585047481628134, + "grad_norm": 2.0286874771118164, + "learning_rate": 8.457265407267572e-05, + "loss": 0.6489, + "step": 8322 + }, + { + "epoch": 0.5586389718465823, + "grad_norm": 1.1897507905960083, + "learning_rate": 8.456480175917713e-05, + "loss": 0.6463, + "step": 8324 + }, + { + "epoch": 0.5587731955303513, + "grad_norm": 1.1002095937728882, + "learning_rate": 8.455694781255963e-05, + "loss": 0.6908, + "step": 8326 + }, + { + "epoch": 0.5589074192141203, + "grad_norm": 1.0195788145065308, + "learning_rate": 8.454909223319429e-05, + "loss": 0.5828, + "step": 8328 + }, + { + "epoch": 0.5590416428978894, + "grad_norm": 1.0193231105804443, + "learning_rate": 8.45412350214523e-05, + "loss": 0.6408, + "step": 8330 + }, + { + "epoch": 0.5591758665816583, + "grad_norm": 1.0430853366851807, + "learning_rate": 8.453337617770486e-05, + "loss": 0.6657, + "step": 8332 + }, + { + "epoch": 0.5593100902654273, + "grad_norm": 1.0074166059494019, + "learning_rate": 8.452551570232331e-05, + "loss": 0.6076, + "step": 8334 + }, + { + "epoch": 0.5594443139491964, + "grad_norm": 0.95444256067276, + "learning_rate": 8.451765359567903e-05, + "loss": 0.6303, + "step": 8336 + }, + { + "epoch": 0.5595785376329654, + "grad_norm": 1.0800069570541382, + "learning_rate": 8.45097898581435e-05, + "loss": 0.652, + "step": 8338 + }, + { + "epoch": 0.5597127613167343, + "grad_norm": 1.1131151914596558, + "learning_rate": 8.450192449008825e-05, + "loss": 0.6429, + "step": 8340 + }, + { + "epoch": 0.5598469850005033, + "grad_norm": 1.0435813665390015, + "learning_rate": 8.449405749188493e-05, + "loss": 0.5579, + "step": 8342 + }, + { + "epoch": 0.5599812086842724, + "grad_norm": 1.211788535118103, + "learning_rate": 8.448618886390522e-05, + "loss": 0.6314, + "step": 8344 + }, + { + "epoch": 0.5601154323680413, + "grad_norm": 1.8600342273712158, + "learning_rate": 8.447831860652089e-05, + "loss": 0.6245, + "step": 8346 + }, + { + "epoch": 0.5602496560518103, + "grad_norm": 1.1609203815460205, + "learning_rate": 8.447044672010382e-05, + "loss": 0.652, + "step": 8348 + }, + { + "epoch": 0.5603838797355793, + "grad_norm": 1.0970147848129272, + "learning_rate": 8.44625732050259e-05, + "loss": 0.6329, + "step": 8350 + }, + { + "epoch": 0.5605181034193484, + "grad_norm": 0.9530807733535767, + "learning_rate": 8.445469806165918e-05, + "loss": 0.6211, + "step": 8352 + }, + { + "epoch": 0.5606523271031173, + "grad_norm": 1.526509165763855, + "learning_rate": 8.444682129037574e-05, + "loss": 0.6128, + "step": 8354 + }, + { + "epoch": 0.5607865507868863, + "grad_norm": 1.1120033264160156, + "learning_rate": 8.44389428915477e-05, + "loss": 0.6429, + "step": 8356 + }, + { + "epoch": 0.5609207744706554, + "grad_norm": 1.0139830112457275, + "learning_rate": 8.443106286554737e-05, + "loss": 0.6424, + "step": 8358 + }, + { + "epoch": 0.5610549981544244, + "grad_norm": 1.029532551765442, + "learning_rate": 8.4423181212747e-05, + "loss": 0.6396, + "step": 8360 + }, + { + "epoch": 0.5611892218381933, + "grad_norm": 1.0851560831069946, + "learning_rate": 8.4415297933519e-05, + "loss": 0.6841, + "step": 8362 + }, + { + "epoch": 0.5613234455219623, + "grad_norm": 1.0855809450149536, + "learning_rate": 8.440741302823585e-05, + "loss": 0.6103, + "step": 8364 + }, + { + "epoch": 0.5614576692057314, + "grad_norm": 0.9500477313995361, + "learning_rate": 8.43995264972701e-05, + "loss": 0.5849, + "step": 8366 + }, + { + "epoch": 0.5615918928895004, + "grad_norm": 1.2251765727996826, + "learning_rate": 8.439163834099437e-05, + "loss": 0.6787, + "step": 8368 + }, + { + "epoch": 0.5617261165732693, + "grad_norm": 1.1215698719024658, + "learning_rate": 8.438374855978136e-05, + "loss": 0.5894, + "step": 8370 + }, + { + "epoch": 0.5618603402570383, + "grad_norm": 0.994908332824707, + "learning_rate": 8.437585715400384e-05, + "loss": 0.5808, + "step": 8372 + }, + { + "epoch": 0.5619945639408074, + "grad_norm": 0.9743037819862366, + "learning_rate": 8.436796412403466e-05, + "loss": 0.5433, + "step": 8374 + }, + { + "epoch": 0.5621287876245764, + "grad_norm": 0.9671458601951599, + "learning_rate": 8.436006947024677e-05, + "loss": 0.5827, + "step": 8376 + }, + { + "epoch": 0.5622630113083453, + "grad_norm": 1.092995047569275, + "learning_rate": 8.435217319301315e-05, + "loss": 0.5866, + "step": 8378 + }, + { + "epoch": 0.5623972349921144, + "grad_norm": 1.0426753759384155, + "learning_rate": 8.43442752927069e-05, + "loss": 0.6717, + "step": 8380 + }, + { + "epoch": 0.5625314586758834, + "grad_norm": 0.9841560125350952, + "learning_rate": 8.433637576970119e-05, + "loss": 0.6382, + "step": 8382 + }, + { + "epoch": 0.5626656823596523, + "grad_norm": 1.1644257307052612, + "learning_rate": 8.432847462436924e-05, + "loss": 0.6793, + "step": 8384 + }, + { + "epoch": 0.5627999060434213, + "grad_norm": 1.0643762350082397, + "learning_rate": 8.432057185708436e-05, + "loss": 0.5893, + "step": 8386 + }, + { + "epoch": 0.5629341297271904, + "grad_norm": 1.0374327898025513, + "learning_rate": 8.431266746821995e-05, + "loss": 0.63, + "step": 8388 + }, + { + "epoch": 0.5630683534109594, + "grad_norm": 0.9825643301010132, + "learning_rate": 8.430476145814948e-05, + "loss": 0.6431, + "step": 8390 + }, + { + "epoch": 0.5632025770947283, + "grad_norm": 1.111504077911377, + "learning_rate": 8.429685382724648e-05, + "loss": 0.6441, + "step": 8392 + }, + { + "epoch": 0.5633368007784973, + "grad_norm": 1.1499359607696533, + "learning_rate": 8.428894457588458e-05, + "loss": 0.6429, + "step": 8394 + }, + { + "epoch": 0.5634710244622664, + "grad_norm": 1.018369197845459, + "learning_rate": 8.428103370443748e-05, + "loss": 0.6267, + "step": 8396 + }, + { + "epoch": 0.5636052481460354, + "grad_norm": 0.9640548229217529, + "learning_rate": 8.427312121327895e-05, + "loss": 0.6659, + "step": 8398 + }, + { + "epoch": 0.5637394718298043, + "grad_norm": 1.0157394409179688, + "learning_rate": 8.426520710278283e-05, + "loss": 0.6143, + "step": 8400 + }, + { + "epoch": 0.5638736955135734, + "grad_norm": 1.0569133758544922, + "learning_rate": 8.425729137332306e-05, + "loss": 0.6107, + "step": 8402 + }, + { + "epoch": 0.5640079191973424, + "grad_norm": 0.9883301854133606, + "learning_rate": 8.424937402527363e-05, + "loss": 0.6164, + "step": 8404 + }, + { + "epoch": 0.5641421428811114, + "grad_norm": 1.2704092264175415, + "learning_rate": 8.424145505900863e-05, + "loss": 0.6932, + "step": 8406 + }, + { + "epoch": 0.5642763665648803, + "grad_norm": 1.6753979921340942, + "learning_rate": 8.423353447490221e-05, + "loss": 0.6067, + "step": 8408 + }, + { + "epoch": 0.5644105902486494, + "grad_norm": 0.9905928373336792, + "learning_rate": 8.422561227332861e-05, + "loss": 0.6036, + "step": 8410 + }, + { + "epoch": 0.5645448139324184, + "grad_norm": 1.1601665019989014, + "learning_rate": 8.421768845466213e-05, + "loss": 0.6253, + "step": 8412 + }, + { + "epoch": 0.5646790376161874, + "grad_norm": 0.9416701197624207, + "learning_rate": 8.420976301927716e-05, + "loss": 0.6642, + "step": 8414 + }, + { + "epoch": 0.5648132612999563, + "grad_norm": 1.2624140977859497, + "learning_rate": 8.420183596754816e-05, + "loss": 0.6034, + "step": 8416 + }, + { + "epoch": 0.5649474849837254, + "grad_norm": 1.0920568704605103, + "learning_rate": 8.419390729984966e-05, + "loss": 0.593, + "step": 8418 + }, + { + "epoch": 0.5650817086674944, + "grad_norm": 0.9582425355911255, + "learning_rate": 8.418597701655628e-05, + "loss": 0.6524, + "step": 8420 + }, + { + "epoch": 0.5652159323512633, + "grad_norm": 1.02560555934906, + "learning_rate": 8.41780451180427e-05, + "loss": 0.6018, + "step": 8422 + }, + { + "epoch": 0.5653501560350324, + "grad_norm": 1.0681962966918945, + "learning_rate": 8.417011160468371e-05, + "loss": 0.5941, + "step": 8424 + }, + { + "epoch": 0.5654843797188014, + "grad_norm": 0.892360270023346, + "learning_rate": 8.416217647685411e-05, + "loss": 0.6703, + "step": 8426 + }, + { + "epoch": 0.5656186034025704, + "grad_norm": 0.9655805826187134, + "learning_rate": 8.415423973492888e-05, + "loss": 0.5849, + "step": 8428 + }, + { + "epoch": 0.5657528270863393, + "grad_norm": 1.1360067129135132, + "learning_rate": 8.414630137928298e-05, + "loss": 0.6583, + "step": 8430 + }, + { + "epoch": 0.5658870507701084, + "grad_norm": 0.9449290633201599, + "learning_rate": 8.413836141029148e-05, + "loss": 0.6052, + "step": 8432 + }, + { + "epoch": 0.5660212744538774, + "grad_norm": 0.9781023263931274, + "learning_rate": 8.413041982832954e-05, + "loss": 0.6945, + "step": 8434 + }, + { + "epoch": 0.5661554981376464, + "grad_norm": 1.038053035736084, + "learning_rate": 8.412247663377237e-05, + "loss": 0.6892, + "step": 8436 + }, + { + "epoch": 0.5662897218214153, + "grad_norm": 1.099488615989685, + "learning_rate": 8.411453182699529e-05, + "loss": 0.6034, + "step": 8438 + }, + { + "epoch": 0.5664239455051844, + "grad_norm": 1.1301482915878296, + "learning_rate": 8.410658540837364e-05, + "loss": 0.6724, + "step": 8440 + }, + { + "epoch": 0.5665581691889534, + "grad_norm": 1.0925681591033936, + "learning_rate": 8.409863737828292e-05, + "loss": 0.6736, + "step": 8442 + }, + { + "epoch": 0.5666923928727224, + "grad_norm": 1.0710530281066895, + "learning_rate": 8.409068773709862e-05, + "loss": 0.7052, + "step": 8444 + }, + { + "epoch": 0.5668266165564914, + "grad_norm": 1.0949926376342773, + "learning_rate": 8.408273648519638e-05, + "loss": 0.697, + "step": 8446 + }, + { + "epoch": 0.5669608402402604, + "grad_norm": 1.0935205221176147, + "learning_rate": 8.407478362295184e-05, + "loss": 0.656, + "step": 8448 + }, + { + "epoch": 0.5670950639240294, + "grad_norm": 1.2679986953735352, + "learning_rate": 8.40668291507408e-05, + "loss": 0.6642, + "step": 8450 + }, + { + "epoch": 0.5672292876077984, + "grad_norm": 1.0844709873199463, + "learning_rate": 8.405887306893906e-05, + "loss": 0.6574, + "step": 8452 + }, + { + "epoch": 0.5673635112915674, + "grad_norm": 1.0213552713394165, + "learning_rate": 8.405091537792253e-05, + "loss": 0.6433, + "step": 8454 + }, + { + "epoch": 0.5674977349753364, + "grad_norm": 0.9217315912246704, + "learning_rate": 8.404295607806723e-05, + "loss": 0.6107, + "step": 8456 + }, + { + "epoch": 0.5676319586591054, + "grad_norm": 1.058948278427124, + "learning_rate": 8.40349951697492e-05, + "loss": 0.7232, + "step": 8458 + }, + { + "epoch": 0.5677661823428743, + "grad_norm": 1.2261768579483032, + "learning_rate": 8.402703265334455e-05, + "loss": 0.658, + "step": 8460 + }, + { + "epoch": 0.5679004060266434, + "grad_norm": 1.1236143112182617, + "learning_rate": 8.401906852922954e-05, + "loss": 0.6335, + "step": 8462 + }, + { + "epoch": 0.5680346297104124, + "grad_norm": 1.1284643411636353, + "learning_rate": 8.401110279778043e-05, + "loss": 0.625, + "step": 8464 + }, + { + "epoch": 0.5681688533941814, + "grad_norm": 1.1031262874603271, + "learning_rate": 8.400313545937359e-05, + "loss": 0.663, + "step": 8466 + }, + { + "epoch": 0.5683030770779504, + "grad_norm": 1.0434975624084473, + "learning_rate": 8.399516651438546e-05, + "loss": 0.6443, + "step": 8468 + }, + { + "epoch": 0.5684373007617194, + "grad_norm": 1.0460388660430908, + "learning_rate": 8.398719596319258e-05, + "loss": 0.6978, + "step": 8470 + }, + { + "epoch": 0.5685715244454884, + "grad_norm": 1.0447012186050415, + "learning_rate": 8.39792238061715e-05, + "loss": 0.6135, + "step": 8472 + }, + { + "epoch": 0.5687057481292574, + "grad_norm": 1.3942762613296509, + "learning_rate": 8.397125004369892e-05, + "loss": 0.5721, + "step": 8474 + }, + { + "epoch": 0.5688399718130264, + "grad_norm": 1.027471661567688, + "learning_rate": 8.396327467615158e-05, + "loss": 0.62, + "step": 8476 + }, + { + "epoch": 0.5689741954967954, + "grad_norm": 0.9808257222175598, + "learning_rate": 8.39552977039063e-05, + "loss": 0.6354, + "step": 8478 + }, + { + "epoch": 0.5691084191805644, + "grad_norm": 1.0226244926452637, + "learning_rate": 8.394731912733998e-05, + "loss": 0.6141, + "step": 8480 + }, + { + "epoch": 0.5692426428643335, + "grad_norm": 1.0700534582138062, + "learning_rate": 8.393933894682957e-05, + "loss": 0.6044, + "step": 8482 + }, + { + "epoch": 0.5693768665481024, + "grad_norm": 0.8504033088684082, + "learning_rate": 8.393135716275215e-05, + "loss": 0.5722, + "step": 8484 + }, + { + "epoch": 0.5695110902318714, + "grad_norm": 1.060200810432434, + "learning_rate": 8.392337377548481e-05, + "loss": 0.6351, + "step": 8486 + }, + { + "epoch": 0.5696453139156404, + "grad_norm": 1.0870810747146606, + "learning_rate": 8.391538878540477e-05, + "loss": 0.622, + "step": 8488 + }, + { + "epoch": 0.5697795375994095, + "grad_norm": 1.059228777885437, + "learning_rate": 8.390740219288931e-05, + "loss": 0.662, + "step": 8490 + }, + { + "epoch": 0.5699137612831784, + "grad_norm": 1.3571457862854004, + "learning_rate": 8.389941399831578e-05, + "loss": 0.6109, + "step": 8492 + }, + { + "epoch": 0.5700479849669474, + "grad_norm": 1.2545329332351685, + "learning_rate": 8.389142420206158e-05, + "loss": 0.729, + "step": 8494 + }, + { + "epoch": 0.5701822086507164, + "grad_norm": 1.0990122556686401, + "learning_rate": 8.388343280450424e-05, + "loss": 0.6821, + "step": 8496 + }, + { + "epoch": 0.5703164323344854, + "grad_norm": 1.2406752109527588, + "learning_rate": 8.387543980602133e-05, + "loss": 0.5928, + "step": 8498 + }, + { + "epoch": 0.5704506560182544, + "grad_norm": 1.21001398563385, + "learning_rate": 8.38674452069905e-05, + "loss": 0.6895, + "step": 8500 + }, + { + "epoch": 0.5705848797020234, + "grad_norm": 1.1252810955047607, + "learning_rate": 8.385944900778948e-05, + "loss": 0.6108, + "step": 8502 + }, + { + "epoch": 0.5707191033857925, + "grad_norm": 1.0943095684051514, + "learning_rate": 8.385145120879607e-05, + "loss": 0.6347, + "step": 8504 + }, + { + "epoch": 0.5708533270695614, + "grad_norm": 1.1311140060424805, + "learning_rate": 8.384345181038818e-05, + "loss": 0.6565, + "step": 8506 + }, + { + "epoch": 0.5709875507533304, + "grad_norm": 1.0514534711837769, + "learning_rate": 8.383545081294371e-05, + "loss": 0.607, + "step": 8508 + }, + { + "epoch": 0.5711217744370994, + "grad_norm": 0.8207479119300842, + "learning_rate": 8.382744821684074e-05, + "loss": 0.5246, + "step": 8510 + }, + { + "epoch": 0.5712559981208685, + "grad_norm": 1.771498680114746, + "learning_rate": 8.381944402245736e-05, + "loss": 0.5949, + "step": 8512 + }, + { + "epoch": 0.5713902218046374, + "grad_norm": 1.1286104917526245, + "learning_rate": 8.381143823017176e-05, + "loss": 0.5539, + "step": 8514 + }, + { + "epoch": 0.5715244454884064, + "grad_norm": 1.0424946546554565, + "learning_rate": 8.380343084036218e-05, + "loss": 0.6896, + "step": 8516 + }, + { + "epoch": 0.5716586691721754, + "grad_norm": 1.213830590248108, + "learning_rate": 8.379542185340696e-05, + "loss": 0.6418, + "step": 8518 + }, + { + "epoch": 0.5717928928559445, + "grad_norm": 1.050347924232483, + "learning_rate": 8.378741126968453e-05, + "loss": 0.6655, + "step": 8520 + }, + { + "epoch": 0.5719271165397134, + "grad_norm": 1.3137097358703613, + "learning_rate": 8.377939908957336e-05, + "loss": 0.6701, + "step": 8522 + }, + { + "epoch": 0.5720613402234824, + "grad_norm": 1.074523687362671, + "learning_rate": 8.377138531345198e-05, + "loss": 0.6345, + "step": 8524 + }, + { + "epoch": 0.5721955639072515, + "grad_norm": 1.134978175163269, + "learning_rate": 8.376336994169908e-05, + "loss": 0.6571, + "step": 8526 + }, + { + "epoch": 0.5723297875910205, + "grad_norm": 1.0185121297836304, + "learning_rate": 8.375535297469332e-05, + "loss": 0.6489, + "step": 8528 + }, + { + "epoch": 0.5724640112747894, + "grad_norm": 1.0336315631866455, + "learning_rate": 8.374733441281353e-05, + "loss": 0.6149, + "step": 8530 + }, + { + "epoch": 0.5725982349585584, + "grad_norm": 1.0753029584884644, + "learning_rate": 8.373931425643855e-05, + "loss": 0.6491, + "step": 8532 + }, + { + "epoch": 0.5727324586423275, + "grad_norm": 1.0328943729400635, + "learning_rate": 8.37312925059473e-05, + "loss": 0.6224, + "step": 8534 + }, + { + "epoch": 0.5728666823260964, + "grad_norm": 1.2001701593399048, + "learning_rate": 8.372326916171882e-05, + "loss": 0.6266, + "step": 8536 + }, + { + "epoch": 0.5730009060098654, + "grad_norm": 1.1070367097854614, + "learning_rate": 8.371524422413218e-05, + "loss": 0.6204, + "step": 8538 + }, + { + "epoch": 0.5731351296936344, + "grad_norm": 1.0270267724990845, + "learning_rate": 8.370721769356656e-05, + "loss": 0.6376, + "step": 8540 + }, + { + "epoch": 0.5732693533774035, + "grad_norm": 1.089913249015808, + "learning_rate": 8.369918957040116e-05, + "loss": 0.6562, + "step": 8542 + }, + { + "epoch": 0.5734035770611724, + "grad_norm": 0.9418351054191589, + "learning_rate": 8.369115985501534e-05, + "loss": 0.5989, + "step": 8544 + }, + { + "epoch": 0.5735378007449414, + "grad_norm": 1.1004632711410522, + "learning_rate": 8.368312854778848e-05, + "loss": 0.604, + "step": 8546 + }, + { + "epoch": 0.5736720244287105, + "grad_norm": 1.0900779962539673, + "learning_rate": 8.367509564910001e-05, + "loss": 0.6368, + "step": 8548 + }, + { + "epoch": 0.5738062481124795, + "grad_norm": 1.1029690504074097, + "learning_rate": 8.366706115932951e-05, + "loss": 0.6536, + "step": 8550 + }, + { + "epoch": 0.5739404717962484, + "grad_norm": 1.3655116558074951, + "learning_rate": 8.365902507885656e-05, + "loss": 0.699, + "step": 8552 + }, + { + "epoch": 0.5740746954800174, + "grad_norm": 0.9136799573898315, + "learning_rate": 8.365098740806086e-05, + "loss": 0.6259, + "step": 8554 + }, + { + "epoch": 0.5742089191637865, + "grad_norm": 1.1151015758514404, + "learning_rate": 8.364294814732218e-05, + "loss": 0.6007, + "step": 8556 + }, + { + "epoch": 0.5743431428475555, + "grad_norm": 0.9604775309562683, + "learning_rate": 8.363490729702034e-05, + "loss": 0.5917, + "step": 8558 + }, + { + "epoch": 0.5744773665313244, + "grad_norm": 1.218360424041748, + "learning_rate": 8.362686485753531e-05, + "loss": 0.6404, + "step": 8560 + }, + { + "epoch": 0.5746115902150934, + "grad_norm": 0.926568329334259, + "learning_rate": 8.361882082924702e-05, + "loss": 0.5643, + "step": 8562 + }, + { + "epoch": 0.5747458138988625, + "grad_norm": 1.341833233833313, + "learning_rate": 8.361077521253553e-05, + "loss": 0.6566, + "step": 8564 + }, + { + "epoch": 0.5748800375826315, + "grad_norm": 1.016913890838623, + "learning_rate": 8.360272800778103e-05, + "loss": 0.6195, + "step": 8566 + }, + { + "epoch": 0.5750142612664004, + "grad_norm": 1.1553874015808105, + "learning_rate": 8.359467921536371e-05, + "loss": 0.6431, + "step": 8568 + }, + { + "epoch": 0.5751484849501695, + "grad_norm": 1.1712065935134888, + "learning_rate": 8.358662883566383e-05, + "loss": 0.6416, + "step": 8570 + }, + { + "epoch": 0.5752827086339385, + "grad_norm": 1.0247400999069214, + "learning_rate": 8.357857686906182e-05, + "loss": 0.692, + "step": 8572 + }, + { + "epoch": 0.5754169323177074, + "grad_norm": 0.9636111855506897, + "learning_rate": 8.357052331593807e-05, + "loss": 0.6063, + "step": 8574 + }, + { + "epoch": 0.5755511560014764, + "grad_norm": 1.0941756963729858, + "learning_rate": 8.35624681766731e-05, + "loss": 0.5797, + "step": 8576 + }, + { + "epoch": 0.5756853796852455, + "grad_norm": 1.0538661479949951, + "learning_rate": 8.355441145164751e-05, + "loss": 0.6479, + "step": 8578 + }, + { + "epoch": 0.5758196033690145, + "grad_norm": 1.0697072744369507, + "learning_rate": 8.354635314124195e-05, + "loss": 0.5541, + "step": 8580 + }, + { + "epoch": 0.5759538270527834, + "grad_norm": 1.0581369400024414, + "learning_rate": 8.35382932458372e-05, + "loss": 0.6651, + "step": 8582 + }, + { + "epoch": 0.5760880507365524, + "grad_norm": 1.2765485048294067, + "learning_rate": 8.353023176581401e-05, + "loss": 0.6432, + "step": 8584 + }, + { + "epoch": 0.5762222744203215, + "grad_norm": 1.1463816165924072, + "learning_rate": 8.352216870155331e-05, + "loss": 0.7451, + "step": 8586 + }, + { + "epoch": 0.5763564981040905, + "grad_norm": 0.9963604211807251, + "learning_rate": 8.351410405343607e-05, + "loss": 0.6129, + "step": 8588 + }, + { + "epoch": 0.5764907217878594, + "grad_norm": 0.9272325038909912, + "learning_rate": 8.350603782184332e-05, + "loss": 0.6512, + "step": 8590 + }, + { + "epoch": 0.5766249454716285, + "grad_norm": 0.9915642738342285, + "learning_rate": 8.349797000715616e-05, + "loss": 0.6561, + "step": 8592 + }, + { + "epoch": 0.5767591691553975, + "grad_norm": 1.0291296243667603, + "learning_rate": 8.348990060975578e-05, + "loss": 0.6414, + "step": 8594 + }, + { + "epoch": 0.5768933928391665, + "grad_norm": 0.9996158480644226, + "learning_rate": 8.348182963002346e-05, + "loss": 0.6249, + "step": 8596 + }, + { + "epoch": 0.5770276165229354, + "grad_norm": 1.1353191137313843, + "learning_rate": 8.347375706834055e-05, + "loss": 0.6113, + "step": 8598 + }, + { + "epoch": 0.5771618402067045, + "grad_norm": 0.9508293271064758, + "learning_rate": 8.34656829250884e-05, + "loss": 0.6033, + "step": 8600 + }, + { + "epoch": 0.5772960638904735, + "grad_norm": 1.0945531129837036, + "learning_rate": 8.345760720064856e-05, + "loss": 0.7605, + "step": 8602 + }, + { + "epoch": 0.5774302875742425, + "grad_norm": 1.0342299938201904, + "learning_rate": 8.344952989540259e-05, + "loss": 0.5899, + "step": 8604 + }, + { + "epoch": 0.5775645112580114, + "grad_norm": 0.990175724029541, + "learning_rate": 8.344145100973209e-05, + "loss": 0.5604, + "step": 8606 + }, + { + "epoch": 0.5776987349417805, + "grad_norm": 1.0384178161621094, + "learning_rate": 8.343337054401878e-05, + "loss": 0.6544, + "step": 8608 + }, + { + "epoch": 0.5778329586255495, + "grad_norm": 1.055012583732605, + "learning_rate": 8.342528849864447e-05, + "loss": 0.624, + "step": 8610 + }, + { + "epoch": 0.5779671823093184, + "grad_norm": 1.0311509370803833, + "learning_rate": 8.3417204873991e-05, + "loss": 0.6355, + "step": 8612 + }, + { + "epoch": 0.5781014059930875, + "grad_norm": 0.9782349467277527, + "learning_rate": 8.340911967044032e-05, + "loss": 0.6929, + "step": 8614 + }, + { + "epoch": 0.5782356296768565, + "grad_norm": 1.315628170967102, + "learning_rate": 8.340103288837443e-05, + "loss": 0.6516, + "step": 8616 + }, + { + "epoch": 0.5783698533606255, + "grad_norm": 1.2841339111328125, + "learning_rate": 8.33929445281754e-05, + "loss": 0.6612, + "step": 8618 + }, + { + "epoch": 0.5785040770443944, + "grad_norm": 1.027443766593933, + "learning_rate": 8.338485459022544e-05, + "loss": 0.5905, + "step": 8620 + }, + { + "epoch": 0.5786383007281635, + "grad_norm": 1.0931084156036377, + "learning_rate": 8.337676307490671e-05, + "loss": 0.6221, + "step": 8622 + }, + { + "epoch": 0.5787725244119325, + "grad_norm": 1.0320110321044922, + "learning_rate": 8.336866998260159e-05, + "loss": 0.6046, + "step": 8624 + }, + { + "epoch": 0.5789067480957015, + "grad_norm": 1.0718752145767212, + "learning_rate": 8.336057531369241e-05, + "loss": 0.5888, + "step": 8626 + }, + { + "epoch": 0.5790409717794704, + "grad_norm": 1.1324846744537354, + "learning_rate": 8.335247906856165e-05, + "loss": 0.6907, + "step": 8628 + }, + { + "epoch": 0.5791751954632395, + "grad_norm": 1.034125804901123, + "learning_rate": 8.334438124759184e-05, + "loss": 0.6191, + "step": 8630 + }, + { + "epoch": 0.5793094191470085, + "grad_norm": 1.0781384706497192, + "learning_rate": 8.333628185116559e-05, + "loss": 0.6483, + "step": 8632 + }, + { + "epoch": 0.5794436428307775, + "grad_norm": 1.0112380981445312, + "learning_rate": 8.332818087966556e-05, + "loss": 0.6759, + "step": 8634 + }, + { + "epoch": 0.5795778665145465, + "grad_norm": 1.0589327812194824, + "learning_rate": 8.332007833347454e-05, + "loss": 0.6546, + "step": 8636 + }, + { + "epoch": 0.5797120901983155, + "grad_norm": 1.160819411277771, + "learning_rate": 8.331197421297534e-05, + "loss": 0.5642, + "step": 8638 + }, + { + "epoch": 0.5798463138820845, + "grad_norm": 1.0162273645401, + "learning_rate": 8.330386851855086e-05, + "loss": 0.5833, + "step": 8640 + }, + { + "epoch": 0.5799805375658535, + "grad_norm": 0.9922532439231873, + "learning_rate": 8.329576125058406e-05, + "loss": 0.672, + "step": 8642 + }, + { + "epoch": 0.5801147612496225, + "grad_norm": 1.1330982446670532, + "learning_rate": 8.328765240945803e-05, + "loss": 0.6178, + "step": 8644 + }, + { + "epoch": 0.5802489849333915, + "grad_norm": 1.0933043956756592, + "learning_rate": 8.327954199555587e-05, + "loss": 0.6216, + "step": 8646 + }, + { + "epoch": 0.5803832086171605, + "grad_norm": 0.939629852771759, + "learning_rate": 8.327143000926082e-05, + "loss": 0.5978, + "step": 8648 + }, + { + "epoch": 0.5805174323009294, + "grad_norm": 1.3238186836242676, + "learning_rate": 8.32633164509561e-05, + "loss": 0.6486, + "step": 8650 + }, + { + "epoch": 0.5806516559846985, + "grad_norm": 0.9212239980697632, + "learning_rate": 8.32552013210251e-05, + "loss": 0.5924, + "step": 8652 + }, + { + "epoch": 0.5807858796684675, + "grad_norm": 1.0109730958938599, + "learning_rate": 8.324708461985124e-05, + "loss": 0.6334, + "step": 8654 + }, + { + "epoch": 0.5809201033522365, + "grad_norm": 1.1129043102264404, + "learning_rate": 8.323896634781799e-05, + "loss": 0.614, + "step": 8656 + }, + { + "epoch": 0.5810543270360055, + "grad_norm": 0.9045149087905884, + "learning_rate": 8.323084650530895e-05, + "loss": 0.6003, + "step": 8658 + }, + { + "epoch": 0.5811885507197745, + "grad_norm": 1.002319097518921, + "learning_rate": 8.322272509270777e-05, + "loss": 0.5334, + "step": 8660 + }, + { + "epoch": 0.5813227744035435, + "grad_norm": 1.0560046434402466, + "learning_rate": 8.321460211039814e-05, + "loss": 0.6468, + "step": 8662 + }, + { + "epoch": 0.5814569980873125, + "grad_norm": 1.061021089553833, + "learning_rate": 8.320647755876389e-05, + "loss": 0.6378, + "step": 8664 + }, + { + "epoch": 0.5815912217710815, + "grad_norm": 1.386268973350525, + "learning_rate": 8.319835143818887e-05, + "loss": 0.5953, + "step": 8666 + }, + { + "epoch": 0.5817254454548505, + "grad_norm": 1.3192063570022583, + "learning_rate": 8.319022374905701e-05, + "loss": 0.6294, + "step": 8668 + }, + { + "epoch": 0.5818596691386195, + "grad_norm": 1.0040688514709473, + "learning_rate": 8.318209449175236e-05, + "loss": 0.5872, + "step": 8670 + }, + { + "epoch": 0.5819938928223886, + "grad_norm": 1.0160542726516724, + "learning_rate": 8.317396366665899e-05, + "loss": 0.589, + "step": 8672 + }, + { + "epoch": 0.5821281165061575, + "grad_norm": 1.5633481740951538, + "learning_rate": 8.316583127416107e-05, + "loss": 0.631, + "step": 8674 + }, + { + "epoch": 0.5822623401899265, + "grad_norm": 1.3659327030181885, + "learning_rate": 8.315769731464284e-05, + "loss": 0.6798, + "step": 8676 + }, + { + "epoch": 0.5823965638736955, + "grad_norm": 1.0467239618301392, + "learning_rate": 8.314956178848861e-05, + "loss": 0.6086, + "step": 8678 + }, + { + "epoch": 0.5825307875574646, + "grad_norm": 1.1177936792373657, + "learning_rate": 8.314142469608278e-05, + "loss": 0.7187, + "step": 8680 + }, + { + "epoch": 0.5826650112412335, + "grad_norm": 1.061246633529663, + "learning_rate": 8.313328603780979e-05, + "loss": 0.6271, + "step": 8682 + }, + { + "epoch": 0.5827992349250025, + "grad_norm": 0.9757559895515442, + "learning_rate": 8.31251458140542e-05, + "loss": 0.6199, + "step": 8684 + }, + { + "epoch": 0.5829334586087715, + "grad_norm": 1.4294708967208862, + "learning_rate": 8.311700402520062e-05, + "loss": 0.6139, + "step": 8686 + }, + { + "epoch": 0.5830676822925405, + "grad_norm": 0.9727804660797119, + "learning_rate": 8.31088606716337e-05, + "loss": 0.637, + "step": 8688 + }, + { + "epoch": 0.5832019059763095, + "grad_norm": 1.0098894834518433, + "learning_rate": 8.310071575373823e-05, + "loss": 0.6181, + "step": 8690 + }, + { + "epoch": 0.5833361296600785, + "grad_norm": 1.1134440898895264, + "learning_rate": 8.309256927189903e-05, + "loss": 0.5976, + "step": 8692 + }, + { + "epoch": 0.5834703533438476, + "grad_norm": 1.059842586517334, + "learning_rate": 8.308442122650099e-05, + "loss": 0.591, + "step": 8694 + }, + { + "epoch": 0.5836045770276165, + "grad_norm": 1.0251708030700684, + "learning_rate": 8.307627161792912e-05, + "loss": 0.616, + "step": 8696 + }, + { + "epoch": 0.5837388007113855, + "grad_norm": 1.0120950937271118, + "learning_rate": 8.306812044656846e-05, + "loss": 0.6321, + "step": 8698 + }, + { + "epoch": 0.5838730243951545, + "grad_norm": 1.1307423114776611, + "learning_rate": 8.305996771280413e-05, + "loss": 0.6375, + "step": 8700 + }, + { + "epoch": 0.5840072480789236, + "grad_norm": 1.4316738843917847, + "learning_rate": 8.305181341702134e-05, + "loss": 0.6423, + "step": 8702 + }, + { + "epoch": 0.5841414717626925, + "grad_norm": 1.0380795001983643, + "learning_rate": 8.304365755960534e-05, + "loss": 0.7367, + "step": 8704 + }, + { + "epoch": 0.5842756954464615, + "grad_norm": 1.0263220071792603, + "learning_rate": 8.303550014094153e-05, + "loss": 0.6002, + "step": 8706 + }, + { + "epoch": 0.5844099191302305, + "grad_norm": 0.9541037678718567, + "learning_rate": 8.302734116141527e-05, + "loss": 0.5746, + "step": 8708 + }, + { + "epoch": 0.5845441428139996, + "grad_norm": 1.034319519996643, + "learning_rate": 8.30191806214121e-05, + "loss": 0.6303, + "step": 8710 + }, + { + "epoch": 0.5846783664977685, + "grad_norm": 1.0026488304138184, + "learning_rate": 8.301101852131757e-05, + "loss": 0.7003, + "step": 8712 + }, + { + "epoch": 0.5848125901815375, + "grad_norm": 1.1234585046768188, + "learning_rate": 8.300285486151734e-05, + "loss": 0.5807, + "step": 8714 + }, + { + "epoch": 0.5849468138653066, + "grad_norm": 1.171585202217102, + "learning_rate": 8.299468964239709e-05, + "loss": 0.643, + "step": 8716 + }, + { + "epoch": 0.5850810375490756, + "grad_norm": 1.1702193021774292, + "learning_rate": 8.298652286434265e-05, + "loss": 0.6512, + "step": 8718 + }, + { + "epoch": 0.5852152612328445, + "grad_norm": 0.9223757982254028, + "learning_rate": 8.297835452773988e-05, + "loss": 0.5944, + "step": 8720 + }, + { + "epoch": 0.5853494849166135, + "grad_norm": 1.0043474435806274, + "learning_rate": 8.297018463297471e-05, + "loss": 0.6106, + "step": 8722 + }, + { + "epoch": 0.5854837086003826, + "grad_norm": 1.0168185234069824, + "learning_rate": 8.296201318043313e-05, + "loss": 0.6452, + "step": 8724 + }, + { + "epoch": 0.5856179322841515, + "grad_norm": 0.9247617721557617, + "learning_rate": 8.295384017050125e-05, + "loss": 0.5894, + "step": 8726 + }, + { + "epoch": 0.5857521559679205, + "grad_norm": 1.0373064279556274, + "learning_rate": 8.294566560356522e-05, + "loss": 0.6591, + "step": 8728 + }, + { + "epoch": 0.5858863796516895, + "grad_norm": 1.0301313400268555, + "learning_rate": 8.293748948001129e-05, + "loss": 0.6364, + "step": 8730 + }, + { + "epoch": 0.5860206033354586, + "grad_norm": 1.2865626811981201, + "learning_rate": 8.292931180022574e-05, + "loss": 0.6155, + "step": 8732 + }, + { + "epoch": 0.5861548270192275, + "grad_norm": 1.0519150495529175, + "learning_rate": 8.292113256459495e-05, + "loss": 0.6381, + "step": 8734 + }, + { + "epoch": 0.5862890507029965, + "grad_norm": 1.0075963735580444, + "learning_rate": 8.29129517735054e-05, + "loss": 0.6134, + "step": 8736 + }, + { + "epoch": 0.5864232743867656, + "grad_norm": 0.9536563754081726, + "learning_rate": 8.29047694273436e-05, + "loss": 0.6308, + "step": 8738 + }, + { + "epoch": 0.5865574980705346, + "grad_norm": 1.0983659029006958, + "learning_rate": 8.289658552649614e-05, + "loss": 0.6543, + "step": 8740 + }, + { + "epoch": 0.5866917217543035, + "grad_norm": 1.0358927249908447, + "learning_rate": 8.28884000713497e-05, + "loss": 0.6165, + "step": 8742 + }, + { + "epoch": 0.5868259454380725, + "grad_norm": 1.001076102256775, + "learning_rate": 8.288021306229103e-05, + "loss": 0.6404, + "step": 8744 + }, + { + "epoch": 0.5869601691218416, + "grad_norm": 1.0301929712295532, + "learning_rate": 8.287202449970695e-05, + "loss": 0.633, + "step": 8746 + }, + { + "epoch": 0.5870943928056106, + "grad_norm": 1.6269079446792603, + "learning_rate": 8.286383438398437e-05, + "loss": 0.6431, + "step": 8748 + }, + { + "epoch": 0.5872286164893795, + "grad_norm": 1.1193408966064453, + "learning_rate": 8.285564271551023e-05, + "loss": 0.7092, + "step": 8750 + }, + { + "epoch": 0.5873628401731485, + "grad_norm": 1.1632122993469238, + "learning_rate": 8.284744949467156e-05, + "loss": 0.5789, + "step": 8752 + }, + { + "epoch": 0.5874970638569176, + "grad_norm": 1.1115137338638306, + "learning_rate": 8.283925472185552e-05, + "loss": 0.6834, + "step": 8754 + }, + { + "epoch": 0.5876312875406866, + "grad_norm": 1.1421805620193481, + "learning_rate": 8.283105839744925e-05, + "loss": 0.5362, + "step": 8756 + }, + { + "epoch": 0.5877655112244555, + "grad_norm": 1.0075037479400635, + "learning_rate": 8.282286052184005e-05, + "loss": 0.5989, + "step": 8758 + }, + { + "epoch": 0.5878997349082246, + "grad_norm": 1.2430760860443115, + "learning_rate": 8.281466109541521e-05, + "loss": 0.624, + "step": 8760 + }, + { + "epoch": 0.5880339585919936, + "grad_norm": 1.0672948360443115, + "learning_rate": 8.280646011856218e-05, + "loss": 0.6325, + "step": 8762 + }, + { + "epoch": 0.5881681822757625, + "grad_norm": 1.056333065032959, + "learning_rate": 8.279825759166842e-05, + "loss": 0.6255, + "step": 8764 + }, + { + "epoch": 0.5883024059595315, + "grad_norm": 1.1188087463378906, + "learning_rate": 8.279005351512146e-05, + "loss": 0.6541, + "step": 8766 + }, + { + "epoch": 0.5884366296433006, + "grad_norm": 1.011813998222351, + "learning_rate": 8.278184788930897e-05, + "loss": 0.6631, + "step": 8768 + }, + { + "epoch": 0.5885708533270696, + "grad_norm": 1.5705772638320923, + "learning_rate": 8.277364071461862e-05, + "loss": 0.6036, + "step": 8770 + }, + { + "epoch": 0.5887050770108385, + "grad_norm": 0.9317624568939209, + "learning_rate": 8.27654319914382e-05, + "loss": 0.6173, + "step": 8772 + }, + { + "epoch": 0.5888393006946075, + "grad_norm": 1.0767476558685303, + "learning_rate": 8.275722172015555e-05, + "loss": 0.6394, + "step": 8774 + }, + { + "epoch": 0.5889735243783766, + "grad_norm": 1.0390230417251587, + "learning_rate": 8.274900990115859e-05, + "loss": 0.6507, + "step": 8776 + }, + { + "epoch": 0.5891077480621456, + "grad_norm": 1.225508451461792, + "learning_rate": 8.274079653483531e-05, + "loss": 0.6498, + "step": 8778 + }, + { + "epoch": 0.5892419717459145, + "grad_norm": 1.0217746496200562, + "learning_rate": 8.273258162157377e-05, + "loss": 0.6289, + "step": 8780 + }, + { + "epoch": 0.5893761954296836, + "grad_norm": 1.1209293603897095, + "learning_rate": 8.272436516176212e-05, + "loss": 0.6353, + "step": 8782 + }, + { + "epoch": 0.5895104191134526, + "grad_norm": 1.0880781412124634, + "learning_rate": 8.271614715578856e-05, + "loss": 0.6749, + "step": 8784 + }, + { + "epoch": 0.5896446427972216, + "grad_norm": 1.2088608741760254, + "learning_rate": 8.270792760404138e-05, + "loss": 0.6776, + "step": 8786 + }, + { + "epoch": 0.5897788664809905, + "grad_norm": 1.173379898071289, + "learning_rate": 8.269970650690894e-05, + "loss": 0.6267, + "step": 8788 + }, + { + "epoch": 0.5899130901647596, + "grad_norm": 1.0878911018371582, + "learning_rate": 8.269148386477968e-05, + "loss": 0.6153, + "step": 8790 + }, + { + "epoch": 0.5900473138485286, + "grad_norm": 0.9524178504943848, + "learning_rate": 8.268325967804207e-05, + "loss": 0.6521, + "step": 8792 + }, + { + "epoch": 0.5901815375322976, + "grad_norm": 0.9811989665031433, + "learning_rate": 8.267503394708472e-05, + "loss": 0.6251, + "step": 8794 + }, + { + "epoch": 0.5903157612160665, + "grad_norm": 1.040177822113037, + "learning_rate": 8.266680667229627e-05, + "loss": 0.6039, + "step": 8796 + }, + { + "epoch": 0.5904499848998356, + "grad_norm": 1.0749226808547974, + "learning_rate": 8.265857785406544e-05, + "loss": 0.5825, + "step": 8798 + }, + { + "epoch": 0.5905842085836046, + "grad_norm": 1.1496376991271973, + "learning_rate": 8.265034749278103e-05, + "loss": 0.7057, + "step": 8800 + }, + { + "epoch": 0.5907184322673735, + "grad_norm": 1.0202300548553467, + "learning_rate": 8.264211558883191e-05, + "loss": 0.6245, + "step": 8802 + }, + { + "epoch": 0.5908526559511426, + "grad_norm": 1.3150079250335693, + "learning_rate": 8.263388214260702e-05, + "loss": 0.6337, + "step": 8804 + }, + { + "epoch": 0.5909868796349116, + "grad_norm": 0.9684760570526123, + "learning_rate": 8.262564715449535e-05, + "loss": 0.637, + "step": 8806 + }, + { + "epoch": 0.5911211033186806, + "grad_norm": 1.236946940422058, + "learning_rate": 8.261741062488602e-05, + "loss": 0.6555, + "step": 8808 + }, + { + "epoch": 0.5912553270024495, + "grad_norm": 0.9326079487800598, + "learning_rate": 8.260917255416817e-05, + "loss": 0.6187, + "step": 8810 + }, + { + "epoch": 0.5913895506862186, + "grad_norm": 1.0765881538391113, + "learning_rate": 8.260093294273103e-05, + "loss": 0.6183, + "step": 8812 + }, + { + "epoch": 0.5915237743699876, + "grad_norm": 1.1676933765411377, + "learning_rate": 8.259269179096393e-05, + "loss": 0.6559, + "step": 8814 + }, + { + "epoch": 0.5916579980537566, + "grad_norm": 1.091220736503601, + "learning_rate": 8.258444909925624e-05, + "loss": 0.6776, + "step": 8816 + }, + { + "epoch": 0.5917922217375255, + "grad_norm": 1.1847777366638184, + "learning_rate": 8.25762048679974e-05, + "loss": 0.5775, + "step": 8818 + }, + { + "epoch": 0.5919264454212946, + "grad_norm": 1.190104603767395, + "learning_rate": 8.256795909757694e-05, + "loss": 0.6238, + "step": 8820 + }, + { + "epoch": 0.5920606691050636, + "grad_norm": 1.1804845333099365, + "learning_rate": 8.255971178838445e-05, + "loss": 0.6297, + "step": 8822 + }, + { + "epoch": 0.5921948927888326, + "grad_norm": 1.1453303098678589, + "learning_rate": 8.25514629408096e-05, + "loss": 0.6756, + "step": 8824 + }, + { + "epoch": 0.5923291164726016, + "grad_norm": 1.0372350215911865, + "learning_rate": 8.254321255524215e-05, + "loss": 0.6554, + "step": 8826 + }, + { + "epoch": 0.5924633401563706, + "grad_norm": 1.1543527841567993, + "learning_rate": 8.25349606320719e-05, + "loss": 0.5751, + "step": 8828 + }, + { + "epoch": 0.5925975638401396, + "grad_norm": 1.1377294063568115, + "learning_rate": 8.252670717168872e-05, + "loss": 0.6465, + "step": 8830 + }, + { + "epoch": 0.5927317875239086, + "grad_norm": 1.1026825904846191, + "learning_rate": 8.25184521744826e-05, + "loss": 0.5892, + "step": 8832 + }, + { + "epoch": 0.5928660112076776, + "grad_norm": 1.012364387512207, + "learning_rate": 8.251019564084355e-05, + "loss": 0.6018, + "step": 8834 + }, + { + "epoch": 0.5930002348914466, + "grad_norm": 0.9433966875076294, + "learning_rate": 8.250193757116169e-05, + "loss": 0.6373, + "step": 8836 + }, + { + "epoch": 0.5931344585752156, + "grad_norm": 1.0152767896652222, + "learning_rate": 8.24936779658272e-05, + "loss": 0.6141, + "step": 8838 + }, + { + "epoch": 0.5932686822589845, + "grad_norm": 1.0709465742111206, + "learning_rate": 8.248541682523032e-05, + "loss": 0.6747, + "step": 8840 + }, + { + "epoch": 0.5934029059427536, + "grad_norm": 1.8665101528167725, + "learning_rate": 8.247715414976136e-05, + "loss": 0.6451, + "step": 8842 + }, + { + "epoch": 0.5935371296265226, + "grad_norm": 1.182285189628601, + "learning_rate": 8.246888993981074e-05, + "loss": 0.6019, + "step": 8844 + }, + { + "epoch": 0.5936713533102916, + "grad_norm": 0.9702306389808655, + "learning_rate": 8.246062419576892e-05, + "loss": 0.6399, + "step": 8846 + }, + { + "epoch": 0.5938055769940606, + "grad_norm": 1.0066275596618652, + "learning_rate": 8.245235691802644e-05, + "loss": 0.6147, + "step": 8848 + }, + { + "epoch": 0.5939398006778296, + "grad_norm": 1.2227391004562378, + "learning_rate": 8.24440881069739e-05, + "loss": 0.5987, + "step": 8850 + }, + { + "epoch": 0.5940740243615986, + "grad_norm": 0.9837579131126404, + "learning_rate": 8.2435817763002e-05, + "loss": 0.6061, + "step": 8852 + }, + { + "epoch": 0.5942082480453676, + "grad_norm": 1.8223912715911865, + "learning_rate": 8.24275458865015e-05, + "loss": 0.6253, + "step": 8854 + }, + { + "epoch": 0.5943424717291366, + "grad_norm": 0.9962928891181946, + "learning_rate": 8.24192724778632e-05, + "loss": 0.6692, + "step": 8856 + }, + { + "epoch": 0.5944766954129056, + "grad_norm": 0.9767764210700989, + "learning_rate": 8.241099753747805e-05, + "loss": 0.6543, + "step": 8858 + }, + { + "epoch": 0.5946109190966746, + "grad_norm": 1.0351556539535522, + "learning_rate": 8.240272106573699e-05, + "loss": 0.5912, + "step": 8860 + }, + { + "epoch": 0.5947451427804437, + "grad_norm": 1.0033786296844482, + "learning_rate": 8.239444306303109e-05, + "loss": 0.6525, + "step": 8862 + }, + { + "epoch": 0.5948793664642126, + "grad_norm": 0.9843260645866394, + "learning_rate": 8.238616352975143e-05, + "loss": 0.6849, + "step": 8864 + }, + { + "epoch": 0.5950135901479816, + "grad_norm": 1.065222144126892, + "learning_rate": 8.237788246628925e-05, + "loss": 0.5955, + "step": 8866 + }, + { + "epoch": 0.5951478138317506, + "grad_norm": 1.0120952129364014, + "learning_rate": 8.236959987303579e-05, + "loss": 0.6793, + "step": 8868 + }, + { + "epoch": 0.5952820375155197, + "grad_norm": 1.0915205478668213, + "learning_rate": 8.236131575038238e-05, + "loss": 0.6915, + "step": 8870 + }, + { + "epoch": 0.5954162611992886, + "grad_norm": 0.9849734902381897, + "learning_rate": 8.235303009872043e-05, + "loss": 0.6771, + "step": 8872 + }, + { + "epoch": 0.5955504848830576, + "grad_norm": 1.1612014770507812, + "learning_rate": 8.234474291844144e-05, + "loss": 0.6502, + "step": 8874 + }, + { + "epoch": 0.5956847085668266, + "grad_norm": 1.1154402494430542, + "learning_rate": 8.233645420993695e-05, + "loss": 0.6228, + "step": 8876 + }, + { + "epoch": 0.5958189322505956, + "grad_norm": 0.9973312020301819, + "learning_rate": 8.232816397359858e-05, + "loss": 0.6254, + "step": 8878 + }, + { + "epoch": 0.5959531559343646, + "grad_norm": 1.0583056211471558, + "learning_rate": 8.231987220981804e-05, + "loss": 0.621, + "step": 8880 + }, + { + "epoch": 0.5960873796181336, + "grad_norm": 1.1791260242462158, + "learning_rate": 8.231157891898708e-05, + "loss": 0.6521, + "step": 8882 + }, + { + "epoch": 0.5962216033019027, + "grad_norm": 1.0589373111724854, + "learning_rate": 8.230328410149756e-05, + "loss": 0.6304, + "step": 8884 + }, + { + "epoch": 0.5963558269856716, + "grad_norm": 0.9476355910301208, + "learning_rate": 8.22949877577414e-05, + "loss": 0.6518, + "step": 8886 + }, + { + "epoch": 0.5964900506694406, + "grad_norm": 1.0454285144805908, + "learning_rate": 8.228668988811055e-05, + "loss": 0.6237, + "step": 8888 + }, + { + "epoch": 0.5966242743532096, + "grad_norm": 1.1258161067962646, + "learning_rate": 8.227839049299711e-05, + "loss": 0.632, + "step": 8890 + }, + { + "epoch": 0.5967584980369787, + "grad_norm": 0.9451509714126587, + "learning_rate": 8.227008957279319e-05, + "loss": 0.6004, + "step": 8892 + }, + { + "epoch": 0.5968927217207476, + "grad_norm": 1.1254085302352905, + "learning_rate": 8.2261787127891e-05, + "loss": 0.5896, + "step": 8894 + }, + { + "epoch": 0.5970269454045166, + "grad_norm": 0.9593334197998047, + "learning_rate": 8.22534831586828e-05, + "loss": 0.5775, + "step": 8896 + }, + { + "epoch": 0.5971611690882856, + "grad_norm": 1.0605380535125732, + "learning_rate": 8.224517766556094e-05, + "loss": 0.5977, + "step": 8898 + }, + { + "epoch": 0.5972953927720547, + "grad_norm": 1.4234702587127686, + "learning_rate": 8.223687064891785e-05, + "loss": 0.6361, + "step": 8900 + }, + { + "epoch": 0.5974296164558236, + "grad_norm": 1.0671988725662231, + "learning_rate": 8.222856210914601e-05, + "loss": 0.643, + "step": 8902 + }, + { + "epoch": 0.5975638401395926, + "grad_norm": 1.0523698329925537, + "learning_rate": 8.222025204663799e-05, + "loss": 0.6824, + "step": 8904 + }, + { + "epoch": 0.5976980638233617, + "grad_norm": 1.166785717010498, + "learning_rate": 8.221194046178641e-05, + "loss": 0.6208, + "step": 8906 + }, + { + "epoch": 0.5978322875071307, + "grad_norm": 1.129096508026123, + "learning_rate": 8.220362735498399e-05, + "loss": 0.5974, + "step": 8908 + }, + { + "epoch": 0.5979665111908996, + "grad_norm": 1.0599874258041382, + "learning_rate": 8.21953127266235e-05, + "loss": 0.671, + "step": 8910 + }, + { + "epoch": 0.5981007348746686, + "grad_norm": 1.0998214483261108, + "learning_rate": 8.21869965770978e-05, + "loss": 0.638, + "step": 8912 + }, + { + "epoch": 0.5982349585584377, + "grad_norm": 1.0187451839447021, + "learning_rate": 8.21786789067998e-05, + "loss": 0.6521, + "step": 8914 + }, + { + "epoch": 0.5983691822422066, + "grad_norm": 1.2205415964126587, + "learning_rate": 8.21703597161225e-05, + "loss": 0.6373, + "step": 8916 + }, + { + "epoch": 0.5985034059259756, + "grad_norm": 1.3533973693847656, + "learning_rate": 8.216203900545895e-05, + "loss": 0.6429, + "step": 8918 + }, + { + "epoch": 0.5986376296097446, + "grad_norm": 0.7839868068695068, + "learning_rate": 8.215371677520231e-05, + "loss": 0.5764, + "step": 8920 + }, + { + "epoch": 0.5987718532935137, + "grad_norm": 1.0047537088394165, + "learning_rate": 8.21453930257458e-05, + "loss": 0.5767, + "step": 8922 + }, + { + "epoch": 0.5989060769772826, + "grad_norm": 0.9989202618598938, + "learning_rate": 8.213706775748265e-05, + "loss": 0.5818, + "step": 8924 + }, + { + "epoch": 0.5990403006610516, + "grad_norm": 1.103752613067627, + "learning_rate": 8.212874097080626e-05, + "loss": 0.7014, + "step": 8926 + }, + { + "epoch": 0.5991745243448207, + "grad_norm": 1.228872299194336, + "learning_rate": 8.212041266611003e-05, + "loss": 0.6309, + "step": 8928 + }, + { + "epoch": 0.5993087480285897, + "grad_norm": 1.09929621219635, + "learning_rate": 8.211208284378747e-05, + "loss": 0.6905, + "step": 8930 + }, + { + "epoch": 0.5994429717123586, + "grad_norm": 1.056538462638855, + "learning_rate": 8.210375150423214e-05, + "loss": 0.6044, + "step": 8932 + }, + { + "epoch": 0.5995771953961276, + "grad_norm": 0.9706535935401917, + "learning_rate": 8.209541864783769e-05, + "loss": 0.6148, + "step": 8934 + }, + { + "epoch": 0.5997114190798967, + "grad_norm": 1.103705883026123, + "learning_rate": 8.20870842749978e-05, + "loss": 0.675, + "step": 8936 + }, + { + "epoch": 0.5998456427636657, + "grad_norm": 1.0811362266540527, + "learning_rate": 8.20787483861063e-05, + "loss": 0.6167, + "step": 8938 + }, + { + "epoch": 0.5999798664474346, + "grad_norm": 1.8568896055221558, + "learning_rate": 8.2070410981557e-05, + "loss": 0.6104, + "step": 8940 + }, + { + "epoch": 0.6001140901312036, + "grad_norm": 1.0807304382324219, + "learning_rate": 8.206207206174386e-05, + "loss": 0.6114, + "step": 8942 + }, + { + "epoch": 0.6002483138149727, + "grad_norm": 1.0534753799438477, + "learning_rate": 8.205373162706085e-05, + "loss": 0.6841, + "step": 8944 + }, + { + "epoch": 0.6003825374987416, + "grad_norm": 1.3214335441589355, + "learning_rate": 8.204538967790205e-05, + "loss": 0.6906, + "step": 8946 + }, + { + "epoch": 0.6005167611825106, + "grad_norm": 1.0081099271774292, + "learning_rate": 8.20370462146616e-05, + "loss": 0.6297, + "step": 8948 + }, + { + "epoch": 0.6006509848662797, + "grad_norm": 1.1878976821899414, + "learning_rate": 8.202870123773371e-05, + "loss": 0.7054, + "step": 8950 + }, + { + "epoch": 0.6007852085500487, + "grad_norm": 1.0634784698486328, + "learning_rate": 8.20203547475127e-05, + "loss": 0.6073, + "step": 8952 + }, + { + "epoch": 0.6009194322338176, + "grad_norm": 1.013996958732605, + "learning_rate": 8.201200674439287e-05, + "loss": 0.5917, + "step": 8954 + }, + { + "epoch": 0.6010536559175866, + "grad_norm": 1.01338529586792, + "learning_rate": 8.200365722876868e-05, + "loss": 0.6146, + "step": 8956 + }, + { + "epoch": 0.6011878796013557, + "grad_norm": 1.0896825790405273, + "learning_rate": 8.199530620103461e-05, + "loss": 0.5908, + "step": 8958 + }, + { + "epoch": 0.6013221032851247, + "grad_norm": 1.0631287097930908, + "learning_rate": 8.198695366158523e-05, + "loss": 0.5892, + "step": 8960 + }, + { + "epoch": 0.6014563269688936, + "grad_norm": 1.2561473846435547, + "learning_rate": 8.197859961081522e-05, + "loss": 0.6779, + "step": 8962 + }, + { + "epoch": 0.6015905506526626, + "grad_norm": 1.100650429725647, + "learning_rate": 8.197024404911924e-05, + "loss": 0.6193, + "step": 8964 + }, + { + "epoch": 0.6017247743364317, + "grad_norm": 1.1069337129592896, + "learning_rate": 8.196188697689209e-05, + "loss": 0.7133, + "step": 8966 + }, + { + "epoch": 0.6018589980202007, + "grad_norm": 1.0815422534942627, + "learning_rate": 8.195352839452863e-05, + "loss": 0.6056, + "step": 8968 + }, + { + "epoch": 0.6019932217039696, + "grad_norm": 0.9499128460884094, + "learning_rate": 8.19451683024238e-05, + "loss": 0.6629, + "step": 8970 + }, + { + "epoch": 0.6021274453877387, + "grad_norm": 1.1803507804870605, + "learning_rate": 8.193680670097257e-05, + "loss": 0.6733, + "step": 8972 + }, + { + "epoch": 0.6022616690715077, + "grad_norm": 1.0044008493423462, + "learning_rate": 8.192844359057003e-05, + "loss": 0.5965, + "step": 8974 + }, + { + "epoch": 0.6023958927552767, + "grad_norm": 0.9643215537071228, + "learning_rate": 8.192007897161133e-05, + "loss": 0.5732, + "step": 8976 + }, + { + "epoch": 0.6025301164390456, + "grad_norm": 0.9642957448959351, + "learning_rate": 8.191171284449166e-05, + "loss": 0.596, + "step": 8978 + }, + { + "epoch": 0.6026643401228147, + "grad_norm": 1.2209882736206055, + "learning_rate": 8.19033452096063e-05, + "loss": 0.6444, + "step": 8980 + }, + { + "epoch": 0.6027985638065837, + "grad_norm": 0.9373955726623535, + "learning_rate": 8.189497606735061e-05, + "loss": 0.6267, + "step": 8982 + }, + { + "epoch": 0.6029327874903526, + "grad_norm": 1.0083324909210205, + "learning_rate": 8.188660541812004e-05, + "loss": 0.6713, + "step": 8984 + }, + { + "epoch": 0.6030670111741216, + "grad_norm": 1.2278261184692383, + "learning_rate": 8.187823326231005e-05, + "loss": 0.6651, + "step": 8986 + }, + { + "epoch": 0.6032012348578907, + "grad_norm": 1.4669445753097534, + "learning_rate": 8.186985960031623e-05, + "loss": 0.7042, + "step": 8988 + }, + { + "epoch": 0.6033354585416597, + "grad_norm": 1.0924478769302368, + "learning_rate": 8.18614844325342e-05, + "loss": 0.6319, + "step": 8990 + }, + { + "epoch": 0.6034696822254286, + "grad_norm": 0.9888027906417847, + "learning_rate": 8.185310775935971e-05, + "loss": 0.6269, + "step": 8992 + }, + { + "epoch": 0.6036039059091977, + "grad_norm": 0.9969624876976013, + "learning_rate": 8.184472958118851e-05, + "loss": 0.6402, + "step": 8994 + }, + { + "epoch": 0.6037381295929667, + "grad_norm": 1.018033742904663, + "learning_rate": 8.183634989841644e-05, + "loss": 0.6584, + "step": 8996 + }, + { + "epoch": 0.6038723532767357, + "grad_norm": 0.8649704456329346, + "learning_rate": 8.182796871143945e-05, + "loss": 0.5438, + "step": 8998 + }, + { + "epoch": 0.6040065769605046, + "grad_norm": 1.122671365737915, + "learning_rate": 8.181958602065351e-05, + "loss": 0.6088, + "step": 9000 + }, + { + "epoch": 0.6041408006442737, + "grad_norm": 1.2516348361968994, + "learning_rate": 8.181120182645473e-05, + "loss": 0.6303, + "step": 9002 + }, + { + "epoch": 0.6042750243280427, + "grad_norm": 1.1585074663162231, + "learning_rate": 8.18028161292392e-05, + "loss": 0.6331, + "step": 9004 + }, + { + "epoch": 0.6044092480118117, + "grad_norm": 0.9821868538856506, + "learning_rate": 8.179442892940314e-05, + "loss": 0.6229, + "step": 9006 + }, + { + "epoch": 0.6045434716955806, + "grad_norm": 1.0640473365783691, + "learning_rate": 8.178604022734287e-05, + "loss": 0.7202, + "step": 9008 + }, + { + "epoch": 0.6046776953793497, + "grad_norm": 1.0368740558624268, + "learning_rate": 8.177765002345466e-05, + "loss": 0.6664, + "step": 9010 + }, + { + "epoch": 0.6048119190631187, + "grad_norm": 0.9344690442085266, + "learning_rate": 8.176925831813499e-05, + "loss": 0.6318, + "step": 9012 + }, + { + "epoch": 0.6049461427468877, + "grad_norm": 0.9644981026649475, + "learning_rate": 8.176086511178034e-05, + "loss": 0.6323, + "step": 9014 + }, + { + "epoch": 0.6050803664306567, + "grad_norm": 1.0489426851272583, + "learning_rate": 8.175247040478727e-05, + "loss": 0.6223, + "step": 9016 + }, + { + "epoch": 0.6052145901144257, + "grad_norm": 1.0508383512496948, + "learning_rate": 8.17440741975524e-05, + "loss": 0.6097, + "step": 9018 + }, + { + "epoch": 0.6053488137981947, + "grad_norm": 1.0071632862091064, + "learning_rate": 8.173567649047246e-05, + "loss": 0.6127, + "step": 9020 + }, + { + "epoch": 0.6054830374819636, + "grad_norm": 1.4123948812484741, + "learning_rate": 8.17272772839442e-05, + "loss": 0.6371, + "step": 9022 + }, + { + "epoch": 0.6056172611657327, + "grad_norm": 1.0234935283660889, + "learning_rate": 8.171887657836448e-05, + "loss": 0.6502, + "step": 9024 + }, + { + "epoch": 0.6057514848495017, + "grad_norm": 1.1688114404678345, + "learning_rate": 8.171047437413022e-05, + "loss": 0.6422, + "step": 9026 + }, + { + "epoch": 0.6058857085332707, + "grad_norm": 0.9891804456710815, + "learning_rate": 8.17020706716384e-05, + "loss": 0.6012, + "step": 9028 + }, + { + "epoch": 0.6060199322170396, + "grad_norm": 1.1169568300247192, + "learning_rate": 8.169366547128607e-05, + "loss": 0.5939, + "step": 9030 + }, + { + "epoch": 0.6061541559008087, + "grad_norm": 1.1597013473510742, + "learning_rate": 8.16852587734704e-05, + "loss": 0.5878, + "step": 9032 + }, + { + "epoch": 0.6062883795845777, + "grad_norm": 1.082456111907959, + "learning_rate": 8.167685057858853e-05, + "loss": 0.6445, + "step": 9034 + }, + { + "epoch": 0.6064226032683467, + "grad_norm": 1.0745278596878052, + "learning_rate": 8.166844088703777e-05, + "loss": 0.5711, + "step": 9036 + }, + { + "epoch": 0.6065568269521157, + "grad_norm": 1.0846468210220337, + "learning_rate": 8.166002969921545e-05, + "loss": 0.5778, + "step": 9038 + }, + { + "epoch": 0.6066910506358847, + "grad_norm": 1.240317463874817, + "learning_rate": 8.165161701551898e-05, + "loss": 0.6438, + "step": 9040 + }, + { + "epoch": 0.6068252743196537, + "grad_norm": 1.1379884481430054, + "learning_rate": 8.164320283634585e-05, + "loss": 0.6407, + "step": 9042 + }, + { + "epoch": 0.6069594980034227, + "grad_norm": 0.9429364800453186, + "learning_rate": 8.16347871620936e-05, + "loss": 0.6703, + "step": 9044 + }, + { + "epoch": 0.6070937216871917, + "grad_norm": 1.0277690887451172, + "learning_rate": 8.162636999315987e-05, + "loss": 0.6008, + "step": 9046 + }, + { + "epoch": 0.6072279453709607, + "grad_norm": 1.5023620128631592, + "learning_rate": 8.161795132994235e-05, + "loss": 0.605, + "step": 9048 + }, + { + "epoch": 0.6073621690547297, + "grad_norm": 1.0926510095596313, + "learning_rate": 8.160953117283881e-05, + "loss": 0.5932, + "step": 9050 + }, + { + "epoch": 0.6074963927384988, + "grad_norm": 1.0439032316207886, + "learning_rate": 8.160110952224707e-05, + "loss": 0.621, + "step": 9052 + }, + { + "epoch": 0.6076306164222677, + "grad_norm": 0.9344532489776611, + "learning_rate": 8.159268637856506e-05, + "loss": 0.631, + "step": 9054 + }, + { + "epoch": 0.6077648401060367, + "grad_norm": 1.1640667915344238, + "learning_rate": 8.158426174219074e-05, + "loss": 0.6677, + "step": 9056 + }, + { + "epoch": 0.6078990637898057, + "grad_norm": 1.0521429777145386, + "learning_rate": 8.157583561352213e-05, + "loss": 0.631, + "step": 9058 + }, + { + "epoch": 0.6080332874735747, + "grad_norm": 1.0165327787399292, + "learning_rate": 8.156740799295741e-05, + "loss": 0.6065, + "step": 9060 + }, + { + "epoch": 0.6081675111573437, + "grad_norm": 0.997218906879425, + "learning_rate": 8.155897888089473e-05, + "loss": 0.5984, + "step": 9062 + }, + { + "epoch": 0.6083017348411127, + "grad_norm": 1.0150151252746582, + "learning_rate": 8.155054827773237e-05, + "loss": 0.612, + "step": 9064 + }, + { + "epoch": 0.6084359585248817, + "grad_norm": 1.089469313621521, + "learning_rate": 8.154211618386862e-05, + "loss": 0.6098, + "step": 9066 + }, + { + "epoch": 0.6085701822086507, + "grad_norm": 1.1109071969985962, + "learning_rate": 8.153368259970191e-05, + "loss": 0.7168, + "step": 9068 + }, + { + "epoch": 0.6087044058924197, + "grad_norm": 0.9952682852745056, + "learning_rate": 8.15252475256307e-05, + "loss": 0.594, + "step": 9070 + }, + { + "epoch": 0.6088386295761887, + "grad_norm": 1.1361262798309326, + "learning_rate": 8.151681096205356e-05, + "loss": 0.673, + "step": 9072 + }, + { + "epoch": 0.6089728532599578, + "grad_norm": 1.187704086303711, + "learning_rate": 8.150837290936905e-05, + "loss": 0.6023, + "step": 9074 + }, + { + "epoch": 0.6091070769437267, + "grad_norm": 1.0512447357177734, + "learning_rate": 8.14999333679759e-05, + "loss": 0.6197, + "step": 9076 + }, + { + "epoch": 0.6092413006274957, + "grad_norm": 1.0476611852645874, + "learning_rate": 8.149149233827285e-05, + "loss": 0.574, + "step": 9078 + }, + { + "epoch": 0.6093755243112647, + "grad_norm": 1.0389647483825684, + "learning_rate": 8.148304982065869e-05, + "loss": 0.6468, + "step": 9080 + }, + { + "epoch": 0.6095097479950338, + "grad_norm": 1.1799911260604858, + "learning_rate": 8.147460581553233e-05, + "loss": 0.601, + "step": 9082 + }, + { + "epoch": 0.6096439716788027, + "grad_norm": 1.0153534412384033, + "learning_rate": 8.146616032329275e-05, + "loss": 0.6727, + "step": 9084 + }, + { + "epoch": 0.6097781953625717, + "grad_norm": 1.1318472623825073, + "learning_rate": 8.145771334433896e-05, + "loss": 0.6342, + "step": 9086 + }, + { + "epoch": 0.6099124190463407, + "grad_norm": 1.0327775478363037, + "learning_rate": 8.144926487907009e-05, + "loss": 0.6956, + "step": 9088 + }, + { + "epoch": 0.6100466427301098, + "grad_norm": 0.922252357006073, + "learning_rate": 8.144081492788528e-05, + "loss": 0.636, + "step": 9090 + }, + { + "epoch": 0.6101808664138787, + "grad_norm": 1.1907691955566406, + "learning_rate": 8.143236349118381e-05, + "loss": 0.6395, + "step": 9092 + }, + { + "epoch": 0.6103150900976477, + "grad_norm": 1.0901716947555542, + "learning_rate": 8.142391056936495e-05, + "loss": 0.7745, + "step": 9094 + }, + { + "epoch": 0.6104493137814168, + "grad_norm": 0.9876933097839355, + "learning_rate": 8.141545616282812e-05, + "loss": 0.6439, + "step": 9096 + }, + { + "epoch": 0.6105835374651857, + "grad_norm": 0.9889509677886963, + "learning_rate": 8.140700027197277e-05, + "loss": 0.6059, + "step": 9098 + }, + { + "epoch": 0.6107177611489547, + "grad_norm": 1.2002860307693481, + "learning_rate": 8.139854289719841e-05, + "loss": 0.6139, + "step": 9100 + }, + { + "epoch": 0.6108519848327237, + "grad_norm": 1.1429078578948975, + "learning_rate": 8.139008403890465e-05, + "loss": 0.6401, + "step": 9102 + }, + { + "epoch": 0.6109862085164928, + "grad_norm": 1.1443973779678345, + "learning_rate": 8.138162369749114e-05, + "loss": 0.6632, + "step": 9104 + }, + { + "epoch": 0.6111204322002617, + "grad_norm": 1.1087549924850464, + "learning_rate": 8.137316187335761e-05, + "loss": 0.653, + "step": 9106 + }, + { + "epoch": 0.6112546558840307, + "grad_norm": 1.0414327383041382, + "learning_rate": 8.136469856690387e-05, + "loss": 0.6801, + "step": 9108 + }, + { + "epoch": 0.6113888795677997, + "grad_norm": 1.0863112211227417, + "learning_rate": 8.135623377852982e-05, + "loss": 0.6728, + "step": 9110 + }, + { + "epoch": 0.6115231032515688, + "grad_norm": 1.082608938217163, + "learning_rate": 8.134776750863535e-05, + "loss": 0.6256, + "step": 9112 + }, + { + "epoch": 0.6116573269353377, + "grad_norm": 1.0916283130645752, + "learning_rate": 8.133929975762053e-05, + "loss": 0.552, + "step": 9114 + }, + { + "epoch": 0.6117915506191067, + "grad_norm": 1.1448688507080078, + "learning_rate": 8.133083052588543e-05, + "loss": 0.652, + "step": 9116 + }, + { + "epoch": 0.6119257743028758, + "grad_norm": 1.1834670305252075, + "learning_rate": 8.132235981383018e-05, + "loss": 0.6892, + "step": 9118 + }, + { + "epoch": 0.6120599979866448, + "grad_norm": 1.07806396484375, + "learning_rate": 8.131388762185503e-05, + "loss": 0.6864, + "step": 9120 + }, + { + "epoch": 0.6121942216704137, + "grad_norm": 0.9594216346740723, + "learning_rate": 8.130541395036027e-05, + "loss": 0.5829, + "step": 9122 + }, + { + "epoch": 0.6123284453541827, + "grad_norm": 1.0681747198104858, + "learning_rate": 8.129693879974626e-05, + "loss": 0.6485, + "step": 9124 + }, + { + "epoch": 0.6124626690379518, + "grad_norm": 1.1206474304199219, + "learning_rate": 8.128846217041344e-05, + "loss": 0.5984, + "step": 9126 + }, + { + "epoch": 0.6125968927217208, + "grad_norm": 1.0924535989761353, + "learning_rate": 8.12799840627623e-05, + "loss": 0.5855, + "step": 9128 + }, + { + "epoch": 0.6127311164054897, + "grad_norm": 1.2383434772491455, + "learning_rate": 8.127150447719342e-05, + "loss": 0.5934, + "step": 9130 + }, + { + "epoch": 0.6128653400892587, + "grad_norm": 1.0007473230361938, + "learning_rate": 8.126302341410744e-05, + "loss": 0.5637, + "step": 9132 + }, + { + "epoch": 0.6129995637730278, + "grad_norm": 1.0839866399765015, + "learning_rate": 8.125454087390509e-05, + "loss": 0.6521, + "step": 9134 + }, + { + "epoch": 0.6131337874567967, + "grad_norm": 1.111741542816162, + "learning_rate": 8.124605685698714e-05, + "loss": 0.6516, + "step": 9136 + }, + { + "epoch": 0.6132680111405657, + "grad_norm": 0.9247532486915588, + "learning_rate": 8.123757136375445e-05, + "loss": 0.6133, + "step": 9138 + }, + { + "epoch": 0.6134022348243348, + "grad_norm": 0.9069095849990845, + "learning_rate": 8.122908439460794e-05, + "loss": 0.6424, + "step": 9140 + }, + { + "epoch": 0.6135364585081038, + "grad_norm": 1.0073060989379883, + "learning_rate": 8.12205959499486e-05, + "loss": 0.6146, + "step": 9142 + }, + { + "epoch": 0.6136706821918727, + "grad_norm": 1.0860568284988403, + "learning_rate": 8.121210603017748e-05, + "loss": 0.5838, + "step": 9144 + }, + { + "epoch": 0.6138049058756417, + "grad_norm": 1.023048996925354, + "learning_rate": 8.120361463569575e-05, + "loss": 0.6019, + "step": 9146 + }, + { + "epoch": 0.6139391295594108, + "grad_norm": 1.0431309938430786, + "learning_rate": 8.119512176690455e-05, + "loss": 0.671, + "step": 9148 + }, + { + "epoch": 0.6140733532431798, + "grad_norm": 1.0847612619400024, + "learning_rate": 8.118662742420523e-05, + "loss": 0.6467, + "step": 9150 + }, + { + "epoch": 0.6142075769269487, + "grad_norm": 0.9644896388053894, + "learning_rate": 8.117813160799908e-05, + "loss": 0.5775, + "step": 9152 + }, + { + "epoch": 0.6143418006107177, + "grad_norm": 1.1092700958251953, + "learning_rate": 8.116963431868751e-05, + "loss": 0.5999, + "step": 9154 + }, + { + "epoch": 0.6144760242944868, + "grad_norm": 1.053810954093933, + "learning_rate": 8.116113555667203e-05, + "loss": 0.7222, + "step": 9156 + }, + { + "epoch": 0.6146102479782558, + "grad_norm": 0.9161946177482605, + "learning_rate": 8.115263532235416e-05, + "loss": 0.6246, + "step": 9158 + }, + { + "epoch": 0.6147444716620247, + "grad_norm": 1.719303011894226, + "learning_rate": 8.114413361613551e-05, + "loss": 0.5393, + "step": 9160 + }, + { + "epoch": 0.6148786953457938, + "grad_norm": 1.1035606861114502, + "learning_rate": 8.113563043841781e-05, + "loss": 0.5999, + "step": 9162 + }, + { + "epoch": 0.6150129190295628, + "grad_norm": 0.940985918045044, + "learning_rate": 8.11271257896028e-05, + "loss": 0.5872, + "step": 9164 + }, + { + "epoch": 0.6151471427133318, + "grad_norm": 1.0824211835861206, + "learning_rate": 8.111861967009229e-05, + "loss": 0.6177, + "step": 9166 + }, + { + "epoch": 0.6152813663971007, + "grad_norm": 1.059979796409607, + "learning_rate": 8.111011208028821e-05, + "loss": 0.6636, + "step": 9168 + }, + { + "epoch": 0.6154155900808698, + "grad_norm": 1.0575546026229858, + "learning_rate": 8.11016030205925e-05, + "loss": 0.6336, + "step": 9170 + }, + { + "epoch": 0.6155498137646388, + "grad_norm": 1.0367028713226318, + "learning_rate": 8.109309249140721e-05, + "loss": 0.7206, + "step": 9172 + }, + { + "epoch": 0.6156840374484077, + "grad_norm": 1.1719602346420288, + "learning_rate": 8.108458049313443e-05, + "loss": 0.6034, + "step": 9174 + }, + { + "epoch": 0.6158182611321767, + "grad_norm": 1.026654839515686, + "learning_rate": 8.107606702617636e-05, + "loss": 0.6252, + "step": 9176 + }, + { + "epoch": 0.6159524848159458, + "grad_norm": 0.9755574464797974, + "learning_rate": 8.106755209093523e-05, + "loss": 0.6031, + "step": 9178 + }, + { + "epoch": 0.6160867084997148, + "grad_norm": 1.0690412521362305, + "learning_rate": 8.105903568781335e-05, + "loss": 0.6279, + "step": 9180 + }, + { + "epoch": 0.6162209321834837, + "grad_norm": 1.1642770767211914, + "learning_rate": 8.105051781721312e-05, + "loss": 0.6367, + "step": 9182 + }, + { + "epoch": 0.6163551558672528, + "grad_norm": 1.4145997762680054, + "learning_rate": 8.104199847953696e-05, + "loss": 0.65, + "step": 9184 + }, + { + "epoch": 0.6164893795510218, + "grad_norm": 1.0357612371444702, + "learning_rate": 8.103347767518743e-05, + "loss": 0.6423, + "step": 9186 + }, + { + "epoch": 0.6166236032347908, + "grad_norm": 0.9822550415992737, + "learning_rate": 8.102495540456711e-05, + "loss": 0.6013, + "step": 9188 + }, + { + "epoch": 0.6167578269185597, + "grad_norm": 1.0211374759674072, + "learning_rate": 8.101643166807864e-05, + "loss": 0.5946, + "step": 9190 + }, + { + "epoch": 0.6168920506023288, + "grad_norm": 1.137711524963379, + "learning_rate": 8.100790646612477e-05, + "loss": 0.5964, + "step": 9192 + }, + { + "epoch": 0.6170262742860978, + "grad_norm": 1.0474289655685425, + "learning_rate": 8.09993797991083e-05, + "loss": 0.626, + "step": 9194 + }, + { + "epoch": 0.6171604979698668, + "grad_norm": 1.2463290691375732, + "learning_rate": 8.099085166743208e-05, + "loss": 0.6328, + "step": 9196 + }, + { + "epoch": 0.6172947216536357, + "grad_norm": 1.0076327323913574, + "learning_rate": 8.098232207149907e-05, + "loss": 0.6683, + "step": 9198 + }, + { + "epoch": 0.6174289453374048, + "grad_norm": 1.053296446800232, + "learning_rate": 8.097379101171225e-05, + "loss": 0.6389, + "step": 9200 + }, + { + "epoch": 0.6175631690211738, + "grad_norm": 1.063968539237976, + "learning_rate": 8.096525848847473e-05, + "loss": 0.6451, + "step": 9202 + }, + { + "epoch": 0.6176973927049428, + "grad_norm": 0.9891154170036316, + "learning_rate": 8.095672450218964e-05, + "loss": 0.5218, + "step": 9204 + }, + { + "epoch": 0.6178316163887118, + "grad_norm": 1.092598795890808, + "learning_rate": 8.094818905326019e-05, + "loss": 0.6316, + "step": 9206 + }, + { + "epoch": 0.6179658400724808, + "grad_norm": 1.0701557397842407, + "learning_rate": 8.093965214208964e-05, + "loss": 0.6268, + "step": 9208 + }, + { + "epoch": 0.6181000637562498, + "grad_norm": 1.4913824796676636, + "learning_rate": 8.09311137690814e-05, + "loss": 0.6582, + "step": 9210 + }, + { + "epoch": 0.6182342874400187, + "grad_norm": 0.962090790271759, + "learning_rate": 8.092257393463882e-05, + "loss": 0.6144, + "step": 9212 + }, + { + "epoch": 0.6183685111237878, + "grad_norm": 1.2915080785751343, + "learning_rate": 8.091403263916546e-05, + "loss": 0.6584, + "step": 9214 + }, + { + "epoch": 0.6185027348075568, + "grad_norm": 0.9153003096580505, + "learning_rate": 8.090548988306483e-05, + "loss": 0.5892, + "step": 9216 + }, + { + "epoch": 0.6186369584913258, + "grad_norm": 1.1019619703292847, + "learning_rate": 8.08969456667406e-05, + "loss": 0.661, + "step": 9218 + }, + { + "epoch": 0.6187711821750947, + "grad_norm": 1.0655659437179565, + "learning_rate": 8.088839999059642e-05, + "loss": 0.705, + "step": 9220 + }, + { + "epoch": 0.6189054058588638, + "grad_norm": 1.0463535785675049, + "learning_rate": 8.087985285503609e-05, + "loss": 0.5788, + "step": 9222 + }, + { + "epoch": 0.6190396295426328, + "grad_norm": 0.9526081681251526, + "learning_rate": 8.087130426046343e-05, + "loss": 0.5708, + "step": 9224 + }, + { + "epoch": 0.6191738532264018, + "grad_norm": 1.0919690132141113, + "learning_rate": 8.086275420728235e-05, + "loss": 0.604, + "step": 9226 + }, + { + "epoch": 0.6193080769101708, + "grad_norm": 0.8936063647270203, + "learning_rate": 8.085420269589682e-05, + "loss": 0.5557, + "step": 9228 + }, + { + "epoch": 0.6194423005939398, + "grad_norm": 1.1682130098342896, + "learning_rate": 8.08456497267109e-05, + "loss": 0.6856, + "step": 9230 + }, + { + "epoch": 0.6195765242777088, + "grad_norm": 1.5496854782104492, + "learning_rate": 8.083709530012867e-05, + "loss": 0.6227, + "step": 9232 + }, + { + "epoch": 0.6197107479614778, + "grad_norm": 1.0385692119598389, + "learning_rate": 8.082853941655433e-05, + "loss": 0.6537, + "step": 9234 + }, + { + "epoch": 0.6198449716452468, + "grad_norm": 0.9992638826370239, + "learning_rate": 8.081998207639212e-05, + "loss": 0.6415, + "step": 9236 + }, + { + "epoch": 0.6199791953290158, + "grad_norm": 1.149176836013794, + "learning_rate": 8.081142328004637e-05, + "loss": 0.6335, + "step": 9238 + }, + { + "epoch": 0.6201134190127848, + "grad_norm": 1.0221976041793823, + "learning_rate": 8.080286302792146e-05, + "loss": 0.687, + "step": 9240 + }, + { + "epoch": 0.6202476426965539, + "grad_norm": 1.1299258470535278, + "learning_rate": 8.079430132042183e-05, + "loss": 0.5745, + "step": 9242 + }, + { + "epoch": 0.6203818663803228, + "grad_norm": 0.9322912096977234, + "learning_rate": 8.078573815795203e-05, + "loss": 0.6768, + "step": 9244 + }, + { + "epoch": 0.6205160900640918, + "grad_norm": 0.8978147506713867, + "learning_rate": 8.077717354091663e-05, + "loss": 0.5811, + "step": 9246 + }, + { + "epoch": 0.6206503137478608, + "grad_norm": 0.9364485144615173, + "learning_rate": 8.07686074697203e-05, + "loss": 0.5774, + "step": 9248 + }, + { + "epoch": 0.6207845374316298, + "grad_norm": 1.059453010559082, + "learning_rate": 8.076003994476778e-05, + "loss": 0.6521, + "step": 9250 + }, + { + "epoch": 0.6209187611153988, + "grad_norm": 1.4061264991760254, + "learning_rate": 8.075147096646387e-05, + "loss": 0.6372, + "step": 9252 + }, + { + "epoch": 0.6210529847991678, + "grad_norm": 1.0206187963485718, + "learning_rate": 8.074290053521341e-05, + "loss": 0.5841, + "step": 9254 + }, + { + "epoch": 0.6211872084829368, + "grad_norm": 1.3406577110290527, + "learning_rate": 8.073432865142135e-05, + "loss": 0.7033, + "step": 9256 + }, + { + "epoch": 0.6213214321667058, + "grad_norm": 1.1175163984298706, + "learning_rate": 8.07257553154927e-05, + "loss": 0.6678, + "step": 9258 + }, + { + "epoch": 0.6214556558504748, + "grad_norm": 1.0556981563568115, + "learning_rate": 8.071718052783253e-05, + "loss": 0.6165, + "step": 9260 + }, + { + "epoch": 0.6215898795342438, + "grad_norm": 1.007433295249939, + "learning_rate": 8.070860428884599e-05, + "loss": 0.588, + "step": 9262 + }, + { + "epoch": 0.6217241032180129, + "grad_norm": 1.1114963293075562, + "learning_rate": 8.070002659893829e-05, + "loss": 0.5717, + "step": 9264 + }, + { + "epoch": 0.6218583269017818, + "grad_norm": 1.250169277191162, + "learning_rate": 8.069144745851469e-05, + "loss": 0.5943, + "step": 9266 + }, + { + "epoch": 0.6219925505855508, + "grad_norm": 1.0131369829177856, + "learning_rate": 8.068286686798054e-05, + "loss": 0.5999, + "step": 9268 + }, + { + "epoch": 0.6221267742693198, + "grad_norm": 0.9430072903633118, + "learning_rate": 8.067428482774129e-05, + "loss": 0.6898, + "step": 9270 + }, + { + "epoch": 0.6222609979530889, + "grad_norm": 1.0309466123580933, + "learning_rate": 8.06657013382024e-05, + "loss": 0.6147, + "step": 9272 + }, + { + "epoch": 0.6223952216368578, + "grad_norm": 1.0565474033355713, + "learning_rate": 8.065711639976939e-05, + "loss": 0.6038, + "step": 9274 + }, + { + "epoch": 0.6225294453206268, + "grad_norm": 1.2949970960617065, + "learning_rate": 8.064853001284793e-05, + "loss": 0.7469, + "step": 9276 + }, + { + "epoch": 0.6226636690043958, + "grad_norm": 1.0799611806869507, + "learning_rate": 8.063994217784372e-05, + "loss": 0.6407, + "step": 9278 + }, + { + "epoch": 0.6227978926881649, + "grad_norm": 1.3754076957702637, + "learning_rate": 8.063135289516245e-05, + "loss": 0.6355, + "step": 9280 + }, + { + "epoch": 0.6229321163719338, + "grad_norm": 1.0370794534683228, + "learning_rate": 8.062276216521003e-05, + "loss": 0.623, + "step": 9282 + }, + { + "epoch": 0.6230663400557028, + "grad_norm": 1.0482323169708252, + "learning_rate": 8.061416998839231e-05, + "loss": 0.6153, + "step": 9284 + }, + { + "epoch": 0.6232005637394719, + "grad_norm": 0.9800313711166382, + "learning_rate": 8.060557636511523e-05, + "loss": 0.5863, + "step": 9286 + }, + { + "epoch": 0.6233347874232408, + "grad_norm": 1.1380650997161865, + "learning_rate": 8.059698129578486e-05, + "loss": 0.6261, + "step": 9288 + }, + { + "epoch": 0.6234690111070098, + "grad_norm": 1.0158618688583374, + "learning_rate": 8.058838478080731e-05, + "loss": 0.6202, + "step": 9290 + }, + { + "epoch": 0.6236032347907788, + "grad_norm": 1.075456976890564, + "learning_rate": 8.05797868205887e-05, + "loss": 0.6662, + "step": 9292 + }, + { + "epoch": 0.6237374584745479, + "grad_norm": 0.8624019622802734, + "learning_rate": 8.057118741553533e-05, + "loss": 0.5554, + "step": 9294 + }, + { + "epoch": 0.6238716821583168, + "grad_norm": 1.1782561540603638, + "learning_rate": 8.056258656605344e-05, + "loss": 0.6102, + "step": 9296 + }, + { + "epoch": 0.6240059058420858, + "grad_norm": 1.1685196161270142, + "learning_rate": 8.055398427254945e-05, + "loss": 0.6215, + "step": 9298 + }, + { + "epoch": 0.6241401295258548, + "grad_norm": 1.2361077070236206, + "learning_rate": 8.054538053542978e-05, + "loss": 0.6379, + "step": 9300 + }, + { + "epoch": 0.6242743532096239, + "grad_norm": 1.1202350854873657, + "learning_rate": 8.053677535510094e-05, + "loss": 0.6236, + "step": 9302 + }, + { + "epoch": 0.6244085768933928, + "grad_norm": 0.9103413820266724, + "learning_rate": 8.052816873196952e-05, + "loss": 0.584, + "step": 9304 + }, + { + "epoch": 0.6245428005771618, + "grad_norm": 1.1568377017974854, + "learning_rate": 8.051956066644217e-05, + "loss": 0.6205, + "step": 9306 + }, + { + "epoch": 0.6246770242609309, + "grad_norm": 0.978415846824646, + "learning_rate": 8.051095115892557e-05, + "loss": 0.675, + "step": 9308 + }, + { + "epoch": 0.6248112479446999, + "grad_norm": 0.939058244228363, + "learning_rate": 8.050234020982653e-05, + "loss": 0.6457, + "step": 9310 + }, + { + "epoch": 0.6249454716284688, + "grad_norm": 0.9567255973815918, + "learning_rate": 8.04937278195519e-05, + "loss": 0.6117, + "step": 9312 + }, + { + "epoch": 0.6250796953122378, + "grad_norm": 0.9711152911186218, + "learning_rate": 8.04851139885086e-05, + "loss": 0.5895, + "step": 9314 + }, + { + "epoch": 0.6252139189960069, + "grad_norm": 0.9670515060424805, + "learning_rate": 8.047649871710359e-05, + "loss": 0.6798, + "step": 9316 + }, + { + "epoch": 0.6253481426797759, + "grad_norm": 1.1726408004760742, + "learning_rate": 8.046788200574395e-05, + "loss": 0.6938, + "step": 9318 + }, + { + "epoch": 0.6254823663635448, + "grad_norm": 0.9561095237731934, + "learning_rate": 8.045926385483682e-05, + "loss": 0.6577, + "step": 9320 + }, + { + "epoch": 0.6256165900473138, + "grad_norm": 0.926373302936554, + "learning_rate": 8.045064426478935e-05, + "loss": 0.6572, + "step": 9322 + }, + { + "epoch": 0.6257508137310829, + "grad_norm": 1.00580894947052, + "learning_rate": 8.044202323600882e-05, + "loss": 0.5591, + "step": 9324 + }, + { + "epoch": 0.6258850374148518, + "grad_norm": 1.028042197227478, + "learning_rate": 8.043340076890256e-05, + "loss": 0.6075, + "step": 9326 + }, + { + "epoch": 0.6260192610986208, + "grad_norm": 1.2293074131011963, + "learning_rate": 8.042477686387794e-05, + "loss": 0.5969, + "step": 9328 + }, + { + "epoch": 0.6261534847823899, + "grad_norm": 0.9517163038253784, + "learning_rate": 8.041615152134247e-05, + "loss": 0.605, + "step": 9330 + }, + { + "epoch": 0.6262877084661589, + "grad_norm": 1.2889013290405273, + "learning_rate": 8.040752474170364e-05, + "loss": 0.5905, + "step": 9332 + }, + { + "epoch": 0.6264219321499278, + "grad_norm": 0.9943846464157104, + "learning_rate": 8.039889652536905e-05, + "loss": 0.5722, + "step": 9334 + }, + { + "epoch": 0.6265561558336968, + "grad_norm": 1.0777925252914429, + "learning_rate": 8.039026687274638e-05, + "loss": 0.6051, + "step": 9336 + }, + { + "epoch": 0.6266903795174659, + "grad_norm": 2.0752015113830566, + "learning_rate": 8.038163578424336e-05, + "loss": 0.6598, + "step": 9338 + }, + { + "epoch": 0.6268246032012349, + "grad_norm": 1.0468372106552124, + "learning_rate": 8.037300326026779e-05, + "loss": 0.6887, + "step": 9340 + }, + { + "epoch": 0.6269588268850038, + "grad_norm": 1.1117974519729614, + "learning_rate": 8.036436930122754e-05, + "loss": 0.6429, + "step": 9342 + }, + { + "epoch": 0.6270930505687728, + "grad_norm": 1.2182691097259521, + "learning_rate": 8.035573390753054e-05, + "loss": 0.6092, + "step": 9344 + }, + { + "epoch": 0.6272272742525419, + "grad_norm": 1.0793417692184448, + "learning_rate": 8.034709707958483e-05, + "loss": 0.7057, + "step": 9346 + }, + { + "epoch": 0.6273614979363109, + "grad_norm": 1.2882823944091797, + "learning_rate": 8.033845881779845e-05, + "loss": 0.6182, + "step": 9348 + }, + { + "epoch": 0.6274957216200798, + "grad_norm": 1.1499978303909302, + "learning_rate": 8.032981912257955e-05, + "loss": 0.6296, + "step": 9350 + }, + { + "epoch": 0.6276299453038489, + "grad_norm": 0.9525328278541565, + "learning_rate": 8.032117799433634e-05, + "loss": 0.594, + "step": 9352 + }, + { + "epoch": 0.6277641689876179, + "grad_norm": 1.1425232887268066, + "learning_rate": 8.031253543347708e-05, + "loss": 0.7363, + "step": 9354 + }, + { + "epoch": 0.6278983926713869, + "grad_norm": 0.9608234167098999, + "learning_rate": 8.030389144041014e-05, + "loss": 0.5717, + "step": 9356 + }, + { + "epoch": 0.6280326163551558, + "grad_norm": 1.0381462574005127, + "learning_rate": 8.029524601554392e-05, + "loss": 0.588, + "step": 9358 + }, + { + "epoch": 0.6281668400389249, + "grad_norm": 1.111899495124817, + "learning_rate": 8.028659915928689e-05, + "loss": 0.6468, + "step": 9360 + }, + { + "epoch": 0.6283010637226939, + "grad_norm": 0.9303768277168274, + "learning_rate": 8.027795087204761e-05, + "loss": 0.599, + "step": 9362 + }, + { + "epoch": 0.6284352874064628, + "grad_norm": 0.9777799844741821, + "learning_rate": 8.026930115423469e-05, + "loss": 0.6209, + "step": 9364 + }, + { + "epoch": 0.6285695110902318, + "grad_norm": 1.0423401594161987, + "learning_rate": 8.026065000625681e-05, + "loss": 0.5481, + "step": 9366 + }, + { + "epoch": 0.6287037347740009, + "grad_norm": 1.0492119789123535, + "learning_rate": 8.025199742852272e-05, + "loss": 0.6571, + "step": 9368 + }, + { + "epoch": 0.6288379584577699, + "grad_norm": 0.9871809482574463, + "learning_rate": 8.024334342144124e-05, + "loss": 0.6433, + "step": 9370 + }, + { + "epoch": 0.6289721821415388, + "grad_norm": 0.9131199717521667, + "learning_rate": 8.023468798542127e-05, + "loss": 0.5715, + "step": 9372 + }, + { + "epoch": 0.6291064058253079, + "grad_norm": 0.976167619228363, + "learning_rate": 8.022603112087174e-05, + "loss": 0.5816, + "step": 9374 + }, + { + "epoch": 0.6292406295090769, + "grad_norm": 1.010567545890808, + "learning_rate": 8.021737282820167e-05, + "loss": 0.6013, + "step": 9376 + }, + { + "epoch": 0.6293748531928459, + "grad_norm": 1.1628448963165283, + "learning_rate": 8.020871310782015e-05, + "loss": 0.667, + "step": 9378 + }, + { + "epoch": 0.6295090768766148, + "grad_norm": 1.045589804649353, + "learning_rate": 8.020005196013636e-05, + "loss": 0.6611, + "step": 9380 + }, + { + "epoch": 0.6296433005603839, + "grad_norm": 0.9522994160652161, + "learning_rate": 8.019138938555947e-05, + "loss": 0.6149, + "step": 9382 + }, + { + "epoch": 0.6297775242441529, + "grad_norm": 2.1528701782226562, + "learning_rate": 8.018272538449882e-05, + "loss": 0.616, + "step": 9384 + }, + { + "epoch": 0.6299117479279219, + "grad_norm": 1.1953459978103638, + "learning_rate": 8.017405995736374e-05, + "loss": 0.6366, + "step": 9386 + }, + { + "epoch": 0.6300459716116908, + "grad_norm": 1.0215733051300049, + "learning_rate": 8.016539310456367e-05, + "loss": 0.5684, + "step": 9388 + }, + { + "epoch": 0.6301801952954599, + "grad_norm": 0.9657465815544128, + "learning_rate": 8.01567248265081e-05, + "loss": 0.6225, + "step": 9390 + }, + { + "epoch": 0.6303144189792289, + "grad_norm": 0.8911698460578918, + "learning_rate": 8.014805512360655e-05, + "loss": 0.5509, + "step": 9392 + }, + { + "epoch": 0.6304486426629979, + "grad_norm": 1.2820117473602295, + "learning_rate": 8.01393839962687e-05, + "loss": 0.6528, + "step": 9394 + }, + { + "epoch": 0.6305828663467669, + "grad_norm": 1.0842069387435913, + "learning_rate": 8.013071144490423e-05, + "loss": 0.5732, + "step": 9396 + }, + { + "epoch": 0.6307170900305359, + "grad_norm": 1.0435683727264404, + "learning_rate": 8.012203746992288e-05, + "loss": 0.6356, + "step": 9398 + }, + { + "epoch": 0.6308513137143049, + "grad_norm": 1.0976462364196777, + "learning_rate": 8.01133620717345e-05, + "loss": 0.5647, + "step": 9400 + }, + { + "epoch": 0.6309855373980738, + "grad_norm": 1.0342777967453003, + "learning_rate": 8.010468525074897e-05, + "loss": 0.6827, + "step": 9402 + }, + { + "epoch": 0.6311197610818429, + "grad_norm": 1.0499624013900757, + "learning_rate": 8.009600700737627e-05, + "loss": 0.5681, + "step": 9404 + }, + { + "epoch": 0.6312539847656119, + "grad_norm": 1.211155652999878, + "learning_rate": 8.008732734202642e-05, + "loss": 0.6753, + "step": 9406 + }, + { + "epoch": 0.6313882084493809, + "grad_norm": 1.012489676475525, + "learning_rate": 8.007864625510951e-05, + "loss": 0.6366, + "step": 9408 + }, + { + "epoch": 0.6315224321331498, + "grad_norm": 1.0493266582489014, + "learning_rate": 8.006996374703572e-05, + "loss": 0.5444, + "step": 9410 + }, + { + "epoch": 0.6316566558169189, + "grad_norm": 1.1480631828308105, + "learning_rate": 8.006127981821526e-05, + "loss": 0.5618, + "step": 9412 + }, + { + "epoch": 0.6317908795006879, + "grad_norm": 1.0542213916778564, + "learning_rate": 8.005259446905845e-05, + "loss": 0.5866, + "step": 9414 + }, + { + "epoch": 0.6319251031844569, + "grad_norm": 1.1040873527526855, + "learning_rate": 8.004390769997565e-05, + "loss": 0.742, + "step": 9416 + }, + { + "epoch": 0.6320593268682259, + "grad_norm": 1.099540114402771, + "learning_rate": 8.003521951137728e-05, + "loss": 0.6227, + "step": 9418 + }, + { + "epoch": 0.6321935505519949, + "grad_norm": 1.1449235677719116, + "learning_rate": 8.002652990367385e-05, + "loss": 0.6253, + "step": 9420 + }, + { + "epoch": 0.6323277742357639, + "grad_norm": 1.0806974172592163, + "learning_rate": 8.001783887727594e-05, + "loss": 0.6733, + "step": 9422 + }, + { + "epoch": 0.632461997919533, + "grad_norm": 1.1577404737472534, + "learning_rate": 8.000914643259416e-05, + "loss": 0.6744, + "step": 9424 + }, + { + "epoch": 0.6325962216033019, + "grad_norm": 1.5830479860305786, + "learning_rate": 8.000045257003923e-05, + "loss": 0.6348, + "step": 9426 + }, + { + "epoch": 0.6327304452870709, + "grad_norm": 1.038620948791504, + "learning_rate": 7.999175729002191e-05, + "loss": 0.6083, + "step": 9428 + }, + { + "epoch": 0.6328646689708399, + "grad_norm": 0.9739518165588379, + "learning_rate": 7.998306059295301e-05, + "loss": 0.6105, + "step": 9430 + }, + { + "epoch": 0.632998892654609, + "grad_norm": 0.9545291662216187, + "learning_rate": 7.997436247924348e-05, + "loss": 0.5806, + "step": 9432 + }, + { + "epoch": 0.6331331163383779, + "grad_norm": 0.987770676612854, + "learning_rate": 7.996566294930428e-05, + "loss": 0.6121, + "step": 9434 + }, + { + "epoch": 0.6332673400221469, + "grad_norm": 1.0466232299804688, + "learning_rate": 7.99569620035464e-05, + "loss": 0.6446, + "step": 9436 + }, + { + "epoch": 0.6334015637059159, + "grad_norm": 0.9641134142875671, + "learning_rate": 7.994825964238099e-05, + "loss": 0.7188, + "step": 9438 + }, + { + "epoch": 0.6335357873896849, + "grad_norm": 1.0057556629180908, + "learning_rate": 7.993955586621918e-05, + "loss": 0.6589, + "step": 9440 + }, + { + "epoch": 0.6336700110734539, + "grad_norm": 1.01612389087677, + "learning_rate": 7.993085067547226e-05, + "loss": 0.6157, + "step": 9442 + }, + { + "epoch": 0.6338042347572229, + "grad_norm": 1.1171088218688965, + "learning_rate": 7.992214407055148e-05, + "loss": 0.6277, + "step": 9444 + }, + { + "epoch": 0.633938458440992, + "grad_norm": 1.028316617012024, + "learning_rate": 7.991343605186826e-05, + "loss": 0.6152, + "step": 9446 + }, + { + "epoch": 0.6340726821247609, + "grad_norm": 0.9967698454856873, + "learning_rate": 7.990472661983398e-05, + "loss": 0.6577, + "step": 9448 + }, + { + "epoch": 0.6342069058085299, + "grad_norm": 1.0741626024246216, + "learning_rate": 7.989601577486017e-05, + "loss": 0.5831, + "step": 9450 + }, + { + "epoch": 0.6343411294922989, + "grad_norm": 1.047202706336975, + "learning_rate": 7.988730351735843e-05, + "loss": 0.6453, + "step": 9452 + }, + { + "epoch": 0.634475353176068, + "grad_norm": 0.9847677946090698, + "learning_rate": 7.987858984774035e-05, + "loss": 0.6171, + "step": 9454 + }, + { + "epoch": 0.6346095768598369, + "grad_norm": 1.1441807746887207, + "learning_rate": 7.986987476641764e-05, + "loss": 0.752, + "step": 9456 + }, + { + "epoch": 0.6347438005436059, + "grad_norm": 1.0286108255386353, + "learning_rate": 7.98611582738021e-05, + "loss": 0.6004, + "step": 9458 + }, + { + "epoch": 0.6348780242273749, + "grad_norm": 1.0483300685882568, + "learning_rate": 7.985244037030556e-05, + "loss": 0.5558, + "step": 9460 + }, + { + "epoch": 0.635012247911144, + "grad_norm": 0.9127335548400879, + "learning_rate": 7.984372105633991e-05, + "loss": 0.6242, + "step": 9462 + }, + { + "epoch": 0.6351464715949129, + "grad_norm": 0.9870041012763977, + "learning_rate": 7.983500033231711e-05, + "loss": 0.5861, + "step": 9464 + }, + { + "epoch": 0.6352806952786819, + "grad_norm": 1.08979070186615, + "learning_rate": 7.982627819864923e-05, + "loss": 0.6495, + "step": 9466 + }, + { + "epoch": 0.635414918962451, + "grad_norm": 1.0896207094192505, + "learning_rate": 7.981755465574834e-05, + "loss": 0.5842, + "step": 9468 + }, + { + "epoch": 0.63554914264622, + "grad_norm": 0.995100736618042, + "learning_rate": 7.980882970402664e-05, + "loss": 0.5988, + "step": 9470 + }, + { + "epoch": 0.6356833663299889, + "grad_norm": 1.0793224573135376, + "learning_rate": 7.980010334389636e-05, + "loss": 0.646, + "step": 9472 + }, + { + "epoch": 0.6358175900137579, + "grad_norm": 1.2096102237701416, + "learning_rate": 7.979137557576978e-05, + "loss": 0.6993, + "step": 9474 + }, + { + "epoch": 0.635951813697527, + "grad_norm": 0.9157567024230957, + "learning_rate": 7.978264640005928e-05, + "loss": 0.5822, + "step": 9476 + }, + { + "epoch": 0.6360860373812959, + "grad_norm": 1.0237129926681519, + "learning_rate": 7.977391581717733e-05, + "loss": 0.5526, + "step": 9478 + }, + { + "epoch": 0.6362202610650649, + "grad_norm": 1.0919779539108276, + "learning_rate": 7.976518382753637e-05, + "loss": 0.5939, + "step": 9480 + }, + { + "epoch": 0.6363544847488339, + "grad_norm": 0.9949460029602051, + "learning_rate": 7.975645043154903e-05, + "loss": 0.6182, + "step": 9482 + }, + { + "epoch": 0.636488708432603, + "grad_norm": 1.152925968170166, + "learning_rate": 7.974771562962791e-05, + "loss": 0.6723, + "step": 9484 + }, + { + "epoch": 0.6366229321163719, + "grad_norm": 0.9883755445480347, + "learning_rate": 7.973897942218573e-05, + "loss": 0.6788, + "step": 9486 + }, + { + "epoch": 0.6367571558001409, + "grad_norm": 0.9843109846115112, + "learning_rate": 7.973024180963526e-05, + "loss": 0.5981, + "step": 9488 + }, + { + "epoch": 0.63689137948391, + "grad_norm": 0.9135860204696655, + "learning_rate": 7.97215027923893e-05, + "loss": 0.6213, + "step": 9490 + }, + { + "epoch": 0.637025603167679, + "grad_norm": 0.9985994696617126, + "learning_rate": 7.97127623708608e-05, + "loss": 0.5922, + "step": 9492 + }, + { + "epoch": 0.6371598268514479, + "grad_norm": 1.1184334754943848, + "learning_rate": 7.97040205454627e-05, + "loss": 0.6606, + "step": 9494 + }, + { + "epoch": 0.6372940505352169, + "grad_norm": 0.9954715371131897, + "learning_rate": 7.969527731660805e-05, + "loss": 0.6347, + "step": 9496 + }, + { + "epoch": 0.637428274218986, + "grad_norm": 1.0252074003219604, + "learning_rate": 7.968653268470992e-05, + "loss": 0.6833, + "step": 9498 + }, + { + "epoch": 0.637562497902755, + "grad_norm": 1.1799784898757935, + "learning_rate": 7.967778665018151e-05, + "loss": 0.5944, + "step": 9500 + }, + { + "epoch": 0.6376967215865239, + "grad_norm": 0.8477250337600708, + "learning_rate": 7.966903921343603e-05, + "loss": 0.588, + "step": 9502 + }, + { + "epoch": 0.6378309452702929, + "grad_norm": 1.1228663921356201, + "learning_rate": 7.966029037488681e-05, + "loss": 0.6696, + "step": 9504 + }, + { + "epoch": 0.637965168954062, + "grad_norm": 0.9697222709655762, + "learning_rate": 7.965154013494717e-05, + "loss": 0.629, + "step": 9506 + }, + { + "epoch": 0.638099392637831, + "grad_norm": 1.421544075012207, + "learning_rate": 7.964278849403057e-05, + "loss": 0.6485, + "step": 9508 + }, + { + "epoch": 0.6382336163215999, + "grad_norm": 1.0229979753494263, + "learning_rate": 7.96340354525505e-05, + "loss": 0.585, + "step": 9510 + }, + { + "epoch": 0.638367840005369, + "grad_norm": 1.0671172142028809, + "learning_rate": 7.962528101092054e-05, + "loss": 0.6059, + "step": 9512 + }, + { + "epoch": 0.638502063689138, + "grad_norm": 0.9892696738243103, + "learning_rate": 7.96165251695543e-05, + "loss": 0.6253, + "step": 9514 + }, + { + "epoch": 0.6386362873729069, + "grad_norm": 1.5110344886779785, + "learning_rate": 7.96077679288655e-05, + "loss": 0.5844, + "step": 9516 + }, + { + "epoch": 0.6387705110566759, + "grad_norm": 1.0938777923583984, + "learning_rate": 7.959900928926788e-05, + "loss": 0.6188, + "step": 9518 + }, + { + "epoch": 0.638904734740445, + "grad_norm": 1.3613741397857666, + "learning_rate": 7.959024925117527e-05, + "loss": 0.6533, + "step": 9520 + }, + { + "epoch": 0.639038958424214, + "grad_norm": 1.0754303932189941, + "learning_rate": 7.958148781500156e-05, + "loss": 0.7255, + "step": 9522 + }, + { + "epoch": 0.6391731821079829, + "grad_norm": 0.9454665184020996, + "learning_rate": 7.957272498116073e-05, + "loss": 0.6007, + "step": 9524 + }, + { + "epoch": 0.6393074057917519, + "grad_norm": 0.860373318195343, + "learning_rate": 7.956396075006681e-05, + "loss": 0.646, + "step": 9526 + }, + { + "epoch": 0.639441629475521, + "grad_norm": 0.944023847579956, + "learning_rate": 7.955519512213386e-05, + "loss": 0.6317, + "step": 9528 + }, + { + "epoch": 0.63957585315929, + "grad_norm": 1.1715887784957886, + "learning_rate": 7.954642809777606e-05, + "loss": 0.6406, + "step": 9530 + }, + { + "epoch": 0.6397100768430589, + "grad_norm": 1.045246958732605, + "learning_rate": 7.953765967740764e-05, + "loss": 0.5636, + "step": 9532 + }, + { + "epoch": 0.639844300526828, + "grad_norm": 1.015788197517395, + "learning_rate": 7.952888986144288e-05, + "loss": 0.6354, + "step": 9534 + }, + { + "epoch": 0.639978524210597, + "grad_norm": 0.9352192878723145, + "learning_rate": 7.952011865029614e-05, + "loss": 0.6232, + "step": 9536 + }, + { + "epoch": 0.640112747894366, + "grad_norm": 1.4465974569320679, + "learning_rate": 7.951134604438183e-05, + "loss": 0.585, + "step": 9538 + }, + { + "epoch": 0.6402469715781349, + "grad_norm": 1.1227067708969116, + "learning_rate": 7.950257204411448e-05, + "loss": 0.6315, + "step": 9540 + }, + { + "epoch": 0.640381195261904, + "grad_norm": 0.9985955357551575, + "learning_rate": 7.949379664990859e-05, + "loss": 0.5605, + "step": 9542 + }, + { + "epoch": 0.640515418945673, + "grad_norm": 0.9917994141578674, + "learning_rate": 7.948501986217883e-05, + "loss": 0.5681, + "step": 9544 + }, + { + "epoch": 0.640649642629442, + "grad_norm": 1.1306357383728027, + "learning_rate": 7.947624168133984e-05, + "loss": 0.695, + "step": 9546 + }, + { + "epoch": 0.6407838663132109, + "grad_norm": 1.1166436672210693, + "learning_rate": 7.94674621078064e-05, + "loss": 0.623, + "step": 9548 + }, + { + "epoch": 0.64091808999698, + "grad_norm": 1.310808539390564, + "learning_rate": 7.945868114199332e-05, + "loss": 0.6501, + "step": 9550 + }, + { + "epoch": 0.641052313680749, + "grad_norm": 1.0382336378097534, + "learning_rate": 7.944989878431548e-05, + "loss": 0.5603, + "step": 9552 + }, + { + "epoch": 0.6411865373645179, + "grad_norm": 1.0245572328567505, + "learning_rate": 7.944111503518782e-05, + "loss": 0.552, + "step": 9554 + }, + { + "epoch": 0.641320761048287, + "grad_norm": 1.0605592727661133, + "learning_rate": 7.943232989502541e-05, + "loss": 0.616, + "step": 9556 + }, + { + "epoch": 0.641454984732056, + "grad_norm": 1.1174709796905518, + "learning_rate": 7.942354336424325e-05, + "loss": 0.6506, + "step": 9558 + }, + { + "epoch": 0.641589208415825, + "grad_norm": 1.1075295209884644, + "learning_rate": 7.941475544325654e-05, + "loss": 0.5926, + "step": 9560 + }, + { + "epoch": 0.6417234320995939, + "grad_norm": 1.1021950244903564, + "learning_rate": 7.940596613248048e-05, + "loss": 0.6594, + "step": 9562 + }, + { + "epoch": 0.641857655783363, + "grad_norm": 1.1112298965454102, + "learning_rate": 7.939717543233034e-05, + "loss": 0.6499, + "step": 9564 + }, + { + "epoch": 0.641991879467132, + "grad_norm": 0.9838439226150513, + "learning_rate": 7.938838334322147e-05, + "loss": 0.6146, + "step": 9566 + }, + { + "epoch": 0.642126103150901, + "grad_norm": 1.048783779144287, + "learning_rate": 7.937958986556927e-05, + "loss": 0.6297, + "step": 9568 + }, + { + "epoch": 0.6422603268346699, + "grad_norm": 1.0209814310073853, + "learning_rate": 7.937079499978923e-05, + "loss": 0.5854, + "step": 9570 + }, + { + "epoch": 0.642394550518439, + "grad_norm": 0.9504020810127258, + "learning_rate": 7.936199874629689e-05, + "loss": 0.645, + "step": 9572 + }, + { + "epoch": 0.642528774202208, + "grad_norm": 0.9745621085166931, + "learning_rate": 7.935320110550783e-05, + "loss": 0.5738, + "step": 9574 + }, + { + "epoch": 0.642662997885977, + "grad_norm": 1.0484421253204346, + "learning_rate": 7.934440207783773e-05, + "loss": 0.6287, + "step": 9576 + }, + { + "epoch": 0.642797221569746, + "grad_norm": 0.9482985138893127, + "learning_rate": 7.933560166370235e-05, + "loss": 0.5943, + "step": 9578 + }, + { + "epoch": 0.642931445253515, + "grad_norm": 1.031992793083191, + "learning_rate": 7.932679986351748e-05, + "loss": 0.6229, + "step": 9580 + }, + { + "epoch": 0.643065668937284, + "grad_norm": 1.068016767501831, + "learning_rate": 7.931799667769899e-05, + "loss": 0.6183, + "step": 9582 + }, + { + "epoch": 0.643199892621053, + "grad_norm": 0.9357966780662537, + "learning_rate": 7.93091921066628e-05, + "loss": 0.635, + "step": 9584 + }, + { + "epoch": 0.643334116304822, + "grad_norm": 0.9679009914398193, + "learning_rate": 7.930038615082491e-05, + "loss": 0.5797, + "step": 9586 + }, + { + "epoch": 0.643468339988591, + "grad_norm": 1.1679974794387817, + "learning_rate": 7.92915788106014e-05, + "loss": 0.6083, + "step": 9588 + }, + { + "epoch": 0.64360256367236, + "grad_norm": 1.0740517377853394, + "learning_rate": 7.928277008640838e-05, + "loss": 0.562, + "step": 9590 + }, + { + "epoch": 0.6437367873561289, + "grad_norm": 1.0679657459259033, + "learning_rate": 7.927395997866207e-05, + "loss": 0.5675, + "step": 9592 + }, + { + "epoch": 0.643871011039898, + "grad_norm": 1.0845129489898682, + "learning_rate": 7.926514848777871e-05, + "loss": 0.6104, + "step": 9594 + }, + { + "epoch": 0.644005234723667, + "grad_norm": 1.0254936218261719, + "learning_rate": 7.925633561417462e-05, + "loss": 0.6037, + "step": 9596 + }, + { + "epoch": 0.644139458407436, + "grad_norm": 1.9534566402435303, + "learning_rate": 7.924752135826623e-05, + "loss": 0.6097, + "step": 9598 + }, + { + "epoch": 0.644273682091205, + "grad_norm": 1.1238148212432861, + "learning_rate": 7.923870572046994e-05, + "loss": 0.5883, + "step": 9600 + }, + { + "epoch": 0.644407905774974, + "grad_norm": 1.180912733078003, + "learning_rate": 7.922988870120232e-05, + "loss": 0.6708, + "step": 9602 + }, + { + "epoch": 0.644542129458743, + "grad_norm": 1.112041711807251, + "learning_rate": 7.922107030087992e-05, + "loss": 0.5837, + "step": 9604 + }, + { + "epoch": 0.644676353142512, + "grad_norm": 1.07122802734375, + "learning_rate": 7.921225051991942e-05, + "loss": 0.616, + "step": 9606 + }, + { + "epoch": 0.644810576826281, + "grad_norm": 1.001197338104248, + "learning_rate": 7.920342935873752e-05, + "loss": 0.6973, + "step": 9608 + }, + { + "epoch": 0.64494480051005, + "grad_norm": 1.050616979598999, + "learning_rate": 7.919460681775101e-05, + "loss": 0.5911, + "step": 9610 + }, + { + "epoch": 0.645079024193819, + "grad_norm": 0.9228583574295044, + "learning_rate": 7.918578289737675e-05, + "loss": 0.6037, + "step": 9612 + }, + { + "epoch": 0.645213247877588, + "grad_norm": 1.0434296131134033, + "learning_rate": 7.917695759803163e-05, + "loss": 0.6193, + "step": 9614 + }, + { + "epoch": 0.645347471561357, + "grad_norm": 0.98051518201828, + "learning_rate": 7.916813092013264e-05, + "loss": 0.5831, + "step": 9616 + }, + { + "epoch": 0.645481695245126, + "grad_norm": 0.9835851192474365, + "learning_rate": 7.91593028640968e-05, + "loss": 0.6135, + "step": 9618 + }, + { + "epoch": 0.645615918928895, + "grad_norm": 1.0886749029159546, + "learning_rate": 7.915047343034128e-05, + "loss": 0.625, + "step": 9620 + }, + { + "epoch": 0.6457501426126641, + "grad_norm": 1.1845823526382446, + "learning_rate": 7.914164261928318e-05, + "loss": 0.5843, + "step": 9622 + }, + { + "epoch": 0.645884366296433, + "grad_norm": 0.809241533279419, + "learning_rate": 7.913281043133978e-05, + "loss": 0.5533, + "step": 9624 + }, + { + "epoch": 0.646018589980202, + "grad_norm": 0.9766884446144104, + "learning_rate": 7.912397686692837e-05, + "loss": 0.6157, + "step": 9626 + }, + { + "epoch": 0.646152813663971, + "grad_norm": 1.074600338935852, + "learning_rate": 7.911514192646632e-05, + "loss": 0.598, + "step": 9628 + }, + { + "epoch": 0.64628703734774, + "grad_norm": 1.175805687904358, + "learning_rate": 7.910630561037109e-05, + "loss": 0.6646, + "step": 9630 + }, + { + "epoch": 0.646421261031509, + "grad_norm": 1.9594348669052124, + "learning_rate": 7.909746791906013e-05, + "loss": 0.621, + "step": 9632 + }, + { + "epoch": 0.646555484715278, + "grad_norm": 1.1910063028335571, + "learning_rate": 7.908862885295103e-05, + "loss": 0.6405, + "step": 9634 + }, + { + "epoch": 0.646689708399047, + "grad_norm": 0.9968884587287903, + "learning_rate": 7.907978841246144e-05, + "loss": 0.5714, + "step": 9636 + }, + { + "epoch": 0.646823932082816, + "grad_norm": 1.0536750555038452, + "learning_rate": 7.907094659800902e-05, + "loss": 0.6174, + "step": 9638 + }, + { + "epoch": 0.646958155766585, + "grad_norm": 2.651026725769043, + "learning_rate": 7.906210341001152e-05, + "loss": 0.5597, + "step": 9640 + }, + { + "epoch": 0.647092379450354, + "grad_norm": 1.0783993005752563, + "learning_rate": 7.905325884888679e-05, + "loss": 0.6645, + "step": 9642 + }, + { + "epoch": 0.6472266031341231, + "grad_norm": 0.9372936487197876, + "learning_rate": 7.90444129150527e-05, + "loss": 0.5837, + "step": 9644 + }, + { + "epoch": 0.647360826817892, + "grad_norm": 1.0757248401641846, + "learning_rate": 7.903556560892724e-05, + "loss": 0.6846, + "step": 9646 + }, + { + "epoch": 0.647495050501661, + "grad_norm": 1.0966620445251465, + "learning_rate": 7.90267169309284e-05, + "loss": 0.6526, + "step": 9648 + }, + { + "epoch": 0.64762927418543, + "grad_norm": 0.9838707447052002, + "learning_rate": 7.901786688147426e-05, + "loss": 0.6245, + "step": 9650 + }, + { + "epoch": 0.6477634978691991, + "grad_norm": 1.0680439472198486, + "learning_rate": 7.900901546098296e-05, + "loss": 0.6486, + "step": 9652 + }, + { + "epoch": 0.647897721552968, + "grad_norm": 1.0925967693328857, + "learning_rate": 7.900016266987274e-05, + "loss": 0.6563, + "step": 9654 + }, + { + "epoch": 0.648031945236737, + "grad_norm": 1.2490516901016235, + "learning_rate": 7.899130850856184e-05, + "loss": 0.6367, + "step": 9656 + }, + { + "epoch": 0.648166168920506, + "grad_norm": 1.0596709251403809, + "learning_rate": 7.898245297746863e-05, + "loss": 0.6069, + "step": 9658 + }, + { + "epoch": 0.6483003926042751, + "grad_norm": 1.0604771375656128, + "learning_rate": 7.897359607701151e-05, + "loss": 0.6104, + "step": 9660 + }, + { + "epoch": 0.648434616288044, + "grad_norm": 1.1038539409637451, + "learning_rate": 7.896473780760895e-05, + "loss": 0.5665, + "step": 9662 + }, + { + "epoch": 0.648568839971813, + "grad_norm": 0.9951624870300293, + "learning_rate": 7.895587816967948e-05, + "loss": 0.6164, + "step": 9664 + }, + { + "epoch": 0.6487030636555821, + "grad_norm": 1.1739658117294312, + "learning_rate": 7.89470171636417e-05, + "loss": 0.6248, + "step": 9666 + }, + { + "epoch": 0.648837287339351, + "grad_norm": 1.0021990537643433, + "learning_rate": 7.89381547899143e-05, + "loss": 0.5915, + "step": 9668 + }, + { + "epoch": 0.64897151102312, + "grad_norm": 1.0784162282943726, + "learning_rate": 7.892929104891598e-05, + "loss": 0.5992, + "step": 9670 + }, + { + "epoch": 0.649105734706889, + "grad_norm": 0.9180759191513062, + "learning_rate": 7.892042594106555e-05, + "loss": 0.5745, + "step": 9672 + }, + { + "epoch": 0.6492399583906581, + "grad_norm": 1.7134888172149658, + "learning_rate": 7.891155946678185e-05, + "loss": 0.6183, + "step": 9674 + }, + { + "epoch": 0.649374182074427, + "grad_norm": 1.005355715751648, + "learning_rate": 7.890269162648382e-05, + "loss": 0.6206, + "step": 9676 + }, + { + "epoch": 0.649508405758196, + "grad_norm": 1.0124843120574951, + "learning_rate": 7.889382242059044e-05, + "loss": 0.5833, + "step": 9678 + }, + { + "epoch": 0.649642629441965, + "grad_norm": 1.091232180595398, + "learning_rate": 7.888495184952079e-05, + "loss": 0.6312, + "step": 9680 + }, + { + "epoch": 0.6497768531257341, + "grad_norm": 0.9182711243629456, + "learning_rate": 7.887607991369396e-05, + "loss": 0.6144, + "step": 9682 + }, + { + "epoch": 0.649911076809503, + "grad_norm": 1.022800326347351, + "learning_rate": 7.886720661352913e-05, + "loss": 0.6135, + "step": 9684 + }, + { + "epoch": 0.650045300493272, + "grad_norm": 0.9301436543464661, + "learning_rate": 7.885833194944555e-05, + "loss": 0.5886, + "step": 9686 + }, + { + "epoch": 0.6501795241770411, + "grad_norm": 1.4711514711380005, + "learning_rate": 7.884945592186255e-05, + "loss": 0.6559, + "step": 9688 + }, + { + "epoch": 0.6503137478608101, + "grad_norm": 1.037255883216858, + "learning_rate": 7.884057853119947e-05, + "loss": 0.5976, + "step": 9690 + }, + { + "epoch": 0.650447971544579, + "grad_norm": 1.082667350769043, + "learning_rate": 7.883169977787577e-05, + "loss": 0.639, + "step": 9692 + }, + { + "epoch": 0.650582195228348, + "grad_norm": 1.0646504163742065, + "learning_rate": 7.882281966231094e-05, + "loss": 0.6241, + "step": 9694 + }, + { + "epoch": 0.6507164189121171, + "grad_norm": 0.9921514391899109, + "learning_rate": 7.881393818492457e-05, + "loss": 0.554, + "step": 9696 + }, + { + "epoch": 0.6508506425958861, + "grad_norm": 1.0990134477615356, + "learning_rate": 7.880505534613629e-05, + "loss": 0.6075, + "step": 9698 + }, + { + "epoch": 0.650984866279655, + "grad_norm": 1.0666207075119019, + "learning_rate": 7.879617114636577e-05, + "loss": 0.6823, + "step": 9700 + }, + { + "epoch": 0.651119089963424, + "grad_norm": 0.94024658203125, + "learning_rate": 7.87872855860328e-05, + "loss": 0.5955, + "step": 9702 + }, + { + "epoch": 0.6512533136471931, + "grad_norm": 1.038899302482605, + "learning_rate": 7.877839866555719e-05, + "loss": 0.6509, + "step": 9704 + }, + { + "epoch": 0.651387537330962, + "grad_norm": 1.012138843536377, + "learning_rate": 7.876951038535883e-05, + "loss": 0.6723, + "step": 9706 + }, + { + "epoch": 0.651521761014731, + "grad_norm": 1.1637332439422607, + "learning_rate": 7.876062074585768e-05, + "loss": 0.6284, + "step": 9708 + }, + { + "epoch": 0.6516559846985001, + "grad_norm": 1.057187557220459, + "learning_rate": 7.875172974747376e-05, + "loss": 0.5865, + "step": 9710 + }, + { + "epoch": 0.6517902083822691, + "grad_norm": 0.8246267437934875, + "learning_rate": 7.874283739062715e-05, + "loss": 0.5282, + "step": 9712 + }, + { + "epoch": 0.651924432066038, + "grad_norm": 1.2830623388290405, + "learning_rate": 7.873394367573798e-05, + "loss": 0.5927, + "step": 9714 + }, + { + "epoch": 0.652058655749807, + "grad_norm": 1.0558857917785645, + "learning_rate": 7.87250486032265e-05, + "loss": 0.5818, + "step": 9716 + }, + { + "epoch": 0.6521928794335761, + "grad_norm": 1.1389917135238647, + "learning_rate": 7.871615217351294e-05, + "loss": 0.6734, + "step": 9718 + }, + { + "epoch": 0.6523271031173451, + "grad_norm": 1.0388612747192383, + "learning_rate": 7.870725438701765e-05, + "loss": 0.6553, + "step": 9720 + }, + { + "epoch": 0.652461326801114, + "grad_norm": 0.9338822364807129, + "learning_rate": 7.869835524416104e-05, + "loss": 0.6292, + "step": 9722 + }, + { + "epoch": 0.652595550484883, + "grad_norm": 1.1659862995147705, + "learning_rate": 7.86894547453636e-05, + "loss": 0.6133, + "step": 9724 + }, + { + "epoch": 0.6527297741686521, + "grad_norm": 0.9935160279273987, + "learning_rate": 7.868055289104581e-05, + "loss": 0.5499, + "step": 9726 + }, + { + "epoch": 0.6528639978524211, + "grad_norm": 1.248638391494751, + "learning_rate": 7.867164968162833e-05, + "loss": 0.6316, + "step": 9728 + }, + { + "epoch": 0.65299822153619, + "grad_norm": 1.0811436176300049, + "learning_rate": 7.866274511753175e-05, + "loss": 0.63, + "step": 9730 + }, + { + "epoch": 0.6531324452199591, + "grad_norm": 1.0859920978546143, + "learning_rate": 7.865383919917683e-05, + "loss": 0.572, + "step": 9732 + }, + { + "epoch": 0.6532666689037281, + "grad_norm": 1.0905475616455078, + "learning_rate": 7.864493192698437e-05, + "loss": 0.5775, + "step": 9734 + }, + { + "epoch": 0.6534008925874971, + "grad_norm": 1.0847246646881104, + "learning_rate": 7.863602330137519e-05, + "loss": 0.6127, + "step": 9736 + }, + { + "epoch": 0.653535116271266, + "grad_norm": 0.9611513614654541, + "learning_rate": 7.862711332277023e-05, + "loss": 0.5731, + "step": 9738 + }, + { + "epoch": 0.6536693399550351, + "grad_norm": 0.94939786195755, + "learning_rate": 7.861820199159044e-05, + "loss": 0.6028, + "step": 9740 + }, + { + "epoch": 0.6538035636388041, + "grad_norm": 0.9335368275642395, + "learning_rate": 7.860928930825691e-05, + "loss": 0.5852, + "step": 9742 + }, + { + "epoch": 0.653937787322573, + "grad_norm": 1.0535403490066528, + "learning_rate": 7.86003752731907e-05, + "loss": 0.6592, + "step": 9744 + }, + { + "epoch": 0.654072011006342, + "grad_norm": 1.0651839971542358, + "learning_rate": 7.8591459886813e-05, + "loss": 0.6449, + "step": 9746 + }, + { + "epoch": 0.6542062346901111, + "grad_norm": 1.0187623500823975, + "learning_rate": 7.858254314954505e-05, + "loss": 0.6007, + "step": 9748 + }, + { + "epoch": 0.6543404583738801, + "grad_norm": 1.020383596420288, + "learning_rate": 7.857362506180813e-05, + "loss": 0.6603, + "step": 9750 + }, + { + "epoch": 0.654474682057649, + "grad_norm": 1.2457799911499023, + "learning_rate": 7.856470562402362e-05, + "loss": 0.5842, + "step": 9752 + }, + { + "epoch": 0.6546089057414181, + "grad_norm": 1.057938814163208, + "learning_rate": 7.855578483661293e-05, + "loss": 0.6104, + "step": 9754 + }, + { + "epoch": 0.6547431294251871, + "grad_norm": 1.1072943210601807, + "learning_rate": 7.854686269999757e-05, + "loss": 0.5934, + "step": 9756 + }, + { + "epoch": 0.6548773531089561, + "grad_norm": 1.2148404121398926, + "learning_rate": 7.853793921459909e-05, + "loss": 0.6323, + "step": 9758 + }, + { + "epoch": 0.655011576792725, + "grad_norm": 1.1100873947143555, + "learning_rate": 7.852901438083908e-05, + "loss": 0.6788, + "step": 9760 + }, + { + "epoch": 0.6551458004764941, + "grad_norm": 0.9000804424285889, + "learning_rate": 7.852008819913925e-05, + "loss": 0.6299, + "step": 9762 + }, + { + "epoch": 0.6552800241602631, + "grad_norm": 1.2299797534942627, + "learning_rate": 7.851116066992133e-05, + "loss": 0.603, + "step": 9764 + }, + { + "epoch": 0.6554142478440321, + "grad_norm": 1.0158828496932983, + "learning_rate": 7.850223179360714e-05, + "loss": 0.6332, + "step": 9766 + }, + { + "epoch": 0.655548471527801, + "grad_norm": 1.0191811323165894, + "learning_rate": 7.849330157061854e-05, + "loss": 0.5966, + "step": 9768 + }, + { + "epoch": 0.6556826952115701, + "grad_norm": 1.0282412767410278, + "learning_rate": 7.848437000137747e-05, + "loss": 0.6104, + "step": 9770 + }, + { + "epoch": 0.6558169188953391, + "grad_norm": 1.0161830186843872, + "learning_rate": 7.847543708630593e-05, + "loss": 0.6041, + "step": 9772 + }, + { + "epoch": 0.6559511425791081, + "grad_norm": 1.0597655773162842, + "learning_rate": 7.8466502825826e-05, + "loss": 0.5792, + "step": 9774 + }, + { + "epoch": 0.6560853662628771, + "grad_norm": 1.132139801979065, + "learning_rate": 7.845756722035978e-05, + "loss": 0.5909, + "step": 9776 + }, + { + "epoch": 0.6562195899466461, + "grad_norm": 0.9694353342056274, + "learning_rate": 7.844863027032945e-05, + "loss": 0.5742, + "step": 9778 + }, + { + "epoch": 0.6563538136304151, + "grad_norm": 1.0666712522506714, + "learning_rate": 7.84396919761573e-05, + "loss": 0.6605, + "step": 9780 + }, + { + "epoch": 0.656488037314184, + "grad_norm": 0.9630760550498962, + "learning_rate": 7.843075233826561e-05, + "loss": 0.5949, + "step": 9782 + }, + { + "epoch": 0.6566222609979531, + "grad_norm": 1.3532354831695557, + "learning_rate": 7.84218113570768e-05, + "loss": 0.5712, + "step": 9784 + }, + { + "epoch": 0.6567564846817221, + "grad_norm": 1.3067315816879272, + "learning_rate": 7.841286903301328e-05, + "loss": 0.5465, + "step": 9786 + }, + { + "epoch": 0.6568907083654911, + "grad_norm": 0.9235031604766846, + "learning_rate": 7.840392536649758e-05, + "loss": 0.5865, + "step": 9788 + }, + { + "epoch": 0.65702493204926, + "grad_norm": 1.137033224105835, + "learning_rate": 7.839498035795224e-05, + "loss": 0.6354, + "step": 9790 + }, + { + "epoch": 0.6571591557330291, + "grad_norm": 1.0483646392822266, + "learning_rate": 7.838603400779993e-05, + "loss": 0.6268, + "step": 9792 + }, + { + "epoch": 0.6572933794167981, + "grad_norm": 1.0721083879470825, + "learning_rate": 7.837708631646333e-05, + "loss": 0.6168, + "step": 9794 + }, + { + "epoch": 0.6574276031005671, + "grad_norm": 1.1923506259918213, + "learning_rate": 7.836813728436521e-05, + "loss": 0.625, + "step": 9796 + }, + { + "epoch": 0.6575618267843361, + "grad_norm": 1.0472887754440308, + "learning_rate": 7.835918691192837e-05, + "loss": 0.6315, + "step": 9798 + }, + { + "epoch": 0.6576960504681051, + "grad_norm": 1.0348756313323975, + "learning_rate": 7.835023519957571e-05, + "loss": 0.6201, + "step": 9800 + }, + { + "epoch": 0.6578302741518741, + "grad_norm": 1.2074779272079468, + "learning_rate": 7.834128214773022e-05, + "loss": 0.5988, + "step": 9802 + }, + { + "epoch": 0.6579644978356431, + "grad_norm": 1.0870145559310913, + "learning_rate": 7.833232775681485e-05, + "loss": 0.6503, + "step": 9804 + }, + { + "epoch": 0.6580987215194121, + "grad_norm": 0.9929474592208862, + "learning_rate": 7.83233720272527e-05, + "loss": 0.6137, + "step": 9806 + }, + { + "epoch": 0.6582329452031811, + "grad_norm": 1.4966610670089722, + "learning_rate": 7.831441495946694e-05, + "loss": 0.6278, + "step": 9808 + }, + { + "epoch": 0.6583671688869501, + "grad_norm": 1.0693825483322144, + "learning_rate": 7.830545655388075e-05, + "loss": 0.5809, + "step": 9810 + }, + { + "epoch": 0.6585013925707192, + "grad_norm": 1.047507405281067, + "learning_rate": 7.829649681091738e-05, + "loss": 0.5508, + "step": 9812 + }, + { + "epoch": 0.6586356162544881, + "grad_norm": 1.130644679069519, + "learning_rate": 7.828753573100019e-05, + "loss": 0.6143, + "step": 9814 + }, + { + "epoch": 0.6587698399382571, + "grad_norm": 0.9263031482696533, + "learning_rate": 7.827857331455256e-05, + "loss": 0.5803, + "step": 9816 + }, + { + "epoch": 0.6589040636220261, + "grad_norm": 1.7844195365905762, + "learning_rate": 7.826960956199794e-05, + "loss": 0.5558, + "step": 9818 + }, + { + "epoch": 0.6590382873057951, + "grad_norm": 1.1864254474639893, + "learning_rate": 7.826064447375987e-05, + "loss": 0.5985, + "step": 9820 + }, + { + "epoch": 0.6591725109895641, + "grad_norm": 1.0454565286636353, + "learning_rate": 7.825167805026193e-05, + "loss": 0.6035, + "step": 9822 + }, + { + "epoch": 0.6593067346733331, + "grad_norm": 0.9805462956428528, + "learning_rate": 7.824271029192773e-05, + "loss": 0.6726, + "step": 9824 + }, + { + "epoch": 0.6594409583571021, + "grad_norm": 0.9167178869247437, + "learning_rate": 7.823374119918103e-05, + "loss": 0.5407, + "step": 9826 + }, + { + "epoch": 0.6595751820408711, + "grad_norm": 0.9714899659156799, + "learning_rate": 7.822477077244557e-05, + "loss": 0.6605, + "step": 9828 + }, + { + "epoch": 0.6597094057246401, + "grad_norm": 1.0378906726837158, + "learning_rate": 7.821579901214518e-05, + "loss": 0.5913, + "step": 9830 + }, + { + "epoch": 0.6598436294084091, + "grad_norm": 1.2218577861785889, + "learning_rate": 7.820682591870378e-05, + "loss": 0.6122, + "step": 9832 + }, + { + "epoch": 0.6599778530921782, + "grad_norm": 1.1689800024032593, + "learning_rate": 7.819785149254532e-05, + "loss": 0.6229, + "step": 9834 + }, + { + "epoch": 0.6601120767759471, + "grad_norm": 1.148080587387085, + "learning_rate": 7.818887573409383e-05, + "loss": 0.6555, + "step": 9836 + }, + { + "epoch": 0.6602463004597161, + "grad_norm": 1.0432515144348145, + "learning_rate": 7.817989864377339e-05, + "loss": 0.5595, + "step": 9838 + }, + { + "epoch": 0.6603805241434851, + "grad_norm": 0.9731805324554443, + "learning_rate": 7.817092022200816e-05, + "loss": 0.5145, + "step": 9840 + }, + { + "epoch": 0.6605147478272542, + "grad_norm": 0.9755486249923706, + "learning_rate": 7.816194046922234e-05, + "loss": 0.5965, + "step": 9842 + }, + { + "epoch": 0.6606489715110231, + "grad_norm": 1.0782371759414673, + "learning_rate": 7.815295938584021e-05, + "loss": 0.6653, + "step": 9844 + }, + { + "epoch": 0.6607831951947921, + "grad_norm": 1.184570550918579, + "learning_rate": 7.81439769722861e-05, + "loss": 0.6332, + "step": 9846 + }, + { + "epoch": 0.6609174188785611, + "grad_norm": 0.9063990116119385, + "learning_rate": 7.813499322898443e-05, + "loss": 0.5187, + "step": 9848 + }, + { + "epoch": 0.6610516425623302, + "grad_norm": 0.9837430715560913, + "learning_rate": 7.812600815635967e-05, + "loss": 0.6289, + "step": 9850 + }, + { + "epoch": 0.6611858662460991, + "grad_norm": 0.9125910401344299, + "learning_rate": 7.81170217548363e-05, + "loss": 0.6033, + "step": 9852 + }, + { + "epoch": 0.6613200899298681, + "grad_norm": 1.0609534978866577, + "learning_rate": 7.810803402483897e-05, + "loss": 0.5779, + "step": 9854 + }, + { + "epoch": 0.6614543136136372, + "grad_norm": 0.9907367825508118, + "learning_rate": 7.80990449667923e-05, + "loss": 0.734, + "step": 9856 + }, + { + "epoch": 0.6615885372974061, + "grad_norm": 0.9950106143951416, + "learning_rate": 7.8090054581121e-05, + "loss": 0.6049, + "step": 9858 + }, + { + "epoch": 0.6617227609811751, + "grad_norm": 1.0695754289627075, + "learning_rate": 7.808106286824986e-05, + "loss": 0.6191, + "step": 9860 + }, + { + "epoch": 0.6618569846649441, + "grad_norm": 1.0091043710708618, + "learning_rate": 7.807206982860371e-05, + "loss": 0.5934, + "step": 9862 + }, + { + "epoch": 0.6619912083487132, + "grad_norm": 1.0125815868377686, + "learning_rate": 7.806307546260748e-05, + "loss": 0.6075, + "step": 9864 + }, + { + "epoch": 0.6621254320324821, + "grad_norm": 1.187417984008789, + "learning_rate": 7.80540797706861e-05, + "loss": 0.6647, + "step": 9866 + }, + { + "epoch": 0.6622596557162511, + "grad_norm": 1.3254411220550537, + "learning_rate": 7.804508275326462e-05, + "loss": 0.7018, + "step": 9868 + }, + { + "epoch": 0.6623938794000201, + "grad_norm": 0.9366840124130249, + "learning_rate": 7.803608441076815e-05, + "loss": 0.551, + "step": 9870 + }, + { + "epoch": 0.6625281030837892, + "grad_norm": 1.0737332105636597, + "learning_rate": 7.80270847436218e-05, + "loss": 0.6484, + "step": 9872 + }, + { + "epoch": 0.6626623267675581, + "grad_norm": 0.9313533902168274, + "learning_rate": 7.801808375225082e-05, + "loss": 0.6453, + "step": 9874 + }, + { + "epoch": 0.6627965504513271, + "grad_norm": 1.0625557899475098, + "learning_rate": 7.800908143708047e-05, + "loss": 0.5741, + "step": 9876 + }, + { + "epoch": 0.6629307741350962, + "grad_norm": 1.0991019010543823, + "learning_rate": 7.800007779853611e-05, + "loss": 0.5746, + "step": 9878 + }, + { + "epoch": 0.6630649978188652, + "grad_norm": 0.9642413258552551, + "learning_rate": 7.799107283704312e-05, + "loss": 0.6528, + "step": 9880 + }, + { + "epoch": 0.6631992215026341, + "grad_norm": 1.091901421546936, + "learning_rate": 7.798206655302698e-05, + "loss": 0.5899, + "step": 9882 + }, + { + "epoch": 0.6633334451864031, + "grad_norm": 1.035330891609192, + "learning_rate": 7.797305894691322e-05, + "loss": 0.6699, + "step": 9884 + }, + { + "epoch": 0.6634676688701722, + "grad_norm": 1.0007174015045166, + "learning_rate": 7.796405001912746e-05, + "loss": 0.5949, + "step": 9886 + }, + { + "epoch": 0.6636018925539412, + "grad_norm": 0.9840034246444702, + "learning_rate": 7.795503977009528e-05, + "loss": 0.6358, + "step": 9888 + }, + { + "epoch": 0.6637361162377101, + "grad_norm": 0.9621015787124634, + "learning_rate": 7.794602820024248e-05, + "loss": 0.5859, + "step": 9890 + }, + { + "epoch": 0.6638703399214791, + "grad_norm": 0.9338459372520447, + "learning_rate": 7.793701530999478e-05, + "loss": 0.58, + "step": 9892 + }, + { + "epoch": 0.6640045636052482, + "grad_norm": 1.088112473487854, + "learning_rate": 7.792800109977804e-05, + "loss": 0.5988, + "step": 9894 + }, + { + "epoch": 0.6641387872890171, + "grad_norm": 1.1450368165969849, + "learning_rate": 7.791898557001818e-05, + "loss": 0.664, + "step": 9896 + }, + { + "epoch": 0.6642730109727861, + "grad_norm": 0.8303860425949097, + "learning_rate": 7.790996872114116e-05, + "loss": 0.6325, + "step": 9898 + }, + { + "epoch": 0.6644072346565552, + "grad_norm": 1.0726555585861206, + "learning_rate": 7.790095055357298e-05, + "loss": 0.6394, + "step": 9900 + }, + { + "epoch": 0.6645414583403242, + "grad_norm": 1.5649341344833374, + "learning_rate": 7.789193106773976e-05, + "loss": 0.6188, + "step": 9902 + }, + { + "epoch": 0.6646756820240931, + "grad_norm": 1.0068936347961426, + "learning_rate": 7.788291026406764e-05, + "loss": 0.5768, + "step": 9904 + }, + { + "epoch": 0.6648099057078621, + "grad_norm": 1.008538007736206, + "learning_rate": 7.787388814298284e-05, + "loss": 0.6108, + "step": 9906 + }, + { + "epoch": 0.6649441293916312, + "grad_norm": 1.0401108264923096, + "learning_rate": 7.786486470491165e-05, + "loss": 0.6475, + "step": 9908 + }, + { + "epoch": 0.6650783530754002, + "grad_norm": 1.1032782793045044, + "learning_rate": 7.785583995028038e-05, + "loss": 0.6025, + "step": 9910 + }, + { + "epoch": 0.6652125767591691, + "grad_norm": 1.1373237371444702, + "learning_rate": 7.784681387951546e-05, + "loss": 0.6146, + "step": 9912 + }, + { + "epoch": 0.6653468004429381, + "grad_norm": 0.8877438306808472, + "learning_rate": 7.783778649304333e-05, + "loss": 0.5687, + "step": 9914 + }, + { + "epoch": 0.6654810241267072, + "grad_norm": 1.1277129650115967, + "learning_rate": 7.782875779129055e-05, + "loss": 0.5518, + "step": 9916 + }, + { + "epoch": 0.6656152478104762, + "grad_norm": 1.2100125551223755, + "learning_rate": 7.781972777468367e-05, + "loss": 0.6513, + "step": 9918 + }, + { + "epoch": 0.6657494714942451, + "grad_norm": 1.1718670129776, + "learning_rate": 7.781069644364936e-05, + "loss": 0.6516, + "step": 9920 + }, + { + "epoch": 0.6658836951780142, + "grad_norm": 0.9917896389961243, + "learning_rate": 7.780166379861432e-05, + "loss": 0.604, + "step": 9922 + }, + { + "epoch": 0.6660179188617832, + "grad_norm": 0.9580020308494568, + "learning_rate": 7.779262984000536e-05, + "loss": 0.6172, + "step": 9924 + }, + { + "epoch": 0.6661521425455522, + "grad_norm": 1.0788674354553223, + "learning_rate": 7.778359456824929e-05, + "loss": 0.6439, + "step": 9926 + }, + { + "epoch": 0.6662863662293211, + "grad_norm": 1.4100440740585327, + "learning_rate": 7.777455798377297e-05, + "loss": 0.6557, + "step": 9928 + }, + { + "epoch": 0.6664205899130902, + "grad_norm": 1.071541428565979, + "learning_rate": 7.776552008700344e-05, + "loss": 0.6275, + "step": 9930 + }, + { + "epoch": 0.6665548135968592, + "grad_norm": 0.9982123374938965, + "learning_rate": 7.775648087836768e-05, + "loss": 0.5679, + "step": 9932 + }, + { + "epoch": 0.6666890372806281, + "grad_norm": 1.0546722412109375, + "learning_rate": 7.774744035829277e-05, + "loss": 0.7266, + "step": 9934 + }, + { + "epoch": 0.6668232609643971, + "grad_norm": 1.2762454748153687, + "learning_rate": 7.773839852720589e-05, + "loss": 0.6516, + "step": 9936 + }, + { + "epoch": 0.6669574846481662, + "grad_norm": 1.0322846174240112, + "learning_rate": 7.772935538553418e-05, + "loss": 0.6761, + "step": 9938 + }, + { + "epoch": 0.6670917083319352, + "grad_norm": 1.0097932815551758, + "learning_rate": 7.772031093370499e-05, + "loss": 0.6756, + "step": 9940 + }, + { + "epoch": 0.6672259320157041, + "grad_norm": 0.9951656460762024, + "learning_rate": 7.771126517214561e-05, + "loss": 0.6605, + "step": 9942 + }, + { + "epoch": 0.6673601556994732, + "grad_norm": 0.9250411987304688, + "learning_rate": 7.770221810128343e-05, + "loss": 0.6245, + "step": 9944 + }, + { + "epoch": 0.6674943793832422, + "grad_norm": 0.9195786714553833, + "learning_rate": 7.769316972154594e-05, + "loss": 0.6032, + "step": 9946 + }, + { + "epoch": 0.6676286030670112, + "grad_norm": 1.2926996946334839, + "learning_rate": 7.768412003336064e-05, + "loss": 0.6591, + "step": 9948 + }, + { + "epoch": 0.6677628267507801, + "grad_norm": 0.9514656066894531, + "learning_rate": 7.767506903715509e-05, + "loss": 0.6087, + "step": 9950 + }, + { + "epoch": 0.6678970504345492, + "grad_norm": 1.1330329179763794, + "learning_rate": 7.766601673335694e-05, + "loss": 0.6591, + "step": 9952 + }, + { + "epoch": 0.6680312741183182, + "grad_norm": 0.8967975974082947, + "learning_rate": 7.765696312239392e-05, + "loss": 0.5153, + "step": 9954 + }, + { + "epoch": 0.6681654978020872, + "grad_norm": 1.05215322971344, + "learning_rate": 7.764790820469377e-05, + "loss": 0.5942, + "step": 9956 + }, + { + "epoch": 0.6682997214858561, + "grad_norm": 1.0084857940673828, + "learning_rate": 7.763885198068433e-05, + "loss": 0.552, + "step": 9958 + }, + { + "epoch": 0.6684339451696252, + "grad_norm": 1.0705190896987915, + "learning_rate": 7.762979445079348e-05, + "loss": 0.6433, + "step": 9960 + }, + { + "epoch": 0.6685681688533942, + "grad_norm": 1.0716350078582764, + "learning_rate": 7.762073561544918e-05, + "loss": 0.6164, + "step": 9962 + }, + { + "epoch": 0.6687023925371632, + "grad_norm": 0.9401054382324219, + "learning_rate": 7.761167547507942e-05, + "loss": 0.6359, + "step": 9964 + }, + { + "epoch": 0.6688366162209322, + "grad_norm": 0.818011999130249, + "learning_rate": 7.760261403011231e-05, + "loss": 0.5992, + "step": 9966 + }, + { + "epoch": 0.6689708399047012, + "grad_norm": 0.9985383152961731, + "learning_rate": 7.759355128097595e-05, + "loss": 0.6641, + "step": 9968 + }, + { + "epoch": 0.6691050635884702, + "grad_norm": 1.016554832458496, + "learning_rate": 7.758448722809856e-05, + "loss": 0.6243, + "step": 9970 + }, + { + "epoch": 0.6692392872722391, + "grad_norm": 1.0722575187683105, + "learning_rate": 7.757542187190838e-05, + "loss": 0.6227, + "step": 9972 + }, + { + "epoch": 0.6693735109560082, + "grad_norm": 1.0455728769302368, + "learning_rate": 7.756635521283375e-05, + "loss": 0.63, + "step": 9974 + }, + { + "epoch": 0.6695077346397772, + "grad_norm": 1.0678956508636475, + "learning_rate": 7.755728725130304e-05, + "loss": 0.5656, + "step": 9976 + }, + { + "epoch": 0.6696419583235462, + "grad_norm": 0.9857407808303833, + "learning_rate": 7.754821798774471e-05, + "loss": 0.6717, + "step": 9978 + }, + { + "epoch": 0.6697761820073151, + "grad_norm": 0.9332596063613892, + "learning_rate": 7.753914742258724e-05, + "loss": 0.5908, + "step": 9980 + }, + { + "epoch": 0.6699104056910842, + "grad_norm": 1.1080443859100342, + "learning_rate": 7.75300755562592e-05, + "loss": 0.67, + "step": 9982 + }, + { + "epoch": 0.6700446293748532, + "grad_norm": 1.1003410816192627, + "learning_rate": 7.752100238918925e-05, + "loss": 0.7029, + "step": 9984 + }, + { + "epoch": 0.6701788530586222, + "grad_norm": 0.9595111608505249, + "learning_rate": 7.751192792180604e-05, + "loss": 0.5828, + "step": 9986 + }, + { + "epoch": 0.6703130767423912, + "grad_norm": 1.0013678073883057, + "learning_rate": 7.750285215453833e-05, + "loss": 0.5596, + "step": 9988 + }, + { + "epoch": 0.6704473004261602, + "grad_norm": 1.0782396793365479, + "learning_rate": 7.749377508781495e-05, + "loss": 0.6067, + "step": 9990 + }, + { + "epoch": 0.6705815241099292, + "grad_norm": 1.0426450967788696, + "learning_rate": 7.748469672206476e-05, + "loss": 0.5675, + "step": 9992 + }, + { + "epoch": 0.6707157477936982, + "grad_norm": 1.0867704153060913, + "learning_rate": 7.747561705771669e-05, + "loss": 0.6423, + "step": 9994 + }, + { + "epoch": 0.6708499714774672, + "grad_norm": 0.949182391166687, + "learning_rate": 7.746653609519973e-05, + "loss": 0.6095, + "step": 9996 + }, + { + "epoch": 0.6709841951612362, + "grad_norm": 0.9397948384284973, + "learning_rate": 7.745745383494296e-05, + "loss": 0.5811, + "step": 9998 + }, + { + "epoch": 0.6711184188450052, + "grad_norm": 0.9894161820411682, + "learning_rate": 7.74483702773755e-05, + "loss": 0.6259, + "step": 10000 + }, + { + "epoch": 0.6712526425287743, + "grad_norm": 1.345678448677063, + "learning_rate": 7.74392854229265e-05, + "loss": 0.5753, + "step": 10002 + }, + { + "epoch": 0.6713868662125432, + "grad_norm": 1.0500710010528564, + "learning_rate": 7.743019927202524e-05, + "loss": 0.6433, + "step": 10004 + }, + { + "epoch": 0.6715210898963122, + "grad_norm": 0.8496576547622681, + "learning_rate": 7.7421111825101e-05, + "loss": 0.5807, + "step": 10006 + }, + { + "epoch": 0.6716553135800812, + "grad_norm": 1.0088499784469604, + "learning_rate": 7.741202308258314e-05, + "loss": 0.5823, + "step": 10008 + }, + { + "epoch": 0.6717895372638502, + "grad_norm": 1.0959690809249878, + "learning_rate": 7.74029330449011e-05, + "loss": 0.5742, + "step": 10010 + }, + { + "epoch": 0.6719237609476192, + "grad_norm": 1.1204447746276855, + "learning_rate": 7.739384171248435e-05, + "loss": 0.665, + "step": 10012 + }, + { + "epoch": 0.6720579846313882, + "grad_norm": 1.0192700624465942, + "learning_rate": 7.738474908576246e-05, + "loss": 0.6114, + "step": 10014 + }, + { + "epoch": 0.6721922083151572, + "grad_norm": 1.0640549659729004, + "learning_rate": 7.737565516516501e-05, + "loss": 0.6024, + "step": 10016 + }, + { + "epoch": 0.6723264319989262, + "grad_norm": 0.9268406629562378, + "learning_rate": 7.73665599511217e-05, + "loss": 0.6366, + "step": 10018 + }, + { + "epoch": 0.6724606556826952, + "grad_norm": 1.0345485210418701, + "learning_rate": 7.735746344406223e-05, + "loss": 0.5904, + "step": 10020 + }, + { + "epoch": 0.6725948793664642, + "grad_norm": 1.0905174016952515, + "learning_rate": 7.734836564441642e-05, + "loss": 0.6618, + "step": 10022 + }, + { + "epoch": 0.6727291030502333, + "grad_norm": 2.176734447479248, + "learning_rate": 7.733926655261411e-05, + "loss": 0.6098, + "step": 10024 + }, + { + "epoch": 0.6728633267340022, + "grad_norm": 1.0297119617462158, + "learning_rate": 7.733016616908521e-05, + "loss": 0.624, + "step": 10026 + }, + { + "epoch": 0.6729975504177712, + "grad_norm": 1.0259052515029907, + "learning_rate": 7.73210644942597e-05, + "loss": 0.6121, + "step": 10028 + }, + { + "epoch": 0.6731317741015402, + "grad_norm": 1.0044418573379517, + "learning_rate": 7.731196152856763e-05, + "loss": 0.5884, + "step": 10030 + }, + { + "epoch": 0.6732659977853093, + "grad_norm": 1.0181063413619995, + "learning_rate": 7.730285727243907e-05, + "loss": 0.6566, + "step": 10032 + }, + { + "epoch": 0.6734002214690782, + "grad_norm": 1.5579719543457031, + "learning_rate": 7.72937517263042e-05, + "loss": 0.5857, + "step": 10034 + }, + { + "epoch": 0.6735344451528472, + "grad_norm": 1.3661072254180908, + "learning_rate": 7.728464489059324e-05, + "loss": 0.6911, + "step": 10036 + }, + { + "epoch": 0.6736686688366162, + "grad_norm": 1.1502236127853394, + "learning_rate": 7.727553676573644e-05, + "loss": 0.6244, + "step": 10038 + }, + { + "epoch": 0.6738028925203853, + "grad_norm": 1.0469461679458618, + "learning_rate": 7.726642735216418e-05, + "loss": 0.5851, + "step": 10040 + }, + { + "epoch": 0.6739371162041542, + "grad_norm": 1.1454188823699951, + "learning_rate": 7.725731665030684e-05, + "loss": 0.626, + "step": 10042 + }, + { + "epoch": 0.6740713398879232, + "grad_norm": 1.0856674909591675, + "learning_rate": 7.72482046605949e-05, + "loss": 0.6999, + "step": 10044 + }, + { + "epoch": 0.6742055635716923, + "grad_norm": 1.0814210176467896, + "learning_rate": 7.723909138345886e-05, + "loss": 0.6791, + "step": 10046 + }, + { + "epoch": 0.6743397872554612, + "grad_norm": 1.538653016090393, + "learning_rate": 7.722997681932931e-05, + "loss": 0.6329, + "step": 10048 + }, + { + "epoch": 0.6744740109392302, + "grad_norm": 0.9768089056015015, + "learning_rate": 7.722086096863692e-05, + "loss": 0.6018, + "step": 10050 + }, + { + "epoch": 0.6746082346229992, + "grad_norm": 1.0232763290405273, + "learning_rate": 7.721174383181235e-05, + "loss": 0.6223, + "step": 10052 + }, + { + "epoch": 0.6747424583067683, + "grad_norm": 1.0074292421340942, + "learning_rate": 7.720262540928641e-05, + "loss": 0.613, + "step": 10054 + }, + { + "epoch": 0.6748766819905372, + "grad_norm": 1.0369144678115845, + "learning_rate": 7.719350570148991e-05, + "loss": 0.5817, + "step": 10056 + }, + { + "epoch": 0.6750109056743062, + "grad_norm": 1.017956256866455, + "learning_rate": 7.718438470885375e-05, + "loss": 0.5936, + "step": 10058 + }, + { + "epoch": 0.6751451293580752, + "grad_norm": 1.0069831609725952, + "learning_rate": 7.717526243180883e-05, + "loss": 0.5982, + "step": 10060 + }, + { + "epoch": 0.6752793530418443, + "grad_norm": 0.9372869729995728, + "learning_rate": 7.716613887078623e-05, + "loss": 0.6255, + "step": 10062 + }, + { + "epoch": 0.6754135767256132, + "grad_norm": 1.2310115098953247, + "learning_rate": 7.715701402621699e-05, + "loss": 0.6804, + "step": 10064 + }, + { + "epoch": 0.6755478004093822, + "grad_norm": 1.140617847442627, + "learning_rate": 7.714788789853223e-05, + "loss": 0.5764, + "step": 10066 + }, + { + "epoch": 0.6756820240931513, + "grad_norm": 1.0454936027526855, + "learning_rate": 7.713876048816316e-05, + "loss": 0.6441, + "step": 10068 + }, + { + "epoch": 0.6758162477769203, + "grad_norm": 1.1380059719085693, + "learning_rate": 7.7129631795541e-05, + "loss": 0.6409, + "step": 10070 + }, + { + "epoch": 0.6759504714606892, + "grad_norm": 1.0036143064498901, + "learning_rate": 7.712050182109711e-05, + "loss": 0.5859, + "step": 10072 + }, + { + "epoch": 0.6760846951444582, + "grad_norm": 1.0466136932373047, + "learning_rate": 7.711137056526283e-05, + "loss": 0.6363, + "step": 10074 + }, + { + "epoch": 0.6762189188282273, + "grad_norm": 1.055659294128418, + "learning_rate": 7.71022380284696e-05, + "loss": 0.6195, + "step": 10076 + }, + { + "epoch": 0.6763531425119963, + "grad_norm": 1.0443370342254639, + "learning_rate": 7.709310421114892e-05, + "loss": 0.6196, + "step": 10078 + }, + { + "epoch": 0.6764873661957652, + "grad_norm": 1.1877312660217285, + "learning_rate": 7.708396911373233e-05, + "loss": 0.5761, + "step": 10080 + }, + { + "epoch": 0.6766215898795342, + "grad_norm": 1.0187946557998657, + "learning_rate": 7.707483273665149e-05, + "loss": 0.6427, + "step": 10082 + }, + { + "epoch": 0.6767558135633033, + "grad_norm": 1.349446177482605, + "learning_rate": 7.706569508033801e-05, + "loss": 0.637, + "step": 10084 + }, + { + "epoch": 0.6768900372470722, + "grad_norm": 1.0880353450775146, + "learning_rate": 7.705655614522367e-05, + "loss": 0.6479, + "step": 10086 + }, + { + "epoch": 0.6770242609308412, + "grad_norm": 0.7679103016853333, + "learning_rate": 7.704741593174026e-05, + "loss": 0.5232, + "step": 10088 + }, + { + "epoch": 0.6771584846146103, + "grad_norm": 1.0523754358291626, + "learning_rate": 7.703827444031963e-05, + "loss": 0.5853, + "step": 10090 + }, + { + "epoch": 0.6772927082983793, + "grad_norm": 1.0461854934692383, + "learning_rate": 7.702913167139372e-05, + "loss": 0.6162, + "step": 10092 + }, + { + "epoch": 0.6774269319821482, + "grad_norm": 0.9754141569137573, + "learning_rate": 7.701998762539446e-05, + "loss": 0.621, + "step": 10094 + }, + { + "epoch": 0.6775611556659172, + "grad_norm": 1.0695887804031372, + "learning_rate": 7.701084230275392e-05, + "loss": 0.6168, + "step": 10096 + }, + { + "epoch": 0.6776953793496863, + "grad_norm": 1.1557365655899048, + "learning_rate": 7.700169570390422e-05, + "loss": 0.6171, + "step": 10098 + }, + { + "epoch": 0.6778296030334553, + "grad_norm": 1.1904844045639038, + "learning_rate": 7.699254782927749e-05, + "loss": 0.6529, + "step": 10100 + }, + { + "epoch": 0.6779638267172242, + "grad_norm": 1.0055497884750366, + "learning_rate": 7.698339867930592e-05, + "loss": 0.6299, + "step": 10102 + }, + { + "epoch": 0.6780980504009932, + "grad_norm": 1.1452466249465942, + "learning_rate": 7.697424825442187e-05, + "loss": 0.5837, + "step": 10104 + }, + { + "epoch": 0.6782322740847623, + "grad_norm": 1.1297428607940674, + "learning_rate": 7.69650965550576e-05, + "loss": 0.6165, + "step": 10106 + }, + { + "epoch": 0.6783664977685313, + "grad_norm": 1.0672310590744019, + "learning_rate": 7.695594358164557e-05, + "loss": 0.6501, + "step": 10108 + }, + { + "epoch": 0.6785007214523002, + "grad_norm": 1.1842536926269531, + "learning_rate": 7.694678933461818e-05, + "loss": 0.595, + "step": 10110 + }, + { + "epoch": 0.6786349451360693, + "grad_norm": 1.150343418121338, + "learning_rate": 7.6937633814408e-05, + "loss": 0.6444, + "step": 10112 + }, + { + "epoch": 0.6787691688198383, + "grad_norm": 0.9357632398605347, + "learning_rate": 7.692847702144759e-05, + "loss": 0.6438, + "step": 10114 + }, + { + "epoch": 0.6789033925036073, + "grad_norm": 1.2471938133239746, + "learning_rate": 7.69193189561696e-05, + "loss": 0.6013, + "step": 10116 + }, + { + "epoch": 0.6790376161873762, + "grad_norm": 1.1025973558425903, + "learning_rate": 7.691015961900671e-05, + "loss": 0.5623, + "step": 10118 + }, + { + "epoch": 0.6791718398711453, + "grad_norm": 1.0556445121765137, + "learning_rate": 7.69009990103917e-05, + "loss": 0.5833, + "step": 10120 + }, + { + "epoch": 0.6793060635549143, + "grad_norm": 1.170667052268982, + "learning_rate": 7.689183713075741e-05, + "loss": 0.6205, + "step": 10122 + }, + { + "epoch": 0.6794402872386832, + "grad_norm": 1.0111286640167236, + "learning_rate": 7.688267398053665e-05, + "loss": 0.626, + "step": 10124 + }, + { + "epoch": 0.6795745109224522, + "grad_norm": 1.1592979431152344, + "learning_rate": 7.687350956016244e-05, + "loss": 0.5957, + "step": 10126 + }, + { + "epoch": 0.6797087346062213, + "grad_norm": 1.0017783641815186, + "learning_rate": 7.686434387006773e-05, + "loss": 0.6221, + "step": 10128 + }, + { + "epoch": 0.6798429582899903, + "grad_norm": 1.0687118768692017, + "learning_rate": 7.685517691068563e-05, + "loss": 0.6079, + "step": 10130 + }, + { + "epoch": 0.6799771819737592, + "grad_norm": 1.0000358819961548, + "learning_rate": 7.68460086824492e-05, + "loss": 0.6036, + "step": 10132 + }, + { + "epoch": 0.6801114056575283, + "grad_norm": 1.069441318511963, + "learning_rate": 7.683683918579165e-05, + "loss": 0.5702, + "step": 10134 + }, + { + "epoch": 0.6802456293412973, + "grad_norm": 1.5402475595474243, + "learning_rate": 7.682766842114622e-05, + "loss": 0.6325, + "step": 10136 + }, + { + "epoch": 0.6803798530250663, + "grad_norm": 0.9602324366569519, + "learning_rate": 7.681849638894623e-05, + "loss": 0.5932, + "step": 10138 + }, + { + "epoch": 0.6805140767088352, + "grad_norm": 1.008661150932312, + "learning_rate": 7.6809323089625e-05, + "loss": 0.63, + "step": 10140 + }, + { + "epoch": 0.6806483003926043, + "grad_norm": 1.1476720571517944, + "learning_rate": 7.680014852361598e-05, + "loss": 0.5783, + "step": 10142 + }, + { + "epoch": 0.6807825240763733, + "grad_norm": 1.0170818567276, + "learning_rate": 7.679097269135266e-05, + "loss": 0.6574, + "step": 10144 + }, + { + "epoch": 0.6809167477601423, + "grad_norm": 1.764005184173584, + "learning_rate": 7.678179559326855e-05, + "loss": 0.6232, + "step": 10146 + }, + { + "epoch": 0.6810509714439112, + "grad_norm": 1.8771541118621826, + "learning_rate": 7.677261722979725e-05, + "loss": 0.5922, + "step": 10148 + }, + { + "epoch": 0.6811851951276803, + "grad_norm": 1.4425020217895508, + "learning_rate": 7.676343760137244e-05, + "loss": 0.62, + "step": 10150 + }, + { + "epoch": 0.6813194188114493, + "grad_norm": 0.840770423412323, + "learning_rate": 7.675425670842783e-05, + "loss": 0.5946, + "step": 10152 + }, + { + "epoch": 0.6814536424952182, + "grad_norm": 1.0578558444976807, + "learning_rate": 7.674507455139721e-05, + "loss": 0.6644, + "step": 10154 + }, + { + "epoch": 0.6815878661789873, + "grad_norm": 1.0931966304779053, + "learning_rate": 7.673589113071442e-05, + "loss": 0.6275, + "step": 10156 + }, + { + "epoch": 0.6817220898627563, + "grad_norm": 1.1474859714508057, + "learning_rate": 7.672670644681332e-05, + "loss": 0.5998, + "step": 10158 + }, + { + "epoch": 0.6818563135465253, + "grad_norm": 0.9686486124992371, + "learning_rate": 7.671752050012792e-05, + "loss": 0.5811, + "step": 10160 + }, + { + "epoch": 0.6819905372302942, + "grad_norm": 1.1089202165603638, + "learning_rate": 7.670833329109219e-05, + "loss": 0.6239, + "step": 10162 + }, + { + "epoch": 0.6821247609140633, + "grad_norm": 1.8264477252960205, + "learning_rate": 7.669914482014025e-05, + "loss": 0.6387, + "step": 10164 + }, + { + "epoch": 0.6822589845978323, + "grad_norm": 1.0332576036453247, + "learning_rate": 7.668995508770621e-05, + "loss": 0.6105, + "step": 10166 + }, + { + "epoch": 0.6823932082816013, + "grad_norm": 1.0819076299667358, + "learning_rate": 7.668076409422427e-05, + "loss": 0.6164, + "step": 10168 + }, + { + "epoch": 0.6825274319653702, + "grad_norm": 1.1721596717834473, + "learning_rate": 7.667157184012871e-05, + "loss": 0.6715, + "step": 10170 + }, + { + "epoch": 0.6826616556491393, + "grad_norm": 1.121983528137207, + "learning_rate": 7.666237832585382e-05, + "loss": 0.5908, + "step": 10172 + }, + { + "epoch": 0.6827958793329083, + "grad_norm": 0.9286911487579346, + "learning_rate": 7.665318355183398e-05, + "loss": 0.5408, + "step": 10174 + }, + { + "epoch": 0.6829301030166773, + "grad_norm": 1.180193305015564, + "learning_rate": 7.664398751850363e-05, + "loss": 0.6882, + "step": 10176 + }, + { + "epoch": 0.6830643267004463, + "grad_norm": 1.121951699256897, + "learning_rate": 7.663479022629727e-05, + "loss": 0.6304, + "step": 10178 + }, + { + "epoch": 0.6831985503842153, + "grad_norm": 1.0333795547485352, + "learning_rate": 7.662559167564944e-05, + "loss": 0.6446, + "step": 10180 + }, + { + "epoch": 0.6833327740679843, + "grad_norm": 0.9968959093093872, + "learning_rate": 7.661639186699474e-05, + "loss": 0.5881, + "step": 10182 + }, + { + "epoch": 0.6834669977517533, + "grad_norm": 0.9462453126907349, + "learning_rate": 7.66071908007679e-05, + "loss": 0.5612, + "step": 10184 + }, + { + "epoch": 0.6836012214355223, + "grad_norm": 1.2085262537002563, + "learning_rate": 7.659798847740359e-05, + "loss": 0.6888, + "step": 10186 + }, + { + "epoch": 0.6837354451192913, + "grad_norm": 1.0166971683502197, + "learning_rate": 7.658878489733664e-05, + "loss": 0.6123, + "step": 10188 + }, + { + "epoch": 0.6838696688030603, + "grad_norm": 1.029884934425354, + "learning_rate": 7.657958006100188e-05, + "loss": 0.6025, + "step": 10190 + }, + { + "epoch": 0.6840038924868292, + "grad_norm": 0.9909621477127075, + "learning_rate": 7.657037396883423e-05, + "loss": 0.6118, + "step": 10192 + }, + { + "epoch": 0.6841381161705983, + "grad_norm": 1.0653865337371826, + "learning_rate": 7.656116662126866e-05, + "loss": 0.6363, + "step": 10194 + }, + { + "epoch": 0.6842723398543673, + "grad_norm": 1.0455743074417114, + "learning_rate": 7.65519580187402e-05, + "loss": 0.5604, + "step": 10196 + }, + { + "epoch": 0.6844065635381363, + "grad_norm": 0.9630902409553528, + "learning_rate": 7.654274816168396e-05, + "loss": 0.5414, + "step": 10198 + }, + { + "epoch": 0.6845407872219053, + "grad_norm": 1.1408867835998535, + "learning_rate": 7.653353705053503e-05, + "loss": 0.5913, + "step": 10200 + }, + { + "epoch": 0.6846750109056743, + "grad_norm": 1.0438798666000366, + "learning_rate": 7.652432468572865e-05, + "loss": 0.5727, + "step": 10202 + }, + { + "epoch": 0.6848092345894433, + "grad_norm": 2.1974589824676514, + "learning_rate": 7.65151110677001e-05, + "loss": 0.5832, + "step": 10204 + }, + { + "epoch": 0.6849434582732123, + "grad_norm": 0.9814490079879761, + "learning_rate": 7.650589619688469e-05, + "loss": 0.6232, + "step": 10206 + }, + { + "epoch": 0.6850776819569813, + "grad_norm": 0.9011107683181763, + "learning_rate": 7.64966800737178e-05, + "loss": 0.5893, + "step": 10208 + }, + { + "epoch": 0.6852119056407503, + "grad_norm": 1.0110008716583252, + "learning_rate": 7.648746269863487e-05, + "loss": 0.583, + "step": 10210 + }, + { + "epoch": 0.6853461293245193, + "grad_norm": 1.6427772045135498, + "learning_rate": 7.647824407207144e-05, + "loss": 0.6113, + "step": 10212 + }, + { + "epoch": 0.6854803530082884, + "grad_norm": 0.9514898657798767, + "learning_rate": 7.646902419446302e-05, + "loss": 0.5717, + "step": 10214 + }, + { + "epoch": 0.6856145766920573, + "grad_norm": 0.8733771443367004, + "learning_rate": 7.645980306624528e-05, + "loss": 0.6078, + "step": 10216 + }, + { + "epoch": 0.6857488003758263, + "grad_norm": 0.9661299586296082, + "learning_rate": 7.645058068785386e-05, + "loss": 0.6116, + "step": 10218 + }, + { + "epoch": 0.6858830240595953, + "grad_norm": 1.0476787090301514, + "learning_rate": 7.64413570597245e-05, + "loss": 0.6379, + "step": 10220 + }, + { + "epoch": 0.6860172477433644, + "grad_norm": 1.0753364562988281, + "learning_rate": 7.643213218229305e-05, + "loss": 0.6188, + "step": 10222 + }, + { + "epoch": 0.6861514714271333, + "grad_norm": 1.123155951499939, + "learning_rate": 7.642290605599531e-05, + "loss": 0.6564, + "step": 10224 + }, + { + "epoch": 0.6862856951109023, + "grad_norm": 1.0590447187423706, + "learning_rate": 7.64136786812672e-05, + "loss": 0.5836, + "step": 10226 + }, + { + "epoch": 0.6864199187946713, + "grad_norm": 0.9150590896606445, + "learning_rate": 7.640445005854475e-05, + "loss": 0.627, + "step": 10228 + }, + { + "epoch": 0.6865541424784403, + "grad_norm": 0.9836304783821106, + "learning_rate": 7.639522018826393e-05, + "loss": 0.6949, + "step": 10230 + }, + { + "epoch": 0.6866883661622093, + "grad_norm": 1.3256171941757202, + "learning_rate": 7.638598907086088e-05, + "loss": 0.6433, + "step": 10232 + }, + { + "epoch": 0.6868225898459783, + "grad_norm": 1.0061068534851074, + "learning_rate": 7.637675670677171e-05, + "loss": 0.6485, + "step": 10234 + }, + { + "epoch": 0.6869568135297474, + "grad_norm": 0.9491173624992371, + "learning_rate": 7.636752309643267e-05, + "loss": 0.6128, + "step": 10236 + }, + { + "epoch": 0.6870910372135163, + "grad_norm": 0.9607385993003845, + "learning_rate": 7.635828824028001e-05, + "loss": 0.6076, + "step": 10238 + }, + { + "epoch": 0.6872252608972853, + "grad_norm": 1.398555040359497, + "learning_rate": 7.634905213875006e-05, + "loss": 0.5525, + "step": 10240 + }, + { + "epoch": 0.6873594845810543, + "grad_norm": 1.0753309726715088, + "learning_rate": 7.633981479227921e-05, + "loss": 0.6379, + "step": 10242 + }, + { + "epoch": 0.6874937082648234, + "grad_norm": 1.0636236667633057, + "learning_rate": 7.633057620130391e-05, + "loss": 0.6578, + "step": 10244 + }, + { + "epoch": 0.6876279319485923, + "grad_norm": 0.9041749238967896, + "learning_rate": 7.632133636626064e-05, + "loss": 0.5808, + "step": 10246 + }, + { + "epoch": 0.6877621556323613, + "grad_norm": 0.9568422436714172, + "learning_rate": 7.6312095287586e-05, + "loss": 0.6274, + "step": 10248 + }, + { + "epoch": 0.6878963793161303, + "grad_norm": 1.0265642404556274, + "learning_rate": 7.630285296571661e-05, + "loss": 0.6536, + "step": 10250 + }, + { + "epoch": 0.6880306029998994, + "grad_norm": 0.9597381353378296, + "learning_rate": 7.629360940108913e-05, + "loss": 0.5919, + "step": 10252 + }, + { + "epoch": 0.6881648266836683, + "grad_norm": 1.0445297956466675, + "learning_rate": 7.62843645941403e-05, + "loss": 0.5497, + "step": 10254 + }, + { + "epoch": 0.6882990503674373, + "grad_norm": 1.0023874044418335, + "learning_rate": 7.627511854530695e-05, + "loss": 0.6252, + "step": 10256 + }, + { + "epoch": 0.6884332740512064, + "grad_norm": 1.0493043661117554, + "learning_rate": 7.62658712550259e-05, + "loss": 0.5709, + "step": 10258 + }, + { + "epoch": 0.6885674977349754, + "grad_norm": 0.9132547378540039, + "learning_rate": 7.625662272373409e-05, + "loss": 0.5709, + "step": 10260 + }, + { + "epoch": 0.6887017214187443, + "grad_norm": 1.17551851272583, + "learning_rate": 7.624737295186849e-05, + "loss": 0.5738, + "step": 10262 + }, + { + "epoch": 0.6888359451025133, + "grad_norm": 1.090474009513855, + "learning_rate": 7.623812193986612e-05, + "loss": 0.647, + "step": 10264 + }, + { + "epoch": 0.6889701687862824, + "grad_norm": 0.9255596995353699, + "learning_rate": 7.62288696881641e-05, + "loss": 0.5732, + "step": 10266 + }, + { + "epoch": 0.6891043924700513, + "grad_norm": 1.0065983533859253, + "learning_rate": 7.621961619719954e-05, + "loss": 0.615, + "step": 10268 + }, + { + "epoch": 0.6892386161538203, + "grad_norm": 2.6161906719207764, + "learning_rate": 7.621036146740972e-05, + "loss": 0.5857, + "step": 10270 + }, + { + "epoch": 0.6893728398375893, + "grad_norm": 1.048517107963562, + "learning_rate": 7.620110549923181e-05, + "loss": 0.6099, + "step": 10272 + }, + { + "epoch": 0.6895070635213584, + "grad_norm": 0.9987454414367676, + "learning_rate": 7.619184829310322e-05, + "loss": 0.6543, + "step": 10274 + }, + { + "epoch": 0.6896412872051273, + "grad_norm": 0.9380282163619995, + "learning_rate": 7.618258984946128e-05, + "loss": 0.5941, + "step": 10276 + }, + { + "epoch": 0.6897755108888963, + "grad_norm": 1.1077454090118408, + "learning_rate": 7.617333016874346e-05, + "loss": 0.6435, + "step": 10278 + }, + { + "epoch": 0.6899097345726654, + "grad_norm": 1.1259806156158447, + "learning_rate": 7.616406925138727e-05, + "loss": 0.6018, + "step": 10280 + }, + { + "epoch": 0.6900439582564344, + "grad_norm": 0.9674026966094971, + "learning_rate": 7.615480709783025e-05, + "loss": 0.5838, + "step": 10282 + }, + { + "epoch": 0.6901781819402033, + "grad_norm": 1.1969586610794067, + "learning_rate": 7.614554370851003e-05, + "loss": 0.6811, + "step": 10284 + }, + { + "epoch": 0.6903124056239723, + "grad_norm": 0.9744465947151184, + "learning_rate": 7.613627908386427e-05, + "loss": 0.6712, + "step": 10286 + }, + { + "epoch": 0.6904466293077414, + "grad_norm": 0.9148718118667603, + "learning_rate": 7.612701322433074e-05, + "loss": 0.5742, + "step": 10288 + }, + { + "epoch": 0.6905808529915104, + "grad_norm": 0.9658867716789246, + "learning_rate": 7.61177461303472e-05, + "loss": 0.627, + "step": 10290 + }, + { + "epoch": 0.6907150766752793, + "grad_norm": 1.0178569555282593, + "learning_rate": 7.610847780235149e-05, + "loss": 0.5996, + "step": 10292 + }, + { + "epoch": 0.6908493003590483, + "grad_norm": 1.414971947669983, + "learning_rate": 7.609920824078157e-05, + "loss": 0.5695, + "step": 10294 + }, + { + "epoch": 0.6909835240428174, + "grad_norm": 1.0085248947143555, + "learning_rate": 7.608993744607538e-05, + "loss": 0.6539, + "step": 10296 + }, + { + "epoch": 0.6911177477265864, + "grad_norm": 1.0939044952392578, + "learning_rate": 7.608066541867093e-05, + "loss": 0.5884, + "step": 10298 + }, + { + "epoch": 0.6912519714103553, + "grad_norm": 1.0116087198257446, + "learning_rate": 7.607139215900636e-05, + "loss": 0.6157, + "step": 10300 + }, + { + "epoch": 0.6913861950941244, + "grad_norm": 0.9540835022926331, + "learning_rate": 7.606211766751976e-05, + "loss": 0.6214, + "step": 10302 + }, + { + "epoch": 0.6915204187778934, + "grad_norm": 1.0693719387054443, + "learning_rate": 7.605284194464934e-05, + "loss": 0.6726, + "step": 10304 + }, + { + "epoch": 0.6916546424616623, + "grad_norm": 1.0814954042434692, + "learning_rate": 7.604356499083338e-05, + "loss": 0.6213, + "step": 10306 + }, + { + "epoch": 0.6917888661454313, + "grad_norm": 1.1147676706314087, + "learning_rate": 7.603428680651019e-05, + "loss": 0.6544, + "step": 10308 + }, + { + "epoch": 0.6919230898292004, + "grad_norm": 1.0701748132705688, + "learning_rate": 7.602500739211813e-05, + "loss": 0.6196, + "step": 10310 + }, + { + "epoch": 0.6920573135129694, + "grad_norm": 1.0478525161743164, + "learning_rate": 7.601572674809565e-05, + "loss": 0.57, + "step": 10312 + }, + { + "epoch": 0.6921915371967383, + "grad_norm": 1.0839003324508667, + "learning_rate": 7.600644487488124e-05, + "loss": 0.6296, + "step": 10314 + }, + { + "epoch": 0.6923257608805073, + "grad_norm": 1.0307375192642212, + "learning_rate": 7.599716177291345e-05, + "loss": 0.5879, + "step": 10316 + }, + { + "epoch": 0.6924599845642764, + "grad_norm": 1.3292783498764038, + "learning_rate": 7.59878774426309e-05, + "loss": 0.5892, + "step": 10318 + }, + { + "epoch": 0.6925942082480454, + "grad_norm": 1.1004843711853027, + "learning_rate": 7.597859188447223e-05, + "loss": 0.6271, + "step": 10320 + }, + { + "epoch": 0.6927284319318143, + "grad_norm": 0.9349532723426819, + "learning_rate": 7.596930509887618e-05, + "loss": 0.5415, + "step": 10322 + }, + { + "epoch": 0.6928626556155834, + "grad_norm": 1.0800589323043823, + "learning_rate": 7.596001708628152e-05, + "loss": 0.645, + "step": 10324 + }, + { + "epoch": 0.6929968792993524, + "grad_norm": 1.141237497329712, + "learning_rate": 7.595072784712711e-05, + "loss": 0.6522, + "step": 10326 + }, + { + "epoch": 0.6931311029831214, + "grad_norm": 1.1993459463119507, + "learning_rate": 7.594143738185184e-05, + "loss": 0.6175, + "step": 10328 + }, + { + "epoch": 0.6932653266668903, + "grad_norm": 1.0714892148971558, + "learning_rate": 7.593214569089467e-05, + "loss": 0.6167, + "step": 10330 + }, + { + "epoch": 0.6933995503506594, + "grad_norm": 1.1327598094940186, + "learning_rate": 7.59228527746946e-05, + "loss": 0.6299, + "step": 10332 + }, + { + "epoch": 0.6935337740344284, + "grad_norm": 1.010339379310608, + "learning_rate": 7.591355863369071e-05, + "loss": 0.6034, + "step": 10334 + }, + { + "epoch": 0.6936679977181974, + "grad_norm": 1.156844139099121, + "learning_rate": 7.590426326832217e-05, + "loss": 0.6784, + "step": 10336 + }, + { + "epoch": 0.6938022214019663, + "grad_norm": 0.9312037229537964, + "learning_rate": 7.589496667902809e-05, + "loss": 0.561, + "step": 10338 + }, + { + "epoch": 0.6939364450857354, + "grad_norm": 1.0371742248535156, + "learning_rate": 7.588566886624774e-05, + "loss": 0.6169, + "step": 10340 + }, + { + "epoch": 0.6940706687695044, + "grad_norm": 1.2180014848709106, + "learning_rate": 7.587636983042048e-05, + "loss": 0.6081, + "step": 10342 + }, + { + "epoch": 0.6942048924532733, + "grad_norm": 1.029098391532898, + "learning_rate": 7.586706957198562e-05, + "loss": 0.6158, + "step": 10344 + }, + { + "epoch": 0.6943391161370424, + "grad_norm": 1.1107252836227417, + "learning_rate": 7.585776809138257e-05, + "loss": 0.6046, + "step": 10346 + }, + { + "epoch": 0.6944733398208114, + "grad_norm": 1.2475100755691528, + "learning_rate": 7.584846538905083e-05, + "loss": 0.6487, + "step": 10348 + }, + { + "epoch": 0.6946075635045804, + "grad_norm": 1.0322431325912476, + "learning_rate": 7.583916146542992e-05, + "loss": 0.649, + "step": 10350 + }, + { + "epoch": 0.6947417871883493, + "grad_norm": 0.919420599937439, + "learning_rate": 7.582985632095946e-05, + "loss": 0.5707, + "step": 10352 + }, + { + "epoch": 0.6948760108721184, + "grad_norm": 1.2439117431640625, + "learning_rate": 7.582054995607908e-05, + "loss": 0.6862, + "step": 10354 + }, + { + "epoch": 0.6950102345558874, + "grad_norm": 1.0442447662353516, + "learning_rate": 7.581124237122848e-05, + "loss": 0.5455, + "step": 10356 + }, + { + "epoch": 0.6951444582396564, + "grad_norm": 0.994797945022583, + "learning_rate": 7.580193356684743e-05, + "loss": 0.5781, + "step": 10358 + }, + { + "epoch": 0.6952786819234253, + "grad_norm": 1.0670169591903687, + "learning_rate": 7.579262354337577e-05, + "loss": 0.5641, + "step": 10360 + }, + { + "epoch": 0.6954129056071944, + "grad_norm": 2.9009644985198975, + "learning_rate": 7.578331230125336e-05, + "loss": 0.6412, + "step": 10362 + }, + { + "epoch": 0.6955471292909634, + "grad_norm": 0.9840713739395142, + "learning_rate": 7.577399984092015e-05, + "loss": 0.5508, + "step": 10364 + }, + { + "epoch": 0.6956813529747324, + "grad_norm": 1.0158319473266602, + "learning_rate": 7.576468616281612e-05, + "loss": 0.5954, + "step": 10366 + }, + { + "epoch": 0.6958155766585014, + "grad_norm": 0.9923294186592102, + "learning_rate": 7.575537126738132e-05, + "loss": 0.5792, + "step": 10368 + }, + { + "epoch": 0.6959498003422704, + "grad_norm": 0.9123964905738831, + "learning_rate": 7.57460551550559e-05, + "loss": 0.5705, + "step": 10370 + }, + { + "epoch": 0.6960840240260394, + "grad_norm": 1.0065155029296875, + "learning_rate": 7.573673782628e-05, + "loss": 0.5463, + "step": 10372 + }, + { + "epoch": 0.6962182477098084, + "grad_norm": 1.0273391008377075, + "learning_rate": 7.572741928149384e-05, + "loss": 0.5893, + "step": 10374 + }, + { + "epoch": 0.6963524713935774, + "grad_norm": 1.121673583984375, + "learning_rate": 7.57180995211377e-05, + "loss": 0.5428, + "step": 10376 + }, + { + "epoch": 0.6964866950773464, + "grad_norm": 1.0862985849380493, + "learning_rate": 7.570877854565195e-05, + "loss": 0.5782, + "step": 10378 + }, + { + "epoch": 0.6966209187611154, + "grad_norm": 1.1917251348495483, + "learning_rate": 7.569945635547695e-05, + "loss": 0.6539, + "step": 10380 + }, + { + "epoch": 0.6967551424448843, + "grad_norm": 0.9577513933181763, + "learning_rate": 7.569013295105318e-05, + "loss": 0.5804, + "step": 10382 + }, + { + "epoch": 0.6968893661286534, + "grad_norm": 1.0312405824661255, + "learning_rate": 7.568080833282114e-05, + "loss": 0.6378, + "step": 10384 + }, + { + "epoch": 0.6970235898124224, + "grad_norm": 2.1493964195251465, + "learning_rate": 7.56714825012214e-05, + "loss": 0.543, + "step": 10386 + }, + { + "epoch": 0.6971578134961914, + "grad_norm": 1.163265347480774, + "learning_rate": 7.566215545669462e-05, + "loss": 0.6308, + "step": 10388 + }, + { + "epoch": 0.6972920371799604, + "grad_norm": 0.9730448722839355, + "learning_rate": 7.565282719968143e-05, + "loss": 0.5708, + "step": 10390 + }, + { + "epoch": 0.6974262608637294, + "grad_norm": 0.9758681058883667, + "learning_rate": 7.564349773062262e-05, + "loss": 0.5878, + "step": 10392 + }, + { + "epoch": 0.6975604845474984, + "grad_norm": 1.1344770193099976, + "learning_rate": 7.563416704995894e-05, + "loss": 0.6225, + "step": 10394 + }, + { + "epoch": 0.6976947082312674, + "grad_norm": 1.0707207918167114, + "learning_rate": 7.56248351581313e-05, + "loss": 0.5813, + "step": 10396 + }, + { + "epoch": 0.6978289319150364, + "grad_norm": 0.972823977470398, + "learning_rate": 7.561550205558058e-05, + "loss": 0.5609, + "step": 10398 + }, + { + "epoch": 0.6979631555988054, + "grad_norm": 0.9138381481170654, + "learning_rate": 7.560616774274775e-05, + "loss": 0.5109, + "step": 10400 + }, + { + "epoch": 0.6980973792825744, + "grad_norm": 1.0473132133483887, + "learning_rate": 7.559683222007386e-05, + "loss": 0.6346, + "step": 10402 + }, + { + "epoch": 0.6982316029663435, + "grad_norm": 1.0219554901123047, + "learning_rate": 7.558749548799997e-05, + "loss": 0.5836, + "step": 10404 + }, + { + "epoch": 0.6983658266501124, + "grad_norm": 1.2376022338867188, + "learning_rate": 7.557815754696724e-05, + "loss": 0.6474, + "step": 10406 + }, + { + "epoch": 0.6985000503338814, + "grad_norm": 1.0176584720611572, + "learning_rate": 7.556881839741687e-05, + "loss": 0.5756, + "step": 10408 + }, + { + "epoch": 0.6986342740176504, + "grad_norm": 1.1369335651397705, + "learning_rate": 7.55594780397901e-05, + "loss": 0.6256, + "step": 10410 + }, + { + "epoch": 0.6987684977014195, + "grad_norm": 0.9506736397743225, + "learning_rate": 7.555013647452826e-05, + "loss": 0.5925, + "step": 10412 + }, + { + "epoch": 0.6989027213851884, + "grad_norm": 0.9491580724716187, + "learning_rate": 7.554079370207269e-05, + "loss": 0.5816, + "step": 10414 + }, + { + "epoch": 0.6990369450689574, + "grad_norm": 1.0840520858764648, + "learning_rate": 7.553144972286488e-05, + "loss": 0.5691, + "step": 10416 + }, + { + "epoch": 0.6991711687527264, + "grad_norm": 1.1048444509506226, + "learning_rate": 7.552210453734625e-05, + "loss": 0.5801, + "step": 10418 + }, + { + "epoch": 0.6993053924364954, + "grad_norm": 1.0872331857681274, + "learning_rate": 7.551275814595837e-05, + "loss": 0.6006, + "step": 10420 + }, + { + "epoch": 0.6994396161202644, + "grad_norm": 1.0584237575531006, + "learning_rate": 7.550341054914284e-05, + "loss": 0.6611, + "step": 10422 + }, + { + "epoch": 0.6995738398040334, + "grad_norm": 1.0477406978607178, + "learning_rate": 7.549406174734132e-05, + "loss": 0.5806, + "step": 10424 + }, + { + "epoch": 0.6997080634878025, + "grad_norm": 1.157261848449707, + "learning_rate": 7.548471174099551e-05, + "loss": 0.5282, + "step": 10426 + }, + { + "epoch": 0.6998422871715714, + "grad_norm": 1.0036718845367432, + "learning_rate": 7.547536053054718e-05, + "loss": 0.606, + "step": 10428 + }, + { + "epoch": 0.6999765108553404, + "grad_norm": 1.0837408304214478, + "learning_rate": 7.546600811643816e-05, + "loss": 0.5896, + "step": 10430 + }, + { + "epoch": 0.7001107345391094, + "grad_norm": 1.0489901304244995, + "learning_rate": 7.545665449911032e-05, + "loss": 0.6062, + "step": 10432 + }, + { + "epoch": 0.7002449582228785, + "grad_norm": 1.1422386169433594, + "learning_rate": 7.544729967900563e-05, + "loss": 0.5789, + "step": 10434 + }, + { + "epoch": 0.7003791819066474, + "grad_norm": 1.064909815788269, + "learning_rate": 7.543794365656609e-05, + "loss": 0.6285, + "step": 10436 + }, + { + "epoch": 0.7005134055904164, + "grad_norm": 1.1166802644729614, + "learning_rate": 7.542858643223369e-05, + "loss": 0.5959, + "step": 10438 + }, + { + "epoch": 0.7006476292741854, + "grad_norm": 1.0952495336532593, + "learning_rate": 7.541922800645061e-05, + "loss": 0.6267, + "step": 10440 + }, + { + "epoch": 0.7007818529579545, + "grad_norm": 1.1388980150222778, + "learning_rate": 7.540986837965899e-05, + "loss": 0.5997, + "step": 10442 + }, + { + "epoch": 0.7009160766417234, + "grad_norm": 1.0324063301086426, + "learning_rate": 7.540050755230104e-05, + "loss": 0.5926, + "step": 10444 + }, + { + "epoch": 0.7010503003254924, + "grad_norm": 1.14800226688385, + "learning_rate": 7.539114552481908e-05, + "loss": 0.5649, + "step": 10446 + }, + { + "epoch": 0.7011845240092615, + "grad_norm": 1.0952218770980835, + "learning_rate": 7.53817822976554e-05, + "loss": 0.6183, + "step": 10448 + }, + { + "epoch": 0.7013187476930305, + "grad_norm": 1.2546967267990112, + "learning_rate": 7.537241787125245e-05, + "loss": 0.6727, + "step": 10450 + }, + { + "epoch": 0.7014529713767994, + "grad_norm": 1.0081515312194824, + "learning_rate": 7.536305224605261e-05, + "loss": 0.6433, + "step": 10452 + }, + { + "epoch": 0.7015871950605684, + "grad_norm": 1.042733907699585, + "learning_rate": 7.535368542249846e-05, + "loss": 0.6368, + "step": 10454 + }, + { + "epoch": 0.7017214187443375, + "grad_norm": 1.0268605947494507, + "learning_rate": 7.534431740103249e-05, + "loss": 0.5982, + "step": 10456 + }, + { + "epoch": 0.7018556424281064, + "grad_norm": 0.9910513162612915, + "learning_rate": 7.53349481820974e-05, + "loss": 0.5713, + "step": 10458 + }, + { + "epoch": 0.7019898661118754, + "grad_norm": 1.038655161857605, + "learning_rate": 7.53255777661358e-05, + "loss": 0.5926, + "step": 10460 + }, + { + "epoch": 0.7021240897956444, + "grad_norm": 1.022022008895874, + "learning_rate": 7.531620615359046e-05, + "loss": 0.6858, + "step": 10462 + }, + { + "epoch": 0.7022583134794135, + "grad_norm": 1.5238577127456665, + "learning_rate": 7.530683334490416e-05, + "loss": 0.5713, + "step": 10464 + }, + { + "epoch": 0.7023925371631824, + "grad_norm": 1.0431960821151733, + "learning_rate": 7.529745934051976e-05, + "loss": 0.6103, + "step": 10466 + }, + { + "epoch": 0.7025267608469514, + "grad_norm": 1.0651239156723022, + "learning_rate": 7.528808414088015e-05, + "loss": 0.5874, + "step": 10468 + }, + { + "epoch": 0.7026609845307205, + "grad_norm": 1.1000802516937256, + "learning_rate": 7.527870774642828e-05, + "loss": 0.5662, + "step": 10470 + }, + { + "epoch": 0.7027952082144895, + "grad_norm": 1.0806845426559448, + "learning_rate": 7.526933015760717e-05, + "loss": 0.5968, + "step": 10472 + }, + { + "epoch": 0.7029294318982584, + "grad_norm": 1.2063080072402954, + "learning_rate": 7.525995137485993e-05, + "loss": 0.6168, + "step": 10474 + }, + { + "epoch": 0.7030636555820274, + "grad_norm": 1.074565052986145, + "learning_rate": 7.525057139862964e-05, + "loss": 0.6239, + "step": 10476 + }, + { + "epoch": 0.7031978792657965, + "grad_norm": 1.032949686050415, + "learning_rate": 7.524119022935949e-05, + "loss": 0.525, + "step": 10478 + }, + { + "epoch": 0.7033321029495655, + "grad_norm": 1.0446885824203491, + "learning_rate": 7.523180786749276e-05, + "loss": 0.5857, + "step": 10480 + }, + { + "epoch": 0.7034663266333344, + "grad_norm": 1.0869170427322388, + "learning_rate": 7.522242431347272e-05, + "loss": 0.5863, + "step": 10482 + }, + { + "epoch": 0.7036005503171034, + "grad_norm": 1.000910758972168, + "learning_rate": 7.521303956774273e-05, + "loss": 0.5666, + "step": 10484 + }, + { + "epoch": 0.7037347740008725, + "grad_norm": 1.1870709657669067, + "learning_rate": 7.52036536307462e-05, + "loss": 0.6281, + "step": 10486 + }, + { + "epoch": 0.7038689976846415, + "grad_norm": 1.1357227563858032, + "learning_rate": 7.519426650292658e-05, + "loss": 0.645, + "step": 10488 + }, + { + "epoch": 0.7040032213684104, + "grad_norm": 0.9073771834373474, + "learning_rate": 7.518487818472743e-05, + "loss": 0.6408, + "step": 10490 + }, + { + "epoch": 0.7041374450521795, + "grad_norm": 1.2567945718765259, + "learning_rate": 7.51754886765923e-05, + "loss": 0.6292, + "step": 10492 + }, + { + "epoch": 0.7042716687359485, + "grad_norm": 1.2820225954055786, + "learning_rate": 7.516609797896483e-05, + "loss": 0.6182, + "step": 10494 + }, + { + "epoch": 0.7044058924197174, + "grad_norm": 1.0083980560302734, + "learning_rate": 7.515670609228873e-05, + "loss": 0.5833, + "step": 10496 + }, + { + "epoch": 0.7045401161034864, + "grad_norm": 0.9857763648033142, + "learning_rate": 7.514731301700773e-05, + "loss": 0.6576, + "step": 10498 + }, + { + "epoch": 0.7046743397872555, + "grad_norm": 1.4885666370391846, + "learning_rate": 7.513791875356564e-05, + "loss": 0.6317, + "step": 10500 + }, + { + "epoch": 0.7048085634710245, + "grad_norm": 1.116719365119934, + "learning_rate": 7.512852330240632e-05, + "loss": 0.6104, + "step": 10502 + }, + { + "epoch": 0.7049427871547934, + "grad_norm": 0.9923556447029114, + "learning_rate": 7.51191266639737e-05, + "loss": 0.6866, + "step": 10504 + }, + { + "epoch": 0.7050770108385624, + "grad_norm": 1.264246940612793, + "learning_rate": 7.51097288387117e-05, + "loss": 0.564, + "step": 10506 + }, + { + "epoch": 0.7052112345223315, + "grad_norm": 1.0330630540847778, + "learning_rate": 7.510032982706443e-05, + "loss": 0.6315, + "step": 10508 + }, + { + "epoch": 0.7053454582061005, + "grad_norm": 1.1161117553710938, + "learning_rate": 7.509092962947591e-05, + "loss": 0.6257, + "step": 10510 + }, + { + "epoch": 0.7054796818898694, + "grad_norm": 1.8677253723144531, + "learning_rate": 7.508152824639032e-05, + "loss": 0.6676, + "step": 10512 + }, + { + "epoch": 0.7056139055736385, + "grad_norm": 1.144211769104004, + "learning_rate": 7.507212567825184e-05, + "loss": 0.6177, + "step": 10514 + }, + { + "epoch": 0.7057481292574075, + "grad_norm": 0.9658868312835693, + "learning_rate": 7.506272192550472e-05, + "loss": 0.6054, + "step": 10516 + }, + { + "epoch": 0.7058823529411765, + "grad_norm": 4.2473978996276855, + "learning_rate": 7.505331698859325e-05, + "loss": 0.5925, + "step": 10518 + }, + { + "epoch": 0.7060165766249454, + "grad_norm": 1.01564621925354, + "learning_rate": 7.504391086796186e-05, + "loss": 0.6168, + "step": 10520 + }, + { + "epoch": 0.7061508003087145, + "grad_norm": 1.3537542819976807, + "learning_rate": 7.50345035640549e-05, + "loss": 0.6144, + "step": 10522 + }, + { + "epoch": 0.7062850239924835, + "grad_norm": 0.8762102723121643, + "learning_rate": 7.502509507731688e-05, + "loss": 0.5816, + "step": 10524 + }, + { + "epoch": 0.7064192476762525, + "grad_norm": 0.9736351370811462, + "learning_rate": 7.501568540819233e-05, + "loss": 0.6564, + "step": 10526 + }, + { + "epoch": 0.7065534713600214, + "grad_norm": 1.4718120098114014, + "learning_rate": 7.500627455712583e-05, + "loss": 0.6406, + "step": 10528 + }, + { + "epoch": 0.7066876950437905, + "grad_norm": 0.9880357384681702, + "learning_rate": 7.499686252456205e-05, + "loss": 0.5654, + "step": 10530 + }, + { + "epoch": 0.7068219187275595, + "grad_norm": 1.083616018295288, + "learning_rate": 7.498744931094565e-05, + "loss": 0.6184, + "step": 10532 + }, + { + "epoch": 0.7069561424113284, + "grad_norm": 1.0512688159942627, + "learning_rate": 7.497803491672141e-05, + "loss": 0.6053, + "step": 10534 + }, + { + "epoch": 0.7070903660950975, + "grad_norm": 1.0200892686843872, + "learning_rate": 7.496861934233414e-05, + "loss": 0.625, + "step": 10536 + }, + { + "epoch": 0.7072245897788665, + "grad_norm": 1.6291522979736328, + "learning_rate": 7.495920258822869e-05, + "loss": 0.6425, + "step": 10538 + }, + { + "epoch": 0.7073588134626355, + "grad_norm": 1.2425224781036377, + "learning_rate": 7.494978465485002e-05, + "loss": 0.6697, + "step": 10540 + }, + { + "epoch": 0.7074930371464044, + "grad_norm": 0.95583575963974, + "learning_rate": 7.494036554264308e-05, + "loss": 0.5765, + "step": 10542 + }, + { + "epoch": 0.7076272608301735, + "grad_norm": 0.9750218987464905, + "learning_rate": 7.49309452520529e-05, + "loss": 0.593, + "step": 10544 + }, + { + "epoch": 0.7077614845139425, + "grad_norm": 1.0550893545150757, + "learning_rate": 7.49215237835246e-05, + "loss": 0.6519, + "step": 10546 + }, + { + "epoch": 0.7078957081977115, + "grad_norm": 0.9797727465629578, + "learning_rate": 7.49121011375033e-05, + "loss": 0.5769, + "step": 10548 + }, + { + "epoch": 0.7080299318814804, + "grad_norm": 3.812112808227539, + "learning_rate": 7.490267731443422e-05, + "loss": 0.5869, + "step": 10550 + }, + { + "epoch": 0.7081641555652495, + "grad_norm": 0.9804056286811829, + "learning_rate": 7.489325231476258e-05, + "loss": 0.5874, + "step": 10552 + }, + { + "epoch": 0.7082983792490185, + "grad_norm": 1.2232451438903809, + "learning_rate": 7.488382613893372e-05, + "loss": 0.6239, + "step": 10554 + }, + { + "epoch": 0.7084326029327875, + "grad_norm": 0.9250823259353638, + "learning_rate": 7.487439878739303e-05, + "loss": 0.6, + "step": 10556 + }, + { + "epoch": 0.7085668266165565, + "grad_norm": 0.9323468804359436, + "learning_rate": 7.486497026058588e-05, + "loss": 0.6055, + "step": 10558 + }, + { + "epoch": 0.7087010503003255, + "grad_norm": 0.9436549544334412, + "learning_rate": 7.485554055895781e-05, + "loss": 0.5185, + "step": 10560 + }, + { + "epoch": 0.7088352739840945, + "grad_norm": 1.084033489227295, + "learning_rate": 7.484610968295431e-05, + "loss": 0.6145, + "step": 10562 + }, + { + "epoch": 0.7089694976678635, + "grad_norm": 1.198091745376587, + "learning_rate": 7.4836677633021e-05, + "loss": 0.5858, + "step": 10564 + }, + { + "epoch": 0.7091037213516325, + "grad_norm": 1.241812825202942, + "learning_rate": 7.48272444096035e-05, + "loss": 0.5898, + "step": 10566 + }, + { + "epoch": 0.7092379450354015, + "grad_norm": 0.9129748344421387, + "learning_rate": 7.481781001314751e-05, + "loss": 0.6265, + "step": 10568 + }, + { + "epoch": 0.7093721687191705, + "grad_norm": 0.901710033416748, + "learning_rate": 7.480837444409882e-05, + "loss": 0.6173, + "step": 10570 + }, + { + "epoch": 0.7095063924029394, + "grad_norm": 1.212648630142212, + "learning_rate": 7.479893770290321e-05, + "loss": 0.6168, + "step": 10572 + }, + { + "epoch": 0.7096406160867085, + "grad_norm": 0.9216916561126709, + "learning_rate": 7.478949979000656e-05, + "loss": 0.5529, + "step": 10574 + }, + { + "epoch": 0.7097748397704775, + "grad_norm": 1.0843908786773682, + "learning_rate": 7.47800607058548e-05, + "loss": 0.5992, + "step": 10576 + }, + { + "epoch": 0.7099090634542465, + "grad_norm": 1.1257072687149048, + "learning_rate": 7.477062045089389e-05, + "loss": 0.639, + "step": 10578 + }, + { + "epoch": 0.7100432871380155, + "grad_norm": 1.2055573463439941, + "learning_rate": 7.476117902556987e-05, + "loss": 0.6759, + "step": 10580 + }, + { + "epoch": 0.7101775108217845, + "grad_norm": 1.157829999923706, + "learning_rate": 7.475173643032882e-05, + "loss": 0.6496, + "step": 10582 + }, + { + "epoch": 0.7103117345055535, + "grad_norm": 1.0404984951019287, + "learning_rate": 7.474229266561692e-05, + "loss": 0.6716, + "step": 10584 + }, + { + "epoch": 0.7104459581893225, + "grad_norm": 1.039047360420227, + "learning_rate": 7.473284773188034e-05, + "loss": 0.641, + "step": 10586 + }, + { + "epoch": 0.7105801818730915, + "grad_norm": 0.9575285315513611, + "learning_rate": 7.472340162956534e-05, + "loss": 0.6802, + "step": 10588 + }, + { + "epoch": 0.7107144055568605, + "grad_norm": 1.1343613862991333, + "learning_rate": 7.471395435911822e-05, + "loss": 0.5732, + "step": 10590 + }, + { + "epoch": 0.7108486292406295, + "grad_norm": 1.0703816413879395, + "learning_rate": 7.470450592098537e-05, + "loss": 0.5905, + "step": 10592 + }, + { + "epoch": 0.7109828529243986, + "grad_norm": 1.05869460105896, + "learning_rate": 7.469505631561317e-05, + "loss": 0.6367, + "step": 10594 + }, + { + "epoch": 0.7111170766081675, + "grad_norm": 1.1756197214126587, + "learning_rate": 7.468560554344814e-05, + "loss": 0.5681, + "step": 10596 + }, + { + "epoch": 0.7112513002919365, + "grad_norm": 0.9795808792114258, + "learning_rate": 7.46761536049368e-05, + "loss": 0.58, + "step": 10598 + }, + { + "epoch": 0.7113855239757055, + "grad_norm": 0.9587852358818054, + "learning_rate": 7.46667005005257e-05, + "loss": 0.6089, + "step": 10600 + }, + { + "epoch": 0.7115197476594746, + "grad_norm": 1.1246082782745361, + "learning_rate": 7.465724623066153e-05, + "loss": 0.5236, + "step": 10602 + }, + { + "epoch": 0.7116539713432435, + "grad_norm": 1.0588661432266235, + "learning_rate": 7.464779079579094e-05, + "loss": 0.5631, + "step": 10604 + }, + { + "epoch": 0.7117881950270125, + "grad_norm": 0.8241974711418152, + "learning_rate": 7.463833419636072e-05, + "loss": 0.5622, + "step": 10606 + }, + { + "epoch": 0.7119224187107815, + "grad_norm": 0.9836342930793762, + "learning_rate": 7.462887643281764e-05, + "loss": 0.5949, + "step": 10608 + }, + { + "epoch": 0.7120566423945505, + "grad_norm": 1.0671802759170532, + "learning_rate": 7.46194175056086e-05, + "loss": 0.5492, + "step": 10610 + }, + { + "epoch": 0.7121908660783195, + "grad_norm": 0.9698143601417542, + "learning_rate": 7.460995741518049e-05, + "loss": 0.6173, + "step": 10612 + }, + { + "epoch": 0.7123250897620885, + "grad_norm": 1.0343726873397827, + "learning_rate": 7.460049616198027e-05, + "loss": 0.5927, + "step": 10614 + }, + { + "epoch": 0.7124593134458576, + "grad_norm": 1.0122747421264648, + "learning_rate": 7.4591033746455e-05, + "loss": 0.5812, + "step": 10616 + }, + { + "epoch": 0.7125935371296265, + "grad_norm": 0.9365114569664001, + "learning_rate": 7.458157016905173e-05, + "loss": 0.593, + "step": 10618 + }, + { + "epoch": 0.7127277608133955, + "grad_norm": 1.0280989408493042, + "learning_rate": 7.45721054302176e-05, + "loss": 0.5717, + "step": 10620 + }, + { + "epoch": 0.7128619844971645, + "grad_norm": 1.0961730480194092, + "learning_rate": 7.456263953039984e-05, + "loss": 0.645, + "step": 10622 + }, + { + "epoch": 0.7129962081809336, + "grad_norm": 1.0248903036117554, + "learning_rate": 7.455317247004563e-05, + "loss": 0.5717, + "step": 10624 + }, + { + "epoch": 0.7131304318647025, + "grad_norm": 0.9453458189964294, + "learning_rate": 7.454370424960231e-05, + "loss": 0.6218, + "step": 10626 + }, + { + "epoch": 0.7132646555484715, + "grad_norm": 1.2950961589813232, + "learning_rate": 7.453423486951723e-05, + "loss": 0.6676, + "step": 10628 + }, + { + "epoch": 0.7133988792322405, + "grad_norm": 1.004745364189148, + "learning_rate": 7.45247643302378e-05, + "loss": 0.6064, + "step": 10630 + }, + { + "epoch": 0.7135331029160096, + "grad_norm": 1.0576225519180298, + "learning_rate": 7.451529263221147e-05, + "loss": 0.647, + "step": 10632 + }, + { + "epoch": 0.7136673265997785, + "grad_norm": 0.8936069011688232, + "learning_rate": 7.450581977588577e-05, + "loss": 0.6206, + "step": 10634 + }, + { + "epoch": 0.7138015502835475, + "grad_norm": 1.0472456216812134, + "learning_rate": 7.44963457617083e-05, + "loss": 0.6276, + "step": 10636 + }, + { + "epoch": 0.7139357739673166, + "grad_norm": 1.0001219511032104, + "learning_rate": 7.448687059012665e-05, + "loss": 0.6069, + "step": 10638 + }, + { + "epoch": 0.7140699976510856, + "grad_norm": 0.9868367314338684, + "learning_rate": 7.44773942615885e-05, + "loss": 0.548, + "step": 10640 + }, + { + "epoch": 0.7142042213348545, + "grad_norm": 1.2530124187469482, + "learning_rate": 7.446791677654162e-05, + "loss": 0.6735, + "step": 10642 + }, + { + "epoch": 0.7143384450186235, + "grad_norm": 2.3150460720062256, + "learning_rate": 7.445843813543379e-05, + "loss": 0.5418, + "step": 10644 + }, + { + "epoch": 0.7144726687023926, + "grad_norm": 1.0615333318710327, + "learning_rate": 7.444895833871283e-05, + "loss": 0.7008, + "step": 10646 + }, + { + "epoch": 0.7146068923861615, + "grad_norm": 1.098844289779663, + "learning_rate": 7.443947738682668e-05, + "loss": 0.7514, + "step": 10648 + }, + { + "epoch": 0.7147411160699305, + "grad_norm": 1.056339979171753, + "learning_rate": 7.44299952802233e-05, + "loss": 0.6633, + "step": 10650 + }, + { + "epoch": 0.7148753397536995, + "grad_norm": 0.9609205722808838, + "learning_rate": 7.442051201935065e-05, + "loss": 0.6155, + "step": 10652 + }, + { + "epoch": 0.7150095634374686, + "grad_norm": 1.1464987993240356, + "learning_rate": 7.441102760465686e-05, + "loss": 0.614, + "step": 10654 + }, + { + "epoch": 0.7151437871212375, + "grad_norm": 1.0660523176193237, + "learning_rate": 7.440154203658999e-05, + "loss": 0.6386, + "step": 10656 + }, + { + "epoch": 0.7152780108050065, + "grad_norm": 1.0465962886810303, + "learning_rate": 7.439205531559825e-05, + "loss": 0.6342, + "step": 10658 + }, + { + "epoch": 0.7154122344887756, + "grad_norm": 0.9891012907028198, + "learning_rate": 7.438256744212987e-05, + "loss": 0.6225, + "step": 10660 + }, + { + "epoch": 0.7155464581725446, + "grad_norm": 0.9910521507263184, + "learning_rate": 7.437307841663312e-05, + "loss": 0.6156, + "step": 10662 + }, + { + "epoch": 0.7156806818563135, + "grad_norm": 0.9643241167068481, + "learning_rate": 7.436358823955634e-05, + "loss": 0.586, + "step": 10664 + }, + { + "epoch": 0.7158149055400825, + "grad_norm": 1.088691234588623, + "learning_rate": 7.435409691134792e-05, + "loss": 0.5797, + "step": 10666 + }, + { + "epoch": 0.7159491292238516, + "grad_norm": 1.0931916236877441, + "learning_rate": 7.434460443245632e-05, + "loss": 0.6551, + "step": 10668 + }, + { + "epoch": 0.7160833529076206, + "grad_norm": 1.0558147430419922, + "learning_rate": 7.433511080333004e-05, + "loss": 0.6085, + "step": 10670 + }, + { + "epoch": 0.7162175765913895, + "grad_norm": 0.9660263657569885, + "learning_rate": 7.43256160244176e-05, + "loss": 0.5556, + "step": 10672 + }, + { + "epoch": 0.7163518002751585, + "grad_norm": 0.9382823705673218, + "learning_rate": 7.431612009616767e-05, + "loss": 0.5926, + "step": 10674 + }, + { + "epoch": 0.7164860239589276, + "grad_norm": 1.1861258745193481, + "learning_rate": 7.430662301902885e-05, + "loss": 0.6212, + "step": 10676 + }, + { + "epoch": 0.7166202476426966, + "grad_norm": 1.098279356956482, + "learning_rate": 7.429712479344992e-05, + "loss": 0.5871, + "step": 10678 + }, + { + "epoch": 0.7167544713264655, + "grad_norm": 1.1667335033416748, + "learning_rate": 7.42876254198796e-05, + "loss": 0.676, + "step": 10680 + }, + { + "epoch": 0.7168886950102346, + "grad_norm": 0.9760928153991699, + "learning_rate": 7.427812489876674e-05, + "loss": 0.5523, + "step": 10682 + }, + { + "epoch": 0.7170229186940036, + "grad_norm": 1.0477968454360962, + "learning_rate": 7.426862323056023e-05, + "loss": 0.6248, + "step": 10684 + }, + { + "epoch": 0.7171571423777725, + "grad_norm": 0.933026909828186, + "learning_rate": 7.425912041570899e-05, + "loss": 0.5802, + "step": 10686 + }, + { + "epoch": 0.7172913660615415, + "grad_norm": 0.999974250793457, + "learning_rate": 7.424961645466202e-05, + "loss": 0.6241, + "step": 10688 + }, + { + "epoch": 0.7174255897453106, + "grad_norm": 1.1083967685699463, + "learning_rate": 7.424011134786835e-05, + "loss": 0.5953, + "step": 10690 + }, + { + "epoch": 0.7175598134290796, + "grad_norm": 1.1214714050292969, + "learning_rate": 7.423060509577707e-05, + "loss": 0.6042, + "step": 10692 + }, + { + "epoch": 0.7176940371128485, + "grad_norm": 0.9524616599082947, + "learning_rate": 7.422109769883738e-05, + "loss": 0.553, + "step": 10694 + }, + { + "epoch": 0.7178282607966175, + "grad_norm": 1.2512476444244385, + "learning_rate": 7.421158915749842e-05, + "loss": 0.5933, + "step": 10696 + }, + { + "epoch": 0.7179624844803866, + "grad_norm": 0.9280952215194702, + "learning_rate": 7.42020794722095e-05, + "loss": 0.56, + "step": 10698 + }, + { + "epoch": 0.7180967081641556, + "grad_norm": 1.0245829820632935, + "learning_rate": 7.419256864341992e-05, + "loss": 0.6062, + "step": 10700 + }, + { + "epoch": 0.7182309318479245, + "grad_norm": 0.9840590953826904, + "learning_rate": 7.418305667157903e-05, + "loss": 0.5775, + "step": 10702 + }, + { + "epoch": 0.7183651555316936, + "grad_norm": 1.1951439380645752, + "learning_rate": 7.417354355713627e-05, + "loss": 0.6861, + "step": 10704 + }, + { + "epoch": 0.7184993792154626, + "grad_norm": 1.0898561477661133, + "learning_rate": 7.416402930054111e-05, + "loss": 0.6109, + "step": 10706 + }, + { + "epoch": 0.7186336028992316, + "grad_norm": 1.06401789188385, + "learning_rate": 7.415451390224309e-05, + "loss": 0.5883, + "step": 10708 + }, + { + "epoch": 0.7187678265830005, + "grad_norm": 0.910780131816864, + "learning_rate": 7.414499736269178e-05, + "loss": 0.5871, + "step": 10710 + }, + { + "epoch": 0.7189020502667696, + "grad_norm": 1.069695234298706, + "learning_rate": 7.413547968233684e-05, + "loss": 0.5824, + "step": 10712 + }, + { + "epoch": 0.7190362739505386, + "grad_norm": 1.0132920742034912, + "learning_rate": 7.412596086162793e-05, + "loss": 0.6835, + "step": 10714 + }, + { + "epoch": 0.7191704976343076, + "grad_norm": 0.9329550862312317, + "learning_rate": 7.411644090101481e-05, + "loss": 0.5653, + "step": 10716 + }, + { + "epoch": 0.7193047213180765, + "grad_norm": 0.8903213739395142, + "learning_rate": 7.410691980094728e-05, + "loss": 0.6081, + "step": 10718 + }, + { + "epoch": 0.7194389450018456, + "grad_norm": 1.0895676612854004, + "learning_rate": 7.409739756187519e-05, + "loss": 0.628, + "step": 10720 + }, + { + "epoch": 0.7195731686856146, + "grad_norm": 1.0903286933898926, + "learning_rate": 7.408787418424848e-05, + "loss": 0.6389, + "step": 10722 + }, + { + "epoch": 0.7197073923693835, + "grad_norm": 1.4827929735183716, + "learning_rate": 7.407834966851705e-05, + "loss": 0.5801, + "step": 10724 + }, + { + "epoch": 0.7198416160531526, + "grad_norm": 0.9832586050033569, + "learning_rate": 7.406882401513096e-05, + "loss": 0.6168, + "step": 10726 + }, + { + "epoch": 0.7199758397369216, + "grad_norm": 1.0722204446792603, + "learning_rate": 7.405929722454026e-05, + "loss": 0.6553, + "step": 10728 + }, + { + "epoch": 0.7201100634206906, + "grad_norm": 1.5461186170578003, + "learning_rate": 7.404976929719507e-05, + "loss": 0.5757, + "step": 10730 + }, + { + "epoch": 0.7202442871044595, + "grad_norm": 0.9481789469718933, + "learning_rate": 7.404024023354558e-05, + "loss": 0.626, + "step": 10732 + }, + { + "epoch": 0.7203785107882286, + "grad_norm": 1.0034617185592651, + "learning_rate": 7.4030710034042e-05, + "loss": 0.5735, + "step": 10734 + }, + { + "epoch": 0.7205127344719976, + "grad_norm": 1.0112169981002808, + "learning_rate": 7.402117869913465e-05, + "loss": 0.6503, + "step": 10736 + }, + { + "epoch": 0.7206469581557666, + "grad_norm": 1.1357429027557373, + "learning_rate": 7.401164622927382e-05, + "loss": 0.6094, + "step": 10738 + }, + { + "epoch": 0.7207811818395355, + "grad_norm": 0.9947372078895569, + "learning_rate": 7.400211262490994e-05, + "loss": 0.6288, + "step": 10740 + }, + { + "epoch": 0.7209154055233046, + "grad_norm": 1.0178544521331787, + "learning_rate": 7.399257788649342e-05, + "loss": 0.6021, + "step": 10742 + }, + { + "epoch": 0.7210496292070736, + "grad_norm": 1.1283224821090698, + "learning_rate": 7.398304201447478e-05, + "loss": 0.5801, + "step": 10744 + }, + { + "epoch": 0.7211838528908426, + "grad_norm": 0.999299943447113, + "learning_rate": 7.397350500930458e-05, + "loss": 0.5839, + "step": 10746 + }, + { + "epoch": 0.7213180765746116, + "grad_norm": 1.0841357707977295, + "learning_rate": 7.39639668714334e-05, + "loss": 0.602, + "step": 10748 + }, + { + "epoch": 0.7214523002583806, + "grad_norm": 1.0852906703948975, + "learning_rate": 7.395442760131192e-05, + "loss": 0.6459, + "step": 10750 + }, + { + "epoch": 0.7215865239421496, + "grad_norm": 1.0680116415023804, + "learning_rate": 7.394488719939081e-05, + "loss": 0.5973, + "step": 10752 + }, + { + "epoch": 0.7217207476259186, + "grad_norm": 1.0845413208007812, + "learning_rate": 7.393534566612089e-05, + "loss": 0.633, + "step": 10754 + }, + { + "epoch": 0.7218549713096876, + "grad_norm": 1.012047290802002, + "learning_rate": 7.392580300195296e-05, + "loss": 0.5649, + "step": 10756 + }, + { + "epoch": 0.7219891949934566, + "grad_norm": 1.0492441654205322, + "learning_rate": 7.391625920733786e-05, + "loss": 0.6044, + "step": 10758 + }, + { + "epoch": 0.7221234186772256, + "grad_norm": 1.1300857067108154, + "learning_rate": 7.390671428272655e-05, + "loss": 0.6455, + "step": 10760 + }, + { + "epoch": 0.7222576423609945, + "grad_norm": 0.9608523845672607, + "learning_rate": 7.389716822857e-05, + "loss": 0.5733, + "step": 10762 + }, + { + "epoch": 0.7223918660447636, + "grad_norm": 1.1491624116897583, + "learning_rate": 7.388762104531925e-05, + "loss": 0.6238, + "step": 10764 + }, + { + "epoch": 0.7225260897285326, + "grad_norm": 0.9276455640792847, + "learning_rate": 7.387807273342539e-05, + "loss": 0.6136, + "step": 10766 + }, + { + "epoch": 0.7226603134123016, + "grad_norm": 1.0152052640914917, + "learning_rate": 7.386852329333953e-05, + "loss": 0.6227, + "step": 10768 + }, + { + "epoch": 0.7227945370960706, + "grad_norm": 1.027574062347412, + "learning_rate": 7.385897272551287e-05, + "loss": 0.5818, + "step": 10770 + }, + { + "epoch": 0.7229287607798396, + "grad_norm": 1.1471387147903442, + "learning_rate": 7.38494210303967e-05, + "loss": 0.6054, + "step": 10772 + }, + { + "epoch": 0.7230629844636086, + "grad_norm": 1.2042551040649414, + "learning_rate": 7.383986820844226e-05, + "loss": 0.6077, + "step": 10774 + }, + { + "epoch": 0.7231972081473776, + "grad_norm": 1.2099021673202515, + "learning_rate": 7.383031426010092e-05, + "loss": 0.6161, + "step": 10776 + }, + { + "epoch": 0.7233314318311466, + "grad_norm": 1.0751736164093018, + "learning_rate": 7.38207591858241e-05, + "loss": 0.5854, + "step": 10778 + }, + { + "epoch": 0.7234656555149156, + "grad_norm": 1.0264619588851929, + "learning_rate": 7.381120298606325e-05, + "loss": 0.5747, + "step": 10780 + }, + { + "epoch": 0.7235998791986846, + "grad_norm": 1.0129170417785645, + "learning_rate": 7.380164566126989e-05, + "loss": 0.5775, + "step": 10782 + }, + { + "epoch": 0.7237341028824537, + "grad_norm": 1.0577455759048462, + "learning_rate": 7.379208721189557e-05, + "loss": 0.6174, + "step": 10784 + }, + { + "epoch": 0.7238683265662226, + "grad_norm": 1.015270471572876, + "learning_rate": 7.37825276383919e-05, + "loss": 0.597, + "step": 10786 + }, + { + "epoch": 0.7240025502499916, + "grad_norm": 1.2538281679153442, + "learning_rate": 7.377296694121058e-05, + "loss": 0.5995, + "step": 10788 + }, + { + "epoch": 0.7241367739337606, + "grad_norm": 1.055049180984497, + "learning_rate": 7.376340512080334e-05, + "loss": 0.6603, + "step": 10790 + }, + { + "epoch": 0.7242709976175297, + "grad_norm": 1.0343923568725586, + "learning_rate": 7.375384217762191e-05, + "loss": 0.5742, + "step": 10792 + }, + { + "epoch": 0.7244052213012986, + "grad_norm": 1.1283427476882935, + "learning_rate": 7.374427811211815e-05, + "loss": 0.6378, + "step": 10794 + }, + { + "epoch": 0.7245394449850676, + "grad_norm": 1.2214810848236084, + "learning_rate": 7.373471292474393e-05, + "loss": 0.6573, + "step": 10796 + }, + { + "epoch": 0.7246736686688366, + "grad_norm": 4.867005348205566, + "learning_rate": 7.372514661595122e-05, + "loss": 0.5434, + "step": 10798 + }, + { + "epoch": 0.7248078923526056, + "grad_norm": 1.8141260147094727, + "learning_rate": 7.371557918619198e-05, + "loss": 0.5713, + "step": 10800 + }, + { + "epoch": 0.7249421160363746, + "grad_norm": 1.0678434371948242, + "learning_rate": 7.370601063591824e-05, + "loss": 0.6123, + "step": 10802 + }, + { + "epoch": 0.7250763397201436, + "grad_norm": 1.0353124141693115, + "learning_rate": 7.369644096558213e-05, + "loss": 0.5903, + "step": 10804 + }, + { + "epoch": 0.7252105634039127, + "grad_norm": 0.9907625317573547, + "learning_rate": 7.368687017563578e-05, + "loss": 0.6341, + "step": 10806 + }, + { + "epoch": 0.7253447870876816, + "grad_norm": 1.0260570049285889, + "learning_rate": 7.36772982665314e-05, + "loss": 0.6229, + "step": 10808 + }, + { + "epoch": 0.7254790107714506, + "grad_norm": 1.0138272047042847, + "learning_rate": 7.366772523872122e-05, + "loss": 0.6119, + "step": 10810 + }, + { + "epoch": 0.7256132344552196, + "grad_norm": 2.370683431625366, + "learning_rate": 7.365815109265757e-05, + "loss": 0.6508, + "step": 10812 + }, + { + "epoch": 0.7257474581389887, + "grad_norm": 0.9320681691169739, + "learning_rate": 7.36485758287928e-05, + "loss": 0.5675, + "step": 10814 + }, + { + "epoch": 0.7258816818227576, + "grad_norm": 1.097596526145935, + "learning_rate": 7.363899944757935e-05, + "loss": 0.5803, + "step": 10816 + }, + { + "epoch": 0.7260159055065266, + "grad_norm": 1.05289888381958, + "learning_rate": 7.362942194946962e-05, + "loss": 0.6129, + "step": 10818 + }, + { + "epoch": 0.7261501291902956, + "grad_norm": 0.9760967493057251, + "learning_rate": 7.361984333491618e-05, + "loss": 0.5784, + "step": 10820 + }, + { + "epoch": 0.7262843528740647, + "grad_norm": 0.9339340925216675, + "learning_rate": 7.361026360437162e-05, + "loss": 0.5642, + "step": 10822 + }, + { + "epoch": 0.7264185765578336, + "grad_norm": 0.8658645153045654, + "learning_rate": 7.360068275828849e-05, + "loss": 0.5677, + "step": 10824 + }, + { + "epoch": 0.7265528002416026, + "grad_norm": 1.546994686126709, + "learning_rate": 7.359110079711953e-05, + "loss": 0.642, + "step": 10826 + }, + { + "epoch": 0.7266870239253717, + "grad_norm": 0.8931393027305603, + "learning_rate": 7.358151772131744e-05, + "loss": 0.6122, + "step": 10828 + }, + { + "epoch": 0.7268212476091407, + "grad_norm": 1.133793830871582, + "learning_rate": 7.357193353133503e-05, + "loss": 0.6548, + "step": 10830 + }, + { + "epoch": 0.7269554712929096, + "grad_norm": 1.0960601568222046, + "learning_rate": 7.356234822762508e-05, + "loss": 0.5956, + "step": 10832 + }, + { + "epoch": 0.7270896949766786, + "grad_norm": 1.1904242038726807, + "learning_rate": 7.355276181064052e-05, + "loss": 0.5764, + "step": 10834 + }, + { + "epoch": 0.7272239186604477, + "grad_norm": 1.222369909286499, + "learning_rate": 7.354317428083428e-05, + "loss": 0.6833, + "step": 10836 + }, + { + "epoch": 0.7273581423442166, + "grad_norm": 1.057033658027649, + "learning_rate": 7.353358563865935e-05, + "loss": 0.5845, + "step": 10838 + }, + { + "epoch": 0.7274923660279856, + "grad_norm": 1.0513136386871338, + "learning_rate": 7.35239958845688e-05, + "loss": 0.574, + "step": 10840 + }, + { + "epoch": 0.7276265897117546, + "grad_norm": 1.9896886348724365, + "learning_rate": 7.351440501901567e-05, + "loss": 0.5445, + "step": 10842 + }, + { + "epoch": 0.7277608133955237, + "grad_norm": 1.1040643453598022, + "learning_rate": 7.350481304245315e-05, + "loss": 0.6422, + "step": 10844 + }, + { + "epoch": 0.7278950370792926, + "grad_norm": 0.9568721055984497, + "learning_rate": 7.349521995533444e-05, + "loss": 0.5501, + "step": 10846 + }, + { + "epoch": 0.7280292607630616, + "grad_norm": 1.1906222105026245, + "learning_rate": 7.348562575811279e-05, + "loss": 0.6471, + "step": 10848 + }, + { + "epoch": 0.7281634844468307, + "grad_norm": 0.911615788936615, + "learning_rate": 7.347603045124149e-05, + "loss": 0.5898, + "step": 10850 + }, + { + "epoch": 0.7282977081305997, + "grad_norm": 1.226242184638977, + "learning_rate": 7.346643403517394e-05, + "loss": 0.6071, + "step": 10852 + }, + { + "epoch": 0.7284319318143686, + "grad_norm": 1.0330523252487183, + "learning_rate": 7.345683651036351e-05, + "loss": 0.6134, + "step": 10854 + }, + { + "epoch": 0.7285661554981376, + "grad_norm": 0.9957926869392395, + "learning_rate": 7.344723787726368e-05, + "loss": 0.5669, + "step": 10856 + }, + { + "epoch": 0.7287003791819067, + "grad_norm": 1.0053691864013672, + "learning_rate": 7.343763813632798e-05, + "loss": 0.5931, + "step": 10858 + }, + { + "epoch": 0.7288346028656757, + "grad_norm": 1.0712882280349731, + "learning_rate": 7.342803728800995e-05, + "loss": 0.6325, + "step": 10860 + }, + { + "epoch": 0.7289688265494446, + "grad_norm": 0.9607504606246948, + "learning_rate": 7.341843533276323e-05, + "loss": 0.5539, + "step": 10862 + }, + { + "epoch": 0.7291030502332136, + "grad_norm": 0.9818548560142517, + "learning_rate": 7.340883227104151e-05, + "loss": 0.5815, + "step": 10864 + }, + { + "epoch": 0.7292372739169827, + "grad_norm": 0.9800499677658081, + "learning_rate": 7.339922810329846e-05, + "loss": 0.5471, + "step": 10866 + }, + { + "epoch": 0.7293714976007517, + "grad_norm": 1.1134047508239746, + "learning_rate": 7.338962282998794e-05, + "loss": 0.584, + "step": 10868 + }, + { + "epoch": 0.7295057212845206, + "grad_norm": 1.0640432834625244, + "learning_rate": 7.33800164515637e-05, + "loss": 0.5663, + "step": 10870 + }, + { + "epoch": 0.7296399449682897, + "grad_norm": 1.0774834156036377, + "learning_rate": 7.337040896847967e-05, + "loss": 0.5586, + "step": 10872 + }, + { + "epoch": 0.7297741686520587, + "grad_norm": 0.9616897702217102, + "learning_rate": 7.336080038118978e-05, + "loss": 0.5472, + "step": 10874 + }, + { + "epoch": 0.7299083923358276, + "grad_norm": 1.0113489627838135, + "learning_rate": 7.335119069014798e-05, + "loss": 0.5974, + "step": 10876 + }, + { + "epoch": 0.7300426160195966, + "grad_norm": 1.2057607173919678, + "learning_rate": 7.334157989580838e-05, + "loss": 0.589, + "step": 10878 + }, + { + "epoch": 0.7301768397033657, + "grad_norm": 0.9961106181144714, + "learning_rate": 7.333196799862499e-05, + "loss": 0.6091, + "step": 10880 + }, + { + "epoch": 0.7303110633871347, + "grad_norm": 0.883374810218811, + "learning_rate": 7.332235499905202e-05, + "loss": 0.6059, + "step": 10882 + }, + { + "epoch": 0.7304452870709036, + "grad_norm": 0.9781058430671692, + "learning_rate": 7.331274089754363e-05, + "loss": 0.5797, + "step": 10884 + }, + { + "epoch": 0.7305795107546726, + "grad_norm": 1.0016698837280273, + "learning_rate": 7.330312569455408e-05, + "loss": 0.5686, + "step": 10886 + }, + { + "epoch": 0.7307137344384417, + "grad_norm": 1.010748267173767, + "learning_rate": 7.329350939053766e-05, + "loss": 0.533, + "step": 10888 + }, + { + "epoch": 0.7308479581222107, + "grad_norm": 1.062753438949585, + "learning_rate": 7.328389198594872e-05, + "loss": 0.6047, + "step": 10890 + }, + { + "epoch": 0.7309821818059796, + "grad_norm": 1.06117844581604, + "learning_rate": 7.327427348124167e-05, + "loss": 0.6129, + "step": 10892 + }, + { + "epoch": 0.7311164054897487, + "grad_norm": 1.3572216033935547, + "learning_rate": 7.326465387687097e-05, + "loss": 0.5429, + "step": 10894 + }, + { + "epoch": 0.7312506291735177, + "grad_norm": 1.013454794883728, + "learning_rate": 7.325503317329112e-05, + "loss": 0.5456, + "step": 10896 + }, + { + "epoch": 0.7313848528572867, + "grad_norm": 1.1997941732406616, + "learning_rate": 7.324541137095669e-05, + "loss": 0.6264, + "step": 10898 + }, + { + "epoch": 0.7315190765410556, + "grad_norm": 1.0833473205566406, + "learning_rate": 7.323578847032226e-05, + "loss": 0.6353, + "step": 10900 + }, + { + "epoch": 0.7316533002248247, + "grad_norm": 1.4310764074325562, + "learning_rate": 7.322616447184254e-05, + "loss": 0.6689, + "step": 10902 + }, + { + "epoch": 0.7317875239085937, + "grad_norm": 1.1005669832229614, + "learning_rate": 7.321653937597222e-05, + "loss": 0.557, + "step": 10904 + }, + { + "epoch": 0.7319217475923627, + "grad_norm": 1.1346960067749023, + "learning_rate": 7.320691318316606e-05, + "loss": 0.6179, + "step": 10906 + }, + { + "epoch": 0.7320559712761316, + "grad_norm": 1.0411349534988403, + "learning_rate": 7.319728589387888e-05, + "loss": 0.6131, + "step": 10908 + }, + { + "epoch": 0.7321901949599007, + "grad_norm": 0.9274797439575195, + "learning_rate": 7.318765750856555e-05, + "loss": 0.6152, + "step": 10910 + }, + { + "epoch": 0.7323244186436697, + "grad_norm": 1.1459671258926392, + "learning_rate": 7.317802802768102e-05, + "loss": 0.6046, + "step": 10912 + }, + { + "epoch": 0.7324586423274386, + "grad_norm": 0.9764459133148193, + "learning_rate": 7.316839745168024e-05, + "loss": 0.5685, + "step": 10914 + }, + { + "epoch": 0.7325928660112077, + "grad_norm": 1.1787185668945312, + "learning_rate": 7.315876578101823e-05, + "loss": 0.61, + "step": 10916 + }, + { + "epoch": 0.7327270896949767, + "grad_norm": 1.0656797885894775, + "learning_rate": 7.314913301615008e-05, + "loss": 0.5772, + "step": 10918 + }, + { + "epoch": 0.7328613133787457, + "grad_norm": 0.9993340969085693, + "learning_rate": 7.313949915753093e-05, + "loss": 0.5461, + "step": 10920 + }, + { + "epoch": 0.7329955370625146, + "grad_norm": 1.1809332370758057, + "learning_rate": 7.312986420561593e-05, + "loss": 0.5383, + "step": 10922 + }, + { + "epoch": 0.7331297607462837, + "grad_norm": 1.0336933135986328, + "learning_rate": 7.312022816086033e-05, + "loss": 0.6016, + "step": 10924 + }, + { + "epoch": 0.7332639844300527, + "grad_norm": 0.8597950339317322, + "learning_rate": 7.311059102371942e-05, + "loss": 0.5386, + "step": 10926 + }, + { + "epoch": 0.7333982081138217, + "grad_norm": 1.0131632089614868, + "learning_rate": 7.310095279464852e-05, + "loss": 0.6432, + "step": 10928 + }, + { + "epoch": 0.7335324317975906, + "grad_norm": 1.0859897136688232, + "learning_rate": 7.309131347410303e-05, + "loss": 0.6246, + "step": 10930 + }, + { + "epoch": 0.7336666554813597, + "grad_norm": 0.9647706747055054, + "learning_rate": 7.308167306253839e-05, + "loss": 0.6463, + "step": 10932 + }, + { + "epoch": 0.7338008791651287, + "grad_norm": 1.065216302871704, + "learning_rate": 7.30720315604101e-05, + "loss": 0.6372, + "step": 10934 + }, + { + "epoch": 0.7339351028488977, + "grad_norm": 0.9903578162193298, + "learning_rate": 7.306238896817366e-05, + "loss": 0.5638, + "step": 10936 + }, + { + "epoch": 0.7340693265326667, + "grad_norm": 1.1092201471328735, + "learning_rate": 7.305274528628469e-05, + "loss": 0.5985, + "step": 10938 + }, + { + "epoch": 0.7342035502164357, + "grad_norm": 1.0268052816390991, + "learning_rate": 7.304310051519886e-05, + "loss": 0.6446, + "step": 10940 + }, + { + "epoch": 0.7343377739002047, + "grad_norm": 1.010176181793213, + "learning_rate": 7.303345465537184e-05, + "loss": 0.5817, + "step": 10942 + }, + { + "epoch": 0.7344719975839737, + "grad_norm": 0.9174301624298096, + "learning_rate": 7.302380770725937e-05, + "loss": 0.6188, + "step": 10944 + }, + { + "epoch": 0.7346062212677427, + "grad_norm": 0.9151535034179688, + "learning_rate": 7.301415967131727e-05, + "loss": 0.6744, + "step": 10946 + }, + { + "epoch": 0.7347404449515117, + "grad_norm": 1.102332353591919, + "learning_rate": 7.300451054800137e-05, + "loss": 0.6855, + "step": 10948 + }, + { + "epoch": 0.7348746686352807, + "grad_norm": 1.0601832866668701, + "learning_rate": 7.29948603377676e-05, + "loss": 0.6343, + "step": 10950 + }, + { + "epoch": 0.7350088923190496, + "grad_norm": 0.9875639081001282, + "learning_rate": 7.298520904107187e-05, + "loss": 0.5326, + "step": 10952 + }, + { + "epoch": 0.7351431160028187, + "grad_norm": 1.0342354774475098, + "learning_rate": 7.297555665837024e-05, + "loss": 0.6317, + "step": 10954 + }, + { + "epoch": 0.7352773396865877, + "grad_norm": 1.122740626335144, + "learning_rate": 7.296590319011871e-05, + "loss": 0.603, + "step": 10956 + }, + { + "epoch": 0.7354115633703567, + "grad_norm": 1.1834608316421509, + "learning_rate": 7.295624863677343e-05, + "loss": 0.6336, + "step": 10958 + }, + { + "epoch": 0.7355457870541257, + "grad_norm": 1.2283573150634766, + "learning_rate": 7.294659299879054e-05, + "loss": 0.569, + "step": 10960 + }, + { + "epoch": 0.7356800107378947, + "grad_norm": 0.9388507604598999, + "learning_rate": 7.293693627662625e-05, + "loss": 0.5259, + "step": 10962 + }, + { + "epoch": 0.7358142344216637, + "grad_norm": 1.1806296110153198, + "learning_rate": 7.292727847073684e-05, + "loss": 0.5792, + "step": 10964 + }, + { + "epoch": 0.7359484581054327, + "grad_norm": 1.0787748098373413, + "learning_rate": 7.29176195815786e-05, + "loss": 0.5581, + "step": 10966 + }, + { + "epoch": 0.7360826817892017, + "grad_norm": 1.2471569776535034, + "learning_rate": 7.29079596096079e-05, + "loss": 0.6241, + "step": 10968 + }, + { + "epoch": 0.7362169054729707, + "grad_norm": 0.9744227528572083, + "learning_rate": 7.289829855528114e-05, + "loss": 0.6011, + "step": 10970 + }, + { + "epoch": 0.7363511291567397, + "grad_norm": 0.9760153889656067, + "learning_rate": 7.288863641905481e-05, + "loss": 0.5663, + "step": 10972 + }, + { + "epoch": 0.7364853528405088, + "grad_norm": 0.8509034514427185, + "learning_rate": 7.287897320138542e-05, + "loss": 0.5625, + "step": 10974 + }, + { + "epoch": 0.7366195765242777, + "grad_norm": 1.7126649618148804, + "learning_rate": 7.286930890272954e-05, + "loss": 0.6058, + "step": 10976 + }, + { + "epoch": 0.7367538002080467, + "grad_norm": 1.046785593032837, + "learning_rate": 7.285964352354378e-05, + "loss": 0.5684, + "step": 10978 + }, + { + "epoch": 0.7368880238918157, + "grad_norm": 1.021963119506836, + "learning_rate": 7.284997706428482e-05, + "loss": 0.6269, + "step": 10980 + }, + { + "epoch": 0.7370222475755848, + "grad_norm": 1.2418005466461182, + "learning_rate": 7.284030952540937e-05, + "loss": 0.6135, + "step": 10982 + }, + { + "epoch": 0.7371564712593537, + "grad_norm": 1.1099112033843994, + "learning_rate": 7.28306409073742e-05, + "loss": 0.6034, + "step": 10984 + }, + { + "epoch": 0.7372906949431227, + "grad_norm": 1.0667752027511597, + "learning_rate": 7.282097121063616e-05, + "loss": 0.6169, + "step": 10986 + }, + { + "epoch": 0.7374249186268917, + "grad_norm": 1.0118855237960815, + "learning_rate": 7.28113004356521e-05, + "loss": 0.5993, + "step": 10988 + }, + { + "epoch": 0.7375591423106607, + "grad_norm": 1.100252628326416, + "learning_rate": 7.280162858287894e-05, + "loss": 0.6263, + "step": 10990 + }, + { + "epoch": 0.7376933659944297, + "grad_norm": 1.1078400611877441, + "learning_rate": 7.279195565277369e-05, + "loss": 0.6074, + "step": 10992 + }, + { + "epoch": 0.7378275896781987, + "grad_norm": 1.0341601371765137, + "learning_rate": 7.278228164579333e-05, + "loss": 0.6238, + "step": 10994 + }, + { + "epoch": 0.7379618133619678, + "grad_norm": 1.0098555088043213, + "learning_rate": 7.277260656239497e-05, + "loss": 0.6187, + "step": 10996 + }, + { + "epoch": 0.7380960370457367, + "grad_norm": 1.0140324831008911, + "learning_rate": 7.276293040303573e-05, + "loss": 0.6115, + "step": 10998 + }, + { + "epoch": 0.7382302607295057, + "grad_norm": 1.0767360925674438, + "learning_rate": 7.275325316817279e-05, + "loss": 0.6177, + "step": 11000 + }, + { + "epoch": 0.7383644844132747, + "grad_norm": 1.2175441980361938, + "learning_rate": 7.274357485826339e-05, + "loss": 0.6105, + "step": 11002 + }, + { + "epoch": 0.7384987080970438, + "grad_norm": 1.2842063903808594, + "learning_rate": 7.27338954737648e-05, + "loss": 0.6284, + "step": 11004 + }, + { + "epoch": 0.7386329317808127, + "grad_norm": 1.1585752964019775, + "learning_rate": 7.272421501513434e-05, + "loss": 0.6264, + "step": 11006 + }, + { + "epoch": 0.7387671554645817, + "grad_norm": 1.0015320777893066, + "learning_rate": 7.27145334828294e-05, + "loss": 0.6096, + "step": 11008 + }, + { + "epoch": 0.7389013791483507, + "grad_norm": 1.0719194412231445, + "learning_rate": 7.270485087730744e-05, + "loss": 0.584, + "step": 11010 + }, + { + "epoch": 0.7390356028321198, + "grad_norm": 1.157904863357544, + "learning_rate": 7.269516719902591e-05, + "loss": 0.5593, + "step": 11012 + }, + { + "epoch": 0.7391698265158887, + "grad_norm": 0.820466935634613, + "learning_rate": 7.268548244844236e-05, + "loss": 0.6092, + "step": 11014 + }, + { + "epoch": 0.7393040501996577, + "grad_norm": 0.9571653604507446, + "learning_rate": 7.267579662601439e-05, + "loss": 0.5796, + "step": 11016 + }, + { + "epoch": 0.7394382738834268, + "grad_norm": 0.9789788126945496, + "learning_rate": 7.266610973219959e-05, + "loss": 0.5665, + "step": 11018 + }, + { + "epoch": 0.7395724975671958, + "grad_norm": 1.0947332382202148, + "learning_rate": 7.265642176745571e-05, + "loss": 0.6398, + "step": 11020 + }, + { + "epoch": 0.7397067212509647, + "grad_norm": 1.1176904439926147, + "learning_rate": 7.264673273224042e-05, + "loss": 0.6227, + "step": 11022 + }, + { + "epoch": 0.7398409449347337, + "grad_norm": 1.037615180015564, + "learning_rate": 7.263704262701157e-05, + "loss": 0.6312, + "step": 11024 + }, + { + "epoch": 0.7399751686185028, + "grad_norm": 1.0402809381484985, + "learning_rate": 7.262735145222696e-05, + "loss": 0.5933, + "step": 11026 + }, + { + "epoch": 0.7401093923022717, + "grad_norm": 0.8275668025016785, + "learning_rate": 7.261765920834447e-05, + "loss": 0.5473, + "step": 11028 + }, + { + "epoch": 0.7402436159860407, + "grad_norm": 0.9896917939186096, + "learning_rate": 7.260796589582208e-05, + "loss": 0.5697, + "step": 11030 + }, + { + "epoch": 0.7403778396698097, + "grad_norm": 1.028495192527771, + "learning_rate": 7.259827151511775e-05, + "loss": 0.629, + "step": 11032 + }, + { + "epoch": 0.7405120633535788, + "grad_norm": 1.251771330833435, + "learning_rate": 7.258857606668951e-05, + "loss": 0.6175, + "step": 11034 + }, + { + "epoch": 0.7406462870373477, + "grad_norm": 1.0744801759719849, + "learning_rate": 7.257887955099551e-05, + "loss": 0.6593, + "step": 11036 + }, + { + "epoch": 0.7407805107211167, + "grad_norm": 1.0244027376174927, + "learning_rate": 7.256918196849382e-05, + "loss": 0.5309, + "step": 11038 + }, + { + "epoch": 0.7409147344048858, + "grad_norm": 1.1488616466522217, + "learning_rate": 7.255948331964268e-05, + "loss": 0.6284, + "step": 11040 + }, + { + "epoch": 0.7410489580886548, + "grad_norm": 0.962162971496582, + "learning_rate": 7.25497836049003e-05, + "loss": 0.6222, + "step": 11042 + }, + { + "epoch": 0.7411831817724237, + "grad_norm": 0.9596481323242188, + "learning_rate": 7.2540082824725e-05, + "loss": 0.6159, + "step": 11044 + }, + { + "epoch": 0.7413174054561927, + "grad_norm": 0.9856430292129517, + "learning_rate": 7.25303809795751e-05, + "loss": 0.5966, + "step": 11046 + }, + { + "epoch": 0.7414516291399618, + "grad_norm": 1.4479644298553467, + "learning_rate": 7.2520678069909e-05, + "loss": 0.5527, + "step": 11048 + }, + { + "epoch": 0.7415858528237308, + "grad_norm": 1.0082017183303833, + "learning_rate": 7.251097409618515e-05, + "loss": 0.6666, + "step": 11050 + }, + { + "epoch": 0.7417200765074997, + "grad_norm": 1.1670246124267578, + "learning_rate": 7.250126905886204e-05, + "loss": 0.575, + "step": 11052 + }, + { + "epoch": 0.7418543001912687, + "grad_norm": 0.9044466018676758, + "learning_rate": 7.249156295839824e-05, + "loss": 0.5328, + "step": 11054 + }, + { + "epoch": 0.7419885238750378, + "grad_norm": 1.0731639862060547, + "learning_rate": 7.248185579525228e-05, + "loss": 0.5571, + "step": 11056 + }, + { + "epoch": 0.7421227475588068, + "grad_norm": 1.0607680082321167, + "learning_rate": 7.247214756988285e-05, + "loss": 0.5882, + "step": 11058 + }, + { + "epoch": 0.7422569712425757, + "grad_norm": 0.9615009427070618, + "learning_rate": 7.246243828274863e-05, + "loss": 0.5327, + "step": 11060 + }, + { + "epoch": 0.7423911949263448, + "grad_norm": 1.143050193786621, + "learning_rate": 7.24527279343084e-05, + "loss": 0.5794, + "step": 11062 + }, + { + "epoch": 0.7425254186101138, + "grad_norm": 1.4304579496383667, + "learning_rate": 7.24430165250209e-05, + "loss": 0.5818, + "step": 11064 + }, + { + "epoch": 0.7426596422938827, + "grad_norm": 0.991141140460968, + "learning_rate": 7.2433304055345e-05, + "loss": 0.6599, + "step": 11066 + }, + { + "epoch": 0.7427938659776517, + "grad_norm": 1.0762946605682373, + "learning_rate": 7.242359052573963e-05, + "loss": 0.6425, + "step": 11068 + }, + { + "epoch": 0.7429280896614208, + "grad_norm": 1.0666600465774536, + "learning_rate": 7.241387593666368e-05, + "loss": 0.5718, + "step": 11070 + }, + { + "epoch": 0.7430623133451898, + "grad_norm": 1.2328325510025024, + "learning_rate": 7.240416028857617e-05, + "loss": 0.5796, + "step": 11072 + }, + { + "epoch": 0.7431965370289587, + "grad_norm": 0.9430737495422363, + "learning_rate": 7.239444358193613e-05, + "loss": 0.6491, + "step": 11074 + }, + { + "epoch": 0.7433307607127277, + "grad_norm": 1.041743278503418, + "learning_rate": 7.238472581720268e-05, + "loss": 0.64, + "step": 11076 + }, + { + "epoch": 0.7434649843964968, + "grad_norm": 1.0096365213394165, + "learning_rate": 7.237500699483495e-05, + "loss": 0.5913, + "step": 11078 + }, + { + "epoch": 0.7435992080802658, + "grad_norm": 1.6502892971038818, + "learning_rate": 7.236528711529213e-05, + "loss": 0.5335, + "step": 11080 + }, + { + "epoch": 0.7437334317640347, + "grad_norm": 1.1586986780166626, + "learning_rate": 7.23555661790335e-05, + "loss": 0.5467, + "step": 11082 + }, + { + "epoch": 0.7438676554478038, + "grad_norm": 0.7920651435852051, + "learning_rate": 7.23458441865183e-05, + "loss": 0.4931, + "step": 11084 + }, + { + "epoch": 0.7440018791315728, + "grad_norm": 1.4115488529205322, + "learning_rate": 7.233612113820592e-05, + "loss": 0.6777, + "step": 11086 + }, + { + "epoch": 0.7441361028153418, + "grad_norm": 1.0248574018478394, + "learning_rate": 7.232639703455573e-05, + "loss": 0.5823, + "step": 11088 + }, + { + "epoch": 0.7442703264991107, + "grad_norm": 1.38546621799469, + "learning_rate": 7.231667187602718e-05, + "loss": 0.6076, + "step": 11090 + }, + { + "epoch": 0.7444045501828798, + "grad_norm": 1.055503487586975, + "learning_rate": 7.230694566307978e-05, + "loss": 0.5366, + "step": 11092 + }, + { + "epoch": 0.7445387738666488, + "grad_norm": 1.0996710062026978, + "learning_rate": 7.229721839617306e-05, + "loss": 0.6016, + "step": 11094 + }, + { + "epoch": 0.7446729975504178, + "grad_norm": 1.1728242635726929, + "learning_rate": 7.228749007576661e-05, + "loss": 0.6528, + "step": 11096 + }, + { + "epoch": 0.7448072212341867, + "grad_norm": 1.067726492881775, + "learning_rate": 7.227776070232008e-05, + "loss": 0.7174, + "step": 11098 + }, + { + "epoch": 0.7449414449179558, + "grad_norm": 1.052566647529602, + "learning_rate": 7.226803027629316e-05, + "loss": 0.5095, + "step": 11100 + }, + { + "epoch": 0.7450756686017248, + "grad_norm": 0.9860169887542725, + "learning_rate": 7.225829879814561e-05, + "loss": 0.5785, + "step": 11102 + }, + { + "epoch": 0.7452098922854937, + "grad_norm": 0.9966432452201843, + "learning_rate": 7.22485662683372e-05, + "loss": 0.5927, + "step": 11104 + }, + { + "epoch": 0.7453441159692628, + "grad_norm": 0.8664472699165344, + "learning_rate": 7.223883268732779e-05, + "loss": 0.5297, + "step": 11106 + }, + { + "epoch": 0.7454783396530318, + "grad_norm": 1.3534940481185913, + "learning_rate": 7.222909805557726e-05, + "loss": 0.6371, + "step": 11108 + }, + { + "epoch": 0.7456125633368008, + "grad_norm": 1.2141648530960083, + "learning_rate": 7.221936237354557e-05, + "loss": 0.5975, + "step": 11110 + }, + { + "epoch": 0.7457467870205697, + "grad_norm": 1.1479947566986084, + "learning_rate": 7.22096256416927e-05, + "loss": 0.5909, + "step": 11112 + }, + { + "epoch": 0.7458810107043388, + "grad_norm": 1.2423744201660156, + "learning_rate": 7.219988786047866e-05, + "loss": 0.5182, + "step": 11114 + }, + { + "epoch": 0.7460152343881078, + "grad_norm": 0.9990307092666626, + "learning_rate": 7.219014903036361e-05, + "loss": 0.608, + "step": 11116 + }, + { + "epoch": 0.7461494580718768, + "grad_norm": 0.9691771268844604, + "learning_rate": 7.218040915180764e-05, + "loss": 0.5972, + "step": 11118 + }, + { + "epoch": 0.7462836817556457, + "grad_norm": 1.0182071924209595, + "learning_rate": 7.217066822527096e-05, + "loss": 0.5767, + "step": 11120 + }, + { + "epoch": 0.7464179054394148, + "grad_norm": 1.1678303480148315, + "learning_rate": 7.216092625121379e-05, + "loss": 0.5906, + "step": 11122 + }, + { + "epoch": 0.7465521291231838, + "grad_norm": 1.0928617715835571, + "learning_rate": 7.215118323009643e-05, + "loss": 0.5542, + "step": 11124 + }, + { + "epoch": 0.7466863528069528, + "grad_norm": 1.1073976755142212, + "learning_rate": 7.214143916237925e-05, + "loss": 0.592, + "step": 11126 + }, + { + "epoch": 0.7468205764907218, + "grad_norm": 1.024496078491211, + "learning_rate": 7.213169404852258e-05, + "loss": 0.6111, + "step": 11128 + }, + { + "epoch": 0.7469548001744908, + "grad_norm": 0.96971195936203, + "learning_rate": 7.21219478889869e-05, + "loss": 0.6242, + "step": 11130 + }, + { + "epoch": 0.7470890238582598, + "grad_norm": 1.0988749265670776, + "learning_rate": 7.211220068423266e-05, + "loss": 0.6015, + "step": 11132 + }, + { + "epoch": 0.7472232475420288, + "grad_norm": 1.0498028993606567, + "learning_rate": 7.210245243472046e-05, + "loss": 0.5652, + "step": 11134 + }, + { + "epoch": 0.7473574712257978, + "grad_norm": 0.9467136859893799, + "learning_rate": 7.209270314091081e-05, + "loss": 0.6386, + "step": 11136 + }, + { + "epoch": 0.7474916949095668, + "grad_norm": 1.2103420495986938, + "learning_rate": 7.208295280326439e-05, + "loss": 0.5894, + "step": 11138 + }, + { + "epoch": 0.7476259185933358, + "grad_norm": 1.0159958600997925, + "learning_rate": 7.207320142224188e-05, + "loss": 0.6403, + "step": 11140 + }, + { + "epoch": 0.7477601422771047, + "grad_norm": 0.9736151695251465, + "learning_rate": 7.206344899830401e-05, + "loss": 0.6065, + "step": 11142 + }, + { + "epoch": 0.7478943659608738, + "grad_norm": 1.1738616228103638, + "learning_rate": 7.205369553191156e-05, + "loss": 0.5054, + "step": 11144 + }, + { + "epoch": 0.7480285896446428, + "grad_norm": 0.9719931483268738, + "learning_rate": 7.204394102352535e-05, + "loss": 0.6456, + "step": 11146 + }, + { + "epoch": 0.7481628133284118, + "grad_norm": 0.9128687381744385, + "learning_rate": 7.20341854736063e-05, + "loss": 0.5473, + "step": 11148 + }, + { + "epoch": 0.7482970370121808, + "grad_norm": 1.0695174932479858, + "learning_rate": 7.20244288826153e-05, + "loss": 0.5721, + "step": 11150 + }, + { + "epoch": 0.7484312606959498, + "grad_norm": 1.0116002559661865, + "learning_rate": 7.201467125101332e-05, + "loss": 0.5793, + "step": 11152 + }, + { + "epoch": 0.7485654843797188, + "grad_norm": 1.1777184009552002, + "learning_rate": 7.200491257926145e-05, + "loss": 0.6363, + "step": 11154 + }, + { + "epoch": 0.7486997080634878, + "grad_norm": 1.085649847984314, + "learning_rate": 7.19951528678207e-05, + "loss": 0.6127, + "step": 11156 + }, + { + "epoch": 0.7488339317472568, + "grad_norm": 1.094965934753418, + "learning_rate": 7.198539211715226e-05, + "loss": 0.6476, + "step": 11158 + }, + { + "epoch": 0.7489681554310258, + "grad_norm": 0.9376707673072815, + "learning_rate": 7.197563032771727e-05, + "loss": 0.5898, + "step": 11160 + }, + { + "epoch": 0.7491023791147948, + "grad_norm": 1.0657490491867065, + "learning_rate": 7.196586749997694e-05, + "loss": 0.6939, + "step": 11162 + }, + { + "epoch": 0.7492366027985639, + "grad_norm": 0.9906362891197205, + "learning_rate": 7.195610363439259e-05, + "loss": 0.5848, + "step": 11164 + }, + { + "epoch": 0.7493708264823328, + "grad_norm": 1.4110857248306274, + "learning_rate": 7.194633873142548e-05, + "loss": 0.6378, + "step": 11166 + }, + { + "epoch": 0.7495050501661018, + "grad_norm": 1.144875168800354, + "learning_rate": 7.193657279153706e-05, + "loss": 0.6205, + "step": 11168 + }, + { + "epoch": 0.7496392738498708, + "grad_norm": 1.0834269523620605, + "learning_rate": 7.19268058151887e-05, + "loss": 0.5946, + "step": 11170 + }, + { + "epoch": 0.7497734975336399, + "grad_norm": 1.1376134157180786, + "learning_rate": 7.191703780284187e-05, + "loss": 0.6001, + "step": 11172 + }, + { + "epoch": 0.7499077212174088, + "grad_norm": 0.9572040438652039, + "learning_rate": 7.190726875495812e-05, + "loss": 0.6463, + "step": 11174 + }, + { + "epoch": 0.7500419449011778, + "grad_norm": 0.9035720229148865, + "learning_rate": 7.189749867199899e-05, + "loss": 0.5924, + "step": 11176 + }, + { + "epoch": 0.7501761685849468, + "grad_norm": 1.004518747329712, + "learning_rate": 7.188772755442611e-05, + "loss": 0.5729, + "step": 11178 + }, + { + "epoch": 0.7503103922687158, + "grad_norm": 1.8473553657531738, + "learning_rate": 7.187795540270114e-05, + "loss": 0.612, + "step": 11180 + }, + { + "epoch": 0.7504446159524848, + "grad_norm": 5.95490026473999, + "learning_rate": 7.18681822172858e-05, + "loss": 0.5695, + "step": 11182 + }, + { + "epoch": 0.7505788396362538, + "grad_norm": 1.2424906492233276, + "learning_rate": 7.185840799864186e-05, + "loss": 0.6821, + "step": 11184 + }, + { + "epoch": 0.7507130633200229, + "grad_norm": 1.002289056777954, + "learning_rate": 7.184863274723111e-05, + "loss": 0.6464, + "step": 11186 + }, + { + "epoch": 0.7508472870037918, + "grad_norm": 1.2320494651794434, + "learning_rate": 7.183885646351542e-05, + "loss": 0.6134, + "step": 11188 + }, + { + "epoch": 0.7509815106875608, + "grad_norm": 1.0252602100372314, + "learning_rate": 7.182907914795672e-05, + "loss": 0.6894, + "step": 11190 + }, + { + "epoch": 0.7511157343713298, + "grad_norm": 1.0166563987731934, + "learning_rate": 7.181930080101696e-05, + "loss": 0.5954, + "step": 11192 + }, + { + "epoch": 0.7512499580550989, + "grad_norm": 1.0519530773162842, + "learning_rate": 7.180952142315813e-05, + "loss": 0.6021, + "step": 11194 + }, + { + "epoch": 0.7513841817388678, + "grad_norm": 1.084174633026123, + "learning_rate": 7.179974101484232e-05, + "loss": 0.5695, + "step": 11196 + }, + { + "epoch": 0.7515184054226368, + "grad_norm": 0.9310356974601746, + "learning_rate": 7.17899595765316e-05, + "loss": 0.4996, + "step": 11198 + }, + { + "epoch": 0.7516526291064058, + "grad_norm": 0.9874021410942078, + "learning_rate": 7.178017710868814e-05, + "loss": 0.5587, + "step": 11200 + }, + { + "epoch": 0.7517868527901749, + "grad_norm": 2.154062032699585, + "learning_rate": 7.177039361177413e-05, + "loss": 0.6281, + "step": 11202 + }, + { + "epoch": 0.7519210764739438, + "grad_norm": 1.0503273010253906, + "learning_rate": 7.176060908625184e-05, + "loss": 0.5806, + "step": 11204 + }, + { + "epoch": 0.7520553001577128, + "grad_norm": 1.2201011180877686, + "learning_rate": 7.175082353258358e-05, + "loss": 0.6996, + "step": 11206 + }, + { + "epoch": 0.7521895238414819, + "grad_norm": 0.8703465461730957, + "learning_rate": 7.174103695123166e-05, + "loss": 0.4985, + "step": 11208 + }, + { + "epoch": 0.7523237475252509, + "grad_norm": 1.0070104598999023, + "learning_rate": 7.17312493426585e-05, + "loss": 0.535, + "step": 11210 + }, + { + "epoch": 0.7524579712090198, + "grad_norm": 1.2064918279647827, + "learning_rate": 7.172146070732652e-05, + "loss": 0.6514, + "step": 11212 + }, + { + "epoch": 0.7525921948927888, + "grad_norm": 1.5008562803268433, + "learning_rate": 7.171167104569826e-05, + "loss": 0.5332, + "step": 11214 + }, + { + "epoch": 0.7527264185765579, + "grad_norm": 1.0957249402999878, + "learning_rate": 7.170188035823624e-05, + "loss": 0.6015, + "step": 11216 + }, + { + "epoch": 0.7528606422603268, + "grad_norm": 1.0030272006988525, + "learning_rate": 7.169208864540303e-05, + "loss": 0.5409, + "step": 11218 + }, + { + "epoch": 0.7529948659440958, + "grad_norm": 0.9358177185058594, + "learning_rate": 7.16822959076613e-05, + "loss": 0.57, + "step": 11220 + }, + { + "epoch": 0.7531290896278648, + "grad_norm": 1.0475953817367554, + "learning_rate": 7.167250214547372e-05, + "loss": 0.6187, + "step": 11222 + }, + { + "epoch": 0.7532633133116339, + "grad_norm": 0.8726040124893188, + "learning_rate": 7.166270735930304e-05, + "loss": 0.5462, + "step": 11224 + }, + { + "epoch": 0.7533975369954028, + "grad_norm": 0.8723013401031494, + "learning_rate": 7.165291154961202e-05, + "loss": 0.5357, + "step": 11226 + }, + { + "epoch": 0.7535317606791718, + "grad_norm": 0.9909194111824036, + "learning_rate": 7.164311471686352e-05, + "loss": 0.6319, + "step": 11228 + }, + { + "epoch": 0.7536659843629409, + "grad_norm": 1.0541589260101318, + "learning_rate": 7.163331686152042e-05, + "loss": 0.6679, + "step": 11230 + }, + { + "epoch": 0.7538002080467099, + "grad_norm": 0.9891774654388428, + "learning_rate": 7.16235179840456e-05, + "loss": 0.6496, + "step": 11232 + }, + { + "epoch": 0.7539344317304788, + "grad_norm": 1.0104056596755981, + "learning_rate": 7.161371808490212e-05, + "loss": 0.5731, + "step": 11234 + }, + { + "epoch": 0.7540686554142478, + "grad_norm": 1.0035972595214844, + "learning_rate": 7.160391716455292e-05, + "loss": 0.6075, + "step": 11236 + }, + { + "epoch": 0.7542028790980169, + "grad_norm": 0.9934561252593994, + "learning_rate": 7.159411522346115e-05, + "loss": 0.5643, + "step": 11238 + }, + { + "epoch": 0.7543371027817859, + "grad_norm": 1.009917140007019, + "learning_rate": 7.158431226208988e-05, + "loss": 0.5825, + "step": 11240 + }, + { + "epoch": 0.7544713264655548, + "grad_norm": 0.9430350661277771, + "learning_rate": 7.157450828090231e-05, + "loss": 0.5723, + "step": 11242 + }, + { + "epoch": 0.7546055501493238, + "grad_norm": 1.1066055297851562, + "learning_rate": 7.156470328036165e-05, + "loss": 0.6028, + "step": 11244 + }, + { + "epoch": 0.7547397738330929, + "grad_norm": 0.9261332154273987, + "learning_rate": 7.155489726093114e-05, + "loss": 0.5163, + "step": 11246 + }, + { + "epoch": 0.7548739975168619, + "grad_norm": 1.0471699237823486, + "learning_rate": 7.154509022307415e-05, + "loss": 0.6726, + "step": 11248 + }, + { + "epoch": 0.7550082212006308, + "grad_norm": 1.1183794736862183, + "learning_rate": 7.1535282167254e-05, + "loss": 0.602, + "step": 11250 + }, + { + "epoch": 0.7551424448843999, + "grad_norm": 1.6070631742477417, + "learning_rate": 7.152547309393411e-05, + "loss": 0.6271, + "step": 11252 + }, + { + "epoch": 0.7552766685681689, + "grad_norm": 2.8954949378967285, + "learning_rate": 7.151566300357796e-05, + "loss": 0.5473, + "step": 11254 + }, + { + "epoch": 0.7554108922519378, + "grad_norm": 1.2352590560913086, + "learning_rate": 7.150585189664902e-05, + "loss": 0.5778, + "step": 11256 + }, + { + "epoch": 0.7555451159357068, + "grad_norm": 1.0116827487945557, + "learning_rate": 7.14960397736109e-05, + "loss": 0.617, + "step": 11258 + }, + { + "epoch": 0.7556793396194759, + "grad_norm": 1.0050992965698242, + "learning_rate": 7.148622663492715e-05, + "loss": 0.5584, + "step": 11260 + }, + { + "epoch": 0.7558135633032449, + "grad_norm": 1.0844353437423706, + "learning_rate": 7.147641248106142e-05, + "loss": 0.5867, + "step": 11262 + }, + { + "epoch": 0.7559477869870138, + "grad_norm": 1.0055772066116333, + "learning_rate": 7.146659731247747e-05, + "loss": 0.6329, + "step": 11264 + }, + { + "epoch": 0.7560820106707828, + "grad_norm": 1.0988073348999023, + "learning_rate": 7.1456781129639e-05, + "loss": 0.5738, + "step": 11266 + }, + { + "epoch": 0.7562162343545519, + "grad_norm": 1.0617798566818237, + "learning_rate": 7.144696393300981e-05, + "loss": 0.5328, + "step": 11268 + }, + { + "epoch": 0.7563504580383209, + "grad_norm": 1.1145955324172974, + "learning_rate": 7.143714572305374e-05, + "loss": 0.5807, + "step": 11270 + }, + { + "epoch": 0.7564846817220898, + "grad_norm": 1.2228416204452515, + "learning_rate": 7.14273265002347e-05, + "loss": 0.6712, + "step": 11272 + }, + { + "epoch": 0.7566189054058589, + "grad_norm": 1.2184854745864868, + "learning_rate": 7.141750626501661e-05, + "loss": 0.6264, + "step": 11274 + }, + { + "epoch": 0.7567531290896279, + "grad_norm": 0.8387179374694824, + "learning_rate": 7.140768501786347e-05, + "loss": 0.5983, + "step": 11276 + }, + { + "epoch": 0.7568873527733969, + "grad_norm": 1.1214351654052734, + "learning_rate": 7.13978627592393e-05, + "loss": 0.576, + "step": 11278 + }, + { + "epoch": 0.7570215764571658, + "grad_norm": 1.0166372060775757, + "learning_rate": 7.13880394896082e-05, + "loss": 0.5419, + "step": 11280 + }, + { + "epoch": 0.7571558001409349, + "grad_norm": 1.0215328931808472, + "learning_rate": 7.13782152094343e-05, + "loss": 0.58, + "step": 11282 + }, + { + "epoch": 0.7572900238247039, + "grad_norm": 2.02290678024292, + "learning_rate": 7.136838991918175e-05, + "loss": 0.7162, + "step": 11284 + }, + { + "epoch": 0.7574242475084729, + "grad_norm": 1.0733305215835571, + "learning_rate": 7.135856361931482e-05, + "loss": 0.6071, + "step": 11286 + }, + { + "epoch": 0.7575584711922418, + "grad_norm": 1.0772255659103394, + "learning_rate": 7.134873631029775e-05, + "loss": 0.5756, + "step": 11288 + }, + { + "epoch": 0.7576926948760109, + "grad_norm": 1.0512280464172363, + "learning_rate": 7.133890799259486e-05, + "loss": 0.6125, + "step": 11290 + }, + { + "epoch": 0.7578269185597799, + "grad_norm": 1.0169273614883423, + "learning_rate": 7.132907866667053e-05, + "loss": 0.5617, + "step": 11292 + }, + { + "epoch": 0.7579611422435488, + "grad_norm": 1.1695070266723633, + "learning_rate": 7.131924833298918e-05, + "loss": 0.6074, + "step": 11294 + }, + { + "epoch": 0.7580953659273179, + "grad_norm": 1.4435280561447144, + "learning_rate": 7.130941699201528e-05, + "loss": 0.619, + "step": 11296 + }, + { + "epoch": 0.7582295896110869, + "grad_norm": 1.000217080116272, + "learning_rate": 7.129958464421331e-05, + "loss": 0.5763, + "step": 11298 + }, + { + "epoch": 0.7583638132948559, + "grad_norm": 1.026188611984253, + "learning_rate": 7.128975129004786e-05, + "loss": 0.5677, + "step": 11300 + }, + { + "epoch": 0.7584980369786248, + "grad_norm": 1.1340270042419434, + "learning_rate": 7.127991692998353e-05, + "loss": 0.621, + "step": 11302 + }, + { + "epoch": 0.7586322606623939, + "grad_norm": 1.1129652261734009, + "learning_rate": 7.127008156448496e-05, + "loss": 0.7227, + "step": 11304 + }, + { + "epoch": 0.7587664843461629, + "grad_norm": 0.9840350151062012, + "learning_rate": 7.126024519401687e-05, + "loss": 0.6042, + "step": 11306 + }, + { + "epoch": 0.7589007080299319, + "grad_norm": 1.0387786626815796, + "learning_rate": 7.1250407819044e-05, + "loss": 0.6167, + "step": 11308 + }, + { + "epoch": 0.7590349317137008, + "grad_norm": 0.9231281876564026, + "learning_rate": 7.124056944003114e-05, + "loss": 0.5275, + "step": 11310 + }, + { + "epoch": 0.7591691553974699, + "grad_norm": 1.3023412227630615, + "learning_rate": 7.123073005744314e-05, + "loss": 0.5872, + "step": 11312 + }, + { + "epoch": 0.7593033790812389, + "grad_norm": 3.0256383419036865, + "learning_rate": 7.12208896717449e-05, + "loss": 0.5885, + "step": 11314 + }, + { + "epoch": 0.7594376027650079, + "grad_norm": 1.1872811317443848, + "learning_rate": 7.121104828340136e-05, + "loss": 0.5464, + "step": 11316 + }, + { + "epoch": 0.7595718264487769, + "grad_norm": 0.9879008531570435, + "learning_rate": 7.120120589287749e-05, + "loss": 0.5875, + "step": 11318 + }, + { + "epoch": 0.7597060501325459, + "grad_norm": 1.045357346534729, + "learning_rate": 7.119136250063833e-05, + "loss": 0.5428, + "step": 11320 + }, + { + "epoch": 0.7598402738163149, + "grad_norm": 1.0363407135009766, + "learning_rate": 7.118151810714896e-05, + "loss": 0.5652, + "step": 11322 + }, + { + "epoch": 0.759974497500084, + "grad_norm": 1.0396844148635864, + "learning_rate": 7.117167271287453e-05, + "loss": 0.661, + "step": 11324 + }, + { + "epoch": 0.7601087211838529, + "grad_norm": 1.00739586353302, + "learning_rate": 7.116182631828016e-05, + "loss": 0.6256, + "step": 11326 + }, + { + "epoch": 0.7602429448676219, + "grad_norm": 1.515653133392334, + "learning_rate": 7.115197892383114e-05, + "loss": 0.6617, + "step": 11328 + }, + { + "epoch": 0.7603771685513909, + "grad_norm": 1.025126338005066, + "learning_rate": 7.11421305299927e-05, + "loss": 0.6959, + "step": 11330 + }, + { + "epoch": 0.7605113922351598, + "grad_norm": 1.0322550535202026, + "learning_rate": 7.113228113723016e-05, + "loss": 0.5964, + "step": 11332 + }, + { + "epoch": 0.7606456159189289, + "grad_norm": 1.1067694425582886, + "learning_rate": 7.11224307460089e-05, + "loss": 0.4972, + "step": 11334 + }, + { + "epoch": 0.7607798396026979, + "grad_norm": 1.0172574520111084, + "learning_rate": 7.111257935679433e-05, + "loss": 0.5196, + "step": 11336 + }, + { + "epoch": 0.7609140632864669, + "grad_norm": 0.9377800822257996, + "learning_rate": 7.110272697005189e-05, + "loss": 0.5904, + "step": 11338 + }, + { + "epoch": 0.7610482869702359, + "grad_norm": 0.9999615550041199, + "learning_rate": 7.10928735862471e-05, + "loss": 0.6526, + "step": 11340 + }, + { + "epoch": 0.7611825106540049, + "grad_norm": 1.1272873878479004, + "learning_rate": 7.108301920584552e-05, + "loss": 0.5765, + "step": 11342 + }, + { + "epoch": 0.7613167343377739, + "grad_norm": 1.0125153064727783, + "learning_rate": 7.107316382931272e-05, + "loss": 0.6562, + "step": 11344 + }, + { + "epoch": 0.761450958021543, + "grad_norm": 1.16325044631958, + "learning_rate": 7.106330745711438e-05, + "loss": 0.61, + "step": 11346 + }, + { + "epoch": 0.7615851817053119, + "grad_norm": 1.0269536972045898, + "learning_rate": 7.105345008971619e-05, + "loss": 0.6078, + "step": 11348 + }, + { + "epoch": 0.7617194053890809, + "grad_norm": 1.2129024267196655, + "learning_rate": 7.104359172758387e-05, + "loss": 0.6532, + "step": 11350 + }, + { + "epoch": 0.7618536290728499, + "grad_norm": 1.1761541366577148, + "learning_rate": 7.103373237118321e-05, + "loss": 0.5589, + "step": 11352 + }, + { + "epoch": 0.761987852756619, + "grad_norm": 1.0910080671310425, + "learning_rate": 7.102387202098008e-05, + "loss": 0.5613, + "step": 11354 + }, + { + "epoch": 0.7621220764403879, + "grad_norm": 1.0277620553970337, + "learning_rate": 7.101401067744033e-05, + "loss": 0.6106, + "step": 11356 + }, + { + "epoch": 0.7622563001241569, + "grad_norm": 0.9003410339355469, + "learning_rate": 7.10041483410299e-05, + "loss": 0.5574, + "step": 11358 + }, + { + "epoch": 0.7623905238079259, + "grad_norm": 1.14950692653656, + "learning_rate": 7.099428501221476e-05, + "loss": 0.5946, + "step": 11360 + }, + { + "epoch": 0.7625247474916949, + "grad_norm": 0.9178588390350342, + "learning_rate": 7.098442069146095e-05, + "loss": 0.5615, + "step": 11362 + }, + { + "epoch": 0.7626589711754639, + "grad_norm": 1.1091738939285278, + "learning_rate": 7.09745553792345e-05, + "loss": 0.6495, + "step": 11364 + }, + { + "epoch": 0.7627931948592329, + "grad_norm": 1.4295737743377686, + "learning_rate": 7.096468907600157e-05, + "loss": 0.6089, + "step": 11366 + }, + { + "epoch": 0.762927418543002, + "grad_norm": 0.9904151558876038, + "learning_rate": 7.095482178222832e-05, + "loss": 0.6365, + "step": 11368 + }, + { + "epoch": 0.7630616422267709, + "grad_norm": 0.9750891923904419, + "learning_rate": 7.094495349838092e-05, + "loss": 0.5672, + "step": 11370 + }, + { + "epoch": 0.7631958659105399, + "grad_norm": 1.06587815284729, + "learning_rate": 7.093508422492568e-05, + "loss": 0.5864, + "step": 11372 + }, + { + "epoch": 0.7633300895943089, + "grad_norm": 0.9094560742378235, + "learning_rate": 7.092521396232887e-05, + "loss": 0.5021, + "step": 11374 + }, + { + "epoch": 0.763464313278078, + "grad_norm": 1.104093074798584, + "learning_rate": 7.091534271105682e-05, + "loss": 0.6175, + "step": 11376 + }, + { + "epoch": 0.7635985369618469, + "grad_norm": 0.9446908831596375, + "learning_rate": 7.090547047157599e-05, + "loss": 0.5092, + "step": 11378 + }, + { + "epoch": 0.7637327606456159, + "grad_norm": 1.130378007888794, + "learning_rate": 7.089559724435277e-05, + "loss": 0.645, + "step": 11380 + }, + { + "epoch": 0.7638669843293849, + "grad_norm": 0.971934974193573, + "learning_rate": 7.088572302985368e-05, + "loss": 0.5984, + "step": 11382 + }, + { + "epoch": 0.764001208013154, + "grad_norm": 1.1072789430618286, + "learning_rate": 7.087584782854525e-05, + "loss": 0.5966, + "step": 11384 + }, + { + "epoch": 0.7641354316969229, + "grad_norm": 0.9358595013618469, + "learning_rate": 7.086597164089404e-05, + "loss": 0.5732, + "step": 11386 + }, + { + "epoch": 0.7642696553806919, + "grad_norm": 0.9158409833908081, + "learning_rate": 7.085609446736671e-05, + "loss": 0.6165, + "step": 11388 + }, + { + "epoch": 0.764403879064461, + "grad_norm": 1.1295926570892334, + "learning_rate": 7.084621630842993e-05, + "loss": 0.6221, + "step": 11390 + }, + { + "epoch": 0.76453810274823, + "grad_norm": 1.6016974449157715, + "learning_rate": 7.083633716455043e-05, + "loss": 0.5171, + "step": 11392 + }, + { + "epoch": 0.7646723264319989, + "grad_norm": 1.0827736854553223, + "learning_rate": 7.082645703619496e-05, + "loss": 0.6444, + "step": 11394 + }, + { + "epoch": 0.7648065501157679, + "grad_norm": 1.1585004329681396, + "learning_rate": 7.081657592383035e-05, + "loss": 0.5721, + "step": 11396 + }, + { + "epoch": 0.764940773799537, + "grad_norm": 1.180953025817871, + "learning_rate": 7.080669382792346e-05, + "loss": 0.5513, + "step": 11398 + }, + { + "epoch": 0.7650749974833059, + "grad_norm": 0.9680315256118774, + "learning_rate": 7.079681074894123e-05, + "loss": 0.5934, + "step": 11400 + }, + { + "epoch": 0.7652092211670749, + "grad_norm": 1.234734058380127, + "learning_rate": 7.078692668735054e-05, + "loss": 0.5883, + "step": 11402 + }, + { + "epoch": 0.7653434448508439, + "grad_norm": 1.1248189210891724, + "learning_rate": 7.077704164361848e-05, + "loss": 0.5944, + "step": 11404 + }, + { + "epoch": 0.765477668534613, + "grad_norm": 1.1900737285614014, + "learning_rate": 7.076715561821204e-05, + "loss": 0.5976, + "step": 11406 + }, + { + "epoch": 0.7656118922183819, + "grad_norm": 1.0213483572006226, + "learning_rate": 7.075726861159832e-05, + "loss": 0.5714, + "step": 11408 + }, + { + "epoch": 0.7657461159021509, + "grad_norm": 1.0772080421447754, + "learning_rate": 7.07473806242445e-05, + "loss": 0.5593, + "step": 11410 + }, + { + "epoch": 0.76588033958592, + "grad_norm": 1.0889390707015991, + "learning_rate": 7.073749165661773e-05, + "loss": 0.6026, + "step": 11412 + }, + { + "epoch": 0.766014563269689, + "grad_norm": 1.0070146322250366, + "learning_rate": 7.072760170918526e-05, + "loss": 0.6509, + "step": 11414 + }, + { + "epoch": 0.7661487869534579, + "grad_norm": 0.9670523405075073, + "learning_rate": 7.071771078241438e-05, + "loss": 0.5696, + "step": 11416 + }, + { + "epoch": 0.7662830106372269, + "grad_norm": 1.8057606220245361, + "learning_rate": 7.070781887677239e-05, + "loss": 0.5914, + "step": 11418 + }, + { + "epoch": 0.766417234320996, + "grad_norm": 0.9283233880996704, + "learning_rate": 7.069792599272669e-05, + "loss": 0.606, + "step": 11420 + }, + { + "epoch": 0.766551458004765, + "grad_norm": 1.1039934158325195, + "learning_rate": 7.068803213074468e-05, + "loss": 0.6239, + "step": 11422 + }, + { + "epoch": 0.7666856816885339, + "grad_norm": 1.1731226444244385, + "learning_rate": 7.067813729129384e-05, + "loss": 0.5942, + "step": 11424 + }, + { + "epoch": 0.7668199053723029, + "grad_norm": 1.0220401287078857, + "learning_rate": 7.066824147484165e-05, + "loss": 0.554, + "step": 11426 + }, + { + "epoch": 0.766954129056072, + "grad_norm": 1.0181552171707153, + "learning_rate": 7.065834468185573e-05, + "loss": 0.6014, + "step": 11428 + }, + { + "epoch": 0.767088352739841, + "grad_norm": 0.9061861038208008, + "learning_rate": 7.064844691280362e-05, + "loss": 0.5985, + "step": 11430 + }, + { + "epoch": 0.7672225764236099, + "grad_norm": 0.8315479159355164, + "learning_rate": 7.063854816815301e-05, + "loss": 0.5446, + "step": 11432 + }, + { + "epoch": 0.767356800107379, + "grad_norm": 1.047680377960205, + "learning_rate": 7.062864844837159e-05, + "loss": 0.6219, + "step": 11434 + }, + { + "epoch": 0.767491023791148, + "grad_norm": 1.0420489311218262, + "learning_rate": 7.061874775392709e-05, + "loss": 0.5858, + "step": 11436 + }, + { + "epoch": 0.7676252474749169, + "grad_norm": 1.1388205289840698, + "learning_rate": 7.060884608528728e-05, + "loss": 0.6973, + "step": 11438 + }, + { + "epoch": 0.7677594711586859, + "grad_norm": 1.8225839138031006, + "learning_rate": 7.059894344292004e-05, + "loss": 0.573, + "step": 11440 + }, + { + "epoch": 0.767893694842455, + "grad_norm": 1.0094853639602661, + "learning_rate": 7.058903982729322e-05, + "loss": 0.6169, + "step": 11442 + }, + { + "epoch": 0.768027918526224, + "grad_norm": 0.9954559803009033, + "learning_rate": 7.057913523887478e-05, + "loss": 0.5874, + "step": 11444 + }, + { + "epoch": 0.7681621422099929, + "grad_norm": 1.2659293413162231, + "learning_rate": 7.056922967813263e-05, + "loss": 0.6087, + "step": 11446 + }, + { + "epoch": 0.7682963658937619, + "grad_norm": 0.9722047448158264, + "learning_rate": 7.055932314553485e-05, + "loss": 0.6462, + "step": 11448 + }, + { + "epoch": 0.768430589577531, + "grad_norm": 1.0264939069747925, + "learning_rate": 7.054941564154946e-05, + "loss": 0.5979, + "step": 11450 + }, + { + "epoch": 0.7685648132613, + "grad_norm": 1.1204912662506104, + "learning_rate": 7.05395071666446e-05, + "loss": 0.6051, + "step": 11452 + }, + { + "epoch": 0.7686990369450689, + "grad_norm": 1.0123697519302368, + "learning_rate": 7.052959772128841e-05, + "loss": 0.6148, + "step": 11454 + }, + { + "epoch": 0.768833260628838, + "grad_norm": 0.9574852585792542, + "learning_rate": 7.051968730594911e-05, + "loss": 0.573, + "step": 11456 + }, + { + "epoch": 0.768967484312607, + "grad_norm": 1.3269104957580566, + "learning_rate": 7.050977592109494e-05, + "loss": 0.5905, + "step": 11458 + }, + { + "epoch": 0.769101707996376, + "grad_norm": 1.0462757349014282, + "learning_rate": 7.049986356719417e-05, + "loss": 0.6053, + "step": 11460 + }, + { + "epoch": 0.7692359316801449, + "grad_norm": 1.1457247734069824, + "learning_rate": 7.048995024471517e-05, + "loss": 0.6514, + "step": 11462 + }, + { + "epoch": 0.769370155363914, + "grad_norm": 1.0068777799606323, + "learning_rate": 7.048003595412632e-05, + "loss": 0.6148, + "step": 11464 + }, + { + "epoch": 0.769504379047683, + "grad_norm": 0.990300714969635, + "learning_rate": 7.047012069589601e-05, + "loss": 0.5711, + "step": 11466 + }, + { + "epoch": 0.769638602731452, + "grad_norm": 1.1073365211486816, + "learning_rate": 7.046020447049277e-05, + "loss": 0.6263, + "step": 11468 + }, + { + "epoch": 0.7697728264152209, + "grad_norm": 1.0013397932052612, + "learning_rate": 7.045028727838511e-05, + "loss": 0.5909, + "step": 11470 + }, + { + "epoch": 0.76990705009899, + "grad_norm": 1.0649570226669312, + "learning_rate": 7.044036912004159e-05, + "loss": 0.6719, + "step": 11472 + }, + { + "epoch": 0.770041273782759, + "grad_norm": 1.0973566770553589, + "learning_rate": 7.04304499959308e-05, + "loss": 0.5559, + "step": 11474 + }, + { + "epoch": 0.7701754974665279, + "grad_norm": 0.9822953343391418, + "learning_rate": 7.042052990652146e-05, + "loss": 0.6049, + "step": 11476 + }, + { + "epoch": 0.7703097211502969, + "grad_norm": 1.1304566860198975, + "learning_rate": 7.041060885228222e-05, + "loss": 0.6276, + "step": 11478 + }, + { + "epoch": 0.770443944834066, + "grad_norm": 1.0694489479064941, + "learning_rate": 7.040068683368181e-05, + "loss": 0.5635, + "step": 11480 + }, + { + "epoch": 0.770578168517835, + "grad_norm": 0.9529529809951782, + "learning_rate": 7.039076385118911e-05, + "loss": 0.5494, + "step": 11482 + }, + { + "epoch": 0.7707123922016039, + "grad_norm": 1.0704907178878784, + "learning_rate": 7.03808399052729e-05, + "loss": 0.6072, + "step": 11484 + }, + { + "epoch": 0.770846615885373, + "grad_norm": 1.017514944076538, + "learning_rate": 7.03709149964021e-05, + "loss": 0.5016, + "step": 11486 + }, + { + "epoch": 0.770980839569142, + "grad_norm": 1.1568889617919922, + "learning_rate": 7.036098912504559e-05, + "loss": 0.5852, + "step": 11488 + }, + { + "epoch": 0.771115063252911, + "grad_norm": 1.095329761505127, + "learning_rate": 7.035106229167241e-05, + "loss": 0.6255, + "step": 11490 + }, + { + "epoch": 0.7712492869366799, + "grad_norm": 1.0314277410507202, + "learning_rate": 7.034113449675154e-05, + "loss": 0.5831, + "step": 11492 + }, + { + "epoch": 0.771383510620449, + "grad_norm": 0.9504003524780273, + "learning_rate": 7.033120574075206e-05, + "loss": 0.6363, + "step": 11494 + }, + { + "epoch": 0.771517734304218, + "grad_norm": 1.0379869937896729, + "learning_rate": 7.032127602414311e-05, + "loss": 0.5973, + "step": 11496 + }, + { + "epoch": 0.771651957987987, + "grad_norm": 1.1097465753555298, + "learning_rate": 7.031134534739381e-05, + "loss": 0.6785, + "step": 11498 + }, + { + "epoch": 0.7717861816717559, + "grad_norm": 0.9349481463432312, + "learning_rate": 7.030141371097339e-05, + "loss": 0.5758, + "step": 11500 + }, + { + "epoch": 0.771920405355525, + "grad_norm": 1.1350663900375366, + "learning_rate": 7.029148111535109e-05, + "loss": 0.5973, + "step": 11502 + }, + { + "epoch": 0.772054629039294, + "grad_norm": 1.1020756959915161, + "learning_rate": 7.02815475609962e-05, + "loss": 0.5947, + "step": 11504 + }, + { + "epoch": 0.772188852723063, + "grad_norm": 0.9822752475738525, + "learning_rate": 7.02716130483781e-05, + "loss": 0.6771, + "step": 11506 + }, + { + "epoch": 0.772323076406832, + "grad_norm": 0.8996891379356384, + "learning_rate": 7.026167757796612e-05, + "loss": 0.5805, + "step": 11508 + }, + { + "epoch": 0.772457300090601, + "grad_norm": 0.9797414541244507, + "learning_rate": 7.025174115022972e-05, + "loss": 0.5829, + "step": 11510 + }, + { + "epoch": 0.77259152377437, + "grad_norm": 1.0092127323150635, + "learning_rate": 7.024180376563838e-05, + "loss": 0.5577, + "step": 11512 + }, + { + "epoch": 0.7727257474581389, + "grad_norm": 0.9922501444816589, + "learning_rate": 7.02318654246616e-05, + "loss": 0.5855, + "step": 11514 + }, + { + "epoch": 0.772859971141908, + "grad_norm": 0.8650910258293152, + "learning_rate": 7.022192612776899e-05, + "loss": 0.6531, + "step": 11516 + }, + { + "epoch": 0.772994194825677, + "grad_norm": 0.9783521890640259, + "learning_rate": 7.021198587543012e-05, + "loss": 0.5521, + "step": 11518 + }, + { + "epoch": 0.773128418509446, + "grad_norm": 1.0634944438934326, + "learning_rate": 7.020204466811467e-05, + "loss": 0.6848, + "step": 11520 + }, + { + "epoch": 0.7732626421932149, + "grad_norm": 1.15195631980896, + "learning_rate": 7.019210250629231e-05, + "loss": 0.5593, + "step": 11522 + }, + { + "epoch": 0.773396865876984, + "grad_norm": 1.0139727592468262, + "learning_rate": 7.018215939043285e-05, + "loss": 0.653, + "step": 11524 + }, + { + "epoch": 0.773531089560753, + "grad_norm": 1.0244086980819702, + "learning_rate": 7.017221532100601e-05, + "loss": 0.5606, + "step": 11526 + }, + { + "epoch": 0.773665313244522, + "grad_norm": 1.0112557411193848, + "learning_rate": 7.016227029848169e-05, + "loss": 0.5724, + "step": 11528 + }, + { + "epoch": 0.773799536928291, + "grad_norm": 1.0652763843536377, + "learning_rate": 7.015232432332974e-05, + "loss": 0.5757, + "step": 11530 + }, + { + "epoch": 0.77393376061206, + "grad_norm": 1.005967617034912, + "learning_rate": 7.014237739602008e-05, + "loss": 0.5454, + "step": 11532 + }, + { + "epoch": 0.774067984295829, + "grad_norm": 1.002595067024231, + "learning_rate": 7.01324295170227e-05, + "loss": 0.582, + "step": 11534 + }, + { + "epoch": 0.774202207979598, + "grad_norm": 0.9543070197105408, + "learning_rate": 7.012248068680762e-05, + "loss": 0.5883, + "step": 11536 + }, + { + "epoch": 0.774336431663367, + "grad_norm": 1.0744119882583618, + "learning_rate": 7.01125309058449e-05, + "loss": 0.6088, + "step": 11538 + }, + { + "epoch": 0.774470655347136, + "grad_norm": 2.2598273754119873, + "learning_rate": 7.010258017460463e-05, + "loss": 0.601, + "step": 11540 + }, + { + "epoch": 0.774604879030905, + "grad_norm": 0.9988208413124084, + "learning_rate": 7.0092628493557e-05, + "loss": 0.6396, + "step": 11542 + }, + { + "epoch": 0.774739102714674, + "grad_norm": 1.0017659664154053, + "learning_rate": 7.008267586317216e-05, + "loss": 0.6025, + "step": 11544 + }, + { + "epoch": 0.774873326398443, + "grad_norm": 1.0925278663635254, + "learning_rate": 7.007272228392039e-05, + "loss": 0.6301, + "step": 11546 + }, + { + "epoch": 0.775007550082212, + "grad_norm": 1.0574859380722046, + "learning_rate": 7.006276775627196e-05, + "loss": 0.6225, + "step": 11548 + }, + { + "epoch": 0.775141773765981, + "grad_norm": 1.3055510520935059, + "learning_rate": 7.005281228069721e-05, + "loss": 0.5825, + "step": 11550 + }, + { + "epoch": 0.77527599744975, + "grad_norm": 1.0374795198440552, + "learning_rate": 7.004285585766651e-05, + "loss": 0.6058, + "step": 11552 + }, + { + "epoch": 0.775410221133519, + "grad_norm": 1.1046421527862549, + "learning_rate": 7.003289848765028e-05, + "loss": 0.5908, + "step": 11554 + }, + { + "epoch": 0.775544444817288, + "grad_norm": 1.1111980676651, + "learning_rate": 7.002294017111899e-05, + "loss": 0.5679, + "step": 11556 + }, + { + "epoch": 0.775678668501057, + "grad_norm": 1.09193754196167, + "learning_rate": 7.001298090854316e-05, + "loss": 0.6047, + "step": 11558 + }, + { + "epoch": 0.775812892184826, + "grad_norm": 0.9132357835769653, + "learning_rate": 7.000302070039332e-05, + "loss": 0.5793, + "step": 11560 + }, + { + "epoch": 0.775947115868595, + "grad_norm": 1.1297905445098877, + "learning_rate": 6.999305954714009e-05, + "loss": 0.6307, + "step": 11562 + }, + { + "epoch": 0.776081339552364, + "grad_norm": 1.0000507831573486, + "learning_rate": 6.998309744925411e-05, + "loss": 0.6287, + "step": 11564 + }, + { + "epoch": 0.776215563236133, + "grad_norm": 1.1495134830474854, + "learning_rate": 6.997313440720608e-05, + "loss": 0.6487, + "step": 11566 + }, + { + "epoch": 0.776349786919902, + "grad_norm": 0.9105957746505737, + "learning_rate": 6.996317042146671e-05, + "loss": 0.5489, + "step": 11568 + }, + { + "epoch": 0.776484010603671, + "grad_norm": 0.9797573685646057, + "learning_rate": 6.995320549250681e-05, + "loss": 0.586, + "step": 11570 + }, + { + "epoch": 0.77661823428744, + "grad_norm": 1.1910641193389893, + "learning_rate": 6.99432396207972e-05, + "loss": 0.5437, + "step": 11572 + }, + { + "epoch": 0.7767524579712091, + "grad_norm": 1.0212557315826416, + "learning_rate": 6.99332728068087e-05, + "loss": 0.5356, + "step": 11574 + }, + { + "epoch": 0.776886681654978, + "grad_norm": 1.0056376457214355, + "learning_rate": 6.992330505101228e-05, + "loss": 0.5565, + "step": 11576 + }, + { + "epoch": 0.777020905338747, + "grad_norm": 0.9811734557151794, + "learning_rate": 6.991333635387886e-05, + "loss": 0.6, + "step": 11578 + }, + { + "epoch": 0.777155129022516, + "grad_norm": 0.9935258030891418, + "learning_rate": 6.990336671587946e-05, + "loss": 0.6373, + "step": 11580 + }, + { + "epoch": 0.7772893527062851, + "grad_norm": 1.10719895362854, + "learning_rate": 6.989339613748512e-05, + "loss": 0.6113, + "step": 11582 + }, + { + "epoch": 0.777423576390054, + "grad_norm": 0.9676934480667114, + "learning_rate": 6.988342461916693e-05, + "loss": 0.589, + "step": 11584 + }, + { + "epoch": 0.777557800073823, + "grad_norm": 1.0333850383758545, + "learning_rate": 6.987345216139604e-05, + "loss": 0.5752, + "step": 11586 + }, + { + "epoch": 0.777692023757592, + "grad_norm": 1.1050474643707275, + "learning_rate": 6.98634787646436e-05, + "loss": 0.615, + "step": 11588 + }, + { + "epoch": 0.777826247441361, + "grad_norm": 1.0079107284545898, + "learning_rate": 6.985350442938084e-05, + "loss": 0.5257, + "step": 11590 + }, + { + "epoch": 0.77796047112513, + "grad_norm": 1.1853986978530884, + "learning_rate": 6.984352915607906e-05, + "loss": 0.576, + "step": 11592 + }, + { + "epoch": 0.778094694808899, + "grad_norm": 0.8857834935188293, + "learning_rate": 6.983355294520952e-05, + "loss": 0.5174, + "step": 11594 + }, + { + "epoch": 0.7782289184926681, + "grad_norm": 1.2821314334869385, + "learning_rate": 6.982357579724364e-05, + "loss": 0.5895, + "step": 11596 + }, + { + "epoch": 0.778363142176437, + "grad_norm": 0.8643916249275208, + "learning_rate": 6.981359771265276e-05, + "loss": 0.5334, + "step": 11598 + }, + { + "epoch": 0.778497365860206, + "grad_norm": 1.3218395709991455, + "learning_rate": 6.980361869190836e-05, + "loss": 0.6501, + "step": 11600 + }, + { + "epoch": 0.778631589543975, + "grad_norm": 1.0390161275863647, + "learning_rate": 6.97936387354819e-05, + "loss": 0.531, + "step": 11602 + }, + { + "epoch": 0.7787658132277441, + "grad_norm": 0.9862406253814697, + "learning_rate": 6.978365784384494e-05, + "loss": 0.5516, + "step": 11604 + }, + { + "epoch": 0.778900036911513, + "grad_norm": 1.0138154029846191, + "learning_rate": 6.977367601746907e-05, + "loss": 0.6473, + "step": 11606 + }, + { + "epoch": 0.779034260595282, + "grad_norm": 0.956520676612854, + "learning_rate": 6.976369325682586e-05, + "loss": 0.568, + "step": 11608 + }, + { + "epoch": 0.779168484279051, + "grad_norm": 1.2190860509872437, + "learning_rate": 6.975370956238703e-05, + "loss": 0.5528, + "step": 11610 + }, + { + "epoch": 0.7793027079628201, + "grad_norm": 1.0392407178878784, + "learning_rate": 6.974372493462427e-05, + "loss": 0.6182, + "step": 11612 + }, + { + "epoch": 0.779436931646589, + "grad_norm": 1.3222901821136475, + "learning_rate": 6.973373937400932e-05, + "loss": 0.5887, + "step": 11614 + }, + { + "epoch": 0.779571155330358, + "grad_norm": 1.0793002843856812, + "learning_rate": 6.9723752881014e-05, + "loss": 0.598, + "step": 11616 + }, + { + "epoch": 0.7797053790141271, + "grad_norm": 1.046127200126648, + "learning_rate": 6.971376545611012e-05, + "loss": 0.5966, + "step": 11618 + }, + { + "epoch": 0.7798396026978961, + "grad_norm": 0.9392060041427612, + "learning_rate": 6.97037770997696e-05, + "loss": 0.568, + "step": 11620 + }, + { + "epoch": 0.779973826381665, + "grad_norm": 1.0161939859390259, + "learning_rate": 6.969378781246436e-05, + "loss": 0.5838, + "step": 11622 + }, + { + "epoch": 0.780108050065434, + "grad_norm": 1.3568873405456543, + "learning_rate": 6.968379759466638e-05, + "loss": 0.6808, + "step": 11624 + }, + { + "epoch": 0.7802422737492031, + "grad_norm": 1.0033005475997925, + "learning_rate": 6.967380644684765e-05, + "loss": 0.5981, + "step": 11626 + }, + { + "epoch": 0.780376497432972, + "grad_norm": 1.0781974792480469, + "learning_rate": 6.966381436948027e-05, + "loss": 0.6977, + "step": 11628 + }, + { + "epoch": 0.780510721116741, + "grad_norm": 1.0803115367889404, + "learning_rate": 6.965382136303632e-05, + "loss": 0.66, + "step": 11630 + }, + { + "epoch": 0.78064494480051, + "grad_norm": 1.0441747903823853, + "learning_rate": 6.964382742798797e-05, + "loss": 0.587, + "step": 11632 + }, + { + "epoch": 0.7807791684842791, + "grad_norm": 1.0430234670639038, + "learning_rate": 6.963383256480738e-05, + "loss": 0.5442, + "step": 11634 + }, + { + "epoch": 0.780913392168048, + "grad_norm": 0.972874641418457, + "learning_rate": 6.962383677396682e-05, + "loss": 0.6236, + "step": 11636 + }, + { + "epoch": 0.781047615851817, + "grad_norm": 1.1453553438186646, + "learning_rate": 6.961384005593856e-05, + "loss": 0.6472, + "step": 11638 + }, + { + "epoch": 0.7811818395355861, + "grad_norm": 1.0501232147216797, + "learning_rate": 6.960384241119494e-05, + "loss": 0.5658, + "step": 11640 + }, + { + "epoch": 0.7813160632193551, + "grad_norm": 1.000161051750183, + "learning_rate": 6.95938438402083e-05, + "loss": 0.5674, + "step": 11642 + }, + { + "epoch": 0.781450286903124, + "grad_norm": 1.0791254043579102, + "learning_rate": 6.958384434345107e-05, + "loss": 0.6232, + "step": 11644 + }, + { + "epoch": 0.781584510586893, + "grad_norm": 0.9885473847389221, + "learning_rate": 6.95738439213957e-05, + "loss": 0.6215, + "step": 11646 + }, + { + "epoch": 0.7817187342706621, + "grad_norm": 0.9427973031997681, + "learning_rate": 6.956384257451471e-05, + "loss": 0.5575, + "step": 11648 + }, + { + "epoch": 0.7818529579544311, + "grad_norm": 0.9807668924331665, + "learning_rate": 6.955384030328063e-05, + "loss": 0.6265, + "step": 11650 + }, + { + "epoch": 0.7819871816382, + "grad_norm": 1.0430842638015747, + "learning_rate": 6.954383710816604e-05, + "loss": 0.5747, + "step": 11652 + }, + { + "epoch": 0.782121405321969, + "grad_norm": 1.0898436307907104, + "learning_rate": 6.953383298964357e-05, + "loss": 0.617, + "step": 11654 + }, + { + "epoch": 0.7822556290057381, + "grad_norm": 1.0856739282608032, + "learning_rate": 6.95238279481859e-05, + "loss": 0.578, + "step": 11656 + }, + { + "epoch": 0.7823898526895071, + "grad_norm": 1.0635261535644531, + "learning_rate": 6.951382198426577e-05, + "loss": 0.6226, + "step": 11658 + }, + { + "epoch": 0.782524076373276, + "grad_norm": 1.0395147800445557, + "learning_rate": 6.95038150983559e-05, + "loss": 0.5009, + "step": 11660 + }, + { + "epoch": 0.7826583000570451, + "grad_norm": 1.0393999814987183, + "learning_rate": 6.949380729092914e-05, + "loss": 0.5967, + "step": 11662 + }, + { + "epoch": 0.7827925237408141, + "grad_norm": 0.9801155924797058, + "learning_rate": 6.948379856245832e-05, + "loss": 0.5237, + "step": 11664 + }, + { + "epoch": 0.782926747424583, + "grad_norm": 1.1089324951171875, + "learning_rate": 6.947378891341631e-05, + "loss": 0.5307, + "step": 11666 + }, + { + "epoch": 0.783060971108352, + "grad_norm": 0.8968009352684021, + "learning_rate": 6.946377834427608e-05, + "loss": 0.5496, + "step": 11668 + }, + { + "epoch": 0.7831951947921211, + "grad_norm": 1.186698317527771, + "learning_rate": 6.945376685551061e-05, + "loss": 0.6179, + "step": 11670 + }, + { + "epoch": 0.7833294184758901, + "grad_norm": 1.0249183177947998, + "learning_rate": 6.94437544475929e-05, + "loss": 0.6164, + "step": 11672 + }, + { + "epoch": 0.783463642159659, + "grad_norm": 0.9646841883659363, + "learning_rate": 6.9433741120996e-05, + "loss": 0.5585, + "step": 11674 + }, + { + "epoch": 0.783597865843428, + "grad_norm": 0.9436026215553284, + "learning_rate": 6.94237268761931e-05, + "loss": 0.5165, + "step": 11676 + }, + { + "epoch": 0.7837320895271971, + "grad_norm": 1.168487310409546, + "learning_rate": 6.941371171365725e-05, + "loss": 0.605, + "step": 11678 + }, + { + "epoch": 0.7838663132109661, + "grad_norm": 1.008285403251648, + "learning_rate": 6.940369563386172e-05, + "loss": 0.5822, + "step": 11680 + }, + { + "epoch": 0.784000536894735, + "grad_norm": 1.0759882926940918, + "learning_rate": 6.939367863727973e-05, + "loss": 0.6026, + "step": 11682 + }, + { + "epoch": 0.7841347605785041, + "grad_norm": 1.0713379383087158, + "learning_rate": 6.938366072438456e-05, + "loss": 0.6155, + "step": 11684 + }, + { + "epoch": 0.7842689842622731, + "grad_norm": 1.0884873867034912, + "learning_rate": 6.937364189564954e-05, + "loss": 0.6289, + "step": 11686 + }, + { + "epoch": 0.7844032079460421, + "grad_norm": 0.9686011672019958, + "learning_rate": 6.936362215154802e-05, + "loss": 0.6258, + "step": 11688 + }, + { + "epoch": 0.784537431629811, + "grad_norm": 1.0788530111312866, + "learning_rate": 6.935360149255345e-05, + "loss": 0.587, + "step": 11690 + }, + { + "epoch": 0.7846716553135801, + "grad_norm": 0.9957976937294006, + "learning_rate": 6.934357991913924e-05, + "loss": 0.6245, + "step": 11692 + }, + { + "epoch": 0.7848058789973491, + "grad_norm": 1.063419222831726, + "learning_rate": 6.933355743177894e-05, + "loss": 0.6002, + "step": 11694 + }, + { + "epoch": 0.7849401026811181, + "grad_norm": 1.0071430206298828, + "learning_rate": 6.932353403094605e-05, + "loss": 0.593, + "step": 11696 + }, + { + "epoch": 0.785074326364887, + "grad_norm": 1.0671712160110474, + "learning_rate": 6.931350971711418e-05, + "loss": 0.5449, + "step": 11698 + }, + { + "epoch": 0.7852085500486561, + "grad_norm": 1.0007480382919312, + "learning_rate": 6.930348449075699e-05, + "loss": 0.6136, + "step": 11700 + }, + { + "epoch": 0.7853427737324251, + "grad_norm": 1.0614603757858276, + "learning_rate": 6.929345835234808e-05, + "loss": 0.6491, + "step": 11702 + }, + { + "epoch": 0.785476997416194, + "grad_norm": 0.9814666509628296, + "learning_rate": 6.928343130236121e-05, + "loss": 0.5912, + "step": 11704 + }, + { + "epoch": 0.7856112210999631, + "grad_norm": 1.0447053909301758, + "learning_rate": 6.927340334127013e-05, + "loss": 0.6188, + "step": 11706 + }, + { + "epoch": 0.7857454447837321, + "grad_norm": 1.0697200298309326, + "learning_rate": 6.926337446954864e-05, + "loss": 0.5908, + "step": 11708 + }, + { + "epoch": 0.7858796684675011, + "grad_norm": 1.0311012268066406, + "learning_rate": 6.92533446876706e-05, + "loss": 0.6419, + "step": 11710 + }, + { + "epoch": 0.78601389215127, + "grad_norm": 1.0517666339874268, + "learning_rate": 6.924331399610986e-05, + "loss": 0.6235, + "step": 11712 + }, + { + "epoch": 0.7861481158350391, + "grad_norm": 1.0834307670593262, + "learning_rate": 6.92332823953404e-05, + "loss": 0.5855, + "step": 11714 + }, + { + "epoch": 0.7862823395188081, + "grad_norm": 0.8779531717300415, + "learning_rate": 6.922324988583616e-05, + "loss": 0.498, + "step": 11716 + }, + { + "epoch": 0.7864165632025771, + "grad_norm": 1.1895356178283691, + "learning_rate": 6.921321646807113e-05, + "loss": 0.5921, + "step": 11718 + }, + { + "epoch": 0.786550786886346, + "grad_norm": 1.094163417816162, + "learning_rate": 6.920318214251945e-05, + "loss": 0.5684, + "step": 11720 + }, + { + "epoch": 0.7866850105701151, + "grad_norm": 1.262536644935608, + "learning_rate": 6.919314690965514e-05, + "loss": 0.5357, + "step": 11722 + }, + { + "epoch": 0.7868192342538841, + "grad_norm": 1.0477513074874878, + "learning_rate": 6.91831107699524e-05, + "loss": 0.5262, + "step": 11724 + }, + { + "epoch": 0.7869534579376531, + "grad_norm": 0.9876933097839355, + "learning_rate": 6.917307372388539e-05, + "loss": 0.5662, + "step": 11726 + }, + { + "epoch": 0.7870876816214221, + "grad_norm": 0.8962420225143433, + "learning_rate": 6.916303577192835e-05, + "loss": 0.5427, + "step": 11728 + }, + { + "epoch": 0.7872219053051911, + "grad_norm": 0.9462629556655884, + "learning_rate": 6.915299691455555e-05, + "loss": 0.6347, + "step": 11730 + }, + { + "epoch": 0.7873561289889601, + "grad_norm": 0.998128354549408, + "learning_rate": 6.914295715224132e-05, + "loss": 0.5965, + "step": 11732 + }, + { + "epoch": 0.7874903526727292, + "grad_norm": 1.0247682332992554, + "learning_rate": 6.913291648546001e-05, + "loss": 0.611, + "step": 11734 + }, + { + "epoch": 0.7876245763564981, + "grad_norm": 0.9905968308448792, + "learning_rate": 6.9122874914686e-05, + "loss": 0.5954, + "step": 11736 + }, + { + "epoch": 0.7877588000402671, + "grad_norm": 1.0264225006103516, + "learning_rate": 6.911283244039377e-05, + "loss": 0.5826, + "step": 11738 + }, + { + "epoch": 0.7878930237240361, + "grad_norm": 0.9632094502449036, + "learning_rate": 6.910278906305778e-05, + "loss": 0.5702, + "step": 11740 + }, + { + "epoch": 0.788027247407805, + "grad_norm": 1.104326605796814, + "learning_rate": 6.909274478315257e-05, + "loss": 0.5727, + "step": 11742 + }, + { + "epoch": 0.7881614710915741, + "grad_norm": 0.9858546257019043, + "learning_rate": 6.908269960115273e-05, + "loss": 0.6141, + "step": 11744 + }, + { + "epoch": 0.7882956947753431, + "grad_norm": 0.9897134900093079, + "learning_rate": 6.907265351753283e-05, + "loss": 0.5913, + "step": 11746 + }, + { + "epoch": 0.7884299184591121, + "grad_norm": 1.038590431213379, + "learning_rate": 6.906260653276758e-05, + "loss": 0.6159, + "step": 11748 + }, + { + "epoch": 0.7885641421428811, + "grad_norm": 1.0222185850143433, + "learning_rate": 6.905255864733164e-05, + "loss": 0.5529, + "step": 11750 + }, + { + "epoch": 0.7886983658266501, + "grad_norm": 1.1252750158309937, + "learning_rate": 6.90425098616998e-05, + "loss": 0.5685, + "step": 11752 + }, + { + "epoch": 0.7888325895104191, + "grad_norm": 0.9735110998153687, + "learning_rate": 6.903246017634677e-05, + "loss": 0.5798, + "step": 11754 + }, + { + "epoch": 0.7889668131941882, + "grad_norm": 1.099358320236206, + "learning_rate": 6.902240959174745e-05, + "loss": 0.6107, + "step": 11756 + }, + { + "epoch": 0.7891010368779571, + "grad_norm": 1.0541526079177856, + "learning_rate": 6.901235810837669e-05, + "loss": 0.5444, + "step": 11758 + }, + { + "epoch": 0.7892352605617261, + "grad_norm": 1.0630377531051636, + "learning_rate": 6.900230572670938e-05, + "loss": 0.6276, + "step": 11760 + }, + { + "epoch": 0.7893694842454951, + "grad_norm": 1.0359013080596924, + "learning_rate": 6.89922524472205e-05, + "loss": 0.5934, + "step": 11762 + }, + { + "epoch": 0.7895037079292642, + "grad_norm": 0.8728708028793335, + "learning_rate": 6.898219827038503e-05, + "loss": 0.5087, + "step": 11764 + }, + { + "epoch": 0.7896379316130331, + "grad_norm": 0.9688023924827576, + "learning_rate": 6.897214319667802e-05, + "loss": 0.5695, + "step": 11766 + }, + { + "epoch": 0.7897721552968021, + "grad_norm": 1.2484382390975952, + "learning_rate": 6.896208722657455e-05, + "loss": 0.6423, + "step": 11768 + }, + { + "epoch": 0.7899063789805711, + "grad_norm": 0.9647979140281677, + "learning_rate": 6.895203036054974e-05, + "loss": 0.5779, + "step": 11770 + }, + { + "epoch": 0.7900406026643402, + "grad_norm": 1.1346417665481567, + "learning_rate": 6.894197259907879e-05, + "loss": 0.5259, + "step": 11772 + }, + { + "epoch": 0.7901748263481091, + "grad_norm": 1.1269654035568237, + "learning_rate": 6.893191394263684e-05, + "loss": 0.5567, + "step": 11774 + }, + { + "epoch": 0.7903090500318781, + "grad_norm": 1.0334393978118896, + "learning_rate": 6.892185439169922e-05, + "loss": 0.6355, + "step": 11776 + }, + { + "epoch": 0.7904432737156472, + "grad_norm": 1.088161826133728, + "learning_rate": 6.891179394674119e-05, + "loss": 0.5314, + "step": 11778 + }, + { + "epoch": 0.7905774973994161, + "grad_norm": 1.1312185525894165, + "learning_rate": 6.890173260823807e-05, + "loss": 0.5725, + "step": 11780 + }, + { + "epoch": 0.7907117210831851, + "grad_norm": 0.9361488819122314, + "learning_rate": 6.889167037666525e-05, + "loss": 0.5429, + "step": 11782 + }, + { + "epoch": 0.7908459447669541, + "grad_norm": 1.098042607307434, + "learning_rate": 6.888160725249816e-05, + "loss": 0.6652, + "step": 11784 + }, + { + "epoch": 0.7909801684507232, + "grad_norm": 1.0583075284957886, + "learning_rate": 6.887154323621225e-05, + "loss": 0.6155, + "step": 11786 + }, + { + "epoch": 0.7911143921344921, + "grad_norm": 0.8496408462524414, + "learning_rate": 6.886147832828303e-05, + "loss": 0.6059, + "step": 11788 + }, + { + "epoch": 0.7912486158182611, + "grad_norm": 1.0364539623260498, + "learning_rate": 6.885141252918607e-05, + "loss": 0.6223, + "step": 11790 + }, + { + "epoch": 0.7913828395020301, + "grad_norm": 0.9641837477684021, + "learning_rate": 6.884134583939692e-05, + "loss": 0.6216, + "step": 11792 + }, + { + "epoch": 0.7915170631857992, + "grad_norm": 1.2971800565719604, + "learning_rate": 6.883127825939122e-05, + "loss": 0.6559, + "step": 11794 + }, + { + "epoch": 0.7916512868695681, + "grad_norm": 1.0435941219329834, + "learning_rate": 6.882120978964466e-05, + "loss": 0.5561, + "step": 11796 + }, + { + "epoch": 0.7917855105533371, + "grad_norm": 0.9846218824386597, + "learning_rate": 6.881114043063296e-05, + "loss": 0.612, + "step": 11798 + }, + { + "epoch": 0.7919197342371062, + "grad_norm": 1.1445139646530151, + "learning_rate": 6.880107018283186e-05, + "loss": 0.5746, + "step": 11800 + }, + { + "epoch": 0.7920539579208752, + "grad_norm": 1.0392295122146606, + "learning_rate": 6.879099904671715e-05, + "loss": 0.5904, + "step": 11802 + }, + { + "epoch": 0.7921881816046441, + "grad_norm": 0.9261115789413452, + "learning_rate": 6.87809270227647e-05, + "loss": 0.4898, + "step": 11804 + }, + { + "epoch": 0.7923224052884131, + "grad_norm": 1.0234037637710571, + "learning_rate": 6.877085411145038e-05, + "loss": 0.5977, + "step": 11806 + }, + { + "epoch": 0.7924566289721822, + "grad_norm": 1.1201796531677246, + "learning_rate": 6.87607803132501e-05, + "loss": 0.5648, + "step": 11808 + }, + { + "epoch": 0.7925908526559512, + "grad_norm": 1.0571445226669312, + "learning_rate": 6.875070562863986e-05, + "loss": 0.6202, + "step": 11810 + }, + { + "epoch": 0.7927250763397201, + "grad_norm": 1.1081063747406006, + "learning_rate": 6.874063005809563e-05, + "loss": 0.56, + "step": 11812 + }, + { + "epoch": 0.7928593000234891, + "grad_norm": 1.0794700384140015, + "learning_rate": 6.87305536020935e-05, + "loss": 0.5982, + "step": 11814 + }, + { + "epoch": 0.7929935237072582, + "grad_norm": 1.0285453796386719, + "learning_rate": 6.872047626110955e-05, + "loss": 0.5858, + "step": 11816 + }, + { + "epoch": 0.7931277473910271, + "grad_norm": 1.135297417640686, + "learning_rate": 6.87103980356199e-05, + "loss": 0.5713, + "step": 11818 + }, + { + "epoch": 0.7932619710747961, + "grad_norm": 1.0952262878417969, + "learning_rate": 6.870031892610073e-05, + "loss": 0.5671, + "step": 11820 + }, + { + "epoch": 0.7933961947585652, + "grad_norm": 1.0698264837265015, + "learning_rate": 6.869023893302826e-05, + "loss": 0.5819, + "step": 11822 + }, + { + "epoch": 0.7935304184423342, + "grad_norm": 0.984652042388916, + "learning_rate": 6.868015805687877e-05, + "loss": 0.5566, + "step": 11824 + }, + { + "epoch": 0.7936646421261031, + "grad_norm": 1.0963594913482666, + "learning_rate": 6.867007629812852e-05, + "loss": 0.7261, + "step": 11826 + }, + { + "epoch": 0.7937988658098721, + "grad_norm": 0.9263544678688049, + "learning_rate": 6.865999365725391e-05, + "loss": 0.576, + "step": 11828 + }, + { + "epoch": 0.7939330894936412, + "grad_norm": 1.1272345781326294, + "learning_rate": 6.864991013473125e-05, + "loss": 0.5807, + "step": 11830 + }, + { + "epoch": 0.7940673131774102, + "grad_norm": 0.939246416091919, + "learning_rate": 6.863982573103704e-05, + "loss": 0.5666, + "step": 11832 + }, + { + "epoch": 0.7942015368611791, + "grad_norm": 1.0075976848602295, + "learning_rate": 6.862974044664772e-05, + "loss": 0.5334, + "step": 11834 + }, + { + "epoch": 0.7943357605449481, + "grad_norm": 1.0745658874511719, + "learning_rate": 6.861965428203978e-05, + "loss": 0.603, + "step": 11836 + }, + { + "epoch": 0.7944699842287172, + "grad_norm": 1.0120166540145874, + "learning_rate": 6.860956723768981e-05, + "loss": 0.6386, + "step": 11838 + }, + { + "epoch": 0.7946042079124862, + "grad_norm": 0.9060064554214478, + "learning_rate": 6.859947931407436e-05, + "loss": 0.5112, + "step": 11840 + }, + { + "epoch": 0.7947384315962551, + "grad_norm": 0.973538339138031, + "learning_rate": 6.858939051167011e-05, + "loss": 0.6096, + "step": 11842 + }, + { + "epoch": 0.7948726552800242, + "grad_norm": 1.0098037719726562, + "learning_rate": 6.85793008309537e-05, + "loss": 0.624, + "step": 11844 + }, + { + "epoch": 0.7950068789637932, + "grad_norm": 1.023622751235962, + "learning_rate": 6.856921027240187e-05, + "loss": 0.5849, + "step": 11846 + }, + { + "epoch": 0.7951411026475622, + "grad_norm": 1.0303845405578613, + "learning_rate": 6.855911883649137e-05, + "loss": 0.616, + "step": 11848 + }, + { + "epoch": 0.7952753263313311, + "grad_norm": 1.1904534101486206, + "learning_rate": 6.854902652369898e-05, + "loss": 0.7265, + "step": 11850 + }, + { + "epoch": 0.7954095500151002, + "grad_norm": 0.9215459823608398, + "learning_rate": 6.853893333450158e-05, + "loss": 0.5722, + "step": 11852 + }, + { + "epoch": 0.7955437736988692, + "grad_norm": 0.9743821620941162, + "learning_rate": 6.852883926937602e-05, + "loss": 0.608, + "step": 11854 + }, + { + "epoch": 0.7956779973826381, + "grad_norm": 1.222190260887146, + "learning_rate": 6.851874432879925e-05, + "loss": 0.6626, + "step": 11856 + }, + { + "epoch": 0.7958122210664071, + "grad_norm": 1.0577031373977661, + "learning_rate": 6.850864851324823e-05, + "loss": 0.5342, + "step": 11858 + }, + { + "epoch": 0.7959464447501762, + "grad_norm": 0.9763243794441223, + "learning_rate": 6.849855182319995e-05, + "loss": 0.6014, + "step": 11860 + }, + { + "epoch": 0.7960806684339452, + "grad_norm": 1.0097098350524902, + "learning_rate": 6.848845425913149e-05, + "loss": 0.6697, + "step": 11862 + }, + { + "epoch": 0.7962148921177141, + "grad_norm": 1.0537142753601074, + "learning_rate": 6.84783558215199e-05, + "loss": 0.6646, + "step": 11864 + }, + { + "epoch": 0.7963491158014832, + "grad_norm": 1.0196220874786377, + "learning_rate": 6.846825651084236e-05, + "loss": 0.6225, + "step": 11866 + }, + { + "epoch": 0.7964833394852522, + "grad_norm": 0.8699311017990112, + "learning_rate": 6.8458156327576e-05, + "loss": 0.5719, + "step": 11868 + }, + { + "epoch": 0.7966175631690212, + "grad_norm": 0.9252583980560303, + "learning_rate": 6.844805527219804e-05, + "loss": 0.5961, + "step": 11870 + }, + { + "epoch": 0.7967517868527901, + "grad_norm": 1.1307587623596191, + "learning_rate": 6.843795334518576e-05, + "loss": 0.5956, + "step": 11872 + }, + { + "epoch": 0.7968860105365592, + "grad_norm": 0.9717182517051697, + "learning_rate": 6.842785054701643e-05, + "loss": 0.6711, + "step": 11874 + }, + { + "epoch": 0.7970202342203282, + "grad_norm": 1.0717628002166748, + "learning_rate": 6.84177468781674e-05, + "loss": 0.5864, + "step": 11876 + }, + { + "epoch": 0.7971544579040972, + "grad_norm": 1.231685996055603, + "learning_rate": 6.840764233911606e-05, + "loss": 0.5761, + "step": 11878 + }, + { + "epoch": 0.7972886815878661, + "grad_norm": 0.9946985840797424, + "learning_rate": 6.83975369303398e-05, + "loss": 0.5445, + "step": 11880 + }, + { + "epoch": 0.7974229052716352, + "grad_norm": 1.0769169330596924, + "learning_rate": 6.838743065231612e-05, + "loss": 0.649, + "step": 11882 + }, + { + "epoch": 0.7975571289554042, + "grad_norm": 0.9944827556610107, + "learning_rate": 6.837732350552249e-05, + "loss": 0.5963, + "step": 11884 + }, + { + "epoch": 0.7976913526391732, + "grad_norm": 1.077628254890442, + "learning_rate": 6.836721549043645e-05, + "loss": 0.5909, + "step": 11886 + }, + { + "epoch": 0.7978255763229422, + "grad_norm": 1.2306236028671265, + "learning_rate": 6.835710660753561e-05, + "loss": 0.5751, + "step": 11888 + }, + { + "epoch": 0.7979598000067112, + "grad_norm": 0.972992479801178, + "learning_rate": 6.834699685729757e-05, + "loss": 0.5889, + "step": 11890 + }, + { + "epoch": 0.7980940236904802, + "grad_norm": 0.9339223504066467, + "learning_rate": 6.83368862402e-05, + "loss": 0.5542, + "step": 11892 + }, + { + "epoch": 0.7982282473742491, + "grad_norm": 1.0817885398864746, + "learning_rate": 6.832677475672063e-05, + "loss": 0.6131, + "step": 11894 + }, + { + "epoch": 0.7983624710580182, + "grad_norm": 1.1673030853271484, + "learning_rate": 6.831666240733718e-05, + "loss": 0.6095, + "step": 11896 + }, + { + "epoch": 0.7984966947417872, + "grad_norm": 1.21226966381073, + "learning_rate": 6.830654919252745e-05, + "loss": 0.6406, + "step": 11898 + }, + { + "epoch": 0.7986309184255562, + "grad_norm": 1.0874611139297485, + "learning_rate": 6.829643511276929e-05, + "loss": 0.6048, + "step": 11900 + }, + { + "epoch": 0.7987651421093251, + "grad_norm": 0.9976110458374023, + "learning_rate": 6.828632016854051e-05, + "loss": 0.5256, + "step": 11902 + }, + { + "epoch": 0.7988993657930942, + "grad_norm": 0.9796349406242371, + "learning_rate": 6.82762043603191e-05, + "loss": 0.5838, + "step": 11904 + }, + { + "epoch": 0.7990335894768632, + "grad_norm": 0.9170714616775513, + "learning_rate": 6.826608768858294e-05, + "loss": 0.5912, + "step": 11906 + }, + { + "epoch": 0.7991678131606322, + "grad_norm": 1.0391581058502197, + "learning_rate": 6.825597015381007e-05, + "loss": 0.5676, + "step": 11908 + }, + { + "epoch": 0.7993020368444012, + "grad_norm": 1.0441548824310303, + "learning_rate": 6.824585175647852e-05, + "loss": 0.6165, + "step": 11910 + }, + { + "epoch": 0.7994362605281702, + "grad_norm": 0.9074247479438782, + "learning_rate": 6.823573249706634e-05, + "loss": 0.5816, + "step": 11912 + }, + { + "epoch": 0.7995704842119392, + "grad_norm": 1.03797447681427, + "learning_rate": 6.822561237605167e-05, + "loss": 0.5578, + "step": 11914 + }, + { + "epoch": 0.7997047078957082, + "grad_norm": 1.083938717842102, + "learning_rate": 6.821549139391264e-05, + "loss": 0.6759, + "step": 11916 + }, + { + "epoch": 0.7998389315794772, + "grad_norm": 0.8666544556617737, + "learning_rate": 6.820536955112747e-05, + "loss": 0.5735, + "step": 11918 + }, + { + "epoch": 0.7999731552632462, + "grad_norm": 1.1249221563339233, + "learning_rate": 6.819524684817438e-05, + "loss": 0.6596, + "step": 11920 + }, + { + "epoch": 0.8001073789470152, + "grad_norm": 1.0319015979766846, + "learning_rate": 6.818512328553166e-05, + "loss": 0.6022, + "step": 11922 + }, + { + "epoch": 0.8002416026307843, + "grad_norm": 1.0916059017181396, + "learning_rate": 6.817499886367763e-05, + "loss": 0.5961, + "step": 11924 + }, + { + "epoch": 0.8003758263145532, + "grad_norm": 1.0699912309646606, + "learning_rate": 6.816487358309064e-05, + "loss": 0.6445, + "step": 11926 + }, + { + "epoch": 0.8005100499983222, + "grad_norm": 1.6955249309539795, + "learning_rate": 6.815474744424908e-05, + "loss": 0.5892, + "step": 11928 + }, + { + "epoch": 0.8006442736820912, + "grad_norm": 1.1446768045425415, + "learning_rate": 6.814462044763143e-05, + "loss": 0.6042, + "step": 11930 + }, + { + "epoch": 0.8007784973658602, + "grad_norm": 0.95870441198349, + "learning_rate": 6.813449259371611e-05, + "loss": 0.6115, + "step": 11932 + }, + { + "epoch": 0.8009127210496292, + "grad_norm": 1.0400450229644775, + "learning_rate": 6.81243638829817e-05, + "loss": 0.6142, + "step": 11934 + }, + { + "epoch": 0.8010469447333982, + "grad_norm": 0.9809623956680298, + "learning_rate": 6.811423431590672e-05, + "loss": 0.6326, + "step": 11936 + }, + { + "epoch": 0.8011811684171672, + "grad_norm": 1.0043615102767944, + "learning_rate": 6.81041038929698e-05, + "loss": 0.5786, + "step": 11938 + }, + { + "epoch": 0.8013153921009362, + "grad_norm": 1.0907158851623535, + "learning_rate": 6.809397261464957e-05, + "loss": 0.5435, + "step": 11940 + }, + { + "epoch": 0.8014496157847052, + "grad_norm": 1.8366481065750122, + "learning_rate": 6.808384048142472e-05, + "loss": 0.5893, + "step": 11942 + }, + { + "epoch": 0.8015838394684742, + "grad_norm": 0.9586851596832275, + "learning_rate": 6.807370749377396e-05, + "loss": 0.5757, + "step": 11944 + }, + { + "epoch": 0.8017180631522433, + "grad_norm": 0.9573516845703125, + "learning_rate": 6.806357365217606e-05, + "loss": 0.5899, + "step": 11946 + }, + { + "epoch": 0.8018522868360122, + "grad_norm": 1.194173812866211, + "learning_rate": 6.805343895710983e-05, + "loss": 0.5861, + "step": 11948 + }, + { + "epoch": 0.8019865105197812, + "grad_norm": 0.9803144931793213, + "learning_rate": 6.80433034090541e-05, + "loss": 0.5297, + "step": 11950 + }, + { + "epoch": 0.8021207342035502, + "grad_norm": 1.0402456521987915, + "learning_rate": 6.803316700848779e-05, + "loss": 0.6527, + "step": 11952 + }, + { + "epoch": 0.8022549578873193, + "grad_norm": 1.0439958572387695, + "learning_rate": 6.802302975588976e-05, + "loss": 0.5892, + "step": 11954 + }, + { + "epoch": 0.8023891815710882, + "grad_norm": 1.0565516948699951, + "learning_rate": 6.801289165173905e-05, + "loss": 0.5179, + "step": 11956 + }, + { + "epoch": 0.8025234052548572, + "grad_norm": 1.2166541814804077, + "learning_rate": 6.800275269651462e-05, + "loss": 0.7249, + "step": 11958 + }, + { + "epoch": 0.8026576289386262, + "grad_norm": 1.0121784210205078, + "learning_rate": 6.79926128906955e-05, + "loss": 0.5717, + "step": 11960 + }, + { + "epoch": 0.8027918526223953, + "grad_norm": 1.14600670337677, + "learning_rate": 6.798247223476084e-05, + "loss": 0.6073, + "step": 11962 + }, + { + "epoch": 0.8029260763061642, + "grad_norm": 1.074245572090149, + "learning_rate": 6.79723307291897e-05, + "loss": 0.6504, + "step": 11964 + }, + { + "epoch": 0.8030602999899332, + "grad_norm": 1.091021180152893, + "learning_rate": 6.79621883744613e-05, + "loss": 0.5784, + "step": 11966 + }, + { + "epoch": 0.8031945236737023, + "grad_norm": 0.9298738837242126, + "learning_rate": 6.79520451710548e-05, + "loss": 0.5882, + "step": 11968 + }, + { + "epoch": 0.8033287473574712, + "grad_norm": 1.0967477560043335, + "learning_rate": 6.794190111944948e-05, + "loss": 0.564, + "step": 11970 + }, + { + "epoch": 0.8034629710412402, + "grad_norm": 0.9110896587371826, + "learning_rate": 6.79317562201246e-05, + "loss": 0.5958, + "step": 11972 + }, + { + "epoch": 0.8035971947250092, + "grad_norm": 1.0178650617599487, + "learning_rate": 6.792161047355951e-05, + "loss": 0.577, + "step": 11974 + }, + { + "epoch": 0.8037314184087783, + "grad_norm": 1.421694040298462, + "learning_rate": 6.791146388023356e-05, + "loss": 0.6112, + "step": 11976 + }, + { + "epoch": 0.8038656420925472, + "grad_norm": 1.0152606964111328, + "learning_rate": 6.790131644062616e-05, + "loss": 0.6533, + "step": 11978 + }, + { + "epoch": 0.8039998657763162, + "grad_norm": 0.9436149001121521, + "learning_rate": 6.789116815521678e-05, + "loss": 0.6309, + "step": 11980 + }, + { + "epoch": 0.8041340894600852, + "grad_norm": 1.0372847318649292, + "learning_rate": 6.788101902448486e-05, + "loss": 0.519, + "step": 11982 + }, + { + "epoch": 0.8042683131438543, + "grad_norm": 0.9170295000076294, + "learning_rate": 6.787086904890998e-05, + "loss": 0.5786, + "step": 11984 + }, + { + "epoch": 0.8044025368276232, + "grad_norm": 1.0059316158294678, + "learning_rate": 6.786071822897166e-05, + "loss": 0.5635, + "step": 11986 + }, + { + "epoch": 0.8045367605113922, + "grad_norm": 0.9609963893890381, + "learning_rate": 6.785056656514953e-05, + "loss": 0.584, + "step": 11988 + }, + { + "epoch": 0.8046709841951613, + "grad_norm": 0.9653461575508118, + "learning_rate": 6.784041405792324e-05, + "loss": 0.5671, + "step": 11990 + }, + { + "epoch": 0.8048052078789303, + "grad_norm": 0.8587554693222046, + "learning_rate": 6.783026070777245e-05, + "loss": 0.5294, + "step": 11992 + }, + { + "epoch": 0.8049394315626992, + "grad_norm": 1.9127076864242554, + "learning_rate": 6.782010651517691e-05, + "loss": 0.5974, + "step": 11994 + }, + { + "epoch": 0.8050736552464682, + "grad_norm": 1.0242037773132324, + "learning_rate": 6.780995148061638e-05, + "loss": 0.5632, + "step": 11996 + }, + { + "epoch": 0.8052078789302373, + "grad_norm": 1.1092298030853271, + "learning_rate": 6.779979560457066e-05, + "loss": 0.6036, + "step": 11998 + }, + { + "epoch": 0.8053421026140063, + "grad_norm": 1.0180758237838745, + "learning_rate": 6.778963888751961e-05, + "loss": 0.6109, + "step": 12000 + }, + { + "epoch": 0.8054763262977752, + "grad_norm": 1.0010555982589722, + "learning_rate": 6.77794813299431e-05, + "loss": 0.5716, + "step": 12002 + }, + { + "epoch": 0.8056105499815442, + "grad_norm": 1.098800539970398, + "learning_rate": 6.776932293232106e-05, + "loss": 0.552, + "step": 12004 + }, + { + "epoch": 0.8057447736653133, + "grad_norm": 0.9850975275039673, + "learning_rate": 6.775916369513344e-05, + "loss": 0.629, + "step": 12006 + }, + { + "epoch": 0.8058789973490822, + "grad_norm": 1.0020534992218018, + "learning_rate": 6.774900361886028e-05, + "loss": 0.5255, + "step": 12008 + }, + { + "epoch": 0.8060132210328512, + "grad_norm": 1.052212119102478, + "learning_rate": 6.773884270398158e-05, + "loss": 0.6533, + "step": 12010 + }, + { + "epoch": 0.8061474447166203, + "grad_norm": 1.0798852443695068, + "learning_rate": 6.772868095097745e-05, + "loss": 0.5716, + "step": 12012 + }, + { + "epoch": 0.8062816684003893, + "grad_norm": 0.8810888528823853, + "learning_rate": 6.771851836032801e-05, + "loss": 0.5025, + "step": 12014 + }, + { + "epoch": 0.8064158920841582, + "grad_norm": 1.1676989793777466, + "learning_rate": 6.770835493251342e-05, + "loss": 0.6193, + "step": 12016 + }, + { + "epoch": 0.8065501157679272, + "grad_norm": 0.9496458172798157, + "learning_rate": 6.769819066801388e-05, + "loss": 0.5612, + "step": 12018 + }, + { + "epoch": 0.8066843394516963, + "grad_norm": 0.988442063331604, + "learning_rate": 6.768802556730964e-05, + "loss": 0.5982, + "step": 12020 + }, + { + "epoch": 0.8068185631354653, + "grad_norm": 1.0732218027114868, + "learning_rate": 6.767785963088096e-05, + "loss": 0.5754, + "step": 12022 + }, + { + "epoch": 0.8069527868192342, + "grad_norm": 1.0232852697372437, + "learning_rate": 6.766769285920819e-05, + "loss": 0.605, + "step": 12024 + }, + { + "epoch": 0.8070870105030032, + "grad_norm": 0.9692088961601257, + "learning_rate": 6.765752525277168e-05, + "loss": 0.5416, + "step": 12026 + }, + { + "epoch": 0.8072212341867723, + "grad_norm": 0.9061393737792969, + "learning_rate": 6.76473568120518e-05, + "loss": 0.5389, + "step": 12028 + }, + { + "epoch": 0.8073554578705413, + "grad_norm": 0.9491109848022461, + "learning_rate": 6.763718753752901e-05, + "loss": 0.5283, + "step": 12030 + }, + { + "epoch": 0.8074896815543102, + "grad_norm": 1.0774158239364624, + "learning_rate": 6.762701742968382e-05, + "loss": 0.6744, + "step": 12032 + }, + { + "epoch": 0.8076239052380793, + "grad_norm": 0.84043949842453, + "learning_rate": 6.761684648899669e-05, + "loss": 0.5901, + "step": 12034 + }, + { + "epoch": 0.8077581289218483, + "grad_norm": 1.1024222373962402, + "learning_rate": 6.760667471594821e-05, + "loss": 0.6299, + "step": 12036 + }, + { + "epoch": 0.8078923526056173, + "grad_norm": 1.4508450031280518, + "learning_rate": 6.7596502111019e-05, + "loss": 0.5974, + "step": 12038 + }, + { + "epoch": 0.8080265762893862, + "grad_norm": 1.3109294176101685, + "learning_rate": 6.758632867468964e-05, + "loss": 0.5406, + "step": 12040 + }, + { + "epoch": 0.8081607999731553, + "grad_norm": 1.1407095193862915, + "learning_rate": 6.757615440744084e-05, + "loss": 0.5489, + "step": 12042 + }, + { + "epoch": 0.8082950236569243, + "grad_norm": 0.943513035774231, + "learning_rate": 6.756597930975331e-05, + "loss": 0.5673, + "step": 12044 + }, + { + "epoch": 0.8084292473406932, + "grad_norm": 1.1445965766906738, + "learning_rate": 6.75558033821078e-05, + "loss": 0.5903, + "step": 12046 + }, + { + "epoch": 0.8085634710244622, + "grad_norm": 1.021998643875122, + "learning_rate": 6.754562662498509e-05, + "loss": 0.5976, + "step": 12048 + }, + { + "epoch": 0.8086976947082313, + "grad_norm": 1.0419927835464478, + "learning_rate": 6.753544903886602e-05, + "loss": 0.5785, + "step": 12050 + }, + { + "epoch": 0.8088319183920003, + "grad_norm": 0.8495463728904724, + "learning_rate": 6.75252706242315e-05, + "loss": 0.5681, + "step": 12052 + }, + { + "epoch": 0.8089661420757692, + "grad_norm": 0.9950043559074402, + "learning_rate": 6.751509138156239e-05, + "loss": 0.5758, + "step": 12054 + }, + { + "epoch": 0.8091003657595383, + "grad_norm": 0.9346325397491455, + "learning_rate": 6.750491131133962e-05, + "loss": 0.6295, + "step": 12056 + }, + { + "epoch": 0.8092345894433073, + "grad_norm": 1.0494962930679321, + "learning_rate": 6.749473041404424e-05, + "loss": 0.5382, + "step": 12058 + }, + { + "epoch": 0.8093688131270763, + "grad_norm": 0.9695845246315002, + "learning_rate": 6.748454869015725e-05, + "loss": 0.5711, + "step": 12060 + }, + { + "epoch": 0.8095030368108452, + "grad_norm": 1.1978754997253418, + "learning_rate": 6.747436614015972e-05, + "loss": 0.6299, + "step": 12062 + }, + { + "epoch": 0.8096372604946143, + "grad_norm": 0.9886444807052612, + "learning_rate": 6.746418276453275e-05, + "loss": 0.5606, + "step": 12064 + }, + { + "epoch": 0.8097714841783833, + "grad_norm": 1.180014729499817, + "learning_rate": 6.745399856375749e-05, + "loss": 0.6357, + "step": 12066 + }, + { + "epoch": 0.8099057078621523, + "grad_norm": 0.8843987584114075, + "learning_rate": 6.74438135383151e-05, + "loss": 0.5641, + "step": 12068 + }, + { + "epoch": 0.8100399315459212, + "grad_norm": 1.0571144819259644, + "learning_rate": 6.743362768868682e-05, + "loss": 0.607, + "step": 12070 + }, + { + "epoch": 0.8101741552296903, + "grad_norm": 0.9392337799072266, + "learning_rate": 6.742344101535394e-05, + "loss": 0.5141, + "step": 12072 + }, + { + "epoch": 0.8103083789134593, + "grad_norm": 1.0465116500854492, + "learning_rate": 6.741325351879771e-05, + "loss": 0.5677, + "step": 12074 + }, + { + "epoch": 0.8104426025972283, + "grad_norm": 0.9749264717102051, + "learning_rate": 6.740306519949952e-05, + "loss": 0.5839, + "step": 12076 + }, + { + "epoch": 0.8105768262809973, + "grad_norm": 1.1746549606323242, + "learning_rate": 6.739287605794069e-05, + "loss": 0.5538, + "step": 12078 + }, + { + "epoch": 0.8107110499647663, + "grad_norm": 1.009050965309143, + "learning_rate": 6.73826860946027e-05, + "loss": 0.7081, + "step": 12080 + }, + { + "epoch": 0.8108452736485353, + "grad_norm": 1.0723981857299805, + "learning_rate": 6.737249530996694e-05, + "loss": 0.5721, + "step": 12082 + }, + { + "epoch": 0.8109794973323042, + "grad_norm": 0.9879979491233826, + "learning_rate": 6.736230370451496e-05, + "loss": 0.6328, + "step": 12084 + }, + { + "epoch": 0.8111137210160733, + "grad_norm": 0.8727061152458191, + "learning_rate": 6.735211127872827e-05, + "loss": 0.5154, + "step": 12086 + }, + { + "epoch": 0.8112479446998423, + "grad_norm": 0.9819492697715759, + "learning_rate": 6.734191803308842e-05, + "loss": 0.5759, + "step": 12088 + }, + { + "epoch": 0.8113821683836113, + "grad_norm": 1.0325849056243896, + "learning_rate": 6.733172396807708e-05, + "loss": 0.5876, + "step": 12090 + }, + { + "epoch": 0.8115163920673802, + "grad_norm": 0.9518097043037415, + "learning_rate": 6.732152908417583e-05, + "loss": 0.556, + "step": 12092 + }, + { + "epoch": 0.8116506157511493, + "grad_norm": 1.0151830911636353, + "learning_rate": 6.731133338186643e-05, + "loss": 0.6942, + "step": 12094 + }, + { + "epoch": 0.8117848394349183, + "grad_norm": 0.9402797222137451, + "learning_rate": 6.730113686163055e-05, + "loss": 0.635, + "step": 12096 + }, + { + "epoch": 0.8119190631186873, + "grad_norm": 1.1566520929336548, + "learning_rate": 6.729093952394996e-05, + "loss": 0.5926, + "step": 12098 + }, + { + "epoch": 0.8120532868024563, + "grad_norm": 1.03131103515625, + "learning_rate": 6.72807413693065e-05, + "loss": 0.635, + "step": 12100 + }, + { + "epoch": 0.8121875104862253, + "grad_norm": 1.0487685203552246, + "learning_rate": 6.727054239818198e-05, + "loss": 0.5959, + "step": 12102 + }, + { + "epoch": 0.8123217341699943, + "grad_norm": 1.0963678359985352, + "learning_rate": 6.72603426110583e-05, + "loss": 0.6084, + "step": 12104 + }, + { + "epoch": 0.8124559578537633, + "grad_norm": 0.9777641296386719, + "learning_rate": 6.725014200841738e-05, + "loss": 0.5647, + "step": 12106 + }, + { + "epoch": 0.8125901815375323, + "grad_norm": 0.9852986931800842, + "learning_rate": 6.723994059074114e-05, + "loss": 0.5841, + "step": 12108 + }, + { + "epoch": 0.8127244052213013, + "grad_norm": 1.1587984561920166, + "learning_rate": 6.722973835851162e-05, + "loss": 0.6071, + "step": 12110 + }, + { + "epoch": 0.8128586289050703, + "grad_norm": 1.3010673522949219, + "learning_rate": 6.721953531221085e-05, + "loss": 0.5806, + "step": 12112 + }, + { + "epoch": 0.8129928525888394, + "grad_norm": 0.9649850130081177, + "learning_rate": 6.720933145232091e-05, + "loss": 0.5731, + "step": 12114 + }, + { + "epoch": 0.8131270762726083, + "grad_norm": 1.01322603225708, + "learning_rate": 6.719912677932389e-05, + "loss": 0.6052, + "step": 12116 + }, + { + "epoch": 0.8132612999563773, + "grad_norm": 1.1387805938720703, + "learning_rate": 6.718892129370195e-05, + "loss": 0.5466, + "step": 12118 + }, + { + "epoch": 0.8133955236401463, + "grad_norm": 1.001020908355713, + "learning_rate": 6.717871499593728e-05, + "loss": 0.6387, + "step": 12120 + }, + { + "epoch": 0.8135297473239153, + "grad_norm": 0.9601315259933472, + "learning_rate": 6.71685078865121e-05, + "loss": 0.5281, + "step": 12122 + }, + { + "epoch": 0.8136639710076843, + "grad_norm": 1.089627981185913, + "learning_rate": 6.71582999659087e-05, + "loss": 0.6077, + "step": 12124 + }, + { + "epoch": 0.8137981946914533, + "grad_norm": 1.0176795721054077, + "learning_rate": 6.714809123460935e-05, + "loss": 0.605, + "step": 12126 + }, + { + "epoch": 0.8139324183752223, + "grad_norm": 0.9597039222717285, + "learning_rate": 6.713788169309641e-05, + "loss": 0.6353, + "step": 12128 + }, + { + "epoch": 0.8140666420589913, + "grad_norm": 1.0546835660934448, + "learning_rate": 6.712767134185228e-05, + "loss": 0.5262, + "step": 12130 + }, + { + "epoch": 0.8142008657427603, + "grad_norm": 1.076861023902893, + "learning_rate": 6.711746018135933e-05, + "loss": 0.5729, + "step": 12132 + }, + { + "epoch": 0.8143350894265293, + "grad_norm": 1.1588343381881714, + "learning_rate": 6.710724821210006e-05, + "loss": 0.606, + "step": 12134 + }, + { + "epoch": 0.8144693131102984, + "grad_norm": 1.027660608291626, + "learning_rate": 6.709703543455695e-05, + "loss": 0.5724, + "step": 12136 + }, + { + "epoch": 0.8146035367940673, + "grad_norm": 1.0974794626235962, + "learning_rate": 6.708682184921255e-05, + "loss": 0.593, + "step": 12138 + }, + { + "epoch": 0.8147377604778363, + "grad_norm": 1.1136621236801147, + "learning_rate": 6.70766074565494e-05, + "loss": 0.6177, + "step": 12140 + }, + { + "epoch": 0.8148719841616053, + "grad_norm": 1.093705654144287, + "learning_rate": 6.706639225705014e-05, + "loss": 0.6727, + "step": 12142 + }, + { + "epoch": 0.8150062078453744, + "grad_norm": 1.0256413221359253, + "learning_rate": 6.705617625119738e-05, + "loss": 0.6204, + "step": 12144 + }, + { + "epoch": 0.8151404315291433, + "grad_norm": 1.0700013637542725, + "learning_rate": 6.704595943947385e-05, + "loss": 0.6268, + "step": 12146 + }, + { + "epoch": 0.8152746552129123, + "grad_norm": 0.9938770532608032, + "learning_rate": 6.703574182236226e-05, + "loss": 0.5993, + "step": 12148 + }, + { + "epoch": 0.8154088788966813, + "grad_norm": 1.084133505821228, + "learning_rate": 6.702552340034535e-05, + "loss": 0.688, + "step": 12150 + }, + { + "epoch": 0.8155431025804504, + "grad_norm": 0.9394397139549255, + "learning_rate": 6.701530417390597e-05, + "loss": 0.5272, + "step": 12152 + }, + { + "epoch": 0.8156773262642193, + "grad_norm": 1.0092073678970337, + "learning_rate": 6.70050841435269e-05, + "loss": 0.585, + "step": 12154 + }, + { + "epoch": 0.8158115499479883, + "grad_norm": 1.026151418685913, + "learning_rate": 6.699486330969106e-05, + "loss": 0.6363, + "step": 12156 + }, + { + "epoch": 0.8159457736317574, + "grad_norm": 0.9925841093063354, + "learning_rate": 6.698464167288133e-05, + "loss": 0.6316, + "step": 12158 + }, + { + "epoch": 0.8160799973155263, + "grad_norm": 0.9733792543411255, + "learning_rate": 6.697441923358068e-05, + "loss": 0.5863, + "step": 12160 + }, + { + "epoch": 0.8162142209992953, + "grad_norm": 0.9599014520645142, + "learning_rate": 6.696419599227213e-05, + "loss": 0.5733, + "step": 12162 + }, + { + "epoch": 0.8163484446830643, + "grad_norm": 0.874563992023468, + "learning_rate": 6.695397194943864e-05, + "loss": 0.5244, + "step": 12164 + }, + { + "epoch": 0.8164826683668334, + "grad_norm": 1.0366476774215698, + "learning_rate": 6.694374710556335e-05, + "loss": 0.6081, + "step": 12166 + }, + { + "epoch": 0.8166168920506023, + "grad_norm": 1.7056056261062622, + "learning_rate": 6.69335214611293e-05, + "loss": 0.6429, + "step": 12168 + }, + { + "epoch": 0.8167511157343713, + "grad_norm": 0.9990048408508301, + "learning_rate": 6.692329501661966e-05, + "loss": 0.6229, + "step": 12170 + }, + { + "epoch": 0.8168853394181403, + "grad_norm": 0.9935768842697144, + "learning_rate": 6.691306777251762e-05, + "loss": 0.6714, + "step": 12172 + }, + { + "epoch": 0.8170195631019094, + "grad_norm": 1.1473873853683472, + "learning_rate": 6.690283972930639e-05, + "loss": 0.6324, + "step": 12174 + }, + { + "epoch": 0.8171537867856783, + "grad_norm": 1.0470446348190308, + "learning_rate": 6.689261088746921e-05, + "loss": 0.643, + "step": 12176 + }, + { + "epoch": 0.8172880104694473, + "grad_norm": 1.36915123462677, + "learning_rate": 6.688238124748939e-05, + "loss": 0.5994, + "step": 12178 + }, + { + "epoch": 0.8174222341532164, + "grad_norm": 1.0217487812042236, + "learning_rate": 6.687215080985025e-05, + "loss": 0.6359, + "step": 12180 + }, + { + "epoch": 0.8175564578369854, + "grad_norm": 1.2322551012039185, + "learning_rate": 6.686191957503517e-05, + "loss": 0.5836, + "step": 12182 + }, + { + "epoch": 0.8176906815207543, + "grad_norm": 1.0810588598251343, + "learning_rate": 6.685168754352754e-05, + "loss": 0.6029, + "step": 12184 + }, + { + "epoch": 0.8178249052045233, + "grad_norm": 0.9595453143119812, + "learning_rate": 6.684145471581081e-05, + "loss": 0.5012, + "step": 12186 + }, + { + "epoch": 0.8179591288882924, + "grad_norm": 1.0888211727142334, + "learning_rate": 6.683122109236845e-05, + "loss": 0.6424, + "step": 12188 + }, + { + "epoch": 0.8180933525720614, + "grad_norm": 1.0043089389801025, + "learning_rate": 6.682098667368403e-05, + "loss": 0.6469, + "step": 12190 + }, + { + "epoch": 0.8182275762558303, + "grad_norm": 1.1146799325942993, + "learning_rate": 6.681075146024104e-05, + "loss": 0.5393, + "step": 12192 + }, + { + "epoch": 0.8183617999395993, + "grad_norm": 1.0782904624938965, + "learning_rate": 6.68005154525231e-05, + "loss": 0.6172, + "step": 12194 + }, + { + "epoch": 0.8184960236233684, + "grad_norm": 0.9525636434555054, + "learning_rate": 6.679027865101383e-05, + "loss": 0.639, + "step": 12196 + }, + { + "epoch": 0.8186302473071373, + "grad_norm": 1.0268981456756592, + "learning_rate": 6.678004105619693e-05, + "loss": 0.5793, + "step": 12198 + }, + { + "epoch": 0.8187644709909063, + "grad_norm": 0.970932126045227, + "learning_rate": 6.676980266855608e-05, + "loss": 0.5664, + "step": 12200 + }, + { + "epoch": 0.8188986946746754, + "grad_norm": 1.0692269802093506, + "learning_rate": 6.675956348857504e-05, + "loss": 0.5385, + "step": 12202 + }, + { + "epoch": 0.8190329183584444, + "grad_norm": 1.116820216178894, + "learning_rate": 6.674932351673758e-05, + "loss": 0.674, + "step": 12204 + }, + { + "epoch": 0.8191671420422133, + "grad_norm": 1.1007301807403564, + "learning_rate": 6.67390827535275e-05, + "loss": 0.5891, + "step": 12206 + }, + { + "epoch": 0.8193013657259823, + "grad_norm": 1.049869418144226, + "learning_rate": 6.672884119942868e-05, + "loss": 0.6069, + "step": 12208 + }, + { + "epoch": 0.8194355894097514, + "grad_norm": 1.1595649719238281, + "learning_rate": 6.671859885492502e-05, + "loss": 0.6151, + "step": 12210 + }, + { + "epoch": 0.8195698130935204, + "grad_norm": 1.0789681673049927, + "learning_rate": 6.670835572050043e-05, + "loss": 0.5684, + "step": 12212 + }, + { + "epoch": 0.8197040367772893, + "grad_norm": 1.0159722566604614, + "learning_rate": 6.669811179663891e-05, + "loss": 0.5849, + "step": 12214 + }, + { + "epoch": 0.8198382604610583, + "grad_norm": 1.06174635887146, + "learning_rate": 6.668786708382441e-05, + "loss": 0.5139, + "step": 12216 + }, + { + "epoch": 0.8199724841448274, + "grad_norm": 1.1024788618087769, + "learning_rate": 6.667762158254104e-05, + "loss": 0.6238, + "step": 12218 + }, + { + "epoch": 0.8201067078285964, + "grad_norm": 1.0661011934280396, + "learning_rate": 6.666737529327282e-05, + "loss": 0.6053, + "step": 12220 + }, + { + "epoch": 0.8202409315123653, + "grad_norm": 1.0743005275726318, + "learning_rate": 6.66571282165039e-05, + "loss": 0.6731, + "step": 12222 + }, + { + "epoch": 0.8203751551961344, + "grad_norm": 1.4202250242233276, + "learning_rate": 6.664688035271843e-05, + "loss": 0.6163, + "step": 12224 + }, + { + "epoch": 0.8205093788799034, + "grad_norm": 0.9662414789199829, + "learning_rate": 6.66366317024006e-05, + "loss": 0.5493, + "step": 12226 + }, + { + "epoch": 0.8206436025636724, + "grad_norm": 0.9612823128700256, + "learning_rate": 6.662638226603463e-05, + "loss": 0.5547, + "step": 12228 + }, + { + "epoch": 0.8207778262474413, + "grad_norm": 1.1255649328231812, + "learning_rate": 6.661613204410479e-05, + "loss": 0.6004, + "step": 12230 + }, + { + "epoch": 0.8209120499312104, + "grad_norm": 1.4618395566940308, + "learning_rate": 6.66058810370954e-05, + "loss": 0.5452, + "step": 12232 + }, + { + "epoch": 0.8210462736149794, + "grad_norm": 1.1108720302581787, + "learning_rate": 6.659562924549076e-05, + "loss": 0.5336, + "step": 12234 + }, + { + "epoch": 0.8211804972987483, + "grad_norm": 1.011399745941162, + "learning_rate": 6.658537666977529e-05, + "loss": 0.6551, + "step": 12236 + }, + { + "epoch": 0.8213147209825173, + "grad_norm": 1.0920634269714355, + "learning_rate": 6.657512331043339e-05, + "loss": 0.6252, + "step": 12238 + }, + { + "epoch": 0.8214489446662864, + "grad_norm": 0.9514212608337402, + "learning_rate": 6.65648691679495e-05, + "loss": 0.5765, + "step": 12240 + }, + { + "epoch": 0.8215831683500554, + "grad_norm": 1.2220063209533691, + "learning_rate": 6.65546142428081e-05, + "loss": 0.6864, + "step": 12242 + }, + { + "epoch": 0.8217173920338243, + "grad_norm": 0.9927895069122314, + "learning_rate": 6.654435853549375e-05, + "loss": 0.5969, + "step": 12244 + }, + { + "epoch": 0.8218516157175934, + "grad_norm": 1.1377277374267578, + "learning_rate": 6.653410204649099e-05, + "loss": 0.6233, + "step": 12246 + }, + { + "epoch": 0.8219858394013624, + "grad_norm": 1.0246647596359253, + "learning_rate": 6.652384477628442e-05, + "loss": 0.6001, + "step": 12248 + }, + { + "epoch": 0.8221200630851314, + "grad_norm": 1.1752510070800781, + "learning_rate": 6.651358672535868e-05, + "loss": 0.6288, + "step": 12250 + }, + { + "epoch": 0.8222542867689003, + "grad_norm": 1.0195872783660889, + "learning_rate": 6.650332789419844e-05, + "loss": 0.6057, + "step": 12252 + }, + { + "epoch": 0.8223885104526694, + "grad_norm": 0.9772490859031677, + "learning_rate": 6.64930682832884e-05, + "loss": 0.6264, + "step": 12254 + }, + { + "epoch": 0.8225227341364384, + "grad_norm": 1.108428716659546, + "learning_rate": 6.648280789311332e-05, + "loss": 0.6569, + "step": 12256 + }, + { + "epoch": 0.8226569578202074, + "grad_norm": 0.9418324828147888, + "learning_rate": 6.6472546724158e-05, + "loss": 0.5641, + "step": 12258 + }, + { + "epoch": 0.8227911815039763, + "grad_norm": 1.1299644708633423, + "learning_rate": 6.646228477690722e-05, + "loss": 0.6731, + "step": 12260 + }, + { + "epoch": 0.8229254051877454, + "grad_norm": 0.8824229836463928, + "learning_rate": 6.645202205184584e-05, + "loss": 0.4883, + "step": 12262 + }, + { + "epoch": 0.8230596288715144, + "grad_norm": 2.0591981410980225, + "learning_rate": 6.644175854945878e-05, + "loss": 0.6361, + "step": 12264 + }, + { + "epoch": 0.8231938525552834, + "grad_norm": 1.0393680334091187, + "learning_rate": 6.643149427023097e-05, + "loss": 0.6191, + "step": 12266 + }, + { + "epoch": 0.8233280762390524, + "grad_norm": 1.083211064338684, + "learning_rate": 6.642122921464736e-05, + "loss": 0.6547, + "step": 12268 + }, + { + "epoch": 0.8234622999228214, + "grad_norm": 1.21903657913208, + "learning_rate": 6.641096338319297e-05, + "loss": 0.5543, + "step": 12270 + }, + { + "epoch": 0.8235965236065904, + "grad_norm": 1.0367486476898193, + "learning_rate": 6.640069677635282e-05, + "loss": 0.5979, + "step": 12272 + }, + { + "epoch": 0.8237307472903593, + "grad_norm": 1.1093127727508545, + "learning_rate": 6.6390429394612e-05, + "loss": 0.6194, + "step": 12274 + }, + { + "epoch": 0.8238649709741284, + "grad_norm": 1.100286841392517, + "learning_rate": 6.638016123845562e-05, + "loss": 0.6025, + "step": 12276 + }, + { + "epoch": 0.8239991946578974, + "grad_norm": 1.0538876056671143, + "learning_rate": 6.636989230836884e-05, + "loss": 0.5919, + "step": 12278 + }, + { + "epoch": 0.8241334183416664, + "grad_norm": 1.1927465200424194, + "learning_rate": 6.635962260483683e-05, + "loss": 0.5939, + "step": 12280 + }, + { + "epoch": 0.8242676420254353, + "grad_norm": 0.8975897431373596, + "learning_rate": 6.634935212834483e-05, + "loss": 0.5064, + "step": 12282 + }, + { + "epoch": 0.8244018657092044, + "grad_norm": 0.9954090118408203, + "learning_rate": 6.63390808793781e-05, + "loss": 0.602, + "step": 12284 + }, + { + "epoch": 0.8245360893929734, + "grad_norm": 0.9890730977058411, + "learning_rate": 6.63288088584219e-05, + "loss": 0.6095, + "step": 12286 + }, + { + "epoch": 0.8246703130767424, + "grad_norm": 1.0482323169708252, + "learning_rate": 6.63185360659616e-05, + "loss": 0.5444, + "step": 12288 + }, + { + "epoch": 0.8248045367605114, + "grad_norm": 1.1700454950332642, + "learning_rate": 6.630826250248256e-05, + "loss": 0.5609, + "step": 12290 + }, + { + "epoch": 0.8249387604442804, + "grad_norm": 0.9571924209594727, + "learning_rate": 6.629798816847019e-05, + "loss": 0.516, + "step": 12292 + }, + { + "epoch": 0.8250729841280494, + "grad_norm": 1.028490662574768, + "learning_rate": 6.628771306440994e-05, + "loss": 0.5647, + "step": 12294 + }, + { + "epoch": 0.8252072078118184, + "grad_norm": 1.81583571434021, + "learning_rate": 6.627743719078725e-05, + "loss": 0.5666, + "step": 12296 + }, + { + "epoch": 0.8253414314955874, + "grad_norm": 1.122802495956421, + "learning_rate": 6.626716054808768e-05, + "loss": 0.6308, + "step": 12298 + }, + { + "epoch": 0.8254756551793564, + "grad_norm": 1.0311051607131958, + "learning_rate": 6.625688313679676e-05, + "loss": 0.5847, + "step": 12300 + }, + { + "epoch": 0.8256098788631254, + "grad_norm": 1.5131393671035767, + "learning_rate": 6.624660495740007e-05, + "loss": 0.6102, + "step": 12302 + }, + { + "epoch": 0.8257441025468945, + "grad_norm": 1.0073790550231934, + "learning_rate": 6.623632601038325e-05, + "loss": 0.5906, + "step": 12304 + }, + { + "epoch": 0.8258783262306634, + "grad_norm": 0.9229928851127625, + "learning_rate": 6.622604629623196e-05, + "loss": 0.6432, + "step": 12306 + }, + { + "epoch": 0.8260125499144324, + "grad_norm": 0.9179010987281799, + "learning_rate": 6.621576581543189e-05, + "loss": 0.5733, + "step": 12308 + }, + { + "epoch": 0.8261467735982014, + "grad_norm": 1.052332878112793, + "learning_rate": 6.620548456846876e-05, + "loss": 0.5641, + "step": 12310 + }, + { + "epoch": 0.8262809972819704, + "grad_norm": 1.1295593976974487, + "learning_rate": 6.619520255582834e-05, + "loss": 0.6259, + "step": 12312 + }, + { + "epoch": 0.8264152209657394, + "grad_norm": 0.9979819059371948, + "learning_rate": 6.618491977799648e-05, + "loss": 0.5714, + "step": 12314 + }, + { + "epoch": 0.8265494446495084, + "grad_norm": 1.22626793384552, + "learning_rate": 6.617463623545895e-05, + "loss": 0.6188, + "step": 12316 + }, + { + "epoch": 0.8266836683332774, + "grad_norm": 1.0156211853027344, + "learning_rate": 6.61643519287017e-05, + "loss": 0.6422, + "step": 12318 + }, + { + "epoch": 0.8268178920170464, + "grad_norm": 1.1920782327651978, + "learning_rate": 6.615406685821058e-05, + "loss": 0.6056, + "step": 12320 + }, + { + "epoch": 0.8269521157008154, + "grad_norm": 1.2539907693862915, + "learning_rate": 6.614378102447158e-05, + "loss": 0.6192, + "step": 12322 + }, + { + "epoch": 0.8270863393845844, + "grad_norm": 1.0377227067947388, + "learning_rate": 6.613349442797066e-05, + "loss": 0.5679, + "step": 12324 + }, + { + "epoch": 0.8272205630683535, + "grad_norm": 1.1849273443222046, + "learning_rate": 6.612320706919387e-05, + "loss": 0.5548, + "step": 12326 + }, + { + "epoch": 0.8273547867521224, + "grad_norm": 1.0158774852752686, + "learning_rate": 6.611291894862726e-05, + "loss": 0.6035, + "step": 12328 + }, + { + "epoch": 0.8274890104358914, + "grad_norm": 1.2449626922607422, + "learning_rate": 6.610263006675688e-05, + "loss": 0.5756, + "step": 12330 + }, + { + "epoch": 0.8276232341196604, + "grad_norm": 1.0021066665649414, + "learning_rate": 6.609234042406892e-05, + "loss": 0.5549, + "step": 12332 + }, + { + "epoch": 0.8277574578034295, + "grad_norm": 1.165596604347229, + "learning_rate": 6.60820500210495e-05, + "loss": 0.6329, + "step": 12334 + }, + { + "epoch": 0.8278916814871984, + "grad_norm": 1.0297205448150635, + "learning_rate": 6.607175885818485e-05, + "loss": 0.6054, + "step": 12336 + }, + { + "epoch": 0.8280259051709674, + "grad_norm": 0.9477143883705139, + "learning_rate": 6.60614669359612e-05, + "loss": 0.615, + "step": 12338 + }, + { + "epoch": 0.8281601288547364, + "grad_norm": 1.0090079307556152, + "learning_rate": 6.605117425486482e-05, + "loss": 0.5564, + "step": 12340 + }, + { + "epoch": 0.8282943525385055, + "grad_norm": 1.0071775913238525, + "learning_rate": 6.604088081538203e-05, + "loss": 0.5843, + "step": 12342 + }, + { + "epoch": 0.8284285762222744, + "grad_norm": 0.9374699592590332, + "learning_rate": 6.603058661799915e-05, + "loss": 0.5506, + "step": 12344 + }, + { + "epoch": 0.8285627999060434, + "grad_norm": 1.0525579452514648, + "learning_rate": 6.602029166320258e-05, + "loss": 0.6223, + "step": 12346 + }, + { + "epoch": 0.8286970235898125, + "grad_norm": 0.9523213505744934, + "learning_rate": 6.600999595147872e-05, + "loss": 0.5676, + "step": 12348 + }, + { + "epoch": 0.8288312472735814, + "grad_norm": 1.093316912651062, + "learning_rate": 6.599969948331403e-05, + "loss": 0.6291, + "step": 12350 + }, + { + "epoch": 0.8289654709573504, + "grad_norm": 1.0761946439743042, + "learning_rate": 6.598940225919504e-05, + "loss": 0.5415, + "step": 12352 + }, + { + "epoch": 0.8290996946411194, + "grad_norm": 1.1536874771118164, + "learning_rate": 6.59791042796082e-05, + "loss": 0.5226, + "step": 12354 + }, + { + "epoch": 0.8292339183248885, + "grad_norm": 0.8335601091384888, + "learning_rate": 6.596880554504011e-05, + "loss": 0.524, + "step": 12356 + }, + { + "epoch": 0.8293681420086574, + "grad_norm": 1.0278328657150269, + "learning_rate": 6.595850605597736e-05, + "loss": 0.5763, + "step": 12358 + }, + { + "epoch": 0.8295023656924264, + "grad_norm": 1.1322448253631592, + "learning_rate": 6.594820581290659e-05, + "loss": 0.7102, + "step": 12360 + }, + { + "epoch": 0.8296365893761954, + "grad_norm": 1.030147910118103, + "learning_rate": 6.593790481631445e-05, + "loss": 0.5821, + "step": 12362 + }, + { + "epoch": 0.8297708130599645, + "grad_norm": 1.0838468074798584, + "learning_rate": 6.592760306668763e-05, + "loss": 0.5579, + "step": 12364 + }, + { + "epoch": 0.8299050367437334, + "grad_norm": 1.009028434753418, + "learning_rate": 6.591730056451292e-05, + "loss": 0.5975, + "step": 12366 + }, + { + "epoch": 0.8300392604275024, + "grad_norm": 1.03667414188385, + "learning_rate": 6.590699731027703e-05, + "loss": 0.5092, + "step": 12368 + }, + { + "epoch": 0.8301734841112715, + "grad_norm": 0.994506299495697, + "learning_rate": 6.589669330446682e-05, + "loss": 0.6129, + "step": 12370 + }, + { + "epoch": 0.8303077077950405, + "grad_norm": 0.9825753569602966, + "learning_rate": 6.58863885475691e-05, + "loss": 0.5808, + "step": 12372 + }, + { + "epoch": 0.8304419314788094, + "grad_norm": 1.4217524528503418, + "learning_rate": 6.587608304007075e-05, + "loss": 0.5751, + "step": 12374 + }, + { + "epoch": 0.8305761551625784, + "grad_norm": 1.0501091480255127, + "learning_rate": 6.586577678245872e-05, + "loss": 0.5963, + "step": 12376 + }, + { + "epoch": 0.8307103788463475, + "grad_norm": 1.012234091758728, + "learning_rate": 6.58554697752199e-05, + "loss": 0.6249, + "step": 12378 + }, + { + "epoch": 0.8308446025301165, + "grad_norm": 1.040315866470337, + "learning_rate": 6.584516201884134e-05, + "loss": 0.5786, + "step": 12380 + }, + { + "epoch": 0.8309788262138854, + "grad_norm": 1.2015894651412964, + "learning_rate": 6.583485351381002e-05, + "loss": 0.6193, + "step": 12382 + }, + { + "epoch": 0.8311130498976544, + "grad_norm": 1.040621280670166, + "learning_rate": 6.582454426061302e-05, + "loss": 0.5689, + "step": 12384 + }, + { + "epoch": 0.8312472735814235, + "grad_norm": 1.1915639638900757, + "learning_rate": 6.581423425973741e-05, + "loss": 0.6585, + "step": 12386 + }, + { + "epoch": 0.8313814972651924, + "grad_norm": 1.0276316404342651, + "learning_rate": 6.580392351167033e-05, + "loss": 0.662, + "step": 12388 + }, + { + "epoch": 0.8315157209489614, + "grad_norm": 1.0397698879241943, + "learning_rate": 6.579361201689895e-05, + "loss": 0.6126, + "step": 12390 + }, + { + "epoch": 0.8316499446327305, + "grad_norm": 1.0907901525497437, + "learning_rate": 6.578329977591045e-05, + "loss": 0.6089, + "step": 12392 + }, + { + "epoch": 0.8317841683164995, + "grad_norm": 0.9839596152305603, + "learning_rate": 6.577298678919209e-05, + "loss": 0.5561, + "step": 12394 + }, + { + "epoch": 0.8319183920002684, + "grad_norm": 0.9547123908996582, + "learning_rate": 6.576267305723111e-05, + "loss": 0.5302, + "step": 12396 + }, + { + "epoch": 0.8320526156840374, + "grad_norm": 0.9410605430603027, + "learning_rate": 6.575235858051481e-05, + "loss": 0.6468, + "step": 12398 + }, + { + "epoch": 0.8321868393678065, + "grad_norm": 0.9823176860809326, + "learning_rate": 6.574204335953056e-05, + "loss": 0.5582, + "step": 12400 + }, + { + "epoch": 0.8323210630515755, + "grad_norm": 1.059046983718872, + "learning_rate": 6.57317273947657e-05, + "loss": 0.6269, + "step": 12402 + }, + { + "epoch": 0.8324552867353444, + "grad_norm": 0.9903417229652405, + "learning_rate": 6.572141068670768e-05, + "loss": 0.5396, + "step": 12404 + }, + { + "epoch": 0.8325895104191134, + "grad_norm": 0.92746502161026, + "learning_rate": 6.57110932358439e-05, + "loss": 0.5885, + "step": 12406 + }, + { + "epoch": 0.8327237341028825, + "grad_norm": 1.1375131607055664, + "learning_rate": 6.570077504266188e-05, + "loss": 0.5735, + "step": 12408 + }, + { + "epoch": 0.8328579577866515, + "grad_norm": 0.9820353984832764, + "learning_rate": 6.569045610764909e-05, + "loss": 0.5607, + "step": 12410 + }, + { + "epoch": 0.8329921814704204, + "grad_norm": 1.0120148658752441, + "learning_rate": 6.568013643129311e-05, + "loss": 0.5833, + "step": 12412 + }, + { + "epoch": 0.8331264051541895, + "grad_norm": 0.9980480074882507, + "learning_rate": 6.566981601408153e-05, + "loss": 0.5451, + "step": 12414 + }, + { + "epoch": 0.8332606288379585, + "grad_norm": 0.9461591243743896, + "learning_rate": 6.565949485650193e-05, + "loss": 0.5984, + "step": 12416 + }, + { + "epoch": 0.8333948525217275, + "grad_norm": 1.4781508445739746, + "learning_rate": 6.564917295904201e-05, + "loss": 0.5972, + "step": 12418 + }, + { + "epoch": 0.8335290762054964, + "grad_norm": 1.0129133462905884, + "learning_rate": 6.563885032218943e-05, + "loss": 0.625, + "step": 12420 + }, + { + "epoch": 0.8336632998892655, + "grad_norm": 1.0137070417404175, + "learning_rate": 6.562852694643194e-05, + "loss": 0.5361, + "step": 12422 + }, + { + "epoch": 0.8337975235730345, + "grad_norm": 1.1191861629486084, + "learning_rate": 6.561820283225726e-05, + "loss": 0.5878, + "step": 12424 + }, + { + "epoch": 0.8339317472568034, + "grad_norm": 0.9222754836082458, + "learning_rate": 6.560787798015323e-05, + "loss": 0.5672, + "step": 12426 + }, + { + "epoch": 0.8340659709405724, + "grad_norm": 1.1135950088500977, + "learning_rate": 6.559755239060765e-05, + "loss": 0.608, + "step": 12428 + }, + { + "epoch": 0.8342001946243415, + "grad_norm": 1.876238465309143, + "learning_rate": 6.558722606410839e-05, + "loss": 0.5573, + "step": 12430 + }, + { + "epoch": 0.8343344183081105, + "grad_norm": 0.9243791699409485, + "learning_rate": 6.557689900114337e-05, + "loss": 0.6266, + "step": 12432 + }, + { + "epoch": 0.8344686419918794, + "grad_norm": 0.975472092628479, + "learning_rate": 6.556657120220046e-05, + "loss": 0.5769, + "step": 12434 + }, + { + "epoch": 0.8346028656756485, + "grad_norm": 1.129416584968567, + "learning_rate": 6.555624266776771e-05, + "loss": 0.5199, + "step": 12436 + }, + { + "epoch": 0.8347370893594175, + "grad_norm": 0.9951165914535522, + "learning_rate": 6.554591339833307e-05, + "loss": 0.4939, + "step": 12438 + }, + { + "epoch": 0.8348713130431865, + "grad_norm": 1.0705358982086182, + "learning_rate": 6.553558339438457e-05, + "loss": 0.5433, + "step": 12440 + }, + { + "epoch": 0.8350055367269554, + "grad_norm": 1.1221299171447754, + "learning_rate": 6.552525265641034e-05, + "loss": 0.5944, + "step": 12442 + }, + { + "epoch": 0.8351397604107245, + "grad_norm": 1.151170015335083, + "learning_rate": 6.551492118489843e-05, + "loss": 0.6213, + "step": 12444 + }, + { + "epoch": 0.8352739840944935, + "grad_norm": 1.3591521978378296, + "learning_rate": 6.550458898033701e-05, + "loss": 0.5648, + "step": 12446 + }, + { + "epoch": 0.8354082077782625, + "grad_norm": 0.9034987688064575, + "learning_rate": 6.549425604321424e-05, + "loss": 0.5845, + "step": 12448 + }, + { + "epoch": 0.8355424314620314, + "grad_norm": 1.1140567064285278, + "learning_rate": 6.548392237401836e-05, + "loss": 0.5704, + "step": 12450 + }, + { + "epoch": 0.8356766551458005, + "grad_norm": 1.1149941682815552, + "learning_rate": 6.547358797323758e-05, + "loss": 0.6356, + "step": 12452 + }, + { + "epoch": 0.8358108788295695, + "grad_norm": 1.0062631368637085, + "learning_rate": 6.546325284136019e-05, + "loss": 0.6261, + "step": 12454 + }, + { + "epoch": 0.8359451025133385, + "grad_norm": 1.1215084791183472, + "learning_rate": 6.545291697887453e-05, + "loss": 0.6334, + "step": 12456 + }, + { + "epoch": 0.8360793261971075, + "grad_norm": 0.9508087038993835, + "learning_rate": 6.544258038626891e-05, + "loss": 0.624, + "step": 12458 + }, + { + "epoch": 0.8362135498808765, + "grad_norm": 0.9110504388809204, + "learning_rate": 6.543224306403174e-05, + "loss": 0.5663, + "step": 12460 + }, + { + "epoch": 0.8363477735646455, + "grad_norm": 0.9410451650619507, + "learning_rate": 6.542190501265144e-05, + "loss": 0.5636, + "step": 12462 + }, + { + "epoch": 0.8364819972484144, + "grad_norm": 0.9118926525115967, + "learning_rate": 6.541156623261646e-05, + "loss": 0.6712, + "step": 12464 + }, + { + "epoch": 0.8366162209321835, + "grad_norm": 1.238031268119812, + "learning_rate": 6.540122672441526e-05, + "loss": 0.6163, + "step": 12466 + }, + { + "epoch": 0.8367504446159525, + "grad_norm": 0.9727116227149963, + "learning_rate": 6.53908864885364e-05, + "loss": 0.6714, + "step": 12468 + }, + { + "epoch": 0.8368846682997215, + "grad_norm": 1.0054668188095093, + "learning_rate": 6.538054552546841e-05, + "loss": 0.6083, + "step": 12470 + }, + { + "epoch": 0.8370188919834904, + "grad_norm": 1.0687994956970215, + "learning_rate": 6.537020383569988e-05, + "loss": 0.5514, + "step": 12472 + }, + { + "epoch": 0.8371531156672595, + "grad_norm": 0.8260632753372192, + "learning_rate": 6.535986141971946e-05, + "loss": 0.514, + "step": 12474 + }, + { + "epoch": 0.8372873393510285, + "grad_norm": 1.081587553024292, + "learning_rate": 6.534951827801579e-05, + "loss": 0.6348, + "step": 12476 + }, + { + "epoch": 0.8374215630347975, + "grad_norm": 1.0609947443008423, + "learning_rate": 6.533917441107755e-05, + "loss": 0.5551, + "step": 12478 + }, + { + "epoch": 0.8375557867185665, + "grad_norm": 1.0872089862823486, + "learning_rate": 6.532882981939349e-05, + "loss": 0.5786, + "step": 12480 + }, + { + "epoch": 0.8376900104023355, + "grad_norm": 1.1435655355453491, + "learning_rate": 6.531848450345236e-05, + "loss": 0.605, + "step": 12482 + }, + { + "epoch": 0.8378242340861045, + "grad_norm": 1.085896611213684, + "learning_rate": 6.530813846374297e-05, + "loss": 0.521, + "step": 12484 + }, + { + "epoch": 0.8379584577698735, + "grad_norm": 1.0298261642456055, + "learning_rate": 6.529779170075413e-05, + "loss": 0.6042, + "step": 12486 + }, + { + "epoch": 0.8380926814536425, + "grad_norm": 1.1587203741073608, + "learning_rate": 6.528744421497471e-05, + "loss": 0.6989, + "step": 12488 + }, + { + "epoch": 0.8382269051374115, + "grad_norm": 0.9626758098602295, + "learning_rate": 6.527709600689363e-05, + "loss": 0.6008, + "step": 12490 + }, + { + "epoch": 0.8383611288211805, + "grad_norm": 0.9409077763557434, + "learning_rate": 6.526674707699979e-05, + "loss": 0.5952, + "step": 12492 + }, + { + "epoch": 0.8384953525049496, + "grad_norm": 0.9995615482330322, + "learning_rate": 6.525639742578218e-05, + "loss": 0.545, + "step": 12494 + }, + { + "epoch": 0.8386295761887185, + "grad_norm": 0.9815777540206909, + "learning_rate": 6.524604705372979e-05, + "loss": 0.5895, + "step": 12496 + }, + { + "epoch": 0.8387637998724875, + "grad_norm": 0.9833337664604187, + "learning_rate": 6.523569596133165e-05, + "loss": 0.5099, + "step": 12498 + }, + { + "epoch": 0.8388980235562565, + "grad_norm": 0.9670937061309814, + "learning_rate": 6.522534414907684e-05, + "loss": 0.5499, + "step": 12500 + }, + { + "epoch": 0.8390322472400255, + "grad_norm": 0.9951319694519043, + "learning_rate": 6.521499161745445e-05, + "loss": 0.5436, + "step": 12502 + }, + { + "epoch": 0.8391664709237945, + "grad_norm": 0.9808186292648315, + "learning_rate": 6.520463836695364e-05, + "loss": 0.6691, + "step": 12504 + }, + { + "epoch": 0.8393006946075635, + "grad_norm": 2.6236653327941895, + "learning_rate": 6.519428439806356e-05, + "loss": 0.6125, + "step": 12506 + }, + { + "epoch": 0.8394349182913325, + "grad_norm": 1.3241273164749146, + "learning_rate": 6.518392971127341e-05, + "loss": 0.5987, + "step": 12508 + }, + { + "epoch": 0.8395691419751015, + "grad_norm": 1.2755435705184937, + "learning_rate": 6.517357430707245e-05, + "loss": 0.6293, + "step": 12510 + }, + { + "epoch": 0.8397033656588705, + "grad_norm": 1.3209245204925537, + "learning_rate": 6.516321818594995e-05, + "loss": 0.6042, + "step": 12512 + }, + { + "epoch": 0.8398375893426395, + "grad_norm": 1.1447761058807373, + "learning_rate": 6.51528613483952e-05, + "loss": 0.6067, + "step": 12514 + }, + { + "epoch": 0.8399718130264086, + "grad_norm": 1.0160657167434692, + "learning_rate": 6.514250379489753e-05, + "loss": 0.5526, + "step": 12516 + }, + { + "epoch": 0.8401060367101775, + "grad_norm": 1.0563714504241943, + "learning_rate": 6.513214552594637e-05, + "loss": 0.5564, + "step": 12518 + }, + { + "epoch": 0.8402402603939465, + "grad_norm": 1.038380742073059, + "learning_rate": 6.512178654203105e-05, + "loss": 0.5417, + "step": 12520 + }, + { + "epoch": 0.8403744840777155, + "grad_norm": 1.062550663948059, + "learning_rate": 6.511142684364109e-05, + "loss": 0.5708, + "step": 12522 + }, + { + "epoch": 0.8405087077614846, + "grad_norm": 0.9445386528968811, + "learning_rate": 6.51010664312659e-05, + "loss": 0.6243, + "step": 12524 + }, + { + "epoch": 0.8406429314452535, + "grad_norm": 0.9992539882659912, + "learning_rate": 6.509070530539502e-05, + "loss": 0.5518, + "step": 12526 + }, + { + "epoch": 0.8407771551290225, + "grad_norm": 3.426658868789673, + "learning_rate": 6.508034346651798e-05, + "loss": 0.6628, + "step": 12528 + }, + { + "epoch": 0.8409113788127915, + "grad_norm": 1.1215256452560425, + "learning_rate": 6.506998091512438e-05, + "loss": 0.5626, + "step": 12530 + }, + { + "epoch": 0.8410456024965605, + "grad_norm": 1.0512257814407349, + "learning_rate": 6.505961765170383e-05, + "loss": 0.4858, + "step": 12532 + }, + { + "epoch": 0.8411798261803295, + "grad_norm": 1.0057384967803955, + "learning_rate": 6.504925367674594e-05, + "loss": 0.6318, + "step": 12534 + }, + { + "epoch": 0.8413140498640985, + "grad_norm": 1.1570310592651367, + "learning_rate": 6.503888899074041e-05, + "loss": 0.5472, + "step": 12536 + }, + { + "epoch": 0.8414482735478676, + "grad_norm": 0.8283010721206665, + "learning_rate": 6.502852359417696e-05, + "loss": 0.4961, + "step": 12538 + }, + { + "epoch": 0.8415824972316365, + "grad_norm": 1.0469348430633545, + "learning_rate": 6.50181574875453e-05, + "loss": 0.5474, + "step": 12540 + }, + { + "epoch": 0.8417167209154055, + "grad_norm": 1.1920287609100342, + "learning_rate": 6.500779067133524e-05, + "loss": 0.5967, + "step": 12542 + }, + { + "epoch": 0.8418509445991745, + "grad_norm": 1.018536925315857, + "learning_rate": 6.499742314603659e-05, + "loss": 0.5903, + "step": 12544 + }, + { + "epoch": 0.8419851682829436, + "grad_norm": 1.0969058275222778, + "learning_rate": 6.498705491213917e-05, + "loss": 0.6266, + "step": 12546 + }, + { + "epoch": 0.8421193919667125, + "grad_norm": 0.929409921169281, + "learning_rate": 6.497668597013289e-05, + "loss": 0.6614, + "step": 12548 + }, + { + "epoch": 0.8422536156504815, + "grad_norm": 1.1739845275878906, + "learning_rate": 6.496631632050763e-05, + "loss": 0.6104, + "step": 12550 + }, + { + "epoch": 0.8423878393342505, + "grad_norm": 1.0775952339172363, + "learning_rate": 6.495594596375338e-05, + "loss": 0.5348, + "step": 12552 + }, + { + "epoch": 0.8425220630180196, + "grad_norm": 1.0988727807998657, + "learning_rate": 6.494557490036009e-05, + "loss": 0.586, + "step": 12554 + }, + { + "epoch": 0.8426562867017885, + "grad_norm": 1.0093157291412354, + "learning_rate": 6.493520313081775e-05, + "loss": 0.6403, + "step": 12556 + }, + { + "epoch": 0.8427905103855575, + "grad_norm": 1.5681852102279663, + "learning_rate": 6.492483065561645e-05, + "loss": 0.5489, + "step": 12558 + }, + { + "epoch": 0.8429247340693266, + "grad_norm": 0.8495414853096008, + "learning_rate": 6.491445747524627e-05, + "loss": 0.5446, + "step": 12560 + }, + { + "epoch": 0.8430589577530956, + "grad_norm": 0.9610116481781006, + "learning_rate": 6.490408359019726e-05, + "loss": 0.6112, + "step": 12562 + }, + { + "epoch": 0.8431931814368645, + "grad_norm": 0.9789390563964844, + "learning_rate": 6.489370900095963e-05, + "loss": 0.5811, + "step": 12564 + }, + { + "epoch": 0.8433274051206335, + "grad_norm": 1.0481964349746704, + "learning_rate": 6.488333370802354e-05, + "loss": 0.6228, + "step": 12566 + }, + { + "epoch": 0.8434616288044026, + "grad_norm": 1.030542254447937, + "learning_rate": 6.487295771187919e-05, + "loss": 0.5577, + "step": 12568 + }, + { + "epoch": 0.8435958524881715, + "grad_norm": 0.9535661935806274, + "learning_rate": 6.486258101301685e-05, + "loss": 0.6148, + "step": 12570 + }, + { + "epoch": 0.8437300761719405, + "grad_norm": 0.8901979923248291, + "learning_rate": 6.485220361192677e-05, + "loss": 0.5021, + "step": 12572 + }, + { + "epoch": 0.8438642998557095, + "grad_norm": 1.042995572090149, + "learning_rate": 6.484182550909927e-05, + "loss": 0.5677, + "step": 12574 + }, + { + "epoch": 0.8439985235394786, + "grad_norm": 1.0647839307785034, + "learning_rate": 6.48314467050247e-05, + "loss": 0.5217, + "step": 12576 + }, + { + "epoch": 0.8441327472232475, + "grad_norm": 1.0358980894088745, + "learning_rate": 6.482106720019344e-05, + "loss": 0.5932, + "step": 12578 + }, + { + "epoch": 0.8442669709070165, + "grad_norm": 0.9523314833641052, + "learning_rate": 6.481068699509591e-05, + "loss": 0.6175, + "step": 12580 + }, + { + "epoch": 0.8444011945907856, + "grad_norm": 1.0470632314682007, + "learning_rate": 6.480030609022253e-05, + "loss": 0.5218, + "step": 12582 + }, + { + "epoch": 0.8445354182745546, + "grad_norm": 1.5232722759246826, + "learning_rate": 6.478992448606381e-05, + "loss": 0.5523, + "step": 12584 + }, + { + "epoch": 0.8446696419583235, + "grad_norm": 1.0707812309265137, + "learning_rate": 6.477954218311021e-05, + "loss": 0.5717, + "step": 12586 + }, + { + "epoch": 0.8448038656420925, + "grad_norm": 1.0655757188796997, + "learning_rate": 6.476915918185234e-05, + "loss": 0.592, + "step": 12588 + }, + { + "epoch": 0.8449380893258616, + "grad_norm": 1.2582905292510986, + "learning_rate": 6.475877548278073e-05, + "loss": 0.5673, + "step": 12590 + }, + { + "epoch": 0.8450723130096306, + "grad_norm": 1.0590770244598389, + "learning_rate": 6.474839108638598e-05, + "loss": 0.6062, + "step": 12592 + }, + { + "epoch": 0.8452065366933995, + "grad_norm": 0.9540066719055176, + "learning_rate": 6.473800599315878e-05, + "loss": 0.5378, + "step": 12594 + }, + { + "epoch": 0.8453407603771685, + "grad_norm": 0.9626739025115967, + "learning_rate": 6.472762020358976e-05, + "loss": 0.5251, + "step": 12596 + }, + { + "epoch": 0.8454749840609376, + "grad_norm": 1.1993706226348877, + "learning_rate": 6.471723371816965e-05, + "loss": 0.6221, + "step": 12598 + }, + { + "epoch": 0.8456092077447066, + "grad_norm": 1.2055251598358154, + "learning_rate": 6.470684653738919e-05, + "loss": 0.5546, + "step": 12600 + }, + { + "epoch": 0.8457434314284755, + "grad_norm": 1.076897144317627, + "learning_rate": 6.469645866173916e-05, + "loss": 0.5965, + "step": 12602 + }, + { + "epoch": 0.8458776551122446, + "grad_norm": 0.9254967570304871, + "learning_rate": 6.468607009171035e-05, + "loss": 0.582, + "step": 12604 + }, + { + "epoch": 0.8460118787960136, + "grad_norm": 1.0577062368392944, + "learning_rate": 6.46756808277936e-05, + "loss": 0.6295, + "step": 12606 + }, + { + "epoch": 0.8461461024797825, + "grad_norm": 1.069785237312317, + "learning_rate": 6.46652908704798e-05, + "loss": 0.6006, + "step": 12608 + }, + { + "epoch": 0.8462803261635515, + "grad_norm": 1.0423345565795898, + "learning_rate": 6.465490022025984e-05, + "loss": 0.5394, + "step": 12610 + }, + { + "epoch": 0.8464145498473206, + "grad_norm": 0.9489768743515015, + "learning_rate": 6.464450887762465e-05, + "loss": 0.5713, + "step": 12612 + }, + { + "epoch": 0.8465487735310896, + "grad_norm": 1.0175904035568237, + "learning_rate": 6.463411684306522e-05, + "loss": 0.5555, + "step": 12614 + }, + { + "epoch": 0.8466829972148585, + "grad_norm": 0.989685595035553, + "learning_rate": 6.462372411707255e-05, + "loss": 0.5475, + "step": 12616 + }, + { + "epoch": 0.8468172208986275, + "grad_norm": 0.989132285118103, + "learning_rate": 6.461333070013767e-05, + "loss": 0.5425, + "step": 12618 + }, + { + "epoch": 0.8469514445823966, + "grad_norm": 0.9655634760856628, + "learning_rate": 6.460293659275165e-05, + "loss": 0.6008, + "step": 12620 + }, + { + "epoch": 0.8470856682661656, + "grad_norm": 1.1773788928985596, + "learning_rate": 6.45925417954056e-05, + "loss": 0.5406, + "step": 12622 + }, + { + "epoch": 0.8472198919499345, + "grad_norm": 1.1722278594970703, + "learning_rate": 6.458214630859064e-05, + "loss": 0.5549, + "step": 12624 + }, + { + "epoch": 0.8473541156337036, + "grad_norm": 0.9735685586929321, + "learning_rate": 6.457175013279792e-05, + "loss": 0.5718, + "step": 12626 + }, + { + "epoch": 0.8474883393174726, + "grad_norm": 1.0683695077896118, + "learning_rate": 6.45613532685187e-05, + "loss": 0.657, + "step": 12628 + }, + { + "epoch": 0.8476225630012416, + "grad_norm": 0.9994304776191711, + "learning_rate": 6.455095571624414e-05, + "loss": 0.6243, + "step": 12630 + }, + { + "epoch": 0.8477567866850105, + "grad_norm": 1.0363658666610718, + "learning_rate": 6.454055747646555e-05, + "loss": 0.609, + "step": 12632 + }, + { + "epoch": 0.8478910103687796, + "grad_norm": 1.0402913093566895, + "learning_rate": 6.45301585496742e-05, + "loss": 0.5772, + "step": 12634 + }, + { + "epoch": 0.8480252340525486, + "grad_norm": 0.9271230101585388, + "learning_rate": 6.451975893636144e-05, + "loss": 0.6032, + "step": 12636 + }, + { + "epoch": 0.8481594577363176, + "grad_norm": 1.1815871000289917, + "learning_rate": 6.450935863701863e-05, + "loss": 0.5725, + "step": 12638 + }, + { + "epoch": 0.8482936814200865, + "grad_norm": 1.3115997314453125, + "learning_rate": 6.449895765213713e-05, + "loss": 0.6197, + "step": 12640 + }, + { + "epoch": 0.8484279051038556, + "grad_norm": 0.9992296695709229, + "learning_rate": 6.448855598220842e-05, + "loss": 0.5907, + "step": 12642 + }, + { + "epoch": 0.8485621287876246, + "grad_norm": 0.9422286748886108, + "learning_rate": 6.447815362772392e-05, + "loss": 0.578, + "step": 12644 + }, + { + "epoch": 0.8486963524713935, + "grad_norm": 1.1738027334213257, + "learning_rate": 6.446775058917514e-05, + "loss": 0.6038, + "step": 12646 + }, + { + "epoch": 0.8488305761551626, + "grad_norm": 1.52314293384552, + "learning_rate": 6.445734686705358e-05, + "loss": 0.6368, + "step": 12648 + }, + { + "epoch": 0.8489647998389316, + "grad_norm": 1.1008228063583374, + "learning_rate": 6.444694246185082e-05, + "loss": 0.6337, + "step": 12650 + }, + { + "epoch": 0.8490990235227006, + "grad_norm": 0.9162123799324036, + "learning_rate": 6.443653737405842e-05, + "loss": 0.5293, + "step": 12652 + }, + { + "epoch": 0.8492332472064695, + "grad_norm": 1.2840626239776611, + "learning_rate": 6.442613160416803e-05, + "loss": 0.5571, + "step": 12654 + }, + { + "epoch": 0.8493674708902386, + "grad_norm": 1.146270513534546, + "learning_rate": 6.44157251526713e-05, + "loss": 0.5627, + "step": 12656 + }, + { + "epoch": 0.8495016945740076, + "grad_norm": 0.9955841898918152, + "learning_rate": 6.440531802005988e-05, + "loss": 0.5282, + "step": 12658 + }, + { + "epoch": 0.8496359182577766, + "grad_norm": 0.9514544606208801, + "learning_rate": 6.439491020682553e-05, + "loss": 0.5606, + "step": 12660 + }, + { + "epoch": 0.8497701419415455, + "grad_norm": 1.0206433534622192, + "learning_rate": 6.438450171345996e-05, + "loss": 0.5182, + "step": 12662 + }, + { + "epoch": 0.8499043656253146, + "grad_norm": 0.9411662817001343, + "learning_rate": 6.437409254045499e-05, + "loss": 0.503, + "step": 12664 + }, + { + "epoch": 0.8500385893090836, + "grad_norm": 0.9811124205589294, + "learning_rate": 6.436368268830241e-05, + "loss": 0.6004, + "step": 12666 + }, + { + "epoch": 0.8501728129928526, + "grad_norm": 0.9797987937927246, + "learning_rate": 6.435327215749405e-05, + "loss": 0.5661, + "step": 12668 + }, + { + "epoch": 0.8503070366766216, + "grad_norm": 1.0006541013717651, + "learning_rate": 6.434286094852184e-05, + "loss": 0.5741, + "step": 12670 + }, + { + "epoch": 0.8504412603603906, + "grad_norm": 1.0398980379104614, + "learning_rate": 6.433244906187763e-05, + "loss": 0.5918, + "step": 12672 + }, + { + "epoch": 0.8505754840441596, + "grad_norm": 1.1455544233322144, + "learning_rate": 6.432203649805337e-05, + "loss": 0.6423, + "step": 12674 + }, + { + "epoch": 0.8507097077279286, + "grad_norm": 1.1146358251571655, + "learning_rate": 6.431162325754108e-05, + "loss": 0.596, + "step": 12676 + }, + { + "epoch": 0.8508439314116976, + "grad_norm": 1.1827868223190308, + "learning_rate": 6.430120934083271e-05, + "loss": 0.6371, + "step": 12678 + }, + { + "epoch": 0.8509781550954666, + "grad_norm": 1.0701404809951782, + "learning_rate": 6.429079474842035e-05, + "loss": 0.5365, + "step": 12680 + }, + { + "epoch": 0.8511123787792356, + "grad_norm": 1.3694677352905273, + "learning_rate": 6.428037948079602e-05, + "loss": 0.5899, + "step": 12682 + }, + { + "epoch": 0.8512466024630045, + "grad_norm": 1.1600438356399536, + "learning_rate": 6.426996353845184e-05, + "loss": 0.5719, + "step": 12684 + }, + { + "epoch": 0.8513808261467736, + "grad_norm": 1.0400867462158203, + "learning_rate": 6.425954692187995e-05, + "loss": 0.5678, + "step": 12686 + }, + { + "epoch": 0.8515150498305426, + "grad_norm": 0.9638028144836426, + "learning_rate": 6.424912963157249e-05, + "loss": 0.5384, + "step": 12688 + }, + { + "epoch": 0.8516492735143116, + "grad_norm": 1.2834314107894897, + "learning_rate": 6.423871166802169e-05, + "loss": 0.5771, + "step": 12690 + }, + { + "epoch": 0.8517834971980806, + "grad_norm": 1.7242958545684814, + "learning_rate": 6.422829303171976e-05, + "loss": 0.5238, + "step": 12692 + }, + { + "epoch": 0.8519177208818496, + "grad_norm": 0.9400202035903931, + "learning_rate": 6.421787372315897e-05, + "loss": 0.615, + "step": 12694 + }, + { + "epoch": 0.8520519445656186, + "grad_norm": 1.0088517665863037, + "learning_rate": 6.420745374283159e-05, + "loss": 0.6427, + "step": 12696 + }, + { + "epoch": 0.8521861682493876, + "grad_norm": 0.9345404505729675, + "learning_rate": 6.419703309122997e-05, + "loss": 0.5377, + "step": 12698 + }, + { + "epoch": 0.8523203919331566, + "grad_norm": 1.0751745700836182, + "learning_rate": 6.418661176884644e-05, + "loss": 0.5821, + "step": 12700 + }, + { + "epoch": 0.8524546156169256, + "grad_norm": 1.0044862031936646, + "learning_rate": 6.417618977617342e-05, + "loss": 0.5448, + "step": 12702 + }, + { + "epoch": 0.8525888393006946, + "grad_norm": 0.9652642011642456, + "learning_rate": 6.41657671137033e-05, + "loss": 0.6212, + "step": 12704 + }, + { + "epoch": 0.8527230629844637, + "grad_norm": 0.835157036781311, + "learning_rate": 6.415534378192853e-05, + "loss": 0.648, + "step": 12706 + }, + { + "epoch": 0.8528572866682326, + "grad_norm": 1.0281325578689575, + "learning_rate": 6.414491978134161e-05, + "loss": 0.5735, + "step": 12708 + }, + { + "epoch": 0.8529915103520016, + "grad_norm": 0.9508633017539978, + "learning_rate": 6.413449511243505e-05, + "loss": 0.5511, + "step": 12710 + }, + { + "epoch": 0.8531257340357706, + "grad_norm": 2.6051974296569824, + "learning_rate": 6.41240697757014e-05, + "loss": 0.6665, + "step": 12712 + }, + { + "epoch": 0.8532599577195397, + "grad_norm": 0.9451747536659241, + "learning_rate": 6.41136437716332e-05, + "loss": 0.6176, + "step": 12714 + }, + { + "epoch": 0.8533941814033086, + "grad_norm": 0.9181392192840576, + "learning_rate": 6.41032171007231e-05, + "loss": 0.5604, + "step": 12716 + }, + { + "epoch": 0.8535284050870776, + "grad_norm": 1.065870761871338, + "learning_rate": 6.409278976346373e-05, + "loss": 0.5658, + "step": 12718 + }, + { + "epoch": 0.8536626287708466, + "grad_norm": 1.0549818277359009, + "learning_rate": 6.408236176034774e-05, + "loss": 0.6213, + "step": 12720 + }, + { + "epoch": 0.8537968524546156, + "grad_norm": 0.9842135906219482, + "learning_rate": 6.407193309186787e-05, + "loss": 0.522, + "step": 12722 + }, + { + "epoch": 0.8539310761383846, + "grad_norm": 1.1176663637161255, + "learning_rate": 6.406150375851682e-05, + "loss": 0.6356, + "step": 12724 + }, + { + "epoch": 0.8540652998221536, + "grad_norm": 0.992957592010498, + "learning_rate": 6.405107376078737e-05, + "loss": 0.6202, + "step": 12726 + }, + { + "epoch": 0.8541995235059227, + "grad_norm": 1.0676531791687012, + "learning_rate": 6.404064309917231e-05, + "loss": 0.5738, + "step": 12728 + }, + { + "epoch": 0.8543337471896916, + "grad_norm": 1.1479809284210205, + "learning_rate": 6.403021177416447e-05, + "loss": 0.6201, + "step": 12730 + }, + { + "epoch": 0.8544679708734606, + "grad_norm": 0.9011659026145935, + "learning_rate": 6.401977978625672e-05, + "loss": 0.5655, + "step": 12732 + }, + { + "epoch": 0.8546021945572296, + "grad_norm": 0.9919249415397644, + "learning_rate": 6.400934713594194e-05, + "loss": 0.6284, + "step": 12734 + }, + { + "epoch": 0.8547364182409987, + "grad_norm": 0.9688147306442261, + "learning_rate": 6.399891382371308e-05, + "loss": 0.543, + "step": 12736 + }, + { + "epoch": 0.8548706419247676, + "grad_norm": 1.1079168319702148, + "learning_rate": 6.398847985006304e-05, + "loss": 0.5324, + "step": 12738 + }, + { + "epoch": 0.8550048656085366, + "grad_norm": 1.1334136724472046, + "learning_rate": 6.397804521548486e-05, + "loss": 0.6161, + "step": 12740 + }, + { + "epoch": 0.8551390892923056, + "grad_norm": 1.054489016532898, + "learning_rate": 6.39676099204715e-05, + "loss": 0.5932, + "step": 12742 + }, + { + "epoch": 0.8552733129760747, + "grad_norm": 1.4015049934387207, + "learning_rate": 6.395717396551606e-05, + "loss": 0.6064, + "step": 12744 + }, + { + "epoch": 0.8554075366598436, + "grad_norm": 1.6162772178649902, + "learning_rate": 6.394673735111158e-05, + "loss": 0.5492, + "step": 12746 + }, + { + "epoch": 0.8555417603436126, + "grad_norm": 1.143741488456726, + "learning_rate": 6.39363000777512e-05, + "loss": 0.6125, + "step": 12748 + }, + { + "epoch": 0.8556759840273817, + "grad_norm": 2.3074216842651367, + "learning_rate": 6.392586214592804e-05, + "loss": 0.6516, + "step": 12750 + }, + { + "epoch": 0.8558102077111507, + "grad_norm": 1.0103665590286255, + "learning_rate": 6.391542355613526e-05, + "loss": 0.591, + "step": 12752 + }, + { + "epoch": 0.8559444313949196, + "grad_norm": 1.2626314163208008, + "learning_rate": 6.390498430886611e-05, + "loss": 0.5789, + "step": 12754 + }, + { + "epoch": 0.8560786550786886, + "grad_norm": 1.0897712707519531, + "learning_rate": 6.389454440461378e-05, + "loss": 0.5849, + "step": 12756 + }, + { + "epoch": 0.8562128787624577, + "grad_norm": 1.0769323110580444, + "learning_rate": 6.388410384387154e-05, + "loss": 0.6044, + "step": 12758 + }, + { + "epoch": 0.8563471024462266, + "grad_norm": 1.125114917755127, + "learning_rate": 6.387366262713272e-05, + "loss": 0.6041, + "step": 12760 + }, + { + "epoch": 0.8564813261299956, + "grad_norm": 0.8260295987129211, + "learning_rate": 6.38632207548906e-05, + "loss": 0.5575, + "step": 12762 + }, + { + "epoch": 0.8566155498137646, + "grad_norm": 1.0780701637268066, + "learning_rate": 6.385277822763857e-05, + "loss": 0.6076, + "step": 12764 + }, + { + "epoch": 0.8567497734975337, + "grad_norm": 0.9742283225059509, + "learning_rate": 6.384233504587001e-05, + "loss": 0.6826, + "step": 12766 + }, + { + "epoch": 0.8568839971813026, + "grad_norm": 0.9774967432022095, + "learning_rate": 6.383189121007833e-05, + "loss": 0.5658, + "step": 12768 + }, + { + "epoch": 0.8570182208650716, + "grad_norm": 1.245073914527893, + "learning_rate": 6.382144672075701e-05, + "loss": 0.5882, + "step": 12770 + }, + { + "epoch": 0.8571524445488407, + "grad_norm": 1.0366885662078857, + "learning_rate": 6.381100157839948e-05, + "loss": 0.5367, + "step": 12772 + }, + { + "epoch": 0.8572866682326097, + "grad_norm": 0.9845640659332275, + "learning_rate": 6.380055578349931e-05, + "loss": 0.6538, + "step": 12774 + }, + { + "epoch": 0.8574208919163786, + "grad_norm": 0.9645702242851257, + "learning_rate": 6.379010933655001e-05, + "loss": 0.5434, + "step": 12776 + }, + { + "epoch": 0.8575551156001476, + "grad_norm": 1.019515872001648, + "learning_rate": 6.377966223804515e-05, + "loss": 0.6037, + "step": 12778 + }, + { + "epoch": 0.8576893392839167, + "grad_norm": 0.8961287140846252, + "learning_rate": 6.376921448847836e-05, + "loss": 0.5209, + "step": 12780 + }, + { + "epoch": 0.8578235629676857, + "grad_norm": 0.9874623417854309, + "learning_rate": 6.375876608834324e-05, + "loss": 0.5967, + "step": 12782 + }, + { + "epoch": 0.8579577866514546, + "grad_norm": 0.9893937110900879, + "learning_rate": 6.374831703813351e-05, + "loss": 0.5701, + "step": 12784 + }, + { + "epoch": 0.8580920103352236, + "grad_norm": 0.9620664715766907, + "learning_rate": 6.37378673383428e-05, + "loss": 0.5999, + "step": 12786 + }, + { + "epoch": 0.8582262340189927, + "grad_norm": 0.9989723563194275, + "learning_rate": 6.372741698946491e-05, + "loss": 0.5446, + "step": 12788 + }, + { + "epoch": 0.8583604577027617, + "grad_norm": 1.1004778146743774, + "learning_rate": 6.371696599199353e-05, + "loss": 0.6531, + "step": 12790 + }, + { + "epoch": 0.8584946813865306, + "grad_norm": 1.4766899347305298, + "learning_rate": 6.370651434642248e-05, + "loss": 0.6686, + "step": 12792 + }, + { + "epoch": 0.8586289050702997, + "grad_norm": 1.1126197576522827, + "learning_rate": 6.36960620532456e-05, + "loss": 0.5572, + "step": 12794 + }, + { + "epoch": 0.8587631287540687, + "grad_norm": 1.0039613246917725, + "learning_rate": 6.368560911295671e-05, + "loss": 0.6369, + "step": 12796 + }, + { + "epoch": 0.8588973524378376, + "grad_norm": 0.9731637239456177, + "learning_rate": 6.367515552604971e-05, + "loss": 0.5513, + "step": 12798 + }, + { + "epoch": 0.8590315761216066, + "grad_norm": 0.9223629832267761, + "learning_rate": 6.36647012930185e-05, + "loss": 0.5354, + "step": 12800 + }, + { + "epoch": 0.8591657998053757, + "grad_norm": 1.0003281831741333, + "learning_rate": 6.365424641435702e-05, + "loss": 0.6371, + "step": 12802 + }, + { + "epoch": 0.8593000234891447, + "grad_norm": 1.3436627388000488, + "learning_rate": 6.364379089055923e-05, + "loss": 0.5978, + "step": 12804 + }, + { + "epoch": 0.8594342471729136, + "grad_norm": 0.9740346074104309, + "learning_rate": 6.363333472211917e-05, + "loss": 0.5426, + "step": 12806 + }, + { + "epoch": 0.8595684708566826, + "grad_norm": 0.9096996784210205, + "learning_rate": 6.362287790953086e-05, + "loss": 0.5132, + "step": 12808 + }, + { + "epoch": 0.8597026945404517, + "grad_norm": 1.1795580387115479, + "learning_rate": 6.361242045328834e-05, + "loss": 0.5476, + "step": 12810 + }, + { + "epoch": 0.8598369182242207, + "grad_norm": 1.140982747077942, + "learning_rate": 6.360196235388574e-05, + "loss": 0.5635, + "step": 12812 + }, + { + "epoch": 0.8599711419079896, + "grad_norm": 1.1558973789215088, + "learning_rate": 6.359150361181715e-05, + "loss": 0.5672, + "step": 12814 + }, + { + "epoch": 0.8601053655917587, + "grad_norm": 1.0639357566833496, + "learning_rate": 6.358104422757673e-05, + "loss": 0.6059, + "step": 12816 + }, + { + "epoch": 0.8602395892755277, + "grad_norm": 1.082284688949585, + "learning_rate": 6.357058420165872e-05, + "loss": 0.5668, + "step": 12818 + }, + { + "epoch": 0.8603738129592967, + "grad_norm": 1.2438132762908936, + "learning_rate": 6.356012353455725e-05, + "loss": 0.5804, + "step": 12820 + }, + { + "epoch": 0.8605080366430656, + "grad_norm": 0.989525318145752, + "learning_rate": 6.354966222676661e-05, + "loss": 0.5388, + "step": 12822 + }, + { + "epoch": 0.8606422603268347, + "grad_norm": 1.0800150632858276, + "learning_rate": 6.353920027878108e-05, + "loss": 0.5278, + "step": 12824 + }, + { + "epoch": 0.8607764840106037, + "grad_norm": 1.059510588645935, + "learning_rate": 6.352873769109498e-05, + "loss": 0.5665, + "step": 12826 + }, + { + "epoch": 0.8609107076943727, + "grad_norm": 1.0791071653366089, + "learning_rate": 6.351827446420261e-05, + "loss": 0.5637, + "step": 12828 + }, + { + "epoch": 0.8610449313781416, + "grad_norm": 1.167945384979248, + "learning_rate": 6.350781059859835e-05, + "loss": 0.6062, + "step": 12830 + }, + { + "epoch": 0.8611791550619107, + "grad_norm": 0.9925161600112915, + "learning_rate": 6.349734609477661e-05, + "loss": 0.5524, + "step": 12832 + }, + { + "epoch": 0.8613133787456797, + "grad_norm": 1.006649374961853, + "learning_rate": 6.348688095323181e-05, + "loss": 0.5375, + "step": 12834 + }, + { + "epoch": 0.8614476024294486, + "grad_norm": 2.9027419090270996, + "learning_rate": 6.34764151744584e-05, + "loss": 0.5353, + "step": 12836 + }, + { + "epoch": 0.8615818261132177, + "grad_norm": 0.9968249797821045, + "learning_rate": 6.346594875895088e-05, + "loss": 0.5111, + "step": 12838 + }, + { + "epoch": 0.8617160497969867, + "grad_norm": 1.0156421661376953, + "learning_rate": 6.345548170720376e-05, + "loss": 0.5524, + "step": 12840 + }, + { + "epoch": 0.8618502734807557, + "grad_norm": 0.8986196517944336, + "learning_rate": 6.344501401971157e-05, + "loss": 0.5402, + "step": 12842 + }, + { + "epoch": 0.8619844971645246, + "grad_norm": 1.3014073371887207, + "learning_rate": 6.343454569696891e-05, + "loss": 0.6122, + "step": 12844 + }, + { + "epoch": 0.8621187208482937, + "grad_norm": 1.0112860202789307, + "learning_rate": 6.34240767394704e-05, + "loss": 0.5491, + "step": 12846 + }, + { + "epoch": 0.8622529445320627, + "grad_norm": 1.1823545694351196, + "learning_rate": 6.341360714771066e-05, + "loss": 0.5835, + "step": 12848 + }, + { + "epoch": 0.8623871682158317, + "grad_norm": 0.9981070756912231, + "learning_rate": 6.340313692218435e-05, + "loss": 0.646, + "step": 12850 + }, + { + "epoch": 0.8625213918996006, + "grad_norm": 1.1531411409378052, + "learning_rate": 6.339266606338619e-05, + "loss": 0.6488, + "step": 12852 + }, + { + "epoch": 0.8626556155833697, + "grad_norm": 0.8666185736656189, + "learning_rate": 6.338219457181088e-05, + "loss": 0.5265, + "step": 12854 + }, + { + "epoch": 0.8627898392671387, + "grad_norm": 0.9972176551818848, + "learning_rate": 6.33717224479532e-05, + "loss": 0.6623, + "step": 12856 + }, + { + "epoch": 0.8629240629509077, + "grad_norm": 1.0785518884658813, + "learning_rate": 6.336124969230792e-05, + "loss": 0.5663, + "step": 12858 + }, + { + "epoch": 0.8630582866346767, + "grad_norm": 2.482130527496338, + "learning_rate": 6.335077630536988e-05, + "loss": 0.5221, + "step": 12860 + }, + { + "epoch": 0.8631925103184457, + "grad_norm": 1.1714829206466675, + "learning_rate": 6.33403022876339e-05, + "loss": 0.5932, + "step": 12862 + }, + { + "epoch": 0.8633267340022147, + "grad_norm": 1.0527064800262451, + "learning_rate": 6.332982763959487e-05, + "loss": 0.5974, + "step": 12864 + }, + { + "epoch": 0.8634609576859837, + "grad_norm": 0.9523221850395203, + "learning_rate": 6.33193523617477e-05, + "loss": 0.5482, + "step": 12866 + }, + { + "epoch": 0.8635951813697527, + "grad_norm": 1.0951387882232666, + "learning_rate": 6.330887645458732e-05, + "loss": 0.5733, + "step": 12868 + }, + { + "epoch": 0.8637294050535217, + "grad_norm": 0.9741978049278259, + "learning_rate": 6.32983999186087e-05, + "loss": 0.5611, + "step": 12870 + }, + { + "epoch": 0.8638636287372907, + "grad_norm": 1.015071988105774, + "learning_rate": 6.328792275430682e-05, + "loss": 0.5379, + "step": 12872 + }, + { + "epoch": 0.8639978524210596, + "grad_norm": 1.0906835794448853, + "learning_rate": 6.327744496217675e-05, + "loss": 0.5316, + "step": 12874 + }, + { + "epoch": 0.8641320761048287, + "grad_norm": 1.0001736879348755, + "learning_rate": 6.32669665427135e-05, + "loss": 0.5752, + "step": 12876 + }, + { + "epoch": 0.8642662997885977, + "grad_norm": 0.9937818646430969, + "learning_rate": 6.325648749641217e-05, + "loss": 0.5676, + "step": 12878 + }, + { + "epoch": 0.8644005234723667, + "grad_norm": 0.9688825607299805, + "learning_rate": 6.324600782376788e-05, + "loss": 0.5647, + "step": 12880 + }, + { + "epoch": 0.8645347471561357, + "grad_norm": 1.5446380376815796, + "learning_rate": 6.323552752527575e-05, + "loss": 0.5929, + "step": 12882 + }, + { + "epoch": 0.8646689708399047, + "grad_norm": 1.0048898458480835, + "learning_rate": 6.322504660143099e-05, + "loss": 0.5474, + "step": 12884 + }, + { + "epoch": 0.8648031945236737, + "grad_norm": 1.2681150436401367, + "learning_rate": 6.321456505272876e-05, + "loss": 0.5835, + "step": 12886 + }, + { + "epoch": 0.8649374182074427, + "grad_norm": 0.9933623671531677, + "learning_rate": 6.320408287966433e-05, + "loss": 0.6576, + "step": 12888 + }, + { + "epoch": 0.8650716418912117, + "grad_norm": 1.0742228031158447, + "learning_rate": 6.319360008273294e-05, + "loss": 0.5655, + "step": 12890 + }, + { + "epoch": 0.8652058655749807, + "grad_norm": 0.95859295129776, + "learning_rate": 6.31831166624299e-05, + "loss": 0.5351, + "step": 12892 + }, + { + "epoch": 0.8653400892587497, + "grad_norm": 1.0715645551681519, + "learning_rate": 6.317263261925052e-05, + "loss": 0.6147, + "step": 12894 + }, + { + "epoch": 0.8654743129425188, + "grad_norm": 1.0217869281768799, + "learning_rate": 6.316214795369016e-05, + "loss": 0.5772, + "step": 12896 + }, + { + "epoch": 0.8656085366262877, + "grad_norm": 0.9852344989776611, + "learning_rate": 6.315166266624418e-05, + "loss": 0.5352, + "step": 12898 + }, + { + "epoch": 0.8657427603100567, + "grad_norm": 1.04218590259552, + "learning_rate": 6.314117675740801e-05, + "loss": 0.6268, + "step": 12900 + }, + { + "epoch": 0.8658769839938257, + "grad_norm": 1.0136308670043945, + "learning_rate": 6.313069022767707e-05, + "loss": 0.6501, + "step": 12902 + }, + { + "epoch": 0.8660112076775948, + "grad_norm": 0.9950689673423767, + "learning_rate": 6.312020307754684e-05, + "loss": 0.632, + "step": 12904 + }, + { + "epoch": 0.8661454313613637, + "grad_norm": 1.0392811298370361, + "learning_rate": 6.31097153075128e-05, + "loss": 0.6059, + "step": 12906 + }, + { + "epoch": 0.8662796550451327, + "grad_norm": 1.0014914274215698, + "learning_rate": 6.30992269180705e-05, + "loss": 0.6234, + "step": 12908 + }, + { + "epoch": 0.8664138787289017, + "grad_norm": 1.0886601209640503, + "learning_rate": 6.30887379097155e-05, + "loss": 0.5978, + "step": 12910 + }, + { + "epoch": 0.8665481024126707, + "grad_norm": 1.0177531242370605, + "learning_rate": 6.307824828294338e-05, + "loss": 0.6496, + "step": 12912 + }, + { + "epoch": 0.8666823260964397, + "grad_norm": 0.9890040159225464, + "learning_rate": 6.306775803824971e-05, + "loss": 0.5529, + "step": 12914 + }, + { + "epoch": 0.8668165497802087, + "grad_norm": 1.1515992879867554, + "learning_rate": 6.30572671761302e-05, + "loss": 0.5655, + "step": 12916 + }, + { + "epoch": 0.8669507734639778, + "grad_norm": 0.9655346274375916, + "learning_rate": 6.304677569708047e-05, + "loss": 0.551, + "step": 12918 + }, + { + "epoch": 0.8670849971477467, + "grad_norm": 1.0232477188110352, + "learning_rate": 6.303628360159625e-05, + "loss": 0.5514, + "step": 12920 + }, + { + "epoch": 0.8672192208315157, + "grad_norm": 0.9819419384002686, + "learning_rate": 6.302579089017327e-05, + "loss": 0.5802, + "step": 12922 + }, + { + "epoch": 0.8673534445152847, + "grad_norm": 1.0228992700576782, + "learning_rate": 6.301529756330728e-05, + "loss": 0.5945, + "step": 12924 + }, + { + "epoch": 0.8674876681990538, + "grad_norm": 1.1202229261398315, + "learning_rate": 6.300480362149409e-05, + "loss": 0.5603, + "step": 12926 + }, + { + "epoch": 0.8676218918828227, + "grad_norm": 0.8481611609458923, + "learning_rate": 6.299430906522948e-05, + "loss": 0.5345, + "step": 12928 + }, + { + "epoch": 0.8677561155665917, + "grad_norm": 0.9928791522979736, + "learning_rate": 6.298381389500933e-05, + "loss": 0.5645, + "step": 12930 + }, + { + "epoch": 0.8678903392503607, + "grad_norm": 1.0613855123519897, + "learning_rate": 6.297331811132951e-05, + "loss": 0.6182, + "step": 12932 + }, + { + "epoch": 0.8680245629341298, + "grad_norm": 0.9913033843040466, + "learning_rate": 6.29628217146859e-05, + "loss": 0.6105, + "step": 12934 + }, + { + "epoch": 0.8681587866178987, + "grad_norm": 0.9503318071365356, + "learning_rate": 6.295232470557447e-05, + "loss": 0.6953, + "step": 12936 + }, + { + "epoch": 0.8682930103016677, + "grad_norm": 1.128720760345459, + "learning_rate": 6.294182708449117e-05, + "loss": 0.5917, + "step": 12938 + }, + { + "epoch": 0.8684272339854368, + "grad_norm": 1.2992370128631592, + "learning_rate": 6.293132885193198e-05, + "loss": 0.6229, + "step": 12940 + }, + { + "epoch": 0.8685614576692058, + "grad_norm": 0.9851817488670349, + "learning_rate": 6.292083000839292e-05, + "loss": 0.5255, + "step": 12942 + }, + { + "epoch": 0.8686956813529747, + "grad_norm": 1.0424292087554932, + "learning_rate": 6.291033055437008e-05, + "loss": 0.5911, + "step": 12944 + }, + { + "epoch": 0.8688299050367437, + "grad_norm": 1.0020283460617065, + "learning_rate": 6.28998304903595e-05, + "loss": 0.5693, + "step": 12946 + }, + { + "epoch": 0.8689641287205128, + "grad_norm": 1.1239510774612427, + "learning_rate": 6.288932981685728e-05, + "loss": 0.6025, + "step": 12948 + }, + { + "epoch": 0.8690983524042817, + "grad_norm": 1.115296721458435, + "learning_rate": 6.287882853435959e-05, + "loss": 0.5697, + "step": 12950 + }, + { + "epoch": 0.8692325760880507, + "grad_norm": 0.9717348217964172, + "learning_rate": 6.286832664336255e-05, + "loss": 0.5958, + "step": 12952 + }, + { + "epoch": 0.8693667997718197, + "grad_norm": 1.0190660953521729, + "learning_rate": 6.28578241443624e-05, + "loss": 0.5641, + "step": 12954 + }, + { + "epoch": 0.8695010234555888, + "grad_norm": 1.0290762186050415, + "learning_rate": 6.284732103785535e-05, + "loss": 0.5396, + "step": 12956 + }, + { + "epoch": 0.8696352471393577, + "grad_norm": 0.993732750415802, + "learning_rate": 6.283681732433762e-05, + "loss": 0.5894, + "step": 12958 + }, + { + "epoch": 0.8697694708231267, + "grad_norm": 0.9604769349098206, + "learning_rate": 6.282631300430554e-05, + "loss": 0.6063, + "step": 12960 + }, + { + "epoch": 0.8699036945068958, + "grad_norm": 1.0046974420547485, + "learning_rate": 6.281580807825538e-05, + "loss": 0.5806, + "step": 12962 + }, + { + "epoch": 0.8700379181906648, + "grad_norm": 1.1138249635696411, + "learning_rate": 6.280530254668349e-05, + "loss": 0.5356, + "step": 12964 + }, + { + "epoch": 0.8701721418744337, + "grad_norm": 1.022376298904419, + "learning_rate": 6.279479641008623e-05, + "loss": 0.6455, + "step": 12966 + }, + { + "epoch": 0.8703063655582027, + "grad_norm": 1.090772032737732, + "learning_rate": 6.278428966896e-05, + "loss": 0.6026, + "step": 12968 + }, + { + "epoch": 0.8704405892419718, + "grad_norm": 0.9981614947319031, + "learning_rate": 6.277378232380123e-05, + "loss": 0.5845, + "step": 12970 + }, + { + "epoch": 0.8705748129257408, + "grad_norm": 1.0998585224151611, + "learning_rate": 6.276327437510636e-05, + "loss": 0.6177, + "step": 12972 + }, + { + "epoch": 0.8707090366095097, + "grad_norm": 0.9015064239501953, + "learning_rate": 6.275276582337188e-05, + "loss": 0.5417, + "step": 12974 + }, + { + "epoch": 0.8708432602932787, + "grad_norm": 1.05171799659729, + "learning_rate": 6.274225666909428e-05, + "loss": 0.5902, + "step": 12976 + }, + { + "epoch": 0.8709774839770478, + "grad_norm": 2.3501741886138916, + "learning_rate": 6.273174691277012e-05, + "loss": 0.5347, + "step": 12978 + }, + { + "epoch": 0.8711117076608168, + "grad_norm": 1.110366940498352, + "learning_rate": 6.272123655489595e-05, + "loss": 0.546, + "step": 12980 + }, + { + "epoch": 0.8712459313445857, + "grad_norm": 1.0878968238830566, + "learning_rate": 6.271072559596836e-05, + "loss": 0.673, + "step": 12982 + }, + { + "epoch": 0.8713801550283548, + "grad_norm": 1.2534525394439697, + "learning_rate": 6.270021403648399e-05, + "loss": 0.5889, + "step": 12984 + }, + { + "epoch": 0.8715143787121238, + "grad_norm": 1.1351861953735352, + "learning_rate": 6.268970187693947e-05, + "loss": 0.5402, + "step": 12986 + }, + { + "epoch": 0.8716486023958927, + "grad_norm": 0.9459824562072754, + "learning_rate": 6.267918911783152e-05, + "loss": 0.5777, + "step": 12988 + }, + { + "epoch": 0.8717828260796617, + "grad_norm": 0.9665629863739014, + "learning_rate": 6.266867575965679e-05, + "loss": 0.5718, + "step": 12990 + }, + { + "epoch": 0.8719170497634308, + "grad_norm": 0.8945533633232117, + "learning_rate": 6.265816180291205e-05, + "loss": 0.5887, + "step": 12992 + }, + { + "epoch": 0.8720512734471998, + "grad_norm": 1.2175227403640747, + "learning_rate": 6.264764724809404e-05, + "loss": 0.5936, + "step": 12994 + }, + { + "epoch": 0.8721854971309687, + "grad_norm": 0.9384693503379822, + "learning_rate": 6.263713209569958e-05, + "loss": 0.5728, + "step": 12996 + }, + { + "epoch": 0.8723197208147377, + "grad_norm": 1.171692132949829, + "learning_rate": 6.262661634622549e-05, + "loss": 0.5447, + "step": 12998 + }, + { + "epoch": 0.8724539444985068, + "grad_norm": 0.9675195813179016, + "learning_rate": 6.26161000001686e-05, + "loss": 0.5513, + "step": 13000 + }, + { + "epoch": 0.8725881681822758, + "grad_norm": 0.999107301235199, + "learning_rate": 6.260558305802578e-05, + "loss": 0.5405, + "step": 13002 + }, + { + "epoch": 0.8727223918660447, + "grad_norm": 1.1518807411193848, + "learning_rate": 6.259506552029396e-05, + "loss": 0.6876, + "step": 13004 + }, + { + "epoch": 0.8728566155498138, + "grad_norm": 1.0599391460418701, + "learning_rate": 6.258454738747006e-05, + "loss": 0.5701, + "step": 13006 + }, + { + "epoch": 0.8729908392335828, + "grad_norm": 1.0742309093475342, + "learning_rate": 6.257402866005105e-05, + "loss": 0.5679, + "step": 13008 + }, + { + "epoch": 0.8731250629173518, + "grad_norm": 1.1204947233200073, + "learning_rate": 6.25635093385339e-05, + "loss": 0.5685, + "step": 13010 + }, + { + "epoch": 0.8732592866011207, + "grad_norm": 0.9725003838539124, + "learning_rate": 6.255298942341564e-05, + "loss": 0.5354, + "step": 13012 + }, + { + "epoch": 0.8733935102848898, + "grad_norm": 1.033810019493103, + "learning_rate": 6.254246891519332e-05, + "loss": 0.6126, + "step": 13014 + }, + { + "epoch": 0.8735277339686588, + "grad_norm": 1.3697429895401, + "learning_rate": 6.2531947814364e-05, + "loss": 0.5436, + "step": 13016 + }, + { + "epoch": 0.8736619576524278, + "grad_norm": 1.0568641424179077, + "learning_rate": 6.25214261214248e-05, + "loss": 0.5932, + "step": 13018 + }, + { + "epoch": 0.8737961813361967, + "grad_norm": 1.1706132888793945, + "learning_rate": 6.251090383687283e-05, + "loss": 0.5681, + "step": 13020 + }, + { + "epoch": 0.8739304050199658, + "grad_norm": 0.9343190789222717, + "learning_rate": 6.250038096120526e-05, + "loss": 0.5876, + "step": 13022 + }, + { + "epoch": 0.8740646287037348, + "grad_norm": 1.0320550203323364, + "learning_rate": 6.248985749491926e-05, + "loss": 0.6106, + "step": 13024 + }, + { + "epoch": 0.8741988523875037, + "grad_norm": 1.1263083219528198, + "learning_rate": 6.247933343851207e-05, + "loss": 0.6176, + "step": 13026 + }, + { + "epoch": 0.8743330760712728, + "grad_norm": 1.0735199451446533, + "learning_rate": 6.246880879248089e-05, + "loss": 0.5905, + "step": 13028 + }, + { + "epoch": 0.8744672997550418, + "grad_norm": 1.0393165349960327, + "learning_rate": 6.245828355732303e-05, + "loss": 0.5268, + "step": 13030 + }, + { + "epoch": 0.8746015234388108, + "grad_norm": 0.9082031846046448, + "learning_rate": 6.244775773353577e-05, + "loss": 0.5749, + "step": 13032 + }, + { + "epoch": 0.8747357471225797, + "grad_norm": 1.1140505075454712, + "learning_rate": 6.243723132161643e-05, + "loss": 0.5703, + "step": 13034 + }, + { + "epoch": 0.8748699708063488, + "grad_norm": 0.9214529991149902, + "learning_rate": 6.242670432206238e-05, + "loss": 0.5326, + "step": 13036 + }, + { + "epoch": 0.8750041944901178, + "grad_norm": 0.938310444355011, + "learning_rate": 6.241617673537097e-05, + "loss": 0.6143, + "step": 13038 + }, + { + "epoch": 0.8751384181738868, + "grad_norm": 1.0400673151016235, + "learning_rate": 6.240564856203964e-05, + "loss": 0.5455, + "step": 13040 + }, + { + "epoch": 0.8752726418576557, + "grad_norm": 0.957240641117096, + "learning_rate": 6.239511980256579e-05, + "loss": 0.6131, + "step": 13042 + }, + { + "epoch": 0.8754068655414248, + "grad_norm": 1.0981582403182983, + "learning_rate": 6.238459045744691e-05, + "loss": 0.5857, + "step": 13044 + }, + { + "epoch": 0.8755410892251938, + "grad_norm": 0.9009153842926025, + "learning_rate": 6.23740605271805e-05, + "loss": 0.5194, + "step": 13046 + }, + { + "epoch": 0.8756753129089628, + "grad_norm": 1.1389052867889404, + "learning_rate": 6.236353001226404e-05, + "loss": 0.5983, + "step": 13048 + }, + { + "epoch": 0.8758095365927318, + "grad_norm": 1.2221170663833618, + "learning_rate": 6.235299891319512e-05, + "loss": 0.5634, + "step": 13050 + }, + { + "epoch": 0.8759437602765008, + "grad_norm": 1.0262128114700317, + "learning_rate": 6.234246723047128e-05, + "loss": 0.5843, + "step": 13052 + }, + { + "epoch": 0.8760779839602698, + "grad_norm": 1.1158145666122437, + "learning_rate": 6.233193496459015e-05, + "loss": 0.6331, + "step": 13054 + }, + { + "epoch": 0.8762122076440388, + "grad_norm": 0.9433540105819702, + "learning_rate": 6.232140211604932e-05, + "loss": 0.5434, + "step": 13056 + }, + { + "epoch": 0.8763464313278078, + "grad_norm": 0.9933052659034729, + "learning_rate": 6.231086868534647e-05, + "loss": 0.5048, + "step": 13058 + }, + { + "epoch": 0.8764806550115768, + "grad_norm": 0.9645562767982483, + "learning_rate": 6.23003346729793e-05, + "loss": 0.5822, + "step": 13060 + }, + { + "epoch": 0.8766148786953458, + "grad_norm": 0.9403725862503052, + "learning_rate": 6.22898000794455e-05, + "loss": 0.5308, + "step": 13062 + }, + { + "epoch": 0.8767491023791147, + "grad_norm": 1.009995937347412, + "learning_rate": 6.22792649052428e-05, + "loss": 0.5839, + "step": 13064 + }, + { + "epoch": 0.8768833260628838, + "grad_norm": 0.9146338105201721, + "learning_rate": 6.226872915086899e-05, + "loss": 0.5643, + "step": 13066 + }, + { + "epoch": 0.8770175497466528, + "grad_norm": 2.256757974624634, + "learning_rate": 6.225819281682186e-05, + "loss": 0.5551, + "step": 13068 + }, + { + "epoch": 0.8771517734304218, + "grad_norm": 1.0844676494598389, + "learning_rate": 6.22476559035992e-05, + "loss": 0.5995, + "step": 13070 + }, + { + "epoch": 0.8772859971141908, + "grad_norm": 1.0413196086883545, + "learning_rate": 6.22371184116989e-05, + "loss": 0.6019, + "step": 13072 + }, + { + "epoch": 0.8774202207979598, + "grad_norm": 1.3758057355880737, + "learning_rate": 6.22265803416188e-05, + "loss": 0.5993, + "step": 13074 + }, + { + "epoch": 0.8775544444817288, + "grad_norm": 0.9704021215438843, + "learning_rate": 6.221604169385685e-05, + "loss": 0.5646, + "step": 13076 + }, + { + "epoch": 0.8776886681654978, + "grad_norm": 1.0348174571990967, + "learning_rate": 6.220550246891095e-05, + "loss": 0.656, + "step": 13078 + }, + { + "epoch": 0.8778228918492668, + "grad_norm": 1.1081150770187378, + "learning_rate": 6.219496266727904e-05, + "loss": 0.6423, + "step": 13080 + }, + { + "epoch": 0.8779571155330358, + "grad_norm": 1.0098882913589478, + "learning_rate": 6.218442228945914e-05, + "loss": 0.5286, + "step": 13082 + }, + { + "epoch": 0.8780913392168048, + "grad_norm": 1.0357965230941772, + "learning_rate": 6.217388133594925e-05, + "loss": 0.6186, + "step": 13084 + }, + { + "epoch": 0.8782255629005739, + "grad_norm": 0.8720493316650391, + "learning_rate": 6.216333980724738e-05, + "loss": 0.5894, + "step": 13086 + }, + { + "epoch": 0.8783597865843428, + "grad_norm": 0.8772887587547302, + "learning_rate": 6.215279770385165e-05, + "loss": 0.5529, + "step": 13088 + }, + { + "epoch": 0.8784940102681118, + "grad_norm": 0.9870423078536987, + "learning_rate": 6.21422550262601e-05, + "loss": 0.6042, + "step": 13090 + }, + { + "epoch": 0.8786282339518808, + "grad_norm": 0.9107111692428589, + "learning_rate": 6.213171177497089e-05, + "loss": 0.5348, + "step": 13092 + }, + { + "epoch": 0.8787624576356499, + "grad_norm": 0.9593957662582397, + "learning_rate": 6.212116795048216e-05, + "loss": 0.5888, + "step": 13094 + }, + { + "epoch": 0.8788966813194188, + "grad_norm": 0.9927076697349548, + "learning_rate": 6.211062355329206e-05, + "loss": 0.5603, + "step": 13096 + }, + { + "epoch": 0.8790309050031878, + "grad_norm": 1.0677847862243652, + "learning_rate": 6.210007858389885e-05, + "loss": 0.5757, + "step": 13098 + }, + { + "epoch": 0.8791651286869568, + "grad_norm": 1.0729224681854248, + "learning_rate": 6.208953304280069e-05, + "loss": 0.6295, + "step": 13100 + }, + { + "epoch": 0.8792993523707258, + "grad_norm": 1.0005987882614136, + "learning_rate": 6.207898693049588e-05, + "loss": 0.5564, + "step": 13102 + }, + { + "epoch": 0.8794335760544948, + "grad_norm": 0.9121020436286926, + "learning_rate": 6.206844024748268e-05, + "loss": 0.5318, + "step": 13104 + }, + { + "epoch": 0.8795677997382638, + "grad_norm": 0.9779648780822754, + "learning_rate": 6.205789299425941e-05, + "loss": 0.5582, + "step": 13106 + }, + { + "epoch": 0.8797020234220329, + "grad_norm": 1.0180506706237793, + "learning_rate": 6.20473451713244e-05, + "loss": 0.5855, + "step": 13108 + }, + { + "epoch": 0.8798362471058018, + "grad_norm": 0.9515188932418823, + "learning_rate": 6.203679677917603e-05, + "loss": 0.5627, + "step": 13110 + }, + { + "epoch": 0.8799704707895708, + "grad_norm": 1.2502775192260742, + "learning_rate": 6.202624781831268e-05, + "loss": 0.6602, + "step": 13112 + }, + { + "epoch": 0.8801046944733398, + "grad_norm": 1.1253612041473389, + "learning_rate": 6.201569828923277e-05, + "loss": 0.6438, + "step": 13114 + }, + { + "epoch": 0.8802389181571089, + "grad_norm": 1.0836929082870483, + "learning_rate": 6.200514819243476e-05, + "loss": 0.5762, + "step": 13116 + }, + { + "epoch": 0.8803731418408778, + "grad_norm": 1.1149053573608398, + "learning_rate": 6.199459752841709e-05, + "loss": 0.5762, + "step": 13118 + }, + { + "epoch": 0.8805073655246468, + "grad_norm": 1.0289145708084106, + "learning_rate": 6.198404629767825e-05, + "loss": 0.5773, + "step": 13120 + }, + { + "epoch": 0.8806415892084158, + "grad_norm": 1.2018251419067383, + "learning_rate": 6.197349450071683e-05, + "loss": 0.5535, + "step": 13122 + }, + { + "epoch": 0.8807758128921849, + "grad_norm": 0.9826356172561646, + "learning_rate": 6.196294213803131e-05, + "loss": 0.5523, + "step": 13124 + }, + { + "epoch": 0.8809100365759538, + "grad_norm": 0.9500182867050171, + "learning_rate": 6.195238921012032e-05, + "loss": 0.5726, + "step": 13126 + }, + { + "epoch": 0.8810442602597228, + "grad_norm": 1.1312055587768555, + "learning_rate": 6.194183571748242e-05, + "loss": 0.5705, + "step": 13128 + }, + { + "epoch": 0.8811784839434919, + "grad_norm": 0.9225857257843018, + "learning_rate": 6.193128166061629e-05, + "loss": 0.5924, + "step": 13130 + }, + { + "epoch": 0.8813127076272609, + "grad_norm": 1.067992925643921, + "learning_rate": 6.192072704002054e-05, + "loss": 0.4947, + "step": 13132 + }, + { + "epoch": 0.8814469313110298, + "grad_norm": 0.8770390748977661, + "learning_rate": 6.191017185619389e-05, + "loss": 0.5097, + "step": 13134 + }, + { + "epoch": 0.8815811549947988, + "grad_norm": 1.0908836126327515, + "learning_rate": 6.189961610963504e-05, + "loss": 0.5721, + "step": 13136 + }, + { + "epoch": 0.8817153786785679, + "grad_norm": 1.0002118349075317, + "learning_rate": 6.188905980084273e-05, + "loss": 0.545, + "step": 13138 + }, + { + "epoch": 0.8818496023623368, + "grad_norm": 1.091551661491394, + "learning_rate": 6.187850293031571e-05, + "loss": 0.6389, + "step": 13140 + }, + { + "epoch": 0.8819838260461058, + "grad_norm": 0.9426711797714233, + "learning_rate": 6.18679454985528e-05, + "loss": 0.5249, + "step": 13142 + }, + { + "epoch": 0.8821180497298748, + "grad_norm": 1.0135663747787476, + "learning_rate": 6.185738750605281e-05, + "loss": 0.5869, + "step": 13144 + }, + { + "epoch": 0.8822522734136439, + "grad_norm": 0.9642865061759949, + "learning_rate": 6.184682895331456e-05, + "loss": 0.5078, + "step": 13146 + }, + { + "epoch": 0.8823864970974128, + "grad_norm": 1.0576778650283813, + "learning_rate": 6.183626984083694e-05, + "loss": 0.5404, + "step": 13148 + }, + { + "epoch": 0.8825207207811818, + "grad_norm": 1.1490014791488647, + "learning_rate": 6.182571016911886e-05, + "loss": 0.6374, + "step": 13150 + }, + { + "epoch": 0.8826549444649509, + "grad_norm": 1.1898303031921387, + "learning_rate": 6.181514993865922e-05, + "loss": 0.5432, + "step": 13152 + }, + { + "epoch": 0.8827891681487199, + "grad_norm": 1.1843942403793335, + "learning_rate": 6.180458914995696e-05, + "loss": 0.6375, + "step": 13154 + }, + { + "epoch": 0.8829233918324888, + "grad_norm": 1.141417145729065, + "learning_rate": 6.17940278035111e-05, + "loss": 0.5919, + "step": 13156 + }, + { + "epoch": 0.8830576155162578, + "grad_norm": 0.9610231518745422, + "learning_rate": 6.178346589982061e-05, + "loss": 0.5513, + "step": 13158 + }, + { + "epoch": 0.8831918392000269, + "grad_norm": 1.13080632686615, + "learning_rate": 6.177290343938452e-05, + "loss": 0.6417, + "step": 13160 + }, + { + "epoch": 0.8833260628837959, + "grad_norm": 1.237656593322754, + "learning_rate": 6.17623404227019e-05, + "loss": 0.5621, + "step": 13162 + }, + { + "epoch": 0.8834602865675648, + "grad_norm": 1.0151150226593018, + "learning_rate": 6.175177685027183e-05, + "loss": 0.4725, + "step": 13164 + }, + { + "epoch": 0.8835945102513338, + "grad_norm": 1.2543600797653198, + "learning_rate": 6.17412127225934e-05, + "loss": 0.568, + "step": 13166 + }, + { + "epoch": 0.8837287339351029, + "grad_norm": 1.0492841005325317, + "learning_rate": 6.173064804016577e-05, + "loss": 0.5572, + "step": 13168 + }, + { + "epoch": 0.8838629576188719, + "grad_norm": 0.9958816170692444, + "learning_rate": 6.172008280348808e-05, + "loss": 0.6348, + "step": 13170 + }, + { + "epoch": 0.8839971813026408, + "grad_norm": 0.9536067843437195, + "learning_rate": 6.170951701305951e-05, + "loss": 0.5068, + "step": 13172 + }, + { + "epoch": 0.8841314049864099, + "grad_norm": 0.9734018445014954, + "learning_rate": 6.169895066937932e-05, + "loss": 0.5774, + "step": 13174 + }, + { + "epoch": 0.8842656286701789, + "grad_norm": 0.978851854801178, + "learning_rate": 6.168838377294669e-05, + "loss": 0.5614, + "step": 13176 + }, + { + "epoch": 0.8843998523539478, + "grad_norm": 0.9927994012832642, + "learning_rate": 6.167781632426093e-05, + "loss": 0.7294, + "step": 13178 + }, + { + "epoch": 0.8845340760377168, + "grad_norm": 1.0294978618621826, + "learning_rate": 6.16672483238213e-05, + "loss": 0.6088, + "step": 13180 + }, + { + "epoch": 0.8846682997214859, + "grad_norm": 1.0444254875183105, + "learning_rate": 6.165667977212712e-05, + "loss": 0.5596, + "step": 13182 + }, + { + "epoch": 0.8848025234052549, + "grad_norm": 0.9945619702339172, + "learning_rate": 6.164611066967775e-05, + "loss": 0.5794, + "step": 13184 + }, + { + "epoch": 0.8849367470890238, + "grad_norm": 1.567557692527771, + "learning_rate": 6.163554101697256e-05, + "loss": 0.6003, + "step": 13186 + }, + { + "epoch": 0.8850709707727928, + "grad_norm": 1.062455177307129, + "learning_rate": 6.162497081451093e-05, + "loss": 0.5674, + "step": 13188 + }, + { + "epoch": 0.8852051944565619, + "grad_norm": 1.147298812866211, + "learning_rate": 6.161440006279228e-05, + "loss": 0.5886, + "step": 13190 + }, + { + "epoch": 0.8853394181403309, + "grad_norm": 1.4330294132232666, + "learning_rate": 6.16038287623161e-05, + "loss": 0.5848, + "step": 13192 + }, + { + "epoch": 0.8854736418240998, + "grad_norm": 0.9991915225982666, + "learning_rate": 6.15932569135818e-05, + "loss": 0.594, + "step": 13194 + }, + { + "epoch": 0.8856078655078689, + "grad_norm": 1.1148768663406372, + "learning_rate": 6.15826845170889e-05, + "loss": 0.6451, + "step": 13196 + }, + { + "epoch": 0.8857420891916379, + "grad_norm": 1.01297128200531, + "learning_rate": 6.157211157333695e-05, + "loss": 0.5847, + "step": 13198 + }, + { + "epoch": 0.8858763128754069, + "grad_norm": 1.046197533607483, + "learning_rate": 6.156153808282548e-05, + "loss": 0.5754, + "step": 13200 + }, + { + "epoch": 0.8860105365591758, + "grad_norm": 2.0050272941589355, + "learning_rate": 6.155096404605408e-05, + "loss": 0.6177, + "step": 13202 + }, + { + "epoch": 0.8861447602429449, + "grad_norm": 1.025641918182373, + "learning_rate": 6.154038946352232e-05, + "loss": 0.5976, + "step": 13204 + }, + { + "epoch": 0.8862789839267139, + "grad_norm": 1.0258253812789917, + "learning_rate": 6.152981433572987e-05, + "loss": 0.5656, + "step": 13206 + }, + { + "epoch": 0.8864132076104829, + "grad_norm": 1.1024008989334106, + "learning_rate": 6.151923866317636e-05, + "loss": 0.6362, + "step": 13208 + }, + { + "epoch": 0.8865474312942518, + "grad_norm": 0.9120305180549622, + "learning_rate": 6.150866244636146e-05, + "loss": 0.6099, + "step": 13210 + }, + { + "epoch": 0.8866816549780209, + "grad_norm": 0.9940522313117981, + "learning_rate": 6.149808568578492e-05, + "loss": 0.6147, + "step": 13212 + }, + { + "epoch": 0.8868158786617899, + "grad_norm": 0.9856640696525574, + "learning_rate": 6.148750838194642e-05, + "loss": 0.559, + "step": 13214 + }, + { + "epoch": 0.8869501023455588, + "grad_norm": 1.1462854146957397, + "learning_rate": 6.147693053534575e-05, + "loss": 0.5442, + "step": 13216 + }, + { + "epoch": 0.8870843260293279, + "grad_norm": 1.1151323318481445, + "learning_rate": 6.146635214648267e-05, + "loss": 0.6204, + "step": 13218 + }, + { + "epoch": 0.8872185497130969, + "grad_norm": 1.0104786157608032, + "learning_rate": 6.145577321585701e-05, + "loss": 0.6074, + "step": 13220 + }, + { + "epoch": 0.8873527733968659, + "grad_norm": 0.9442871809005737, + "learning_rate": 6.144519374396859e-05, + "loss": 0.5557, + "step": 13222 + }, + { + "epoch": 0.8874869970806348, + "grad_norm": 1.0627042055130005, + "learning_rate": 6.143461373131725e-05, + "loss": 0.5659, + "step": 13224 + }, + { + "epoch": 0.8876212207644039, + "grad_norm": 0.9679484963417053, + "learning_rate": 6.142403317840292e-05, + "loss": 0.5486, + "step": 13226 + }, + { + "epoch": 0.8877554444481729, + "grad_norm": 1.0377732515335083, + "learning_rate": 6.14134520857255e-05, + "loss": 0.6623, + "step": 13228 + }, + { + "epoch": 0.8878896681319419, + "grad_norm": 0.8365923762321472, + "learning_rate": 6.140287045378488e-05, + "loss": 0.5505, + "step": 13230 + }, + { + "epoch": 0.8880238918157108, + "grad_norm": 1.0632461309432983, + "learning_rate": 6.139228828308107e-05, + "loss": 0.585, + "step": 13232 + }, + { + "epoch": 0.8881581154994799, + "grad_norm": 0.8229578733444214, + "learning_rate": 6.138170557411403e-05, + "loss": 0.6767, + "step": 13234 + }, + { + "epoch": 0.8882923391832489, + "grad_norm": 0.9627971649169922, + "learning_rate": 6.13711223273838e-05, + "loss": 0.5549, + "step": 13236 + }, + { + "epoch": 0.8884265628670179, + "grad_norm": 1.4396005868911743, + "learning_rate": 6.136053854339039e-05, + "loss": 0.5491, + "step": 13238 + }, + { + "epoch": 0.8885607865507869, + "grad_norm": 1.045028805732727, + "learning_rate": 6.134995422263388e-05, + "loss": 0.6166, + "step": 13240 + }, + { + "epoch": 0.8886950102345559, + "grad_norm": 1.0063236951828003, + "learning_rate": 6.133936936561432e-05, + "loss": 0.5512, + "step": 13242 + }, + { + "epoch": 0.8888292339183249, + "grad_norm": 1.044127345085144, + "learning_rate": 6.132878397283189e-05, + "loss": 0.5556, + "step": 13244 + }, + { + "epoch": 0.8889634576020939, + "grad_norm": 1.0469932556152344, + "learning_rate": 6.131819804478669e-05, + "loss": 0.5538, + "step": 13246 + }, + { + "epoch": 0.8890976812858629, + "grad_norm": 0.9266387224197388, + "learning_rate": 6.130761158197888e-05, + "loss": 0.5829, + "step": 13248 + }, + { + "epoch": 0.8892319049696319, + "grad_norm": 1.0345057249069214, + "learning_rate": 6.129702458490867e-05, + "loss": 0.5726, + "step": 13250 + }, + { + "epoch": 0.8893661286534009, + "grad_norm": 1.0493334531784058, + "learning_rate": 6.128643705407625e-05, + "loss": 0.5445, + "step": 13252 + }, + { + "epoch": 0.8895003523371698, + "grad_norm": 1.1147289276123047, + "learning_rate": 6.127584898998188e-05, + "loss": 0.643, + "step": 13254 + }, + { + "epoch": 0.8896345760209389, + "grad_norm": 1.0473716259002686, + "learning_rate": 6.126526039312581e-05, + "loss": 0.6176, + "step": 13256 + }, + { + "epoch": 0.8897687997047079, + "grad_norm": 1.0527476072311401, + "learning_rate": 6.125467126400835e-05, + "loss": 0.6052, + "step": 13258 + }, + { + "epoch": 0.8899030233884769, + "grad_norm": 1.0677533149719238, + "learning_rate": 6.12440816031298e-05, + "loss": 0.6273, + "step": 13260 + }, + { + "epoch": 0.8900372470722459, + "grad_norm": 1.0257129669189453, + "learning_rate": 6.12334914109905e-05, + "loss": 0.606, + "step": 13262 + }, + { + "epoch": 0.8901714707560149, + "grad_norm": 1.0978323221206665, + "learning_rate": 6.122290068809083e-05, + "loss": 0.6126, + "step": 13264 + }, + { + "epoch": 0.8903056944397839, + "grad_norm": 1.1365458965301514, + "learning_rate": 6.121230943493117e-05, + "loss": 0.5748, + "step": 13266 + }, + { + "epoch": 0.8904399181235529, + "grad_norm": 0.9947246313095093, + "learning_rate": 6.120171765201194e-05, + "loss": 0.5155, + "step": 13268 + }, + { + "epoch": 0.8905741418073219, + "grad_norm": 1.1812353134155273, + "learning_rate": 6.119112533983355e-05, + "loss": 0.6175, + "step": 13270 + }, + { + "epoch": 0.8907083654910909, + "grad_norm": 1.1084691286087036, + "learning_rate": 6.118053249889652e-05, + "loss": 0.5496, + "step": 13272 + }, + { + "epoch": 0.8908425891748599, + "grad_norm": 1.022020936012268, + "learning_rate": 6.116993912970132e-05, + "loss": 0.5587, + "step": 13274 + }, + { + "epoch": 0.890976812858629, + "grad_norm": 0.9779108166694641, + "learning_rate": 6.115934523274845e-05, + "loss": 0.5614, + "step": 13276 + }, + { + "epoch": 0.8911110365423979, + "grad_norm": 1.0704686641693115, + "learning_rate": 6.114875080853846e-05, + "loss": 0.5601, + "step": 13278 + }, + { + "epoch": 0.8912452602261669, + "grad_norm": 1.691609263420105, + "learning_rate": 6.113815585757192e-05, + "loss": 0.5689, + "step": 13280 + }, + { + "epoch": 0.8913794839099359, + "grad_norm": 1.0745458602905273, + "learning_rate": 6.112756038034942e-05, + "loss": 0.585, + "step": 13282 + }, + { + "epoch": 0.891513707593705, + "grad_norm": 1.0736366510391235, + "learning_rate": 6.111696437737157e-05, + "loss": 0.573, + "step": 13284 + }, + { + "epoch": 0.8916479312774739, + "grad_norm": 0.9195927977561951, + "learning_rate": 6.110636784913901e-05, + "loss": 0.5705, + "step": 13286 + }, + { + "epoch": 0.8917821549612429, + "grad_norm": 0.9990829825401306, + "learning_rate": 6.109577079615243e-05, + "loss": 0.6391, + "step": 13288 + }, + { + "epoch": 0.8919163786450119, + "grad_norm": 0.964238703250885, + "learning_rate": 6.108517321891247e-05, + "loss": 0.6249, + "step": 13290 + }, + { + "epoch": 0.8920506023287809, + "grad_norm": 1.1120233535766602, + "learning_rate": 6.10745751179199e-05, + "loss": 0.5978, + "step": 13292 + }, + { + "epoch": 0.8921848260125499, + "grad_norm": 0.9856030941009521, + "learning_rate": 6.106397649367541e-05, + "loss": 0.5742, + "step": 13294 + }, + { + "epoch": 0.8923190496963189, + "grad_norm": 1.4189636707305908, + "learning_rate": 6.10533773466798e-05, + "loss": 0.554, + "step": 13296 + }, + { + "epoch": 0.892453273380088, + "grad_norm": 1.0698367357254028, + "learning_rate": 6.104277767743385e-05, + "loss": 0.5626, + "step": 13298 + }, + { + "epoch": 0.8925874970638569, + "grad_norm": 0.9755097031593323, + "learning_rate": 6.1032177486438355e-05, + "loss": 0.5531, + "step": 13300 + }, + { + "epoch": 0.8927217207476259, + "grad_norm": 1.0601940155029297, + "learning_rate": 6.102157677419418e-05, + "loss": 0.5307, + "step": 13302 + }, + { + "epoch": 0.8928559444313949, + "grad_norm": 1.0407485961914062, + "learning_rate": 6.101097554120216e-05, + "loss": 0.544, + "step": 13304 + }, + { + "epoch": 0.892990168115164, + "grad_norm": 1.1194372177124023, + "learning_rate": 6.100037378796321e-05, + "loss": 0.6405, + "step": 13306 + }, + { + "epoch": 0.8931243917989329, + "grad_norm": 1.0852293968200684, + "learning_rate": 6.0989771514978235e-05, + "loss": 0.6079, + "step": 13308 + }, + { + "epoch": 0.8932586154827019, + "grad_norm": 0.9562682509422302, + "learning_rate": 6.097916872274815e-05, + "loss": 0.5276, + "step": 13310 + }, + { + "epoch": 0.8933928391664709, + "grad_norm": 1.0830674171447754, + "learning_rate": 6.096856541177395e-05, + "loss": 0.5368, + "step": 13312 + }, + { + "epoch": 0.89352706285024, + "grad_norm": 1.0606974363327026, + "learning_rate": 6.095796158255659e-05, + "loss": 0.5347, + "step": 13314 + }, + { + "epoch": 0.8936612865340089, + "grad_norm": 1.005293846130371, + "learning_rate": 6.09473572355971e-05, + "loss": 0.5241, + "step": 13316 + }, + { + "epoch": 0.8937955102177779, + "grad_norm": 1.0626174211502075, + "learning_rate": 6.09367523713965e-05, + "loss": 0.6296, + "step": 13318 + }, + { + "epoch": 0.893929733901547, + "grad_norm": 1.0723947286605835, + "learning_rate": 6.092614699045587e-05, + "loss": 0.5691, + "step": 13320 + }, + { + "epoch": 0.894063957585316, + "grad_norm": 1.091525673866272, + "learning_rate": 6.091554109327626e-05, + "loss": 0.6199, + "step": 13322 + }, + { + "epoch": 0.8941981812690849, + "grad_norm": 1.0702077150344849, + "learning_rate": 6.090493468035882e-05, + "loss": 0.5419, + "step": 13324 + }, + { + "epoch": 0.8943324049528539, + "grad_norm": 1.0261483192443848, + "learning_rate": 6.089432775220465e-05, + "loss": 0.6493, + "step": 13326 + }, + { + "epoch": 0.894466628636623, + "grad_norm": 0.816967785358429, + "learning_rate": 6.088372030931491e-05, + "loss": 0.5044, + "step": 13328 + }, + { + "epoch": 0.8946008523203919, + "grad_norm": 1.0482172966003418, + "learning_rate": 6.0873112352190795e-05, + "loss": 0.6522, + "step": 13330 + }, + { + "epoch": 0.8947350760041609, + "grad_norm": 1.0886648893356323, + "learning_rate": 6.0862503881333496e-05, + "loss": 0.5491, + "step": 13332 + }, + { + "epoch": 0.8948692996879299, + "grad_norm": 1.7727330923080444, + "learning_rate": 6.085189489724426e-05, + "loss": 0.5805, + "step": 13334 + }, + { + "epoch": 0.895003523371699, + "grad_norm": 0.9612252712249756, + "learning_rate": 6.084128540042432e-05, + "loss": 0.5659, + "step": 13336 + }, + { + "epoch": 0.8951377470554679, + "grad_norm": 0.930299699306488, + "learning_rate": 6.083067539137497e-05, + "loss": 0.5867, + "step": 13338 + }, + { + "epoch": 0.8952719707392369, + "grad_norm": 0.8667044639587402, + "learning_rate": 6.0820064870597504e-05, + "loss": 0.5225, + "step": 13340 + }, + { + "epoch": 0.895406194423006, + "grad_norm": 1.191062092781067, + "learning_rate": 6.0809453838593246e-05, + "loss": 0.6223, + "step": 13342 + }, + { + "epoch": 0.895540418106775, + "grad_norm": 0.9350284934043884, + "learning_rate": 6.079884229586355e-05, + "loss": 0.5615, + "step": 13344 + }, + { + "epoch": 0.8956746417905439, + "grad_norm": 1.1160637140274048, + "learning_rate": 6.0788230242909795e-05, + "loss": 0.5936, + "step": 13346 + }, + { + "epoch": 0.8958088654743129, + "grad_norm": 0.9354270100593567, + "learning_rate": 6.077761768023337e-05, + "loss": 0.5426, + "step": 13348 + }, + { + "epoch": 0.895943089158082, + "grad_norm": 0.9717685580253601, + "learning_rate": 6.076700460833571e-05, + "loss": 0.5489, + "step": 13350 + }, + { + "epoch": 0.896077312841851, + "grad_norm": 1.0427420139312744, + "learning_rate": 6.075639102771824e-05, + "loss": 0.4835, + "step": 13352 + }, + { + "epoch": 0.8962115365256199, + "grad_norm": 1.0258979797363281, + "learning_rate": 6.074577693888246e-05, + "loss": 0.5937, + "step": 13354 + }, + { + "epoch": 0.8963457602093889, + "grad_norm": 0.9997778534889221, + "learning_rate": 6.073516234232985e-05, + "loss": 0.594, + "step": 13356 + }, + { + "epoch": 0.896479983893158, + "grad_norm": 0.9338006973266602, + "learning_rate": 6.072454723856192e-05, + "loss": 0.5378, + "step": 13358 + }, + { + "epoch": 0.896614207576927, + "grad_norm": 1.1041646003723145, + "learning_rate": 6.0713931628080236e-05, + "loss": 0.5606, + "step": 13360 + }, + { + "epoch": 0.8967484312606959, + "grad_norm": 1.0136053562164307, + "learning_rate": 6.0703315511386336e-05, + "loss": 0.5599, + "step": 13362 + }, + { + "epoch": 0.896882654944465, + "grad_norm": 0.8840245604515076, + "learning_rate": 6.069269888898184e-05, + "loss": 0.5809, + "step": 13364 + }, + { + "epoch": 0.897016878628234, + "grad_norm": 0.9898674488067627, + "learning_rate": 6.0682081761368325e-05, + "loss": 0.5791, + "step": 13366 + }, + { + "epoch": 0.8971511023120029, + "grad_norm": 1.0291520357131958, + "learning_rate": 6.0671464129047474e-05, + "loss": 0.5538, + "step": 13368 + }, + { + "epoch": 0.8972853259957719, + "grad_norm": 0.9146878123283386, + "learning_rate": 6.066084599252092e-05, + "loss": 0.5134, + "step": 13370 + }, + { + "epoch": 0.897419549679541, + "grad_norm": 1.0117591619491577, + "learning_rate": 6.0650227352290345e-05, + "loss": 0.5649, + "step": 13372 + }, + { + "epoch": 0.89755377336331, + "grad_norm": 1.171174168586731, + "learning_rate": 6.063960820885749e-05, + "loss": 0.5771, + "step": 13374 + }, + { + "epoch": 0.8976879970470789, + "grad_norm": 1.0561974048614502, + "learning_rate": 6.0628988562724054e-05, + "loss": 0.5866, + "step": 13376 + }, + { + "epoch": 0.8978222207308479, + "grad_norm": 1.0028769969940186, + "learning_rate": 6.061836841439182e-05, + "loss": 0.5668, + "step": 13378 + }, + { + "epoch": 0.897956444414617, + "grad_norm": 1.009484887123108, + "learning_rate": 6.060774776436255e-05, + "loss": 0.5868, + "step": 13380 + }, + { + "epoch": 0.898090668098386, + "grad_norm": 1.0336222648620605, + "learning_rate": 6.059712661313807e-05, + "loss": 0.5579, + "step": 13382 + }, + { + "epoch": 0.8982248917821549, + "grad_norm": 0.9472618103027344, + "learning_rate": 6.058650496122018e-05, + "loss": 0.5292, + "step": 13384 + }, + { + "epoch": 0.898359115465924, + "grad_norm": 1.9172183275222778, + "learning_rate": 6.057588280911075e-05, + "loss": 0.5298, + "step": 13386 + }, + { + "epoch": 0.898493339149693, + "grad_norm": 1.035021424293518, + "learning_rate": 6.056526015731166e-05, + "loss": 0.5675, + "step": 13388 + }, + { + "epoch": 0.898627562833462, + "grad_norm": 0.9642541408538818, + "learning_rate": 6.05546370063248e-05, + "loss": 0.4864, + "step": 13390 + }, + { + "epoch": 0.8987617865172309, + "grad_norm": 2.293220043182373, + "learning_rate": 6.054401335665211e-05, + "loss": 0.563, + "step": 13392 + }, + { + "epoch": 0.898896010201, + "grad_norm": 1.1741487979888916, + "learning_rate": 6.05333892087955e-05, + "loss": 0.5971, + "step": 13394 + }, + { + "epoch": 0.899030233884769, + "grad_norm": 1.116698980331421, + "learning_rate": 6.0522764563256985e-05, + "loss": 0.6076, + "step": 13396 + }, + { + "epoch": 0.899164457568538, + "grad_norm": 1.0133905410766602, + "learning_rate": 6.0512139420538515e-05, + "loss": 0.6022, + "step": 13398 + }, + { + "epoch": 0.8992986812523069, + "grad_norm": 0.9247488975524902, + "learning_rate": 6.050151378114214e-05, + "loss": 0.6255, + "step": 13400 + }, + { + "epoch": 0.899432904936076, + "grad_norm": 1.062780737876892, + "learning_rate": 6.049088764556989e-05, + "loss": 0.6129, + "step": 13402 + }, + { + "epoch": 0.899567128619845, + "grad_norm": 1.0486704111099243, + "learning_rate": 6.0480261014323826e-05, + "loss": 0.5423, + "step": 13404 + }, + { + "epoch": 0.8997013523036139, + "grad_norm": 1.0059887170791626, + "learning_rate": 6.046963388790604e-05, + "loss": 0.5905, + "step": 13406 + }, + { + "epoch": 0.899835575987383, + "grad_norm": 3.0994739532470703, + "learning_rate": 6.045900626681864e-05, + "loss": 0.5409, + "step": 13408 + }, + { + "epoch": 0.899969799671152, + "grad_norm": 1.0882151126861572, + "learning_rate": 6.044837815156377e-05, + "loss": 0.5289, + "step": 13410 + }, + { + "epoch": 0.900104023354921, + "grad_norm": 1.0200914144515991, + "learning_rate": 6.043774954264355e-05, + "loss": 0.5814, + "step": 13412 + }, + { + "epoch": 0.9002382470386899, + "grad_norm": 1.0724719762802124, + "learning_rate": 6.0427120440560204e-05, + "loss": 0.5838, + "step": 13414 + }, + { + "epoch": 0.900372470722459, + "grad_norm": 0.8868136405944824, + "learning_rate": 6.041649084581593e-05, + "loss": 0.5186, + "step": 13416 + }, + { + "epoch": 0.900506694406228, + "grad_norm": 1.1544818878173828, + "learning_rate": 6.040586075891293e-05, + "loss": 0.6216, + "step": 13418 + }, + { + "epoch": 0.900640918089997, + "grad_norm": 0.9154133200645447, + "learning_rate": 6.0395230180353504e-05, + "loss": 0.6012, + "step": 13420 + }, + { + "epoch": 0.9007751417737659, + "grad_norm": 1.0481500625610352, + "learning_rate": 6.038459911063986e-05, + "loss": 0.5635, + "step": 13422 + }, + { + "epoch": 0.900909365457535, + "grad_norm": 0.9151366353034973, + "learning_rate": 6.0373967550274336e-05, + "loss": 0.5563, + "step": 13424 + }, + { + "epoch": 0.901043589141304, + "grad_norm": 1.0725773572921753, + "learning_rate": 6.0363335499759265e-05, + "loss": 0.5719, + "step": 13426 + }, + { + "epoch": 0.901177812825073, + "grad_norm": 1.158782958984375, + "learning_rate": 6.035270295959695e-05, + "loss": 0.5878, + "step": 13428 + }, + { + "epoch": 0.901312036508842, + "grad_norm": 1.0959148406982422, + "learning_rate": 6.034206993028979e-05, + "loss": 0.5669, + "step": 13430 + }, + { + "epoch": 0.901446260192611, + "grad_norm": 0.9470241069793701, + "learning_rate": 6.0331436412340147e-05, + "loss": 0.6154, + "step": 13432 + }, + { + "epoch": 0.90158048387638, + "grad_norm": 1.0807803869247437, + "learning_rate": 6.032080240625045e-05, + "loss": 0.5643, + "step": 13434 + }, + { + "epoch": 0.901714707560149, + "grad_norm": 0.996346652507782, + "learning_rate": 6.031016791252315e-05, + "loss": 0.6318, + "step": 13436 + }, + { + "epoch": 0.901848931243918, + "grad_norm": 0.9364991188049316, + "learning_rate": 6.0299532931660675e-05, + "loss": 0.5382, + "step": 13438 + }, + { + "epoch": 0.901983154927687, + "grad_norm": 1.0150824785232544, + "learning_rate": 6.028889746416553e-05, + "loss": 0.5568, + "step": 13440 + }, + { + "epoch": 0.902117378611456, + "grad_norm": 1.0036660432815552, + "learning_rate": 6.0278261510540214e-05, + "loss": 0.5679, + "step": 13442 + }, + { + "epoch": 0.9022516022952249, + "grad_norm": 0.9905140995979309, + "learning_rate": 6.026762507128725e-05, + "loss": 0.5577, + "step": 13444 + }, + { + "epoch": 0.902385825978994, + "grad_norm": 1.0355485677719116, + "learning_rate": 6.025698814690919e-05, + "loss": 0.5882, + "step": 13446 + }, + { + "epoch": 0.902520049662763, + "grad_norm": 1.0384221076965332, + "learning_rate": 6.0246350737908605e-05, + "loss": 0.6298, + "step": 13448 + }, + { + "epoch": 0.902654273346532, + "grad_norm": 0.9836850762367249, + "learning_rate": 6.02357128447881e-05, + "loss": 0.5671, + "step": 13450 + }, + { + "epoch": 0.902788497030301, + "grad_norm": 0.9347818493843079, + "learning_rate": 6.022507446805029e-05, + "loss": 0.5293, + "step": 13452 + }, + { + "epoch": 0.90292272071407, + "grad_norm": 0.9558823704719543, + "learning_rate": 6.0214435608197825e-05, + "loss": 0.5222, + "step": 13454 + }, + { + "epoch": 0.903056944397839, + "grad_norm": 7.560023307800293, + "learning_rate": 6.020379626573336e-05, + "loss": 0.5481, + "step": 13456 + }, + { + "epoch": 0.903191168081608, + "grad_norm": 1.25991690158844, + "learning_rate": 6.019315644115959e-05, + "loss": 0.5572, + "step": 13458 + }, + { + "epoch": 0.903325391765377, + "grad_norm": 0.9077351093292236, + "learning_rate": 6.018251613497922e-05, + "loss": 0.595, + "step": 13460 + }, + { + "epoch": 0.903459615449146, + "grad_norm": 1.0868170261383057, + "learning_rate": 6.0171875347694974e-05, + "loss": 0.5016, + "step": 13462 + }, + { + "epoch": 0.903593839132915, + "grad_norm": 0.8331081867218018, + "learning_rate": 6.0161234079809635e-05, + "loss": 0.53, + "step": 13464 + }, + { + "epoch": 0.903728062816684, + "grad_norm": 1.0237597227096558, + "learning_rate": 6.015059233182596e-05, + "loss": 0.6166, + "step": 13466 + }, + { + "epoch": 0.903862286500453, + "grad_norm": 0.9407981038093567, + "learning_rate": 6.013995010424676e-05, + "loss": 0.5412, + "step": 13468 + }, + { + "epoch": 0.903996510184222, + "grad_norm": 0.8044424057006836, + "learning_rate": 6.012930739757485e-05, + "loss": 0.5489, + "step": 13470 + }, + { + "epoch": 0.904130733867991, + "grad_norm": 1.1213793754577637, + "learning_rate": 6.011866421231309e-05, + "loss": 0.5933, + "step": 13472 + }, + { + "epoch": 0.9042649575517601, + "grad_norm": 1.1691128015518188, + "learning_rate": 6.010802054896435e-05, + "loss": 0.6067, + "step": 13474 + }, + { + "epoch": 0.904399181235529, + "grad_norm": 1.0560898780822754, + "learning_rate": 6.0097376408031504e-05, + "loss": 0.5358, + "step": 13476 + }, + { + "epoch": 0.904533404919298, + "grad_norm": 1.1528799533843994, + "learning_rate": 6.008673179001748e-05, + "loss": 0.6544, + "step": 13478 + }, + { + "epoch": 0.904667628603067, + "grad_norm": 0.9637367129325867, + "learning_rate": 6.007608669542522e-05, + "loss": 0.5733, + "step": 13480 + }, + { + "epoch": 0.904801852286836, + "grad_norm": 1.0823876857757568, + "learning_rate": 6.006544112475767e-05, + "loss": 0.5851, + "step": 13482 + }, + { + "epoch": 0.904936075970605, + "grad_norm": 1.0340442657470703, + "learning_rate": 6.0054795078517826e-05, + "loss": 0.6026, + "step": 13484 + }, + { + "epoch": 0.905070299654374, + "grad_norm": 1.4215126037597656, + "learning_rate": 6.004414855720869e-05, + "loss": 0.5864, + "step": 13486 + }, + { + "epoch": 0.905204523338143, + "grad_norm": 1.04523503780365, + "learning_rate": 6.003350156133327e-05, + "loss": 0.5672, + "step": 13488 + }, + { + "epoch": 0.905338747021912, + "grad_norm": 0.9753003716468811, + "learning_rate": 6.002285409139464e-05, + "loss": 0.6201, + "step": 13490 + }, + { + "epoch": 0.905472970705681, + "grad_norm": 1.0781129598617554, + "learning_rate": 6.001220614789587e-05, + "loss": 0.6049, + "step": 13492 + }, + { + "epoch": 0.90560719438945, + "grad_norm": 0.9975566267967224, + "learning_rate": 6.000155773134005e-05, + "loss": 0.5173, + "step": 13494 + }, + { + "epoch": 0.9057414180732191, + "grad_norm": 1.2302011251449585, + "learning_rate": 5.999090884223029e-05, + "loss": 0.6304, + "step": 13496 + }, + { + "epoch": 0.905875641756988, + "grad_norm": 1.0563715696334839, + "learning_rate": 5.998025948106973e-05, + "loss": 0.5148, + "step": 13498 + }, + { + "epoch": 0.906009865440757, + "grad_norm": 1.1134657859802246, + "learning_rate": 5.9969609648361526e-05, + "loss": 0.588, + "step": 13500 + }, + { + "epoch": 0.906144089124526, + "grad_norm": 1.0876972675323486, + "learning_rate": 5.995895934460889e-05, + "loss": 0.6915, + "step": 13502 + }, + { + "epoch": 0.9062783128082951, + "grad_norm": 0.9595404267311096, + "learning_rate": 5.994830857031499e-05, + "loss": 0.4885, + "step": 13504 + }, + { + "epoch": 0.906412536492064, + "grad_norm": 0.8634196519851685, + "learning_rate": 5.9937657325983086e-05, + "loss": 0.4999, + "step": 13506 + }, + { + "epoch": 0.906546760175833, + "grad_norm": 1.1488295793533325, + "learning_rate": 5.992700561211641e-05, + "loss": 0.6347, + "step": 13508 + }, + { + "epoch": 0.906680983859602, + "grad_norm": 1.0268058776855469, + "learning_rate": 5.991635342921823e-05, + "loss": 0.5978, + "step": 13510 + }, + { + "epoch": 0.9068152075433711, + "grad_norm": 1.0089184045791626, + "learning_rate": 5.9905700777791864e-05, + "loss": 0.5651, + "step": 13512 + }, + { + "epoch": 0.90694943122714, + "grad_norm": 1.0301094055175781, + "learning_rate": 5.989504765834061e-05, + "loss": 0.5734, + "step": 13514 + }, + { + "epoch": 0.907083654910909, + "grad_norm": 0.872610867023468, + "learning_rate": 5.9884394071367814e-05, + "loss": 0.5492, + "step": 13516 + }, + { + "epoch": 0.9072178785946781, + "grad_norm": 0.9029110074043274, + "learning_rate": 5.9873740017376825e-05, + "loss": 0.5531, + "step": 13518 + }, + { + "epoch": 0.907352102278447, + "grad_norm": 1.2001795768737793, + "learning_rate": 5.986308549687105e-05, + "loss": 0.5644, + "step": 13520 + }, + { + "epoch": 0.907486325962216, + "grad_norm": 1.0047690868377686, + "learning_rate": 5.9852430510353876e-05, + "loss": 0.5968, + "step": 13522 + }, + { + "epoch": 0.907620549645985, + "grad_norm": 0.9836839437484741, + "learning_rate": 5.984177505832872e-05, + "loss": 0.6166, + "step": 13524 + }, + { + "epoch": 0.9077547733297541, + "grad_norm": 1.0714616775512695, + "learning_rate": 5.9831119141299064e-05, + "loss": 0.608, + "step": 13526 + }, + { + "epoch": 0.907888997013523, + "grad_norm": 1.010029911994934, + "learning_rate": 5.982046275976836e-05, + "loss": 0.5505, + "step": 13528 + }, + { + "epoch": 0.908023220697292, + "grad_norm": 1.1323999166488647, + "learning_rate": 5.9809805914240104e-05, + "loss": 0.5769, + "step": 13530 + }, + { + "epoch": 0.908157444381061, + "grad_norm": 1.1161619424819946, + "learning_rate": 5.979914860521779e-05, + "loss": 0.6016, + "step": 13532 + }, + { + "epoch": 0.9082916680648301, + "grad_norm": 0.9982798099517822, + "learning_rate": 5.978849083320499e-05, + "loss": 0.5088, + "step": 13534 + }, + { + "epoch": 0.908425891748599, + "grad_norm": 0.9984214305877686, + "learning_rate": 5.977783259870524e-05, + "loss": 0.561, + "step": 13536 + }, + { + "epoch": 0.908560115432368, + "grad_norm": 1.203289270401001, + "learning_rate": 5.9767173902222116e-05, + "loss": 0.6009, + "step": 13538 + }, + { + "epoch": 0.9086943391161371, + "grad_norm": 1.0529379844665527, + "learning_rate": 5.975651474425925e-05, + "loss": 0.5708, + "step": 13540 + }, + { + "epoch": 0.9088285627999061, + "grad_norm": 0.941807746887207, + "learning_rate": 5.9745855125320236e-05, + "loss": 0.5475, + "step": 13542 + }, + { + "epoch": 0.908962786483675, + "grad_norm": 0.9625012278556824, + "learning_rate": 5.973519504590874e-05, + "loss": 0.5515, + "step": 13544 + }, + { + "epoch": 0.909097010167444, + "grad_norm": 1.1061104536056519, + "learning_rate": 5.972453450652842e-05, + "loss": 0.6741, + "step": 13546 + }, + { + "epoch": 0.9092312338512131, + "grad_norm": 1.1080636978149414, + "learning_rate": 5.971387350768297e-05, + "loss": 0.5872, + "step": 13548 + }, + { + "epoch": 0.9093654575349821, + "grad_norm": 1.0175831317901611, + "learning_rate": 5.970321204987609e-05, + "loss": 0.5422, + "step": 13550 + }, + { + "epoch": 0.909499681218751, + "grad_norm": 1.123096227645874, + "learning_rate": 5.969255013361153e-05, + "loss": 0.6282, + "step": 13552 + }, + { + "epoch": 0.90963390490252, + "grad_norm": 0.9660599827766418, + "learning_rate": 5.968188775939303e-05, + "loss": 0.5618, + "step": 13554 + }, + { + "epoch": 0.9097681285862891, + "grad_norm": 1.6226638555526733, + "learning_rate": 5.9671224927724366e-05, + "loss": 0.5281, + "step": 13556 + }, + { + "epoch": 0.909902352270058, + "grad_norm": 1.0370644330978394, + "learning_rate": 5.966056163910936e-05, + "loss": 0.6161, + "step": 13558 + }, + { + "epoch": 0.910036575953827, + "grad_norm": 0.9805968403816223, + "learning_rate": 5.9649897894051785e-05, + "loss": 0.5821, + "step": 13560 + }, + { + "epoch": 0.9101707996375961, + "grad_norm": 1.1472930908203125, + "learning_rate": 5.963923369305554e-05, + "loss": 0.5705, + "step": 13562 + }, + { + "epoch": 0.9103050233213651, + "grad_norm": 1.0938692092895508, + "learning_rate": 5.9628569036624446e-05, + "loss": 0.5824, + "step": 13564 + }, + { + "epoch": 0.910439247005134, + "grad_norm": 1.1195660829544067, + "learning_rate": 5.9617903925262395e-05, + "loss": 0.5666, + "step": 13566 + }, + { + "epoch": 0.910573470688903, + "grad_norm": 1.0558995008468628, + "learning_rate": 5.960723835947331e-05, + "loss": 0.6008, + "step": 13568 + }, + { + "epoch": 0.9107076943726721, + "grad_norm": 1.1444767713546753, + "learning_rate": 5.959657233976108e-05, + "loss": 0.5366, + "step": 13570 + }, + { + "epoch": 0.9108419180564411, + "grad_norm": 0.8425387144088745, + "learning_rate": 5.9585905866629687e-05, + "loss": 0.5151, + "step": 13572 + }, + { + "epoch": 0.91097614174021, + "grad_norm": 1.035324215888977, + "learning_rate": 5.95752389405831e-05, + "loss": 0.5506, + "step": 13574 + }, + { + "epoch": 0.911110365423979, + "grad_norm": 1.067969560623169, + "learning_rate": 5.9564571562125294e-05, + "loss": 0.5854, + "step": 13576 + }, + { + "epoch": 0.9112445891077481, + "grad_norm": 1.0317015647888184, + "learning_rate": 5.9553903731760295e-05, + "loss": 0.5845, + "step": 13578 + }, + { + "epoch": 0.9113788127915171, + "grad_norm": 1.012310266494751, + "learning_rate": 5.954323544999213e-05, + "loss": 0.5982, + "step": 13580 + }, + { + "epoch": 0.911513036475286, + "grad_norm": 0.8723168969154358, + "learning_rate": 5.953256671732487e-05, + "loss": 0.5747, + "step": 13582 + }, + { + "epoch": 0.9116472601590551, + "grad_norm": 0.97370445728302, + "learning_rate": 5.952189753426255e-05, + "loss": 0.5772, + "step": 13584 + }, + { + "epoch": 0.9117814838428241, + "grad_norm": 1.5741316080093384, + "learning_rate": 5.951122790130931e-05, + "loss": 0.5825, + "step": 13586 + }, + { + "epoch": 0.9119157075265931, + "grad_norm": 0.9457995891571045, + "learning_rate": 5.950055781896926e-05, + "loss": 0.5331, + "step": 13588 + }, + { + "epoch": 0.912049931210362, + "grad_norm": 0.8744380474090576, + "learning_rate": 5.948988728774652e-05, + "loss": 0.5643, + "step": 13590 + }, + { + "epoch": 0.9121841548941311, + "grad_norm": 0.9800755381584167, + "learning_rate": 5.947921630814528e-05, + "loss": 0.5018, + "step": 13592 + }, + { + "epoch": 0.9123183785779001, + "grad_norm": 1.0435247421264648, + "learning_rate": 5.94685448806697e-05, + "loss": 0.6186, + "step": 13594 + }, + { + "epoch": 0.912452602261669, + "grad_norm": 1.0551824569702148, + "learning_rate": 5.945787300582401e-05, + "loss": 0.6018, + "step": 13596 + }, + { + "epoch": 0.912586825945438, + "grad_norm": 0.974545955657959, + "learning_rate": 5.944720068411239e-05, + "loss": 0.5904, + "step": 13598 + }, + { + "epoch": 0.9127210496292071, + "grad_norm": 1.0696234703063965, + "learning_rate": 5.943652791603913e-05, + "loss": 0.5361, + "step": 13600 + }, + { + "epoch": 0.9128552733129761, + "grad_norm": 1.0218651294708252, + "learning_rate": 5.9425854702108494e-05, + "loss": 0.5753, + "step": 13602 + }, + { + "epoch": 0.912989496996745, + "grad_norm": 0.9226881861686707, + "learning_rate": 5.941518104282474e-05, + "loss": 0.5257, + "step": 13604 + }, + { + "epoch": 0.9131237206805141, + "grad_norm": 0.9766383171081543, + "learning_rate": 5.940450693869222e-05, + "loss": 0.5771, + "step": 13606 + }, + { + "epoch": 0.9132579443642831, + "grad_norm": 1.0304303169250488, + "learning_rate": 5.939383239021521e-05, + "loss": 0.5821, + "step": 13608 + }, + { + "epoch": 0.9133921680480521, + "grad_norm": 1.0369923114776611, + "learning_rate": 5.938315739789812e-05, + "loss": 0.5398, + "step": 13610 + }, + { + "epoch": 0.913526391731821, + "grad_norm": 0.9135134816169739, + "learning_rate": 5.9372481962245274e-05, + "loss": 0.5373, + "step": 13612 + }, + { + "epoch": 0.9136606154155901, + "grad_norm": 1.0946376323699951, + "learning_rate": 5.9361806083761084e-05, + "loss": 0.5835, + "step": 13614 + }, + { + "epoch": 0.9137948390993591, + "grad_norm": 0.9748871922492981, + "learning_rate": 5.935112976294997e-05, + "loss": 0.6036, + "step": 13616 + }, + { + "epoch": 0.9139290627831281, + "grad_norm": 1.154877781867981, + "learning_rate": 5.934045300031637e-05, + "loss": 0.5261, + "step": 13618 + }, + { + "epoch": 0.914063286466897, + "grad_norm": 1.0215812921524048, + "learning_rate": 5.932977579636474e-05, + "loss": 0.5244, + "step": 13620 + }, + { + "epoch": 0.9141975101506661, + "grad_norm": 0.9126201868057251, + "learning_rate": 5.9319098151599525e-05, + "loss": 0.5494, + "step": 13622 + }, + { + "epoch": 0.9143317338344351, + "grad_norm": 1.1465352773666382, + "learning_rate": 5.9308420066525274e-05, + "loss": 0.5866, + "step": 13624 + }, + { + "epoch": 0.9144659575182041, + "grad_norm": 1.2223682403564453, + "learning_rate": 5.9297741541646465e-05, + "loss": 0.5492, + "step": 13626 + }, + { + "epoch": 0.9146001812019731, + "grad_norm": 0.9171653985977173, + "learning_rate": 5.9287062577467646e-05, + "loss": 0.5318, + "step": 13628 + }, + { + "epoch": 0.9147344048857421, + "grad_norm": 0.9185870289802551, + "learning_rate": 5.92763831744934e-05, + "loss": 0.6155, + "step": 13630 + }, + { + "epoch": 0.9148686285695111, + "grad_norm": 1.004479169845581, + "learning_rate": 5.926570333322828e-05, + "loss": 0.5913, + "step": 13632 + }, + { + "epoch": 0.91500285225328, + "grad_norm": 1.0511237382888794, + "learning_rate": 5.9255023054176895e-05, + "loss": 0.5884, + "step": 13634 + }, + { + "epoch": 0.9151370759370491, + "grad_norm": 1.1387178897857666, + "learning_rate": 5.924434233784388e-05, + "loss": 0.6764, + "step": 13636 + }, + { + "epoch": 0.9152712996208181, + "grad_norm": 0.9619917273521423, + "learning_rate": 5.9233661184733856e-05, + "loss": 0.5727, + "step": 13638 + }, + { + "epoch": 0.9154055233045871, + "grad_norm": 0.9369518160820007, + "learning_rate": 5.92229795953515e-05, + "loss": 0.5183, + "step": 13640 + }, + { + "epoch": 0.915539746988356, + "grad_norm": 0.9075946807861328, + "learning_rate": 5.92122975702015e-05, + "loss": 0.469, + "step": 13642 + }, + { + "epoch": 0.9156739706721251, + "grad_norm": 0.9302048087120056, + "learning_rate": 5.9201615109788555e-05, + "loss": 0.5152, + "step": 13644 + }, + { + "epoch": 0.9158081943558941, + "grad_norm": 1.0235142707824707, + "learning_rate": 5.9190932214617376e-05, + "loss": 0.5364, + "step": 13646 + }, + { + "epoch": 0.9159424180396631, + "grad_norm": 1.1028952598571777, + "learning_rate": 5.9180248885192735e-05, + "loss": 0.6569, + "step": 13648 + }, + { + "epoch": 0.9160766417234321, + "grad_norm": 1.1462273597717285, + "learning_rate": 5.9169565122019386e-05, + "loss": 0.5516, + "step": 13650 + }, + { + "epoch": 0.9162108654072011, + "grad_norm": 1.035347819328308, + "learning_rate": 5.915888092560212e-05, + "loss": 0.5598, + "step": 13652 + }, + { + "epoch": 0.9163450890909701, + "grad_norm": 0.9631198048591614, + "learning_rate": 5.914819629644574e-05, + "loss": 0.5827, + "step": 13654 + }, + { + "epoch": 0.9164793127747392, + "grad_norm": 0.9795407652854919, + "learning_rate": 5.913751123505506e-05, + "loss": 0.606, + "step": 13656 + }, + { + "epoch": 0.9166135364585081, + "grad_norm": 1.05812406539917, + "learning_rate": 5.912682574193497e-05, + "loss": 0.6198, + "step": 13658 + }, + { + "epoch": 0.9167477601422771, + "grad_norm": 0.9368353486061096, + "learning_rate": 5.911613981759029e-05, + "loss": 0.5508, + "step": 13660 + }, + { + "epoch": 0.9168819838260461, + "grad_norm": 1.0112839937210083, + "learning_rate": 5.9105453462525915e-05, + "loss": 0.5814, + "step": 13662 + }, + { + "epoch": 0.9170162075098152, + "grad_norm": 1.1456836462020874, + "learning_rate": 5.90947666772468e-05, + "loss": 0.5903, + "step": 13664 + }, + { + "epoch": 0.9171504311935841, + "grad_norm": 0.9879085421562195, + "learning_rate": 5.9084079462257824e-05, + "loss": 0.5769, + "step": 13666 + }, + { + "epoch": 0.9172846548773531, + "grad_norm": 1.080277919769287, + "learning_rate": 5.907339181806397e-05, + "loss": 0.5787, + "step": 13668 + }, + { + "epoch": 0.9174188785611221, + "grad_norm": 0.9881849884986877, + "learning_rate": 5.906270374517019e-05, + "loss": 0.5421, + "step": 13670 + }, + { + "epoch": 0.9175531022448911, + "grad_norm": 1.0958977937698364, + "learning_rate": 5.905201524408148e-05, + "loss": 0.5934, + "step": 13672 + }, + { + "epoch": 0.9176873259286601, + "grad_norm": 0.9930013418197632, + "learning_rate": 5.9041326315302835e-05, + "loss": 0.5825, + "step": 13674 + }, + { + "epoch": 0.9178215496124291, + "grad_norm": 1.0344637632369995, + "learning_rate": 5.903063695933931e-05, + "loss": 0.6249, + "step": 13676 + }, + { + "epoch": 0.9179557732961982, + "grad_norm": 1.037278413772583, + "learning_rate": 5.9019947176695954e-05, + "loss": 0.6541, + "step": 13678 + }, + { + "epoch": 0.9180899969799671, + "grad_norm": 1.1754828691482544, + "learning_rate": 5.900925696787783e-05, + "loss": 0.5919, + "step": 13680 + }, + { + "epoch": 0.9182242206637361, + "grad_norm": 0.9693192839622498, + "learning_rate": 5.899856633339003e-05, + "loss": 0.5956, + "step": 13682 + }, + { + "epoch": 0.9183584443475051, + "grad_norm": 1.4770317077636719, + "learning_rate": 5.898787527373766e-05, + "loss": 0.5479, + "step": 13684 + }, + { + "epoch": 0.9184926680312742, + "grad_norm": 1.070414662361145, + "learning_rate": 5.897718378942586e-05, + "loss": 0.5375, + "step": 13686 + }, + { + "epoch": 0.9186268917150431, + "grad_norm": 1.149235725402832, + "learning_rate": 5.8966491880959775e-05, + "loss": 0.5922, + "step": 13688 + }, + { + "epoch": 0.9187611153988121, + "grad_norm": 0.9528571367263794, + "learning_rate": 5.895579954884458e-05, + "loss": 0.5335, + "step": 13690 + }, + { + "epoch": 0.9188953390825811, + "grad_norm": 1.0059806108474731, + "learning_rate": 5.894510679358547e-05, + "loss": 0.5089, + "step": 13692 + }, + { + "epoch": 0.9190295627663502, + "grad_norm": 0.9775877594947815, + "learning_rate": 5.8934413615687655e-05, + "loss": 0.547, + "step": 13694 + }, + { + "epoch": 0.9191637864501191, + "grad_norm": 1.1173903942108154, + "learning_rate": 5.892372001565637e-05, + "loss": 0.5916, + "step": 13696 + }, + { + "epoch": 0.9192980101338881, + "grad_norm": 1.368699550628662, + "learning_rate": 5.891302599399685e-05, + "loss": 0.5281, + "step": 13698 + }, + { + "epoch": 0.9194322338176572, + "grad_norm": 0.9711874127388, + "learning_rate": 5.8902331551214387e-05, + "loss": 0.5552, + "step": 13700 + }, + { + "epoch": 0.9195664575014262, + "grad_norm": 1.1150704622268677, + "learning_rate": 5.8891636687814276e-05, + "loss": 0.5741, + "step": 13702 + }, + { + "epoch": 0.9197006811851951, + "grad_norm": 1.1072198152542114, + "learning_rate": 5.8880941404301795e-05, + "loss": 0.5677, + "step": 13704 + }, + { + "epoch": 0.9198349048689641, + "grad_norm": 1.1020877361297607, + "learning_rate": 5.887024570118231e-05, + "loss": 0.5875, + "step": 13706 + }, + { + "epoch": 0.9199691285527332, + "grad_norm": 0.941753089427948, + "learning_rate": 5.885954957896115e-05, + "loss": 0.533, + "step": 13708 + }, + { + "epoch": 0.9201033522365021, + "grad_norm": 0.9490309357643127, + "learning_rate": 5.884885303814369e-05, + "loss": 0.5541, + "step": 13710 + }, + { + "epoch": 0.9202375759202711, + "grad_norm": 0.9945219159126282, + "learning_rate": 5.8838156079235326e-05, + "loss": 0.5433, + "step": 13712 + }, + { + "epoch": 0.9203717996040401, + "grad_norm": 1.004615068435669, + "learning_rate": 5.8827458702741465e-05, + "loss": 0.5921, + "step": 13714 + }, + { + "epoch": 0.9205060232878092, + "grad_norm": 0.942687451839447, + "learning_rate": 5.881676090916756e-05, + "loss": 0.5633, + "step": 13716 + }, + { + "epoch": 0.9206402469715781, + "grad_norm": 1.033555507659912, + "learning_rate": 5.880606269901902e-05, + "loss": 0.5705, + "step": 13718 + }, + { + "epoch": 0.9207744706553471, + "grad_norm": 1.065773367881775, + "learning_rate": 5.879536407280134e-05, + "loss": 0.5556, + "step": 13720 + }, + { + "epoch": 0.9209086943391162, + "grad_norm": 1.1345205307006836, + "learning_rate": 5.8784665031020004e-05, + "loss": 0.6009, + "step": 13722 + }, + { + "epoch": 0.9210429180228852, + "grad_norm": 1.0204540491104126, + "learning_rate": 5.87739655741805e-05, + "loss": 0.5334, + "step": 13724 + }, + { + "epoch": 0.9211771417066541, + "grad_norm": 0.9084629416465759, + "learning_rate": 5.8763265702788404e-05, + "loss": 0.4941, + "step": 13726 + }, + { + "epoch": 0.9213113653904231, + "grad_norm": 1.0169886350631714, + "learning_rate": 5.8752565417349215e-05, + "loss": 0.5576, + "step": 13728 + }, + { + "epoch": 0.9214455890741922, + "grad_norm": 0.9902750849723816, + "learning_rate": 5.874186471836854e-05, + "loss": 0.6228, + "step": 13730 + }, + { + "epoch": 0.9215798127579612, + "grad_norm": 1.0019904375076294, + "learning_rate": 5.8731163606351933e-05, + "loss": 0.6261, + "step": 13732 + }, + { + "epoch": 0.9217140364417301, + "grad_norm": 0.9995107650756836, + "learning_rate": 5.872046208180503e-05, + "loss": 0.5553, + "step": 13734 + }, + { + "epoch": 0.9218482601254991, + "grad_norm": 1.1056007146835327, + "learning_rate": 5.8709760145233416e-05, + "loss": 0.5882, + "step": 13736 + }, + { + "epoch": 0.9219824838092682, + "grad_norm": 0.9298726916313171, + "learning_rate": 5.869905779714278e-05, + "loss": 0.5271, + "step": 13738 + }, + { + "epoch": 0.9221167074930371, + "grad_norm": 1.219878911972046, + "learning_rate": 5.868835503803877e-05, + "loss": 0.6204, + "step": 13740 + }, + { + "epoch": 0.9222509311768061, + "grad_norm": 1.0206249952316284, + "learning_rate": 5.867765186842706e-05, + "loss": 0.5647, + "step": 13742 + }, + { + "epoch": 0.9223851548605752, + "grad_norm": 0.9591622948646545, + "learning_rate": 5.866694828881337e-05, + "loss": 0.5728, + "step": 13744 + }, + { + "epoch": 0.9225193785443442, + "grad_norm": 1.1038897037506104, + "learning_rate": 5.8656244299703414e-05, + "loss": 0.591, + "step": 13746 + }, + { + "epoch": 0.9226536022281131, + "grad_norm": 1.1192244291305542, + "learning_rate": 5.864553990160294e-05, + "loss": 0.6176, + "step": 13748 + }, + { + "epoch": 0.9227878259118821, + "grad_norm": 1.0274847745895386, + "learning_rate": 5.86348350950177e-05, + "loss": 0.6095, + "step": 13750 + }, + { + "epoch": 0.9229220495956512, + "grad_norm": 0.9790551662445068, + "learning_rate": 5.8624129880453485e-05, + "loss": 0.5717, + "step": 13752 + }, + { + "epoch": 0.9230562732794202, + "grad_norm": 1.0101349353790283, + "learning_rate": 5.8613424258416094e-05, + "loss": 0.6067, + "step": 13754 + }, + { + "epoch": 0.9231904969631891, + "grad_norm": 1.0220857858657837, + "learning_rate": 5.8602718229411335e-05, + "loss": 0.5313, + "step": 13756 + }, + { + "epoch": 0.9233247206469581, + "grad_norm": 0.9489063620567322, + "learning_rate": 5.859201179394508e-05, + "loss": 0.5373, + "step": 13758 + }, + { + "epoch": 0.9234589443307272, + "grad_norm": 0.9276555180549622, + "learning_rate": 5.858130495252314e-05, + "loss": 0.5868, + "step": 13760 + }, + { + "epoch": 0.9235931680144962, + "grad_norm": 1.358279824256897, + "learning_rate": 5.857059770565142e-05, + "loss": 0.5832, + "step": 13762 + }, + { + "epoch": 0.9237273916982651, + "grad_norm": 1.062290072441101, + "learning_rate": 5.855989005383581e-05, + "loss": 0.5398, + "step": 13764 + }, + { + "epoch": 0.9238616153820342, + "grad_norm": 1.0557093620300293, + "learning_rate": 5.8549181997582225e-05, + "loss": 0.6448, + "step": 13766 + }, + { + "epoch": 0.9239958390658032, + "grad_norm": 1.2883045673370361, + "learning_rate": 5.853847353739663e-05, + "loss": 0.5191, + "step": 13768 + }, + { + "epoch": 0.9241300627495722, + "grad_norm": 1.1146330833435059, + "learning_rate": 5.852776467378492e-05, + "loss": 0.6236, + "step": 13770 + }, + { + "epoch": 0.9242642864333411, + "grad_norm": 1.016501784324646, + "learning_rate": 5.8517055407253115e-05, + "loss": 0.5872, + "step": 13772 + }, + { + "epoch": 0.9243985101171102, + "grad_norm": 0.9463066458702087, + "learning_rate": 5.850634573830718e-05, + "loss": 0.5085, + "step": 13774 + }, + { + "epoch": 0.9245327338008792, + "grad_norm": 0.9416144490242004, + "learning_rate": 5.849563566745313e-05, + "loss": 0.567, + "step": 13776 + }, + { + "epoch": 0.9246669574846481, + "grad_norm": 1.0586539506912231, + "learning_rate": 5.8484925195197016e-05, + "loss": 0.5298, + "step": 13778 + }, + { + "epoch": 0.9248011811684171, + "grad_norm": 0.986104428768158, + "learning_rate": 5.847421432204486e-05, + "loss": 0.5542, + "step": 13780 + }, + { + "epoch": 0.9249354048521862, + "grad_norm": 1.1029913425445557, + "learning_rate": 5.846350304850274e-05, + "loss": 0.5755, + "step": 13782 + }, + { + "epoch": 0.9250696285359552, + "grad_norm": 0.9016444087028503, + "learning_rate": 5.845279137507675e-05, + "loss": 0.5321, + "step": 13784 + }, + { + "epoch": 0.9252038522197241, + "grad_norm": 0.9199962615966797, + "learning_rate": 5.8442079302273e-05, + "loss": 0.5333, + "step": 13786 + }, + { + "epoch": 0.9253380759034932, + "grad_norm": 0.9962599277496338, + "learning_rate": 5.843136683059758e-05, + "loss": 0.5798, + "step": 13788 + }, + { + "epoch": 0.9254722995872622, + "grad_norm": 1.060033917427063, + "learning_rate": 5.842065396055667e-05, + "loss": 0.573, + "step": 13790 + }, + { + "epoch": 0.9256065232710312, + "grad_norm": 1.0801200866699219, + "learning_rate": 5.8409940692656416e-05, + "loss": 0.5885, + "step": 13792 + }, + { + "epoch": 0.9257407469548001, + "grad_norm": 1.0969904661178589, + "learning_rate": 5.8399227027403e-05, + "loss": 0.5462, + "step": 13794 + }, + { + "epoch": 0.9258749706385692, + "grad_norm": 1.4458987712860107, + "learning_rate": 5.838851296530263e-05, + "loss": 0.5204, + "step": 13796 + }, + { + "epoch": 0.9260091943223382, + "grad_norm": 0.9598033428192139, + "learning_rate": 5.837779850686152e-05, + "loss": 0.5761, + "step": 13798 + }, + { + "epoch": 0.9261434180061072, + "grad_norm": 1.0324113368988037, + "learning_rate": 5.836708365258589e-05, + "loss": 0.6112, + "step": 13800 + }, + { + "epoch": 0.9262776416898761, + "grad_norm": 0.886147141456604, + "learning_rate": 5.835636840298202e-05, + "loss": 0.5548, + "step": 13802 + }, + { + "epoch": 0.9264118653736452, + "grad_norm": 1.0079913139343262, + "learning_rate": 5.834565275855617e-05, + "loss": 0.55, + "step": 13804 + }, + { + "epoch": 0.9265460890574142, + "grad_norm": 1.0281902551651, + "learning_rate": 5.833493671981465e-05, + "loss": 0.5206, + "step": 13806 + }, + { + "epoch": 0.9266803127411832, + "grad_norm": 0.8789565563201904, + "learning_rate": 5.832422028726375e-05, + "loss": 0.4976, + "step": 13808 + }, + { + "epoch": 0.9268145364249522, + "grad_norm": 1.0310882329940796, + "learning_rate": 5.8313503461409826e-05, + "loss": 0.5597, + "step": 13810 + }, + { + "epoch": 0.9269487601087212, + "grad_norm": 1.0418280363082886, + "learning_rate": 5.830278624275919e-05, + "loss": 0.5402, + "step": 13812 + }, + { + "epoch": 0.9270829837924902, + "grad_norm": 1.0995814800262451, + "learning_rate": 5.829206863181823e-05, + "loss": 0.5538, + "step": 13814 + }, + { + "epoch": 0.9272172074762591, + "grad_norm": 0.9341369867324829, + "learning_rate": 5.8281350629093346e-05, + "loss": 0.5543, + "step": 13816 + }, + { + "epoch": 0.9273514311600282, + "grad_norm": 1.0317434072494507, + "learning_rate": 5.8270632235090916e-05, + "loss": 0.5834, + "step": 13818 + }, + { + "epoch": 0.9274856548437972, + "grad_norm": 1.0883063077926636, + "learning_rate": 5.825991345031739e-05, + "loss": 0.5688, + "step": 13820 + }, + { + "epoch": 0.9276198785275662, + "grad_norm": 1.025431752204895, + "learning_rate": 5.82491942752792e-05, + "loss": 0.6176, + "step": 13822 + }, + { + "epoch": 0.9277541022113351, + "grad_norm": 0.9473832845687866, + "learning_rate": 5.823847471048279e-05, + "loss": 0.5329, + "step": 13824 + }, + { + "epoch": 0.9278883258951042, + "grad_norm": 0.9456143379211426, + "learning_rate": 5.822775475643465e-05, + "loss": 0.7219, + "step": 13826 + }, + { + "epoch": 0.9280225495788732, + "grad_norm": 0.8828773498535156, + "learning_rate": 5.821703441364128e-05, + "loss": 0.6389, + "step": 13828 + }, + { + "epoch": 0.9281567732626422, + "grad_norm": 0.9927616715431213, + "learning_rate": 5.820631368260919e-05, + "loss": 0.5574, + "step": 13830 + }, + { + "epoch": 0.9282909969464112, + "grad_norm": 1.0232497453689575, + "learning_rate": 5.819559256384492e-05, + "loss": 0.5249, + "step": 13832 + }, + { + "epoch": 0.9284252206301802, + "grad_norm": 0.9136453866958618, + "learning_rate": 5.818487105785502e-05, + "loss": 0.5522, + "step": 13834 + }, + { + "epoch": 0.9285594443139492, + "grad_norm": 1.032614827156067, + "learning_rate": 5.8174149165146044e-05, + "loss": 0.5188, + "step": 13836 + }, + { + "epoch": 0.9286936679977182, + "grad_norm": 1.0510674715042114, + "learning_rate": 5.816342688622462e-05, + "loss": 0.604, + "step": 13838 + }, + { + "epoch": 0.9288278916814872, + "grad_norm": 1.0441523790359497, + "learning_rate": 5.815270422159731e-05, + "loss": 0.556, + "step": 13840 + }, + { + "epoch": 0.9289621153652562, + "grad_norm": 1.0888450145721436, + "learning_rate": 5.8141981171770755e-05, + "loss": 0.6212, + "step": 13842 + }, + { + "epoch": 0.9290963390490252, + "grad_norm": 0.9734417796134949, + "learning_rate": 5.81312577372516e-05, + "loss": 0.6006, + "step": 13844 + }, + { + "epoch": 0.9292305627327943, + "grad_norm": 1.0175516605377197, + "learning_rate": 5.8120533918546506e-05, + "loss": 0.5665, + "step": 13846 + }, + { + "epoch": 0.9293647864165632, + "grad_norm": 1.0254477262496948, + "learning_rate": 5.8109809716162164e-05, + "loss": 0.5918, + "step": 13848 + }, + { + "epoch": 0.9294990101003322, + "grad_norm": 0.9375723600387573, + "learning_rate": 5.809908513060524e-05, + "loss": 0.5815, + "step": 13850 + }, + { + "epoch": 0.9296332337841012, + "grad_norm": 0.9494642019271851, + "learning_rate": 5.8088360162382486e-05, + "loss": 0.5414, + "step": 13852 + }, + { + "epoch": 0.9297674574678702, + "grad_norm": 1.0334968566894531, + "learning_rate": 5.8077634812000614e-05, + "loss": 0.566, + "step": 13854 + }, + { + "epoch": 0.9299016811516392, + "grad_norm": 1.381908893585205, + "learning_rate": 5.806690907996638e-05, + "loss": 0.5649, + "step": 13856 + }, + { + "epoch": 0.9300359048354082, + "grad_norm": 1.1403449773788452, + "learning_rate": 5.8056182966786566e-05, + "loss": 0.6084, + "step": 13858 + }, + { + "epoch": 0.9301701285191772, + "grad_norm": 1.0421216487884521, + "learning_rate": 5.804545647296793e-05, + "loss": 0.5579, + "step": 13860 + }, + { + "epoch": 0.9303043522029462, + "grad_norm": 1.0358749628067017, + "learning_rate": 5.803472959901731e-05, + "loss": 0.591, + "step": 13862 + }, + { + "epoch": 0.9304385758867152, + "grad_norm": 1.1746695041656494, + "learning_rate": 5.802400234544152e-05, + "loss": 0.6525, + "step": 13864 + }, + { + "epoch": 0.9305727995704842, + "grad_norm": 1.0326679944992065, + "learning_rate": 5.801327471274738e-05, + "loss": 0.5865, + "step": 13866 + }, + { + "epoch": 0.9307070232542533, + "grad_norm": 1.0468127727508545, + "learning_rate": 5.8002546701441785e-05, + "loss": 0.6163, + "step": 13868 + }, + { + "epoch": 0.9308412469380222, + "grad_norm": 0.9539506435394287, + "learning_rate": 5.7991818312031575e-05, + "loss": 0.5444, + "step": 13870 + }, + { + "epoch": 0.9309754706217912, + "grad_norm": 0.9731017351150513, + "learning_rate": 5.798108954502368e-05, + "loss": 0.5387, + "step": 13872 + }, + { + "epoch": 0.9311096943055602, + "grad_norm": 0.9137818813323975, + "learning_rate": 5.7970360400924994e-05, + "loss": 0.5162, + "step": 13874 + }, + { + "epoch": 0.9312439179893293, + "grad_norm": 1.7687433958053589, + "learning_rate": 5.795963088024247e-05, + "loss": 0.584, + "step": 13876 + }, + { + "epoch": 0.9313781416730982, + "grad_norm": 1.2483657598495483, + "learning_rate": 5.794890098348301e-05, + "loss": 0.597, + "step": 13878 + }, + { + "epoch": 0.9315123653568672, + "grad_norm": 1.0281192064285278, + "learning_rate": 5.7938170711153614e-05, + "loss": 0.5919, + "step": 13880 + }, + { + "epoch": 0.9316465890406362, + "grad_norm": 1.559960126876831, + "learning_rate": 5.792744006376127e-05, + "loss": 0.5408, + "step": 13882 + }, + { + "epoch": 0.9317808127244053, + "grad_norm": 1.0883233547210693, + "learning_rate": 5.791670904181297e-05, + "loss": 0.546, + "step": 13884 + }, + { + "epoch": 0.9319150364081742, + "grad_norm": 1.0710606575012207, + "learning_rate": 5.7905977645815745e-05, + "loss": 0.595, + "step": 13886 + }, + { + "epoch": 0.9320492600919432, + "grad_norm": 1.2496391534805298, + "learning_rate": 5.789524587627661e-05, + "loss": 0.5146, + "step": 13888 + }, + { + "epoch": 0.9321834837757123, + "grad_norm": 0.9544695615768433, + "learning_rate": 5.788451373370263e-05, + "loss": 0.583, + "step": 13890 + }, + { + "epoch": 0.9323177074594812, + "grad_norm": 1.0018619298934937, + "learning_rate": 5.78737812186009e-05, + "loss": 0.615, + "step": 13892 + }, + { + "epoch": 0.9324519311432502, + "grad_norm": 1.0168393850326538, + "learning_rate": 5.7863048331478466e-05, + "loss": 0.5867, + "step": 13894 + }, + { + "epoch": 0.9325861548270192, + "grad_norm": 0.9545914530754089, + "learning_rate": 5.785231507284248e-05, + "loss": 0.6116, + "step": 13896 + }, + { + "epoch": 0.9327203785107883, + "grad_norm": 1.0262305736541748, + "learning_rate": 5.7841581443200035e-05, + "loss": 0.5901, + "step": 13898 + }, + { + "epoch": 0.9328546021945572, + "grad_norm": 0.9550249576568604, + "learning_rate": 5.783084744305829e-05, + "loss": 0.5452, + "step": 13900 + }, + { + "epoch": 0.9329888258783262, + "grad_norm": 0.8978448510169983, + "learning_rate": 5.7820113072924395e-05, + "loss": 0.5158, + "step": 13902 + }, + { + "epoch": 0.9331230495620952, + "grad_norm": 1.11091148853302, + "learning_rate": 5.780937833330554e-05, + "loss": 0.5729, + "step": 13904 + }, + { + "epoch": 0.9332572732458643, + "grad_norm": 0.9957830905914307, + "learning_rate": 5.779864322470894e-05, + "loss": 0.6102, + "step": 13906 + }, + { + "epoch": 0.9333914969296332, + "grad_norm": 1.2966653108596802, + "learning_rate": 5.778790774764176e-05, + "loss": 0.5261, + "step": 13908 + }, + { + "epoch": 0.9335257206134022, + "grad_norm": 1.0487574338912964, + "learning_rate": 5.777717190261125e-05, + "loss": 0.5847, + "step": 13910 + }, + { + "epoch": 0.9336599442971713, + "grad_norm": 1.1108134984970093, + "learning_rate": 5.7766435690124667e-05, + "loss": 0.5592, + "step": 13912 + }, + { + "epoch": 0.9337941679809403, + "grad_norm": 0.9877544045448303, + "learning_rate": 5.775569911068925e-05, + "loss": 0.4801, + "step": 13914 + }, + { + "epoch": 0.9339283916647092, + "grad_norm": 1.107453465461731, + "learning_rate": 5.774496216481233e-05, + "loss": 0.5566, + "step": 13916 + }, + { + "epoch": 0.9340626153484782, + "grad_norm": 1.3392894268035889, + "learning_rate": 5.773422485300116e-05, + "loss": 0.5968, + "step": 13918 + }, + { + "epoch": 0.9341968390322473, + "grad_norm": 1.0226224660873413, + "learning_rate": 5.772348717576309e-05, + "loss": 0.6067, + "step": 13920 + }, + { + "epoch": 0.9343310627160163, + "grad_norm": 1.0401580333709717, + "learning_rate": 5.771274913360543e-05, + "loss": 0.5746, + "step": 13922 + }, + { + "epoch": 0.9344652863997852, + "grad_norm": 1.0642728805541992, + "learning_rate": 5.7702010727035536e-05, + "loss": 0.5197, + "step": 13924 + }, + { + "epoch": 0.9345995100835542, + "grad_norm": 1.141188383102417, + "learning_rate": 5.769127195656079e-05, + "loss": 0.593, + "step": 13926 + }, + { + "epoch": 0.9347337337673233, + "grad_norm": 1.0031726360321045, + "learning_rate": 5.768053282268855e-05, + "loss": 0.5746, + "step": 13928 + }, + { + "epoch": 0.9348679574510922, + "grad_norm": 0.9465569853782654, + "learning_rate": 5.766979332592626e-05, + "loss": 0.5283, + "step": 13930 + }, + { + "epoch": 0.9350021811348612, + "grad_norm": 1.2472999095916748, + "learning_rate": 5.76590534667813e-05, + "loss": 0.522, + "step": 13932 + }, + { + "epoch": 0.9351364048186303, + "grad_norm": 1.7484946250915527, + "learning_rate": 5.764831324576113e-05, + "loss": 0.5881, + "step": 13934 + }, + { + "epoch": 0.9352706285023993, + "grad_norm": 1.056246042251587, + "learning_rate": 5.7637572663373194e-05, + "loss": 0.5656, + "step": 13936 + }, + { + "epoch": 0.9354048521861682, + "grad_norm": 0.9071460366249084, + "learning_rate": 5.762683172012498e-05, + "loss": 0.4649, + "step": 13938 + }, + { + "epoch": 0.9355390758699372, + "grad_norm": 1.0678855180740356, + "learning_rate": 5.761609041652396e-05, + "loss": 0.5451, + "step": 13940 + }, + { + "epoch": 0.9356732995537063, + "grad_norm": 0.9900966286659241, + "learning_rate": 5.7605348753077634e-05, + "loss": 0.5007, + "step": 13942 + }, + { + "epoch": 0.9358075232374753, + "grad_norm": 1.0609707832336426, + "learning_rate": 5.7594606730293554e-05, + "loss": 0.5787, + "step": 13944 + }, + { + "epoch": 0.9359417469212442, + "grad_norm": 1.087057113647461, + "learning_rate": 5.7583864348679226e-05, + "loss": 0.6216, + "step": 13946 + }, + { + "epoch": 0.9360759706050132, + "grad_norm": 1.0868579149246216, + "learning_rate": 5.7573121608742234e-05, + "loss": 0.6006, + "step": 13948 + }, + { + "epoch": 0.9362101942887823, + "grad_norm": 1.0762981176376343, + "learning_rate": 5.7562378510990125e-05, + "loss": 0.58, + "step": 13950 + }, + { + "epoch": 0.9363444179725513, + "grad_norm": 1.1625175476074219, + "learning_rate": 5.755163505593051e-05, + "loss": 0.5484, + "step": 13952 + }, + { + "epoch": 0.9364786416563202, + "grad_norm": 1.0515735149383545, + "learning_rate": 5.754089124407097e-05, + "loss": 0.5732, + "step": 13954 + }, + { + "epoch": 0.9366128653400893, + "grad_norm": 0.9707502126693726, + "learning_rate": 5.753014707591916e-05, + "loss": 0.5438, + "step": 13956 + }, + { + "epoch": 0.9367470890238583, + "grad_norm": 1.2039716243743896, + "learning_rate": 5.751940255198272e-05, + "loss": 0.6074, + "step": 13958 + }, + { + "epoch": 0.9368813127076273, + "grad_norm": 1.8390384912490845, + "learning_rate": 5.750865767276927e-05, + "loss": 0.5314, + "step": 13960 + }, + { + "epoch": 0.9370155363913962, + "grad_norm": 1.0250873565673828, + "learning_rate": 5.7497912438786536e-05, + "loss": 0.6513, + "step": 13962 + }, + { + "epoch": 0.9371497600751653, + "grad_norm": 1.1512795686721802, + "learning_rate": 5.7487166850542165e-05, + "loss": 0.564, + "step": 13964 + }, + { + "epoch": 0.9372839837589343, + "grad_norm": 1.3163058757781982, + "learning_rate": 5.74764209085439e-05, + "loss": 0.5528, + "step": 13966 + }, + { + "epoch": 0.9374182074427032, + "grad_norm": 1.0196266174316406, + "learning_rate": 5.746567461329943e-05, + "loss": 0.5459, + "step": 13968 + }, + { + "epoch": 0.9375524311264722, + "grad_norm": 1.0521317720413208, + "learning_rate": 5.7454927965316516e-05, + "loss": 0.5021, + "step": 13970 + }, + { + "epoch": 0.9376866548102413, + "grad_norm": 0.9628283381462097, + "learning_rate": 5.7444180965102936e-05, + "loss": 0.5569, + "step": 13972 + }, + { + "epoch": 0.9378208784940103, + "grad_norm": 1.032279372215271, + "learning_rate": 5.743343361316644e-05, + "loss": 0.6062, + "step": 13974 + }, + { + "epoch": 0.9379551021777792, + "grad_norm": 1.5314966440200806, + "learning_rate": 5.742268591001481e-05, + "loss": 0.6062, + "step": 13976 + }, + { + "epoch": 0.9380893258615483, + "grad_norm": 1.048116683959961, + "learning_rate": 5.741193785615587e-05, + "loss": 0.5767, + "step": 13978 + }, + { + "epoch": 0.9382235495453173, + "grad_norm": 1.027986764907837, + "learning_rate": 5.740118945209744e-05, + "loss": 0.5113, + "step": 13980 + }, + { + "epoch": 0.9383577732290863, + "grad_norm": 0.9534016847610474, + "learning_rate": 5.739044069834737e-05, + "loss": 0.5451, + "step": 13982 + }, + { + "epoch": 0.9384919969128552, + "grad_norm": 0.9986965656280518, + "learning_rate": 5.73796915954135e-05, + "loss": 0.5946, + "step": 13984 + }, + { + "epoch": 0.9386262205966243, + "grad_norm": 0.9526494145393372, + "learning_rate": 5.7368942143803725e-05, + "loss": 0.616, + "step": 13986 + }, + { + "epoch": 0.9387604442803933, + "grad_norm": 0.9984648823738098, + "learning_rate": 5.735819234402591e-05, + "loss": 0.5458, + "step": 13988 + }, + { + "epoch": 0.9388946679641623, + "grad_norm": 1.1599220037460327, + "learning_rate": 5.7347442196587986e-05, + "loss": 0.5715, + "step": 13990 + }, + { + "epoch": 0.9390288916479312, + "grad_norm": 1.0007301568984985, + "learning_rate": 5.7336691701997866e-05, + "loss": 0.5822, + "step": 13992 + }, + { + "epoch": 0.9391631153317003, + "grad_norm": 1.06472647190094, + "learning_rate": 5.732594086076348e-05, + "loss": 0.6006, + "step": 13994 + }, + { + "epoch": 0.9392973390154693, + "grad_norm": 0.8864287734031677, + "learning_rate": 5.731518967339281e-05, + "loss": 0.6629, + "step": 13996 + }, + { + "epoch": 0.9394315626992383, + "grad_norm": 1.0640459060668945, + "learning_rate": 5.730443814039379e-05, + "loss": 0.5738, + "step": 13998 + }, + { + "epoch": 0.9395657863830073, + "grad_norm": 1.1999764442443848, + "learning_rate": 5.729368626227446e-05, + "loss": 0.5897, + "step": 14000 + }, + { + "epoch": 0.9397000100667763, + "grad_norm": 0.9244864583015442, + "learning_rate": 5.728293403954278e-05, + "loss": 0.5637, + "step": 14002 + }, + { + "epoch": 0.9398342337505453, + "grad_norm": 1.094678282737732, + "learning_rate": 5.727218147270678e-05, + "loss": 0.6543, + "step": 14004 + }, + { + "epoch": 0.9399684574343142, + "grad_norm": 1.1416062116622925, + "learning_rate": 5.726142856227452e-05, + "loss": 0.5896, + "step": 14006 + }, + { + "epoch": 0.9401026811180833, + "grad_norm": 1.0241265296936035, + "learning_rate": 5.725067530875403e-05, + "loss": 0.578, + "step": 14008 + }, + { + "epoch": 0.9402369048018523, + "grad_norm": 1.0072925090789795, + "learning_rate": 5.72399217126534e-05, + "loss": 0.539, + "step": 14010 + }, + { + "epoch": 0.9403711284856213, + "grad_norm": 0.9036150574684143, + "learning_rate": 5.722916777448069e-05, + "loss": 0.5875, + "step": 14012 + }, + { + "epoch": 0.9405053521693902, + "grad_norm": 0.9075189828872681, + "learning_rate": 5.721841349474404e-05, + "loss": 0.5424, + "step": 14014 + }, + { + "epoch": 0.9406395758531593, + "grad_norm": 0.9381653070449829, + "learning_rate": 5.720765887395153e-05, + "loss": 0.5986, + "step": 14016 + }, + { + "epoch": 0.9407737995369283, + "grad_norm": 1.1812331676483154, + "learning_rate": 5.719690391261131e-05, + "loss": 0.586, + "step": 14018 + }, + { + "epoch": 0.9409080232206973, + "grad_norm": 1.0900657176971436, + "learning_rate": 5.718614861123155e-05, + "loss": 0.5922, + "step": 14020 + }, + { + "epoch": 0.9410422469044663, + "grad_norm": 1.0432283878326416, + "learning_rate": 5.717539297032039e-05, + "loss": 0.6569, + "step": 14022 + }, + { + "epoch": 0.9411764705882353, + "grad_norm": 1.0058923959732056, + "learning_rate": 5.716463699038602e-05, + "loss": 0.5696, + "step": 14024 + }, + { + "epoch": 0.9413106942720043, + "grad_norm": 1.1516404151916504, + "learning_rate": 5.7153880671936635e-05, + "loss": 0.6802, + "step": 14026 + }, + { + "epoch": 0.9414449179557733, + "grad_norm": 1.3103623390197754, + "learning_rate": 5.7143124015480466e-05, + "loss": 0.6526, + "step": 14028 + }, + { + "epoch": 0.9415791416395423, + "grad_norm": 0.9740951657295227, + "learning_rate": 5.713236702152572e-05, + "loss": 0.6088, + "step": 14030 + }, + { + "epoch": 0.9417133653233113, + "grad_norm": 0.9947975873947144, + "learning_rate": 5.7121609690580666e-05, + "loss": 0.5089, + "step": 14032 + }, + { + "epoch": 0.9418475890070803, + "grad_norm": 1.1044875383377075, + "learning_rate": 5.711085202315356e-05, + "loss": 0.5827, + "step": 14034 + }, + { + "epoch": 0.9419818126908494, + "grad_norm": 1.043334722518921, + "learning_rate": 5.710009401975268e-05, + "loss": 0.5639, + "step": 14036 + }, + { + "epoch": 0.9421160363746183, + "grad_norm": 0.9688626527786255, + "learning_rate": 5.708933568088632e-05, + "loss": 0.5751, + "step": 14038 + }, + { + "epoch": 0.9422502600583873, + "grad_norm": 1.048002004623413, + "learning_rate": 5.707857700706278e-05, + "loss": 0.5697, + "step": 14040 + }, + { + "epoch": 0.9423844837421563, + "grad_norm": 1.013647198677063, + "learning_rate": 5.706781799879041e-05, + "loss": 0.5997, + "step": 14042 + }, + { + "epoch": 0.9425187074259253, + "grad_norm": 0.9789992570877075, + "learning_rate": 5.705705865657753e-05, + "loss": 0.5131, + "step": 14044 + }, + { + "epoch": 0.9426529311096943, + "grad_norm": 0.9202343821525574, + "learning_rate": 5.704629898093251e-05, + "loss": 0.4936, + "step": 14046 + }, + { + "epoch": 0.9427871547934633, + "grad_norm": 1.0707128047943115, + "learning_rate": 5.703553897236372e-05, + "loss": 0.6015, + "step": 14048 + }, + { + "epoch": 0.9429213784772323, + "grad_norm": 1.1972709894180298, + "learning_rate": 5.702477863137954e-05, + "loss": 0.5278, + "step": 14050 + }, + { + "epoch": 0.9430556021610013, + "grad_norm": 0.9796330332756042, + "learning_rate": 5.7014017958488375e-05, + "loss": 0.5181, + "step": 14052 + }, + { + "epoch": 0.9431898258447703, + "grad_norm": 1.0746116638183594, + "learning_rate": 5.700325695419868e-05, + "loss": 0.5927, + "step": 14054 + }, + { + "epoch": 0.9433240495285393, + "grad_norm": 0.8973516821861267, + "learning_rate": 5.699249561901884e-05, + "loss": 0.5777, + "step": 14056 + }, + { + "epoch": 0.9434582732123084, + "grad_norm": 1.0385633707046509, + "learning_rate": 5.698173395345735e-05, + "loss": 0.6789, + "step": 14058 + }, + { + "epoch": 0.9435924968960773, + "grad_norm": 1.1208579540252686, + "learning_rate": 5.6970971958022644e-05, + "loss": 0.5817, + "step": 14060 + }, + { + "epoch": 0.9437267205798463, + "grad_norm": 1.0253500938415527, + "learning_rate": 5.696020963322324e-05, + "loss": 0.5461, + "step": 14062 + }, + { + "epoch": 0.9438609442636153, + "grad_norm": 0.8999269604682922, + "learning_rate": 5.69494469795676e-05, + "loss": 0.5908, + "step": 14064 + }, + { + "epoch": 0.9439951679473844, + "grad_norm": 1.0557160377502441, + "learning_rate": 5.693868399756426e-05, + "loss": 0.6127, + "step": 14066 + }, + { + "epoch": 0.9441293916311533, + "grad_norm": 0.8857409954071045, + "learning_rate": 5.692792068772176e-05, + "loss": 0.5033, + "step": 14068 + }, + { + "epoch": 0.9442636153149223, + "grad_norm": 0.8972952365875244, + "learning_rate": 5.691715705054861e-05, + "loss": 0.549, + "step": 14070 + }, + { + "epoch": 0.9443978389986913, + "grad_norm": 1.5294030904769897, + "learning_rate": 5.69063930865534e-05, + "loss": 0.6064, + "step": 14072 + }, + { + "epoch": 0.9445320626824604, + "grad_norm": 0.9840405583381653, + "learning_rate": 5.689562879624469e-05, + "loss": 0.5188, + "step": 14074 + }, + { + "epoch": 0.9446662863662293, + "grad_norm": 1.0034092664718628, + "learning_rate": 5.68848641801311e-05, + "loss": 0.5347, + "step": 14076 + }, + { + "epoch": 0.9448005100499983, + "grad_norm": 1.0162858963012695, + "learning_rate": 5.68740992387212e-05, + "loss": 0.5567, + "step": 14078 + }, + { + "epoch": 0.9449347337337674, + "grad_norm": 1.1284905672073364, + "learning_rate": 5.686333397252363e-05, + "loss": 0.5736, + "step": 14080 + }, + { + "epoch": 0.9450689574175363, + "grad_norm": 0.9935622811317444, + "learning_rate": 5.6852568382047026e-05, + "loss": 0.5967, + "step": 14082 + }, + { + "epoch": 0.9452031811013053, + "grad_norm": 0.9152777791023254, + "learning_rate": 5.684180246780004e-05, + "loss": 0.5198, + "step": 14084 + }, + { + "epoch": 0.9453374047850743, + "grad_norm": 1.104828953742981, + "learning_rate": 5.683103623029135e-05, + "loss": 0.5803, + "step": 14086 + }, + { + "epoch": 0.9454716284688434, + "grad_norm": 0.9986449480056763, + "learning_rate": 5.6820269670029615e-05, + "loss": 0.5851, + "step": 14088 + }, + { + "epoch": 0.9456058521526123, + "grad_norm": 0.8983079791069031, + "learning_rate": 5.680950278752356e-05, + "loss": 0.5776, + "step": 14090 + }, + { + "epoch": 0.9457400758363813, + "grad_norm": 1.0416613817214966, + "learning_rate": 5.679873558328188e-05, + "loss": 0.5336, + "step": 14092 + }, + { + "epoch": 0.9458742995201503, + "grad_norm": 0.8977410793304443, + "learning_rate": 5.6787968057813324e-05, + "loss": 0.6275, + "step": 14094 + }, + { + "epoch": 0.9460085232039194, + "grad_norm": 1.108778953552246, + "learning_rate": 5.6777200211626624e-05, + "loss": 0.6378, + "step": 14096 + }, + { + "epoch": 0.9461427468876883, + "grad_norm": 1.1451404094696045, + "learning_rate": 5.6766432045230536e-05, + "loss": 0.631, + "step": 14098 + }, + { + "epoch": 0.9462769705714573, + "grad_norm": 1.8046618700027466, + "learning_rate": 5.675566355913384e-05, + "loss": 0.6229, + "step": 14100 + }, + { + "epoch": 0.9464111942552264, + "grad_norm": 0.9053519368171692, + "learning_rate": 5.6744894753845326e-05, + "loss": 0.4854, + "step": 14102 + }, + { + "epoch": 0.9465454179389954, + "grad_norm": 1.076835036277771, + "learning_rate": 5.67341256298738e-05, + "loss": 0.5935, + "step": 14104 + }, + { + "epoch": 0.9466796416227643, + "grad_norm": 1.1154593229293823, + "learning_rate": 5.672335618772808e-05, + "loss": 0.6073, + "step": 14106 + }, + { + "epoch": 0.9468138653065333, + "grad_norm": 0.9786189794540405, + "learning_rate": 5.671258642791699e-05, + "loss": 0.5601, + "step": 14108 + }, + { + "epoch": 0.9469480889903024, + "grad_norm": 0.944292426109314, + "learning_rate": 5.670181635094941e-05, + "loss": 0.5244, + "step": 14110 + }, + { + "epoch": 0.9470823126740714, + "grad_norm": 0.9197020530700684, + "learning_rate": 5.669104595733419e-05, + "loss": 0.5687, + "step": 14112 + }, + { + "epoch": 0.9472165363578403, + "grad_norm": 1.0978754758834839, + "learning_rate": 5.668027524758019e-05, + "loss": 0.6745, + "step": 14114 + }, + { + "epoch": 0.9473507600416093, + "grad_norm": 1.1043089628219604, + "learning_rate": 5.6669504222196327e-05, + "loss": 0.5164, + "step": 14116 + }, + { + "epoch": 0.9474849837253784, + "grad_norm": 0.9744171500205994, + "learning_rate": 5.665873288169149e-05, + "loss": 0.5805, + "step": 14118 + }, + { + "epoch": 0.9476192074091473, + "grad_norm": 0.8978694081306458, + "learning_rate": 5.664796122657463e-05, + "loss": 0.6074, + "step": 14120 + }, + { + "epoch": 0.9477534310929163, + "grad_norm": 1.1278657913208008, + "learning_rate": 5.6637189257354675e-05, + "loss": 0.6419, + "step": 14122 + }, + { + "epoch": 0.9478876547766854, + "grad_norm": 0.9175683259963989, + "learning_rate": 5.6626416974540585e-05, + "loss": 0.4927, + "step": 14124 + }, + { + "epoch": 0.9480218784604544, + "grad_norm": 1.0380014181137085, + "learning_rate": 5.6615644378641306e-05, + "loss": 0.547, + "step": 14126 + }, + { + "epoch": 0.9481561021442233, + "grad_norm": 1.1204023361206055, + "learning_rate": 5.660487147016584e-05, + "loss": 0.5597, + "step": 14128 + }, + { + "epoch": 0.9482903258279923, + "grad_norm": 1.1506484746932983, + "learning_rate": 5.659409824962321e-05, + "loss": 0.523, + "step": 14130 + }, + { + "epoch": 0.9484245495117614, + "grad_norm": 1.0465366840362549, + "learning_rate": 5.658332471752239e-05, + "loss": 0.5114, + "step": 14132 + }, + { + "epoch": 0.9485587731955304, + "grad_norm": 0.9276459813117981, + "learning_rate": 5.6572550874372424e-05, + "loss": 0.6087, + "step": 14134 + }, + { + "epoch": 0.9486929968792993, + "grad_norm": 1.038851261138916, + "learning_rate": 5.656177672068235e-05, + "loss": 0.6461, + "step": 14136 + }, + { + "epoch": 0.9488272205630683, + "grad_norm": 0.9065415263175964, + "learning_rate": 5.655100225696123e-05, + "loss": 0.5719, + "step": 14138 + }, + { + "epoch": 0.9489614442468374, + "grad_norm": 1.2550922632217407, + "learning_rate": 5.654022748371813e-05, + "loss": 0.5503, + "step": 14140 + }, + { + "epoch": 0.9490956679306064, + "grad_norm": 1.0652170181274414, + "learning_rate": 5.6529452401462145e-05, + "loss": 0.5942, + "step": 14142 + }, + { + "epoch": 0.9492298916143753, + "grad_norm": 1.002002239227295, + "learning_rate": 5.651867701070238e-05, + "loss": 0.6353, + "step": 14144 + }, + { + "epoch": 0.9493641152981444, + "grad_norm": 0.9674058556556702, + "learning_rate": 5.650790131194794e-05, + "loss": 0.5837, + "step": 14146 + }, + { + "epoch": 0.9494983389819134, + "grad_norm": 1.1481109857559204, + "learning_rate": 5.649712530570797e-05, + "loss": 0.5894, + "step": 14148 + }, + { + "epoch": 0.9496325626656824, + "grad_norm": 1.0205858945846558, + "learning_rate": 5.648634899249159e-05, + "loss": 0.5757, + "step": 14150 + }, + { + "epoch": 0.9497667863494513, + "grad_norm": 0.9188188314437866, + "learning_rate": 5.647557237280798e-05, + "loss": 0.5892, + "step": 14152 + }, + { + "epoch": 0.9499010100332204, + "grad_norm": 1.0300636291503906, + "learning_rate": 5.6464795447166306e-05, + "loss": 0.5722, + "step": 14154 + }, + { + "epoch": 0.9500352337169894, + "grad_norm": 1.0198545455932617, + "learning_rate": 5.645401821607577e-05, + "loss": 0.5481, + "step": 14156 + }, + { + "epoch": 0.9501694574007583, + "grad_norm": 1.0702418088912964, + "learning_rate": 5.644324068004556e-05, + "loss": 0.6041, + "step": 14158 + }, + { + "epoch": 0.9503036810845273, + "grad_norm": 0.919014573097229, + "learning_rate": 5.6432462839584886e-05, + "loss": 0.5284, + "step": 14160 + }, + { + "epoch": 0.9504379047682964, + "grad_norm": 1.0594698190689087, + "learning_rate": 5.6421684695203015e-05, + "loss": 0.551, + "step": 14162 + }, + { + "epoch": 0.9505721284520654, + "grad_norm": 1.2445323467254639, + "learning_rate": 5.6410906247409144e-05, + "loss": 0.5357, + "step": 14164 + }, + { + "epoch": 0.9507063521358343, + "grad_norm": 0.9506213665008545, + "learning_rate": 5.6400127496712585e-05, + "loss": 0.5968, + "step": 14166 + }, + { + "epoch": 0.9508405758196034, + "grad_norm": 1.0716971158981323, + "learning_rate": 5.6389348443622566e-05, + "loss": 0.5523, + "step": 14168 + }, + { + "epoch": 0.9509747995033724, + "grad_norm": 1.083978295326233, + "learning_rate": 5.6378569088648406e-05, + "loss": 0.5733, + "step": 14170 + }, + { + "epoch": 0.9511090231871414, + "grad_norm": 1.1101188659667969, + "learning_rate": 5.63677894322994e-05, + "loss": 0.576, + "step": 14172 + }, + { + "epoch": 0.9512432468709103, + "grad_norm": 0.9841271638870239, + "learning_rate": 5.635700947508486e-05, + "loss": 0.5641, + "step": 14174 + }, + { + "epoch": 0.9513774705546794, + "grad_norm": 0.9433690309524536, + "learning_rate": 5.6346229217514136e-05, + "loss": 0.5221, + "step": 14176 + }, + { + "epoch": 0.9515116942384484, + "grad_norm": 1.0756908655166626, + "learning_rate": 5.6335448660096544e-05, + "loss": 0.5917, + "step": 14178 + }, + { + "epoch": 0.9516459179222174, + "grad_norm": 1.024658203125, + "learning_rate": 5.632466780334148e-05, + "loss": 0.5805, + "step": 14180 + }, + { + "epoch": 0.9517801416059863, + "grad_norm": 1.002120018005371, + "learning_rate": 5.6313886647758294e-05, + "loss": 0.5101, + "step": 14182 + }, + { + "epoch": 0.9519143652897554, + "grad_norm": 1.178074598312378, + "learning_rate": 5.6303105193856365e-05, + "loss": 0.5554, + "step": 14184 + }, + { + "epoch": 0.9520485889735244, + "grad_norm": 1.1281471252441406, + "learning_rate": 5.6292323442145126e-05, + "loss": 0.5906, + "step": 14186 + }, + { + "epoch": 0.9521828126572934, + "grad_norm": 1.529188871383667, + "learning_rate": 5.6281541393133976e-05, + "loss": 0.5639, + "step": 14188 + }, + { + "epoch": 0.9523170363410624, + "grad_norm": 1.301809549331665, + "learning_rate": 5.6270759047332334e-05, + "loss": 0.6673, + "step": 14190 + }, + { + "epoch": 0.9524512600248314, + "grad_norm": 1.0367953777313232, + "learning_rate": 5.625997640524967e-05, + "loss": 0.5203, + "step": 14192 + }, + { + "epoch": 0.9525854837086004, + "grad_norm": 0.9069284200668335, + "learning_rate": 5.624919346739542e-05, + "loss": 0.6059, + "step": 14194 + }, + { + "epoch": 0.9527197073923693, + "grad_norm": 0.9075580835342407, + "learning_rate": 5.623841023427908e-05, + "loss": 0.544, + "step": 14196 + }, + { + "epoch": 0.9528539310761384, + "grad_norm": 1.0331097841262817, + "learning_rate": 5.622762670641011e-05, + "loss": 0.4943, + "step": 14198 + }, + { + "epoch": 0.9529881547599074, + "grad_norm": 0.9860792756080627, + "learning_rate": 5.6216842884298035e-05, + "loss": 0.5721, + "step": 14200 + }, + { + "epoch": 0.9531223784436764, + "grad_norm": 0.9954896569252014, + "learning_rate": 5.6206058768452355e-05, + "loss": 0.5632, + "step": 14202 + }, + { + "epoch": 0.9532566021274453, + "grad_norm": 1.2916309833526611, + "learning_rate": 5.61952743593826e-05, + "loss": 0.6002, + "step": 14204 + }, + { + "epoch": 0.9533908258112144, + "grad_norm": 1.2309260368347168, + "learning_rate": 5.618448965759832e-05, + "loss": 0.6084, + "step": 14206 + }, + { + "epoch": 0.9535250494949834, + "grad_norm": 1.2513889074325562, + "learning_rate": 5.617370466360905e-05, + "loss": 0.5453, + "step": 14208 + }, + { + "epoch": 0.9536592731787524, + "grad_norm": 1.052930474281311, + "learning_rate": 5.616291937792439e-05, + "loss": 0.5588, + "step": 14210 + }, + { + "epoch": 0.9537934968625214, + "grad_norm": 0.9992319941520691, + "learning_rate": 5.615213380105391e-05, + "loss": 0.5131, + "step": 14212 + }, + { + "epoch": 0.9539277205462904, + "grad_norm": 0.9948887825012207, + "learning_rate": 5.6141347933507204e-05, + "loss": 0.5847, + "step": 14214 + }, + { + "epoch": 0.9540619442300594, + "grad_norm": 0.9883646368980408, + "learning_rate": 5.613056177579388e-05, + "loss": 0.5812, + "step": 14216 + }, + { + "epoch": 0.9541961679138284, + "grad_norm": 0.9431707262992859, + "learning_rate": 5.611977532842355e-05, + "loss": 0.5341, + "step": 14218 + }, + { + "epoch": 0.9543303915975974, + "grad_norm": 0.9274953007698059, + "learning_rate": 5.61089885919059e-05, + "loss": 0.5018, + "step": 14220 + }, + { + "epoch": 0.9544646152813664, + "grad_norm": 0.9960140585899353, + "learning_rate": 5.609820156675053e-05, + "loss": 0.6009, + "step": 14222 + }, + { + "epoch": 0.9545988389651354, + "grad_norm": 1.1744269132614136, + "learning_rate": 5.608741425346714e-05, + "loss": 0.5912, + "step": 14224 + }, + { + "epoch": 0.9547330626489045, + "grad_norm": 1.336578369140625, + "learning_rate": 5.607662665256539e-05, + "loss": 0.5872, + "step": 14226 + }, + { + "epoch": 0.9548672863326734, + "grad_norm": 1.1415321826934814, + "learning_rate": 5.606583876455499e-05, + "loss": 0.5724, + "step": 14228 + }, + { + "epoch": 0.9550015100164424, + "grad_norm": 1.2327148914337158, + "learning_rate": 5.605505058994562e-05, + "loss": 0.5643, + "step": 14230 + }, + { + "epoch": 0.9551357337002114, + "grad_norm": 1.079979658126831, + "learning_rate": 5.604426212924703e-05, + "loss": 0.5841, + "step": 14232 + }, + { + "epoch": 0.9552699573839804, + "grad_norm": 1.0726792812347412, + "learning_rate": 5.6033473382968936e-05, + "loss": 0.6108, + "step": 14234 + }, + { + "epoch": 0.9554041810677494, + "grad_norm": 1.0051898956298828, + "learning_rate": 5.6022684351621094e-05, + "loss": 0.5602, + "step": 14236 + }, + { + "epoch": 0.9555384047515184, + "grad_norm": 0.9825963973999023, + "learning_rate": 5.601189503571326e-05, + "loss": 0.5093, + "step": 14238 + }, + { + "epoch": 0.9556726284352874, + "grad_norm": 1.103543758392334, + "learning_rate": 5.6001105435755194e-05, + "loss": 0.5362, + "step": 14240 + }, + { + "epoch": 0.9558068521190564, + "grad_norm": 0.9990962147712708, + "learning_rate": 5.599031555225671e-05, + "loss": 0.5821, + "step": 14242 + }, + { + "epoch": 0.9559410758028254, + "grad_norm": 1.0620709657669067, + "learning_rate": 5.597952538572758e-05, + "loss": 0.5948, + "step": 14244 + }, + { + "epoch": 0.9560752994865944, + "grad_norm": 0.8622420430183411, + "learning_rate": 5.596873493667765e-05, + "loss": 0.5126, + "step": 14246 + }, + { + "epoch": 0.9562095231703635, + "grad_norm": 0.9423746466636658, + "learning_rate": 5.595794420561673e-05, + "loss": 0.4744, + "step": 14248 + }, + { + "epoch": 0.9563437468541324, + "grad_norm": 0.9781889319419861, + "learning_rate": 5.5947153193054655e-05, + "loss": 0.5743, + "step": 14250 + }, + { + "epoch": 0.9564779705379014, + "grad_norm": 1.4569517374038696, + "learning_rate": 5.59363618995013e-05, + "loss": 0.5426, + "step": 14252 + }, + { + "epoch": 0.9566121942216704, + "grad_norm": 1.0532081127166748, + "learning_rate": 5.5925570325466504e-05, + "loss": 0.5646, + "step": 14254 + }, + { + "epoch": 0.9567464179054395, + "grad_norm": 1.0339252948760986, + "learning_rate": 5.591477847146016e-05, + "loss": 0.6094, + "step": 14256 + }, + { + "epoch": 0.9568806415892084, + "grad_norm": 1.1620696783065796, + "learning_rate": 5.5903986337992174e-05, + "loss": 0.5773, + "step": 14258 + }, + { + "epoch": 0.9570148652729774, + "grad_norm": 1.038197636604309, + "learning_rate": 5.589319392557244e-05, + "loss": 0.5293, + "step": 14260 + }, + { + "epoch": 0.9571490889567464, + "grad_norm": 1.1082844734191895, + "learning_rate": 5.588240123471088e-05, + "loss": 0.5785, + "step": 14262 + }, + { + "epoch": 0.9572833126405155, + "grad_norm": 1.0165228843688965, + "learning_rate": 5.587160826591743e-05, + "loss": 0.568, + "step": 14264 + }, + { + "epoch": 0.9574175363242844, + "grad_norm": 0.986341655254364, + "learning_rate": 5.586081501970203e-05, + "loss": 0.5844, + "step": 14266 + }, + { + "epoch": 0.9575517600080534, + "grad_norm": 1.1666079759597778, + "learning_rate": 5.585002149657466e-05, + "loss": 0.5321, + "step": 14268 + }, + { + "epoch": 0.9576859836918225, + "grad_norm": 1.0912145376205444, + "learning_rate": 5.583922769704526e-05, + "loss": 0.5657, + "step": 14270 + }, + { + "epoch": 0.9578202073755914, + "grad_norm": 1.0317710638046265, + "learning_rate": 5.5828433621623845e-05, + "loss": 0.5876, + "step": 14272 + }, + { + "epoch": 0.9579544310593604, + "grad_norm": 1.0638327598571777, + "learning_rate": 5.5817639270820386e-05, + "loss": 0.583, + "step": 14274 + }, + { + "epoch": 0.9580886547431294, + "grad_norm": 0.9866347908973694, + "learning_rate": 5.580684464514494e-05, + "loss": 0.5578, + "step": 14276 + }, + { + "epoch": 0.9582228784268985, + "grad_norm": 0.9564821124076843, + "learning_rate": 5.579604974510748e-05, + "loss": 0.52, + "step": 14278 + }, + { + "epoch": 0.9583571021106674, + "grad_norm": 0.9945246577262878, + "learning_rate": 5.578525457121807e-05, + "loss": 0.6157, + "step": 14280 + }, + { + "epoch": 0.9584913257944364, + "grad_norm": 1.0693161487579346, + "learning_rate": 5.5774459123986766e-05, + "loss": 0.5211, + "step": 14282 + }, + { + "epoch": 0.9586255494782054, + "grad_norm": 0.8919063806533813, + "learning_rate": 5.5763663403923614e-05, + "loss": 0.5595, + "step": 14284 + }, + { + "epoch": 0.9587597731619745, + "grad_norm": 1.1274627447128296, + "learning_rate": 5.575286741153871e-05, + "loss": 0.578, + "step": 14286 + }, + { + "epoch": 0.9588939968457434, + "grad_norm": 0.9995314478874207, + "learning_rate": 5.574207114734212e-05, + "loss": 0.5078, + "step": 14288 + }, + { + "epoch": 0.9590282205295124, + "grad_norm": 0.99176424741745, + "learning_rate": 5.573127461184398e-05, + "loss": 0.6396, + "step": 14290 + }, + { + "epoch": 0.9591624442132815, + "grad_norm": 1.0251235961914062, + "learning_rate": 5.5720477805554374e-05, + "loss": 0.5035, + "step": 14292 + }, + { + "epoch": 0.9592966678970505, + "grad_norm": 1.1442151069641113, + "learning_rate": 5.570968072898343e-05, + "loss": 0.5987, + "step": 14294 + }, + { + "epoch": 0.9594308915808194, + "grad_norm": 0.9356967806816101, + "learning_rate": 5.569888338264131e-05, + "loss": 0.6108, + "step": 14296 + }, + { + "epoch": 0.9595651152645884, + "grad_norm": 0.9598607420921326, + "learning_rate": 5.568808576703816e-05, + "loss": 0.5645, + "step": 14298 + }, + { + "epoch": 0.9596993389483575, + "grad_norm": 1.0034985542297363, + "learning_rate": 5.567728788268414e-05, + "loss": 0.5511, + "step": 14300 + }, + { + "epoch": 0.9598335626321265, + "grad_norm": 1.1254874467849731, + "learning_rate": 5.566648973008942e-05, + "loss": 0.5516, + "step": 14302 + }, + { + "epoch": 0.9599677863158954, + "grad_norm": 1.0316601991653442, + "learning_rate": 5.565569130976422e-05, + "loss": 0.5374, + "step": 14304 + }, + { + "epoch": 0.9601020099996644, + "grad_norm": 0.9164896011352539, + "learning_rate": 5.564489262221871e-05, + "loss": 0.5546, + "step": 14306 + }, + { + "epoch": 0.9602362336834335, + "grad_norm": 0.9961039423942566, + "learning_rate": 5.563409366796314e-05, + "loss": 0.5399, + "step": 14308 + }, + { + "epoch": 0.9603704573672024, + "grad_norm": 1.1713188886642456, + "learning_rate": 5.562329444750771e-05, + "loss": 0.5298, + "step": 14310 + }, + { + "epoch": 0.9605046810509714, + "grad_norm": 1.0865498781204224, + "learning_rate": 5.561249496136268e-05, + "loss": 0.6398, + "step": 14312 + }, + { + "epoch": 0.9606389047347405, + "grad_norm": 0.9350690245628357, + "learning_rate": 5.5601695210038306e-05, + "loss": 0.5508, + "step": 14314 + }, + { + "epoch": 0.9607731284185095, + "grad_norm": 1.061317801475525, + "learning_rate": 5.559089519404484e-05, + "loss": 0.5835, + "step": 14316 + }, + { + "epoch": 0.9609073521022784, + "grad_norm": 1.0611228942871094, + "learning_rate": 5.5580094913892575e-05, + "loss": 0.5825, + "step": 14318 + }, + { + "epoch": 0.9610415757860474, + "grad_norm": 1.0069859027862549, + "learning_rate": 5.556929437009181e-05, + "loss": 0.5683, + "step": 14320 + }, + { + "epoch": 0.9611757994698165, + "grad_norm": 1.1307488679885864, + "learning_rate": 5.555849356315281e-05, + "loss": 0.5182, + "step": 14322 + }, + { + "epoch": 0.9613100231535855, + "grad_norm": 1.115998387336731, + "learning_rate": 5.554769249358595e-05, + "loss": 0.5551, + "step": 14324 + }, + { + "epoch": 0.9614442468373544, + "grad_norm": 1.0153923034667969, + "learning_rate": 5.553689116190151e-05, + "loss": 0.5525, + "step": 14326 + }, + { + "epoch": 0.9615784705211234, + "grad_norm": 1.0945372581481934, + "learning_rate": 5.552608956860985e-05, + "loss": 0.5292, + "step": 14328 + }, + { + "epoch": 0.9617126942048925, + "grad_norm": 0.8611571192741394, + "learning_rate": 5.551528771422133e-05, + "loss": 0.5004, + "step": 14330 + }, + { + "epoch": 0.9618469178886615, + "grad_norm": 1.0803242921829224, + "learning_rate": 5.55044855992463e-05, + "loss": 0.5747, + "step": 14332 + }, + { + "epoch": 0.9619811415724304, + "grad_norm": 0.9125853776931763, + "learning_rate": 5.549368322419517e-05, + "loss": 0.5362, + "step": 14334 + }, + { + "epoch": 0.9621153652561995, + "grad_norm": 1.0588202476501465, + "learning_rate": 5.548288058957829e-05, + "loss": 0.5546, + "step": 14336 + }, + { + "epoch": 0.9622495889399685, + "grad_norm": 0.9635347723960876, + "learning_rate": 5.54720776959061e-05, + "loss": 0.506, + "step": 14338 + }, + { + "epoch": 0.9623838126237375, + "grad_norm": 1.0806838274002075, + "learning_rate": 5.546127454368898e-05, + "loss": 0.5233, + "step": 14340 + }, + { + "epoch": 0.9625180363075064, + "grad_norm": 1.066923975944519, + "learning_rate": 5.54504711334374e-05, + "loss": 0.5789, + "step": 14342 + }, + { + "epoch": 0.9626522599912755, + "grad_norm": 1.0439131259918213, + "learning_rate": 5.543966746566176e-05, + "loss": 0.5167, + "step": 14344 + }, + { + "epoch": 0.9627864836750445, + "grad_norm": 0.9134027361869812, + "learning_rate": 5.542886354087252e-05, + "loss": 0.5634, + "step": 14346 + }, + { + "epoch": 0.9629207073588134, + "grad_norm": 0.922843337059021, + "learning_rate": 5.5418059359580175e-05, + "loss": 0.5137, + "step": 14348 + }, + { + "epoch": 0.9630549310425824, + "grad_norm": 0.9720866680145264, + "learning_rate": 5.5407254922295174e-05, + "loss": 0.6233, + "step": 14350 + }, + { + "epoch": 0.9631891547263515, + "grad_norm": 0.9502121210098267, + "learning_rate": 5.5396450229528006e-05, + "loss": 0.5707, + "step": 14352 + }, + { + "epoch": 0.9633233784101205, + "grad_norm": 0.9958463311195374, + "learning_rate": 5.5385645281789176e-05, + "loss": 0.5451, + "step": 14354 + }, + { + "epoch": 0.9634576020938894, + "grad_norm": 0.9970313310623169, + "learning_rate": 5.537484007958921e-05, + "loss": 0.5506, + "step": 14356 + }, + { + "epoch": 0.9635918257776585, + "grad_norm": 0.9940379858016968, + "learning_rate": 5.53640346234386e-05, + "loss": 0.561, + "step": 14358 + }, + { + "epoch": 0.9637260494614275, + "grad_norm": 0.8887693285942078, + "learning_rate": 5.535322891384791e-05, + "loss": 0.5119, + "step": 14360 + }, + { + "epoch": 0.9638602731451965, + "grad_norm": 1.103327989578247, + "learning_rate": 5.534242295132769e-05, + "loss": 0.5505, + "step": 14362 + }, + { + "epoch": 0.9639944968289654, + "grad_norm": 0.986727237701416, + "learning_rate": 5.533161673638847e-05, + "loss": 0.5266, + "step": 14364 + }, + { + "epoch": 0.9641287205127345, + "grad_norm": 1.071113109588623, + "learning_rate": 5.532081026954087e-05, + "loss": 0.5083, + "step": 14366 + }, + { + "epoch": 0.9642629441965035, + "grad_norm": 1.021314263343811, + "learning_rate": 5.531000355129543e-05, + "loss": 0.5189, + "step": 14368 + }, + { + "epoch": 0.9643971678802725, + "grad_norm": 1.0355725288391113, + "learning_rate": 5.529919658216276e-05, + "loss": 0.6481, + "step": 14370 + }, + { + "epoch": 0.9645313915640414, + "grad_norm": 1.4715527296066284, + "learning_rate": 5.5288389362653484e-05, + "loss": 0.5981, + "step": 14372 + }, + { + "epoch": 0.9646656152478105, + "grad_norm": 1.6728663444519043, + "learning_rate": 5.52775818932782e-05, + "loss": 0.5561, + "step": 14374 + }, + { + "epoch": 0.9647998389315795, + "grad_norm": 0.9639118313789368, + "learning_rate": 5.5266774174547564e-05, + "loss": 0.6246, + "step": 14376 + }, + { + "epoch": 0.9649340626153485, + "grad_norm": 1.2170542478561401, + "learning_rate": 5.525596620697219e-05, + "loss": 0.529, + "step": 14378 + }, + { + "epoch": 0.9650682862991175, + "grad_norm": 1.3546918630599976, + "learning_rate": 5.5245157991062755e-05, + "loss": 0.4982, + "step": 14380 + }, + { + "epoch": 0.9652025099828865, + "grad_norm": 0.9821697473526001, + "learning_rate": 5.523434952732991e-05, + "loss": 0.5468, + "step": 14382 + }, + { + "epoch": 0.9653367336666555, + "grad_norm": 1.0652328729629517, + "learning_rate": 5.522354081628435e-05, + "loss": 0.5706, + "step": 14384 + }, + { + "epoch": 0.9654709573504244, + "grad_norm": 1.3068857192993164, + "learning_rate": 5.5212731858436774e-05, + "loss": 0.5942, + "step": 14386 + }, + { + "epoch": 0.9656051810341935, + "grad_norm": 0.922817587852478, + "learning_rate": 5.520192265429784e-05, + "loss": 0.5204, + "step": 14388 + }, + { + "epoch": 0.9657394047179625, + "grad_norm": 1.0467361211776733, + "learning_rate": 5.519111320437832e-05, + "loss": 0.6672, + "step": 14390 + }, + { + "epoch": 0.9658736284017315, + "grad_norm": 1.0627771615982056, + "learning_rate": 5.518030350918888e-05, + "loss": 0.5945, + "step": 14392 + }, + { + "epoch": 0.9660078520855004, + "grad_norm": 1.2258409261703491, + "learning_rate": 5.5169493569240295e-05, + "loss": 0.6113, + "step": 14394 + }, + { + "epoch": 0.9661420757692695, + "grad_norm": 0.9771142601966858, + "learning_rate": 5.5158683385043307e-05, + "loss": 0.629, + "step": 14396 + }, + { + "epoch": 0.9662762994530385, + "grad_norm": 1.0257612466812134, + "learning_rate": 5.514787295710867e-05, + "loss": 0.5419, + "step": 14398 + }, + { + "epoch": 0.9664105231368075, + "grad_norm": 1.0048645734786987, + "learning_rate": 5.513706228594717e-05, + "loss": 0.51, + "step": 14400 + }, + { + "epoch": 0.9665447468205765, + "grad_norm": 0.9502382278442383, + "learning_rate": 5.512625137206957e-05, + "loss": 0.5781, + "step": 14402 + }, + { + "epoch": 0.9666789705043455, + "grad_norm": 0.9471548199653625, + "learning_rate": 5.5115440215986666e-05, + "loss": 0.5382, + "step": 14404 + }, + { + "epoch": 0.9668131941881145, + "grad_norm": 0.8059168457984924, + "learning_rate": 5.510462881820928e-05, + "loss": 0.5111, + "step": 14406 + }, + { + "epoch": 0.9669474178718835, + "grad_norm": 1.1791728734970093, + "learning_rate": 5.509381717924822e-05, + "loss": 0.5111, + "step": 14408 + }, + { + "epoch": 0.9670816415556525, + "grad_norm": 1.0287694931030273, + "learning_rate": 5.508300529961431e-05, + "loss": 0.5906, + "step": 14410 + }, + { + "epoch": 0.9672158652394215, + "grad_norm": 1.289045810699463, + "learning_rate": 5.50721931798184e-05, + "loss": 0.5613, + "step": 14412 + }, + { + "epoch": 0.9673500889231905, + "grad_norm": 1.0925270318984985, + "learning_rate": 5.506138082037133e-05, + "loss": 0.6117, + "step": 14414 + }, + { + "epoch": 0.9674843126069596, + "grad_norm": 0.950714111328125, + "learning_rate": 5.505056822178397e-05, + "loss": 0.5338, + "step": 14416 + }, + { + "epoch": 0.9676185362907285, + "grad_norm": 0.9248968362808228, + "learning_rate": 5.5039755384567207e-05, + "loss": 0.6021, + "step": 14418 + }, + { + "epoch": 0.9677527599744975, + "grad_norm": 0.9010964632034302, + "learning_rate": 5.502894230923189e-05, + "loss": 0.5709, + "step": 14420 + }, + { + "epoch": 0.9678869836582665, + "grad_norm": 1.3293532133102417, + "learning_rate": 5.501812899628895e-05, + "loss": 0.5097, + "step": 14422 + }, + { + "epoch": 0.9680212073420355, + "grad_norm": 1.0516278743743896, + "learning_rate": 5.5007315446249285e-05, + "loss": 0.6134, + "step": 14424 + }, + { + "epoch": 0.9681554310258045, + "grad_norm": 0.9383218288421631, + "learning_rate": 5.49965016596238e-05, + "loss": 0.5647, + "step": 14426 + }, + { + "epoch": 0.9682896547095735, + "grad_norm": 0.9845718145370483, + "learning_rate": 5.498568763692345e-05, + "loss": 0.6191, + "step": 14428 + }, + { + "epoch": 0.9684238783933425, + "grad_norm": 0.9353554844856262, + "learning_rate": 5.497487337865916e-05, + "loss": 0.5185, + "step": 14430 + }, + { + "epoch": 0.9685581020771115, + "grad_norm": 1.1654001474380493, + "learning_rate": 5.4964058885341886e-05, + "loss": 0.5456, + "step": 14432 + }, + { + "epoch": 0.9686923257608805, + "grad_norm": 1.0446298122406006, + "learning_rate": 5.495324415748259e-05, + "loss": 0.5631, + "step": 14434 + }, + { + "epoch": 0.9688265494446495, + "grad_norm": 1.0155998468399048, + "learning_rate": 5.494242919559224e-05, + "loss": 0.5683, + "step": 14436 + }, + { + "epoch": 0.9689607731284186, + "grad_norm": 1.6033741235733032, + "learning_rate": 5.493161400018184e-05, + "loss": 0.5528, + "step": 14438 + }, + { + "epoch": 0.9690949968121875, + "grad_norm": 1.1973978281021118, + "learning_rate": 5.492079857176236e-05, + "loss": 0.5742, + "step": 14440 + }, + { + "epoch": 0.9692292204959565, + "grad_norm": 1.5843385457992554, + "learning_rate": 5.490998291084485e-05, + "loss": 0.6305, + "step": 14442 + }, + { + "epoch": 0.9693634441797255, + "grad_norm": 0.9376939535140991, + "learning_rate": 5.489916701794028e-05, + "loss": 0.5295, + "step": 14444 + }, + { + "epoch": 0.9694976678634946, + "grad_norm": 1.0814796686172485, + "learning_rate": 5.488835089355971e-05, + "loss": 0.5446, + "step": 14446 + }, + { + "epoch": 0.9696318915472635, + "grad_norm": 1.1428366899490356, + "learning_rate": 5.487753453821418e-05, + "loss": 0.6206, + "step": 14448 + }, + { + "epoch": 0.9697661152310325, + "grad_norm": 1.3953572511672974, + "learning_rate": 5.4866717952414716e-05, + "loss": 0.5119, + "step": 14450 + }, + { + "epoch": 0.9699003389148015, + "grad_norm": 0.9352095723152161, + "learning_rate": 5.485590113667242e-05, + "loss": 0.5478, + "step": 14452 + }, + { + "epoch": 0.9700345625985706, + "grad_norm": 1.2032496929168701, + "learning_rate": 5.484508409149833e-05, + "loss": 0.5643, + "step": 14454 + }, + { + "epoch": 0.9701687862823395, + "grad_norm": 1.0390560626983643, + "learning_rate": 5.483426681740356e-05, + "loss": 0.5129, + "step": 14456 + }, + { + "epoch": 0.9703030099661085, + "grad_norm": 1.0590581893920898, + "learning_rate": 5.482344931489918e-05, + "loss": 0.4824, + "step": 14458 + }, + { + "epoch": 0.9704372336498776, + "grad_norm": 1.2861895561218262, + "learning_rate": 5.48126315844963e-05, + "loss": 0.525, + "step": 14460 + }, + { + "epoch": 0.9705714573336465, + "grad_norm": 0.983880877494812, + "learning_rate": 5.480181362670605e-05, + "loss": 0.5981, + "step": 14462 + }, + { + "epoch": 0.9707056810174155, + "grad_norm": 0.8727746605873108, + "learning_rate": 5.4790995442039537e-05, + "loss": 0.5825, + "step": 14464 + }, + { + "epoch": 0.9708399047011845, + "grad_norm": 0.8624457716941833, + "learning_rate": 5.4780177031007916e-05, + "loss": 0.5023, + "step": 14466 + }, + { + "epoch": 0.9709741283849536, + "grad_norm": 1.4579488039016724, + "learning_rate": 5.4769358394122326e-05, + "loss": 0.5932, + "step": 14468 + }, + { + "epoch": 0.9711083520687225, + "grad_norm": 1.1287567615509033, + "learning_rate": 5.475853953189393e-05, + "loss": 0.5976, + "step": 14470 + }, + { + "epoch": 0.9712425757524915, + "grad_norm": 1.1077579259872437, + "learning_rate": 5.474772044483391e-05, + "loss": 0.5965, + "step": 14472 + }, + { + "epoch": 0.9713767994362605, + "grad_norm": 0.9818617701530457, + "learning_rate": 5.473690113345342e-05, + "loss": 0.5347, + "step": 14474 + }, + { + "epoch": 0.9715110231200296, + "grad_norm": 1.0572015047073364, + "learning_rate": 5.472608159826368e-05, + "loss": 0.5764, + "step": 14476 + }, + { + "epoch": 0.9716452468037985, + "grad_norm": 1.1121615171432495, + "learning_rate": 5.471526183977587e-05, + "loss": 0.586, + "step": 14478 + }, + { + "epoch": 0.9717794704875675, + "grad_norm": 0.9781137108802795, + "learning_rate": 5.470444185850121e-05, + "loss": 0.5684, + "step": 14480 + }, + { + "epoch": 0.9719136941713366, + "grad_norm": 0.9827112555503845, + "learning_rate": 5.4693621654950925e-05, + "loss": 0.6294, + "step": 14482 + }, + { + "epoch": 0.9720479178551056, + "grad_norm": 0.9477865099906921, + "learning_rate": 5.4682801229636236e-05, + "loss": 0.6169, + "step": 14484 + }, + { + "epoch": 0.9721821415388745, + "grad_norm": 1.0273088216781616, + "learning_rate": 5.467198058306842e-05, + "loss": 0.482, + "step": 14486 + }, + { + "epoch": 0.9723163652226435, + "grad_norm": 1.1097997426986694, + "learning_rate": 5.466115971575869e-05, + "loss": 0.5889, + "step": 14488 + }, + { + "epoch": 0.9724505889064126, + "grad_norm": 1.070196509361267, + "learning_rate": 5.465033862821835e-05, + "loss": 0.6092, + "step": 14490 + }, + { + "epoch": 0.9725848125901816, + "grad_norm": 1.0266796350479126, + "learning_rate": 5.463951732095862e-05, + "loss": 0.5122, + "step": 14492 + }, + { + "epoch": 0.9727190362739505, + "grad_norm": 0.8820559978485107, + "learning_rate": 5.462869579449085e-05, + "loss": 0.5148, + "step": 14494 + }, + { + "epoch": 0.9728532599577195, + "grad_norm": 0.9386069178581238, + "learning_rate": 5.461787404932629e-05, + "loss": 0.6387, + "step": 14496 + }, + { + "epoch": 0.9729874836414886, + "grad_norm": 1.0852800607681274, + "learning_rate": 5.460705208597626e-05, + "loss": 0.5331, + "step": 14498 + }, + { + "epoch": 0.9731217073252575, + "grad_norm": 1.0661139488220215, + "learning_rate": 5.459622990495209e-05, + "loss": 0.6374, + "step": 14500 + }, + { + "epoch": 0.9732559310090265, + "grad_norm": 1.011678695678711, + "learning_rate": 5.458540750676509e-05, + "loss": 0.5799, + "step": 14502 + }, + { + "epoch": 0.9733901546927956, + "grad_norm": 0.9369387626647949, + "learning_rate": 5.457458489192661e-05, + "loss": 0.551, + "step": 14504 + }, + { + "epoch": 0.9735243783765646, + "grad_norm": 1.2200850248336792, + "learning_rate": 5.4563762060947975e-05, + "loss": 0.5718, + "step": 14506 + }, + { + "epoch": 0.9736586020603335, + "grad_norm": 0.9318606853485107, + "learning_rate": 5.455293901434056e-05, + "loss": 0.6158, + "step": 14508 + }, + { + "epoch": 0.9737928257441025, + "grad_norm": 1.0478731393814087, + "learning_rate": 5.454211575261574e-05, + "loss": 0.5957, + "step": 14510 + }, + { + "epoch": 0.9739270494278716, + "grad_norm": 1.01186203956604, + "learning_rate": 5.453129227628487e-05, + "loss": 0.5676, + "step": 14512 + }, + { + "epoch": 0.9740612731116406, + "grad_norm": 0.862026572227478, + "learning_rate": 5.4520468585859364e-05, + "loss": 0.5009, + "step": 14514 + }, + { + "epoch": 0.9741954967954095, + "grad_norm": 0.9080526828765869, + "learning_rate": 5.450964468185059e-05, + "loss": 0.5186, + "step": 14516 + }, + { + "epoch": 0.9743297204791785, + "grad_norm": 0.9445207715034485, + "learning_rate": 5.4498820564769994e-05, + "loss": 0.5707, + "step": 14518 + }, + { + "epoch": 0.9744639441629476, + "grad_norm": 1.0916472673416138, + "learning_rate": 5.4487996235128946e-05, + "loss": 0.6095, + "step": 14520 + }, + { + "epoch": 0.9745981678467166, + "grad_norm": 1.0981478691101074, + "learning_rate": 5.447717169343892e-05, + "loss": 0.5759, + "step": 14522 + }, + { + "epoch": 0.9747323915304855, + "grad_norm": 0.9555610418319702, + "learning_rate": 5.4466346940211345e-05, + "loss": 0.4984, + "step": 14524 + }, + { + "epoch": 0.9748666152142546, + "grad_norm": 0.9845288991928101, + "learning_rate": 5.4455521975957635e-05, + "loss": 0.5628, + "step": 14526 + }, + { + "epoch": 0.9750008388980236, + "grad_norm": 0.9806883335113525, + "learning_rate": 5.444469680118929e-05, + "loss": 0.6304, + "step": 14528 + }, + { + "epoch": 0.9751350625817926, + "grad_norm": 1.305885910987854, + "learning_rate": 5.443387141641775e-05, + "loss": 0.494, + "step": 14530 + }, + { + "epoch": 0.9752692862655615, + "grad_norm": 1.145750641822815, + "learning_rate": 5.4423045822154506e-05, + "loss": 0.55, + "step": 14532 + }, + { + "epoch": 0.9754035099493306, + "grad_norm": 1.3389184474945068, + "learning_rate": 5.4412220018911056e-05, + "loss": 0.6095, + "step": 14534 + }, + { + "epoch": 0.9755377336330996, + "grad_norm": 1.0017625093460083, + "learning_rate": 5.4401394007198866e-05, + "loss": 0.5358, + "step": 14536 + }, + { + "epoch": 0.9756719573168685, + "grad_norm": 1.0230153799057007, + "learning_rate": 5.4390567787529476e-05, + "loss": 0.563, + "step": 14538 + }, + { + "epoch": 0.9758061810006375, + "grad_norm": 0.8892274498939514, + "learning_rate": 5.437974136041439e-05, + "loss": 0.5384, + "step": 14540 + }, + { + "epoch": 0.9759404046844066, + "grad_norm": 1.2023080587387085, + "learning_rate": 5.436891472636515e-05, + "loss": 0.6027, + "step": 14542 + }, + { + "epoch": 0.9760746283681756, + "grad_norm": 1.0622117519378662, + "learning_rate": 5.435808788589326e-05, + "loss": 0.5884, + "step": 14544 + }, + { + "epoch": 0.9762088520519445, + "grad_norm": 0.9289963841438293, + "learning_rate": 5.4347260839510296e-05, + "loss": 0.5231, + "step": 14546 + }, + { + "epoch": 0.9763430757357136, + "grad_norm": 0.9408954381942749, + "learning_rate": 5.433643358772781e-05, + "loss": 0.5371, + "step": 14548 + }, + { + "epoch": 0.9764772994194826, + "grad_norm": 1.0064287185668945, + "learning_rate": 5.432560613105736e-05, + "loss": 0.6041, + "step": 14550 + }, + { + "epoch": 0.9766115231032516, + "grad_norm": 1.038871169090271, + "learning_rate": 5.431477847001053e-05, + "loss": 0.5369, + "step": 14552 + }, + { + "epoch": 0.9767457467870205, + "grad_norm": 1.119463324546814, + "learning_rate": 5.4303950605098894e-05, + "loss": 0.5582, + "step": 14554 + }, + { + "epoch": 0.9768799704707896, + "grad_norm": 0.9053308963775635, + "learning_rate": 5.429312253683406e-05, + "loss": 0.5107, + "step": 14556 + }, + { + "epoch": 0.9770141941545586, + "grad_norm": 1.0307782888412476, + "learning_rate": 5.428229426572763e-05, + "loss": 0.5846, + "step": 14558 + }, + { + "epoch": 0.9771484178383276, + "grad_norm": 1.3395590782165527, + "learning_rate": 5.427146579229121e-05, + "loss": 0.5869, + "step": 14560 + }, + { + "epoch": 0.9772826415220965, + "grad_norm": 1.043879747390747, + "learning_rate": 5.426063711703644e-05, + "loss": 0.5227, + "step": 14562 + }, + { + "epoch": 0.9774168652058656, + "grad_norm": 1.0789127349853516, + "learning_rate": 5.424980824047493e-05, + "loss": 0.4961, + "step": 14564 + }, + { + "epoch": 0.9775510888896346, + "grad_norm": 0.843816876411438, + "learning_rate": 5.4238979163118355e-05, + "loss": 0.5498, + "step": 14566 + }, + { + "epoch": 0.9776853125734036, + "grad_norm": 1.064896821975708, + "learning_rate": 5.422814988547834e-05, + "loss": 0.5129, + "step": 14568 + }, + { + "epoch": 0.9778195362571726, + "grad_norm": 1.0035399198532104, + "learning_rate": 5.421732040806656e-05, + "loss": 0.5442, + "step": 14570 + }, + { + "epoch": 0.9779537599409416, + "grad_norm": 0.9078490734100342, + "learning_rate": 5.420649073139469e-05, + "loss": 0.5404, + "step": 14572 + }, + { + "epoch": 0.9780879836247106, + "grad_norm": 1.3465280532836914, + "learning_rate": 5.419566085597439e-05, + "loss": 0.5681, + "step": 14574 + }, + { + "epoch": 0.9782222073084795, + "grad_norm": 1.1271129846572876, + "learning_rate": 5.418483078231737e-05, + "loss": 0.5544, + "step": 14576 + }, + { + "epoch": 0.9783564309922486, + "grad_norm": 1.011531114578247, + "learning_rate": 5.417400051093533e-05, + "loss": 0.5734, + "step": 14578 + }, + { + "epoch": 0.9784906546760176, + "grad_norm": 1.0569415092468262, + "learning_rate": 5.416317004233997e-05, + "loss": 0.5964, + "step": 14580 + }, + { + "epoch": 0.9786248783597866, + "grad_norm": 1.0962992906570435, + "learning_rate": 5.4152339377043015e-05, + "loss": 0.6113, + "step": 14582 + }, + { + "epoch": 0.9787591020435555, + "grad_norm": 1.5367470979690552, + "learning_rate": 5.41415085155562e-05, + "loss": 0.57, + "step": 14584 + }, + { + "epoch": 0.9788933257273246, + "grad_norm": 0.9857184290885925, + "learning_rate": 5.413067745839123e-05, + "loss": 0.5527, + "step": 14586 + }, + { + "epoch": 0.9790275494110936, + "grad_norm": 1.040051817893982, + "learning_rate": 5.4119846206059876e-05, + "loss": 0.4739, + "step": 14588 + }, + { + "epoch": 0.9791617730948626, + "grad_norm": 1.4327423572540283, + "learning_rate": 5.410901475907391e-05, + "loss": 0.6003, + "step": 14590 + }, + { + "epoch": 0.9792959967786316, + "grad_norm": 1.0983351469039917, + "learning_rate": 5.409818311794506e-05, + "loss": 0.5514, + "step": 14592 + }, + { + "epoch": 0.9794302204624006, + "grad_norm": 0.9713186025619507, + "learning_rate": 5.4087351283185116e-05, + "loss": 0.5475, + "step": 14594 + }, + { + "epoch": 0.9795644441461696, + "grad_norm": 1.1389487981796265, + "learning_rate": 5.407651925530587e-05, + "loss": 0.5809, + "step": 14596 + }, + { + "epoch": 0.9796986678299386, + "grad_norm": 1.1033782958984375, + "learning_rate": 5.406568703481909e-05, + "loss": 0.5278, + "step": 14598 + }, + { + "epoch": 0.9798328915137076, + "grad_norm": 1.0787391662597656, + "learning_rate": 5.4054854622236614e-05, + "loss": 0.6168, + "step": 14600 + }, + { + "epoch": 0.9799671151974766, + "grad_norm": 1.0263839960098267, + "learning_rate": 5.4044022018070214e-05, + "loss": 0.6141, + "step": 14602 + }, + { + "epoch": 0.9801013388812456, + "grad_norm": 1.0619577169418335, + "learning_rate": 5.4033189222831735e-05, + "loss": 0.5731, + "step": 14604 + }, + { + "epoch": 0.9802355625650147, + "grad_norm": 0.9042320847511292, + "learning_rate": 5.402235623703299e-05, + "loss": 0.5621, + "step": 14606 + }, + { + "epoch": 0.9803697862487836, + "grad_norm": 1.096408724784851, + "learning_rate": 5.401152306118582e-05, + "loss": 0.5749, + "step": 14608 + }, + { + "epoch": 0.9805040099325526, + "grad_norm": 0.9234085083007812, + "learning_rate": 5.400068969580209e-05, + "loss": 0.5746, + "step": 14610 + }, + { + "epoch": 0.9806382336163216, + "grad_norm": 1.074016809463501, + "learning_rate": 5.398985614139361e-05, + "loss": 0.5535, + "step": 14612 + }, + { + "epoch": 0.9807724573000905, + "grad_norm": 1.464599609375, + "learning_rate": 5.3979022398472304e-05, + "loss": 0.5655, + "step": 14614 + }, + { + "epoch": 0.9809066809838596, + "grad_norm": 1.0339241027832031, + "learning_rate": 5.396818846754999e-05, + "loss": 0.5225, + "step": 14616 + }, + { + "epoch": 0.9810409046676286, + "grad_norm": 1.2816940546035767, + "learning_rate": 5.39573543491386e-05, + "loss": 0.5979, + "step": 14618 + }, + { + "epoch": 0.9811751283513976, + "grad_norm": 1.0052075386047363, + "learning_rate": 5.3946520043749974e-05, + "loss": 0.5497, + "step": 14620 + }, + { + "epoch": 0.9813093520351666, + "grad_norm": 1.0185097455978394, + "learning_rate": 5.3935685551896045e-05, + "loss": 0.5869, + "step": 14622 + }, + { + "epoch": 0.9814435757189356, + "grad_norm": 1.0448133945465088, + "learning_rate": 5.392485087408872e-05, + "loss": 0.5521, + "step": 14624 + }, + { + "epoch": 0.9815777994027046, + "grad_norm": 1.0752489566802979, + "learning_rate": 5.39140160108399e-05, + "loss": 0.5742, + "step": 14626 + }, + { + "epoch": 0.9817120230864737, + "grad_norm": 1.1034209728240967, + "learning_rate": 5.390318096266152e-05, + "loss": 0.5859, + "step": 14628 + }, + { + "epoch": 0.9818462467702426, + "grad_norm": 0.8905879855155945, + "learning_rate": 5.3892345730065506e-05, + "loss": 0.5794, + "step": 14630 + }, + { + "epoch": 0.9819804704540116, + "grad_norm": 0.9774551391601562, + "learning_rate": 5.3881510313563835e-05, + "loss": 0.5619, + "step": 14632 + }, + { + "epoch": 0.9821146941377806, + "grad_norm": 0.9077772498130798, + "learning_rate": 5.387067471366841e-05, + "loss": 0.5547, + "step": 14634 + }, + { + "epoch": 0.9822489178215497, + "grad_norm": 0.9890025854110718, + "learning_rate": 5.385983893089123e-05, + "loss": 0.5386, + "step": 14636 + }, + { + "epoch": 0.9823831415053186, + "grad_norm": 1.1164836883544922, + "learning_rate": 5.384900296574425e-05, + "loss": 0.6223, + "step": 14638 + }, + { + "epoch": 0.9825173651890876, + "grad_norm": 0.9294412732124329, + "learning_rate": 5.383816681873944e-05, + "loss": 0.5361, + "step": 14640 + }, + { + "epoch": 0.9826515888728566, + "grad_norm": 1.1060147285461426, + "learning_rate": 5.3827330490388804e-05, + "loss": 0.5258, + "step": 14642 + }, + { + "epoch": 0.9827858125566257, + "grad_norm": 1.018481969833374, + "learning_rate": 5.381649398120433e-05, + "loss": 0.533, + "step": 14644 + }, + { + "epoch": 0.9829200362403946, + "grad_norm": 1.081281304359436, + "learning_rate": 5.380565729169802e-05, + "loss": 0.5637, + "step": 14646 + }, + { + "epoch": 0.9830542599241636, + "grad_norm": 1.0529273748397827, + "learning_rate": 5.379482042238187e-05, + "loss": 0.5198, + "step": 14648 + }, + { + "epoch": 0.9831884836079327, + "grad_norm": 1.2811602354049683, + "learning_rate": 5.378398337376793e-05, + "loss": 0.6008, + "step": 14650 + }, + { + "epoch": 0.9833227072917016, + "grad_norm": 0.9522740840911865, + "learning_rate": 5.377314614636822e-05, + "loss": 0.5655, + "step": 14652 + }, + { + "epoch": 0.9834569309754706, + "grad_norm": 0.9688047170639038, + "learning_rate": 5.3762308740694765e-05, + "loss": 0.5304, + "step": 14654 + }, + { + "epoch": 0.9835911546592396, + "grad_norm": 0.9964549541473389, + "learning_rate": 5.375147115725964e-05, + "loss": 0.5505, + "step": 14656 + }, + { + "epoch": 0.9837253783430087, + "grad_norm": 1.0041929483413696, + "learning_rate": 5.374063339657486e-05, + "loss": 0.5451, + "step": 14658 + }, + { + "epoch": 0.9838596020267776, + "grad_norm": 0.9628466367721558, + "learning_rate": 5.372979545915252e-05, + "loss": 0.4683, + "step": 14660 + }, + { + "epoch": 0.9839938257105466, + "grad_norm": 0.9024425148963928, + "learning_rate": 5.371895734550467e-05, + "loss": 0.5437, + "step": 14662 + }, + { + "epoch": 0.9841280493943156, + "grad_norm": 0.9593817591667175, + "learning_rate": 5.3708119056143404e-05, + "loss": 0.6366, + "step": 14664 + }, + { + "epoch": 0.9842622730780847, + "grad_norm": 1.0373190641403198, + "learning_rate": 5.3697280591580804e-05, + "loss": 0.5662, + "step": 14666 + }, + { + "epoch": 0.9843964967618536, + "grad_norm": 1.1384583711624146, + "learning_rate": 5.368644195232896e-05, + "loss": 0.5668, + "step": 14668 + }, + { + "epoch": 0.9845307204456226, + "grad_norm": 0.8748607039451599, + "learning_rate": 5.367560313889998e-05, + "loss": 0.5586, + "step": 14670 + }, + { + "epoch": 0.9846649441293917, + "grad_norm": 0.9103529453277588, + "learning_rate": 5.366476415180599e-05, + "loss": 0.5621, + "step": 14672 + }, + { + "epoch": 0.9847991678131607, + "grad_norm": 1.006399393081665, + "learning_rate": 5.365392499155909e-05, + "loss": 0.548, + "step": 14674 + }, + { + "epoch": 0.9849333914969296, + "grad_norm": 1.0762403011322021, + "learning_rate": 5.3643085658671435e-05, + "loss": 0.5084, + "step": 14676 + }, + { + "epoch": 0.9850676151806986, + "grad_norm": 1.1388792991638184, + "learning_rate": 5.363224615365513e-05, + "loss": 0.5364, + "step": 14678 + }, + { + "epoch": 0.9852018388644677, + "grad_norm": 0.9625356793403625, + "learning_rate": 5.362140647702235e-05, + "loss": 0.5392, + "step": 14680 + }, + { + "epoch": 0.9853360625482367, + "grad_norm": 1.1637482643127441, + "learning_rate": 5.361056662928522e-05, + "loss": 0.5696, + "step": 14682 + }, + { + "epoch": 0.9854702862320056, + "grad_norm": 0.9725779294967651, + "learning_rate": 5.3599726610955926e-05, + "loss": 0.5755, + "step": 14684 + }, + { + "epoch": 0.9856045099157746, + "grad_norm": 1.0699341297149658, + "learning_rate": 5.358888642254663e-05, + "loss": 0.5043, + "step": 14686 + }, + { + "epoch": 0.9857387335995437, + "grad_norm": 1.067064881324768, + "learning_rate": 5.35780460645695e-05, + "loss": 0.4789, + "step": 14688 + }, + { + "epoch": 0.9858729572833126, + "grad_norm": 0.9795387387275696, + "learning_rate": 5.3567205537536735e-05, + "loss": 0.6033, + "step": 14690 + }, + { + "epoch": 0.9860071809670816, + "grad_norm": 1.0278023481369019, + "learning_rate": 5.3556364841960524e-05, + "loss": 0.5999, + "step": 14692 + }, + { + "epoch": 0.9861414046508507, + "grad_norm": 1.0064140558242798, + "learning_rate": 5.3545523978353073e-05, + "loss": 0.5623, + "step": 14694 + }, + { + "epoch": 0.9862756283346197, + "grad_norm": 1.255136251449585, + "learning_rate": 5.3534682947226576e-05, + "loss": 0.5479, + "step": 14696 + }, + { + "epoch": 0.9864098520183886, + "grad_norm": 0.9306426644325256, + "learning_rate": 5.352384174909325e-05, + "loss": 0.5063, + "step": 14698 + }, + { + "epoch": 0.9865440757021576, + "grad_norm": 1.0653096437454224, + "learning_rate": 5.351300038446535e-05, + "loss": 0.5808, + "step": 14700 + }, + { + "epoch": 0.9866782993859267, + "grad_norm": 1.0382238626480103, + "learning_rate": 5.350215885385509e-05, + "loss": 0.6025, + "step": 14702 + }, + { + "epoch": 0.9868125230696957, + "grad_norm": 1.0285108089447021, + "learning_rate": 5.34913171577747e-05, + "loss": 0.565, + "step": 14704 + }, + { + "epoch": 0.9869467467534646, + "grad_norm": 1.061592936515808, + "learning_rate": 5.348047529673644e-05, + "loss": 0.5551, + "step": 14706 + }, + { + "epoch": 0.9870809704372336, + "grad_norm": 1.0288335084915161, + "learning_rate": 5.346963327125258e-05, + "loss": 0.6314, + "step": 14708 + }, + { + "epoch": 0.9872151941210027, + "grad_norm": 0.976237416267395, + "learning_rate": 5.3458791081835356e-05, + "loss": 0.5668, + "step": 14710 + }, + { + "epoch": 0.9873494178047717, + "grad_norm": 1.0939053297042847, + "learning_rate": 5.3447948728997054e-05, + "loss": 0.6221, + "step": 14712 + }, + { + "epoch": 0.9874836414885406, + "grad_norm": 0.9784379005432129, + "learning_rate": 5.3437106213249964e-05, + "loss": 0.6294, + "step": 14714 + }, + { + "epoch": 0.9876178651723097, + "grad_norm": 1.2871649265289307, + "learning_rate": 5.3426263535106355e-05, + "loss": 0.6035, + "step": 14716 + }, + { + "epoch": 0.9877520888560787, + "grad_norm": 1.0066332817077637, + "learning_rate": 5.3415420695078555e-05, + "loss": 0.5382, + "step": 14718 + }, + { + "epoch": 0.9878863125398477, + "grad_norm": 0.8645907640457153, + "learning_rate": 5.3404577693678815e-05, + "loss": 0.5568, + "step": 14720 + }, + { + "epoch": 0.9880205362236166, + "grad_norm": 0.8626660108566284, + "learning_rate": 5.3393734531419496e-05, + "loss": 0.5174, + "step": 14722 + }, + { + "epoch": 0.9881547599073857, + "grad_norm": 1.0427008867263794, + "learning_rate": 5.338289120881287e-05, + "loss": 0.5293, + "step": 14724 + }, + { + "epoch": 0.9882889835911547, + "grad_norm": 1.114475131034851, + "learning_rate": 5.33720477263713e-05, + "loss": 0.6314, + "step": 14726 + }, + { + "epoch": 0.9884232072749236, + "grad_norm": 0.9445342421531677, + "learning_rate": 5.336120408460711e-05, + "loss": 0.5164, + "step": 14728 + }, + { + "epoch": 0.9885574309586926, + "grad_norm": 0.9713104963302612, + "learning_rate": 5.335036028403264e-05, + "loss": 0.4932, + "step": 14730 + }, + { + "epoch": 0.9886916546424617, + "grad_norm": 0.9602975249290466, + "learning_rate": 5.333951632516023e-05, + "loss": 0.5653, + "step": 14732 + }, + { + "epoch": 0.9888258783262307, + "grad_norm": 1.1221202611923218, + "learning_rate": 5.332867220850223e-05, + "loss": 0.5752, + "step": 14734 + }, + { + "epoch": 0.9889601020099996, + "grad_norm": 1.0036790370941162, + "learning_rate": 5.331782793457102e-05, + "loss": 0.5924, + "step": 14736 + }, + { + "epoch": 0.9890943256937687, + "grad_norm": 1.0514169931411743, + "learning_rate": 5.330698350387897e-05, + "loss": 0.4869, + "step": 14738 + }, + { + "epoch": 0.9892285493775377, + "grad_norm": 0.9604849815368652, + "learning_rate": 5.329613891693843e-05, + "loss": 0.5585, + "step": 14740 + }, + { + "epoch": 0.9893627730613067, + "grad_norm": 0.91641765832901, + "learning_rate": 5.3285294174261836e-05, + "loss": 0.5653, + "step": 14742 + }, + { + "epoch": 0.9894969967450756, + "grad_norm": 0.9412071108818054, + "learning_rate": 5.3274449276361535e-05, + "loss": 0.5276, + "step": 14744 + }, + { + "epoch": 0.9896312204288447, + "grad_norm": 1.0002440214157104, + "learning_rate": 5.3263604223749954e-05, + "loss": 0.5431, + "step": 14746 + }, + { + "epoch": 0.9897654441126137, + "grad_norm": 1.19967520236969, + "learning_rate": 5.3252759016939494e-05, + "loss": 0.5119, + "step": 14748 + }, + { + "epoch": 0.9898996677963827, + "grad_norm": 1.0427453517913818, + "learning_rate": 5.3241913656442557e-05, + "loss": 0.5676, + "step": 14750 + }, + { + "epoch": 0.9900338914801516, + "grad_norm": 0.9265989661216736, + "learning_rate": 5.3231068142771576e-05, + "loss": 0.525, + "step": 14752 + }, + { + "epoch": 0.9901681151639207, + "grad_norm": 1.0798743963241577, + "learning_rate": 5.322022247643897e-05, + "loss": 0.5615, + "step": 14754 + }, + { + "epoch": 0.9903023388476897, + "grad_norm": 1.1865473985671997, + "learning_rate": 5.320937665795721e-05, + "loss": 0.5714, + "step": 14756 + }, + { + "epoch": 0.9904365625314587, + "grad_norm": 0.9494701027870178, + "learning_rate": 5.31985306878387e-05, + "loss": 0.5346, + "step": 14758 + }, + { + "epoch": 0.9905707862152276, + "grad_norm": 1.1785624027252197, + "learning_rate": 5.31876845665959e-05, + "loss": 0.5902, + "step": 14760 + }, + { + "epoch": 0.9907050098989967, + "grad_norm": 1.032488465309143, + "learning_rate": 5.317683829474127e-05, + "loss": 0.5525, + "step": 14762 + }, + { + "epoch": 0.9908392335827657, + "grad_norm": 1.1259870529174805, + "learning_rate": 5.3165991872787276e-05, + "loss": 0.5442, + "step": 14764 + }, + { + "epoch": 0.9909734572665346, + "grad_norm": 0.9362767934799194, + "learning_rate": 5.315514530124641e-05, + "loss": 0.5118, + "step": 14766 + }, + { + "epoch": 0.9911076809503037, + "grad_norm": 0.9072252511978149, + "learning_rate": 5.3144298580631124e-05, + "loss": 0.5172, + "step": 14768 + }, + { + "epoch": 0.9912419046340727, + "grad_norm": 1.0078966617584229, + "learning_rate": 5.313345171145392e-05, + "loss": 0.5842, + "step": 14770 + }, + { + "epoch": 0.9913761283178417, + "grad_norm": 1.0020571947097778, + "learning_rate": 5.3122604694227265e-05, + "loss": 0.5355, + "step": 14772 + }, + { + "epoch": 0.9915103520016106, + "grad_norm": 1.0428295135498047, + "learning_rate": 5.311175752946369e-05, + "loss": 0.5712, + "step": 14774 + }, + { + "epoch": 0.9916445756853797, + "grad_norm": 0.929182767868042, + "learning_rate": 5.310091021767569e-05, + "loss": 0.5704, + "step": 14776 + }, + { + "epoch": 0.9917787993691487, + "grad_norm": 1.0764765739440918, + "learning_rate": 5.309006275937578e-05, + "loss": 0.5831, + "step": 14778 + }, + { + "epoch": 0.9919130230529177, + "grad_norm": 0.9372965693473816, + "learning_rate": 5.307921515507649e-05, + "loss": 0.5835, + "step": 14780 + }, + { + "epoch": 0.9920472467366866, + "grad_norm": 0.9608304500579834, + "learning_rate": 5.306836740529032e-05, + "loss": 0.5846, + "step": 14782 + }, + { + "epoch": 0.9921814704204557, + "grad_norm": 1.067068099975586, + "learning_rate": 5.305751951052984e-05, + "loss": 0.5471, + "step": 14784 + }, + { + "epoch": 0.9923156941042247, + "grad_norm": 0.9308376908302307, + "learning_rate": 5.3046671471307566e-05, + "loss": 0.5986, + "step": 14786 + }, + { + "epoch": 0.9924499177879937, + "grad_norm": 1.5615354776382446, + "learning_rate": 5.303582328813605e-05, + "loss": 0.5178, + "step": 14788 + }, + { + "epoch": 0.9925841414717627, + "grad_norm": 1.149412751197815, + "learning_rate": 5.302497496152787e-05, + "loss": 0.5592, + "step": 14790 + }, + { + "epoch": 0.9927183651555317, + "grad_norm": 1.004371166229248, + "learning_rate": 5.3014126491995566e-05, + "loss": 0.5396, + "step": 14792 + }, + { + "epoch": 0.9928525888393007, + "grad_norm": 0.9424512386322021, + "learning_rate": 5.300327788005171e-05, + "loss": 0.5271, + "step": 14794 + }, + { + "epoch": 0.9929868125230698, + "grad_norm": 0.9469589591026306, + "learning_rate": 5.2992429126208875e-05, + "loss": 0.5848, + "step": 14796 + }, + { + "epoch": 0.9931210362068387, + "grad_norm": 1.0328859090805054, + "learning_rate": 5.2981580230979664e-05, + "loss": 0.5462, + "step": 14798 + }, + { + "epoch": 0.9932552598906077, + "grad_norm": 1.0615562200546265, + "learning_rate": 5.297073119487664e-05, + "loss": 0.5663, + "step": 14800 + }, + { + "epoch": 0.9933894835743767, + "grad_norm": 0.840920627117157, + "learning_rate": 5.295988201841241e-05, + "loss": 0.5381, + "step": 14802 + }, + { + "epoch": 0.9935237072581456, + "grad_norm": 0.9971423149108887, + "learning_rate": 5.294903270209958e-05, + "loss": 0.6138, + "step": 14804 + }, + { + "epoch": 0.9936579309419147, + "grad_norm": 0.92223060131073, + "learning_rate": 5.2938183246450735e-05, + "loss": 0.5215, + "step": 14806 + }, + { + "epoch": 0.9937921546256837, + "grad_norm": 1.0310229063034058, + "learning_rate": 5.2927333651978526e-05, + "loss": 0.513, + "step": 14808 + }, + { + "epoch": 0.9939263783094527, + "grad_norm": 0.995556116104126, + "learning_rate": 5.291648391919555e-05, + "loss": 0.5637, + "step": 14810 + }, + { + "epoch": 0.9940606019932217, + "grad_norm": 1.0456836223602295, + "learning_rate": 5.290563404861445e-05, + "loss": 0.5778, + "step": 14812 + }, + { + "epoch": 0.9941948256769907, + "grad_norm": 1.004275918006897, + "learning_rate": 5.289478404074786e-05, + "loss": 0.5476, + "step": 14814 + }, + { + "epoch": 0.9943290493607597, + "grad_norm": 0.9248266220092773, + "learning_rate": 5.28839338961084e-05, + "loss": 0.4775, + "step": 14816 + }, + { + "epoch": 0.9944632730445288, + "grad_norm": 0.8902478218078613, + "learning_rate": 5.287308361520875e-05, + "loss": 0.5171, + "step": 14818 + }, + { + "epoch": 0.9945974967282977, + "grad_norm": 0.9084435105323792, + "learning_rate": 5.2862233198561525e-05, + "loss": 0.5864, + "step": 14820 + }, + { + "epoch": 0.9947317204120667, + "grad_norm": 0.9576614499092102, + "learning_rate": 5.285138264667944e-05, + "loss": 0.5086, + "step": 14822 + }, + { + "epoch": 0.9948659440958357, + "grad_norm": 0.906507134437561, + "learning_rate": 5.284053196007511e-05, + "loss": 0.5744, + "step": 14824 + }, + { + "epoch": 0.9950001677796048, + "grad_norm": 0.9690229296684265, + "learning_rate": 5.282968113926123e-05, + "loss": 0.5425, + "step": 14826 + }, + { + "epoch": 0.9951343914633737, + "grad_norm": 1.0923490524291992, + "learning_rate": 5.28188301847505e-05, + "loss": 0.5831, + "step": 14828 + }, + { + "epoch": 0.9952686151471427, + "grad_norm": 0.9221930503845215, + "learning_rate": 5.280797909705557e-05, + "loss": 0.5376, + "step": 14830 + }, + { + "epoch": 0.9954028388309117, + "grad_norm": 1.0290145874023438, + "learning_rate": 5.279712787668917e-05, + "loss": 0.5997, + "step": 14832 + }, + { + "epoch": 0.9955370625146808, + "grad_norm": 1.0450056791305542, + "learning_rate": 5.278627652416396e-05, + "loss": 0.5372, + "step": 14834 + }, + { + "epoch": 0.9956712861984497, + "grad_norm": 1.0815494060516357, + "learning_rate": 5.277542503999265e-05, + "loss": 0.6227, + "step": 14836 + }, + { + "epoch": 0.9958055098822187, + "grad_norm": 1.0937553644180298, + "learning_rate": 5.2764573424688e-05, + "loss": 0.5485, + "step": 14838 + }, + { + "epoch": 0.9959397335659878, + "grad_norm": 1.1829911470413208, + "learning_rate": 5.275372167876267e-05, + "loss": 0.5357, + "step": 14840 + }, + { + "epoch": 0.9960739572497567, + "grad_norm": 0.966904878616333, + "learning_rate": 5.274286980272941e-05, + "loss": 0.5506, + "step": 14842 + }, + { + "epoch": 0.9962081809335257, + "grad_norm": 1.0377508401870728, + "learning_rate": 5.273201779710094e-05, + "loss": 0.5892, + "step": 14844 + }, + { + "epoch": 0.9963424046172947, + "grad_norm": 1.8484457731246948, + "learning_rate": 5.2721165662390014e-05, + "loss": 0.5935, + "step": 14846 + }, + { + "epoch": 0.9964766283010638, + "grad_norm": 0.9310474991798401, + "learning_rate": 5.2710313399109346e-05, + "loss": 0.5466, + "step": 14848 + }, + { + "epoch": 0.9966108519848327, + "grad_norm": 1.152030348777771, + "learning_rate": 5.26994610077717e-05, + "loss": 0.5936, + "step": 14850 + }, + { + "epoch": 0.9967450756686017, + "grad_norm": 1.1568481922149658, + "learning_rate": 5.2688608488889844e-05, + "loss": 0.5835, + "step": 14852 + }, + { + "epoch": 0.9968792993523707, + "grad_norm": 2.007110834121704, + "learning_rate": 5.2677755842976516e-05, + "loss": 0.5801, + "step": 14854 + }, + { + "epoch": 0.9970135230361398, + "grad_norm": 1.1071622371673584, + "learning_rate": 5.266690307054449e-05, + "loss": 0.4839, + "step": 14856 + }, + { + "epoch": 0.9971477467199087, + "grad_norm": 1.1917561292648315, + "learning_rate": 5.265605017210654e-05, + "loss": 0.6797, + "step": 14858 + }, + { + "epoch": 0.9972819704036777, + "grad_norm": 0.9590668082237244, + "learning_rate": 5.264519714817544e-05, + "loss": 0.5584, + "step": 14860 + }, + { + "epoch": 0.9974161940874468, + "grad_norm": 1.018331527709961, + "learning_rate": 5.263434399926398e-05, + "loss": 0.5471, + "step": 14862 + }, + { + "epoch": 0.9975504177712158, + "grad_norm": 1.0494593381881714, + "learning_rate": 5.2623490725884936e-05, + "loss": 0.629, + "step": 14864 + }, + { + "epoch": 0.9976846414549847, + "grad_norm": 1.1049399375915527, + "learning_rate": 5.261263732855113e-05, + "loss": 0.5519, + "step": 14866 + }, + { + "epoch": 0.9978188651387537, + "grad_norm": 1.0918635129928589, + "learning_rate": 5.260178380777534e-05, + "loss": 0.5383, + "step": 14868 + }, + { + "epoch": 0.9979530888225228, + "grad_norm": 1.036096453666687, + "learning_rate": 5.2590930164070384e-05, + "loss": 0.5483, + "step": 14870 + }, + { + "epoch": 0.9980873125062918, + "grad_norm": 1.629939317703247, + "learning_rate": 5.258007639794907e-05, + "loss": 0.5799, + "step": 14872 + }, + { + "epoch": 0.9982215361900607, + "grad_norm": 0.9630006551742554, + "learning_rate": 5.256922250992421e-05, + "loss": 0.5127, + "step": 14874 + }, + { + "epoch": 0.9983557598738297, + "grad_norm": 1.0506491661071777, + "learning_rate": 5.255836850050866e-05, + "loss": 0.5702, + "step": 14876 + }, + { + "epoch": 0.9984899835575988, + "grad_norm": 1.0084110498428345, + "learning_rate": 5.2547514370215214e-05, + "loss": 0.547, + "step": 14878 + }, + { + "epoch": 0.9986242072413677, + "grad_norm": 0.9572237133979797, + "learning_rate": 5.253666011955673e-05, + "loss": 0.5034, + "step": 14880 + }, + { + "epoch": 0.9987584309251367, + "grad_norm": 0.914639949798584, + "learning_rate": 5.252580574904603e-05, + "loss": 0.5491, + "step": 14882 + }, + { + "epoch": 0.9988926546089058, + "grad_norm": 1.0369341373443604, + "learning_rate": 5.2514951259195985e-05, + "loss": 0.5681, + "step": 14884 + }, + { + "epoch": 0.9990268782926748, + "grad_norm": 1.4023351669311523, + "learning_rate": 5.250409665051944e-05, + "loss": 0.6777, + "step": 14886 + }, + { + "epoch": 0.9991611019764437, + "grad_norm": 1.0324677228927612, + "learning_rate": 5.2493241923529245e-05, + "loss": 0.607, + "step": 14888 + }, + { + "epoch": 0.9992953256602127, + "grad_norm": 0.9444645643234253, + "learning_rate": 5.248238707873828e-05, + "loss": 0.5395, + "step": 14890 + }, + { + "epoch": 0.9994295493439818, + "grad_norm": 1.0191946029663086, + "learning_rate": 5.24715321166594e-05, + "loss": 0.5444, + "step": 14892 + }, + { + "epoch": 0.9995637730277508, + "grad_norm": 1.0185309648513794, + "learning_rate": 5.246067703780551e-05, + "loss": 0.5728, + "step": 14894 + }, + { + "epoch": 0.9996979967115197, + "grad_norm": 1.0221612453460693, + "learning_rate": 5.244982184268944e-05, + "loss": 0.5277, + "step": 14896 + }, + { + "epoch": 0.9998322203952887, + "grad_norm": 0.8939405083656311, + "learning_rate": 5.243896653182412e-05, + "loss": 0.5152, + "step": 14898 + }, + { + "epoch": 0.9999664440790578, + "grad_norm": 0.9363802075386047, + "learning_rate": 5.242811110572242e-05, + "loss": 0.5334, + "step": 14900 + }, + { + "epoch": 1.0000671118418845, + "grad_norm": 0.926179051399231, + "learning_rate": 5.241725556489724e-05, + "loss": 0.3938, + "step": 14902 + }, + { + "epoch": 1.0002013355256536, + "grad_norm": 1.053067684173584, + "learning_rate": 5.2406399909861505e-05, + "loss": 0.5814, + "step": 14904 + }, + { + "epoch": 1.0003355592094225, + "grad_norm": 1.0768216848373413, + "learning_rate": 5.239554414112809e-05, + "loss": 0.539, + "step": 14906 + }, + { + "epoch": 1.0004697828931914, + "grad_norm": 0.8976664543151855, + "learning_rate": 5.238468825920994e-05, + "loss": 0.5091, + "step": 14908 + }, + { + "epoch": 1.0006040065769606, + "grad_norm": 1.118083119392395, + "learning_rate": 5.237383226461995e-05, + "loss": 0.5734, + "step": 14910 + }, + { + "epoch": 1.0007382302607295, + "grad_norm": 0.9272267818450928, + "learning_rate": 5.236297615787106e-05, + "loss": 0.5075, + "step": 14912 + }, + { + "epoch": 1.0008724539444984, + "grad_norm": 0.9571478366851807, + "learning_rate": 5.235211993947617e-05, + "loss": 0.5533, + "step": 14914 + }, + { + "epoch": 1.0010066776282676, + "grad_norm": 1.372754693031311, + "learning_rate": 5.2341263609948254e-05, + "loss": 0.5162, + "step": 14916 + }, + { + "epoch": 1.0011409013120365, + "grad_norm": 0.9124377965927124, + "learning_rate": 5.2330407169800245e-05, + "loss": 0.4917, + "step": 14918 + }, + { + "epoch": 1.0012751249958054, + "grad_norm": 1.0203857421875, + "learning_rate": 5.2319550619545074e-05, + "loss": 0.5309, + "step": 14920 + }, + { + "epoch": 1.0014093486795745, + "grad_norm": 1.0425175428390503, + "learning_rate": 5.23086939596957e-05, + "loss": 0.539, + "step": 14922 + }, + { + "epoch": 1.0015435723633435, + "grad_norm": 0.9315015077590942, + "learning_rate": 5.229783719076506e-05, + "loss": 0.5044, + "step": 14924 + }, + { + "epoch": 1.0016777960471126, + "grad_norm": 1.1324846744537354, + "learning_rate": 5.2286980313266145e-05, + "loss": 0.5346, + "step": 14926 + }, + { + "epoch": 1.0018120197308815, + "grad_norm": 1.1176323890686035, + "learning_rate": 5.227612332771191e-05, + "loss": 0.5452, + "step": 14928 + }, + { + "epoch": 1.0019462434146504, + "grad_norm": 0.9105541706085205, + "learning_rate": 5.2265266234615315e-05, + "loss": 0.5263, + "step": 14930 + }, + { + "epoch": 1.0020804670984196, + "grad_norm": 0.9921098351478577, + "learning_rate": 5.225440903448936e-05, + "loss": 0.4861, + "step": 14932 + }, + { + "epoch": 1.0022146907821885, + "grad_norm": 0.9832185506820679, + "learning_rate": 5.2243551727847e-05, + "loss": 0.5168, + "step": 14934 + }, + { + "epoch": 1.0023489144659574, + "grad_norm": 0.9794383645057678, + "learning_rate": 5.223269431520124e-05, + "loss": 0.5213, + "step": 14936 + }, + { + "epoch": 1.0024831381497266, + "grad_norm": 1.0107759237289429, + "learning_rate": 5.222183679706506e-05, + "loss": 0.5449, + "step": 14938 + }, + { + "epoch": 1.0026173618334955, + "grad_norm": 0.9340489506721497, + "learning_rate": 5.221097917395146e-05, + "loss": 0.5418, + "step": 14940 + }, + { + "epoch": 1.0027515855172646, + "grad_norm": 1.0198230743408203, + "learning_rate": 5.220012144637344e-05, + "loss": 0.4777, + "step": 14942 + }, + { + "epoch": 1.0028858092010335, + "grad_norm": 1.0641562938690186, + "learning_rate": 5.2189263614843995e-05, + "loss": 0.592, + "step": 14944 + }, + { + "epoch": 1.0030200328848025, + "grad_norm": 1.0694611072540283, + "learning_rate": 5.2178405679876174e-05, + "loss": 0.4984, + "step": 14946 + }, + { + "epoch": 1.0031542565685716, + "grad_norm": 1.0537960529327393, + "learning_rate": 5.216754764198295e-05, + "loss": 0.541, + "step": 14948 + }, + { + "epoch": 1.0032884802523405, + "grad_norm": 1.0442579984664917, + "learning_rate": 5.215668950167737e-05, + "loss": 0.5575, + "step": 14950 + }, + { + "epoch": 1.0034227039361094, + "grad_norm": 0.9863391518592834, + "learning_rate": 5.214583125947245e-05, + "loss": 0.5007, + "step": 14952 + }, + { + "epoch": 1.0035569276198786, + "grad_norm": 0.8672400712966919, + "learning_rate": 5.213497291588122e-05, + "loss": 0.5325, + "step": 14954 + }, + { + "epoch": 1.0036911513036475, + "grad_norm": 1.105807900428772, + "learning_rate": 5.212411447141673e-05, + "loss": 0.5258, + "step": 14956 + }, + { + "epoch": 1.0038253749874164, + "grad_norm": 1.0071481466293335, + "learning_rate": 5.2113255926592e-05, + "loss": 0.5268, + "step": 14958 + }, + { + "epoch": 1.0039595986711856, + "grad_norm": 1.0454081296920776, + "learning_rate": 5.210239728192008e-05, + "loss": 0.5541, + "step": 14960 + }, + { + "epoch": 1.0040938223549545, + "grad_norm": 0.9871719479560852, + "learning_rate": 5.209153853791402e-05, + "loss": 0.5713, + "step": 14962 + }, + { + "epoch": 1.0042280460387236, + "grad_norm": 1.0889171361923218, + "learning_rate": 5.208067969508688e-05, + "loss": 0.5697, + "step": 14964 + }, + { + "epoch": 1.0043622697224925, + "grad_norm": 1.007730484008789, + "learning_rate": 5.2069820753951726e-05, + "loss": 0.5457, + "step": 14966 + }, + { + "epoch": 1.0044964934062615, + "grad_norm": 1.1622997522354126, + "learning_rate": 5.205896171502159e-05, + "loss": 0.6189, + "step": 14968 + }, + { + "epoch": 1.0046307170900306, + "grad_norm": 1.2704561948776245, + "learning_rate": 5.204810257880959e-05, + "loss": 0.477, + "step": 14970 + }, + { + "epoch": 1.0047649407737995, + "grad_norm": 1.12416672706604, + "learning_rate": 5.203724334582875e-05, + "loss": 0.5957, + "step": 14972 + }, + { + "epoch": 1.0048991644575684, + "grad_norm": 1.3205918073654175, + "learning_rate": 5.202638401659218e-05, + "loss": 0.5413, + "step": 14974 + }, + { + "epoch": 1.0050333881413376, + "grad_norm": 1.1126925945281982, + "learning_rate": 5.201552459161293e-05, + "loss": 0.5216, + "step": 14976 + }, + { + "epoch": 1.0051676118251065, + "grad_norm": 1.7548534870147705, + "learning_rate": 5.200466507140411e-05, + "loss": 0.5176, + "step": 14978 + }, + { + "epoch": 1.0053018355088756, + "grad_norm": 1.009682536125183, + "learning_rate": 5.1993805456478826e-05, + "loss": 0.5166, + "step": 14980 + }, + { + "epoch": 1.0054360591926446, + "grad_norm": 1.3446884155273438, + "learning_rate": 5.198294574735014e-05, + "loss": 0.5367, + "step": 14982 + }, + { + "epoch": 1.0055702828764135, + "grad_norm": 1.5239311456680298, + "learning_rate": 5.197208594453118e-05, + "loss": 0.5185, + "step": 14984 + }, + { + "epoch": 1.0057045065601826, + "grad_norm": 1.0466574430465698, + "learning_rate": 5.1961226048535026e-05, + "loss": 0.5689, + "step": 14986 + }, + { + "epoch": 1.0058387302439515, + "grad_norm": 0.9727661609649658, + "learning_rate": 5.195036605987481e-05, + "loss": 0.5662, + "step": 14988 + }, + { + "epoch": 1.0059729539277205, + "grad_norm": 1.0696989297866821, + "learning_rate": 5.193950597906362e-05, + "loss": 0.6016, + "step": 14990 + }, + { + "epoch": 1.0061071776114896, + "grad_norm": 0.9576994776725769, + "learning_rate": 5.192864580661459e-05, + "loss": 0.4973, + "step": 14992 + }, + { + "epoch": 1.0062414012952585, + "grad_norm": 1.0753369331359863, + "learning_rate": 5.1917785543040844e-05, + "loss": 0.5364, + "step": 14994 + }, + { + "epoch": 1.0063756249790274, + "grad_norm": 1.042697548866272, + "learning_rate": 5.1906925188855494e-05, + "loss": 0.5297, + "step": 14996 + }, + { + "epoch": 1.0065098486627966, + "grad_norm": 0.9751782417297363, + "learning_rate": 5.189606474457168e-05, + "loss": 0.546, + "step": 14998 + }, + { + "epoch": 1.0066440723465655, + "grad_norm": 0.9660375714302063, + "learning_rate": 5.188520421070253e-05, + "loss": 0.5445, + "step": 15000 + }, + { + "epoch": 1.0067782960303346, + "grad_norm": 1.259419560432434, + "learning_rate": 5.187434358776121e-05, + "loss": 0.5468, + "step": 15002 + }, + { + "epoch": 1.0069125197141036, + "grad_norm": 1.0952138900756836, + "learning_rate": 5.186348287626083e-05, + "loss": 0.5825, + "step": 15004 + }, + { + "epoch": 1.0070467433978725, + "grad_norm": 1.0939810276031494, + "learning_rate": 5.185262207671454e-05, + "loss": 0.5217, + "step": 15006 + }, + { + "epoch": 1.0071809670816416, + "grad_norm": 1.4713177680969238, + "learning_rate": 5.184176118963551e-05, + "loss": 0.5227, + "step": 15008 + }, + { + "epoch": 1.0073151907654105, + "grad_norm": 0.9648658633232117, + "learning_rate": 5.183090021553687e-05, + "loss": 0.4656, + "step": 15010 + }, + { + "epoch": 1.0074494144491795, + "grad_norm": 1.2975503206253052, + "learning_rate": 5.182003915493181e-05, + "loss": 0.4875, + "step": 15012 + }, + { + "epoch": 1.0075836381329486, + "grad_norm": 1.0438868999481201, + "learning_rate": 5.180917800833347e-05, + "loss": 0.5622, + "step": 15014 + }, + { + "epoch": 1.0077178618167175, + "grad_norm": 0.9114359021186829, + "learning_rate": 5.179831677625503e-05, + "loss": 0.5518, + "step": 15016 + }, + { + "epoch": 1.0078520855004867, + "grad_norm": 0.9837536215782166, + "learning_rate": 5.1787455459209665e-05, + "loss": 0.5371, + "step": 15018 + }, + { + "epoch": 1.0079863091842556, + "grad_norm": 1.137822151184082, + "learning_rate": 5.1776594057710535e-05, + "loss": 0.5564, + "step": 15020 + }, + { + "epoch": 1.0081205328680245, + "grad_norm": 1.026039958000183, + "learning_rate": 5.176573257227084e-05, + "loss": 0.493, + "step": 15022 + }, + { + "epoch": 1.0082547565517936, + "grad_norm": 0.9626202583312988, + "learning_rate": 5.175487100340374e-05, + "loss": 0.4941, + "step": 15024 + }, + { + "epoch": 1.0083889802355626, + "grad_norm": 0.8714377284049988, + "learning_rate": 5.1744009351622434e-05, + "loss": 0.4707, + "step": 15026 + }, + { + "epoch": 1.0085232039193315, + "grad_norm": 0.9464874267578125, + "learning_rate": 5.173314761744014e-05, + "loss": 0.4727, + "step": 15028 + }, + { + "epoch": 1.0086574276031006, + "grad_norm": 0.9182068705558777, + "learning_rate": 5.172228580137001e-05, + "loss": 0.5544, + "step": 15030 + }, + { + "epoch": 1.0087916512868695, + "grad_norm": 0.984942615032196, + "learning_rate": 5.171142390392526e-05, + "loss": 0.5588, + "step": 15032 + }, + { + "epoch": 1.0089258749706387, + "grad_norm": 0.9684733748435974, + "learning_rate": 5.1700561925619096e-05, + "loss": 0.5343, + "step": 15034 + }, + { + "epoch": 1.0090600986544076, + "grad_norm": 0.8601505160331726, + "learning_rate": 5.1689699866964735e-05, + "loss": 0.5005, + "step": 15036 + }, + { + "epoch": 1.0091943223381765, + "grad_norm": 1.0310051441192627, + "learning_rate": 5.167883772847537e-05, + "loss": 0.4845, + "step": 15038 + }, + { + "epoch": 1.0093285460219457, + "grad_norm": 1.2362431287765503, + "learning_rate": 5.166797551066422e-05, + "loss": 0.5716, + "step": 15040 + }, + { + "epoch": 1.0094627697057146, + "grad_norm": 0.9385756850242615, + "learning_rate": 5.165711321404453e-05, + "loss": 0.5015, + "step": 15042 + }, + { + "epoch": 1.0095969933894835, + "grad_norm": 0.9549235701560974, + "learning_rate": 5.1646250839129476e-05, + "loss": 0.5317, + "step": 15044 + }, + { + "epoch": 1.0097312170732526, + "grad_norm": 1.0057629346847534, + "learning_rate": 5.163538838643232e-05, + "loss": 0.511, + "step": 15046 + }, + { + "epoch": 1.0098654407570216, + "grad_norm": 1.0021158456802368, + "learning_rate": 5.162452585646627e-05, + "loss": 0.5398, + "step": 15048 + }, + { + "epoch": 1.0099996644407905, + "grad_norm": 1.0679807662963867, + "learning_rate": 5.161366324974457e-05, + "loss": 0.5315, + "step": 15050 + }, + { + "epoch": 1.0101338881245596, + "grad_norm": 0.9807422161102295, + "learning_rate": 5.160280056678045e-05, + "loss": 0.5698, + "step": 15052 + }, + { + "epoch": 1.0102681118083285, + "grad_norm": 0.9693793058395386, + "learning_rate": 5.159193780808716e-05, + "loss": 0.4875, + "step": 15054 + }, + { + "epoch": 1.0104023354920977, + "grad_norm": 0.9912816286087036, + "learning_rate": 5.158107497417795e-05, + "loss": 0.5324, + "step": 15056 + }, + { + "epoch": 1.0105365591758666, + "grad_norm": 0.9343096017837524, + "learning_rate": 5.157021206556604e-05, + "loss": 0.5687, + "step": 15058 + }, + { + "epoch": 1.0106707828596355, + "grad_norm": 0.8785936832427979, + "learning_rate": 5.155934908276471e-05, + "loss": 0.4924, + "step": 15060 + }, + { + "epoch": 1.0108050065434047, + "grad_norm": 1.003708004951477, + "learning_rate": 5.154848602628719e-05, + "loss": 0.5324, + "step": 15062 + }, + { + "epoch": 1.0109392302271736, + "grad_norm": 0.9341418147087097, + "learning_rate": 5.153762289664676e-05, + "loss": 0.5417, + "step": 15064 + }, + { + "epoch": 1.0110734539109425, + "grad_norm": 0.9692851901054382, + "learning_rate": 5.1526759694356665e-05, + "loss": 0.5643, + "step": 15066 + }, + { + "epoch": 1.0112076775947116, + "grad_norm": 0.9379444122314453, + "learning_rate": 5.151589641993018e-05, + "loss": 0.55, + "step": 15068 + }, + { + "epoch": 1.0113419012784806, + "grad_norm": 1.1084202527999878, + "learning_rate": 5.150503307388057e-05, + "loss": 0.5434, + "step": 15070 + }, + { + "epoch": 1.0114761249622495, + "grad_norm": 0.9925508499145508, + "learning_rate": 5.1494169656721104e-05, + "loss": 0.5256, + "step": 15072 + }, + { + "epoch": 1.0116103486460186, + "grad_norm": 1.0956413745880127, + "learning_rate": 5.148330616896506e-05, + "loss": 0.5348, + "step": 15074 + }, + { + "epoch": 1.0117445723297875, + "grad_norm": 1.0557767152786255, + "learning_rate": 5.147244261112573e-05, + "loss": 0.5522, + "step": 15076 + }, + { + "epoch": 1.0118787960135567, + "grad_norm": 1.0028269290924072, + "learning_rate": 5.1461578983716365e-05, + "loss": 0.4744, + "step": 15078 + }, + { + "epoch": 1.0120130196973256, + "grad_norm": 1.0762214660644531, + "learning_rate": 5.145071528725027e-05, + "loss": 0.5087, + "step": 15080 + }, + { + "epoch": 1.0121472433810945, + "grad_norm": 0.9658424258232117, + "learning_rate": 5.143985152224073e-05, + "loss": 0.5149, + "step": 15082 + }, + { + "epoch": 1.0122814670648637, + "grad_norm": 1.0201389789581299, + "learning_rate": 5.1428987689201046e-05, + "loss": 0.5152, + "step": 15084 + }, + { + "epoch": 1.0124156907486326, + "grad_norm": 1.068712830543518, + "learning_rate": 5.1418123788644504e-05, + "loss": 0.5835, + "step": 15086 + }, + { + "epoch": 1.0125499144324015, + "grad_norm": 1.046278715133667, + "learning_rate": 5.1407259821084384e-05, + "loss": 0.5451, + "step": 15088 + }, + { + "epoch": 1.0126841381161706, + "grad_norm": 1.3016585111618042, + "learning_rate": 5.139639578703403e-05, + "loss": 0.5872, + "step": 15090 + }, + { + "epoch": 1.0128183617999396, + "grad_norm": 1.1036945581436157, + "learning_rate": 5.1385531687006695e-05, + "loss": 0.5226, + "step": 15092 + }, + { + "epoch": 1.0129525854837087, + "grad_norm": 1.1816788911819458, + "learning_rate": 5.137466752151573e-05, + "loss": 0.5442, + "step": 15094 + }, + { + "epoch": 1.0130868091674776, + "grad_norm": 1.0297253131866455, + "learning_rate": 5.136380329107442e-05, + "loss": 0.5106, + "step": 15096 + }, + { + "epoch": 1.0132210328512465, + "grad_norm": 1.035953164100647, + "learning_rate": 5.1352938996196084e-05, + "loss": 0.5429, + "step": 15098 + }, + { + "epoch": 1.0133552565350157, + "grad_norm": 0.8848439455032349, + "learning_rate": 5.134207463739404e-05, + "loss": 0.5334, + "step": 15100 + }, + { + "epoch": 1.0134894802187846, + "grad_norm": 1.233180046081543, + "learning_rate": 5.1331210215181613e-05, + "loss": 0.5403, + "step": 15102 + }, + { + "epoch": 1.0136237039025535, + "grad_norm": 0.961600124835968, + "learning_rate": 5.132034573007213e-05, + "loss": 0.4906, + "step": 15104 + }, + { + "epoch": 1.0137579275863227, + "grad_norm": 0.9835965037345886, + "learning_rate": 5.130948118257889e-05, + "loss": 0.478, + "step": 15106 + }, + { + "epoch": 1.0138921512700916, + "grad_norm": 1.0651036500930786, + "learning_rate": 5.129861657321525e-05, + "loss": 0.553, + "step": 15108 + }, + { + "epoch": 1.0140263749538605, + "grad_norm": 1.226927399635315, + "learning_rate": 5.128775190249453e-05, + "loss": 0.5391, + "step": 15110 + }, + { + "epoch": 1.0141605986376296, + "grad_norm": 1.0278208255767822, + "learning_rate": 5.1276887170930064e-05, + "loss": 0.5731, + "step": 15112 + }, + { + "epoch": 1.0142948223213986, + "grad_norm": 1.1334031820297241, + "learning_rate": 5.1266022379035196e-05, + "loss": 0.5337, + "step": 15114 + }, + { + "epoch": 1.0144290460051677, + "grad_norm": 1.5725743770599365, + "learning_rate": 5.125515752732324e-05, + "loss": 0.522, + "step": 15116 + }, + { + "epoch": 1.0145632696889366, + "grad_norm": 1.0932778120040894, + "learning_rate": 5.1244292616307574e-05, + "loss": 0.5713, + "step": 15118 + }, + { + "epoch": 1.0146974933727055, + "grad_norm": 0.8721126317977905, + "learning_rate": 5.123342764650153e-05, + "loss": 0.4632, + "step": 15120 + }, + { + "epoch": 1.0148317170564747, + "grad_norm": 0.9035164713859558, + "learning_rate": 5.122256261841846e-05, + "loss": 0.4953, + "step": 15122 + }, + { + "epoch": 1.0149659407402436, + "grad_norm": 1.0134105682373047, + "learning_rate": 5.12116975325717e-05, + "loss": 0.5629, + "step": 15124 + }, + { + "epoch": 1.0151001644240125, + "grad_norm": 0.9002230763435364, + "learning_rate": 5.120083238947462e-05, + "loss": 0.497, + "step": 15126 + }, + { + "epoch": 1.0152343881077817, + "grad_norm": 1.014272928237915, + "learning_rate": 5.1189967189640566e-05, + "loss": 0.4951, + "step": 15128 + }, + { + "epoch": 1.0153686117915506, + "grad_norm": 1.0982106924057007, + "learning_rate": 5.11791019335829e-05, + "loss": 0.5601, + "step": 15130 + }, + { + "epoch": 1.0155028354753197, + "grad_norm": 1.020804762840271, + "learning_rate": 5.116823662181499e-05, + "loss": 0.5075, + "step": 15132 + }, + { + "epoch": 1.0156370591590886, + "grad_norm": 0.7800991535186768, + "learning_rate": 5.11573712548502e-05, + "loss": 0.4879, + "step": 15134 + }, + { + "epoch": 1.0157712828428576, + "grad_norm": 1.0576050281524658, + "learning_rate": 5.11465058332019e-05, + "loss": 0.5168, + "step": 15136 + }, + { + "epoch": 1.0159055065266267, + "grad_norm": 0.9641649723052979, + "learning_rate": 5.113564035738343e-05, + "loss": 0.5333, + "step": 15138 + }, + { + "epoch": 1.0160397302103956, + "grad_norm": 1.0422450304031372, + "learning_rate": 5.1124774827908215e-05, + "loss": 0.4981, + "step": 15140 + }, + { + "epoch": 1.0161739538941645, + "grad_norm": 0.9134299755096436, + "learning_rate": 5.111390924528959e-05, + "loss": 0.5266, + "step": 15142 + }, + { + "epoch": 1.0163081775779337, + "grad_norm": 0.8555437326431274, + "learning_rate": 5.1103043610040934e-05, + "loss": 0.5251, + "step": 15144 + }, + { + "epoch": 1.0164424012617026, + "grad_norm": 0.8712595701217651, + "learning_rate": 5.1092177922675655e-05, + "loss": 0.5532, + "step": 15146 + }, + { + "epoch": 1.0165766249454715, + "grad_norm": 0.9501728415489197, + "learning_rate": 5.1081312183707095e-05, + "loss": 0.5305, + "step": 15148 + }, + { + "epoch": 1.0167108486292407, + "grad_norm": 1.0660793781280518, + "learning_rate": 5.107044639364867e-05, + "loss": 0.5201, + "step": 15150 + }, + { + "epoch": 1.0168450723130096, + "grad_norm": 1.0098044872283936, + "learning_rate": 5.1059580553013764e-05, + "loss": 0.516, + "step": 15152 + }, + { + "epoch": 1.0169792959967787, + "grad_norm": 1.2454719543457031, + "learning_rate": 5.104871466231575e-05, + "loss": 0.511, + "step": 15154 + }, + { + "epoch": 1.0171135196805476, + "grad_norm": 0.9541716575622559, + "learning_rate": 5.103784872206805e-05, + "loss": 0.5125, + "step": 15156 + }, + { + "epoch": 1.0172477433643166, + "grad_norm": 1.0367459058761597, + "learning_rate": 5.1026982732784014e-05, + "loss": 0.5279, + "step": 15158 + }, + { + "epoch": 1.0173819670480857, + "grad_norm": 1.1213152408599854, + "learning_rate": 5.101611669497708e-05, + "loss": 0.6091, + "step": 15160 + }, + { + "epoch": 1.0175161907318546, + "grad_norm": 0.9246627688407898, + "learning_rate": 5.1005250609160625e-05, + "loss": 0.4449, + "step": 15162 + }, + { + "epoch": 1.0176504144156235, + "grad_norm": 0.9593459963798523, + "learning_rate": 5.099438447584804e-05, + "loss": 0.5165, + "step": 15164 + }, + { + "epoch": 1.0177846380993927, + "grad_norm": 1.5162698030471802, + "learning_rate": 5.098351829555277e-05, + "loss": 0.5046, + "step": 15166 + }, + { + "epoch": 1.0179188617831616, + "grad_norm": 0.9656062126159668, + "learning_rate": 5.097265206878817e-05, + "loss": 0.5115, + "step": 15168 + }, + { + "epoch": 1.0180530854669307, + "grad_norm": 1.1321715116500854, + "learning_rate": 5.0961785796067696e-05, + "loss": 0.5514, + "step": 15170 + }, + { + "epoch": 1.0181873091506997, + "grad_norm": 0.9948189854621887, + "learning_rate": 5.095091947790472e-05, + "loss": 0.552, + "step": 15172 + }, + { + "epoch": 1.0183215328344686, + "grad_norm": 1.1939868927001953, + "learning_rate": 5.094005311481268e-05, + "loss": 0.6173, + "step": 15174 + }, + { + "epoch": 1.0184557565182377, + "grad_norm": 1.0231444835662842, + "learning_rate": 5.0929186707304964e-05, + "loss": 0.5605, + "step": 15176 + }, + { + "epoch": 1.0185899802020066, + "grad_norm": 0.9155842065811157, + "learning_rate": 5.0918320255895014e-05, + "loss": 0.5419, + "step": 15178 + }, + { + "epoch": 1.0187242038857756, + "grad_norm": 1.0757856369018555, + "learning_rate": 5.0907453761096235e-05, + "loss": 0.6102, + "step": 15180 + }, + { + "epoch": 1.0188584275695447, + "grad_norm": 0.9137722253799438, + "learning_rate": 5.0896587223422044e-05, + "loss": 0.538, + "step": 15182 + }, + { + "epoch": 1.0189926512533136, + "grad_norm": 0.923180341720581, + "learning_rate": 5.088572064338588e-05, + "loss": 0.5489, + "step": 15184 + }, + { + "epoch": 1.0191268749370828, + "grad_norm": 1.0486664772033691, + "learning_rate": 5.087485402150115e-05, + "loss": 0.5017, + "step": 15186 + }, + { + "epoch": 1.0192610986208517, + "grad_norm": 0.9409403204917908, + "learning_rate": 5.086398735828129e-05, + "loss": 0.5356, + "step": 15188 + }, + { + "epoch": 1.0193953223046206, + "grad_norm": 1.0220710039138794, + "learning_rate": 5.085312065423973e-05, + "loss": 0.5132, + "step": 15190 + }, + { + "epoch": 1.0195295459883897, + "grad_norm": 0.9534757733345032, + "learning_rate": 5.084225390988989e-05, + "loss": 0.479, + "step": 15192 + }, + { + "epoch": 1.0196637696721587, + "grad_norm": 1.086157202720642, + "learning_rate": 5.083138712574522e-05, + "loss": 0.4957, + "step": 15194 + }, + { + "epoch": 1.0197979933559276, + "grad_norm": 1.0429636240005493, + "learning_rate": 5.082052030231913e-05, + "loss": 0.5382, + "step": 15196 + }, + { + "epoch": 1.0199322170396967, + "grad_norm": 1.0217652320861816, + "learning_rate": 5.080965344012508e-05, + "loss": 0.5078, + "step": 15198 + }, + { + "epoch": 1.0200664407234656, + "grad_norm": 1.0648549795150757, + "learning_rate": 5.079878653967649e-05, + "loss": 0.5477, + "step": 15200 + }, + { + "epoch": 1.0202006644072346, + "grad_norm": 1.0137075185775757, + "learning_rate": 5.078791960148681e-05, + "loss": 0.5905, + "step": 15202 + }, + { + "epoch": 1.0203348880910037, + "grad_norm": 0.9725910425186157, + "learning_rate": 5.0777052626069476e-05, + "loss": 0.4865, + "step": 15204 + }, + { + "epoch": 1.0204691117747726, + "grad_norm": 0.9284847974777222, + "learning_rate": 5.076618561393793e-05, + "loss": 0.5589, + "step": 15206 + }, + { + "epoch": 1.0206033354585418, + "grad_norm": 0.9524595737457275, + "learning_rate": 5.075531856560563e-05, + "loss": 0.584, + "step": 15208 + }, + { + "epoch": 1.0207375591423107, + "grad_norm": 1.1114261150360107, + "learning_rate": 5.074445148158601e-05, + "loss": 0.5155, + "step": 15210 + }, + { + "epoch": 1.0208717828260796, + "grad_norm": 0.8127722144126892, + "learning_rate": 5.073358436239252e-05, + "loss": 0.5009, + "step": 15212 + }, + { + "epoch": 1.0210060065098487, + "grad_norm": 1.8480273485183716, + "learning_rate": 5.072271720853862e-05, + "loss": 0.5675, + "step": 15214 + }, + { + "epoch": 1.0211402301936177, + "grad_norm": 1.0490953922271729, + "learning_rate": 5.071185002053773e-05, + "loss": 0.5311, + "step": 15216 + }, + { + "epoch": 1.0212744538773866, + "grad_norm": 0.9619256258010864, + "learning_rate": 5.0700982798903355e-05, + "loss": 0.5182, + "step": 15218 + }, + { + "epoch": 1.0214086775611557, + "grad_norm": 1.031275987625122, + "learning_rate": 5.06901155441489e-05, + "loss": 0.5755, + "step": 15220 + }, + { + "epoch": 1.0215429012449246, + "grad_norm": 0.938673734664917, + "learning_rate": 5.067924825678785e-05, + "loss": 0.4703, + "step": 15222 + }, + { + "epoch": 1.0216771249286936, + "grad_norm": 1.615932822227478, + "learning_rate": 5.066838093733365e-05, + "loss": 0.5154, + "step": 15224 + }, + { + "epoch": 1.0218113486124627, + "grad_norm": 0.8763120770454407, + "learning_rate": 5.065751358629976e-05, + "loss": 0.4773, + "step": 15226 + }, + { + "epoch": 1.0219455722962316, + "grad_norm": 1.0280568599700928, + "learning_rate": 5.0646646204199663e-05, + "loss": 0.4761, + "step": 15228 + }, + { + "epoch": 1.0220797959800008, + "grad_norm": 0.9118609428405762, + "learning_rate": 5.0635778791546794e-05, + "loss": 0.4487, + "step": 15230 + }, + { + "epoch": 1.0222140196637697, + "grad_norm": 1.0934200286865234, + "learning_rate": 5.062491134885463e-05, + "loss": 0.5314, + "step": 15232 + }, + { + "epoch": 1.0223482433475386, + "grad_norm": 0.9376586079597473, + "learning_rate": 5.061404387663663e-05, + "loss": 0.4847, + "step": 15234 + }, + { + "epoch": 1.0224824670313077, + "grad_norm": 1.2409019470214844, + "learning_rate": 5.060317637540628e-05, + "loss": 0.5537, + "step": 15236 + }, + { + "epoch": 1.0226166907150767, + "grad_norm": 1.0247794389724731, + "learning_rate": 5.059230884567702e-05, + "loss": 0.51, + "step": 15238 + }, + { + "epoch": 1.0227509143988456, + "grad_norm": 0.9596056938171387, + "learning_rate": 5.058144128796232e-05, + "loss": 0.5794, + "step": 15240 + }, + { + "epoch": 1.0228851380826147, + "grad_norm": 1.0091639757156372, + "learning_rate": 5.057057370277568e-05, + "loss": 0.5205, + "step": 15242 + }, + { + "epoch": 1.0230193617663836, + "grad_norm": 1.0142924785614014, + "learning_rate": 5.0559706090630546e-05, + "loss": 0.5319, + "step": 15244 + }, + { + "epoch": 1.0231535854501528, + "grad_norm": 1.8155488967895508, + "learning_rate": 5.0548838452040407e-05, + "loss": 0.5499, + "step": 15246 + }, + { + "epoch": 1.0232878091339217, + "grad_norm": 1.1085665225982666, + "learning_rate": 5.053797078751872e-05, + "loss": 0.5811, + "step": 15248 + }, + { + "epoch": 1.0234220328176906, + "grad_norm": 1.546950101852417, + "learning_rate": 5.052710309757899e-05, + "loss": 0.5264, + "step": 15250 + }, + { + "epoch": 1.0235562565014598, + "grad_norm": 1.0630773305892944, + "learning_rate": 5.0516235382734654e-05, + "loss": 0.5004, + "step": 15252 + }, + { + "epoch": 1.0236904801852287, + "grad_norm": 1.086484432220459, + "learning_rate": 5.050536764349921e-05, + "loss": 0.5448, + "step": 15254 + }, + { + "epoch": 1.0238247038689976, + "grad_norm": 0.962790846824646, + "learning_rate": 5.049449988038616e-05, + "loss": 0.4521, + "step": 15256 + }, + { + "epoch": 1.0239589275527667, + "grad_norm": 1.0335437059402466, + "learning_rate": 5.048363209390895e-05, + "loss": 0.5506, + "step": 15258 + }, + { + "epoch": 1.0240931512365357, + "grad_norm": 0.9776301980018616, + "learning_rate": 5.047276428458108e-05, + "loss": 0.5172, + "step": 15260 + }, + { + "epoch": 1.0242273749203046, + "grad_norm": 1.1775176525115967, + "learning_rate": 5.046189645291601e-05, + "loss": 0.6449, + "step": 15262 + }, + { + "epoch": 1.0243615986040737, + "grad_norm": 1.103438138961792, + "learning_rate": 5.0451028599427265e-05, + "loss": 0.5563, + "step": 15264 + }, + { + "epoch": 1.0244958222878426, + "grad_norm": 3.5545434951782227, + "learning_rate": 5.044016072462829e-05, + "loss": 0.565, + "step": 15266 + }, + { + "epoch": 1.0246300459716118, + "grad_norm": 0.9224105477333069, + "learning_rate": 5.0429292829032596e-05, + "loss": 0.5518, + "step": 15268 + }, + { + "epoch": 1.0247642696553807, + "grad_norm": 0.9298921823501587, + "learning_rate": 5.041842491315365e-05, + "loss": 0.5029, + "step": 15270 + }, + { + "epoch": 1.0248984933391496, + "grad_norm": 1.0152496099472046, + "learning_rate": 5.040755697750496e-05, + "loss": 0.5202, + "step": 15272 + }, + { + "epoch": 1.0250327170229188, + "grad_norm": 1.5045030117034912, + "learning_rate": 5.03966890226e-05, + "loss": 0.5188, + "step": 15274 + }, + { + "epoch": 1.0251669407066877, + "grad_norm": 1.1592066287994385, + "learning_rate": 5.0385821048952255e-05, + "loss": 0.5209, + "step": 15276 + }, + { + "epoch": 1.0253011643904566, + "grad_norm": 0.8823243379592896, + "learning_rate": 5.037495305707524e-05, + "loss": 0.5019, + "step": 15278 + }, + { + "epoch": 1.0254353880742257, + "grad_norm": 0.9502339363098145, + "learning_rate": 5.0364085047482424e-05, + "loss": 0.5374, + "step": 15280 + }, + { + "epoch": 1.0255696117579947, + "grad_norm": 1.0712730884552002, + "learning_rate": 5.035321702068731e-05, + "loss": 0.5501, + "step": 15282 + }, + { + "epoch": 1.0257038354417638, + "grad_norm": 0.9488974213600159, + "learning_rate": 5.034234897720338e-05, + "loss": 0.5536, + "step": 15284 + }, + { + "epoch": 1.0258380591255327, + "grad_norm": 1.146383285522461, + "learning_rate": 5.0331480917544126e-05, + "loss": 0.5045, + "step": 15286 + }, + { + "epoch": 1.0259722828093016, + "grad_norm": 1.014808177947998, + "learning_rate": 5.032061284222307e-05, + "loss": 0.5006, + "step": 15288 + }, + { + "epoch": 1.0261065064930708, + "grad_norm": 1.062474012374878, + "learning_rate": 5.0309744751753664e-05, + "loss": 0.537, + "step": 15290 + }, + { + "epoch": 1.0262407301768397, + "grad_norm": 1.1122550964355469, + "learning_rate": 5.029887664664944e-05, + "loss": 0.5578, + "step": 15292 + }, + { + "epoch": 1.0263749538606086, + "grad_norm": 0.9269875288009644, + "learning_rate": 5.028800852742388e-05, + "loss": 0.4753, + "step": 15294 + }, + { + "epoch": 1.0265091775443778, + "grad_norm": 1.066814661026001, + "learning_rate": 5.0277140394590484e-05, + "loss": 0.5387, + "step": 15296 + }, + { + "epoch": 1.0266434012281467, + "grad_norm": 1.0602961778640747, + "learning_rate": 5.026627224866276e-05, + "loss": 0.5202, + "step": 15298 + }, + { + "epoch": 1.0267776249119156, + "grad_norm": 1.077146053314209, + "learning_rate": 5.025540409015418e-05, + "loss": 0.5219, + "step": 15300 + }, + { + "epoch": 1.0269118485956847, + "grad_norm": 1.0115227699279785, + "learning_rate": 5.024453591957826e-05, + "loss": 0.6292, + "step": 15302 + }, + { + "epoch": 1.0270460722794537, + "grad_norm": 1.0494619607925415, + "learning_rate": 5.023366773744851e-05, + "loss": 0.539, + "step": 15304 + }, + { + "epoch": 1.0271802959632228, + "grad_norm": 1.0594550371170044, + "learning_rate": 5.022279954427841e-05, + "loss": 0.5376, + "step": 15306 + }, + { + "epoch": 1.0273145196469917, + "grad_norm": 0.9713077545166016, + "learning_rate": 5.0211931340581476e-05, + "loss": 0.5806, + "step": 15308 + }, + { + "epoch": 1.0274487433307606, + "grad_norm": 0.9390246272087097, + "learning_rate": 5.020106312687119e-05, + "loss": 0.5121, + "step": 15310 + }, + { + "epoch": 1.0275829670145298, + "grad_norm": 1.2264317274093628, + "learning_rate": 5.0190194903661094e-05, + "loss": 0.5358, + "step": 15312 + }, + { + "epoch": 1.0277171906982987, + "grad_norm": 2.0339279174804688, + "learning_rate": 5.0179326671464634e-05, + "loss": 0.5839, + "step": 15314 + }, + { + "epoch": 1.0278514143820676, + "grad_norm": 0.9309272766113281, + "learning_rate": 5.016845843079535e-05, + "loss": 0.5548, + "step": 15316 + }, + { + "epoch": 1.0279856380658368, + "grad_norm": 0.9557270407676697, + "learning_rate": 5.015759018216675e-05, + "loss": 0.5128, + "step": 15318 + }, + { + "epoch": 1.0281198617496057, + "grad_norm": 1.0359398126602173, + "learning_rate": 5.01467219260923e-05, + "loss": 0.5008, + "step": 15320 + }, + { + "epoch": 1.0282540854333748, + "grad_norm": 0.9772992134094238, + "learning_rate": 5.0135853663085554e-05, + "loss": 0.4431, + "step": 15322 + }, + { + "epoch": 1.0283883091171437, + "grad_norm": 1.0367960929870605, + "learning_rate": 5.012498539365996e-05, + "loss": 0.498, + "step": 15324 + }, + { + "epoch": 1.0285225328009127, + "grad_norm": 0.9311704635620117, + "learning_rate": 5.0114117118329075e-05, + "loss": 0.5364, + "step": 15326 + }, + { + "epoch": 1.0286567564846818, + "grad_norm": 1.004698634147644, + "learning_rate": 5.0103248837606375e-05, + "loss": 0.523, + "step": 15328 + }, + { + "epoch": 1.0287909801684507, + "grad_norm": 0.9973604083061218, + "learning_rate": 5.0092380552005355e-05, + "loss": 0.5805, + "step": 15330 + }, + { + "epoch": 1.0289252038522196, + "grad_norm": 1.0765000581741333, + "learning_rate": 5.008151226203957e-05, + "loss": 0.584, + "step": 15332 + }, + { + "epoch": 1.0290594275359888, + "grad_norm": 0.9316189289093018, + "learning_rate": 5.007064396822245e-05, + "loss": 0.5076, + "step": 15334 + }, + { + "epoch": 1.0291936512197577, + "grad_norm": 0.9343737959861755, + "learning_rate": 5.0059775671067575e-05, + "loss": 0.5298, + "step": 15336 + }, + { + "epoch": 1.0293278749035266, + "grad_norm": 0.8994764685630798, + "learning_rate": 5.0048907371088394e-05, + "loss": 0.5493, + "step": 15338 + }, + { + "epoch": 1.0294620985872958, + "grad_norm": 0.9307758212089539, + "learning_rate": 5.0038039068798446e-05, + "loss": 0.5207, + "step": 15340 + }, + { + "epoch": 1.0295963222710647, + "grad_norm": 1.009665846824646, + "learning_rate": 5.002717076471123e-05, + "loss": 0.5141, + "step": 15342 + }, + { + "epoch": 1.0297305459548338, + "grad_norm": 1.037687063217163, + "learning_rate": 5.001630245934024e-05, + "loss": 0.4918, + "step": 15344 + }, + { + "epoch": 1.0298647696386027, + "grad_norm": 0.9489524364471436, + "learning_rate": 5.000543415319901e-05, + "loss": 0.5337, + "step": 15346 + }, + { + "epoch": 1.0299989933223717, + "grad_norm": 0.9218322038650513, + "learning_rate": 4.999456584680101e-05, + "loss": 0.441, + "step": 15348 + }, + { + "epoch": 1.0301332170061408, + "grad_norm": 1.0558072328567505, + "learning_rate": 4.998369754065976e-05, + "loss": 0.5311, + "step": 15350 + }, + { + "epoch": 1.0302674406899097, + "grad_norm": 0.8926640152931213, + "learning_rate": 4.997282923528878e-05, + "loss": 0.487, + "step": 15352 + }, + { + "epoch": 1.0304016643736786, + "grad_norm": 0.9766799807548523, + "learning_rate": 4.996196093120156e-05, + "loss": 0.4965, + "step": 15354 + }, + { + "epoch": 1.0305358880574478, + "grad_norm": 0.9748830199241638, + "learning_rate": 4.995109262891162e-05, + "loss": 0.498, + "step": 15356 + }, + { + "epoch": 1.0306701117412167, + "grad_norm": 1.2828388214111328, + "learning_rate": 4.994022432893243e-05, + "loss": 0.5085, + "step": 15358 + }, + { + "epoch": 1.0308043354249858, + "grad_norm": 1.0764567852020264, + "learning_rate": 4.9929356031777555e-05, + "loss": 0.4781, + "step": 15360 + }, + { + "epoch": 1.0309385591087548, + "grad_norm": 1.0225658416748047, + "learning_rate": 4.991848773796044e-05, + "loss": 0.5247, + "step": 15362 + }, + { + "epoch": 1.0310727827925237, + "grad_norm": 1.0597457885742188, + "learning_rate": 4.990761944799465e-05, + "loss": 0.5318, + "step": 15364 + }, + { + "epoch": 1.0312070064762928, + "grad_norm": 0.9888409972190857, + "learning_rate": 4.989675116239363e-05, + "loss": 0.5106, + "step": 15366 + }, + { + "epoch": 1.0313412301600617, + "grad_norm": 1.002458095550537, + "learning_rate": 4.988588288167093e-05, + "loss": 0.5645, + "step": 15368 + }, + { + "epoch": 1.0314754538438307, + "grad_norm": 1.1779125928878784, + "learning_rate": 4.9875014606340035e-05, + "loss": 0.6012, + "step": 15370 + }, + { + "epoch": 1.0316096775275998, + "grad_norm": 1.0960805416107178, + "learning_rate": 4.9864146336914465e-05, + "loss": 0.5152, + "step": 15372 + }, + { + "epoch": 1.0317439012113687, + "grad_norm": 1.0984114408493042, + "learning_rate": 4.98532780739077e-05, + "loss": 0.5857, + "step": 15374 + }, + { + "epoch": 1.0318781248951376, + "grad_norm": 1.0565415620803833, + "learning_rate": 4.984240981783327e-05, + "loss": 0.5191, + "step": 15376 + }, + { + "epoch": 1.0320123485789068, + "grad_norm": 0.964955747127533, + "learning_rate": 4.9831541569204654e-05, + "loss": 0.5259, + "step": 15378 + }, + { + "epoch": 1.0321465722626757, + "grad_norm": 0.8490574359893799, + "learning_rate": 4.982067332853538e-05, + "loss": 0.5113, + "step": 15380 + }, + { + "epoch": 1.0322807959464448, + "grad_norm": 1.0339528322219849, + "learning_rate": 4.980980509633894e-05, + "loss": 0.5321, + "step": 15382 + }, + { + "epoch": 1.0324150196302138, + "grad_norm": 1.0278650522232056, + "learning_rate": 4.979893687312881e-05, + "loss": 0.5367, + "step": 15384 + }, + { + "epoch": 1.0325492433139827, + "grad_norm": 0.9023627638816833, + "learning_rate": 4.978806865941855e-05, + "loss": 0.5279, + "step": 15386 + }, + { + "epoch": 1.0326834669977518, + "grad_norm": 1.0649601221084595, + "learning_rate": 4.97772004557216e-05, + "loss": 0.5181, + "step": 15388 + }, + { + "epoch": 1.0328176906815207, + "grad_norm": 0.95375657081604, + "learning_rate": 4.9766332262551515e-05, + "loss": 0.5993, + "step": 15390 + }, + { + "epoch": 1.0329519143652897, + "grad_norm": 1.07741379737854, + "learning_rate": 4.975546408042175e-05, + "loss": 0.5013, + "step": 15392 + }, + { + "epoch": 1.0330861380490588, + "grad_norm": 1.0350271463394165, + "learning_rate": 4.9744595909845845e-05, + "loss": 0.5081, + "step": 15394 + }, + { + "epoch": 1.0332203617328277, + "grad_norm": 0.9855200052261353, + "learning_rate": 4.973372775133725e-05, + "loss": 0.6056, + "step": 15396 + }, + { + "epoch": 1.0333545854165969, + "grad_norm": 1.0609577894210815, + "learning_rate": 4.972285960540954e-05, + "loss": 0.49, + "step": 15398 + }, + { + "epoch": 1.0334888091003658, + "grad_norm": 0.8649885058403015, + "learning_rate": 4.9711991472576124e-05, + "loss": 0.4648, + "step": 15400 + }, + { + "epoch": 1.0336230327841347, + "grad_norm": 1.088198184967041, + "learning_rate": 4.9701123353350575e-05, + "loss": 0.5857, + "step": 15402 + }, + { + "epoch": 1.0337572564679038, + "grad_norm": 0.9336490631103516, + "learning_rate": 4.969025524824634e-05, + "loss": 0.5333, + "step": 15404 + }, + { + "epoch": 1.0338914801516728, + "grad_norm": 0.9601561427116394, + "learning_rate": 4.967938715777696e-05, + "loss": 0.5048, + "step": 15406 + }, + { + "epoch": 1.0340257038354417, + "grad_norm": 1.0431386232376099, + "learning_rate": 4.9668519082455885e-05, + "loss": 0.4623, + "step": 15408 + }, + { + "epoch": 1.0341599275192108, + "grad_norm": 1.0056498050689697, + "learning_rate": 4.965765102279664e-05, + "loss": 0.5427, + "step": 15410 + }, + { + "epoch": 1.0342941512029797, + "grad_norm": 1.0134047269821167, + "learning_rate": 4.964678297931271e-05, + "loss": 0.5011, + "step": 15412 + }, + { + "epoch": 1.0344283748867489, + "grad_norm": 1.0183725357055664, + "learning_rate": 4.9635914952517594e-05, + "loss": 0.5092, + "step": 15414 + }, + { + "epoch": 1.0345625985705178, + "grad_norm": 1.4094773530960083, + "learning_rate": 4.9625046942924774e-05, + "loss": 0.6091, + "step": 15416 + }, + { + "epoch": 1.0346968222542867, + "grad_norm": 0.8718349933624268, + "learning_rate": 4.9614178951047756e-05, + "loss": 0.5464, + "step": 15418 + }, + { + "epoch": 1.0348310459380559, + "grad_norm": 0.9513201117515564, + "learning_rate": 4.9603310977400006e-05, + "loss": 0.4983, + "step": 15420 + }, + { + "epoch": 1.0349652696218248, + "grad_norm": 1.1781314611434937, + "learning_rate": 4.959244302249506e-05, + "loss": 0.5312, + "step": 15422 + }, + { + "epoch": 1.0350994933055937, + "grad_norm": 1.0829561948776245, + "learning_rate": 4.958157508684635e-05, + "loss": 0.4857, + "step": 15424 + }, + { + "epoch": 1.0352337169893628, + "grad_norm": 1.045479416847229, + "learning_rate": 4.957070717096742e-05, + "loss": 0.4792, + "step": 15426 + }, + { + "epoch": 1.0353679406731318, + "grad_norm": 0.9355472326278687, + "learning_rate": 4.955983927537171e-05, + "loss": 0.514, + "step": 15428 + }, + { + "epoch": 1.0355021643569007, + "grad_norm": 1.0759092569351196, + "learning_rate": 4.9548971400572747e-05, + "loss": 0.5504, + "step": 15430 + }, + { + "epoch": 1.0356363880406698, + "grad_norm": 0.847141683101654, + "learning_rate": 4.9538103547083985e-05, + "loss": 0.5492, + "step": 15432 + }, + { + "epoch": 1.0357706117244387, + "grad_norm": 0.9926348328590393, + "learning_rate": 4.952723571541893e-05, + "loss": 0.5228, + "step": 15434 + }, + { + "epoch": 1.0359048354082079, + "grad_norm": 1.4465930461883545, + "learning_rate": 4.9516367906091055e-05, + "loss": 0.5384, + "step": 15436 + }, + { + "epoch": 1.0360390590919768, + "grad_norm": 0.8559713959693909, + "learning_rate": 4.950550011961385e-05, + "loss": 0.5147, + "step": 15438 + }, + { + "epoch": 1.0361732827757457, + "grad_norm": 1.0673739910125732, + "learning_rate": 4.9494632356500785e-05, + "loss": 0.5402, + "step": 15440 + }, + { + "epoch": 1.0363075064595149, + "grad_norm": 0.8725389838218689, + "learning_rate": 4.948376461726536e-05, + "loss": 0.5465, + "step": 15442 + }, + { + "epoch": 1.0364417301432838, + "grad_norm": 1.2083971500396729, + "learning_rate": 4.947289690242102e-05, + "loss": 0.565, + "step": 15444 + }, + { + "epoch": 1.0365759538270527, + "grad_norm": 1.0119904279708862, + "learning_rate": 4.9462029212481286e-05, + "loss": 0.5557, + "step": 15446 + }, + { + "epoch": 1.0367101775108218, + "grad_norm": 1.090747594833374, + "learning_rate": 4.945116154795959e-05, + "loss": 0.5061, + "step": 15448 + }, + { + "epoch": 1.0368444011945908, + "grad_norm": 0.9607612490653992, + "learning_rate": 4.9440293909369465e-05, + "loss": 0.477, + "step": 15450 + }, + { + "epoch": 1.0369786248783597, + "grad_norm": 0.9650788307189941, + "learning_rate": 4.942942629722434e-05, + "loss": 0.4996, + "step": 15452 + }, + { + "epoch": 1.0371128485621288, + "grad_norm": 0.9474275708198547, + "learning_rate": 4.941855871203769e-05, + "loss": 0.5445, + "step": 15454 + }, + { + "epoch": 1.0372470722458977, + "grad_norm": 1.009928584098816, + "learning_rate": 4.940769115432301e-05, + "loss": 0.5052, + "step": 15456 + }, + { + "epoch": 1.0373812959296669, + "grad_norm": 1.0166889429092407, + "learning_rate": 4.939682362459374e-05, + "loss": 0.5236, + "step": 15458 + }, + { + "epoch": 1.0375155196134358, + "grad_norm": 1.0336209535598755, + "learning_rate": 4.938595612336339e-05, + "loss": 0.5955, + "step": 15460 + }, + { + "epoch": 1.0376497432972047, + "grad_norm": 0.9604952931404114, + "learning_rate": 4.937508865114537e-05, + "loss": 0.4671, + "step": 15462 + }, + { + "epoch": 1.0377839669809739, + "grad_norm": 0.8933176398277283, + "learning_rate": 4.936422120845322e-05, + "loss": 0.5717, + "step": 15464 + }, + { + "epoch": 1.0379181906647428, + "grad_norm": 1.7485871315002441, + "learning_rate": 4.935335379580034e-05, + "loss": 0.4836, + "step": 15466 + }, + { + "epoch": 1.0380524143485117, + "grad_norm": 0.983931303024292, + "learning_rate": 4.934248641370025e-05, + "loss": 0.5091, + "step": 15468 + }, + { + "epoch": 1.0381866380322808, + "grad_norm": 0.856342077255249, + "learning_rate": 4.933161906266636e-05, + "loss": 0.4919, + "step": 15470 + }, + { + "epoch": 1.0383208617160498, + "grad_norm": 0.995201051235199, + "learning_rate": 4.9320751743212176e-05, + "loss": 0.5338, + "step": 15472 + }, + { + "epoch": 1.038455085399819, + "grad_norm": 0.9791795015335083, + "learning_rate": 4.930988445585112e-05, + "loss": 0.5643, + "step": 15474 + }, + { + "epoch": 1.0385893090835878, + "grad_norm": 0.9542690515518188, + "learning_rate": 4.929901720109668e-05, + "loss": 0.5248, + "step": 15476 + }, + { + "epoch": 1.0387235327673567, + "grad_norm": 0.95274817943573, + "learning_rate": 4.928814997946228e-05, + "loss": 0.5349, + "step": 15478 + }, + { + "epoch": 1.0388577564511259, + "grad_norm": 1.0679212808609009, + "learning_rate": 4.927728279146141e-05, + "loss": 0.5314, + "step": 15480 + }, + { + "epoch": 1.0389919801348948, + "grad_norm": 1.0120261907577515, + "learning_rate": 4.9266415637607494e-05, + "loss": 0.5584, + "step": 15482 + }, + { + "epoch": 1.0391262038186637, + "grad_norm": 0.9586861729621887, + "learning_rate": 4.925554851841401e-05, + "loss": 0.5243, + "step": 15484 + }, + { + "epoch": 1.0392604275024329, + "grad_norm": 0.850306510925293, + "learning_rate": 4.924468143439437e-05, + "loss": 0.4676, + "step": 15486 + }, + { + "epoch": 1.0393946511862018, + "grad_norm": 0.9537341594696045, + "learning_rate": 4.923381438606208e-05, + "loss": 0.4883, + "step": 15488 + }, + { + "epoch": 1.039528874869971, + "grad_norm": 0.8852509260177612, + "learning_rate": 4.922294737393053e-05, + "loss": 0.532, + "step": 15490 + }, + { + "epoch": 1.0396630985537398, + "grad_norm": 0.929013192653656, + "learning_rate": 4.92120803985132e-05, + "loss": 0.5263, + "step": 15492 + }, + { + "epoch": 1.0397973222375088, + "grad_norm": 1.0598328113555908, + "learning_rate": 4.920121346032351e-05, + "loss": 0.5332, + "step": 15494 + }, + { + "epoch": 1.039931545921278, + "grad_norm": 1.0350052118301392, + "learning_rate": 4.919034655987493e-05, + "loss": 0.494, + "step": 15496 + }, + { + "epoch": 1.0400657696050468, + "grad_norm": 1.0005286931991577, + "learning_rate": 4.917947969768087e-05, + "loss": 0.5654, + "step": 15498 + }, + { + "epoch": 1.0401999932888157, + "grad_norm": 1.1021864414215088, + "learning_rate": 4.916861287425479e-05, + "loss": 0.4904, + "step": 15500 + }, + { + "epoch": 1.0403342169725849, + "grad_norm": 4.215618133544922, + "learning_rate": 4.915774609011011e-05, + "loss": 0.5055, + "step": 15502 + }, + { + "epoch": 1.0404684406563538, + "grad_norm": 1.0950462818145752, + "learning_rate": 4.914687934576028e-05, + "loss": 0.6111, + "step": 15504 + }, + { + "epoch": 1.0406026643401227, + "grad_norm": 2.1147496700286865, + "learning_rate": 4.913601264171871e-05, + "loss": 0.5393, + "step": 15506 + }, + { + "epoch": 1.0407368880238919, + "grad_norm": 1.1324349641799927, + "learning_rate": 4.912514597849886e-05, + "loss": 0.5148, + "step": 15508 + }, + { + "epoch": 1.0408711117076608, + "grad_norm": 1.0787025690078735, + "learning_rate": 4.911427935661411e-05, + "loss": 0.4627, + "step": 15510 + }, + { + "epoch": 1.04100533539143, + "grad_norm": 1.0233891010284424, + "learning_rate": 4.910341277657796e-05, + "loss": 0.5605, + "step": 15512 + }, + { + "epoch": 1.0411395590751988, + "grad_norm": 1.246164321899414, + "learning_rate": 4.909254623890376e-05, + "loss": 0.5491, + "step": 15514 + }, + { + "epoch": 1.0412737827589678, + "grad_norm": 0.9711535573005676, + "learning_rate": 4.9081679744105e-05, + "loss": 0.5522, + "step": 15516 + }, + { + "epoch": 1.041408006442737, + "grad_norm": 0.989448606967926, + "learning_rate": 4.907081329269503e-05, + "loss": 0.4817, + "step": 15518 + }, + { + "epoch": 1.0415422301265058, + "grad_norm": 1.0142803192138672, + "learning_rate": 4.905994688518733e-05, + "loss": 0.5904, + "step": 15520 + }, + { + "epoch": 1.0416764538102747, + "grad_norm": 0.8768779039382935, + "learning_rate": 4.9049080522095296e-05, + "loss": 0.5051, + "step": 15522 + }, + { + "epoch": 1.0418106774940439, + "grad_norm": 0.9926018118858337, + "learning_rate": 4.903821420393231e-05, + "loss": 0.4733, + "step": 15524 + }, + { + "epoch": 1.0419449011778128, + "grad_norm": 0.9845841526985168, + "learning_rate": 4.902734793121183e-05, + "loss": 0.5254, + "step": 15526 + }, + { + "epoch": 1.0420791248615817, + "grad_norm": 0.9354475736618042, + "learning_rate": 4.9016481704447245e-05, + "loss": 0.4868, + "step": 15528 + }, + { + "epoch": 1.0422133485453509, + "grad_norm": 1.0436205863952637, + "learning_rate": 4.900561552415197e-05, + "loss": 0.5086, + "step": 15530 + }, + { + "epoch": 1.0423475722291198, + "grad_norm": 1.009522557258606, + "learning_rate": 4.899474939083939e-05, + "loss": 0.528, + "step": 15532 + }, + { + "epoch": 1.042481795912889, + "grad_norm": 1.0397756099700928, + "learning_rate": 4.8983883305022945e-05, + "loss": 0.6065, + "step": 15534 + }, + { + "epoch": 1.0426160195966578, + "grad_norm": 1.0646541118621826, + "learning_rate": 4.8973017267216e-05, + "loss": 0.5222, + "step": 15536 + }, + { + "epoch": 1.0427502432804268, + "grad_norm": 1.0897085666656494, + "learning_rate": 4.896215127793198e-05, + "loss": 0.4996, + "step": 15538 + }, + { + "epoch": 1.042884466964196, + "grad_norm": 0.9040811657905579, + "learning_rate": 4.895128533768426e-05, + "loss": 0.5078, + "step": 15540 + }, + { + "epoch": 1.0430186906479648, + "grad_norm": 1.010197639465332, + "learning_rate": 4.894041944698626e-05, + "loss": 0.5786, + "step": 15542 + }, + { + "epoch": 1.0431529143317337, + "grad_norm": 1.1387351751327515, + "learning_rate": 4.892955360635134e-05, + "loss": 0.4843, + "step": 15544 + }, + { + "epoch": 1.0432871380155029, + "grad_norm": 1.2616219520568848, + "learning_rate": 4.891868781629292e-05, + "loss": 0.5211, + "step": 15546 + }, + { + "epoch": 1.0434213616992718, + "grad_norm": 0.9757934808731079, + "learning_rate": 4.890782207732436e-05, + "loss": 0.5223, + "step": 15548 + }, + { + "epoch": 1.043555585383041, + "grad_norm": 1.0289428234100342, + "learning_rate": 4.8896956389959084e-05, + "loss": 0.5326, + "step": 15550 + }, + { + "epoch": 1.0436898090668099, + "grad_norm": 1.0326473712921143, + "learning_rate": 4.888609075471042e-05, + "loss": 0.5655, + "step": 15552 + }, + { + "epoch": 1.0438240327505788, + "grad_norm": 0.9574761390686035, + "learning_rate": 4.88752251720918e-05, + "loss": 0.4885, + "step": 15554 + }, + { + "epoch": 1.043958256434348, + "grad_norm": 1.0071873664855957, + "learning_rate": 4.886435964261656e-05, + "loss": 0.5466, + "step": 15556 + }, + { + "epoch": 1.0440924801181168, + "grad_norm": 0.9385173916816711, + "learning_rate": 4.885349416679812e-05, + "loss": 0.5591, + "step": 15558 + }, + { + "epoch": 1.0442267038018858, + "grad_norm": 1.1447536945343018, + "learning_rate": 4.884262874514981e-05, + "loss": 0.5402, + "step": 15560 + }, + { + "epoch": 1.044360927485655, + "grad_norm": 1.043013095855713, + "learning_rate": 4.883176337818502e-05, + "loss": 0.4843, + "step": 15562 + }, + { + "epoch": 1.0444951511694238, + "grad_norm": 0.9053693413734436, + "learning_rate": 4.88208980664171e-05, + "loss": 0.5234, + "step": 15564 + }, + { + "epoch": 1.044629374853193, + "grad_norm": 1.0035746097564697, + "learning_rate": 4.8810032810359446e-05, + "loss": 0.5079, + "step": 15566 + }, + { + "epoch": 1.0447635985369619, + "grad_norm": 1.0047979354858398, + "learning_rate": 4.879916761052539e-05, + "loss": 0.5525, + "step": 15568 + }, + { + "epoch": 1.0448978222207308, + "grad_norm": 1.0252243280410767, + "learning_rate": 4.8788302467428316e-05, + "loss": 0.5254, + "step": 15570 + }, + { + "epoch": 1.0450320459045, + "grad_norm": 1.0482876300811768, + "learning_rate": 4.877743738158155e-05, + "loss": 0.5176, + "step": 15572 + }, + { + "epoch": 1.0451662695882689, + "grad_norm": 1.153091311454773, + "learning_rate": 4.876657235349848e-05, + "loss": 0.564, + "step": 15574 + }, + { + "epoch": 1.0453004932720378, + "grad_norm": 0.9788694977760315, + "learning_rate": 4.875570738369242e-05, + "loss": 0.482, + "step": 15576 + }, + { + "epoch": 1.045434716955807, + "grad_norm": 1.1284476518630981, + "learning_rate": 4.8744842472676764e-05, + "loss": 0.5772, + "step": 15578 + }, + { + "epoch": 1.0455689406395758, + "grad_norm": 0.9172835350036621, + "learning_rate": 4.873397762096481e-05, + "loss": 0.504, + "step": 15580 + }, + { + "epoch": 1.0457031643233448, + "grad_norm": 0.9032467007637024, + "learning_rate": 4.872311282906994e-05, + "loss": 0.5281, + "step": 15582 + }, + { + "epoch": 1.045837388007114, + "grad_norm": 1.1164354085922241, + "learning_rate": 4.871224809750547e-05, + "loss": 0.4898, + "step": 15584 + }, + { + "epoch": 1.0459716116908828, + "grad_norm": 1.059873104095459, + "learning_rate": 4.8701383426784755e-05, + "loss": 0.4963, + "step": 15586 + }, + { + "epoch": 1.046105835374652, + "grad_norm": 1.0384563207626343, + "learning_rate": 4.869051881742112e-05, + "loss": 0.5859, + "step": 15588 + }, + { + "epoch": 1.0462400590584209, + "grad_norm": 1.0289987325668335, + "learning_rate": 4.867965426992788e-05, + "loss": 0.5095, + "step": 15590 + }, + { + "epoch": 1.0463742827421898, + "grad_norm": 0.8439013957977295, + "learning_rate": 4.86687897848184e-05, + "loss": 0.5785, + "step": 15592 + }, + { + "epoch": 1.046508506425959, + "grad_norm": 0.995211124420166, + "learning_rate": 4.8657925362605964e-05, + "loss": 0.5559, + "step": 15594 + }, + { + "epoch": 1.0466427301097279, + "grad_norm": 0.8534039258956909, + "learning_rate": 4.8647061003803934e-05, + "loss": 0.5191, + "step": 15596 + }, + { + "epoch": 1.0467769537934968, + "grad_norm": 0.8605908751487732, + "learning_rate": 4.86361967089256e-05, + "loss": 0.5426, + "step": 15598 + }, + { + "epoch": 1.046911177477266, + "grad_norm": 0.9673829078674316, + "learning_rate": 4.8625332478484295e-05, + "loss": 0.4796, + "step": 15600 + }, + { + "epoch": 1.0470454011610348, + "grad_norm": 1.0090407133102417, + "learning_rate": 4.861446831299332e-05, + "loss": 0.5467, + "step": 15602 + }, + { + "epoch": 1.0471796248448038, + "grad_norm": 0.9353126287460327, + "learning_rate": 4.8603604212966005e-05, + "loss": 0.5637, + "step": 15604 + }, + { + "epoch": 1.047313848528573, + "grad_norm": 1.1131153106689453, + "learning_rate": 4.859274017891563e-05, + "loss": 0.5538, + "step": 15606 + }, + { + "epoch": 1.0474480722123418, + "grad_norm": 0.9784945845603943, + "learning_rate": 4.858187621135553e-05, + "loss": 0.5159, + "step": 15608 + }, + { + "epoch": 1.047582295896111, + "grad_norm": 2.116542100906372, + "learning_rate": 4.8571012310798966e-05, + "loss": 0.4381, + "step": 15610 + }, + { + "epoch": 1.0477165195798799, + "grad_norm": 0.9533230066299438, + "learning_rate": 4.8560148477759285e-05, + "loss": 0.5108, + "step": 15612 + }, + { + "epoch": 1.0478507432636488, + "grad_norm": 0.8793781399726868, + "learning_rate": 4.854928471274974e-05, + "loss": 0.4968, + "step": 15614 + }, + { + "epoch": 1.047984966947418, + "grad_norm": 1.0545098781585693, + "learning_rate": 4.8538421016283646e-05, + "loss": 0.5109, + "step": 15616 + }, + { + "epoch": 1.0481191906311869, + "grad_norm": 0.9535578489303589, + "learning_rate": 4.8527557388874285e-05, + "loss": 0.5156, + "step": 15618 + }, + { + "epoch": 1.0482534143149558, + "grad_norm": 0.9787843823432922, + "learning_rate": 4.851669383103495e-05, + "loss": 0.5216, + "step": 15620 + }, + { + "epoch": 1.048387637998725, + "grad_norm": 1.1390225887298584, + "learning_rate": 4.85058303432789e-05, + "loss": 0.5536, + "step": 15622 + }, + { + "epoch": 1.0485218616824938, + "grad_norm": 0.929733395576477, + "learning_rate": 4.849496692611945e-05, + "loss": 0.5519, + "step": 15624 + }, + { + "epoch": 1.048656085366263, + "grad_norm": 0.8988843560218811, + "learning_rate": 4.8484103580069826e-05, + "loss": 0.5773, + "step": 15626 + }, + { + "epoch": 1.048790309050032, + "grad_norm": 0.9410921335220337, + "learning_rate": 4.8473240305643346e-05, + "loss": 0.5147, + "step": 15628 + }, + { + "epoch": 1.0489245327338008, + "grad_norm": 1.0105018615722656, + "learning_rate": 4.846237710335325e-05, + "loss": 0.559, + "step": 15630 + }, + { + "epoch": 1.04905875641757, + "grad_norm": 1.3036748170852661, + "learning_rate": 4.845151397371283e-05, + "loss": 0.5141, + "step": 15632 + }, + { + "epoch": 1.0491929801013389, + "grad_norm": 0.9526950716972351, + "learning_rate": 4.8440650917235304e-05, + "loss": 0.497, + "step": 15634 + }, + { + "epoch": 1.0493272037851078, + "grad_norm": 1.0017216205596924, + "learning_rate": 4.8429787934433974e-05, + "loss": 0.5319, + "step": 15636 + }, + { + "epoch": 1.049461427468877, + "grad_norm": 0.922514796257019, + "learning_rate": 4.841892502582206e-05, + "loss": 0.5426, + "step": 15638 + }, + { + "epoch": 1.0495956511526459, + "grad_norm": 0.8724059462547302, + "learning_rate": 4.840806219191285e-05, + "loss": 0.4598, + "step": 15640 + }, + { + "epoch": 1.0497298748364148, + "grad_norm": 0.9697558879852295, + "learning_rate": 4.839719943321954e-05, + "loss": 0.4628, + "step": 15642 + }, + { + "epoch": 1.049864098520184, + "grad_norm": 0.9984668493270874, + "learning_rate": 4.8386336750255435e-05, + "loss": 0.4958, + "step": 15644 + }, + { + "epoch": 1.0499983222039528, + "grad_norm": 1.0505653619766235, + "learning_rate": 4.837547414353373e-05, + "loss": 0.5643, + "step": 15646 + }, + { + "epoch": 1.050132545887722, + "grad_norm": 1.0480382442474365, + "learning_rate": 4.8364611613567686e-05, + "loss": 0.5502, + "step": 15648 + }, + { + "epoch": 1.050266769571491, + "grad_norm": 0.9821338057518005, + "learning_rate": 4.835374916087053e-05, + "loss": 0.5248, + "step": 15650 + }, + { + "epoch": 1.0504009932552598, + "grad_norm": 1.1076489686965942, + "learning_rate": 4.8342886785955484e-05, + "loss": 0.5948, + "step": 15652 + }, + { + "epoch": 1.050535216939029, + "grad_norm": 0.9364720582962036, + "learning_rate": 4.8332024489335775e-05, + "loss": 0.5359, + "step": 15654 + }, + { + "epoch": 1.0506694406227979, + "grad_norm": 0.8961271643638611, + "learning_rate": 4.8321162271524634e-05, + "loss": 0.5273, + "step": 15656 + }, + { + "epoch": 1.0508036643065668, + "grad_norm": 0.9937961101531982, + "learning_rate": 4.831030013303528e-05, + "loss": 0.5199, + "step": 15658 + }, + { + "epoch": 1.050937887990336, + "grad_norm": 1.0300461053848267, + "learning_rate": 4.829943807438091e-05, + "loss": 0.498, + "step": 15660 + }, + { + "epoch": 1.0510721116741049, + "grad_norm": 0.9919794797897339, + "learning_rate": 4.828857609607476e-05, + "loss": 0.5201, + "step": 15662 + }, + { + "epoch": 1.051206335357874, + "grad_norm": 0.9712582230567932, + "learning_rate": 4.827771419863001e-05, + "loss": 0.5073, + "step": 15664 + }, + { + "epoch": 1.051340559041643, + "grad_norm": 0.9613664746284485, + "learning_rate": 4.826685238255989e-05, + "loss": 0.4974, + "step": 15666 + }, + { + "epoch": 1.0514747827254118, + "grad_norm": 1.1144733428955078, + "learning_rate": 4.825599064837758e-05, + "loss": 0.573, + "step": 15668 + }, + { + "epoch": 1.051609006409181, + "grad_norm": 1.1156517267227173, + "learning_rate": 4.824512899659628e-05, + "loss": 0.5103, + "step": 15670 + }, + { + "epoch": 1.05174323009295, + "grad_norm": 0.9424161911010742, + "learning_rate": 4.823426742772917e-05, + "loss": 0.5447, + "step": 15672 + }, + { + "epoch": 1.0518774537767188, + "grad_norm": 1.1218960285186768, + "learning_rate": 4.822340594228948e-05, + "loss": 0.4991, + "step": 15674 + }, + { + "epoch": 1.052011677460488, + "grad_norm": 1.262618899345398, + "learning_rate": 4.821254454079035e-05, + "loss": 0.5725, + "step": 15676 + }, + { + "epoch": 1.0521459011442569, + "grad_norm": 1.4871364831924438, + "learning_rate": 4.820168322374498e-05, + "loss": 0.5486, + "step": 15678 + }, + { + "epoch": 1.0522801248280258, + "grad_norm": 0.847819447517395, + "learning_rate": 4.819082199166654e-05, + "loss": 0.4634, + "step": 15680 + }, + { + "epoch": 1.052414348511795, + "grad_norm": 1.0007290840148926, + "learning_rate": 4.8179960845068204e-05, + "loss": 0.5471, + "step": 15682 + }, + { + "epoch": 1.0525485721955639, + "grad_norm": 0.8427039384841919, + "learning_rate": 4.816909978446314e-05, + "loss": 0.4646, + "step": 15684 + }, + { + "epoch": 1.052682795879333, + "grad_norm": 1.072413444519043, + "learning_rate": 4.815823881036451e-05, + "loss": 0.4995, + "step": 15686 + }, + { + "epoch": 1.052817019563102, + "grad_norm": 0.9217433929443359, + "learning_rate": 4.814737792328547e-05, + "loss": 0.5128, + "step": 15688 + }, + { + "epoch": 1.0529512432468708, + "grad_norm": 0.9816290140151978, + "learning_rate": 4.8136517123739197e-05, + "loss": 0.5168, + "step": 15690 + }, + { + "epoch": 1.05308546693064, + "grad_norm": 1.0718098878860474, + "learning_rate": 4.8125656412238805e-05, + "loss": 0.5763, + "step": 15692 + }, + { + "epoch": 1.053219690614409, + "grad_norm": 0.9611361026763916, + "learning_rate": 4.8114795789297476e-05, + "loss": 0.4981, + "step": 15694 + }, + { + "epoch": 1.0533539142981778, + "grad_norm": 1.0672252178192139, + "learning_rate": 4.810393525542833e-05, + "loss": 0.5391, + "step": 15696 + }, + { + "epoch": 1.053488137981947, + "grad_norm": 1.0438896417617798, + "learning_rate": 4.8093074811144525e-05, + "loss": 0.5363, + "step": 15698 + }, + { + "epoch": 1.0536223616657159, + "grad_norm": 1.0529118776321411, + "learning_rate": 4.808221445695916e-05, + "loss": 0.5853, + "step": 15700 + }, + { + "epoch": 1.053756585349485, + "grad_norm": 0.7470693588256836, + "learning_rate": 4.8071354193385424e-05, + "loss": 0.4655, + "step": 15702 + }, + { + "epoch": 1.053890809033254, + "grad_norm": 1.0209118127822876, + "learning_rate": 4.806049402093638e-05, + "loss": 0.4974, + "step": 15704 + }, + { + "epoch": 1.0540250327170229, + "grad_norm": 0.9176881909370422, + "learning_rate": 4.80496339401252e-05, + "loss": 0.5345, + "step": 15706 + }, + { + "epoch": 1.054159256400792, + "grad_norm": 0.9125534296035767, + "learning_rate": 4.803877395146497e-05, + "loss": 0.4683, + "step": 15708 + }, + { + "epoch": 1.054293480084561, + "grad_norm": 1.0628985166549683, + "learning_rate": 4.8027914055468826e-05, + "loss": 0.565, + "step": 15710 + }, + { + "epoch": 1.0544277037683298, + "grad_norm": 1.0448428392410278, + "learning_rate": 4.8017054252649856e-05, + "loss": 0.5217, + "step": 15712 + }, + { + "epoch": 1.054561927452099, + "grad_norm": 1.2996362447738647, + "learning_rate": 4.800619454352118e-05, + "loss": 0.5223, + "step": 15714 + }, + { + "epoch": 1.054696151135868, + "grad_norm": 0.9758515954017639, + "learning_rate": 4.799533492859588e-05, + "loss": 0.5306, + "step": 15716 + }, + { + "epoch": 1.0548303748196368, + "grad_norm": 0.9228617548942566, + "learning_rate": 4.798447540838708e-05, + "loss": 0.5784, + "step": 15718 + }, + { + "epoch": 1.054964598503406, + "grad_norm": 0.9225476384162903, + "learning_rate": 4.797361598340784e-05, + "loss": 0.5606, + "step": 15720 + }, + { + "epoch": 1.0550988221871749, + "grad_norm": 0.8702499866485596, + "learning_rate": 4.7962756654171264e-05, + "loss": 0.4912, + "step": 15722 + }, + { + "epoch": 1.055233045870944, + "grad_norm": 0.9524227976799011, + "learning_rate": 4.795189742119043e-05, + "loss": 0.4948, + "step": 15724 + }, + { + "epoch": 1.055367269554713, + "grad_norm": 0.9143248796463013, + "learning_rate": 4.794103828497842e-05, + "loss": 0.5853, + "step": 15726 + }, + { + "epoch": 1.0555014932384819, + "grad_norm": 1.1705297231674194, + "learning_rate": 4.79301792460483e-05, + "loss": 0.4788, + "step": 15728 + }, + { + "epoch": 1.055635716922251, + "grad_norm": 0.9776636362075806, + "learning_rate": 4.791932030491313e-05, + "loss": 0.5441, + "step": 15730 + }, + { + "epoch": 1.05576994060602, + "grad_norm": 1.2618513107299805, + "learning_rate": 4.7908461462086e-05, + "loss": 0.5419, + "step": 15732 + }, + { + "epoch": 1.0559041642897888, + "grad_norm": 0.7628543972969055, + "learning_rate": 4.789760271807993e-05, + "loss": 0.5262, + "step": 15734 + }, + { + "epoch": 1.056038387973558, + "grad_norm": 1.0558035373687744, + "learning_rate": 4.788674407340803e-05, + "loss": 0.5234, + "step": 15736 + }, + { + "epoch": 1.056172611657327, + "grad_norm": 1.4487251043319702, + "learning_rate": 4.787588552858328e-05, + "loss": 0.5369, + "step": 15738 + }, + { + "epoch": 1.056306835341096, + "grad_norm": 0.9106961488723755, + "learning_rate": 4.7865027084118805e-05, + "loss": 0.4481, + "step": 15740 + }, + { + "epoch": 1.056441059024865, + "grad_norm": 1.0827953815460205, + "learning_rate": 4.785416874052756e-05, + "loss": 0.4801, + "step": 15742 + }, + { + "epoch": 1.0565752827086339, + "grad_norm": 1.0268560647964478, + "learning_rate": 4.7843310498322646e-05, + "loss": 0.5095, + "step": 15744 + }, + { + "epoch": 1.056709506392403, + "grad_norm": 1.1408530473709106, + "learning_rate": 4.783245235801706e-05, + "loss": 0.5395, + "step": 15746 + }, + { + "epoch": 1.056843730076172, + "grad_norm": 1.4307409524917603, + "learning_rate": 4.7821594320123844e-05, + "loss": 0.4855, + "step": 15748 + }, + { + "epoch": 1.0569779537599409, + "grad_norm": 1.0708088874816895, + "learning_rate": 4.781073638515601e-05, + "loss": 0.521, + "step": 15750 + }, + { + "epoch": 1.05711217744371, + "grad_norm": 1.113745927810669, + "learning_rate": 4.779987855362658e-05, + "loss": 0.5703, + "step": 15752 + }, + { + "epoch": 1.057246401127479, + "grad_norm": 1.1250548362731934, + "learning_rate": 4.778902082604856e-05, + "loss": 0.5019, + "step": 15754 + }, + { + "epoch": 1.0573806248112478, + "grad_norm": 0.9557268619537354, + "learning_rate": 4.777816320293496e-05, + "loss": 0.4917, + "step": 15756 + }, + { + "epoch": 1.057514848495017, + "grad_norm": 0.9044797420501709, + "learning_rate": 4.7767305684798775e-05, + "loss": 0.5486, + "step": 15758 + }, + { + "epoch": 1.057649072178786, + "grad_norm": 0.9962960481643677, + "learning_rate": 4.7756448272153014e-05, + "loss": 0.533, + "step": 15760 + }, + { + "epoch": 1.057783295862555, + "grad_norm": 0.907252848148346, + "learning_rate": 4.774559096551064e-05, + "loss": 0.5375, + "step": 15762 + }, + { + "epoch": 1.057917519546324, + "grad_norm": 1.0109272003173828, + "learning_rate": 4.773473376538469e-05, + "loss": 0.5223, + "step": 15764 + }, + { + "epoch": 1.0580517432300929, + "grad_norm": 1.1772733926773071, + "learning_rate": 4.772387667228808e-05, + "loss": 0.521, + "step": 15766 + }, + { + "epoch": 1.058185966913862, + "grad_norm": 1.018127202987671, + "learning_rate": 4.771301968673386e-05, + "loss": 0.4875, + "step": 15768 + }, + { + "epoch": 1.058320190597631, + "grad_norm": 1.1795014142990112, + "learning_rate": 4.770216280923493e-05, + "loss": 0.5322, + "step": 15770 + }, + { + "epoch": 1.0584544142813999, + "grad_norm": 0.9482703804969788, + "learning_rate": 4.7691306040304306e-05, + "loss": 0.4502, + "step": 15772 + }, + { + "epoch": 1.058588637965169, + "grad_norm": 0.9544832110404968, + "learning_rate": 4.768044938045493e-05, + "loss": 0.5145, + "step": 15774 + }, + { + "epoch": 1.058722861648938, + "grad_norm": 1.0192888975143433, + "learning_rate": 4.766959283019976e-05, + "loss": 0.522, + "step": 15776 + }, + { + "epoch": 1.058857085332707, + "grad_norm": 1.009538173675537, + "learning_rate": 4.765873639005174e-05, + "loss": 0.5029, + "step": 15778 + }, + { + "epoch": 1.058991309016476, + "grad_norm": 1.0669879913330078, + "learning_rate": 4.764788006052383e-05, + "loss": 0.5343, + "step": 15780 + }, + { + "epoch": 1.059125532700245, + "grad_norm": 0.9777652025222778, + "learning_rate": 4.763702384212896e-05, + "loss": 0.5338, + "step": 15782 + }, + { + "epoch": 1.059259756384014, + "grad_norm": 1.0392273664474487, + "learning_rate": 4.762616773538007e-05, + "loss": 0.5249, + "step": 15784 + }, + { + "epoch": 1.059393980067783, + "grad_norm": 0.9199720025062561, + "learning_rate": 4.7615311740790074e-05, + "loss": 0.4959, + "step": 15786 + }, + { + "epoch": 1.0595282037515519, + "grad_norm": 0.9195120334625244, + "learning_rate": 4.760445585887192e-05, + "loss": 0.5123, + "step": 15788 + }, + { + "epoch": 1.059662427435321, + "grad_norm": 1.0891872644424438, + "learning_rate": 4.759360009013849e-05, + "loss": 0.4989, + "step": 15790 + }, + { + "epoch": 1.05979665111909, + "grad_norm": 0.9747539162635803, + "learning_rate": 4.7582744435102774e-05, + "loss": 0.4794, + "step": 15792 + }, + { + "epoch": 1.0599308748028589, + "grad_norm": 1.2288579940795898, + "learning_rate": 4.7571888894277604e-05, + "loss": 0.5171, + "step": 15794 + }, + { + "epoch": 1.060065098486628, + "grad_norm": 1.1269773244857788, + "learning_rate": 4.756103346817589e-05, + "loss": 0.5108, + "step": 15796 + }, + { + "epoch": 1.060199322170397, + "grad_norm": 1.2069462537765503, + "learning_rate": 4.755017815731059e-05, + "loss": 0.5832, + "step": 15798 + }, + { + "epoch": 1.060333545854166, + "grad_norm": 1.3380992412567139, + "learning_rate": 4.753932296219451e-05, + "loss": 0.5867, + "step": 15800 + }, + { + "epoch": 1.060467769537935, + "grad_norm": 1.0082257986068726, + "learning_rate": 4.7528467883340615e-05, + "loss": 0.5294, + "step": 15802 + }, + { + "epoch": 1.060601993221704, + "grad_norm": 1.0993695259094238, + "learning_rate": 4.751761292126173e-05, + "loss": 0.5601, + "step": 15804 + }, + { + "epoch": 1.060736216905473, + "grad_norm": 1.235148310661316, + "learning_rate": 4.750675807647077e-05, + "loss": 0.5, + "step": 15806 + }, + { + "epoch": 1.060870440589242, + "grad_norm": 0.9587987661361694, + "learning_rate": 4.749590334948057e-05, + "loss": 0.521, + "step": 15808 + }, + { + "epoch": 1.0610046642730109, + "grad_norm": 1.074412226676941, + "learning_rate": 4.7485048740804026e-05, + "loss": 0.5009, + "step": 15810 + }, + { + "epoch": 1.06113888795678, + "grad_norm": 1.0490909814834595, + "learning_rate": 4.747419425095398e-05, + "loss": 0.5007, + "step": 15812 + }, + { + "epoch": 1.061273111640549, + "grad_norm": 1.0220736265182495, + "learning_rate": 4.74633398804433e-05, + "loss": 0.4692, + "step": 15814 + }, + { + "epoch": 1.061407335324318, + "grad_norm": 0.9102790355682373, + "learning_rate": 4.7452485629784804e-05, + "loss": 0.4877, + "step": 15816 + }, + { + "epoch": 1.061541559008087, + "grad_norm": 1.1566509008407593, + "learning_rate": 4.744163149949136e-05, + "loss": 0.5361, + "step": 15818 + }, + { + "epoch": 1.061675782691856, + "grad_norm": 1.052136778831482, + "learning_rate": 4.74307774900758e-05, + "loss": 0.5177, + "step": 15820 + }, + { + "epoch": 1.061810006375625, + "grad_norm": 1.163494348526001, + "learning_rate": 4.7419923602050954e-05, + "loss": 0.5434, + "step": 15822 + }, + { + "epoch": 1.061944230059394, + "grad_norm": 1.0164376497268677, + "learning_rate": 4.740906983592962e-05, + "loss": 0.5123, + "step": 15824 + }, + { + "epoch": 1.062078453743163, + "grad_norm": 1.0817575454711914, + "learning_rate": 4.739821619222468e-05, + "loss": 0.506, + "step": 15826 + }, + { + "epoch": 1.062212677426932, + "grad_norm": 1.079735517501831, + "learning_rate": 4.738736267144887e-05, + "loss": 0.5029, + "step": 15828 + }, + { + "epoch": 1.062346901110701, + "grad_norm": 1.0035247802734375, + "learning_rate": 4.7376509274115075e-05, + "loss": 0.5282, + "step": 15830 + }, + { + "epoch": 1.0624811247944699, + "grad_norm": 0.8570907115936279, + "learning_rate": 4.736565600073602e-05, + "loss": 0.4403, + "step": 15832 + }, + { + "epoch": 1.062615348478239, + "grad_norm": 1.206586241722107, + "learning_rate": 4.735480285182457e-05, + "loss": 0.5351, + "step": 15834 + }, + { + "epoch": 1.062749572162008, + "grad_norm": 0.960978627204895, + "learning_rate": 4.7343949827893465e-05, + "loss": 0.5426, + "step": 15836 + }, + { + "epoch": 1.062883795845777, + "grad_norm": 1.1588352918624878, + "learning_rate": 4.733309692945552e-05, + "loss": 0.5676, + "step": 15838 + }, + { + "epoch": 1.063018019529546, + "grad_norm": 1.031522274017334, + "learning_rate": 4.732224415702349e-05, + "loss": 0.5163, + "step": 15840 + }, + { + "epoch": 1.063152243213315, + "grad_norm": 0.9539149403572083, + "learning_rate": 4.731139151111017e-05, + "loss": 0.5197, + "step": 15842 + }, + { + "epoch": 1.063286466897084, + "grad_norm": 0.9566182494163513, + "learning_rate": 4.7300538992228295e-05, + "loss": 0.5026, + "step": 15844 + }, + { + "epoch": 1.063420690580853, + "grad_norm": 0.9178019762039185, + "learning_rate": 4.7289686600890666e-05, + "loss": 0.5097, + "step": 15846 + }, + { + "epoch": 1.063554914264622, + "grad_norm": 1.049057126045227, + "learning_rate": 4.727883433761e-05, + "loss": 0.546, + "step": 15848 + }, + { + "epoch": 1.063689137948391, + "grad_norm": 1.0182324647903442, + "learning_rate": 4.726798220289907e-05, + "loss": 0.5669, + "step": 15850 + }, + { + "epoch": 1.06382336163216, + "grad_norm": 0.9539602994918823, + "learning_rate": 4.7257130197270586e-05, + "loss": 0.5234, + "step": 15852 + }, + { + "epoch": 1.063957585315929, + "grad_norm": 0.981003999710083, + "learning_rate": 4.724627832123735e-05, + "loss": 0.5796, + "step": 15854 + }, + { + "epoch": 1.064091808999698, + "grad_norm": 0.9285516142845154, + "learning_rate": 4.7235426575312004e-05, + "loss": 0.5216, + "step": 15856 + }, + { + "epoch": 1.064226032683467, + "grad_norm": 1.002652883529663, + "learning_rate": 4.722457496000735e-05, + "loss": 0.4857, + "step": 15858 + }, + { + "epoch": 1.064360256367236, + "grad_norm": 1.072216510772705, + "learning_rate": 4.721372347583604e-05, + "loss": 0.508, + "step": 15860 + }, + { + "epoch": 1.064494480051005, + "grad_norm": 0.9234919548034668, + "learning_rate": 4.7202872123310845e-05, + "loss": 0.51, + "step": 15862 + }, + { + "epoch": 1.064628703734774, + "grad_norm": 0.9576901197433472, + "learning_rate": 4.7192020902944446e-05, + "loss": 0.5355, + "step": 15864 + }, + { + "epoch": 1.064762927418543, + "grad_norm": 0.9486761689186096, + "learning_rate": 4.71811698152495e-05, + "loss": 0.4915, + "step": 15866 + }, + { + "epoch": 1.064897151102312, + "grad_norm": 1.001057744026184, + "learning_rate": 4.7170318860738774e-05, + "loss": 0.5064, + "step": 15868 + }, + { + "epoch": 1.0650313747860811, + "grad_norm": 1.1781103610992432, + "learning_rate": 4.71594680399249e-05, + "loss": 0.551, + "step": 15870 + }, + { + "epoch": 1.06516559846985, + "grad_norm": 0.9904471635818481, + "learning_rate": 4.714861735332058e-05, + "loss": 0.481, + "step": 15872 + }, + { + "epoch": 1.065299822153619, + "grad_norm": 0.966191291809082, + "learning_rate": 4.713776680143848e-05, + "loss": 0.5074, + "step": 15874 + }, + { + "epoch": 1.065434045837388, + "grad_norm": 1.0541632175445557, + "learning_rate": 4.7126916384791274e-05, + "loss": 0.4729, + "step": 15876 + }, + { + "epoch": 1.065568269521157, + "grad_norm": 1.1319295167922974, + "learning_rate": 4.7116066103891615e-05, + "loss": 0.5112, + "step": 15878 + }, + { + "epoch": 1.065702493204926, + "grad_norm": 1.2292240858078003, + "learning_rate": 4.710521595925217e-05, + "loss": 0.534, + "step": 15880 + }, + { + "epoch": 1.065836716888695, + "grad_norm": 1.2403215169906616, + "learning_rate": 4.709436595138557e-05, + "loss": 0.5141, + "step": 15882 + }, + { + "epoch": 1.065970940572464, + "grad_norm": 1.0871567726135254, + "learning_rate": 4.708351608080447e-05, + "loss": 0.5905, + "step": 15884 + }, + { + "epoch": 1.066105164256233, + "grad_norm": 0.9792987108230591, + "learning_rate": 4.7072666348021485e-05, + "loss": 0.5363, + "step": 15886 + }, + { + "epoch": 1.066239387940002, + "grad_norm": 1.1306519508361816, + "learning_rate": 4.706181675354929e-05, + "loss": 0.5359, + "step": 15888 + }, + { + "epoch": 1.066373611623771, + "grad_norm": 0.9030818343162537, + "learning_rate": 4.7050967297900434e-05, + "loss": 0.4874, + "step": 15890 + }, + { + "epoch": 1.0665078353075401, + "grad_norm": 1.0452027320861816, + "learning_rate": 4.704011798158762e-05, + "loss": 0.5392, + "step": 15892 + }, + { + "epoch": 1.066642058991309, + "grad_norm": 1.0378111600875854, + "learning_rate": 4.702926880512337e-05, + "loss": 0.5701, + "step": 15894 + }, + { + "epoch": 1.066776282675078, + "grad_norm": 0.915477454662323, + "learning_rate": 4.701841976902035e-05, + "loss": 0.4822, + "step": 15896 + }, + { + "epoch": 1.066910506358847, + "grad_norm": 0.9939043521881104, + "learning_rate": 4.700757087379112e-05, + "loss": 0.4608, + "step": 15898 + }, + { + "epoch": 1.067044730042616, + "grad_norm": 0.9939762949943542, + "learning_rate": 4.69967221199483e-05, + "loss": 0.6015, + "step": 15900 + }, + { + "epoch": 1.067178953726385, + "grad_norm": 0.9916800856590271, + "learning_rate": 4.698587350800444e-05, + "loss": 0.5067, + "step": 15902 + }, + { + "epoch": 1.067313177410154, + "grad_norm": 0.9965695738792419, + "learning_rate": 4.697502503847214e-05, + "loss": 0.4918, + "step": 15904 + }, + { + "epoch": 1.067447401093923, + "grad_norm": 0.9177051782608032, + "learning_rate": 4.696417671186395e-05, + "loss": 0.5107, + "step": 15906 + }, + { + "epoch": 1.067581624777692, + "grad_norm": 1.0016969442367554, + "learning_rate": 4.6953328528692446e-05, + "loss": 0.5405, + "step": 15908 + }, + { + "epoch": 1.067715848461461, + "grad_norm": 1.0871280431747437, + "learning_rate": 4.694248048947017e-05, + "loss": 0.4547, + "step": 15910 + }, + { + "epoch": 1.06785007214523, + "grad_norm": 0.9711535573005676, + "learning_rate": 4.693163259470969e-05, + "loss": 0.5125, + "step": 15912 + }, + { + "epoch": 1.0679842958289991, + "grad_norm": 1.0115195512771606, + "learning_rate": 4.692078484492351e-05, + "loss": 0.5022, + "step": 15914 + }, + { + "epoch": 1.068118519512768, + "grad_norm": 1.0912082195281982, + "learning_rate": 4.6909937240624236e-05, + "loss": 0.5468, + "step": 15916 + }, + { + "epoch": 1.068252743196537, + "grad_norm": 1.298017978668213, + "learning_rate": 4.689908978232431e-05, + "loss": 0.5389, + "step": 15918 + }, + { + "epoch": 1.068386966880306, + "grad_norm": 1.0747569799423218, + "learning_rate": 4.688824247053633e-05, + "loss": 0.4934, + "step": 15920 + }, + { + "epoch": 1.068521190564075, + "grad_norm": 0.9096031188964844, + "learning_rate": 4.6877395305772733e-05, + "loss": 0.4718, + "step": 15922 + }, + { + "epoch": 1.068655414247844, + "grad_norm": 1.0086485147476196, + "learning_rate": 4.68665482885461e-05, + "loss": 0.5542, + "step": 15924 + }, + { + "epoch": 1.068789637931613, + "grad_norm": 1.085276484489441, + "learning_rate": 4.685570141936888e-05, + "loss": 0.585, + "step": 15926 + }, + { + "epoch": 1.068923861615382, + "grad_norm": 1.0500186681747437, + "learning_rate": 4.684485469875359e-05, + "loss": 0.5445, + "step": 15928 + }, + { + "epoch": 1.069058085299151, + "grad_norm": 1.138278841972351, + "learning_rate": 4.6834008127212715e-05, + "loss": 0.4745, + "step": 15930 + }, + { + "epoch": 1.06919230898292, + "grad_norm": 1.0815253257751465, + "learning_rate": 4.682316170525874e-05, + "loss": 0.5361, + "step": 15932 + }, + { + "epoch": 1.069326532666689, + "grad_norm": 0.9723836183547974, + "learning_rate": 4.681231543340412e-05, + "loss": 0.5012, + "step": 15934 + }, + { + "epoch": 1.0694607563504581, + "grad_norm": 0.908646821975708, + "learning_rate": 4.680146931216132e-05, + "loss": 0.5123, + "step": 15936 + }, + { + "epoch": 1.069594980034227, + "grad_norm": 1.2489705085754395, + "learning_rate": 4.679062334204281e-05, + "loss": 0.5196, + "step": 15938 + }, + { + "epoch": 1.069729203717996, + "grad_norm": 1.1271929740905762, + "learning_rate": 4.677977752356103e-05, + "loss": 0.5203, + "step": 15940 + }, + { + "epoch": 1.069863427401765, + "grad_norm": 1.1925686597824097, + "learning_rate": 4.676893185722844e-05, + "loss": 0.5434, + "step": 15942 + }, + { + "epoch": 1.069997651085534, + "grad_norm": 1.0192997455596924, + "learning_rate": 4.675808634355746e-05, + "loss": 0.5289, + "step": 15944 + }, + { + "epoch": 1.0701318747693032, + "grad_norm": 1.0960651636123657, + "learning_rate": 4.674724098306053e-05, + "loss": 0.5115, + "step": 15946 + }, + { + "epoch": 1.070266098453072, + "grad_norm": 0.9219456315040588, + "learning_rate": 4.6736395776250065e-05, + "loss": 0.4884, + "step": 15948 + }, + { + "epoch": 1.070400322136841, + "grad_norm": 1.0570820569992065, + "learning_rate": 4.6725550723638484e-05, + "loss": 0.5441, + "step": 15950 + }, + { + "epoch": 1.0705345458206101, + "grad_norm": 0.9883645176887512, + "learning_rate": 4.671470582573817e-05, + "loss": 0.5217, + "step": 15952 + }, + { + "epoch": 1.070668769504379, + "grad_norm": 1.0507816076278687, + "learning_rate": 4.670386108306159e-05, + "loss": 0.5517, + "step": 15954 + }, + { + "epoch": 1.070802993188148, + "grad_norm": 0.9801546335220337, + "learning_rate": 4.669301649612105e-05, + "loss": 0.494, + "step": 15956 + }, + { + "epoch": 1.0709372168719171, + "grad_norm": 1.0489400625228882, + "learning_rate": 4.6682172065428994e-05, + "loss": 0.5596, + "step": 15958 + }, + { + "epoch": 1.071071440555686, + "grad_norm": 0.9993674159049988, + "learning_rate": 4.667132779149778e-05, + "loss": 0.5661, + "step": 15960 + }, + { + "epoch": 1.071205664239455, + "grad_norm": 1.1793206930160522, + "learning_rate": 4.6660483674839794e-05, + "loss": 0.504, + "step": 15962 + }, + { + "epoch": 1.071339887923224, + "grad_norm": 0.9190689325332642, + "learning_rate": 4.664963971596737e-05, + "loss": 0.5407, + "step": 15964 + }, + { + "epoch": 1.071474111606993, + "grad_norm": 0.9717457890510559, + "learning_rate": 4.66387959153929e-05, + "loss": 0.5289, + "step": 15966 + }, + { + "epoch": 1.0716083352907622, + "grad_norm": 1.0649359226226807, + "learning_rate": 4.6627952273628705e-05, + "loss": 0.5045, + "step": 15968 + }, + { + "epoch": 1.071742558974531, + "grad_norm": 1.1034852266311646, + "learning_rate": 4.6617108791187137e-05, + "loss": 0.5771, + "step": 15970 + }, + { + "epoch": 1.0718767826583, + "grad_norm": 0.9633535742759705, + "learning_rate": 4.6606265468580516e-05, + "loss": 0.5035, + "step": 15972 + }, + { + "epoch": 1.0720110063420691, + "grad_norm": 0.9629459977149963, + "learning_rate": 4.6595422306321196e-05, + "loss": 0.5197, + "step": 15974 + }, + { + "epoch": 1.072145230025838, + "grad_norm": 0.9145321249961853, + "learning_rate": 4.658457930492145e-05, + "loss": 0.5586, + "step": 15976 + }, + { + "epoch": 1.072279453709607, + "grad_norm": 1.001462697982788, + "learning_rate": 4.657373646489365e-05, + "loss": 0.5125, + "step": 15978 + }, + { + "epoch": 1.0724136773933761, + "grad_norm": 1.06826913356781, + "learning_rate": 4.656289378675003e-05, + "loss": 0.4817, + "step": 15980 + }, + { + "epoch": 1.072547901077145, + "grad_norm": 0.9796193838119507, + "learning_rate": 4.655205127100296e-05, + "loss": 0.4744, + "step": 15982 + }, + { + "epoch": 1.072682124760914, + "grad_norm": 1.1441547870635986, + "learning_rate": 4.654120891816464e-05, + "loss": 0.559, + "step": 15984 + }, + { + "epoch": 1.072816348444683, + "grad_norm": 1.1235517263412476, + "learning_rate": 4.653036672874743e-05, + "loss": 0.5606, + "step": 15986 + }, + { + "epoch": 1.072950572128452, + "grad_norm": 0.9840055108070374, + "learning_rate": 4.651952470326356e-05, + "loss": 0.4739, + "step": 15988 + }, + { + "epoch": 1.0730847958122212, + "grad_norm": 0.9808876514434814, + "learning_rate": 4.6508682842225305e-05, + "loss": 0.5439, + "step": 15990 + }, + { + "epoch": 1.07321901949599, + "grad_norm": 1.0697896480560303, + "learning_rate": 4.6497841146144916e-05, + "loss": 0.5498, + "step": 15992 + }, + { + "epoch": 1.073353243179759, + "grad_norm": 1.1211960315704346, + "learning_rate": 4.6486999615534656e-05, + "loss": 0.5357, + "step": 15994 + }, + { + "epoch": 1.0734874668635281, + "grad_norm": 1.0424667596817017, + "learning_rate": 4.6476158250906743e-05, + "loss": 0.5504, + "step": 15996 + }, + { + "epoch": 1.073621690547297, + "grad_norm": 1.003092646598816, + "learning_rate": 4.6465317052773436e-05, + "loss": 0.5114, + "step": 15998 + }, + { + "epoch": 1.073755914231066, + "grad_norm": 1.044110894203186, + "learning_rate": 4.645447602164695e-05, + "loss": 0.511, + "step": 16000 + }, + { + "epoch": 1.0738901379148351, + "grad_norm": 1.0264644622802734, + "learning_rate": 4.644363515803949e-05, + "loss": 0.4896, + "step": 16002 + }, + { + "epoch": 1.074024361598604, + "grad_norm": 1.0796425342559814, + "learning_rate": 4.643279446246328e-05, + "loss": 0.5787, + "step": 16004 + }, + { + "epoch": 1.074158585282373, + "grad_norm": 1.070082187652588, + "learning_rate": 4.6421953935430514e-05, + "loss": 0.4995, + "step": 16006 + }, + { + "epoch": 1.074292808966142, + "grad_norm": 1.3933830261230469, + "learning_rate": 4.641111357745339e-05, + "loss": 0.5326, + "step": 16008 + }, + { + "epoch": 1.074427032649911, + "grad_norm": 0.9243556261062622, + "learning_rate": 4.6400273389044086e-05, + "loss": 0.5082, + "step": 16010 + }, + { + "epoch": 1.0745612563336802, + "grad_norm": 1.024209976196289, + "learning_rate": 4.6389433370714795e-05, + "loss": 0.5286, + "step": 16012 + }, + { + "epoch": 1.074695480017449, + "grad_norm": 0.9765271544456482, + "learning_rate": 4.637859352297766e-05, + "loss": 0.5442, + "step": 16014 + }, + { + "epoch": 1.074829703701218, + "grad_norm": 1.0625931024551392, + "learning_rate": 4.636775384634489e-05, + "loss": 0.4908, + "step": 16016 + }, + { + "epoch": 1.0749639273849871, + "grad_norm": 1.0464589595794678, + "learning_rate": 4.635691434132858e-05, + "loss": 0.5058, + "step": 16018 + }, + { + "epoch": 1.075098151068756, + "grad_norm": 0.9293957948684692, + "learning_rate": 4.634607500844092e-05, + "loss": 0.5146, + "step": 16020 + }, + { + "epoch": 1.0752323747525252, + "grad_norm": 1.012489676475525, + "learning_rate": 4.633523584819402e-05, + "loss": 0.5348, + "step": 16022 + }, + { + "epoch": 1.0753665984362941, + "grad_norm": 1.014211654663086, + "learning_rate": 4.632439686110003e-05, + "loss": 0.4705, + "step": 16024 + }, + { + "epoch": 1.075500822120063, + "grad_norm": 0.9691089987754822, + "learning_rate": 4.631355804767105e-05, + "loss": 0.4996, + "step": 16026 + }, + { + "epoch": 1.0756350458038322, + "grad_norm": 0.902036190032959, + "learning_rate": 4.6302719408419214e-05, + "loss": 0.4969, + "step": 16028 + }, + { + "epoch": 1.075769269487601, + "grad_norm": 1.0455721616744995, + "learning_rate": 4.629188094385661e-05, + "loss": 0.5705, + "step": 16030 + }, + { + "epoch": 1.07590349317137, + "grad_norm": 0.8780853748321533, + "learning_rate": 4.6281042654495346e-05, + "loss": 0.5202, + "step": 16032 + }, + { + "epoch": 1.0760377168551392, + "grad_norm": 0.939189076423645, + "learning_rate": 4.627020454084749e-05, + "loss": 0.5331, + "step": 16034 + }, + { + "epoch": 1.076171940538908, + "grad_norm": 1.0931113958358765, + "learning_rate": 4.6259366603425153e-05, + "loss": 0.4951, + "step": 16036 + }, + { + "epoch": 1.076306164222677, + "grad_norm": 1.035915732383728, + "learning_rate": 4.624852884274037e-05, + "loss": 0.5259, + "step": 16038 + }, + { + "epoch": 1.0764403879064461, + "grad_norm": 2.711751937866211, + "learning_rate": 4.623769125930524e-05, + "loss": 0.5106, + "step": 16040 + }, + { + "epoch": 1.076574611590215, + "grad_norm": 1.0253950357437134, + "learning_rate": 4.6226853853631775e-05, + "loss": 0.5397, + "step": 16042 + }, + { + "epoch": 1.0767088352739842, + "grad_norm": 1.050638198852539, + "learning_rate": 4.621601662623208e-05, + "loss": 0.519, + "step": 16044 + }, + { + "epoch": 1.0768430589577531, + "grad_norm": 0.9499335885047913, + "learning_rate": 4.6205179577618125e-05, + "loss": 0.5042, + "step": 16046 + }, + { + "epoch": 1.076977282641522, + "grad_norm": 1.057031273841858, + "learning_rate": 4.619434270830199e-05, + "loss": 0.4846, + "step": 16048 + }, + { + "epoch": 1.0771115063252912, + "grad_norm": 0.9463146328926086, + "learning_rate": 4.6183506018795674e-05, + "loss": 0.4907, + "step": 16050 + }, + { + "epoch": 1.07724573000906, + "grad_norm": 0.9645826816558838, + "learning_rate": 4.61726695096112e-05, + "loss": 0.4425, + "step": 16052 + }, + { + "epoch": 1.077379953692829, + "grad_norm": 1.0076628923416138, + "learning_rate": 4.616183318126056e-05, + "loss": 0.4796, + "step": 16054 + }, + { + "epoch": 1.0775141773765982, + "grad_norm": 0.9495932459831238, + "learning_rate": 4.615099703425576e-05, + "loss": 0.5142, + "step": 16056 + }, + { + "epoch": 1.077648401060367, + "grad_norm": 1.01508367061615, + "learning_rate": 4.614016106910877e-05, + "loss": 0.544, + "step": 16058 + }, + { + "epoch": 1.077782624744136, + "grad_norm": 0.9466344714164734, + "learning_rate": 4.6129325286331594e-05, + "loss": 0.529, + "step": 16060 + }, + { + "epoch": 1.0779168484279051, + "grad_norm": 1.0745893716812134, + "learning_rate": 4.611848968643618e-05, + "loss": 0.5167, + "step": 16062 + }, + { + "epoch": 1.078051072111674, + "grad_norm": 1.092829942703247, + "learning_rate": 4.61076542699345e-05, + "loss": 0.5296, + "step": 16064 + }, + { + "epoch": 1.0781852957954432, + "grad_norm": 1.1403659582138062, + "learning_rate": 4.609681903733848e-05, + "loss": 0.5268, + "step": 16066 + }, + { + "epoch": 1.0783195194792121, + "grad_norm": 0.9573639631271362, + "learning_rate": 4.608598398916012e-05, + "loss": 0.5114, + "step": 16068 + }, + { + "epoch": 1.078453743162981, + "grad_norm": 1.0285440683364868, + "learning_rate": 4.607514912591131e-05, + "loss": 0.5266, + "step": 16070 + }, + { + "epoch": 1.0785879668467502, + "grad_norm": 1.112507939338684, + "learning_rate": 4.6064314448103974e-05, + "loss": 0.4869, + "step": 16072 + }, + { + "epoch": 1.078722190530519, + "grad_norm": 0.9259080290794373, + "learning_rate": 4.605347995625005e-05, + "loss": 0.4942, + "step": 16074 + }, + { + "epoch": 1.078856414214288, + "grad_norm": 1.1185704469680786, + "learning_rate": 4.604264565086141e-05, + "loss": 0.4717, + "step": 16076 + }, + { + "epoch": 1.0789906378980572, + "grad_norm": 1.0561480522155762, + "learning_rate": 4.6031811532450025e-05, + "loss": 0.6153, + "step": 16078 + }, + { + "epoch": 1.079124861581826, + "grad_norm": 1.0150277614593506, + "learning_rate": 4.602097760152771e-05, + "loss": 0.5489, + "step": 16080 + }, + { + "epoch": 1.079259085265595, + "grad_norm": 0.9486995935440063, + "learning_rate": 4.601014385860639e-05, + "loss": 0.4631, + "step": 16082 + }, + { + "epoch": 1.0793933089493641, + "grad_norm": 0.9660173058509827, + "learning_rate": 4.599931030419793e-05, + "loss": 0.5461, + "step": 16084 + }, + { + "epoch": 1.079527532633133, + "grad_norm": 1.0332201719284058, + "learning_rate": 4.598847693881419e-05, + "loss": 0.5661, + "step": 16086 + }, + { + "epoch": 1.0796617563169022, + "grad_norm": 0.9627061486244202, + "learning_rate": 4.5977643762967024e-05, + "loss": 0.4992, + "step": 16088 + }, + { + "epoch": 1.0797959800006711, + "grad_norm": 1.7318940162658691, + "learning_rate": 4.596681077716828e-05, + "loss": 0.5161, + "step": 16090 + }, + { + "epoch": 1.07993020368444, + "grad_norm": 1.0354950428009033, + "learning_rate": 4.59559779819298e-05, + "loss": 0.5276, + "step": 16092 + }, + { + "epoch": 1.0800644273682092, + "grad_norm": 1.1396992206573486, + "learning_rate": 4.594514537776341e-05, + "loss": 0.5819, + "step": 16094 + }, + { + "epoch": 1.080198651051978, + "grad_norm": 0.9875621199607849, + "learning_rate": 4.5934312965180916e-05, + "loss": 0.5414, + "step": 16096 + }, + { + "epoch": 1.0803328747357472, + "grad_norm": 0.8946545124053955, + "learning_rate": 4.5923480744694156e-05, + "loss": 0.5294, + "step": 16098 + }, + { + "epoch": 1.0804670984195162, + "grad_norm": 0.9404687285423279, + "learning_rate": 4.5912648716814896e-05, + "loss": 0.5864, + "step": 16100 + }, + { + "epoch": 1.080601322103285, + "grad_norm": 0.9598170518875122, + "learning_rate": 4.590181688205496e-05, + "loss": 0.4466, + "step": 16102 + }, + { + "epoch": 1.0807355457870542, + "grad_norm": 0.9472155570983887, + "learning_rate": 4.58909852409261e-05, + "loss": 0.5081, + "step": 16104 + }, + { + "epoch": 1.0808697694708231, + "grad_norm": 1.1002511978149414, + "learning_rate": 4.5880153793940136e-05, + "loss": 0.5322, + "step": 16106 + }, + { + "epoch": 1.081003993154592, + "grad_norm": 1.1255414485931396, + "learning_rate": 4.5869322541608775e-05, + "loss": 0.5168, + "step": 16108 + }, + { + "epoch": 1.0811382168383612, + "grad_norm": 1.0046885013580322, + "learning_rate": 4.5858491484443815e-05, + "loss": 0.4487, + "step": 16110 + }, + { + "epoch": 1.0812724405221301, + "grad_norm": 0.956312894821167, + "learning_rate": 4.584766062295698e-05, + "loss": 0.5109, + "step": 16112 + }, + { + "epoch": 1.081406664205899, + "grad_norm": 1.0430505275726318, + "learning_rate": 4.583682995766003e-05, + "loss": 0.5431, + "step": 16114 + }, + { + "epoch": 1.0815408878896682, + "grad_norm": 1.0224894285202026, + "learning_rate": 4.5825999489064676e-05, + "loss": 0.5609, + "step": 16116 + }, + { + "epoch": 1.081675111573437, + "grad_norm": 1.0758253335952759, + "learning_rate": 4.581516921768263e-05, + "loss": 0.529, + "step": 16118 + }, + { + "epoch": 1.0818093352572062, + "grad_norm": 1.0846209526062012, + "learning_rate": 4.580433914402562e-05, + "loss": 0.5035, + "step": 16120 + }, + { + "epoch": 1.0819435589409752, + "grad_norm": 1.0508264303207397, + "learning_rate": 4.579350926860533e-05, + "loss": 0.5633, + "step": 16122 + }, + { + "epoch": 1.082077782624744, + "grad_norm": 0.9724158048629761, + "learning_rate": 4.5782679591933444e-05, + "loss": 0.4924, + "step": 16124 + }, + { + "epoch": 1.0822120063085132, + "grad_norm": 0.9624977111816406, + "learning_rate": 4.577185011452168e-05, + "loss": 0.5157, + "step": 16126 + }, + { + "epoch": 1.0823462299922821, + "grad_norm": 1.0579854249954224, + "learning_rate": 4.576102083688165e-05, + "loss": 0.4689, + "step": 16128 + }, + { + "epoch": 1.082480453676051, + "grad_norm": 1.0702515840530396, + "learning_rate": 4.575019175952508e-05, + "loss": 0.5714, + "step": 16130 + }, + { + "epoch": 1.0826146773598202, + "grad_norm": 0.9868396520614624, + "learning_rate": 4.573936288296356e-05, + "loss": 0.5183, + "step": 16132 + }, + { + "epoch": 1.0827489010435891, + "grad_norm": 1.0361624956130981, + "learning_rate": 4.57285342077088e-05, + "loss": 0.4942, + "step": 16134 + }, + { + "epoch": 1.082883124727358, + "grad_norm": 1.1705342531204224, + "learning_rate": 4.5717705734272375e-05, + "loss": 0.5181, + "step": 16136 + }, + { + "epoch": 1.0830173484111272, + "grad_norm": 0.8735315203666687, + "learning_rate": 4.570687746316595e-05, + "loss": 0.5328, + "step": 16138 + }, + { + "epoch": 1.083151572094896, + "grad_norm": 0.9336830973625183, + "learning_rate": 4.569604939490113e-05, + "loss": 0.5143, + "step": 16140 + }, + { + "epoch": 1.0832857957786652, + "grad_norm": 1.2821564674377441, + "learning_rate": 4.568522152998948e-05, + "loss": 0.5512, + "step": 16142 + }, + { + "epoch": 1.0834200194624342, + "grad_norm": 1.0232316255569458, + "learning_rate": 4.5674393868942666e-05, + "loss": 0.504, + "step": 16144 + }, + { + "epoch": 1.083554243146203, + "grad_norm": 0.9223902225494385, + "learning_rate": 4.56635664122722e-05, + "loss": 0.5187, + "step": 16146 + }, + { + "epoch": 1.0836884668299722, + "grad_norm": 0.9724869728088379, + "learning_rate": 4.5652739160489715e-05, + "loss": 0.5811, + "step": 16148 + }, + { + "epoch": 1.0838226905137411, + "grad_norm": 1.07927668094635, + "learning_rate": 4.564191211410675e-05, + "loss": 0.5569, + "step": 16150 + }, + { + "epoch": 1.08395691419751, + "grad_norm": 0.8426041603088379, + "learning_rate": 4.563108527363487e-05, + "loss": 0.4821, + "step": 16152 + }, + { + "epoch": 1.0840911378812792, + "grad_norm": 0.8651317954063416, + "learning_rate": 4.562025863958562e-05, + "loss": 0.5163, + "step": 16154 + }, + { + "epoch": 1.0842253615650481, + "grad_norm": 1.1290720701217651, + "learning_rate": 4.5609432212470536e-05, + "loss": 0.6004, + "step": 16156 + }, + { + "epoch": 1.084359585248817, + "grad_norm": 0.8488070368766785, + "learning_rate": 4.5598605992801145e-05, + "loss": 0.5177, + "step": 16158 + }, + { + "epoch": 1.0844938089325862, + "grad_norm": 1.019907832145691, + "learning_rate": 4.5587779981088976e-05, + "loss": 0.5181, + "step": 16160 + }, + { + "epoch": 1.084628032616355, + "grad_norm": 0.998149573802948, + "learning_rate": 4.557695417784551e-05, + "loss": 0.4685, + "step": 16162 + }, + { + "epoch": 1.0847622563001242, + "grad_norm": 1.00678288936615, + "learning_rate": 4.5566128583582276e-05, + "loss": 0.4771, + "step": 16164 + }, + { + "epoch": 1.0848964799838932, + "grad_norm": 0.9872289299964905, + "learning_rate": 4.5555303198810724e-05, + "loss": 0.4704, + "step": 16166 + }, + { + "epoch": 1.085030703667662, + "grad_norm": 1.347158432006836, + "learning_rate": 4.5544478024042384e-05, + "loss": 0.5665, + "step": 16168 + }, + { + "epoch": 1.0851649273514312, + "grad_norm": 2.2895302772521973, + "learning_rate": 4.553365305978867e-05, + "loss": 0.5762, + "step": 16170 + }, + { + "epoch": 1.0852991510352001, + "grad_norm": 0.9067912697792053, + "learning_rate": 4.5522828306561085e-05, + "loss": 0.4757, + "step": 16172 + }, + { + "epoch": 1.0854333747189693, + "grad_norm": 1.1870911121368408, + "learning_rate": 4.551200376487105e-05, + "loss": 0.5064, + "step": 16174 + }, + { + "epoch": 1.0855675984027382, + "grad_norm": 0.9698526859283447, + "learning_rate": 4.5501179435230024e-05, + "loss": 0.5167, + "step": 16176 + }, + { + "epoch": 1.0857018220865071, + "grad_norm": 1.0006768703460693, + "learning_rate": 4.549035531814941e-05, + "loss": 0.4965, + "step": 16178 + }, + { + "epoch": 1.0858360457702763, + "grad_norm": 1.1461081504821777, + "learning_rate": 4.5479531414140654e-05, + "loss": 0.5515, + "step": 16180 + }, + { + "epoch": 1.0859702694540452, + "grad_norm": 0.9489409327507019, + "learning_rate": 4.5468707723715135e-05, + "loss": 0.5092, + "step": 16182 + }, + { + "epoch": 1.086104493137814, + "grad_norm": 1.11704421043396, + "learning_rate": 4.545788424738427e-05, + "loss": 0.5481, + "step": 16184 + }, + { + "epoch": 1.0862387168215832, + "grad_norm": 0.8226884007453918, + "learning_rate": 4.544706098565944e-05, + "loss": 0.5124, + "step": 16186 + }, + { + "epoch": 1.0863729405053522, + "grad_norm": 1.157618522644043, + "learning_rate": 4.543623793905204e-05, + "loss": 0.543, + "step": 16188 + }, + { + "epoch": 1.086507164189121, + "grad_norm": 1.1042851209640503, + "learning_rate": 4.54254151080734e-05, + "loss": 0.5354, + "step": 16190 + }, + { + "epoch": 1.0866413878728902, + "grad_norm": 0.9790997505187988, + "learning_rate": 4.541459249323492e-05, + "loss": 0.5428, + "step": 16192 + }, + { + "epoch": 1.0867756115566591, + "grad_norm": 0.9773795008659363, + "learning_rate": 4.54037700950479e-05, + "loss": 0.4837, + "step": 16194 + }, + { + "epoch": 1.0869098352404283, + "grad_norm": 1.0129271745681763, + "learning_rate": 4.5392947914023745e-05, + "loss": 0.5219, + "step": 16196 + }, + { + "epoch": 1.0870440589241972, + "grad_norm": 0.9165695309638977, + "learning_rate": 4.538212595067371e-05, + "loss": 0.4536, + "step": 16198 + }, + { + "epoch": 1.0871782826079661, + "grad_norm": 0.8758540153503418, + "learning_rate": 4.5371304205509154e-05, + "loss": 0.4542, + "step": 16200 + }, + { + "epoch": 1.0873125062917353, + "grad_norm": 1.1202819347381592, + "learning_rate": 4.5360482679041375e-05, + "loss": 0.5094, + "step": 16202 + }, + { + "epoch": 1.0874467299755042, + "grad_norm": 1.1250008344650269, + "learning_rate": 4.534966137178167e-05, + "loss": 0.5187, + "step": 16204 + }, + { + "epoch": 1.087580953659273, + "grad_norm": 1.0680208206176758, + "learning_rate": 4.533884028424133e-05, + "loss": 0.5682, + "step": 16206 + }, + { + "epoch": 1.0877151773430422, + "grad_norm": 1.0380007028579712, + "learning_rate": 4.5328019416931594e-05, + "loss": 0.4994, + "step": 16208 + }, + { + "epoch": 1.0878494010268112, + "grad_norm": 0.9769343733787537, + "learning_rate": 4.531719877036377e-05, + "loss": 0.4653, + "step": 16210 + }, + { + "epoch": 1.08798362471058, + "grad_norm": 0.9960572123527527, + "learning_rate": 4.530637834504908e-05, + "loss": 0.5058, + "step": 16212 + }, + { + "epoch": 1.0881178483943492, + "grad_norm": 1.506523609161377, + "learning_rate": 4.5295558141498804e-05, + "loss": 0.5711, + "step": 16214 + }, + { + "epoch": 1.0882520720781181, + "grad_norm": 1.0053727626800537, + "learning_rate": 4.528473816022414e-05, + "loss": 0.4923, + "step": 16216 + }, + { + "epoch": 1.0883862957618873, + "grad_norm": 0.912634015083313, + "learning_rate": 4.5273918401736346e-05, + "loss": 0.4597, + "step": 16218 + }, + { + "epoch": 1.0885205194456562, + "grad_norm": 0.9932961463928223, + "learning_rate": 4.5263098866546586e-05, + "loss": 0.5446, + "step": 16220 + }, + { + "epoch": 1.0886547431294251, + "grad_norm": 0.943209171295166, + "learning_rate": 4.525227955516612e-05, + "loss": 0.5379, + "step": 16222 + }, + { + "epoch": 1.0887889668131943, + "grad_norm": 1.0366032123565674, + "learning_rate": 4.524146046810608e-05, + "loss": 0.5927, + "step": 16224 + }, + { + "epoch": 1.0889231904969632, + "grad_norm": 0.9892827868461609, + "learning_rate": 4.523064160587769e-05, + "loss": 0.5305, + "step": 16226 + }, + { + "epoch": 1.089057414180732, + "grad_norm": 1.046505331993103, + "learning_rate": 4.5219822968992096e-05, + "loss": 0.5157, + "step": 16228 + }, + { + "epoch": 1.0891916378645012, + "grad_norm": 0.9887049794197083, + "learning_rate": 4.520900455796049e-05, + "loss": 0.5115, + "step": 16230 + }, + { + "epoch": 1.0893258615482702, + "grad_norm": 1.1368025541305542, + "learning_rate": 4.5198186373293964e-05, + "loss": 0.4765, + "step": 16232 + }, + { + "epoch": 1.089460085232039, + "grad_norm": 0.9224391579627991, + "learning_rate": 4.518736841550373e-05, + "loss": 0.4793, + "step": 16234 + }, + { + "epoch": 1.0895943089158082, + "grad_norm": 1.2007701396942139, + "learning_rate": 4.5176550685100834e-05, + "loss": 0.5259, + "step": 16236 + }, + { + "epoch": 1.0897285325995771, + "grad_norm": 0.918589174747467, + "learning_rate": 4.516573318259646e-05, + "loss": 0.5098, + "step": 16238 + }, + { + "epoch": 1.0898627562833463, + "grad_norm": 0.9159785509109497, + "learning_rate": 4.515491590850167e-05, + "loss": 0.5297, + "step": 16240 + }, + { + "epoch": 1.0899969799671152, + "grad_norm": 1.0669053792953491, + "learning_rate": 4.5144098863327586e-05, + "loss": 0.5326, + "step": 16242 + }, + { + "epoch": 1.0901312036508841, + "grad_norm": 0.9650025367736816, + "learning_rate": 4.513328204758528e-05, + "loss": 0.5523, + "step": 16244 + }, + { + "epoch": 1.0902654273346533, + "grad_norm": 1.2968016862869263, + "learning_rate": 4.5122465461785833e-05, + "loss": 0.5294, + "step": 16246 + }, + { + "epoch": 1.0903996510184222, + "grad_norm": 1.3808320760726929, + "learning_rate": 4.511164910644029e-05, + "loss": 0.5688, + "step": 16248 + }, + { + "epoch": 1.0905338747021913, + "grad_norm": 0.9764894843101501, + "learning_rate": 4.5100832982059724e-05, + "loss": 0.5248, + "step": 16250 + }, + { + "epoch": 1.0906680983859602, + "grad_norm": 1.1142827272415161, + "learning_rate": 4.509001708915516e-05, + "loss": 0.5147, + "step": 16252 + }, + { + "epoch": 1.0908023220697292, + "grad_norm": 0.7973035573959351, + "learning_rate": 4.507920142823764e-05, + "loss": 0.4832, + "step": 16254 + }, + { + "epoch": 1.0909365457534983, + "grad_norm": 0.8636682033538818, + "learning_rate": 4.506838599981816e-05, + "loss": 0.5067, + "step": 16256 + }, + { + "epoch": 1.0910707694372672, + "grad_norm": 0.9508370757102966, + "learning_rate": 4.5057570804407774e-05, + "loss": 0.4957, + "step": 16258 + }, + { + "epoch": 1.0912049931210361, + "grad_norm": 1.341615080833435, + "learning_rate": 4.5046755842517415e-05, + "loss": 0.5189, + "step": 16260 + }, + { + "epoch": 1.0913392168048053, + "grad_norm": 1.0907329320907593, + "learning_rate": 4.503594111465812e-05, + "loss": 0.4795, + "step": 16262 + }, + { + "epoch": 1.0914734404885742, + "grad_norm": 1.1239118576049805, + "learning_rate": 4.5025126621340844e-05, + "loss": 0.5052, + "step": 16264 + }, + { + "epoch": 1.0916076641723431, + "grad_norm": 1.1593395471572876, + "learning_rate": 4.501431236307655e-05, + "loss": 0.5134, + "step": 16266 + }, + { + "epoch": 1.0917418878561123, + "grad_norm": 1.0171058177947998, + "learning_rate": 4.500349834037619e-05, + "loss": 0.581, + "step": 16268 + }, + { + "epoch": 1.0918761115398812, + "grad_norm": 1.9492132663726807, + "learning_rate": 4.499268455375073e-05, + "loss": 0.5196, + "step": 16270 + }, + { + "epoch": 1.0920103352236503, + "grad_norm": 0.9708751440048218, + "learning_rate": 4.498187100371105e-05, + "loss": 0.5253, + "step": 16272 + }, + { + "epoch": 1.0921445589074192, + "grad_norm": 0.8562864661216736, + "learning_rate": 4.497105769076812e-05, + "loss": 0.4797, + "step": 16274 + }, + { + "epoch": 1.0922787825911882, + "grad_norm": 0.9457457661628723, + "learning_rate": 4.496024461543282e-05, + "loss": 0.5226, + "step": 16276 + }, + { + "epoch": 1.0924130062749573, + "grad_norm": 1.0705337524414062, + "learning_rate": 4.494943177821604e-05, + "loss": 0.5482, + "step": 16278 + }, + { + "epoch": 1.0925472299587262, + "grad_norm": 0.9535156488418579, + "learning_rate": 4.493861917962869e-05, + "loss": 0.5088, + "step": 16280 + }, + { + "epoch": 1.0926814536424951, + "grad_norm": 0.9982197284698486, + "learning_rate": 4.492780682018162e-05, + "loss": 0.5319, + "step": 16282 + }, + { + "epoch": 1.0928156773262643, + "grad_norm": 0.9514666795730591, + "learning_rate": 4.4916994700385714e-05, + "loss": 0.4807, + "step": 16284 + }, + { + "epoch": 1.0929499010100332, + "grad_norm": 0.948200523853302, + "learning_rate": 4.4906182820751796e-05, + "loss": 0.4654, + "step": 16286 + }, + { + "epoch": 1.0930841246938021, + "grad_norm": 0.9668101072311401, + "learning_rate": 4.489537118179074e-05, + "loss": 0.5109, + "step": 16288 + }, + { + "epoch": 1.0932183483775713, + "grad_norm": 1.168709635734558, + "learning_rate": 4.488455978401334e-05, + "loss": 0.5406, + "step": 16290 + }, + { + "epoch": 1.0933525720613402, + "grad_norm": 0.963259220123291, + "learning_rate": 4.4873748627930455e-05, + "loss": 0.5907, + "step": 16292 + }, + { + "epoch": 1.0934867957451093, + "grad_norm": 0.9700908660888672, + "learning_rate": 4.4862937714052835e-05, + "loss": 0.548, + "step": 16294 + }, + { + "epoch": 1.0936210194288782, + "grad_norm": 0.9730653762817383, + "learning_rate": 4.4852127042891354e-05, + "loss": 0.5575, + "step": 16296 + }, + { + "epoch": 1.0937552431126472, + "grad_norm": 0.8968258500099182, + "learning_rate": 4.48413166149567e-05, + "loss": 0.4486, + "step": 16298 + }, + { + "epoch": 1.0938894667964163, + "grad_norm": 0.8358473181724548, + "learning_rate": 4.483050643075972e-05, + "loss": 0.4455, + "step": 16300 + }, + { + "epoch": 1.0940236904801852, + "grad_norm": 0.9382853507995605, + "learning_rate": 4.481969649081112e-05, + "loss": 0.5231, + "step": 16302 + }, + { + "epoch": 1.0941579141639541, + "grad_norm": 0.9947872161865234, + "learning_rate": 4.4808886795621705e-05, + "loss": 0.5522, + "step": 16304 + }, + { + "epoch": 1.0942921378477233, + "grad_norm": 1.0441114902496338, + "learning_rate": 4.479807734570216e-05, + "loss": 0.5215, + "step": 16306 + }, + { + "epoch": 1.0944263615314922, + "grad_norm": 0.883138120174408, + "learning_rate": 4.478726814156325e-05, + "loss": 0.5197, + "step": 16308 + }, + { + "epoch": 1.0945605852152611, + "grad_norm": 0.8907055854797363, + "learning_rate": 4.4776459183715654e-05, + "loss": 0.4668, + "step": 16310 + }, + { + "epoch": 1.0946948088990303, + "grad_norm": 1.0268926620483398, + "learning_rate": 4.4765650472670094e-05, + "loss": 0.4907, + "step": 16312 + }, + { + "epoch": 1.0948290325827992, + "grad_norm": 0.9086816310882568, + "learning_rate": 4.475484200893725e-05, + "loss": 0.4902, + "step": 16314 + }, + { + "epoch": 1.0949632562665683, + "grad_norm": 0.9753834009170532, + "learning_rate": 4.474403379302782e-05, + "loss": 0.4988, + "step": 16316 + }, + { + "epoch": 1.0950974799503372, + "grad_norm": 0.960116982460022, + "learning_rate": 4.473322582545244e-05, + "loss": 0.5771, + "step": 16318 + }, + { + "epoch": 1.0952317036341062, + "grad_norm": 1.0215414762496948, + "learning_rate": 4.472241810672181e-05, + "loss": 0.5681, + "step": 16320 + }, + { + "epoch": 1.0953659273178753, + "grad_norm": 1.0304012298583984, + "learning_rate": 4.471161063734652e-05, + "loss": 0.5528, + "step": 16322 + }, + { + "epoch": 1.0955001510016442, + "grad_norm": 1.0913125276565552, + "learning_rate": 4.470080341783726e-05, + "loss": 0.5906, + "step": 16324 + }, + { + "epoch": 1.0956343746854131, + "grad_norm": 1.086877465248108, + "learning_rate": 4.4689996448704574e-05, + "loss": 0.5078, + "step": 16326 + }, + { + "epoch": 1.0957685983691823, + "grad_norm": 0.9057826995849609, + "learning_rate": 4.4679189730459144e-05, + "loss": 0.5856, + "step": 16328 + }, + { + "epoch": 1.0959028220529512, + "grad_norm": 0.9654415845870972, + "learning_rate": 4.466838326361152e-05, + "loss": 0.5095, + "step": 16330 + }, + { + "epoch": 1.0960370457367203, + "grad_norm": 0.9577851295471191, + "learning_rate": 4.4657577048672324e-05, + "loss": 0.5384, + "step": 16332 + }, + { + "epoch": 1.0961712694204893, + "grad_norm": 0.8881108164787292, + "learning_rate": 4.464677108615209e-05, + "loss": 0.49, + "step": 16334 + }, + { + "epoch": 1.0963054931042582, + "grad_norm": 0.9594892263412476, + "learning_rate": 4.463596537656141e-05, + "loss": 0.5315, + "step": 16336 + }, + { + "epoch": 1.0964397167880273, + "grad_norm": 1.1049425601959229, + "learning_rate": 4.4625159920410804e-05, + "loss": 0.5076, + "step": 16338 + }, + { + "epoch": 1.0965739404717962, + "grad_norm": 1.0198605060577393, + "learning_rate": 4.4614354718210835e-05, + "loss": 0.5602, + "step": 16340 + }, + { + "epoch": 1.0967081641555652, + "grad_norm": 1.053015112876892, + "learning_rate": 4.4603549770472e-05, + "loss": 0.5425, + "step": 16342 + }, + { + "epoch": 1.0968423878393343, + "grad_norm": 1.0289908647537231, + "learning_rate": 4.459274507770484e-05, + "loss": 0.5039, + "step": 16344 + }, + { + "epoch": 1.0969766115231032, + "grad_norm": 0.9137513637542725, + "learning_rate": 4.4581940640419836e-05, + "loss": 0.4879, + "step": 16346 + }, + { + "epoch": 1.0971108352068724, + "grad_norm": 0.9542897343635559, + "learning_rate": 4.457113645912748e-05, + "loss": 0.5342, + "step": 16348 + }, + { + "epoch": 1.0972450588906413, + "grad_norm": 1.230817198753357, + "learning_rate": 4.456033253433826e-05, + "loss": 0.534, + "step": 16350 + }, + { + "epoch": 1.0973792825744102, + "grad_norm": 1.0108160972595215, + "learning_rate": 4.4549528866562614e-05, + "loss": 0.5424, + "step": 16352 + }, + { + "epoch": 1.0975135062581793, + "grad_norm": 1.118992567062378, + "learning_rate": 4.453872545631104e-05, + "loss": 0.5265, + "step": 16354 + }, + { + "epoch": 1.0976477299419483, + "grad_norm": 1.1093859672546387, + "learning_rate": 4.452792230409392e-05, + "loss": 0.5922, + "step": 16356 + }, + { + "epoch": 1.0977819536257172, + "grad_norm": 0.9274280071258545, + "learning_rate": 4.451711941042173e-05, + "loss": 0.4781, + "step": 16358 + }, + { + "epoch": 1.0979161773094863, + "grad_norm": 0.9606037735939026, + "learning_rate": 4.4506316775804844e-05, + "loss": 0.4916, + "step": 16360 + }, + { + "epoch": 1.0980504009932552, + "grad_norm": 0.9684455394744873, + "learning_rate": 4.449551440075371e-05, + "loss": 0.4674, + "step": 16362 + }, + { + "epoch": 1.0981846246770242, + "grad_norm": 0.9659847021102905, + "learning_rate": 4.448471228577868e-05, + "loss": 0.4583, + "step": 16364 + }, + { + "epoch": 1.0983188483607933, + "grad_norm": 0.9780763387680054, + "learning_rate": 4.447391043139016e-05, + "loss": 0.4875, + "step": 16366 + }, + { + "epoch": 1.0984530720445622, + "grad_norm": 0.9492866396903992, + "learning_rate": 4.4463108838098504e-05, + "loss": 0.5307, + "step": 16368 + }, + { + "epoch": 1.0985872957283314, + "grad_norm": 1.2335330247879028, + "learning_rate": 4.445230750641407e-05, + "loss": 0.5269, + "step": 16370 + }, + { + "epoch": 1.0987215194121003, + "grad_norm": 1.0171606540679932, + "learning_rate": 4.4441506436847194e-05, + "loss": 0.6064, + "step": 16372 + }, + { + "epoch": 1.0988557430958692, + "grad_norm": 0.92994624376297, + "learning_rate": 4.443070562990821e-05, + "loss": 0.4774, + "step": 16374 + }, + { + "epoch": 1.0989899667796383, + "grad_norm": 0.9542266130447388, + "learning_rate": 4.441990508610743e-05, + "loss": 0.531, + "step": 16376 + }, + { + "epoch": 1.0991241904634073, + "grad_norm": 0.9981623888015747, + "learning_rate": 4.440910480595517e-05, + "loss": 0.4934, + "step": 16378 + }, + { + "epoch": 1.0992584141471762, + "grad_norm": 1.0491749048233032, + "learning_rate": 4.439830478996169e-05, + "loss": 0.4824, + "step": 16380 + }, + { + "epoch": 1.0993926378309453, + "grad_norm": 1.1588420867919922, + "learning_rate": 4.438750503863733e-05, + "loss": 0.5313, + "step": 16382 + }, + { + "epoch": 1.0995268615147142, + "grad_norm": 1.062170147895813, + "learning_rate": 4.437670555249228e-05, + "loss": 0.5637, + "step": 16384 + }, + { + "epoch": 1.0996610851984832, + "grad_norm": 1.0424307584762573, + "learning_rate": 4.436590633203688e-05, + "loss": 0.5015, + "step": 16386 + }, + { + "epoch": 1.0997953088822523, + "grad_norm": 0.9287711977958679, + "learning_rate": 4.4355107377781286e-05, + "loss": 0.5714, + "step": 16388 + }, + { + "epoch": 1.0999295325660212, + "grad_norm": 1.1476563215255737, + "learning_rate": 4.434430869023579e-05, + "loss": 0.5265, + "step": 16390 + }, + { + "epoch": 1.1000637562497904, + "grad_norm": 0.9120633006095886, + "learning_rate": 4.433351026991058e-05, + "loss": 0.5154, + "step": 16392 + }, + { + "epoch": 1.1001979799335593, + "grad_norm": 1.005014419555664, + "learning_rate": 4.4322712117315874e-05, + "loss": 0.5392, + "step": 16394 + }, + { + "epoch": 1.1003322036173282, + "grad_norm": 0.8479301333427429, + "learning_rate": 4.4311914232961847e-05, + "loss": 0.433, + "step": 16396 + }, + { + "epoch": 1.1004664273010973, + "grad_norm": 2.0558910369873047, + "learning_rate": 4.4301116617358695e-05, + "loss": 0.4687, + "step": 16398 + }, + { + "epoch": 1.1006006509848663, + "grad_norm": 0.9969558119773865, + "learning_rate": 4.429031927101657e-05, + "loss": 0.5175, + "step": 16400 + }, + { + "epoch": 1.1007348746686352, + "grad_norm": 0.9618621468544006, + "learning_rate": 4.4279522194445644e-05, + "loss": 0.4884, + "step": 16402 + }, + { + "epoch": 1.1008690983524043, + "grad_norm": 1.015475869178772, + "learning_rate": 4.4268725388156026e-05, + "loss": 0.5621, + "step": 16404 + }, + { + "epoch": 1.1010033220361732, + "grad_norm": 1.1162946224212646, + "learning_rate": 4.425792885265789e-05, + "loss": 0.5391, + "step": 16406 + }, + { + "epoch": 1.1011375457199424, + "grad_norm": 1.1964268684387207, + "learning_rate": 4.4247132588461286e-05, + "loss": 0.5575, + "step": 16408 + }, + { + "epoch": 1.1012717694037113, + "grad_norm": 0.8786594867706299, + "learning_rate": 4.423633659607639e-05, + "loss": 0.5216, + "step": 16410 + }, + { + "epoch": 1.1014059930874802, + "grad_norm": 1.0007182359695435, + "learning_rate": 4.422554087601325e-05, + "loss": 0.4838, + "step": 16412 + }, + { + "epoch": 1.1015402167712494, + "grad_norm": 1.1021301746368408, + "learning_rate": 4.421474542878195e-05, + "loss": 0.5776, + "step": 16414 + }, + { + "epoch": 1.1016744404550183, + "grad_norm": 0.9514111876487732, + "learning_rate": 4.4203950254892545e-05, + "loss": 0.4923, + "step": 16416 + }, + { + "epoch": 1.1018086641387872, + "grad_norm": 1.0056926012039185, + "learning_rate": 4.4193155354855074e-05, + "loss": 0.4776, + "step": 16418 + }, + { + "epoch": 1.1019428878225563, + "grad_norm": 1.1442005634307861, + "learning_rate": 4.4182360729179625e-05, + "loss": 0.4783, + "step": 16420 + }, + { + "epoch": 1.1020771115063253, + "grad_norm": 1.4579870700836182, + "learning_rate": 4.4171566378376166e-05, + "loss": 0.4795, + "step": 16422 + }, + { + "epoch": 1.1022113351900944, + "grad_norm": 1.0166319608688354, + "learning_rate": 4.416077230295475e-05, + "loss": 0.4865, + "step": 16424 + }, + { + "epoch": 1.1023455588738633, + "grad_norm": 1.166882872581482, + "learning_rate": 4.414997850342536e-05, + "loss": 0.5292, + "step": 16426 + }, + { + "epoch": 1.1024797825576322, + "grad_norm": 0.9497777819633484, + "learning_rate": 4.4139184980297985e-05, + "loss": 0.4938, + "step": 16428 + }, + { + "epoch": 1.1026140062414014, + "grad_norm": 0.9077328443527222, + "learning_rate": 4.4128391734082586e-05, + "loss": 0.4966, + "step": 16430 + }, + { + "epoch": 1.1027482299251703, + "grad_norm": 0.9687386155128479, + "learning_rate": 4.4117598765289145e-05, + "loss": 0.5106, + "step": 16432 + }, + { + "epoch": 1.1028824536089392, + "grad_norm": 1.446250081062317, + "learning_rate": 4.410680607442758e-05, + "loss": 0.5049, + "step": 16434 + }, + { + "epoch": 1.1030166772927084, + "grad_norm": 0.9839146137237549, + "learning_rate": 4.4096013662007844e-05, + "loss": 0.4981, + "step": 16436 + }, + { + "epoch": 1.1031509009764773, + "grad_norm": 0.9283409714698792, + "learning_rate": 4.408522152853985e-05, + "loss": 0.5569, + "step": 16438 + }, + { + "epoch": 1.1032851246602462, + "grad_norm": 0.952659547328949, + "learning_rate": 4.407442967453352e-05, + "loss": 0.538, + "step": 16440 + }, + { + "epoch": 1.1034193483440153, + "grad_norm": 1.0071378946304321, + "learning_rate": 4.406363810049871e-05, + "loss": 0.5718, + "step": 16442 + }, + { + "epoch": 1.1035535720277843, + "grad_norm": 1.1124968528747559, + "learning_rate": 4.405284680694536e-05, + "loss": 0.5117, + "step": 16444 + }, + { + "epoch": 1.1036877957115534, + "grad_norm": 0.8670393824577332, + "learning_rate": 4.4042055794383277e-05, + "loss": 0.4914, + "step": 16446 + }, + { + "epoch": 1.1038220193953223, + "grad_norm": 1.0304452180862427, + "learning_rate": 4.4031265063322364e-05, + "loss": 0.5183, + "step": 16448 + }, + { + "epoch": 1.1039562430790912, + "grad_norm": 0.9968014359474182, + "learning_rate": 4.402047461427241e-05, + "loss": 0.508, + "step": 16450 + }, + { + "epoch": 1.1040904667628604, + "grad_norm": 1.0950051546096802, + "learning_rate": 4.40096844477433e-05, + "loss": 0.545, + "step": 16452 + }, + { + "epoch": 1.1042246904466293, + "grad_norm": 0.7988759279251099, + "learning_rate": 4.399889456424481e-05, + "loss": 0.4289, + "step": 16454 + }, + { + "epoch": 1.1043589141303982, + "grad_norm": 1.0534387826919556, + "learning_rate": 4.398810496428675e-05, + "loss": 0.5391, + "step": 16456 + }, + { + "epoch": 1.1044931378141674, + "grad_norm": 1.1303116083145142, + "learning_rate": 4.397731564837891e-05, + "loss": 0.5349, + "step": 16458 + }, + { + "epoch": 1.1046273614979363, + "grad_norm": 1.1211031675338745, + "learning_rate": 4.396652661703107e-05, + "loss": 0.5482, + "step": 16460 + }, + { + "epoch": 1.1047615851817052, + "grad_norm": 1.0421147346496582, + "learning_rate": 4.395573787075297e-05, + "loss": 0.6276, + "step": 16462 + }, + { + "epoch": 1.1048958088654743, + "grad_norm": 1.1967524290084839, + "learning_rate": 4.394494941005438e-05, + "loss": 0.5523, + "step": 16464 + }, + { + "epoch": 1.1050300325492433, + "grad_norm": 0.9994180202484131, + "learning_rate": 4.3934161235445015e-05, + "loss": 0.5339, + "step": 16466 + }, + { + "epoch": 1.1051642562330124, + "grad_norm": 0.9488910436630249, + "learning_rate": 4.392337334743461e-05, + "loss": 0.4396, + "step": 16468 + }, + { + "epoch": 1.1052984799167813, + "grad_norm": 0.9103826284408569, + "learning_rate": 4.391258574653285e-05, + "loss": 0.4677, + "step": 16470 + }, + { + "epoch": 1.1054327036005502, + "grad_norm": 0.8869611620903015, + "learning_rate": 4.390179843324947e-05, + "loss": 0.4818, + "step": 16472 + }, + { + "epoch": 1.1055669272843194, + "grad_norm": 0.9392869472503662, + "learning_rate": 4.3891011408094105e-05, + "loss": 0.5036, + "step": 16474 + }, + { + "epoch": 1.1057011509680883, + "grad_norm": 0.919829785823822, + "learning_rate": 4.3880224671576455e-05, + "loss": 0.4985, + "step": 16476 + }, + { + "epoch": 1.1058353746518572, + "grad_norm": 1.0271228551864624, + "learning_rate": 4.3869438224206126e-05, + "loss": 0.587, + "step": 16478 + }, + { + "epoch": 1.1059695983356264, + "grad_norm": 1.1444569826126099, + "learning_rate": 4.3858652066492814e-05, + "loss": 0.5378, + "step": 16480 + }, + { + "epoch": 1.1061038220193953, + "grad_norm": 0.974545955657959, + "learning_rate": 4.3847866198946116e-05, + "loss": 0.4844, + "step": 16482 + }, + { + "epoch": 1.1062380457031644, + "grad_norm": 1.0419704914093018, + "learning_rate": 4.383708062207562e-05, + "loss": 0.5338, + "step": 16484 + }, + { + "epoch": 1.1063722693869333, + "grad_norm": 3.126567840576172, + "learning_rate": 4.382629533639095e-05, + "loss": 0.5039, + "step": 16486 + }, + { + "epoch": 1.1065064930707023, + "grad_norm": 0.9070227146148682, + "learning_rate": 4.381551034240169e-05, + "loss": 0.4685, + "step": 16488 + }, + { + "epoch": 1.1066407167544714, + "grad_norm": 1.186069130897522, + "learning_rate": 4.3804725640617414e-05, + "loss": 0.4914, + "step": 16490 + }, + { + "epoch": 1.1067749404382403, + "grad_norm": 0.9902550578117371, + "learning_rate": 4.379394123154766e-05, + "loss": 0.5757, + "step": 16492 + }, + { + "epoch": 1.1069091641220092, + "grad_norm": 0.9856970906257629, + "learning_rate": 4.3783157115701984e-05, + "loss": 0.5759, + "step": 16494 + }, + { + "epoch": 1.1070433878057784, + "grad_norm": 1.0601118803024292, + "learning_rate": 4.3772373293589894e-05, + "loss": 0.5112, + "step": 16496 + }, + { + "epoch": 1.1071776114895473, + "grad_norm": 1.0865708589553833, + "learning_rate": 4.376158976572094e-05, + "loss": 0.5128, + "step": 16498 + }, + { + "epoch": 1.1073118351733164, + "grad_norm": 1.0044316053390503, + "learning_rate": 4.375080653260459e-05, + "loss": 0.4788, + "step": 16500 + }, + { + "epoch": 1.1074460588570854, + "grad_norm": 0.95487380027771, + "learning_rate": 4.374002359475036e-05, + "loss": 0.4666, + "step": 16502 + }, + { + "epoch": 1.1075802825408543, + "grad_norm": 0.9853495955467224, + "learning_rate": 4.3729240952667684e-05, + "loss": 0.4803, + "step": 16504 + }, + { + "epoch": 1.1077145062246234, + "grad_norm": 1.0185718536376953, + "learning_rate": 4.3718458606866056e-05, + "loss": 0.4928, + "step": 16506 + }, + { + "epoch": 1.1078487299083923, + "grad_norm": 1.0792577266693115, + "learning_rate": 4.370767655785488e-05, + "loss": 0.5683, + "step": 16508 + }, + { + "epoch": 1.1079829535921613, + "grad_norm": 1.0750395059585571, + "learning_rate": 4.3696894806143653e-05, + "loss": 0.49, + "step": 16510 + }, + { + "epoch": 1.1081171772759304, + "grad_norm": 0.9474500417709351, + "learning_rate": 4.368611335224172e-05, + "loss": 0.5574, + "step": 16512 + }, + { + "epoch": 1.1082514009596993, + "grad_norm": 1.19668447971344, + "learning_rate": 4.367533219665853e-05, + "loss": 0.5625, + "step": 16514 + }, + { + "epoch": 1.1083856246434682, + "grad_norm": 0.8541036248207092, + "learning_rate": 4.3664551339903454e-05, + "loss": 0.4387, + "step": 16516 + }, + { + "epoch": 1.1085198483272374, + "grad_norm": 1.0199828147888184, + "learning_rate": 4.3653770782485876e-05, + "loss": 0.5202, + "step": 16518 + }, + { + "epoch": 1.1086540720110063, + "grad_norm": 1.1608493328094482, + "learning_rate": 4.364299052491514e-05, + "loss": 0.4858, + "step": 16520 + }, + { + "epoch": 1.1087882956947754, + "grad_norm": 0.9933175444602966, + "learning_rate": 4.36322105677006e-05, + "loss": 0.4959, + "step": 16522 + }, + { + "epoch": 1.1089225193785444, + "grad_norm": 1.8374931812286377, + "learning_rate": 4.36214309113516e-05, + "loss": 0.4993, + "step": 16524 + }, + { + "epoch": 1.1090567430623133, + "grad_norm": 1.0183755159378052, + "learning_rate": 4.361065155637744e-05, + "loss": 0.5429, + "step": 16526 + }, + { + "epoch": 1.1091909667460824, + "grad_norm": 1.0194748640060425, + "learning_rate": 4.359987250328743e-05, + "loss": 0.5263, + "step": 16528 + }, + { + "epoch": 1.1093251904298513, + "grad_norm": 1.099562644958496, + "learning_rate": 4.358909375259086e-05, + "loss": 0.5132, + "step": 16530 + }, + { + "epoch": 1.1094594141136203, + "grad_norm": 0.9151814579963684, + "learning_rate": 4.357831530479699e-05, + "loss": 0.4958, + "step": 16532 + }, + { + "epoch": 1.1095936377973894, + "grad_norm": 0.92188560962677, + "learning_rate": 4.3567537160415125e-05, + "loss": 0.4762, + "step": 16534 + }, + { + "epoch": 1.1097278614811583, + "grad_norm": 1.0499858856201172, + "learning_rate": 4.355675931995444e-05, + "loss": 0.5955, + "step": 16536 + }, + { + "epoch": 1.1098620851649272, + "grad_norm": 1.0483999252319336, + "learning_rate": 4.354598178392425e-05, + "loss": 0.5233, + "step": 16538 + }, + { + "epoch": 1.1099963088486964, + "grad_norm": 1.0125541687011719, + "learning_rate": 4.353520455283369e-05, + "loss": 0.5648, + "step": 16540 + }, + { + "epoch": 1.1101305325324653, + "grad_norm": 0.927888035774231, + "learning_rate": 4.3524427627192024e-05, + "loss": 0.4779, + "step": 16542 + }, + { + "epoch": 1.1102647562162344, + "grad_norm": 0.8777483701705933, + "learning_rate": 4.3513651007508414e-05, + "loss": 0.4826, + "step": 16544 + }, + { + "epoch": 1.1103989799000034, + "grad_norm": 0.905463457107544, + "learning_rate": 4.3502874694292045e-05, + "loss": 0.514, + "step": 16546 + }, + { + "epoch": 1.1105332035837723, + "grad_norm": 1.000179648399353, + "learning_rate": 4.3492098688052064e-05, + "loss": 0.494, + "step": 16548 + }, + { + "epoch": 1.1106674272675414, + "grad_norm": 0.9666675925254822, + "learning_rate": 4.3481322989297626e-05, + "loss": 0.5082, + "step": 16550 + }, + { + "epoch": 1.1108016509513103, + "grad_norm": 1.0104807615280151, + "learning_rate": 4.347054759853787e-05, + "loss": 0.4707, + "step": 16552 + }, + { + "epoch": 1.1109358746350793, + "grad_norm": 2.4930474758148193, + "learning_rate": 4.3459772516281886e-05, + "loss": 0.5219, + "step": 16554 + }, + { + "epoch": 1.1110700983188484, + "grad_norm": 0.97373366355896, + "learning_rate": 4.344899774303879e-05, + "loss": 0.4301, + "step": 16556 + }, + { + "epoch": 1.1112043220026173, + "grad_norm": 0.9810713529586792, + "learning_rate": 4.343822327931767e-05, + "loss": 0.5474, + "step": 16558 + }, + { + "epoch": 1.1113385456863865, + "grad_norm": 1.073265790939331, + "learning_rate": 4.34274491256276e-05, + "loss": 0.5042, + "step": 16560 + }, + { + "epoch": 1.1114727693701554, + "grad_norm": 1.1238216161727905, + "learning_rate": 4.3416675282477626e-05, + "loss": 0.513, + "step": 16562 + }, + { + "epoch": 1.1116069930539243, + "grad_norm": 0.9564653038978577, + "learning_rate": 4.340590175037681e-05, + "loss": 0.5194, + "step": 16564 + }, + { + "epoch": 1.1117412167376934, + "grad_norm": 1.396801233291626, + "learning_rate": 4.339512852983416e-05, + "loss": 0.5107, + "step": 16566 + }, + { + "epoch": 1.1118754404214624, + "grad_norm": 1.0099092721939087, + "learning_rate": 4.3384355621358706e-05, + "loss": 0.4657, + "step": 16568 + }, + { + "epoch": 1.1120096641052313, + "grad_norm": 0.8993440270423889, + "learning_rate": 4.3373583025459426e-05, + "loss": 0.5107, + "step": 16570 + }, + { + "epoch": 1.1121438877890004, + "grad_norm": 1.0395410060882568, + "learning_rate": 4.3362810742645344e-05, + "loss": 0.5569, + "step": 16572 + }, + { + "epoch": 1.1122781114727693, + "grad_norm": 1.009219765663147, + "learning_rate": 4.3352038773425376e-05, + "loss": 0.4735, + "step": 16574 + }, + { + "epoch": 1.1124123351565385, + "grad_norm": 0.9954773783683777, + "learning_rate": 4.334126711830852e-05, + "loss": 0.4676, + "step": 16576 + }, + { + "epoch": 1.1125465588403074, + "grad_norm": 1.03559148311615, + "learning_rate": 4.3330495777803685e-05, + "loss": 0.5046, + "step": 16578 + }, + { + "epoch": 1.1126807825240763, + "grad_norm": 0.972273588180542, + "learning_rate": 4.3319724752419825e-05, + "loss": 0.4735, + "step": 16580 + }, + { + "epoch": 1.1128150062078455, + "grad_norm": 0.8986105918884277, + "learning_rate": 4.330895404266583e-05, + "loss": 0.4736, + "step": 16582 + }, + { + "epoch": 1.1129492298916144, + "grad_norm": 1.0649678707122803, + "learning_rate": 4.32981836490506e-05, + "loss": 0.5084, + "step": 16584 + }, + { + "epoch": 1.1130834535753833, + "grad_norm": 1.1590815782546997, + "learning_rate": 4.3287413572083e-05, + "loss": 0.4913, + "step": 16586 + }, + { + "epoch": 1.1132176772591524, + "grad_norm": 1.0562185049057007, + "learning_rate": 4.327664381227193e-05, + "loss": 0.5497, + "step": 16588 + }, + { + "epoch": 1.1133519009429214, + "grad_norm": 0.97037672996521, + "learning_rate": 4.32658743701262e-05, + "loss": 0.5467, + "step": 16590 + }, + { + "epoch": 1.1134861246266903, + "grad_norm": 1.028220772743225, + "learning_rate": 4.325510524615468e-05, + "loss": 0.472, + "step": 16592 + }, + { + "epoch": 1.1136203483104594, + "grad_norm": 1.1340383291244507, + "learning_rate": 4.324433644086617e-05, + "loss": 0.5538, + "step": 16594 + }, + { + "epoch": 1.1137545719942283, + "grad_norm": 1.131611704826355, + "learning_rate": 4.3233567954769475e-05, + "loss": 0.4957, + "step": 16596 + }, + { + "epoch": 1.1138887956779975, + "grad_norm": 1.2661018371582031, + "learning_rate": 4.3222799788373374e-05, + "loss": 0.4914, + "step": 16598 + }, + { + "epoch": 1.1140230193617664, + "grad_norm": 1.0229700803756714, + "learning_rate": 4.3212031942186694e-05, + "loss": 0.5814, + "step": 16600 + }, + { + "epoch": 1.1141572430455353, + "grad_norm": 0.9082956314086914, + "learning_rate": 4.320126441671811e-05, + "loss": 0.5003, + "step": 16602 + }, + { + "epoch": 1.1142914667293045, + "grad_norm": 1.0306313037872314, + "learning_rate": 4.3190497212476444e-05, + "loss": 0.4976, + "step": 16604 + }, + { + "epoch": 1.1144256904130734, + "grad_norm": 0.9325722455978394, + "learning_rate": 4.317973032997038e-05, + "loss": 0.4844, + "step": 16606 + }, + { + "epoch": 1.1145599140968423, + "grad_norm": 1.0862069129943848, + "learning_rate": 4.316896376970866e-05, + "loss": 0.5539, + "step": 16608 + }, + { + "epoch": 1.1146941377806114, + "grad_norm": 0.9598913192749023, + "learning_rate": 4.3158197532199964e-05, + "loss": 0.4889, + "step": 16610 + }, + { + "epoch": 1.1148283614643804, + "grad_norm": 1.1408271789550781, + "learning_rate": 4.314743161795298e-05, + "loss": 0.5388, + "step": 16612 + }, + { + "epoch": 1.1149625851481493, + "grad_norm": 0.9544355273246765, + "learning_rate": 4.313666602747638e-05, + "loss": 0.5197, + "step": 16614 + }, + { + "epoch": 1.1150968088319184, + "grad_norm": 1.1572375297546387, + "learning_rate": 4.3125900761278817e-05, + "loss": 0.4982, + "step": 16616 + }, + { + "epoch": 1.1152310325156873, + "grad_norm": 0.9330697655677795, + "learning_rate": 4.3115135819868925e-05, + "loss": 0.4749, + "step": 16618 + }, + { + "epoch": 1.1153652561994565, + "grad_norm": 0.9616682529449463, + "learning_rate": 4.3104371203755315e-05, + "loss": 0.4508, + "step": 16620 + }, + { + "epoch": 1.1154994798832254, + "grad_norm": 1.175607681274414, + "learning_rate": 4.309360691344662e-05, + "loss": 0.5284, + "step": 16622 + }, + { + "epoch": 1.1156337035669943, + "grad_norm": 1.0348509550094604, + "learning_rate": 4.30828429494514e-05, + "loss": 0.4945, + "step": 16624 + }, + { + "epoch": 1.1157679272507635, + "grad_norm": 1.0511419773101807, + "learning_rate": 4.3072079312278266e-05, + "loss": 0.5344, + "step": 16626 + }, + { + "epoch": 1.1159021509345324, + "grad_norm": 1.7016751766204834, + "learning_rate": 4.306131600243575e-05, + "loss": 0.5863, + "step": 16628 + }, + { + "epoch": 1.1160363746183013, + "grad_norm": 1.1384695768356323, + "learning_rate": 4.305055302043242e-05, + "loss": 0.5492, + "step": 16630 + }, + { + "epoch": 1.1161705983020704, + "grad_norm": 1.0545936822891235, + "learning_rate": 4.303979036677677e-05, + "loss": 0.507, + "step": 16632 + }, + { + "epoch": 1.1163048219858394, + "grad_norm": 0.8321439623832703, + "learning_rate": 4.302902804197737e-05, + "loss": 0.4136, + "step": 16634 + }, + { + "epoch": 1.1164390456696085, + "grad_norm": 1.3442049026489258, + "learning_rate": 4.301826604654266e-05, + "loss": 0.5373, + "step": 16636 + }, + { + "epoch": 1.1165732693533774, + "grad_norm": 0.9786088466644287, + "learning_rate": 4.3007504380981165e-05, + "loss": 0.5555, + "step": 16638 + }, + { + "epoch": 1.1167074930371463, + "grad_norm": 1.044503927230835, + "learning_rate": 4.2996743045801334e-05, + "loss": 0.499, + "step": 16640 + }, + { + "epoch": 1.1168417167209155, + "grad_norm": 1.028637170791626, + "learning_rate": 4.298598204151163e-05, + "loss": 0.5545, + "step": 16642 + }, + { + "epoch": 1.1169759404046844, + "grad_norm": 1.0345284938812256, + "learning_rate": 4.297522136862047e-05, + "loss": 0.5329, + "step": 16644 + }, + { + "epoch": 1.1171101640884533, + "grad_norm": 0.9361871480941772, + "learning_rate": 4.2964461027636305e-05, + "loss": 0.5188, + "step": 16646 + }, + { + "epoch": 1.1172443877722225, + "grad_norm": 1.0022510290145874, + "learning_rate": 4.29537010190675e-05, + "loss": 0.5554, + "step": 16648 + }, + { + "epoch": 1.1173786114559914, + "grad_norm": 1.1015814542770386, + "learning_rate": 4.2942941343422485e-05, + "loss": 0.5083, + "step": 16650 + }, + { + "epoch": 1.1175128351397605, + "grad_norm": 1.0244308710098267, + "learning_rate": 4.2932182001209596e-05, + "loss": 0.5631, + "step": 16652 + }, + { + "epoch": 1.1176470588235294, + "grad_norm": 0.9587509632110596, + "learning_rate": 4.2921422992937224e-05, + "loss": 0.4849, + "step": 16654 + }, + { + "epoch": 1.1177812825072984, + "grad_norm": 1.068928837776184, + "learning_rate": 4.291066431911369e-05, + "loss": 0.4937, + "step": 16656 + }, + { + "epoch": 1.1179155061910675, + "grad_norm": 0.9799166321754456, + "learning_rate": 4.2899905980247336e-05, + "loss": 0.4927, + "step": 16658 + }, + { + "epoch": 1.1180497298748364, + "grad_norm": 0.970655620098114, + "learning_rate": 4.2889147976846436e-05, + "loss": 0.5538, + "step": 16660 + }, + { + "epoch": 1.1181839535586053, + "grad_norm": 1.1147160530090332, + "learning_rate": 4.287839030941934e-05, + "loss": 0.5111, + "step": 16662 + }, + { + "epoch": 1.1183181772423745, + "grad_norm": 1.0592046976089478, + "learning_rate": 4.286763297847428e-05, + "loss": 0.6146, + "step": 16664 + }, + { + "epoch": 1.1184524009261434, + "grad_norm": 1.1099165678024292, + "learning_rate": 4.2856875984519545e-05, + "loss": 0.5323, + "step": 16666 + }, + { + "epoch": 1.1185866246099123, + "grad_norm": 0.9070066809654236, + "learning_rate": 4.2846119328063363e-05, + "loss": 0.5229, + "step": 16668 + }, + { + "epoch": 1.1187208482936815, + "grad_norm": 1.1833245754241943, + "learning_rate": 4.283536300961399e-05, + "loss": 0.5909, + "step": 16670 + }, + { + "epoch": 1.1188550719774504, + "grad_norm": 0.9487655162811279, + "learning_rate": 4.282460702967962e-05, + "loss": 0.4418, + "step": 16672 + }, + { + "epoch": 1.1189892956612195, + "grad_norm": 0.9981588125228882, + "learning_rate": 4.2813851388768464e-05, + "loss": 0.5104, + "step": 16674 + }, + { + "epoch": 1.1191235193449884, + "grad_norm": 0.9488283395767212, + "learning_rate": 4.2803096087388687e-05, + "loss": 0.523, + "step": 16676 + }, + { + "epoch": 1.1192577430287574, + "grad_norm": 0.9185371994972229, + "learning_rate": 4.279234112604848e-05, + "loss": 0.5018, + "step": 16678 + }, + { + "epoch": 1.1193919667125265, + "grad_norm": 0.9361635446548462, + "learning_rate": 4.2781586505255965e-05, + "loss": 0.5053, + "step": 16680 + }, + { + "epoch": 1.1195261903962954, + "grad_norm": 1.1580199003219604, + "learning_rate": 4.277083222551931e-05, + "loss": 0.5772, + "step": 16682 + }, + { + "epoch": 1.1196604140800643, + "grad_norm": 1.0087647438049316, + "learning_rate": 4.276007828734661e-05, + "loss": 0.5337, + "step": 16684 + }, + { + "epoch": 1.1197946377638335, + "grad_norm": 0.9543098211288452, + "learning_rate": 4.2749324691245977e-05, + "loss": 0.5617, + "step": 16686 + }, + { + "epoch": 1.1199288614476024, + "grad_norm": 0.9907855987548828, + "learning_rate": 4.27385714377255e-05, + "loss": 0.4678, + "step": 16688 + }, + { + "epoch": 1.1200630851313713, + "grad_norm": 0.9831494092941284, + "learning_rate": 4.2727818527293226e-05, + "loss": 0.489, + "step": 16690 + }, + { + "epoch": 1.1201973088151405, + "grad_norm": 1.1383953094482422, + "learning_rate": 4.271706596045725e-05, + "loss": 0.5409, + "step": 16692 + }, + { + "epoch": 1.1203315324989094, + "grad_norm": 1.3033020496368408, + "learning_rate": 4.270631373772556e-05, + "loss": 0.4914, + "step": 16694 + }, + { + "epoch": 1.1204657561826785, + "grad_norm": 1.1383192539215088, + "learning_rate": 4.269556185960623e-05, + "loss": 0.5127, + "step": 16696 + }, + { + "epoch": 1.1205999798664474, + "grad_norm": 1.1375019550323486, + "learning_rate": 4.2684810326607206e-05, + "loss": 0.5282, + "step": 16698 + }, + { + "epoch": 1.1207342035502164, + "grad_norm": 0.9572421312332153, + "learning_rate": 4.267405913923654e-05, + "loss": 0.4612, + "step": 16700 + }, + { + "epoch": 1.1208684272339855, + "grad_norm": 1.217170238494873, + "learning_rate": 4.2663308298002146e-05, + "loss": 0.5451, + "step": 16702 + }, + { + "epoch": 1.1210026509177544, + "grad_norm": 0.9934802055358887, + "learning_rate": 4.265255780341203e-05, + "loss": 0.4905, + "step": 16704 + }, + { + "epoch": 1.1211368746015233, + "grad_norm": 1.1445648670196533, + "learning_rate": 4.2641807655974096e-05, + "loss": 0.604, + "step": 16706 + }, + { + "epoch": 1.1212710982852925, + "grad_norm": 1.0030955076217651, + "learning_rate": 4.26310578561963e-05, + "loss": 0.5055, + "step": 16708 + }, + { + "epoch": 1.1214053219690614, + "grad_norm": 1.2229998111724854, + "learning_rate": 4.262030840458651e-05, + "loss": 0.6362, + "step": 16710 + }, + { + "epoch": 1.1215395456528305, + "grad_norm": 0.9906282424926758, + "learning_rate": 4.260955930165265e-05, + "loss": 0.4983, + "step": 16712 + }, + { + "epoch": 1.1216737693365995, + "grad_norm": 0.9919167757034302, + "learning_rate": 4.259881054790257e-05, + "loss": 0.5159, + "step": 16714 + }, + { + "epoch": 1.1218079930203684, + "grad_norm": 0.9842211604118347, + "learning_rate": 4.258806214384415e-05, + "loss": 0.5157, + "step": 16716 + }, + { + "epoch": 1.1219422167041375, + "grad_norm": 1.0635043382644653, + "learning_rate": 4.2577314089985204e-05, + "loss": 0.5433, + "step": 16718 + }, + { + "epoch": 1.1220764403879064, + "grad_norm": 1.0264695882797241, + "learning_rate": 4.2566566386833584e-05, + "loss": 0.5344, + "step": 16720 + }, + { + "epoch": 1.1222106640716754, + "grad_norm": 0.9740486145019531, + "learning_rate": 4.255581903489707e-05, + "loss": 0.5472, + "step": 16722 + }, + { + "epoch": 1.1223448877554445, + "grad_norm": 0.8539181351661682, + "learning_rate": 4.254507203468349e-05, + "loss": 0.4583, + "step": 16724 + }, + { + "epoch": 1.1224791114392134, + "grad_norm": 1.0089354515075684, + "learning_rate": 4.253432538670057e-05, + "loss": 0.4837, + "step": 16726 + }, + { + "epoch": 1.1226133351229826, + "grad_norm": 0.9615651965141296, + "learning_rate": 4.252357909145611e-05, + "loss": 0.5052, + "step": 16728 + }, + { + "epoch": 1.1227475588067515, + "grad_norm": 1.0212202072143555, + "learning_rate": 4.251283314945783e-05, + "loss": 0.5558, + "step": 16730 + }, + { + "epoch": 1.1228817824905204, + "grad_norm": 1.0308191776275635, + "learning_rate": 4.2502087561213475e-05, + "loss": 0.4758, + "step": 16732 + }, + { + "epoch": 1.1230160061742895, + "grad_norm": 0.9926826357841492, + "learning_rate": 4.249134232723073e-05, + "loss": 0.5044, + "step": 16734 + }, + { + "epoch": 1.1231502298580585, + "grad_norm": 1.4401243925094604, + "learning_rate": 4.2480597448017294e-05, + "loss": 0.4973, + "step": 16736 + }, + { + "epoch": 1.1232844535418274, + "grad_norm": 0.9778569340705872, + "learning_rate": 4.246985292408084e-05, + "loss": 0.5164, + "step": 16738 + }, + { + "epoch": 1.1234186772255965, + "grad_norm": 0.9825350642204285, + "learning_rate": 4.2459108755929034e-05, + "loss": 0.5223, + "step": 16740 + }, + { + "epoch": 1.1235529009093654, + "grad_norm": 1.7382341623306274, + "learning_rate": 4.244836494406951e-05, + "loss": 0.4909, + "step": 16742 + }, + { + "epoch": 1.1236871245931344, + "grad_norm": 1.0226861238479614, + "learning_rate": 4.2437621489009894e-05, + "loss": 0.5395, + "step": 16744 + }, + { + "epoch": 1.1238213482769035, + "grad_norm": 0.9868961572647095, + "learning_rate": 4.242687839125778e-05, + "loss": 0.5436, + "step": 16746 + }, + { + "epoch": 1.1239555719606724, + "grad_norm": 1.6000854969024658, + "learning_rate": 4.2416135651320785e-05, + "loss": 0.5686, + "step": 16748 + }, + { + "epoch": 1.1240897956444416, + "grad_norm": 0.9828076958656311, + "learning_rate": 4.2405393269706444e-05, + "loss": 0.5938, + "step": 16750 + }, + { + "epoch": 1.1242240193282105, + "grad_norm": 0.9150723814964294, + "learning_rate": 4.239465124692237e-05, + "loss": 0.443, + "step": 16752 + }, + { + "epoch": 1.1243582430119794, + "grad_norm": 0.915022075176239, + "learning_rate": 4.238390958347604e-05, + "loss": 0.4964, + "step": 16754 + }, + { + "epoch": 1.1244924666957485, + "grad_norm": 0.9428569674491882, + "learning_rate": 4.237316827987502e-05, + "loss": 0.4481, + "step": 16756 + }, + { + "epoch": 1.1246266903795175, + "grad_norm": 0.7897739410400391, + "learning_rate": 4.236242733662682e-05, + "loss": 0.5133, + "step": 16758 + }, + { + "epoch": 1.1247609140632864, + "grad_norm": 0.892500638961792, + "learning_rate": 4.235168675423887e-05, + "loss": 0.5001, + "step": 16760 + }, + { + "epoch": 1.1248951377470555, + "grad_norm": 0.92838454246521, + "learning_rate": 4.2340946533218725e-05, + "loss": 0.4946, + "step": 16762 + }, + { + "epoch": 1.1250293614308244, + "grad_norm": 0.9659767150878906, + "learning_rate": 4.233020667407375e-05, + "loss": 0.5315, + "step": 16764 + }, + { + "epoch": 1.1251635851145934, + "grad_norm": 1.129931926727295, + "learning_rate": 4.2319467177311463e-05, + "loss": 0.5051, + "step": 16766 + }, + { + "epoch": 1.1252978087983625, + "grad_norm": 0.9765834212303162, + "learning_rate": 4.2308728043439224e-05, + "loss": 0.4413, + "step": 16768 + }, + { + "epoch": 1.1254320324821314, + "grad_norm": 1.0590866804122925, + "learning_rate": 4.2297989272964476e-05, + "loss": 0.508, + "step": 16770 + }, + { + "epoch": 1.1255662561659006, + "grad_norm": 0.9979098439216614, + "learning_rate": 4.228725086639458e-05, + "loss": 0.5216, + "step": 16772 + }, + { + "epoch": 1.1257004798496695, + "grad_norm": 0.987686812877655, + "learning_rate": 4.227651282423693e-05, + "loss": 0.5149, + "step": 16774 + }, + { + "epoch": 1.1258347035334384, + "grad_norm": 0.9693307876586914, + "learning_rate": 4.2265775146998844e-05, + "loss": 0.5459, + "step": 16776 + }, + { + "epoch": 1.1259689272172075, + "grad_norm": 1.0481395721435547, + "learning_rate": 4.225503783518769e-05, + "loss": 0.5663, + "step": 16778 + }, + { + "epoch": 1.1261031509009765, + "grad_norm": 0.9288170337677002, + "learning_rate": 4.224430088931075e-05, + "loss": 0.493, + "step": 16780 + }, + { + "epoch": 1.1262373745847456, + "grad_norm": 1.0299285650253296, + "learning_rate": 4.223356430987536e-05, + "loss": 0.5827, + "step": 16782 + }, + { + "epoch": 1.1263715982685145, + "grad_norm": 1.0432701110839844, + "learning_rate": 4.222282809738875e-05, + "loss": 0.463, + "step": 16784 + }, + { + "epoch": 1.1265058219522834, + "grad_norm": 1.214779257774353, + "learning_rate": 4.221209225235827e-05, + "loss": 0.5146, + "step": 16786 + }, + { + "epoch": 1.1266400456360526, + "grad_norm": 1.0165801048278809, + "learning_rate": 4.220135677529107e-05, + "loss": 0.487, + "step": 16788 + }, + { + "epoch": 1.1267742693198215, + "grad_norm": 0.9639692306518555, + "learning_rate": 4.219062166669447e-05, + "loss": 0.5206, + "step": 16790 + }, + { + "epoch": 1.1269084930035904, + "grad_norm": 1.0271775722503662, + "learning_rate": 4.21798869270756e-05, + "loss": 0.5052, + "step": 16792 + }, + { + "epoch": 1.1270427166873596, + "grad_norm": 1.0103228092193604, + "learning_rate": 4.216915255694172e-05, + "loss": 0.5101, + "step": 16794 + }, + { + "epoch": 1.1271769403711285, + "grad_norm": 1.2222870588302612, + "learning_rate": 4.215841855679997e-05, + "loss": 0.5024, + "step": 16796 + }, + { + "epoch": 1.1273111640548974, + "grad_norm": 0.9785430431365967, + "learning_rate": 4.214768492715754e-05, + "loss": 0.5736, + "step": 16798 + }, + { + "epoch": 1.1274453877386665, + "grad_norm": 1.064740538597107, + "learning_rate": 4.213695166852153e-05, + "loss": 0.5272, + "step": 16800 + }, + { + "epoch": 1.1275796114224355, + "grad_norm": 1.0236676931381226, + "learning_rate": 4.212621878139912e-05, + "loss": 0.6104, + "step": 16802 + }, + { + "epoch": 1.1277138351062046, + "grad_norm": 0.9074214696884155, + "learning_rate": 4.211548626629737e-05, + "loss": 0.5048, + "step": 16804 + }, + { + "epoch": 1.1278480587899735, + "grad_norm": 0.8793216943740845, + "learning_rate": 4.21047541237234e-05, + "loss": 0.513, + "step": 16806 + }, + { + "epoch": 1.1279822824737424, + "grad_norm": 1.0216366052627563, + "learning_rate": 4.2094022354184266e-05, + "loss": 0.471, + "step": 16808 + }, + { + "epoch": 1.1281165061575116, + "grad_norm": 1.0820353031158447, + "learning_rate": 4.208329095818704e-05, + "loss": 0.5181, + "step": 16810 + }, + { + "epoch": 1.1282507298412805, + "grad_norm": 0.9416481852531433, + "learning_rate": 4.207255993623872e-05, + "loss": 0.552, + "step": 16812 + }, + { + "epoch": 1.1283849535250494, + "grad_norm": 0.9594340920448303, + "learning_rate": 4.206182928884639e-05, + "loss": 0.4678, + "step": 16814 + }, + { + "epoch": 1.1285191772088186, + "grad_norm": 0.9467558264732361, + "learning_rate": 4.205109901651699e-05, + "loss": 0.4932, + "step": 16816 + }, + { + "epoch": 1.1286534008925875, + "grad_norm": 0.9780734181404114, + "learning_rate": 4.204036911975755e-05, + "loss": 0.4848, + "step": 16818 + }, + { + "epoch": 1.1287876245763564, + "grad_norm": 1.0107637643814087, + "learning_rate": 4.2029639599075004e-05, + "loss": 0.5356, + "step": 16820 + }, + { + "epoch": 1.1289218482601255, + "grad_norm": 1.1758426427841187, + "learning_rate": 4.201891045497633e-05, + "loss": 0.5661, + "step": 16822 + }, + { + "epoch": 1.1290560719438945, + "grad_norm": 1.074581503868103, + "learning_rate": 4.200818168796844e-05, + "loss": 0.498, + "step": 16824 + }, + { + "epoch": 1.1291902956276636, + "grad_norm": 1.0495214462280273, + "learning_rate": 4.199745329855823e-05, + "loss": 0.4953, + "step": 16826 + }, + { + "epoch": 1.1293245193114325, + "grad_norm": 1.0303970575332642, + "learning_rate": 4.1986725287252634e-05, + "loss": 0.5049, + "step": 16828 + }, + { + "epoch": 1.1294587429952014, + "grad_norm": 0.956167459487915, + "learning_rate": 4.19759976545585e-05, + "loss": 0.531, + "step": 16830 + }, + { + "epoch": 1.1295929666789706, + "grad_norm": 1.0434627532958984, + "learning_rate": 4.19652704009827e-05, + "loss": 0.5046, + "step": 16832 + }, + { + "epoch": 1.1297271903627395, + "grad_norm": 1.7044645547866821, + "learning_rate": 4.1954543527032076e-05, + "loss": 0.3955, + "step": 16834 + }, + { + "epoch": 1.1298614140465084, + "grad_norm": 1.1061593294143677, + "learning_rate": 4.194381703321346e-05, + "loss": 0.5206, + "step": 16836 + }, + { + "epoch": 1.1299956377302776, + "grad_norm": 0.972310483455658, + "learning_rate": 4.193309092003363e-05, + "loss": 0.5309, + "step": 16838 + }, + { + "epoch": 1.1301298614140465, + "grad_norm": 1.1154252290725708, + "learning_rate": 4.1922365187999404e-05, + "loss": 0.5935, + "step": 16840 + }, + { + "epoch": 1.1302640850978154, + "grad_norm": 1.104824185371399, + "learning_rate": 4.1911639837617526e-05, + "loss": 0.5234, + "step": 16842 + }, + { + "epoch": 1.1303983087815845, + "grad_norm": 1.0318411588668823, + "learning_rate": 4.1900914869394775e-05, + "loss": 0.4876, + "step": 16844 + }, + { + "epoch": 1.1305325324653535, + "grad_norm": 0.9598281383514404, + "learning_rate": 4.189019028383785e-05, + "loss": 0.4643, + "step": 16846 + }, + { + "epoch": 1.1306667561491226, + "grad_norm": 1.2466892004013062, + "learning_rate": 4.187946608145351e-05, + "loss": 0.5751, + "step": 16848 + }, + { + "epoch": 1.1308009798328915, + "grad_norm": 1.0190337896347046, + "learning_rate": 4.18687422627484e-05, + "loss": 0.512, + "step": 16850 + }, + { + "epoch": 1.1309352035166604, + "grad_norm": 1.062680721282959, + "learning_rate": 4.185801882822927e-05, + "loss": 0.524, + "step": 16852 + }, + { + "epoch": 1.1310694272004296, + "grad_norm": 1.2178183794021606, + "learning_rate": 4.184729577840271e-05, + "loss": 0.5675, + "step": 16854 + }, + { + "epoch": 1.1312036508841985, + "grad_norm": 1.0182420015335083, + "learning_rate": 4.18365731137754e-05, + "loss": 0.5095, + "step": 16856 + }, + { + "epoch": 1.1313378745679676, + "grad_norm": 0.9893097877502441, + "learning_rate": 4.1825850834853954e-05, + "loss": 0.476, + "step": 16858 + }, + { + "epoch": 1.1314720982517366, + "grad_norm": 1.0728018283843994, + "learning_rate": 4.181512894214499e-05, + "loss": 0.5199, + "step": 16860 + }, + { + "epoch": 1.1316063219355055, + "grad_norm": 1.1170811653137207, + "learning_rate": 4.180440743615508e-05, + "loss": 0.547, + "step": 16862 + }, + { + "epoch": 1.1317405456192744, + "grad_norm": 1.0509802103042603, + "learning_rate": 4.1793686317390815e-05, + "loss": 0.4813, + "step": 16864 + }, + { + "epoch": 1.1318747693030435, + "grad_norm": 0.9971725344657898, + "learning_rate": 4.178296558635873e-05, + "loss": 0.4501, + "step": 16866 + }, + { + "epoch": 1.1320089929868125, + "grad_norm": 0.957833468914032, + "learning_rate": 4.1772245243565364e-05, + "loss": 0.5182, + "step": 16868 + }, + { + "epoch": 1.1321432166705816, + "grad_norm": 1.008810043334961, + "learning_rate": 4.1761525289517215e-05, + "loss": 0.4993, + "step": 16870 + }, + { + "epoch": 1.1322774403543505, + "grad_norm": 0.941467821598053, + "learning_rate": 4.175080572472082e-05, + "loss": 0.5912, + "step": 16872 + }, + { + "epoch": 1.1324116640381194, + "grad_norm": 1.0539593696594238, + "learning_rate": 4.1740086549682606e-05, + "loss": 0.4971, + "step": 16874 + }, + { + "epoch": 1.1325458877218886, + "grad_norm": 1.0735893249511719, + "learning_rate": 4.172936776490909e-05, + "loss": 0.5694, + "step": 16876 + }, + { + "epoch": 1.1326801114056575, + "grad_norm": 0.9888498783111572, + "learning_rate": 4.171864937090665e-05, + "loss": 0.4602, + "step": 16878 + }, + { + "epoch": 1.1328143350894266, + "grad_norm": 0.9963198900222778, + "learning_rate": 4.170793136818178e-05, + "loss": 0.5226, + "step": 16880 + }, + { + "epoch": 1.1329485587731956, + "grad_norm": 2.036478042602539, + "learning_rate": 4.1697213757240814e-05, + "loss": 0.4859, + "step": 16882 + }, + { + "epoch": 1.1330827824569645, + "grad_norm": 0.9562026858329773, + "learning_rate": 4.168649653859019e-05, + "loss": 0.4906, + "step": 16884 + }, + { + "epoch": 1.1332170061407336, + "grad_norm": 1.1929901838302612, + "learning_rate": 4.1675779712736254e-05, + "loss": 0.5306, + "step": 16886 + }, + { + "epoch": 1.1333512298245025, + "grad_norm": 1.0257091522216797, + "learning_rate": 4.1665063280185354e-05, + "loss": 0.5154, + "step": 16888 + }, + { + "epoch": 1.1334854535082715, + "grad_norm": 1.0640220642089844, + "learning_rate": 4.165434724144383e-05, + "loss": 0.4901, + "step": 16890 + }, + { + "epoch": 1.1336196771920406, + "grad_norm": 1.0287139415740967, + "learning_rate": 4.1643631597017987e-05, + "loss": 0.5058, + "step": 16892 + }, + { + "epoch": 1.1337539008758095, + "grad_norm": 1.0756909847259521, + "learning_rate": 4.1632916347414124e-05, + "loss": 0.4885, + "step": 16894 + }, + { + "epoch": 1.1338881245595784, + "grad_norm": 1.0119127035140991, + "learning_rate": 4.1622201493138495e-05, + "loss": 0.4996, + "step": 16896 + }, + { + "epoch": 1.1340223482433476, + "grad_norm": 1.1204429864883423, + "learning_rate": 4.161148703469739e-05, + "loss": 0.5635, + "step": 16898 + }, + { + "epoch": 1.1341565719271165, + "grad_norm": 1.0327421426773071, + "learning_rate": 4.160077297259701e-05, + "loss": 0.4881, + "step": 16900 + }, + { + "epoch": 1.1342907956108856, + "grad_norm": 1.0640689134597778, + "learning_rate": 4.1590059307343596e-05, + "loss": 0.5416, + "step": 16902 + }, + { + "epoch": 1.1344250192946546, + "grad_norm": 1.1616973876953125, + "learning_rate": 4.1579346039443346e-05, + "loss": 0.4643, + "step": 16904 + }, + { + "epoch": 1.1345592429784235, + "grad_norm": 1.0026849508285522, + "learning_rate": 4.156863316940244e-05, + "loss": 0.5394, + "step": 16906 + }, + { + "epoch": 1.1346934666621926, + "grad_norm": 1.00323486328125, + "learning_rate": 4.155792069772702e-05, + "loss": 0.5543, + "step": 16908 + }, + { + "epoch": 1.1348276903459615, + "grad_norm": 1.1087974309921265, + "learning_rate": 4.1547208624923266e-05, + "loss": 0.5363, + "step": 16910 + }, + { + "epoch": 1.1349619140297305, + "grad_norm": 0.9458073377609253, + "learning_rate": 4.153649695149726e-05, + "loss": 0.5073, + "step": 16912 + }, + { + "epoch": 1.1350961377134996, + "grad_norm": 1.2469216585159302, + "learning_rate": 4.152578567795516e-05, + "loss": 0.5397, + "step": 16914 + }, + { + "epoch": 1.1352303613972685, + "grad_norm": 0.915696382522583, + "learning_rate": 4.1515074804802995e-05, + "loss": 0.5485, + "step": 16916 + }, + { + "epoch": 1.1353645850810374, + "grad_norm": 0.8987810611724854, + "learning_rate": 4.150436433254688e-05, + "loss": 0.4696, + "step": 16918 + }, + { + "epoch": 1.1354988087648066, + "grad_norm": 0.939405620098114, + "learning_rate": 4.1493654261692834e-05, + "loss": 0.494, + "step": 16920 + }, + { + "epoch": 1.1356330324485755, + "grad_norm": 0.9830738306045532, + "learning_rate": 4.148294459274691e-05, + "loss": 0.4773, + "step": 16922 + }, + { + "epoch": 1.1357672561323446, + "grad_norm": 1.1729371547698975, + "learning_rate": 4.147223532621508e-05, + "loss": 0.4847, + "step": 16924 + }, + { + "epoch": 1.1359014798161136, + "grad_norm": 1.1334617137908936, + "learning_rate": 4.146152646260339e-05, + "loss": 0.5156, + "step": 16926 + }, + { + "epoch": 1.1360357034998825, + "grad_norm": 1.2039645910263062, + "learning_rate": 4.1450818002417766e-05, + "loss": 0.5845, + "step": 16928 + }, + { + "epoch": 1.1361699271836516, + "grad_norm": 1.044554591178894, + "learning_rate": 4.1440109946164196e-05, + "loss": 0.4928, + "step": 16930 + }, + { + "epoch": 1.1363041508674205, + "grad_norm": 0.9057065844535828, + "learning_rate": 4.142940229434858e-05, + "loss": 0.46, + "step": 16932 + }, + { + "epoch": 1.1364383745511897, + "grad_norm": 0.9855564832687378, + "learning_rate": 4.141869504747687e-05, + "loss": 0.5202, + "step": 16934 + }, + { + "epoch": 1.1365725982349586, + "grad_norm": 0.9572040438652039, + "learning_rate": 4.140798820605493e-05, + "loss": 0.4946, + "step": 16936 + }, + { + "epoch": 1.1367068219187275, + "grad_norm": 0.9390221238136292, + "learning_rate": 4.139728177058867e-05, + "loss": 0.5045, + "step": 16938 + }, + { + "epoch": 1.1368410456024964, + "grad_norm": 0.9028573632240295, + "learning_rate": 4.1386575741583904e-05, + "loss": 0.5411, + "step": 16940 + }, + { + "epoch": 1.1369752692862656, + "grad_norm": 1.138342261314392, + "learning_rate": 4.1375870119546526e-05, + "loss": 0.5993, + "step": 16942 + }, + { + "epoch": 1.1371094929700345, + "grad_norm": 0.9852227568626404, + "learning_rate": 4.1365164904982296e-05, + "loss": 0.5267, + "step": 16944 + }, + { + "epoch": 1.1372437166538036, + "grad_norm": 0.9474890232086182, + "learning_rate": 4.135446009839707e-05, + "loss": 0.5092, + "step": 16946 + }, + { + "epoch": 1.1373779403375726, + "grad_norm": 0.9510831832885742, + "learning_rate": 4.1343755700296584e-05, + "loss": 0.5578, + "step": 16948 + }, + { + "epoch": 1.1375121640213415, + "grad_norm": 1.0347990989685059, + "learning_rate": 4.1333051711186635e-05, + "loss": 0.555, + "step": 16950 + }, + { + "epoch": 1.1376463877051106, + "grad_norm": 1.0550622940063477, + "learning_rate": 4.132234813157294e-05, + "loss": 0.5184, + "step": 16952 + }, + { + "epoch": 1.1377806113888795, + "grad_norm": 1.0261003971099854, + "learning_rate": 4.1311644961961244e-05, + "loss": 0.523, + "step": 16954 + }, + { + "epoch": 1.1379148350726487, + "grad_norm": 0.9454745054244995, + "learning_rate": 4.1300942202857226e-05, + "loss": 0.5207, + "step": 16956 + }, + { + "epoch": 1.1380490587564176, + "grad_norm": 0.9956864714622498, + "learning_rate": 4.129023985476659e-05, + "loss": 0.564, + "step": 16958 + }, + { + "epoch": 1.1381832824401865, + "grad_norm": 1.1226762533187866, + "learning_rate": 4.127953791819499e-05, + "loss": 0.4966, + "step": 16960 + }, + { + "epoch": 1.1383175061239557, + "grad_norm": 1.004565954208374, + "learning_rate": 4.126883639364808e-05, + "loss": 0.4772, + "step": 16962 + }, + { + "epoch": 1.1384517298077246, + "grad_norm": 1.0866649150848389, + "learning_rate": 4.125813528163148e-05, + "loss": 0.5378, + "step": 16964 + }, + { + "epoch": 1.1385859534914935, + "grad_norm": 0.9746032357215881, + "learning_rate": 4.124743458265079e-05, + "loss": 0.4925, + "step": 16966 + }, + { + "epoch": 1.1387201771752626, + "grad_norm": 0.9077035784721375, + "learning_rate": 4.123673429721162e-05, + "loss": 0.5254, + "step": 16968 + }, + { + "epoch": 1.1388544008590316, + "grad_norm": 1.0778063535690308, + "learning_rate": 4.12260344258195e-05, + "loss": 0.6036, + "step": 16970 + }, + { + "epoch": 1.1389886245428005, + "grad_norm": 1.1656830310821533, + "learning_rate": 4.121533496898002e-05, + "loss": 0.4911, + "step": 16972 + }, + { + "epoch": 1.1391228482265696, + "grad_norm": 1.0001049041748047, + "learning_rate": 4.120463592719867e-05, + "loss": 0.4811, + "step": 16974 + }, + { + "epoch": 1.1392570719103385, + "grad_norm": 0.9660854935646057, + "learning_rate": 4.119393730098101e-05, + "loss": 0.5281, + "step": 16976 + }, + { + "epoch": 1.1393912955941077, + "grad_norm": 0.9911681413650513, + "learning_rate": 4.1183239090832455e-05, + "loss": 0.5248, + "step": 16978 + }, + { + "epoch": 1.1395255192778766, + "grad_norm": 1.0164984464645386, + "learning_rate": 4.117254129725854e-05, + "loss": 0.5099, + "step": 16980 + }, + { + "epoch": 1.1396597429616455, + "grad_norm": 1.0523251295089722, + "learning_rate": 4.116184392076468e-05, + "loss": 0.5631, + "step": 16982 + }, + { + "epoch": 1.1397939666454147, + "grad_norm": 1.0661886930465698, + "learning_rate": 4.1151146961856326e-05, + "loss": 0.5161, + "step": 16984 + }, + { + "epoch": 1.1399281903291836, + "grad_norm": 0.9028160572052002, + "learning_rate": 4.114045042103887e-05, + "loss": 0.4946, + "step": 16986 + }, + { + "epoch": 1.1400624140129525, + "grad_norm": 1.2710407972335815, + "learning_rate": 4.1129754298817714e-05, + "loss": 0.5375, + "step": 16988 + }, + { + "epoch": 1.1401966376967216, + "grad_norm": 1.012258529663086, + "learning_rate": 4.1119058595698217e-05, + "loss": 0.5062, + "step": 16990 + }, + { + "epoch": 1.1403308613804906, + "grad_norm": 0.9147583842277527, + "learning_rate": 4.110836331218575e-05, + "loss": 0.5091, + "step": 16992 + }, + { + "epoch": 1.1404650850642595, + "grad_norm": 1.1260439157485962, + "learning_rate": 4.109766844878562e-05, + "loss": 0.5854, + "step": 16994 + }, + { + "epoch": 1.1405993087480286, + "grad_norm": 0.9493597149848938, + "learning_rate": 4.108697400600316e-05, + "loss": 0.4336, + "step": 16996 + }, + { + "epoch": 1.1407335324317975, + "grad_norm": 0.9938174486160278, + "learning_rate": 4.1076279984343636e-05, + "loss": 0.5017, + "step": 16998 + }, + { + "epoch": 1.1408677561155667, + "grad_norm": 1.0089324712753296, + "learning_rate": 4.106558638431236e-05, + "loss": 0.5049, + "step": 17000 + }, + { + "epoch": 1.1410019797993356, + "grad_norm": 1.0204837322235107, + "learning_rate": 4.105489320641452e-05, + "loss": 0.5761, + "step": 17002 + }, + { + "epoch": 1.1411362034831045, + "grad_norm": 1.0670700073242188, + "learning_rate": 4.104420045115543e-05, + "loss": 0.5781, + "step": 17004 + }, + { + "epoch": 1.1412704271668737, + "grad_norm": 0.95860755443573, + "learning_rate": 4.103350811904022e-05, + "loss": 0.4447, + "step": 17006 + }, + { + "epoch": 1.1414046508506426, + "grad_norm": 0.926114559173584, + "learning_rate": 4.102281621057415e-05, + "loss": 0.4874, + "step": 17008 + }, + { + "epoch": 1.1415388745344117, + "grad_norm": 0.9847375154495239, + "learning_rate": 4.1012124726262345e-05, + "loss": 0.5116, + "step": 17010 + }, + { + "epoch": 1.1416730982181806, + "grad_norm": 1.06489896774292, + "learning_rate": 4.100143366660998e-05, + "loss": 0.5412, + "step": 17012 + }, + { + "epoch": 1.1418073219019496, + "grad_norm": 1.0089995861053467, + "learning_rate": 4.099074303212218e-05, + "loss": 0.523, + "step": 17014 + }, + { + "epoch": 1.1419415455857185, + "grad_norm": 1.2394014596939087, + "learning_rate": 4.098005282330406e-05, + "loss": 0.6122, + "step": 17016 + }, + { + "epoch": 1.1420757692694876, + "grad_norm": 1.0862969160079956, + "learning_rate": 4.0969363040660685e-05, + "loss": 0.5264, + "step": 17018 + }, + { + "epoch": 1.1422099929532565, + "grad_norm": 1.0558875799179077, + "learning_rate": 4.095867368469717e-05, + "loss": 0.5508, + "step": 17020 + }, + { + "epoch": 1.1423442166370257, + "grad_norm": 1.0416982173919678, + "learning_rate": 4.094798475591853e-05, + "loss": 0.512, + "step": 17022 + }, + { + "epoch": 1.1424784403207946, + "grad_norm": 0.9815146327018738, + "learning_rate": 4.093729625482983e-05, + "loss": 0.5085, + "step": 17024 + }, + { + "epoch": 1.1426126640045635, + "grad_norm": 0.936656653881073, + "learning_rate": 4.092660818193603e-05, + "loss": 0.512, + "step": 17026 + }, + { + "epoch": 1.1427468876883327, + "grad_norm": 1.0042047500610352, + "learning_rate": 4.091592053774219e-05, + "loss": 0.4615, + "step": 17028 + }, + { + "epoch": 1.1428811113721016, + "grad_norm": 1.1232362985610962, + "learning_rate": 4.0905233322753224e-05, + "loss": 0.5795, + "step": 17030 + }, + { + "epoch": 1.1430153350558707, + "grad_norm": 1.0395944118499756, + "learning_rate": 4.089454653747409e-05, + "loss": 0.5091, + "step": 17032 + }, + { + "epoch": 1.1431495587396396, + "grad_norm": 1.0014408826828003, + "learning_rate": 4.088386018240974e-05, + "loss": 0.506, + "step": 17034 + }, + { + "epoch": 1.1432837824234086, + "grad_norm": 1.0939862728118896, + "learning_rate": 4.087317425806504e-05, + "loss": 0.5409, + "step": 17036 + }, + { + "epoch": 1.1434180061071777, + "grad_norm": 1.2131550312042236, + "learning_rate": 4.0862488764944956e-05, + "loss": 0.5304, + "step": 17038 + }, + { + "epoch": 1.1435522297909466, + "grad_norm": 1.020034909248352, + "learning_rate": 4.085180370355427e-05, + "loss": 0.5231, + "step": 17040 + }, + { + "epoch": 1.1436864534747155, + "grad_norm": 1.0369129180908203, + "learning_rate": 4.0841119074397895e-05, + "loss": 0.592, + "step": 17042 + }, + { + "epoch": 1.1438206771584847, + "grad_norm": 1.0240205526351929, + "learning_rate": 4.083043487798062e-05, + "loss": 0.5103, + "step": 17044 + }, + { + "epoch": 1.1439549008422536, + "grad_norm": 0.9685449004173279, + "learning_rate": 4.081975111480728e-05, + "loss": 0.5502, + "step": 17046 + }, + { + "epoch": 1.1440891245260225, + "grad_norm": 0.9529426693916321, + "learning_rate": 4.080906778538263e-05, + "loss": 0.4839, + "step": 17048 + }, + { + "epoch": 1.1442233482097917, + "grad_norm": 0.9597352147102356, + "learning_rate": 4.079838489021146e-05, + "loss": 0.5054, + "step": 17050 + }, + { + "epoch": 1.1443575718935606, + "grad_norm": 0.9930350184440613, + "learning_rate": 4.078770242979851e-05, + "loss": 0.5735, + "step": 17052 + }, + { + "epoch": 1.1444917955773297, + "grad_norm": 0.8961820006370544, + "learning_rate": 4.077702040464851e-05, + "loss": 0.4948, + "step": 17054 + }, + { + "epoch": 1.1446260192610986, + "grad_norm": 1.060964822769165, + "learning_rate": 4.0766338815266156e-05, + "loss": 0.4675, + "step": 17056 + }, + { + "epoch": 1.1447602429448676, + "grad_norm": 1.0138031244277954, + "learning_rate": 4.0755657662156144e-05, + "loss": 0.4757, + "step": 17058 + }, + { + "epoch": 1.1448944666286367, + "grad_norm": 0.8255618214607239, + "learning_rate": 4.0744976945823116e-05, + "loss": 0.5061, + "step": 17060 + }, + { + "epoch": 1.1450286903124056, + "grad_norm": 1.083939790725708, + "learning_rate": 4.073429666677174e-05, + "loss": 0.5133, + "step": 17062 + }, + { + "epoch": 1.1451629139961745, + "grad_norm": 1.0115509033203125, + "learning_rate": 4.07236168255066e-05, + "loss": 0.5213, + "step": 17064 + }, + { + "epoch": 1.1452971376799437, + "grad_norm": 1.0404711961746216, + "learning_rate": 4.0712937422532366e-05, + "loss": 0.4725, + "step": 17066 + }, + { + "epoch": 1.1454313613637126, + "grad_norm": 1.0284857749938965, + "learning_rate": 4.070225845835354e-05, + "loss": 0.5641, + "step": 17068 + }, + { + "epoch": 1.1455655850474815, + "grad_norm": 0.8678731322288513, + "learning_rate": 4.069157993347474e-05, + "loss": 0.4719, + "step": 17070 + }, + { + "epoch": 1.1456998087312507, + "grad_norm": 0.972980260848999, + "learning_rate": 4.068090184840047e-05, + "loss": 0.4771, + "step": 17072 + }, + { + "epoch": 1.1458340324150196, + "grad_norm": 0.9753967523574829, + "learning_rate": 4.0670224203635274e-05, + "loss": 0.5157, + "step": 17074 + }, + { + "epoch": 1.1459682560987887, + "grad_norm": 0.9964821934700012, + "learning_rate": 4.065954699968363e-05, + "loss": 0.4685, + "step": 17076 + }, + { + "epoch": 1.1461024797825576, + "grad_norm": 1.0178292989730835, + "learning_rate": 4.064887023705003e-05, + "loss": 0.4768, + "step": 17078 + }, + { + "epoch": 1.1462367034663266, + "grad_norm": 1.0333075523376465, + "learning_rate": 4.0638193916238914e-05, + "loss": 0.5122, + "step": 17080 + }, + { + "epoch": 1.1463709271500957, + "grad_norm": 0.8620358109474182, + "learning_rate": 4.0627518037754745e-05, + "loss": 0.4573, + "step": 17082 + }, + { + "epoch": 1.1465051508338646, + "grad_norm": 0.8973593711853027, + "learning_rate": 4.0616842602101894e-05, + "loss": 0.4458, + "step": 17084 + }, + { + "epoch": 1.1466393745176338, + "grad_norm": 1.081467628479004, + "learning_rate": 4.060616760978479e-05, + "loss": 0.4875, + "step": 17086 + }, + { + "epoch": 1.1467735982014027, + "grad_norm": 1.004448413848877, + "learning_rate": 4.05954930613078e-05, + "loss": 0.5093, + "step": 17088 + }, + { + "epoch": 1.1469078218851716, + "grad_norm": 0.8252903819084167, + "learning_rate": 4.058481895717527e-05, + "loss": 0.5308, + "step": 17090 + }, + { + "epoch": 1.1470420455689405, + "grad_norm": 0.9397380352020264, + "learning_rate": 4.057414529789151e-05, + "loss": 0.4919, + "step": 17092 + }, + { + "epoch": 1.1471762692527097, + "grad_norm": 1.4505244493484497, + "learning_rate": 4.056347208396088e-05, + "loss": 0.53, + "step": 17094 + }, + { + "epoch": 1.1473104929364786, + "grad_norm": 0.9114947319030762, + "learning_rate": 4.05527993158876e-05, + "loss": 0.4595, + "step": 17096 + }, + { + "epoch": 1.1474447166202477, + "grad_norm": 1.0862253904342651, + "learning_rate": 4.0542126994176005e-05, + "loss": 0.4962, + "step": 17098 + }, + { + "epoch": 1.1475789403040166, + "grad_norm": 1.0601911544799805, + "learning_rate": 4.053145511933032e-05, + "loss": 0.526, + "step": 17100 + }, + { + "epoch": 1.1477131639877856, + "grad_norm": 0.9861023426055908, + "learning_rate": 4.052078369185473e-05, + "loss": 0.5379, + "step": 17102 + }, + { + "epoch": 1.1478473876715547, + "grad_norm": 0.9312143325805664, + "learning_rate": 4.051011271225349e-05, + "loss": 0.4921, + "step": 17104 + }, + { + "epoch": 1.1479816113553236, + "grad_norm": 1.0780442953109741, + "learning_rate": 4.049944218103076e-05, + "loss": 0.5544, + "step": 17106 + }, + { + "epoch": 1.1481158350390928, + "grad_norm": 1.0807689428329468, + "learning_rate": 4.048877209869071e-05, + "loss": 0.5142, + "step": 17108 + }, + { + "epoch": 1.1482500587228617, + "grad_norm": 1.092450737953186, + "learning_rate": 4.047810246573746e-05, + "loss": 0.5646, + "step": 17110 + }, + { + "epoch": 1.1483842824066306, + "grad_norm": 1.3242136240005493, + "learning_rate": 4.046743328267516e-05, + "loss": 0.5701, + "step": 17112 + }, + { + "epoch": 1.1485185060903997, + "grad_norm": 1.155165195465088, + "learning_rate": 4.045676455000788e-05, + "loss": 0.5041, + "step": 17114 + }, + { + "epoch": 1.1486527297741687, + "grad_norm": 0.9011285901069641, + "learning_rate": 4.0446096268239716e-05, + "loss": 0.4813, + "step": 17116 + }, + { + "epoch": 1.1487869534579376, + "grad_norm": 0.9024450778961182, + "learning_rate": 4.043542843787472e-05, + "loss": 0.5612, + "step": 17118 + }, + { + "epoch": 1.1489211771417067, + "grad_norm": 0.931702196598053, + "learning_rate": 4.0424761059416915e-05, + "loss": 0.5496, + "step": 17120 + }, + { + "epoch": 1.1490554008254756, + "grad_norm": 0.9446266293525696, + "learning_rate": 4.0414094133370325e-05, + "loss": 0.5196, + "step": 17122 + }, + { + "epoch": 1.1491896245092446, + "grad_norm": 0.907971203327179, + "learning_rate": 4.040342766023894e-05, + "loss": 0.46, + "step": 17124 + }, + { + "epoch": 1.1493238481930137, + "grad_norm": 1.0521533489227295, + "learning_rate": 4.0392761640526705e-05, + "loss": 0.4844, + "step": 17126 + }, + { + "epoch": 1.1494580718767826, + "grad_norm": 0.9352225661277771, + "learning_rate": 4.038209607473763e-05, + "loss": 0.5457, + "step": 17128 + }, + { + "epoch": 1.1495922955605518, + "grad_norm": 1.092270016670227, + "learning_rate": 4.037143096337556e-05, + "loss": 0.5504, + "step": 17130 + }, + { + "epoch": 1.1497265192443207, + "grad_norm": 0.9602010250091553, + "learning_rate": 4.0360766306944474e-05, + "loss": 0.5331, + "step": 17132 + }, + { + "epoch": 1.1498607429280896, + "grad_norm": 0.9188903570175171, + "learning_rate": 4.035010210594821e-05, + "loss": 0.4765, + "step": 17134 + }, + { + "epoch": 1.1499949666118587, + "grad_norm": 1.1557133197784424, + "learning_rate": 4.033943836089066e-05, + "loss": 0.4902, + "step": 17136 + }, + { + "epoch": 1.1501291902956277, + "grad_norm": 0.8777872323989868, + "learning_rate": 4.032877507227564e-05, + "loss": 0.4979, + "step": 17138 + }, + { + "epoch": 1.1502634139793966, + "grad_norm": 0.9906134009361267, + "learning_rate": 4.031811224060699e-05, + "loss": 0.5023, + "step": 17140 + }, + { + "epoch": 1.1503976376631657, + "grad_norm": 1.0566591024398804, + "learning_rate": 4.030744986638848e-05, + "loss": 0.456, + "step": 17142 + }, + { + "epoch": 1.1505318613469346, + "grad_norm": 1.0089061260223389, + "learning_rate": 4.029678795012392e-05, + "loss": 0.4734, + "step": 17144 + }, + { + "epoch": 1.1506660850307036, + "grad_norm": 0.9829335808753967, + "learning_rate": 4.028612649231704e-05, + "loss": 0.5027, + "step": 17146 + }, + { + "epoch": 1.1508003087144727, + "grad_norm": 1.0287353992462158, + "learning_rate": 4.027546549347159e-05, + "loss": 0.5008, + "step": 17148 + }, + { + "epoch": 1.1509345323982416, + "grad_norm": 1.08994722366333, + "learning_rate": 4.026480495409127e-05, + "loss": 0.5361, + "step": 17150 + }, + { + "epoch": 1.1510687560820108, + "grad_norm": 1.085174322128296, + "learning_rate": 4.025414487467977e-05, + "loss": 0.5735, + "step": 17152 + }, + { + "epoch": 1.1512029797657797, + "grad_norm": 1.1141557693481445, + "learning_rate": 4.024348525574075e-05, + "loss": 0.521, + "step": 17154 + }, + { + "epoch": 1.1513372034495486, + "grad_norm": 1.0411978960037231, + "learning_rate": 4.0232826097777895e-05, + "loss": 0.4969, + "step": 17156 + }, + { + "epoch": 1.1514714271333177, + "grad_norm": 1.0179568529129028, + "learning_rate": 4.0222167401294766e-05, + "loss": 0.4433, + "step": 17158 + }, + { + "epoch": 1.1516056508170867, + "grad_norm": 0.9610757231712341, + "learning_rate": 4.021150916679502e-05, + "loss": 0.4584, + "step": 17160 + }, + { + "epoch": 1.1517398745008558, + "grad_norm": 1.2858824729919434, + "learning_rate": 4.020085139478221e-05, + "loss": 0.5452, + "step": 17162 + }, + { + "epoch": 1.1518740981846247, + "grad_norm": 1.0095821619033813, + "learning_rate": 4.0190194085759915e-05, + "loss": 0.476, + "step": 17164 + }, + { + "epoch": 1.1520083218683936, + "grad_norm": 0.9941380023956299, + "learning_rate": 4.017953724023165e-05, + "loss": 0.475, + "step": 17166 + }, + { + "epoch": 1.1521425455521626, + "grad_norm": 1.1222082376480103, + "learning_rate": 4.016888085870094e-05, + "loss": 0.4566, + "step": 17168 + }, + { + "epoch": 1.1522767692359317, + "grad_norm": 1.0464144945144653, + "learning_rate": 4.0158224941671285e-05, + "loss": 0.5358, + "step": 17170 + }, + { + "epoch": 1.1524109929197006, + "grad_norm": 0.8921653032302856, + "learning_rate": 4.0147569489646135e-05, + "loss": 0.5072, + "step": 17172 + }, + { + "epoch": 1.1525452166034698, + "grad_norm": 0.8786313533782959, + "learning_rate": 4.013691450312897e-05, + "loss": 0.6054, + "step": 17174 + }, + { + "epoch": 1.1526794402872387, + "grad_norm": 0.9894219636917114, + "learning_rate": 4.012625998262318e-05, + "loss": 0.4763, + "step": 17176 + }, + { + "epoch": 1.1528136639710076, + "grad_norm": 1.0877069234848022, + "learning_rate": 4.0115605928632205e-05, + "loss": 0.4868, + "step": 17178 + }, + { + "epoch": 1.1529478876547767, + "grad_norm": 1.1121222972869873, + "learning_rate": 4.0104952341659405e-05, + "loss": 0.5178, + "step": 17180 + }, + { + "epoch": 1.1530821113385457, + "grad_norm": 1.0284626483917236, + "learning_rate": 4.0094299222208154e-05, + "loss": 0.5683, + "step": 17182 + }, + { + "epoch": 1.1532163350223148, + "grad_norm": 0.9250473976135254, + "learning_rate": 4.008364657078178e-05, + "loss": 0.4824, + "step": 17184 + }, + { + "epoch": 1.1533505587060837, + "grad_norm": 1.0787099599838257, + "learning_rate": 4.007299438788362e-05, + "loss": 0.524, + "step": 17186 + }, + { + "epoch": 1.1534847823898526, + "grad_norm": 0.9786050319671631, + "learning_rate": 4.0062342674016926e-05, + "loss": 0.5332, + "step": 17188 + }, + { + "epoch": 1.1536190060736218, + "grad_norm": 0.860023558139801, + "learning_rate": 4.005169142968503e-05, + "loss": 0.5021, + "step": 17190 + }, + { + "epoch": 1.1537532297573907, + "grad_norm": 1.308913230895996, + "learning_rate": 4.004104065539112e-05, + "loss": 0.5735, + "step": 17192 + }, + { + "epoch": 1.1538874534411596, + "grad_norm": 1.022002935409546, + "learning_rate": 4.003039035163848e-05, + "loss": 0.5111, + "step": 17194 + }, + { + "epoch": 1.1540216771249288, + "grad_norm": 1.0221171379089355, + "learning_rate": 4.0019740518930285e-05, + "loss": 0.5244, + "step": 17196 + }, + { + "epoch": 1.1541559008086977, + "grad_norm": 1.025030493736267, + "learning_rate": 4.000909115776973e-05, + "loss": 0.5237, + "step": 17198 + }, + { + "epoch": 1.1542901244924666, + "grad_norm": 1.0764104127883911, + "learning_rate": 3.999844226865996e-05, + "loss": 0.5109, + "step": 17200 + }, + { + "epoch": 1.1544243481762357, + "grad_norm": 1.0224589109420776, + "learning_rate": 3.998779385210414e-05, + "loss": 0.4532, + "step": 17202 + }, + { + "epoch": 1.1545585718600047, + "grad_norm": 1.0146963596343994, + "learning_rate": 3.9977145908605355e-05, + "loss": 0.4632, + "step": 17204 + }, + { + "epoch": 1.1546927955437738, + "grad_norm": 1.0280194282531738, + "learning_rate": 3.9966498438666734e-05, + "loss": 0.5198, + "step": 17206 + }, + { + "epoch": 1.1548270192275427, + "grad_norm": 1.2473407983779907, + "learning_rate": 3.995585144279132e-05, + "loss": 0.6126, + "step": 17208 + }, + { + "epoch": 1.1549612429113116, + "grad_norm": 1.049044132232666, + "learning_rate": 3.994520492148218e-05, + "loss": 0.5154, + "step": 17210 + }, + { + "epoch": 1.1550954665950808, + "grad_norm": 0.9378153085708618, + "learning_rate": 3.9934558875242337e-05, + "loss": 0.528, + "step": 17212 + }, + { + "epoch": 1.1552296902788497, + "grad_norm": 1.051458716392517, + "learning_rate": 3.9923913304574794e-05, + "loss": 0.5225, + "step": 17214 + }, + { + "epoch": 1.1553639139626186, + "grad_norm": 0.990528404712677, + "learning_rate": 3.991326820998251e-05, + "loss": 0.5219, + "step": 17216 + }, + { + "epoch": 1.1554981376463878, + "grad_norm": 1.0346277952194214, + "learning_rate": 3.9902623591968514e-05, + "loss": 0.4648, + "step": 17218 + }, + { + "epoch": 1.1556323613301567, + "grad_norm": 0.9941197633743286, + "learning_rate": 3.989197945103565e-05, + "loss": 0.5467, + "step": 17220 + }, + { + "epoch": 1.1557665850139256, + "grad_norm": 1.0439867973327637, + "learning_rate": 3.9881335787686915e-05, + "loss": 0.5672, + "step": 17222 + }, + { + "epoch": 1.1559008086976947, + "grad_norm": 0.9574592709541321, + "learning_rate": 3.987069260242515e-05, + "loss": 0.506, + "step": 17224 + }, + { + "epoch": 1.1560350323814637, + "grad_norm": 0.9048669934272766, + "learning_rate": 3.986004989575325e-05, + "loss": 0.4766, + "step": 17226 + }, + { + "epoch": 1.1561692560652328, + "grad_norm": 1.0663636922836304, + "learning_rate": 3.9849407668174044e-05, + "loss": 0.541, + "step": 17228 + }, + { + "epoch": 1.1563034797490017, + "grad_norm": 1.2869949340820312, + "learning_rate": 3.9838765920190376e-05, + "loss": 0.4894, + "step": 17230 + }, + { + "epoch": 1.1564377034327706, + "grad_norm": 0.9784888029098511, + "learning_rate": 3.9828124652305024e-05, + "loss": 0.4924, + "step": 17232 + }, + { + "epoch": 1.1565719271165398, + "grad_norm": 1.0746228694915771, + "learning_rate": 3.98174838650208e-05, + "loss": 0.5297, + "step": 17234 + }, + { + "epoch": 1.1567061508003087, + "grad_norm": 0.8882965445518494, + "learning_rate": 3.9806843558840427e-05, + "loss": 0.4856, + "step": 17236 + }, + { + "epoch": 1.1568403744840778, + "grad_norm": 1.1055036783218384, + "learning_rate": 3.979620373426665e-05, + "loss": 0.513, + "step": 17238 + }, + { + "epoch": 1.1569745981678468, + "grad_norm": 1.1486135721206665, + "learning_rate": 3.978556439180219e-05, + "loss": 0.5003, + "step": 17240 + }, + { + "epoch": 1.1571088218516157, + "grad_norm": 1.0955274105072021, + "learning_rate": 3.9774925531949714e-05, + "loss": 0.5072, + "step": 17242 + }, + { + "epoch": 1.1572430455353846, + "grad_norm": 1.2183367013931274, + "learning_rate": 3.9764287155211916e-05, + "loss": 0.5716, + "step": 17244 + }, + { + "epoch": 1.1573772692191537, + "grad_norm": 1.0250377655029297, + "learning_rate": 3.975364926209141e-05, + "loss": 0.5355, + "step": 17246 + }, + { + "epoch": 1.1575114929029227, + "grad_norm": 1.3676515817642212, + "learning_rate": 3.974301185309083e-05, + "loss": 0.5002, + "step": 17248 + }, + { + "epoch": 1.1576457165866918, + "grad_norm": 1.0352555513381958, + "learning_rate": 3.973237492871276e-05, + "loss": 0.5515, + "step": 17250 + }, + { + "epoch": 1.1577799402704607, + "grad_norm": 1.0024343729019165, + "learning_rate": 3.972173848945981e-05, + "loss": 0.5135, + "step": 17252 + }, + { + "epoch": 1.1579141639542296, + "grad_norm": 0.9455167055130005, + "learning_rate": 3.971110253583448e-05, + "loss": 0.54, + "step": 17254 + }, + { + "epoch": 1.1580483876379988, + "grad_norm": 1.19954252243042, + "learning_rate": 3.970046706833935e-05, + "loss": 0.5259, + "step": 17256 + }, + { + "epoch": 1.1581826113217677, + "grad_norm": 0.9135200381278992, + "learning_rate": 3.968983208747686e-05, + "loss": 0.4701, + "step": 17258 + }, + { + "epoch": 1.1583168350055368, + "grad_norm": 1.0010595321655273, + "learning_rate": 3.967919759374956e-05, + "loss": 0.5032, + "step": 17260 + }, + { + "epoch": 1.1584510586893058, + "grad_norm": 0.9441422820091248, + "learning_rate": 3.9668563587659865e-05, + "loss": 0.4644, + "step": 17262 + }, + { + "epoch": 1.1585852823730747, + "grad_norm": 1.0468347072601318, + "learning_rate": 3.9657930069710236e-05, + "loss": 0.5695, + "step": 17264 + }, + { + "epoch": 1.1587195060568438, + "grad_norm": 1.0343860387802124, + "learning_rate": 3.9647297040403066e-05, + "loss": 0.5416, + "step": 17266 + }, + { + "epoch": 1.1588537297406127, + "grad_norm": 0.8644536137580872, + "learning_rate": 3.963666450024075e-05, + "loss": 0.4487, + "step": 17268 + }, + { + "epoch": 1.1589879534243817, + "grad_norm": 0.938911497592926, + "learning_rate": 3.962603244972566e-05, + "loss": 0.4666, + "step": 17270 + }, + { + "epoch": 1.1591221771081508, + "grad_norm": 0.8899960517883301, + "learning_rate": 3.9615400889360146e-05, + "loss": 0.4681, + "step": 17272 + }, + { + "epoch": 1.1592564007919197, + "grad_norm": 1.0408364534378052, + "learning_rate": 3.9604769819646515e-05, + "loss": 0.4553, + "step": 17274 + }, + { + "epoch": 1.1593906244756886, + "grad_norm": 1.0575523376464844, + "learning_rate": 3.959413924108707e-05, + "loss": 0.5335, + "step": 17276 + }, + { + "epoch": 1.1595248481594578, + "grad_norm": 1.0249072313308716, + "learning_rate": 3.958350915418407e-05, + "loss": 0.512, + "step": 17278 + }, + { + "epoch": 1.1596590718432267, + "grad_norm": 0.9576659202575684, + "learning_rate": 3.95728795594398e-05, + "loss": 0.5155, + "step": 17280 + }, + { + "epoch": 1.1597932955269958, + "grad_norm": 1.0557490587234497, + "learning_rate": 3.9562250457356445e-05, + "loss": 0.5466, + "step": 17282 + }, + { + "epoch": 1.1599275192107648, + "grad_norm": 0.9900712370872498, + "learning_rate": 3.955162184843625e-05, + "loss": 0.5618, + "step": 17284 + }, + { + "epoch": 1.1600617428945337, + "grad_norm": 0.9188636541366577, + "learning_rate": 3.9540993733181366e-05, + "loss": 0.4764, + "step": 17286 + }, + { + "epoch": 1.1601959665783028, + "grad_norm": 0.980354905128479, + "learning_rate": 3.953036611209397e-05, + "loss": 0.4352, + "step": 17288 + }, + { + "epoch": 1.1603301902620717, + "grad_norm": 0.9077704548835754, + "learning_rate": 3.951973898567617e-05, + "loss": 0.467, + "step": 17290 + }, + { + "epoch": 1.1604644139458407, + "grad_norm": 1.2130227088928223, + "learning_rate": 3.950911235443012e-05, + "loss": 0.5626, + "step": 17292 + }, + { + "epoch": 1.1605986376296098, + "grad_norm": 1.041313648223877, + "learning_rate": 3.949848621885786e-05, + "loss": 0.4997, + "step": 17294 + }, + { + "epoch": 1.1607328613133787, + "grad_norm": 1.0547500848770142, + "learning_rate": 3.948786057946149e-05, + "loss": 0.5183, + "step": 17296 + }, + { + "epoch": 1.1608670849971476, + "grad_norm": 1.0261955261230469, + "learning_rate": 3.9477235436743026e-05, + "loss": 0.5048, + "step": 17298 + }, + { + "epoch": 1.1610013086809168, + "grad_norm": 0.8961619734764099, + "learning_rate": 3.9466610791204504e-05, + "loss": 0.4825, + "step": 17300 + }, + { + "epoch": 1.1611355323646857, + "grad_norm": 0.8076702356338501, + "learning_rate": 3.94559866433479e-05, + "loss": 0.4813, + "step": 17302 + }, + { + "epoch": 1.1612697560484548, + "grad_norm": 0.9887005686759949, + "learning_rate": 3.9445362993675204e-05, + "loss": 0.525, + "step": 17304 + }, + { + "epoch": 1.1614039797322238, + "grad_norm": 1.0277241468429565, + "learning_rate": 3.9434739842688356e-05, + "loss": 0.5315, + "step": 17306 + }, + { + "epoch": 1.1615382034159927, + "grad_norm": 1.036078929901123, + "learning_rate": 3.942411719088926e-05, + "loss": 0.5651, + "step": 17308 + }, + { + "epoch": 1.1616724270997618, + "grad_norm": 1.1413941383361816, + "learning_rate": 3.941349503877984e-05, + "loss": 0.5602, + "step": 17310 + }, + { + "epoch": 1.1618066507835307, + "grad_norm": 1.0826722383499146, + "learning_rate": 3.9402873386861944e-05, + "loss": 0.5312, + "step": 17312 + }, + { + "epoch": 1.1619408744672999, + "grad_norm": 1.1246724128723145, + "learning_rate": 3.939225223563747e-05, + "loss": 0.5123, + "step": 17314 + }, + { + "epoch": 1.1620750981510688, + "grad_norm": 0.9715235829353333, + "learning_rate": 3.938163158560819e-05, + "loss": 0.4429, + "step": 17316 + }, + { + "epoch": 1.1622093218348377, + "grad_norm": 0.9817221164703369, + "learning_rate": 3.9371011437275965e-05, + "loss": 0.4916, + "step": 17318 + }, + { + "epoch": 1.1623435455186066, + "grad_norm": 0.9534182548522949, + "learning_rate": 3.936039179114253e-05, + "loss": 0.5323, + "step": 17320 + }, + { + "epoch": 1.1624777692023758, + "grad_norm": 1.2011499404907227, + "learning_rate": 3.9349772647709666e-05, + "loss": 0.4955, + "step": 17322 + }, + { + "epoch": 1.1626119928861447, + "grad_norm": 0.9890552163124084, + "learning_rate": 3.9339154007479096e-05, + "loss": 0.5058, + "step": 17324 + }, + { + "epoch": 1.1627462165699138, + "grad_norm": 0.9952134490013123, + "learning_rate": 3.9328535870952545e-05, + "loss": 0.5537, + "step": 17326 + }, + { + "epoch": 1.1628804402536828, + "grad_norm": 1.0444145202636719, + "learning_rate": 3.931791823863168e-05, + "loss": 0.5134, + "step": 17328 + }, + { + "epoch": 1.1630146639374517, + "grad_norm": 1.0761719942092896, + "learning_rate": 3.930730111101819e-05, + "loss": 0.5787, + "step": 17330 + }, + { + "epoch": 1.1631488876212208, + "grad_norm": 1.03164803981781, + "learning_rate": 3.9296684488613675e-05, + "loss": 0.5072, + "step": 17332 + }, + { + "epoch": 1.1632831113049897, + "grad_norm": 0.9586412906646729, + "learning_rate": 3.928606837191979e-05, + "loss": 0.4944, + "step": 17334 + }, + { + "epoch": 1.1634173349887589, + "grad_norm": 0.9396413564682007, + "learning_rate": 3.9275452761438095e-05, + "loss": 0.5199, + "step": 17336 + }, + { + "epoch": 1.1635515586725278, + "grad_norm": 1.0381124019622803, + "learning_rate": 3.926483765767017e-05, + "loss": 0.5051, + "step": 17338 + }, + { + "epoch": 1.1636857823562967, + "grad_norm": 1.0771284103393555, + "learning_rate": 3.925422306111754e-05, + "loss": 0.5278, + "step": 17340 + }, + { + "epoch": 1.1638200060400659, + "grad_norm": 0.9608303904533386, + "learning_rate": 3.924360897228177e-05, + "loss": 0.5329, + "step": 17342 + }, + { + "epoch": 1.1639542297238348, + "grad_norm": 1.1120320558547974, + "learning_rate": 3.9232995391664296e-05, + "loss": 0.484, + "step": 17344 + }, + { + "epoch": 1.1640884534076037, + "grad_norm": 0.8794587254524231, + "learning_rate": 3.922238231976665e-05, + "loss": 0.4966, + "step": 17346 + }, + { + "epoch": 1.1642226770913728, + "grad_norm": 1.0041494369506836, + "learning_rate": 3.921176975709021e-05, + "loss": 0.4705, + "step": 17348 + }, + { + "epoch": 1.1643569007751418, + "grad_norm": 0.9785823225975037, + "learning_rate": 3.920115770413646e-05, + "loss": 0.6067, + "step": 17350 + }, + { + "epoch": 1.1644911244589107, + "grad_norm": 1.7183648347854614, + "learning_rate": 3.919054616140676e-05, + "loss": 0.5145, + "step": 17352 + }, + { + "epoch": 1.1646253481426798, + "grad_norm": 1.084672212600708, + "learning_rate": 3.917993512940251e-05, + "loss": 0.493, + "step": 17354 + }, + { + "epoch": 1.1647595718264487, + "grad_norm": 0.9071468710899353, + "learning_rate": 3.916932460862504e-05, + "loss": 0.4254, + "step": 17356 + }, + { + "epoch": 1.1648937955102179, + "grad_norm": 0.8470814824104309, + "learning_rate": 3.915871459957569e-05, + "loss": 0.5063, + "step": 17358 + }, + { + "epoch": 1.1650280191939868, + "grad_norm": 0.9796807765960693, + "learning_rate": 3.914810510275575e-05, + "loss": 0.492, + "step": 17360 + }, + { + "epoch": 1.1651622428777557, + "grad_norm": 0.9707605838775635, + "learning_rate": 3.913749611866651e-05, + "loss": 0.519, + "step": 17362 + }, + { + "epoch": 1.1652964665615249, + "grad_norm": 0.9335168600082397, + "learning_rate": 3.912688764780921e-05, + "loss": 0.4455, + "step": 17364 + }, + { + "epoch": 1.1654306902452938, + "grad_norm": 0.9438152313232422, + "learning_rate": 3.91162796906851e-05, + "loss": 0.4892, + "step": 17366 + }, + { + "epoch": 1.1655649139290627, + "grad_norm": 1.0461525917053223, + "learning_rate": 3.910567224779535e-05, + "loss": 0.5111, + "step": 17368 + }, + { + "epoch": 1.1656991376128318, + "grad_norm": 0.9843353629112244, + "learning_rate": 3.90950653196412e-05, + "loss": 0.4903, + "step": 17370 + }, + { + "epoch": 1.1658333612966008, + "grad_norm": 1.0085928440093994, + "learning_rate": 3.908445890672373e-05, + "loss": 0.5351, + "step": 17372 + }, + { + "epoch": 1.1659675849803697, + "grad_norm": 0.9849995970726013, + "learning_rate": 3.907385300954414e-05, + "loss": 0.4564, + "step": 17374 + }, + { + "epoch": 1.1661018086641388, + "grad_norm": 1.2877445220947266, + "learning_rate": 3.906324762860352e-05, + "loss": 0.5493, + "step": 17376 + }, + { + "epoch": 1.1662360323479077, + "grad_norm": 0.9882004261016846, + "learning_rate": 3.9052642764402906e-05, + "loss": 0.5426, + "step": 17378 + }, + { + "epoch": 1.1663702560316769, + "grad_norm": 1.2059866189956665, + "learning_rate": 3.904203841744343e-05, + "loss": 0.5313, + "step": 17380 + }, + { + "epoch": 1.1665044797154458, + "grad_norm": 1.0301743745803833, + "learning_rate": 3.903143458822606e-05, + "loss": 0.4965, + "step": 17382 + }, + { + "epoch": 1.1666387033992147, + "grad_norm": 0.9556608200073242, + "learning_rate": 3.9020831277251863e-05, + "loss": 0.5347, + "step": 17384 + }, + { + "epoch": 1.1667729270829839, + "grad_norm": 0.8305768370628357, + "learning_rate": 3.9010228485021784e-05, + "loss": 0.4965, + "step": 17386 + }, + { + "epoch": 1.1669071507667528, + "grad_norm": 1.1022326946258545, + "learning_rate": 3.8999626212036805e-05, + "loss": 0.4865, + "step": 17388 + }, + { + "epoch": 1.167041374450522, + "grad_norm": 0.9911981225013733, + "learning_rate": 3.898902445879784e-05, + "loss": 0.4622, + "step": 17390 + }, + { + "epoch": 1.1671755981342908, + "grad_norm": 1.1780292987823486, + "learning_rate": 3.897842322580584e-05, + "loss": 0.5141, + "step": 17392 + }, + { + "epoch": 1.1673098218180598, + "grad_norm": 1.2555780410766602, + "learning_rate": 3.8967822513561656e-05, + "loss": 0.5295, + "step": 17394 + }, + { + "epoch": 1.1674440455018287, + "grad_norm": 1.005345106124878, + "learning_rate": 3.895722232256618e-05, + "loss": 0.5085, + "step": 17396 + }, + { + "epoch": 1.1675782691855978, + "grad_norm": 0.9374826550483704, + "learning_rate": 3.8946622653320216e-05, + "loss": 0.4729, + "step": 17398 + }, + { + "epoch": 1.1677124928693667, + "grad_norm": 1.1584421396255493, + "learning_rate": 3.89360235063246e-05, + "loss": 0.5027, + "step": 17400 + }, + { + "epoch": 1.1678467165531359, + "grad_norm": 1.0123322010040283, + "learning_rate": 3.8925424882080105e-05, + "loss": 0.487, + "step": 17402 + }, + { + "epoch": 1.1679809402369048, + "grad_norm": 1.1311192512512207, + "learning_rate": 3.891482678108754e-05, + "loss": 0.5161, + "step": 17404 + }, + { + "epoch": 1.1681151639206737, + "grad_norm": 1.065680980682373, + "learning_rate": 3.890422920384758e-05, + "loss": 0.553, + "step": 17406 + }, + { + "epoch": 1.1682493876044429, + "grad_norm": 0.8839694261550903, + "learning_rate": 3.8893632150861004e-05, + "loss": 0.4933, + "step": 17408 + }, + { + "epoch": 1.1683836112882118, + "grad_norm": 0.9866513013839722, + "learning_rate": 3.888303562262843e-05, + "loss": 0.5221, + "step": 17410 + }, + { + "epoch": 1.168517834971981, + "grad_norm": 1.0440144538879395, + "learning_rate": 3.887243961965059e-05, + "loss": 0.499, + "step": 17412 + }, + { + "epoch": 1.1686520586557498, + "grad_norm": 0.9592545628547668, + "learning_rate": 3.8861844142428083e-05, + "loss": 0.4699, + "step": 17414 + }, + { + "epoch": 1.1687862823395188, + "grad_norm": 0.9670283198356628, + "learning_rate": 3.8851249191461544e-05, + "loss": 0.4649, + "step": 17416 + }, + { + "epoch": 1.168920506023288, + "grad_norm": 1.100111961364746, + "learning_rate": 3.884065476725156e-05, + "loss": 0.4961, + "step": 17418 + }, + { + "epoch": 1.1690547297070568, + "grad_norm": 1.0785095691680908, + "learning_rate": 3.8830060870298696e-05, + "loss": 0.5124, + "step": 17420 + }, + { + "epoch": 1.1691889533908257, + "grad_norm": 1.0409162044525146, + "learning_rate": 3.881946750110348e-05, + "loss": 0.5206, + "step": 17422 + }, + { + "epoch": 1.1693231770745949, + "grad_norm": 1.0038281679153442, + "learning_rate": 3.880887466016645e-05, + "loss": 0.5282, + "step": 17424 + }, + { + "epoch": 1.1694574007583638, + "grad_norm": 0.9651196002960205, + "learning_rate": 3.879828234798808e-05, + "loss": 0.4753, + "step": 17426 + }, + { + "epoch": 1.1695916244421327, + "grad_norm": 1.3088198900222778, + "learning_rate": 3.8787690565068846e-05, + "loss": 0.547, + "step": 17428 + }, + { + "epoch": 1.1697258481259019, + "grad_norm": 1.2230758666992188, + "learning_rate": 3.877709931190917e-05, + "loss": 0.4999, + "step": 17430 + }, + { + "epoch": 1.1698600718096708, + "grad_norm": 0.9876248836517334, + "learning_rate": 3.876650858900951e-05, + "loss": 0.5295, + "step": 17432 + }, + { + "epoch": 1.16999429549344, + "grad_norm": 1.0522669553756714, + "learning_rate": 3.87559183968702e-05, + "loss": 0.5396, + "step": 17434 + }, + { + "epoch": 1.1701285191772088, + "grad_norm": 1.0726021528244019, + "learning_rate": 3.8745328735991664e-05, + "loss": 0.4984, + "step": 17436 + }, + { + "epoch": 1.1702627428609778, + "grad_norm": 1.0583510398864746, + "learning_rate": 3.8734739606874184e-05, + "loss": 0.4875, + "step": 17438 + }, + { + "epoch": 1.170396966544747, + "grad_norm": 1.0662593841552734, + "learning_rate": 3.872415101001813e-05, + "loss": 0.5118, + "step": 17440 + }, + { + "epoch": 1.1705311902285158, + "grad_norm": 0.8635338544845581, + "learning_rate": 3.871356294592377e-05, + "loss": 0.4636, + "step": 17442 + }, + { + "epoch": 1.1706654139122847, + "grad_norm": 0.9625066518783569, + "learning_rate": 3.870297541509134e-05, + "loss": 0.4783, + "step": 17444 + }, + { + "epoch": 1.1707996375960539, + "grad_norm": 1.0663460493087769, + "learning_rate": 3.869238841802113e-05, + "loss": 0.5439, + "step": 17446 + }, + { + "epoch": 1.1709338612798228, + "grad_norm": 1.0197993516921997, + "learning_rate": 3.868180195521332e-05, + "loss": 0.4795, + "step": 17448 + }, + { + "epoch": 1.1710680849635917, + "grad_norm": 0.9224337935447693, + "learning_rate": 3.867121602716812e-05, + "loss": 0.494, + "step": 17450 + }, + { + "epoch": 1.1712023086473609, + "grad_norm": 1.0983052253723145, + "learning_rate": 3.8660630634385674e-05, + "loss": 0.5593, + "step": 17452 + }, + { + "epoch": 1.1713365323311298, + "grad_norm": 0.97585529088974, + "learning_rate": 3.865004577736615e-05, + "loss": 0.453, + "step": 17454 + }, + { + "epoch": 1.171470756014899, + "grad_norm": 0.9972829818725586, + "learning_rate": 3.8639461456609626e-05, + "loss": 0.5571, + "step": 17456 + }, + { + "epoch": 1.1716049796986678, + "grad_norm": 1.0324832201004028, + "learning_rate": 3.862887767261623e-05, + "loss": 0.5233, + "step": 17458 + }, + { + "epoch": 1.1717392033824368, + "grad_norm": 1.1102116107940674, + "learning_rate": 3.861829442588598e-05, + "loss": 0.5954, + "step": 17460 + }, + { + "epoch": 1.171873427066206, + "grad_norm": 1.0274536609649658, + "learning_rate": 3.860771171691895e-05, + "loss": 0.5099, + "step": 17462 + }, + { + "epoch": 1.1720076507499748, + "grad_norm": 0.940142035484314, + "learning_rate": 3.8597129546215135e-05, + "loss": 0.5394, + "step": 17464 + }, + { + "epoch": 1.172141874433744, + "grad_norm": 0.8878188729286194, + "learning_rate": 3.858654791427453e-05, + "loss": 0.4924, + "step": 17466 + }, + { + "epoch": 1.1722760981175129, + "grad_norm": 0.9084573984146118, + "learning_rate": 3.8575966821597075e-05, + "loss": 0.4914, + "step": 17468 + }, + { + "epoch": 1.1724103218012818, + "grad_norm": 1.108896017074585, + "learning_rate": 3.856538626868276e-05, + "loss": 0.5614, + "step": 17470 + }, + { + "epoch": 1.1725445454850507, + "grad_norm": 3.371222496032715, + "learning_rate": 3.855480625603142e-05, + "loss": 0.5331, + "step": 17472 + }, + { + "epoch": 1.1726787691688199, + "grad_norm": 0.939922571182251, + "learning_rate": 3.8544226784143e-05, + "loss": 0.5053, + "step": 17474 + }, + { + "epoch": 1.1728129928525888, + "grad_norm": 0.8527662754058838, + "learning_rate": 3.853364785351733e-05, + "loss": 0.4852, + "step": 17476 + }, + { + "epoch": 1.172947216536358, + "grad_norm": 0.8310046792030334, + "learning_rate": 3.852306946465426e-05, + "loss": 0.4621, + "step": 17478 + }, + { + "epoch": 1.1730814402201268, + "grad_norm": 0.9874976277351379, + "learning_rate": 3.851249161805358e-05, + "loss": 0.5087, + "step": 17480 + }, + { + "epoch": 1.1732156639038958, + "grad_norm": 1.1040178537368774, + "learning_rate": 3.8501914314215095e-05, + "loss": 0.4731, + "step": 17482 + }, + { + "epoch": 1.173349887587665, + "grad_norm": 1.460074543952942, + "learning_rate": 3.849133755363853e-05, + "loss": 0.4691, + "step": 17484 + }, + { + "epoch": 1.1734841112714338, + "grad_norm": 0.9203589558601379, + "learning_rate": 3.8480761336823654e-05, + "loss": 0.4678, + "step": 17486 + }, + { + "epoch": 1.173618334955203, + "grad_norm": 1.0377877950668335, + "learning_rate": 3.8470185664270134e-05, + "loss": 0.5609, + "step": 17488 + }, + { + "epoch": 1.1737525586389719, + "grad_norm": 1.0666682720184326, + "learning_rate": 3.845961053647768e-05, + "loss": 0.491, + "step": 17490 + }, + { + "epoch": 1.1738867823227408, + "grad_norm": 0.9724449515342712, + "learning_rate": 3.844903595394592e-05, + "loss": 0.5039, + "step": 17492 + }, + { + "epoch": 1.17402100600651, + "grad_norm": 1.1078293323516846, + "learning_rate": 3.843846191717453e-05, + "loss": 0.5721, + "step": 17494 + }, + { + "epoch": 1.1741552296902789, + "grad_norm": 1.0953867435455322, + "learning_rate": 3.8427888426663044e-05, + "loss": 0.5078, + "step": 17496 + }, + { + "epoch": 1.1742894533740478, + "grad_norm": 0.9513339400291443, + "learning_rate": 3.841731548291111e-05, + "loss": 0.5298, + "step": 17498 + }, + { + "epoch": 1.174423677057817, + "grad_norm": 1.0575677156448364, + "learning_rate": 3.840674308641821e-05, + "loss": 0.5594, + "step": 17500 + }, + { + "epoch": 1.1745579007415858, + "grad_norm": 1.0335451364517212, + "learning_rate": 3.839617123768392e-05, + "loss": 0.4774, + "step": 17502 + }, + { + "epoch": 1.1746921244253548, + "grad_norm": 0.9178796410560608, + "learning_rate": 3.8385599937207713e-05, + "loss": 0.5096, + "step": 17504 + }, + { + "epoch": 1.174826348109124, + "grad_norm": 1.166209101676941, + "learning_rate": 3.837502918548907e-05, + "loss": 0.4891, + "step": 17506 + }, + { + "epoch": 1.1749605717928928, + "grad_norm": 0.899621307849884, + "learning_rate": 3.836445898302745e-05, + "loss": 0.4807, + "step": 17508 + }, + { + "epoch": 1.175094795476662, + "grad_norm": 1.127547025680542, + "learning_rate": 3.835388933032226e-05, + "loss": 0.5472, + "step": 17510 + }, + { + "epoch": 1.1752290191604309, + "grad_norm": 0.9545964002609253, + "learning_rate": 3.8343320227872894e-05, + "loss": 0.4765, + "step": 17512 + }, + { + "epoch": 1.1753632428441998, + "grad_norm": 1.27057945728302, + "learning_rate": 3.833275167617872e-05, + "loss": 0.5113, + "step": 17514 + }, + { + "epoch": 1.175497466527969, + "grad_norm": 0.8805884718894958, + "learning_rate": 3.83221836757391e-05, + "loss": 0.4793, + "step": 17516 + }, + { + "epoch": 1.1756316902117379, + "grad_norm": 1.0999140739440918, + "learning_rate": 3.831161622705332e-05, + "loss": 0.5474, + "step": 17518 + }, + { + "epoch": 1.1757659138955068, + "grad_norm": 0.852262020111084, + "learning_rate": 3.83010493306207e-05, + "loss": 0.4854, + "step": 17520 + }, + { + "epoch": 1.175900137579276, + "grad_norm": 0.9464651346206665, + "learning_rate": 3.829048298694049e-05, + "loss": 0.524, + "step": 17522 + }, + { + "epoch": 1.1760343612630448, + "grad_norm": 0.8718568682670593, + "learning_rate": 3.827991719651194e-05, + "loss": 0.4508, + "step": 17524 + }, + { + "epoch": 1.1761685849468138, + "grad_norm": 0.8984642028808594, + "learning_rate": 3.826935195983425e-05, + "loss": 0.49, + "step": 17526 + }, + { + "epoch": 1.176302808630583, + "grad_norm": 0.9936639070510864, + "learning_rate": 3.8258787277406615e-05, + "loss": 0.5017, + "step": 17528 + }, + { + "epoch": 1.1764370323143518, + "grad_norm": 1.0150092840194702, + "learning_rate": 3.824822314972818e-05, + "loss": 0.5155, + "step": 17530 + }, + { + "epoch": 1.176571255998121, + "grad_norm": 0.9806212782859802, + "learning_rate": 3.823765957729813e-05, + "loss": 0.4745, + "step": 17532 + }, + { + "epoch": 1.1767054796818899, + "grad_norm": 0.9721527695655823, + "learning_rate": 3.8227096560615486e-05, + "loss": 0.4945, + "step": 17534 + }, + { + "epoch": 1.1768397033656588, + "grad_norm": 0.9306116104125977, + "learning_rate": 3.821653410017941e-05, + "loss": 0.531, + "step": 17536 + }, + { + "epoch": 1.176973927049428, + "grad_norm": 1.0064274072647095, + "learning_rate": 3.820597219648891e-05, + "loss": 0.5429, + "step": 17538 + }, + { + "epoch": 1.1771081507331969, + "grad_norm": 1.1770319938659668, + "learning_rate": 3.819541085004304e-05, + "loss": 0.515, + "step": 17540 + }, + { + "epoch": 1.177242374416966, + "grad_norm": 0.8967007994651794, + "learning_rate": 3.81848500613408e-05, + "loss": 0.4666, + "step": 17542 + }, + { + "epoch": 1.177376598100735, + "grad_norm": 1.0089564323425293, + "learning_rate": 3.817428983088116e-05, + "loss": 0.4986, + "step": 17544 + }, + { + "epoch": 1.1775108217845038, + "grad_norm": 0.8636226654052734, + "learning_rate": 3.8163730159163065e-05, + "loss": 0.4436, + "step": 17546 + }, + { + "epoch": 1.1776450454682728, + "grad_norm": 1.1310228109359741, + "learning_rate": 3.8153171046685454e-05, + "loss": 0.5249, + "step": 17548 + }, + { + "epoch": 1.177779269152042, + "grad_norm": 0.976316511631012, + "learning_rate": 3.8142612493947205e-05, + "loss": 0.4639, + "step": 17550 + }, + { + "epoch": 1.1779134928358108, + "grad_norm": 1.0009357929229736, + "learning_rate": 3.8132054501447206e-05, + "loss": 0.5191, + "step": 17552 + }, + { + "epoch": 1.17804771651958, + "grad_norm": 0.9338767528533936, + "learning_rate": 3.8121497069684285e-05, + "loss": 0.5434, + "step": 17554 + }, + { + "epoch": 1.1781819402033489, + "grad_norm": 0.9037238359451294, + "learning_rate": 3.8110940199157284e-05, + "loss": 0.4996, + "step": 17556 + }, + { + "epoch": 1.1783161638871178, + "grad_norm": 1.027163028717041, + "learning_rate": 3.8100383890364956e-05, + "loss": 0.5148, + "step": 17558 + }, + { + "epoch": 1.178450387570887, + "grad_norm": 1.1842451095581055, + "learning_rate": 3.808982814380612e-05, + "loss": 0.5143, + "step": 17560 + }, + { + "epoch": 1.1785846112546559, + "grad_norm": 1.0291099548339844, + "learning_rate": 3.8079272959979454e-05, + "loss": 0.5255, + "step": 17562 + }, + { + "epoch": 1.178718834938425, + "grad_norm": 0.9508962035179138, + "learning_rate": 3.8068718339383726e-05, + "loss": 0.4743, + "step": 17564 + }, + { + "epoch": 1.178853058622194, + "grad_norm": 1.0064043998718262, + "learning_rate": 3.805816428251757e-05, + "loss": 0.496, + "step": 17566 + }, + { + "epoch": 1.1789872823059628, + "grad_norm": 0.9309611320495605, + "learning_rate": 3.8047610789879696e-05, + "loss": 0.511, + "step": 17568 + }, + { + "epoch": 1.179121505989732, + "grad_norm": 1.0128785371780396, + "learning_rate": 3.8037057861968684e-05, + "loss": 0.5054, + "step": 17570 + }, + { + "epoch": 1.179255729673501, + "grad_norm": 1.043404221534729, + "learning_rate": 3.8026505499283184e-05, + "loss": 0.5251, + "step": 17572 + }, + { + "epoch": 1.1793899533572698, + "grad_norm": 1.0299290418624878, + "learning_rate": 3.801595370232174e-05, + "loss": 0.5318, + "step": 17574 + }, + { + "epoch": 1.179524177041039, + "grad_norm": 0.9969633221626282, + "learning_rate": 3.800540247158293e-05, + "loss": 0.4817, + "step": 17576 + }, + { + "epoch": 1.1796584007248079, + "grad_norm": 1.067553997039795, + "learning_rate": 3.7994851807565254e-05, + "loss": 0.4643, + "step": 17578 + }, + { + "epoch": 1.1797926244085768, + "grad_norm": 1.044662356376648, + "learning_rate": 3.798430171076723e-05, + "loss": 0.5075, + "step": 17580 + }, + { + "epoch": 1.179926848092346, + "grad_norm": 1.7358472347259521, + "learning_rate": 3.7973752181687335e-05, + "loss": 0.5194, + "step": 17582 + }, + { + "epoch": 1.1800610717761149, + "grad_norm": 1.1128405332565308, + "learning_rate": 3.796320322082398e-05, + "loss": 0.5602, + "step": 17584 + }, + { + "epoch": 1.180195295459884, + "grad_norm": 1.165319800376892, + "learning_rate": 3.7952654828675616e-05, + "loss": 0.5301, + "step": 17586 + }, + { + "epoch": 1.180329519143653, + "grad_norm": 1.0974600315093994, + "learning_rate": 3.794210700574061e-05, + "loss": 0.6221, + "step": 17588 + }, + { + "epoch": 1.1804637428274218, + "grad_norm": 0.8745746612548828, + "learning_rate": 3.7931559752517354e-05, + "loss": 0.5553, + "step": 17590 + }, + { + "epoch": 1.180597966511191, + "grad_norm": 1.0914452075958252, + "learning_rate": 3.792101306950414e-05, + "loss": 0.5834, + "step": 17592 + }, + { + "epoch": 1.18073219019496, + "grad_norm": 0.9152393937110901, + "learning_rate": 3.7910466957199336e-05, + "loss": 0.4696, + "step": 17594 + }, + { + "epoch": 1.1808664138787288, + "grad_norm": 1.0173288583755493, + "learning_rate": 3.789992141610117e-05, + "loss": 0.4779, + "step": 17596 + }, + { + "epoch": 1.181000637562498, + "grad_norm": 1.014234185218811, + "learning_rate": 3.788937644670794e-05, + "loss": 0.5314, + "step": 17598 + }, + { + "epoch": 1.1811348612462669, + "grad_norm": 1.069540023803711, + "learning_rate": 3.7878832049517846e-05, + "loss": 0.526, + "step": 17600 + }, + { + "epoch": 1.1812690849300358, + "grad_norm": 1.0426692962646484, + "learning_rate": 3.786828822502912e-05, + "loss": 0.4956, + "step": 17602 + }, + { + "epoch": 1.181403308613805, + "grad_norm": 1.066277265548706, + "learning_rate": 3.78577449737399e-05, + "loss": 0.5578, + "step": 17604 + }, + { + "epoch": 1.1815375322975739, + "grad_norm": 0.8797937631607056, + "learning_rate": 3.784720229614838e-05, + "loss": 0.5221, + "step": 17606 + }, + { + "epoch": 1.181671755981343, + "grad_norm": 1.0879156589508057, + "learning_rate": 3.783666019275263e-05, + "loss": 0.5703, + "step": 17608 + }, + { + "epoch": 1.181805979665112, + "grad_norm": 0.9450921416282654, + "learning_rate": 3.782611866405078e-05, + "loss": 0.4624, + "step": 17610 + }, + { + "epoch": 1.1819402033488808, + "grad_norm": 1.0539926290512085, + "learning_rate": 3.7815577710540874e-05, + "loss": 0.5211, + "step": 17612 + }, + { + "epoch": 1.18207442703265, + "grad_norm": 0.9046851992607117, + "learning_rate": 3.7805037332720975e-05, + "loss": 0.5124, + "step": 17614 + }, + { + "epoch": 1.182208650716419, + "grad_norm": 1.0149006843566895, + "learning_rate": 3.7794497531089066e-05, + "loss": 0.4943, + "step": 17616 + }, + { + "epoch": 1.182342874400188, + "grad_norm": 0.942730188369751, + "learning_rate": 3.778395830614317e-05, + "loss": 0.4962, + "step": 17618 + }, + { + "epoch": 1.182477098083957, + "grad_norm": 1.0554370880126953, + "learning_rate": 3.7773419658381185e-05, + "loss": 0.4686, + "step": 17620 + }, + { + "epoch": 1.1826113217677259, + "grad_norm": 0.9436022043228149, + "learning_rate": 3.7762881588301115e-05, + "loss": 0.472, + "step": 17622 + }, + { + "epoch": 1.1827455454514948, + "grad_norm": 0.913763165473938, + "learning_rate": 3.7752344096400794e-05, + "loss": 0.5218, + "step": 17624 + }, + { + "epoch": 1.182879769135264, + "grad_norm": 1.0532349348068237, + "learning_rate": 3.774180718317816e-05, + "loss": 0.4568, + "step": 17626 + }, + { + "epoch": 1.1830139928190329, + "grad_norm": 1.3988685607910156, + "learning_rate": 3.7731270849131014e-05, + "loss": 0.4813, + "step": 17628 + }, + { + "epoch": 1.183148216502802, + "grad_norm": 0.9777106642723083, + "learning_rate": 3.772073509475721e-05, + "loss": 0.5098, + "step": 17630 + }, + { + "epoch": 1.183282440186571, + "grad_norm": 1.157128930091858, + "learning_rate": 3.7710199920554513e-05, + "loss": 0.5584, + "step": 17632 + }, + { + "epoch": 1.1834166638703398, + "grad_norm": 1.0951553583145142, + "learning_rate": 3.769966532702071e-05, + "loss": 0.4952, + "step": 17634 + }, + { + "epoch": 1.183550887554109, + "grad_norm": 1.0618690252304077, + "learning_rate": 3.7689131314653525e-05, + "loss": 0.4437, + "step": 17636 + }, + { + "epoch": 1.183685111237878, + "grad_norm": 0.9891539216041565, + "learning_rate": 3.7678597883950696e-05, + "loss": 0.4713, + "step": 17638 + }, + { + "epoch": 1.183819334921647, + "grad_norm": 1.0724852085113525, + "learning_rate": 3.7668065035409864e-05, + "loss": 0.528, + "step": 17640 + }, + { + "epoch": 1.183953558605416, + "grad_norm": 0.9540714621543884, + "learning_rate": 3.765753276952873e-05, + "loss": 0.5084, + "step": 17642 + }, + { + "epoch": 1.1840877822891849, + "grad_norm": 1.026107907295227, + "learning_rate": 3.764700108680489e-05, + "loss": 0.6074, + "step": 17644 + }, + { + "epoch": 1.184222005972954, + "grad_norm": 1.0301669836044312, + "learning_rate": 3.763646998773596e-05, + "loss": 0.4662, + "step": 17646 + }, + { + "epoch": 1.184356229656723, + "grad_norm": 0.8550382852554321, + "learning_rate": 3.762593947281952e-05, + "loss": 0.415, + "step": 17648 + }, + { + "epoch": 1.1844904533404919, + "grad_norm": 1.0265635251998901, + "learning_rate": 3.7615409542553094e-05, + "loss": 0.5387, + "step": 17650 + }, + { + "epoch": 1.184624677024261, + "grad_norm": 0.789578914642334, + "learning_rate": 3.760488019743422e-05, + "loss": 0.459, + "step": 17652 + }, + { + "epoch": 1.18475890070803, + "grad_norm": 2.789158344268799, + "learning_rate": 3.759435143796037e-05, + "loss": 0.5348, + "step": 17654 + }, + { + "epoch": 1.1848931243917988, + "grad_norm": 1.1224685907363892, + "learning_rate": 3.7583823264629045e-05, + "loss": 0.5749, + "step": 17656 + }, + { + "epoch": 1.185027348075568, + "grad_norm": 1.0600457191467285, + "learning_rate": 3.7573295677937633e-05, + "loss": 0.5241, + "step": 17658 + }, + { + "epoch": 1.185161571759337, + "grad_norm": 0.9870001077651978, + "learning_rate": 3.7562768678383576e-05, + "loss": 0.4698, + "step": 17660 + }, + { + "epoch": 1.185295795443106, + "grad_norm": 1.0906633138656616, + "learning_rate": 3.755224226646423e-05, + "loss": 0.4884, + "step": 17662 + }, + { + "epoch": 1.185430019126875, + "grad_norm": 1.013982892036438, + "learning_rate": 3.754171644267698e-05, + "loss": 0.5416, + "step": 17664 + }, + { + "epoch": 1.1855642428106439, + "grad_norm": 1.0187809467315674, + "learning_rate": 3.7531191207519114e-05, + "loss": 0.5265, + "step": 17666 + }, + { + "epoch": 1.185698466494413, + "grad_norm": 1.1359308958053589, + "learning_rate": 3.752066656148796e-05, + "loss": 0.5757, + "step": 17668 + }, + { + "epoch": 1.185832690178182, + "grad_norm": 0.9246326088905334, + "learning_rate": 3.7510142505080746e-05, + "loss": 0.5185, + "step": 17670 + }, + { + "epoch": 1.1859669138619509, + "grad_norm": 1.0252164602279663, + "learning_rate": 3.749961903879477e-05, + "loss": 0.5504, + "step": 17672 + }, + { + "epoch": 1.18610113754572, + "grad_norm": 1.0328184366226196, + "learning_rate": 3.748909616312718e-05, + "loss": 0.4842, + "step": 17674 + }, + { + "epoch": 1.186235361229489, + "grad_norm": 0.8675329685211182, + "learning_rate": 3.747857387857522e-05, + "loss": 0.4546, + "step": 17676 + }, + { + "epoch": 1.1863695849132578, + "grad_norm": 0.9986277222633362, + "learning_rate": 3.7468052185636e-05, + "loss": 0.5553, + "step": 17678 + }, + { + "epoch": 1.186503808597027, + "grad_norm": 1.0309573411941528, + "learning_rate": 3.74575310848067e-05, + "loss": 0.5489, + "step": 17680 + }, + { + "epoch": 1.186638032280796, + "grad_norm": 1.0841816663742065, + "learning_rate": 3.744701057658436e-05, + "loss": 0.4927, + "step": 17682 + }, + { + "epoch": 1.186772255964565, + "grad_norm": 1.0273849964141846, + "learning_rate": 3.743649066146612e-05, + "loss": 0.5874, + "step": 17684 + }, + { + "epoch": 1.186906479648334, + "grad_norm": 1.7728848457336426, + "learning_rate": 3.742597133994896e-05, + "loss": 0.5942, + "step": 17686 + }, + { + "epoch": 1.1870407033321029, + "grad_norm": 0.9678778052330017, + "learning_rate": 3.741545261252994e-05, + "loss": 0.4759, + "step": 17688 + }, + { + "epoch": 1.187174927015872, + "grad_norm": 0.9566680788993835, + "learning_rate": 3.740493447970604e-05, + "loss": 0.5046, + "step": 17690 + }, + { + "epoch": 1.187309150699641, + "grad_norm": 1.0579038858413696, + "learning_rate": 3.739441694197422e-05, + "loss": 0.516, + "step": 17692 + }, + { + "epoch": 1.18744337438341, + "grad_norm": 0.9175028800964355, + "learning_rate": 3.7383899999831406e-05, + "loss": 0.4448, + "step": 17694 + }, + { + "epoch": 1.187577598067179, + "grad_norm": 0.9152856469154358, + "learning_rate": 3.737338365377452e-05, + "loss": 0.4691, + "step": 17696 + }, + { + "epoch": 1.187711821750948, + "grad_norm": 1.0249311923980713, + "learning_rate": 3.7362867904300415e-05, + "loss": 0.5694, + "step": 17698 + }, + { + "epoch": 1.1878460454347168, + "grad_norm": 1.0089964866638184, + "learning_rate": 3.7352352751905964e-05, + "loss": 0.4709, + "step": 17700 + }, + { + "epoch": 1.187980269118486, + "grad_norm": 1.1326613426208496, + "learning_rate": 3.7341838197087964e-05, + "loss": 0.539, + "step": 17702 + }, + { + "epoch": 1.188114492802255, + "grad_norm": 0.9482781887054443, + "learning_rate": 3.7331324240343226e-05, + "loss": 0.5831, + "step": 17704 + }, + { + "epoch": 1.188248716486024, + "grad_norm": 0.9407302141189575, + "learning_rate": 3.7320810882168495e-05, + "loss": 0.5088, + "step": 17706 + }, + { + "epoch": 1.188382940169793, + "grad_norm": 1.0773608684539795, + "learning_rate": 3.7310298123060527e-05, + "loss": 0.6095, + "step": 17708 + }, + { + "epoch": 1.1885171638535619, + "grad_norm": 1.024487853050232, + "learning_rate": 3.7299785963516e-05, + "loss": 0.5277, + "step": 17710 + }, + { + "epoch": 1.188651387537331, + "grad_norm": 0.9741911888122559, + "learning_rate": 3.728927440403165e-05, + "loss": 0.5577, + "step": 17712 + }, + { + "epoch": 1.1887856112211, + "grad_norm": 0.9415025115013123, + "learning_rate": 3.727876344510405e-05, + "loss": 0.4607, + "step": 17714 + }, + { + "epoch": 1.188919834904869, + "grad_norm": 1.1595509052276611, + "learning_rate": 3.726825308722989e-05, + "loss": 0.5347, + "step": 17716 + }, + { + "epoch": 1.189054058588638, + "grad_norm": 1.1408499479293823, + "learning_rate": 3.725774333090574e-05, + "loss": 0.4462, + "step": 17718 + }, + { + "epoch": 1.189188282272407, + "grad_norm": 1.05015230178833, + "learning_rate": 3.724723417662813e-05, + "loss": 0.5859, + "step": 17720 + }, + { + "epoch": 1.189322505956176, + "grad_norm": 1.0466431379318237, + "learning_rate": 3.723672562489366e-05, + "loss": 0.4653, + "step": 17722 + }, + { + "epoch": 1.189456729639945, + "grad_norm": 1.0187463760375977, + "learning_rate": 3.7226217676198775e-05, + "loss": 0.555, + "step": 17724 + }, + { + "epoch": 1.189590953323714, + "grad_norm": 1.1777077913284302, + "learning_rate": 3.721571033104001e-05, + "loss": 0.5346, + "step": 17726 + }, + { + "epoch": 1.189725177007483, + "grad_norm": 0.9326265454292297, + "learning_rate": 3.720520358991378e-05, + "loss": 0.44, + "step": 17728 + }, + { + "epoch": 1.189859400691252, + "grad_norm": 1.1017358303070068, + "learning_rate": 3.719469745331653e-05, + "loss": 0.5862, + "step": 17730 + }, + { + "epoch": 1.1899936243750209, + "grad_norm": 1.0286736488342285, + "learning_rate": 3.718419192174464e-05, + "loss": 0.4303, + "step": 17732 + }, + { + "epoch": 1.19012784805879, + "grad_norm": 0.9467357993125916, + "learning_rate": 3.717368699569448e-05, + "loss": 0.5062, + "step": 17734 + }, + { + "epoch": 1.190262071742559, + "grad_norm": 0.9172195196151733, + "learning_rate": 3.716318267566238e-05, + "loss": 0.4467, + "step": 17736 + }, + { + "epoch": 1.190396295426328, + "grad_norm": 1.2418681383132935, + "learning_rate": 3.715267896214467e-05, + "loss": 0.5656, + "step": 17738 + }, + { + "epoch": 1.190530519110097, + "grad_norm": 0.925208330154419, + "learning_rate": 3.7142175855637606e-05, + "loss": 0.4698, + "step": 17740 + }, + { + "epoch": 1.190664742793866, + "grad_norm": 0.93855881690979, + "learning_rate": 3.7131673356637464e-05, + "loss": 0.4955, + "step": 17742 + }, + { + "epoch": 1.190798966477635, + "grad_norm": 1.0195592641830444, + "learning_rate": 3.7121171465640425e-05, + "loss": 0.5211, + "step": 17744 + }, + { + "epoch": 1.190933190161404, + "grad_norm": 0.9680092930793762, + "learning_rate": 3.7110670183142735e-05, + "loss": 0.5084, + "step": 17746 + }, + { + "epoch": 1.191067413845173, + "grad_norm": 0.9840260744094849, + "learning_rate": 3.7100169509640517e-05, + "loss": 0.5218, + "step": 17748 + }, + { + "epoch": 1.191201637528942, + "grad_norm": 1.0281223058700562, + "learning_rate": 3.708966944562993e-05, + "loss": 0.5306, + "step": 17750 + }, + { + "epoch": 1.191335861212711, + "grad_norm": 0.9090124368667603, + "learning_rate": 3.707916999160706e-05, + "loss": 0.4871, + "step": 17752 + }, + { + "epoch": 1.1914700848964799, + "grad_norm": 1.0749856233596802, + "learning_rate": 3.706867114806803e-05, + "loss": 0.4986, + "step": 17754 + }, + { + "epoch": 1.191604308580249, + "grad_norm": 0.9622864127159119, + "learning_rate": 3.705817291550884e-05, + "loss": 0.5304, + "step": 17756 + }, + { + "epoch": 1.191738532264018, + "grad_norm": 1.1090136766433716, + "learning_rate": 3.7047675294425535e-05, + "loss": 0.574, + "step": 17758 + }, + { + "epoch": 1.191872755947787, + "grad_norm": 1.0562835931777954, + "learning_rate": 3.70371782853141e-05, + "loss": 0.4782, + "step": 17760 + }, + { + "epoch": 1.192006979631556, + "grad_norm": 1.0050835609436035, + "learning_rate": 3.702668188867051e-05, + "loss": 0.5709, + "step": 17762 + }, + { + "epoch": 1.192141203315325, + "grad_norm": 0.9055975675582886, + "learning_rate": 3.701618610499068e-05, + "loss": 0.4767, + "step": 17764 + }, + { + "epoch": 1.192275426999094, + "grad_norm": 1.0100618600845337, + "learning_rate": 3.7005690934770534e-05, + "loss": 0.4838, + "step": 17766 + }, + { + "epoch": 1.192409650682863, + "grad_norm": 1.1675429344177246, + "learning_rate": 3.6995196378505926e-05, + "loss": 0.5449, + "step": 17768 + }, + { + "epoch": 1.192543874366632, + "grad_norm": 1.0675374269485474, + "learning_rate": 3.6984702436692726e-05, + "loss": 0.5021, + "step": 17770 + }, + { + "epoch": 1.192678098050401, + "grad_norm": 1.0601750612258911, + "learning_rate": 3.6974209109826726e-05, + "loss": 0.5486, + "step": 17772 + }, + { + "epoch": 1.19281232173417, + "grad_norm": 1.0038232803344727, + "learning_rate": 3.696371639840376e-05, + "loss": 0.4954, + "step": 17774 + }, + { + "epoch": 1.1929465454179389, + "grad_norm": 1.0992913246154785, + "learning_rate": 3.695322430291953e-05, + "loss": 0.5561, + "step": 17776 + }, + { + "epoch": 1.193080769101708, + "grad_norm": 0.9731093049049377, + "learning_rate": 3.694273282386981e-05, + "loss": 0.4909, + "step": 17778 + }, + { + "epoch": 1.193214992785477, + "grad_norm": 1.0533447265625, + "learning_rate": 3.693224196175029e-05, + "loss": 0.5425, + "step": 17780 + }, + { + "epoch": 1.193349216469246, + "grad_norm": 0.852382481098175, + "learning_rate": 3.692175171705664e-05, + "loss": 0.5027, + "step": 17782 + }, + { + "epoch": 1.193483440153015, + "grad_norm": 1.2404472827911377, + "learning_rate": 3.6911262090284506e-05, + "loss": 0.5075, + "step": 17784 + }, + { + "epoch": 1.193617663836784, + "grad_norm": 1.0792241096496582, + "learning_rate": 3.6900773081929496e-05, + "loss": 0.4673, + "step": 17786 + }, + { + "epoch": 1.193751887520553, + "grad_norm": 0.8949421048164368, + "learning_rate": 3.689028469248721e-05, + "loss": 0.4906, + "step": 17788 + }, + { + "epoch": 1.193886111204322, + "grad_norm": 1.0198348760604858, + "learning_rate": 3.687979692245318e-05, + "loss": 0.5639, + "step": 17790 + }, + { + "epoch": 1.1940203348880911, + "grad_norm": 1.091529369354248, + "learning_rate": 3.6869309772322955e-05, + "loss": 0.5408, + "step": 17792 + }, + { + "epoch": 1.19415455857186, + "grad_norm": 0.9910798072814941, + "learning_rate": 3.685882324259201e-05, + "loss": 0.5494, + "step": 17794 + }, + { + "epoch": 1.194288782255629, + "grad_norm": 1.0622038841247559, + "learning_rate": 3.684833733375584e-05, + "loss": 0.4839, + "step": 17796 + }, + { + "epoch": 1.194423005939398, + "grad_norm": 0.9485304355621338, + "learning_rate": 3.683785204630986e-05, + "loss": 0.5038, + "step": 17798 + }, + { + "epoch": 1.194557229623167, + "grad_norm": 1.4749153852462769, + "learning_rate": 3.68273673807495e-05, + "loss": 0.5175, + "step": 17800 + }, + { + "epoch": 1.194691453306936, + "grad_norm": 1.138503909111023, + "learning_rate": 3.681688333757011e-05, + "loss": 0.5002, + "step": 17802 + }, + { + "epoch": 1.194825676990705, + "grad_norm": 0.9076254963874817, + "learning_rate": 3.6806399917267074e-05, + "loss": 0.5075, + "step": 17804 + }, + { + "epoch": 1.194959900674474, + "grad_norm": 1.083914875984192, + "learning_rate": 3.679591712033568e-05, + "loss": 0.4778, + "step": 17806 + }, + { + "epoch": 1.195094124358243, + "grad_norm": 1.0164463520050049, + "learning_rate": 3.678543494727126e-05, + "loss": 0.5803, + "step": 17808 + }, + { + "epoch": 1.195228348042012, + "grad_norm": 0.9743123650550842, + "learning_rate": 3.677495339856903e-05, + "loss": 0.5415, + "step": 17810 + }, + { + "epoch": 1.195362571725781, + "grad_norm": 1.047502875328064, + "learning_rate": 3.676447247472428e-05, + "loss": 0.5183, + "step": 17812 + }, + { + "epoch": 1.1954967954095501, + "grad_norm": 1.0990514755249023, + "learning_rate": 3.6753992176232136e-05, + "loss": 0.4921, + "step": 17814 + }, + { + "epoch": 1.195631019093319, + "grad_norm": 0.9854269623756409, + "learning_rate": 3.6743512503587844e-05, + "loss": 0.4958, + "step": 17816 + }, + { + "epoch": 1.195765242777088, + "grad_norm": 1.0102423429489136, + "learning_rate": 3.673303345728651e-05, + "loss": 0.4884, + "step": 17818 + }, + { + "epoch": 1.195899466460857, + "grad_norm": 1.0501770973205566, + "learning_rate": 3.672255503782326e-05, + "loss": 0.5247, + "step": 17820 + }, + { + "epoch": 1.196033690144626, + "grad_norm": 0.9007217288017273, + "learning_rate": 3.671207724569317e-05, + "loss": 0.5295, + "step": 17822 + }, + { + "epoch": 1.196167913828395, + "grad_norm": 0.9345261454582214, + "learning_rate": 3.670160008139131e-05, + "loss": 0.472, + "step": 17824 + }, + { + "epoch": 1.196302137512164, + "grad_norm": 1.0468041896820068, + "learning_rate": 3.669112354541269e-05, + "loss": 0.4921, + "step": 17826 + }, + { + "epoch": 1.196436361195933, + "grad_norm": 1.0646734237670898, + "learning_rate": 3.668064763825231e-05, + "loss": 0.5565, + "step": 17828 + }, + { + "epoch": 1.196570584879702, + "grad_norm": 0.9364396929740906, + "learning_rate": 3.6670172360405136e-05, + "loss": 0.4596, + "step": 17830 + }, + { + "epoch": 1.196704808563471, + "grad_norm": 0.9761340022087097, + "learning_rate": 3.6659697712366116e-05, + "loss": 0.523, + "step": 17832 + }, + { + "epoch": 1.19683903224724, + "grad_norm": 0.9941263198852539, + "learning_rate": 3.664922369463012e-05, + "loss": 0.4505, + "step": 17834 + }, + { + "epoch": 1.1969732559310091, + "grad_norm": 0.9420185089111328, + "learning_rate": 3.663875030769209e-05, + "loss": 0.4849, + "step": 17836 + }, + { + "epoch": 1.197107479614778, + "grad_norm": 1.0919946432113647, + "learning_rate": 3.66282775520468e-05, + "loss": 0.4925, + "step": 17838 + }, + { + "epoch": 1.197241703298547, + "grad_norm": 1.6442017555236816, + "learning_rate": 3.6617805428189135e-05, + "loss": 0.5326, + "step": 17840 + }, + { + "epoch": 1.197375926982316, + "grad_norm": 0.9553593993186951, + "learning_rate": 3.660733393661381e-05, + "loss": 0.5323, + "step": 17842 + }, + { + "epoch": 1.197510150666085, + "grad_norm": 0.9864370226860046, + "learning_rate": 3.6596863077815644e-05, + "loss": 0.5602, + "step": 17844 + }, + { + "epoch": 1.197644374349854, + "grad_norm": 1.0713117122650146, + "learning_rate": 3.658639285228934e-05, + "loss": 0.6159, + "step": 17846 + }, + { + "epoch": 1.197778598033623, + "grad_norm": 0.8509141802787781, + "learning_rate": 3.65759232605296e-05, + "loss": 0.4617, + "step": 17848 + }, + { + "epoch": 1.197912821717392, + "grad_norm": 0.9234373569488525, + "learning_rate": 3.656545430303108e-05, + "loss": 0.4666, + "step": 17850 + }, + { + "epoch": 1.198047045401161, + "grad_norm": 0.9886850118637085, + "learning_rate": 3.655498598028844e-05, + "loss": 0.5268, + "step": 17852 + }, + { + "epoch": 1.19818126908493, + "grad_norm": 0.9016915559768677, + "learning_rate": 3.654451829279626e-05, + "loss": 0.4801, + "step": 17854 + }, + { + "epoch": 1.198315492768699, + "grad_norm": 1.187814474105835, + "learning_rate": 3.6534051241049137e-05, + "loss": 0.5369, + "step": 17856 + }, + { + "epoch": 1.1984497164524681, + "grad_norm": 0.9405223727226257, + "learning_rate": 3.652358482554162e-05, + "loss": 0.5279, + "step": 17858 + }, + { + "epoch": 1.198583940136237, + "grad_norm": 0.956626296043396, + "learning_rate": 3.6513119046768206e-05, + "loss": 0.5491, + "step": 17860 + }, + { + "epoch": 1.198718163820006, + "grad_norm": 1.0439075231552124, + "learning_rate": 3.65026539052234e-05, + "loss": 0.5631, + "step": 17862 + }, + { + "epoch": 1.198852387503775, + "grad_norm": 0.9437693357467651, + "learning_rate": 3.6492189401401656e-05, + "loss": 0.5267, + "step": 17864 + }, + { + "epoch": 1.198986611187544, + "grad_norm": 1.1598464250564575, + "learning_rate": 3.6481725535797415e-05, + "loss": 0.5814, + "step": 17866 + }, + { + "epoch": 1.1991208348713132, + "grad_norm": 0.9224652051925659, + "learning_rate": 3.647126230890503e-05, + "loss": 0.5349, + "step": 17868 + }, + { + "epoch": 1.199255058555082, + "grad_norm": 0.9480599164962769, + "learning_rate": 3.646079972121893e-05, + "loss": 0.5187, + "step": 17870 + }, + { + "epoch": 1.199389282238851, + "grad_norm": 1.0072935819625854, + "learning_rate": 3.645033777323339e-05, + "loss": 0.4771, + "step": 17872 + }, + { + "epoch": 1.1995235059226201, + "grad_norm": 0.9206082224845886, + "learning_rate": 3.643987646544278e-05, + "loss": 0.568, + "step": 17874 + }, + { + "epoch": 1.199657729606389, + "grad_norm": 0.9697580337524414, + "learning_rate": 3.642941579834131e-05, + "loss": 0.4322, + "step": 17876 + }, + { + "epoch": 1.199791953290158, + "grad_norm": 0.9863641262054443, + "learning_rate": 3.641895577242327e-05, + "loss": 0.5506, + "step": 17878 + }, + { + "epoch": 1.1999261769739271, + "grad_norm": 0.8394425511360168, + "learning_rate": 3.640849638818286e-05, + "loss": 0.4906, + "step": 17880 + }, + { + "epoch": 1.200060400657696, + "grad_norm": 0.9852147698402405, + "learning_rate": 3.639803764611428e-05, + "loss": 0.4999, + "step": 17882 + }, + { + "epoch": 1.200194624341465, + "grad_norm": 1.240363597869873, + "learning_rate": 3.6387579546711656e-05, + "loss": 0.5641, + "step": 17884 + }, + { + "epoch": 1.200328848025234, + "grad_norm": 1.0349351167678833, + "learning_rate": 3.637712209046915e-05, + "loss": 0.4971, + "step": 17886 + }, + { + "epoch": 1.200463071709003, + "grad_norm": 0.9062631726264954, + "learning_rate": 3.636666527788083e-05, + "loss": 0.4782, + "step": 17888 + }, + { + "epoch": 1.2005972953927722, + "grad_norm": 1.0077486038208008, + "learning_rate": 3.635620910944077e-05, + "loss": 0.4889, + "step": 17890 + }, + { + "epoch": 1.200731519076541, + "grad_norm": 1.0355712175369263, + "learning_rate": 3.6345753585642996e-05, + "loss": 0.5339, + "step": 17892 + }, + { + "epoch": 1.20086574276031, + "grad_norm": 0.9321998357772827, + "learning_rate": 3.6335298706981517e-05, + "loss": 0.5161, + "step": 17894 + }, + { + "epoch": 1.2009999664440791, + "grad_norm": 0.967764675617218, + "learning_rate": 3.632484447395029e-05, + "loss": 0.6395, + "step": 17896 + }, + { + "epoch": 1.201134190127848, + "grad_norm": 1.0610249042510986, + "learning_rate": 3.63143908870433e-05, + "loss": 0.4735, + "step": 17898 + }, + { + "epoch": 1.201268413811617, + "grad_norm": 0.9627799391746521, + "learning_rate": 3.63039379467544e-05, + "loss": 0.518, + "step": 17900 + }, + { + "epoch": 1.2014026374953861, + "grad_norm": 0.9245826005935669, + "learning_rate": 3.629348565357752e-05, + "loss": 0.4635, + "step": 17902 + }, + { + "epoch": 1.201536861179155, + "grad_norm": 0.9800781011581421, + "learning_rate": 3.6283034008006465e-05, + "loss": 0.4852, + "step": 17904 + }, + { + "epoch": 1.201671084862924, + "grad_norm": 1.0018424987792969, + "learning_rate": 3.627258301053511e-05, + "loss": 0.5053, + "step": 17906 + }, + { + "epoch": 1.201805308546693, + "grad_norm": 1.0354385375976562, + "learning_rate": 3.626213266165719e-05, + "loss": 0.5597, + "step": 17908 + }, + { + "epoch": 1.201939532230462, + "grad_norm": 2.0732288360595703, + "learning_rate": 3.6251682961866506e-05, + "loss": 0.5155, + "step": 17910 + }, + { + "epoch": 1.2020737559142312, + "grad_norm": 0.9796667098999023, + "learning_rate": 3.6241233911656755e-05, + "loss": 0.4653, + "step": 17912 + }, + { + "epoch": 1.202207979598, + "grad_norm": 1.0895826816558838, + "learning_rate": 3.6230785511521656e-05, + "loss": 0.4678, + "step": 17914 + }, + { + "epoch": 1.202342203281769, + "grad_norm": 1.074872612953186, + "learning_rate": 3.6220337761954855e-05, + "loss": 0.4601, + "step": 17916 + }, + { + "epoch": 1.2024764269655381, + "grad_norm": 1.085161805152893, + "learning_rate": 3.620989066345001e-05, + "loss": 0.5764, + "step": 17918 + }, + { + "epoch": 1.202610650649307, + "grad_norm": 1.0547800064086914, + "learning_rate": 3.61994442165007e-05, + "loss": 0.4965, + "step": 17920 + }, + { + "epoch": 1.202744874333076, + "grad_norm": 0.9630693793296814, + "learning_rate": 3.618899842160053e-05, + "loss": 0.4983, + "step": 17922 + }, + { + "epoch": 1.2028790980168451, + "grad_norm": 0.9729456901550293, + "learning_rate": 3.617855327924302e-05, + "loss": 0.5085, + "step": 17924 + }, + { + "epoch": 1.203013321700614, + "grad_norm": 1.0706628561019897, + "learning_rate": 3.616810878992168e-05, + "loss": 0.5246, + "step": 17926 + }, + { + "epoch": 1.203147545384383, + "grad_norm": 0.9929572939872742, + "learning_rate": 3.6157664954130014e-05, + "loss": 0.4808, + "step": 17928 + }, + { + "epoch": 1.203281769068152, + "grad_norm": 1.0386592149734497, + "learning_rate": 3.6147221772361446e-05, + "loss": 0.5013, + "step": 17930 + }, + { + "epoch": 1.203415992751921, + "grad_norm": 1.1223738193511963, + "learning_rate": 3.613677924510942e-05, + "loss": 0.4968, + "step": 17932 + }, + { + "epoch": 1.2035502164356902, + "grad_norm": 1.0150068998336792, + "learning_rate": 3.612633737286729e-05, + "loss": 0.5536, + "step": 17934 + }, + { + "epoch": 1.203684440119459, + "grad_norm": 0.9773173332214355, + "learning_rate": 3.611589615612847e-05, + "loss": 0.5046, + "step": 17936 + }, + { + "epoch": 1.203818663803228, + "grad_norm": 0.9851709604263306, + "learning_rate": 3.610545559538623e-05, + "loss": 0.5024, + "step": 17938 + }, + { + "epoch": 1.2039528874869971, + "grad_norm": 0.9655515551567078, + "learning_rate": 3.6095015691133904e-05, + "loss": 0.5187, + "step": 17940 + }, + { + "epoch": 1.204087111170766, + "grad_norm": 1.0774941444396973, + "learning_rate": 3.608457644386474e-05, + "loss": 0.5763, + "step": 17942 + }, + { + "epoch": 1.2042213348545352, + "grad_norm": 1.0457144975662231, + "learning_rate": 3.6074137854071976e-05, + "loss": 0.5097, + "step": 17944 + }, + { + "epoch": 1.2043555585383041, + "grad_norm": 0.9518623352050781, + "learning_rate": 3.606369992224882e-05, + "loss": 0.4483, + "step": 17946 + }, + { + "epoch": 1.204489782222073, + "grad_norm": 0.9342052936553955, + "learning_rate": 3.605326264888843e-05, + "loss": 0.5088, + "step": 17948 + }, + { + "epoch": 1.2046240059058422, + "grad_norm": 1.091081142425537, + "learning_rate": 3.6042826034483956e-05, + "loss": 0.5486, + "step": 17950 + }, + { + "epoch": 1.204758229589611, + "grad_norm": 1.0078307390213013, + "learning_rate": 3.603239007952851e-05, + "loss": 0.5193, + "step": 17952 + }, + { + "epoch": 1.20489245327338, + "grad_norm": 0.965714156627655, + "learning_rate": 3.602195478451516e-05, + "loss": 0.519, + "step": 17954 + }, + { + "epoch": 1.2050266769571492, + "grad_norm": 0.8684780597686768, + "learning_rate": 3.6011520149936975e-05, + "loss": 0.4679, + "step": 17956 + }, + { + "epoch": 1.205160900640918, + "grad_norm": 0.9268240928649902, + "learning_rate": 3.600108617628692e-05, + "loss": 0.5211, + "step": 17958 + }, + { + "epoch": 1.205295124324687, + "grad_norm": 1.1090248823165894, + "learning_rate": 3.599065286405806e-05, + "loss": 0.5528, + "step": 17960 + }, + { + "epoch": 1.2054293480084561, + "grad_norm": 1.0047954320907593, + "learning_rate": 3.598022021374327e-05, + "loss": 0.496, + "step": 17962 + }, + { + "epoch": 1.205563571692225, + "grad_norm": 1.0637043714523315, + "learning_rate": 3.596978822583554e-05, + "loss": 0.4953, + "step": 17964 + }, + { + "epoch": 1.2056977953759942, + "grad_norm": 1.0383027791976929, + "learning_rate": 3.595935690082769e-05, + "loss": 0.5041, + "step": 17966 + }, + { + "epoch": 1.2058320190597631, + "grad_norm": 0.9820025563240051, + "learning_rate": 3.594892623921264e-05, + "loss": 0.5287, + "step": 17968 + }, + { + "epoch": 1.205966242743532, + "grad_norm": 1.118740200996399, + "learning_rate": 3.593849624148319e-05, + "loss": 0.5667, + "step": 17970 + }, + { + "epoch": 1.2061004664273012, + "grad_norm": 1.0724904537200928, + "learning_rate": 3.5928066908132144e-05, + "loss": 0.5043, + "step": 17972 + }, + { + "epoch": 1.20623469011107, + "grad_norm": 1.0753625631332397, + "learning_rate": 3.591763823965226e-05, + "loss": 0.5786, + "step": 17974 + }, + { + "epoch": 1.206368913794839, + "grad_norm": 1.0068565607070923, + "learning_rate": 3.590721023653628e-05, + "loss": 0.5206, + "step": 17976 + }, + { + "epoch": 1.2065031374786082, + "grad_norm": 0.9540721774101257, + "learning_rate": 3.5896782899276905e-05, + "loss": 0.4794, + "step": 17978 + }, + { + "epoch": 1.206637361162377, + "grad_norm": 1.0614107847213745, + "learning_rate": 3.58863562283668e-05, + "loss": 0.5097, + "step": 17980 + }, + { + "epoch": 1.206771584846146, + "grad_norm": 1.1896748542785645, + "learning_rate": 3.5875930224298616e-05, + "loss": 0.4741, + "step": 17982 + }, + { + "epoch": 1.2069058085299151, + "grad_norm": 1.0033884048461914, + "learning_rate": 3.586550488756496e-05, + "loss": 0.5514, + "step": 17984 + }, + { + "epoch": 1.207040032213684, + "grad_norm": 1.0701922178268433, + "learning_rate": 3.585508021865838e-05, + "loss": 0.54, + "step": 17986 + }, + { + "epoch": 1.2071742558974532, + "grad_norm": 0.9692116379737854, + "learning_rate": 3.584465621807148e-05, + "loss": 0.5118, + "step": 17988 + }, + { + "epoch": 1.2073084795812221, + "grad_norm": 0.9265748262405396, + "learning_rate": 3.58342328862967e-05, + "loss": 0.5253, + "step": 17990 + }, + { + "epoch": 1.207442703264991, + "grad_norm": 1.0179463624954224, + "learning_rate": 3.58238102238266e-05, + "loss": 0.5352, + "step": 17992 + }, + { + "epoch": 1.2075769269487602, + "grad_norm": 1.0018413066864014, + "learning_rate": 3.5813388231153576e-05, + "loss": 0.5197, + "step": 17994 + }, + { + "epoch": 1.207711150632529, + "grad_norm": 0.993140697479248, + "learning_rate": 3.5802966908770044e-05, + "loss": 0.4556, + "step": 17996 + }, + { + "epoch": 1.207845374316298, + "grad_norm": 0.9674150347709656, + "learning_rate": 3.579254625716844e-05, + "loss": 0.4942, + "step": 17998 + }, + { + "epoch": 1.2079795980000672, + "grad_norm": 1.0136640071868896, + "learning_rate": 3.5782126276841045e-05, + "loss": 0.5252, + "step": 18000 + }, + { + "epoch": 1.208113821683836, + "grad_norm": 1.0575333833694458, + "learning_rate": 3.577170696828026e-05, + "loss": 0.4999, + "step": 18002 + }, + { + "epoch": 1.208248045367605, + "grad_norm": 0.998807430267334, + "learning_rate": 3.576128833197832e-05, + "loss": 0.478, + "step": 18004 + }, + { + "epoch": 1.2083822690513741, + "grad_norm": 1.1003563404083252, + "learning_rate": 3.5750870368427526e-05, + "loss": 0.5106, + "step": 18006 + }, + { + "epoch": 1.208516492735143, + "grad_norm": 0.9529297947883606, + "learning_rate": 3.574045307812007e-05, + "loss": 0.4817, + "step": 18008 + }, + { + "epoch": 1.2086507164189122, + "grad_norm": 0.9703645706176758, + "learning_rate": 3.573003646154819e-05, + "loss": 0.5092, + "step": 18010 + }, + { + "epoch": 1.2087849401026811, + "grad_norm": 0.9716821312904358, + "learning_rate": 3.5719620519204e-05, + "loss": 0.4638, + "step": 18012 + }, + { + "epoch": 1.20891916378645, + "grad_norm": 1.0448511838912964, + "learning_rate": 3.570920525157968e-05, + "loss": 0.5714, + "step": 18014 + }, + { + "epoch": 1.2090533874702192, + "grad_norm": 1.0114126205444336, + "learning_rate": 3.569879065916729e-05, + "loss": 0.5992, + "step": 18016 + }, + { + "epoch": 1.209187611153988, + "grad_norm": 0.994085431098938, + "learning_rate": 3.568837674245894e-05, + "loss": 0.4827, + "step": 18018 + }, + { + "epoch": 1.2093218348377572, + "grad_norm": 0.9882687330245972, + "learning_rate": 3.5677963501946634e-05, + "loss": 0.5636, + "step": 18020 + }, + { + "epoch": 1.2094560585215262, + "grad_norm": 1.0266276597976685, + "learning_rate": 3.5667550938122396e-05, + "loss": 0.5164, + "step": 18022 + }, + { + "epoch": 1.209590282205295, + "grad_norm": 1.452270746231079, + "learning_rate": 3.565713905147817e-05, + "loss": 0.5519, + "step": 18024 + }, + { + "epoch": 1.2097245058890642, + "grad_norm": 0.9671774506568909, + "learning_rate": 3.5646727842505955e-05, + "loss": 0.5172, + "step": 18026 + }, + { + "epoch": 1.2098587295728331, + "grad_norm": 1.0094542503356934, + "learning_rate": 3.56363173116976e-05, + "loss": 0.5207, + "step": 18028 + }, + { + "epoch": 1.209992953256602, + "grad_norm": 1.1328052282333374, + "learning_rate": 3.562590745954501e-05, + "loss": 0.5359, + "step": 18030 + }, + { + "epoch": 1.2101271769403712, + "grad_norm": 1.0051718950271606, + "learning_rate": 3.561549828654003e-05, + "loss": 0.5685, + "step": 18032 + }, + { + "epoch": 1.2102614006241401, + "grad_norm": 1.0846080780029297, + "learning_rate": 3.5605089793174485e-05, + "loss": 0.5447, + "step": 18034 + }, + { + "epoch": 1.210395624307909, + "grad_norm": 1.0940159559249878, + "learning_rate": 3.559468197994012e-05, + "loss": 0.469, + "step": 18036 + }, + { + "epoch": 1.2105298479916782, + "grad_norm": 1.0636924505233765, + "learning_rate": 3.558427484732872e-05, + "loss": 0.4979, + "step": 18038 + }, + { + "epoch": 1.210664071675447, + "grad_norm": 1.1106228828430176, + "learning_rate": 3.5573868395831975e-05, + "loss": 0.4764, + "step": 18040 + }, + { + "epoch": 1.2107982953592162, + "grad_norm": 1.08779776096344, + "learning_rate": 3.556346262594159e-05, + "loss": 0.5086, + "step": 18042 + }, + { + "epoch": 1.2109325190429852, + "grad_norm": 0.8930120468139648, + "learning_rate": 3.5553057538149196e-05, + "loss": 0.5473, + "step": 18044 + }, + { + "epoch": 1.211066742726754, + "grad_norm": 3.5846095085144043, + "learning_rate": 3.5542653132946435e-05, + "loss": 0.5397, + "step": 18046 + }, + { + "epoch": 1.2112009664105232, + "grad_norm": 1.1256413459777832, + "learning_rate": 3.553224941082487e-05, + "loss": 0.587, + "step": 18048 + }, + { + "epoch": 1.2113351900942921, + "grad_norm": 1.0590184926986694, + "learning_rate": 3.552184637227609e-05, + "loss": 0.5121, + "step": 18050 + }, + { + "epoch": 1.211469413778061, + "grad_norm": 0.9384714961051941, + "learning_rate": 3.5511444017791575e-05, + "loss": 0.5786, + "step": 18052 + }, + { + "epoch": 1.2116036374618302, + "grad_norm": 1.056459665298462, + "learning_rate": 3.550104234786287e-05, + "loss": 0.5226, + "step": 18054 + }, + { + "epoch": 1.2117378611455991, + "grad_norm": 1.0418684482574463, + "learning_rate": 3.5490641362981374e-05, + "loss": 0.5002, + "step": 18056 + }, + { + "epoch": 1.211872084829368, + "grad_norm": 1.3959248065948486, + "learning_rate": 3.5480241063638566e-05, + "loss": 0.5453, + "step": 18058 + }, + { + "epoch": 1.2120063085131372, + "grad_norm": 0.9421146512031555, + "learning_rate": 3.546984145032582e-05, + "loss": 0.4287, + "step": 18060 + }, + { + "epoch": 1.212140532196906, + "grad_norm": 0.9701901078224182, + "learning_rate": 3.545944252353446e-05, + "loss": 0.4897, + "step": 18062 + }, + { + "epoch": 1.2122747558806752, + "grad_norm": 1.0220611095428467, + "learning_rate": 3.5449044283755874e-05, + "loss": 0.5086, + "step": 18064 + }, + { + "epoch": 1.2124089795644442, + "grad_norm": 1.0941497087478638, + "learning_rate": 3.543864673148132e-05, + "loss": 0.4985, + "step": 18066 + }, + { + "epoch": 1.212543203248213, + "grad_norm": 0.967139720916748, + "learning_rate": 3.542824986720209e-05, + "loss": 0.5037, + "step": 18068 + }, + { + "epoch": 1.2126774269319822, + "grad_norm": 1.2793251276016235, + "learning_rate": 3.541785369140938e-05, + "loss": 0.5264, + "step": 18070 + }, + { + "epoch": 1.2128116506157511, + "grad_norm": 1.0796058177947998, + "learning_rate": 3.5407458204594426e-05, + "loss": 0.5904, + "step": 18072 + }, + { + "epoch": 1.21294587429952, + "grad_norm": 1.0878667831420898, + "learning_rate": 3.5397063407248365e-05, + "loss": 0.5369, + "step": 18074 + }, + { + "epoch": 1.2130800979832892, + "grad_norm": 1.0877381563186646, + "learning_rate": 3.5386669299862355e-05, + "loss": 0.5289, + "step": 18076 + }, + { + "epoch": 1.2132143216670581, + "grad_norm": 1.190150260925293, + "learning_rate": 3.5376275882927466e-05, + "loss": 0.5311, + "step": 18078 + }, + { + "epoch": 1.213348545350827, + "grad_norm": 0.9739012718200684, + "learning_rate": 3.5365883156934795e-05, + "loss": 0.4665, + "step": 18080 + }, + { + "epoch": 1.2134827690345962, + "grad_norm": 0.9805138111114502, + "learning_rate": 3.535549112237537e-05, + "loss": 0.505, + "step": 18082 + }, + { + "epoch": 1.213616992718365, + "grad_norm": 0.8651003241539001, + "learning_rate": 3.534509977974019e-05, + "loss": 0.4986, + "step": 18084 + }, + { + "epoch": 1.2137512164021342, + "grad_norm": 1.2091014385223389, + "learning_rate": 3.533470912952022e-05, + "loss": 0.5289, + "step": 18086 + }, + { + "epoch": 1.2138854400859032, + "grad_norm": 0.9878976941108704, + "learning_rate": 3.532431917220642e-05, + "loss": 0.4601, + "step": 18088 + }, + { + "epoch": 1.214019663769672, + "grad_norm": 0.9306507706642151, + "learning_rate": 3.5313929908289665e-05, + "loss": 0.5849, + "step": 18090 + }, + { + "epoch": 1.2141538874534412, + "grad_norm": 0.9268865585327148, + "learning_rate": 3.5303541338260856e-05, + "loss": 0.479, + "step": 18092 + }, + { + "epoch": 1.2142881111372101, + "grad_norm": 1.228854775428772, + "learning_rate": 3.529315346261081e-05, + "loss": 0.509, + "step": 18094 + }, + { + "epoch": 1.2144223348209793, + "grad_norm": 0.8898868560791016, + "learning_rate": 3.528276628183035e-05, + "loss": 0.4212, + "step": 18096 + }, + { + "epoch": 1.2145565585047482, + "grad_norm": 0.9570024013519287, + "learning_rate": 3.527237979641024e-05, + "loss": 0.502, + "step": 18098 + }, + { + "epoch": 1.2146907821885171, + "grad_norm": 1.054985523223877, + "learning_rate": 3.526199400684124e-05, + "loss": 0.4765, + "step": 18100 + }, + { + "epoch": 1.2148250058722863, + "grad_norm": 1.0664186477661133, + "learning_rate": 3.5251608913614014e-05, + "loss": 0.5318, + "step": 18102 + }, + { + "epoch": 1.2149592295560552, + "grad_norm": 0.9964617490768433, + "learning_rate": 3.5241224517219286e-05, + "loss": 0.5151, + "step": 18104 + }, + { + "epoch": 1.215093453239824, + "grad_norm": 0.9645223021507263, + "learning_rate": 3.523084081814767e-05, + "loss": 0.5289, + "step": 18106 + }, + { + "epoch": 1.2152276769235932, + "grad_norm": 1.088595986366272, + "learning_rate": 3.5220457816889784e-05, + "loss": 0.4541, + "step": 18108 + }, + { + "epoch": 1.2153619006073622, + "grad_norm": 1.0221766233444214, + "learning_rate": 3.5210075513936206e-05, + "loss": 0.5183, + "step": 18110 + }, + { + "epoch": 1.215496124291131, + "grad_norm": 0.8846173882484436, + "learning_rate": 3.519969390977748e-05, + "loss": 0.4248, + "step": 18112 + }, + { + "epoch": 1.2156303479749002, + "grad_norm": 1.0733497142791748, + "learning_rate": 3.518931300490409e-05, + "loss": 0.5247, + "step": 18114 + }, + { + "epoch": 1.2157645716586691, + "grad_norm": 1.0206555128097534, + "learning_rate": 3.517893279980656e-05, + "loss": 0.5308, + "step": 18116 + }, + { + "epoch": 1.2158987953424383, + "grad_norm": 1.1100720167160034, + "learning_rate": 3.516855329497529e-05, + "loss": 0.4776, + "step": 18118 + }, + { + "epoch": 1.2160330190262072, + "grad_norm": 0.9295620918273926, + "learning_rate": 3.5158174490900744e-05, + "loss": 0.4824, + "step": 18120 + }, + { + "epoch": 1.2161672427099761, + "grad_norm": 1.0154778957366943, + "learning_rate": 3.514779638807324e-05, + "loss": 0.5329, + "step": 18122 + }, + { + "epoch": 1.2163014663937453, + "grad_norm": 1.0279161930084229, + "learning_rate": 3.513741898698317e-05, + "loss": 0.4524, + "step": 18124 + }, + { + "epoch": 1.2164356900775142, + "grad_norm": 1.1146999597549438, + "learning_rate": 3.512704228812082e-05, + "loss": 0.4705, + "step": 18126 + }, + { + "epoch": 1.216569913761283, + "grad_norm": 1.072011113166809, + "learning_rate": 3.5116666291976474e-05, + "loss": 0.5498, + "step": 18128 + }, + { + "epoch": 1.2167041374450522, + "grad_norm": 1.0384671688079834, + "learning_rate": 3.510629099904038e-05, + "loss": 0.4668, + "step": 18130 + }, + { + "epoch": 1.2168383611288212, + "grad_norm": 0.9664222598075867, + "learning_rate": 3.509591640980274e-05, + "loss": 0.4974, + "step": 18132 + }, + { + "epoch": 1.21697258481259, + "grad_norm": 1.051636815071106, + "learning_rate": 3.508554252475376e-05, + "loss": 0.5368, + "step": 18134 + }, + { + "epoch": 1.2171068084963592, + "grad_norm": 0.8387489914894104, + "learning_rate": 3.507516934438355e-05, + "loss": 0.4544, + "step": 18136 + }, + { + "epoch": 1.2172410321801281, + "grad_norm": 1.011088490486145, + "learning_rate": 3.506479686918226e-05, + "loss": 0.5168, + "step": 18138 + }, + { + "epoch": 1.2173752558638973, + "grad_norm": 0.9723619222640991, + "learning_rate": 3.505442509963993e-05, + "loss": 0.4749, + "step": 18140 + }, + { + "epoch": 1.2175094795476662, + "grad_norm": 0.9728646874427795, + "learning_rate": 3.504405403624664e-05, + "loss": 0.494, + "step": 18142 + }, + { + "epoch": 1.2176437032314351, + "grad_norm": 1.091109275817871, + "learning_rate": 3.503368367949237e-05, + "loss": 0.5355, + "step": 18144 + }, + { + "epoch": 1.2177779269152043, + "grad_norm": 1.0038992166519165, + "learning_rate": 3.502331402986713e-05, + "loss": 0.5109, + "step": 18146 + }, + { + "epoch": 1.2179121505989732, + "grad_norm": 1.120638132095337, + "learning_rate": 3.501294508786084e-05, + "loss": 0.4706, + "step": 18148 + }, + { + "epoch": 1.218046374282742, + "grad_norm": 0.9986687898635864, + "learning_rate": 3.500257685396344e-05, + "loss": 0.5479, + "step": 18150 + }, + { + "epoch": 1.2181805979665112, + "grad_norm": 1.1052039861679077, + "learning_rate": 3.4992209328664774e-05, + "loss": 0.5399, + "step": 18152 + }, + { + "epoch": 1.2183148216502802, + "grad_norm": 0.9729425311088562, + "learning_rate": 3.4981842512454716e-05, + "loss": 0.505, + "step": 18154 + }, + { + "epoch": 1.218449045334049, + "grad_norm": 1.1042462587356567, + "learning_rate": 3.497147640582306e-05, + "loss": 0.4806, + "step": 18156 + }, + { + "epoch": 1.2185832690178182, + "grad_norm": 1.1737509965896606, + "learning_rate": 3.4961111009259605e-05, + "loss": 0.5463, + "step": 18158 + }, + { + "epoch": 1.2187174927015871, + "grad_norm": 0.889025092124939, + "learning_rate": 3.495074632325407e-05, + "loss": 0.5417, + "step": 18160 + }, + { + "epoch": 1.2188517163853563, + "grad_norm": 1.0269781351089478, + "learning_rate": 3.4940382348296186e-05, + "loss": 0.5851, + "step": 18162 + }, + { + "epoch": 1.2189859400691252, + "grad_norm": 0.9341825842857361, + "learning_rate": 3.493001908487561e-05, + "loss": 0.5085, + "step": 18164 + }, + { + "epoch": 1.2191201637528941, + "grad_norm": 0.9956706166267395, + "learning_rate": 3.491965653348202e-05, + "loss": 0.5383, + "step": 18166 + }, + { + "epoch": 1.2192543874366633, + "grad_norm": 1.0724912881851196, + "learning_rate": 3.490929469460499e-05, + "loss": 0.5229, + "step": 18168 + }, + { + "epoch": 1.2193886111204322, + "grad_norm": 0.9348374605178833, + "learning_rate": 3.4898933568734115e-05, + "loss": 0.492, + "step": 18170 + }, + { + "epoch": 1.2195228348042013, + "grad_norm": 0.9576309323310852, + "learning_rate": 3.488857315635893e-05, + "loss": 0.4348, + "step": 18172 + }, + { + "epoch": 1.2196570584879702, + "grad_norm": 1.3052387237548828, + "learning_rate": 3.487821345796895e-05, + "loss": 0.4717, + "step": 18174 + }, + { + "epoch": 1.2197912821717392, + "grad_norm": 1.044317364692688, + "learning_rate": 3.4867854474053644e-05, + "loss": 0.4311, + "step": 18176 + }, + { + "epoch": 1.2199255058555083, + "grad_norm": 1.0891656875610352, + "learning_rate": 3.4857496205102474e-05, + "loss": 0.5037, + "step": 18178 + }, + { + "epoch": 1.2200597295392772, + "grad_norm": 1.0032423734664917, + "learning_rate": 3.4847138651604805e-05, + "loss": 0.5302, + "step": 18180 + }, + { + "epoch": 1.2201939532230461, + "grad_norm": 0.8056291937828064, + "learning_rate": 3.483678181405006e-05, + "loss": 0.4807, + "step": 18182 + }, + { + "epoch": 1.2203281769068153, + "grad_norm": 1.104245662689209, + "learning_rate": 3.4826425692927545e-05, + "loss": 0.5011, + "step": 18184 + }, + { + "epoch": 1.2204624005905842, + "grad_norm": 0.9866969585418701, + "learning_rate": 3.481607028872659e-05, + "loss": 0.5141, + "step": 18186 + }, + { + "epoch": 1.2205966242743531, + "grad_norm": 1.0253126621246338, + "learning_rate": 3.480571560193645e-05, + "loss": 0.5778, + "step": 18188 + }, + { + "epoch": 1.2207308479581223, + "grad_norm": 0.9861544370651245, + "learning_rate": 3.479536163304637e-05, + "loss": 0.4452, + "step": 18190 + }, + { + "epoch": 1.2208650716418912, + "grad_norm": 0.9421694874763489, + "learning_rate": 3.4785008382545546e-05, + "loss": 0.5036, + "step": 18192 + }, + { + "epoch": 1.2209992953256603, + "grad_norm": 1.0168511867523193, + "learning_rate": 3.4774655850923174e-05, + "loss": 0.5127, + "step": 18194 + }, + { + "epoch": 1.2211335190094292, + "grad_norm": 1.1280962228775024, + "learning_rate": 3.476430403866836e-05, + "loss": 0.5206, + "step": 18196 + }, + { + "epoch": 1.2212677426931982, + "grad_norm": 0.909195065498352, + "learning_rate": 3.475395294627023e-05, + "loss": 0.4774, + "step": 18198 + }, + { + "epoch": 1.2214019663769673, + "grad_norm": 1.0700228214263916, + "learning_rate": 3.474360257421784e-05, + "loss": 0.4393, + "step": 18200 + }, + { + "epoch": 1.2215361900607362, + "grad_norm": 1.0085256099700928, + "learning_rate": 3.4733252923000226e-05, + "loss": 0.4778, + "step": 18202 + }, + { + "epoch": 1.2216704137445051, + "grad_norm": 0.963640570640564, + "learning_rate": 3.4722903993106395e-05, + "loss": 0.5338, + "step": 18204 + }, + { + "epoch": 1.2218046374282743, + "grad_norm": 1.1290463209152222, + "learning_rate": 3.4712555785025305e-05, + "loss": 0.5166, + "step": 18206 + }, + { + "epoch": 1.2219388611120432, + "grad_norm": 0.9944272637367249, + "learning_rate": 3.4702208299245895e-05, + "loss": 0.4689, + "step": 18208 + }, + { + "epoch": 1.2220730847958121, + "grad_norm": 0.8961957693099976, + "learning_rate": 3.469186153625704e-05, + "loss": 0.5015, + "step": 18210 + }, + { + "epoch": 1.2222073084795813, + "grad_norm": 0.9712990522384644, + "learning_rate": 3.468151549654766e-05, + "loss": 0.4926, + "step": 18212 + }, + { + "epoch": 1.2223415321633502, + "grad_norm": 0.9702814817428589, + "learning_rate": 3.467117018060652e-05, + "loss": 0.4775, + "step": 18214 + }, + { + "epoch": 1.2224757558471193, + "grad_norm": 0.9661044478416443, + "learning_rate": 3.466082558892247e-05, + "loss": 0.5065, + "step": 18216 + }, + { + "epoch": 1.2226099795308882, + "grad_norm": 0.9817913174629211, + "learning_rate": 3.465048172198423e-05, + "loss": 0.4714, + "step": 18218 + }, + { + "epoch": 1.2227442032146572, + "grad_norm": 1.1519465446472168, + "learning_rate": 3.464013858028056e-05, + "loss": 0.5162, + "step": 18220 + }, + { + "epoch": 1.2228784268984263, + "grad_norm": 0.944974958896637, + "learning_rate": 3.462979616430012e-05, + "loss": 0.5193, + "step": 18222 + }, + { + "epoch": 1.2230126505821952, + "grad_norm": 1.0148048400878906, + "learning_rate": 3.461945447453161e-05, + "loss": 0.4923, + "step": 18224 + }, + { + "epoch": 1.2231468742659641, + "grad_norm": 1.0852816104888916, + "learning_rate": 3.4609113511463616e-05, + "loss": 0.4819, + "step": 18226 + }, + { + "epoch": 1.2232810979497333, + "grad_norm": 1.0011029243469238, + "learning_rate": 3.459877327558475e-05, + "loss": 0.5228, + "step": 18228 + }, + { + "epoch": 1.2234153216335022, + "grad_norm": 1.0051935911178589, + "learning_rate": 3.458843376738355e-05, + "loss": 0.5644, + "step": 18230 + }, + { + "epoch": 1.2235495453172711, + "grad_norm": 1.0228527784347534, + "learning_rate": 3.457809498734857e-05, + "loss": 0.478, + "step": 18232 + }, + { + "epoch": 1.2236837690010403, + "grad_norm": 0.8186175227165222, + "learning_rate": 3.456775693596825e-05, + "loss": 0.4832, + "step": 18234 + }, + { + "epoch": 1.2238179926848092, + "grad_norm": 0.9312829971313477, + "learning_rate": 3.455741961373109e-05, + "loss": 0.52, + "step": 18236 + }, + { + "epoch": 1.2239522163685783, + "grad_norm": 1.4396902322769165, + "learning_rate": 3.454708302112547e-05, + "loss": 0.5446, + "step": 18238 + }, + { + "epoch": 1.2240864400523472, + "grad_norm": 0.9807689785957336, + "learning_rate": 3.453674715863982e-05, + "loss": 0.4864, + "step": 18240 + }, + { + "epoch": 1.2242206637361162, + "grad_norm": 0.9718871116638184, + "learning_rate": 3.452641202676242e-05, + "loss": 0.4939, + "step": 18242 + }, + { + "epoch": 1.2243548874198853, + "grad_norm": 1.0906612873077393, + "learning_rate": 3.4516077625981644e-05, + "loss": 0.5018, + "step": 18244 + }, + { + "epoch": 1.2244891111036542, + "grad_norm": 1.0393626689910889, + "learning_rate": 3.450574395678575e-05, + "loss": 0.5002, + "step": 18246 + }, + { + "epoch": 1.2246233347874234, + "grad_norm": 1.05021071434021, + "learning_rate": 3.4495411019662994e-05, + "loss": 0.4996, + "step": 18248 + }, + { + "epoch": 1.2247575584711923, + "grad_norm": 1.0882809162139893, + "learning_rate": 3.4485078815101566e-05, + "loss": 0.5134, + "step": 18250 + }, + { + "epoch": 1.2248917821549612, + "grad_norm": 1.0747679471969604, + "learning_rate": 3.447474734358967e-05, + "loss": 0.4203, + "step": 18252 + }, + { + "epoch": 1.2250260058387303, + "grad_norm": 1.0003612041473389, + "learning_rate": 3.4464416605615424e-05, + "loss": 0.4996, + "step": 18254 + }, + { + "epoch": 1.2251602295224993, + "grad_norm": 1.214212417602539, + "learning_rate": 3.4454086601666945e-05, + "loss": 0.5272, + "step": 18256 + }, + { + "epoch": 1.2252944532062682, + "grad_norm": 1.1573606729507446, + "learning_rate": 3.44437573322323e-05, + "loss": 0.4412, + "step": 18258 + }, + { + "epoch": 1.2254286768900373, + "grad_norm": 0.9980739951133728, + "learning_rate": 3.443342879779954e-05, + "loss": 0.5256, + "step": 18260 + }, + { + "epoch": 1.2255629005738062, + "grad_norm": 1.5467917919158936, + "learning_rate": 3.442310099885665e-05, + "loss": 0.482, + "step": 18262 + }, + { + "epoch": 1.2256971242575752, + "grad_norm": 2.0626678466796875, + "learning_rate": 3.441277393589162e-05, + "loss": 0.528, + "step": 18264 + }, + { + "epoch": 1.2258313479413443, + "grad_norm": 0.8996275663375854, + "learning_rate": 3.440244760939236e-05, + "loss": 0.4628, + "step": 18266 + }, + { + "epoch": 1.2259655716251132, + "grad_norm": 0.7829926609992981, + "learning_rate": 3.439212201984679e-05, + "loss": 0.5047, + "step": 18268 + }, + { + "epoch": 1.2260997953088824, + "grad_norm": 1.0295792818069458, + "learning_rate": 3.438179716774275e-05, + "loss": 0.5633, + "step": 18270 + }, + { + "epoch": 1.2262340189926513, + "grad_norm": 0.9000246524810791, + "learning_rate": 3.437147305356807e-05, + "loss": 0.4807, + "step": 18272 + }, + { + "epoch": 1.2263682426764202, + "grad_norm": 0.8799797892570496, + "learning_rate": 3.436114967781059e-05, + "loss": 0.5576, + "step": 18274 + }, + { + "epoch": 1.2265024663601893, + "grad_norm": 0.9025986790657043, + "learning_rate": 3.4350827040958e-05, + "loss": 0.4866, + "step": 18276 + }, + { + "epoch": 1.2266366900439583, + "grad_norm": 1.0311486721038818, + "learning_rate": 3.434050514349809e-05, + "loss": 0.5643, + "step": 18278 + }, + { + "epoch": 1.2267709137277272, + "grad_norm": 0.9919482469558716, + "learning_rate": 3.43301839859185e-05, + "loss": 0.4589, + "step": 18280 + }, + { + "epoch": 1.2269051374114963, + "grad_norm": 1.0341503620147705, + "learning_rate": 3.431986356870691e-05, + "loss": 0.4944, + "step": 18282 + }, + { + "epoch": 1.2270393610952652, + "grad_norm": 0.968529224395752, + "learning_rate": 3.430954389235092e-05, + "loss": 0.4862, + "step": 18284 + }, + { + "epoch": 1.2271735847790342, + "grad_norm": 1.0049059391021729, + "learning_rate": 3.429922495733815e-05, + "loss": 0.5348, + "step": 18286 + }, + { + "epoch": 1.2273078084628033, + "grad_norm": 0.8700787425041199, + "learning_rate": 3.4288906764156106e-05, + "loss": 0.4803, + "step": 18288 + }, + { + "epoch": 1.2274420321465722, + "grad_norm": 1.1416560411453247, + "learning_rate": 3.4278589313292345e-05, + "loss": 0.5415, + "step": 18290 + }, + { + "epoch": 1.2275762558303414, + "grad_norm": 1.0105998516082764, + "learning_rate": 3.4268272605234296e-05, + "loss": 0.4932, + "step": 18292 + }, + { + "epoch": 1.2277104795141103, + "grad_norm": 0.9765534400939941, + "learning_rate": 3.425795664046946e-05, + "loss": 0.5506, + "step": 18294 + }, + { + "epoch": 1.2278447031978792, + "grad_norm": 1.060732364654541, + "learning_rate": 3.42476414194852e-05, + "loss": 0.5005, + "step": 18296 + }, + { + "epoch": 1.2279789268816483, + "grad_norm": 1.036497712135315, + "learning_rate": 3.423732694276891e-05, + "loss": 0.5465, + "step": 18298 + }, + { + "epoch": 1.2281131505654173, + "grad_norm": 1.1618542671203613, + "learning_rate": 3.422701321080791e-05, + "loss": 0.5404, + "step": 18300 + }, + { + "epoch": 1.2282473742491862, + "grad_norm": 0.8722122311592102, + "learning_rate": 3.421670022408956e-05, + "loss": 0.49, + "step": 18302 + }, + { + "epoch": 1.2283815979329553, + "grad_norm": 0.9689675569534302, + "learning_rate": 3.420638798310105e-05, + "loss": 0.497, + "step": 18304 + }, + { + "epoch": 1.2285158216167242, + "grad_norm": 1.1279895305633545, + "learning_rate": 3.419607648832968e-05, + "loss": 0.5396, + "step": 18306 + }, + { + "epoch": 1.2286500453004932, + "grad_norm": 0.9110646843910217, + "learning_rate": 3.4185765740262586e-05, + "loss": 0.485, + "step": 18308 + }, + { + "epoch": 1.2287842689842623, + "grad_norm": 0.9667492508888245, + "learning_rate": 3.417545573938699e-05, + "loss": 0.5364, + "step": 18310 + }, + { + "epoch": 1.2289184926680312, + "grad_norm": 1.0547741651535034, + "learning_rate": 3.416514648618998e-05, + "loss": 0.4854, + "step": 18312 + }, + { + "epoch": 1.2290527163518004, + "grad_norm": 1.0077714920043945, + "learning_rate": 3.415483798115867e-05, + "loss": 0.4669, + "step": 18314 + }, + { + "epoch": 1.2291869400355693, + "grad_norm": 1.0353007316589355, + "learning_rate": 3.4144530224780095e-05, + "loss": 0.551, + "step": 18316 + }, + { + "epoch": 1.2293211637193382, + "grad_norm": 1.0020809173583984, + "learning_rate": 3.41342232175413e-05, + "loss": 0.5244, + "step": 18318 + }, + { + "epoch": 1.2294553874031073, + "grad_norm": 1.1620763540267944, + "learning_rate": 3.412391695992925e-05, + "loss": 0.5181, + "step": 18320 + }, + { + "epoch": 1.2295896110868763, + "grad_norm": 1.1190643310546875, + "learning_rate": 3.4113611452430914e-05, + "loss": 0.5102, + "step": 18322 + }, + { + "epoch": 1.2297238347706454, + "grad_norm": 1.0470566749572754, + "learning_rate": 3.410330669553319e-05, + "loss": 0.5045, + "step": 18324 + }, + { + "epoch": 1.2298580584544143, + "grad_norm": 1.0887559652328491, + "learning_rate": 3.409300268972298e-05, + "loss": 0.4794, + "step": 18326 + }, + { + "epoch": 1.2299922821381832, + "grad_norm": 1.129022240638733, + "learning_rate": 3.4082699435487084e-05, + "loss": 0.5229, + "step": 18328 + }, + { + "epoch": 1.2301265058219524, + "grad_norm": 0.9262421727180481, + "learning_rate": 3.4072396933312365e-05, + "loss": 0.5555, + "step": 18330 + }, + { + "epoch": 1.2302607295057213, + "grad_norm": 1.0439683198928833, + "learning_rate": 3.406209518368555e-05, + "loss": 0.517, + "step": 18332 + }, + { + "epoch": 1.2303949531894902, + "grad_norm": 1.0444456338882446, + "learning_rate": 3.405179418709342e-05, + "loss": 0.5042, + "step": 18334 + }, + { + "epoch": 1.2305291768732594, + "grad_norm": 1.017127275466919, + "learning_rate": 3.4041493944022655e-05, + "loss": 0.4798, + "step": 18336 + }, + { + "epoch": 1.2306634005570283, + "grad_norm": 1.0182819366455078, + "learning_rate": 3.40311944549599e-05, + "loss": 0.5153, + "step": 18338 + }, + { + "epoch": 1.2307976242407972, + "grad_norm": 0.9072675704956055, + "learning_rate": 3.402089572039183e-05, + "loss": 0.4493, + "step": 18340 + }, + { + "epoch": 1.2309318479245663, + "grad_norm": 0.9794762134552002, + "learning_rate": 3.401059774080498e-05, + "loss": 0.467, + "step": 18342 + }, + { + "epoch": 1.2310660716083353, + "grad_norm": 1.4422526359558105, + "learning_rate": 3.4000300516685965e-05, + "loss": 0.5094, + "step": 18344 + }, + { + "epoch": 1.2312002952921044, + "grad_norm": 1.0380927324295044, + "learning_rate": 3.399000404852129e-05, + "loss": 0.5298, + "step": 18346 + }, + { + "epoch": 1.2313345189758733, + "grad_norm": 1.0192526578903198, + "learning_rate": 3.397970833679744e-05, + "loss": 0.4584, + "step": 18348 + }, + { + "epoch": 1.2314687426596422, + "grad_norm": 1.0050543546676636, + "learning_rate": 3.396941338200087e-05, + "loss": 0.4855, + "step": 18350 + }, + { + "epoch": 1.2316029663434114, + "grad_norm": 1.0599147081375122, + "learning_rate": 3.3959119184618e-05, + "loss": 0.5053, + "step": 18352 + }, + { + "epoch": 1.2317371900271803, + "grad_norm": 0.9676487445831299, + "learning_rate": 3.394882574513519e-05, + "loss": 0.4528, + "step": 18354 + }, + { + "epoch": 1.2318714137109492, + "grad_norm": 1.147121548652649, + "learning_rate": 3.393853306403881e-05, + "loss": 0.5288, + "step": 18356 + }, + { + "epoch": 1.2320056373947184, + "grad_norm": 0.8826063871383667, + "learning_rate": 3.392824114181516e-05, + "loss": 0.5134, + "step": 18358 + }, + { + "epoch": 1.2321398610784873, + "grad_norm": 1.051510214805603, + "learning_rate": 3.391794997895051e-05, + "loss": 0.4756, + "step": 18360 + }, + { + "epoch": 1.2322740847622562, + "grad_norm": 0.9662967920303345, + "learning_rate": 3.3907659575931096e-05, + "loss": 0.4593, + "step": 18362 + }, + { + "epoch": 1.2324083084460253, + "grad_norm": 1.0281141996383667, + "learning_rate": 3.389736993324314e-05, + "loss": 0.5, + "step": 18364 + }, + { + "epoch": 1.2325425321297943, + "grad_norm": 0.966301441192627, + "learning_rate": 3.388708105137276e-05, + "loss": 0.477, + "step": 18366 + }, + { + "epoch": 1.2326767558135634, + "grad_norm": 0.9663693308830261, + "learning_rate": 3.3876792930806144e-05, + "loss": 0.5562, + "step": 18368 + }, + { + "epoch": 1.2328109794973323, + "grad_norm": 0.892840564250946, + "learning_rate": 3.3866505572029334e-05, + "loss": 0.4397, + "step": 18370 + }, + { + "epoch": 1.2329452031811012, + "grad_norm": 1.1237777471542358, + "learning_rate": 3.3856218975528434e-05, + "loss": 0.5463, + "step": 18372 + }, + { + "epoch": 1.2330794268648704, + "grad_norm": 0.975803554058075, + "learning_rate": 3.384593314178942e-05, + "loss": 0.4984, + "step": 18374 + }, + { + "epoch": 1.2332136505486393, + "grad_norm": 0.9467883110046387, + "learning_rate": 3.383564807129832e-05, + "loss": 0.5388, + "step": 18376 + }, + { + "epoch": 1.2333478742324082, + "grad_norm": 1.12064528465271, + "learning_rate": 3.382536376454104e-05, + "loss": 0.5373, + "step": 18378 + }, + { + "epoch": 1.2334820979161774, + "grad_norm": 0.9428995251655579, + "learning_rate": 3.3815080222003533e-05, + "loss": 0.4736, + "step": 18380 + }, + { + "epoch": 1.2336163215999463, + "grad_norm": 0.9213509559631348, + "learning_rate": 3.3804797444171654e-05, + "loss": 0.4904, + "step": 18382 + }, + { + "epoch": 1.2337505452837152, + "grad_norm": 1.0931910276412964, + "learning_rate": 3.379451543153126e-05, + "loss": 0.5332, + "step": 18384 + }, + { + "epoch": 1.2338847689674843, + "grad_norm": 1.0226064920425415, + "learning_rate": 3.378423418456813e-05, + "loss": 0.5754, + "step": 18386 + }, + { + "epoch": 1.2340189926512533, + "grad_norm": 1.019049882888794, + "learning_rate": 3.3773953703768055e-05, + "loss": 0.4714, + "step": 18388 + }, + { + "epoch": 1.2341532163350224, + "grad_norm": 1.0632864236831665, + "learning_rate": 3.376367398961674e-05, + "loss": 0.5205, + "step": 18390 + }, + { + "epoch": 1.2342874400187913, + "grad_norm": 1.2359843254089355, + "learning_rate": 3.375339504259994e-05, + "loss": 0.506, + "step": 18392 + }, + { + "epoch": 1.2344216637025602, + "grad_norm": 1.0137488842010498, + "learning_rate": 3.3743116863203236e-05, + "loss": 0.5279, + "step": 18394 + }, + { + "epoch": 1.2345558873863294, + "grad_norm": 0.8961823582649231, + "learning_rate": 3.3732839451912335e-05, + "loss": 0.4639, + "step": 18396 + }, + { + "epoch": 1.2346901110700983, + "grad_norm": 0.9982473850250244, + "learning_rate": 3.372256280921274e-05, + "loss": 0.5054, + "step": 18398 + }, + { + "epoch": 1.2348243347538674, + "grad_norm": 1.088832974433899, + "learning_rate": 3.371228693559007e-05, + "loss": 0.5486, + "step": 18400 + }, + { + "epoch": 1.2349585584376364, + "grad_norm": 0.926771342754364, + "learning_rate": 3.3702011831529805e-05, + "loss": 0.5207, + "step": 18402 + }, + { + "epoch": 1.2350927821214053, + "grad_norm": 1.06533944606781, + "learning_rate": 3.369173749751744e-05, + "loss": 0.4763, + "step": 18404 + }, + { + "epoch": 1.2352270058051744, + "grad_norm": 1.282995343208313, + "learning_rate": 3.3681463934038415e-05, + "loss": 0.4892, + "step": 18406 + }, + { + "epoch": 1.2353612294889433, + "grad_norm": 0.9002024531364441, + "learning_rate": 3.3671191141578114e-05, + "loss": 0.4841, + "step": 18408 + }, + { + "epoch": 1.2354954531727123, + "grad_norm": 1.3077336549758911, + "learning_rate": 3.366091912062194e-05, + "loss": 0.4384, + "step": 18410 + }, + { + "epoch": 1.2356296768564814, + "grad_norm": 0.9488261938095093, + "learning_rate": 3.365064787165518e-05, + "loss": 0.4489, + "step": 18412 + }, + { + "epoch": 1.2357639005402503, + "grad_norm": 1.1497156620025635, + "learning_rate": 3.3640377395163185e-05, + "loss": 0.5817, + "step": 18414 + }, + { + "epoch": 1.2358981242240192, + "grad_norm": 1.0375531911849976, + "learning_rate": 3.3630107691631174e-05, + "loss": 0.567, + "step": 18416 + }, + { + "epoch": 1.2360323479077884, + "grad_norm": 0.9430070519447327, + "learning_rate": 3.361983876154439e-05, + "loss": 0.4803, + "step": 18418 + }, + { + "epoch": 1.2361665715915573, + "grad_norm": 0.9468973278999329, + "learning_rate": 3.360957060538801e-05, + "loss": 0.4333, + "step": 18420 + }, + { + "epoch": 1.2363007952753264, + "grad_norm": 0.9819739460945129, + "learning_rate": 3.35993032236472e-05, + "loss": 0.5326, + "step": 18422 + }, + { + "epoch": 1.2364350189590954, + "grad_norm": 0.9785555601119995, + "learning_rate": 3.358903661680703e-05, + "loss": 0.4747, + "step": 18424 + }, + { + "epoch": 1.2365692426428643, + "grad_norm": 0.9492508769035339, + "learning_rate": 3.3578770785352654e-05, + "loss": 0.44, + "step": 18426 + }, + { + "epoch": 1.2367034663266334, + "grad_norm": 0.9501255750656128, + "learning_rate": 3.356850572976903e-05, + "loss": 0.5529, + "step": 18428 + }, + { + "epoch": 1.2368376900104023, + "grad_norm": 1.0544824600219727, + "learning_rate": 3.3558241450541235e-05, + "loss": 0.4882, + "step": 18430 + }, + { + "epoch": 1.2369719136941713, + "grad_norm": 0.9859625101089478, + "learning_rate": 3.354797794815416e-05, + "loss": 0.5196, + "step": 18432 + }, + { + "epoch": 1.2371061373779404, + "grad_norm": 0.9515497088432312, + "learning_rate": 3.35377152230928e-05, + "loss": 0.4984, + "step": 18434 + }, + { + "epoch": 1.2372403610617093, + "grad_norm": 1.6617851257324219, + "learning_rate": 3.352745327584202e-05, + "loss": 0.5168, + "step": 18436 + }, + { + "epoch": 1.2373745847454782, + "grad_norm": 0.9807457327842712, + "learning_rate": 3.3517192106886684e-05, + "loss": 0.5494, + "step": 18438 + }, + { + "epoch": 1.2375088084292474, + "grad_norm": 1.1575415134429932, + "learning_rate": 3.3506931716711596e-05, + "loss": 0.4799, + "step": 18440 + }, + { + "epoch": 1.2376430321130163, + "grad_norm": 1.086361289024353, + "learning_rate": 3.3496672105801575e-05, + "loss": 0.4879, + "step": 18442 + }, + { + "epoch": 1.2377772557967854, + "grad_norm": 1.0192978382110596, + "learning_rate": 3.3486413274641327e-05, + "loss": 0.5117, + "step": 18444 + }, + { + "epoch": 1.2379114794805544, + "grad_norm": 1.3816787004470825, + "learning_rate": 3.347615522371559e-05, + "loss": 0.519, + "step": 18446 + }, + { + "epoch": 1.2380457031643233, + "grad_norm": 1.0699265003204346, + "learning_rate": 3.3465897953509006e-05, + "loss": 0.5066, + "step": 18448 + }, + { + "epoch": 1.2381799268480924, + "grad_norm": 1.1121770143508911, + "learning_rate": 3.345564146450625e-05, + "loss": 0.5127, + "step": 18450 + }, + { + "epoch": 1.2383141505318613, + "grad_norm": 1.0726310014724731, + "learning_rate": 3.3445385757191885e-05, + "loss": 0.508, + "step": 18452 + }, + { + "epoch": 1.2384483742156303, + "grad_norm": 1.135878562927246, + "learning_rate": 3.3435130832050517e-05, + "loss": 0.5387, + "step": 18454 + }, + { + "epoch": 1.2385825978993994, + "grad_norm": 0.9350636005401611, + "learning_rate": 3.342487668956661e-05, + "loss": 0.5215, + "step": 18456 + }, + { + "epoch": 1.2387168215831683, + "grad_norm": 1.0614064931869507, + "learning_rate": 3.341462333022472e-05, + "loss": 0.5485, + "step": 18458 + }, + { + "epoch": 1.2388510452669372, + "grad_norm": 1.1100685596466064, + "learning_rate": 3.340437075450923e-05, + "loss": 0.5167, + "step": 18460 + }, + { + "epoch": 1.2389852689507064, + "grad_norm": 1.136499285697937, + "learning_rate": 3.339411896290462e-05, + "loss": 0.553, + "step": 18462 + }, + { + "epoch": 1.2391194926344753, + "grad_norm": 1.1006940603256226, + "learning_rate": 3.338386795589521e-05, + "loss": 0.5592, + "step": 18464 + }, + { + "epoch": 1.2392537163182444, + "grad_norm": 0.9916535019874573, + "learning_rate": 3.337361773396538e-05, + "loss": 0.4779, + "step": 18466 + }, + { + "epoch": 1.2393879400020134, + "grad_norm": 1.1026976108551025, + "learning_rate": 3.336336829759941e-05, + "loss": 0.5057, + "step": 18468 + }, + { + "epoch": 1.2395221636857823, + "grad_norm": 1.1068087816238403, + "learning_rate": 3.3353119647281575e-05, + "loss": 0.4868, + "step": 18470 + }, + { + "epoch": 1.2396563873695514, + "grad_norm": 1.0291463136672974, + "learning_rate": 3.334287178349611e-05, + "loss": 0.5074, + "step": 18472 + }, + { + "epoch": 1.2397906110533203, + "grad_norm": 0.9090981483459473, + "learning_rate": 3.333262470672719e-05, + "loss": 0.4529, + "step": 18474 + }, + { + "epoch": 1.2399248347370895, + "grad_norm": 1.090463638305664, + "learning_rate": 3.332237841745898e-05, + "loss": 0.5083, + "step": 18476 + }, + { + "epoch": 1.2400590584208584, + "grad_norm": 1.8354182243347168, + "learning_rate": 3.3312132916175586e-05, + "loss": 0.4737, + "step": 18478 + }, + { + "epoch": 1.2401932821046273, + "grad_norm": 1.238089919090271, + "learning_rate": 3.330188820336111e-05, + "loss": 0.599, + "step": 18480 + }, + { + "epoch": 1.2403275057883965, + "grad_norm": 1.2819405794143677, + "learning_rate": 3.329164427949957e-05, + "loss": 0.5218, + "step": 18482 + }, + { + "epoch": 1.2404617294721654, + "grad_norm": 0.9651596546173096, + "learning_rate": 3.3281401145075e-05, + "loss": 0.5906, + "step": 18484 + }, + { + "epoch": 1.2405959531559343, + "grad_norm": 0.9273151159286499, + "learning_rate": 3.3271158800571326e-05, + "loss": 0.5921, + "step": 18486 + }, + { + "epoch": 1.2407301768397034, + "grad_norm": 1.0414173603057861, + "learning_rate": 3.3260917246472525e-05, + "loss": 0.5502, + "step": 18488 + }, + { + "epoch": 1.2408644005234724, + "grad_norm": 1.1283290386199951, + "learning_rate": 3.325067648326244e-05, + "loss": 0.5242, + "step": 18490 + }, + { + "epoch": 1.2409986242072413, + "grad_norm": 0.974013090133667, + "learning_rate": 3.324043651142499e-05, + "loss": 0.4687, + "step": 18492 + }, + { + "epoch": 1.2411328478910104, + "grad_norm": 0.985163688659668, + "learning_rate": 3.323019733144392e-05, + "loss": 0.5497, + "step": 18494 + }, + { + "epoch": 1.2412670715747793, + "grad_norm": 0.972523033618927, + "learning_rate": 3.3219958943803076e-05, + "loss": 0.5479, + "step": 18496 + }, + { + "epoch": 1.2414012952585485, + "grad_norm": 0.9983100891113281, + "learning_rate": 3.3209721348986166e-05, + "loss": 0.4886, + "step": 18498 + }, + { + "epoch": 1.2415355189423174, + "grad_norm": 1.0459272861480713, + "learning_rate": 3.3199484547476915e-05, + "loss": 0.5291, + "step": 18500 + }, + { + "epoch": 1.2416697426260863, + "grad_norm": 1.0113303661346436, + "learning_rate": 3.3189248539758974e-05, + "loss": 0.462, + "step": 18502 + }, + { + "epoch": 1.2418039663098555, + "grad_norm": 0.879380464553833, + "learning_rate": 3.317901332631599e-05, + "loss": 0.4726, + "step": 18504 + }, + { + "epoch": 1.2419381899936244, + "grad_norm": 1.1066372394561768, + "learning_rate": 3.3168778907631534e-05, + "loss": 0.4816, + "step": 18506 + }, + { + "epoch": 1.2420724136773933, + "grad_norm": 1.1736201047897339, + "learning_rate": 3.3158545284189204e-05, + "loss": 0.5234, + "step": 18508 + }, + { + "epoch": 1.2422066373611624, + "grad_norm": 1.0361710786819458, + "learning_rate": 3.314831245647247e-05, + "loss": 0.486, + "step": 18510 + }, + { + "epoch": 1.2423408610449314, + "grad_norm": 1.0722993612289429, + "learning_rate": 3.3138080424964845e-05, + "loss": 0.5198, + "step": 18512 + }, + { + "epoch": 1.2424750847287003, + "grad_norm": 1.145302176475525, + "learning_rate": 3.312784919014974e-05, + "loss": 0.4976, + "step": 18514 + }, + { + "epoch": 1.2426093084124694, + "grad_norm": 0.9962978363037109, + "learning_rate": 3.311761875251062e-05, + "loss": 0.5858, + "step": 18516 + }, + { + "epoch": 1.2427435320962383, + "grad_norm": 1.1119589805603027, + "learning_rate": 3.3107389112530784e-05, + "loss": 0.5064, + "step": 18518 + }, + { + "epoch": 1.2428777557800075, + "grad_norm": 1.1276476383209229, + "learning_rate": 3.309716027069362e-05, + "loss": 0.5459, + "step": 18520 + }, + { + "epoch": 1.2430119794637764, + "grad_norm": 0.9385380744934082, + "learning_rate": 3.308693222748237e-05, + "loss": 0.485, + "step": 18522 + }, + { + "epoch": 1.2431462031475453, + "grad_norm": 1.0502797365188599, + "learning_rate": 3.3076704983380334e-05, + "loss": 0.5509, + "step": 18524 + }, + { + "epoch": 1.2432804268313145, + "grad_norm": 0.9831997156143188, + "learning_rate": 3.30664785388707e-05, + "loss": 0.5499, + "step": 18526 + }, + { + "epoch": 1.2434146505150834, + "grad_norm": 0.9537842273712158, + "learning_rate": 3.305625289443667e-05, + "loss": 0.4943, + "step": 18528 + }, + { + "epoch": 1.2435488741988523, + "grad_norm": 0.9327796697616577, + "learning_rate": 3.304602805056135e-05, + "loss": 0.5067, + "step": 18530 + }, + { + "epoch": 1.2436830978826214, + "grad_norm": 0.793362021446228, + "learning_rate": 3.3035804007727885e-05, + "loss": 0.4489, + "step": 18532 + }, + { + "epoch": 1.2438173215663904, + "grad_norm": 0.9797016382217407, + "learning_rate": 3.302558076641931e-05, + "loss": 0.4644, + "step": 18534 + }, + { + "epoch": 1.2439515452501593, + "grad_norm": 1.111334204673767, + "learning_rate": 3.301535832711867e-05, + "loss": 0.5596, + "step": 18536 + }, + { + "epoch": 1.2440857689339284, + "grad_norm": 1.0660854578018188, + "learning_rate": 3.300513669030895e-05, + "loss": 0.5534, + "step": 18538 + }, + { + "epoch": 1.2442199926176973, + "grad_norm": 1.0276641845703125, + "learning_rate": 3.299491585647311e-05, + "loss": 0.4905, + "step": 18540 + }, + { + "epoch": 1.2443542163014665, + "grad_norm": 1.1686550378799438, + "learning_rate": 3.298469582609405e-05, + "loss": 0.5589, + "step": 18542 + }, + { + "epoch": 1.2444884399852354, + "grad_norm": 1.0222535133361816, + "learning_rate": 3.2974476599654646e-05, + "loss": 0.5572, + "step": 18544 + }, + { + "epoch": 1.2446226636690043, + "grad_norm": 1.0813424587249756, + "learning_rate": 3.296425817763776e-05, + "loss": 0.5217, + "step": 18546 + }, + { + "epoch": 1.2447568873527735, + "grad_norm": 1.1131865978240967, + "learning_rate": 3.295404056052616e-05, + "loss": 0.5526, + "step": 18548 + }, + { + "epoch": 1.2448911110365424, + "grad_norm": 0.9508737921714783, + "learning_rate": 3.2943823748802635e-05, + "loss": 0.4891, + "step": 18550 + }, + { + "epoch": 1.2450253347203115, + "grad_norm": 1.0114701986312866, + "learning_rate": 3.2933607742949876e-05, + "loss": 0.5116, + "step": 18552 + }, + { + "epoch": 1.2451595584040804, + "grad_norm": 1.0921638011932373, + "learning_rate": 3.292339254345063e-05, + "loss": 0.4719, + "step": 18554 + }, + { + "epoch": 1.2452937820878494, + "grad_norm": 1.1859817504882812, + "learning_rate": 3.2913178150787465e-05, + "loss": 0.546, + "step": 18556 + }, + { + "epoch": 1.2454280057716185, + "grad_norm": 1.0289673805236816, + "learning_rate": 3.290296456544306e-05, + "loss": 0.4925, + "step": 18558 + }, + { + "epoch": 1.2455622294553874, + "grad_norm": 0.9060097932815552, + "learning_rate": 3.289275178789995e-05, + "loss": 0.5271, + "step": 18560 + }, + { + "epoch": 1.2456964531391563, + "grad_norm": 0.955970048904419, + "learning_rate": 3.288253981864068e-05, + "loss": 0.5207, + "step": 18562 + }, + { + "epoch": 1.2458306768229255, + "grad_norm": 0.9810526967048645, + "learning_rate": 3.287232865814774e-05, + "loss": 0.4588, + "step": 18564 + }, + { + "epoch": 1.2459649005066944, + "grad_norm": 0.9905154705047607, + "learning_rate": 3.28621183069036e-05, + "loss": 0.5414, + "step": 18566 + }, + { + "epoch": 1.2460991241904633, + "grad_norm": 1.0671234130859375, + "learning_rate": 3.285190876539066e-05, + "loss": 0.5407, + "step": 18568 + }, + { + "epoch": 1.2462333478742325, + "grad_norm": 1.082667350769043, + "learning_rate": 3.284170003409132e-05, + "loss": 0.5051, + "step": 18570 + }, + { + "epoch": 1.2463675715580014, + "grad_norm": 1.0688824653625488, + "learning_rate": 3.2831492113487904e-05, + "loss": 0.5403, + "step": 18572 + }, + { + "epoch": 1.2465017952417705, + "grad_norm": 0.9907695055007935, + "learning_rate": 3.282128500406273e-05, + "loss": 0.5392, + "step": 18574 + }, + { + "epoch": 1.2466360189255394, + "grad_norm": 0.8497958183288574, + "learning_rate": 3.281107870629806e-05, + "loss": 0.4663, + "step": 18576 + }, + { + "epoch": 1.2467702426093084, + "grad_norm": 0.9073302745819092, + "learning_rate": 3.280087322067612e-05, + "loss": 0.5011, + "step": 18578 + }, + { + "epoch": 1.2469044662930775, + "grad_norm": 1.1064773797988892, + "learning_rate": 3.279066854767908e-05, + "loss": 0.5251, + "step": 18580 + }, + { + "epoch": 1.2470386899768464, + "grad_norm": 1.1607712507247925, + "learning_rate": 3.2780464687789144e-05, + "loss": 0.4841, + "step": 18582 + }, + { + "epoch": 1.2471729136606153, + "grad_norm": 1.0058317184448242, + "learning_rate": 3.277026164148836e-05, + "loss": 0.5717, + "step": 18584 + }, + { + "epoch": 1.2473071373443845, + "grad_norm": 1.129913091659546, + "learning_rate": 3.2760059409258855e-05, + "loss": 0.5153, + "step": 18586 + }, + { + "epoch": 1.2474413610281534, + "grad_norm": 0.946969211101532, + "learning_rate": 3.2749857991582635e-05, + "loss": 0.4985, + "step": 18588 + }, + { + "epoch": 1.2475755847119223, + "grad_norm": 1.0105154514312744, + "learning_rate": 3.273965738894171e-05, + "loss": 0.5492, + "step": 18590 + }, + { + "epoch": 1.2477098083956915, + "grad_norm": 0.8665843605995178, + "learning_rate": 3.272945760181802e-05, + "loss": 0.442, + "step": 18592 + }, + { + "epoch": 1.2478440320794604, + "grad_norm": 1.0189673900604248, + "learning_rate": 3.271925863069351e-05, + "loss": 0.511, + "step": 18594 + }, + { + "epoch": 1.2479782557632295, + "grad_norm": 1.064059853553772, + "learning_rate": 3.270906047605003e-05, + "loss": 0.4635, + "step": 18596 + }, + { + "epoch": 1.2481124794469984, + "grad_norm": 0.9782739281654358, + "learning_rate": 3.269886313836947e-05, + "loss": 0.4735, + "step": 18598 + }, + { + "epoch": 1.2482467031307674, + "grad_norm": 0.9779257774353027, + "learning_rate": 3.268866661813358e-05, + "loss": 0.4897, + "step": 18600 + }, + { + "epoch": 1.2483809268145365, + "grad_norm": 0.9636147022247314, + "learning_rate": 3.2678470915824166e-05, + "loss": 0.4963, + "step": 18602 + }, + { + "epoch": 1.2485151504983054, + "grad_norm": 1.0301083326339722, + "learning_rate": 3.266827603192292e-05, + "loss": 0.5127, + "step": 18604 + }, + { + "epoch": 1.2486493741820743, + "grad_norm": 0.9167841076850891, + "learning_rate": 3.265808196691158e-05, + "loss": 0.4677, + "step": 18606 + }, + { + "epoch": 1.2487835978658435, + "grad_norm": 1.0668751001358032, + "learning_rate": 3.264788872127173e-05, + "loss": 0.5269, + "step": 18608 + }, + { + "epoch": 1.2489178215496124, + "grad_norm": 0.9439718127250671, + "learning_rate": 3.2637696295485055e-05, + "loss": 0.5401, + "step": 18610 + }, + { + "epoch": 1.2490520452333813, + "grad_norm": 0.9846333861351013, + "learning_rate": 3.262750469003307e-05, + "loss": 0.4892, + "step": 18612 + }, + { + "epoch": 1.2491862689171505, + "grad_norm": 1.0669738054275513, + "learning_rate": 3.261731390539732e-05, + "loss": 0.5529, + "step": 18614 + }, + { + "epoch": 1.2493204926009194, + "grad_norm": 0.9945108294487, + "learning_rate": 3.260712394205933e-05, + "loss": 0.473, + "step": 18616 + }, + { + "epoch": 1.2494547162846885, + "grad_norm": 0.9489200711250305, + "learning_rate": 3.25969348005005e-05, + "loss": 0.5114, + "step": 18618 + }, + { + "epoch": 1.2495889399684574, + "grad_norm": 0.9898130893707275, + "learning_rate": 3.2586746481202294e-05, + "loss": 0.5466, + "step": 18620 + }, + { + "epoch": 1.2497231636522264, + "grad_norm": 1.1003575325012207, + "learning_rate": 3.2576558984646075e-05, + "loss": 0.5652, + "step": 18622 + }, + { + "epoch": 1.2498573873359955, + "grad_norm": 0.9520372152328491, + "learning_rate": 3.256637231131319e-05, + "loss": 0.5103, + "step": 18624 + }, + { + "epoch": 1.2499916110197644, + "grad_norm": 0.9153604507446289, + "learning_rate": 3.255618646168491e-05, + "loss": 0.533, + "step": 18626 + }, + { + "epoch": 1.2501258347035336, + "grad_norm": 1.1091843843460083, + "learning_rate": 3.2546001436242547e-05, + "loss": 0.5138, + "step": 18628 + }, + { + "epoch": 1.2502600583873025, + "grad_norm": 0.919145941734314, + "learning_rate": 3.253581723546727e-05, + "loss": 0.4221, + "step": 18630 + }, + { + "epoch": 1.2503942820710714, + "grad_norm": 1.1235380172729492, + "learning_rate": 3.25256338598403e-05, + "loss": 0.5764, + "step": 18632 + }, + { + "epoch": 1.2505285057548403, + "grad_norm": 1.2690304517745972, + "learning_rate": 3.251545130984276e-05, + "loss": 0.4321, + "step": 18634 + }, + { + "epoch": 1.2506627294386095, + "grad_norm": 1.2638576030731201, + "learning_rate": 3.250526958595578e-05, + "loss": 0.533, + "step": 18636 + }, + { + "epoch": 1.2507969531223784, + "grad_norm": 0.9857358932495117, + "learning_rate": 3.249508868866039e-05, + "loss": 0.4782, + "step": 18638 + }, + { + "epoch": 1.2509311768061475, + "grad_norm": 1.0567487478256226, + "learning_rate": 3.248490861843765e-05, + "loss": 0.5068, + "step": 18640 + }, + { + "epoch": 1.2510654004899164, + "grad_norm": 1.0033124685287476, + "learning_rate": 3.247472937576852e-05, + "loss": 0.5441, + "step": 18642 + }, + { + "epoch": 1.2511996241736854, + "grad_norm": 1.047810435295105, + "learning_rate": 3.246455096113399e-05, + "loss": 0.5456, + "step": 18644 + }, + { + "epoch": 1.2513338478574545, + "grad_norm": 1.0675344467163086, + "learning_rate": 3.245437337501491e-05, + "loss": 0.5262, + "step": 18646 + }, + { + "epoch": 1.2514680715412234, + "grad_norm": 1.068942666053772, + "learning_rate": 3.244419661789222e-05, + "loss": 0.5746, + "step": 18648 + }, + { + "epoch": 1.2516022952249926, + "grad_norm": 0.9902501106262207, + "learning_rate": 3.24340206902467e-05, + "loss": 0.4986, + "step": 18650 + }, + { + "epoch": 1.2517365189087615, + "grad_norm": 0.9147794246673584, + "learning_rate": 3.242384559255917e-05, + "loss": 0.445, + "step": 18652 + }, + { + "epoch": 1.2518707425925304, + "grad_norm": 1.0337930917739868, + "learning_rate": 3.241367132531036e-05, + "loss": 0.525, + "step": 18654 + }, + { + "epoch": 1.2520049662762995, + "grad_norm": 0.9941875338554382, + "learning_rate": 3.2403497888981015e-05, + "loss": 0.478, + "step": 18656 + }, + { + "epoch": 1.2521391899600685, + "grad_norm": 1.101161241531372, + "learning_rate": 3.2393325284051776e-05, + "loss": 0.5018, + "step": 18658 + }, + { + "epoch": 1.2522734136438374, + "grad_norm": 0.95537269115448, + "learning_rate": 3.238315351100332e-05, + "loss": 0.5611, + "step": 18660 + }, + { + "epoch": 1.2524076373276065, + "grad_norm": 1.0222139358520508, + "learning_rate": 3.237298257031619e-05, + "loss": 0.5171, + "step": 18662 + }, + { + "epoch": 1.2525418610113754, + "grad_norm": 0.8889300227165222, + "learning_rate": 3.236281246247099e-05, + "loss": 0.4837, + "step": 18664 + }, + { + "epoch": 1.2526760846951444, + "grad_norm": 1.0732251405715942, + "learning_rate": 3.235264318794821e-05, + "loss": 0.5226, + "step": 18666 + }, + { + "epoch": 1.2528103083789135, + "grad_norm": 1.0490918159484863, + "learning_rate": 3.234247474722835e-05, + "loss": 0.4917, + "step": 18668 + }, + { + "epoch": 1.2529445320626824, + "grad_norm": 1.081900954246521, + "learning_rate": 3.2332307140791805e-05, + "loss": 0.5014, + "step": 18670 + }, + { + "epoch": 1.2530787557464516, + "grad_norm": 0.9502929449081421, + "learning_rate": 3.2322140369119045e-05, + "loss": 0.5165, + "step": 18672 + }, + { + "epoch": 1.2532129794302205, + "grad_norm": 1.0244060754776, + "learning_rate": 3.231197443269036e-05, + "loss": 0.4838, + "step": 18674 + }, + { + "epoch": 1.2533472031139894, + "grad_norm": 1.018920660018921, + "learning_rate": 3.230180933198612e-05, + "loss": 0.4737, + "step": 18676 + }, + { + "epoch": 1.2534814267977585, + "grad_norm": 2.8826301097869873, + "learning_rate": 3.2291645067486595e-05, + "loss": 0.559, + "step": 18678 + }, + { + "epoch": 1.2536156504815275, + "grad_norm": 1.0019567012786865, + "learning_rate": 3.2281481639672e-05, + "loss": 0.4896, + "step": 18680 + }, + { + "epoch": 1.2537498741652966, + "grad_norm": 1.0178946256637573, + "learning_rate": 3.227131904902257e-05, + "loss": 0.4947, + "step": 18682 + }, + { + "epoch": 1.2538840978490655, + "grad_norm": 0.9037721157073975, + "learning_rate": 3.226115729601843e-05, + "loss": 0.4808, + "step": 18684 + }, + { + "epoch": 1.2540183215328344, + "grad_norm": 1.0834360122680664, + "learning_rate": 3.225099638113974e-05, + "loss": 0.547, + "step": 18686 + }, + { + "epoch": 1.2541525452166034, + "grad_norm": 0.911413311958313, + "learning_rate": 3.224083630486656e-05, + "loss": 0.5445, + "step": 18688 + }, + { + "epoch": 1.2542867689003725, + "grad_norm": 0.995908796787262, + "learning_rate": 3.223067706767896e-05, + "loss": 0.5148, + "step": 18690 + }, + { + "epoch": 1.2544209925841414, + "grad_norm": 1.0455437898635864, + "learning_rate": 3.2220518670056916e-05, + "loss": 0.5453, + "step": 18692 + }, + { + "epoch": 1.2545552162679106, + "grad_norm": 0.9652132987976074, + "learning_rate": 3.221036111248041e-05, + "loss": 0.4771, + "step": 18694 + }, + { + "epoch": 1.2546894399516795, + "grad_norm": 2.6709609031677246, + "learning_rate": 3.2200204395429344e-05, + "loss": 0.5278, + "step": 18696 + }, + { + "epoch": 1.2548236636354484, + "grad_norm": 0.9316728711128235, + "learning_rate": 3.2190048519383635e-05, + "loss": 0.4673, + "step": 18698 + }, + { + "epoch": 1.2549578873192175, + "grad_norm": 0.9168113470077515, + "learning_rate": 3.2179893484823106e-05, + "loss": 0.4929, + "step": 18700 + }, + { + "epoch": 1.2550921110029865, + "grad_norm": 1.005332589149475, + "learning_rate": 3.216973929222757e-05, + "loss": 0.4858, + "step": 18702 + }, + { + "epoch": 1.2552263346867556, + "grad_norm": 1.124756097793579, + "learning_rate": 3.2159585942076775e-05, + "loss": 0.5492, + "step": 18704 + }, + { + "epoch": 1.2553605583705245, + "grad_norm": 1.0058296918869019, + "learning_rate": 3.214943343485049e-05, + "loss": 0.4753, + "step": 18706 + }, + { + "epoch": 1.2554947820542934, + "grad_norm": 1.2010509967803955, + "learning_rate": 3.213928177102834e-05, + "loss": 0.4831, + "step": 18708 + }, + { + "epoch": 1.2556290057380624, + "grad_norm": 1.0022845268249512, + "learning_rate": 3.2129130951090036e-05, + "loss": 0.4934, + "step": 18710 + }, + { + "epoch": 1.2557632294218315, + "grad_norm": 1.1425693035125732, + "learning_rate": 3.2118980975515134e-05, + "loss": 0.4571, + "step": 18712 + }, + { + "epoch": 1.2558974531056004, + "grad_norm": 0.9820336699485779, + "learning_rate": 3.210883184478323e-05, + "loss": 0.4552, + "step": 18714 + }, + { + "epoch": 1.2560316767893696, + "grad_norm": 0.9208889603614807, + "learning_rate": 3.2098683559373835e-05, + "loss": 0.5166, + "step": 18716 + }, + { + "epoch": 1.2561659004731385, + "grad_norm": 0.9718737602233887, + "learning_rate": 3.208853611976644e-05, + "loss": 0.4928, + "step": 18718 + }, + { + "epoch": 1.2563001241569074, + "grad_norm": 1.2219953536987305, + "learning_rate": 3.2078389526440495e-05, + "loss": 0.6289, + "step": 18720 + }, + { + "epoch": 1.2564343478406765, + "grad_norm": 1.119890570640564, + "learning_rate": 3.206824377987541e-05, + "loss": 0.5022, + "step": 18722 + }, + { + "epoch": 1.2565685715244455, + "grad_norm": 1.0400066375732422, + "learning_rate": 3.205809888055053e-05, + "loss": 0.5057, + "step": 18724 + }, + { + "epoch": 1.2567027952082146, + "grad_norm": 0.9680889844894409, + "learning_rate": 3.204795482894521e-05, + "loss": 0.4882, + "step": 18726 + }, + { + "epoch": 1.2568370188919835, + "grad_norm": 1.02334463596344, + "learning_rate": 3.203781162553871e-05, + "loss": 0.4791, + "step": 18728 + }, + { + "epoch": 1.2569712425757524, + "grad_norm": 1.0177057981491089, + "learning_rate": 3.20276692708103e-05, + "loss": 0.4628, + "step": 18730 + }, + { + "epoch": 1.2571054662595216, + "grad_norm": 0.9529568552970886, + "learning_rate": 3.201752776523917e-05, + "loss": 0.4753, + "step": 18732 + }, + { + "epoch": 1.2572396899432905, + "grad_norm": 1.061122179031372, + "learning_rate": 3.20073871093045e-05, + "loss": 0.4968, + "step": 18734 + }, + { + "epoch": 1.2573739136270594, + "grad_norm": 1.1048305034637451, + "learning_rate": 3.199724730348539e-05, + "loss": 0.48, + "step": 18736 + }, + { + "epoch": 1.2575081373108286, + "grad_norm": 1.2485487461090088, + "learning_rate": 3.198710834826096e-05, + "loss": 0.5333, + "step": 18738 + }, + { + "epoch": 1.2576423609945975, + "grad_norm": 0.9257562756538391, + "learning_rate": 3.1976970244110234e-05, + "loss": 0.4765, + "step": 18740 + }, + { + "epoch": 1.2577765846783664, + "grad_norm": 1.0418267250061035, + "learning_rate": 3.196683299151223e-05, + "loss": 0.4768, + "step": 18742 + }, + { + "epoch": 1.2579108083621355, + "grad_norm": 1.0285526514053345, + "learning_rate": 3.19566965909459e-05, + "loss": 0.5373, + "step": 18744 + }, + { + "epoch": 1.2580450320459045, + "grad_norm": 0.8784838914871216, + "learning_rate": 3.1946561042890174e-05, + "loss": 0.4501, + "step": 18746 + }, + { + "epoch": 1.2581792557296736, + "grad_norm": 0.9794188737869263, + "learning_rate": 3.1936426347823955e-05, + "loss": 0.4713, + "step": 18748 + }, + { + "epoch": 1.2583134794134425, + "grad_norm": 0.976508378982544, + "learning_rate": 3.1926292506226054e-05, + "loss": 0.5099, + "step": 18750 + }, + { + "epoch": 1.2584477030972114, + "grad_norm": 0.9098379015922546, + "learning_rate": 3.1916159518575294e-05, + "loss": 0.5241, + "step": 18752 + }, + { + "epoch": 1.2585819267809806, + "grad_norm": 1.3595097064971924, + "learning_rate": 3.190602738535043e-05, + "loss": 0.5093, + "step": 18754 + }, + { + "epoch": 1.2587161504647495, + "grad_norm": 1.1624239683151245, + "learning_rate": 3.189589610703021e-05, + "loss": 0.4938, + "step": 18756 + }, + { + "epoch": 1.2588503741485186, + "grad_norm": 1.031873106956482, + "learning_rate": 3.188576568409328e-05, + "loss": 0.5355, + "step": 18758 + }, + { + "epoch": 1.2589845978322876, + "grad_norm": 1.002721905708313, + "learning_rate": 3.187563611701832e-05, + "loss": 0.4826, + "step": 18760 + }, + { + "epoch": 1.2591188215160565, + "grad_norm": 1.8391629457473755, + "learning_rate": 3.18655074062839e-05, + "loss": 0.4936, + "step": 18762 + }, + { + "epoch": 1.2592530451998254, + "grad_norm": 1.0190123319625854, + "learning_rate": 3.1855379552368605e-05, + "loss": 0.612, + "step": 18764 + }, + { + "epoch": 1.2593872688835945, + "grad_norm": 0.82330721616745, + "learning_rate": 3.184525255575092e-05, + "loss": 0.4679, + "step": 18766 + }, + { + "epoch": 1.2595214925673635, + "grad_norm": 1.0846850872039795, + "learning_rate": 3.183512641690939e-05, + "loss": 0.4938, + "step": 18768 + }, + { + "epoch": 1.2596557162511326, + "grad_norm": 1.0195902585983276, + "learning_rate": 3.182500113632238e-05, + "loss": 0.506, + "step": 18770 + }, + { + "epoch": 1.2597899399349015, + "grad_norm": 1.0187736749649048, + "learning_rate": 3.181487671446836e-05, + "loss": 0.5466, + "step": 18772 + }, + { + "epoch": 1.2599241636186704, + "grad_norm": 1.1696752309799194, + "learning_rate": 3.180475315182563e-05, + "loss": 0.4986, + "step": 18774 + }, + { + "epoch": 1.2600583873024396, + "grad_norm": 0.9804502725601196, + "learning_rate": 3.1794630448872545e-05, + "loss": 0.4734, + "step": 18776 + }, + { + "epoch": 1.2601926109862085, + "grad_norm": 1.367415189743042, + "learning_rate": 3.1784508606087374e-05, + "loss": 0.4764, + "step": 18778 + }, + { + "epoch": 1.2603268346699776, + "grad_norm": 0.9473814368247986, + "learning_rate": 3.177438762394835e-05, + "loss": 0.5156, + "step": 18780 + }, + { + "epoch": 1.2604610583537466, + "grad_norm": 1.252681851387024, + "learning_rate": 3.1764267502933666e-05, + "loss": 0.4782, + "step": 18782 + }, + { + "epoch": 1.2605952820375155, + "grad_norm": 1.0670397281646729, + "learning_rate": 3.17541482435215e-05, + "loss": 0.5068, + "step": 18784 + }, + { + "epoch": 1.2607295057212844, + "grad_norm": 0.9958670139312744, + "learning_rate": 3.1744029846189925e-05, + "loss": 0.5183, + "step": 18786 + }, + { + "epoch": 1.2608637294050535, + "grad_norm": 0.9645728468894958, + "learning_rate": 3.173391231141706e-05, + "loss": 0.4881, + "step": 18788 + }, + { + "epoch": 1.2609979530888225, + "grad_norm": 1.0699021816253662, + "learning_rate": 3.172379563968092e-05, + "loss": 0.5038, + "step": 18790 + }, + { + "epoch": 1.2611321767725916, + "grad_norm": 0.9757247567176819, + "learning_rate": 3.171367983145949e-05, + "loss": 0.5218, + "step": 18792 + }, + { + "epoch": 1.2612664004563605, + "grad_norm": 0.9883769154548645, + "learning_rate": 3.170356488723072e-05, + "loss": 0.4885, + "step": 18794 + }, + { + "epoch": 1.2614006241401294, + "grad_norm": 0.9620486497879028, + "learning_rate": 3.169345080747256e-05, + "loss": 0.5286, + "step": 18796 + }, + { + "epoch": 1.2615348478238986, + "grad_norm": 1.0763846635818481, + "learning_rate": 3.168333759266282e-05, + "loss": 0.5452, + "step": 18798 + }, + { + "epoch": 1.2616690715076675, + "grad_norm": 1.1380316019058228, + "learning_rate": 3.167322524327938e-05, + "loss": 0.5408, + "step": 18800 + }, + { + "epoch": 1.2618032951914366, + "grad_norm": 1.1167099475860596, + "learning_rate": 3.166311375979999e-05, + "loss": 0.4681, + "step": 18802 + }, + { + "epoch": 1.2619375188752056, + "grad_norm": 0.990744948387146, + "learning_rate": 3.1653003142702444e-05, + "loss": 0.5582, + "step": 18804 + }, + { + "epoch": 1.2620717425589745, + "grad_norm": 0.7896711230278015, + "learning_rate": 3.1642893392464404e-05, + "loss": 0.4751, + "step": 18806 + }, + { + "epoch": 1.2622059662427434, + "grad_norm": 1.1622854471206665, + "learning_rate": 3.1632784509563565e-05, + "loss": 0.501, + "step": 18808 + }, + { + "epoch": 1.2623401899265125, + "grad_norm": 0.9174784421920776, + "learning_rate": 3.162267649447752e-05, + "loss": 0.5044, + "step": 18810 + }, + { + "epoch": 1.2624744136102815, + "grad_norm": 1.0127310752868652, + "learning_rate": 3.1612569347683895e-05, + "loss": 0.4747, + "step": 18812 + }, + { + "epoch": 1.2626086372940506, + "grad_norm": 1.002046823501587, + "learning_rate": 3.160246306966019e-05, + "loss": 0.5464, + "step": 18814 + }, + { + "epoch": 1.2627428609778195, + "grad_norm": 0.9624472856521606, + "learning_rate": 3.1592357660883954e-05, + "loss": 0.466, + "step": 18816 + }, + { + "epoch": 1.2628770846615884, + "grad_norm": 1.0694067478179932, + "learning_rate": 3.158225312183261e-05, + "loss": 0.5229, + "step": 18818 + }, + { + "epoch": 1.2630113083453576, + "grad_norm": 0.891897439956665, + "learning_rate": 3.157214945298358e-05, + "loss": 0.4503, + "step": 18820 + }, + { + "epoch": 1.2631455320291265, + "grad_norm": 0.9390667676925659, + "learning_rate": 3.156204665481426e-05, + "loss": 0.4689, + "step": 18822 + }, + { + "epoch": 1.2632797557128956, + "grad_norm": 1.0218350887298584, + "learning_rate": 3.155194472780197e-05, + "loss": 0.5263, + "step": 18824 + }, + { + "epoch": 1.2634139793966646, + "grad_norm": 1.1957632303237915, + "learning_rate": 3.154184367242403e-05, + "loss": 0.5302, + "step": 18826 + }, + { + "epoch": 1.2635482030804335, + "grad_norm": 1.0991414785385132, + "learning_rate": 3.1531743489157664e-05, + "loss": 0.5263, + "step": 18828 + }, + { + "epoch": 1.2636824267642026, + "grad_norm": 1.0328730344772339, + "learning_rate": 3.152164417848012e-05, + "loss": 0.5064, + "step": 18830 + }, + { + "epoch": 1.2638166504479715, + "grad_norm": 1.0690430402755737, + "learning_rate": 3.151154574086853e-05, + "loss": 0.4402, + "step": 18832 + }, + { + "epoch": 1.2639508741317407, + "grad_norm": 1.054821491241455, + "learning_rate": 3.150144817680008e-05, + "loss": 0.4937, + "step": 18834 + }, + { + "epoch": 1.2640850978155096, + "grad_norm": 1.02252197265625, + "learning_rate": 3.149135148675179e-05, + "loss": 0.5199, + "step": 18836 + }, + { + "epoch": 1.2642193214992785, + "grad_norm": 0.9816807508468628, + "learning_rate": 3.148125567120076e-05, + "loss": 0.488, + "step": 18838 + }, + { + "epoch": 1.2643535451830474, + "grad_norm": 0.902094841003418, + "learning_rate": 3.147116073062399e-05, + "loss": 0.4548, + "step": 18840 + }, + { + "epoch": 1.2644877688668166, + "grad_norm": 1.0778688192367554, + "learning_rate": 3.146106666549844e-05, + "loss": 0.5177, + "step": 18842 + }, + { + "epoch": 1.2646219925505855, + "grad_norm": 0.8925482630729675, + "learning_rate": 3.1450973476301026e-05, + "loss": 0.4492, + "step": 18844 + }, + { + "epoch": 1.2647562162343546, + "grad_norm": 0.9562098383903503, + "learning_rate": 3.144088116350866e-05, + "loss": 0.4942, + "step": 18846 + }, + { + "epoch": 1.2648904399181236, + "grad_norm": 0.9487055540084839, + "learning_rate": 3.1430789727598145e-05, + "loss": 0.5028, + "step": 18848 + }, + { + "epoch": 1.2650246636018925, + "grad_norm": 1.0395872592926025, + "learning_rate": 3.142069916904631e-05, + "loss": 0.4787, + "step": 18850 + }, + { + "epoch": 1.2651588872856616, + "grad_norm": 0.9606418013572693, + "learning_rate": 3.14106094883299e-05, + "loss": 0.5069, + "step": 18852 + }, + { + "epoch": 1.2652931109694305, + "grad_norm": 1.210218071937561, + "learning_rate": 3.140052068592564e-05, + "loss": 0.5159, + "step": 18854 + }, + { + "epoch": 1.2654273346531997, + "grad_norm": 1.0063384771347046, + "learning_rate": 3.139043276231019e-05, + "loss": 0.4747, + "step": 18856 + }, + { + "epoch": 1.2655615583369686, + "grad_norm": 0.8223646283149719, + "learning_rate": 3.138034571796022e-05, + "loss": 0.4376, + "step": 18858 + }, + { + "epoch": 1.2656957820207375, + "grad_norm": 1.190323829650879, + "learning_rate": 3.1370259553352274e-05, + "loss": 0.5213, + "step": 18860 + }, + { + "epoch": 1.2658300057045064, + "grad_norm": 1.1596482992172241, + "learning_rate": 3.1360174268962973e-05, + "loss": 0.4962, + "step": 18862 + }, + { + "epoch": 1.2659642293882756, + "grad_norm": 0.9268326163291931, + "learning_rate": 3.135008986526874e-05, + "loss": 0.4507, + "step": 18864 + }, + { + "epoch": 1.2660984530720445, + "grad_norm": 1.0648741722106934, + "learning_rate": 3.1340006342746107e-05, + "loss": 0.5527, + "step": 18866 + }, + { + "epoch": 1.2662326767558136, + "grad_norm": 1.0832303762435913, + "learning_rate": 3.132992370187148e-05, + "loss": 0.5718, + "step": 18868 + }, + { + "epoch": 1.2663669004395826, + "grad_norm": 1.0149818658828735, + "learning_rate": 3.131984194312125e-05, + "loss": 0.5047, + "step": 18870 + }, + { + "epoch": 1.2665011241233515, + "grad_norm": 0.9199078679084778, + "learning_rate": 3.130976106697174e-05, + "loss": 0.4241, + "step": 18872 + }, + { + "epoch": 1.2666353478071206, + "grad_norm": 1.0143941640853882, + "learning_rate": 3.1299681073899284e-05, + "loss": 0.4366, + "step": 18874 + }, + { + "epoch": 1.2667695714908895, + "grad_norm": 1.0422862768173218, + "learning_rate": 3.128960196438011e-05, + "loss": 0.4941, + "step": 18876 + }, + { + "epoch": 1.2669037951746587, + "grad_norm": 1.0042037963867188, + "learning_rate": 3.127952373889046e-05, + "loss": 0.4664, + "step": 18878 + }, + { + "epoch": 1.2670380188584276, + "grad_norm": 2.3638975620269775, + "learning_rate": 3.12694463979065e-05, + "loss": 0.4863, + "step": 18880 + }, + { + "epoch": 1.2671722425421965, + "grad_norm": 1.1389439105987549, + "learning_rate": 3.1259369941904374e-05, + "loss": 0.5407, + "step": 18882 + }, + { + "epoch": 1.2673064662259654, + "grad_norm": 1.0771986246109009, + "learning_rate": 3.124929437136016e-05, + "loss": 0.4671, + "step": 18884 + }, + { + "epoch": 1.2674406899097346, + "grad_norm": 0.9763202667236328, + "learning_rate": 3.123921968674991e-05, + "loss": 0.5225, + "step": 18886 + }, + { + "epoch": 1.2675749135935035, + "grad_norm": 0.9527340531349182, + "learning_rate": 3.122914588854964e-05, + "loss": 0.4885, + "step": 18888 + }, + { + "epoch": 1.2677091372772726, + "grad_norm": 1.0093185901641846, + "learning_rate": 3.1219072977235305e-05, + "loss": 0.5397, + "step": 18890 + }, + { + "epoch": 1.2678433609610416, + "grad_norm": 0.91331547498703, + "learning_rate": 3.120900095328286e-05, + "loss": 0.5074, + "step": 18892 + }, + { + "epoch": 1.2679775846448105, + "grad_norm": 1.0820149183273315, + "learning_rate": 3.1198929817168154e-05, + "loss": 0.5128, + "step": 18894 + }, + { + "epoch": 1.2681118083285796, + "grad_norm": 1.0016778707504272, + "learning_rate": 3.118885956936706e-05, + "loss": 0.5063, + "step": 18896 + }, + { + "epoch": 1.2682460320123485, + "grad_norm": 0.9243509769439697, + "learning_rate": 3.117879021035534e-05, + "loss": 0.5134, + "step": 18898 + }, + { + "epoch": 1.2683802556961177, + "grad_norm": 1.0297565460205078, + "learning_rate": 3.116872174060878e-05, + "loss": 0.5065, + "step": 18900 + }, + { + "epoch": 1.2685144793798866, + "grad_norm": 1.028468132019043, + "learning_rate": 3.1158654160603096e-05, + "loss": 0.5198, + "step": 18902 + }, + { + "epoch": 1.2686487030636555, + "grad_norm": 1.0754987001419067, + "learning_rate": 3.114858747081395e-05, + "loss": 0.5308, + "step": 18904 + }, + { + "epoch": 1.2687829267474247, + "grad_norm": 0.9786719083786011, + "learning_rate": 3.113852167171697e-05, + "loss": 0.4841, + "step": 18906 + }, + { + "epoch": 1.2689171504311936, + "grad_norm": 0.9417799115180969, + "learning_rate": 3.112845676378776e-05, + "loss": 0.4579, + "step": 18908 + }, + { + "epoch": 1.2690513741149627, + "grad_norm": 1.0206085443496704, + "learning_rate": 3.1118392747501854e-05, + "loss": 0.5222, + "step": 18910 + }, + { + "epoch": 1.2691855977987316, + "grad_norm": 0.9408379197120667, + "learning_rate": 3.1108329623334766e-05, + "loss": 0.494, + "step": 18912 + }, + { + "epoch": 1.2693198214825006, + "grad_norm": 0.9950227737426758, + "learning_rate": 3.109826739176195e-05, + "loss": 0.4814, + "step": 18914 + }, + { + "epoch": 1.2694540451662695, + "grad_norm": 1.0771654844284058, + "learning_rate": 3.108820605325883e-05, + "loss": 0.5509, + "step": 18916 + }, + { + "epoch": 1.2695882688500386, + "grad_norm": 1.1590389013290405, + "learning_rate": 3.1078145608300776e-05, + "loss": 0.5334, + "step": 18918 + }, + { + "epoch": 1.2697224925338075, + "grad_norm": 0.8306449055671692, + "learning_rate": 3.1068086057363155e-05, + "loss": 0.4865, + "step": 18920 + }, + { + "epoch": 1.2698567162175767, + "grad_norm": 1.0557630062103271, + "learning_rate": 3.105802740092122e-05, + "loss": 0.507, + "step": 18922 + }, + { + "epoch": 1.2699909399013456, + "grad_norm": 1.056467890739441, + "learning_rate": 3.104796963945027e-05, + "loss": 0.5766, + "step": 18924 + }, + { + "epoch": 1.2701251635851145, + "grad_norm": 1.1036617755889893, + "learning_rate": 3.103791277342545e-05, + "loss": 0.489, + "step": 18926 + }, + { + "epoch": 1.2702593872688837, + "grad_norm": 1.1211748123168945, + "learning_rate": 3.102785680332199e-05, + "loss": 0.5598, + "step": 18928 + }, + { + "epoch": 1.2703936109526526, + "grad_norm": 1.1051607131958008, + "learning_rate": 3.101780172961497e-05, + "loss": 0.5387, + "step": 18930 + }, + { + "epoch": 1.2705278346364217, + "grad_norm": 1.058781623840332, + "learning_rate": 3.1007747552779515e-05, + "loss": 0.5104, + "step": 18932 + }, + { + "epoch": 1.2706620583201906, + "grad_norm": 0.948963463306427, + "learning_rate": 3.099769427329062e-05, + "loss": 0.5391, + "step": 18934 + }, + { + "epoch": 1.2707962820039596, + "grad_norm": 1.0071995258331299, + "learning_rate": 3.098764189162332e-05, + "loss": 0.4627, + "step": 18936 + }, + { + "epoch": 1.2709305056877285, + "grad_norm": 1.0042026042938232, + "learning_rate": 3.0977590408252546e-05, + "loss": 0.5054, + "step": 18938 + }, + { + "epoch": 1.2710647293714976, + "grad_norm": 0.9338868856430054, + "learning_rate": 3.096753982365323e-05, + "loss": 0.5491, + "step": 18940 + }, + { + "epoch": 1.2711989530552665, + "grad_norm": 1.0394976139068604, + "learning_rate": 3.0957490138300215e-05, + "loss": 0.5156, + "step": 18942 + }, + { + "epoch": 1.2713331767390357, + "grad_norm": 0.9292700886726379, + "learning_rate": 3.094744135266836e-05, + "loss": 0.4393, + "step": 18944 + }, + { + "epoch": 1.2714674004228046, + "grad_norm": 1.1042059659957886, + "learning_rate": 3.093739346723242e-05, + "loss": 0.4668, + "step": 18946 + }, + { + "epoch": 1.2716016241065735, + "grad_norm": 1.0075424909591675, + "learning_rate": 3.092734648246718e-05, + "loss": 0.5124, + "step": 18948 + }, + { + "epoch": 1.2717358477903427, + "grad_norm": 1.0175364017486572, + "learning_rate": 3.091730039884728e-05, + "loss": 0.5205, + "step": 18950 + }, + { + "epoch": 1.2718700714741116, + "grad_norm": 1.0737347602844238, + "learning_rate": 3.090725521684744e-05, + "loss": 0.4697, + "step": 18952 + }, + { + "epoch": 1.2720042951578807, + "grad_norm": 0.8391603231430054, + "learning_rate": 3.089721093694224e-05, + "loss": 0.4156, + "step": 18954 + }, + { + "epoch": 1.2721385188416496, + "grad_norm": 0.9844167232513428, + "learning_rate": 3.088716755960624e-05, + "loss": 0.5035, + "step": 18956 + }, + { + "epoch": 1.2722727425254186, + "grad_norm": 1.1497169733047485, + "learning_rate": 3.087712508531402e-05, + "loss": 0.5162, + "step": 18958 + }, + { + "epoch": 1.2724069662091875, + "grad_norm": 1.0359950065612793, + "learning_rate": 3.0867083514540004e-05, + "loss": 0.5119, + "step": 18960 + }, + { + "epoch": 1.2725411898929566, + "grad_norm": 1.0207481384277344, + "learning_rate": 3.085704284775869e-05, + "loss": 0.457, + "step": 18962 + }, + { + "epoch": 1.2726754135767255, + "grad_norm": 1.1166799068450928, + "learning_rate": 3.084700308544445e-05, + "loss": 0.5362, + "step": 18964 + }, + { + "epoch": 1.2728096372604947, + "grad_norm": 1.011577844619751, + "learning_rate": 3.083696422807166e-05, + "loss": 0.5093, + "step": 18966 + }, + { + "epoch": 1.2729438609442636, + "grad_norm": 0.7684579491615295, + "learning_rate": 3.082692627611462e-05, + "loss": 0.4695, + "step": 18968 + }, + { + "epoch": 1.2730780846280325, + "grad_norm": 1.02020263671875, + "learning_rate": 3.081688923004762e-05, + "loss": 0.5184, + "step": 18970 + }, + { + "epoch": 1.2732123083118017, + "grad_norm": 1.0614569187164307, + "learning_rate": 3.080685309034487e-05, + "loss": 0.5676, + "step": 18972 + }, + { + "epoch": 1.2733465319955706, + "grad_norm": 1.0228722095489502, + "learning_rate": 3.079681785748057e-05, + "loss": 0.5098, + "step": 18974 + }, + { + "epoch": 1.2734807556793397, + "grad_norm": 0.8926354050636292, + "learning_rate": 3.078678353192887e-05, + "loss": 0.4811, + "step": 18976 + }, + { + "epoch": 1.2736149793631086, + "grad_norm": 0.8983010053634644, + "learning_rate": 3.0776750114163876e-05, + "loss": 0.4778, + "step": 18978 + }, + { + "epoch": 1.2737492030468776, + "grad_norm": 1.0334417819976807, + "learning_rate": 3.076671760465961e-05, + "loss": 0.5023, + "step": 18980 + }, + { + "epoch": 1.2738834267306467, + "grad_norm": 1.8472157716751099, + "learning_rate": 3.0756686003890156e-05, + "loss": 0.5451, + "step": 18982 + }, + { + "epoch": 1.2740176504144156, + "grad_norm": 0.9782081842422485, + "learning_rate": 3.074665531232941e-05, + "loss": 0.4861, + "step": 18984 + }, + { + "epoch": 1.2741518740981848, + "grad_norm": 0.9048913717269897, + "learning_rate": 3.073662553045138e-05, + "loss": 0.4748, + "step": 18986 + }, + { + "epoch": 1.2742860977819537, + "grad_norm": 1.0437411069869995, + "learning_rate": 3.072659665872988e-05, + "loss": 0.4714, + "step": 18988 + }, + { + "epoch": 1.2744203214657226, + "grad_norm": 0.9941788911819458, + "learning_rate": 3.07165686976388e-05, + "loss": 0.5814, + "step": 18990 + }, + { + "epoch": 1.2745545451494915, + "grad_norm": 1.147536277770996, + "learning_rate": 3.070654164765193e-05, + "loss": 0.5174, + "step": 18992 + }, + { + "epoch": 1.2746887688332607, + "grad_norm": 1.0783060789108276, + "learning_rate": 3.0696515509243026e-05, + "loss": 0.487, + "step": 18994 + }, + { + "epoch": 1.2748229925170296, + "grad_norm": 0.9300075173377991, + "learning_rate": 3.068649028288581e-05, + "loss": 0.5002, + "step": 18996 + }, + { + "epoch": 1.2749572162007987, + "grad_norm": 1.0361125469207764, + "learning_rate": 3.0676465969053946e-05, + "loss": 0.5958, + "step": 18998 + }, + { + "epoch": 1.2750914398845676, + "grad_norm": 1.1151272058486938, + "learning_rate": 3.0666442568221065e-05, + "loss": 0.5107, + "step": 19000 + }, + { + "epoch": 1.2752256635683366, + "grad_norm": 0.9330003261566162, + "learning_rate": 3.065642008086076e-05, + "loss": 0.4949, + "step": 19002 + }, + { + "epoch": 1.2753598872521057, + "grad_norm": 0.9569537043571472, + "learning_rate": 3.064639850744657e-05, + "loss": 0.5493, + "step": 19004 + }, + { + "epoch": 1.2754941109358746, + "grad_norm": 1.061761736869812, + "learning_rate": 3.063637784845199e-05, + "loss": 0.5463, + "step": 19006 + }, + { + "epoch": 1.2756283346196438, + "grad_norm": 1.055582880973816, + "learning_rate": 3.0626358104350475e-05, + "loss": 0.4809, + "step": 19008 + }, + { + "epoch": 1.2757625583034127, + "grad_norm": 1.039762020111084, + "learning_rate": 3.0616339275615455e-05, + "loss": 0.5304, + "step": 19010 + }, + { + "epoch": 1.2758967819871816, + "grad_norm": 1.0182931423187256, + "learning_rate": 3.0606321362720267e-05, + "loss": 0.4901, + "step": 19012 + }, + { + "epoch": 1.2760310056709505, + "grad_norm": 0.9698779582977295, + "learning_rate": 3.059630436613829e-05, + "loss": 0.4836, + "step": 19014 + }, + { + "epoch": 1.2761652293547197, + "grad_norm": 0.9517755508422852, + "learning_rate": 3.058628828634274e-05, + "loss": 0.482, + "step": 19016 + }, + { + "epoch": 1.2762994530384886, + "grad_norm": 1.0961393117904663, + "learning_rate": 3.057627312380692e-05, + "loss": 0.5077, + "step": 19018 + }, + { + "epoch": 1.2764336767222577, + "grad_norm": 1.094814419746399, + "learning_rate": 3.0566258879003986e-05, + "loss": 0.5784, + "step": 19020 + }, + { + "epoch": 1.2765679004060266, + "grad_norm": 1.0417190790176392, + "learning_rate": 3.055624555240711e-05, + "loss": 0.4749, + "step": 19022 + }, + { + "epoch": 1.2767021240897956, + "grad_norm": 1.007354497909546, + "learning_rate": 3.0546233144489414e-05, + "loss": 0.4932, + "step": 19024 + }, + { + "epoch": 1.2768363477735647, + "grad_norm": 1.0553150177001953, + "learning_rate": 3.053622165572392e-05, + "loss": 0.4851, + "step": 19026 + }, + { + "epoch": 1.2769705714573336, + "grad_norm": 0.872083306312561, + "learning_rate": 3.05262110865837e-05, + "loss": 0.4629, + "step": 19028 + }, + { + "epoch": 1.2771047951411028, + "grad_norm": 0.9458115100860596, + "learning_rate": 3.0516201437541693e-05, + "loss": 0.5173, + "step": 19030 + }, + { + "epoch": 1.2772390188248717, + "grad_norm": 1.1611723899841309, + "learning_rate": 3.0506192709070868e-05, + "loss": 0.5351, + "step": 19032 + }, + { + "epoch": 1.2773732425086406, + "grad_norm": 1.0064661502838135, + "learning_rate": 3.04961849016441e-05, + "loss": 0.4785, + "step": 19034 + }, + { + "epoch": 1.2775074661924095, + "grad_norm": 1.0834848880767822, + "learning_rate": 3.0486178015734246e-05, + "loss": 0.5189, + "step": 19036 + }, + { + "epoch": 1.2776416898761787, + "grad_norm": 1.0904122591018677, + "learning_rate": 3.04761720518141e-05, + "loss": 0.559, + "step": 19038 + }, + { + "epoch": 1.2777759135599476, + "grad_norm": 0.9285237193107605, + "learning_rate": 3.046616701035645e-05, + "loss": 0.4663, + "step": 19040 + }, + { + "epoch": 1.2779101372437167, + "grad_norm": 1.1051727533340454, + "learning_rate": 3.0456162891833978e-05, + "loss": 0.4522, + "step": 19042 + }, + { + "epoch": 1.2780443609274856, + "grad_norm": 0.9298257231712341, + "learning_rate": 3.044615969671939e-05, + "loss": 0.5165, + "step": 19044 + }, + { + "epoch": 1.2781785846112546, + "grad_norm": 0.9940264225006104, + "learning_rate": 3.043615742548529e-05, + "loss": 0.4633, + "step": 19046 + }, + { + "epoch": 1.2783128082950237, + "grad_norm": 1.0981922149658203, + "learning_rate": 3.0426156078604305e-05, + "loss": 0.4873, + "step": 19048 + }, + { + "epoch": 1.2784470319787926, + "grad_norm": 1.0183321237564087, + "learning_rate": 3.0416155656548928e-05, + "loss": 0.5701, + "step": 19050 + }, + { + "epoch": 1.2785812556625618, + "grad_norm": 1.0703001022338867, + "learning_rate": 3.040615615979171e-05, + "loss": 0.5107, + "step": 19052 + }, + { + "epoch": 1.2787154793463307, + "grad_norm": 1.0372713804244995, + "learning_rate": 3.0396157588805068e-05, + "loss": 0.4989, + "step": 19054 + }, + { + "epoch": 1.2788497030300996, + "grad_norm": 0.9407340288162231, + "learning_rate": 3.0386159944061444e-05, + "loss": 0.4823, + "step": 19056 + }, + { + "epoch": 1.2789839267138687, + "grad_norm": 1.0311121940612793, + "learning_rate": 3.0376163226033183e-05, + "loss": 0.485, + "step": 19058 + }, + { + "epoch": 1.2791181503976377, + "grad_norm": 1.3180968761444092, + "learning_rate": 3.0366167435192627e-05, + "loss": 0.5411, + "step": 19060 + }, + { + "epoch": 1.2792523740814068, + "grad_norm": 1.1690349578857422, + "learning_rate": 3.035617257201204e-05, + "loss": 0.4954, + "step": 19062 + }, + { + "epoch": 1.2793865977651757, + "grad_norm": 1.024618148803711, + "learning_rate": 3.034617863696369e-05, + "loss": 0.5332, + "step": 19064 + }, + { + "epoch": 1.2795208214489446, + "grad_norm": 1.085595726966858, + "learning_rate": 3.0336185630519737e-05, + "loss": 0.5139, + "step": 19066 + }, + { + "epoch": 1.2796550451327136, + "grad_norm": 0.8518025875091553, + "learning_rate": 3.032619355315236e-05, + "loss": 0.4813, + "step": 19068 + }, + { + "epoch": 1.2797892688164827, + "grad_norm": 1.0916320085525513, + "learning_rate": 3.0316202405333626e-05, + "loss": 0.5583, + "step": 19070 + }, + { + "epoch": 1.2799234925002516, + "grad_norm": 1.020693063735962, + "learning_rate": 3.0306212187535653e-05, + "loss": 0.4914, + "step": 19072 + }, + { + "epoch": 1.2800577161840208, + "grad_norm": 1.0341733694076538, + "learning_rate": 3.0296222900230397e-05, + "loss": 0.4996, + "step": 19074 + }, + { + "epoch": 1.2801919398677897, + "grad_norm": 1.2896177768707275, + "learning_rate": 3.0286234543889892e-05, + "loss": 0.6399, + "step": 19076 + }, + { + "epoch": 1.2803261635515586, + "grad_norm": 0.9945932626724243, + "learning_rate": 3.0276247118986013e-05, + "loss": 0.5082, + "step": 19078 + }, + { + "epoch": 1.2804603872353277, + "grad_norm": 0.8634927272796631, + "learning_rate": 3.026626062599069e-05, + "loss": 0.4289, + "step": 19080 + }, + { + "epoch": 1.2805946109190967, + "grad_norm": 1.334445834159851, + "learning_rate": 3.025627506537574e-05, + "loss": 0.5118, + "step": 19082 + }, + { + "epoch": 1.2807288346028658, + "grad_norm": 0.8973318934440613, + "learning_rate": 3.0246290437612974e-05, + "loss": 0.4608, + "step": 19084 + }, + { + "epoch": 1.2808630582866347, + "grad_norm": 1.1655054092407227, + "learning_rate": 3.0236306743174135e-05, + "loss": 0.5179, + "step": 19086 + }, + { + "epoch": 1.2809972819704036, + "grad_norm": 1.0608400106430054, + "learning_rate": 3.0226323982530946e-05, + "loss": 0.4694, + "step": 19088 + }, + { + "epoch": 1.2811315056541726, + "grad_norm": 1.2491742372512817, + "learning_rate": 3.0216342156155063e-05, + "loss": 0.5309, + "step": 19090 + }, + { + "epoch": 1.2812657293379417, + "grad_norm": 1.1221964359283447, + "learning_rate": 3.0206361264518106e-05, + "loss": 0.5641, + "step": 19092 + }, + { + "epoch": 1.2813999530217106, + "grad_norm": 0.9746096730232239, + "learning_rate": 3.019638130809167e-05, + "loss": 0.4834, + "step": 19094 + }, + { + "epoch": 1.2815341767054798, + "grad_norm": 1.0384272336959839, + "learning_rate": 3.018640228734726e-05, + "loss": 0.5042, + "step": 19096 + }, + { + "epoch": 1.2816684003892487, + "grad_norm": 0.9357865452766418, + "learning_rate": 3.0176424202756388e-05, + "loss": 0.4443, + "step": 19098 + }, + { + "epoch": 1.2818026240730176, + "grad_norm": 0.948830783367157, + "learning_rate": 3.016644705479048e-05, + "loss": 0.5083, + "step": 19100 + }, + { + "epoch": 1.2819368477567867, + "grad_norm": 0.9100587368011475, + "learning_rate": 3.0156470843920965e-05, + "loss": 0.4692, + "step": 19102 + }, + { + "epoch": 1.2820710714405557, + "grad_norm": 1.0029624700546265, + "learning_rate": 3.014649557061917e-05, + "loss": 0.4883, + "step": 19104 + }, + { + "epoch": 1.2822052951243248, + "grad_norm": 1.0393444299697876, + "learning_rate": 3.0136521235356425e-05, + "loss": 0.5089, + "step": 19106 + }, + { + "epoch": 1.2823395188080937, + "grad_norm": 1.058200716972351, + "learning_rate": 3.0126547838603976e-05, + "loss": 0.5405, + "step": 19108 + }, + { + "epoch": 1.2824737424918626, + "grad_norm": 1.0188109874725342, + "learning_rate": 3.0116575380833088e-05, + "loss": 0.5337, + "step": 19110 + }, + { + "epoch": 1.2826079661756316, + "grad_norm": 1.1087276935577393, + "learning_rate": 3.0106603862514882e-05, + "loss": 0.4425, + "step": 19112 + }, + { + "epoch": 1.2827421898594007, + "grad_norm": 0.9219624400138855, + "learning_rate": 3.0096633284120556e-05, + "loss": 0.4702, + "step": 19114 + }, + { + "epoch": 1.2828764135431696, + "grad_norm": 0.9535489082336426, + "learning_rate": 3.008666364612115e-05, + "loss": 0.4899, + "step": 19116 + }, + { + "epoch": 1.2830106372269388, + "grad_norm": 0.8020244240760803, + "learning_rate": 3.0076694948987738e-05, + "loss": 0.5104, + "step": 19118 + }, + { + "epoch": 1.2831448609107077, + "grad_norm": 0.9598153233528137, + "learning_rate": 3.0066727193191307e-05, + "loss": 0.4854, + "step": 19120 + }, + { + "epoch": 1.2832790845944766, + "grad_norm": 1.084541916847229, + "learning_rate": 3.0056760379202824e-05, + "loss": 0.5471, + "step": 19122 + }, + { + "epoch": 1.2834133082782457, + "grad_norm": 1.1065452098846436, + "learning_rate": 3.0046794507493192e-05, + "loss": 0.4721, + "step": 19124 + }, + { + "epoch": 1.2835475319620147, + "grad_norm": 0.9663994312286377, + "learning_rate": 3.0036829578533287e-05, + "loss": 0.4458, + "step": 19126 + }, + { + "epoch": 1.2836817556457838, + "grad_norm": 0.9064396619796753, + "learning_rate": 3.0026865592793928e-05, + "loss": 0.4571, + "step": 19128 + }, + { + "epoch": 1.2838159793295527, + "grad_norm": 1.063199520111084, + "learning_rate": 3.0016902550745897e-05, + "loss": 0.5314, + "step": 19130 + }, + { + "epoch": 1.2839502030133216, + "grad_norm": 0.9132474064826965, + "learning_rate": 3.0006940452859916e-05, + "loss": 0.5467, + "step": 19132 + }, + { + "epoch": 1.2840844266970908, + "grad_norm": 1.0902132987976074, + "learning_rate": 2.9996979299606697e-05, + "loss": 0.5334, + "step": 19134 + }, + { + "epoch": 1.2842186503808597, + "grad_norm": 0.9519376158714294, + "learning_rate": 2.9987019091456846e-05, + "loss": 0.4686, + "step": 19136 + }, + { + "epoch": 1.2843528740646288, + "grad_norm": 1.072738766670227, + "learning_rate": 2.9977059828881027e-05, + "loss": 0.5224, + "step": 19138 + }, + { + "epoch": 1.2844870977483978, + "grad_norm": 1.204823613166809, + "learning_rate": 2.996710151234972e-05, + "loss": 0.6014, + "step": 19140 + }, + { + "epoch": 1.2846213214321667, + "grad_norm": 1.0488771200180054, + "learning_rate": 2.9957144142333494e-05, + "loss": 0.4384, + "step": 19142 + }, + { + "epoch": 1.2847555451159356, + "grad_norm": 0.9874991774559021, + "learning_rate": 2.9947187719302792e-05, + "loss": 0.4718, + "step": 19144 + }, + { + "epoch": 1.2848897687997047, + "grad_norm": 1.00027334690094, + "learning_rate": 2.993723224372804e-05, + "loss": 0.5419, + "step": 19146 + }, + { + "epoch": 1.2850239924834737, + "grad_norm": 1.0105254650115967, + "learning_rate": 2.9927277716079605e-05, + "loss": 0.5151, + "step": 19148 + }, + { + "epoch": 1.2851582161672428, + "grad_norm": 0.9950233697891235, + "learning_rate": 2.991732413682784e-05, + "loss": 0.487, + "step": 19150 + }, + { + "epoch": 1.2852924398510117, + "grad_norm": 1.175166130065918, + "learning_rate": 2.9907371506443003e-05, + "loss": 0.5578, + "step": 19152 + }, + { + "epoch": 1.2854266635347806, + "grad_norm": 1.0153224468231201, + "learning_rate": 2.9897419825395367e-05, + "loss": 0.4839, + "step": 19154 + }, + { + "epoch": 1.2855608872185498, + "grad_norm": 1.0609861612319946, + "learning_rate": 2.9887469094155108e-05, + "loss": 0.5828, + "step": 19156 + }, + { + "epoch": 1.2856951109023187, + "grad_norm": 1.1728836297988892, + "learning_rate": 2.9877519313192386e-05, + "loss": 0.5589, + "step": 19158 + }, + { + "epoch": 1.2858293345860878, + "grad_norm": 1.1383951902389526, + "learning_rate": 2.9867570482977316e-05, + "loss": 0.5306, + "step": 19160 + }, + { + "epoch": 1.2859635582698568, + "grad_norm": 1.0976660251617432, + "learning_rate": 2.9857622603979933e-05, + "loss": 0.5036, + "step": 19162 + }, + { + "epoch": 1.2860977819536257, + "grad_norm": 0.9632569551467896, + "learning_rate": 2.9847675676670285e-05, + "loss": 0.5693, + "step": 19164 + }, + { + "epoch": 1.2862320056373946, + "grad_norm": 1.1621909141540527, + "learning_rate": 2.9837729701518325e-05, + "loss": 0.4986, + "step": 19166 + }, + { + "epoch": 1.2863662293211637, + "grad_norm": 1.0787842273712158, + "learning_rate": 2.9827784678994003e-05, + "loss": 0.4265, + "step": 19168 + }, + { + "epoch": 1.2865004530049327, + "grad_norm": 1.0132031440734863, + "learning_rate": 2.9817840609567166e-05, + "loss": 0.5004, + "step": 19170 + }, + { + "epoch": 1.2866346766887018, + "grad_norm": 1.047211766242981, + "learning_rate": 2.9807897493707703e-05, + "loss": 0.4887, + "step": 19172 + }, + { + "epoch": 1.2867689003724707, + "grad_norm": 1.1641210317611694, + "learning_rate": 2.9797955331885346e-05, + "loss": 0.5006, + "step": 19174 + }, + { + "epoch": 1.2869031240562396, + "grad_norm": 1.0389865636825562, + "learning_rate": 2.9788014124569895e-05, + "loss": 0.527, + "step": 19176 + }, + { + "epoch": 1.2870373477400088, + "grad_norm": 1.0260602235794067, + "learning_rate": 2.977807387223102e-05, + "loss": 0.4668, + "step": 19178 + }, + { + "epoch": 1.2871715714237777, + "grad_norm": 1.4412453174591064, + "learning_rate": 2.9768134575338402e-05, + "loss": 0.5256, + "step": 19180 + }, + { + "epoch": 1.2873057951075468, + "grad_norm": 0.9133895635604858, + "learning_rate": 2.975819623436163e-05, + "loss": 0.5167, + "step": 19182 + }, + { + "epoch": 1.2874400187913158, + "grad_norm": 0.9934803247451782, + "learning_rate": 2.9748258849770293e-05, + "loss": 0.5346, + "step": 19184 + }, + { + "epoch": 1.2875742424750847, + "grad_norm": 0.9523311257362366, + "learning_rate": 2.9738322422033895e-05, + "loss": 0.4901, + "step": 19186 + }, + { + "epoch": 1.2877084661588536, + "grad_norm": 1.0600826740264893, + "learning_rate": 2.9728386951621923e-05, + "loss": 0.522, + "step": 19188 + }, + { + "epoch": 1.2878426898426227, + "grad_norm": 1.0519115924835205, + "learning_rate": 2.9718452439003796e-05, + "loss": 0.4582, + "step": 19190 + }, + { + "epoch": 1.2879769135263917, + "grad_norm": 1.139541745185852, + "learning_rate": 2.9708518884648923e-05, + "loss": 0.5583, + "step": 19192 + }, + { + "epoch": 1.2881111372101608, + "grad_norm": 1.0064599514007568, + "learning_rate": 2.969858628902662e-05, + "loss": 0.5957, + "step": 19194 + }, + { + "epoch": 1.2882453608939297, + "grad_norm": 1.003114938735962, + "learning_rate": 2.9688654652606207e-05, + "loss": 0.4909, + "step": 19196 + }, + { + "epoch": 1.2883795845776986, + "grad_norm": 0.9284659028053284, + "learning_rate": 2.967872397585689e-05, + "loss": 0.5212, + "step": 19198 + }, + { + "epoch": 1.2885138082614678, + "grad_norm": 1.0107605457305908, + "learning_rate": 2.9668794259247945e-05, + "loss": 0.5099, + "step": 19200 + }, + { + "epoch": 1.2886480319452367, + "grad_norm": 0.9981799721717834, + "learning_rate": 2.9658865503248463e-05, + "loss": 0.5337, + "step": 19202 + }, + { + "epoch": 1.2887822556290058, + "grad_norm": 1.0402655601501465, + "learning_rate": 2.96489377083276e-05, + "loss": 0.5138, + "step": 19204 + }, + { + "epoch": 1.2889164793127748, + "grad_norm": 0.9867530465126038, + "learning_rate": 2.963901087495441e-05, + "loss": 0.5181, + "step": 19206 + }, + { + "epoch": 1.2890507029965437, + "grad_norm": 0.9896501898765564, + "learning_rate": 2.9629085003597918e-05, + "loss": 0.5025, + "step": 19208 + }, + { + "epoch": 1.2891849266803128, + "grad_norm": 0.958095371723175, + "learning_rate": 2.9619160094727093e-05, + "loss": 0.502, + "step": 19210 + }, + { + "epoch": 1.2893191503640817, + "grad_norm": 1.1133953332901, + "learning_rate": 2.960923614881089e-05, + "loss": 0.5428, + "step": 19212 + }, + { + "epoch": 1.2894533740478509, + "grad_norm": 1.0463382005691528, + "learning_rate": 2.9599313166318177e-05, + "loss": 0.5092, + "step": 19214 + }, + { + "epoch": 1.2895875977316198, + "grad_norm": 0.9142300486564636, + "learning_rate": 2.9589391147717803e-05, + "loss": 0.4992, + "step": 19216 + }, + { + "epoch": 1.2897218214153887, + "grad_norm": 1.1417711973190308, + "learning_rate": 2.9579470093478558e-05, + "loss": 0.4988, + "step": 19218 + }, + { + "epoch": 1.2898560450991576, + "grad_norm": 1.090346097946167, + "learning_rate": 2.9569550004069203e-05, + "loss": 0.4902, + "step": 19220 + }, + { + "epoch": 1.2899902687829268, + "grad_norm": 0.9615224599838257, + "learning_rate": 2.9559630879958422e-05, + "loss": 0.426, + "step": 19222 + }, + { + "epoch": 1.2901244924666957, + "grad_norm": 1.0067687034606934, + "learning_rate": 2.9549712721614902e-05, + "loss": 0.5547, + "step": 19224 + }, + { + "epoch": 1.2902587161504648, + "grad_norm": 1.2727751731872559, + "learning_rate": 2.953979552950722e-05, + "loss": 0.546, + "step": 19226 + }, + { + "epoch": 1.2903929398342338, + "grad_norm": 1.0097273588180542, + "learning_rate": 2.9529879304103997e-05, + "loss": 0.4805, + "step": 19228 + }, + { + "epoch": 1.2905271635180027, + "grad_norm": 1.0178848505020142, + "learning_rate": 2.9519964045873716e-05, + "loss": 0.5131, + "step": 19230 + }, + { + "epoch": 1.2906613872017718, + "grad_norm": 1.038375973701477, + "learning_rate": 2.951004975528484e-05, + "loss": 0.4862, + "step": 19232 + }, + { + "epoch": 1.2907956108855407, + "grad_norm": 0.9703474044799805, + "learning_rate": 2.9500136432805848e-05, + "loss": 0.5114, + "step": 19234 + }, + { + "epoch": 1.2909298345693099, + "grad_norm": 1.1284089088439941, + "learning_rate": 2.949022407890507e-05, + "loss": 0.5053, + "step": 19236 + }, + { + "epoch": 1.2910640582530788, + "grad_norm": 1.1331610679626465, + "learning_rate": 2.9480312694050905e-05, + "loss": 0.4618, + "step": 19238 + }, + { + "epoch": 1.2911982819368477, + "grad_norm": 1.259121060371399, + "learning_rate": 2.9470402278711584e-05, + "loss": 0.5041, + "step": 19240 + }, + { + "epoch": 1.2913325056206166, + "grad_norm": 0.8875788450241089, + "learning_rate": 2.9460492833355407e-05, + "loss": 0.4756, + "step": 19242 + }, + { + "epoch": 1.2914667293043858, + "grad_norm": 1.0083508491516113, + "learning_rate": 2.945058435845054e-05, + "loss": 0.5571, + "step": 19244 + }, + { + "epoch": 1.2916009529881547, + "grad_norm": 1.2376141548156738, + "learning_rate": 2.9440676854465165e-05, + "loss": 0.5199, + "step": 19246 + }, + { + "epoch": 1.2917351766719238, + "grad_norm": 0.9780316948890686, + "learning_rate": 2.9430770321867374e-05, + "loss": 0.501, + "step": 19248 + }, + { + "epoch": 1.2918694003556928, + "grad_norm": 0.9901347756385803, + "learning_rate": 2.942086476112525e-05, + "loss": 0.4607, + "step": 19250 + }, + { + "epoch": 1.2920036240394617, + "grad_norm": 0.9821428656578064, + "learning_rate": 2.941096017270678e-05, + "loss": 0.5326, + "step": 19252 + }, + { + "epoch": 1.2921378477232308, + "grad_norm": 0.9778178334236145, + "learning_rate": 2.9401056557079975e-05, + "loss": 0.4725, + "step": 19254 + }, + { + "epoch": 1.2922720714069997, + "grad_norm": 1.065484881401062, + "learning_rate": 2.9391153914712722e-05, + "loss": 0.5688, + "step": 19256 + }, + { + "epoch": 1.2924062950907689, + "grad_norm": 1.1318145990371704, + "learning_rate": 2.938125224607294e-05, + "loss": 0.5338, + "step": 19258 + }, + { + "epoch": 1.2925405187745378, + "grad_norm": 1.0310784578323364, + "learning_rate": 2.937135155162842e-05, + "loss": 0.5378, + "step": 19260 + }, + { + "epoch": 1.2926747424583067, + "grad_norm": 1.0071526765823364, + "learning_rate": 2.9361451831847004e-05, + "loss": 0.4794, + "step": 19262 + }, + { + "epoch": 1.2928089661420756, + "grad_norm": 1.001133918762207, + "learning_rate": 2.935155308719637e-05, + "loss": 0.5911, + "step": 19264 + }, + { + "epoch": 1.2929431898258448, + "grad_norm": 1.0991449356079102, + "learning_rate": 2.9341655318144278e-05, + "loss": 0.505, + "step": 19266 + }, + { + "epoch": 1.2930774135096137, + "grad_norm": 1.0803444385528564, + "learning_rate": 2.9331758525158338e-05, + "loss": 0.5172, + "step": 19268 + }, + { + "epoch": 1.2932116371933828, + "grad_norm": 1.080772876739502, + "learning_rate": 2.9321862708706172e-05, + "loss": 0.5521, + "step": 19270 + }, + { + "epoch": 1.2933458608771518, + "grad_norm": 1.0210713148117065, + "learning_rate": 2.9311967869255324e-05, + "loss": 0.533, + "step": 19272 + }, + { + "epoch": 1.2934800845609207, + "grad_norm": 1.0075637102127075, + "learning_rate": 2.9302074007273317e-05, + "loss": 0.5399, + "step": 19274 + }, + { + "epoch": 1.2936143082446898, + "grad_norm": 1.0783876180648804, + "learning_rate": 2.9292181123227612e-05, + "loss": 0.5131, + "step": 19276 + }, + { + "epoch": 1.2937485319284587, + "grad_norm": 1.3644986152648926, + "learning_rate": 2.9282289217585633e-05, + "loss": 0.4527, + "step": 19278 + }, + { + "epoch": 1.2938827556122279, + "grad_norm": 0.9839290380477905, + "learning_rate": 2.927239829081474e-05, + "loss": 0.502, + "step": 19280 + }, + { + "epoch": 1.2940169792959968, + "grad_norm": 0.9688754677772522, + "learning_rate": 2.9262508343382276e-05, + "loss": 0.4621, + "step": 19282 + }, + { + "epoch": 1.2941512029797657, + "grad_norm": 1.1157327890396118, + "learning_rate": 2.9252619375755508e-05, + "loss": 0.447, + "step": 19284 + }, + { + "epoch": 1.2942854266635349, + "grad_norm": 0.9797788858413696, + "learning_rate": 2.9242731388401685e-05, + "loss": 0.473, + "step": 19286 + }, + { + "epoch": 1.2944196503473038, + "grad_norm": 1.1115068197250366, + "learning_rate": 2.9232844381787967e-05, + "loss": 0.5085, + "step": 19288 + }, + { + "epoch": 1.294553874031073, + "grad_norm": 0.9349199533462524, + "learning_rate": 2.9222958356381547e-05, + "loss": 0.561, + "step": 19290 + }, + { + "epoch": 1.2946880977148418, + "grad_norm": 1.0068470239639282, + "learning_rate": 2.9213073312649452e-05, + "loss": 0.4471, + "step": 19292 + }, + { + "epoch": 1.2948223213986108, + "grad_norm": 0.9416777491569519, + "learning_rate": 2.9203189251058792e-05, + "loss": 0.4911, + "step": 19294 + }, + { + "epoch": 1.2949565450823797, + "grad_norm": 1.0156686305999756, + "learning_rate": 2.9193306172076553e-05, + "loss": 0.5114, + "step": 19296 + }, + { + "epoch": 1.2950907687661488, + "grad_norm": 1.1254180669784546, + "learning_rate": 2.9183424076169653e-05, + "loss": 0.4796, + "step": 19298 + }, + { + "epoch": 1.2952249924499177, + "grad_norm": 0.9957659840583801, + "learning_rate": 2.9173542963805058e-05, + "loss": 0.4872, + "step": 19300 + }, + { + "epoch": 1.2953592161336869, + "grad_norm": 0.9471935629844666, + "learning_rate": 2.916366283544959e-05, + "loss": 0.4351, + "step": 19302 + }, + { + "epoch": 1.2954934398174558, + "grad_norm": 1.0207605361938477, + "learning_rate": 2.915378369157009e-05, + "loss": 0.5233, + "step": 19304 + }, + { + "epoch": 1.2956276635012247, + "grad_norm": 1.0082496404647827, + "learning_rate": 2.91439055326333e-05, + "loss": 0.4257, + "step": 19306 + }, + { + "epoch": 1.2957618871849939, + "grad_norm": 1.0640394687652588, + "learning_rate": 2.913402835910598e-05, + "loss": 0.5308, + "step": 19308 + }, + { + "epoch": 1.2958961108687628, + "grad_norm": 1.0955893993377686, + "learning_rate": 2.9124152171454766e-05, + "loss": 0.5193, + "step": 19310 + }, + { + "epoch": 1.296030334552532, + "grad_norm": 1.0826829671859741, + "learning_rate": 2.9114276970146355e-05, + "loss": 0.5126, + "step": 19312 + }, + { + "epoch": 1.2961645582363008, + "grad_norm": 3.258791923522949, + "learning_rate": 2.910440275564724e-05, + "loss": 0.4691, + "step": 19314 + }, + { + "epoch": 1.2962987819200698, + "grad_norm": 0.9478439688682556, + "learning_rate": 2.9094529528424032e-05, + "loss": 0.5007, + "step": 19316 + }, + { + "epoch": 1.2964330056038387, + "grad_norm": 1.0688092708587646, + "learning_rate": 2.9084657288943174e-05, + "loss": 0.523, + "step": 19318 + }, + { + "epoch": 1.2965672292876078, + "grad_norm": 1.0586588382720947, + "learning_rate": 2.9074786037671153e-05, + "loss": 0.5582, + "step": 19320 + }, + { + "epoch": 1.2967014529713767, + "grad_norm": 1.011034369468689, + "learning_rate": 2.9064915775074342e-05, + "loss": 0.5154, + "step": 19322 + }, + { + "epoch": 1.2968356766551459, + "grad_norm": 1.160041093826294, + "learning_rate": 2.905504650161909e-05, + "loss": 0.4864, + "step": 19324 + }, + { + "epoch": 1.2969699003389148, + "grad_norm": 1.1135926246643066, + "learning_rate": 2.9045178217771684e-05, + "loss": 0.5105, + "step": 19326 + }, + { + "epoch": 1.2971041240226837, + "grad_norm": 2.92923903465271, + "learning_rate": 2.9035310923998427e-05, + "loss": 0.4478, + "step": 19328 + }, + { + "epoch": 1.2972383477064529, + "grad_norm": 0.9935290813446045, + "learning_rate": 2.9025444620765502e-05, + "loss": 0.4581, + "step": 19330 + }, + { + "epoch": 1.2973725713902218, + "grad_norm": 1.0143234729766846, + "learning_rate": 2.901557930853907e-05, + "loss": 0.4711, + "step": 19332 + }, + { + "epoch": 1.297506795073991, + "grad_norm": 0.9148914217948914, + "learning_rate": 2.9005714987785236e-05, + "loss": 0.5094, + "step": 19334 + }, + { + "epoch": 1.2976410187577598, + "grad_norm": 0.9044680595397949, + "learning_rate": 2.8995851658970102e-05, + "loss": 0.4913, + "step": 19336 + }, + { + "epoch": 1.2977752424415288, + "grad_norm": 1.0622960329055786, + "learning_rate": 2.898598932255966e-05, + "loss": 0.5493, + "step": 19338 + }, + { + "epoch": 1.2979094661252977, + "grad_norm": 1.059946060180664, + "learning_rate": 2.8976127979019934e-05, + "loss": 0.5537, + "step": 19340 + }, + { + "epoch": 1.2980436898090668, + "grad_norm": 1.0593196153640747, + "learning_rate": 2.896626762881678e-05, + "loss": 0.5263, + "step": 19342 + }, + { + "epoch": 1.2981779134928357, + "grad_norm": 0.8505914807319641, + "learning_rate": 2.8956408272416148e-05, + "loss": 0.4574, + "step": 19344 + }, + { + "epoch": 1.2983121371766049, + "grad_norm": 0.9502744674682617, + "learning_rate": 2.8946549910283817e-05, + "loss": 0.531, + "step": 19346 + }, + { + "epoch": 1.2984463608603738, + "grad_norm": 1.0109035968780518, + "learning_rate": 2.8936692542885617e-05, + "loss": 0.5213, + "step": 19348 + }, + { + "epoch": 1.2985805845441427, + "grad_norm": 1.015946626663208, + "learning_rate": 2.8926836170687284e-05, + "loss": 0.4924, + "step": 19350 + }, + { + "epoch": 1.2987148082279119, + "grad_norm": 1.1477079391479492, + "learning_rate": 2.8916980794154503e-05, + "loss": 0.526, + "step": 19352 + }, + { + "epoch": 1.2988490319116808, + "grad_norm": 1.1128402948379517, + "learning_rate": 2.8907126413752895e-05, + "loss": 0.5568, + "step": 19354 + }, + { + "epoch": 1.29898325559545, + "grad_norm": 1.0364487171173096, + "learning_rate": 2.889727302994811e-05, + "loss": 0.6144, + "step": 19356 + }, + { + "epoch": 1.2991174792792188, + "grad_norm": 0.9741474986076355, + "learning_rate": 2.8887420643205682e-05, + "loss": 0.4971, + "step": 19358 + }, + { + "epoch": 1.2992517029629878, + "grad_norm": 1.0670655965805054, + "learning_rate": 2.887756925399111e-05, + "loss": 0.4898, + "step": 19360 + }, + { + "epoch": 1.299385926646757, + "grad_norm": 0.8790115714073181, + "learning_rate": 2.8867718862769834e-05, + "loss": 0.4672, + "step": 19362 + }, + { + "epoch": 1.2995201503305258, + "grad_norm": 1.0886361598968506, + "learning_rate": 2.8857869470007302e-05, + "loss": 0.4906, + "step": 19364 + }, + { + "epoch": 1.299654374014295, + "grad_norm": 0.8426548838615417, + "learning_rate": 2.8848021076168875e-05, + "loss": 0.4661, + "step": 19366 + }, + { + "epoch": 1.2997885976980639, + "grad_norm": 1.0217941999435425, + "learning_rate": 2.883817368171985e-05, + "loss": 0.504, + "step": 19368 + }, + { + "epoch": 1.2999228213818328, + "grad_norm": 1.404261827468872, + "learning_rate": 2.882832728712551e-05, + "loss": 0.554, + "step": 19370 + }, + { + "epoch": 1.3000570450656017, + "grad_norm": 0.9625460505485535, + "learning_rate": 2.881848189285105e-05, + "loss": 0.5089, + "step": 19372 + }, + { + "epoch": 1.3001912687493709, + "grad_norm": 0.9874593615531921, + "learning_rate": 2.880863749936169e-05, + "loss": 0.5137, + "step": 19374 + }, + { + "epoch": 1.3003254924331398, + "grad_norm": 1.07978093624115, + "learning_rate": 2.879879410712252e-05, + "loss": 0.4855, + "step": 19376 + }, + { + "epoch": 1.300459716116909, + "grad_norm": 0.9420575499534607, + "learning_rate": 2.8788951716598656e-05, + "loss": 0.4991, + "step": 19378 + }, + { + "epoch": 1.3005939398006778, + "grad_norm": 0.9697789549827576, + "learning_rate": 2.877911032825511e-05, + "loss": 0.4976, + "step": 19380 + }, + { + "epoch": 1.3007281634844468, + "grad_norm": 1.011218547821045, + "learning_rate": 2.8769269942556875e-05, + "loss": 0.5941, + "step": 19382 + }, + { + "epoch": 1.300862387168216, + "grad_norm": 1.0524111986160278, + "learning_rate": 2.875943055996887e-05, + "loss": 0.5196, + "step": 19384 + }, + { + "epoch": 1.3009966108519848, + "grad_norm": 1.0156795978546143, + "learning_rate": 2.874959218095602e-05, + "loss": 0.4749, + "step": 19386 + }, + { + "epoch": 1.301130834535754, + "grad_norm": 0.8876588940620422, + "learning_rate": 2.873975480598315e-05, + "loss": 0.4607, + "step": 19388 + }, + { + "epoch": 1.3012650582195229, + "grad_norm": 0.9444760680198669, + "learning_rate": 2.8729918435515058e-05, + "loss": 0.4715, + "step": 19390 + }, + { + "epoch": 1.3013992819032918, + "grad_norm": 1.092552661895752, + "learning_rate": 2.872008307001648e-05, + "loss": 0.4848, + "step": 19392 + }, + { + "epoch": 1.3015335055870607, + "grad_norm": 1.0454710721969604, + "learning_rate": 2.8710248709952147e-05, + "loss": 0.5212, + "step": 19394 + }, + { + "epoch": 1.3016677292708299, + "grad_norm": 1.009968876838684, + "learning_rate": 2.8700415355786704e-05, + "loss": 0.5196, + "step": 19396 + }, + { + "epoch": 1.3018019529545988, + "grad_norm": 0.9652559161186218, + "learning_rate": 2.8690583007984745e-05, + "loss": 0.5146, + "step": 19398 + }, + { + "epoch": 1.301936176638368, + "grad_norm": 0.9651011228561401, + "learning_rate": 2.868075166701082e-05, + "loss": 0.4049, + "step": 19400 + }, + { + "epoch": 1.3020704003221368, + "grad_norm": 1.0508171319961548, + "learning_rate": 2.867092133332947e-05, + "loss": 0.5036, + "step": 19402 + }, + { + "epoch": 1.3022046240059058, + "grad_norm": 0.9191226959228516, + "learning_rate": 2.8661092007405132e-05, + "loss": 0.4603, + "step": 19404 + }, + { + "epoch": 1.302338847689675, + "grad_norm": 0.8831053972244263, + "learning_rate": 2.8651263689702256e-05, + "loss": 0.4623, + "step": 19406 + }, + { + "epoch": 1.3024730713734438, + "grad_norm": 1.0167165994644165, + "learning_rate": 2.8641436380685184e-05, + "loss": 0.4737, + "step": 19408 + }, + { + "epoch": 1.302607295057213, + "grad_norm": 0.9504034519195557, + "learning_rate": 2.863161008081825e-05, + "loss": 0.4663, + "step": 19410 + }, + { + "epoch": 1.3027415187409819, + "grad_norm": 1.011612892150879, + "learning_rate": 2.8621784790565696e-05, + "loss": 0.4308, + "step": 19412 + }, + { + "epoch": 1.3028757424247508, + "grad_norm": 1.1726253032684326, + "learning_rate": 2.8611960510391795e-05, + "loss": 0.5226, + "step": 19414 + }, + { + "epoch": 1.3030099661085197, + "grad_norm": 1.0991902351379395, + "learning_rate": 2.8602137240760695e-05, + "loss": 0.5698, + "step": 19416 + }, + { + "epoch": 1.3031441897922889, + "grad_norm": 0.9891123175621033, + "learning_rate": 2.859231498213654e-05, + "loss": 0.4947, + "step": 19418 + }, + { + "epoch": 1.3032784134760578, + "grad_norm": 0.8932777643203735, + "learning_rate": 2.8582493734983384e-05, + "loss": 0.511, + "step": 19420 + }, + { + "epoch": 1.303412637159827, + "grad_norm": 0.9994890689849854, + "learning_rate": 2.8572673499765307e-05, + "loss": 0.4924, + "step": 19422 + }, + { + "epoch": 1.3035468608435958, + "grad_norm": 1.1127361059188843, + "learning_rate": 2.856285427694627e-05, + "loss": 0.5761, + "step": 19424 + }, + { + "epoch": 1.3036810845273648, + "grad_norm": 0.9759175777435303, + "learning_rate": 2.8553036066990214e-05, + "loss": 0.4683, + "step": 19426 + }, + { + "epoch": 1.303815308211134, + "grad_norm": 1.0085070133209229, + "learning_rate": 2.854321887036101e-05, + "loss": 0.4728, + "step": 19428 + }, + { + "epoch": 1.3039495318949028, + "grad_norm": 1.080127239227295, + "learning_rate": 2.8533402687522538e-05, + "loss": 0.4812, + "step": 19430 + }, + { + "epoch": 1.304083755578672, + "grad_norm": 1.0481758117675781, + "learning_rate": 2.8523587518938554e-05, + "loss": 0.4898, + "step": 19432 + }, + { + "epoch": 1.3042179792624409, + "grad_norm": 0.9496031403541565, + "learning_rate": 2.851377336507286e-05, + "loss": 0.4965, + "step": 19434 + }, + { + "epoch": 1.3043522029462098, + "grad_norm": 1.1129904985427856, + "learning_rate": 2.8503960226389136e-05, + "loss": 0.5361, + "step": 19436 + }, + { + "epoch": 1.304486426629979, + "grad_norm": 0.9533309936523438, + "learning_rate": 2.8494148103350983e-05, + "loss": 0.4836, + "step": 19438 + }, + { + "epoch": 1.3046206503137479, + "grad_norm": 1.0634078979492188, + "learning_rate": 2.8484336996422057e-05, + "loss": 0.6061, + "step": 19440 + }, + { + "epoch": 1.304754873997517, + "grad_norm": 1.0369467735290527, + "learning_rate": 2.847452690606589e-05, + "loss": 0.5382, + "step": 19442 + }, + { + "epoch": 1.304889097681286, + "grad_norm": 0.9479460716247559, + "learning_rate": 2.8464717832746014e-05, + "loss": 0.5047, + "step": 19444 + }, + { + "epoch": 1.3050233213650548, + "grad_norm": 1.0582891702651978, + "learning_rate": 2.8454909776925865e-05, + "loss": 0.5063, + "step": 19446 + }, + { + "epoch": 1.3051575450488238, + "grad_norm": 0.9738956689834595, + "learning_rate": 2.8445102739068873e-05, + "loss": 0.4998, + "step": 19448 + }, + { + "epoch": 1.305291768732593, + "grad_norm": 1.0738046169281006, + "learning_rate": 2.8435296719638366e-05, + "loss": 0.5153, + "step": 19450 + }, + { + "epoch": 1.3054259924163618, + "grad_norm": 0.8883415460586548, + "learning_rate": 2.8425491719097707e-05, + "loss": 0.4196, + "step": 19452 + }, + { + "epoch": 1.305560216100131, + "grad_norm": 1.0990755558013916, + "learning_rate": 2.8415687737910134e-05, + "loss": 0.4818, + "step": 19454 + }, + { + "epoch": 1.3056944397838999, + "grad_norm": 1.189955234527588, + "learning_rate": 2.8405884776538876e-05, + "loss": 0.5433, + "step": 19456 + }, + { + "epoch": 1.3058286634676688, + "grad_norm": 1.0909309387207031, + "learning_rate": 2.839608283544708e-05, + "loss": 0.5031, + "step": 19458 + }, + { + "epoch": 1.305962887151438, + "grad_norm": 0.9963827729225159, + "learning_rate": 2.8386281915097907e-05, + "loss": 0.4778, + "step": 19460 + }, + { + "epoch": 1.3060971108352069, + "grad_norm": 0.9648716449737549, + "learning_rate": 2.8376482015954387e-05, + "loss": 0.437, + "step": 19462 + }, + { + "epoch": 1.306231334518976, + "grad_norm": 1.0067102909088135, + "learning_rate": 2.836668313847962e-05, + "loss": 0.5187, + "step": 19464 + }, + { + "epoch": 1.306365558202745, + "grad_norm": 1.1235438585281372, + "learning_rate": 2.8356885283136485e-05, + "loss": 0.5222, + "step": 19466 + }, + { + "epoch": 1.3064997818865138, + "grad_norm": 1.0012166500091553, + "learning_rate": 2.8347088450387986e-05, + "loss": 0.4864, + "step": 19468 + }, + { + "epoch": 1.3066340055702828, + "grad_norm": 1.0444602966308594, + "learning_rate": 2.833729264069696e-05, + "loss": 0.447, + "step": 19470 + }, + { + "epoch": 1.306768229254052, + "grad_norm": 0.9028139114379883, + "learning_rate": 2.8327497854526276e-05, + "loss": 0.4706, + "step": 19472 + }, + { + "epoch": 1.3069024529378208, + "grad_norm": 0.9326144456863403, + "learning_rate": 2.8317704092338703e-05, + "loss": 0.3933, + "step": 19474 + }, + { + "epoch": 1.30703667662159, + "grad_norm": 0.9908169507980347, + "learning_rate": 2.8307911354596978e-05, + "loss": 0.5367, + "step": 19476 + }, + { + "epoch": 1.3071709003053589, + "grad_norm": 0.9125736355781555, + "learning_rate": 2.8298119641763763e-05, + "loss": 0.4801, + "step": 19478 + }, + { + "epoch": 1.3073051239891278, + "grad_norm": 0.9635627269744873, + "learning_rate": 2.828832895430174e-05, + "loss": 0.5259, + "step": 19480 + }, + { + "epoch": 1.307439347672897, + "grad_norm": 1.1029691696166992, + "learning_rate": 2.827853929267348e-05, + "loss": 0.4381, + "step": 19482 + }, + { + "epoch": 1.3075735713566659, + "grad_norm": 0.9907110333442688, + "learning_rate": 2.8268750657341524e-05, + "loss": 0.5017, + "step": 19484 + }, + { + "epoch": 1.307707795040435, + "grad_norm": 0.8806270360946655, + "learning_rate": 2.825896304876835e-05, + "loss": 0.4635, + "step": 19486 + }, + { + "epoch": 1.307842018724204, + "grad_norm": 1.082257628440857, + "learning_rate": 2.8249176467416438e-05, + "loss": 0.5017, + "step": 19488 + }, + { + "epoch": 1.3079762424079728, + "grad_norm": 0.9463380575180054, + "learning_rate": 2.8239390913748144e-05, + "loss": 0.483, + "step": 19490 + }, + { + "epoch": 1.3081104660917418, + "grad_norm": 0.9221293926239014, + "learning_rate": 2.822960638822588e-05, + "loss": 0.4758, + "step": 19492 + }, + { + "epoch": 1.308244689775511, + "grad_norm": 1.0206997394561768, + "learning_rate": 2.8219822891311863e-05, + "loss": 0.4737, + "step": 19494 + }, + { + "epoch": 1.3083789134592798, + "grad_norm": 1.2292686700820923, + "learning_rate": 2.8210040423468408e-05, + "loss": 0.5189, + "step": 19496 + }, + { + "epoch": 1.308513137143049, + "grad_norm": 0.9695619940757751, + "learning_rate": 2.820025898515768e-05, + "loss": 0.4991, + "step": 19498 + }, + { + "epoch": 1.3086473608268179, + "grad_norm": 1.0157265663146973, + "learning_rate": 2.8190478576841862e-05, + "loss": 0.4968, + "step": 19500 + }, + { + "epoch": 1.3087815845105868, + "grad_norm": 1.0575307607650757, + "learning_rate": 2.8180699198983062e-05, + "loss": 0.5299, + "step": 19502 + }, + { + "epoch": 1.308915808194356, + "grad_norm": 1.0781140327453613, + "learning_rate": 2.8170920852043286e-05, + "loss": 0.5328, + "step": 19504 + }, + { + "epoch": 1.3090500318781249, + "grad_norm": 1.0960808992385864, + "learning_rate": 2.8161143536484592e-05, + "loss": 0.5449, + "step": 19506 + }, + { + "epoch": 1.309184255561894, + "grad_norm": 1.0679954290390015, + "learning_rate": 2.81513672527689e-05, + "loss": 0.4821, + "step": 19508 + }, + { + "epoch": 1.309318479245663, + "grad_norm": 0.8787037134170532, + "learning_rate": 2.8141592001358163e-05, + "loss": 0.4442, + "step": 19510 + }, + { + "epoch": 1.3094527029294318, + "grad_norm": 0.9036421179771423, + "learning_rate": 2.813181778271422e-05, + "loss": 0.4954, + "step": 19512 + }, + { + "epoch": 1.309586926613201, + "grad_norm": 1.6593456268310547, + "learning_rate": 2.8122044597298886e-05, + "loss": 0.5353, + "step": 19514 + }, + { + "epoch": 1.30972115029697, + "grad_norm": 0.9559675455093384, + "learning_rate": 2.8112272445573905e-05, + "loss": 0.5036, + "step": 19516 + }, + { + "epoch": 1.309855373980739, + "grad_norm": 1.2117623090744019, + "learning_rate": 2.810250132800103e-05, + "loss": 0.51, + "step": 19518 + }, + { + "epoch": 1.309989597664508, + "grad_norm": 1.0451873540878296, + "learning_rate": 2.8092731245041903e-05, + "loss": 0.5105, + "step": 19520 + }, + { + "epoch": 1.3101238213482769, + "grad_norm": 0.9599100351333618, + "learning_rate": 2.8082962197158148e-05, + "loss": 0.4935, + "step": 19522 + }, + { + "epoch": 1.3102580450320458, + "grad_norm": 0.9579646587371826, + "learning_rate": 2.8073194184811314e-05, + "loss": 0.5055, + "step": 19524 + }, + { + "epoch": 1.310392268715815, + "grad_norm": 0.9336437582969666, + "learning_rate": 2.8063427208462957e-05, + "loss": 0.5059, + "step": 19526 + }, + { + "epoch": 1.3105264923995839, + "grad_norm": 0.8615583777427673, + "learning_rate": 2.8053661268574505e-05, + "loss": 0.4374, + "step": 19528 + }, + { + "epoch": 1.310660716083353, + "grad_norm": 1.0524922609329224, + "learning_rate": 2.8043896365607447e-05, + "loss": 0.5419, + "step": 19530 + }, + { + "epoch": 1.310794939767122, + "grad_norm": 0.9826512336730957, + "learning_rate": 2.803413250002307e-05, + "loss": 0.513, + "step": 19532 + }, + { + "epoch": 1.3109291634508908, + "grad_norm": 0.9838429093360901, + "learning_rate": 2.8024369672282756e-05, + "loss": 0.4585, + "step": 19534 + }, + { + "epoch": 1.31106338713466, + "grad_norm": 1.0887281894683838, + "learning_rate": 2.8014607882847743e-05, + "loss": 0.4766, + "step": 19536 + }, + { + "epoch": 1.311197610818429, + "grad_norm": 1.03814697265625, + "learning_rate": 2.800484713217929e-05, + "loss": 0.5152, + "step": 19538 + }, + { + "epoch": 1.311331834502198, + "grad_norm": 1.0641275644302368, + "learning_rate": 2.7995087420738565e-05, + "loss": 0.4898, + "step": 19540 + }, + { + "epoch": 1.311466058185967, + "grad_norm": 2.263648509979248, + "learning_rate": 2.7985328748986682e-05, + "loss": 0.5257, + "step": 19542 + }, + { + "epoch": 1.3116002818697359, + "grad_norm": 0.9365549683570862, + "learning_rate": 2.7975571117384713e-05, + "loss": 0.4544, + "step": 19544 + }, + { + "epoch": 1.3117345055535048, + "grad_norm": 0.9181808829307556, + "learning_rate": 2.7965814526393718e-05, + "loss": 0.4883, + "step": 19546 + }, + { + "epoch": 1.311868729237274, + "grad_norm": 1.0153909921646118, + "learning_rate": 2.795605897647466e-05, + "loss": 0.5232, + "step": 19548 + }, + { + "epoch": 1.3120029529210429, + "grad_norm": 1.0753653049468994, + "learning_rate": 2.7946304468088463e-05, + "loss": 0.4537, + "step": 19550 + }, + { + "epoch": 1.312137176604812, + "grad_norm": 1.0972576141357422, + "learning_rate": 2.7936551001695992e-05, + "loss": 0.5292, + "step": 19552 + }, + { + "epoch": 1.312271400288581, + "grad_norm": 0.9783473014831543, + "learning_rate": 2.792679857775813e-05, + "loss": 0.4966, + "step": 19554 + }, + { + "epoch": 1.3124056239723498, + "grad_norm": 0.9188259840011597, + "learning_rate": 2.7917047196735602e-05, + "loss": 0.5125, + "step": 19556 + }, + { + "epoch": 1.312539847656119, + "grad_norm": 1.1747207641601562, + "learning_rate": 2.790729685908919e-05, + "loss": 0.526, + "step": 19558 + }, + { + "epoch": 1.312674071339888, + "grad_norm": 1.0269527435302734, + "learning_rate": 2.7897547565279557e-05, + "loss": 0.557, + "step": 19560 + }, + { + "epoch": 1.312808295023657, + "grad_norm": 1.1212568283081055, + "learning_rate": 2.788779931576734e-05, + "loss": 0.5312, + "step": 19562 + }, + { + "epoch": 1.312942518707426, + "grad_norm": 0.9822105169296265, + "learning_rate": 2.787805211101311e-05, + "loss": 0.5002, + "step": 19564 + }, + { + "epoch": 1.3130767423911949, + "grad_norm": 1.115074872970581, + "learning_rate": 2.7868305951477425e-05, + "loss": 0.5395, + "step": 19566 + }, + { + "epoch": 1.3132109660749638, + "grad_norm": 1.0336101055145264, + "learning_rate": 2.7858560837620773e-05, + "loss": 0.4518, + "step": 19568 + }, + { + "epoch": 1.313345189758733, + "grad_norm": 0.9578868746757507, + "learning_rate": 2.7848816769903574e-05, + "loss": 0.536, + "step": 19570 + }, + { + "epoch": 1.3134794134425019, + "grad_norm": 1.0108757019042969, + "learning_rate": 2.783907374878623e-05, + "loss": 0.495, + "step": 19572 + }, + { + "epoch": 1.313613637126271, + "grad_norm": 0.9814471006393433, + "learning_rate": 2.7829331774729056e-05, + "loss": 0.4853, + "step": 19574 + }, + { + "epoch": 1.31374786081004, + "grad_norm": 0.9435616731643677, + "learning_rate": 2.781959084819238e-05, + "loss": 0.4488, + "step": 19576 + }, + { + "epoch": 1.3138820844938088, + "grad_norm": 1.0283944606781006, + "learning_rate": 2.780985096963641e-05, + "loss": 0.4549, + "step": 19578 + }, + { + "epoch": 1.314016308177578, + "grad_norm": 0.9773117303848267, + "learning_rate": 2.780011213952135e-05, + "loss": 0.5221, + "step": 19580 + }, + { + "epoch": 1.314150531861347, + "grad_norm": 1.0373390913009644, + "learning_rate": 2.7790374358307327e-05, + "loss": 0.4952, + "step": 19582 + }, + { + "epoch": 1.314284755545116, + "grad_norm": 1.02969491481781, + "learning_rate": 2.7780637626454452e-05, + "loss": 0.5134, + "step": 19584 + }, + { + "epoch": 1.314418979228885, + "grad_norm": 1.0415964126586914, + "learning_rate": 2.7770901944422744e-05, + "loss": 0.4908, + "step": 19586 + }, + { + "epoch": 1.3145532029126539, + "grad_norm": 0.9986392855644226, + "learning_rate": 2.7761167312672242e-05, + "loss": 0.5128, + "step": 19588 + }, + { + "epoch": 1.314687426596423, + "grad_norm": 1.0504610538482666, + "learning_rate": 2.775143373166281e-05, + "loss": 0.4669, + "step": 19590 + }, + { + "epoch": 1.314821650280192, + "grad_norm": 1.1573312282562256, + "learning_rate": 2.7741701201854414e-05, + "loss": 0.4936, + "step": 19592 + }, + { + "epoch": 1.314955873963961, + "grad_norm": 1.3695416450500488, + "learning_rate": 2.773196972370684e-05, + "loss": 0.4605, + "step": 19594 + }, + { + "epoch": 1.31509009764773, + "grad_norm": 1.1023151874542236, + "learning_rate": 2.772223929767993e-05, + "loss": 0.5058, + "step": 19596 + }, + { + "epoch": 1.315224321331499, + "grad_norm": 0.9775272607803345, + "learning_rate": 2.771250992423341e-05, + "loss": 0.4713, + "step": 19598 + }, + { + "epoch": 1.3153585450152678, + "grad_norm": 1.0285531282424927, + "learning_rate": 2.7702781603826965e-05, + "loss": 0.5197, + "step": 19600 + }, + { + "epoch": 1.315492768699037, + "grad_norm": 1.1167120933532715, + "learning_rate": 2.7693054336920228e-05, + "loss": 0.5114, + "step": 19602 + }, + { + "epoch": 1.315626992382806, + "grad_norm": 0.9397265315055847, + "learning_rate": 2.7683328123972823e-05, + "loss": 0.4821, + "step": 19604 + }, + { + "epoch": 1.315761216066575, + "grad_norm": 0.9107526540756226, + "learning_rate": 2.7673602965444285e-05, + "loss": 0.5193, + "step": 19606 + }, + { + "epoch": 1.315895439750344, + "grad_norm": 0.9846282005310059, + "learning_rate": 2.76638788617941e-05, + "loss": 0.5116, + "step": 19608 + }, + { + "epoch": 1.3160296634341129, + "grad_norm": 0.9590386152267456, + "learning_rate": 2.76541558134817e-05, + "loss": 0.4889, + "step": 19610 + }, + { + "epoch": 1.316163887117882, + "grad_norm": 0.9941418766975403, + "learning_rate": 2.764443382096652e-05, + "loss": 0.5371, + "step": 19612 + }, + { + "epoch": 1.316298110801651, + "grad_norm": 1.1155186891555786, + "learning_rate": 2.7634712884707852e-05, + "loss": 0.4739, + "step": 19614 + }, + { + "epoch": 1.31643233448542, + "grad_norm": 1.1311336755752563, + "learning_rate": 2.7624993005165066e-05, + "loss": 0.5331, + "step": 19616 + }, + { + "epoch": 1.316566558169189, + "grad_norm": 0.9965952038764954, + "learning_rate": 2.7615274182797325e-05, + "loss": 0.478, + "step": 19618 + }, + { + "epoch": 1.316700781852958, + "grad_norm": 0.9219123721122742, + "learning_rate": 2.7605556418063877e-05, + "loss": 0.4746, + "step": 19620 + }, + { + "epoch": 1.3168350055367268, + "grad_norm": 1.1114145517349243, + "learning_rate": 2.759583971142383e-05, + "loss": 0.5478, + "step": 19622 + }, + { + "epoch": 1.316969229220496, + "grad_norm": 0.9328547120094299, + "learning_rate": 2.758612406333633e-05, + "loss": 0.506, + "step": 19624 + }, + { + "epoch": 1.317103452904265, + "grad_norm": 0.8412764072418213, + "learning_rate": 2.7576409474260378e-05, + "loss": 0.5242, + "step": 19626 + }, + { + "epoch": 1.317237676588034, + "grad_norm": 0.966641902923584, + "learning_rate": 2.7566695944654997e-05, + "loss": 0.4647, + "step": 19628 + }, + { + "epoch": 1.317371900271803, + "grad_norm": 0.9684114456176758, + "learning_rate": 2.7556983474979093e-05, + "loss": 0.4965, + "step": 19630 + }, + { + "epoch": 1.3175061239555719, + "grad_norm": 0.9587264060974121, + "learning_rate": 2.754727206569161e-05, + "loss": 0.5264, + "step": 19632 + }, + { + "epoch": 1.317640347639341, + "grad_norm": 1.021653652191162, + "learning_rate": 2.753756171725137e-05, + "loss": 0.508, + "step": 19634 + }, + { + "epoch": 1.31777457132311, + "grad_norm": 0.9952699542045593, + "learning_rate": 2.7527852430117167e-05, + "loss": 0.4698, + "step": 19636 + }, + { + "epoch": 1.317908795006879, + "grad_norm": 0.9360623359680176, + "learning_rate": 2.7518144204747732e-05, + "loss": 0.4765, + "step": 19638 + }, + { + "epoch": 1.318043018690648, + "grad_norm": 0.9587939381599426, + "learning_rate": 2.7508437041601786e-05, + "loss": 0.4329, + "step": 19640 + }, + { + "epoch": 1.318177242374417, + "grad_norm": 1.0178338289260864, + "learning_rate": 2.749873094113797e-05, + "loss": 0.5465, + "step": 19642 + }, + { + "epoch": 1.3183114660581858, + "grad_norm": 1.1787763833999634, + "learning_rate": 2.748902590381487e-05, + "loss": 0.4357, + "step": 19644 + }, + { + "epoch": 1.318445689741955, + "grad_norm": 1.0718886852264404, + "learning_rate": 2.7479321930091023e-05, + "loss": 0.4981, + "step": 19646 + }, + { + "epoch": 1.318579913425724, + "grad_norm": 1.336592435836792, + "learning_rate": 2.7469619020424913e-05, + "loss": 0.453, + "step": 19648 + }, + { + "epoch": 1.318714137109493, + "grad_norm": 1.0587416887283325, + "learning_rate": 2.7459917175275018e-05, + "loss": 0.4984, + "step": 19650 + }, + { + "epoch": 1.318848360793262, + "grad_norm": 1.0716010332107544, + "learning_rate": 2.7450216395099705e-05, + "loss": 0.5693, + "step": 19652 + }, + { + "epoch": 1.3189825844770309, + "grad_norm": 0.9708546996116638, + "learning_rate": 2.7440516680357354e-05, + "loss": 0.5115, + "step": 19654 + }, + { + "epoch": 1.3191168081608, + "grad_norm": 1.0340770483016968, + "learning_rate": 2.7430818031506188e-05, + "loss": 0.494, + "step": 19656 + }, + { + "epoch": 1.319251031844569, + "grad_norm": 1.0604135990142822, + "learning_rate": 2.742112044900451e-05, + "loss": 0.5184, + "step": 19658 + }, + { + "epoch": 1.319385255528338, + "grad_norm": 0.9252541065216064, + "learning_rate": 2.7411423933310476e-05, + "loss": 0.4121, + "step": 19660 + }, + { + "epoch": 1.319519479212107, + "grad_norm": 1.0229439735412598, + "learning_rate": 2.740172848488226e-05, + "loss": 0.4907, + "step": 19662 + }, + { + "epoch": 1.319653702895876, + "grad_norm": 0.9297716021537781, + "learning_rate": 2.739203410417794e-05, + "loss": 0.5034, + "step": 19664 + }, + { + "epoch": 1.319787926579645, + "grad_norm": 0.9669817090034485, + "learning_rate": 2.738234079165555e-05, + "loss": 0.5169, + "step": 19666 + }, + { + "epoch": 1.319922150263414, + "grad_norm": 1.0199397802352905, + "learning_rate": 2.737264854777306e-05, + "loss": 0.4526, + "step": 19668 + }, + { + "epoch": 1.3200563739471831, + "grad_norm": 0.8246620893478394, + "learning_rate": 2.7362957372988452e-05, + "loss": 0.4251, + "step": 19670 + }, + { + "epoch": 1.320190597630952, + "grad_norm": 1.0211527347564697, + "learning_rate": 2.7353267267759587e-05, + "loss": 0.5057, + "step": 19672 + }, + { + "epoch": 1.320324821314721, + "grad_norm": 1.0440993309020996, + "learning_rate": 2.734357823254432e-05, + "loss": 0.5367, + "step": 19674 + }, + { + "epoch": 1.3204590449984899, + "grad_norm": 1.2122056484222412, + "learning_rate": 2.7333890267800412e-05, + "loss": 0.5195, + "step": 19676 + }, + { + "epoch": 1.320593268682259, + "grad_norm": 1.0193716287612915, + "learning_rate": 2.7324203373985626e-05, + "loss": 0.5795, + "step": 19678 + }, + { + "epoch": 1.320727492366028, + "grad_norm": 0.9539496898651123, + "learning_rate": 2.7314517551557627e-05, + "loss": 0.4247, + "step": 19680 + }, + { + "epoch": 1.320861716049797, + "grad_norm": 1.0349043607711792, + "learning_rate": 2.7304832800974105e-05, + "loss": 0.5161, + "step": 19682 + }, + { + "epoch": 1.320995939733566, + "grad_norm": 0.9099202156066895, + "learning_rate": 2.7295149122692566e-05, + "loss": 0.4728, + "step": 19684 + }, + { + "epoch": 1.321130163417335, + "grad_norm": 0.9715734720230103, + "learning_rate": 2.7285466517170605e-05, + "loss": 0.5016, + "step": 19686 + }, + { + "epoch": 1.321264387101104, + "grad_norm": 1.1370961666107178, + "learning_rate": 2.727578498486566e-05, + "loss": 0.4813, + "step": 19688 + }, + { + "epoch": 1.321398610784873, + "grad_norm": 1.0140807628631592, + "learning_rate": 2.7266104526235215e-05, + "loss": 0.5288, + "step": 19690 + }, + { + "epoch": 1.3215328344686421, + "grad_norm": 1.0584005117416382, + "learning_rate": 2.725642514173662e-05, + "loss": 0.4736, + "step": 19692 + }, + { + "epoch": 1.321667058152411, + "grad_norm": 0.9137961268424988, + "learning_rate": 2.724674683182722e-05, + "loss": 0.4887, + "step": 19694 + }, + { + "epoch": 1.32180128183618, + "grad_norm": 1.1081452369689941, + "learning_rate": 2.7237069596964266e-05, + "loss": 0.4869, + "step": 19696 + }, + { + "epoch": 1.3219355055199489, + "grad_norm": 1.1327177286148071, + "learning_rate": 2.722739343760503e-05, + "loss": 0.5325, + "step": 19698 + }, + { + "epoch": 1.322069729203718, + "grad_norm": 0.9909605979919434, + "learning_rate": 2.721771835420668e-05, + "loss": 0.5747, + "step": 19700 + }, + { + "epoch": 1.322203952887487, + "grad_norm": 1.0612339973449707, + "learning_rate": 2.7208044347226335e-05, + "loss": 0.4779, + "step": 19702 + }, + { + "epoch": 1.322338176571256, + "grad_norm": 0.8930952548980713, + "learning_rate": 2.719837141712106e-05, + "loss": 0.4119, + "step": 19704 + }, + { + "epoch": 1.322472400255025, + "grad_norm": 0.9589643478393555, + "learning_rate": 2.718869956434791e-05, + "loss": 0.4433, + "step": 19706 + }, + { + "epoch": 1.322606623938794, + "grad_norm": 1.1247379779815674, + "learning_rate": 2.717902878936386e-05, + "loss": 0.5554, + "step": 19708 + }, + { + "epoch": 1.322740847622563, + "grad_norm": 1.1248613595962524, + "learning_rate": 2.7169359092625813e-05, + "loss": 0.5025, + "step": 19710 + }, + { + "epoch": 1.322875071306332, + "grad_norm": 1.054959774017334, + "learning_rate": 2.715969047459066e-05, + "loss": 0.5498, + "step": 19712 + }, + { + "epoch": 1.3230092949901011, + "grad_norm": 1.0124552249908447, + "learning_rate": 2.7150022935715196e-05, + "loss": 0.5236, + "step": 19714 + }, + { + "epoch": 1.32314351867387, + "grad_norm": 1.1516460180282593, + "learning_rate": 2.714035647645624e-05, + "loss": 0.5068, + "step": 19716 + }, + { + "epoch": 1.323277742357639, + "grad_norm": 0.8048339486122131, + "learning_rate": 2.7130691097270468e-05, + "loss": 0.4566, + "step": 19718 + }, + { + "epoch": 1.3234119660414079, + "grad_norm": 0.946758508682251, + "learning_rate": 2.7121026798614583e-05, + "loss": 0.4247, + "step": 19720 + }, + { + "epoch": 1.323546189725177, + "grad_norm": 1.184579849243164, + "learning_rate": 2.7111363580945202e-05, + "loss": 0.542, + "step": 19722 + }, + { + "epoch": 1.323680413408946, + "grad_norm": 1.1211533546447754, + "learning_rate": 2.710170144471888e-05, + "loss": 0.4784, + "step": 19724 + }, + { + "epoch": 1.323814637092715, + "grad_norm": 1.029905080795288, + "learning_rate": 2.7092040390392115e-05, + "loss": 0.5168, + "step": 19726 + }, + { + "epoch": 1.323948860776484, + "grad_norm": 0.9693290591239929, + "learning_rate": 2.7082380418421417e-05, + "loss": 0.5317, + "step": 19728 + }, + { + "epoch": 1.324083084460253, + "grad_norm": 1.0033167600631714, + "learning_rate": 2.7072721529263177e-05, + "loss": 0.4559, + "step": 19730 + }, + { + "epoch": 1.324217308144022, + "grad_norm": 1.0830321311950684, + "learning_rate": 2.706306372337376e-05, + "loss": 0.4772, + "step": 19732 + }, + { + "epoch": 1.324351531827791, + "grad_norm": 1.0198765993118286, + "learning_rate": 2.7053407001209465e-05, + "loss": 0.5029, + "step": 19734 + }, + { + "epoch": 1.3244857555115601, + "grad_norm": 1.114503264427185, + "learning_rate": 2.7043751363226575e-05, + "loss": 0.4737, + "step": 19736 + }, + { + "epoch": 1.324619979195329, + "grad_norm": 1.0865391492843628, + "learning_rate": 2.703409680988128e-05, + "loss": 0.4728, + "step": 19738 + }, + { + "epoch": 1.324754202879098, + "grad_norm": 1.020459771156311, + "learning_rate": 2.702444334162979e-05, + "loss": 0.4673, + "step": 19740 + }, + { + "epoch": 1.324888426562867, + "grad_norm": 0.9951459765434265, + "learning_rate": 2.7014790958928138e-05, + "loss": 0.4764, + "step": 19742 + }, + { + "epoch": 1.325022650246636, + "grad_norm": 0.8151239156723022, + "learning_rate": 2.7005139662232425e-05, + "loss": 0.4571, + "step": 19744 + }, + { + "epoch": 1.3251568739304052, + "grad_norm": 1.0674978494644165, + "learning_rate": 2.699548945199863e-05, + "loss": 0.5049, + "step": 19746 + }, + { + "epoch": 1.325291097614174, + "grad_norm": 1.6327846050262451, + "learning_rate": 2.6985840328682737e-05, + "loss": 0.5086, + "step": 19748 + }, + { + "epoch": 1.325425321297943, + "grad_norm": 1.055503487586975, + "learning_rate": 2.6976192292740637e-05, + "loss": 0.5274, + "step": 19750 + }, + { + "epoch": 1.325559544981712, + "grad_norm": 0.9699060916900635, + "learning_rate": 2.696654534462818e-05, + "loss": 0.5392, + "step": 19752 + }, + { + "epoch": 1.325693768665481, + "grad_norm": 0.9717655181884766, + "learning_rate": 2.6956899484801134e-05, + "loss": 0.5691, + "step": 19754 + }, + { + "epoch": 1.32582799234925, + "grad_norm": 0.9502808451652527, + "learning_rate": 2.6947254713715304e-05, + "loss": 0.4602, + "step": 19756 + }, + { + "epoch": 1.3259622160330191, + "grad_norm": 0.837290346622467, + "learning_rate": 2.693761103182635e-05, + "loss": 0.4719, + "step": 19758 + }, + { + "epoch": 1.326096439716788, + "grad_norm": 0.9364441633224487, + "learning_rate": 2.692796843958993e-05, + "loss": 0.4502, + "step": 19760 + }, + { + "epoch": 1.326230663400557, + "grad_norm": 1.1107641458511353, + "learning_rate": 2.691832693746161e-05, + "loss": 0.565, + "step": 19762 + }, + { + "epoch": 1.326364887084326, + "grad_norm": 1.2745716571807861, + "learning_rate": 2.6908686525896977e-05, + "loss": 0.4906, + "step": 19764 + }, + { + "epoch": 1.326499110768095, + "grad_norm": 1.05853271484375, + "learning_rate": 2.689904720535147e-05, + "loss": 0.5468, + "step": 19766 + }, + { + "epoch": 1.3266333344518642, + "grad_norm": 1.0703691244125366, + "learning_rate": 2.68894089762806e-05, + "loss": 0.4954, + "step": 19768 + }, + { + "epoch": 1.326767558135633, + "grad_norm": 0.9572019577026367, + "learning_rate": 2.6879771839139678e-05, + "loss": 0.4597, + "step": 19770 + }, + { + "epoch": 1.326901781819402, + "grad_norm": 1.1199136972427368, + "learning_rate": 2.6870135794384084e-05, + "loss": 0.6664, + "step": 19772 + }, + { + "epoch": 1.327036005503171, + "grad_norm": 1.0304343700408936, + "learning_rate": 2.686050084246907e-05, + "loss": 0.4733, + "step": 19774 + }, + { + "epoch": 1.32717022918694, + "grad_norm": 0.8788327574729919, + "learning_rate": 2.6850866983849915e-05, + "loss": 0.4735, + "step": 19776 + }, + { + "epoch": 1.327304452870709, + "grad_norm": 0.9124723672866821, + "learning_rate": 2.684123421898179e-05, + "loss": 0.5136, + "step": 19778 + }, + { + "epoch": 1.3274386765544781, + "grad_norm": 1.0493173599243164, + "learning_rate": 2.6831602548319773e-05, + "loss": 0.4791, + "step": 19780 + }, + { + "epoch": 1.327572900238247, + "grad_norm": 1.0095930099487305, + "learning_rate": 2.6821971972318992e-05, + "loss": 0.4779, + "step": 19782 + }, + { + "epoch": 1.327707123922016, + "grad_norm": 2.1133108139038086, + "learning_rate": 2.6812342491434444e-05, + "loss": 0.4133, + "step": 19784 + }, + { + "epoch": 1.327841347605785, + "grad_norm": 2.8399405479431152, + "learning_rate": 2.680271410612113e-05, + "loss": 0.5151, + "step": 19786 + }, + { + "epoch": 1.327975571289554, + "grad_norm": 0.8534837961196899, + "learning_rate": 2.6793086816833967e-05, + "loss": 0.4798, + "step": 19788 + }, + { + "epoch": 1.3281097949733232, + "grad_norm": 1.0487769842147827, + "learning_rate": 2.6783460624027813e-05, + "loss": 0.5263, + "step": 19790 + }, + { + "epoch": 1.328244018657092, + "grad_norm": 0.954577624797821, + "learning_rate": 2.6773835528157464e-05, + "loss": 0.4473, + "step": 19792 + }, + { + "epoch": 1.328378242340861, + "grad_norm": 1.010227084159851, + "learning_rate": 2.676421152967774e-05, + "loss": 0.5166, + "step": 19794 + }, + { + "epoch": 1.32851246602463, + "grad_norm": 1.0502123832702637, + "learning_rate": 2.675458862904333e-05, + "loss": 0.4875, + "step": 19796 + }, + { + "epoch": 1.328646689708399, + "grad_norm": 1.1167939901351929, + "learning_rate": 2.6744966826708906e-05, + "loss": 0.5347, + "step": 19798 + }, + { + "epoch": 1.328780913392168, + "grad_norm": 1.131696343421936, + "learning_rate": 2.673534612312904e-05, + "loss": 0.5007, + "step": 19800 + }, + { + "epoch": 1.3289151370759371, + "grad_norm": 0.9861233234405518, + "learning_rate": 2.6725726518758344e-05, + "loss": 0.5568, + "step": 19802 + }, + { + "epoch": 1.329049360759706, + "grad_norm": 0.9725451469421387, + "learning_rate": 2.6716108014051282e-05, + "loss": 0.5044, + "step": 19804 + }, + { + "epoch": 1.329183584443475, + "grad_norm": 0.9981563091278076, + "learning_rate": 2.670649060946237e-05, + "loss": 0.5231, + "step": 19806 + }, + { + "epoch": 1.329317808127244, + "grad_norm": 1.1433606147766113, + "learning_rate": 2.6696874305445936e-05, + "loss": 0.5069, + "step": 19808 + }, + { + "epoch": 1.329452031811013, + "grad_norm": 1.113306999206543, + "learning_rate": 2.6687259102456386e-05, + "loss": 0.4659, + "step": 19810 + }, + { + "epoch": 1.3295862554947822, + "grad_norm": 1.0019900798797607, + "learning_rate": 2.6677645000947982e-05, + "loss": 0.4865, + "step": 19812 + }, + { + "epoch": 1.329720479178551, + "grad_norm": 0.9664418697357178, + "learning_rate": 2.6668032001375002e-05, + "loss": 0.4992, + "step": 19814 + }, + { + "epoch": 1.32985470286232, + "grad_norm": 1.1979312896728516, + "learning_rate": 2.665842010419164e-05, + "loss": 0.4842, + "step": 19816 + }, + { + "epoch": 1.3299889265460891, + "grad_norm": 1.013890027999878, + "learning_rate": 2.6648809309852017e-05, + "loss": 0.5108, + "step": 19818 + }, + { + "epoch": 1.330123150229858, + "grad_norm": 0.8482784032821655, + "learning_rate": 2.6639199618810228e-05, + "loss": 0.4415, + "step": 19820 + }, + { + "epoch": 1.3302573739136272, + "grad_norm": 1.0452470779418945, + "learning_rate": 2.6629591031520334e-05, + "loss": 0.5919, + "step": 19822 + }, + { + "epoch": 1.3303915975973961, + "grad_norm": 1.0971609354019165, + "learning_rate": 2.6619983548436313e-05, + "loss": 0.4799, + "step": 19824 + }, + { + "epoch": 1.330525821281165, + "grad_norm": 1.0110511779785156, + "learning_rate": 2.661037717001209e-05, + "loss": 0.5498, + "step": 19826 + }, + { + "epoch": 1.330660044964934, + "grad_norm": 1.140190601348877, + "learning_rate": 2.660077189670153e-05, + "loss": 0.4567, + "step": 19828 + }, + { + "epoch": 1.330794268648703, + "grad_norm": 1.0503569841384888, + "learning_rate": 2.659116772895851e-05, + "loss": 0.5217, + "step": 19830 + }, + { + "epoch": 1.330928492332472, + "grad_norm": 1.0813215970993042, + "learning_rate": 2.6581564667236758e-05, + "loss": 0.4438, + "step": 19832 + }, + { + "epoch": 1.3310627160162412, + "grad_norm": 0.8762857913970947, + "learning_rate": 2.6571962711990073e-05, + "loss": 0.5069, + "step": 19834 + }, + { + "epoch": 1.33119693970001, + "grad_norm": 1.0669820308685303, + "learning_rate": 2.6562361863672037e-05, + "loss": 0.4461, + "step": 19836 + }, + { + "epoch": 1.331331163383779, + "grad_norm": 1.0361627340316772, + "learning_rate": 2.655276212273633e-05, + "loss": 0.489, + "step": 19838 + }, + { + "epoch": 1.3314653870675481, + "grad_norm": 1.078757405281067, + "learning_rate": 2.654316348963649e-05, + "loss": 0.4468, + "step": 19840 + }, + { + "epoch": 1.331599610751317, + "grad_norm": 1.0235754251480103, + "learning_rate": 2.653356596482607e-05, + "loss": 0.5024, + "step": 19842 + }, + { + "epoch": 1.3317338344350862, + "grad_norm": 1.0501896142959595, + "learning_rate": 2.652396954875851e-05, + "loss": 0.5146, + "step": 19844 + }, + { + "epoch": 1.3318680581188551, + "grad_norm": 0.8992530107498169, + "learning_rate": 2.651437424188723e-05, + "loss": 0.5311, + "step": 19846 + }, + { + "epoch": 1.332002281802624, + "grad_norm": 1.1598973274230957, + "learning_rate": 2.6504780044665578e-05, + "loss": 0.5134, + "step": 19848 + }, + { + "epoch": 1.332136505486393, + "grad_norm": 0.9470548629760742, + "learning_rate": 2.6495186957546852e-05, + "loss": 0.5206, + "step": 19850 + }, + { + "epoch": 1.332270729170162, + "grad_norm": 1.2048908472061157, + "learning_rate": 2.648559498098434e-05, + "loss": 0.5537, + "step": 19852 + }, + { + "epoch": 1.332404952853931, + "grad_norm": 0.9466072916984558, + "learning_rate": 2.647600411543123e-05, + "loss": 0.4738, + "step": 19854 + }, + { + "epoch": 1.3325391765377002, + "grad_norm": 1.1737912893295288, + "learning_rate": 2.6466414361340663e-05, + "loss": 0.5499, + "step": 19856 + }, + { + "epoch": 1.332673400221469, + "grad_norm": 1.0548076629638672, + "learning_rate": 2.6456825719165723e-05, + "loss": 0.5891, + "step": 19858 + }, + { + "epoch": 1.332807623905238, + "grad_norm": 0.9503657817840576, + "learning_rate": 2.644723818935949e-05, + "loss": 0.5022, + "step": 19860 + }, + { + "epoch": 1.3329418475890071, + "grad_norm": 1.0045098066329956, + "learning_rate": 2.6437651772374937e-05, + "loss": 0.4418, + "step": 19862 + }, + { + "epoch": 1.333076071272776, + "grad_norm": 0.9361532926559448, + "learning_rate": 2.6428066468665013e-05, + "loss": 0.4746, + "step": 19864 + }, + { + "epoch": 1.3332102949565452, + "grad_norm": 1.0541296005249023, + "learning_rate": 2.6418482278682566e-05, + "loss": 0.4461, + "step": 19866 + }, + { + "epoch": 1.3333445186403141, + "grad_norm": 0.9788280129432678, + "learning_rate": 2.640889920288049e-05, + "loss": 0.5003, + "step": 19868 + }, + { + "epoch": 1.333478742324083, + "grad_norm": 1.0545464754104614, + "learning_rate": 2.6399317241711507e-05, + "loss": 0.5072, + "step": 19870 + }, + { + "epoch": 1.333612966007852, + "grad_norm": 1.0186614990234375, + "learning_rate": 2.63897363956284e-05, + "loss": 0.4442, + "step": 19872 + }, + { + "epoch": 1.333747189691621, + "grad_norm": 1.0518426895141602, + "learning_rate": 2.638015666508382e-05, + "loss": 0.4598, + "step": 19874 + }, + { + "epoch": 1.33388141337539, + "grad_norm": 1.012442946434021, + "learning_rate": 2.6370578050530392e-05, + "loss": 0.544, + "step": 19876 + }, + { + "epoch": 1.3340156370591592, + "grad_norm": 1.0759896039962769, + "learning_rate": 2.636100055242067e-05, + "loss": 0.4876, + "step": 19878 + }, + { + "epoch": 1.334149860742928, + "grad_norm": 1.1002321243286133, + "learning_rate": 2.63514241712072e-05, + "loss": 0.5496, + "step": 19880 + }, + { + "epoch": 1.334284084426697, + "grad_norm": 1.0845890045166016, + "learning_rate": 2.6341848907342436e-05, + "loss": 0.5002, + "step": 19882 + }, + { + "epoch": 1.3344183081104661, + "grad_norm": 0.9589855074882507, + "learning_rate": 2.6332274761278797e-05, + "loss": 0.5202, + "step": 19884 + }, + { + "epoch": 1.334552531794235, + "grad_norm": 0.9928118586540222, + "learning_rate": 2.6322701733468608e-05, + "loss": 0.4507, + "step": 19886 + }, + { + "epoch": 1.3346867554780042, + "grad_norm": 0.9842401146888733, + "learning_rate": 2.6313129824364224e-05, + "loss": 0.555, + "step": 19888 + }, + { + "epoch": 1.3348209791617731, + "grad_norm": 1.1226009130477905, + "learning_rate": 2.630355903441788e-05, + "loss": 0.5192, + "step": 19890 + }, + { + "epoch": 1.334955202845542, + "grad_norm": 1.1191109418869019, + "learning_rate": 2.6293989364081774e-05, + "loss": 0.5041, + "step": 19892 + }, + { + "epoch": 1.3350894265293112, + "grad_norm": 0.9476625323295593, + "learning_rate": 2.628442081380803e-05, + "loss": 0.4581, + "step": 19894 + }, + { + "epoch": 1.33522365021308, + "grad_norm": 0.902819812297821, + "learning_rate": 2.6274853384048793e-05, + "loss": 0.4951, + "step": 19896 + }, + { + "epoch": 1.3353578738968492, + "grad_norm": 0.9704957604408264, + "learning_rate": 2.6265287075256057e-05, + "loss": 0.4322, + "step": 19898 + }, + { + "epoch": 1.3354920975806182, + "grad_norm": 1.0102123022079468, + "learning_rate": 2.6255721887881857e-05, + "loss": 0.4701, + "step": 19900 + }, + { + "epoch": 1.335626321264387, + "grad_norm": 0.9383575916290283, + "learning_rate": 2.6246157822378104e-05, + "loss": 0.4717, + "step": 19902 + }, + { + "epoch": 1.335760544948156, + "grad_norm": 1.066762924194336, + "learning_rate": 2.6236594879196685e-05, + "loss": 0.4914, + "step": 19904 + }, + { + "epoch": 1.3358947686319251, + "grad_norm": 0.9062620401382446, + "learning_rate": 2.6227033058789408e-05, + "loss": 0.5052, + "step": 19906 + }, + { + "epoch": 1.336028992315694, + "grad_norm": 1.233633041381836, + "learning_rate": 2.6217472361608094e-05, + "loss": 0.4843, + "step": 19908 + }, + { + "epoch": 1.3361632159994632, + "grad_norm": 1.0275330543518066, + "learning_rate": 2.6207912788104438e-05, + "loss": 0.5502, + "step": 19910 + }, + { + "epoch": 1.3362974396832321, + "grad_norm": 0.9860764145851135, + "learning_rate": 2.6198354338730123e-05, + "loss": 0.4641, + "step": 19912 + }, + { + "epoch": 1.336431663367001, + "grad_norm": 0.9409505128860474, + "learning_rate": 2.6188797013936762e-05, + "loss": 0.5787, + "step": 19914 + }, + { + "epoch": 1.3365658870507702, + "grad_norm": 1.118374228477478, + "learning_rate": 2.6179240814175897e-05, + "loss": 0.4895, + "step": 19916 + }, + { + "epoch": 1.336700110734539, + "grad_norm": 0.9939571022987366, + "learning_rate": 2.6169685739899085e-05, + "loss": 0.5625, + "step": 19918 + }, + { + "epoch": 1.3368343344183082, + "grad_norm": 0.8455119132995605, + "learning_rate": 2.6160131791557763e-05, + "loss": 0.4674, + "step": 19920 + }, + { + "epoch": 1.3369685581020772, + "grad_norm": 1.0985572338104248, + "learning_rate": 2.6150578969603334e-05, + "loss": 0.5249, + "step": 19922 + }, + { + "epoch": 1.337102781785846, + "grad_norm": 1.0769293308258057, + "learning_rate": 2.614102727448713e-05, + "loss": 0.5309, + "step": 19924 + }, + { + "epoch": 1.337237005469615, + "grad_norm": 1.0132557153701782, + "learning_rate": 2.6131476706660485e-05, + "loss": 0.4845, + "step": 19926 + }, + { + "epoch": 1.3373712291533841, + "grad_norm": 1.0289278030395508, + "learning_rate": 2.612192726657462e-05, + "loss": 0.4253, + "step": 19928 + }, + { + "epoch": 1.337505452837153, + "grad_norm": 1.099208950996399, + "learning_rate": 2.6112378954680773e-05, + "loss": 0.4906, + "step": 19930 + }, + { + "epoch": 1.3376396765209222, + "grad_norm": 0.9392862319946289, + "learning_rate": 2.6102831771430003e-05, + "loss": 0.4784, + "step": 19932 + }, + { + "epoch": 1.3377739002046911, + "grad_norm": 1.1693406105041504, + "learning_rate": 2.6093285717273465e-05, + "loss": 0.4967, + "step": 19934 + }, + { + "epoch": 1.33790812388846, + "grad_norm": 1.182614803314209, + "learning_rate": 2.6083740792662148e-05, + "loss": 0.5382, + "step": 19936 + }, + { + "epoch": 1.3380423475722292, + "grad_norm": 1.1358389854431152, + "learning_rate": 2.607419699804706e-05, + "loss": 0.587, + "step": 19938 + }, + { + "epoch": 1.338176571255998, + "grad_norm": 0.9437689185142517, + "learning_rate": 2.6064654333879123e-05, + "loss": 0.5281, + "step": 19940 + }, + { + "epoch": 1.3383107949397672, + "grad_norm": 1.1665679216384888, + "learning_rate": 2.6055112800609206e-05, + "loss": 0.5308, + "step": 19942 + }, + { + "epoch": 1.3384450186235362, + "grad_norm": 1.0980030298233032, + "learning_rate": 2.6045572398688095e-05, + "loss": 0.4539, + "step": 19944 + }, + { + "epoch": 1.338579242307305, + "grad_norm": 1.079601764678955, + "learning_rate": 2.6036033128566608e-05, + "loss": 0.5316, + "step": 19946 + }, + { + "epoch": 1.338713465991074, + "grad_norm": 1.1160887479782104, + "learning_rate": 2.6026494990695426e-05, + "loss": 0.5248, + "step": 19948 + }, + { + "epoch": 1.3388476896748431, + "grad_norm": 1.0072522163391113, + "learning_rate": 2.6016957985525227e-05, + "loss": 0.4889, + "step": 19950 + }, + { + "epoch": 1.338981913358612, + "grad_norm": 0.8797785639762878, + "learning_rate": 2.6007422113506573e-05, + "loss": 0.5538, + "step": 19952 + }, + { + "epoch": 1.3391161370423812, + "grad_norm": 0.938903272151947, + "learning_rate": 2.599788737509007e-05, + "loss": 0.5266, + "step": 19954 + }, + { + "epoch": 1.3392503607261501, + "grad_norm": 0.9888512492179871, + "learning_rate": 2.5988353770726166e-05, + "loss": 0.4562, + "step": 19956 + }, + { + "epoch": 1.339384584409919, + "grad_norm": 0.9281738996505737, + "learning_rate": 2.5978821300865368e-05, + "loss": 0.4834, + "step": 19958 + }, + { + "epoch": 1.3395188080936882, + "grad_norm": 0.9474917650222778, + "learning_rate": 2.5969289965957988e-05, + "loss": 0.382, + "step": 19960 + }, + { + "epoch": 1.339653031777457, + "grad_norm": 1.0445213317871094, + "learning_rate": 2.5959759766454428e-05, + "loss": 0.4968, + "step": 19962 + }, + { + "epoch": 1.3397872554612262, + "grad_norm": 1.082736611366272, + "learning_rate": 2.5950230702804923e-05, + "loss": 0.4946, + "step": 19964 + }, + { + "epoch": 1.3399214791449952, + "grad_norm": 1.0125893354415894, + "learning_rate": 2.5940702775459747e-05, + "loss": 0.4805, + "step": 19966 + }, + { + "epoch": 1.340055702828764, + "grad_norm": 0.8601479530334473, + "learning_rate": 2.593117598486905e-05, + "loss": 0.5045, + "step": 19968 + }, + { + "epoch": 1.3401899265125332, + "grad_norm": 0.9454371929168701, + "learning_rate": 2.5921650331482962e-05, + "loss": 0.4602, + "step": 19970 + }, + { + "epoch": 1.3403241501963021, + "grad_norm": 1.1329702138900757, + "learning_rate": 2.591212581575153e-05, + "loss": 0.4194, + "step": 19972 + }, + { + "epoch": 1.3404583738800713, + "grad_norm": 0.9190551042556763, + "learning_rate": 2.59026024381248e-05, + "loss": 0.491, + "step": 19974 + }, + { + "epoch": 1.3405925975638402, + "grad_norm": 0.9530711770057678, + "learning_rate": 2.589308019905273e-05, + "loss": 0.5468, + "step": 19976 + }, + { + "epoch": 1.3407268212476091, + "grad_norm": 1.036069393157959, + "learning_rate": 2.5883559098985204e-05, + "loss": 0.5485, + "step": 19978 + }, + { + "epoch": 1.340861044931378, + "grad_norm": 0.9575065970420837, + "learning_rate": 2.5874039138372075e-05, + "loss": 0.4476, + "step": 19980 + }, + { + "epoch": 1.3409952686151472, + "grad_norm": 0.9909329414367676, + "learning_rate": 2.586452031766317e-05, + "loss": 0.5144, + "step": 19982 + }, + { + "epoch": 1.341129492298916, + "grad_norm": 1.0757123231887817, + "learning_rate": 2.5855002637308224e-05, + "loss": 0.5254, + "step": 19984 + }, + { + "epoch": 1.3412637159826852, + "grad_norm": 0.9575532674789429, + "learning_rate": 2.584548609775692e-05, + "loss": 0.4818, + "step": 19986 + }, + { + "epoch": 1.3413979396664542, + "grad_norm": 0.9985975623130798, + "learning_rate": 2.5835970699458906e-05, + "loss": 0.4669, + "step": 19988 + }, + { + "epoch": 1.341532163350223, + "grad_norm": 0.9868609309196472, + "learning_rate": 2.582645644286374e-05, + "loss": 0.4987, + "step": 19990 + }, + { + "epoch": 1.3416663870339922, + "grad_norm": 0.9942282438278198, + "learning_rate": 2.5816943328420984e-05, + "loss": 0.5489, + "step": 19992 + }, + { + "epoch": 1.3418006107177611, + "grad_norm": 1.0065810680389404, + "learning_rate": 2.5807431356580092e-05, + "loss": 0.4594, + "step": 19994 + }, + { + "epoch": 1.3419348344015303, + "grad_norm": 1.054410457611084, + "learning_rate": 2.5797920527790526e-05, + "loss": 0.5011, + "step": 19996 + }, + { + "epoch": 1.3420690580852992, + "grad_norm": 0.9025766253471375, + "learning_rate": 2.5788410842501588e-05, + "loss": 0.5257, + "step": 19998 + }, + { + "epoch": 1.3422032817690681, + "grad_norm": 0.9478522539138794, + "learning_rate": 2.577890230116265e-05, + "loss": 0.4517, + "step": 20000 + }, + { + "epoch": 1.342337505452837, + "grad_norm": 0.9698958396911621, + "learning_rate": 2.5769394904222926e-05, + "loss": 0.4975, + "step": 20002 + }, + { + "epoch": 1.3424717291366062, + "grad_norm": 1.0644009113311768, + "learning_rate": 2.575988865213167e-05, + "loss": 0.4529, + "step": 20004 + }, + { + "epoch": 1.342605952820375, + "grad_norm": 1.110205888748169, + "learning_rate": 2.5750383545338002e-05, + "loss": 0.5225, + "step": 20006 + }, + { + "epoch": 1.3427401765041442, + "grad_norm": 1.094616174697876, + "learning_rate": 2.5740879584291034e-05, + "loss": 0.4981, + "step": 20008 + }, + { + "epoch": 1.3428744001879132, + "grad_norm": 0.9325581789016724, + "learning_rate": 2.5731376769439776e-05, + "loss": 0.5181, + "step": 20010 + }, + { + "epoch": 1.343008623871682, + "grad_norm": 1.0330641269683838, + "learning_rate": 2.572187510123326e-05, + "loss": 0.499, + "step": 20012 + }, + { + "epoch": 1.3431428475554512, + "grad_norm": 1.054914951324463, + "learning_rate": 2.5712374580120414e-05, + "loss": 0.5062, + "step": 20014 + }, + { + "epoch": 1.3432770712392201, + "grad_norm": 0.9782353043556213, + "learning_rate": 2.5702875206550105e-05, + "loss": 0.5673, + "step": 20016 + }, + { + "epoch": 1.3434112949229893, + "grad_norm": 1.109312891960144, + "learning_rate": 2.5693376980971145e-05, + "loss": 0.4858, + "step": 20018 + }, + { + "epoch": 1.3435455186067582, + "grad_norm": 1.1009002923965454, + "learning_rate": 2.5683879903832347e-05, + "loss": 0.4806, + "step": 20020 + }, + { + "epoch": 1.3436797422905271, + "grad_norm": 1.0390616655349731, + "learning_rate": 2.567438397558239e-05, + "loss": 0.5129, + "step": 20022 + }, + { + "epoch": 1.343813965974296, + "grad_norm": 0.9547844529151917, + "learning_rate": 2.5664889196669973e-05, + "loss": 0.4629, + "step": 20024 + }, + { + "epoch": 1.3439481896580652, + "grad_norm": 1.1426942348480225, + "learning_rate": 2.5655395567543684e-05, + "loss": 0.5165, + "step": 20026 + }, + { + "epoch": 1.344082413341834, + "grad_norm": 1.0668246746063232, + "learning_rate": 2.564590308865209e-05, + "loss": 0.4642, + "step": 20028 + }, + { + "epoch": 1.3442166370256032, + "grad_norm": 1.1261327266693115, + "learning_rate": 2.5636411760443657e-05, + "loss": 0.5446, + "step": 20030 + }, + { + "epoch": 1.3443508607093722, + "grad_norm": 1.0175567865371704, + "learning_rate": 2.5626921583366886e-05, + "loss": 0.476, + "step": 20032 + }, + { + "epoch": 1.344485084393141, + "grad_norm": 0.9739302396774292, + "learning_rate": 2.561743255787014e-05, + "loss": 0.5262, + "step": 20034 + }, + { + "epoch": 1.3446193080769102, + "grad_norm": 0.9953427314758301, + "learning_rate": 2.5607944684401764e-05, + "loss": 0.4766, + "step": 20036 + }, + { + "epoch": 1.3447535317606791, + "grad_norm": 0.981567919254303, + "learning_rate": 2.559845796341001e-05, + "loss": 0.4691, + "step": 20038 + }, + { + "epoch": 1.3448877554444483, + "grad_norm": 1.0105043649673462, + "learning_rate": 2.5588972395343157e-05, + "loss": 0.4665, + "step": 20040 + }, + { + "epoch": 1.3450219791282172, + "grad_norm": 1.0410128831863403, + "learning_rate": 2.557948798064935e-05, + "loss": 0.5859, + "step": 20042 + }, + { + "epoch": 1.3451562028119861, + "grad_norm": 1.1123363971710205, + "learning_rate": 2.557000471977673e-05, + "loss": 0.4977, + "step": 20044 + }, + { + "epoch": 1.3452904264957553, + "grad_norm": 1.0192078351974487, + "learning_rate": 2.5560522613173314e-05, + "loss": 0.446, + "step": 20046 + }, + { + "epoch": 1.3454246501795242, + "grad_norm": 1.1729594469070435, + "learning_rate": 2.555104166128717e-05, + "loss": 0.476, + "step": 20048 + }, + { + "epoch": 1.3455588738632933, + "grad_norm": 1.0416300296783447, + "learning_rate": 2.554156186456621e-05, + "loss": 0.5405, + "step": 20050 + }, + { + "epoch": 1.3456930975470622, + "grad_norm": 1.0445678234100342, + "learning_rate": 2.553208322345838e-05, + "loss": 0.4817, + "step": 20052 + }, + { + "epoch": 1.3458273212308312, + "grad_norm": 1.0217093229293823, + "learning_rate": 2.552260573841152e-05, + "loss": 0.5277, + "step": 20054 + }, + { + "epoch": 1.3459615449146, + "grad_norm": 1.1795514822006226, + "learning_rate": 2.5513129409873372e-05, + "loss": 0.4549, + "step": 20056 + }, + { + "epoch": 1.3460957685983692, + "grad_norm": 0.9328517317771912, + "learning_rate": 2.550365423829172e-05, + "loss": 0.4461, + "step": 20058 + }, + { + "epoch": 1.3462299922821381, + "grad_norm": 1.020693302154541, + "learning_rate": 2.5494180224114217e-05, + "loss": 0.5468, + "step": 20060 + }, + { + "epoch": 1.3463642159659073, + "grad_norm": 0.8491286039352417, + "learning_rate": 2.548470736778853e-05, + "loss": 0.4652, + "step": 20062 + }, + { + "epoch": 1.3464984396496762, + "grad_norm": 0.8926975727081299, + "learning_rate": 2.547523566976222e-05, + "loss": 0.513, + "step": 20064 + }, + { + "epoch": 1.3466326633334451, + "grad_norm": 1.0468785762786865, + "learning_rate": 2.546576513048279e-05, + "loss": 0.5602, + "step": 20066 + }, + { + "epoch": 1.3467668870172143, + "grad_norm": 1.0157579183578491, + "learning_rate": 2.5456295750397698e-05, + "loss": 0.4753, + "step": 20068 + }, + { + "epoch": 1.3469011107009832, + "grad_norm": 0.9623618721961975, + "learning_rate": 2.5446827529954385e-05, + "loss": 0.5044, + "step": 20070 + }, + { + "epoch": 1.3470353343847523, + "grad_norm": 1.1392110586166382, + "learning_rate": 2.543736046960019e-05, + "loss": 0.4971, + "step": 20072 + }, + { + "epoch": 1.3471695580685212, + "grad_norm": 1.296471357345581, + "learning_rate": 2.5427894569782412e-05, + "loss": 0.4338, + "step": 20074 + }, + { + "epoch": 1.3473037817522902, + "grad_norm": 1.1017106771469116, + "learning_rate": 2.5418429830948276e-05, + "loss": 0.4988, + "step": 20076 + }, + { + "epoch": 1.347438005436059, + "grad_norm": 1.0007191896438599, + "learning_rate": 2.540896625354502e-05, + "loss": 0.5375, + "step": 20078 + }, + { + "epoch": 1.3475722291198282, + "grad_norm": 1.0098400115966797, + "learning_rate": 2.5399503838019722e-05, + "loss": 0.4733, + "step": 20080 + }, + { + "epoch": 1.3477064528035971, + "grad_norm": 0.9135158061981201, + "learning_rate": 2.539004258481954e-05, + "loss": 0.4741, + "step": 20082 + }, + { + "epoch": 1.3478406764873663, + "grad_norm": 0.9549738764762878, + "learning_rate": 2.5380582494391413e-05, + "loss": 0.5202, + "step": 20084 + }, + { + "epoch": 1.3479749001711352, + "grad_norm": 1.1197456121444702, + "learning_rate": 2.5371123567182363e-05, + "loss": 0.4726, + "step": 20086 + }, + { + "epoch": 1.3481091238549041, + "grad_norm": 0.9651446342468262, + "learning_rate": 2.5361665803639282e-05, + "loss": 0.4997, + "step": 20088 + }, + { + "epoch": 1.3482433475386733, + "grad_norm": 1.2065728902816772, + "learning_rate": 2.5352209204209065e-05, + "loss": 0.5219, + "step": 20090 + }, + { + "epoch": 1.3483775712224422, + "grad_norm": 1.1290955543518066, + "learning_rate": 2.534275376933849e-05, + "loss": 0.5273, + "step": 20092 + }, + { + "epoch": 1.3485117949062113, + "grad_norm": 1.0234849452972412, + "learning_rate": 2.533329949947431e-05, + "loss": 0.4433, + "step": 20094 + }, + { + "epoch": 1.3486460185899802, + "grad_norm": 0.9947075247764587, + "learning_rate": 2.532384639506321e-05, + "loss": 0.4801, + "step": 20096 + }, + { + "epoch": 1.3487802422737492, + "grad_norm": 1.044743299484253, + "learning_rate": 2.531439445655186e-05, + "loss": 0.4686, + "step": 20098 + }, + { + "epoch": 1.348914465957518, + "grad_norm": 1.1770745515823364, + "learning_rate": 2.530494368438683e-05, + "loss": 0.4632, + "step": 20100 + }, + { + "epoch": 1.3490486896412872, + "grad_norm": 1.1431256532669067, + "learning_rate": 2.5295494079014647e-05, + "loss": 0.5243, + "step": 20102 + }, + { + "epoch": 1.3491829133250561, + "grad_norm": 1.1603831052780151, + "learning_rate": 2.528604564088178e-05, + "loss": 0.5558, + "step": 20104 + }, + { + "epoch": 1.3493171370088253, + "grad_norm": 1.0677740573883057, + "learning_rate": 2.5276598370434666e-05, + "loss": 0.4388, + "step": 20106 + }, + { + "epoch": 1.3494513606925942, + "grad_norm": 0.952692449092865, + "learning_rate": 2.5267152268119653e-05, + "loss": 0.5032, + "step": 20108 + }, + { + "epoch": 1.3495855843763631, + "grad_norm": 1.0014994144439697, + "learning_rate": 2.5257707334383095e-05, + "loss": 0.531, + "step": 20110 + }, + { + "epoch": 1.3497198080601323, + "grad_norm": 0.9802749156951904, + "learning_rate": 2.524826356967117e-05, + "loss": 0.4783, + "step": 20112 + }, + { + "epoch": 1.3498540317439012, + "grad_norm": 1.054930329322815, + "learning_rate": 2.5238820974430143e-05, + "loss": 0.4785, + "step": 20114 + }, + { + "epoch": 1.3499882554276703, + "grad_norm": 1.3713618516921997, + "learning_rate": 2.5229379549106115e-05, + "loss": 0.5262, + "step": 20116 + }, + { + "epoch": 1.3501224791114392, + "grad_norm": 1.056983232498169, + "learning_rate": 2.521993929414521e-05, + "loss": 0.5007, + "step": 20118 + }, + { + "epoch": 1.3502567027952082, + "grad_norm": 0.9778099060058594, + "learning_rate": 2.521050020999347e-05, + "loss": 0.4966, + "step": 20120 + }, + { + "epoch": 1.3503909264789773, + "grad_norm": 1.049523115158081, + "learning_rate": 2.5201062297096807e-05, + "loss": 0.5833, + "step": 20122 + }, + { + "epoch": 1.3505251501627462, + "grad_norm": 0.9236955046653748, + "learning_rate": 2.51916255559012e-05, + "loss": 0.5228, + "step": 20124 + }, + { + "epoch": 1.3506593738465154, + "grad_norm": 1.0097215175628662, + "learning_rate": 2.518218998685249e-05, + "loss": 0.5528, + "step": 20126 + }, + { + "epoch": 1.3507935975302843, + "grad_norm": 0.9391131401062012, + "learning_rate": 2.5172755590396517e-05, + "loss": 0.4982, + "step": 20128 + }, + { + "epoch": 1.3509278212140532, + "grad_norm": 0.8988783955574036, + "learning_rate": 2.5163322366979024e-05, + "loss": 0.5132, + "step": 20130 + }, + { + "epoch": 1.3510620448978221, + "grad_norm": 0.9671209454536438, + "learning_rate": 2.5153890317045702e-05, + "loss": 0.4777, + "step": 20132 + }, + { + "epoch": 1.3511962685815913, + "grad_norm": 1.1700220108032227, + "learning_rate": 2.5144459441042196e-05, + "loss": 0.5518, + "step": 20134 + }, + { + "epoch": 1.3513304922653602, + "grad_norm": 1.0199360847473145, + "learning_rate": 2.5135029739414116e-05, + "loss": 0.4867, + "step": 20136 + }, + { + "epoch": 1.3514647159491293, + "grad_norm": 0.9349581003189087, + "learning_rate": 2.5125601212606992e-05, + "loss": 0.4588, + "step": 20138 + }, + { + "epoch": 1.3515989396328982, + "grad_norm": 1.1096220016479492, + "learning_rate": 2.5116173861066296e-05, + "loss": 0.458, + "step": 20140 + }, + { + "epoch": 1.3517331633166672, + "grad_norm": 0.8157988786697388, + "learning_rate": 2.510674768523743e-05, + "loss": 0.4541, + "step": 20142 + }, + { + "epoch": 1.3518673870004363, + "grad_norm": 1.1094350814819336, + "learning_rate": 2.509732268556581e-05, + "loss": 0.4659, + "step": 20144 + }, + { + "epoch": 1.3520016106842052, + "grad_norm": 1.0682333707809448, + "learning_rate": 2.5087898862496705e-05, + "loss": 0.4778, + "step": 20146 + }, + { + "epoch": 1.3521358343679744, + "grad_norm": 0.9747878313064575, + "learning_rate": 2.5078476216475422e-05, + "loss": 0.4584, + "step": 20148 + }, + { + "epoch": 1.3522700580517433, + "grad_norm": 1.086549162864685, + "learning_rate": 2.5069054747947102e-05, + "loss": 0.5574, + "step": 20150 + }, + { + "epoch": 1.3524042817355122, + "grad_norm": 1.0562546253204346, + "learning_rate": 2.5059634457356933e-05, + "loss": 0.4842, + "step": 20152 + }, + { + "epoch": 1.3525385054192811, + "grad_norm": 1.0438446998596191, + "learning_rate": 2.5050215345149975e-05, + "loss": 0.5118, + "step": 20154 + }, + { + "epoch": 1.3526727291030503, + "grad_norm": 1.0002363920211792, + "learning_rate": 2.5040797411771306e-05, + "loss": 0.4894, + "step": 20156 + }, + { + "epoch": 1.3528069527868192, + "grad_norm": 1.0864993333816528, + "learning_rate": 2.503138065766587e-05, + "loss": 0.476, + "step": 20158 + }, + { + "epoch": 1.3529411764705883, + "grad_norm": 1.0022969245910645, + "learning_rate": 2.502196508327861e-05, + "loss": 0.4776, + "step": 20160 + }, + { + "epoch": 1.3530754001543572, + "grad_norm": 1.067216396331787, + "learning_rate": 2.5012550689054355e-05, + "loss": 0.4521, + "step": 20162 + }, + { + "epoch": 1.3532096238381262, + "grad_norm": 0.8796282410621643, + "learning_rate": 2.5003137475437966e-05, + "loss": 0.454, + "step": 20164 + }, + { + "epoch": 1.3533438475218953, + "grad_norm": 1.033448576927185, + "learning_rate": 2.499372544287417e-05, + "loss": 0.5156, + "step": 20166 + }, + { + "epoch": 1.3534780712056642, + "grad_norm": 1.0945485830307007, + "learning_rate": 2.4984314591807682e-05, + "loss": 0.4867, + "step": 20168 + }, + { + "epoch": 1.3536122948894334, + "grad_norm": 1.0850149393081665, + "learning_rate": 2.4974904922683118e-05, + "loss": 0.5128, + "step": 20170 + }, + { + "epoch": 1.3537465185732023, + "grad_norm": 1.0059179067611694, + "learning_rate": 2.4965496435945106e-05, + "loss": 0.462, + "step": 20172 + }, + { + "epoch": 1.3538807422569712, + "grad_norm": 1.287937045097351, + "learning_rate": 2.495608913203814e-05, + "loss": 0.4408, + "step": 20174 + }, + { + "epoch": 1.3540149659407401, + "grad_norm": 1.053192377090454, + "learning_rate": 2.4946683011406752e-05, + "loss": 0.5038, + "step": 20176 + }, + { + "epoch": 1.3541491896245093, + "grad_norm": 1.0209565162658691, + "learning_rate": 2.493727807449529e-05, + "loss": 0.4562, + "step": 20178 + }, + { + "epoch": 1.3542834133082782, + "grad_norm": 1.0276302099227905, + "learning_rate": 2.4927874321748173e-05, + "loss": 0.4979, + "step": 20180 + }, + { + "epoch": 1.3544176369920473, + "grad_norm": 0.7771276831626892, + "learning_rate": 2.4918471753609675e-05, + "loss": 0.4448, + "step": 20182 + }, + { + "epoch": 1.3545518606758162, + "grad_norm": 0.8976522088050842, + "learning_rate": 2.4909070370524083e-05, + "loss": 0.4867, + "step": 20184 + }, + { + "epoch": 1.3546860843595852, + "grad_norm": 0.9804712533950806, + "learning_rate": 2.489967017293558e-05, + "loss": 0.5131, + "step": 20186 + }, + { + "epoch": 1.3548203080433543, + "grad_norm": 1.0005247592926025, + "learning_rate": 2.4890271161288302e-05, + "loss": 0.4038, + "step": 20188 + }, + { + "epoch": 1.3549545317271232, + "grad_norm": 1.1950335502624512, + "learning_rate": 2.4880873336026332e-05, + "loss": 0.5083, + "step": 20190 + }, + { + "epoch": 1.3550887554108924, + "grad_norm": 1.0356762409210205, + "learning_rate": 2.4871476697593687e-05, + "loss": 0.4998, + "step": 20192 + }, + { + "epoch": 1.3552229790946613, + "grad_norm": 1.0134941339492798, + "learning_rate": 2.4862081246434378e-05, + "loss": 0.5141, + "step": 20194 + }, + { + "epoch": 1.3553572027784302, + "grad_norm": 1.083375096321106, + "learning_rate": 2.485268698299229e-05, + "loss": 0.5193, + "step": 20196 + }, + { + "epoch": 1.3554914264621993, + "grad_norm": 0.9247611165046692, + "learning_rate": 2.4843293907711295e-05, + "loss": 0.4712, + "step": 20198 + }, + { + "epoch": 1.3556256501459683, + "grad_norm": 1.007331371307373, + "learning_rate": 2.4833902021035177e-05, + "loss": 0.4537, + "step": 20200 + }, + { + "epoch": 1.3557598738297374, + "grad_norm": 0.9057896733283997, + "learning_rate": 2.482451132340772e-05, + "loss": 0.48, + "step": 20202 + }, + { + "epoch": 1.3558940975135063, + "grad_norm": 0.9654235243797302, + "learning_rate": 2.4815121815272573e-05, + "loss": 0.5029, + "step": 20204 + }, + { + "epoch": 1.3560283211972752, + "grad_norm": 0.9667559862136841, + "learning_rate": 2.480573349707344e-05, + "loss": 0.463, + "step": 20206 + }, + { + "epoch": 1.3561625448810442, + "grad_norm": 1.0404314994812012, + "learning_rate": 2.479634636925382e-05, + "loss": 0.4783, + "step": 20208 + }, + { + "epoch": 1.3562967685648133, + "grad_norm": 1.0094547271728516, + "learning_rate": 2.478696043225729e-05, + "loss": 0.4635, + "step": 20210 + }, + { + "epoch": 1.3564309922485822, + "grad_norm": 1.0778117179870605, + "learning_rate": 2.477757568652728e-05, + "loss": 0.4946, + "step": 20212 + }, + { + "epoch": 1.3565652159323514, + "grad_norm": 0.9366231560707092, + "learning_rate": 2.4768192132507245e-05, + "loss": 0.5015, + "step": 20214 + }, + { + "epoch": 1.3566994396161203, + "grad_norm": 1.0686755180358887, + "learning_rate": 2.475880977064051e-05, + "loss": 0.473, + "step": 20216 + }, + { + "epoch": 1.3568336632998892, + "grad_norm": 0.9536647200584412, + "learning_rate": 2.4749428601370384e-05, + "loss": 0.4982, + "step": 20218 + }, + { + "epoch": 1.3569678869836583, + "grad_norm": 0.9279668927192688, + "learning_rate": 2.4740048625140078e-05, + "loss": 0.4641, + "step": 20220 + }, + { + "epoch": 1.3571021106674273, + "grad_norm": 1.2315804958343506, + "learning_rate": 2.473066984239282e-05, + "loss": 0.4984, + "step": 20222 + }, + { + "epoch": 1.3572363343511964, + "grad_norm": 0.9835600256919861, + "learning_rate": 2.472129225357173e-05, + "loss": 0.4598, + "step": 20224 + }, + { + "epoch": 1.3573705580349653, + "grad_norm": 0.9124851226806641, + "learning_rate": 2.471191585911987e-05, + "loss": 0.494, + "step": 20226 + }, + { + "epoch": 1.3575047817187342, + "grad_norm": 1.033633828163147, + "learning_rate": 2.4702540659480238e-05, + "loss": 0.472, + "step": 20228 + }, + { + "epoch": 1.3576390054025032, + "grad_norm": 0.9878078699111938, + "learning_rate": 2.469316665509584e-05, + "loss": 0.4974, + "step": 20230 + }, + { + "epoch": 1.3577732290862723, + "grad_norm": 0.9498612284660339, + "learning_rate": 2.4683793846409524e-05, + "loss": 0.4954, + "step": 20232 + }, + { + "epoch": 1.3579074527700412, + "grad_norm": 1.0005344152450562, + "learning_rate": 2.4674422233864213e-05, + "loss": 0.4534, + "step": 20234 + }, + { + "epoch": 1.3580416764538104, + "grad_norm": 1.2731600999832153, + "learning_rate": 2.466505181790261e-05, + "loss": 0.4823, + "step": 20236 + }, + { + "epoch": 1.3581759001375793, + "grad_norm": 0.9503934979438782, + "learning_rate": 2.465568259896751e-05, + "loss": 0.5024, + "step": 20238 + }, + { + "epoch": 1.3583101238213482, + "grad_norm": 0.9364110827445984, + "learning_rate": 2.464631457750155e-05, + "loss": 0.4845, + "step": 20240 + }, + { + "epoch": 1.3584443475051173, + "grad_norm": 1.1178056001663208, + "learning_rate": 2.4636947753947387e-05, + "loss": 0.4674, + "step": 20242 + }, + { + "epoch": 1.3585785711888863, + "grad_norm": 1.00879967212677, + "learning_rate": 2.4627582128747567e-05, + "loss": 0.4675, + "step": 20244 + }, + { + "epoch": 1.3587127948726554, + "grad_norm": 0.9615558981895447, + "learning_rate": 2.4618217702344603e-05, + "loss": 0.481, + "step": 20246 + }, + { + "epoch": 1.3588470185564243, + "grad_norm": 1.1916038990020752, + "learning_rate": 2.460885447518092e-05, + "loss": 0.5446, + "step": 20248 + }, + { + "epoch": 1.3589812422401932, + "grad_norm": 1.065394401550293, + "learning_rate": 2.4599492447698958e-05, + "loss": 0.5024, + "step": 20250 + }, + { + "epoch": 1.3591154659239622, + "grad_norm": 1.1207654476165771, + "learning_rate": 2.4590131620341023e-05, + "loss": 0.4775, + "step": 20252 + }, + { + "epoch": 1.3592496896077313, + "grad_norm": 0.9586560726165771, + "learning_rate": 2.4580771993549407e-05, + "loss": 0.4822, + "step": 20254 + }, + { + "epoch": 1.3593839132915002, + "grad_norm": 0.9638738036155701, + "learning_rate": 2.4571413567766305e-05, + "loss": 0.4469, + "step": 20256 + }, + { + "epoch": 1.3595181369752694, + "grad_norm": 1.106136679649353, + "learning_rate": 2.4562056343433933e-05, + "loss": 0.5251, + "step": 20258 + }, + { + "epoch": 1.3596523606590383, + "grad_norm": 1.168971061706543, + "learning_rate": 2.4552700320994377e-05, + "loss": 0.4528, + "step": 20260 + }, + { + "epoch": 1.3597865843428072, + "grad_norm": 1.069299578666687, + "learning_rate": 2.4543345500889687e-05, + "loss": 0.5055, + "step": 20262 + }, + { + "epoch": 1.3599208080265763, + "grad_norm": 1.0085757970809937, + "learning_rate": 2.4533991883561868e-05, + "loss": 0.4194, + "step": 20264 + }, + { + "epoch": 1.3600550317103453, + "grad_norm": 0.9998415112495422, + "learning_rate": 2.4524639469452838e-05, + "loss": 0.4773, + "step": 20266 + }, + { + "epoch": 1.3601892553941144, + "grad_norm": 1.1032497882843018, + "learning_rate": 2.451528825900451e-05, + "loss": 0.5662, + "step": 20268 + }, + { + "epoch": 1.3603234790778833, + "grad_norm": 0.9551237225532532, + "learning_rate": 2.4505938252658682e-05, + "loss": 0.5417, + "step": 20270 + }, + { + "epoch": 1.3604577027616522, + "grad_norm": 1.0504215955734253, + "learning_rate": 2.449658945085718e-05, + "loss": 0.4674, + "step": 20272 + }, + { + "epoch": 1.3605919264454214, + "grad_norm": 1.1610937118530273, + "learning_rate": 2.4487241854041637e-05, + "loss": 0.5049, + "step": 20274 + }, + { + "epoch": 1.3607261501291903, + "grad_norm": 0.9728044271469116, + "learning_rate": 2.4477895462653764e-05, + "loss": 0.4661, + "step": 20276 + }, + { + "epoch": 1.3608603738129594, + "grad_norm": 0.8766085505485535, + "learning_rate": 2.4468550277135128e-05, + "loss": 0.4712, + "step": 20278 + }, + { + "epoch": 1.3609945974967284, + "grad_norm": 1.001575231552124, + "learning_rate": 2.4459206297927306e-05, + "loss": 0.4446, + "step": 20280 + }, + { + "epoch": 1.3611288211804973, + "grad_norm": 1.0977261066436768, + "learning_rate": 2.4449863525471763e-05, + "loss": 0.4841, + "step": 20282 + }, + { + "epoch": 1.3612630448642662, + "grad_norm": 1.1366521120071411, + "learning_rate": 2.4440521960209924e-05, + "loss": 0.5129, + "step": 20284 + }, + { + "epoch": 1.3613972685480353, + "grad_norm": 1.0719273090362549, + "learning_rate": 2.4431181602583137e-05, + "loss": 0.5287, + "step": 20286 + }, + { + "epoch": 1.3615314922318043, + "grad_norm": 0.9917060732841492, + "learning_rate": 2.442184245303277e-05, + "loss": 0.4884, + "step": 20288 + }, + { + "epoch": 1.3616657159155734, + "grad_norm": 1.0406906604766846, + "learning_rate": 2.441250451200004e-05, + "loss": 0.4561, + "step": 20290 + }, + { + "epoch": 1.3617999395993423, + "grad_norm": 0.9567008018493652, + "learning_rate": 2.4403167779926162e-05, + "loss": 0.5076, + "step": 20292 + }, + { + "epoch": 1.3619341632831112, + "grad_norm": 0.8407172560691833, + "learning_rate": 2.4393832257252252e-05, + "loss": 0.4622, + "step": 20294 + }, + { + "epoch": 1.3620683869668804, + "grad_norm": 0.9681487679481506, + "learning_rate": 2.438449794441943e-05, + "loss": 0.4714, + "step": 20296 + }, + { + "epoch": 1.3622026106506493, + "grad_norm": 0.9833024144172668, + "learning_rate": 2.4375164841868697e-05, + "loss": 0.4349, + "step": 20298 + }, + { + "epoch": 1.3623368343344184, + "grad_norm": 1.0088200569152832, + "learning_rate": 2.4365832950041072e-05, + "loss": 0.443, + "step": 20300 + }, + { + "epoch": 1.3624710580181874, + "grad_norm": 1.1553938388824463, + "learning_rate": 2.4356502269377397e-05, + "loss": 0.5802, + "step": 20302 + }, + { + "epoch": 1.3626052817019563, + "grad_norm": 0.9181938171386719, + "learning_rate": 2.434717280031858e-05, + "loss": 0.4581, + "step": 20304 + }, + { + "epoch": 1.3627395053857252, + "grad_norm": 1.0648936033248901, + "learning_rate": 2.4337844543305387e-05, + "loss": 0.5063, + "step": 20306 + }, + { + "epoch": 1.3628737290694943, + "grad_norm": 1.046919822692871, + "learning_rate": 2.4328517498778596e-05, + "loss": 0.5244, + "step": 20308 + }, + { + "epoch": 1.3630079527532633, + "grad_norm": 1.006034016609192, + "learning_rate": 2.431919166717887e-05, + "loss": 0.4525, + "step": 20310 + }, + { + "epoch": 1.3631421764370324, + "grad_norm": 0.9807382225990295, + "learning_rate": 2.4309867048946838e-05, + "loss": 0.4511, + "step": 20312 + }, + { + "epoch": 1.3632764001208013, + "grad_norm": 1.055794596672058, + "learning_rate": 2.4300543644523056e-05, + "loss": 0.4722, + "step": 20314 + }, + { + "epoch": 1.3634106238045702, + "grad_norm": 1.0888453722000122, + "learning_rate": 2.429122145434807e-05, + "loss": 0.4709, + "step": 20316 + }, + { + "epoch": 1.3635448474883394, + "grad_norm": 1.0583890676498413, + "learning_rate": 2.4281900478862306e-05, + "loss": 0.467, + "step": 20318 + }, + { + "epoch": 1.3636790711721083, + "grad_norm": 1.1082299947738647, + "learning_rate": 2.427258071850618e-05, + "loss": 0.533, + "step": 20320 + }, + { + "epoch": 1.3638132948558774, + "grad_norm": 0.9727580547332764, + "learning_rate": 2.4263262173720008e-05, + "loss": 0.5385, + "step": 20322 + }, + { + "epoch": 1.3639475185396464, + "grad_norm": 1.0671035051345825, + "learning_rate": 2.4253944844944105e-05, + "loss": 0.485, + "step": 20324 + }, + { + "epoch": 1.3640817422234153, + "grad_norm": 0.9661836624145508, + "learning_rate": 2.424462873261868e-05, + "loss": 0.4603, + "step": 20326 + }, + { + "epoch": 1.3642159659071842, + "grad_norm": 0.9973102807998657, + "learning_rate": 2.42353138371839e-05, + "loss": 0.5054, + "step": 20328 + }, + { + "epoch": 1.3643501895909533, + "grad_norm": 1.0182770490646362, + "learning_rate": 2.422600015907988e-05, + "loss": 0.4084, + "step": 20330 + }, + { + "epoch": 1.3644844132747223, + "grad_norm": 0.9462610483169556, + "learning_rate": 2.421668769874665e-05, + "loss": 0.4351, + "step": 20332 + }, + { + "epoch": 1.3646186369584914, + "grad_norm": 1.179429292678833, + "learning_rate": 2.4207376456624243e-05, + "loss": 0.5364, + "step": 20334 + }, + { + "epoch": 1.3647528606422603, + "grad_norm": 1.2717273235321045, + "learning_rate": 2.419806643315256e-05, + "loss": 0.5074, + "step": 20336 + }, + { + "epoch": 1.3648870843260292, + "grad_norm": 1.2078745365142822, + "learning_rate": 2.4188757628771524e-05, + "loss": 0.4458, + "step": 20338 + }, + { + "epoch": 1.3650213080097984, + "grad_norm": 0.9432024955749512, + "learning_rate": 2.4179450043920933e-05, + "loss": 0.4646, + "step": 20340 + }, + { + "epoch": 1.3651555316935673, + "grad_norm": 1.0819599628448486, + "learning_rate": 2.4170143679040552e-05, + "loss": 0.5303, + "step": 20342 + }, + { + "epoch": 1.3652897553773364, + "grad_norm": 1.0137512683868408, + "learning_rate": 2.4160838534570074e-05, + "loss": 0.4371, + "step": 20344 + }, + { + "epoch": 1.3654239790611054, + "grad_norm": 1.136479139328003, + "learning_rate": 2.4151534610949178e-05, + "loss": 0.4898, + "step": 20346 + }, + { + "epoch": 1.3655582027448743, + "grad_norm": 0.8670485615730286, + "learning_rate": 2.4142231908617445e-05, + "loss": 0.477, + "step": 20348 + }, + { + "epoch": 1.3656924264286434, + "grad_norm": 1.0853685140609741, + "learning_rate": 2.413293042801441e-05, + "loss": 0.534, + "step": 20350 + }, + { + "epoch": 1.3658266501124123, + "grad_norm": 1.0450760126113892, + "learning_rate": 2.4123630169579526e-05, + "loss": 0.4518, + "step": 20352 + }, + { + "epoch": 1.3659608737961815, + "grad_norm": 1.060826063156128, + "learning_rate": 2.4114331133752254e-05, + "loss": 0.4867, + "step": 20354 + }, + { + "epoch": 1.3660950974799504, + "grad_norm": 1.0384961366653442, + "learning_rate": 2.410503332097193e-05, + "loss": 0.4835, + "step": 20356 + }, + { + "epoch": 1.3662293211637193, + "grad_norm": 1.0962611436843872, + "learning_rate": 2.4095736731677863e-05, + "loss": 0.5329, + "step": 20358 + }, + { + "epoch": 1.3663635448474882, + "grad_norm": 1.1338086128234863, + "learning_rate": 2.4086441366309277e-05, + "loss": 0.4969, + "step": 20360 + }, + { + "epoch": 1.3664977685312574, + "grad_norm": 1.0492395162582397, + "learning_rate": 2.4077147225305403e-05, + "loss": 0.4232, + "step": 20362 + }, + { + "epoch": 1.3666319922150263, + "grad_norm": 0.9534843564033508, + "learning_rate": 2.4067854309105326e-05, + "loss": 0.4584, + "step": 20364 + }, + { + "epoch": 1.3667662158987954, + "grad_norm": 1.0854427814483643, + "learning_rate": 2.4058562618148155e-05, + "loss": 0.4982, + "step": 20366 + }, + { + "epoch": 1.3669004395825644, + "grad_norm": 1.072067379951477, + "learning_rate": 2.4049272152872893e-05, + "loss": 0.511, + "step": 20368 + }, + { + "epoch": 1.3670346632663333, + "grad_norm": 1.0549293756484985, + "learning_rate": 2.403998291371849e-05, + "loss": 0.5146, + "step": 20370 + }, + { + "epoch": 1.3671688869501024, + "grad_norm": 1.0453256368637085, + "learning_rate": 2.4030694901123825e-05, + "loss": 0.5173, + "step": 20372 + }, + { + "epoch": 1.3673031106338713, + "grad_norm": 1.1417920589447021, + "learning_rate": 2.402140811552778e-05, + "loss": 0.4684, + "step": 20374 + }, + { + "epoch": 1.3674373343176405, + "grad_norm": 0.9018839001655579, + "learning_rate": 2.401212255736912e-05, + "loss": 0.4742, + "step": 20376 + }, + { + "epoch": 1.3675715580014094, + "grad_norm": 1.0004795789718628, + "learning_rate": 2.400283822708656e-05, + "loss": 0.5078, + "step": 20378 + }, + { + "epoch": 1.3677057816851783, + "grad_norm": 0.8867558836936951, + "learning_rate": 2.399355512511876e-05, + "loss": 0.4474, + "step": 20380 + }, + { + "epoch": 1.3678400053689472, + "grad_norm": 1.1225106716156006, + "learning_rate": 2.3984273251904356e-05, + "loss": 0.4629, + "step": 20382 + }, + { + "epoch": 1.3679742290527164, + "grad_norm": 0.963327169418335, + "learning_rate": 2.3974992607881887e-05, + "loss": 0.4835, + "step": 20384 + }, + { + "epoch": 1.3681084527364853, + "grad_norm": 1.0399508476257324, + "learning_rate": 2.3965713193489836e-05, + "loss": 0.4706, + "step": 20386 + }, + { + "epoch": 1.3682426764202544, + "grad_norm": 1.0206793546676636, + "learning_rate": 2.3956435009166627e-05, + "loss": 0.4832, + "step": 20388 + }, + { + "epoch": 1.3683769001040234, + "grad_norm": 1.1822404861450195, + "learning_rate": 2.394715805535067e-05, + "loss": 0.5164, + "step": 20390 + }, + { + "epoch": 1.3685111237877923, + "grad_norm": 1.8236244916915894, + "learning_rate": 2.3937882332480243e-05, + "loss": 0.4687, + "step": 20392 + }, + { + "epoch": 1.3686453474715614, + "grad_norm": 1.1084983348846436, + "learning_rate": 2.392860784099365e-05, + "loss": 0.5216, + "step": 20394 + }, + { + "epoch": 1.3687795711553303, + "grad_norm": 1.001785159111023, + "learning_rate": 2.3919334581329082e-05, + "loss": 0.5059, + "step": 20396 + }, + { + "epoch": 1.3689137948390995, + "grad_norm": 1.0605645179748535, + "learning_rate": 2.391006255392464e-05, + "loss": 0.5077, + "step": 20398 + }, + { + "epoch": 1.3690480185228684, + "grad_norm": 0.8846589922904968, + "learning_rate": 2.390079175921845e-05, + "loss": 0.4692, + "step": 20400 + }, + { + "epoch": 1.3691822422066373, + "grad_norm": 1.018807053565979, + "learning_rate": 2.3891522197648508e-05, + "loss": 0.498, + "step": 20402 + }, + { + "epoch": 1.3693164658904062, + "grad_norm": 1.0309909582138062, + "learning_rate": 2.3882253869652828e-05, + "loss": 0.5042, + "step": 20404 + }, + { + "epoch": 1.3694506895741754, + "grad_norm": 1.1537896394729614, + "learning_rate": 2.3872986775669288e-05, + "loss": 0.4564, + "step": 20406 + }, + { + "epoch": 1.3695849132579443, + "grad_norm": 0.9856048822402954, + "learning_rate": 2.386372091613575e-05, + "loss": 0.4525, + "step": 20408 + }, + { + "epoch": 1.3697191369417134, + "grad_norm": 1.110657811164856, + "learning_rate": 2.3854456291489987e-05, + "loss": 0.4641, + "step": 20410 + }, + { + "epoch": 1.3698533606254824, + "grad_norm": 0.9858686327934265, + "learning_rate": 2.3845192902169766e-05, + "loss": 0.4832, + "step": 20412 + }, + { + "epoch": 1.3699875843092513, + "grad_norm": 1.1257051229476929, + "learning_rate": 2.383593074861275e-05, + "loss": 0.4913, + "step": 20414 + }, + { + "epoch": 1.3701218079930204, + "grad_norm": 0.9121658802032471, + "learning_rate": 2.3826669831256554e-05, + "loss": 0.4668, + "step": 20416 + }, + { + "epoch": 1.3702560316767893, + "grad_norm": 1.0087146759033203, + "learning_rate": 2.3817410150538728e-05, + "loss": 0.5393, + "step": 20418 + }, + { + "epoch": 1.3703902553605585, + "grad_norm": 1.0078567266464233, + "learning_rate": 2.3808151706896804e-05, + "loss": 0.5155, + "step": 20420 + }, + { + "epoch": 1.3705244790443274, + "grad_norm": 1.0872535705566406, + "learning_rate": 2.3798894500768183e-05, + "loss": 0.4565, + "step": 20422 + }, + { + "epoch": 1.3706587027280963, + "grad_norm": 1.031997561454773, + "learning_rate": 2.378963853259032e-05, + "loss": 0.4866, + "step": 20424 + }, + { + "epoch": 1.3707929264118655, + "grad_norm": 1.018782615661621, + "learning_rate": 2.3780383802800453e-05, + "loss": 0.4596, + "step": 20426 + }, + { + "epoch": 1.3709271500956344, + "grad_norm": 1.0133177042007446, + "learning_rate": 2.3771130311835914e-05, + "loss": 0.4945, + "step": 20428 + }, + { + "epoch": 1.3710613737794035, + "grad_norm": 1.014116883277893, + "learning_rate": 2.3761878060133868e-05, + "loss": 0.4044, + "step": 20430 + }, + { + "epoch": 1.3711955974631724, + "grad_norm": 1.083739995956421, + "learning_rate": 2.3752627048131516e-05, + "loss": 0.4973, + "step": 20432 + }, + { + "epoch": 1.3713298211469414, + "grad_norm": 1.0590863227844238, + "learning_rate": 2.3743377276265916e-05, + "loss": 0.4977, + "step": 20434 + }, + { + "epoch": 1.3714640448307103, + "grad_norm": 1.184765100479126, + "learning_rate": 2.373412874497411e-05, + "loss": 0.4348, + "step": 20436 + }, + { + "epoch": 1.3715982685144794, + "grad_norm": 1.0506863594055176, + "learning_rate": 2.3724881454693048e-05, + "loss": 0.4796, + "step": 20438 + }, + { + "epoch": 1.3717324921982483, + "grad_norm": 1.010758399963379, + "learning_rate": 2.3715635405859692e-05, + "loss": 0.4563, + "step": 20440 + }, + { + "epoch": 1.3718667158820175, + "grad_norm": 0.9689465761184692, + "learning_rate": 2.3706390598910878e-05, + "loss": 0.4726, + "step": 20442 + }, + { + "epoch": 1.3720009395657864, + "grad_norm": 0.9940888285636902, + "learning_rate": 2.3697147034283407e-05, + "loss": 0.4485, + "step": 20444 + }, + { + "epoch": 1.3721351632495553, + "grad_norm": 0.9998622536659241, + "learning_rate": 2.3687904712413996e-05, + "loss": 0.4643, + "step": 20446 + }, + { + "epoch": 1.3722693869333245, + "grad_norm": 1.1568114757537842, + "learning_rate": 2.367866363373936e-05, + "loss": 0.5178, + "step": 20448 + }, + { + "epoch": 1.3724036106170934, + "grad_norm": 0.9747192859649658, + "learning_rate": 2.3669423798696094e-05, + "loss": 0.5501, + "step": 20450 + }, + { + "epoch": 1.3725378343008625, + "grad_norm": 0.9922822117805481, + "learning_rate": 2.3660185207720813e-05, + "loss": 0.4787, + "step": 20452 + }, + { + "epoch": 1.3726720579846314, + "grad_norm": 0.961625337600708, + "learning_rate": 2.365094786124995e-05, + "loss": 0.4688, + "step": 20454 + }, + { + "epoch": 1.3728062816684004, + "grad_norm": 1.0886564254760742, + "learning_rate": 2.3641711759720003e-05, + "loss": 0.4928, + "step": 20456 + }, + { + "epoch": 1.3729405053521693, + "grad_norm": 1.1044477224349976, + "learning_rate": 2.363247690356733e-05, + "loss": 0.4975, + "step": 20458 + }, + { + "epoch": 1.3730747290359384, + "grad_norm": 1.0163726806640625, + "learning_rate": 2.3623243293228287e-05, + "loss": 0.5184, + "step": 20460 + }, + { + "epoch": 1.3732089527197073, + "grad_norm": 0.9921759963035583, + "learning_rate": 2.3614010929139136e-05, + "loss": 0.4297, + "step": 20462 + }, + { + "epoch": 1.3733431764034765, + "grad_norm": 1.009995937347412, + "learning_rate": 2.360477981173608e-05, + "loss": 0.4722, + "step": 20464 + }, + { + "epoch": 1.3734774000872454, + "grad_norm": 0.9856315851211548, + "learning_rate": 2.3595549941455276e-05, + "loss": 0.5022, + "step": 20466 + }, + { + "epoch": 1.3736116237710143, + "grad_norm": 0.8814576864242554, + "learning_rate": 2.358632131873279e-05, + "loss": 0.4634, + "step": 20468 + }, + { + "epoch": 1.3737458474547835, + "grad_norm": 1.0409995317459106, + "learning_rate": 2.3577093944004707e-05, + "loss": 0.5505, + "step": 20470 + }, + { + "epoch": 1.3738800711385524, + "grad_norm": 1.1030895709991455, + "learning_rate": 2.3567867817706974e-05, + "loss": 0.5467, + "step": 20472 + }, + { + "epoch": 1.3740142948223215, + "grad_norm": 0.9939549565315247, + "learning_rate": 2.355864294027551e-05, + "loss": 0.4681, + "step": 20474 + }, + { + "epoch": 1.3741485185060904, + "grad_norm": 1.004815697669983, + "learning_rate": 2.3549419312146153e-05, + "loss": 0.4403, + "step": 20476 + }, + { + "epoch": 1.3742827421898594, + "grad_norm": 1.0654927492141724, + "learning_rate": 2.3540196933754743e-05, + "loss": 0.4571, + "step": 20478 + }, + { + "epoch": 1.3744169658736283, + "grad_norm": 1.0886633396148682, + "learning_rate": 2.3530975805536993e-05, + "loss": 0.4856, + "step": 20480 + }, + { + "epoch": 1.3745511895573974, + "grad_norm": 1.0018386840820312, + "learning_rate": 2.352175592792859e-05, + "loss": 0.4564, + "step": 20482 + }, + { + "epoch": 1.3746854132411663, + "grad_norm": 0.9411597847938538, + "learning_rate": 2.3512537301365134e-05, + "loss": 0.4969, + "step": 20484 + }, + { + "epoch": 1.3748196369249355, + "grad_norm": 0.9695761203765869, + "learning_rate": 2.3503319926282218e-05, + "loss": 0.4395, + "step": 20486 + }, + { + "epoch": 1.3749538606087044, + "grad_norm": 1.0185626745224, + "learning_rate": 2.349410380311532e-05, + "loss": 0.5126, + "step": 20488 + }, + { + "epoch": 1.3750880842924733, + "grad_norm": 1.1450790166854858, + "learning_rate": 2.348488893229991e-05, + "loss": 0.5268, + "step": 20490 + }, + { + "epoch": 1.3752223079762425, + "grad_norm": 1.0380408763885498, + "learning_rate": 2.3475675314271363e-05, + "loss": 0.4382, + "step": 20492 + }, + { + "epoch": 1.3753565316600114, + "grad_norm": 1.0181218385696411, + "learning_rate": 2.3466462949464996e-05, + "loss": 0.5113, + "step": 20494 + }, + { + "epoch": 1.3754907553437805, + "grad_norm": 0.9403045773506165, + "learning_rate": 2.345725183831606e-05, + "loss": 0.5087, + "step": 20496 + }, + { + "epoch": 1.3756249790275494, + "grad_norm": 1.0722616910934448, + "learning_rate": 2.34480419812598e-05, + "loss": 0.4588, + "step": 20498 + }, + { + "epoch": 1.3757592027113184, + "grad_norm": 0.9313488006591797, + "learning_rate": 2.3438833378731338e-05, + "loss": 0.4626, + "step": 20500 + }, + { + "epoch": 1.3758934263950875, + "grad_norm": 1.1240397691726685, + "learning_rate": 2.342962603116578e-05, + "loss": 0.5289, + "step": 20502 + }, + { + "epoch": 1.3760276500788564, + "grad_norm": 1.1276453733444214, + "learning_rate": 2.342041993899812e-05, + "loss": 0.5325, + "step": 20504 + }, + { + "epoch": 1.3761618737626256, + "grad_norm": 1.1331275701522827, + "learning_rate": 2.341121510266337e-05, + "loss": 0.4912, + "step": 20506 + }, + { + "epoch": 1.3762960974463945, + "grad_norm": 0.7722598910331726, + "learning_rate": 2.340201152259642e-05, + "loss": 0.4556, + "step": 20508 + }, + { + "epoch": 1.3764303211301634, + "grad_norm": 1.0412757396697998, + "learning_rate": 2.3392809199232123e-05, + "loss": 0.4943, + "step": 20510 + }, + { + "epoch": 1.3765645448139323, + "grad_norm": 0.9970412254333496, + "learning_rate": 2.3383608133005253e-05, + "loss": 0.4934, + "step": 20512 + }, + { + "epoch": 1.3766987684977015, + "grad_norm": 0.9749853014945984, + "learning_rate": 2.337440832435058e-05, + "loss": 0.4486, + "step": 20514 + }, + { + "epoch": 1.3768329921814704, + "grad_norm": 1.1657603979110718, + "learning_rate": 2.3365209773702736e-05, + "loss": 0.5406, + "step": 20516 + }, + { + "epoch": 1.3769672158652395, + "grad_norm": 0.8795127868652344, + "learning_rate": 2.335601248149637e-05, + "loss": 0.4899, + "step": 20518 + }, + { + "epoch": 1.3771014395490084, + "grad_norm": 0.8475943207740784, + "learning_rate": 2.3346816448166025e-05, + "loss": 0.47, + "step": 20520 + }, + { + "epoch": 1.3772356632327774, + "grad_norm": 1.1030148267745972, + "learning_rate": 2.3337621674146193e-05, + "loss": 0.5703, + "step": 20522 + }, + { + "epoch": 1.3773698869165465, + "grad_norm": 1.0147314071655273, + "learning_rate": 2.332842815987129e-05, + "loss": 0.5035, + "step": 20524 + }, + { + "epoch": 1.3775041106003154, + "grad_norm": 1.1151841878890991, + "learning_rate": 2.3319235905775727e-05, + "loss": 0.4712, + "step": 20526 + }, + { + "epoch": 1.3776383342840846, + "grad_norm": 0.8732739090919495, + "learning_rate": 2.33100449122938e-05, + "loss": 0.5501, + "step": 20528 + }, + { + "epoch": 1.3777725579678535, + "grad_norm": 1.0400519371032715, + "learning_rate": 2.3300855179859765e-05, + "loss": 0.4494, + "step": 20530 + }, + { + "epoch": 1.3779067816516224, + "grad_norm": 1.0130256414413452, + "learning_rate": 2.3291666708907826e-05, + "loss": 0.4888, + "step": 20532 + }, + { + "epoch": 1.3780410053353913, + "grad_norm": 0.9467426538467407, + "learning_rate": 2.3282479499872095e-05, + "loss": 0.4929, + "step": 20534 + }, + { + "epoch": 1.3781752290191605, + "grad_norm": 0.9923089742660522, + "learning_rate": 2.3273293553186688e-05, + "loss": 0.5216, + "step": 20536 + }, + { + "epoch": 1.3783094527029294, + "grad_norm": 1.347734808921814, + "learning_rate": 2.3264108869285606e-05, + "loss": 0.5083, + "step": 20538 + }, + { + "epoch": 1.3784436763866985, + "grad_norm": 0.9640908241271973, + "learning_rate": 2.3254925448602806e-05, + "loss": 0.5134, + "step": 20540 + }, + { + "epoch": 1.3785779000704674, + "grad_norm": 1.0231845378875732, + "learning_rate": 2.3245743291572164e-05, + "loss": 0.4919, + "step": 20542 + }, + { + "epoch": 1.3787121237542364, + "grad_norm": 1.092300534248352, + "learning_rate": 2.3236562398627566e-05, + "loss": 0.5318, + "step": 20544 + }, + { + "epoch": 1.3788463474380055, + "grad_norm": 1.0548770427703857, + "learning_rate": 2.3227382770202744e-05, + "loss": 0.4472, + "step": 20546 + }, + { + "epoch": 1.3789805711217744, + "grad_norm": 0.9882622957229614, + "learning_rate": 2.3218204406731474e-05, + "loss": 0.518, + "step": 20548 + }, + { + "epoch": 1.3791147948055436, + "grad_norm": 0.9383532404899597, + "learning_rate": 2.3209027308647352e-05, + "loss": 0.4538, + "step": 20550 + }, + { + "epoch": 1.3792490184893125, + "grad_norm": 1.078097939491272, + "learning_rate": 2.3199851476384022e-05, + "loss": 0.535, + "step": 20552 + }, + { + "epoch": 1.3793832421730814, + "grad_norm": 1.1034539937973022, + "learning_rate": 2.319067691037499e-05, + "loss": 0.4945, + "step": 20554 + }, + { + "epoch": 1.3795174658568503, + "grad_norm": 0.9837169051170349, + "learning_rate": 2.318150361105378e-05, + "loss": 0.5057, + "step": 20556 + }, + { + "epoch": 1.3796516895406195, + "grad_norm": 0.9580581188201904, + "learning_rate": 2.3172331578853785e-05, + "loss": 0.4469, + "step": 20558 + }, + { + "epoch": 1.3797859132243884, + "grad_norm": 1.0174978971481323, + "learning_rate": 2.3163160814208367e-05, + "loss": 0.5027, + "step": 20560 + }, + { + "epoch": 1.3799201369081575, + "grad_norm": 1.1509947776794434, + "learning_rate": 2.315399131755081e-05, + "loss": 0.4928, + "step": 20562 + }, + { + "epoch": 1.3800543605919264, + "grad_norm": 0.9773992300033569, + "learning_rate": 2.3144823089314388e-05, + "loss": 0.4518, + "step": 20564 + }, + { + "epoch": 1.3801885842756954, + "grad_norm": 0.983257532119751, + "learning_rate": 2.313565612993227e-05, + "loss": 0.5034, + "step": 20566 + }, + { + "epoch": 1.3803228079594645, + "grad_norm": 1.0288691520690918, + "learning_rate": 2.3126490439837578e-05, + "loss": 0.5071, + "step": 20568 + }, + { + "epoch": 1.3804570316432334, + "grad_norm": 1.0755268335342407, + "learning_rate": 2.311732601946334e-05, + "loss": 0.4987, + "step": 20570 + }, + { + "epoch": 1.3805912553270026, + "grad_norm": 1.0262552499771118, + "learning_rate": 2.310816286924261e-05, + "loss": 0.5021, + "step": 20572 + }, + { + "epoch": 1.3807254790107715, + "grad_norm": 1.126383662223816, + "learning_rate": 2.3099000989608288e-05, + "loss": 0.4603, + "step": 20574 + }, + { + "epoch": 1.3808597026945404, + "grad_norm": 1.0954885482788086, + "learning_rate": 2.3089840380993305e-05, + "loss": 0.4856, + "step": 20576 + }, + { + "epoch": 1.3809939263783095, + "grad_norm": 1.0396054983139038, + "learning_rate": 2.308068104383041e-05, + "loss": 0.4617, + "step": 20578 + }, + { + "epoch": 1.3811281500620785, + "grad_norm": 1.0048110485076904, + "learning_rate": 2.3071522978552418e-05, + "loss": 0.4569, + "step": 20580 + }, + { + "epoch": 1.3812623737458474, + "grad_norm": 0.8580957651138306, + "learning_rate": 2.3062366185591995e-05, + "loss": 0.4575, + "step": 20582 + }, + { + "epoch": 1.3813965974296165, + "grad_norm": 0.9834871888160706, + "learning_rate": 2.3053210665381818e-05, + "loss": 0.4724, + "step": 20584 + }, + { + "epoch": 1.3815308211133854, + "grad_norm": 1.0720840692520142, + "learning_rate": 2.304405641835445e-05, + "loss": 0.5053, + "step": 20586 + }, + { + "epoch": 1.3816650447971544, + "grad_norm": 0.9133841395378113, + "learning_rate": 2.3034903444942406e-05, + "loss": 0.4478, + "step": 20588 + }, + { + "epoch": 1.3817992684809235, + "grad_norm": 1.0490212440490723, + "learning_rate": 2.3025751745578134e-05, + "loss": 0.5149, + "step": 20590 + }, + { + "epoch": 1.3819334921646924, + "grad_norm": 1.0795177221298218, + "learning_rate": 2.3016601320694065e-05, + "loss": 0.5253, + "step": 20592 + }, + { + "epoch": 1.3820677158484616, + "grad_norm": 0.8143784999847412, + "learning_rate": 2.3007452170722528e-05, + "loss": 0.4588, + "step": 20594 + }, + { + "epoch": 1.3822019395322305, + "grad_norm": 1.0718694925308228, + "learning_rate": 2.299830429609579e-05, + "loss": 0.4899, + "step": 20596 + }, + { + "epoch": 1.3823361632159994, + "grad_norm": 1.0476452112197876, + "learning_rate": 2.2989157697246066e-05, + "loss": 0.5106, + "step": 20598 + }, + { + "epoch": 1.3824703868997685, + "grad_norm": 0.9038621187210083, + "learning_rate": 2.2980012374605542e-05, + "loss": 0.4656, + "step": 20600 + }, + { + "epoch": 1.3826046105835375, + "grad_norm": 1.0184696912765503, + "learning_rate": 2.29708683286063e-05, + "loss": 0.5373, + "step": 20602 + }, + { + "epoch": 1.3827388342673066, + "grad_norm": 0.9636785387992859, + "learning_rate": 2.296172555968038e-05, + "loss": 0.5063, + "step": 20604 + }, + { + "epoch": 1.3828730579510755, + "grad_norm": 0.9802426099777222, + "learning_rate": 2.2952584068259757e-05, + "loss": 0.4836, + "step": 20606 + }, + { + "epoch": 1.3830072816348444, + "grad_norm": 1.0625511407852173, + "learning_rate": 2.2943443854776335e-05, + "loss": 0.486, + "step": 20608 + }, + { + "epoch": 1.3831415053186134, + "grad_norm": 1.0863850116729736, + "learning_rate": 2.2934304919662003e-05, + "loss": 0.4964, + "step": 20610 + }, + { + "epoch": 1.3832757290023825, + "grad_norm": 0.8596870303153992, + "learning_rate": 2.2925167263348525e-05, + "loss": 0.4592, + "step": 20612 + }, + { + "epoch": 1.3834099526861514, + "grad_norm": 0.9515671133995056, + "learning_rate": 2.2916030886267687e-05, + "loss": 0.4575, + "step": 20614 + }, + { + "epoch": 1.3835441763699206, + "grad_norm": 1.1263680458068848, + "learning_rate": 2.2906895788851097e-05, + "loss": 0.5034, + "step": 20616 + }, + { + "epoch": 1.3836784000536895, + "grad_norm": 1.0509198904037476, + "learning_rate": 2.2897761971530425e-05, + "loss": 0.4805, + "step": 20618 + }, + { + "epoch": 1.3838126237374584, + "grad_norm": 0.9480360746383667, + "learning_rate": 2.288862943473718e-05, + "loss": 0.5098, + "step": 20620 + }, + { + "epoch": 1.3839468474212275, + "grad_norm": 0.9155513048171997, + "learning_rate": 2.2879498178902904e-05, + "loss": 0.4373, + "step": 20622 + }, + { + "epoch": 1.3840810711049965, + "grad_norm": 0.9924333095550537, + "learning_rate": 2.2870368204459008e-05, + "loss": 0.4545, + "step": 20624 + }, + { + "epoch": 1.3842152947887656, + "grad_norm": 1.1303173303604126, + "learning_rate": 2.2861239511836867e-05, + "loss": 0.4909, + "step": 20626 + }, + { + "epoch": 1.3843495184725345, + "grad_norm": 0.9560405611991882, + "learning_rate": 2.2852112101467778e-05, + "loss": 0.4618, + "step": 20628 + }, + { + "epoch": 1.3844837421563034, + "grad_norm": 1.2171663045883179, + "learning_rate": 2.2842985973783026e-05, + "loss": 0.4482, + "step": 20630 + }, + { + "epoch": 1.3846179658400724, + "grad_norm": 0.9853360056877136, + "learning_rate": 2.283386112921378e-05, + "loss": 0.515, + "step": 20632 + }, + { + "epoch": 1.3847521895238415, + "grad_norm": 0.9441184997558594, + "learning_rate": 2.2824737568191178e-05, + "loss": 0.5091, + "step": 20634 + }, + { + "epoch": 1.3848864132076104, + "grad_norm": 1.2589091062545776, + "learning_rate": 2.281561529114627e-05, + "loss": 0.5033, + "step": 20636 + }, + { + "epoch": 1.3850206368913796, + "grad_norm": 0.9849038124084473, + "learning_rate": 2.28064942985101e-05, + "loss": 0.5216, + "step": 20638 + }, + { + "epoch": 1.3851548605751485, + "grad_norm": 0.9936666488647461, + "learning_rate": 2.279737459071359e-05, + "loss": 0.4874, + "step": 20640 + }, + { + "epoch": 1.3852890842589174, + "grad_norm": 1.0241390466690063, + "learning_rate": 2.2788256168187665e-05, + "loss": 0.4866, + "step": 20642 + }, + { + "epoch": 1.3854233079426865, + "grad_norm": 0.9956193566322327, + "learning_rate": 2.2779139031363096e-05, + "loss": 0.4914, + "step": 20644 + }, + { + "epoch": 1.3855575316264555, + "grad_norm": 1.0614651441574097, + "learning_rate": 2.2770023180670698e-05, + "loss": 0.5134, + "step": 20646 + }, + { + "epoch": 1.3856917553102246, + "grad_norm": 0.9981801509857178, + "learning_rate": 2.2760908616541136e-05, + "loss": 0.515, + "step": 20648 + }, + { + "epoch": 1.3858259789939935, + "grad_norm": 0.9139376878738403, + "learning_rate": 2.2751795339405103e-05, + "loss": 0.4629, + "step": 20650 + }, + { + "epoch": 1.3859602026777624, + "grad_norm": 0.9527412056922913, + "learning_rate": 2.274268334969316e-05, + "loss": 0.4909, + "step": 20652 + }, + { + "epoch": 1.3860944263615316, + "grad_norm": 1.0902884006500244, + "learning_rate": 2.2733572647835828e-05, + "loss": 0.5259, + "step": 20654 + }, + { + "epoch": 1.3862286500453005, + "grad_norm": 0.9523167014122009, + "learning_rate": 2.2724463234263548e-05, + "loss": 0.4923, + "step": 20656 + }, + { + "epoch": 1.3863628737290694, + "grad_norm": 1.0305126905441284, + "learning_rate": 2.2715355109406766e-05, + "loss": 0.4921, + "step": 20658 + }, + { + "epoch": 1.3864970974128386, + "grad_norm": 1.0861053466796875, + "learning_rate": 2.27062482736958e-05, + "loss": 0.5463, + "step": 20660 + }, + { + "epoch": 1.3866313210966075, + "grad_norm": 1.0559502840042114, + "learning_rate": 2.2697142727560937e-05, + "loss": 0.5093, + "step": 20662 + }, + { + "epoch": 1.3867655447803764, + "grad_norm": 0.9299308657646179, + "learning_rate": 2.2688038471432373e-05, + "loss": 0.564, + "step": 20664 + }, + { + "epoch": 1.3868997684641455, + "grad_norm": 1.0282251834869385, + "learning_rate": 2.26789355057403e-05, + "loss": 0.4818, + "step": 20666 + }, + { + "epoch": 1.3870339921479145, + "grad_norm": 0.9290526509284973, + "learning_rate": 2.2669833830914782e-05, + "loss": 0.4723, + "step": 20668 + }, + { + "epoch": 1.3871682158316836, + "grad_norm": 0.9008547067642212, + "learning_rate": 2.2660733447385894e-05, + "loss": 0.4255, + "step": 20670 + }, + { + "epoch": 1.3873024395154525, + "grad_norm": 0.8995382189750671, + "learning_rate": 2.2651634355583606e-05, + "loss": 0.5014, + "step": 20672 + }, + { + "epoch": 1.3874366631992214, + "grad_norm": 1.4763998985290527, + "learning_rate": 2.2642536555937783e-05, + "loss": 0.493, + "step": 20674 + }, + { + "epoch": 1.3875708868829906, + "grad_norm": 1.1050307750701904, + "learning_rate": 2.2633440048878323e-05, + "loss": 0.4752, + "step": 20676 + }, + { + "epoch": 1.3877051105667595, + "grad_norm": 0.9923198819160461, + "learning_rate": 2.2624344834834992e-05, + "loss": 0.4421, + "step": 20678 + }, + { + "epoch": 1.3878393342505286, + "grad_norm": 1.04277765750885, + "learning_rate": 2.2615250914237558e-05, + "loss": 0.5183, + "step": 20680 + }, + { + "epoch": 1.3879735579342976, + "grad_norm": 1.0581557750701904, + "learning_rate": 2.260615828751566e-05, + "loss": 0.4809, + "step": 20682 + }, + { + "epoch": 1.3881077816180665, + "grad_norm": 1.04483962059021, + "learning_rate": 2.2597066955098923e-05, + "loss": 0.4696, + "step": 20684 + }, + { + "epoch": 1.3882420053018354, + "grad_norm": 0.9544757604598999, + "learning_rate": 2.2587976917416864e-05, + "loss": 0.45, + "step": 20686 + }, + { + "epoch": 1.3883762289856045, + "grad_norm": 1.074668049812317, + "learning_rate": 2.2578888174899014e-05, + "loss": 0.5675, + "step": 20688 + }, + { + "epoch": 1.3885104526693735, + "grad_norm": 0.9719125032424927, + "learning_rate": 2.2569800727974777e-05, + "loss": 0.4975, + "step": 20690 + }, + { + "epoch": 1.3886446763531426, + "grad_norm": 0.8937525749206543, + "learning_rate": 2.2560714577073512e-05, + "loss": 0.4317, + "step": 20692 + }, + { + "epoch": 1.3887789000369115, + "grad_norm": 1.0742558240890503, + "learning_rate": 2.255162972262451e-05, + "loss": 0.457, + "step": 20694 + }, + { + "epoch": 1.3889131237206804, + "grad_norm": 0.9840453267097473, + "learning_rate": 2.254254616505705e-05, + "loss": 0.5375, + "step": 20696 + }, + { + "epoch": 1.3890473474044496, + "grad_norm": 1.455519676208496, + "learning_rate": 2.2533463904800268e-05, + "loss": 0.4567, + "step": 20698 + }, + { + "epoch": 1.3891815710882185, + "grad_norm": 1.0274455547332764, + "learning_rate": 2.2524382942283345e-05, + "loss": 0.5004, + "step": 20700 + }, + { + "epoch": 1.3893157947719876, + "grad_norm": 1.2613234519958496, + "learning_rate": 2.2515303277935257e-05, + "loss": 0.5372, + "step": 20702 + }, + { + "epoch": 1.3894500184557566, + "grad_norm": 0.9007759690284729, + "learning_rate": 2.250622491218507e-05, + "loss": 0.4797, + "step": 20704 + }, + { + "epoch": 1.3895842421395255, + "grad_norm": 1.0045149326324463, + "learning_rate": 2.2497147845461668e-05, + "loss": 0.4741, + "step": 20706 + }, + { + "epoch": 1.3897184658232944, + "grad_norm": 1.0059312582015991, + "learning_rate": 2.2488072078193968e-05, + "loss": 0.4911, + "step": 20708 + }, + { + "epoch": 1.3898526895070635, + "grad_norm": 0.9289558529853821, + "learning_rate": 2.2478997610810764e-05, + "loss": 0.4067, + "step": 20710 + }, + { + "epoch": 1.3899869131908325, + "grad_norm": 1.004197120666504, + "learning_rate": 2.2469924443740804e-05, + "loss": 0.5426, + "step": 20712 + }, + { + "epoch": 1.3901211368746016, + "grad_norm": 1.1435902118682861, + "learning_rate": 2.2460852577412754e-05, + "loss": 0.5566, + "step": 20714 + }, + { + "epoch": 1.3902553605583705, + "grad_norm": 1.0519390106201172, + "learning_rate": 2.2451782012255294e-05, + "loss": 0.4759, + "step": 20716 + }, + { + "epoch": 1.3903895842421394, + "grad_norm": 0.9619807600975037, + "learning_rate": 2.244271274869696e-05, + "loss": 0.4767, + "step": 20718 + }, + { + "epoch": 1.3905238079259086, + "grad_norm": 1.1317015886306763, + "learning_rate": 2.243364478716626e-05, + "loss": 0.4815, + "step": 20720 + }, + { + "epoch": 1.3906580316096775, + "grad_norm": 1.0051021575927734, + "learning_rate": 2.2424578128091617e-05, + "loss": 0.5831, + "step": 20722 + }, + { + "epoch": 1.3907922552934466, + "grad_norm": 1.0220969915390015, + "learning_rate": 2.241551277190145e-05, + "loss": 0.4742, + "step": 20724 + }, + { + "epoch": 1.3909264789772156, + "grad_norm": 0.9811525344848633, + "learning_rate": 2.2406448719024044e-05, + "loss": 0.5178, + "step": 20726 + }, + { + "epoch": 1.3910607026609845, + "grad_norm": 1.10579514503479, + "learning_rate": 2.2397385969887713e-05, + "loss": 0.5329, + "step": 20728 + }, + { + "epoch": 1.3911949263447536, + "grad_norm": 1.003940463066101, + "learning_rate": 2.2388324524920577e-05, + "loss": 0.4809, + "step": 20730 + }, + { + "epoch": 1.3913291500285225, + "grad_norm": 0.9447029829025269, + "learning_rate": 2.2379264384550836e-05, + "loss": 0.472, + "step": 20732 + }, + { + "epoch": 1.3914633737122915, + "grad_norm": 0.9570050835609436, + "learning_rate": 2.237020554920652e-05, + "loss": 0.4414, + "step": 20734 + }, + { + "epoch": 1.3915975973960606, + "grad_norm": 0.9190820455551147, + "learning_rate": 2.236114801931567e-05, + "loss": 0.4616, + "step": 20736 + }, + { + "epoch": 1.3917318210798295, + "grad_norm": 1.0760704278945923, + "learning_rate": 2.2352091795306252e-05, + "loss": 0.4619, + "step": 20738 + }, + { + "epoch": 1.3918660447635984, + "grad_norm": 0.981931209564209, + "learning_rate": 2.2343036877606094e-05, + "loss": 0.5765, + "step": 20740 + }, + { + "epoch": 1.3920002684473676, + "grad_norm": 0.9680532813072205, + "learning_rate": 2.233398326664307e-05, + "loss": 0.4447, + "step": 20742 + }, + { + "epoch": 1.3921344921311365, + "grad_norm": 1.0856574773788452, + "learning_rate": 2.232493096284492e-05, + "loss": 0.4907, + "step": 20744 + }, + { + "epoch": 1.3922687158149056, + "grad_norm": 1.0441486835479736, + "learning_rate": 2.231587996663938e-05, + "loss": 0.5341, + "step": 20746 + }, + { + "epoch": 1.3924029394986746, + "grad_norm": 1.130039095878601, + "learning_rate": 2.2306830278454072e-05, + "loss": 0.5279, + "step": 20748 + }, + { + "epoch": 1.3925371631824435, + "grad_norm": 0.9704619646072388, + "learning_rate": 2.229778189871658e-05, + "loss": 0.4829, + "step": 20750 + }, + { + "epoch": 1.3926713868662126, + "grad_norm": 1.6788395643234253, + "learning_rate": 2.2288734827854395e-05, + "loss": 0.4921, + "step": 20752 + }, + { + "epoch": 1.3928056105499815, + "grad_norm": 0.9546937346458435, + "learning_rate": 2.2279689066295022e-05, + "loss": 0.4837, + "step": 20754 + }, + { + "epoch": 1.3929398342337507, + "grad_norm": 1.1602784395217896, + "learning_rate": 2.227064461446583e-05, + "loss": 0.553, + "step": 20756 + }, + { + "epoch": 1.3930740579175196, + "grad_norm": 1.0249425172805786, + "learning_rate": 2.2261601472794147e-05, + "loss": 0.5028, + "step": 20758 + }, + { + "epoch": 1.3932082816012885, + "grad_norm": 1.065847396850586, + "learning_rate": 2.2252559641707237e-05, + "loss": 0.4242, + "step": 20760 + }, + { + "epoch": 1.3933425052850574, + "grad_norm": 0.9325951933860779, + "learning_rate": 2.2243519121632332e-05, + "loss": 0.4736, + "step": 20762 + }, + { + "epoch": 1.3934767289688266, + "grad_norm": 1.0492970943450928, + "learning_rate": 2.2234479912996557e-05, + "loss": 0.5372, + "step": 20764 + }, + { + "epoch": 1.3936109526525955, + "grad_norm": 1.1325078010559082, + "learning_rate": 2.2225442016227044e-05, + "loss": 0.5958, + "step": 20766 + }, + { + "epoch": 1.3937451763363646, + "grad_norm": 1.083055853843689, + "learning_rate": 2.2216405431750735e-05, + "loss": 0.5259, + "step": 20768 + }, + { + "epoch": 1.3938794000201336, + "grad_norm": 0.9872101545333862, + "learning_rate": 2.2207370159994662e-05, + "loss": 0.5425, + "step": 20770 + }, + { + "epoch": 1.3940136237039025, + "grad_norm": 0.9818466305732727, + "learning_rate": 2.2198336201385674e-05, + "loss": 0.4414, + "step": 20772 + }, + { + "epoch": 1.3941478473876716, + "grad_norm": 0.9530790448188782, + "learning_rate": 2.218930355635065e-05, + "loss": 0.4587, + "step": 20774 + }, + { + "epoch": 1.3942820710714405, + "grad_norm": 1.1509958505630493, + "learning_rate": 2.2180272225316346e-05, + "loss": 0.4853, + "step": 20776 + }, + { + "epoch": 1.3944162947552097, + "grad_norm": 0.9824808239936829, + "learning_rate": 2.2171242208709474e-05, + "loss": 0.464, + "step": 20778 + }, + { + "epoch": 1.3945505184389786, + "grad_norm": 1.0254404544830322, + "learning_rate": 2.216221350695667e-05, + "loss": 0.516, + "step": 20780 + }, + { + "epoch": 1.3946847421227475, + "grad_norm": 1.113036870956421, + "learning_rate": 2.2153186120484546e-05, + "loss": 0.5511, + "step": 20782 + }, + { + "epoch": 1.3948189658065164, + "grad_norm": 1.0598375797271729, + "learning_rate": 2.2144160049719632e-05, + "loss": 0.496, + "step": 20784 + }, + { + "epoch": 1.3949531894902856, + "grad_norm": 1.0639839172363281, + "learning_rate": 2.2135135295088373e-05, + "loss": 0.5056, + "step": 20786 + }, + { + "epoch": 1.3950874131740545, + "grad_norm": 1.0893691778182983, + "learning_rate": 2.2126111857017157e-05, + "loss": 0.5265, + "step": 20788 + }, + { + "epoch": 1.3952216368578236, + "grad_norm": 1.1033680438995361, + "learning_rate": 2.2117089735932367e-05, + "loss": 0.5766, + "step": 20790 + }, + { + "epoch": 1.3953558605415926, + "grad_norm": 1.1397067308425903, + "learning_rate": 2.2108068932260235e-05, + "loss": 0.5154, + "step": 20792 + }, + { + "epoch": 1.3954900842253615, + "grad_norm": 1.0490998029708862, + "learning_rate": 2.2099049446427038e-05, + "loss": 0.5106, + "step": 20794 + }, + { + "epoch": 1.3956243079091306, + "grad_norm": 1.205463171005249, + "learning_rate": 2.2090031278858854e-05, + "loss": 0.506, + "step": 20796 + }, + { + "epoch": 1.3957585315928995, + "grad_norm": 1.0334906578063965, + "learning_rate": 2.208101442998182e-05, + "loss": 0.4647, + "step": 20798 + }, + { + "epoch": 1.3958927552766687, + "grad_norm": 1.0384498834609985, + "learning_rate": 2.2071998900221947e-05, + "loss": 0.486, + "step": 20800 + }, + { + "epoch": 1.3960269789604376, + "grad_norm": 1.1685583591461182, + "learning_rate": 2.2062984690005224e-05, + "loss": 0.4676, + "step": 20802 + }, + { + "epoch": 1.3961612026442065, + "grad_norm": 0.9502678513526917, + "learning_rate": 2.2053971799757535e-05, + "loss": 0.486, + "step": 20804 + }, + { + "epoch": 1.3962954263279757, + "grad_norm": 0.962885320186615, + "learning_rate": 2.2044960229904722e-05, + "loss": 0.5062, + "step": 20806 + }, + { + "epoch": 1.3964296500117446, + "grad_norm": 1.1103259325027466, + "learning_rate": 2.2035949980872573e-05, + "loss": 0.5132, + "step": 20808 + }, + { + "epoch": 1.3965638736955135, + "grad_norm": 1.0677655935287476, + "learning_rate": 2.2026941053086773e-05, + "loss": 0.5633, + "step": 20810 + }, + { + "epoch": 1.3966980973792826, + "grad_norm": 1.0549362897872925, + "learning_rate": 2.2017933446973027e-05, + "loss": 0.4588, + "step": 20812 + }, + { + "epoch": 1.3968323210630516, + "grad_norm": 0.9541140198707581, + "learning_rate": 2.20089271629569e-05, + "loss": 0.4411, + "step": 20814 + }, + { + "epoch": 1.3969665447468205, + "grad_norm": 0.9641833305358887, + "learning_rate": 2.199992220146392e-05, + "loss": 0.488, + "step": 20816 + }, + { + "epoch": 1.3971007684305896, + "grad_norm": 1.637166142463684, + "learning_rate": 2.199091856291954e-05, + "loss": 0.4531, + "step": 20818 + }, + { + "epoch": 1.3972349921143585, + "grad_norm": 1.167707920074463, + "learning_rate": 2.1981916247749196e-05, + "loss": 0.4867, + "step": 20820 + }, + { + "epoch": 1.3973692157981277, + "grad_norm": 0.9572454690933228, + "learning_rate": 2.1972915256378217e-05, + "loss": 0.4576, + "step": 20822 + }, + { + "epoch": 1.3975034394818966, + "grad_norm": 1.1723190546035767, + "learning_rate": 2.1963915589231876e-05, + "loss": 0.4989, + "step": 20824 + }, + { + "epoch": 1.3976376631656655, + "grad_norm": 0.9696754217147827, + "learning_rate": 2.195491724673538e-05, + "loss": 0.4452, + "step": 20826 + }, + { + "epoch": 1.3977718868494347, + "grad_norm": 1.079972505569458, + "learning_rate": 2.194592022931391e-05, + "loss": 0.498, + "step": 20828 + }, + { + "epoch": 1.3979061105332036, + "grad_norm": 0.9583383202552795, + "learning_rate": 2.1936924537392527e-05, + "loss": 0.4416, + "step": 20830 + }, + { + "epoch": 1.3980403342169727, + "grad_norm": 1.1776984930038452, + "learning_rate": 2.1927930171396293e-05, + "loss": 0.5538, + "step": 20832 + }, + { + "epoch": 1.3981745579007416, + "grad_norm": 1.0068161487579346, + "learning_rate": 2.1918937131750157e-05, + "loss": 0.5406, + "step": 20834 + }, + { + "epoch": 1.3983087815845106, + "grad_norm": 1.1186065673828125, + "learning_rate": 2.1909945418879017e-05, + "loss": 0.4743, + "step": 20836 + }, + { + "epoch": 1.3984430052682795, + "grad_norm": 1.0533246994018555, + "learning_rate": 2.190095503320771e-05, + "loss": 0.5541, + "step": 20838 + }, + { + "epoch": 1.3985772289520486, + "grad_norm": 0.8219181299209595, + "learning_rate": 2.1891965975161037e-05, + "loss": 0.4716, + "step": 20840 + }, + { + "epoch": 1.3987114526358175, + "grad_norm": 1.1751853227615356, + "learning_rate": 2.1882978245163705e-05, + "loss": 0.4849, + "step": 20842 + }, + { + "epoch": 1.3988456763195867, + "grad_norm": 0.9777979254722595, + "learning_rate": 2.187399184364035e-05, + "loss": 0.4917, + "step": 20844 + }, + { + "epoch": 1.3989799000033556, + "grad_norm": 1.089253306388855, + "learning_rate": 2.1865006771015567e-05, + "loss": 0.4869, + "step": 20846 + }, + { + "epoch": 1.3991141236871245, + "grad_norm": 0.8422405123710632, + "learning_rate": 2.18560230277139e-05, + "loss": 0.4439, + "step": 20848 + }, + { + "epoch": 1.3992483473708937, + "grad_norm": 1.0621827840805054, + "learning_rate": 2.1847040614159804e-05, + "loss": 0.527, + "step": 20850 + }, + { + "epoch": 1.3993825710546626, + "grad_norm": 1.1477723121643066, + "learning_rate": 2.1838059530777677e-05, + "loss": 0.476, + "step": 20852 + }, + { + "epoch": 1.3995167947384317, + "grad_norm": 0.860033392906189, + "learning_rate": 2.1829079777991844e-05, + "loss": 0.481, + "step": 20854 + }, + { + "epoch": 1.3996510184222006, + "grad_norm": 0.8550235033035278, + "learning_rate": 2.1820101356226612e-05, + "loss": 0.466, + "step": 20856 + }, + { + "epoch": 1.3997852421059696, + "grad_norm": 0.9906622767448425, + "learning_rate": 2.1811124265906157e-05, + "loss": 0.4855, + "step": 20858 + }, + { + "epoch": 1.3999194657897385, + "grad_norm": 0.8928864002227783, + "learning_rate": 2.180214850745467e-05, + "loss": 0.5128, + "step": 20860 + }, + { + "epoch": 1.4000536894735076, + "grad_norm": 0.9970691800117493, + "learning_rate": 2.179317408129622e-05, + "loss": 0.5149, + "step": 20862 + }, + { + "epoch": 1.4001879131572765, + "grad_norm": 0.9386903047561646, + "learning_rate": 2.1784200987854825e-05, + "loss": 0.4421, + "step": 20864 + }, + { + "epoch": 1.4003221368410457, + "grad_norm": 0.8827176094055176, + "learning_rate": 2.1775229227554433e-05, + "loss": 0.4363, + "step": 20866 + }, + { + "epoch": 1.4004563605248146, + "grad_norm": 0.9380989670753479, + "learning_rate": 2.1766258800818972e-05, + "loss": 0.5722, + "step": 20868 + }, + { + "epoch": 1.4005905842085835, + "grad_norm": 1.1577523946762085, + "learning_rate": 2.1757289708072272e-05, + "loss": 0.4723, + "step": 20870 + }, + { + "epoch": 1.4007248078923527, + "grad_norm": 0.931845486164093, + "learning_rate": 2.1748321949738088e-05, + "loss": 0.4809, + "step": 20872 + }, + { + "epoch": 1.4008590315761216, + "grad_norm": 1.0608102083206177, + "learning_rate": 2.1739355526240124e-05, + "loss": 0.5016, + "step": 20874 + }, + { + "epoch": 1.4009932552598907, + "grad_norm": 1.084637999534607, + "learning_rate": 2.173039043800206e-05, + "loss": 0.4945, + "step": 20876 + }, + { + "epoch": 1.4011274789436596, + "grad_norm": 1.0683140754699707, + "learning_rate": 2.172142668544745e-05, + "loss": 0.4652, + "step": 20878 + }, + { + "epoch": 1.4012617026274286, + "grad_norm": 1.0307785272598267, + "learning_rate": 2.1712464268999826e-05, + "loss": 0.4868, + "step": 20880 + }, + { + "epoch": 1.4013959263111977, + "grad_norm": 1.1262508630752563, + "learning_rate": 2.1703503189082634e-05, + "loss": 0.5312, + "step": 20882 + }, + { + "epoch": 1.4015301499949666, + "grad_norm": 0.9825021028518677, + "learning_rate": 2.1694543446119265e-05, + "loss": 0.4495, + "step": 20884 + }, + { + "epoch": 1.4016643736787355, + "grad_norm": 0.9609806537628174, + "learning_rate": 2.1685585040533075e-05, + "loss": 0.5075, + "step": 20886 + }, + { + "epoch": 1.4017985973625047, + "grad_norm": 1.0035909414291382, + "learning_rate": 2.167662797274729e-05, + "loss": 0.5096, + "step": 20888 + }, + { + "epoch": 1.4019328210462736, + "grad_norm": 0.925609290599823, + "learning_rate": 2.1667672243185178e-05, + "loss": 0.4324, + "step": 20890 + }, + { + "epoch": 1.4020670447300425, + "grad_norm": 0.98987877368927, + "learning_rate": 2.16587178522698e-05, + "loss": 0.4612, + "step": 20892 + }, + { + "epoch": 1.4022012684138117, + "grad_norm": 0.9895521998405457, + "learning_rate": 2.1649764800424294e-05, + "loss": 0.5452, + "step": 20894 + }, + { + "epoch": 1.4023354920975806, + "grad_norm": 1.0596760511398315, + "learning_rate": 2.1640813088071633e-05, + "loss": 0.5384, + "step": 20896 + }, + { + "epoch": 1.4024697157813497, + "grad_norm": 0.9556659460067749, + "learning_rate": 2.163186271563481e-05, + "loss": 0.4398, + "step": 20898 + }, + { + "epoch": 1.4026039394651186, + "grad_norm": 1.061688780784607, + "learning_rate": 2.1622913683536684e-05, + "loss": 0.4534, + "step": 20900 + }, + { + "epoch": 1.4027381631488876, + "grad_norm": 1.1543545722961426, + "learning_rate": 2.1613965992200087e-05, + "loss": 0.4923, + "step": 20902 + }, + { + "epoch": 1.4028723868326567, + "grad_norm": 1.1429158449172974, + "learning_rate": 2.1605019642047765e-05, + "loss": 0.5098, + "step": 20904 + }, + { + "epoch": 1.4030066105164256, + "grad_norm": 1.0910015106201172, + "learning_rate": 2.159607463350244e-05, + "loss": 0.4345, + "step": 20906 + }, + { + "epoch": 1.4031408342001948, + "grad_norm": 1.1227449178695679, + "learning_rate": 2.1587130966986735e-05, + "loss": 0.4461, + "step": 20908 + }, + { + "epoch": 1.4032750578839637, + "grad_norm": 1.3753069639205933, + "learning_rate": 2.1578188642923225e-05, + "loss": 0.5575, + "step": 20910 + }, + { + "epoch": 1.4034092815677326, + "grad_norm": 1.1705611944198608, + "learning_rate": 2.1569247661734392e-05, + "loss": 0.5633, + "step": 20912 + }, + { + "epoch": 1.4035435052515015, + "grad_norm": 0.9577829241752625, + "learning_rate": 2.1560308023842717e-05, + "loss": 0.4806, + "step": 20914 + }, + { + "epoch": 1.4036777289352707, + "grad_norm": 1.0731220245361328, + "learning_rate": 2.1551369729670546e-05, + "loss": 0.4483, + "step": 20916 + }, + { + "epoch": 1.4038119526190396, + "grad_norm": 0.9627668857574463, + "learning_rate": 2.1542432779640255e-05, + "loss": 0.4701, + "step": 20918 + }, + { + "epoch": 1.4039461763028087, + "grad_norm": 1.0893503427505493, + "learning_rate": 2.153349717417401e-05, + "loss": 0.4905, + "step": 20920 + }, + { + "epoch": 1.4040803999865776, + "grad_norm": 0.9726977348327637, + "learning_rate": 2.1524562913694074e-05, + "loss": 0.4862, + "step": 20922 + }, + { + "epoch": 1.4042146236703466, + "grad_norm": 0.8646559119224548, + "learning_rate": 2.151562999862252e-05, + "loss": 0.4984, + "step": 20924 + }, + { + "epoch": 1.4043488473541157, + "grad_norm": 1.0155696868896484, + "learning_rate": 2.150669842938146e-05, + "loss": 0.5444, + "step": 20926 + }, + { + "epoch": 1.4044830710378846, + "grad_norm": 1.0304315090179443, + "learning_rate": 2.1497768206392864e-05, + "loss": 0.5069, + "step": 20928 + }, + { + "epoch": 1.4046172947216538, + "grad_norm": 0.9800500869750977, + "learning_rate": 2.148883933007868e-05, + "loss": 0.4949, + "step": 20930 + }, + { + "epoch": 1.4047515184054227, + "grad_norm": 0.8693739771842957, + "learning_rate": 2.1479911800860752e-05, + "loss": 0.4786, + "step": 20932 + }, + { + "epoch": 1.4048857420891916, + "grad_norm": 1.0007768869400024, + "learning_rate": 2.1470985619160926e-05, + "loss": 0.444, + "step": 20934 + }, + { + "epoch": 1.4050199657729605, + "grad_norm": 0.8693901300430298, + "learning_rate": 2.146206078540093e-05, + "loss": 0.4498, + "step": 20936 + }, + { + "epoch": 1.4051541894567297, + "grad_norm": 0.9130423665046692, + "learning_rate": 2.1453137300002445e-05, + "loss": 0.4634, + "step": 20938 + }, + { + "epoch": 1.4052884131404986, + "grad_norm": 1.1779518127441406, + "learning_rate": 2.1444215163387067e-05, + "loss": 0.5016, + "step": 20940 + }, + { + "epoch": 1.4054226368242677, + "grad_norm": 1.0261186361312866, + "learning_rate": 2.143529437597639e-05, + "loss": 0.524, + "step": 20942 + }, + { + "epoch": 1.4055568605080366, + "grad_norm": 1.036157488822937, + "learning_rate": 2.1426374938191884e-05, + "loss": 0.5194, + "step": 20944 + }, + { + "epoch": 1.4056910841918056, + "grad_norm": 1.007396936416626, + "learning_rate": 2.1417456850454976e-05, + "loss": 0.4604, + "step": 20946 + }, + { + "epoch": 1.4058253078755747, + "grad_norm": 0.9180713295936584, + "learning_rate": 2.1408540113187027e-05, + "loss": 0.4729, + "step": 20948 + }, + { + "epoch": 1.4059595315593436, + "grad_norm": 1.113799810409546, + "learning_rate": 2.139962472680931e-05, + "loss": 0.4736, + "step": 20950 + }, + { + "epoch": 1.4060937552431128, + "grad_norm": 1.0503449440002441, + "learning_rate": 2.139071069174311e-05, + "loss": 0.4933, + "step": 20952 + }, + { + "epoch": 1.4062279789268817, + "grad_norm": 1.029120922088623, + "learning_rate": 2.1381798008409547e-05, + "loss": 0.4722, + "step": 20954 + }, + { + "epoch": 1.4063622026106506, + "grad_norm": 1.1079305410385132, + "learning_rate": 2.1372886677229775e-05, + "loss": 0.4751, + "step": 20956 + }, + { + "epoch": 1.4064964262944197, + "grad_norm": 1.1420440673828125, + "learning_rate": 2.1363976698624815e-05, + "loss": 0.5038, + "step": 20958 + }, + { + "epoch": 1.4066306499781887, + "grad_norm": 1.000579595565796, + "learning_rate": 2.135506807301565e-05, + "loss": 0.4933, + "step": 20960 + }, + { + "epoch": 1.4067648736619576, + "grad_norm": 0.9605050683021545, + "learning_rate": 2.1346160800823166e-05, + "loss": 0.4728, + "step": 20962 + }, + { + "epoch": 1.4068990973457267, + "grad_norm": 1.0144368410110474, + "learning_rate": 2.133725488246826e-05, + "loss": 0.4661, + "step": 20964 + }, + { + "epoch": 1.4070333210294956, + "grad_norm": 0.853814423084259, + "learning_rate": 2.13283503183717e-05, + "loss": 0.4718, + "step": 20966 + }, + { + "epoch": 1.4071675447132646, + "grad_norm": 0.8243944644927979, + "learning_rate": 2.13194471089542e-05, + "loss": 0.451, + "step": 20968 + }, + { + "epoch": 1.4073017683970337, + "grad_norm": 0.9485942125320435, + "learning_rate": 2.1310545254636412e-05, + "loss": 0.4761, + "step": 20970 + }, + { + "epoch": 1.4074359920808026, + "grad_norm": 0.95245361328125, + "learning_rate": 2.130164475583896e-05, + "loss": 0.5054, + "step": 20972 + }, + { + "epoch": 1.4075702157645718, + "grad_norm": 1.0146323442459106, + "learning_rate": 2.129274561298237e-05, + "loss": 0.5032, + "step": 20974 + }, + { + "epoch": 1.4077044394483407, + "grad_norm": 1.0180444717407227, + "learning_rate": 2.1283847826487092e-05, + "loss": 0.5308, + "step": 20976 + }, + { + "epoch": 1.4078386631321096, + "grad_norm": 0.9613975882530212, + "learning_rate": 2.1274951396773517e-05, + "loss": 0.4744, + "step": 20978 + }, + { + "epoch": 1.4079728868158787, + "grad_norm": 1.1436967849731445, + "learning_rate": 2.1266056324262023e-05, + "loss": 0.5142, + "step": 20980 + }, + { + "epoch": 1.4081071104996477, + "grad_norm": 0.9791481494903564, + "learning_rate": 2.125716260937285e-05, + "loss": 0.4777, + "step": 20982 + }, + { + "epoch": 1.4082413341834168, + "grad_norm": 1.1079564094543457, + "learning_rate": 2.1248270252526237e-05, + "loss": 0.4608, + "step": 20984 + }, + { + "epoch": 1.4083755578671857, + "grad_norm": 1.0518410205841064, + "learning_rate": 2.1239379254142323e-05, + "loss": 0.4951, + "step": 20986 + }, + { + "epoch": 1.4085097815509546, + "grad_norm": 0.8152596950531006, + "learning_rate": 2.1230489614641174e-05, + "loss": 0.4868, + "step": 20988 + }, + { + "epoch": 1.4086440052347236, + "grad_norm": 1.0190281867980957, + "learning_rate": 2.122160133444281e-05, + "loss": 0.5007, + "step": 20990 + }, + { + "epoch": 1.4087782289184927, + "grad_norm": 1.1111348867416382, + "learning_rate": 2.1212714413967206e-05, + "loss": 0.5638, + "step": 20992 + }, + { + "epoch": 1.4089124526022616, + "grad_norm": 1.0240122079849243, + "learning_rate": 2.120382885363424e-05, + "loss": 0.4973, + "step": 20994 + }, + { + "epoch": 1.4090466762860308, + "grad_norm": 1.0007127523422241, + "learning_rate": 2.119494465386373e-05, + "loss": 0.4804, + "step": 20996 + }, + { + "epoch": 1.4091808999697997, + "grad_norm": 0.9426189064979553, + "learning_rate": 2.1186061815075424e-05, + "loss": 0.5261, + "step": 20998 + }, + { + "epoch": 1.4093151236535686, + "grad_norm": 0.9891473054885864, + "learning_rate": 2.117718033768906e-05, + "loss": 0.5182, + "step": 21000 + }, + { + "epoch": 1.4094493473373377, + "grad_norm": 1.1265980005264282, + "learning_rate": 2.1168300222124248e-05, + "loss": 0.5239, + "step": 21002 + }, + { + "epoch": 1.4095835710211067, + "grad_norm": 0.9827826023101807, + "learning_rate": 2.1159421468800556e-05, + "loss": 0.4719, + "step": 21004 + }, + { + "epoch": 1.4097177947048758, + "grad_norm": 1.1680190563201904, + "learning_rate": 2.115054407813747e-05, + "loss": 0.5362, + "step": 21006 + }, + { + "epoch": 1.4098520183886447, + "grad_norm": 1.0358234643936157, + "learning_rate": 2.1141668050554457e-05, + "loss": 0.5439, + "step": 21008 + }, + { + "epoch": 1.4099862420724136, + "grad_norm": 1.0975145101547241, + "learning_rate": 2.1132793386470867e-05, + "loss": 0.4816, + "step": 21010 + }, + { + "epoch": 1.4101204657561826, + "grad_norm": 0.9598863124847412, + "learning_rate": 2.112392008630604e-05, + "loss": 0.5637, + "step": 21012 + }, + { + "epoch": 1.4102546894399517, + "grad_norm": 1.0278030633926392, + "learning_rate": 2.1115048150479233e-05, + "loss": 0.4563, + "step": 21014 + }, + { + "epoch": 1.4103889131237206, + "grad_norm": 1.017699956893921, + "learning_rate": 2.110617757940956e-05, + "loss": 0.5587, + "step": 21016 + }, + { + "epoch": 1.4105231368074898, + "grad_norm": 0.8422526121139526, + "learning_rate": 2.1097308373516194e-05, + "loss": 0.4379, + "step": 21018 + }, + { + "epoch": 1.4106573604912587, + "grad_norm": 0.9482269287109375, + "learning_rate": 2.1088440533218156e-05, + "loss": 0.419, + "step": 21020 + }, + { + "epoch": 1.4107915841750276, + "grad_norm": 1.0598201751708984, + "learning_rate": 2.107957405893447e-05, + "loss": 0.4747, + "step": 21022 + }, + { + "epoch": 1.4109258078587967, + "grad_norm": 1.0010422468185425, + "learning_rate": 2.107070895108404e-05, + "loss": 0.4989, + "step": 21024 + }, + { + "epoch": 1.4110600315425657, + "grad_norm": 0.8625784516334534, + "learning_rate": 2.1061845210085728e-05, + "loss": 0.4603, + "step": 21026 + }, + { + "epoch": 1.4111942552263348, + "grad_norm": 1.1818901300430298, + "learning_rate": 2.1052982836358303e-05, + "loss": 0.4763, + "step": 21028 + }, + { + "epoch": 1.4113284789101037, + "grad_norm": 1.0573731660842896, + "learning_rate": 2.1044121830320535e-05, + "loss": 0.4394, + "step": 21030 + }, + { + "epoch": 1.4114627025938726, + "grad_norm": 1.0304356813430786, + "learning_rate": 2.1035262192391077e-05, + "loss": 0.4655, + "step": 21032 + }, + { + "epoch": 1.4115969262776418, + "grad_norm": 0.9857569336891174, + "learning_rate": 2.102640392298852e-05, + "loss": 0.4512, + "step": 21034 + }, + { + "epoch": 1.4117311499614107, + "grad_norm": 1.1267962455749512, + "learning_rate": 2.101754702253138e-05, + "loss": 0.5103, + "step": 21036 + }, + { + "epoch": 1.4118653736451796, + "grad_norm": 0.9692539572715759, + "learning_rate": 2.1008691491438175e-05, + "loss": 0.5152, + "step": 21038 + }, + { + "epoch": 1.4119995973289488, + "grad_norm": 1.0739535093307495, + "learning_rate": 2.099983733012727e-05, + "loss": 0.4613, + "step": 21040 + }, + { + "epoch": 1.4121338210127177, + "grad_norm": 1.01617431640625, + "learning_rate": 2.0990984539017062e-05, + "loss": 0.5276, + "step": 21042 + }, + { + "epoch": 1.4122680446964866, + "grad_norm": 1.0768928527832031, + "learning_rate": 2.0982133118525755e-05, + "loss": 0.4781, + "step": 21044 + }, + { + "epoch": 1.4124022683802557, + "grad_norm": 1.089753270149231, + "learning_rate": 2.0973283069071608e-05, + "loss": 0.5187, + "step": 21046 + }, + { + "epoch": 1.4125364920640247, + "grad_norm": 1.1406461000442505, + "learning_rate": 2.0964434391072745e-05, + "loss": 0.4936, + "step": 21048 + }, + { + "epoch": 1.4126707157477938, + "grad_norm": 1.0846813917160034, + "learning_rate": 2.0955587084947282e-05, + "loss": 0.4411, + "step": 21050 + }, + { + "epoch": 1.4128049394315627, + "grad_norm": 0.9188988208770752, + "learning_rate": 2.0946741151113215e-05, + "loss": 0.4442, + "step": 21052 + }, + { + "epoch": 1.4129391631153316, + "grad_norm": 1.0002262592315674, + "learning_rate": 2.0937896589988494e-05, + "loss": 0.4859, + "step": 21054 + }, + { + "epoch": 1.4130733867991008, + "grad_norm": 0.9710934162139893, + "learning_rate": 2.0929053401990995e-05, + "loss": 0.5224, + "step": 21056 + }, + { + "epoch": 1.4132076104828697, + "grad_norm": 0.9674138426780701, + "learning_rate": 2.0920211587538573e-05, + "loss": 0.5951, + "step": 21058 + }, + { + "epoch": 1.4133418341666388, + "grad_norm": 1.0755289793014526, + "learning_rate": 2.091137114704897e-05, + "loss": 0.4637, + "step": 21060 + }, + { + "epoch": 1.4134760578504078, + "grad_norm": 0.8497282266616821, + "learning_rate": 2.0902532080939886e-05, + "loss": 0.4851, + "step": 21062 + }, + { + "epoch": 1.4136102815341767, + "grad_norm": 0.9734933972358704, + "learning_rate": 2.089369438962892e-05, + "loss": 0.4962, + "step": 21064 + }, + { + "epoch": 1.4137445052179456, + "grad_norm": 0.9598963856697083, + "learning_rate": 2.0884858073533674e-05, + "loss": 0.5254, + "step": 21066 + }, + { + "epoch": 1.4138787289017147, + "grad_norm": 1.0404390096664429, + "learning_rate": 2.087602313307162e-05, + "loss": 0.5203, + "step": 21068 + }, + { + "epoch": 1.4140129525854837, + "grad_norm": 0.8939399719238281, + "learning_rate": 2.086718956866024e-05, + "loss": 0.5101, + "step": 21070 + }, + { + "epoch": 1.4141471762692528, + "grad_norm": 0.8685775399208069, + "learning_rate": 2.0858357380716826e-05, + "loss": 0.4528, + "step": 21072 + }, + { + "epoch": 1.4142813999530217, + "grad_norm": 1.0026849508285522, + "learning_rate": 2.084952656965874e-05, + "loss": 0.5312, + "step": 21074 + }, + { + "epoch": 1.4144156236367906, + "grad_norm": 1.0558052062988281, + "learning_rate": 2.084069713590318e-05, + "loss": 0.4603, + "step": 21076 + }, + { + "epoch": 1.4145498473205598, + "grad_norm": 1.0284819602966309, + "learning_rate": 2.0831869079867368e-05, + "loss": 0.4986, + "step": 21078 + }, + { + "epoch": 1.4146840710043287, + "grad_norm": 1.1609539985656738, + "learning_rate": 2.0823042401968374e-05, + "loss": 0.504, + "step": 21080 + }, + { + "epoch": 1.4148182946880978, + "grad_norm": 1.082035779953003, + "learning_rate": 2.0814217102623264e-05, + "loss": 0.5068, + "step": 21082 + }, + { + "epoch": 1.4149525183718668, + "grad_norm": 1.0110067129135132, + "learning_rate": 2.0805393182248995e-05, + "loss": 0.4733, + "step": 21084 + }, + { + "epoch": 1.4150867420556357, + "grad_norm": 1.0480561256408691, + "learning_rate": 2.0796570641262476e-05, + "loss": 0.4388, + "step": 21086 + }, + { + "epoch": 1.4152209657394046, + "grad_norm": 0.9707655310630798, + "learning_rate": 2.0787749480080587e-05, + "loss": 0.4946, + "step": 21088 + }, + { + "epoch": 1.4153551894231737, + "grad_norm": 0.9844449162483215, + "learning_rate": 2.0778929699120088e-05, + "loss": 0.5224, + "step": 21090 + }, + { + "epoch": 1.4154894131069427, + "grad_norm": 2.456658124923706, + "learning_rate": 2.0770111298797702e-05, + "loss": 0.5376, + "step": 21092 + }, + { + "epoch": 1.4156236367907118, + "grad_norm": 1.0447827577590942, + "learning_rate": 2.076129427953006e-05, + "loss": 0.5156, + "step": 21094 + }, + { + "epoch": 1.4157578604744807, + "grad_norm": 1.0012128353118896, + "learning_rate": 2.0752478641733785e-05, + "loss": 0.4915, + "step": 21096 + }, + { + "epoch": 1.4158920841582496, + "grad_norm": 1.5245203971862793, + "learning_rate": 2.0743664385825385e-05, + "loss": 0.6781, + "step": 21098 + }, + { + "epoch": 1.4160263078420188, + "grad_norm": 1.121191382408142, + "learning_rate": 2.073485151222131e-05, + "loss": 0.5262, + "step": 21100 + }, + { + "epoch": 1.4161605315257877, + "grad_norm": 0.9988177418708801, + "learning_rate": 2.072604002133794e-05, + "loss": 0.4631, + "step": 21102 + }, + { + "epoch": 1.4162947552095568, + "grad_norm": 1.0113762617111206, + "learning_rate": 2.071722991359163e-05, + "loss": 0.4545, + "step": 21104 + }, + { + "epoch": 1.4164289788933258, + "grad_norm": 0.9512695670127869, + "learning_rate": 2.0708421189398604e-05, + "loss": 0.4821, + "step": 21106 + }, + { + "epoch": 1.4165632025770947, + "grad_norm": 1.1240195035934448, + "learning_rate": 2.0699613849175114e-05, + "loss": 0.4688, + "step": 21108 + }, + { + "epoch": 1.4166974262608638, + "grad_norm": 0.9319685101509094, + "learning_rate": 2.0690807893337217e-05, + "loss": 0.4531, + "step": 21110 + }, + { + "epoch": 1.4168316499446327, + "grad_norm": 0.9338365197181702, + "learning_rate": 2.068200332230103e-05, + "loss": 0.4689, + "step": 21112 + }, + { + "epoch": 1.4169658736284017, + "grad_norm": 1.0448791980743408, + "learning_rate": 2.0673200136482518e-05, + "loss": 0.461, + "step": 21114 + }, + { + "epoch": 1.4171000973121708, + "grad_norm": 1.1320430040359497, + "learning_rate": 2.0664398336297646e-05, + "loss": 0.5149, + "step": 21116 + }, + { + "epoch": 1.4172343209959397, + "grad_norm": 0.934168815612793, + "learning_rate": 2.0655597922162274e-05, + "loss": 0.4847, + "step": 21118 + }, + { + "epoch": 1.4173685446797086, + "grad_norm": 1.030619740486145, + "learning_rate": 2.064679889449219e-05, + "loss": 0.4648, + "step": 21120 + }, + { + "epoch": 1.4175027683634778, + "grad_norm": 1.2354577779769897, + "learning_rate": 2.063800125370312e-05, + "loss": 0.4735, + "step": 21122 + }, + { + "epoch": 1.4176369920472467, + "grad_norm": 0.9623546600341797, + "learning_rate": 2.062920500021078e-05, + "loss": 0.5488, + "step": 21124 + }, + { + "epoch": 1.4177712157310158, + "grad_norm": 0.9811288118362427, + "learning_rate": 2.062041013443074e-05, + "loss": 0.509, + "step": 21126 + }, + { + "epoch": 1.4179054394147848, + "grad_norm": 1.1046497821807861, + "learning_rate": 2.0611616656778547e-05, + "loss": 0.4268, + "step": 21128 + }, + { + "epoch": 1.4180396630985537, + "grad_norm": 1.0107942819595337, + "learning_rate": 2.0602824567669664e-05, + "loss": 0.5002, + "step": 21130 + }, + { + "epoch": 1.4181738867823228, + "grad_norm": 1.008588194847107, + "learning_rate": 2.059403386751953e-05, + "loss": 0.4647, + "step": 21132 + }, + { + "epoch": 1.4183081104660917, + "grad_norm": 0.8859527111053467, + "learning_rate": 2.0585244556743448e-05, + "loss": 0.4628, + "step": 21134 + }, + { + "epoch": 1.4184423341498609, + "grad_norm": 1.128159523010254, + "learning_rate": 2.0576456635756762e-05, + "loss": 0.5023, + "step": 21136 + }, + { + "epoch": 1.4185765578336298, + "grad_norm": 0.9821992516517639, + "learning_rate": 2.0567670104974607e-05, + "loss": 0.5079, + "step": 21138 + }, + { + "epoch": 1.4187107815173987, + "grad_norm": 1.0721951723098755, + "learning_rate": 2.0558884964812174e-05, + "loss": 0.5065, + "step": 21140 + }, + { + "epoch": 1.4188450052011676, + "grad_norm": 0.9656804203987122, + "learning_rate": 2.055010121568452e-05, + "loss": 0.4979, + "step": 21142 + }, + { + "epoch": 1.4189792288849368, + "grad_norm": 0.9861791133880615, + "learning_rate": 2.0541318858006685e-05, + "loss": 0.4696, + "step": 21144 + }, + { + "epoch": 1.4191134525687057, + "grad_norm": 0.8539104461669922, + "learning_rate": 2.0532537892193615e-05, + "loss": 0.4893, + "step": 21146 + }, + { + "epoch": 1.4192476762524748, + "grad_norm": 1.1190543174743652, + "learning_rate": 2.052375831866018e-05, + "loss": 0.5314, + "step": 21148 + }, + { + "epoch": 1.4193818999362438, + "grad_norm": 1.093247652053833, + "learning_rate": 2.05149801378212e-05, + "loss": 0.493, + "step": 21150 + }, + { + "epoch": 1.4195161236200127, + "grad_norm": 0.9135884642601013, + "learning_rate": 2.0506203350091414e-05, + "loss": 0.4278, + "step": 21152 + }, + { + "epoch": 1.4196503473037818, + "grad_norm": 1.005814790725708, + "learning_rate": 2.0497427955885533e-05, + "loss": 0.5174, + "step": 21154 + }, + { + "epoch": 1.4197845709875507, + "grad_norm": 1.111291527748108, + "learning_rate": 2.0488653955618174e-05, + "loss": 0.4829, + "step": 21156 + }, + { + "epoch": 1.4199187946713199, + "grad_norm": 0.9510667324066162, + "learning_rate": 2.0479881349703883e-05, + "loss": 0.4606, + "step": 21158 + }, + { + "epoch": 1.4200530183550888, + "grad_norm": 1.0444616079330444, + "learning_rate": 2.047111013855713e-05, + "loss": 0.4972, + "step": 21160 + }, + { + "epoch": 1.4201872420388577, + "grad_norm": 1.0435190200805664, + "learning_rate": 2.0462340322592378e-05, + "loss": 0.4813, + "step": 21162 + }, + { + "epoch": 1.4203214657226266, + "grad_norm": 0.8452038764953613, + "learning_rate": 2.0453571902223938e-05, + "loss": 0.4287, + "step": 21164 + }, + { + "epoch": 1.4204556894063958, + "grad_norm": 0.8310977816581726, + "learning_rate": 2.044480487786617e-05, + "loss": 0.3898, + "step": 21166 + }, + { + "epoch": 1.4205899130901647, + "grad_norm": 1.0622636079788208, + "learning_rate": 2.0436039249933208e-05, + "loss": 0.4817, + "step": 21168 + }, + { + "epoch": 1.4207241367739338, + "grad_norm": 1.041193962097168, + "learning_rate": 2.042727501883928e-05, + "loss": 0.5178, + "step": 21170 + }, + { + "epoch": 1.4208583604577028, + "grad_norm": 1.0471394062042236, + "learning_rate": 2.041851218499844e-05, + "loss": 0.5063, + "step": 21172 + }, + { + "epoch": 1.4209925841414717, + "grad_norm": 1.1910574436187744, + "learning_rate": 2.040975074882474e-05, + "loss": 0.5231, + "step": 21174 + }, + { + "epoch": 1.4211268078252408, + "grad_norm": 1.1655224561691284, + "learning_rate": 2.040099071073214e-05, + "loss": 0.5138, + "step": 21176 + }, + { + "epoch": 1.4212610315090097, + "grad_norm": 1.1665756702423096, + "learning_rate": 2.0392232071134517e-05, + "loss": 0.4695, + "step": 21178 + }, + { + "epoch": 1.4213952551927789, + "grad_norm": 1.0353599786758423, + "learning_rate": 2.0383474830445692e-05, + "loss": 0.481, + "step": 21180 + }, + { + "epoch": 1.4215294788765478, + "grad_norm": 1.0987224578857422, + "learning_rate": 2.037471898907946e-05, + "loss": 0.4692, + "step": 21182 + }, + { + "epoch": 1.4216637025603167, + "grad_norm": 1.1034603118896484, + "learning_rate": 2.0365964547449502e-05, + "loss": 0.4786, + "step": 21184 + }, + { + "epoch": 1.4217979262440856, + "grad_norm": 0.8870950937271118, + "learning_rate": 2.035721150596944e-05, + "loss": 0.4515, + "step": 21186 + }, + { + "epoch": 1.4219321499278548, + "grad_norm": 0.970938503742218, + "learning_rate": 2.0348459865052832e-05, + "loss": 0.5585, + "step": 21188 + }, + { + "epoch": 1.4220663736116237, + "grad_norm": 1.1109589338302612, + "learning_rate": 2.0339709625113208e-05, + "loss": 0.4898, + "step": 21190 + }, + { + "epoch": 1.4222005972953928, + "grad_norm": 1.042851448059082, + "learning_rate": 2.0330960786563963e-05, + "loss": 0.4403, + "step": 21192 + }, + { + "epoch": 1.4223348209791618, + "grad_norm": 1.2549662590026855, + "learning_rate": 2.0322213349818507e-05, + "loss": 0.4719, + "step": 21194 + }, + { + "epoch": 1.4224690446629307, + "grad_norm": 0.9628381729125977, + "learning_rate": 2.031346731529008e-05, + "loss": 0.4539, + "step": 21196 + }, + { + "epoch": 1.4226032683466998, + "grad_norm": 1.042144775390625, + "learning_rate": 2.0304722683391964e-05, + "loss": 0.4781, + "step": 21198 + }, + { + "epoch": 1.4227374920304687, + "grad_norm": 1.130577564239502, + "learning_rate": 2.029597945453729e-05, + "loss": 0.5006, + "step": 21200 + }, + { + "epoch": 1.4228717157142379, + "grad_norm": 1.0895018577575684, + "learning_rate": 2.0287237629139192e-05, + "loss": 0.493, + "step": 21202 + }, + { + "epoch": 1.4230059393980068, + "grad_norm": 0.9593971967697144, + "learning_rate": 2.0278497207610697e-05, + "loss": 0.4971, + "step": 21204 + }, + { + "epoch": 1.4231401630817757, + "grad_norm": 1.0314475297927856, + "learning_rate": 2.0269758190364758e-05, + "loss": 0.4966, + "step": 21206 + }, + { + "epoch": 1.4232743867655449, + "grad_norm": 1.0766124725341797, + "learning_rate": 2.026102057781426e-05, + "loss": 0.4696, + "step": 21208 + }, + { + "epoch": 1.4234086104493138, + "grad_norm": 0.9447959065437317, + "learning_rate": 2.0252284370372087e-05, + "loss": 0.4751, + "step": 21210 + }, + { + "epoch": 1.423542834133083, + "grad_norm": 0.9836453199386597, + "learning_rate": 2.0243549568450974e-05, + "loss": 0.5005, + "step": 21212 + }, + { + "epoch": 1.4236770578168518, + "grad_norm": 1.0962803363800049, + "learning_rate": 2.023481617246363e-05, + "loss": 0.4753, + "step": 21214 + }, + { + "epoch": 1.4238112815006208, + "grad_norm": 1.0244379043579102, + "learning_rate": 2.022608418282268e-05, + "loss": 0.4903, + "step": 21216 + }, + { + "epoch": 1.4239455051843897, + "grad_norm": 0.9502069354057312, + "learning_rate": 2.021735359994072e-05, + "loss": 0.4726, + "step": 21218 + }, + { + "epoch": 1.4240797288681588, + "grad_norm": 1.0028274059295654, + "learning_rate": 2.020862442423024e-05, + "loss": 0.4824, + "step": 21220 + }, + { + "epoch": 1.4242139525519277, + "grad_norm": 1.1517637968063354, + "learning_rate": 2.0199896656103666e-05, + "loss": 0.498, + "step": 21222 + }, + { + "epoch": 1.4243481762356969, + "grad_norm": 0.9527151584625244, + "learning_rate": 2.0191170295973382e-05, + "loss": 0.4545, + "step": 21224 + }, + { + "epoch": 1.4244823999194658, + "grad_norm": 1.1030168533325195, + "learning_rate": 2.0182445344251667e-05, + "loss": 0.5273, + "step": 21226 + }, + { + "epoch": 1.4246166236032347, + "grad_norm": 0.9871190786361694, + "learning_rate": 2.0173721801350787e-05, + "loss": 0.4946, + "step": 21228 + }, + { + "epoch": 1.4247508472870039, + "grad_norm": 1.0557881593704224, + "learning_rate": 2.0164999667682887e-05, + "loss": 0.5172, + "step": 21230 + }, + { + "epoch": 1.4248850709707728, + "grad_norm": 1.0631259679794312, + "learning_rate": 2.0156278943660123e-05, + "loss": 0.4653, + "step": 21232 + }, + { + "epoch": 1.425019294654542, + "grad_norm": 0.9954543113708496, + "learning_rate": 2.0147559629694455e-05, + "loss": 0.4941, + "step": 21234 + }, + { + "epoch": 1.4251535183383108, + "grad_norm": 0.9826318621635437, + "learning_rate": 2.013884172619791e-05, + "loss": 0.4995, + "step": 21236 + }, + { + "epoch": 1.4252877420220798, + "grad_norm": 1.0855181217193604, + "learning_rate": 2.013012523358236e-05, + "loss": 0.4978, + "step": 21238 + }, + { + "epoch": 1.4254219657058487, + "grad_norm": 0.9283624291419983, + "learning_rate": 2.012141015225967e-05, + "loss": 0.4722, + "step": 21240 + }, + { + "epoch": 1.4255561893896178, + "grad_norm": 1.0445226430892944, + "learning_rate": 2.0112696482641596e-05, + "loss": 0.5212, + "step": 21242 + }, + { + "epoch": 1.4256904130733867, + "grad_norm": 1.0525439977645874, + "learning_rate": 2.0103984225139845e-05, + "loss": 0.4703, + "step": 21244 + }, + { + "epoch": 1.4258246367571559, + "grad_norm": 0.9168682098388672, + "learning_rate": 2.0095273380166036e-05, + "loss": 0.4566, + "step": 21246 + }, + { + "epoch": 1.4259588604409248, + "grad_norm": 1.0187290906906128, + "learning_rate": 2.0086563948131765e-05, + "loss": 0.493, + "step": 21248 + }, + { + "epoch": 1.4260930841246937, + "grad_norm": 1.0676566362380981, + "learning_rate": 2.007785592944853e-05, + "loss": 0.4371, + "step": 21250 + }, + { + "epoch": 1.4262273078084629, + "grad_norm": 1.172776699066162, + "learning_rate": 2.0069149324527763e-05, + "loss": 0.4856, + "step": 21252 + }, + { + "epoch": 1.4263615314922318, + "grad_norm": 0.997073233127594, + "learning_rate": 2.0060444133780816e-05, + "loss": 0.4938, + "step": 21254 + }, + { + "epoch": 1.426495755176001, + "grad_norm": 1.026848316192627, + "learning_rate": 2.0051740357619024e-05, + "loss": 0.4413, + "step": 21256 + }, + { + "epoch": 1.4266299788597698, + "grad_norm": 1.0656684637069702, + "learning_rate": 2.0043037996453597e-05, + "loss": 0.5141, + "step": 21258 + }, + { + "epoch": 1.4267642025435388, + "grad_norm": 0.9959386587142944, + "learning_rate": 2.0034337050695752e-05, + "loss": 0.4947, + "step": 21260 + }, + { + "epoch": 1.4268984262273077, + "grad_norm": 0.9422990083694458, + "learning_rate": 2.0025637520756518e-05, + "loss": 0.4286, + "step": 21262 + }, + { + "epoch": 1.4270326499110768, + "grad_norm": 1.0149545669555664, + "learning_rate": 2.0016939407046987e-05, + "loss": 0.555, + "step": 21264 + }, + { + "epoch": 1.4271668735948457, + "grad_norm": 1.128385066986084, + "learning_rate": 2.000824270997809e-05, + "loss": 0.4297, + "step": 21266 + }, + { + "epoch": 1.4273010972786149, + "grad_norm": 1.0133975744247437, + "learning_rate": 1.9999547429960774e-05, + "loss": 0.4741, + "step": 21268 + }, + { + "epoch": 1.4274353209623838, + "grad_norm": 1.1769696474075317, + "learning_rate": 1.9990853567405843e-05, + "loss": 0.5217, + "step": 21270 + }, + { + "epoch": 1.4275695446461527, + "grad_norm": 1.0363843441009521, + "learning_rate": 1.998216112272407e-05, + "loss": 0.477, + "step": 21272 + }, + { + "epoch": 1.4277037683299219, + "grad_norm": 0.99687260389328, + "learning_rate": 1.997347009632614e-05, + "loss": 0.5187, + "step": 21274 + }, + { + "epoch": 1.4278379920136908, + "grad_norm": 1.007660150527954, + "learning_rate": 1.996478048862272e-05, + "loss": 0.4331, + "step": 21276 + }, + { + "epoch": 1.42797221569746, + "grad_norm": 0.9654832482337952, + "learning_rate": 1.9956092300024364e-05, + "loss": 0.4636, + "step": 21278 + }, + { + "epoch": 1.4281064393812288, + "grad_norm": 1.1670583486557007, + "learning_rate": 1.9947405530941565e-05, + "loss": 0.514, + "step": 21280 + }, + { + "epoch": 1.4282406630649978, + "grad_norm": 0.8719428777694702, + "learning_rate": 1.9938720181784743e-05, + "loss": 0.4627, + "step": 21282 + }, + { + "epoch": 1.428374886748767, + "grad_norm": 1.3694096803665161, + "learning_rate": 1.9930036252964295e-05, + "loss": 0.4919, + "step": 21284 + }, + { + "epoch": 1.4285091104325358, + "grad_norm": 1.0464633703231812, + "learning_rate": 1.9921353744890488e-05, + "loss": 0.4696, + "step": 21286 + }, + { + "epoch": 1.428643334116305, + "grad_norm": 0.984787106513977, + "learning_rate": 1.99126726579736e-05, + "loss": 0.5129, + "step": 21288 + }, + { + "epoch": 1.4287775578000739, + "grad_norm": 1.0427882671356201, + "learning_rate": 1.9903992992623755e-05, + "loss": 0.4904, + "step": 21290 + }, + { + "epoch": 1.4289117814838428, + "grad_norm": 1.1129236221313477, + "learning_rate": 1.9895314749251038e-05, + "loss": 0.4712, + "step": 21292 + }, + { + "epoch": 1.4290460051676117, + "grad_norm": 1.205096960067749, + "learning_rate": 1.9886637928265516e-05, + "loss": 0.5254, + "step": 21294 + }, + { + "epoch": 1.4291802288513809, + "grad_norm": 1.0272270441055298, + "learning_rate": 1.9877962530077123e-05, + "loss": 0.4748, + "step": 21296 + }, + { + "epoch": 1.4293144525351498, + "grad_norm": 1.138121247291565, + "learning_rate": 1.986928855509578e-05, + "loss": 0.4687, + "step": 21298 + }, + { + "epoch": 1.429448676218919, + "grad_norm": 1.0026321411132812, + "learning_rate": 1.9860616003731307e-05, + "loss": 0.4459, + "step": 21300 + }, + { + "epoch": 1.4295828999026878, + "grad_norm": 0.9387633204460144, + "learning_rate": 1.9851944876393463e-05, + "loss": 0.4546, + "step": 21302 + }, + { + "epoch": 1.4297171235864568, + "grad_norm": 1.1133606433868408, + "learning_rate": 1.984327517349192e-05, + "loss": 0.4649, + "step": 21304 + }, + { + "epoch": 1.429851347270226, + "grad_norm": 0.9046503901481628, + "learning_rate": 1.983460689543634e-05, + "loss": 0.537, + "step": 21306 + }, + { + "epoch": 1.4299855709539948, + "grad_norm": 0.9311491847038269, + "learning_rate": 1.982594004263627e-05, + "loss": 0.4314, + "step": 21308 + }, + { + "epoch": 1.430119794637764, + "grad_norm": 1.051015853881836, + "learning_rate": 1.9817274615501204e-05, + "loss": 0.5037, + "step": 21310 + }, + { + "epoch": 1.4302540183215329, + "grad_norm": 1.1479281187057495, + "learning_rate": 1.9808610614440532e-05, + "loss": 0.4923, + "step": 21312 + }, + { + "epoch": 1.4303882420053018, + "grad_norm": 1.0123236179351807, + "learning_rate": 1.9799948039863666e-05, + "loss": 0.4856, + "step": 21314 + }, + { + "epoch": 1.4305224656890707, + "grad_norm": 0.9054355621337891, + "learning_rate": 1.979128689217986e-05, + "loss": 0.4404, + "step": 21316 + }, + { + "epoch": 1.4306566893728399, + "grad_norm": 1.3398513793945312, + "learning_rate": 1.9782627171798353e-05, + "loss": 0.4903, + "step": 21318 + }, + { + "epoch": 1.4307909130566088, + "grad_norm": 0.9859731793403625, + "learning_rate": 1.9773968879128272e-05, + "loss": 0.4297, + "step": 21320 + }, + { + "epoch": 1.430925136740378, + "grad_norm": 0.9070358276367188, + "learning_rate": 1.976531201457874e-05, + "loss": 0.4997, + "step": 21322 + }, + { + "epoch": 1.4310593604241468, + "grad_norm": 1.1439933776855469, + "learning_rate": 1.9756656578558746e-05, + "loss": 0.4739, + "step": 21324 + }, + { + "epoch": 1.4311935841079158, + "grad_norm": 0.9748851656913757, + "learning_rate": 1.9748002571477277e-05, + "loss": 0.4601, + "step": 21326 + }, + { + "epoch": 1.431327807791685, + "grad_norm": 1.045695424079895, + "learning_rate": 1.9739349993743194e-05, + "loss": 0.4841, + "step": 21328 + }, + { + "epoch": 1.4314620314754538, + "grad_norm": 0.9456130266189575, + "learning_rate": 1.973069884576532e-05, + "loss": 0.4816, + "step": 21330 + }, + { + "epoch": 1.431596255159223, + "grad_norm": 1.1043306589126587, + "learning_rate": 1.972204912795239e-05, + "loss": 0.4705, + "step": 21332 + }, + { + "epoch": 1.4317304788429919, + "grad_norm": 0.9588850140571594, + "learning_rate": 1.9713400840713116e-05, + "loss": 0.4906, + "step": 21334 + }, + { + "epoch": 1.4318647025267608, + "grad_norm": 1.0731778144836426, + "learning_rate": 1.9704753984456094e-05, + "loss": 0.5061, + "step": 21336 + }, + { + "epoch": 1.4319989262105297, + "grad_norm": 1.1636130809783936, + "learning_rate": 1.9696108559589875e-05, + "loss": 0.4515, + "step": 21338 + }, + { + "epoch": 1.4321331498942989, + "grad_norm": 0.8831275701522827, + "learning_rate": 1.968746456652292e-05, + "loss": 0.4468, + "step": 21340 + }, + { + "epoch": 1.4322673735780678, + "grad_norm": 1.1297000646591187, + "learning_rate": 1.967882200566367e-05, + "loss": 0.5227, + "step": 21342 + }, + { + "epoch": 1.432401597261837, + "grad_norm": 0.9723655581474304, + "learning_rate": 1.967018087742044e-05, + "loss": 0.4799, + "step": 21344 + }, + { + "epoch": 1.4325358209456058, + "grad_norm": 0.8861226439476013, + "learning_rate": 1.966154118220156e-05, + "loss": 0.4335, + "step": 21346 + }, + { + "epoch": 1.4326700446293748, + "grad_norm": 1.075437307357788, + "learning_rate": 1.9652902920415167e-05, + "loss": 0.4959, + "step": 21348 + }, + { + "epoch": 1.432804268313144, + "grad_norm": 1.0862118005752563, + "learning_rate": 1.964426609246945e-05, + "loss": 0.5556, + "step": 21350 + }, + { + "epoch": 1.4329384919969128, + "grad_norm": 1.2012962102890015, + "learning_rate": 1.9635630698772455e-05, + "loss": 0.5405, + "step": 21352 + }, + { + "epoch": 1.433072715680682, + "grad_norm": 1.0934321880340576, + "learning_rate": 1.9626996739732212e-05, + "loss": 0.4795, + "step": 21354 + }, + { + "epoch": 1.4332069393644509, + "grad_norm": 0.8701134920120239, + "learning_rate": 1.961836421575667e-05, + "loss": 0.445, + "step": 21356 + }, + { + "epoch": 1.4333411630482198, + "grad_norm": 1.0055067539215088, + "learning_rate": 1.960973312725364e-05, + "loss": 0.5013, + "step": 21358 + }, + { + "epoch": 1.433475386731989, + "grad_norm": 0.9744080305099487, + "learning_rate": 1.960110347463097e-05, + "loss": 0.4911, + "step": 21360 + }, + { + "epoch": 1.4336096104157579, + "grad_norm": 1.0852227210998535, + "learning_rate": 1.959247525829638e-05, + "loss": 0.5211, + "step": 21362 + }, + { + "epoch": 1.433743834099527, + "grad_norm": 0.905018150806427, + "learning_rate": 1.9583848478657544e-05, + "loss": 0.5219, + "step": 21364 + }, + { + "epoch": 1.433878057783296, + "grad_norm": 0.9459779262542725, + "learning_rate": 1.957522313612207e-05, + "loss": 0.5036, + "step": 21366 + }, + { + "epoch": 1.4340122814670648, + "grad_norm": 1.0793790817260742, + "learning_rate": 1.9566599231097467e-05, + "loss": 0.4279, + "step": 21368 + }, + { + "epoch": 1.4341465051508338, + "grad_norm": 0.9977946281433105, + "learning_rate": 1.9557976763991188e-05, + "loss": 0.4438, + "step": 21370 + }, + { + "epoch": 1.434280728834603, + "grad_norm": 1.025618314743042, + "learning_rate": 1.9549355735210663e-05, + "loss": 0.5266, + "step": 21372 + }, + { + "epoch": 1.4344149525183718, + "grad_norm": 1.0904392004013062, + "learning_rate": 1.95407361451632e-05, + "loss": 0.5137, + "step": 21374 + }, + { + "epoch": 1.434549176202141, + "grad_norm": 1.053537130355835, + "learning_rate": 1.953211799425606e-05, + "loss": 0.4902, + "step": 21376 + }, + { + "epoch": 1.4346833998859099, + "grad_norm": 1.0914679765701294, + "learning_rate": 1.9523501282896417e-05, + "loss": 0.5094, + "step": 21378 + }, + { + "epoch": 1.4348176235696788, + "grad_norm": 0.898318350315094, + "learning_rate": 1.9514886011491422e-05, + "loss": 0.4391, + "step": 21380 + }, + { + "epoch": 1.434951847253448, + "grad_norm": 0.9205227494239807, + "learning_rate": 1.9506272180448104e-05, + "loss": 0.4686, + "step": 21382 + }, + { + "epoch": 1.4350860709372169, + "grad_norm": 1.480831503868103, + "learning_rate": 1.9497659790173495e-05, + "loss": 0.489, + "step": 21384 + }, + { + "epoch": 1.435220294620986, + "grad_norm": 0.9175587892532349, + "learning_rate": 1.9489048841074444e-05, + "loss": 0.4238, + "step": 21386 + }, + { + "epoch": 1.435354518304755, + "grad_norm": 1.1250370740890503, + "learning_rate": 1.9480439333557855e-05, + "loss": 0.4379, + "step": 21388 + }, + { + "epoch": 1.4354887419885238, + "grad_norm": 1.0786669254302979, + "learning_rate": 1.9471831268030478e-05, + "loss": 0.5135, + "step": 21390 + }, + { + "epoch": 1.4356229656722928, + "grad_norm": 0.9290297031402588, + "learning_rate": 1.9463224644899057e-05, + "loss": 0.3903, + "step": 21392 + }, + { + "epoch": 1.435757189356062, + "grad_norm": 0.8532508015632629, + "learning_rate": 1.945461946457023e-05, + "loss": 0.3723, + "step": 21394 + }, + { + "epoch": 1.4358914130398308, + "grad_norm": 1.1052356958389282, + "learning_rate": 1.944601572745056e-05, + "loss": 0.5119, + "step": 21396 + }, + { + "epoch": 1.4360256367236, + "grad_norm": 1.059980034828186, + "learning_rate": 1.9437413433946556e-05, + "loss": 0.538, + "step": 21398 + }, + { + "epoch": 1.4361598604073689, + "grad_norm": 0.925700843334198, + "learning_rate": 1.942881258446468e-05, + "loss": 0.4563, + "step": 21400 + }, + { + "epoch": 1.4362940840911378, + "grad_norm": 1.0991933345794678, + "learning_rate": 1.9420213179411295e-05, + "loss": 0.5543, + "step": 21402 + }, + { + "epoch": 1.436428307774907, + "grad_norm": 1.028889536857605, + "learning_rate": 1.941161521919271e-05, + "loss": 0.5328, + "step": 21404 + }, + { + "epoch": 1.4365625314586759, + "grad_norm": 1.015349268913269, + "learning_rate": 1.940301870421513e-05, + "loss": 0.4829, + "step": 21406 + }, + { + "epoch": 1.436696755142445, + "grad_norm": 1.159861445426941, + "learning_rate": 1.9394423634884772e-05, + "loss": 0.4627, + "step": 21408 + }, + { + "epoch": 1.436830978826214, + "grad_norm": 1.0292245149612427, + "learning_rate": 1.9385830011607698e-05, + "loss": 0.4361, + "step": 21410 + }, + { + "epoch": 1.4369652025099828, + "grad_norm": 1.166775107383728, + "learning_rate": 1.9377237834789987e-05, + "loss": 0.4988, + "step": 21412 + }, + { + "epoch": 1.4370994261937517, + "grad_norm": 1.0735512971878052, + "learning_rate": 1.936864710483754e-05, + "loss": 0.5168, + "step": 21414 + }, + { + "epoch": 1.437233649877521, + "grad_norm": 1.0026241540908813, + "learning_rate": 1.9360057822156298e-05, + "loss": 0.4931, + "step": 21416 + }, + { + "epoch": 1.4373678735612898, + "grad_norm": 0.9932929277420044, + "learning_rate": 1.9351469987152056e-05, + "loss": 0.5066, + "step": 21418 + }, + { + "epoch": 1.437502097245059, + "grad_norm": 1.0896600484848022, + "learning_rate": 1.934288360023061e-05, + "loss": 0.4603, + "step": 21420 + }, + { + "epoch": 1.4376363209288279, + "grad_norm": 1.0472451448440552, + "learning_rate": 1.9334298661797623e-05, + "loss": 0.4696, + "step": 21422 + }, + { + "epoch": 1.4377705446125968, + "grad_norm": 1.1072280406951904, + "learning_rate": 1.932571517225873e-05, + "loss": 0.51, + "step": 21424 + }, + { + "epoch": 1.437904768296366, + "grad_norm": 1.0572043657302856, + "learning_rate": 1.9317133132019472e-05, + "loss": 0.4363, + "step": 21426 + }, + { + "epoch": 1.4380389919801349, + "grad_norm": 0.983066201210022, + "learning_rate": 1.930855254148532e-05, + "loss": 0.4479, + "step": 21428 + }, + { + "epoch": 1.438173215663904, + "grad_norm": 0.9452990293502808, + "learning_rate": 1.9299973401061726e-05, + "loss": 0.4991, + "step": 21430 + }, + { + "epoch": 1.438307439347673, + "grad_norm": 1.0300242900848389, + "learning_rate": 1.9291395711154025e-05, + "loss": 0.5181, + "step": 21432 + }, + { + "epoch": 1.4384416630314418, + "grad_norm": 1.1217944622039795, + "learning_rate": 1.9282819472167484e-05, + "loss": 0.5238, + "step": 21434 + }, + { + "epoch": 1.438575886715211, + "grad_norm": 0.9125789403915405, + "learning_rate": 1.9274244684507307e-05, + "loss": 0.4463, + "step": 21436 + }, + { + "epoch": 1.43871011039898, + "grad_norm": 1.03207266330719, + "learning_rate": 1.9265671348578667e-05, + "loss": 0.5244, + "step": 21438 + }, + { + "epoch": 1.438844334082749, + "grad_norm": 1.0886263847351074, + "learning_rate": 1.9257099464786615e-05, + "loss": 0.4924, + "step": 21440 + }, + { + "epoch": 1.438978557766518, + "grad_norm": 0.8789401650428772, + "learning_rate": 1.9248529033536162e-05, + "loss": 0.5126, + "step": 21442 + }, + { + "epoch": 1.4391127814502869, + "grad_norm": 1.0428028106689453, + "learning_rate": 1.9239960055232226e-05, + "loss": 0.4548, + "step": 21444 + }, + { + "epoch": 1.4392470051340558, + "grad_norm": 1.0682729482650757, + "learning_rate": 1.9231392530279706e-05, + "loss": 0.4665, + "step": 21446 + }, + { + "epoch": 1.439381228817825, + "grad_norm": 0.9879252314567566, + "learning_rate": 1.922282645908337e-05, + "loss": 0.5348, + "step": 21448 + }, + { + "epoch": 1.4395154525015939, + "grad_norm": 1.1306565999984741, + "learning_rate": 1.9214261842047982e-05, + "loss": 0.4947, + "step": 21450 + }, + { + "epoch": 1.439649676185363, + "grad_norm": 1.0321546792984009, + "learning_rate": 1.920569867957818e-05, + "loss": 0.482, + "step": 21452 + }, + { + "epoch": 1.439783899869132, + "grad_norm": 1.0249627828598022, + "learning_rate": 1.9197136972078563e-05, + "loss": 0.4649, + "step": 21454 + }, + { + "epoch": 1.4399181235529008, + "grad_norm": 1.0898289680480957, + "learning_rate": 1.9188576719953633e-05, + "loss": 0.5052, + "step": 21456 + }, + { + "epoch": 1.44005234723667, + "grad_norm": 0.9657506942749023, + "learning_rate": 1.9180017923607886e-05, + "loss": 0.428, + "step": 21458 + }, + { + "epoch": 1.440186570920439, + "grad_norm": 0.9635042548179626, + "learning_rate": 1.9171460583445684e-05, + "loss": 0.3944, + "step": 21460 + }, + { + "epoch": 1.440320794604208, + "grad_norm": 0.8400821089744568, + "learning_rate": 1.9162904699871347e-05, + "loss": 0.4102, + "step": 21462 + }, + { + "epoch": 1.440455018287977, + "grad_norm": 1.024552583694458, + "learning_rate": 1.9154350273289113e-05, + "loss": 0.5273, + "step": 21464 + }, + { + "epoch": 1.4405892419717459, + "grad_norm": 0.8710123300552368, + "learning_rate": 1.9145797304103186e-05, + "loss": 0.5207, + "step": 21466 + }, + { + "epoch": 1.4407234656555148, + "grad_norm": 1.0175509452819824, + "learning_rate": 1.9137245792717668e-05, + "loss": 0.4739, + "step": 21468 + }, + { + "epoch": 1.440857689339284, + "grad_norm": 1.0981627702713013, + "learning_rate": 1.912869573953659e-05, + "loss": 0.4928, + "step": 21470 + }, + { + "epoch": 1.4409919130230529, + "grad_norm": 1.0668026208877563, + "learning_rate": 1.9120147144963918e-05, + "loss": 0.5096, + "step": 21472 + }, + { + "epoch": 1.441126136706822, + "grad_norm": 0.9977552890777588, + "learning_rate": 1.9111600009403592e-05, + "loss": 0.4646, + "step": 21474 + }, + { + "epoch": 1.441260360390591, + "grad_norm": 1.1157846450805664, + "learning_rate": 1.910305433325941e-05, + "loss": 0.5493, + "step": 21476 + }, + { + "epoch": 1.4413945840743598, + "grad_norm": 1.1162174940109253, + "learning_rate": 1.9094510116935167e-05, + "loss": 0.4907, + "step": 21478 + }, + { + "epoch": 1.441528807758129, + "grad_norm": 1.1077864170074463, + "learning_rate": 1.9085967360834544e-05, + "loss": 0.5888, + "step": 21480 + }, + { + "epoch": 1.441663031441898, + "grad_norm": 0.912291944026947, + "learning_rate": 1.907742606536118e-05, + "loss": 0.421, + "step": 21482 + }, + { + "epoch": 1.441797255125667, + "grad_norm": 1.0846011638641357, + "learning_rate": 1.9068886230918608e-05, + "loss": 0.4901, + "step": 21484 + }, + { + "epoch": 1.441931478809436, + "grad_norm": 1.0536340475082397, + "learning_rate": 1.906034785791036e-05, + "loss": 0.5485, + "step": 21486 + }, + { + "epoch": 1.4420657024932049, + "grad_norm": 1.0287697315216064, + "learning_rate": 1.9051810946739828e-05, + "loss": 0.5602, + "step": 21488 + }, + { + "epoch": 1.4421999261769738, + "grad_norm": 0.951287567615509, + "learning_rate": 1.9043275497810377e-05, + "loss": 0.4924, + "step": 21490 + }, + { + "epoch": 1.442334149860743, + "grad_norm": 0.9405043721199036, + "learning_rate": 1.903474151152527e-05, + "loss": 0.4611, + "step": 21492 + }, + { + "epoch": 1.4424683735445119, + "grad_norm": 0.859656035900116, + "learning_rate": 1.9026208988287746e-05, + "loss": 0.4543, + "step": 21494 + }, + { + "epoch": 1.442602597228281, + "grad_norm": 0.9674857258796692, + "learning_rate": 1.901767792850094e-05, + "loss": 0.5117, + "step": 21496 + }, + { + "epoch": 1.44273682091205, + "grad_norm": 1.019100546836853, + "learning_rate": 1.9009148332567932e-05, + "loss": 0.4471, + "step": 21498 + }, + { + "epoch": 1.4428710445958188, + "grad_norm": 0.9753833413124084, + "learning_rate": 1.900062020089173e-05, + "loss": 0.5191, + "step": 21500 + }, + { + "epoch": 1.443005268279588, + "grad_norm": 0.9421480894088745, + "learning_rate": 1.899209353387524e-05, + "loss": 0.419, + "step": 21502 + }, + { + "epoch": 1.443139491963357, + "grad_norm": 0.9760271310806274, + "learning_rate": 1.8983568331921375e-05, + "loss": 0.4939, + "step": 21504 + }, + { + "epoch": 1.443273715647126, + "grad_norm": 0.9397803544998169, + "learning_rate": 1.89750445954329e-05, + "loss": 0.4635, + "step": 21506 + }, + { + "epoch": 1.443407939330895, + "grad_norm": 1.3463327884674072, + "learning_rate": 1.896652232481259e-05, + "loss": 0.5534, + "step": 21508 + }, + { + "epoch": 1.4435421630146639, + "grad_norm": 1.047634243965149, + "learning_rate": 1.8958001520463043e-05, + "loss": 0.5032, + "step": 21510 + }, + { + "epoch": 1.443676386698433, + "grad_norm": 1.0969831943511963, + "learning_rate": 1.8949482182786904e-05, + "loss": 0.4102, + "step": 21512 + }, + { + "epoch": 1.443810610382202, + "grad_norm": 0.9968197345733643, + "learning_rate": 1.8940964312186653e-05, + "loss": 0.4793, + "step": 21514 + }, + { + "epoch": 1.443944834065971, + "grad_norm": 1.0413786172866821, + "learning_rate": 1.8932447909064775e-05, + "loss": 0.4534, + "step": 21516 + }, + { + "epoch": 1.44407905774974, + "grad_norm": 0.8649364709854126, + "learning_rate": 1.892393297382365e-05, + "loss": 0.4605, + "step": 21518 + }, + { + "epoch": 1.444213281433509, + "grad_norm": 1.0816700458526611, + "learning_rate": 1.8915419506865583e-05, + "loss": 0.4633, + "step": 21520 + }, + { + "epoch": 1.4443475051172778, + "grad_norm": 1.1534231901168823, + "learning_rate": 1.8906907508592798e-05, + "loss": 0.5159, + "step": 21522 + }, + { + "epoch": 1.444481728801047, + "grad_norm": 0.9998898506164551, + "learning_rate": 1.889839697940751e-05, + "loss": 0.4935, + "step": 21524 + }, + { + "epoch": 1.444615952484816, + "grad_norm": 1.0361120700836182, + "learning_rate": 1.8889887919711808e-05, + "loss": 0.5168, + "step": 21526 + }, + { + "epoch": 1.444750176168585, + "grad_norm": 0.9519708156585693, + "learning_rate": 1.8881380329907722e-05, + "loss": 0.4998, + "step": 21528 + }, + { + "epoch": 1.444884399852354, + "grad_norm": 1.0586285591125488, + "learning_rate": 1.8872874210397213e-05, + "loss": 0.4256, + "step": 21530 + }, + { + "epoch": 1.4450186235361229, + "grad_norm": 1.076627254486084, + "learning_rate": 1.88643695615822e-05, + "loss": 0.5524, + "step": 21532 + }, + { + "epoch": 1.445152847219892, + "grad_norm": 0.9264193177223206, + "learning_rate": 1.8855866383864483e-05, + "loss": 0.4912, + "step": 21534 + }, + { + "epoch": 1.445287070903661, + "grad_norm": 1.0108731985092163, + "learning_rate": 1.8847364677645873e-05, + "loss": 0.5871, + "step": 21536 + }, + { + "epoch": 1.44542129458743, + "grad_norm": 1.1152682304382324, + "learning_rate": 1.8838864443327986e-05, + "loss": 0.4531, + "step": 21538 + }, + { + "epoch": 1.445555518271199, + "grad_norm": 1.0338091850280762, + "learning_rate": 1.8830365681312502e-05, + "loss": 0.5172, + "step": 21540 + }, + { + "epoch": 1.445689741954968, + "grad_norm": 1.0477180480957031, + "learning_rate": 1.8821868392000925e-05, + "loss": 0.4847, + "step": 21542 + }, + { + "epoch": 1.4458239656387368, + "grad_norm": 1.0267977714538574, + "learning_rate": 1.8813372575794768e-05, + "loss": 0.5348, + "step": 21544 + }, + { + "epoch": 1.445958189322506, + "grad_norm": 0.9160050749778748, + "learning_rate": 1.880487823309544e-05, + "loss": 0.5006, + "step": 21546 + }, + { + "epoch": 1.446092413006275, + "grad_norm": 0.9413403868675232, + "learning_rate": 1.879638536430427e-05, + "loss": 0.4925, + "step": 21548 + }, + { + "epoch": 1.446226636690044, + "grad_norm": 1.0063854455947876, + "learning_rate": 1.8787893969822517e-05, + "loss": 0.468, + "step": 21550 + }, + { + "epoch": 1.446360860373813, + "grad_norm": 0.9902949929237366, + "learning_rate": 1.8779404050051413e-05, + "loss": 0.4644, + "step": 21552 + }, + { + "epoch": 1.4464950840575819, + "grad_norm": 0.7921978235244751, + "learning_rate": 1.8770915605392077e-05, + "loss": 0.4609, + "step": 21554 + }, + { + "epoch": 1.446629307741351, + "grad_norm": 1.0576163530349731, + "learning_rate": 1.8762428636245564e-05, + "loss": 0.4959, + "step": 21556 + }, + { + "epoch": 1.44676353142512, + "grad_norm": 0.8750025033950806, + "learning_rate": 1.8753943143012864e-05, + "loss": 0.4538, + "step": 21558 + }, + { + "epoch": 1.446897755108889, + "grad_norm": 0.9878489971160889, + "learning_rate": 1.8745459126094915e-05, + "loss": 0.5063, + "step": 21560 + }, + { + "epoch": 1.447031978792658, + "grad_norm": 1.0368151664733887, + "learning_rate": 1.8736976585892572e-05, + "loss": 0.4613, + "step": 21562 + }, + { + "epoch": 1.447166202476427, + "grad_norm": 1.1094083786010742, + "learning_rate": 1.8728495522806606e-05, + "loss": 0.4851, + "step": 21564 + }, + { + "epoch": 1.4473004261601958, + "grad_norm": 0.9020206928253174, + "learning_rate": 1.8720015937237733e-05, + "loss": 0.4275, + "step": 21566 + }, + { + "epoch": 1.447434649843965, + "grad_norm": 1.0404009819030762, + "learning_rate": 1.871153782958658e-05, + "loss": 0.4766, + "step": 21568 + }, + { + "epoch": 1.447568873527734, + "grad_norm": 1.043221354484558, + "learning_rate": 1.870306120025375e-05, + "loss": 0.4849, + "step": 21570 + }, + { + "epoch": 1.447703097211503, + "grad_norm": 0.9704181551933289, + "learning_rate": 1.869458604963973e-05, + "loss": 0.446, + "step": 21572 + }, + { + "epoch": 1.447837320895272, + "grad_norm": 0.889318585395813, + "learning_rate": 1.8686112378144988e-05, + "loss": 0.4764, + "step": 21574 + }, + { + "epoch": 1.4479715445790409, + "grad_norm": 0.9584651589393616, + "learning_rate": 1.867764018616982e-05, + "loss": 0.5097, + "step": 21576 + }, + { + "epoch": 1.44810576826281, + "grad_norm": 1.079387903213501, + "learning_rate": 1.8669169474114582e-05, + "loss": 0.5277, + "step": 21578 + }, + { + "epoch": 1.448239991946579, + "grad_norm": 0.998833954334259, + "learning_rate": 1.8660700242379463e-05, + "loss": 0.4426, + "step": 21580 + }, + { + "epoch": 1.448374215630348, + "grad_norm": 0.927009642124176, + "learning_rate": 1.8652232491364648e-05, + "loss": 0.5055, + "step": 21582 + }, + { + "epoch": 1.448508439314117, + "grad_norm": 1.0209903717041016, + "learning_rate": 1.8643766221470204e-05, + "loss": 0.5451, + "step": 21584 + }, + { + "epoch": 1.448642662997886, + "grad_norm": 1.0737221240997314, + "learning_rate": 1.863530143309615e-05, + "loss": 0.488, + "step": 21586 + }, + { + "epoch": 1.448776886681655, + "grad_norm": 1.0311989784240723, + "learning_rate": 1.8626838126642403e-05, + "loss": 0.4725, + "step": 21588 + }, + { + "epoch": 1.448911110365424, + "grad_norm": 1.0966482162475586, + "learning_rate": 1.861837630250888e-05, + "loss": 0.5068, + "step": 21590 + }, + { + "epoch": 1.4490453340491931, + "grad_norm": 1.0530413389205933, + "learning_rate": 1.860991596109537e-05, + "loss": 0.5359, + "step": 21592 + }, + { + "epoch": 1.449179557732962, + "grad_norm": 0.965501070022583, + "learning_rate": 1.8601457102801608e-05, + "loss": 0.4393, + "step": 21594 + }, + { + "epoch": 1.449313781416731, + "grad_norm": 1.1146641969680786, + "learning_rate": 1.8592999728027234e-05, + "loss": 0.5026, + "step": 21596 + }, + { + "epoch": 1.4494480051004999, + "grad_norm": 1.087504267692566, + "learning_rate": 1.858454383717188e-05, + "loss": 0.5362, + "step": 21598 + }, + { + "epoch": 1.449582228784269, + "grad_norm": 0.9007117748260498, + "learning_rate": 1.857608943063504e-05, + "loss": 0.4667, + "step": 21600 + }, + { + "epoch": 1.449716452468038, + "grad_norm": 1.2094337940216064, + "learning_rate": 1.8567636508816212e-05, + "loss": 0.587, + "step": 21602 + }, + { + "epoch": 1.449850676151807, + "grad_norm": 1.0666933059692383, + "learning_rate": 1.855918507211472e-05, + "loss": 0.4908, + "step": 21604 + }, + { + "epoch": 1.449984899835576, + "grad_norm": 0.9486268758773804, + "learning_rate": 1.8550735120929926e-05, + "loss": 0.4853, + "step": 21606 + }, + { + "epoch": 1.450119123519345, + "grad_norm": 1.0211158990859985, + "learning_rate": 1.8542286655661027e-05, + "loss": 0.5098, + "step": 21608 + }, + { + "epoch": 1.450253347203114, + "grad_norm": 1.0262904167175293, + "learning_rate": 1.8533839676707255e-05, + "loss": 0.5266, + "step": 21610 + }, + { + "epoch": 1.450387570886883, + "grad_norm": 1.088341474533081, + "learning_rate": 1.8525394184467676e-05, + "loss": 0.4646, + "step": 21612 + }, + { + "epoch": 1.4505217945706521, + "grad_norm": 1.045545220375061, + "learning_rate": 1.851695017934133e-05, + "loss": 0.5073, + "step": 21614 + }, + { + "epoch": 1.450656018254421, + "grad_norm": 1.011045217514038, + "learning_rate": 1.8508507661727163e-05, + "loss": 0.5643, + "step": 21616 + }, + { + "epoch": 1.45079024193819, + "grad_norm": 0.9788286685943604, + "learning_rate": 1.85000666320241e-05, + "loss": 0.5108, + "step": 21618 + }, + { + "epoch": 1.4509244656219589, + "grad_norm": 1.0276010036468506, + "learning_rate": 1.8491627090630948e-05, + "loss": 0.4665, + "step": 21620 + }, + { + "epoch": 1.451058689305728, + "grad_norm": 0.9841398000717163, + "learning_rate": 1.848318903794646e-05, + "loss": 0.4942, + "step": 21622 + }, + { + "epoch": 1.451192912989497, + "grad_norm": 1.0520919561386108, + "learning_rate": 1.8474752474369296e-05, + "loss": 0.4844, + "step": 21624 + }, + { + "epoch": 1.451327136673266, + "grad_norm": 0.9467720985412598, + "learning_rate": 1.84663174002981e-05, + "loss": 0.4829, + "step": 21626 + }, + { + "epoch": 1.451461360357035, + "grad_norm": 0.9637964963912964, + "learning_rate": 1.845788381613138e-05, + "loss": 0.4828, + "step": 21628 + }, + { + "epoch": 1.451595584040804, + "grad_norm": 1.0604698657989502, + "learning_rate": 1.8449451722267646e-05, + "loss": 0.492, + "step": 21630 + }, + { + "epoch": 1.451729807724573, + "grad_norm": 1.139906406402588, + "learning_rate": 1.844102111910529e-05, + "loss": 0.532, + "step": 21632 + }, + { + "epoch": 1.451864031408342, + "grad_norm": 1.1570615768432617, + "learning_rate": 1.84325920070426e-05, + "loss": 0.4849, + "step": 21634 + }, + { + "epoch": 1.4519982550921111, + "grad_norm": 1.039461374282837, + "learning_rate": 1.8424164386477878e-05, + "loss": 0.4266, + "step": 21636 + }, + { + "epoch": 1.45213247877588, + "grad_norm": 1.10236394405365, + "learning_rate": 1.8415738257809275e-05, + "loss": 0.5226, + "step": 21638 + }, + { + "epoch": 1.452266702459649, + "grad_norm": 1.047585129737854, + "learning_rate": 1.8407313621434952e-05, + "loss": 0.4525, + "step": 21640 + }, + { + "epoch": 1.4524009261434179, + "grad_norm": 1.1584211587905884, + "learning_rate": 1.839889047775294e-05, + "loss": 0.5172, + "step": 21642 + }, + { + "epoch": 1.452535149827187, + "grad_norm": 1.0993553400039673, + "learning_rate": 1.8390468827161207e-05, + "loss": 0.5157, + "step": 21644 + }, + { + "epoch": 1.452669373510956, + "grad_norm": 1.1074891090393066, + "learning_rate": 1.838204867005765e-05, + "loss": 0.5238, + "step": 21646 + }, + { + "epoch": 1.452803597194725, + "grad_norm": 0.9702417850494385, + "learning_rate": 1.837363000684013e-05, + "loss": 0.4804, + "step": 21648 + }, + { + "epoch": 1.452937820878494, + "grad_norm": 1.046984076499939, + "learning_rate": 1.836521283790641e-05, + "loss": 0.4525, + "step": 21650 + }, + { + "epoch": 1.453072044562263, + "grad_norm": 1.0746935606002808, + "learning_rate": 1.835679716365417e-05, + "loss": 0.4884, + "step": 21652 + }, + { + "epoch": 1.453206268246032, + "grad_norm": 1.0536320209503174, + "learning_rate": 1.8348382984481034e-05, + "loss": 0.4848, + "step": 21654 + }, + { + "epoch": 1.453340491929801, + "grad_norm": 1.8641531467437744, + "learning_rate": 1.833997030078457e-05, + "loss": 0.5027, + "step": 21656 + }, + { + "epoch": 1.4534747156135701, + "grad_norm": 1.090524435043335, + "learning_rate": 1.8331559112962232e-05, + "loss": 0.4615, + "step": 21658 + }, + { + "epoch": 1.453608939297339, + "grad_norm": 1.1034295558929443, + "learning_rate": 1.8323149421411497e-05, + "loss": 0.5082, + "step": 21660 + }, + { + "epoch": 1.453743162981108, + "grad_norm": 1.0330030918121338, + "learning_rate": 1.831474122652962e-05, + "loss": 0.4812, + "step": 21662 + }, + { + "epoch": 1.453877386664877, + "grad_norm": 0.9497284293174744, + "learning_rate": 1.830633452871393e-05, + "loss": 0.4553, + "step": 21664 + }, + { + "epoch": 1.454011610348646, + "grad_norm": 0.9797527194023132, + "learning_rate": 1.8297929328361594e-05, + "loss": 0.4597, + "step": 21666 + }, + { + "epoch": 1.4541458340324152, + "grad_norm": 0.9925074577331543, + "learning_rate": 1.828952562586978e-05, + "loss": 0.4644, + "step": 21668 + }, + { + "epoch": 1.454280057716184, + "grad_norm": 0.9066833853721619, + "learning_rate": 1.828112342163552e-05, + "loss": 0.4259, + "step": 21670 + }, + { + "epoch": 1.454414281399953, + "grad_norm": 1.0383461713790894, + "learning_rate": 1.827272271605581e-05, + "loss": 0.4895, + "step": 21672 + }, + { + "epoch": 1.454548505083722, + "grad_norm": 1.1868698596954346, + "learning_rate": 1.8264323509527543e-05, + "loss": 0.4826, + "step": 21674 + }, + { + "epoch": 1.454682728767491, + "grad_norm": 0.9397610425949097, + "learning_rate": 1.8255925802447604e-05, + "loss": 0.4659, + "step": 21676 + }, + { + "epoch": 1.45481695245126, + "grad_norm": 0.8566227555274963, + "learning_rate": 1.8247529595212743e-05, + "loss": 0.4663, + "step": 21678 + }, + { + "epoch": 1.4549511761350291, + "grad_norm": 1.125922441482544, + "learning_rate": 1.8239134888219676e-05, + "loss": 0.5208, + "step": 21680 + }, + { + "epoch": 1.455085399818798, + "grad_norm": 1.0006755590438843, + "learning_rate": 1.823074168186501e-05, + "loss": 0.5075, + "step": 21682 + }, + { + "epoch": 1.455219623502567, + "grad_norm": 1.1265826225280762, + "learning_rate": 1.8222349976545343e-05, + "loss": 0.5357, + "step": 21684 + }, + { + "epoch": 1.455353847186336, + "grad_norm": 1.0235995054244995, + "learning_rate": 1.8213959772657142e-05, + "loss": 0.4938, + "step": 21686 + }, + { + "epoch": 1.455488070870105, + "grad_norm": 0.8493221998214722, + "learning_rate": 1.8205571070596867e-05, + "loss": 0.4864, + "step": 21688 + }, + { + "epoch": 1.4556222945538742, + "grad_norm": 1.1292208433151245, + "learning_rate": 1.8197183870760803e-05, + "loss": 0.4735, + "step": 21690 + }, + { + "epoch": 1.455756518237643, + "grad_norm": 1.0905543565750122, + "learning_rate": 1.818879817354528e-05, + "loss": 0.4877, + "step": 21692 + }, + { + "epoch": 1.455890741921412, + "grad_norm": 0.9483680725097656, + "learning_rate": 1.818041397934647e-05, + "loss": 0.4189, + "step": 21694 + }, + { + "epoch": 1.456024965605181, + "grad_norm": 0.9809787273406982, + "learning_rate": 1.8172031288560554e-05, + "loss": 0.4545, + "step": 21696 + }, + { + "epoch": 1.45615918928895, + "grad_norm": 0.9923301935195923, + "learning_rate": 1.8163650101583567e-05, + "loss": 0.5628, + "step": 21698 + }, + { + "epoch": 1.456293412972719, + "grad_norm": 1.0795013904571533, + "learning_rate": 1.815527041881151e-05, + "loss": 0.4643, + "step": 21700 + }, + { + "epoch": 1.4564276366564881, + "grad_norm": 1.05745530128479, + "learning_rate": 1.8146892240640307e-05, + "loss": 0.4644, + "step": 21702 + }, + { + "epoch": 1.456561860340257, + "grad_norm": 1.1342967748641968, + "learning_rate": 1.8138515567465793e-05, + "loss": 0.5214, + "step": 21704 + }, + { + "epoch": 1.456696084024026, + "grad_norm": 1.1231117248535156, + "learning_rate": 1.8130140399683776e-05, + "loss": 0.429, + "step": 21706 + }, + { + "epoch": 1.456830307707795, + "grad_norm": 0.9478837847709656, + "learning_rate": 1.812176673768996e-05, + "loss": 0.5035, + "step": 21708 + }, + { + "epoch": 1.456964531391564, + "grad_norm": 1.055141806602478, + "learning_rate": 1.8113394581879985e-05, + "loss": 0.4912, + "step": 21710 + }, + { + "epoch": 1.4570987550753332, + "grad_norm": 0.8722517490386963, + "learning_rate": 1.810502393264939e-05, + "loss": 0.425, + "step": 21712 + }, + { + "epoch": 1.457232978759102, + "grad_norm": 1.0529042482376099, + "learning_rate": 1.8096654790393714e-05, + "loss": 0.4594, + "step": 21714 + }, + { + "epoch": 1.457367202442871, + "grad_norm": 1.0669209957122803, + "learning_rate": 1.808828715550836e-05, + "loss": 0.5019, + "step": 21716 + }, + { + "epoch": 1.45750142612664, + "grad_norm": 0.9976727366447449, + "learning_rate": 1.8079921028388692e-05, + "loss": 0.4755, + "step": 21718 + }, + { + "epoch": 1.457635649810409, + "grad_norm": 1.0423046350479126, + "learning_rate": 1.8071556409429968e-05, + "loss": 0.4495, + "step": 21720 + }, + { + "epoch": 1.457769873494178, + "grad_norm": 1.0592374801635742, + "learning_rate": 1.8063193299027436e-05, + "loss": 0.5846, + "step": 21722 + }, + { + "epoch": 1.4579040971779471, + "grad_norm": 1.0062223672866821, + "learning_rate": 1.8054831697576203e-05, + "loss": 0.4799, + "step": 21724 + }, + { + "epoch": 1.458038320861716, + "grad_norm": 0.9556512832641602, + "learning_rate": 1.8046471605471393e-05, + "loss": 0.4769, + "step": 21726 + }, + { + "epoch": 1.458172544545485, + "grad_norm": 0.8796365261077881, + "learning_rate": 1.8038113023107928e-05, + "loss": 0.4946, + "step": 21728 + }, + { + "epoch": 1.458306768229254, + "grad_norm": 1.0229606628417969, + "learning_rate": 1.8029755950880787e-05, + "loss": 0.4577, + "step": 21730 + }, + { + "epoch": 1.458440991913023, + "grad_norm": 1.0130605697631836, + "learning_rate": 1.8021400389184795e-05, + "loss": 0.5172, + "step": 21732 + }, + { + "epoch": 1.4585752155967922, + "grad_norm": 2.721203088760376, + "learning_rate": 1.801304633841477e-05, + "loss": 0.5908, + "step": 21734 + }, + { + "epoch": 1.458709439280561, + "grad_norm": 1.1897770166397095, + "learning_rate": 1.8004693798965404e-05, + "loss": 0.5511, + "step": 21736 + }, + { + "epoch": 1.45884366296433, + "grad_norm": 1.0730276107788086, + "learning_rate": 1.799634277123134e-05, + "loss": 0.474, + "step": 21738 + }, + { + "epoch": 1.4589778866480991, + "grad_norm": 1.1049566268920898, + "learning_rate": 1.7987993255607132e-05, + "loss": 0.5343, + "step": 21740 + }, + { + "epoch": 1.459112110331868, + "grad_norm": 0.9912712574005127, + "learning_rate": 1.797964525248731e-05, + "loss": 0.4348, + "step": 21742 + }, + { + "epoch": 1.4592463340156372, + "grad_norm": 1.0087511539459229, + "learning_rate": 1.7971298762266287e-05, + "loss": 0.5086, + "step": 21744 + }, + { + "epoch": 1.4593805576994061, + "grad_norm": 1.1275442838668823, + "learning_rate": 1.796295378533841e-05, + "loss": 0.4361, + "step": 21746 + }, + { + "epoch": 1.459514781383175, + "grad_norm": 0.9750168323516846, + "learning_rate": 1.7954610322097953e-05, + "loss": 0.4603, + "step": 21748 + }, + { + "epoch": 1.459649005066944, + "grad_norm": 1.0766751766204834, + "learning_rate": 1.794626837293916e-05, + "loss": 0.503, + "step": 21750 + }, + { + "epoch": 1.459783228750713, + "grad_norm": 0.9707712531089783, + "learning_rate": 1.7937927938256143e-05, + "loss": 0.4528, + "step": 21752 + }, + { + "epoch": 1.459917452434482, + "grad_norm": 1.1949578523635864, + "learning_rate": 1.7929589018443016e-05, + "loss": 0.4502, + "step": 21754 + }, + { + "epoch": 1.4600516761182512, + "grad_norm": 1.1604313850402832, + "learning_rate": 1.7921251613893715e-05, + "loss": 0.5451, + "step": 21756 + }, + { + "epoch": 1.46018589980202, + "grad_norm": 1.018720269203186, + "learning_rate": 1.7912915725002205e-05, + "loss": 0.4937, + "step": 21758 + }, + { + "epoch": 1.460320123485789, + "grad_norm": 0.9990526437759399, + "learning_rate": 1.7904581352162312e-05, + "loss": 0.529, + "step": 21760 + }, + { + "epoch": 1.4604543471695581, + "grad_norm": 0.8207942247390747, + "learning_rate": 1.789624849576786e-05, + "loss": 0.4089, + "step": 21762 + }, + { + "epoch": 1.460588570853327, + "grad_norm": 1.1114898920059204, + "learning_rate": 1.7887917156212532e-05, + "loss": 0.5099, + "step": 21764 + }, + { + "epoch": 1.4607227945370962, + "grad_norm": 1.045689582824707, + "learning_rate": 1.7879587333889975e-05, + "loss": 0.4647, + "step": 21766 + }, + { + "epoch": 1.4608570182208651, + "grad_norm": 1.1530375480651855, + "learning_rate": 1.7871259029193754e-05, + "loss": 0.4636, + "step": 21768 + }, + { + "epoch": 1.460991241904634, + "grad_norm": 1.0641634464263916, + "learning_rate": 1.786293224251735e-05, + "loss": 0.456, + "step": 21770 + }, + { + "epoch": 1.461125465588403, + "grad_norm": 0.8483332991600037, + "learning_rate": 1.785460697425422e-05, + "loss": 0.4027, + "step": 21772 + }, + { + "epoch": 1.461259689272172, + "grad_norm": 1.3350237607955933, + "learning_rate": 1.7846283224797698e-05, + "loss": 0.4661, + "step": 21774 + }, + { + "epoch": 1.461393912955941, + "grad_norm": 1.0425379276275635, + "learning_rate": 1.7837960994541063e-05, + "loss": 0.4859, + "step": 21776 + }, + { + "epoch": 1.4615281366397102, + "grad_norm": 1.0305285453796387, + "learning_rate": 1.7829640283877514e-05, + "loss": 0.4955, + "step": 21778 + }, + { + "epoch": 1.461662360323479, + "grad_norm": 0.9458545446395874, + "learning_rate": 1.7821321093200217e-05, + "loss": 0.4683, + "step": 21780 + }, + { + "epoch": 1.461796584007248, + "grad_norm": 0.9453576803207397, + "learning_rate": 1.7813003422902224e-05, + "loss": 0.4282, + "step": 21782 + }, + { + "epoch": 1.4619308076910171, + "grad_norm": 0.9296939969062805, + "learning_rate": 1.7804687273376526e-05, + "loss": 0.425, + "step": 21784 + }, + { + "epoch": 1.462065031374786, + "grad_norm": 1.0925058126449585, + "learning_rate": 1.7796372645016024e-05, + "loss": 0.5591, + "step": 21786 + }, + { + "epoch": 1.4621992550585552, + "grad_norm": 1.0471255779266357, + "learning_rate": 1.778805953821361e-05, + "loss": 0.4467, + "step": 21788 + }, + { + "epoch": 1.462333478742324, + "grad_norm": 0.9129785299301147, + "learning_rate": 1.777974795336202e-05, + "loss": 0.5041, + "step": 21790 + }, + { + "epoch": 1.462467702426093, + "grad_norm": 0.9907031655311584, + "learning_rate": 1.7771437890854e-05, + "loss": 0.4398, + "step": 21792 + }, + { + "epoch": 1.462601926109862, + "grad_norm": 1.1017745733261108, + "learning_rate": 1.7763129351082165e-05, + "loss": 0.4732, + "step": 21794 + }, + { + "epoch": 1.462736149793631, + "grad_norm": 1.0758988857269287, + "learning_rate": 1.7754822334439075e-05, + "loss": 0.4765, + "step": 21796 + }, + { + "epoch": 1.4628703734774, + "grad_norm": 0.9974637627601624, + "learning_rate": 1.7746516841317207e-05, + "loss": 0.4644, + "step": 21798 + }, + { + "epoch": 1.4630045971611692, + "grad_norm": 0.8979473114013672, + "learning_rate": 1.773821287210901e-05, + "loss": 0.4921, + "step": 21800 + }, + { + "epoch": 1.463138820844938, + "grad_norm": 1.1205793619155884, + "learning_rate": 1.772991042720682e-05, + "loss": 0.5132, + "step": 21802 + }, + { + "epoch": 1.463273044528707, + "grad_norm": 0.9702668190002441, + "learning_rate": 1.77216095070029e-05, + "loss": 0.4651, + "step": 21804 + }, + { + "epoch": 1.4634072682124761, + "grad_norm": 1.0302200317382812, + "learning_rate": 1.7713310111889443e-05, + "loss": 0.5258, + "step": 21806 + }, + { + "epoch": 1.463541491896245, + "grad_norm": 0.9386501312255859, + "learning_rate": 1.7705012242258613e-05, + "loss": 0.5057, + "step": 21808 + }, + { + "epoch": 1.4636757155800142, + "grad_norm": 0.9929801225662231, + "learning_rate": 1.769671589850243e-05, + "loss": 0.4907, + "step": 21810 + }, + { + "epoch": 1.463809939263783, + "grad_norm": 0.9299373626708984, + "learning_rate": 1.768842108101293e-05, + "loss": 0.4233, + "step": 21812 + }, + { + "epoch": 1.463944162947552, + "grad_norm": 0.9028799533843994, + "learning_rate": 1.7680127790181967e-05, + "loss": 0.451, + "step": 21814 + }, + { + "epoch": 1.4640783866313212, + "grad_norm": 1.065677523612976, + "learning_rate": 1.767183602640143e-05, + "loss": 0.5507, + "step": 21816 + }, + { + "epoch": 1.46421261031509, + "grad_norm": 0.7841746211051941, + "learning_rate": 1.7663545790063047e-05, + "loss": 0.4868, + "step": 21818 + }, + { + "epoch": 1.4643468339988592, + "grad_norm": 0.9826595783233643, + "learning_rate": 1.765525708155856e-05, + "loss": 0.4955, + "step": 21820 + }, + { + "epoch": 1.4644810576826282, + "grad_norm": 0.8774657249450684, + "learning_rate": 1.764696990127957e-05, + "loss": 0.4306, + "step": 21822 + }, + { + "epoch": 1.464615281366397, + "grad_norm": 0.9303463697433472, + "learning_rate": 1.7638684249617632e-05, + "loss": 0.4163, + "step": 21824 + }, + { + "epoch": 1.464749505050166, + "grad_norm": 1.0600500106811523, + "learning_rate": 1.7630400126964214e-05, + "loss": 0.4468, + "step": 21826 + }, + { + "epoch": 1.4648837287339351, + "grad_norm": 1.0781666040420532, + "learning_rate": 1.7622117533710752e-05, + "loss": 0.4558, + "step": 21828 + }, + { + "epoch": 1.465017952417704, + "grad_norm": 0.9212883710861206, + "learning_rate": 1.7613836470248575e-05, + "loss": 0.4128, + "step": 21830 + }, + { + "epoch": 1.4651521761014732, + "grad_norm": 1.0324337482452393, + "learning_rate": 1.760555693696893e-05, + "loss": 0.4796, + "step": 21832 + }, + { + "epoch": 1.465286399785242, + "grad_norm": 0.9544461369514465, + "learning_rate": 1.7597278934263007e-05, + "loss": 0.5213, + "step": 21834 + }, + { + "epoch": 1.465420623469011, + "grad_norm": 1.010330080986023, + "learning_rate": 1.7589002462521954e-05, + "loss": 0.4643, + "step": 21836 + }, + { + "epoch": 1.4655548471527802, + "grad_norm": 0.9494339227676392, + "learning_rate": 1.7580727522136804e-05, + "loss": 0.4516, + "step": 21838 + }, + { + "epoch": 1.465689070836549, + "grad_norm": 1.0267274379730225, + "learning_rate": 1.757245411349852e-05, + "loss": 0.5265, + "step": 21840 + }, + { + "epoch": 1.4658232945203182, + "grad_norm": 1.125027060508728, + "learning_rate": 1.7564182236998024e-05, + "loss": 0.5227, + "step": 21842 + }, + { + "epoch": 1.4659575182040872, + "grad_norm": 1.034874439239502, + "learning_rate": 1.755591189302611e-05, + "loss": 0.5408, + "step": 21844 + }, + { + "epoch": 1.466091741887856, + "grad_norm": 1.0111058950424194, + "learning_rate": 1.754764308197358e-05, + "loss": 0.5175, + "step": 21846 + }, + { + "epoch": 1.466225965571625, + "grad_norm": 0.921784520149231, + "learning_rate": 1.7539375804231083e-05, + "loss": 0.4944, + "step": 21848 + }, + { + "epoch": 1.4663601892553941, + "grad_norm": 1.0426305532455444, + "learning_rate": 1.7531110060189283e-05, + "loss": 0.5336, + "step": 21850 + }, + { + "epoch": 1.466494412939163, + "grad_norm": 1.0855768918991089, + "learning_rate": 1.752284585023865e-05, + "loss": 0.4675, + "step": 21852 + }, + { + "epoch": 1.4666286366229322, + "grad_norm": 1.0991650819778442, + "learning_rate": 1.75145831747697e-05, + "loss": 0.4888, + "step": 21854 + }, + { + "epoch": 1.466762860306701, + "grad_norm": 1.0111898183822632, + "learning_rate": 1.7506322034172808e-05, + "loss": 0.537, + "step": 21856 + }, + { + "epoch": 1.46689708399047, + "grad_norm": 0.8977757096290588, + "learning_rate": 1.7498062428838314e-05, + "loss": 0.4458, + "step": 21858 + }, + { + "epoch": 1.4670313076742392, + "grad_norm": 0.9779512882232666, + "learning_rate": 1.7489804359156458e-05, + "loss": 0.5115, + "step": 21860 + }, + { + "epoch": 1.467165531358008, + "grad_norm": 1.0374683141708374, + "learning_rate": 1.748154782551742e-05, + "loss": 0.5079, + "step": 21862 + }, + { + "epoch": 1.4672997550417772, + "grad_norm": 1.0324270725250244, + "learning_rate": 1.7473292828311282e-05, + "loss": 0.5083, + "step": 21864 + }, + { + "epoch": 1.4674339787255462, + "grad_norm": 1.0627511739730835, + "learning_rate": 1.7465039367928117e-05, + "loss": 0.476, + "step": 21866 + }, + { + "epoch": 1.467568202409315, + "grad_norm": 1.0572152137756348, + "learning_rate": 1.745678744475786e-05, + "loss": 0.4595, + "step": 21868 + }, + { + "epoch": 1.467702426093084, + "grad_norm": 1.024880051612854, + "learning_rate": 1.7448537059190407e-05, + "loss": 0.4777, + "step": 21870 + }, + { + "epoch": 1.4678366497768531, + "grad_norm": 1.0588107109069824, + "learning_rate": 1.7440288211615553e-05, + "loss": 0.5856, + "step": 21872 + }, + { + "epoch": 1.467970873460622, + "grad_norm": 1.052733302116394, + "learning_rate": 1.743204090242307e-05, + "loss": 0.4873, + "step": 21874 + }, + { + "epoch": 1.4681050971443912, + "grad_norm": 1.1226145029067993, + "learning_rate": 1.742379513200259e-05, + "loss": 0.536, + "step": 21876 + }, + { + "epoch": 1.46823932082816, + "grad_norm": 0.8805328011512756, + "learning_rate": 1.741555090074377e-05, + "loss": 0.4685, + "step": 21878 + }, + { + "epoch": 1.468373544511929, + "grad_norm": 0.8982160687446594, + "learning_rate": 1.7407308209036066e-05, + "loss": 0.4502, + "step": 21880 + }, + { + "epoch": 1.4685077681956982, + "grad_norm": 0.996573269367218, + "learning_rate": 1.739906705726897e-05, + "loss": 0.492, + "step": 21882 + }, + { + "epoch": 1.468641991879467, + "grad_norm": 1.0391132831573486, + "learning_rate": 1.739082744583183e-05, + "loss": 0.4687, + "step": 21884 + }, + { + "epoch": 1.4687762155632362, + "grad_norm": 0.8928671479225159, + "learning_rate": 1.738258937511399e-05, + "loss": 0.4398, + "step": 21886 + }, + { + "epoch": 1.4689104392470052, + "grad_norm": 1.0958343744277954, + "learning_rate": 1.737435284550466e-05, + "loss": 0.4913, + "step": 21888 + }, + { + "epoch": 1.469044662930774, + "grad_norm": 1.0855900049209595, + "learning_rate": 1.7366117857393007e-05, + "loss": 0.4774, + "step": 21890 + }, + { + "epoch": 1.4691788866145432, + "grad_norm": 0.9509583115577698, + "learning_rate": 1.7357884411168097e-05, + "loss": 0.4433, + "step": 21892 + }, + { + "epoch": 1.4693131102983121, + "grad_norm": 1.227495551109314, + "learning_rate": 1.734965250721897e-05, + "loss": 0.514, + "step": 21894 + }, + { + "epoch": 1.4694473339820813, + "grad_norm": 0.9190131425857544, + "learning_rate": 1.7341422145934568e-05, + "loss": 0.4864, + "step": 21896 + }, + { + "epoch": 1.4695815576658502, + "grad_norm": 1.0067321062088013, + "learning_rate": 1.7333193327703738e-05, + "loss": 0.4545, + "step": 21898 + }, + { + "epoch": 1.469715781349619, + "grad_norm": 0.9455028772354126, + "learning_rate": 1.7324966052915274e-05, + "loss": 0.4393, + "step": 21900 + }, + { + "epoch": 1.469850005033388, + "grad_norm": 1.0392711162567139, + "learning_rate": 1.7316740321957937e-05, + "loss": 0.5109, + "step": 21902 + }, + { + "epoch": 1.4699842287171572, + "grad_norm": 1.030715823173523, + "learning_rate": 1.7308516135220325e-05, + "loss": 0.5246, + "step": 21904 + }, + { + "epoch": 1.470118452400926, + "grad_norm": 0.9867245554924011, + "learning_rate": 1.7300293493091074e-05, + "loss": 0.4647, + "step": 21906 + }, + { + "epoch": 1.4702526760846952, + "grad_norm": 0.9754754900932312, + "learning_rate": 1.7292072395958643e-05, + "loss": 0.4182, + "step": 21908 + }, + { + "epoch": 1.4703868997684642, + "grad_norm": 1.0586143732070923, + "learning_rate": 1.728385284421145e-05, + "loss": 0.5041, + "step": 21910 + }, + { + "epoch": 1.470521123452233, + "grad_norm": 1.1397920846939087, + "learning_rate": 1.7275634838237897e-05, + "loss": 0.5264, + "step": 21912 + }, + { + "epoch": 1.4706553471360022, + "grad_norm": 1.073852777481079, + "learning_rate": 1.7267418378426232e-05, + "loss": 0.4827, + "step": 21914 + }, + { + "epoch": 1.4707895708197711, + "grad_norm": 1.0655097961425781, + "learning_rate": 1.72592034651647e-05, + "loss": 0.5417, + "step": 21916 + }, + { + "epoch": 1.4709237945035403, + "grad_norm": 0.9804315567016602, + "learning_rate": 1.7250990098841417e-05, + "loss": 0.4795, + "step": 21918 + }, + { + "epoch": 1.4710580181873092, + "grad_norm": 0.9913179874420166, + "learning_rate": 1.7242778279844463e-05, + "loss": 0.509, + "step": 21920 + }, + { + "epoch": 1.471192241871078, + "grad_norm": 0.9221985936164856, + "learning_rate": 1.7234568008561797e-05, + "loss": 0.4097, + "step": 21922 + }, + { + "epoch": 1.471326465554847, + "grad_norm": 0.860504150390625, + "learning_rate": 1.722635928538138e-05, + "loss": 0.4928, + "step": 21924 + }, + { + "epoch": 1.4714606892386162, + "grad_norm": 1.0933024883270264, + "learning_rate": 1.7218152110691044e-05, + "loss": 0.4646, + "step": 21926 + }, + { + "epoch": 1.471594912922385, + "grad_norm": 0.9904744625091553, + "learning_rate": 1.7209946484878554e-05, + "loss": 0.4596, + "step": 21928 + }, + { + "epoch": 1.4717291366061542, + "grad_norm": 1.0993026494979858, + "learning_rate": 1.72017424083316e-05, + "loss": 0.4669, + "step": 21930 + }, + { + "epoch": 1.4718633602899232, + "grad_norm": 0.9404697418212891, + "learning_rate": 1.7193539881437837e-05, + "loss": 0.4174, + "step": 21932 + }, + { + "epoch": 1.471997583973692, + "grad_norm": 1.056062936782837, + "learning_rate": 1.71853389045848e-05, + "loss": 0.4875, + "step": 21934 + }, + { + "epoch": 1.4721318076574612, + "grad_norm": 0.9709733128547668, + "learning_rate": 1.7177139478159977e-05, + "loss": 0.4813, + "step": 21936 + }, + { + "epoch": 1.4722660313412301, + "grad_norm": 0.9787045121192932, + "learning_rate": 1.7168941602550754e-05, + "loss": 0.5001, + "step": 21938 + }, + { + "epoch": 1.4724002550249993, + "grad_norm": 1.0178070068359375, + "learning_rate": 1.7160745278144498e-05, + "loss": 0.4499, + "step": 21940 + }, + { + "epoch": 1.4725344787087682, + "grad_norm": 1.0335322618484497, + "learning_rate": 1.715255050532843e-05, + "loss": 0.4557, + "step": 21942 + }, + { + "epoch": 1.472668702392537, + "grad_norm": 1.0430026054382324, + "learning_rate": 1.7144357284489782e-05, + "loss": 0.4808, + "step": 21944 + }, + { + "epoch": 1.472802926076306, + "grad_norm": 1.0066791772842407, + "learning_rate": 1.7136165616015636e-05, + "loss": 0.516, + "step": 21946 + }, + { + "epoch": 1.4729371497600752, + "grad_norm": 1.0189493894577026, + "learning_rate": 1.712797550029305e-05, + "loss": 0.4766, + "step": 21948 + }, + { + "epoch": 1.473071373443844, + "grad_norm": 1.0610727071762085, + "learning_rate": 1.711978693770896e-05, + "loss": 0.5111, + "step": 21950 + }, + { + "epoch": 1.4732055971276132, + "grad_norm": 0.9555049538612366, + "learning_rate": 1.7111599928650302e-05, + "loss": 0.4599, + "step": 21952 + }, + { + "epoch": 1.4733398208113822, + "grad_norm": 1.0025465488433838, + "learning_rate": 1.710341447350387e-05, + "loss": 0.5389, + "step": 21954 + }, + { + "epoch": 1.473474044495151, + "grad_norm": 1.009305477142334, + "learning_rate": 1.7095230572656418e-05, + "loss": 0.515, + "step": 21956 + }, + { + "epoch": 1.4736082681789202, + "grad_norm": 0.9866924285888672, + "learning_rate": 1.70870482264946e-05, + "loss": 0.4518, + "step": 21958 + }, + { + "epoch": 1.4737424918626891, + "grad_norm": 1.1503498554229736, + "learning_rate": 1.707886743540505e-05, + "loss": 0.5341, + "step": 21960 + }, + { + "epoch": 1.4738767155464583, + "grad_norm": 1.3890480995178223, + "learning_rate": 1.7070688199774275e-05, + "loss": 0.4966, + "step": 21962 + }, + { + "epoch": 1.4740109392302272, + "grad_norm": 1.063452959060669, + "learning_rate": 1.7062510519988727e-05, + "loss": 0.4917, + "step": 21964 + }, + { + "epoch": 1.474145162913996, + "grad_norm": 1.2469446659088135, + "learning_rate": 1.705433439643478e-05, + "loss": 0.4613, + "step": 21966 + }, + { + "epoch": 1.4742793865977653, + "grad_norm": 1.0755219459533691, + "learning_rate": 1.704615982949876e-05, + "loss": 0.4906, + "step": 21968 + }, + { + "epoch": 1.4744136102815342, + "grad_norm": 1.1139458417892456, + "learning_rate": 1.703798681956687e-05, + "loss": 0.4987, + "step": 21970 + }, + { + "epoch": 1.4745478339653033, + "grad_norm": 0.9543841481208801, + "learning_rate": 1.7029815367025304e-05, + "loss": 0.4754, + "step": 21972 + }, + { + "epoch": 1.4746820576490722, + "grad_norm": 1.0988030433654785, + "learning_rate": 1.7021645472260146e-05, + "loss": 0.5489, + "step": 21974 + }, + { + "epoch": 1.4748162813328412, + "grad_norm": 0.9184737801551819, + "learning_rate": 1.701347713565735e-05, + "loss": 0.5104, + "step": 21976 + }, + { + "epoch": 1.47495050501661, + "grad_norm": 1.095395565032959, + "learning_rate": 1.7005310357602916e-05, + "loss": 0.4972, + "step": 21978 + }, + { + "epoch": 1.4750847287003792, + "grad_norm": 1.0471221208572388, + "learning_rate": 1.6997145138482674e-05, + "loss": 0.479, + "step": 21980 + }, + { + "epoch": 1.4752189523841481, + "grad_norm": 1.0793814659118652, + "learning_rate": 1.6988981478682436e-05, + "loss": 0.5019, + "step": 21982 + }, + { + "epoch": 1.4753531760679173, + "grad_norm": 0.9716320633888245, + "learning_rate": 1.6980819378587915e-05, + "loss": 0.4508, + "step": 21984 + }, + { + "epoch": 1.4754873997516862, + "grad_norm": 1.0645676851272583, + "learning_rate": 1.697265883858475e-05, + "loss": 0.5126, + "step": 21986 + }, + { + "epoch": 1.475621623435455, + "grad_norm": 1.1257740259170532, + "learning_rate": 1.6964499859058485e-05, + "loss": 0.525, + "step": 21988 + }, + { + "epoch": 1.4757558471192243, + "grad_norm": 1.1400830745697021, + "learning_rate": 1.6956342440394663e-05, + "loss": 0.4899, + "step": 21990 + }, + { + "epoch": 1.4758900708029932, + "grad_norm": 1.0845667123794556, + "learning_rate": 1.6948186582978683e-05, + "loss": 0.5192, + "step": 21992 + }, + { + "epoch": 1.4760242944867623, + "grad_norm": 0.9828360080718994, + "learning_rate": 1.694003228719589e-05, + "loss": 0.4691, + "step": 21994 + }, + { + "epoch": 1.4761585181705312, + "grad_norm": 1.1349446773529053, + "learning_rate": 1.693187955343155e-05, + "loss": 0.5609, + "step": 21996 + }, + { + "epoch": 1.4762927418543002, + "grad_norm": 0.8819553852081299, + "learning_rate": 1.6923728382070886e-05, + "loss": 0.4321, + "step": 21998 + }, + { + "epoch": 1.476426965538069, + "grad_norm": 0.9215646386146545, + "learning_rate": 1.6915578773499003e-05, + "loss": 0.5229, + "step": 22000 + }, + { + "epoch": 1.4765611892218382, + "grad_norm": 1.0390326976776123, + "learning_rate": 1.6907430728101e-05, + "loss": 0.4688, + "step": 22002 + }, + { + "epoch": 1.4766954129056071, + "grad_norm": 0.9488692879676819, + "learning_rate": 1.689928424626178e-05, + "loss": 0.5094, + "step": 22004 + }, + { + "epoch": 1.4768296365893763, + "grad_norm": 1.4182480573654175, + "learning_rate": 1.6891139328366313e-05, + "loss": 0.5009, + "step": 22006 + }, + { + "epoch": 1.4769638602731452, + "grad_norm": 1.041573405265808, + "learning_rate": 1.6882995974799387e-05, + "loss": 0.474, + "step": 22008 + }, + { + "epoch": 1.477098083956914, + "grad_norm": 1.0602591037750244, + "learning_rate": 1.68748541859458e-05, + "loss": 0.496, + "step": 22010 + }, + { + "epoch": 1.4772323076406833, + "grad_norm": 0.9753900170326233, + "learning_rate": 1.686671396219021e-05, + "loss": 0.5225, + "step": 22012 + }, + { + "epoch": 1.4773665313244522, + "grad_norm": 1.1660841703414917, + "learning_rate": 1.6858575303917234e-05, + "loss": 0.4788, + "step": 22014 + }, + { + "epoch": 1.4775007550082213, + "grad_norm": 0.9335696697235107, + "learning_rate": 1.685043821151139e-05, + "loss": 0.4973, + "step": 22016 + }, + { + "epoch": 1.4776349786919902, + "grad_norm": 1.10598886013031, + "learning_rate": 1.6842302685357165e-05, + "loss": 0.4726, + "step": 22018 + }, + { + "epoch": 1.4777692023757591, + "grad_norm": 0.9315410256385803, + "learning_rate": 1.683416872583894e-05, + "loss": 0.4898, + "step": 22020 + }, + { + "epoch": 1.477903426059528, + "grad_norm": 1.0776035785675049, + "learning_rate": 1.6826036333341027e-05, + "loss": 0.5003, + "step": 22022 + }, + { + "epoch": 1.4780376497432972, + "grad_norm": 0.8986199498176575, + "learning_rate": 1.6817905508247643e-05, + "loss": 0.5129, + "step": 22024 + }, + { + "epoch": 1.4781718734270661, + "grad_norm": 1.0102131366729736, + "learning_rate": 1.6809776250942995e-05, + "loss": 0.4638, + "step": 22026 + }, + { + "epoch": 1.4783060971108353, + "grad_norm": 1.0412551164627075, + "learning_rate": 1.6801648561811133e-05, + "loss": 0.4676, + "step": 22028 + }, + { + "epoch": 1.4784403207946042, + "grad_norm": 0.9449609518051147, + "learning_rate": 1.679352244123613e-05, + "loss": 0.5366, + "step": 22030 + }, + { + "epoch": 1.478574544478373, + "grad_norm": 1.1150097846984863, + "learning_rate": 1.678539788960186e-05, + "loss": 0.4236, + "step": 22032 + }, + { + "epoch": 1.4787087681621423, + "grad_norm": 1.0405479669570923, + "learning_rate": 1.6777274907292245e-05, + "loss": 0.5033, + "step": 22034 + }, + { + "epoch": 1.4788429918459112, + "grad_norm": 1.066287875175476, + "learning_rate": 1.6769153494691043e-05, + "loss": 0.5122, + "step": 22036 + }, + { + "epoch": 1.4789772155296803, + "grad_norm": 1.1990296840667725, + "learning_rate": 1.6761033652182008e-05, + "loss": 0.5075, + "step": 22038 + }, + { + "epoch": 1.4791114392134492, + "grad_norm": 1.2186036109924316, + "learning_rate": 1.675291538014877e-05, + "loss": 0.4539, + "step": 22040 + }, + { + "epoch": 1.4792456628972181, + "grad_norm": 1.1169241666793823, + "learning_rate": 1.6744798678974906e-05, + "loss": 0.4672, + "step": 22042 + }, + { + "epoch": 1.4793798865809873, + "grad_norm": 1.2180031538009644, + "learning_rate": 1.673668354904391e-05, + "loss": 0.4853, + "step": 22044 + }, + { + "epoch": 1.4795141102647562, + "grad_norm": 1.022362232208252, + "learning_rate": 1.672856999073919e-05, + "loss": 0.4604, + "step": 22046 + }, + { + "epoch": 1.4796483339485254, + "grad_norm": 2.732006311416626, + "learning_rate": 1.672045800444413e-05, + "loss": 0.4882, + "step": 22048 + }, + { + "epoch": 1.4797825576322943, + "grad_norm": 1.036034345626831, + "learning_rate": 1.6712347590541983e-05, + "loss": 0.5039, + "step": 22050 + }, + { + "epoch": 1.4799167813160632, + "grad_norm": 0.9300113916397095, + "learning_rate": 1.6704238749415957e-05, + "loss": 0.4571, + "step": 22052 + }, + { + "epoch": 1.480051004999832, + "grad_norm": 1.0871950387954712, + "learning_rate": 1.6696131481449162e-05, + "loss": 0.5488, + "step": 22054 + }, + { + "epoch": 1.4801852286836013, + "grad_norm": 1.0557066202163696, + "learning_rate": 1.668802578702468e-05, + "loss": 0.477, + "step": 22056 + }, + { + "epoch": 1.4803194523673702, + "grad_norm": 1.0013145208358765, + "learning_rate": 1.6679921666525476e-05, + "loss": 0.4437, + "step": 22058 + }, + { + "epoch": 1.4804536760511393, + "grad_norm": 0.9436190128326416, + "learning_rate": 1.6671819120334454e-05, + "loss": 0.4432, + "step": 22060 + }, + { + "epoch": 1.4805878997349082, + "grad_norm": 1.2498823404312134, + "learning_rate": 1.6663718148834423e-05, + "loss": 0.5072, + "step": 22062 + }, + { + "epoch": 1.4807221234186771, + "grad_norm": 1.0579122304916382, + "learning_rate": 1.6655618752408176e-05, + "loss": 0.408, + "step": 22064 + }, + { + "epoch": 1.4808563471024463, + "grad_norm": 1.0340778827667236, + "learning_rate": 1.6647520931438355e-05, + "loss": 0.5089, + "step": 22066 + }, + { + "epoch": 1.4809905707862152, + "grad_norm": 1.081233263015747, + "learning_rate": 1.6639424686307613e-05, + "loss": 0.5374, + "step": 22068 + }, + { + "epoch": 1.4811247944699844, + "grad_norm": 0.8616448640823364, + "learning_rate": 1.663133001739843e-05, + "loss": 0.4137, + "step": 22070 + }, + { + "epoch": 1.4812590181537533, + "grad_norm": 0.9921708106994629, + "learning_rate": 1.6623236925093293e-05, + "loss": 0.5115, + "step": 22072 + }, + { + "epoch": 1.4813932418375222, + "grad_norm": 0.9645135402679443, + "learning_rate": 1.661514540977457e-05, + "loss": 0.4393, + "step": 22074 + }, + { + "epoch": 1.481527465521291, + "grad_norm": 1.0653139352798462, + "learning_rate": 1.6607055471824595e-05, + "loss": 0.4785, + "step": 22076 + }, + { + "epoch": 1.4816616892050603, + "grad_norm": 0.9894845485687256, + "learning_rate": 1.659896711162558e-05, + "loss": 0.4881, + "step": 22078 + }, + { + "epoch": 1.4817959128888292, + "grad_norm": 0.9394354820251465, + "learning_rate": 1.6590880329559692e-05, + "loss": 0.531, + "step": 22080 + }, + { + "epoch": 1.4819301365725983, + "grad_norm": 0.9779031872749329, + "learning_rate": 1.658279512600899e-05, + "loss": 0.4629, + "step": 22082 + }, + { + "epoch": 1.4820643602563672, + "grad_norm": 1.0840893983840942, + "learning_rate": 1.6574711501355534e-05, + "loss": 0.4881, + "step": 22084 + }, + { + "epoch": 1.4821985839401361, + "grad_norm": 1.404558777809143, + "learning_rate": 1.6566629455981218e-05, + "loss": 0.4181, + "step": 22086 + }, + { + "epoch": 1.4823328076239053, + "grad_norm": 0.9741535186767578, + "learning_rate": 1.6558548990267925e-05, + "loss": 0.4326, + "step": 22088 + }, + { + "epoch": 1.4824670313076742, + "grad_norm": 0.9967256784439087, + "learning_rate": 1.6550470104597417e-05, + "loss": 0.4607, + "step": 22090 + }, + { + "epoch": 1.4826012549914434, + "grad_norm": 1.3023165464401245, + "learning_rate": 1.654239279935143e-05, + "loss": 0.4302, + "step": 22092 + }, + { + "epoch": 1.4827354786752123, + "grad_norm": 0.9432088732719421, + "learning_rate": 1.6534317074911582e-05, + "loss": 0.4792, + "step": 22094 + }, + { + "epoch": 1.4828697023589812, + "grad_norm": 1.0441328287124634, + "learning_rate": 1.652624293165946e-05, + "loss": 0.4332, + "step": 22096 + }, + { + "epoch": 1.48300392604275, + "grad_norm": 1.0635572671890259, + "learning_rate": 1.6518170369976532e-05, + "loss": 0.466, + "step": 22098 + }, + { + "epoch": 1.4831381497265193, + "grad_norm": 0.9838083386421204, + "learning_rate": 1.651009939024422e-05, + "loss": 0.4276, + "step": 22100 + }, + { + "epoch": 1.4832723734102882, + "grad_norm": 1.1172091960906982, + "learning_rate": 1.6502029992843836e-05, + "loss": 0.4367, + "step": 22102 + }, + { + "epoch": 1.4834065970940573, + "grad_norm": 1.242810606956482, + "learning_rate": 1.6493962178156685e-05, + "loss": 0.4783, + "step": 22104 + }, + { + "epoch": 1.4835408207778262, + "grad_norm": 1.1038495302200317, + "learning_rate": 1.6485895946563927e-05, + "loss": 0.4885, + "step": 22106 + }, + { + "epoch": 1.4836750444615951, + "grad_norm": 0.980767548084259, + "learning_rate": 1.647783129844669e-05, + "loss": 0.5184, + "step": 22108 + }, + { + "epoch": 1.4838092681453643, + "grad_norm": 0.9729624390602112, + "learning_rate": 1.646976823418599e-05, + "loss": 0.5311, + "step": 22110 + }, + { + "epoch": 1.4839434918291332, + "grad_norm": 1.0298727750778198, + "learning_rate": 1.646170675416282e-05, + "loss": 0.4504, + "step": 22112 + }, + { + "epoch": 1.4840777155129024, + "grad_norm": 1.0202456712722778, + "learning_rate": 1.6453646858758055e-05, + "loss": 0.5316, + "step": 22114 + }, + { + "epoch": 1.4842119391966713, + "grad_norm": 1.0979762077331543, + "learning_rate": 1.644558854835251e-05, + "loss": 0.5239, + "step": 22116 + }, + { + "epoch": 1.4843461628804402, + "grad_norm": 0.9957572221755981, + "learning_rate": 1.6437531823326922e-05, + "loss": 0.4982, + "step": 22118 + }, + { + "epoch": 1.4844803865642093, + "grad_norm": 0.9726836085319519, + "learning_rate": 1.6429476684061945e-05, + "loss": 0.4833, + "step": 22120 + }, + { + "epoch": 1.4846146102479783, + "grad_norm": 1.643353819847107, + "learning_rate": 1.6421423130938196e-05, + "loss": 0.4703, + "step": 22122 + }, + { + "epoch": 1.4847488339317474, + "grad_norm": 0.9165441989898682, + "learning_rate": 1.6413371164336154e-05, + "loss": 0.4287, + "step": 22124 + }, + { + "epoch": 1.4848830576155163, + "grad_norm": 0.9640105366706848, + "learning_rate": 1.6405320784636318e-05, + "loss": 0.4866, + "step": 22126 + }, + { + "epoch": 1.4850172812992852, + "grad_norm": 0.8589420318603516, + "learning_rate": 1.6397271992218982e-05, + "loss": 0.4146, + "step": 22128 + }, + { + "epoch": 1.4851515049830541, + "grad_norm": 1.0952810049057007, + "learning_rate": 1.638922478746448e-05, + "loss": 0.5149, + "step": 22130 + }, + { + "epoch": 1.4852857286668233, + "grad_norm": 0.9134923815727234, + "learning_rate": 1.6381179170752997e-05, + "loss": 0.5017, + "step": 22132 + }, + { + "epoch": 1.4854199523505922, + "grad_norm": 0.9884053468704224, + "learning_rate": 1.6373135142464707e-05, + "loss": 0.4738, + "step": 22134 + }, + { + "epoch": 1.4855541760343614, + "grad_norm": 1.1389747858047485, + "learning_rate": 1.636509270297966e-05, + "loss": 0.4923, + "step": 22136 + }, + { + "epoch": 1.4856883997181303, + "grad_norm": 1.044872522354126, + "learning_rate": 1.635705185267784e-05, + "loss": 0.4902, + "step": 22138 + }, + { + "epoch": 1.4858226234018992, + "grad_norm": 1.0619035959243774, + "learning_rate": 1.634901259193915e-05, + "loss": 0.5227, + "step": 22140 + }, + { + "epoch": 1.4859568470856683, + "grad_norm": 0.9813964366912842, + "learning_rate": 1.6340974921143455e-05, + "loss": 0.4091, + "step": 22142 + }, + { + "epoch": 1.4860910707694373, + "grad_norm": 1.0402342081069946, + "learning_rate": 1.6332938840670508e-05, + "loss": 0.4614, + "step": 22144 + }, + { + "epoch": 1.4862252944532064, + "grad_norm": 0.9421043395996094, + "learning_rate": 1.63249043509e-05, + "loss": 0.5051, + "step": 22146 + }, + { + "epoch": 1.4863595181369753, + "grad_norm": 1.063668131828308, + "learning_rate": 1.6316871452211524e-05, + "loss": 0.4944, + "step": 22148 + }, + { + "epoch": 1.4864937418207442, + "grad_norm": 1.1434531211853027, + "learning_rate": 1.6308840144984656e-05, + "loss": 0.5422, + "step": 22150 + }, + { + "epoch": 1.4866279655045131, + "grad_norm": 0.9865716099739075, + "learning_rate": 1.6300810429598823e-05, + "loss": 0.497, + "step": 22152 + }, + { + "epoch": 1.4867621891882823, + "grad_norm": 1.030135989189148, + "learning_rate": 1.6292782306433462e-05, + "loss": 0.4521, + "step": 22154 + }, + { + "epoch": 1.4868964128720512, + "grad_norm": 1.0379871129989624, + "learning_rate": 1.628475577586782e-05, + "loss": 0.4689, + "step": 22156 + }, + { + "epoch": 1.4870306365558204, + "grad_norm": 1.1710013151168823, + "learning_rate": 1.627673083828119e-05, + "loss": 0.499, + "step": 22158 + }, + { + "epoch": 1.4871648602395893, + "grad_norm": 0.8846861124038696, + "learning_rate": 1.62687074940527e-05, + "loss": 0.4483, + "step": 22160 + }, + { + "epoch": 1.4872990839233582, + "grad_norm": 0.9259506464004517, + "learning_rate": 1.626068574356146e-05, + "loss": 0.4439, + "step": 22162 + }, + { + "epoch": 1.4874333076071273, + "grad_norm": 1.6477583646774292, + "learning_rate": 1.6252665587186477e-05, + "loss": 0.4561, + "step": 22164 + }, + { + "epoch": 1.4875675312908962, + "grad_norm": 0.9675223231315613, + "learning_rate": 1.624464702530668e-05, + "loss": 0.4391, + "step": 22166 + }, + { + "epoch": 1.4877017549746654, + "grad_norm": 0.9893090724945068, + "learning_rate": 1.6236630058300924e-05, + "loss": 0.4569, + "step": 22168 + }, + { + "epoch": 1.4878359786584343, + "grad_norm": 0.9624961614608765, + "learning_rate": 1.6228614686548018e-05, + "loss": 0.474, + "step": 22170 + }, + { + "epoch": 1.4879702023422032, + "grad_norm": 0.9343721270561218, + "learning_rate": 1.622060091042666e-05, + "loss": 0.481, + "step": 22172 + }, + { + "epoch": 1.4881044260259721, + "grad_norm": 1.0352492332458496, + "learning_rate": 1.621258873031548e-05, + "loss": 0.5404, + "step": 22174 + }, + { + "epoch": 1.4882386497097413, + "grad_norm": 1.141648769378662, + "learning_rate": 1.620457814659303e-05, + "loss": 0.4337, + "step": 22176 + }, + { + "epoch": 1.4883728733935102, + "grad_norm": 0.9643620848655701, + "learning_rate": 1.6196569159637825e-05, + "loss": 0.529, + "step": 22178 + }, + { + "epoch": 1.4885070970772794, + "grad_norm": 1.030743956565857, + "learning_rate": 1.6188561769828252e-05, + "loss": 0.422, + "step": 22180 + }, + { + "epoch": 1.4886413207610483, + "grad_norm": 1.0663295984268188, + "learning_rate": 1.6180555977542655e-05, + "loss": 0.5016, + "step": 22182 + }, + { + "epoch": 1.4887755444448172, + "grad_norm": 0.8622217774391174, + "learning_rate": 1.6172551783159278e-05, + "loss": 0.4585, + "step": 22184 + }, + { + "epoch": 1.4889097681285863, + "grad_norm": 1.099914312362671, + "learning_rate": 1.6164549187056294e-05, + "loss": 0.5106, + "step": 22186 + }, + { + "epoch": 1.4890439918123552, + "grad_norm": 0.9957973957061768, + "learning_rate": 1.6156548189611847e-05, + "loss": 0.4241, + "step": 22188 + }, + { + "epoch": 1.4891782154961244, + "grad_norm": 0.8730512857437134, + "learning_rate": 1.6148548791203926e-05, + "loss": 0.4678, + "step": 22190 + }, + { + "epoch": 1.4893124391798933, + "grad_norm": 1.048096776008606, + "learning_rate": 1.6140550992210545e-05, + "loss": 0.5086, + "step": 22192 + }, + { + "epoch": 1.4894466628636622, + "grad_norm": 0.9279083609580994, + "learning_rate": 1.6132554793009514e-05, + "loss": 0.4457, + "step": 22194 + }, + { + "epoch": 1.4895808865474314, + "grad_norm": 1.15582275390625, + "learning_rate": 1.612456019397869e-05, + "loss": 0.5107, + "step": 22196 + }, + { + "epoch": 1.4897151102312003, + "grad_norm": 1.0184177160263062, + "learning_rate": 1.6116567195495767e-05, + "loss": 0.5176, + "step": 22198 + }, + { + "epoch": 1.4898493339149694, + "grad_norm": 1.0488437414169312, + "learning_rate": 1.610857579793843e-05, + "loss": 0.4874, + "step": 22200 + }, + { + "epoch": 1.4899835575987384, + "grad_norm": 1.0049000978469849, + "learning_rate": 1.610058600168424e-05, + "loss": 0.4562, + "step": 22202 + }, + { + "epoch": 1.4901177812825073, + "grad_norm": 0.9642466306686401, + "learning_rate": 1.6092597807110703e-05, + "loss": 0.4392, + "step": 22204 + }, + { + "epoch": 1.4902520049662762, + "grad_norm": 1.002695918083191, + "learning_rate": 1.608461121459523e-05, + "loss": 0.4841, + "step": 22206 + }, + { + "epoch": 1.4903862286500453, + "grad_norm": 1.1459479331970215, + "learning_rate": 1.6076626224515196e-05, + "loss": 0.4775, + "step": 22208 + }, + { + "epoch": 1.4905204523338142, + "grad_norm": 1.008082628250122, + "learning_rate": 1.6068642837247872e-05, + "loss": 0.456, + "step": 22210 + }, + { + "epoch": 1.4906546760175834, + "grad_norm": 0.9226283431053162, + "learning_rate": 1.6060661053170444e-05, + "loss": 0.4293, + "step": 22212 + }, + { + "epoch": 1.4907888997013523, + "grad_norm": 1.091553807258606, + "learning_rate": 1.6052680872660032e-05, + "loss": 0.4298, + "step": 22214 + }, + { + "epoch": 1.4909231233851212, + "grad_norm": 1.1263148784637451, + "learning_rate": 1.6044702296093707e-05, + "loss": 0.5747, + "step": 22216 + }, + { + "epoch": 1.4910573470688904, + "grad_norm": 0.9613867998123169, + "learning_rate": 1.603672532384841e-05, + "loss": 0.4229, + "step": 22218 + }, + { + "epoch": 1.4911915707526593, + "grad_norm": 1.1696033477783203, + "learning_rate": 1.6028749956301094e-05, + "loss": 0.4472, + "step": 22220 + }, + { + "epoch": 1.4913257944364284, + "grad_norm": 0.8974775671958923, + "learning_rate": 1.6020776193828506e-05, + "loss": 0.4678, + "step": 22222 + }, + { + "epoch": 1.4914600181201974, + "grad_norm": 1.1368629932403564, + "learning_rate": 1.601280403680744e-05, + "loss": 0.5267, + "step": 22224 + }, + { + "epoch": 1.4915942418039663, + "grad_norm": 0.9700947999954224, + "learning_rate": 1.6004833485614536e-05, + "loss": 0.4524, + "step": 22226 + }, + { + "epoch": 1.4917284654877352, + "grad_norm": 1.0958696603775024, + "learning_rate": 1.5996864540626416e-05, + "loss": 0.5448, + "step": 22228 + }, + { + "epoch": 1.4918626891715043, + "grad_norm": 0.923024594783783, + "learning_rate": 1.5988897202219582e-05, + "loss": 0.4223, + "step": 22230 + }, + { + "epoch": 1.4919969128552732, + "grad_norm": 0.9782229065895081, + "learning_rate": 1.5980931470770476e-05, + "loss": 0.4957, + "step": 22232 + }, + { + "epoch": 1.4921311365390424, + "grad_norm": 1.0762877464294434, + "learning_rate": 1.5972967346655448e-05, + "loss": 0.483, + "step": 22234 + }, + { + "epoch": 1.4922653602228113, + "grad_norm": 1.0898056030273438, + "learning_rate": 1.5965004830250814e-05, + "loss": 0.5673, + "step": 22236 + }, + { + "epoch": 1.4923995839065802, + "grad_norm": 0.9333029985427856, + "learning_rate": 1.595704392193278e-05, + "loss": 0.5022, + "step": 22238 + }, + { + "epoch": 1.4925338075903494, + "grad_norm": 1.0515496730804443, + "learning_rate": 1.5949084622077472e-05, + "loss": 0.5338, + "step": 22240 + }, + { + "epoch": 1.4926680312741183, + "grad_norm": 0.9734745025634766, + "learning_rate": 1.5941126931060947e-05, + "loss": 0.482, + "step": 22242 + }, + { + "epoch": 1.4928022549578874, + "grad_norm": 0.9645838737487793, + "learning_rate": 1.593317084925921e-05, + "loss": 0.4808, + "step": 22244 + }, + { + "epoch": 1.4929364786416564, + "grad_norm": 1.1264073848724365, + "learning_rate": 1.5925216377048153e-05, + "loss": 0.4864, + "step": 22246 + }, + { + "epoch": 1.4930707023254253, + "grad_norm": 0.9437454342842102, + "learning_rate": 1.5917263514803643e-05, + "loss": 0.4782, + "step": 22248 + }, + { + "epoch": 1.4932049260091942, + "grad_norm": 0.9792948365211487, + "learning_rate": 1.5909312262901395e-05, + "loss": 0.469, + "step": 22250 + }, + { + "epoch": 1.4933391496929633, + "grad_norm": 1.0361741781234741, + "learning_rate": 1.590136262171709e-05, + "loss": 0.5108, + "step": 22252 + }, + { + "epoch": 1.4934733733767322, + "grad_norm": 1.7859095335006714, + "learning_rate": 1.5893414591626372e-05, + "loss": 0.4976, + "step": 22254 + }, + { + "epoch": 1.4936075970605014, + "grad_norm": 1.0242053270339966, + "learning_rate": 1.5885468173004726e-05, + "loss": 0.4742, + "step": 22256 + }, + { + "epoch": 1.4937418207442703, + "grad_norm": 1.0312488079071045, + "learning_rate": 1.5877523366227635e-05, + "loss": 0.5077, + "step": 22258 + }, + { + "epoch": 1.4938760444280392, + "grad_norm": 1.1025859117507935, + "learning_rate": 1.586958017167047e-05, + "loss": 0.4627, + "step": 22260 + }, + { + "epoch": 1.4940102681118084, + "grad_norm": 1.0530234575271606, + "learning_rate": 1.5861638589708534e-05, + "loss": 0.4536, + "step": 22262 + }, + { + "epoch": 1.4941444917955773, + "grad_norm": 1.0849812030792236, + "learning_rate": 1.585369862071702e-05, + "loss": 0.501, + "step": 22264 + }, + { + "epoch": 1.4942787154793464, + "grad_norm": 0.8678374886512756, + "learning_rate": 1.5845760265071125e-05, + "loss": 0.4688, + "step": 22266 + }, + { + "epoch": 1.4944129391631154, + "grad_norm": 0.9843668937683105, + "learning_rate": 1.583782352314589e-05, + "loss": 0.4795, + "step": 22268 + }, + { + "epoch": 1.4945471628468843, + "grad_norm": 1.0516777038574219, + "learning_rate": 1.5829888395316317e-05, + "loss": 0.4421, + "step": 22270 + }, + { + "epoch": 1.4946813865306534, + "grad_norm": 1.0073456764221191, + "learning_rate": 1.582195488195731e-05, + "loss": 0.4981, + "step": 22272 + }, + { + "epoch": 1.4948156102144223, + "grad_norm": 0.9583938121795654, + "learning_rate": 1.5814022983443743e-05, + "loss": 0.4775, + "step": 22274 + }, + { + "epoch": 1.4949498338981915, + "grad_norm": 1.054783582687378, + "learning_rate": 1.580609270015035e-05, + "loss": 0.5537, + "step": 22276 + }, + { + "epoch": 1.4950840575819604, + "grad_norm": 0.8921435475349426, + "learning_rate": 1.5798164032451867e-05, + "loss": 0.4924, + "step": 22278 + }, + { + "epoch": 1.4952182812657293, + "grad_norm": 1.042488694190979, + "learning_rate": 1.579023698072285e-05, + "loss": 0.4765, + "step": 22280 + }, + { + "epoch": 1.4953525049494982, + "grad_norm": 1.041841983795166, + "learning_rate": 1.578231154533788e-05, + "loss": 0.4815, + "step": 22282 + }, + { + "epoch": 1.4954867286332674, + "grad_norm": 1.0122069120407104, + "learning_rate": 1.5774387726671385e-05, + "loss": 0.4836, + "step": 22284 + }, + { + "epoch": 1.4956209523170363, + "grad_norm": 1.0777119398117065, + "learning_rate": 1.5766465525097784e-05, + "loss": 0.4443, + "step": 22286 + }, + { + "epoch": 1.4957551760008054, + "grad_norm": 0.9867185354232788, + "learning_rate": 1.575854494099137e-05, + "loss": 0.4634, + "step": 22288 + }, + { + "epoch": 1.4958893996845744, + "grad_norm": 1.0630228519439697, + "learning_rate": 1.575062597472638e-05, + "loss": 0.5764, + "step": 22290 + }, + { + "epoch": 1.4960236233683433, + "grad_norm": 1.1396502256393433, + "learning_rate": 1.5742708626676943e-05, + "loss": 0.4397, + "step": 22292 + }, + { + "epoch": 1.4961578470521124, + "grad_norm": 0.9992977976799011, + "learning_rate": 1.5734792897217178e-05, + "loss": 0.5039, + "step": 22294 + }, + { + "epoch": 1.4962920707358813, + "grad_norm": 1.0234202146530151, + "learning_rate": 1.5726878786721067e-05, + "loss": 0.4732, + "step": 22296 + }, + { + "epoch": 1.4964262944196505, + "grad_norm": 1.0589426755905151, + "learning_rate": 1.5718966295562538e-05, + "loss": 0.4807, + "step": 22298 + }, + { + "epoch": 1.4965605181034194, + "grad_norm": 0.9448699951171875, + "learning_rate": 1.5711055424115424e-05, + "loss": 0.4663, + "step": 22300 + }, + { + "epoch": 1.4966947417871883, + "grad_norm": 1.0552382469177246, + "learning_rate": 1.570314617275353e-05, + "loss": 0.4461, + "step": 22302 + }, + { + "epoch": 1.4968289654709572, + "grad_norm": 1.1160061359405518, + "learning_rate": 1.5695238541850526e-05, + "loss": 0.5051, + "step": 22304 + }, + { + "epoch": 1.4969631891547264, + "grad_norm": 0.9328937530517578, + "learning_rate": 1.568733253178007e-05, + "loss": 0.4966, + "step": 22306 + }, + { + "epoch": 1.4970974128384953, + "grad_norm": 0.9663674831390381, + "learning_rate": 1.5679428142915652e-05, + "loss": 0.4867, + "step": 22308 + }, + { + "epoch": 1.4972316365222644, + "grad_norm": 1.0076817274093628, + "learning_rate": 1.567152537563078e-05, + "loss": 0.527, + "step": 22310 + }, + { + "epoch": 1.4973658602060334, + "grad_norm": 0.8996386528015137, + "learning_rate": 1.566362423029881e-05, + "loss": 0.4739, + "step": 22312 + }, + { + "epoch": 1.4975000838898023, + "grad_norm": 1.0612578392028809, + "learning_rate": 1.5655724707293096e-05, + "loss": 0.4517, + "step": 22314 + }, + { + "epoch": 1.4976343075735714, + "grad_norm": 0.9685842990875244, + "learning_rate": 1.5647826806986853e-05, + "loss": 0.4856, + "step": 22316 + }, + { + "epoch": 1.4977685312573403, + "grad_norm": 1.0554558038711548, + "learning_rate": 1.5639930529753244e-05, + "loss": 0.4535, + "step": 22318 + }, + { + "epoch": 1.4979027549411095, + "grad_norm": 1.0142149925231934, + "learning_rate": 1.563203587596535e-05, + "loss": 0.4464, + "step": 22320 + }, + { + "epoch": 1.4980369786248784, + "grad_norm": 1.0849847793579102, + "learning_rate": 1.5624142845996165e-05, + "loss": 0.497, + "step": 22322 + }, + { + "epoch": 1.4981712023086473, + "grad_norm": 1.0530681610107422, + "learning_rate": 1.561625144021865e-05, + "loss": 0.5238, + "step": 22324 + }, + { + "epoch": 1.4983054259924162, + "grad_norm": 1.2183722257614136, + "learning_rate": 1.5608361659005637e-05, + "loss": 0.4877, + "step": 22326 + }, + { + "epoch": 1.4984396496761854, + "grad_norm": 1.033240795135498, + "learning_rate": 1.560047350272991e-05, + "loss": 0.511, + "step": 22328 + }, + { + "epoch": 1.4985738733599543, + "grad_norm": 0.9958361983299255, + "learning_rate": 1.559258697176415e-05, + "loss": 0.4966, + "step": 22330 + }, + { + "epoch": 1.4987080970437234, + "grad_norm": 0.8929667472839355, + "learning_rate": 1.5584702066481016e-05, + "loss": 0.3835, + "step": 22332 + }, + { + "epoch": 1.4988423207274923, + "grad_norm": 1.1883955001831055, + "learning_rate": 1.5576818787253027e-05, + "loss": 0.5171, + "step": 22334 + }, + { + "epoch": 1.4989765444112613, + "grad_norm": 1.052208423614502, + "learning_rate": 1.5568937134452664e-05, + "loss": 0.4719, + "step": 22336 + }, + { + "epoch": 1.4991107680950304, + "grad_norm": 0.8502485752105713, + "learning_rate": 1.5561057108452304e-05, + "loss": 0.4843, + "step": 22338 + }, + { + "epoch": 1.4992449917787993, + "grad_norm": 0.9813560247421265, + "learning_rate": 1.5553178709624283e-05, + "loss": 0.4513, + "step": 22340 + }, + { + "epoch": 1.4993792154625685, + "grad_norm": 1.0747052431106567, + "learning_rate": 1.5545301938340817e-05, + "loss": 0.5191, + "step": 22342 + }, + { + "epoch": 1.4995134391463374, + "grad_norm": 1.043516993522644, + "learning_rate": 1.553742679497412e-05, + "loss": 0.542, + "step": 22344 + }, + { + "epoch": 1.4996476628301063, + "grad_norm": 1.3962687253952026, + "learning_rate": 1.5529553279896202e-05, + "loss": 0.4693, + "step": 22346 + }, + { + "epoch": 1.4997818865138755, + "grad_norm": 1.1393555402755737, + "learning_rate": 1.5521681393479126e-05, + "loss": 0.5109, + "step": 22348 + }, + { + "epoch": 1.4999161101976444, + "grad_norm": 1.1023056507110596, + "learning_rate": 1.5513811136094787e-05, + "loss": 0.4991, + "step": 22350 + }, + { + "epoch": 1.5000503338814135, + "grad_norm": 1.081140160560608, + "learning_rate": 1.5505942508115073e-05, + "loss": 0.4659, + "step": 22352 + }, + { + "epoch": 1.5001845575651824, + "grad_norm": 1.0859403610229492, + "learning_rate": 1.5498075509911745e-05, + "loss": 0.4666, + "step": 22354 + }, + { + "epoch": 1.5003187812489513, + "grad_norm": 1.193873405456543, + "learning_rate": 1.5490210141856508e-05, + "loss": 0.5178, + "step": 22356 + }, + { + "epoch": 1.5004530049327203, + "grad_norm": 1.0530229806900024, + "learning_rate": 1.5482346404320967e-05, + "loss": 0.4997, + "step": 22358 + }, + { + "epoch": 1.5005872286164894, + "grad_norm": 1.0302584171295166, + "learning_rate": 1.5474484297676694e-05, + "loss": 0.4758, + "step": 22360 + }, + { + "epoch": 1.5007214523002583, + "grad_norm": 1.1433876752853394, + "learning_rate": 1.546662382229515e-05, + "loss": 0.5066, + "step": 22362 + }, + { + "epoch": 1.5008556759840275, + "grad_norm": 1.156192421913147, + "learning_rate": 1.5458764978547718e-05, + "loss": 0.5427, + "step": 22364 + }, + { + "epoch": 1.5009898996677964, + "grad_norm": 1.0290182828903198, + "learning_rate": 1.545090776680571e-05, + "loss": 0.4451, + "step": 22366 + }, + { + "epoch": 1.5011241233515653, + "grad_norm": 0.9857996702194214, + "learning_rate": 1.544305218744038e-05, + "loss": 0.4972, + "step": 22368 + }, + { + "epoch": 1.5012583470353342, + "grad_norm": 1.1398378610610962, + "learning_rate": 1.5435198240822873e-05, + "loss": 0.4968, + "step": 22370 + }, + { + "epoch": 1.5013925707191034, + "grad_norm": 0.9778451323509216, + "learning_rate": 1.5427345927324305e-05, + "loss": 0.4606, + "step": 22372 + }, + { + "epoch": 1.5015267944028725, + "grad_norm": 1.0325839519500732, + "learning_rate": 1.541949524731563e-05, + "loss": 0.5229, + "step": 22374 + }, + { + "epoch": 1.5016610180866414, + "grad_norm": 1.1207289695739746, + "learning_rate": 1.5411646201167817e-05, + "loss": 0.5016, + "step": 22376 + }, + { + "epoch": 1.5017952417704103, + "grad_norm": 1.1190552711486816, + "learning_rate": 1.5403798789251695e-05, + "loss": 0.5034, + "step": 22378 + }, + { + "epoch": 1.5019294654541793, + "grad_norm": 0.903392493724823, + "learning_rate": 1.5395953011938063e-05, + "loss": 0.4447, + "step": 22380 + }, + { + "epoch": 1.5020636891379484, + "grad_norm": 0.9386855959892273, + "learning_rate": 1.5388108869597605e-05, + "loss": 0.4992, + "step": 22382 + }, + { + "epoch": 1.5021979128217176, + "grad_norm": 0.9510099291801453, + "learning_rate": 1.5380266362600943e-05, + "loss": 0.4438, + "step": 22384 + }, + { + "epoch": 1.5023321365054865, + "grad_norm": 1.1823641061782837, + "learning_rate": 1.5372425491318615e-05, + "loss": 0.5579, + "step": 22386 + }, + { + "epoch": 1.5024663601892554, + "grad_norm": 1.0836889743804932, + "learning_rate": 1.5364586256121076e-05, + "loss": 0.5182, + "step": 22388 + }, + { + "epoch": 1.5026005838730243, + "grad_norm": 0.9870573878288269, + "learning_rate": 1.535674865737875e-05, + "loss": 0.4395, + "step": 22390 + }, + { + "epoch": 1.5027348075567935, + "grad_norm": 1.028084397315979, + "learning_rate": 1.534891269546192e-05, + "loss": 0.519, + "step": 22392 + }, + { + "epoch": 1.5028690312405624, + "grad_norm": 0.9285738468170166, + "learning_rate": 1.534107837074083e-05, + "loss": 0.4851, + "step": 22394 + }, + { + "epoch": 1.5030032549243315, + "grad_norm": 0.9531024098396301, + "learning_rate": 1.5333245683585618e-05, + "loss": 0.4493, + "step": 22396 + }, + { + "epoch": 1.5031374786081004, + "grad_norm": 0.9427856206893921, + "learning_rate": 1.5325414634366392e-05, + "loss": 0.4815, + "step": 22398 + }, + { + "epoch": 1.5032717022918693, + "grad_norm": 1.107208251953125, + "learning_rate": 1.531758522345314e-05, + "loss": 0.496, + "step": 22400 + }, + { + "epoch": 1.5034059259756383, + "grad_norm": 0.9453824162483215, + "learning_rate": 1.530975745121579e-05, + "loss": 0.5032, + "step": 22402 + }, + { + "epoch": 1.5035401496594074, + "grad_norm": 1.037557601928711, + "learning_rate": 1.5301931318024166e-05, + "loss": 0.5256, + "step": 22404 + }, + { + "epoch": 1.5036743733431766, + "grad_norm": 0.9742701649665833, + "learning_rate": 1.5294106824248065e-05, + "loss": 0.4578, + "step": 22406 + }, + { + "epoch": 1.5038085970269455, + "grad_norm": 1.0854389667510986, + "learning_rate": 1.5286283970257166e-05, + "loss": 0.4644, + "step": 22408 + }, + { + "epoch": 1.5039428207107144, + "grad_norm": 1.0835214853286743, + "learning_rate": 1.52784627564211e-05, + "loss": 0.4463, + "step": 22410 + }, + { + "epoch": 1.5040770443944833, + "grad_norm": 1.093213677406311, + "learning_rate": 1.527064318310939e-05, + "loss": 0.5273, + "step": 22412 + }, + { + "epoch": 1.5042112680782525, + "grad_norm": 1.007411003112793, + "learning_rate": 1.5262825250691497e-05, + "loss": 0.4554, + "step": 22414 + }, + { + "epoch": 1.5043454917620214, + "grad_norm": 1.095778465270996, + "learning_rate": 1.5255008959536787e-05, + "loss": 0.4905, + "step": 22416 + }, + { + "epoch": 1.5044797154457905, + "grad_norm": 1.0032175779342651, + "learning_rate": 1.5247194310014601e-05, + "loss": 0.4516, + "step": 22418 + }, + { + "epoch": 1.5046139391295594, + "grad_norm": 1.103121042251587, + "learning_rate": 1.5239381302494144e-05, + "loss": 0.4593, + "step": 22420 + }, + { + "epoch": 1.5047481628133283, + "grad_norm": 1.0171631574630737, + "learning_rate": 1.5231569937344564e-05, + "loss": 0.5937, + "step": 22422 + }, + { + "epoch": 1.5048823864970973, + "grad_norm": 1.1355721950531006, + "learning_rate": 1.5223760214934917e-05, + "loss": 0.5546, + "step": 22424 + }, + { + "epoch": 1.5050166101808664, + "grad_norm": 0.8735546469688416, + "learning_rate": 1.5215952135634237e-05, + "loss": 0.436, + "step": 22426 + }, + { + "epoch": 1.5051508338646356, + "grad_norm": 1.1134432554244995, + "learning_rate": 1.5208145699811415e-05, + "loss": 0.4899, + "step": 22428 + }, + { + "epoch": 1.5052850575484045, + "grad_norm": 1.0774012804031372, + "learning_rate": 1.5200340907835298e-05, + "loss": 0.4783, + "step": 22430 + }, + { + "epoch": 1.5054192812321734, + "grad_norm": 1.0947202444076538, + "learning_rate": 1.519253776007462e-05, + "loss": 0.462, + "step": 22432 + }, + { + "epoch": 1.5055535049159423, + "grad_norm": 0.9434705972671509, + "learning_rate": 1.5184736256898107e-05, + "loss": 0.474, + "step": 22434 + }, + { + "epoch": 1.5056877285997115, + "grad_norm": 1.147038221359253, + "learning_rate": 1.517693639867433e-05, + "loss": 0.5078, + "step": 22436 + }, + { + "epoch": 1.5058219522834804, + "grad_norm": 1.1056617498397827, + "learning_rate": 1.5169138185771841e-05, + "loss": 0.4686, + "step": 22438 + }, + { + "epoch": 1.5059561759672495, + "grad_norm": 0.9053165316581726, + "learning_rate": 1.5161341618559088e-05, + "loss": 0.4727, + "step": 22440 + }, + { + "epoch": 1.5060903996510184, + "grad_norm": 1.0582969188690186, + "learning_rate": 1.515354669740443e-05, + "loss": 0.4678, + "step": 22442 + }, + { + "epoch": 1.5062246233347873, + "grad_norm": 0.9727494120597839, + "learning_rate": 1.5145753422676157e-05, + "loss": 0.4288, + "step": 22444 + }, + { + "epoch": 1.5063588470185563, + "grad_norm": 0.9582500457763672, + "learning_rate": 1.5137961794742511e-05, + "loss": 0.4534, + "step": 22446 + }, + { + "epoch": 1.5064930707023254, + "grad_norm": 1.2127413749694824, + "learning_rate": 1.5130171813971617e-05, + "loss": 0.4995, + "step": 22448 + }, + { + "epoch": 1.5066272943860946, + "grad_norm": 0.9376811385154724, + "learning_rate": 1.5122383480731533e-05, + "loss": 0.459, + "step": 22450 + }, + { + "epoch": 1.5067615180698635, + "grad_norm": 1.0548505783081055, + "learning_rate": 1.5114596795390234e-05, + "loss": 0.4784, + "step": 22452 + }, + { + "epoch": 1.5068957417536324, + "grad_norm": 1.0130548477172852, + "learning_rate": 1.5106811758315654e-05, + "loss": 0.5134, + "step": 22454 + }, + { + "epoch": 1.5070299654374013, + "grad_norm": 1.0775259733200073, + "learning_rate": 1.50990283698756e-05, + "loss": 0.5642, + "step": 22456 + }, + { + "epoch": 1.5071641891211705, + "grad_norm": 1.0384955406188965, + "learning_rate": 1.5091246630437827e-05, + "loss": 0.5506, + "step": 22458 + }, + { + "epoch": 1.5072984128049396, + "grad_norm": 1.0623197555541992, + "learning_rate": 1.5083466540370006e-05, + "loss": 0.4947, + "step": 22460 + }, + { + "epoch": 1.5074326364887085, + "grad_norm": 1.0416051149368286, + "learning_rate": 1.5075688100039715e-05, + "loss": 0.4979, + "step": 22462 + }, + { + "epoch": 1.5075668601724774, + "grad_norm": 1.0377678871154785, + "learning_rate": 1.5067911309814503e-05, + "loss": 0.4405, + "step": 22464 + }, + { + "epoch": 1.5077010838562463, + "grad_norm": 0.9336119890213013, + "learning_rate": 1.5060136170061773e-05, + "loss": 0.4473, + "step": 22466 + }, + { + "epoch": 1.5078353075400155, + "grad_norm": 0.8464655876159668, + "learning_rate": 1.5052362681148935e-05, + "loss": 0.4769, + "step": 22468 + }, + { + "epoch": 1.5079695312237844, + "grad_norm": 1.09187650680542, + "learning_rate": 1.5044590843443207e-05, + "loss": 0.4561, + "step": 22470 + }, + { + "epoch": 1.5081037549075536, + "grad_norm": 0.9741097092628479, + "learning_rate": 1.5036820657311839e-05, + "loss": 0.4661, + "step": 22472 + }, + { + "epoch": 1.5082379785913225, + "grad_norm": 0.9991913437843323, + "learning_rate": 1.5029052123121928e-05, + "loss": 0.5184, + "step": 22474 + }, + { + "epoch": 1.5083722022750914, + "grad_norm": 0.9883608222007751, + "learning_rate": 1.5021285241240552e-05, + "loss": 0.4325, + "step": 22476 + }, + { + "epoch": 1.5085064259588603, + "grad_norm": 0.9305347204208374, + "learning_rate": 1.5013520012034665e-05, + "loss": 0.508, + "step": 22478 + }, + { + "epoch": 1.5086406496426294, + "grad_norm": 1.0459718704223633, + "learning_rate": 1.5005756435871165e-05, + "loss": 0.4844, + "step": 22480 + }, + { + "epoch": 1.5087748733263986, + "grad_norm": 1.0773015022277832, + "learning_rate": 1.4997994513116836e-05, + "loss": 0.6178, + "step": 22482 + }, + { + "epoch": 1.5089090970101675, + "grad_norm": 1.004028558731079, + "learning_rate": 1.4990234244138457e-05, + "loss": 0.4982, + "step": 22484 + }, + { + "epoch": 1.5090433206939364, + "grad_norm": 0.8696219325065613, + "learning_rate": 1.4982475629302667e-05, + "loss": 0.4384, + "step": 22486 + }, + { + "epoch": 1.5091775443777053, + "grad_norm": 0.9148178696632385, + "learning_rate": 1.4974718668976046e-05, + "loss": 0.464, + "step": 22488 + }, + { + "epoch": 1.5093117680614745, + "grad_norm": 1.0425653457641602, + "learning_rate": 1.4966963363525077e-05, + "loss": 0.428, + "step": 22490 + }, + { + "epoch": 1.5094459917452434, + "grad_norm": 0.8826553821563721, + "learning_rate": 1.4959209713316213e-05, + "loss": 0.4278, + "step": 22492 + }, + { + "epoch": 1.5095802154290126, + "grad_norm": 1.0622694492340088, + "learning_rate": 1.4951457718715772e-05, + "loss": 0.477, + "step": 22494 + }, + { + "epoch": 1.5097144391127815, + "grad_norm": 1.084007740020752, + "learning_rate": 1.4943707380090067e-05, + "loss": 0.5435, + "step": 22496 + }, + { + "epoch": 1.5098486627965504, + "grad_norm": 1.0633143186569214, + "learning_rate": 1.493595869780522e-05, + "loss": 0.56, + "step": 22498 + }, + { + "epoch": 1.5099828864803193, + "grad_norm": 1.3784126043319702, + "learning_rate": 1.4928211672227387e-05, + "loss": 0.4376, + "step": 22500 + }, + { + "epoch": 1.5101171101640884, + "grad_norm": 1.055816888809204, + "learning_rate": 1.4920466303722568e-05, + "loss": 0.4615, + "step": 22502 + }, + { + "epoch": 1.5102513338478576, + "grad_norm": 1.0887091159820557, + "learning_rate": 1.4912722592656758e-05, + "loss": 0.511, + "step": 22504 + }, + { + "epoch": 1.5103855575316265, + "grad_norm": 1.1009773015975952, + "learning_rate": 1.4904980539395808e-05, + "loss": 0.4833, + "step": 22506 + }, + { + "epoch": 1.5105197812153954, + "grad_norm": 1.0618146657943726, + "learning_rate": 1.4897240144305513e-05, + "loss": 0.4793, + "step": 22508 + }, + { + "epoch": 1.5106540048991643, + "grad_norm": 1.0707080364227295, + "learning_rate": 1.4889501407751583e-05, + "loss": 0.5317, + "step": 22510 + }, + { + "epoch": 1.5107882285829335, + "grad_norm": 0.9482539296150208, + "learning_rate": 1.4881764330099685e-05, + "loss": 0.515, + "step": 22512 + }, + { + "epoch": 1.5109224522667024, + "grad_norm": 0.8839782476425171, + "learning_rate": 1.487402891171536e-05, + "loss": 0.4213, + "step": 22514 + }, + { + "epoch": 1.5110566759504716, + "grad_norm": 0.9936320781707764, + "learning_rate": 1.4866295152964104e-05, + "loss": 0.4851, + "step": 22516 + }, + { + "epoch": 1.5111908996342405, + "grad_norm": 0.9398537278175354, + "learning_rate": 1.4858563054211294e-05, + "loss": 0.4829, + "step": 22518 + }, + { + "epoch": 1.5113251233180094, + "grad_norm": 1.0250434875488281, + "learning_rate": 1.4850832615822297e-05, + "loss": 0.475, + "step": 22520 + }, + { + "epoch": 1.5114593470017783, + "grad_norm": 1.359682559967041, + "learning_rate": 1.484310383816232e-05, + "loss": 0.5049, + "step": 22522 + }, + { + "epoch": 1.5115935706855474, + "grad_norm": 1.0129345655441284, + "learning_rate": 1.4835376721596588e-05, + "loss": 0.4854, + "step": 22524 + }, + { + "epoch": 1.5117277943693166, + "grad_norm": 1.093867540359497, + "learning_rate": 1.482765126649014e-05, + "loss": 0.5006, + "step": 22526 + }, + { + "epoch": 1.5118620180530855, + "grad_norm": 0.9952700138092041, + "learning_rate": 1.481992747320799e-05, + "loss": 0.5021, + "step": 22528 + }, + { + "epoch": 1.5119962417368544, + "grad_norm": 0.9017931222915649, + "learning_rate": 1.4812205342115104e-05, + "loss": 0.4701, + "step": 22530 + }, + { + "epoch": 1.5121304654206233, + "grad_norm": 1.0444830656051636, + "learning_rate": 1.4804484873576314e-05, + "loss": 0.4944, + "step": 22532 + }, + { + "epoch": 1.5122646891043925, + "grad_norm": 1.1205003261566162, + "learning_rate": 1.4796766067956424e-05, + "loss": 0.5356, + "step": 22534 + }, + { + "epoch": 1.5123989127881616, + "grad_norm": 1.121140480041504, + "learning_rate": 1.4789048925620091e-05, + "loss": 0.4984, + "step": 22536 + }, + { + "epoch": 1.5125331364719306, + "grad_norm": 0.7770562171936035, + "learning_rate": 1.4781333446931966e-05, + "loss": 0.3848, + "step": 22538 + }, + { + "epoch": 1.5126673601556995, + "grad_norm": 0.9135148525238037, + "learning_rate": 1.4773619632256568e-05, + "loss": 0.4568, + "step": 22540 + }, + { + "epoch": 1.5128015838394684, + "grad_norm": 0.8369489908218384, + "learning_rate": 1.476590748195839e-05, + "loss": 0.4315, + "step": 22542 + }, + { + "epoch": 1.5129358075232375, + "grad_norm": 1.1326509714126587, + "learning_rate": 1.4758196996401796e-05, + "loss": 0.429, + "step": 22544 + }, + { + "epoch": 1.5130700312070064, + "grad_norm": 1.1862406730651855, + "learning_rate": 1.4750488175951099e-05, + "loss": 0.4852, + "step": 22546 + }, + { + "epoch": 1.5132042548907756, + "grad_norm": 1.0056285858154297, + "learning_rate": 1.4742781020970498e-05, + "loss": 0.4809, + "step": 22548 + }, + { + "epoch": 1.5133384785745445, + "grad_norm": 0.997708261013031, + "learning_rate": 1.4735075531824183e-05, + "loss": 0.4832, + "step": 22550 + }, + { + "epoch": 1.5134727022583134, + "grad_norm": 1.0812458992004395, + "learning_rate": 1.4727371708876203e-05, + "loss": 0.5126, + "step": 22552 + }, + { + "epoch": 1.5136069259420823, + "grad_norm": 1.117914080619812, + "learning_rate": 1.4719669552490551e-05, + "loss": 0.4548, + "step": 22554 + }, + { + "epoch": 1.5137411496258515, + "grad_norm": 1.194248914718628, + "learning_rate": 1.4711969063031117e-05, + "loss": 0.5727, + "step": 22556 + }, + { + "epoch": 1.5138753733096206, + "grad_norm": 0.9323875308036804, + "learning_rate": 1.4704270240861773e-05, + "loss": 0.4637, + "step": 22558 + }, + { + "epoch": 1.5140095969933896, + "grad_norm": 1.0517685413360596, + "learning_rate": 1.4696573086346237e-05, + "loss": 0.4656, + "step": 22560 + }, + { + "epoch": 1.5141438206771585, + "grad_norm": 0.9463817477226257, + "learning_rate": 1.4688877599848216e-05, + "loss": 0.4301, + "step": 22562 + }, + { + "epoch": 1.5142780443609274, + "grad_norm": 1.0993897914886475, + "learning_rate": 1.4681183781731294e-05, + "loss": 0.4534, + "step": 22564 + }, + { + "epoch": 1.5144122680446965, + "grad_norm": 1.0667643547058105, + "learning_rate": 1.4673491632358987e-05, + "loss": 0.4823, + "step": 22566 + }, + { + "epoch": 1.5145464917284654, + "grad_norm": 1.004535436630249, + "learning_rate": 1.4665801152094715e-05, + "loss": 0.4562, + "step": 22568 + }, + { + "epoch": 1.5146807154122346, + "grad_norm": 1.5485291481018066, + "learning_rate": 1.4658112341301878e-05, + "loss": 0.4401, + "step": 22570 + }, + { + "epoch": 1.5148149390960035, + "grad_norm": 1.1256248950958252, + "learning_rate": 1.4650425200343732e-05, + "loss": 0.4754, + "step": 22572 + }, + { + "epoch": 1.5149491627797724, + "grad_norm": 1.0484271049499512, + "learning_rate": 1.4642739729583482e-05, + "loss": 0.5324, + "step": 22574 + }, + { + "epoch": 1.5150833864635413, + "grad_norm": 0.9325432777404785, + "learning_rate": 1.463505592938424e-05, + "loss": 0.4541, + "step": 22576 + }, + { + "epoch": 1.5152176101473105, + "grad_norm": 0.9707252383232117, + "learning_rate": 1.4627373800109078e-05, + "loss": 0.4435, + "step": 22578 + }, + { + "epoch": 1.5153518338310796, + "grad_norm": 1.595065951347351, + "learning_rate": 1.4619693342120943e-05, + "loss": 0.4888, + "step": 22580 + }, + { + "epoch": 1.5154860575148486, + "grad_norm": 1.0771530866622925, + "learning_rate": 1.4612014555782733e-05, + "loss": 0.4859, + "step": 22582 + }, + { + "epoch": 1.5156202811986175, + "grad_norm": 0.9721912741661072, + "learning_rate": 1.4604337441457227e-05, + "loss": 0.4599, + "step": 22584 + }, + { + "epoch": 1.5157545048823864, + "grad_norm": 0.9945108294487, + "learning_rate": 1.459666199950719e-05, + "loss": 0.4889, + "step": 22586 + }, + { + "epoch": 1.5158887285661555, + "grad_norm": 0.9683254957199097, + "learning_rate": 1.4588988230295242e-05, + "loss": 0.4873, + "step": 22588 + }, + { + "epoch": 1.5160229522499244, + "grad_norm": 1.6426539421081543, + "learning_rate": 1.458131613418398e-05, + "loss": 0.4694, + "step": 22590 + }, + { + "epoch": 1.5161571759336936, + "grad_norm": 1.1235136985778809, + "learning_rate": 1.4573645711535899e-05, + "loss": 0.4858, + "step": 22592 + }, + { + "epoch": 1.5162913996174625, + "grad_norm": 1.0458322763442993, + "learning_rate": 1.4565976962713369e-05, + "loss": 0.5101, + "step": 22594 + }, + { + "epoch": 1.5164256233012314, + "grad_norm": 0.9468684196472168, + "learning_rate": 1.4558309888078758e-05, + "loss": 0.4137, + "step": 22596 + }, + { + "epoch": 1.5165598469850003, + "grad_norm": 1.0720895528793335, + "learning_rate": 1.4550644487994303e-05, + "loss": 0.469, + "step": 22598 + }, + { + "epoch": 1.5166940706687695, + "grad_norm": 0.8745453357696533, + "learning_rate": 1.45429807628222e-05, + "loss": 0.4279, + "step": 22600 + }, + { + "epoch": 1.5168282943525386, + "grad_norm": 1.0765306949615479, + "learning_rate": 1.4535318712924529e-05, + "loss": 0.4812, + "step": 22602 + }, + { + "epoch": 1.5169625180363076, + "grad_norm": 0.9445536136627197, + "learning_rate": 1.4527658338663313e-05, + "loss": 0.4573, + "step": 22604 + }, + { + "epoch": 1.5170967417200765, + "grad_norm": 1.193177342414856, + "learning_rate": 1.4519999640400478e-05, + "loss": 0.5268, + "step": 22606 + }, + { + "epoch": 1.5172309654038454, + "grad_norm": 1.078736424446106, + "learning_rate": 1.4512342618497898e-05, + "loss": 0.5245, + "step": 22608 + }, + { + "epoch": 1.5173651890876145, + "grad_norm": 1.1003497838974, + "learning_rate": 1.4504687273317352e-05, + "loss": 0.4627, + "step": 22610 + }, + { + "epoch": 1.5174994127713837, + "grad_norm": 1.018603801727295, + "learning_rate": 1.4497033605220529e-05, + "loss": 0.4807, + "step": 22612 + }, + { + "epoch": 1.5176336364551526, + "grad_norm": 0.9425628781318665, + "learning_rate": 1.4489381614569043e-05, + "loss": 0.436, + "step": 22614 + }, + { + "epoch": 1.5177678601389215, + "grad_norm": 1.0539040565490723, + "learning_rate": 1.4481731301724461e-05, + "loss": 0.4773, + "step": 22616 + }, + { + "epoch": 1.5179020838226904, + "grad_norm": 1.076253890991211, + "learning_rate": 1.4474082667048217e-05, + "loss": 0.4542, + "step": 22618 + }, + { + "epoch": 1.5180363075064596, + "grad_norm": 1.0536024570465088, + "learning_rate": 1.4466435710901737e-05, + "loss": 0.5279, + "step": 22620 + }, + { + "epoch": 1.5181705311902285, + "grad_norm": 1.0149723291397095, + "learning_rate": 1.4458790433646263e-05, + "loss": 0.5016, + "step": 22622 + }, + { + "epoch": 1.5183047548739976, + "grad_norm": 4.249631881713867, + "learning_rate": 1.445114683564307e-05, + "loss": 0.5644, + "step": 22624 + }, + { + "epoch": 1.5184389785577665, + "grad_norm": 0.9433045983314514, + "learning_rate": 1.4443504917253275e-05, + "loss": 0.4131, + "step": 22626 + }, + { + "epoch": 1.5185732022415355, + "grad_norm": 1.0704541206359863, + "learning_rate": 1.4435864678837962e-05, + "loss": 0.4985, + "step": 22628 + }, + { + "epoch": 1.5187074259253044, + "grad_norm": 1.1655170917510986, + "learning_rate": 1.4428226120758114e-05, + "loss": 0.5102, + "step": 22630 + }, + { + "epoch": 1.5188416496090735, + "grad_norm": 0.979566216468811, + "learning_rate": 1.4420589243374626e-05, + "loss": 0.4862, + "step": 22632 + }, + { + "epoch": 1.5189758732928427, + "grad_norm": 0.9434431791305542, + "learning_rate": 1.4412954047048326e-05, + "loss": 0.4373, + "step": 22634 + }, + { + "epoch": 1.5191100969766116, + "grad_norm": 1.0647971630096436, + "learning_rate": 1.4405320532139976e-05, + "loss": 0.4929, + "step": 22636 + }, + { + "epoch": 1.5192443206603805, + "grad_norm": 0.9849785566329956, + "learning_rate": 1.4397688699010241e-05, + "loss": 0.4674, + "step": 22638 + }, + { + "epoch": 1.5193785443441494, + "grad_norm": 1.1018036603927612, + "learning_rate": 1.4390058548019708e-05, + "loss": 0.5187, + "step": 22640 + }, + { + "epoch": 1.5195127680279186, + "grad_norm": 1.1140117645263672, + "learning_rate": 1.4382430079528864e-05, + "loss": 0.4234, + "step": 22642 + }, + { + "epoch": 1.5196469917116875, + "grad_norm": 1.0272324085235596, + "learning_rate": 1.4374803293898181e-05, + "loss": 0.4543, + "step": 22644 + }, + { + "epoch": 1.5197812153954566, + "grad_norm": 1.0630650520324707, + "learning_rate": 1.4367178191487968e-05, + "loss": 0.5103, + "step": 22646 + }, + { + "epoch": 1.5199154390792255, + "grad_norm": 0.9447331428527832, + "learning_rate": 1.4359554772658552e-05, + "loss": 0.482, + "step": 22648 + }, + { + "epoch": 1.5200496627629945, + "grad_norm": 1.052130103111267, + "learning_rate": 1.4351933037770049e-05, + "loss": 0.5365, + "step": 22650 + }, + { + "epoch": 1.5201838864467634, + "grad_norm": 1.0532441139221191, + "learning_rate": 1.4344312987182635e-05, + "loss": 0.521, + "step": 22652 + }, + { + "epoch": 1.5203181101305325, + "grad_norm": 1.086188554763794, + "learning_rate": 1.4336694621256303e-05, + "loss": 0.5064, + "step": 22654 + }, + { + "epoch": 1.5204523338143017, + "grad_norm": 1.1076205968856812, + "learning_rate": 1.4329077940351031e-05, + "loss": 0.4888, + "step": 22656 + }, + { + "epoch": 1.5205865574980706, + "grad_norm": 1.1020424365997314, + "learning_rate": 1.4321462944826685e-05, + "loss": 0.4586, + "step": 22658 + }, + { + "epoch": 1.5207207811818395, + "grad_norm": 0.9768147468566895, + "learning_rate": 1.4313849635043058e-05, + "loss": 0.4714, + "step": 22660 + }, + { + "epoch": 1.5208550048656084, + "grad_norm": 0.9518616795539856, + "learning_rate": 1.4306238011359863e-05, + "loss": 0.4481, + "step": 22662 + }, + { + "epoch": 1.5209892285493776, + "grad_norm": 0.9238994717597961, + "learning_rate": 1.4298628074136717e-05, + "loss": 0.4517, + "step": 22664 + }, + { + "epoch": 1.5211234522331465, + "grad_norm": 0.9104543328285217, + "learning_rate": 1.4291019823733203e-05, + "loss": 0.4736, + "step": 22666 + }, + { + "epoch": 1.5212576759169156, + "grad_norm": 1.152108907699585, + "learning_rate": 1.4283413260508788e-05, + "loss": 0.5382, + "step": 22668 + }, + { + "epoch": 1.5213918996006845, + "grad_norm": 1.0099257230758667, + "learning_rate": 1.4275808384822858e-05, + "loss": 0.4768, + "step": 22670 + }, + { + "epoch": 1.5215261232844535, + "grad_norm": 1.1115846633911133, + "learning_rate": 1.4268205197034717e-05, + "loss": 0.4871, + "step": 22672 + }, + { + "epoch": 1.5216603469682224, + "grad_norm": 0.9350446462631226, + "learning_rate": 1.4260603697503638e-05, + "loss": 0.4816, + "step": 22674 + }, + { + "epoch": 1.5217945706519915, + "grad_norm": 1.0540435314178467, + "learning_rate": 1.4253003886588751e-05, + "loss": 0.4726, + "step": 22676 + }, + { + "epoch": 1.5219287943357607, + "grad_norm": 0.9183579683303833, + "learning_rate": 1.4245405764649138e-05, + "loss": 0.4748, + "step": 22678 + }, + { + "epoch": 1.5220630180195296, + "grad_norm": 1.0179717540740967, + "learning_rate": 1.4237809332043784e-05, + "loss": 0.4622, + "step": 22680 + }, + { + "epoch": 1.5221972417032985, + "grad_norm": 1.0048412084579468, + "learning_rate": 1.4230214589131625e-05, + "loss": 0.4691, + "step": 22682 + }, + { + "epoch": 1.5223314653870674, + "grad_norm": 1.0105338096618652, + "learning_rate": 1.4222621536271474e-05, + "loss": 0.4468, + "step": 22684 + }, + { + "epoch": 1.5224656890708366, + "grad_norm": 0.9902668595314026, + "learning_rate": 1.4215030173822136e-05, + "loss": 0.4491, + "step": 22686 + }, + { + "epoch": 1.5225999127546057, + "grad_norm": 1.0982518196105957, + "learning_rate": 1.4207440502142222e-05, + "loss": 0.4566, + "step": 22688 + }, + { + "epoch": 1.5227341364383746, + "grad_norm": 1.0153194665908813, + "learning_rate": 1.4199852521590373e-05, + "loss": 0.4912, + "step": 22690 + }, + { + "epoch": 1.5228683601221435, + "grad_norm": 0.9769749045372009, + "learning_rate": 1.419226623252508e-05, + "loss": 0.4997, + "step": 22692 + }, + { + "epoch": 1.5230025838059125, + "grad_norm": 0.9678530097007751, + "learning_rate": 1.4184681635304814e-05, + "loss": 0.4807, + "step": 22694 + }, + { + "epoch": 1.5231368074896814, + "grad_norm": 0.9360699653625488, + "learning_rate": 1.4177098730287914e-05, + "loss": 0.4441, + "step": 22696 + }, + { + "epoch": 1.5232710311734505, + "grad_norm": 1.1069527864456177, + "learning_rate": 1.4169517517832654e-05, + "loss": 0.4684, + "step": 22698 + }, + { + "epoch": 1.5234052548572197, + "grad_norm": 1.3689544200897217, + "learning_rate": 1.4161937998297214e-05, + "loss": 0.5736, + "step": 22700 + }, + { + "epoch": 1.5235394785409886, + "grad_norm": 0.9946498274803162, + "learning_rate": 1.4154360172039749e-05, + "loss": 0.5132, + "step": 22702 + }, + { + "epoch": 1.5236737022247575, + "grad_norm": 0.970460832118988, + "learning_rate": 1.4146784039418282e-05, + "loss": 0.4773, + "step": 22704 + }, + { + "epoch": 1.5238079259085264, + "grad_norm": 0.9970105886459351, + "learning_rate": 1.4139209600790766e-05, + "loss": 0.4294, + "step": 22706 + }, + { + "epoch": 1.5239421495922956, + "grad_norm": 1.1605664491653442, + "learning_rate": 1.4131636856515056e-05, + "loss": 0.455, + "step": 22708 + }, + { + "epoch": 1.5240763732760647, + "grad_norm": 1.0581333637237549, + "learning_rate": 1.4124065806948994e-05, + "loss": 0.4962, + "step": 22710 + }, + { + "epoch": 1.5242105969598336, + "grad_norm": 0.9241947531700134, + "learning_rate": 1.4116496452450262e-05, + "loss": 0.3972, + "step": 22712 + }, + { + "epoch": 1.5243448206436025, + "grad_norm": 1.2357085943222046, + "learning_rate": 1.4108928793376542e-05, + "loss": 0.441, + "step": 22714 + }, + { + "epoch": 1.5244790443273715, + "grad_norm": 1.1028046607971191, + "learning_rate": 1.4101362830085324e-05, + "loss": 0.4704, + "step": 22716 + }, + { + "epoch": 1.5246132680111406, + "grad_norm": 0.9638956189155579, + "learning_rate": 1.4093798562934141e-05, + "loss": 0.4772, + "step": 22718 + }, + { + "epoch": 1.5247474916949095, + "grad_norm": 1.06745445728302, + "learning_rate": 1.408623599228035e-05, + "loss": 0.4541, + "step": 22720 + }, + { + "epoch": 1.5248817153786787, + "grad_norm": 1.1657145023345947, + "learning_rate": 1.4078675118481305e-05, + "loss": 0.5244, + "step": 22722 + }, + { + "epoch": 1.5250159390624476, + "grad_norm": 1.0456702709197998, + "learning_rate": 1.407111594189422e-05, + "loss": 0.4596, + "step": 22724 + }, + { + "epoch": 1.5251501627462165, + "grad_norm": 1.0020138025283813, + "learning_rate": 1.4063558462876259e-05, + "loss": 0.5044, + "step": 22726 + }, + { + "epoch": 1.5252843864299854, + "grad_norm": 0.9499043822288513, + "learning_rate": 1.4056002681784496e-05, + "loss": 0.4474, + "step": 22728 + }, + { + "epoch": 1.5254186101137546, + "grad_norm": 0.9909380674362183, + "learning_rate": 1.4048448598975905e-05, + "loss": 0.4658, + "step": 22730 + }, + { + "epoch": 1.5255528337975237, + "grad_norm": 1.068764090538025, + "learning_rate": 1.4040896214807442e-05, + "loss": 0.4693, + "step": 22732 + }, + { + "epoch": 1.5256870574812926, + "grad_norm": 0.9431993365287781, + "learning_rate": 1.4033345529635922e-05, + "loss": 0.4253, + "step": 22734 + }, + { + "epoch": 1.5258212811650615, + "grad_norm": 1.0659258365631104, + "learning_rate": 1.4025796543818098e-05, + "loss": 0.5087, + "step": 22736 + }, + { + "epoch": 1.5259555048488305, + "grad_norm": 1.0343708992004395, + "learning_rate": 1.4018249257710636e-05, + "loss": 0.4301, + "step": 22738 + }, + { + "epoch": 1.5260897285325996, + "grad_norm": 1.1548924446105957, + "learning_rate": 1.4010703671670155e-05, + "loss": 0.4605, + "step": 22740 + }, + { + "epoch": 1.5262239522163685, + "grad_norm": 1.0230505466461182, + "learning_rate": 1.4003159786053139e-05, + "loss": 0.4107, + "step": 22742 + }, + { + "epoch": 1.5263581759001377, + "grad_norm": 1.0957962274551392, + "learning_rate": 1.3995617601216066e-05, + "loss": 0.488, + "step": 22744 + }, + { + "epoch": 1.5264923995839066, + "grad_norm": 0.9980576038360596, + "learning_rate": 1.3988077117515236e-05, + "loss": 0.4498, + "step": 22746 + }, + { + "epoch": 1.5266266232676755, + "grad_norm": 0.8710897564888, + "learning_rate": 1.3980538335306964e-05, + "loss": 0.5065, + "step": 22748 + }, + { + "epoch": 1.5267608469514444, + "grad_norm": 1.0267164707183838, + "learning_rate": 1.3973001254947404e-05, + "loss": 0.4917, + "step": 22750 + }, + { + "epoch": 1.5268950706352136, + "grad_norm": 1.0731956958770752, + "learning_rate": 1.3965465876792716e-05, + "loss": 0.5146, + "step": 22752 + }, + { + "epoch": 1.5270292943189827, + "grad_norm": 0.9476215839385986, + "learning_rate": 1.3957932201198897e-05, + "loss": 0.4683, + "step": 22754 + }, + { + "epoch": 1.5271635180027516, + "grad_norm": 0.9859458208084106, + "learning_rate": 1.3950400228521914e-05, + "loss": 0.4408, + "step": 22756 + }, + { + "epoch": 1.5272977416865205, + "grad_norm": 0.9938817620277405, + "learning_rate": 1.3942869959117616e-05, + "loss": 0.4758, + "step": 22758 + }, + { + "epoch": 1.5274319653702895, + "grad_norm": 1.087532877922058, + "learning_rate": 1.3935341393341822e-05, + "loss": 0.5726, + "step": 22760 + }, + { + "epoch": 1.5275661890540586, + "grad_norm": 1.059373378753662, + "learning_rate": 1.3927814531550226e-05, + "loss": 0.4676, + "step": 22762 + }, + { + "epoch": 1.5277004127378278, + "grad_norm": 1.0563775300979614, + "learning_rate": 1.3920289374098467e-05, + "loss": 0.4405, + "step": 22764 + }, + { + "epoch": 1.5278346364215967, + "grad_norm": 0.9323393702507019, + "learning_rate": 1.391276592134207e-05, + "loss": 0.4398, + "step": 22766 + }, + { + "epoch": 1.5279688601053656, + "grad_norm": 0.9752359390258789, + "learning_rate": 1.3905244173636534e-05, + "loss": 0.4928, + "step": 22768 + }, + { + "epoch": 1.5281030837891345, + "grad_norm": 0.9924960732460022, + "learning_rate": 1.389772413133722e-05, + "loss": 0.5054, + "step": 22770 + }, + { + "epoch": 1.5282373074729034, + "grad_norm": 1.0688010454177856, + "learning_rate": 1.3890205794799476e-05, + "loss": 0.517, + "step": 22772 + }, + { + "epoch": 1.5283715311566726, + "grad_norm": 1.0437699556350708, + "learning_rate": 1.3882689164378477e-05, + "loss": 0.482, + "step": 22774 + }, + { + "epoch": 1.5285057548404417, + "grad_norm": 1.0509988069534302, + "learning_rate": 1.3875174240429405e-05, + "loss": 0.4936, + "step": 22776 + }, + { + "epoch": 1.5286399785242106, + "grad_norm": 1.0045628547668457, + "learning_rate": 1.3867661023307298e-05, + "loss": 0.474, + "step": 22778 + }, + { + "epoch": 1.5287742022079795, + "grad_norm": 1.0354044437408447, + "learning_rate": 1.3860149513367171e-05, + "loss": 0.507, + "step": 22780 + }, + { + "epoch": 1.5289084258917485, + "grad_norm": 1.0339772701263428, + "learning_rate": 1.3852639710963916e-05, + "loss": 0.4641, + "step": 22782 + }, + { + "epoch": 1.5290426495755176, + "grad_norm": 1.1421676874160767, + "learning_rate": 1.384513161645235e-05, + "loss": 0.5663, + "step": 22784 + }, + { + "epoch": 1.5291768732592868, + "grad_norm": 1.1045042276382446, + "learning_rate": 1.3837625230187212e-05, + "loss": 0.4658, + "step": 22786 + }, + { + "epoch": 1.5293110969430557, + "grad_norm": 1.1218268871307373, + "learning_rate": 1.383012055252318e-05, + "loss": 0.4573, + "step": 22788 + }, + { + "epoch": 1.5294453206268246, + "grad_norm": 0.9117696285247803, + "learning_rate": 1.3822617583814828e-05, + "loss": 0.4597, + "step": 22790 + }, + { + "epoch": 1.5295795443105935, + "grad_norm": 0.9316728711128235, + "learning_rate": 1.3815116324416654e-05, + "loss": 0.4361, + "step": 22792 + }, + { + "epoch": 1.5297137679943626, + "grad_norm": 0.9336127042770386, + "learning_rate": 1.380761677468307e-05, + "loss": 0.455, + "step": 22794 + }, + { + "epoch": 1.5298479916781316, + "grad_norm": 0.9192490577697754, + "learning_rate": 1.3800118934968436e-05, + "loss": 0.4215, + "step": 22796 + }, + { + "epoch": 1.5299822153619007, + "grad_norm": 1.175935983657837, + "learning_rate": 1.3792622805626998e-05, + "loss": 0.4492, + "step": 22798 + }, + { + "epoch": 1.5301164390456696, + "grad_norm": 1.0037862062454224, + "learning_rate": 1.378512838701293e-05, + "loss": 0.4946, + "step": 22800 + }, + { + "epoch": 1.5302506627294385, + "grad_norm": 1.108101725578308, + "learning_rate": 1.377763567948034e-05, + "loss": 0.4989, + "step": 22802 + }, + { + "epoch": 1.5303848864132075, + "grad_norm": 1.0684829950332642, + "learning_rate": 1.3770144683383213e-05, + "loss": 0.506, + "step": 22804 + }, + { + "epoch": 1.5305191100969766, + "grad_norm": 0.9487861394882202, + "learning_rate": 1.3762655399075524e-05, + "loss": 0.4489, + "step": 22806 + }, + { + "epoch": 1.5306533337807458, + "grad_norm": 1.1091814041137695, + "learning_rate": 1.3755167826911097e-05, + "loss": 0.4866, + "step": 22808 + }, + { + "epoch": 1.5307875574645147, + "grad_norm": 1.2332231998443604, + "learning_rate": 1.3747681967243748e-05, + "loss": 0.4827, + "step": 22810 + }, + { + "epoch": 1.5309217811482836, + "grad_norm": 1.0854482650756836, + "learning_rate": 1.3740197820427108e-05, + "loss": 0.4717, + "step": 22812 + }, + { + "epoch": 1.5310560048320525, + "grad_norm": 1.0635013580322266, + "learning_rate": 1.3732715386814832e-05, + "loss": 0.4622, + "step": 22814 + }, + { + "epoch": 1.5311902285158216, + "grad_norm": 1.0756031274795532, + "learning_rate": 1.3725234666760428e-05, + "loss": 0.4947, + "step": 22816 + }, + { + "epoch": 1.5313244521995906, + "grad_norm": 1.1334630250930786, + "learning_rate": 1.3717755660617365e-05, + "loss": 0.4707, + "step": 22818 + }, + { + "epoch": 1.5314586758833597, + "grad_norm": 0.9022170305252075, + "learning_rate": 1.3710278368738998e-05, + "loss": 0.4441, + "step": 22820 + }, + { + "epoch": 1.5315928995671286, + "grad_norm": 0.9382498264312744, + "learning_rate": 1.3702802791478625e-05, + "loss": 0.4669, + "step": 22822 + }, + { + "epoch": 1.5317271232508975, + "grad_norm": 1.1221257448196411, + "learning_rate": 1.3695328929189426e-05, + "loss": 0.5197, + "step": 22824 + }, + { + "epoch": 1.5318613469346665, + "grad_norm": 0.9527440071105957, + "learning_rate": 1.3687856782224562e-05, + "loss": 0.4608, + "step": 22826 + }, + { + "epoch": 1.5319955706184356, + "grad_norm": 0.9356170892715454, + "learning_rate": 1.3680386350937058e-05, + "loss": 0.5009, + "step": 22828 + }, + { + "epoch": 1.5321297943022048, + "grad_norm": 1.005391001701355, + "learning_rate": 1.3672917635679877e-05, + "loss": 0.4767, + "step": 22830 + }, + { + "epoch": 1.5322640179859737, + "grad_norm": 0.804820716381073, + "learning_rate": 1.3665450636805893e-05, + "loss": 0.4131, + "step": 22832 + }, + { + "epoch": 1.5323982416697426, + "grad_norm": 0.9496227502822876, + "learning_rate": 1.3657985354667929e-05, + "loss": 0.5042, + "step": 22834 + }, + { + "epoch": 1.5325324653535115, + "grad_norm": 1.1346501111984253, + "learning_rate": 1.365052178961868e-05, + "loss": 0.5019, + "step": 22836 + }, + { + "epoch": 1.5326666890372806, + "grad_norm": 1.0846158266067505, + "learning_rate": 1.3643059942010832e-05, + "loss": 0.4948, + "step": 22838 + }, + { + "epoch": 1.5328009127210498, + "grad_norm": 1.0905262231826782, + "learning_rate": 1.3635599812196875e-05, + "loss": 0.4883, + "step": 22840 + }, + { + "epoch": 1.5329351364048187, + "grad_norm": 0.9363589882850647, + "learning_rate": 1.3628141400529337e-05, + "loss": 0.494, + "step": 22842 + }, + { + "epoch": 1.5330693600885876, + "grad_norm": 1.1014615297317505, + "learning_rate": 1.3620684707360576e-05, + "loss": 0.4778, + "step": 22844 + }, + { + "epoch": 1.5332035837723565, + "grad_norm": 1.0319340229034424, + "learning_rate": 1.3613229733042943e-05, + "loss": 0.4386, + "step": 22846 + }, + { + "epoch": 1.5333378074561255, + "grad_norm": 1.0542309284210205, + "learning_rate": 1.3605776477928656e-05, + "loss": 0.4262, + "step": 22848 + }, + { + "epoch": 1.5334720311398946, + "grad_norm": 1.1086591482162476, + "learning_rate": 1.359832494236986e-05, + "loss": 0.4776, + "step": 22850 + }, + { + "epoch": 1.5336062548236638, + "grad_norm": 0.8893815875053406, + "learning_rate": 1.3590875126718621e-05, + "loss": 0.4307, + "step": 22852 + }, + { + "epoch": 1.5337404785074327, + "grad_norm": 1.0791600942611694, + "learning_rate": 1.3583427031326946e-05, + "loss": 0.4935, + "step": 22854 + }, + { + "epoch": 1.5338747021912016, + "grad_norm": 0.8816147446632385, + "learning_rate": 1.3575980656546732e-05, + "loss": 0.444, + "step": 22856 + }, + { + "epoch": 1.5340089258749705, + "grad_norm": 1.0307036638259888, + "learning_rate": 1.3568536002729815e-05, + "loss": 0.4971, + "step": 22858 + }, + { + "epoch": 1.5341431495587396, + "grad_norm": 1.0213000774383545, + "learning_rate": 1.3561093070227909e-05, + "loss": 0.4832, + "step": 22860 + }, + { + "epoch": 1.5342773732425088, + "grad_norm": 1.0029380321502686, + "learning_rate": 1.3553651859392725e-05, + "loss": 0.4448, + "step": 22862 + }, + { + "epoch": 1.5344115969262777, + "grad_norm": 0.9618352651596069, + "learning_rate": 1.35462123705758e-05, + "loss": 0.4694, + "step": 22864 + }, + { + "epoch": 1.5345458206100466, + "grad_norm": 1.014799952507019, + "learning_rate": 1.353877460412869e-05, + "loss": 0.4678, + "step": 22866 + }, + { + "epoch": 1.5346800442938155, + "grad_norm": 1.0170156955718994, + "learning_rate": 1.3531338560402767e-05, + "loss": 0.4646, + "step": 22868 + }, + { + "epoch": 1.5348142679775847, + "grad_norm": 0.9131841659545898, + "learning_rate": 1.352390423974937e-05, + "loss": 0.4263, + "step": 22870 + }, + { + "epoch": 1.5349484916613536, + "grad_norm": 1.0565253496170044, + "learning_rate": 1.3516471642519784e-05, + "loss": 0.4633, + "step": 22872 + }, + { + "epoch": 1.5350827153451228, + "grad_norm": 1.152448058128357, + "learning_rate": 1.3509040769065157e-05, + "loss": 0.5157, + "step": 22874 + }, + { + "epoch": 1.5352169390288917, + "grad_norm": 0.9575986862182617, + "learning_rate": 1.3501611619736609e-05, + "loss": 0.4592, + "step": 22876 + }, + { + "epoch": 1.5353511627126606, + "grad_norm": 1.0137993097305298, + "learning_rate": 1.3494184194885146e-05, + "loss": 0.4733, + "step": 22878 + }, + { + "epoch": 1.5354853863964295, + "grad_norm": 1.088549256324768, + "learning_rate": 1.3486758494861696e-05, + "loss": 0.4896, + "step": 22880 + }, + { + "epoch": 1.5356196100801986, + "grad_norm": 0.902563214302063, + "learning_rate": 1.3479334520017084e-05, + "loss": 0.4272, + "step": 22882 + }, + { + "epoch": 1.5357538337639678, + "grad_norm": 0.9075942039489746, + "learning_rate": 1.347191227070212e-05, + "loss": 0.4132, + "step": 22884 + }, + { + "epoch": 1.5358880574477367, + "grad_norm": 1.0854324102401733, + "learning_rate": 1.3464491747267472e-05, + "loss": 0.4918, + "step": 22886 + }, + { + "epoch": 1.5360222811315056, + "grad_norm": 0.8842144012451172, + "learning_rate": 1.3457072950063748e-05, + "loss": 0.5058, + "step": 22888 + }, + { + "epoch": 1.5361565048152745, + "grad_norm": 1.0799119472503662, + "learning_rate": 1.3449655879441447e-05, + "loss": 0.5155, + "step": 22890 + }, + { + "epoch": 1.5362907284990437, + "grad_norm": 1.0394150018692017, + "learning_rate": 1.3442240535751049e-05, + "loss": 0.4633, + "step": 22892 + }, + { + "epoch": 1.5364249521828126, + "grad_norm": 0.9008004069328308, + "learning_rate": 1.3434826919342901e-05, + "loss": 0.439, + "step": 22894 + }, + { + "epoch": 1.5365591758665818, + "grad_norm": 1.079035997390747, + "learning_rate": 1.342741503056728e-05, + "loss": 0.4822, + "step": 22896 + }, + { + "epoch": 1.5366933995503507, + "grad_norm": 1.0192890167236328, + "learning_rate": 1.3420004869774367e-05, + "loss": 0.4212, + "step": 22898 + }, + { + "epoch": 1.5368276232341196, + "grad_norm": 0.8894795775413513, + "learning_rate": 1.3412596437314302e-05, + "loss": 0.499, + "step": 22900 + }, + { + "epoch": 1.5369618469178885, + "grad_norm": 1.1050899028778076, + "learning_rate": 1.3405189733537098e-05, + "loss": 0.419, + "step": 22902 + }, + { + "epoch": 1.5370960706016576, + "grad_norm": 0.9837499260902405, + "learning_rate": 1.3397784758792731e-05, + "loss": 0.4764, + "step": 22904 + }, + { + "epoch": 1.5372302942854268, + "grad_norm": 1.1928001642227173, + "learning_rate": 1.339038151343106e-05, + "loss": 0.5221, + "step": 22906 + }, + { + "epoch": 1.5373645179691957, + "grad_norm": 1.0429415702819824, + "learning_rate": 1.3382979997801876e-05, + "loss": 0.4655, + "step": 22908 + }, + { + "epoch": 1.5374987416529646, + "grad_norm": 1.0762752294540405, + "learning_rate": 1.3375580212254862e-05, + "loss": 0.4728, + "step": 22910 + }, + { + "epoch": 1.5376329653367335, + "grad_norm": 1.103614091873169, + "learning_rate": 1.3368182157139686e-05, + "loss": 0.4927, + "step": 22912 + }, + { + "epoch": 1.5377671890205027, + "grad_norm": 1.0256223678588867, + "learning_rate": 1.3360785832805867e-05, + "loss": 0.5132, + "step": 22914 + }, + { + "epoch": 1.5379014127042718, + "grad_norm": 1.0534138679504395, + "learning_rate": 1.335339123960287e-05, + "loss": 0.4351, + "step": 22916 + }, + { + "epoch": 1.5380356363880407, + "grad_norm": 1.0414409637451172, + "learning_rate": 1.3345998377880053e-05, + "loss": 0.5019, + "step": 22918 + }, + { + "epoch": 1.5381698600718097, + "grad_norm": 0.9938732981681824, + "learning_rate": 1.3338607247986756e-05, + "loss": 0.5434, + "step": 22920 + }, + { + "epoch": 1.5383040837555786, + "grad_norm": 1.0894412994384766, + "learning_rate": 1.3331217850272176e-05, + "loss": 0.4831, + "step": 22922 + }, + { + "epoch": 1.5384383074393475, + "grad_norm": 0.8352552652359009, + "learning_rate": 1.3323830185085445e-05, + "loss": 0.405, + "step": 22924 + }, + { + "epoch": 1.5385725311231166, + "grad_norm": 1.027392864227295, + "learning_rate": 1.3316444252775601e-05, + "loss": 0.4724, + "step": 22926 + }, + { + "epoch": 1.5387067548068858, + "grad_norm": 1.2071126699447632, + "learning_rate": 1.3309060053691646e-05, + "loss": 0.4669, + "step": 22928 + }, + { + "epoch": 1.5388409784906547, + "grad_norm": 1.0582983493804932, + "learning_rate": 1.3301677588182442e-05, + "loss": 0.4437, + "step": 22930 + }, + { + "epoch": 1.5389752021744236, + "grad_norm": 1.1398028135299683, + "learning_rate": 1.3294296856596821e-05, + "loss": 0.5096, + "step": 22932 + }, + { + "epoch": 1.5391094258581925, + "grad_norm": 1.09684419631958, + "learning_rate": 1.328691785928351e-05, + "loss": 0.5131, + "step": 22934 + }, + { + "epoch": 1.5392436495419617, + "grad_norm": 1.0865932703018188, + "learning_rate": 1.3279540596591105e-05, + "loss": 0.4987, + "step": 22936 + }, + { + "epoch": 1.5393778732257308, + "grad_norm": 1.120383381843567, + "learning_rate": 1.327216506886822e-05, + "loss": 0.4713, + "step": 22938 + }, + { + "epoch": 1.5395120969094997, + "grad_norm": 0.9693374037742615, + "learning_rate": 1.326479127646329e-05, + "loss": 0.4503, + "step": 22940 + }, + { + "epoch": 1.5396463205932687, + "grad_norm": 0.9845932722091675, + "learning_rate": 1.325741921972476e-05, + "loss": 0.5201, + "step": 22942 + }, + { + "epoch": 1.5397805442770376, + "grad_norm": 1.0494850873947144, + "learning_rate": 1.325004889900091e-05, + "loss": 0.4806, + "step": 22944 + }, + { + "epoch": 1.5399147679608067, + "grad_norm": 1.0713417530059814, + "learning_rate": 1.3242680314639993e-05, + "loss": 0.5529, + "step": 22946 + }, + { + "epoch": 1.5400489916445756, + "grad_norm": 0.932652473449707, + "learning_rate": 1.3235313466990129e-05, + "loss": 0.5251, + "step": 22948 + }, + { + "epoch": 1.5401832153283448, + "grad_norm": 1.5986257791519165, + "learning_rate": 1.3227948356399428e-05, + "loss": 0.5229, + "step": 22950 + }, + { + "epoch": 1.5403174390121137, + "grad_norm": 1.0638889074325562, + "learning_rate": 1.3220584983215855e-05, + "loss": 0.4649, + "step": 22952 + }, + { + "epoch": 1.5404516626958826, + "grad_norm": 1.0206716060638428, + "learning_rate": 1.3213223347787324e-05, + "loss": 0.446, + "step": 22954 + }, + { + "epoch": 1.5405858863796515, + "grad_norm": 1.1154367923736572, + "learning_rate": 1.320586345046163e-05, + "loss": 0.485, + "step": 22956 + }, + { + "epoch": 1.5407201100634207, + "grad_norm": 0.9781999588012695, + "learning_rate": 1.3198505291586555e-05, + "loss": 0.4917, + "step": 22958 + }, + { + "epoch": 1.5408543337471898, + "grad_norm": 0.9744774699211121, + "learning_rate": 1.3191148871509724e-05, + "loss": 0.5047, + "step": 22960 + }, + { + "epoch": 1.5409885574309587, + "grad_norm": 1.0790060758590698, + "learning_rate": 1.3183794190578752e-05, + "loss": 0.4558, + "step": 22962 + }, + { + "epoch": 1.5411227811147277, + "grad_norm": 1.945683479309082, + "learning_rate": 1.317644124914108e-05, + "loss": 0.5305, + "step": 22964 + }, + { + "epoch": 1.5412570047984966, + "grad_norm": 1.2510796785354614, + "learning_rate": 1.3169090047544164e-05, + "loss": 0.4542, + "step": 22966 + }, + { + "epoch": 1.5413912284822657, + "grad_norm": 1.0230785608291626, + "learning_rate": 1.3161740586135308e-05, + "loss": 0.4038, + "step": 22968 + }, + { + "epoch": 1.5415254521660346, + "grad_norm": 1.131100058555603, + "learning_rate": 1.3154392865261778e-05, + "loss": 0.4862, + "step": 22970 + }, + { + "epoch": 1.5416596758498038, + "grad_norm": 1.1616319417953491, + "learning_rate": 1.3147046885270736e-05, + "loss": 0.4893, + "step": 22972 + }, + { + "epoch": 1.5417938995335727, + "grad_norm": 0.9876658320426941, + "learning_rate": 1.3139702646509255e-05, + "loss": 0.437, + "step": 22974 + }, + { + "epoch": 1.5419281232173416, + "grad_norm": 0.9533802270889282, + "learning_rate": 1.313236014932433e-05, + "loss": 0.4538, + "step": 22976 + }, + { + "epoch": 1.5420623469011105, + "grad_norm": 0.9141824841499329, + "learning_rate": 1.312501939406291e-05, + "loss": 0.5134, + "step": 22978 + }, + { + "epoch": 1.5421965705848797, + "grad_norm": 0.8759487271308899, + "learning_rate": 1.3117680381071807e-05, + "loss": 0.4456, + "step": 22980 + }, + { + "epoch": 1.5423307942686488, + "grad_norm": 1.0206835269927979, + "learning_rate": 1.3110343110697782e-05, + "loss": 0.4491, + "step": 22982 + }, + { + "epoch": 1.5424650179524177, + "grad_norm": 1.1610170602798462, + "learning_rate": 1.3103007583287486e-05, + "loss": 0.476, + "step": 22984 + }, + { + "epoch": 1.5425992416361867, + "grad_norm": 1.086210012435913, + "learning_rate": 1.3095673799187546e-05, + "loss": 0.4752, + "step": 22986 + }, + { + "epoch": 1.5427334653199556, + "grad_norm": 1.1827083826065063, + "learning_rate": 1.3088341758744432e-05, + "loss": 0.4967, + "step": 22988 + }, + { + "epoch": 1.5428676890037247, + "grad_norm": 1.1646783351898193, + "learning_rate": 1.3081011462304615e-05, + "loss": 0.4897, + "step": 22990 + }, + { + "epoch": 1.5430019126874939, + "grad_norm": 0.93353670835495, + "learning_rate": 1.3073682910214386e-05, + "loss": 0.4634, + "step": 22992 + }, + { + "epoch": 1.5431361363712628, + "grad_norm": 1.0688941478729248, + "learning_rate": 1.3066356102820032e-05, + "loss": 0.5364, + "step": 22994 + }, + { + "epoch": 1.5432703600550317, + "grad_norm": 0.9025923609733582, + "learning_rate": 1.3059031040467717e-05, + "loss": 0.4392, + "step": 22996 + }, + { + "epoch": 1.5434045837388006, + "grad_norm": 1.08487069606781, + "learning_rate": 1.305170772350356e-05, + "loss": 0.457, + "step": 22998 + }, + { + "epoch": 1.5435388074225695, + "grad_norm": 0.9283562302589417, + "learning_rate": 1.3044386152273552e-05, + "loss": 0.4543, + "step": 23000 + }, + { + "epoch": 1.5436730311063387, + "grad_norm": 1.0412951707839966, + "learning_rate": 1.303706632712363e-05, + "loss": 0.5077, + "step": 23002 + }, + { + "epoch": 1.5438072547901078, + "grad_norm": 1.063987374305725, + "learning_rate": 1.3029748248399642e-05, + "loss": 0.4588, + "step": 23004 + }, + { + "epoch": 1.5439414784738767, + "grad_norm": 1.0530637502670288, + "learning_rate": 1.302243191644733e-05, + "loss": 0.4607, + "step": 23006 + }, + { + "epoch": 1.5440757021576457, + "grad_norm": 0.9383824467658997, + "learning_rate": 1.3015117331612415e-05, + "loss": 0.4724, + "step": 23008 + }, + { + "epoch": 1.5442099258414146, + "grad_norm": 1.0184880495071411, + "learning_rate": 1.3007804494240478e-05, + "loss": 0.5048, + "step": 23010 + }, + { + "epoch": 1.5443441495251837, + "grad_norm": 0.978066623210907, + "learning_rate": 1.3000493404677039e-05, + "loss": 0.4338, + "step": 23012 + }, + { + "epoch": 1.5444783732089529, + "grad_norm": 1.1344259977340698, + "learning_rate": 1.2993184063267516e-05, + "loss": 0.4571, + "step": 23014 + }, + { + "epoch": 1.5446125968927218, + "grad_norm": 1.0928936004638672, + "learning_rate": 1.2985876470357284e-05, + "loss": 0.5347, + "step": 23016 + }, + { + "epoch": 1.5447468205764907, + "grad_norm": 0.9233540296554565, + "learning_rate": 1.2978570626291609e-05, + "loss": 0.4887, + "step": 23018 + }, + { + "epoch": 1.5448810442602596, + "grad_norm": 0.9370667338371277, + "learning_rate": 1.2971266531415672e-05, + "loss": 0.4634, + "step": 23020 + }, + { + "epoch": 1.5450152679440288, + "grad_norm": 1.0324769020080566, + "learning_rate": 1.2963964186074563e-05, + "loss": 0.5203, + "step": 23022 + }, + { + "epoch": 1.5451494916277977, + "grad_norm": 1.148537039756775, + "learning_rate": 1.2956663590613327e-05, + "loss": 0.5763, + "step": 23024 + }, + { + "epoch": 1.5452837153115668, + "grad_norm": 1.030452013015747, + "learning_rate": 1.2949364745376884e-05, + "loss": 0.4983, + "step": 23026 + }, + { + "epoch": 1.5454179389953357, + "grad_norm": 0.9592177271842957, + "learning_rate": 1.2942067650710116e-05, + "loss": 0.4368, + "step": 23028 + }, + { + "epoch": 1.5455521626791047, + "grad_norm": 0.7843078970909119, + "learning_rate": 1.2934772306957776e-05, + "loss": 0.4538, + "step": 23030 + }, + { + "epoch": 1.5456863863628736, + "grad_norm": 1.0391765832901, + "learning_rate": 1.2927478714464559e-05, + "loss": 0.4873, + "step": 23032 + }, + { + "epoch": 1.5458206100466427, + "grad_norm": 1.0403895378112793, + "learning_rate": 1.2920186873575064e-05, + "loss": 0.4878, + "step": 23034 + }, + { + "epoch": 1.5459548337304119, + "grad_norm": 1.0897223949432373, + "learning_rate": 1.2912896784633833e-05, + "loss": 0.5198, + "step": 23036 + }, + { + "epoch": 1.5460890574141808, + "grad_norm": 1.030831217765808, + "learning_rate": 1.2905608447985301e-05, + "loss": 0.4858, + "step": 23038 + }, + { + "epoch": 1.5462232810979497, + "grad_norm": 1.1313307285308838, + "learning_rate": 1.2898321863973829e-05, + "loss": 0.5108, + "step": 23040 + }, + { + "epoch": 1.5463575047817186, + "grad_norm": 0.9930167198181152, + "learning_rate": 1.2891037032943676e-05, + "loss": 0.51, + "step": 23042 + }, + { + "epoch": 1.5464917284654878, + "grad_norm": 0.9952108860015869, + "learning_rate": 1.288375395523907e-05, + "loss": 0.4509, + "step": 23044 + }, + { + "epoch": 1.5466259521492567, + "grad_norm": 1.1315922737121582, + "learning_rate": 1.2876472631204095e-05, + "loss": 0.5316, + "step": 23046 + }, + { + "epoch": 1.5467601758330258, + "grad_norm": 1.0469512939453125, + "learning_rate": 1.2869193061182794e-05, + "loss": 0.492, + "step": 23048 + }, + { + "epoch": 1.5468943995167947, + "grad_norm": 1.2232224941253662, + "learning_rate": 1.286191524551909e-05, + "loss": 0.4547, + "step": 23050 + }, + { + "epoch": 1.5470286232005637, + "grad_norm": 1.02517831325531, + "learning_rate": 1.2854639184556877e-05, + "loss": 0.4744, + "step": 23052 + }, + { + "epoch": 1.5471628468843326, + "grad_norm": 1.056570053100586, + "learning_rate": 1.2847364878639905e-05, + "loss": 0.5329, + "step": 23054 + }, + { + "epoch": 1.5472970705681017, + "grad_norm": 0.9619842171669006, + "learning_rate": 1.2840092328111903e-05, + "loss": 0.5317, + "step": 23056 + }, + { + "epoch": 1.5474312942518709, + "grad_norm": 1.1084604263305664, + "learning_rate": 1.2832821533316464e-05, + "loss": 0.4555, + "step": 23058 + }, + { + "epoch": 1.5475655179356398, + "grad_norm": 0.9658335447311401, + "learning_rate": 1.282555249459712e-05, + "loss": 0.5221, + "step": 23060 + }, + { + "epoch": 1.5476997416194087, + "grad_norm": 0.9604290127754211, + "learning_rate": 1.2818285212297303e-05, + "loss": 0.4113, + "step": 23062 + }, + { + "epoch": 1.5478339653031776, + "grad_norm": 1.0296306610107422, + "learning_rate": 1.2811019686760412e-05, + "loss": 0.4564, + "step": 23064 + }, + { + "epoch": 1.5479681889869468, + "grad_norm": 1.0469577312469482, + "learning_rate": 1.2803755918329712e-05, + "loss": 0.4442, + "step": 23066 + }, + { + "epoch": 1.548102412670716, + "grad_norm": 1.0704536437988281, + "learning_rate": 1.2796493907348406e-05, + "loss": 0.5484, + "step": 23068 + }, + { + "epoch": 1.5482366363544848, + "grad_norm": 0.9309592247009277, + "learning_rate": 1.2789233654159582e-05, + "loss": 0.4486, + "step": 23070 + }, + { + "epoch": 1.5483708600382537, + "grad_norm": 1.0256258249282837, + "learning_rate": 1.2781975159106319e-05, + "loss": 0.476, + "step": 23072 + }, + { + "epoch": 1.5485050837220227, + "grad_norm": 0.9599533081054688, + "learning_rate": 1.277471842253154e-05, + "loss": 0.4906, + "step": 23074 + }, + { + "epoch": 1.5486393074057916, + "grad_norm": 0.8366543650627136, + "learning_rate": 1.2767463444778117e-05, + "loss": 0.4661, + "step": 23076 + }, + { + "epoch": 1.5487735310895607, + "grad_norm": 0.9993776082992554, + "learning_rate": 1.276021022618883e-05, + "loss": 0.4571, + "step": 23078 + }, + { + "epoch": 1.5489077547733299, + "grad_norm": 1.1004295349121094, + "learning_rate": 1.2752958767106366e-05, + "loss": 0.482, + "step": 23080 + }, + { + "epoch": 1.5490419784570988, + "grad_norm": 0.9002885222434998, + "learning_rate": 1.2745709067873369e-05, + "loss": 0.4416, + "step": 23082 + }, + { + "epoch": 1.5491762021408677, + "grad_norm": 1.0409138202667236, + "learning_rate": 1.2738461128832353e-05, + "loss": 0.468, + "step": 23084 + }, + { + "epoch": 1.5493104258246366, + "grad_norm": 1.0249680280685425, + "learning_rate": 1.2731214950325799e-05, + "loss": 0.4723, + "step": 23086 + }, + { + "epoch": 1.5494446495084058, + "grad_norm": 1.1449722051620483, + "learning_rate": 1.2723970532696027e-05, + "loss": 0.4917, + "step": 23088 + }, + { + "epoch": 1.549578873192175, + "grad_norm": 1.0396345853805542, + "learning_rate": 1.271672787628536e-05, + "loss": 0.4533, + "step": 23090 + }, + { + "epoch": 1.5497130968759438, + "grad_norm": 0.9743292927742004, + "learning_rate": 1.2709486981435976e-05, + "loss": 0.457, + "step": 23092 + }, + { + "epoch": 1.5498473205597127, + "grad_norm": 0.9149103760719299, + "learning_rate": 1.2702247848490012e-05, + "loss": 0.4378, + "step": 23094 + }, + { + "epoch": 1.5499815442434817, + "grad_norm": 1.0200872421264648, + "learning_rate": 1.2695010477789498e-05, + "loss": 0.4957, + "step": 23096 + }, + { + "epoch": 1.5501157679272508, + "grad_norm": 0.9662615060806274, + "learning_rate": 1.2687774869676384e-05, + "loss": 0.4914, + "step": 23098 + }, + { + "epoch": 1.5502499916110197, + "grad_norm": 0.8781123161315918, + "learning_rate": 1.268054102449252e-05, + "loss": 0.3764, + "step": 23100 + }, + { + "epoch": 1.5503842152947889, + "grad_norm": 0.9832098484039307, + "learning_rate": 1.267330894257972e-05, + "loss": 0.4445, + "step": 23102 + }, + { + "epoch": 1.5505184389785578, + "grad_norm": 1.0590614080429077, + "learning_rate": 1.2666078624279676e-05, + "loss": 0.4758, + "step": 23104 + }, + { + "epoch": 1.5506526626623267, + "grad_norm": 0.986400842666626, + "learning_rate": 1.2658850069933998e-05, + "loss": 0.4434, + "step": 23106 + }, + { + "epoch": 1.5507868863460956, + "grad_norm": 1.0914992094039917, + "learning_rate": 1.2651623279884211e-05, + "loss": 0.5054, + "step": 23108 + }, + { + "epoch": 1.5509211100298648, + "grad_norm": 0.9587206840515137, + "learning_rate": 1.2644398254471801e-05, + "loss": 0.4642, + "step": 23110 + }, + { + "epoch": 1.551055333713634, + "grad_norm": 0.964069664478302, + "learning_rate": 1.2637174994038104e-05, + "loss": 0.4808, + "step": 23112 + }, + { + "epoch": 1.5511895573974028, + "grad_norm": 0.9514328837394714, + "learning_rate": 1.262995349892444e-05, + "loss": 0.4808, + "step": 23114 + }, + { + "epoch": 1.5513237810811717, + "grad_norm": 0.950463056564331, + "learning_rate": 1.2622733769471962e-05, + "loss": 0.4498, + "step": 23116 + }, + { + "epoch": 1.5514580047649407, + "grad_norm": 1.109168291091919, + "learning_rate": 1.2615515806021827e-05, + "loss": 0.4549, + "step": 23118 + }, + { + "epoch": 1.5515922284487098, + "grad_norm": 0.9191727638244629, + "learning_rate": 1.2608299608915047e-05, + "loss": 0.492, + "step": 23120 + }, + { + "epoch": 1.5517264521324787, + "grad_norm": 0.8966755270957947, + "learning_rate": 1.2601085178492594e-05, + "loss": 0.4732, + "step": 23122 + }, + { + "epoch": 1.5518606758162479, + "grad_norm": 1.1773089170455933, + "learning_rate": 1.2593872515095323e-05, + "loss": 0.4923, + "step": 23124 + }, + { + "epoch": 1.5519948995000168, + "grad_norm": 1.02364182472229, + "learning_rate": 1.2586661619064022e-05, + "loss": 0.5059, + "step": 23126 + }, + { + "epoch": 1.5521291231837857, + "grad_norm": 1.1185792684555054, + "learning_rate": 1.2579452490739374e-05, + "loss": 0.528, + "step": 23128 + }, + { + "epoch": 1.5522633468675546, + "grad_norm": 0.9567335844039917, + "learning_rate": 1.257224513046203e-05, + "loss": 0.4232, + "step": 23130 + }, + { + "epoch": 1.5523975705513238, + "grad_norm": 1.0627226829528809, + "learning_rate": 1.2565039538572499e-05, + "loss": 0.5006, + "step": 23132 + }, + { + "epoch": 1.552531794235093, + "grad_norm": 0.9607114791870117, + "learning_rate": 1.2557835715411237e-05, + "loss": 0.5347, + "step": 23134 + }, + { + "epoch": 1.5526660179188618, + "grad_norm": 1.0322315692901611, + "learning_rate": 1.25506336613186e-05, + "loss": 0.5135, + "step": 23136 + }, + { + "epoch": 1.5528002416026307, + "grad_norm": 1.1543500423431396, + "learning_rate": 1.2543433376634894e-05, + "loss": 0.5045, + "step": 23138 + }, + { + "epoch": 1.5529344652863997, + "grad_norm": 1.1164662837982178, + "learning_rate": 1.2536234861700307e-05, + "loss": 0.5522, + "step": 23140 + }, + { + "epoch": 1.5530686889701688, + "grad_norm": 1.0598642826080322, + "learning_rate": 1.2529038116854951e-05, + "loss": 0.4391, + "step": 23142 + }, + { + "epoch": 1.553202912653938, + "grad_norm": 0.9816542267799377, + "learning_rate": 1.2521843142438867e-05, + "loss": 0.4365, + "step": 23144 + }, + { + "epoch": 1.5533371363377069, + "grad_norm": 0.9278079867362976, + "learning_rate": 1.2514649938791978e-05, + "loss": 0.4136, + "step": 23146 + }, + { + "epoch": 1.5534713600214758, + "grad_norm": 0.929638147354126, + "learning_rate": 1.2507458506254178e-05, + "loss": 0.4902, + "step": 23148 + }, + { + "epoch": 1.5536055837052447, + "grad_norm": 0.9098523259162903, + "learning_rate": 1.250026884516523e-05, + "loss": 0.4483, + "step": 23150 + }, + { + "epoch": 1.5537398073890136, + "grad_norm": 1.2403209209442139, + "learning_rate": 1.249308095586486e-05, + "loss": 0.5198, + "step": 23152 + }, + { + "epoch": 1.5538740310727828, + "grad_norm": 0.8855780363082886, + "learning_rate": 1.2485894838692641e-05, + "loss": 0.4567, + "step": 23154 + }, + { + "epoch": 1.554008254756552, + "grad_norm": 1.0350875854492188, + "learning_rate": 1.2478710493988133e-05, + "loss": 0.4493, + "step": 23156 + }, + { + "epoch": 1.5541424784403208, + "grad_norm": 0.9332207441329956, + "learning_rate": 1.2471527922090754e-05, + "loss": 0.5363, + "step": 23158 + }, + { + "epoch": 1.5542767021240897, + "grad_norm": 0.9449737668037415, + "learning_rate": 1.2464347123339904e-05, + "loss": 0.5492, + "step": 23160 + }, + { + "epoch": 1.5544109258078587, + "grad_norm": 1.0071675777435303, + "learning_rate": 1.2457168098074839e-05, + "loss": 0.5195, + "step": 23162 + }, + { + "epoch": 1.5545451494916278, + "grad_norm": 1.085398554801941, + "learning_rate": 1.2449990846634757e-05, + "loss": 0.4877, + "step": 23164 + }, + { + "epoch": 1.554679373175397, + "grad_norm": 1.099508285522461, + "learning_rate": 1.244281536935875e-05, + "loss": 0.5213, + "step": 23166 + }, + { + "epoch": 1.5548135968591659, + "grad_norm": 1.03630793094635, + "learning_rate": 1.2435641666585884e-05, + "loss": 0.4711, + "step": 23168 + }, + { + "epoch": 1.5549478205429348, + "grad_norm": 0.9316409826278687, + "learning_rate": 1.242846973865508e-05, + "loss": 0.4284, + "step": 23170 + }, + { + "epoch": 1.5550820442267037, + "grad_norm": 0.9820182919502258, + "learning_rate": 1.24212995859052e-05, + "loss": 0.504, + "step": 23172 + }, + { + "epoch": 1.5552162679104728, + "grad_norm": 1.2011853456497192, + "learning_rate": 1.2414131208675006e-05, + "loss": 0.5168, + "step": 23174 + }, + { + "epoch": 1.5553504915942418, + "grad_norm": 0.9982286691665649, + "learning_rate": 1.2406964607303212e-05, + "loss": 0.4828, + "step": 23176 + }, + { + "epoch": 1.555484715278011, + "grad_norm": 1.0185546875, + "learning_rate": 1.2399799782128407e-05, + "loss": 0.4556, + "step": 23178 + }, + { + "epoch": 1.5556189389617798, + "grad_norm": 0.9433539509773254, + "learning_rate": 1.2392636733489155e-05, + "loss": 0.5052, + "step": 23180 + }, + { + "epoch": 1.5557531626455487, + "grad_norm": 1.020642876625061, + "learning_rate": 1.2385475461723833e-05, + "loss": 0.5012, + "step": 23182 + }, + { + "epoch": 1.5558873863293177, + "grad_norm": 0.9557784795761108, + "learning_rate": 1.2378315967170845e-05, + "loss": 0.4644, + "step": 23184 + }, + { + "epoch": 1.5560216100130868, + "grad_norm": 1.135974645614624, + "learning_rate": 1.2371158250168436e-05, + "loss": 0.5054, + "step": 23186 + }, + { + "epoch": 1.556155833696856, + "grad_norm": 0.9932438731193542, + "learning_rate": 1.2364002311054818e-05, + "loss": 0.4858, + "step": 23188 + }, + { + "epoch": 1.5562900573806249, + "grad_norm": 1.077278971672058, + "learning_rate": 1.2356848150168082e-05, + "loss": 0.4898, + "step": 23190 + }, + { + "epoch": 1.5564242810643938, + "grad_norm": 1.1726289987564087, + "learning_rate": 1.2349695767846247e-05, + "loss": 0.491, + "step": 23192 + }, + { + "epoch": 1.5565585047481627, + "grad_norm": 1.0936367511749268, + "learning_rate": 1.2342545164427244e-05, + "loss": 0.4557, + "step": 23194 + }, + { + "epoch": 1.5566927284319318, + "grad_norm": 1.0036332607269287, + "learning_rate": 1.233539634024894e-05, + "loss": 0.4511, + "step": 23196 + }, + { + "epoch": 1.5568269521157008, + "grad_norm": 1.285355806350708, + "learning_rate": 1.2328249295649097e-05, + "loss": 0.4976, + "step": 23198 + }, + { + "epoch": 1.55696117579947, + "grad_norm": 1.0038576126098633, + "learning_rate": 1.232110403096539e-05, + "loss": 0.4419, + "step": 23200 + }, + { + "epoch": 1.5570953994832388, + "grad_norm": 0.9209941029548645, + "learning_rate": 1.2313960546535419e-05, + "loss": 0.4389, + "step": 23202 + }, + { + "epoch": 1.5572296231670077, + "grad_norm": 1.1058903932571411, + "learning_rate": 1.2306818842696716e-05, + "loss": 0.51, + "step": 23204 + }, + { + "epoch": 1.5573638468507767, + "grad_norm": 1.073699712753296, + "learning_rate": 1.2299678919786689e-05, + "loss": 0.5003, + "step": 23206 + }, + { + "epoch": 1.5574980705345458, + "grad_norm": 1.0794081687927246, + "learning_rate": 1.229254077814273e-05, + "loss": 0.5043, + "step": 23208 + }, + { + "epoch": 1.557632294218315, + "grad_norm": 0.9014645218849182, + "learning_rate": 1.2285404418102053e-05, + "loss": 0.4542, + "step": 23210 + }, + { + "epoch": 1.5577665179020839, + "grad_norm": 0.9933221936225891, + "learning_rate": 1.2278269840001838e-05, + "loss": 0.4572, + "step": 23212 + }, + { + "epoch": 1.5579007415858528, + "grad_norm": 1.0882283449172974, + "learning_rate": 1.2271137044179215e-05, + "loss": 0.4931, + "step": 23214 + }, + { + "epoch": 1.5580349652696217, + "grad_norm": 1.047158122062683, + "learning_rate": 1.226400603097116e-05, + "loss": 0.4966, + "step": 23216 + }, + { + "epoch": 1.5581691889533908, + "grad_norm": 1.0260157585144043, + "learning_rate": 1.2256876800714633e-05, + "loss": 0.4813, + "step": 23218 + }, + { + "epoch": 1.55830341263716, + "grad_norm": 0.9270790219306946, + "learning_rate": 1.2249749353746454e-05, + "loss": 0.4752, + "step": 23220 + }, + { + "epoch": 1.558437636320929, + "grad_norm": 0.8672318458557129, + "learning_rate": 1.2242623690403387e-05, + "loss": 0.4314, + "step": 23222 + }, + { + "epoch": 1.5585718600046978, + "grad_norm": 0.9131848216056824, + "learning_rate": 1.223549981102209e-05, + "loss": 0.4286, + "step": 23224 + }, + { + "epoch": 1.5587060836884667, + "grad_norm": 1.1069892644882202, + "learning_rate": 1.2228377715939183e-05, + "loss": 0.4977, + "step": 23226 + }, + { + "epoch": 1.5588403073722357, + "grad_norm": 1.1032196283340454, + "learning_rate": 1.2221257405491148e-05, + "loss": 0.4791, + "step": 23228 + }, + { + "epoch": 1.5589745310560048, + "grad_norm": 1.0574241876602173, + "learning_rate": 1.2214138880014414e-05, + "loss": 0.5097, + "step": 23230 + }, + { + "epoch": 1.559108754739774, + "grad_norm": 0.9770507216453552, + "learning_rate": 1.2207022139845303e-05, + "loss": 0.4435, + "step": 23232 + }, + { + "epoch": 1.5592429784235429, + "grad_norm": 1.1765968799591064, + "learning_rate": 1.2199907185320091e-05, + "loss": 0.4514, + "step": 23234 + }, + { + "epoch": 1.5593772021073118, + "grad_norm": 1.059708595275879, + "learning_rate": 1.2192794016774922e-05, + "loss": 0.3983, + "step": 23236 + }, + { + "epoch": 1.5595114257910807, + "grad_norm": 1.7546675205230713, + "learning_rate": 1.218568263454592e-05, + "loss": 0.4223, + "step": 23238 + }, + { + "epoch": 1.5596456494748498, + "grad_norm": 1.036199927330017, + "learning_rate": 1.2178573038969026e-05, + "loss": 0.4571, + "step": 23240 + }, + { + "epoch": 1.559779873158619, + "grad_norm": 1.1719951629638672, + "learning_rate": 1.2171465230380196e-05, + "loss": 0.4786, + "step": 23242 + }, + { + "epoch": 1.559914096842388, + "grad_norm": 0.9336151480674744, + "learning_rate": 1.2164359209115234e-05, + "loss": 0.45, + "step": 23244 + }, + { + "epoch": 1.5600483205261568, + "grad_norm": 0.9916703104972839, + "learning_rate": 1.2157254975509912e-05, + "loss": 0.4858, + "step": 23246 + }, + { + "epoch": 1.5601825442099257, + "grad_norm": 0.9358370304107666, + "learning_rate": 1.2150152529899878e-05, + "loss": 0.4647, + "step": 23248 + }, + { + "epoch": 1.560316767893695, + "grad_norm": 1.0746142864227295, + "learning_rate": 1.2143051872620708e-05, + "loss": 0.4617, + "step": 23250 + }, + { + "epoch": 1.5604509915774638, + "grad_norm": 1.036489486694336, + "learning_rate": 1.2135953004007882e-05, + "loss": 0.486, + "step": 23252 + }, + { + "epoch": 1.560585215261233, + "grad_norm": 1.023247241973877, + "learning_rate": 1.2128855924396837e-05, + "loss": 0.5029, + "step": 23254 + }, + { + "epoch": 1.5607194389450019, + "grad_norm": 0.9809913039207458, + "learning_rate": 1.2121760634122875e-05, + "loss": 0.4899, + "step": 23256 + }, + { + "epoch": 1.5608536626287708, + "grad_norm": 0.9748601317405701, + "learning_rate": 1.211466713352124e-05, + "loss": 0.4277, + "step": 23258 + }, + { + "epoch": 1.5609878863125397, + "grad_norm": 1.083095908164978, + "learning_rate": 1.210757542292707e-05, + "loss": 0.5058, + "step": 23260 + }, + { + "epoch": 1.5611221099963088, + "grad_norm": 1.1028939485549927, + "learning_rate": 1.2100485502675463e-05, + "loss": 0.5041, + "step": 23262 + }, + { + "epoch": 1.561256333680078, + "grad_norm": 1.063094139099121, + "learning_rate": 1.2093397373101373e-05, + "loss": 0.5128, + "step": 23264 + }, + { + "epoch": 1.561390557363847, + "grad_norm": 1.0166436433792114, + "learning_rate": 1.2086311034539749e-05, + "loss": 0.5278, + "step": 23266 + }, + { + "epoch": 1.5615247810476158, + "grad_norm": 1.1538267135620117, + "learning_rate": 1.2079226487325347e-05, + "loss": 0.4639, + "step": 23268 + }, + { + "epoch": 1.5616590047313847, + "grad_norm": 1.1826285123825073, + "learning_rate": 1.2072143731792934e-05, + "loss": 0.4692, + "step": 23270 + }, + { + "epoch": 1.561793228415154, + "grad_norm": 0.9804865717887878, + "learning_rate": 1.2065062768277135e-05, + "loss": 0.4773, + "step": 23272 + }, + { + "epoch": 1.5619274520989228, + "grad_norm": 1.1251349449157715, + "learning_rate": 1.2057983597112543e-05, + "loss": 0.4574, + "step": 23274 + }, + { + "epoch": 1.562061675782692, + "grad_norm": 1.0487046241760254, + "learning_rate": 1.2050906218633617e-05, + "loss": 0.5116, + "step": 23276 + }, + { + "epoch": 1.5621958994664609, + "grad_norm": 1.1605690717697144, + "learning_rate": 1.2043830633174747e-05, + "loss": 0.4846, + "step": 23278 + }, + { + "epoch": 1.5623301231502298, + "grad_norm": 0.9828898310661316, + "learning_rate": 1.203675684107024e-05, + "loss": 0.4914, + "step": 23280 + }, + { + "epoch": 1.5624643468339987, + "grad_norm": 1.118901252746582, + "learning_rate": 1.2029684842654304e-05, + "loss": 0.4927, + "step": 23282 + }, + { + "epoch": 1.5625985705177678, + "grad_norm": 1.0152097940444946, + "learning_rate": 1.2022614638261109e-05, + "loss": 0.5329, + "step": 23284 + }, + { + "epoch": 1.562732794201537, + "grad_norm": 1.0980987548828125, + "learning_rate": 1.20155462282247e-05, + "loss": 0.4771, + "step": 23286 + }, + { + "epoch": 1.562867017885306, + "grad_norm": 1.060423731803894, + "learning_rate": 1.2008479612879032e-05, + "loss": 0.5249, + "step": 23288 + }, + { + "epoch": 1.5630012415690748, + "grad_norm": 1.042399287223816, + "learning_rate": 1.200141479255798e-05, + "loss": 0.5566, + "step": 23290 + }, + { + "epoch": 1.5631354652528437, + "grad_norm": 1.145865797996521, + "learning_rate": 1.1994351767595374e-05, + "loss": 0.5369, + "step": 23292 + }, + { + "epoch": 1.5632696889366129, + "grad_norm": 1.0151365995407104, + "learning_rate": 1.1987290538324914e-05, + "loss": 0.4685, + "step": 23294 + }, + { + "epoch": 1.563403912620382, + "grad_norm": 1.0187491178512573, + "learning_rate": 1.1980231105080225e-05, + "loss": 0.5318, + "step": 23296 + }, + { + "epoch": 1.563538136304151, + "grad_norm": 0.933469295501709, + "learning_rate": 1.1973173468194843e-05, + "loss": 0.4409, + "step": 23298 + }, + { + "epoch": 1.5636723599879199, + "grad_norm": 1.0887510776519775, + "learning_rate": 1.1966117628002249e-05, + "loss": 0.521, + "step": 23300 + }, + { + "epoch": 1.5638065836716888, + "grad_norm": 0.9163106083869934, + "learning_rate": 1.1959063584835795e-05, + "loss": 0.4308, + "step": 23302 + }, + { + "epoch": 1.5639408073554577, + "grad_norm": 0.9544594287872314, + "learning_rate": 1.1952011339028807e-05, + "loss": 0.48, + "step": 23304 + }, + { + "epoch": 1.5640750310392268, + "grad_norm": 1.0309133529663086, + "learning_rate": 1.1944960890914442e-05, + "loss": 0.4892, + "step": 23306 + }, + { + "epoch": 1.564209254722996, + "grad_norm": 1.001133918762207, + "learning_rate": 1.1937912240825849e-05, + "loss": 0.4814, + "step": 23308 + }, + { + "epoch": 1.564343478406765, + "grad_norm": 0.9937286972999573, + "learning_rate": 1.1930865389096052e-05, + "loss": 0.5181, + "step": 23310 + }, + { + "epoch": 1.5644777020905338, + "grad_norm": 0.9627927541732788, + "learning_rate": 1.1923820336058012e-05, + "loss": 0.4482, + "step": 23312 + }, + { + "epoch": 1.5646119257743027, + "grad_norm": 1.146790623664856, + "learning_rate": 1.1916777082044588e-05, + "loss": 0.4795, + "step": 23314 + }, + { + "epoch": 1.5647461494580719, + "grad_norm": 1.0631041526794434, + "learning_rate": 1.1909735627388562e-05, + "loss": 0.4174, + "step": 23316 + }, + { + "epoch": 1.564880373141841, + "grad_norm": 1.1096659898757935, + "learning_rate": 1.1902695972422618e-05, + "loss": 0.4753, + "step": 23318 + }, + { + "epoch": 1.56501459682561, + "grad_norm": 1.0855131149291992, + "learning_rate": 1.189565811747938e-05, + "loss": 0.5567, + "step": 23320 + }, + { + "epoch": 1.5651488205093789, + "grad_norm": 1.2113926410675049, + "learning_rate": 1.188862206289137e-05, + "loss": 0.4981, + "step": 23322 + }, + { + "epoch": 1.5652830441931478, + "grad_norm": 0.9407163858413696, + "learning_rate": 1.1881587808991023e-05, + "loss": 0.5004, + "step": 23324 + }, + { + "epoch": 1.565417267876917, + "grad_norm": 1.029340386390686, + "learning_rate": 1.1874555356110684e-05, + "loss": 0.5376, + "step": 23326 + }, + { + "epoch": 1.5655514915606858, + "grad_norm": 1.0241527557373047, + "learning_rate": 1.1867524704582644e-05, + "loss": 0.4501, + "step": 23328 + }, + { + "epoch": 1.565685715244455, + "grad_norm": 1.0108983516693115, + "learning_rate": 1.1860495854739067e-05, + "loss": 0.4492, + "step": 23330 + }, + { + "epoch": 1.565819938928224, + "grad_norm": 1.048672080039978, + "learning_rate": 1.1853468806912088e-05, + "loss": 0.4283, + "step": 23332 + }, + { + "epoch": 1.5659541626119928, + "grad_norm": 1.066028356552124, + "learning_rate": 1.1846443561433673e-05, + "loss": 0.427, + "step": 23334 + }, + { + "epoch": 1.5660883862957617, + "grad_norm": 0.9435437917709351, + "learning_rate": 1.1839420118635785e-05, + "loss": 0.4447, + "step": 23336 + }, + { + "epoch": 1.5662226099795309, + "grad_norm": 0.912183940410614, + "learning_rate": 1.1832398478850243e-05, + "loss": 0.4356, + "step": 23338 + }, + { + "epoch": 1.5663568336633, + "grad_norm": 0.9043480157852173, + "learning_rate": 1.182537864240883e-05, + "loss": 0.4056, + "step": 23340 + }, + { + "epoch": 1.566491057347069, + "grad_norm": 0.9847351312637329, + "learning_rate": 1.1818360609643215e-05, + "loss": 0.4958, + "step": 23342 + }, + { + "epoch": 1.5666252810308379, + "grad_norm": 1.0925720930099487, + "learning_rate": 1.1811344380884975e-05, + "loss": 0.5238, + "step": 23344 + }, + { + "epoch": 1.5667595047146068, + "grad_norm": 1.146865963935852, + "learning_rate": 1.180432995646562e-05, + "loss": 0.4531, + "step": 23346 + }, + { + "epoch": 1.566893728398376, + "grad_norm": 0.9299564361572266, + "learning_rate": 1.1797317336716552e-05, + "loss": 0.5076, + "step": 23348 + }, + { + "epoch": 1.5670279520821448, + "grad_norm": 1.033992886543274, + "learning_rate": 1.179030652196913e-05, + "loss": 0.4787, + "step": 23350 + }, + { + "epoch": 1.567162175765914, + "grad_norm": 1.011711835861206, + "learning_rate": 1.1783297512554591e-05, + "loss": 0.5177, + "step": 23352 + }, + { + "epoch": 1.567296399449683, + "grad_norm": 0.9572144746780396, + "learning_rate": 1.1776290308804095e-05, + "loss": 0.4684, + "step": 23354 + }, + { + "epoch": 1.5674306231334518, + "grad_norm": 1.0200315713882446, + "learning_rate": 1.1769284911048706e-05, + "loss": 0.4596, + "step": 23356 + }, + { + "epoch": 1.5675648468172207, + "grad_norm": 0.9919617772102356, + "learning_rate": 1.176228131961944e-05, + "loss": 0.451, + "step": 23358 + }, + { + "epoch": 1.5676990705009899, + "grad_norm": 1.0359539985656738, + "learning_rate": 1.1755279534847186e-05, + "loss": 0.4625, + "step": 23360 + }, + { + "epoch": 1.567833294184759, + "grad_norm": 1.0746967792510986, + "learning_rate": 1.1748279557062775e-05, + "loss": 0.4682, + "step": 23362 + }, + { + "epoch": 1.567967517868528, + "grad_norm": 1.1900519132614136, + "learning_rate": 1.1741281386596913e-05, + "loss": 0.4418, + "step": 23364 + }, + { + "epoch": 1.5681017415522969, + "grad_norm": 0.9637075066566467, + "learning_rate": 1.1734285023780295e-05, + "loss": 0.4454, + "step": 23366 + }, + { + "epoch": 1.5682359652360658, + "grad_norm": 1.0022313594818115, + "learning_rate": 1.1727290468943447e-05, + "loss": 0.4795, + "step": 23368 + }, + { + "epoch": 1.568370188919835, + "grad_norm": 1.1569160223007202, + "learning_rate": 1.172029772241688e-05, + "loss": 0.5055, + "step": 23370 + }, + { + "epoch": 1.568504412603604, + "grad_norm": 0.9874303936958313, + "learning_rate": 1.171330678453097e-05, + "loss": 0.4996, + "step": 23372 + }, + { + "epoch": 1.568638636287373, + "grad_norm": 1.0189257860183716, + "learning_rate": 1.1706317655616029e-05, + "loss": 0.427, + "step": 23374 + }, + { + "epoch": 1.568772859971142, + "grad_norm": 0.8920593857765198, + "learning_rate": 1.1699330336002261e-05, + "loss": 0.4697, + "step": 23376 + }, + { + "epoch": 1.5689070836549108, + "grad_norm": 1.0679326057434082, + "learning_rate": 1.1692344826019835e-05, + "loss": 0.4625, + "step": 23378 + }, + { + "epoch": 1.5690413073386797, + "grad_norm": 1.0527321100234985, + "learning_rate": 1.168536112599879e-05, + "loss": 0.5308, + "step": 23380 + }, + { + "epoch": 1.5691755310224489, + "grad_norm": 1.0946030616760254, + "learning_rate": 1.1678379236269082e-05, + "loss": 0.4846, + "step": 23382 + }, + { + "epoch": 1.569309754706218, + "grad_norm": 1.0754079818725586, + "learning_rate": 1.1671399157160589e-05, + "loss": 0.417, + "step": 23384 + }, + { + "epoch": 1.569443978389987, + "grad_norm": 1.1126391887664795, + "learning_rate": 1.1664420889003131e-05, + "loss": 0.4788, + "step": 23386 + }, + { + "epoch": 1.5695782020737559, + "grad_norm": 0.9987939596176147, + "learning_rate": 1.1657444432126403e-05, + "loss": 0.4408, + "step": 23388 + }, + { + "epoch": 1.5697124257575248, + "grad_norm": 1.1336610317230225, + "learning_rate": 1.1650469786860025e-05, + "loss": 0.5384, + "step": 23390 + }, + { + "epoch": 1.569846649441294, + "grad_norm": 0.9724118709564209, + "learning_rate": 1.1643496953533522e-05, + "loss": 0.459, + "step": 23392 + }, + { + "epoch": 1.569980873125063, + "grad_norm": 0.9910349249839783, + "learning_rate": 1.1636525932476382e-05, + "loss": 0.4318, + "step": 23394 + }, + { + "epoch": 1.570115096808832, + "grad_norm": 0.9914292693138123, + "learning_rate": 1.1629556724017936e-05, + "loss": 0.4409, + "step": 23396 + }, + { + "epoch": 1.570249320492601, + "grad_norm": 0.9634431600570679, + "learning_rate": 1.1622589328487504e-05, + "loss": 0.5045, + "step": 23398 + }, + { + "epoch": 1.5703835441763698, + "grad_norm": 1.0638256072998047, + "learning_rate": 1.1615623746214255e-05, + "loss": 0.4842, + "step": 23400 + }, + { + "epoch": 1.570517767860139, + "grad_norm": 0.990250825881958, + "learning_rate": 1.1608659977527303e-05, + "loss": 0.4729, + "step": 23402 + }, + { + "epoch": 1.5706519915439079, + "grad_norm": 0.9492442607879639, + "learning_rate": 1.1601698022755658e-05, + "loss": 0.4816, + "step": 23404 + }, + { + "epoch": 1.570786215227677, + "grad_norm": 1.1385966539382935, + "learning_rate": 1.159473788222829e-05, + "loss": 0.4708, + "step": 23406 + }, + { + "epoch": 1.570920438911446, + "grad_norm": 1.0014199018478394, + "learning_rate": 1.1587779556274042e-05, + "loss": 0.4839, + "step": 23408 + }, + { + "epoch": 1.5710546625952149, + "grad_norm": 0.9885188341140747, + "learning_rate": 1.1580823045221673e-05, + "loss": 0.4584, + "step": 23410 + }, + { + "epoch": 1.5711888862789838, + "grad_norm": 0.9447250366210938, + "learning_rate": 1.1573868349399848e-05, + "loss": 0.4484, + "step": 23412 + }, + { + "epoch": 1.571323109962753, + "grad_norm": 0.9757140278816223, + "learning_rate": 1.1566915469137201e-05, + "loss": 0.4863, + "step": 23414 + }, + { + "epoch": 1.571457333646522, + "grad_norm": 1.0615960359573364, + "learning_rate": 1.155996440476222e-05, + "loss": 0.4747, + "step": 23416 + }, + { + "epoch": 1.571591557330291, + "grad_norm": 0.9987040162086487, + "learning_rate": 1.1553015156603331e-05, + "loss": 0.4632, + "step": 23418 + }, + { + "epoch": 1.57172578101406, + "grad_norm": 1.0166984796524048, + "learning_rate": 1.1546067724988873e-05, + "loss": 0.4359, + "step": 23420 + }, + { + "epoch": 1.5718600046978288, + "grad_norm": 0.9812948107719421, + "learning_rate": 1.1539122110247086e-05, + "loss": 0.4624, + "step": 23422 + }, + { + "epoch": 1.571994228381598, + "grad_norm": 1.099785327911377, + "learning_rate": 1.1532178312706166e-05, + "loss": 0.4843, + "step": 23424 + }, + { + "epoch": 1.5721284520653669, + "grad_norm": 1.042169213294983, + "learning_rate": 1.1525236332694162e-05, + "loss": 0.5418, + "step": 23426 + }, + { + "epoch": 1.572262675749136, + "grad_norm": 1.0316282510757446, + "learning_rate": 1.1518296170539105e-05, + "loss": 0.4221, + "step": 23428 + }, + { + "epoch": 1.572396899432905, + "grad_norm": 1.0206512212753296, + "learning_rate": 1.1511357826568864e-05, + "loss": 0.4744, + "step": 23430 + }, + { + "epoch": 1.5725311231166739, + "grad_norm": 1.0751756429672241, + "learning_rate": 1.150442130111129e-05, + "loss": 0.532, + "step": 23432 + }, + { + "epoch": 1.5726653468004428, + "grad_norm": 1.0585800409317017, + "learning_rate": 1.14974865944941e-05, + "loss": 0.4456, + "step": 23434 + }, + { + "epoch": 1.572799570484212, + "grad_norm": 1.1338976621627808, + "learning_rate": 1.1490553707044965e-05, + "loss": 0.5337, + "step": 23436 + }, + { + "epoch": 1.572933794167981, + "grad_norm": 0.9720898270606995, + "learning_rate": 1.1483622639091446e-05, + "loss": 0.5049, + "step": 23438 + }, + { + "epoch": 1.57306801785175, + "grad_norm": 1.065352439880371, + "learning_rate": 1.1476693390961019e-05, + "loss": 0.4294, + "step": 23440 + }, + { + "epoch": 1.573202241535519, + "grad_norm": 0.9880411028862, + "learning_rate": 1.146976596298106e-05, + "loss": 0.4533, + "step": 23442 + }, + { + "epoch": 1.5733364652192878, + "grad_norm": 1.2746909856796265, + "learning_rate": 1.146284035547891e-05, + "loss": 0.4535, + "step": 23444 + }, + { + "epoch": 1.573470688903057, + "grad_norm": 0.8341091275215149, + "learning_rate": 1.1455916568781772e-05, + "loss": 0.4299, + "step": 23446 + }, + { + "epoch": 1.573604912586826, + "grad_norm": 1.156506061553955, + "learning_rate": 1.1448994603216779e-05, + "loss": 0.4808, + "step": 23448 + }, + { + "epoch": 1.573739136270595, + "grad_norm": 1.1684461832046509, + "learning_rate": 1.1442074459110974e-05, + "loss": 0.4736, + "step": 23450 + }, + { + "epoch": 1.573873359954364, + "grad_norm": 1.2712520360946655, + "learning_rate": 1.1435156136791342e-05, + "loss": 0.5259, + "step": 23452 + }, + { + "epoch": 1.5740075836381329, + "grad_norm": 1.3483922481536865, + "learning_rate": 1.1428239636584732e-05, + "loss": 0.509, + "step": 23454 + }, + { + "epoch": 1.5741418073219018, + "grad_norm": 1.0553919076919556, + "learning_rate": 1.142132495881798e-05, + "loss": 0.4452, + "step": 23456 + }, + { + "epoch": 1.574276031005671, + "grad_norm": 1.154910683631897, + "learning_rate": 1.141441210381774e-05, + "loss": 0.5168, + "step": 23458 + }, + { + "epoch": 1.57441025468944, + "grad_norm": 0.8882455229759216, + "learning_rate": 1.140750107191066e-05, + "loss": 0.4366, + "step": 23460 + }, + { + "epoch": 1.574544478373209, + "grad_norm": 1.1063082218170166, + "learning_rate": 1.1400591863423254e-05, + "loss": 0.5155, + "step": 23462 + }, + { + "epoch": 1.574678702056978, + "grad_norm": 1.195164442062378, + "learning_rate": 1.1393684478681993e-05, + "loss": 0.4859, + "step": 23464 + }, + { + "epoch": 1.5748129257407468, + "grad_norm": 1.1390354633331299, + "learning_rate": 1.1386778918013225e-05, + "loss": 0.5171, + "step": 23466 + }, + { + "epoch": 1.574947149424516, + "grad_norm": 1.1175415515899658, + "learning_rate": 1.1379875181743222e-05, + "loss": 0.477, + "step": 23468 + }, + { + "epoch": 1.575081373108285, + "grad_norm": 1.0612244606018066, + "learning_rate": 1.1372973270198163e-05, + "loss": 0.5073, + "step": 23470 + }, + { + "epoch": 1.575215596792054, + "grad_norm": 1.0456383228302002, + "learning_rate": 1.1366073183704167e-05, + "loss": 0.4893, + "step": 23472 + }, + { + "epoch": 1.575349820475823, + "grad_norm": 1.1216797828674316, + "learning_rate": 1.1359174922587251e-05, + "loss": 0.5365, + "step": 23474 + }, + { + "epoch": 1.5754840441595919, + "grad_norm": 0.8650084137916565, + "learning_rate": 1.1352278487173328e-05, + "loss": 0.4372, + "step": 23476 + }, + { + "epoch": 1.575618267843361, + "grad_norm": 1.2271318435668945, + "learning_rate": 1.1345383877788241e-05, + "loss": 0.4567, + "step": 23478 + }, + { + "epoch": 1.57575249152713, + "grad_norm": 1.0574833154678345, + "learning_rate": 1.133849109475777e-05, + "loss": 0.5015, + "step": 23480 + }, + { + "epoch": 1.575886715210899, + "grad_norm": 1.1210253238677979, + "learning_rate": 1.133160013840755e-05, + "loss": 0.4641, + "step": 23482 + }, + { + "epoch": 1.576020938894668, + "grad_norm": 0.9176300168037415, + "learning_rate": 1.132471100906322e-05, + "loss": 0.4495, + "step": 23484 + }, + { + "epoch": 1.576155162578437, + "grad_norm": 1.1437222957611084, + "learning_rate": 1.1317823707050224e-05, + "loss": 0.5005, + "step": 23486 + }, + { + "epoch": 1.5762893862622058, + "grad_norm": 0.9332422018051147, + "learning_rate": 1.131093823269398e-05, + "loss": 0.4629, + "step": 23488 + }, + { + "epoch": 1.576423609945975, + "grad_norm": 1.0111188888549805, + "learning_rate": 1.130405458631984e-05, + "loss": 0.4381, + "step": 23490 + }, + { + "epoch": 1.576557833629744, + "grad_norm": 1.0104448795318604, + "learning_rate": 1.1297172768253018e-05, + "loss": 0.4717, + "step": 23492 + }, + { + "epoch": 1.576692057313513, + "grad_norm": 1.1603244543075562, + "learning_rate": 1.1290292778818684e-05, + "loss": 0.4849, + "step": 23494 + }, + { + "epoch": 1.576826280997282, + "grad_norm": 1.1314606666564941, + "learning_rate": 1.1283414618341903e-05, + "loss": 0.5394, + "step": 23496 + }, + { + "epoch": 1.5769605046810509, + "grad_norm": 1.1823599338531494, + "learning_rate": 1.1276538287147642e-05, + "loss": 0.4741, + "step": 23498 + }, + { + "epoch": 1.57709472836482, + "grad_norm": 1.140844702720642, + "learning_rate": 1.1269663785560792e-05, + "loss": 0.5258, + "step": 23500 + }, + { + "epoch": 1.577228952048589, + "grad_norm": 1.1149667501449585, + "learning_rate": 1.1262791113906179e-05, + "loss": 0.5388, + "step": 23502 + }, + { + "epoch": 1.577363175732358, + "grad_norm": 0.9421783089637756, + "learning_rate": 1.1255920272508519e-05, + "loss": 0.4208, + "step": 23504 + }, + { + "epoch": 1.577497399416127, + "grad_norm": 1.1093504428863525, + "learning_rate": 1.1249051261692433e-05, + "loss": 0.511, + "step": 23506 + }, + { + "epoch": 1.577631623099896, + "grad_norm": 1.1066049337387085, + "learning_rate": 1.124218408178246e-05, + "loss": 0.4689, + "step": 23508 + }, + { + "epoch": 1.5777658467836648, + "grad_norm": 1.0475133657455444, + "learning_rate": 1.1235318733103089e-05, + "loss": 0.5488, + "step": 23510 + }, + { + "epoch": 1.577900070467434, + "grad_norm": 1.338498830795288, + "learning_rate": 1.1228455215978683e-05, + "loss": 0.448, + "step": 23512 + }, + { + "epoch": 1.578034294151203, + "grad_norm": 0.8669447302818298, + "learning_rate": 1.1221593530733527e-05, + "loss": 0.4772, + "step": 23514 + }, + { + "epoch": 1.578168517834972, + "grad_norm": 1.027613878250122, + "learning_rate": 1.1214733677691808e-05, + "loss": 0.5029, + "step": 23516 + }, + { + "epoch": 1.578302741518741, + "grad_norm": 0.9990667700767517, + "learning_rate": 1.1207875657177663e-05, + "loss": 0.4554, + "step": 23518 + }, + { + "epoch": 1.5784369652025099, + "grad_norm": 1.090484380722046, + "learning_rate": 1.1201019469515106e-05, + "loss": 0.4944, + "step": 23520 + }, + { + "epoch": 1.578571188886279, + "grad_norm": 0.8236868381500244, + "learning_rate": 1.1194165115028094e-05, + "loss": 0.453, + "step": 23522 + }, + { + "epoch": 1.5787054125700481, + "grad_norm": 0.9789796471595764, + "learning_rate": 1.1187312594040472e-05, + "loss": 0.5029, + "step": 23524 + }, + { + "epoch": 1.578839636253817, + "grad_norm": 0.8634899258613586, + "learning_rate": 1.118046190687601e-05, + "loss": 0.4639, + "step": 23526 + }, + { + "epoch": 1.578973859937586, + "grad_norm": 1.1111963987350464, + "learning_rate": 1.1173613053858373e-05, + "loss": 0.5827, + "step": 23528 + }, + { + "epoch": 1.579108083621355, + "grad_norm": 0.9517291784286499, + "learning_rate": 1.1166766035311182e-05, + "loss": 0.4568, + "step": 23530 + }, + { + "epoch": 1.5792423073051238, + "grad_norm": 1.070487380027771, + "learning_rate": 1.1159920851557937e-05, + "loss": 0.4755, + "step": 23532 + }, + { + "epoch": 1.579376530988893, + "grad_norm": 1.080098271369934, + "learning_rate": 1.1153077502922055e-05, + "loss": 0.4911, + "step": 23534 + }, + { + "epoch": 1.579510754672662, + "grad_norm": 1.1324330568313599, + "learning_rate": 1.1146235989726856e-05, + "loss": 0.5065, + "step": 23536 + }, + { + "epoch": 1.579644978356431, + "grad_norm": 1.0070408582687378, + "learning_rate": 1.113939631229562e-05, + "loss": 0.4729, + "step": 23538 + }, + { + "epoch": 1.5797792020402, + "grad_norm": 1.0733522176742554, + "learning_rate": 1.1132558470951487e-05, + "loss": 0.5337, + "step": 23540 + }, + { + "epoch": 1.5799134257239689, + "grad_norm": 1.0693459510803223, + "learning_rate": 1.1125722466017547e-05, + "loss": 0.4989, + "step": 23542 + }, + { + "epoch": 1.580047649407738, + "grad_norm": 1.1518325805664062, + "learning_rate": 1.1118888297816754e-05, + "loss": 0.4878, + "step": 23544 + }, + { + "epoch": 1.5801818730915071, + "grad_norm": 1.1432942152023315, + "learning_rate": 1.1112055966672053e-05, + "loss": 0.461, + "step": 23546 + }, + { + "epoch": 1.580316096775276, + "grad_norm": 0.9708120822906494, + "learning_rate": 1.1105225472906216e-05, + "loss": 0.4498, + "step": 23548 + }, + { + "epoch": 1.580450320459045, + "grad_norm": 0.8708906173706055, + "learning_rate": 1.109839681684201e-05, + "loss": 0.4934, + "step": 23550 + }, + { + "epoch": 1.580584544142814, + "grad_norm": 0.9430227279663086, + "learning_rate": 1.1091569998802075e-05, + "loss": 0.4438, + "step": 23552 + }, + { + "epoch": 1.580718767826583, + "grad_norm": 0.990928590297699, + "learning_rate": 1.1084745019108916e-05, + "loss": 0.4344, + "step": 23554 + }, + { + "epoch": 1.580852991510352, + "grad_norm": 1.0264774560928345, + "learning_rate": 1.1077921878085047e-05, + "loss": 0.4769, + "step": 23556 + }, + { + "epoch": 1.580987215194121, + "grad_norm": 1.0623157024383545, + "learning_rate": 1.1071100576052818e-05, + "loss": 0.5127, + "step": 23558 + }, + { + "epoch": 1.58112143887789, + "grad_norm": 1.053694486618042, + "learning_rate": 1.1064281113334546e-05, + "loss": 0.4593, + "step": 23560 + }, + { + "epoch": 1.581255662561659, + "grad_norm": 1.0246875286102295, + "learning_rate": 1.1057463490252434e-05, + "loss": 0.5409, + "step": 23562 + }, + { + "epoch": 1.5813898862454279, + "grad_norm": 1.1175333261489868, + "learning_rate": 1.1050647707128592e-05, + "loss": 0.5502, + "step": 23564 + }, + { + "epoch": 1.581524109929197, + "grad_norm": 0.8779823184013367, + "learning_rate": 1.104383376428504e-05, + "loss": 0.4051, + "step": 23566 + }, + { + "epoch": 1.5816583336129661, + "grad_norm": 1.0429855585098267, + "learning_rate": 1.1037021662043756e-05, + "loss": 0.497, + "step": 23568 + }, + { + "epoch": 1.581792557296735, + "grad_norm": 1.1017266511917114, + "learning_rate": 1.1030211400726576e-05, + "loss": 0.4846, + "step": 23570 + }, + { + "epoch": 1.581926780980504, + "grad_norm": 1.1748697757720947, + "learning_rate": 1.1023402980655279e-05, + "loss": 0.4775, + "step": 23572 + }, + { + "epoch": 1.582061004664273, + "grad_norm": 1.116491675376892, + "learning_rate": 1.1016596402151536e-05, + "loss": 0.5414, + "step": 23574 + }, + { + "epoch": 1.582195228348042, + "grad_norm": 1.145172357559204, + "learning_rate": 1.1009791665536968e-05, + "loss": 0.5369, + "step": 23576 + }, + { + "epoch": 1.582329452031811, + "grad_norm": 1.0083330869674683, + "learning_rate": 1.1002988771133054e-05, + "loss": 0.4583, + "step": 23578 + }, + { + "epoch": 1.58246367571558, + "grad_norm": 0.9352924823760986, + "learning_rate": 1.0996187719261269e-05, + "loss": 0.4951, + "step": 23580 + }, + { + "epoch": 1.582597899399349, + "grad_norm": 1.0740264654159546, + "learning_rate": 1.0989388510242887e-05, + "loss": 0.4206, + "step": 23582 + }, + { + "epoch": 1.582732123083118, + "grad_norm": 1.1314347982406616, + "learning_rate": 1.0982591144399201e-05, + "loss": 0.451, + "step": 23584 + }, + { + "epoch": 1.5828663467668869, + "grad_norm": 1.1169883012771606, + "learning_rate": 1.097579562205135e-05, + "loss": 0.5466, + "step": 23586 + }, + { + "epoch": 1.583000570450656, + "grad_norm": 1.0791853666305542, + "learning_rate": 1.0969001943520424e-05, + "loss": 0.4515, + "step": 23588 + }, + { + "epoch": 1.5831347941344251, + "grad_norm": 1.1795690059661865, + "learning_rate": 1.0962210109127407e-05, + "loss": 0.4691, + "step": 23590 + }, + { + "epoch": 1.583269017818194, + "grad_norm": 1.0334551334381104, + "learning_rate": 1.0955420119193199e-05, + "loss": 0.4492, + "step": 23592 + }, + { + "epoch": 1.583403241501963, + "grad_norm": 0.8433769941329956, + "learning_rate": 1.0948631974038604e-05, + "loss": 0.3866, + "step": 23594 + }, + { + "epoch": 1.583537465185732, + "grad_norm": 1.1193410158157349, + "learning_rate": 1.0941845673984364e-05, + "loss": 0.4858, + "step": 23596 + }, + { + "epoch": 1.583671688869501, + "grad_norm": 0.9980464577674866, + "learning_rate": 1.0935061219351111e-05, + "loss": 0.4543, + "step": 23598 + }, + { + "epoch": 1.5838059125532702, + "grad_norm": 0.8986864686012268, + "learning_rate": 1.0928278610459398e-05, + "loss": 0.4081, + "step": 23600 + }, + { + "epoch": 1.583940136237039, + "grad_norm": 1.094754695892334, + "learning_rate": 1.092149784762968e-05, + "loss": 0.4822, + "step": 23602 + }, + { + "epoch": 1.584074359920808, + "grad_norm": 0.8928935527801514, + "learning_rate": 1.0914718931182355e-05, + "loss": 0.4363, + "step": 23604 + }, + { + "epoch": 1.584208583604577, + "grad_norm": 1.0659420490264893, + "learning_rate": 1.0907941861437688e-05, + "loss": 0.503, + "step": 23606 + }, + { + "epoch": 1.5843428072883459, + "grad_norm": 1.1274199485778809, + "learning_rate": 1.0901166638715926e-05, + "loss": 0.5279, + "step": 23608 + }, + { + "epoch": 1.584477030972115, + "grad_norm": 1.0495855808258057, + "learning_rate": 1.0894393263337127e-05, + "loss": 0.5288, + "step": 23610 + }, + { + "epoch": 1.5846112546558841, + "grad_norm": 1.0033987760543823, + "learning_rate": 1.0887621735621362e-05, + "loss": 0.4518, + "step": 23612 + }, + { + "epoch": 1.584745478339653, + "grad_norm": 0.9234064221382141, + "learning_rate": 1.0880852055888547e-05, + "loss": 0.4399, + "step": 23614 + }, + { + "epoch": 1.584879702023422, + "grad_norm": 0.9354954957962036, + "learning_rate": 1.0874084224458558e-05, + "loss": 0.4442, + "step": 23616 + }, + { + "epoch": 1.585013925707191, + "grad_norm": 0.978236973285675, + "learning_rate": 1.0867318241651154e-05, + "loss": 0.4884, + "step": 23618 + }, + { + "epoch": 1.58514814939096, + "grad_norm": 0.9640846848487854, + "learning_rate": 1.0860554107786015e-05, + "loss": 0.52, + "step": 23620 + }, + { + "epoch": 1.5852823730747292, + "grad_norm": 1.090498447418213, + "learning_rate": 1.0853791823182723e-05, + "loss": 0.4359, + "step": 23622 + }, + { + "epoch": 1.585416596758498, + "grad_norm": 1.0842543840408325, + "learning_rate": 1.0847031388160777e-05, + "loss": 0.4861, + "step": 23624 + }, + { + "epoch": 1.585550820442267, + "grad_norm": 0.9848273396492004, + "learning_rate": 1.0840272803039625e-05, + "loss": 0.4802, + "step": 23626 + }, + { + "epoch": 1.585685044126036, + "grad_norm": 0.8913401365280151, + "learning_rate": 1.0833516068138577e-05, + "loss": 0.3745, + "step": 23628 + }, + { + "epoch": 1.585819267809805, + "grad_norm": 0.9378396272659302, + "learning_rate": 1.0826761183776879e-05, + "loss": 0.4543, + "step": 23630 + }, + { + "epoch": 1.585953491493574, + "grad_norm": 1.1678546667099, + "learning_rate": 1.0820008150273669e-05, + "loss": 0.4911, + "step": 23632 + }, + { + "epoch": 1.5860877151773431, + "grad_norm": 1.0119303464889526, + "learning_rate": 1.0813256967948044e-05, + "loss": 0.4856, + "step": 23634 + }, + { + "epoch": 1.586221938861112, + "grad_norm": 1.1532948017120361, + "learning_rate": 1.0806507637118967e-05, + "loss": 0.482, + "step": 23636 + }, + { + "epoch": 1.586356162544881, + "grad_norm": 1.0109304189682007, + "learning_rate": 1.0799760158105337e-05, + "loss": 0.5013, + "step": 23638 + }, + { + "epoch": 1.58649038622865, + "grad_norm": 1.0066771507263184, + "learning_rate": 1.0793014531225937e-05, + "loss": 0.4438, + "step": 23640 + }, + { + "epoch": 1.586624609912419, + "grad_norm": 1.0120000839233398, + "learning_rate": 1.0786270756799522e-05, + "loss": 0.4615, + "step": 23642 + }, + { + "epoch": 1.5867588335961882, + "grad_norm": 1.0950095653533936, + "learning_rate": 1.0779528835144686e-05, + "loss": 0.4697, + "step": 23644 + }, + { + "epoch": 1.586893057279957, + "grad_norm": 0.864580512046814, + "learning_rate": 1.0772788766580022e-05, + "loss": 0.4016, + "step": 23646 + }, + { + "epoch": 1.587027280963726, + "grad_norm": 0.970075249671936, + "learning_rate": 1.0766050551423917e-05, + "loss": 0.5004, + "step": 23648 + }, + { + "epoch": 1.587161504647495, + "grad_norm": 1.0387160778045654, + "learning_rate": 1.075931418999479e-05, + "loss": 0.4854, + "step": 23650 + }, + { + "epoch": 1.587295728331264, + "grad_norm": 0.913539469242096, + "learning_rate": 1.0752579682610891e-05, + "loss": 0.4409, + "step": 23652 + }, + { + "epoch": 1.587429952015033, + "grad_norm": 1.9731390476226807, + "learning_rate": 1.0745847029590439e-05, + "loss": 0.5259, + "step": 23654 + }, + { + "epoch": 1.5875641756988021, + "grad_norm": 0.968388557434082, + "learning_rate": 1.0739116231251523e-05, + "loss": 0.4799, + "step": 23656 + }, + { + "epoch": 1.587698399382571, + "grad_norm": 0.9847589135169983, + "learning_rate": 1.0732387287912166e-05, + "loss": 0.4816, + "step": 23658 + }, + { + "epoch": 1.58783262306634, + "grad_norm": 1.0640642642974854, + "learning_rate": 1.0725660199890275e-05, + "loss": 0.4835, + "step": 23660 + }, + { + "epoch": 1.587966846750109, + "grad_norm": 0.9867693781852722, + "learning_rate": 1.0718934967503724e-05, + "loss": 0.4932, + "step": 23662 + }, + { + "epoch": 1.588101070433878, + "grad_norm": 1.11772882938385, + "learning_rate": 1.0712211591070254e-05, + "loss": 0.4645, + "step": 23664 + }, + { + "epoch": 1.5882352941176472, + "grad_norm": 1.0750466585159302, + "learning_rate": 1.070549007090753e-05, + "loss": 0.4911, + "step": 23666 + }, + { + "epoch": 1.588369517801416, + "grad_norm": 1.048567771911621, + "learning_rate": 1.069877040733312e-05, + "loss": 0.4705, + "step": 23668 + }, + { + "epoch": 1.588503741485185, + "grad_norm": 1.1166553497314453, + "learning_rate": 1.0692052600664537e-05, + "loss": 0.4695, + "step": 23670 + }, + { + "epoch": 1.588637965168954, + "grad_norm": 1.2060980796813965, + "learning_rate": 1.0685336651219158e-05, + "loss": 0.4708, + "step": 23672 + }, + { + "epoch": 1.588772188852723, + "grad_norm": 1.026139259338379, + "learning_rate": 1.0678622559314344e-05, + "loss": 0.4734, + "step": 23674 + }, + { + "epoch": 1.5889064125364922, + "grad_norm": 1.078305721282959, + "learning_rate": 1.067191032526726e-05, + "loss": 0.4849, + "step": 23676 + }, + { + "epoch": 1.5890406362202611, + "grad_norm": 0.9637233018875122, + "learning_rate": 1.0665199949395093e-05, + "loss": 0.4589, + "step": 23678 + }, + { + "epoch": 1.58917485990403, + "grad_norm": 1.070265531539917, + "learning_rate": 1.0658491432014867e-05, + "loss": 0.5019, + "step": 23680 + }, + { + "epoch": 1.589309083587799, + "grad_norm": 1.0106004476547241, + "learning_rate": 1.0651784773443573e-05, + "loss": 0.4511, + "step": 23682 + }, + { + "epoch": 1.589443307271568, + "grad_norm": 0.9935443997383118, + "learning_rate": 1.0645079973998073e-05, + "loss": 0.4821, + "step": 23684 + }, + { + "epoch": 1.589577530955337, + "grad_norm": 1.0848814249038696, + "learning_rate": 1.0638377033995156e-05, + "loss": 0.4612, + "step": 23686 + }, + { + "epoch": 1.5897117546391062, + "grad_norm": 1.0245535373687744, + "learning_rate": 1.0631675953751508e-05, + "loss": 0.5094, + "step": 23688 + }, + { + "epoch": 1.589845978322875, + "grad_norm": 1.1181143522262573, + "learning_rate": 1.0624976733583775e-05, + "loss": 0.4962, + "step": 23690 + }, + { + "epoch": 1.589980202006644, + "grad_norm": 1.008202314376831, + "learning_rate": 1.0618279373808459e-05, + "loss": 0.4608, + "step": 23692 + }, + { + "epoch": 1.590114425690413, + "grad_norm": 1.0486910343170166, + "learning_rate": 1.0611583874742004e-05, + "loss": 0.5284, + "step": 23694 + }, + { + "epoch": 1.590248649374182, + "grad_norm": 1.1446926593780518, + "learning_rate": 1.0604890236700753e-05, + "loss": 0.4721, + "step": 23696 + }, + { + "epoch": 1.5903828730579512, + "grad_norm": 1.0472633838653564, + "learning_rate": 1.0598198460000963e-05, + "loss": 0.4813, + "step": 23698 + }, + { + "epoch": 1.5905170967417201, + "grad_norm": 1.083215355873108, + "learning_rate": 1.0591508544958823e-05, + "loss": 0.4383, + "step": 23700 + }, + { + "epoch": 1.590651320425489, + "grad_norm": 1.0399837493896484, + "learning_rate": 1.0584820491890402e-05, + "loss": 0.4771, + "step": 23702 + }, + { + "epoch": 1.590785544109258, + "grad_norm": 1.2904225587844849, + "learning_rate": 1.0578134301111731e-05, + "loss": 0.472, + "step": 23704 + }, + { + "epoch": 1.5909197677930271, + "grad_norm": 1.0123722553253174, + "learning_rate": 1.0571449972938668e-05, + "loss": 0.4639, + "step": 23706 + }, + { + "epoch": 1.591053991476796, + "grad_norm": 1.1034024953842163, + "learning_rate": 1.0564767507687078e-05, + "loss": 0.5012, + "step": 23708 + }, + { + "epoch": 1.5911882151605652, + "grad_norm": 0.9454026222229004, + "learning_rate": 1.055808690567266e-05, + "loss": 0.4836, + "step": 23710 + }, + { + "epoch": 1.591322438844334, + "grad_norm": 0.9937888979911804, + "learning_rate": 1.055140816721109e-05, + "loss": 0.432, + "step": 23712 + }, + { + "epoch": 1.591456662528103, + "grad_norm": 1.0560818910598755, + "learning_rate": 1.054473129261791e-05, + "loss": 0.4543, + "step": 23714 + }, + { + "epoch": 1.591590886211872, + "grad_norm": 1.067166805267334, + "learning_rate": 1.0538056282208598e-05, + "loss": 0.4555, + "step": 23716 + }, + { + "epoch": 1.591725109895641, + "grad_norm": 1.1397252082824707, + "learning_rate": 1.0531383136298507e-05, + "loss": 0.5167, + "step": 23718 + }, + { + "epoch": 1.5918593335794102, + "grad_norm": 0.9467281699180603, + "learning_rate": 1.0524711855202967e-05, + "loss": 0.532, + "step": 23720 + }, + { + "epoch": 1.5919935572631791, + "grad_norm": 1.033605933189392, + "learning_rate": 1.051804243923717e-05, + "loss": 0.4741, + "step": 23722 + }, + { + "epoch": 1.592127780946948, + "grad_norm": 0.9673511385917664, + "learning_rate": 1.0511374888716224e-05, + "loss": 0.5159, + "step": 23724 + }, + { + "epoch": 1.592262004630717, + "grad_norm": 0.8267554044723511, + "learning_rate": 1.0504709203955155e-05, + "loss": 0.4169, + "step": 23726 + }, + { + "epoch": 1.5923962283144861, + "grad_norm": 1.0662051439285278, + "learning_rate": 1.0498045385268923e-05, + "loss": 0.4861, + "step": 23728 + }, + { + "epoch": 1.592530451998255, + "grad_norm": 1.031984806060791, + "learning_rate": 1.0491383432972357e-05, + "loss": 0.4379, + "step": 23730 + }, + { + "epoch": 1.5926646756820242, + "grad_norm": 1.0029823780059814, + "learning_rate": 1.048472334738026e-05, + "loss": 0.4919, + "step": 23732 + }, + { + "epoch": 1.592798899365793, + "grad_norm": 1.0263690948486328, + "learning_rate": 1.0478065128807251e-05, + "loss": 0.5601, + "step": 23734 + }, + { + "epoch": 1.592933123049562, + "grad_norm": 0.8970464468002319, + "learning_rate": 1.0471408777567965e-05, + "loss": 0.4435, + "step": 23736 + }, + { + "epoch": 1.593067346733331, + "grad_norm": 1.0362054109573364, + "learning_rate": 1.0464754293976875e-05, + "loss": 0.4712, + "step": 23738 + }, + { + "epoch": 1.5932015704171, + "grad_norm": 1.1117849349975586, + "learning_rate": 1.045810167834841e-05, + "loss": 0.5359, + "step": 23740 + }, + { + "epoch": 1.5933357941008692, + "grad_norm": 1.1014670133590698, + "learning_rate": 1.0451450930996887e-05, + "loss": 0.4716, + "step": 23742 + }, + { + "epoch": 1.5934700177846381, + "grad_norm": 0.8141437768936157, + "learning_rate": 1.0444802052236535e-05, + "loss": 0.4373, + "step": 23744 + }, + { + "epoch": 1.593604241468407, + "grad_norm": 0.9930576682090759, + "learning_rate": 1.0438155042381491e-05, + "loss": 0.4522, + "step": 23746 + }, + { + "epoch": 1.593738465152176, + "grad_norm": 0.8991205096244812, + "learning_rate": 1.0431509901745846e-05, + "loss": 0.461, + "step": 23748 + }, + { + "epoch": 1.5938726888359451, + "grad_norm": 1.0563803911209106, + "learning_rate": 1.0424866630643542e-05, + "loss": 0.4278, + "step": 23750 + }, + { + "epoch": 1.5940069125197143, + "grad_norm": 1.1721192598342896, + "learning_rate": 1.0418225229388479e-05, + "loss": 0.452, + "step": 23752 + }, + { + "epoch": 1.5941411362034832, + "grad_norm": 1.0518708229064941, + "learning_rate": 1.0411585698294418e-05, + "loss": 0.4441, + "step": 23754 + }, + { + "epoch": 1.594275359887252, + "grad_norm": 1.1242610216140747, + "learning_rate": 1.04049480376751e-05, + "loss": 0.4785, + "step": 23756 + }, + { + "epoch": 1.594409583571021, + "grad_norm": 1.1258399486541748, + "learning_rate": 1.0398312247844127e-05, + "loss": 0.5362, + "step": 23758 + }, + { + "epoch": 1.59454380725479, + "grad_norm": 1.0580230951309204, + "learning_rate": 1.0391678329115028e-05, + "loss": 0.4615, + "step": 23760 + }, + { + "epoch": 1.594678030938559, + "grad_norm": 1.0128093957901, + "learning_rate": 1.0385046281801242e-05, + "loss": 0.494, + "step": 23762 + }, + { + "epoch": 1.5948122546223282, + "grad_norm": 1.00709068775177, + "learning_rate": 1.0378416106216105e-05, + "loss": 0.5047, + "step": 23764 + }, + { + "epoch": 1.5949464783060971, + "grad_norm": 1.025793194770813, + "learning_rate": 1.0371787802672906e-05, + "loss": 0.4645, + "step": 23766 + }, + { + "epoch": 1.595080701989866, + "grad_norm": 1.0294287204742432, + "learning_rate": 1.03651613714848e-05, + "loss": 0.5148, + "step": 23768 + }, + { + "epoch": 1.595214925673635, + "grad_norm": 1.018163800239563, + "learning_rate": 1.0358536812964904e-05, + "loss": 0.457, + "step": 23770 + }, + { + "epoch": 1.5953491493574041, + "grad_norm": 1.067983627319336, + "learning_rate": 1.0351914127426166e-05, + "loss": 0.4174, + "step": 23772 + }, + { + "epoch": 1.5954833730411733, + "grad_norm": 1.0045945644378662, + "learning_rate": 1.034529331518153e-05, + "loss": 0.4896, + "step": 23774 + }, + { + "epoch": 1.5956175967249422, + "grad_norm": 1.0037025213241577, + "learning_rate": 1.03386743765438e-05, + "loss": 0.4742, + "step": 23776 + }, + { + "epoch": 1.595751820408711, + "grad_norm": 1.0527470111846924, + "learning_rate": 1.0332057311825726e-05, + "loss": 0.4628, + "step": 23778 + }, + { + "epoch": 1.59588604409248, + "grad_norm": 1.3437060117721558, + "learning_rate": 1.032544212133994e-05, + "loss": 0.4746, + "step": 23780 + }, + { + "epoch": 1.5960202677762492, + "grad_norm": 1.1082535982131958, + "learning_rate": 1.0318828805398994e-05, + "loss": 0.5561, + "step": 23782 + }, + { + "epoch": 1.596154491460018, + "grad_norm": 1.046541690826416, + "learning_rate": 1.0312217364315351e-05, + "loss": 0.5194, + "step": 23784 + }, + { + "epoch": 1.5962887151437872, + "grad_norm": 1.1011675596237183, + "learning_rate": 1.03056077984014e-05, + "loss": 0.5635, + "step": 23786 + }, + { + "epoch": 1.5964229388275561, + "grad_norm": 1.070259928703308, + "learning_rate": 1.029900010796943e-05, + "loss": 0.4599, + "step": 23788 + }, + { + "epoch": 1.596557162511325, + "grad_norm": 1.0323693752288818, + "learning_rate": 1.0292394293331637e-05, + "loss": 0.4672, + "step": 23790 + }, + { + "epoch": 1.596691386195094, + "grad_norm": 0.905518651008606, + "learning_rate": 1.0285790354800118e-05, + "loss": 0.4294, + "step": 23792 + }, + { + "epoch": 1.5968256098788631, + "grad_norm": 1.046562671661377, + "learning_rate": 1.027918829268692e-05, + "loss": 0.4776, + "step": 23794 + }, + { + "epoch": 1.5969598335626323, + "grad_norm": 0.9571558833122253, + "learning_rate": 1.027258810730396e-05, + "loss": 0.4368, + "step": 23796 + }, + { + "epoch": 1.5970940572464012, + "grad_norm": 1.0320895910263062, + "learning_rate": 1.0265989798963116e-05, + "loss": 0.4882, + "step": 23798 + }, + { + "epoch": 1.59722828093017, + "grad_norm": 0.9254117012023926, + "learning_rate": 1.025939336797609e-05, + "loss": 0.4507, + "step": 23800 + }, + { + "epoch": 1.597362504613939, + "grad_norm": 1.1301331520080566, + "learning_rate": 1.0252798814654597e-05, + "loss": 0.4758, + "step": 23802 + }, + { + "epoch": 1.5974967282977082, + "grad_norm": 1.0007113218307495, + "learning_rate": 1.0246206139310187e-05, + "loss": 0.4802, + "step": 23804 + }, + { + "epoch": 1.597630951981477, + "grad_norm": 1.1954693794250488, + "learning_rate": 1.0239615342254378e-05, + "loss": 0.4899, + "step": 23806 + }, + { + "epoch": 1.5977651756652462, + "grad_norm": 0.9859517812728882, + "learning_rate": 1.0233026423798559e-05, + "loss": 0.5094, + "step": 23808 + }, + { + "epoch": 1.5978993993490151, + "grad_norm": 1.156038522720337, + "learning_rate": 1.0226439384254044e-05, + "loss": 0.5374, + "step": 23810 + }, + { + "epoch": 1.598033623032784, + "grad_norm": 1.1276850700378418, + "learning_rate": 1.0219854223932041e-05, + "loss": 0.5342, + "step": 23812 + }, + { + "epoch": 1.598167846716553, + "grad_norm": 1.047963261604309, + "learning_rate": 1.0213270943143716e-05, + "loss": 0.4327, + "step": 23814 + }, + { + "epoch": 1.5983020704003221, + "grad_norm": 1.077088475227356, + "learning_rate": 1.02066895422001e-05, + "loss": 0.453, + "step": 23816 + }, + { + "epoch": 1.5984362940840913, + "grad_norm": 1.0303544998168945, + "learning_rate": 1.020011002141215e-05, + "loss": 0.5176, + "step": 23818 + }, + { + "epoch": 1.5985705177678602, + "grad_norm": 0.9186301827430725, + "learning_rate": 1.0193532381090732e-05, + "loss": 0.4509, + "step": 23820 + }, + { + "epoch": 1.598704741451629, + "grad_norm": 1.0409269332885742, + "learning_rate": 1.0186956621546645e-05, + "loss": 0.5433, + "step": 23822 + }, + { + "epoch": 1.598838965135398, + "grad_norm": 1.086869239807129, + "learning_rate": 1.0180382743090555e-05, + "loss": 0.4621, + "step": 23824 + }, + { + "epoch": 1.5989731888191672, + "grad_norm": 0.9349266886711121, + "learning_rate": 1.0173810746033103e-05, + "loss": 0.4674, + "step": 23826 + }, + { + "epoch": 1.5991074125029363, + "grad_norm": 0.9937050938606262, + "learning_rate": 1.0167240630684765e-05, + "loss": 0.4908, + "step": 23828 + }, + { + "epoch": 1.5992416361867052, + "grad_norm": 0.9550259709358215, + "learning_rate": 1.0160672397355969e-05, + "loss": 0.4843, + "step": 23830 + }, + { + "epoch": 1.5993758598704741, + "grad_norm": 1.048781156539917, + "learning_rate": 1.0154106046357071e-05, + "loss": 0.5059, + "step": 23832 + }, + { + "epoch": 1.599510083554243, + "grad_norm": 0.9172594547271729, + "learning_rate": 1.0147541577998298e-05, + "loss": 0.4637, + "step": 23834 + }, + { + "epoch": 1.599644307238012, + "grad_norm": 1.0190455913543701, + "learning_rate": 1.0140978992589833e-05, + "loss": 0.4182, + "step": 23836 + }, + { + "epoch": 1.5997785309217811, + "grad_norm": 0.9304594993591309, + "learning_rate": 1.0134418290441728e-05, + "loss": 0.4749, + "step": 23838 + }, + { + "epoch": 1.5999127546055503, + "grad_norm": 0.9567098021507263, + "learning_rate": 1.012785947186397e-05, + "loss": 0.4793, + "step": 23840 + }, + { + "epoch": 1.6000469782893192, + "grad_norm": 1.0866206884384155, + "learning_rate": 1.0121302537166433e-05, + "loss": 0.4727, + "step": 23842 + }, + { + "epoch": 1.600181201973088, + "grad_norm": 1.1256355047225952, + "learning_rate": 1.0114747486658943e-05, + "loss": 0.5181, + "step": 23844 + }, + { + "epoch": 1.600315425656857, + "grad_norm": 1.1933351755142212, + "learning_rate": 1.0108194320651205e-05, + "loss": 0.5037, + "step": 23846 + }, + { + "epoch": 1.6004496493406262, + "grad_norm": 1.0349985361099243, + "learning_rate": 1.010164303945284e-05, + "loss": 0.5087, + "step": 23848 + }, + { + "epoch": 1.6005838730243953, + "grad_norm": 0.9755839109420776, + "learning_rate": 1.0095093643373377e-05, + "loss": 0.5202, + "step": 23850 + }, + { + "epoch": 1.6007180967081642, + "grad_norm": 1.032321810722351, + "learning_rate": 1.0088546132722276e-05, + "loss": 0.5722, + "step": 23852 + }, + { + "epoch": 1.6008523203919331, + "grad_norm": 0.9793102145195007, + "learning_rate": 1.0082000507808892e-05, + "loss": 0.5009, + "step": 23854 + }, + { + "epoch": 1.600986544075702, + "grad_norm": 1.093875765800476, + "learning_rate": 1.0075456768942488e-05, + "loss": 0.5367, + "step": 23856 + }, + { + "epoch": 1.6011207677594712, + "grad_norm": 1.1801968812942505, + "learning_rate": 1.0068914916432231e-05, + "loss": 0.5578, + "step": 23858 + }, + { + "epoch": 1.6012549914432401, + "grad_norm": 0.9556602835655212, + "learning_rate": 1.0062374950587234e-05, + "loss": 0.4843, + "step": 23860 + }, + { + "epoch": 1.6013892151270093, + "grad_norm": 1.0572758913040161, + "learning_rate": 1.0055836871716473e-05, + "loss": 0.4997, + "step": 23862 + }, + { + "epoch": 1.6015234388107782, + "grad_norm": 1.0739139318466187, + "learning_rate": 1.0049300680128887e-05, + "loss": 0.5168, + "step": 23864 + }, + { + "epoch": 1.601657662494547, + "grad_norm": 1.0001834630966187, + "learning_rate": 1.004276637613329e-05, + "loss": 0.4611, + "step": 23866 + }, + { + "epoch": 1.601791886178316, + "grad_norm": 0.9812644720077515, + "learning_rate": 1.0036233960038399e-05, + "loss": 0.5198, + "step": 23868 + }, + { + "epoch": 1.6019261098620852, + "grad_norm": 1.0067847967147827, + "learning_rate": 1.0029703432152859e-05, + "loss": 0.4752, + "step": 23870 + }, + { + "epoch": 1.6020603335458543, + "grad_norm": 1.0018914937973022, + "learning_rate": 1.002317479278525e-05, + "loss": 0.4786, + "step": 23872 + }, + { + "epoch": 1.6021945572296232, + "grad_norm": 1.049195408821106, + "learning_rate": 1.001664804224402e-05, + "loss": 0.4891, + "step": 23874 + }, + { + "epoch": 1.6023287809133921, + "grad_norm": 1.0486708879470825, + "learning_rate": 1.0010123180837544e-05, + "loss": 0.4587, + "step": 23876 + }, + { + "epoch": 1.602463004597161, + "grad_norm": 0.9982337355613708, + "learning_rate": 1.0003600208874098e-05, + "loss": 0.466, + "step": 23878 + }, + { + "epoch": 1.6025972282809302, + "grad_norm": 1.1180343627929688, + "learning_rate": 9.997079126661907e-06, + "loss": 0.4489, + "step": 23880 + }, + { + "epoch": 1.6027314519646991, + "grad_norm": 1.0404959917068481, + "learning_rate": 9.990559934509053e-06, + "loss": 0.4944, + "step": 23882 + }, + { + "epoch": 1.6028656756484683, + "grad_norm": 1.0978857278823853, + "learning_rate": 9.984042632723589e-06, + "loss": 0.532, + "step": 23884 + }, + { + "epoch": 1.6029998993322372, + "grad_norm": 1.035715103149414, + "learning_rate": 9.977527221613397e-06, + "loss": 0.5087, + "step": 23886 + }, + { + "epoch": 1.603134123016006, + "grad_norm": 1.1274858713150024, + "learning_rate": 9.97101370148636e-06, + "loss": 0.5453, + "step": 23888 + }, + { + "epoch": 1.603268346699775, + "grad_norm": 1.0599273443222046, + "learning_rate": 9.964502072650195e-06, + "loss": 0.4714, + "step": 23890 + }, + { + "epoch": 1.6034025703835442, + "grad_norm": 1.046848177909851, + "learning_rate": 9.957992335412597e-06, + "loss": 0.4718, + "step": 23892 + }, + { + "epoch": 1.6035367940673133, + "grad_norm": 1.0222798585891724, + "learning_rate": 9.95148449008112e-06, + "loss": 0.4411, + "step": 23894 + }, + { + "epoch": 1.6036710177510822, + "grad_norm": 1.1524306535720825, + "learning_rate": 9.944978536963246e-06, + "loss": 0.4696, + "step": 23896 + }, + { + "epoch": 1.6038052414348511, + "grad_norm": 0.9545031189918518, + "learning_rate": 9.938474476366378e-06, + "loss": 0.4327, + "step": 23898 + }, + { + "epoch": 1.60393946511862, + "grad_norm": 0.9605699777603149, + "learning_rate": 9.931972308597792e-06, + "loss": 0.4458, + "step": 23900 + }, + { + "epoch": 1.6040736888023892, + "grad_norm": 0.9968399405479431, + "learning_rate": 9.925472033964744e-06, + "loss": 0.4516, + "step": 23902 + }, + { + "epoch": 1.6042079124861581, + "grad_norm": 1.0261261463165283, + "learning_rate": 9.918973652774339e-06, + "loss": 0.5041, + "step": 23904 + }, + { + "epoch": 1.6043421361699273, + "grad_norm": 1.03523850440979, + "learning_rate": 9.912477165333612e-06, + "loss": 0.4782, + "step": 23906 + }, + { + "epoch": 1.6044763598536962, + "grad_norm": 1.0323772430419922, + "learning_rate": 9.905982571949491e-06, + "loss": 0.4811, + "step": 23908 + }, + { + "epoch": 1.604610583537465, + "grad_norm": 0.9295010566711426, + "learning_rate": 9.899489872928874e-06, + "loss": 0.4352, + "step": 23910 + }, + { + "epoch": 1.604744807221234, + "grad_norm": 1.0068014860153198, + "learning_rate": 9.892999068578502e-06, + "loss": 0.4196, + "step": 23912 + }, + { + "epoch": 1.6048790309050032, + "grad_norm": 0.9357097744941711, + "learning_rate": 9.886510159205053e-06, + "loss": 0.4026, + "step": 23914 + }, + { + "epoch": 1.6050132545887723, + "grad_norm": 0.9535422921180725, + "learning_rate": 9.880023145115114e-06, + "loss": 0.4632, + "step": 23916 + }, + { + "epoch": 1.6051474782725412, + "grad_norm": 1.1852741241455078, + "learning_rate": 9.873538026615198e-06, + "loss": 0.4818, + "step": 23918 + }, + { + "epoch": 1.6052817019563101, + "grad_norm": 1.123728632926941, + "learning_rate": 9.867054804011693e-06, + "loss": 0.4643, + "step": 23920 + }, + { + "epoch": 1.605415925640079, + "grad_norm": 0.9284471869468689, + "learning_rate": 9.860573477610952e-06, + "loss": 0.484, + "step": 23922 + }, + { + "epoch": 1.6055501493238482, + "grad_norm": 1.1171112060546875, + "learning_rate": 9.854094047719164e-06, + "loss": 0.457, + "step": 23924 + }, + { + "epoch": 1.6056843730076173, + "grad_norm": 1.0885698795318604, + "learning_rate": 9.8476165146425e-06, + "loss": 0.4984, + "step": 23926 + }, + { + "epoch": 1.6058185966913863, + "grad_norm": 1.0228629112243652, + "learning_rate": 9.841140878686983e-06, + "loss": 0.4652, + "step": 23928 + }, + { + "epoch": 1.6059528203751552, + "grad_norm": 0.9912495017051697, + "learning_rate": 9.834667140158605e-06, + "loss": 0.5383, + "step": 23930 + }, + { + "epoch": 1.606087044058924, + "grad_norm": 1.0985013246536255, + "learning_rate": 9.828195299363224e-06, + "loss": 0.4345, + "step": 23932 + }, + { + "epoch": 1.6062212677426932, + "grad_norm": 1.0537720918655396, + "learning_rate": 9.82172535660662e-06, + "loss": 0.4558, + "step": 23934 + }, + { + "epoch": 1.6063554914264622, + "grad_norm": 0.8829530477523804, + "learning_rate": 9.81525731219447e-06, + "loss": 0.4826, + "step": 23936 + }, + { + "epoch": 1.6064897151102313, + "grad_norm": 1.0897839069366455, + "learning_rate": 9.808791166432413e-06, + "loss": 0.4997, + "step": 23938 + }, + { + "epoch": 1.6066239387940002, + "grad_norm": 1.4052565097808838, + "learning_rate": 9.802326919625931e-06, + "loss": 0.4729, + "step": 23940 + }, + { + "epoch": 1.6067581624777691, + "grad_norm": 1.0854122638702393, + "learning_rate": 9.795864572080466e-06, + "loss": 0.4899, + "step": 23942 + }, + { + "epoch": 1.606892386161538, + "grad_norm": 0.8918828368186951, + "learning_rate": 9.78940412410132e-06, + "loss": 0.4255, + "step": 23944 + }, + { + "epoch": 1.6070266098453072, + "grad_norm": 1.6517807245254517, + "learning_rate": 9.782945575993779e-06, + "loss": 0.421, + "step": 23946 + }, + { + "epoch": 1.6071608335290763, + "grad_norm": 0.9856716990470886, + "learning_rate": 9.77648892806296e-06, + "loss": 0.4596, + "step": 23948 + }, + { + "epoch": 1.6072950572128453, + "grad_norm": 1.058132290840149, + "learning_rate": 9.770034180613968e-06, + "loss": 0.514, + "step": 23950 + }, + { + "epoch": 1.6074292808966142, + "grad_norm": 0.9935703277587891, + "learning_rate": 9.763581333951727e-06, + "loss": 0.4523, + "step": 23952 + }, + { + "epoch": 1.607563504580383, + "grad_norm": 0.9994638562202454, + "learning_rate": 9.757130388381158e-06, + "loss": 0.4745, + "step": 23954 + }, + { + "epoch": 1.6076977282641522, + "grad_norm": 1.0030862092971802, + "learning_rate": 9.750681344207035e-06, + "loss": 0.4799, + "step": 23956 + }, + { + "epoch": 1.6078319519479212, + "grad_norm": 1.0018229484558105, + "learning_rate": 9.744234201734082e-06, + "loss": 0.4709, + "step": 23958 + }, + { + "epoch": 1.6079661756316903, + "grad_norm": 0.9601824879646301, + "learning_rate": 9.737788961266903e-06, + "loss": 0.4588, + "step": 23960 + }, + { + "epoch": 1.6081003993154592, + "grad_norm": 1.0627379417419434, + "learning_rate": 9.731345623110022e-06, + "loss": 0.4816, + "step": 23962 + }, + { + "epoch": 1.6082346229992281, + "grad_norm": 1.1042258739471436, + "learning_rate": 9.724904187567879e-06, + "loss": 0.533, + "step": 23964 + }, + { + "epoch": 1.608368846682997, + "grad_norm": 1.1249784231185913, + "learning_rate": 9.718464654944798e-06, + "loss": 0.4703, + "step": 23966 + }, + { + "epoch": 1.6085030703667662, + "grad_norm": 0.8959927558898926, + "learning_rate": 9.712027025545067e-06, + "loss": 0.4851, + "step": 23968 + }, + { + "epoch": 1.6086372940505353, + "grad_norm": 1.1174625158309937, + "learning_rate": 9.70559129967284e-06, + "loss": 0.4609, + "step": 23970 + }, + { + "epoch": 1.6087715177343043, + "grad_norm": 1.1044148206710815, + "learning_rate": 9.699157477632181e-06, + "loss": 0.455, + "step": 23972 + }, + { + "epoch": 1.6089057414180732, + "grad_norm": 1.0346393585205078, + "learning_rate": 9.692725559727072e-06, + "loss": 0.5299, + "step": 23974 + }, + { + "epoch": 1.609039965101842, + "grad_norm": 1.0218905210494995, + "learning_rate": 9.686295546261436e-06, + "loss": 0.4798, + "step": 23976 + }, + { + "epoch": 1.6091741887856112, + "grad_norm": 0.9859256148338318, + "learning_rate": 9.679867437539063e-06, + "loss": 0.4614, + "step": 23978 + }, + { + "epoch": 1.6093084124693802, + "grad_norm": 1.0233776569366455, + "learning_rate": 9.673441233863662e-06, + "loss": 0.4599, + "step": 23980 + }, + { + "epoch": 1.6094426361531493, + "grad_norm": 1.170893907546997, + "learning_rate": 9.667016935538859e-06, + "loss": 0.4715, + "step": 23982 + }, + { + "epoch": 1.6095768598369182, + "grad_norm": 1.0487552881240845, + "learning_rate": 9.6605945428682e-06, + "loss": 0.4594, + "step": 23984 + }, + { + "epoch": 1.6097110835206871, + "grad_norm": 1.0074599981307983, + "learning_rate": 9.654174056155113e-06, + "loss": 0.4731, + "step": 23986 + }, + { + "epoch": 1.609845307204456, + "grad_norm": 0.9991166591644287, + "learning_rate": 9.64775547570298e-06, + "loss": 0.4205, + "step": 23988 + }, + { + "epoch": 1.6099795308882252, + "grad_norm": 0.9638586640357971, + "learning_rate": 9.641338801815048e-06, + "loss": 0.463, + "step": 23990 + }, + { + "epoch": 1.6101137545719943, + "grad_norm": 1.2994322776794434, + "learning_rate": 9.634924034794501e-06, + "loss": 0.4747, + "step": 23992 + }, + { + "epoch": 1.6102479782557633, + "grad_norm": 1.1774928569793701, + "learning_rate": 9.628511174944404e-06, + "loss": 0.5106, + "step": 23994 + }, + { + "epoch": 1.6103822019395322, + "grad_norm": 1.0639322996139526, + "learning_rate": 9.622100222567775e-06, + "loss": 0.4535, + "step": 23996 + }, + { + "epoch": 1.610516425623301, + "grad_norm": 1.0801434516906738, + "learning_rate": 9.615691177967518e-06, + "loss": 0.4782, + "step": 23998 + }, + { + "epoch": 1.6106506493070702, + "grad_norm": 0.9594919085502625, + "learning_rate": 9.609284041446438e-06, + "loss": 0.5241, + "step": 24000 + }, + { + "epoch": 1.6107848729908394, + "grad_norm": 2.1199254989624023, + "learning_rate": 9.602878813307249e-06, + "loss": 0.4488, + "step": 24002 + }, + { + "epoch": 1.6109190966746083, + "grad_norm": 0.9908333420753479, + "learning_rate": 9.596475493852608e-06, + "loss": 0.481, + "step": 24004 + }, + { + "epoch": 1.6110533203583772, + "grad_norm": 1.0409706830978394, + "learning_rate": 9.590074083385053e-06, + "loss": 0.457, + "step": 24006 + }, + { + "epoch": 1.6111875440421461, + "grad_norm": 1.0337355136871338, + "learning_rate": 9.583674582207036e-06, + "loss": 0.4516, + "step": 24008 + }, + { + "epoch": 1.6113217677259153, + "grad_norm": 0.9755725860595703, + "learning_rate": 9.577276990620903e-06, + "loss": 0.4258, + "step": 24010 + }, + { + "epoch": 1.6114559914096842, + "grad_norm": 1.0953158140182495, + "learning_rate": 9.570881308928958e-06, + "loss": 0.504, + "step": 24012 + }, + { + "epoch": 1.6115902150934533, + "grad_norm": 1.0453033447265625, + "learning_rate": 9.564487537433365e-06, + "loss": 0.4813, + "step": 24014 + }, + { + "epoch": 1.6117244387772223, + "grad_norm": 0.9101317524909973, + "learning_rate": 9.55809567643623e-06, + "loss": 0.4578, + "step": 24016 + }, + { + "epoch": 1.6118586624609912, + "grad_norm": 1.0483723878860474, + "learning_rate": 9.551705726239546e-06, + "loss": 0.4429, + "step": 24018 + }, + { + "epoch": 1.61199288614476, + "grad_norm": 1.000253438949585, + "learning_rate": 9.545317687145232e-06, + "loss": 0.4399, + "step": 24020 + }, + { + "epoch": 1.6121271098285292, + "grad_norm": 0.8763706088066101, + "learning_rate": 9.538931559455095e-06, + "loss": 0.4658, + "step": 24022 + }, + { + "epoch": 1.6122613335122984, + "grad_norm": 0.9557889103889465, + "learning_rate": 9.532547343470889e-06, + "loss": 0.4444, + "step": 24024 + }, + { + "epoch": 1.6123955571960673, + "grad_norm": 1.0888750553131104, + "learning_rate": 9.526165039494244e-06, + "loss": 0.4544, + "step": 24026 + }, + { + "epoch": 1.6125297808798362, + "grad_norm": 1.0589221715927124, + "learning_rate": 9.519784647826713e-06, + "loss": 0.4658, + "step": 24028 + }, + { + "epoch": 1.6126640045636051, + "grad_norm": 1.1489931344985962, + "learning_rate": 9.513406168769745e-06, + "loss": 0.4601, + "step": 24030 + }, + { + "epoch": 1.6127982282473743, + "grad_norm": 0.9511253833770752, + "learning_rate": 9.507029602624734e-06, + "loss": 0.4467, + "step": 24032 + }, + { + "epoch": 1.6129324519311432, + "grad_norm": 1.063330888748169, + "learning_rate": 9.50065494969295e-06, + "loss": 0.4993, + "step": 24034 + }, + { + "epoch": 1.6130666756149123, + "grad_norm": 0.9531103372573853, + "learning_rate": 9.494282210275579e-06, + "loss": 0.5202, + "step": 24036 + }, + { + "epoch": 1.6132008992986813, + "grad_norm": 1.212205410003662, + "learning_rate": 9.48791138467372e-06, + "loss": 0.5128, + "step": 24038 + }, + { + "epoch": 1.6133351229824502, + "grad_norm": 1.049813985824585, + "learning_rate": 9.481542473188377e-06, + "loss": 0.5109, + "step": 24040 + }, + { + "epoch": 1.613469346666219, + "grad_norm": 1.0941503047943115, + "learning_rate": 9.475175476120484e-06, + "loss": 0.4704, + "step": 24042 + }, + { + "epoch": 1.6136035703499882, + "grad_norm": 0.9977181553840637, + "learning_rate": 9.468810393770856e-06, + "loss": 0.4586, + "step": 24044 + }, + { + "epoch": 1.6137377940337574, + "grad_norm": 1.051984190940857, + "learning_rate": 9.462447226440252e-06, + "loss": 0.4489, + "step": 24046 + }, + { + "epoch": 1.6138720177175263, + "grad_norm": 1.109163761138916, + "learning_rate": 9.456085974429286e-06, + "loss": 0.5115, + "step": 24048 + }, + { + "epoch": 1.6140062414012952, + "grad_norm": 0.9957579374313354, + "learning_rate": 9.449726638038536e-06, + "loss": 0.4676, + "step": 24050 + }, + { + "epoch": 1.6141404650850641, + "grad_norm": 1.0215259790420532, + "learning_rate": 9.443369217568455e-06, + "loss": 0.4786, + "step": 24052 + }, + { + "epoch": 1.6142746887688333, + "grad_norm": 0.995062530040741, + "learning_rate": 9.437013713319437e-06, + "loss": 0.4429, + "step": 24054 + }, + { + "epoch": 1.6144089124526022, + "grad_norm": 0.9971685409545898, + "learning_rate": 9.430660125591762e-06, + "loss": 0.4857, + "step": 24056 + }, + { + "epoch": 1.6145431361363713, + "grad_norm": 0.99956876039505, + "learning_rate": 9.424308454685622e-06, + "loss": 0.4322, + "step": 24058 + }, + { + "epoch": 1.6146773598201403, + "grad_norm": 1.0475598573684692, + "learning_rate": 9.417958700901097e-06, + "loss": 0.4886, + "step": 24060 + }, + { + "epoch": 1.6148115835039092, + "grad_norm": 1.055567979812622, + "learning_rate": 9.411610864538239e-06, + "loss": 0.4814, + "step": 24062 + }, + { + "epoch": 1.614945807187678, + "grad_norm": 1.0338261127471924, + "learning_rate": 9.405264945896959e-06, + "loss": 0.4787, + "step": 24064 + }, + { + "epoch": 1.6150800308714472, + "grad_norm": 1.0155824422836304, + "learning_rate": 9.398920945277079e-06, + "loss": 0.4907, + "step": 24066 + }, + { + "epoch": 1.6152142545552164, + "grad_norm": 1.0634373426437378, + "learning_rate": 9.392578862978334e-06, + "loss": 0.4434, + "step": 24068 + }, + { + "epoch": 1.6153484782389853, + "grad_norm": 1.0663130283355713, + "learning_rate": 9.386238699300398e-06, + "loss": 0.4752, + "step": 24070 + }, + { + "epoch": 1.6154827019227542, + "grad_norm": 1.0315016508102417, + "learning_rate": 9.37990045454281e-06, + "loss": 0.4798, + "step": 24072 + }, + { + "epoch": 1.6156169256065231, + "grad_norm": 0.9357467293739319, + "learning_rate": 9.373564129005074e-06, + "loss": 0.5043, + "step": 24074 + }, + { + "epoch": 1.6157511492902923, + "grad_norm": 0.9481315016746521, + "learning_rate": 9.36722972298652e-06, + "loss": 0.526, + "step": 24076 + }, + { + "epoch": 1.6158853729740614, + "grad_norm": 0.9957374930381775, + "learning_rate": 9.36089723678647e-06, + "loss": 0.4725, + "step": 24078 + }, + { + "epoch": 1.6160195966578303, + "grad_norm": 1.7620972394943237, + "learning_rate": 9.354566670704102e-06, + "loss": 0.4567, + "step": 24080 + }, + { + "epoch": 1.6161538203415993, + "grad_norm": 1.176615834236145, + "learning_rate": 9.348238025038547e-06, + "loss": 0.4465, + "step": 24082 + }, + { + "epoch": 1.6162880440253682, + "grad_norm": 0.9767720699310303, + "learning_rate": 9.341911300088807e-06, + "loss": 0.3975, + "step": 24084 + }, + { + "epoch": 1.6164222677091373, + "grad_norm": 0.9831427335739136, + "learning_rate": 9.335586496153804e-06, + "loss": 0.4522, + "step": 24086 + }, + { + "epoch": 1.6165564913929062, + "grad_norm": 1.091477394104004, + "learning_rate": 9.329263613532363e-06, + "loss": 0.4849, + "step": 24088 + }, + { + "epoch": 1.6166907150766754, + "grad_norm": 1.0562587976455688, + "learning_rate": 9.322942652523259e-06, + "loss": 0.4562, + "step": 24090 + }, + { + "epoch": 1.6168249387604443, + "grad_norm": 0.8952846527099609, + "learning_rate": 9.316623613425119e-06, + "loss": 0.4132, + "step": 24092 + }, + { + "epoch": 1.6169591624442132, + "grad_norm": 1.0176690816879272, + "learning_rate": 9.310306496536519e-06, + "loss": 0.4767, + "step": 24094 + }, + { + "epoch": 1.6170933861279821, + "grad_norm": 1.0066777467727661, + "learning_rate": 9.303991302155907e-06, + "loss": 0.4267, + "step": 24096 + }, + { + "epoch": 1.6172276098117513, + "grad_norm": 0.9026097655296326, + "learning_rate": 9.297678030581697e-06, + "loss": 0.4117, + "step": 24098 + }, + { + "epoch": 1.6173618334955204, + "grad_norm": 0.9847307801246643, + "learning_rate": 9.291366682112152e-06, + "loss": 0.4353, + "step": 24100 + }, + { + "epoch": 1.6174960571792893, + "grad_norm": 0.9909016489982605, + "learning_rate": 9.285057257045498e-06, + "loss": 0.4191, + "step": 24102 + }, + { + "epoch": 1.6176302808630583, + "grad_norm": 1.0340805053710938, + "learning_rate": 9.278749755679823e-06, + "loss": 0.4293, + "step": 24104 + }, + { + "epoch": 1.6177645045468272, + "grad_norm": 1.0646992921829224, + "learning_rate": 9.272444178313127e-06, + "loss": 0.4774, + "step": 24106 + }, + { + "epoch": 1.6178987282305963, + "grad_norm": 0.9061216115951538, + "learning_rate": 9.266140525243373e-06, + "loss": 0.3749, + "step": 24108 + }, + { + "epoch": 1.6180329519143652, + "grad_norm": 1.1270420551300049, + "learning_rate": 9.259838796768367e-06, + "loss": 0.4538, + "step": 24110 + }, + { + "epoch": 1.6181671755981344, + "grad_norm": 0.9613937735557556, + "learning_rate": 9.25353899318589e-06, + "loss": 0.4073, + "step": 24112 + }, + { + "epoch": 1.6183013992819033, + "grad_norm": 1.1744015216827393, + "learning_rate": 9.247241114793543e-06, + "loss": 0.523, + "step": 24114 + }, + { + "epoch": 1.6184356229656722, + "grad_norm": 0.8820990324020386, + "learning_rate": 9.240945161888931e-06, + "loss": 0.4776, + "step": 24116 + }, + { + "epoch": 1.6185698466494411, + "grad_norm": 0.978901743888855, + "learning_rate": 9.2346511347695e-06, + "loss": 0.4847, + "step": 24118 + }, + { + "epoch": 1.6187040703332103, + "grad_norm": 1.159218192100525, + "learning_rate": 9.228359033732653e-06, + "loss": 0.4507, + "step": 24120 + }, + { + "epoch": 1.6188382940169794, + "grad_norm": 1.0147674083709717, + "learning_rate": 9.222068859075667e-06, + "loss": 0.4269, + "step": 24122 + }, + { + "epoch": 1.6189725177007483, + "grad_norm": 1.0214875936508179, + "learning_rate": 9.215780611095743e-06, + "loss": 0.4861, + "step": 24124 + }, + { + "epoch": 1.6191067413845173, + "grad_norm": 8.175945281982422, + "learning_rate": 9.209494290089971e-06, + "loss": 0.5368, + "step": 24126 + }, + { + "epoch": 1.6192409650682862, + "grad_norm": 1.0697429180145264, + "learning_rate": 9.203209896355398e-06, + "loss": 0.5083, + "step": 24128 + }, + { + "epoch": 1.6193751887520553, + "grad_norm": 1.0672308206558228, + "learning_rate": 9.196927430188929e-06, + "loss": 0.5229, + "step": 24130 + }, + { + "epoch": 1.6195094124358242, + "grad_norm": 0.8495045304298401, + "learning_rate": 9.190646891887405e-06, + "loss": 0.4541, + "step": 24132 + }, + { + "epoch": 1.6196436361195934, + "grad_norm": 1.102070689201355, + "learning_rate": 9.184368281747557e-06, + "loss": 0.5542, + "step": 24134 + }, + { + "epoch": 1.6197778598033623, + "grad_norm": 1.296924114227295, + "learning_rate": 9.178091600066063e-06, + "loss": 0.5097, + "step": 24136 + }, + { + "epoch": 1.6199120834871312, + "grad_norm": 1.095659852027893, + "learning_rate": 9.171816847139448e-06, + "loss": 0.4479, + "step": 24138 + }, + { + "epoch": 1.6200463071709001, + "grad_norm": 1.1159011125564575, + "learning_rate": 9.165544023264233e-06, + "loss": 0.4538, + "step": 24140 + }, + { + "epoch": 1.6201805308546693, + "grad_norm": 0.9968121647834778, + "learning_rate": 9.159273128736734e-06, + "loss": 0.4992, + "step": 24142 + }, + { + "epoch": 1.6203147545384384, + "grad_norm": 1.04972243309021, + "learning_rate": 9.15300416385329e-06, + "loss": 0.5159, + "step": 24144 + }, + { + "epoch": 1.6204489782222073, + "grad_norm": 1.0225540399551392, + "learning_rate": 9.14673712891006e-06, + "loss": 0.4906, + "step": 24146 + }, + { + "epoch": 1.6205832019059763, + "grad_norm": 1.0914386510849, + "learning_rate": 9.140472024203179e-06, + "loss": 0.4838, + "step": 24148 + }, + { + "epoch": 1.6207174255897452, + "grad_norm": 0.9849649667739868, + "learning_rate": 9.134208850028647e-06, + "loss": 0.4953, + "step": 24150 + }, + { + "epoch": 1.6208516492735143, + "grad_norm": 1.0372711420059204, + "learning_rate": 9.127947606682391e-06, + "loss": 0.4938, + "step": 24152 + }, + { + "epoch": 1.6209858729572835, + "grad_norm": 1.1415741443634033, + "learning_rate": 9.121688294460223e-06, + "loss": 0.4554, + "step": 24154 + }, + { + "epoch": 1.6211200966410524, + "grad_norm": 1.046114206314087, + "learning_rate": 9.115430913657912e-06, + "loss": 0.5092, + "step": 24156 + }, + { + "epoch": 1.6212543203248213, + "grad_norm": 1.081341028213501, + "learning_rate": 9.109175464571096e-06, + "loss": 0.469, + "step": 24158 + }, + { + "epoch": 1.6213885440085902, + "grad_norm": 0.9061887264251709, + "learning_rate": 9.10292194749533e-06, + "loss": 0.426, + "step": 24160 + }, + { + "epoch": 1.6215227676923594, + "grad_norm": 1.0150114297866821, + "learning_rate": 9.096670362726073e-06, + "loss": 0.4992, + "step": 24162 + }, + { + "epoch": 1.6216569913761283, + "grad_norm": 1.1075929403305054, + "learning_rate": 9.090420710558718e-06, + "loss": 0.4694, + "step": 24164 + }, + { + "epoch": 1.6217912150598974, + "grad_norm": 1.0864084959030151, + "learning_rate": 9.084172991288525e-06, + "loss": 0.5109, + "step": 24166 + }, + { + "epoch": 1.6219254387436663, + "grad_norm": 1.1024622917175293, + "learning_rate": 9.077927205210712e-06, + "loss": 0.4718, + "step": 24168 + }, + { + "epoch": 1.6220596624274353, + "grad_norm": 1.1553581953048706, + "learning_rate": 9.071683352620385e-06, + "loss": 0.4834, + "step": 24170 + }, + { + "epoch": 1.6221938861112042, + "grad_norm": 1.1230138540267944, + "learning_rate": 9.06544143381251e-06, + "loss": 0.4815, + "step": 24172 + }, + { + "epoch": 1.6223281097949733, + "grad_norm": 0.9074229001998901, + "learning_rate": 9.059201449082045e-06, + "loss": 0.4646, + "step": 24174 + }, + { + "epoch": 1.6224623334787425, + "grad_norm": 0.9188847541809082, + "learning_rate": 9.052963398723796e-06, + "loss": 0.4414, + "step": 24176 + }, + { + "epoch": 1.6225965571625114, + "grad_norm": 1.0981385707855225, + "learning_rate": 9.046727283032519e-06, + "loss": 0.4777, + "step": 24178 + }, + { + "epoch": 1.6227307808462803, + "grad_norm": 1.112444519996643, + "learning_rate": 9.040493102302844e-06, + "loss": 0.4312, + "step": 24180 + }, + { + "epoch": 1.6228650045300492, + "grad_norm": 0.9772467613220215, + "learning_rate": 9.03426085682933e-06, + "loss": 0.4385, + "step": 24182 + }, + { + "epoch": 1.6229992282138184, + "grad_norm": 0.9029442071914673, + "learning_rate": 9.028030546906419e-06, + "loss": 0.4033, + "step": 24184 + }, + { + "epoch": 1.6231334518975873, + "grad_norm": 1.0854231119155884, + "learning_rate": 9.021802172828509e-06, + "loss": 0.4883, + "step": 24186 + }, + { + "epoch": 1.6232676755813564, + "grad_norm": 0.9764794707298279, + "learning_rate": 9.01557573488987e-06, + "loss": 0.474, + "step": 24188 + }, + { + "epoch": 1.6234018992651253, + "grad_norm": 1.0908037424087524, + "learning_rate": 9.009351233384684e-06, + "loss": 0.4767, + "step": 24190 + }, + { + "epoch": 1.6235361229488943, + "grad_norm": 1.0757347345352173, + "learning_rate": 9.003128668607031e-06, + "loss": 0.5152, + "step": 24192 + }, + { + "epoch": 1.6236703466326632, + "grad_norm": 1.0436749458312988, + "learning_rate": 8.99690804085095e-06, + "loss": 0.4605, + "step": 24194 + }, + { + "epoch": 1.6238045703164323, + "grad_norm": 1.0184334516525269, + "learning_rate": 8.990689350410314e-06, + "loss": 0.4616, + "step": 24196 + }, + { + "epoch": 1.6239387940002015, + "grad_norm": 1.022641658782959, + "learning_rate": 8.984472597578997e-06, + "loss": 0.4667, + "step": 24198 + }, + { + "epoch": 1.6240730176839704, + "grad_norm": 0.9187766313552856, + "learning_rate": 8.978257782650668e-06, + "loss": 0.4686, + "step": 24200 + }, + { + "epoch": 1.6242072413677393, + "grad_norm": 0.908483624458313, + "learning_rate": 8.972044905919008e-06, + "loss": 0.4349, + "step": 24202 + }, + { + "epoch": 1.6243414650515082, + "grad_norm": 1.0387719869613647, + "learning_rate": 8.965833967677534e-06, + "loss": 0.4219, + "step": 24204 + }, + { + "epoch": 1.6244756887352774, + "grad_norm": 0.919877827167511, + "learning_rate": 8.959624968219732e-06, + "loss": 0.4656, + "step": 24206 + }, + { + "epoch": 1.6246099124190463, + "grad_norm": 1.0006935596466064, + "learning_rate": 8.95341790783894e-06, + "loss": 0.4791, + "step": 24208 + }, + { + "epoch": 1.6247441361028154, + "grad_norm": 1.0531569719314575, + "learning_rate": 8.94721278682844e-06, + "loss": 0.4467, + "step": 24210 + }, + { + "epoch": 1.6248783597865843, + "grad_norm": 0.9406517744064331, + "learning_rate": 8.941009605481398e-06, + "loss": 0.4366, + "step": 24212 + }, + { + "epoch": 1.6250125834703533, + "grad_norm": 0.979747474193573, + "learning_rate": 8.934808364090924e-06, + "loss": 0.4619, + "step": 24214 + }, + { + "epoch": 1.6251468071541222, + "grad_norm": 0.9257476925849915, + "learning_rate": 8.928609062950005e-06, + "loss": 0.5277, + "step": 24216 + }, + { + "epoch": 1.6252810308378913, + "grad_norm": 1.0857957601547241, + "learning_rate": 8.922411702351546e-06, + "loss": 0.4364, + "step": 24218 + }, + { + "epoch": 1.6254152545216605, + "grad_norm": 0.9188510775566101, + "learning_rate": 8.916216282588341e-06, + "loss": 0.4383, + "step": 24220 + }, + { + "epoch": 1.6255494782054294, + "grad_norm": 1.0596593618392944, + "learning_rate": 8.910022803953144e-06, + "loss": 0.4746, + "step": 24222 + }, + { + "epoch": 1.6256837018891983, + "grad_norm": 1.073279857635498, + "learning_rate": 8.90383126673856e-06, + "loss": 0.4814, + "step": 24224 + }, + { + "epoch": 1.6258179255729672, + "grad_norm": 1.12808096408844, + "learning_rate": 8.897641671237156e-06, + "loss": 0.4481, + "step": 24226 + }, + { + "epoch": 1.6259521492567364, + "grad_norm": 0.8533748984336853, + "learning_rate": 8.891454017741341e-06, + "loss": 0.4077, + "step": 24228 + }, + { + "epoch": 1.6260863729405055, + "grad_norm": 1.05867338180542, + "learning_rate": 8.885268306543492e-06, + "loss": 0.5289, + "step": 24230 + }, + { + "epoch": 1.6262205966242744, + "grad_norm": 0.8957057595252991, + "learning_rate": 8.879084537935855e-06, + "loss": 0.4112, + "step": 24232 + }, + { + "epoch": 1.6263548203080433, + "grad_norm": 1.0482287406921387, + "learning_rate": 8.872902712210628e-06, + "loss": 0.4893, + "step": 24234 + }, + { + "epoch": 1.6264890439918123, + "grad_norm": 1.076281189918518, + "learning_rate": 8.86672282965988e-06, + "loss": 0.434, + "step": 24236 + }, + { + "epoch": 1.6266232676755814, + "grad_norm": 1.1521581411361694, + "learning_rate": 8.860544890575584e-06, + "loss": 0.4854, + "step": 24238 + }, + { + "epoch": 1.6267574913593503, + "grad_norm": 0.9764477014541626, + "learning_rate": 8.85436889524965e-06, + "loss": 0.4432, + "step": 24240 + }, + { + "epoch": 1.6268917150431195, + "grad_norm": 1.2391307353973389, + "learning_rate": 8.848194843973862e-06, + "loss": 0.5098, + "step": 24242 + }, + { + "epoch": 1.6270259387268884, + "grad_norm": 0.8947152495384216, + "learning_rate": 8.842022737039957e-06, + "loss": 0.4335, + "step": 24244 + }, + { + "epoch": 1.6271601624106573, + "grad_norm": 1.1010911464691162, + "learning_rate": 8.835852574739544e-06, + "loss": 0.5048, + "step": 24246 + }, + { + "epoch": 1.6272943860944262, + "grad_norm": 1.032392978668213, + "learning_rate": 8.829684357364153e-06, + "loss": 0.4737, + "step": 24248 + }, + { + "epoch": 1.6274286097781954, + "grad_norm": 1.0642104148864746, + "learning_rate": 8.823518085205206e-06, + "loss": 0.519, + "step": 24250 + }, + { + "epoch": 1.6275628334619645, + "grad_norm": 0.9958031177520752, + "learning_rate": 8.817353758554075e-06, + "loss": 0.4677, + "step": 24252 + }, + { + "epoch": 1.6276970571457334, + "grad_norm": 0.9674162268638611, + "learning_rate": 8.811191377701995e-06, + "loss": 0.4031, + "step": 24254 + }, + { + "epoch": 1.6278312808295023, + "grad_norm": 1.0566951036453247, + "learning_rate": 8.805030942940123e-06, + "loss": 0.4425, + "step": 24256 + }, + { + "epoch": 1.6279655045132713, + "grad_norm": 1.0692909955978394, + "learning_rate": 8.79887245455952e-06, + "loss": 0.4803, + "step": 24258 + }, + { + "epoch": 1.6280997281970404, + "grad_norm": 0.8982757925987244, + "learning_rate": 8.792715912851196e-06, + "loss": 0.4266, + "step": 24260 + }, + { + "epoch": 1.6282339518808093, + "grad_norm": 1.057623028755188, + "learning_rate": 8.786561318105996e-06, + "loss": 0.4831, + "step": 24262 + }, + { + "epoch": 1.6283681755645785, + "grad_norm": 1.1221034526824951, + "learning_rate": 8.780408670614753e-06, + "loss": 0.5228, + "step": 24264 + }, + { + "epoch": 1.6285023992483474, + "grad_norm": 0.9700652360916138, + "learning_rate": 8.774257970668127e-06, + "loss": 0.4641, + "step": 24266 + }, + { + "epoch": 1.6286366229321163, + "grad_norm": 1.0552066564559937, + "learning_rate": 8.768109218556753e-06, + "loss": 0.485, + "step": 24268 + }, + { + "epoch": 1.6287708466158852, + "grad_norm": 1.1140886545181274, + "learning_rate": 8.76196241457113e-06, + "loss": 0.4931, + "step": 24270 + }, + { + "epoch": 1.6289050702996544, + "grad_norm": 0.9525741338729858, + "learning_rate": 8.7558175590017e-06, + "loss": 0.4955, + "step": 24272 + }, + { + "epoch": 1.6290392939834235, + "grad_norm": 0.972881555557251, + "learning_rate": 8.74967465213879e-06, + "loss": 0.4408, + "step": 24274 + }, + { + "epoch": 1.6291735176671924, + "grad_norm": 1.0415549278259277, + "learning_rate": 8.743533694272638e-06, + "loss": 0.4838, + "step": 24276 + }, + { + "epoch": 1.6293077413509613, + "grad_norm": 1.0316137075424194, + "learning_rate": 8.737394685693378e-06, + "loss": 0.4813, + "step": 24278 + }, + { + "epoch": 1.6294419650347303, + "grad_norm": 1.128528356552124, + "learning_rate": 8.731257626691092e-06, + "loss": 0.4627, + "step": 24280 + }, + { + "epoch": 1.6295761887184994, + "grad_norm": 0.9621889591217041, + "learning_rate": 8.725122517555734e-06, + "loss": 0.443, + "step": 24282 + }, + { + "epoch": 1.6297104124022683, + "grad_norm": 1.0137451887130737, + "learning_rate": 8.718989358577167e-06, + "loss": 0.503, + "step": 24284 + }, + { + "epoch": 1.6298446360860375, + "grad_norm": 1.077551245689392, + "learning_rate": 8.712858150045172e-06, + "loss": 0.4421, + "step": 24286 + }, + { + "epoch": 1.6299788597698064, + "grad_norm": 1.0536868572235107, + "learning_rate": 8.706728892249449e-06, + "loss": 0.4551, + "step": 24288 + }, + { + "epoch": 1.6301130834535753, + "grad_norm": 1.1465702056884766, + "learning_rate": 8.700601585479579e-06, + "loss": 0.4962, + "step": 24290 + }, + { + "epoch": 1.6302473071373442, + "grad_norm": 0.9732354283332825, + "learning_rate": 8.694476230025094e-06, + "loss": 0.4252, + "step": 24292 + }, + { + "epoch": 1.6303815308211134, + "grad_norm": 0.9033833146095276, + "learning_rate": 8.68835282617536e-06, + "loss": 0.4583, + "step": 24294 + }, + { + "epoch": 1.6305157545048825, + "grad_norm": 1.121772050857544, + "learning_rate": 8.68223137421973e-06, + "loss": 0.4867, + "step": 24296 + }, + { + "epoch": 1.6306499781886514, + "grad_norm": 1.051953911781311, + "learning_rate": 8.676111874447407e-06, + "loss": 0.4999, + "step": 24298 + }, + { + "epoch": 1.6307842018724203, + "grad_norm": 1.007792592048645, + "learning_rate": 8.669994327147552e-06, + "loss": 0.541, + "step": 24300 + }, + { + "epoch": 1.6309184255561893, + "grad_norm": 0.9095066785812378, + "learning_rate": 8.663878732609187e-06, + "loss": 0.4635, + "step": 24302 + }, + { + "epoch": 1.6310526492399584, + "grad_norm": 0.9789410829544067, + "learning_rate": 8.657765091121273e-06, + "loss": 0.4391, + "step": 24304 + }, + { + "epoch": 1.6311868729237275, + "grad_norm": 1.049037218093872, + "learning_rate": 8.651653402972654e-06, + "loss": 0.4506, + "step": 24306 + }, + { + "epoch": 1.6313210966074965, + "grad_norm": 0.935494065284729, + "learning_rate": 8.645543668452112e-06, + "loss": 0.4854, + "step": 24308 + }, + { + "epoch": 1.6314553202912654, + "grad_norm": 1.0862587690353394, + "learning_rate": 8.639435887848307e-06, + "loss": 0.4681, + "step": 24310 + }, + { + "epoch": 1.6315895439750343, + "grad_norm": 0.9944185614585876, + "learning_rate": 8.63333006144983e-06, + "loss": 0.436, + "step": 24312 + }, + { + "epoch": 1.6317237676588034, + "grad_norm": 1.0487542152404785, + "learning_rate": 8.627226189545162e-06, + "loss": 0.4708, + "step": 24314 + }, + { + "epoch": 1.6318579913425724, + "grad_norm": 1.168813705444336, + "learning_rate": 8.621124272422688e-06, + "loss": 0.4989, + "step": 24316 + }, + { + "epoch": 1.6319922150263415, + "grad_norm": 1.0781807899475098, + "learning_rate": 8.61502431037074e-06, + "loss": 0.5246, + "step": 24318 + }, + { + "epoch": 1.6321264387101104, + "grad_norm": 1.1735255718231201, + "learning_rate": 8.608926303677506e-06, + "loss": 0.5324, + "step": 24320 + }, + { + "epoch": 1.6322606623938793, + "grad_norm": 1.054585337638855, + "learning_rate": 8.602830252631116e-06, + "loss": 0.4316, + "step": 24322 + }, + { + "epoch": 1.6323948860776483, + "grad_norm": 1.0092709064483643, + "learning_rate": 8.59673615751958e-06, + "loss": 0.4439, + "step": 24324 + }, + { + "epoch": 1.6325291097614174, + "grad_norm": 0.9695367813110352, + "learning_rate": 8.590644018630861e-06, + "loss": 0.4566, + "step": 24326 + }, + { + "epoch": 1.6326633334451865, + "grad_norm": 1.1469132900238037, + "learning_rate": 8.584553836252768e-06, + "loss": 0.503, + "step": 24328 + }, + { + "epoch": 1.6327975571289555, + "grad_norm": 0.9535269737243652, + "learning_rate": 8.578465610673076e-06, + "loss": 0.4698, + "step": 24330 + }, + { + "epoch": 1.6329317808127244, + "grad_norm": 1.1265140771865845, + "learning_rate": 8.572379342179437e-06, + "loss": 0.4746, + "step": 24332 + }, + { + "epoch": 1.6330660044964933, + "grad_norm": 0.9789575934410095, + "learning_rate": 8.566295031059407e-06, + "loss": 0.4916, + "step": 24334 + }, + { + "epoch": 1.6332002281802624, + "grad_norm": 1.0417834520339966, + "learning_rate": 8.560212677600448e-06, + "loss": 0.4428, + "step": 24336 + }, + { + "epoch": 1.6333344518640314, + "grad_norm": 1.0458998680114746, + "learning_rate": 8.554132282089967e-06, + "loss": 0.4488, + "step": 24338 + }, + { + "epoch": 1.6334686755478005, + "grad_norm": 1.9805901050567627, + "learning_rate": 8.548053844815236e-06, + "loss": 0.4567, + "step": 24340 + }, + { + "epoch": 1.6336028992315694, + "grad_norm": 1.0965425968170166, + "learning_rate": 8.541977366063448e-06, + "loss": 0.4532, + "step": 24342 + }, + { + "epoch": 1.6337371229153383, + "grad_norm": 0.953021764755249, + "learning_rate": 8.53590284612169e-06, + "loss": 0.497, + "step": 24344 + }, + { + "epoch": 1.6338713465991073, + "grad_norm": 1.0408341884613037, + "learning_rate": 8.529830285277001e-06, + "loss": 0.4347, + "step": 24346 + }, + { + "epoch": 1.6340055702828764, + "grad_norm": 1.0203832387924194, + "learning_rate": 8.523759683816274e-06, + "loss": 0.491, + "step": 24348 + }, + { + "epoch": 1.6341397939666455, + "grad_norm": 0.9675636887550354, + "learning_rate": 8.517691042026365e-06, + "loss": 0.5657, + "step": 24350 + }, + { + "epoch": 1.6342740176504145, + "grad_norm": 0.9288319945335388, + "learning_rate": 8.51162436019396e-06, + "loss": 0.4978, + "step": 24352 + }, + { + "epoch": 1.6344082413341834, + "grad_norm": 1.0077799558639526, + "learning_rate": 8.505559638605732e-06, + "loss": 0.4438, + "step": 24354 + }, + { + "epoch": 1.6345424650179523, + "grad_norm": 0.9253379106521606, + "learning_rate": 8.499496877548202e-06, + "loss": 0.4669, + "step": 24356 + }, + { + "epoch": 1.6346766887017214, + "grad_norm": 1.0141042470932007, + "learning_rate": 8.493436077307848e-06, + "loss": 0.4449, + "step": 24358 + }, + { + "epoch": 1.6348109123854904, + "grad_norm": 0.9649704098701477, + "learning_rate": 8.487377238171024e-06, + "loss": 0.4532, + "step": 24360 + }, + { + "epoch": 1.6349451360692595, + "grad_norm": 1.0539889335632324, + "learning_rate": 8.481320360423994e-06, + "loss": 0.4502, + "step": 24362 + }, + { + "epoch": 1.6350793597530284, + "grad_norm": 1.0667953491210938, + "learning_rate": 8.47526544435292e-06, + "loss": 0.4614, + "step": 24364 + }, + { + "epoch": 1.6352135834367973, + "grad_norm": 1.1892738342285156, + "learning_rate": 8.469212490243911e-06, + "loss": 0.4995, + "step": 24366 + }, + { + "epoch": 1.6353478071205663, + "grad_norm": 0.9447067379951477, + "learning_rate": 8.463161498382948e-06, + "loss": 0.4993, + "step": 24368 + }, + { + "epoch": 1.6354820308043354, + "grad_norm": 0.9807162880897522, + "learning_rate": 8.457112469055923e-06, + "loss": 0.5024, + "step": 24370 + }, + { + "epoch": 1.6356162544881045, + "grad_norm": 1.0578080415725708, + "learning_rate": 8.45106540254863e-06, + "loss": 0.4119, + "step": 24372 + }, + { + "epoch": 1.6357504781718735, + "grad_norm": 0.9289562702178955, + "learning_rate": 8.445020299146812e-06, + "loss": 0.4189, + "step": 24374 + }, + { + "epoch": 1.6358847018556424, + "grad_norm": 0.8946725726127625, + "learning_rate": 8.438977159136063e-06, + "loss": 0.443, + "step": 24376 + }, + { + "epoch": 1.6360189255394113, + "grad_norm": 1.1086225509643555, + "learning_rate": 8.432935982801921e-06, + "loss": 0.477, + "step": 24378 + }, + { + "epoch": 1.6361531492231804, + "grad_norm": 1.088677167892456, + "learning_rate": 8.426896770429815e-06, + "loss": 0.4793, + "step": 24380 + }, + { + "epoch": 1.6362873729069496, + "grad_norm": 1.0079419612884521, + "learning_rate": 8.420859522305075e-06, + "loss": 0.4715, + "step": 24382 + }, + { + "epoch": 1.6364215965907185, + "grad_norm": 0.9994024634361267, + "learning_rate": 8.41482423871297e-06, + "loss": 0.4382, + "step": 24384 + }, + { + "epoch": 1.6365558202744874, + "grad_norm": 1.1206235885620117, + "learning_rate": 8.408790919938636e-06, + "loss": 0.4492, + "step": 24386 + }, + { + "epoch": 1.6366900439582563, + "grad_norm": 1.0920964479446411, + "learning_rate": 8.402759566267171e-06, + "loss": 0.4001, + "step": 24388 + }, + { + "epoch": 1.6368242676420255, + "grad_norm": 1.0045682191848755, + "learning_rate": 8.396730177983497e-06, + "loss": 0.543, + "step": 24390 + }, + { + "epoch": 1.6369584913257944, + "grad_norm": 1.0271632671356201, + "learning_rate": 8.39070275537252e-06, + "loss": 0.4419, + "step": 24392 + }, + { + "epoch": 1.6370927150095635, + "grad_norm": 0.9291853904724121, + "learning_rate": 8.384677298719006e-06, + "loss": 0.5088, + "step": 24394 + }, + { + "epoch": 1.6372269386933325, + "grad_norm": 1.037255883216858, + "learning_rate": 8.378653808307668e-06, + "loss": 0.4374, + "step": 24396 + }, + { + "epoch": 1.6373611623771014, + "grad_norm": 1.1152249574661255, + "learning_rate": 8.372632284423094e-06, + "loss": 0.4607, + "step": 24398 + }, + { + "epoch": 1.6374953860608703, + "grad_norm": 0.9363716244697571, + "learning_rate": 8.366612727349787e-06, + "loss": 0.503, + "step": 24400 + }, + { + "epoch": 1.6376296097446394, + "grad_norm": 1.0684161186218262, + "learning_rate": 8.360595137372151e-06, + "loss": 0.5139, + "step": 24402 + }, + { + "epoch": 1.6377638334284086, + "grad_norm": 0.9547727704048157, + "learning_rate": 8.354579514774525e-06, + "loss": 0.4621, + "step": 24404 + }, + { + "epoch": 1.6378980571121775, + "grad_norm": 1.076237440109253, + "learning_rate": 8.348565859841123e-06, + "loss": 0.5022, + "step": 24406 + }, + { + "epoch": 1.6380322807959464, + "grad_norm": 1.1423180103302002, + "learning_rate": 8.342554172856076e-06, + "loss": 0.5077, + "step": 24408 + }, + { + "epoch": 1.6381665044797153, + "grad_norm": 0.9638442397117615, + "learning_rate": 8.336544454103424e-06, + "loss": 0.4551, + "step": 24410 + }, + { + "epoch": 1.6383007281634845, + "grad_norm": 0.9665816426277161, + "learning_rate": 8.330536703867126e-06, + "loss": 0.4097, + "step": 24412 + }, + { + "epoch": 1.6384349518472534, + "grad_norm": 0.9710249900817871, + "learning_rate": 8.324530922431018e-06, + "loss": 0.454, + "step": 24414 + }, + { + "epoch": 1.6385691755310225, + "grad_norm": 0.9155290126800537, + "learning_rate": 8.3185271100789e-06, + "loss": 0.4525, + "step": 24416 + }, + { + "epoch": 1.6387033992147915, + "grad_norm": 1.1133970022201538, + "learning_rate": 8.312525267094385e-06, + "loss": 0.4567, + "step": 24418 + }, + { + "epoch": 1.6388376228985604, + "grad_norm": 1.2855124473571777, + "learning_rate": 8.306525393761095e-06, + "loss": 0.4322, + "step": 24420 + }, + { + "epoch": 1.6389718465823293, + "grad_norm": 1.0277609825134277, + "learning_rate": 8.300527490362476e-06, + "loss": 0.4718, + "step": 24422 + }, + { + "epoch": 1.6391060702660984, + "grad_norm": 1.107844591140747, + "learning_rate": 8.294531557181945e-06, + "loss": 0.518, + "step": 24424 + }, + { + "epoch": 1.6392402939498676, + "grad_norm": 1.0599548816680908, + "learning_rate": 8.288537594502788e-06, + "loss": 0.4354, + "step": 24426 + }, + { + "epoch": 1.6393745176336365, + "grad_norm": 1.0279606580734253, + "learning_rate": 8.282545602608211e-06, + "loss": 0.4515, + "step": 24428 + }, + { + "epoch": 1.6395087413174054, + "grad_norm": 1.03935968875885, + "learning_rate": 8.276555581781303e-06, + "loss": 0.4835, + "step": 24430 + }, + { + "epoch": 1.6396429650011743, + "grad_norm": 1.0039790868759155, + "learning_rate": 8.270567532305118e-06, + "loss": 0.461, + "step": 24432 + }, + { + "epoch": 1.6397771886849435, + "grad_norm": 0.9484822154045105, + "learning_rate": 8.264581454462555e-06, + "loss": 0.4623, + "step": 24434 + }, + { + "epoch": 1.6399114123687124, + "grad_norm": 0.9461115002632141, + "learning_rate": 8.25859734853645e-06, + "loss": 0.4257, + "step": 24436 + }, + { + "epoch": 1.6400456360524815, + "grad_norm": 1.037699818611145, + "learning_rate": 8.252615214809528e-06, + "loss": 0.511, + "step": 24438 + }, + { + "epoch": 1.6401798597362505, + "grad_norm": 1.0340139865875244, + "learning_rate": 8.246635053564455e-06, + "loss": 0.5021, + "step": 24440 + }, + { + "epoch": 1.6403140834200194, + "grad_norm": 1.0507210493087769, + "learning_rate": 8.240656865083757e-06, + "loss": 0.4525, + "step": 24442 + }, + { + "epoch": 1.6404483071037883, + "grad_norm": 0.9403718113899231, + "learning_rate": 8.234680649649935e-06, + "loss": 0.4646, + "step": 24444 + }, + { + "epoch": 1.6405825307875574, + "grad_norm": 0.8724382519721985, + "learning_rate": 8.228706407545306e-06, + "loss": 0.467, + "step": 24446 + }, + { + "epoch": 1.6407167544713266, + "grad_norm": 1.0470285415649414, + "learning_rate": 8.222734139052152e-06, + "loss": 0.5314, + "step": 24448 + }, + { + "epoch": 1.6408509781550955, + "grad_norm": 0.9047724604606628, + "learning_rate": 8.21676384445267e-06, + "loss": 0.4392, + "step": 24450 + }, + { + "epoch": 1.6409852018388644, + "grad_norm": 0.9392143487930298, + "learning_rate": 8.210795524028924e-06, + "loss": 0.5069, + "step": 24452 + }, + { + "epoch": 1.6411194255226333, + "grad_norm": 1.0471280813217163, + "learning_rate": 8.204829178062923e-06, + "loss": 0.4784, + "step": 24454 + }, + { + "epoch": 1.6412536492064025, + "grad_norm": 1.150148868560791, + "learning_rate": 8.198864806836553e-06, + "loss": 0.4711, + "step": 24456 + }, + { + "epoch": 1.6413878728901716, + "grad_norm": 1.0669646263122559, + "learning_rate": 8.192902410631632e-06, + "loss": 0.4197, + "step": 24458 + }, + { + "epoch": 1.6415220965739405, + "grad_norm": 1.0547282695770264, + "learning_rate": 8.18694198972984e-06, + "loss": 0.4555, + "step": 24460 + }, + { + "epoch": 1.6416563202577095, + "grad_norm": 1.076779842376709, + "learning_rate": 8.180983544412834e-06, + "loss": 0.4827, + "step": 24462 + }, + { + "epoch": 1.6417905439414784, + "grad_norm": 0.961208701133728, + "learning_rate": 8.175027074962122e-06, + "loss": 0.4254, + "step": 24464 + }, + { + "epoch": 1.6419247676252475, + "grad_norm": 1.114872932434082, + "learning_rate": 8.169072581659137e-06, + "loss": 0.4758, + "step": 24466 + }, + { + "epoch": 1.6420589913090164, + "grad_norm": 1.0683403015136719, + "learning_rate": 8.163120064785201e-06, + "loss": 0.4474, + "step": 24468 + }, + { + "epoch": 1.6421932149927856, + "grad_norm": 0.8533875942230225, + "learning_rate": 8.157169524621583e-06, + "loss": 0.501, + "step": 24470 + }, + { + "epoch": 1.6423274386765545, + "grad_norm": 1.0208030939102173, + "learning_rate": 8.15122096144943e-06, + "loss": 0.4891, + "step": 24472 + }, + { + "epoch": 1.6424616623603234, + "grad_norm": 0.895160973072052, + "learning_rate": 8.145274375549794e-06, + "loss": 0.4192, + "step": 24474 + }, + { + "epoch": 1.6425958860440923, + "grad_norm": 0.9571309089660645, + "learning_rate": 8.139329767203624e-06, + "loss": 0.4343, + "step": 24476 + }, + { + "epoch": 1.6427301097278615, + "grad_norm": 1.2458082437515259, + "learning_rate": 8.133387136691822e-06, + "loss": 0.4861, + "step": 24478 + }, + { + "epoch": 1.6428643334116306, + "grad_norm": 1.117855191230774, + "learning_rate": 8.127446484295137e-06, + "loss": 0.4511, + "step": 24480 + }, + { + "epoch": 1.6429985570953995, + "grad_norm": 0.957265317440033, + "learning_rate": 8.121507810294276e-06, + "loss": 0.4072, + "step": 24482 + }, + { + "epoch": 1.6431327807791685, + "grad_norm": 0.9388834238052368, + "learning_rate": 8.115571114969828e-06, + "loss": 0.4677, + "step": 24484 + }, + { + "epoch": 1.6432670044629374, + "grad_norm": 1.0905842781066895, + "learning_rate": 8.109636398602276e-06, + "loss": 0.5119, + "step": 24486 + }, + { + "epoch": 1.6434012281467065, + "grad_norm": 1.0503604412078857, + "learning_rate": 8.103703661472023e-06, + "loss": 0.4732, + "step": 24488 + }, + { + "epoch": 1.6435354518304754, + "grad_norm": 1.1269357204437256, + "learning_rate": 8.09777290385939e-06, + "loss": 0.5235, + "step": 24490 + }, + { + "epoch": 1.6436696755142446, + "grad_norm": 0.9764650464057922, + "learning_rate": 8.091844126044601e-06, + "loss": 0.5161, + "step": 24492 + }, + { + "epoch": 1.6438038991980135, + "grad_norm": 1.036859393119812, + "learning_rate": 8.085917328307763e-06, + "loss": 0.5158, + "step": 24494 + }, + { + "epoch": 1.6439381228817824, + "grad_norm": 1.1827311515808105, + "learning_rate": 8.079992510928897e-06, + "loss": 0.4658, + "step": 24496 + }, + { + "epoch": 1.6440723465655513, + "grad_norm": 1.084754228591919, + "learning_rate": 8.074069674187968e-06, + "loss": 0.4841, + "step": 24498 + }, + { + "epoch": 1.6442065702493205, + "grad_norm": 0.7590323686599731, + "learning_rate": 8.068148818364802e-06, + "loss": 0.3998, + "step": 24500 + }, + { + "epoch": 1.6443407939330896, + "grad_norm": 0.8609249591827393, + "learning_rate": 8.062229943739148e-06, + "loss": 0.4364, + "step": 24502 + }, + { + "epoch": 1.6444750176168585, + "grad_norm": 1.0397735834121704, + "learning_rate": 8.056313050590652e-06, + "loss": 0.4556, + "step": 24504 + }, + { + "epoch": 1.6446092413006275, + "grad_norm": 0.9933218359947205, + "learning_rate": 8.050398139198895e-06, + "loss": 0.4548, + "step": 24506 + }, + { + "epoch": 1.6447434649843964, + "grad_norm": 1.1229078769683838, + "learning_rate": 8.044485209843327e-06, + "loss": 0.4544, + "step": 24508 + }, + { + "epoch": 1.6448776886681655, + "grad_norm": 1.0707441568374634, + "learning_rate": 8.03857426280334e-06, + "loss": 0.5081, + "step": 24510 + }, + { + "epoch": 1.6450119123519344, + "grad_norm": 1.0590848922729492, + "learning_rate": 8.032665298358205e-06, + "loss": 0.4747, + "step": 24512 + }, + { + "epoch": 1.6451461360357036, + "grad_norm": 0.9965388774871826, + "learning_rate": 8.026758316787108e-06, + "loss": 0.4613, + "step": 24514 + }, + { + "epoch": 1.6452803597194725, + "grad_norm": 1.0541249513626099, + "learning_rate": 8.020853318369149e-06, + "loss": 0.4614, + "step": 24516 + }, + { + "epoch": 1.6454145834032414, + "grad_norm": 1.0630964040756226, + "learning_rate": 8.014950303383305e-06, + "loss": 0.4383, + "step": 24518 + }, + { + "epoch": 1.6455488070870103, + "grad_norm": 1.1048094034194946, + "learning_rate": 8.009049272108516e-06, + "loss": 0.4591, + "step": 24520 + }, + { + "epoch": 1.6456830307707795, + "grad_norm": 1.175954818725586, + "learning_rate": 8.003150224823574e-06, + "loss": 0.4995, + "step": 24522 + }, + { + "epoch": 1.6458172544545486, + "grad_norm": 0.9277005195617676, + "learning_rate": 7.997253161807205e-06, + "loss": 0.414, + "step": 24524 + }, + { + "epoch": 1.6459514781383175, + "grad_norm": 0.9522114396095276, + "learning_rate": 7.991358083338013e-06, + "loss": 0.4443, + "step": 24526 + }, + { + "epoch": 1.6460857018220865, + "grad_norm": 0.9481935501098633, + "learning_rate": 7.985464989694558e-06, + "loss": 0.4982, + "step": 24528 + }, + { + "epoch": 1.6462199255058554, + "grad_norm": 1.0402849912643433, + "learning_rate": 7.979573881155261e-06, + "loss": 0.4713, + "step": 24530 + }, + { + "epoch": 1.6463541491896245, + "grad_norm": 1.096053957939148, + "learning_rate": 7.973684757998472e-06, + "loss": 0.5413, + "step": 24532 + }, + { + "epoch": 1.6464883728733937, + "grad_norm": 1.1752251386642456, + "learning_rate": 7.967797620502426e-06, + "loss": 0.4956, + "step": 24534 + }, + { + "epoch": 1.6466225965571626, + "grad_norm": 1.226892113685608, + "learning_rate": 7.961912468945304e-06, + "loss": 0.4958, + "step": 24536 + }, + { + "epoch": 1.6467568202409315, + "grad_norm": 1.0735069513320923, + "learning_rate": 7.956029303605134e-06, + "loss": 0.4744, + "step": 24538 + }, + { + "epoch": 1.6468910439247004, + "grad_norm": 0.8980748057365417, + "learning_rate": 7.950148124759932e-06, + "loss": 0.4323, + "step": 24540 + }, + { + "epoch": 1.6470252676084696, + "grad_norm": 1.162540316581726, + "learning_rate": 7.944268932687521e-06, + "loss": 0.5048, + "step": 24542 + }, + { + "epoch": 1.6471594912922385, + "grad_norm": 0.966513454914093, + "learning_rate": 7.938391727665712e-06, + "loss": 0.4728, + "step": 24544 + }, + { + "epoch": 1.6472937149760076, + "grad_norm": 1.059916615486145, + "learning_rate": 7.93251650997217e-06, + "loss": 0.4523, + "step": 24546 + }, + { + "epoch": 1.6474279386597765, + "grad_norm": 1.071812391281128, + "learning_rate": 7.926643279884521e-06, + "loss": 0.533, + "step": 24548 + }, + { + "epoch": 1.6475621623435455, + "grad_norm": 0.8966371417045593, + "learning_rate": 7.920772037680236e-06, + "loss": 0.4953, + "step": 24550 + }, + { + "epoch": 1.6476963860273144, + "grad_norm": 0.9865714907646179, + "learning_rate": 7.914902783636729e-06, + "loss": 0.4738, + "step": 24552 + }, + { + "epoch": 1.6478306097110835, + "grad_norm": 0.9470037221908569, + "learning_rate": 7.909035518031299e-06, + "loss": 0.4447, + "step": 24554 + }, + { + "epoch": 1.6479648333948527, + "grad_norm": 1.0400514602661133, + "learning_rate": 7.903170241141178e-06, + "loss": 0.4488, + "step": 24556 + }, + { + "epoch": 1.6480990570786216, + "grad_norm": 0.9073011875152588, + "learning_rate": 7.897306953243489e-06, + "loss": 0.4574, + "step": 24558 + }, + { + "epoch": 1.6482332807623905, + "grad_norm": 0.8621808886528015, + "learning_rate": 7.891445654615254e-06, + "loss": 0.4469, + "step": 24560 + }, + { + "epoch": 1.6483675044461594, + "grad_norm": 0.9729363918304443, + "learning_rate": 7.885586345533397e-06, + "loss": 0.4661, + "step": 24562 + }, + { + "epoch": 1.6485017281299286, + "grad_norm": 1.0364904403686523, + "learning_rate": 7.879729026274779e-06, + "loss": 0.4614, + "step": 24564 + }, + { + "epoch": 1.6486359518136975, + "grad_norm": 0.9890289902687073, + "learning_rate": 7.873873697116129e-06, + "loss": 0.4435, + "step": 24566 + }, + { + "epoch": 1.6487701754974666, + "grad_norm": 0.9436256885528564, + "learning_rate": 7.868020358334139e-06, + "loss": 0.4559, + "step": 24568 + }, + { + "epoch": 1.6489043991812355, + "grad_norm": 1.0989784002304077, + "learning_rate": 7.862169010205306e-06, + "loss": 0.456, + "step": 24570 + }, + { + "epoch": 1.6490386228650045, + "grad_norm": 1.0259724855422974, + "learning_rate": 7.856319653006144e-06, + "loss": 0.4262, + "step": 24572 + }, + { + "epoch": 1.6491728465487734, + "grad_norm": 1.0732414722442627, + "learning_rate": 7.850472287012988e-06, + "loss": 0.4381, + "step": 24574 + }, + { + "epoch": 1.6493070702325425, + "grad_norm": 1.0366039276123047, + "learning_rate": 7.84462691250215e-06, + "loss": 0.5088, + "step": 24576 + }, + { + "epoch": 1.6494412939163117, + "grad_norm": 1.1138050556182861, + "learning_rate": 7.8387835297498e-06, + "loss": 0.4757, + "step": 24578 + }, + { + "epoch": 1.6495755176000806, + "grad_norm": 1.0513672828674316, + "learning_rate": 7.832942139032012e-06, + "loss": 0.4919, + "step": 24580 + }, + { + "epoch": 1.6497097412838495, + "grad_norm": 0.8169397711753845, + "learning_rate": 7.827102740624798e-06, + "loss": 0.4306, + "step": 24582 + }, + { + "epoch": 1.6498439649676184, + "grad_norm": 1.0516151189804077, + "learning_rate": 7.821265334804028e-06, + "loss": 0.458, + "step": 24584 + }, + { + "epoch": 1.6499781886513876, + "grad_norm": 0.8732035160064697, + "learning_rate": 7.815429921845553e-06, + "loss": 0.3974, + "step": 24586 + }, + { + "epoch": 1.6501124123351565, + "grad_norm": 0.8667218685150146, + "learning_rate": 7.809596502025057e-06, + "loss": 0.4782, + "step": 24588 + }, + { + "epoch": 1.6502466360189256, + "grad_norm": 0.9133517742156982, + "learning_rate": 7.803765075618164e-06, + "loss": 0.4389, + "step": 24590 + }, + { + "epoch": 1.6503808597026945, + "grad_norm": 1.0821890830993652, + "learning_rate": 7.797935642900378e-06, + "loss": 0.465, + "step": 24592 + }, + { + "epoch": 1.6505150833864635, + "grad_norm": 0.9128795862197876, + "learning_rate": 7.792108204147164e-06, + "loss": 0.4588, + "step": 24594 + }, + { + "epoch": 1.6506493070702324, + "grad_norm": 1.2393850088119507, + "learning_rate": 7.786282759633839e-06, + "loss": 0.4522, + "step": 24596 + }, + { + "epoch": 1.6507835307540015, + "grad_norm": 1.0295946598052979, + "learning_rate": 7.78045930963564e-06, + "loss": 0.511, + "step": 24598 + }, + { + "epoch": 1.6509177544377707, + "grad_norm": 1.0016932487487793, + "learning_rate": 7.774637854427707e-06, + "loss": 0.443, + "step": 24600 + }, + { + "epoch": 1.6510519781215396, + "grad_norm": 0.9255552887916565, + "learning_rate": 7.768818394285116e-06, + "loss": 0.4247, + "step": 24602 + }, + { + "epoch": 1.6511862018053085, + "grad_norm": 1.1796424388885498, + "learning_rate": 7.7630009294828e-06, + "loss": 0.4073, + "step": 24604 + }, + { + "epoch": 1.6513204254890774, + "grad_norm": 0.8748919367790222, + "learning_rate": 7.757185460295657e-06, + "loss": 0.3719, + "step": 24606 + }, + { + "epoch": 1.6514546491728466, + "grad_norm": 1.1171672344207764, + "learning_rate": 7.751371986998412e-06, + "loss": 0.4661, + "step": 24608 + }, + { + "epoch": 1.6515888728566157, + "grad_norm": 1.030329704284668, + "learning_rate": 7.745560509865773e-06, + "loss": 0.4457, + "step": 24610 + }, + { + "epoch": 1.6517230965403846, + "grad_norm": 1.0265859365463257, + "learning_rate": 7.739751029172298e-06, + "loss": 0.4772, + "step": 24612 + }, + { + "epoch": 1.6518573202241535, + "grad_norm": 1.0171829462051392, + "learning_rate": 7.7339435451925e-06, + "loss": 0.4653, + "step": 24614 + }, + { + "epoch": 1.6519915439079225, + "grad_norm": 1.1498104333877563, + "learning_rate": 7.728138058200757e-06, + "loss": 0.4804, + "step": 24616 + }, + { + "epoch": 1.6521257675916916, + "grad_norm": 1.0023661851882935, + "learning_rate": 7.722334568471362e-06, + "loss": 0.4265, + "step": 24618 + }, + { + "epoch": 1.6522599912754605, + "grad_norm": 0.9971412420272827, + "learning_rate": 7.716533076278515e-06, + "loss": 0.44, + "step": 24620 + }, + { + "epoch": 1.6523942149592297, + "grad_norm": 1.0238125324249268, + "learning_rate": 7.710733581896345e-06, + "loss": 0.4527, + "step": 24622 + }, + { + "epoch": 1.6525284386429986, + "grad_norm": 0.9874955415725708, + "learning_rate": 7.704936085598852e-06, + "loss": 0.466, + "step": 24624 + }, + { + "epoch": 1.6526626623267675, + "grad_norm": 1.1266205310821533, + "learning_rate": 7.699140587659965e-06, + "loss": 0.5204, + "step": 24626 + }, + { + "epoch": 1.6527968860105364, + "grad_norm": 0.9751604795455933, + "learning_rate": 7.693347088353487e-06, + "loss": 0.4418, + "step": 24628 + }, + { + "epoch": 1.6529311096943056, + "grad_norm": 0.9777958393096924, + "learning_rate": 7.687555587953176e-06, + "loss": 0.452, + "step": 24630 + }, + { + "epoch": 1.6530653333780747, + "grad_norm": 1.0757713317871094, + "learning_rate": 7.681766086732655e-06, + "loss": 0.4294, + "step": 24632 + }, + { + "epoch": 1.6531995570618436, + "grad_norm": 0.9893507957458496, + "learning_rate": 7.675978584965482e-06, + "loss": 0.4502, + "step": 24634 + }, + { + "epoch": 1.6533337807456125, + "grad_norm": 1.0750385522842407, + "learning_rate": 7.670193082925092e-06, + "loss": 0.474, + "step": 24636 + }, + { + "epoch": 1.6534680044293815, + "grad_norm": 0.9043234586715698, + "learning_rate": 7.66440958088484e-06, + "loss": 0.4602, + "step": 24638 + }, + { + "epoch": 1.6536022281131506, + "grad_norm": 1.0703434944152832, + "learning_rate": 7.658628079117974e-06, + "loss": 0.4676, + "step": 24640 + }, + { + "epoch": 1.6537364517969195, + "grad_norm": 0.8634180426597595, + "learning_rate": 7.652848577897681e-06, + "loss": 0.4362, + "step": 24642 + }, + { + "epoch": 1.6538706754806887, + "grad_norm": 1.051302433013916, + "learning_rate": 7.647071077497025e-06, + "loss": 0.5006, + "step": 24644 + }, + { + "epoch": 1.6540048991644576, + "grad_norm": 1.3736884593963623, + "learning_rate": 7.641295578188973e-06, + "loss": 0.5256, + "step": 24646 + }, + { + "epoch": 1.6541391228482265, + "grad_norm": 0.9008198380470276, + "learning_rate": 7.635522080246404e-06, + "loss": 0.4089, + "step": 24648 + }, + { + "epoch": 1.6542733465319954, + "grad_norm": 0.9291314482688904, + "learning_rate": 7.62975058394212e-06, + "loss": 0.4096, + "step": 24650 + }, + { + "epoch": 1.6544075702157646, + "grad_norm": 1.099732756614685, + "learning_rate": 7.623981089548798e-06, + "loss": 0.4814, + "step": 24652 + }, + { + "epoch": 1.6545417938995337, + "grad_norm": 1.053756833076477, + "learning_rate": 7.618213597339046e-06, + "loss": 0.5251, + "step": 24654 + }, + { + "epoch": 1.6546760175833026, + "grad_norm": 1.060849905014038, + "learning_rate": 7.612448107585363e-06, + "loss": 0.4829, + "step": 24656 + }, + { + "epoch": 1.6548102412670715, + "grad_norm": 1.047171950340271, + "learning_rate": 7.6066846205601396e-06, + "loss": 0.4898, + "step": 24658 + }, + { + "epoch": 1.6549444649508405, + "grad_norm": 0.9501131772994995, + "learning_rate": 7.600923136535715e-06, + "loss": 0.5317, + "step": 24660 + }, + { + "epoch": 1.6550786886346096, + "grad_norm": 1.0560083389282227, + "learning_rate": 7.595163655784293e-06, + "loss": 0.5048, + "step": 24662 + }, + { + "epoch": 1.6552129123183785, + "grad_norm": 1.015169620513916, + "learning_rate": 7.589406178578018e-06, + "loss": 0.4694, + "step": 24664 + }, + { + "epoch": 1.6553471360021477, + "grad_norm": 0.9952278137207031, + "learning_rate": 7.5836507051888885e-06, + "loss": 0.445, + "step": 24666 + }, + { + "epoch": 1.6554813596859166, + "grad_norm": 0.9039380550384521, + "learning_rate": 7.577897235888865e-06, + "loss": 0.4914, + "step": 24668 + }, + { + "epoch": 1.6556155833696855, + "grad_norm": 1.1089794635772705, + "learning_rate": 7.5721457709497625e-06, + "loss": 0.6241, + "step": 24670 + }, + { + "epoch": 1.6557498070534544, + "grad_norm": 1.0138074159622192, + "learning_rate": 7.566396310643353e-06, + "loss": 0.4944, + "step": 24672 + }, + { + "epoch": 1.6558840307372236, + "grad_norm": 1.1758661270141602, + "learning_rate": 7.560648855241276e-06, + "loss": 0.4968, + "step": 24674 + }, + { + "epoch": 1.6560182544209927, + "grad_norm": 0.9619575142860413, + "learning_rate": 7.5549034050150906e-06, + "loss": 0.4951, + "step": 24676 + }, + { + "epoch": 1.6561524781047616, + "grad_norm": 0.926916778087616, + "learning_rate": 7.549159960236241e-06, + "loss": 0.437, + "step": 24678 + }, + { + "epoch": 1.6562867017885305, + "grad_norm": 0.969592273235321, + "learning_rate": 7.543418521176121e-06, + "loss": 0.4218, + "step": 24680 + }, + { + "epoch": 1.6564209254722995, + "grad_norm": 1.0959103107452393, + "learning_rate": 7.537679088105987e-06, + "loss": 0.4791, + "step": 24682 + }, + { + "epoch": 1.6565551491560686, + "grad_norm": 1.1442915201187134, + "learning_rate": 7.531941661297015e-06, + "loss": 0.546, + "step": 24684 + }, + { + "epoch": 1.6566893728398377, + "grad_norm": 1.033138394355774, + "learning_rate": 7.526206241020278e-06, + "loss": 0.478, + "step": 24686 + }, + { + "epoch": 1.6568235965236067, + "grad_norm": 1.0051465034484863, + "learning_rate": 7.520472827546793e-06, + "loss": 0.4685, + "step": 24688 + }, + { + "epoch": 1.6569578202073756, + "grad_norm": 1.0380859375, + "learning_rate": 7.514741421147415e-06, + "loss": 0.4148, + "step": 24690 + }, + { + "epoch": 1.6570920438911445, + "grad_norm": 1.1546108722686768, + "learning_rate": 7.509012022092993e-06, + "loss": 0.5254, + "step": 24692 + }, + { + "epoch": 1.6572262675749136, + "grad_norm": 1.1542062759399414, + "learning_rate": 7.503284630654167e-06, + "loss": 0.5089, + "step": 24694 + }, + { + "epoch": 1.6573604912586826, + "grad_norm": 1.0694364309310913, + "learning_rate": 7.497559247101593e-06, + "loss": 0.4592, + "step": 24696 + }, + { + "epoch": 1.6574947149424517, + "grad_norm": 0.8872901201248169, + "learning_rate": 7.491835871705749e-06, + "loss": 0.4382, + "step": 24698 + }, + { + "epoch": 1.6576289386262206, + "grad_norm": 1.0182710886001587, + "learning_rate": 7.486114504737091e-06, + "loss": 0.5335, + "step": 24700 + }, + { + "epoch": 1.6577631623099895, + "grad_norm": 1.0655882358551025, + "learning_rate": 7.4803951464659185e-06, + "loss": 0.4844, + "step": 24702 + }, + { + "epoch": 1.6578973859937585, + "grad_norm": 1.042083740234375, + "learning_rate": 7.4746777971624605e-06, + "loss": 0.4664, + "step": 24704 + }, + { + "epoch": 1.6580316096775276, + "grad_norm": 0.9128807187080383, + "learning_rate": 7.468962457096845e-06, + "loss": 0.4326, + "step": 24706 + }, + { + "epoch": 1.6581658333612967, + "grad_norm": 0.8535100817680359, + "learning_rate": 7.4632491265391344e-06, + "loss": 0.4522, + "step": 24708 + }, + { + "epoch": 1.6583000570450657, + "grad_norm": 0.9464491605758667, + "learning_rate": 7.4575378057592574e-06, + "loss": 0.4201, + "step": 24710 + }, + { + "epoch": 1.6584342807288346, + "grad_norm": 1.014621615409851, + "learning_rate": 7.451828495027058e-06, + "loss": 0.472, + "step": 24712 + }, + { + "epoch": 1.6585685044126035, + "grad_norm": 0.9806382656097412, + "learning_rate": 7.446121194612288e-06, + "loss": 0.4546, + "step": 24714 + }, + { + "epoch": 1.6587027280963726, + "grad_norm": 0.9317238330841064, + "learning_rate": 7.440415904784625e-06, + "loss": 0.5104, + "step": 24716 + }, + { + "epoch": 1.6588369517801416, + "grad_norm": 1.0521879196166992, + "learning_rate": 7.434712625813605e-06, + "loss": 0.5019, + "step": 24718 + }, + { + "epoch": 1.6589711754639107, + "grad_norm": 1.0747883319854736, + "learning_rate": 7.4290113579687405e-06, + "loss": 0.4667, + "step": 24720 + }, + { + "epoch": 1.6591053991476796, + "grad_norm": 1.1273607015609741, + "learning_rate": 7.4233121015193586e-06, + "loss": 0.4902, + "step": 24722 + }, + { + "epoch": 1.6592396228314485, + "grad_norm": 0.9528072476387024, + "learning_rate": 7.417614856734751e-06, + "loss": 0.4835, + "step": 24724 + }, + { + "epoch": 1.6593738465152175, + "grad_norm": 1.1074789762496948, + "learning_rate": 7.411919623884117e-06, + "loss": 0.4096, + "step": 24726 + }, + { + "epoch": 1.6595080701989866, + "grad_norm": 0.948462188243866, + "learning_rate": 7.406226403236522e-06, + "loss": 0.4344, + "step": 24728 + }, + { + "epoch": 1.6596422938827557, + "grad_norm": 0.9927400350570679, + "learning_rate": 7.400535195060998e-06, + "loss": 0.4515, + "step": 24730 + }, + { + "epoch": 1.6597765175665247, + "grad_norm": 0.9371539354324341, + "learning_rate": 7.394845999626393e-06, + "loss": 0.3937, + "step": 24732 + }, + { + "epoch": 1.6599107412502936, + "grad_norm": 1.1667990684509277, + "learning_rate": 7.389158817201542e-06, + "loss": 0.5165, + "step": 24734 + }, + { + "epoch": 1.6600449649340625, + "grad_norm": 1.1483110189437866, + "learning_rate": 7.383473648055139e-06, + "loss": 0.424, + "step": 24736 + }, + { + "epoch": 1.6601791886178316, + "grad_norm": 1.0563877820968628, + "learning_rate": 7.377790492455816e-06, + "loss": 0.5473, + "step": 24738 + }, + { + "epoch": 1.6603134123016006, + "grad_norm": 1.0585153102874756, + "learning_rate": 7.3721093506720795e-06, + "loss": 0.42, + "step": 24740 + }, + { + "epoch": 1.6604476359853697, + "grad_norm": 0.9234572052955627, + "learning_rate": 7.366430222972353e-06, + "loss": 0.4157, + "step": 24742 + }, + { + "epoch": 1.6605818596691386, + "grad_norm": 0.918400764465332, + "learning_rate": 7.360753109624952e-06, + "loss": 0.4457, + "step": 24744 + }, + { + "epoch": 1.6607160833529075, + "grad_norm": 0.9188838005065918, + "learning_rate": 7.355078010898131e-06, + "loss": 0.5161, + "step": 24746 + }, + { + "epoch": 1.6608503070366765, + "grad_norm": 1.0078390836715698, + "learning_rate": 7.349404927060011e-06, + "loss": 0.4592, + "step": 24748 + }, + { + "epoch": 1.6609845307204456, + "grad_norm": 1.0019540786743164, + "learning_rate": 7.343733858378643e-06, + "loss": 0.4762, + "step": 24750 + }, + { + "epoch": 1.6611187544042147, + "grad_norm": 0.9697957634925842, + "learning_rate": 7.338064805121964e-06, + "loss": 0.4374, + "step": 24752 + }, + { + "epoch": 1.6612529780879837, + "grad_norm": 1.0653742551803589, + "learning_rate": 7.332397767557836e-06, + "loss": 0.4935, + "step": 24754 + }, + { + "epoch": 1.6613872017717526, + "grad_norm": 0.9930585622787476, + "learning_rate": 7.3267327459540015e-06, + "loss": 0.4151, + "step": 24756 + }, + { + "epoch": 1.6615214254555215, + "grad_norm": 1.0316599607467651, + "learning_rate": 7.321069740578157e-06, + "loss": 0.4034, + "step": 24758 + }, + { + "epoch": 1.6616556491392906, + "grad_norm": 0.927063524723053, + "learning_rate": 7.315408751697817e-06, + "loss": 0.4999, + "step": 24760 + }, + { + "epoch": 1.6617898728230598, + "grad_norm": 1.085348129272461, + "learning_rate": 7.309749779580494e-06, + "loss": 0.5474, + "step": 24762 + }, + { + "epoch": 1.6619240965068287, + "grad_norm": 0.9205760955810547, + "learning_rate": 7.30409282449353e-06, + "loss": 0.4622, + "step": 24764 + }, + { + "epoch": 1.6620583201905976, + "grad_norm": 0.935708224773407, + "learning_rate": 7.298437886704235e-06, + "loss": 0.4502, + "step": 24766 + }, + { + "epoch": 1.6621925438743665, + "grad_norm": 0.9033759832382202, + "learning_rate": 7.292784966479782e-06, + "loss": 0.4189, + "step": 24768 + }, + { + "epoch": 1.6623267675581357, + "grad_norm": 0.9785996079444885, + "learning_rate": 7.287134064087259e-06, + "loss": 0.4686, + "step": 24770 + }, + { + "epoch": 1.6624609912419046, + "grad_norm": 1.0184147357940674, + "learning_rate": 7.2814851797936455e-06, + "loss": 0.4644, + "step": 24772 + }, + { + "epoch": 1.6625952149256737, + "grad_norm": 0.9698885679244995, + "learning_rate": 7.275838313865874e-06, + "loss": 0.472, + "step": 24774 + }, + { + "epoch": 1.6627294386094427, + "grad_norm": 0.9263590574264526, + "learning_rate": 7.270193466570724e-06, + "loss": 0.4197, + "step": 24776 + }, + { + "epoch": 1.6628636622932116, + "grad_norm": 1.024530053138733, + "learning_rate": 7.264550638174905e-06, + "loss": 0.4021, + "step": 24778 + }, + { + "epoch": 1.6629978859769805, + "grad_norm": 0.9894106388092041, + "learning_rate": 7.258909828945027e-06, + "loss": 0.4817, + "step": 24780 + }, + { + "epoch": 1.6631321096607496, + "grad_norm": 0.9556596279144287, + "learning_rate": 7.2532710391476185e-06, + "loss": 0.4675, + "step": 24782 + }, + { + "epoch": 1.6632663333445188, + "grad_norm": 0.9056533575057983, + "learning_rate": 7.247634269049092e-06, + "loss": 0.4585, + "step": 24784 + }, + { + "epoch": 1.6634005570282877, + "grad_norm": 1.0582910776138306, + "learning_rate": 7.241999518915793e-06, + "loss": 0.4851, + "step": 24786 + }, + { + "epoch": 1.6635347807120566, + "grad_norm": 1.0909756422042847, + "learning_rate": 7.2363667890139265e-06, + "loss": 0.4517, + "step": 24788 + }, + { + "epoch": 1.6636690043958255, + "grad_norm": 1.0948517322540283, + "learning_rate": 7.230736079609629e-06, + "loss": 0.4894, + "step": 24790 + }, + { + "epoch": 1.6638032280795947, + "grad_norm": 0.9200692772865295, + "learning_rate": 7.22510739096896e-06, + "loss": 0.5111, + "step": 24792 + }, + { + "epoch": 1.6639374517633636, + "grad_norm": 1.0163975954055786, + "learning_rate": 7.219480723357841e-06, + "loss": 0.4542, + "step": 24794 + }, + { + "epoch": 1.6640716754471327, + "grad_norm": 1.4413422346115112, + "learning_rate": 7.213856077042147e-06, + "loss": 0.4463, + "step": 24796 + }, + { + "epoch": 1.6642058991309017, + "grad_norm": 1.1777487993240356, + "learning_rate": 7.208233452287616e-06, + "loss": 0.4462, + "step": 24798 + }, + { + "epoch": 1.6643401228146706, + "grad_norm": 1.193180799484253, + "learning_rate": 7.202612849359913e-06, + "loss": 0.4186, + "step": 24800 + }, + { + "epoch": 1.6644743464984395, + "grad_norm": 1.006433129310608, + "learning_rate": 7.196994268524576e-06, + "loss": 0.4139, + "step": 24802 + }, + { + "epoch": 1.6646085701822086, + "grad_norm": 0.9738898277282715, + "learning_rate": 7.1913777100471096e-06, + "loss": 0.4432, + "step": 24804 + }, + { + "epoch": 1.6647427938659778, + "grad_norm": 1.082888126373291, + "learning_rate": 7.185763174192861e-06, + "loss": 0.4252, + "step": 24806 + }, + { + "epoch": 1.6648770175497467, + "grad_norm": 1.2457809448242188, + "learning_rate": 7.18015066122712e-06, + "loss": 0.5175, + "step": 24808 + }, + { + "epoch": 1.6650112412335156, + "grad_norm": 0.948826014995575, + "learning_rate": 7.174540171415039e-06, + "loss": 0.4626, + "step": 24810 + }, + { + "epoch": 1.6651454649172845, + "grad_norm": 0.8969342112541199, + "learning_rate": 7.168931705021737e-06, + "loss": 0.4844, + "step": 24812 + }, + { + "epoch": 1.6652796886010537, + "grad_norm": 0.9707522988319397, + "learning_rate": 7.163325262312176e-06, + "loss": 0.4855, + "step": 24814 + }, + { + "epoch": 1.6654139122848226, + "grad_norm": 0.9882151484489441, + "learning_rate": 7.157720843551286e-06, + "loss": 0.4367, + "step": 24816 + }, + { + "epoch": 1.6655481359685917, + "grad_norm": 1.0631093978881836, + "learning_rate": 7.152118449003819e-06, + "loss": 0.478, + "step": 24818 + }, + { + "epoch": 1.6656823596523607, + "grad_norm": 1.2123793363571167, + "learning_rate": 7.146518078934505e-06, + "loss": 0.527, + "step": 24820 + }, + { + "epoch": 1.6658165833361296, + "grad_norm": 0.9780756831169128, + "learning_rate": 7.140919733607937e-06, + "loss": 0.3846, + "step": 24822 + }, + { + "epoch": 1.6659508070198985, + "grad_norm": 1.3662919998168945, + "learning_rate": 7.135323413288641e-06, + "loss": 0.4805, + "step": 24824 + }, + { + "epoch": 1.6660850307036676, + "grad_norm": 0.8742290139198303, + "learning_rate": 7.1297291182410246e-06, + "loss": 0.4277, + "step": 24826 + }, + { + "epoch": 1.6662192543874368, + "grad_norm": 1.0050386190414429, + "learning_rate": 7.124136848729407e-06, + "loss": 0.476, + "step": 24828 + }, + { + "epoch": 1.6663534780712057, + "grad_norm": 1.0886958837509155, + "learning_rate": 7.118546605018e-06, + "loss": 0.4916, + "step": 24830 + }, + { + "epoch": 1.6664877017549746, + "grad_norm": 1.1330604553222656, + "learning_rate": 7.11295838737096e-06, + "loss": 0.4312, + "step": 24832 + }, + { + "epoch": 1.6666219254387435, + "grad_norm": 1.0930204391479492, + "learning_rate": 7.107372196052298e-06, + "loss": 0.5102, + "step": 24834 + }, + { + "epoch": 1.6667561491225127, + "grad_norm": 1.0614120960235596, + "learning_rate": 7.101788031325951e-06, + "loss": 0.4445, + "step": 24836 + }, + { + "epoch": 1.6668903728062818, + "grad_norm": 0.9910305738449097, + "learning_rate": 7.09620589345576e-06, + "loss": 0.4942, + "step": 24838 + }, + { + "epoch": 1.6670245964900507, + "grad_norm": 0.9669861197471619, + "learning_rate": 7.0906257827054786e-06, + "loss": 0.4726, + "step": 24840 + }, + { + "epoch": 1.6671588201738197, + "grad_norm": 1.4137396812438965, + "learning_rate": 7.085047699338742e-06, + "loss": 0.5131, + "step": 24842 + }, + { + "epoch": 1.6672930438575886, + "grad_norm": 0.9933016300201416, + "learning_rate": 7.079471643619135e-06, + "loss": 0.4637, + "step": 24844 + }, + { + "epoch": 1.6674272675413577, + "grad_norm": 1.0396409034729004, + "learning_rate": 7.0738976158100715e-06, + "loss": 0.5158, + "step": 24846 + }, + { + "epoch": 1.6675614912251266, + "grad_norm": 1.098038673400879, + "learning_rate": 7.068325616174948e-06, + "loss": 0.5014, + "step": 24848 + }, + { + "epoch": 1.6676957149088958, + "grad_norm": 1.0721778869628906, + "learning_rate": 7.062755644977004e-06, + "loss": 0.4014, + "step": 24850 + }, + { + "epoch": 1.6678299385926647, + "grad_norm": 1.0025501251220703, + "learning_rate": 7.05718770247944e-06, + "loss": 0.4526, + "step": 24852 + }, + { + "epoch": 1.6679641622764336, + "grad_norm": 1.0482416152954102, + "learning_rate": 7.051621788945306e-06, + "loss": 0.4261, + "step": 24854 + }, + { + "epoch": 1.6680983859602025, + "grad_norm": 1.5061668157577515, + "learning_rate": 7.046057904637593e-06, + "loss": 0.4502, + "step": 24856 + }, + { + "epoch": 1.6682326096439717, + "grad_norm": 1.000950813293457, + "learning_rate": 7.040496049819178e-06, + "loss": 0.4135, + "step": 24858 + }, + { + "epoch": 1.6683668333277408, + "grad_norm": 1.0769424438476562, + "learning_rate": 7.034936224752836e-06, + "loss": 0.5046, + "step": 24860 + }, + { + "epoch": 1.6685010570115097, + "grad_norm": 1.056839942932129, + "learning_rate": 7.029378429701278e-06, + "loss": 0.4646, + "step": 24862 + }, + { + "epoch": 1.6686352806952787, + "grad_norm": 0.93470698595047, + "learning_rate": 7.023822664927099e-06, + "loss": 0.5146, + "step": 24864 + }, + { + "epoch": 1.6687695043790476, + "grad_norm": 1.0354546308517456, + "learning_rate": 7.018268930692784e-06, + "loss": 0.4397, + "step": 24866 + }, + { + "epoch": 1.6689037280628167, + "grad_norm": 0.8594359159469604, + "learning_rate": 7.012717227260734e-06, + "loss": 0.5207, + "step": 24868 + }, + { + "epoch": 1.6690379517465856, + "grad_norm": 1.2969167232513428, + "learning_rate": 7.007167554893274e-06, + "loss": 0.4622, + "step": 24870 + }, + { + "epoch": 1.6691721754303548, + "grad_norm": 1.0026137828826904, + "learning_rate": 7.001619913852603e-06, + "loss": 0.4298, + "step": 24872 + }, + { + "epoch": 1.6693063991141237, + "grad_norm": 1.0497329235076904, + "learning_rate": 6.99607430440084e-06, + "loss": 0.4532, + "step": 24874 + }, + { + "epoch": 1.6694406227978926, + "grad_norm": 1.0968456268310547, + "learning_rate": 6.990530726799993e-06, + "loss": 0.4998, + "step": 24876 + }, + { + "epoch": 1.6695748464816615, + "grad_norm": 1.0032005310058594, + "learning_rate": 6.984989181312007e-06, + "loss": 0.5401, + "step": 24878 + }, + { + "epoch": 1.6697090701654307, + "grad_norm": 1.1241943836212158, + "learning_rate": 6.979449668198679e-06, + "loss": 0.486, + "step": 24880 + }, + { + "epoch": 1.6698432938491998, + "grad_norm": 1.1395469903945923, + "learning_rate": 6.973912187721787e-06, + "loss": 0.4375, + "step": 24882 + }, + { + "epoch": 1.6699775175329687, + "grad_norm": 1.0104833841323853, + "learning_rate": 6.968376740142912e-06, + "loss": 0.4592, + "step": 24884 + }, + { + "epoch": 1.6701117412167377, + "grad_norm": 1.0128306150436401, + "learning_rate": 6.962843325723628e-06, + "loss": 0.492, + "step": 24886 + }, + { + "epoch": 1.6702459649005066, + "grad_norm": 1.1043764352798462, + "learning_rate": 6.957311944725359e-06, + "loss": 0.4791, + "step": 24888 + }, + { + "epoch": 1.6703801885842757, + "grad_norm": 0.9430697560310364, + "learning_rate": 6.951782597409473e-06, + "loss": 0.3964, + "step": 24890 + }, + { + "epoch": 1.6705144122680446, + "grad_norm": 1.0144459009170532, + "learning_rate": 6.946255284037207e-06, + "loss": 0.407, + "step": 24892 + }, + { + "epoch": 1.6706486359518138, + "grad_norm": 1.1235939264297485, + "learning_rate": 6.940730004869717e-06, + "loss": 0.4974, + "step": 24894 + }, + { + "epoch": 1.6707828596355827, + "grad_norm": 0.9710730910301208, + "learning_rate": 6.9352067601680506e-06, + "loss": 0.4397, + "step": 24896 + }, + { + "epoch": 1.6709170833193516, + "grad_norm": 0.9254615902900696, + "learning_rate": 6.929685550193193e-06, + "loss": 0.4578, + "step": 24898 + }, + { + "epoch": 1.6710513070031205, + "grad_norm": 1.1053893566131592, + "learning_rate": 6.924166375205998e-06, + "loss": 0.4728, + "step": 24900 + }, + { + "epoch": 1.6711855306868897, + "grad_norm": 0.9544780850410461, + "learning_rate": 6.918649235467234e-06, + "loss": 0.5049, + "step": 24902 + }, + { + "epoch": 1.6713197543706588, + "grad_norm": 0.983664333820343, + "learning_rate": 6.913134131237575e-06, + "loss": 0.4404, + "step": 24904 + }, + { + "epoch": 1.6714539780544277, + "grad_norm": 1.0222305059432983, + "learning_rate": 6.9076210627776085e-06, + "loss": 0.4881, + "step": 24906 + }, + { + "epoch": 1.6715882017381967, + "grad_norm": 1.106665015220642, + "learning_rate": 6.902110030347797e-06, + "loss": 0.5302, + "step": 24908 + }, + { + "epoch": 1.6717224254219656, + "grad_norm": 0.9527423977851868, + "learning_rate": 6.896601034208561e-06, + "loss": 0.4498, + "step": 24910 + }, + { + "epoch": 1.6718566491057347, + "grad_norm": 1.1243098974227905, + "learning_rate": 6.891094074620141e-06, + "loss": 0.4221, + "step": 24912 + }, + { + "epoch": 1.6719908727895039, + "grad_norm": 1.0397752523422241, + "learning_rate": 6.88558915184277e-06, + "loss": 0.4972, + "step": 24914 + }, + { + "epoch": 1.6721250964732728, + "grad_norm": 0.9265273809432983, + "learning_rate": 6.880086266136521e-06, + "loss": 0.4661, + "step": 24916 + }, + { + "epoch": 1.6722593201570417, + "grad_norm": 1.0777788162231445, + "learning_rate": 6.874585417761414e-06, + "loss": 0.452, + "step": 24918 + }, + { + "epoch": 1.6723935438408106, + "grad_norm": 1.089951992034912, + "learning_rate": 6.869086606977349e-06, + "loss": 0.4876, + "step": 24920 + }, + { + "epoch": 1.6725277675245798, + "grad_norm": 0.9507853984832764, + "learning_rate": 6.863589834044121e-06, + "loss": 0.4291, + "step": 24922 + }, + { + "epoch": 1.6726619912083487, + "grad_norm": 1.1321830749511719, + "learning_rate": 6.8580950992214395e-06, + "loss": 0.4382, + "step": 24924 + }, + { + "epoch": 1.6727962148921178, + "grad_norm": 0.9954093098640442, + "learning_rate": 6.852602402768943e-06, + "loss": 0.4493, + "step": 24926 + }, + { + "epoch": 1.6729304385758867, + "grad_norm": 1.0333951711654663, + "learning_rate": 6.847111744946139e-06, + "loss": 0.5084, + "step": 24928 + }, + { + "epoch": 1.6730646622596557, + "grad_norm": 0.9585345387458801, + "learning_rate": 6.841623126012442e-06, + "loss": 0.4525, + "step": 24930 + }, + { + "epoch": 1.6731988859434246, + "grad_norm": 1.0446515083312988, + "learning_rate": 6.83613654622719e-06, + "loss": 0.5496, + "step": 24932 + }, + { + "epoch": 1.6733331096271937, + "grad_norm": 1.0466399192810059, + "learning_rate": 6.830652005849597e-06, + "loss": 0.5162, + "step": 24934 + }, + { + "epoch": 1.6734673333109629, + "grad_norm": 0.9950196743011475, + "learning_rate": 6.825169505138818e-06, + "loss": 0.4763, + "step": 24936 + }, + { + "epoch": 1.6736015569947318, + "grad_norm": 0.9856505393981934, + "learning_rate": 6.819689044353877e-06, + "loss": 0.5168, + "step": 24938 + }, + { + "epoch": 1.6737357806785007, + "grad_norm": 1.1162148714065552, + "learning_rate": 6.8142106237537255e-06, + "loss": 0.4407, + "step": 24940 + }, + { + "epoch": 1.6738700043622696, + "grad_norm": 1.1244986057281494, + "learning_rate": 6.8087342435971815e-06, + "loss": 0.4785, + "step": 24942 + }, + { + "epoch": 1.6740042280460388, + "grad_norm": 0.8998849391937256, + "learning_rate": 6.803259904143027e-06, + "loss": 0.4196, + "step": 24944 + }, + { + "epoch": 1.6741384517298077, + "grad_norm": 1.0995714664459229, + "learning_rate": 6.797787605649891e-06, + "loss": 0.5114, + "step": 24946 + }, + { + "epoch": 1.6742726754135768, + "grad_norm": 0.9137680530548096, + "learning_rate": 6.792317348376348e-06, + "loss": 0.422, + "step": 24948 + }, + { + "epoch": 1.6744068990973457, + "grad_norm": 0.9242937564849854, + "learning_rate": 6.786849132580842e-06, + "loss": 0.521, + "step": 24950 + }, + { + "epoch": 1.6745411227811147, + "grad_norm": 0.8841946125030518, + "learning_rate": 6.781382958521743e-06, + "loss": 0.4557, + "step": 24952 + }, + { + "epoch": 1.6746753464648836, + "grad_norm": 0.9266219139099121, + "learning_rate": 6.7759188264573005e-06, + "loss": 0.4404, + "step": 24954 + }, + { + "epoch": 1.6748095701486527, + "grad_norm": 0.9643049836158752, + "learning_rate": 6.770456736645708e-06, + "loss": 0.456, + "step": 24956 + }, + { + "epoch": 1.6749437938324219, + "grad_norm": 1.0734564065933228, + "learning_rate": 6.764996689345027e-06, + "loss": 0.4664, + "step": 24958 + }, + { + "epoch": 1.6750780175161908, + "grad_norm": 1.0203007459640503, + "learning_rate": 6.7595386848132325e-06, + "loss": 0.4589, + "step": 24960 + }, + { + "epoch": 1.6752122411999597, + "grad_norm": 1.0745856761932373, + "learning_rate": 6.754082723308203e-06, + "loss": 0.536, + "step": 24962 + }, + { + "epoch": 1.6753464648837286, + "grad_norm": 1.0680867433547974, + "learning_rate": 6.748628805087731e-06, + "loss": 0.5385, + "step": 24964 + }, + { + "epoch": 1.6754806885674978, + "grad_norm": 1.0987399816513062, + "learning_rate": 6.743176930409495e-06, + "loss": 0.5095, + "step": 24966 + }, + { + "epoch": 1.6756149122512667, + "grad_norm": 1.0225719213485718, + "learning_rate": 6.737727099531094e-06, + "loss": 0.4711, + "step": 24968 + }, + { + "epoch": 1.6757491359350358, + "grad_norm": 1.1248780488967896, + "learning_rate": 6.732279312709999e-06, + "loss": 0.4256, + "step": 24970 + }, + { + "epoch": 1.6758833596188047, + "grad_norm": 1.1062440872192383, + "learning_rate": 6.726833570203639e-06, + "loss": 0.4524, + "step": 24972 + }, + { + "epoch": 1.6760175833025737, + "grad_norm": 1.0149747133255005, + "learning_rate": 6.721389872269291e-06, + "loss": 0.4915, + "step": 24974 + }, + { + "epoch": 1.6761518069863426, + "grad_norm": 1.0408467054367065, + "learning_rate": 6.715948219164175e-06, + "loss": 0.4601, + "step": 24976 + }, + { + "epoch": 1.6762860306701117, + "grad_norm": 0.9350197911262512, + "learning_rate": 6.7105086111453944e-06, + "loss": 0.437, + "step": 24978 + }, + { + "epoch": 1.6764202543538809, + "grad_norm": 0.9445664286613464, + "learning_rate": 6.705071048469952e-06, + "loss": 0.4797, + "step": 24980 + }, + { + "epoch": 1.6765544780376498, + "grad_norm": 1.073475956916809, + "learning_rate": 6.69963553139476e-06, + "loss": 0.4725, + "step": 24982 + }, + { + "epoch": 1.6766887017214187, + "grad_norm": 1.0421797037124634, + "learning_rate": 6.694202060176652e-06, + "loss": 0.4289, + "step": 24984 + }, + { + "epoch": 1.6768229254051876, + "grad_norm": 2.2069389820098877, + "learning_rate": 6.688770635072339e-06, + "loss": 0.4576, + "step": 24986 + }, + { + "epoch": 1.6769571490889568, + "grad_norm": 0.9449515342712402, + "learning_rate": 6.683341256338444e-06, + "loss": 0.4131, + "step": 24988 + }, + { + "epoch": 1.677091372772726, + "grad_norm": 0.94171142578125, + "learning_rate": 6.677913924231488e-06, + "loss": 0.4726, + "step": 24990 + }, + { + "epoch": 1.6772255964564948, + "grad_norm": 1.0046991109848022, + "learning_rate": 6.672488639007918e-06, + "loss": 0.4717, + "step": 24992 + }, + { + "epoch": 1.6773598201402637, + "grad_norm": 0.9732492566108704, + "learning_rate": 6.667065400924066e-06, + "loss": 0.4163, + "step": 24994 + }, + { + "epoch": 1.6774940438240327, + "grad_norm": 0.9594755172729492, + "learning_rate": 6.661644210236156e-06, + "loss": 0.4575, + "step": 24996 + }, + { + "epoch": 1.6776282675078018, + "grad_norm": 1.0889017581939697, + "learning_rate": 6.656225067200339e-06, + "loss": 0.4944, + "step": 24998 + }, + { + "epoch": 1.6777624911915707, + "grad_norm": 1.1145284175872803, + "learning_rate": 6.650807972072648e-06, + "loss": 0.4343, + "step": 25000 + }, + { + "epoch": 1.6778967148753399, + "grad_norm": 1.0912227630615234, + "learning_rate": 6.645392925109045e-06, + "loss": 0.4931, + "step": 25002 + }, + { + "epoch": 1.6780309385591088, + "grad_norm": 1.1241024732589722, + "learning_rate": 6.639979926565359e-06, + "loss": 0.4578, + "step": 25004 + }, + { + "epoch": 1.6781651622428777, + "grad_norm": 1.0750898122787476, + "learning_rate": 6.634568976697386e-06, + "loss": 0.4653, + "step": 25006 + }, + { + "epoch": 1.6782993859266466, + "grad_norm": 1.055790901184082, + "learning_rate": 6.629160075760726e-06, + "loss": 0.5403, + "step": 25008 + }, + { + "epoch": 1.6784336096104158, + "grad_norm": 1.0737273693084717, + "learning_rate": 6.623753224010987e-06, + "loss": 0.5264, + "step": 25010 + }, + { + "epoch": 1.678567833294185, + "grad_norm": 1.0886459350585938, + "learning_rate": 6.618348421703596e-06, + "loss": 0.4444, + "step": 25012 + }, + { + "epoch": 1.6787020569779538, + "grad_norm": 0.9762519001960754, + "learning_rate": 6.6129456690939455e-06, + "loss": 0.4607, + "step": 25014 + }, + { + "epoch": 1.6788362806617227, + "grad_norm": 1.086861491203308, + "learning_rate": 6.607544966437296e-06, + "loss": 0.4689, + "step": 25016 + }, + { + "epoch": 1.6789705043454917, + "grad_norm": 1.0281087160110474, + "learning_rate": 6.602146313988822e-06, + "loss": 0.4209, + "step": 25018 + }, + { + "epoch": 1.6791047280292608, + "grad_norm": 0.918687641620636, + "learning_rate": 6.59674971200358e-06, + "loss": 0.4127, + "step": 25020 + }, + { + "epoch": 1.6792389517130297, + "grad_norm": 1.1233757734298706, + "learning_rate": 6.591355160736578e-06, + "loss": 0.4873, + "step": 25022 + }, + { + "epoch": 1.6793731753967989, + "grad_norm": 0.9721964001655579, + "learning_rate": 6.5859626604426905e-06, + "loss": 0.4487, + "step": 25024 + }, + { + "epoch": 1.6795073990805678, + "grad_norm": 1.129554033279419, + "learning_rate": 6.580572211376695e-06, + "loss": 0.5156, + "step": 25026 + }, + { + "epoch": 1.6796416227643367, + "grad_norm": 0.9984642863273621, + "learning_rate": 6.575183813793267e-06, + "loss": 0.4914, + "step": 25028 + }, + { + "epoch": 1.6797758464481056, + "grad_norm": 0.9877324104309082, + "learning_rate": 6.569797467947026e-06, + "loss": 0.4615, + "step": 25030 + }, + { + "epoch": 1.6799100701318748, + "grad_norm": 1.579860806465149, + "learning_rate": 6.564413174092443e-06, + "loss": 0.5255, + "step": 25032 + }, + { + "epoch": 1.680044293815644, + "grad_norm": 1.0377464294433594, + "learning_rate": 6.559030932483945e-06, + "loss": 0.5348, + "step": 25034 + }, + { + "epoch": 1.6801785174994128, + "grad_norm": 1.0144003629684448, + "learning_rate": 6.553650743375794e-06, + "loss": 0.4118, + "step": 25036 + }, + { + "epoch": 1.6803127411831817, + "grad_norm": 1.0023353099822998, + "learning_rate": 6.548272607022221e-06, + "loss": 0.4661, + "step": 25038 + }, + { + "epoch": 1.6804469648669507, + "grad_norm": 0.9578472971916199, + "learning_rate": 6.542896523677311e-06, + "loss": 0.5483, + "step": 25040 + }, + { + "epoch": 1.6805811885507198, + "grad_norm": 1.0725754499435425, + "learning_rate": 6.537522493595094e-06, + "loss": 0.511, + "step": 25042 + }, + { + "epoch": 1.6807154122344887, + "grad_norm": 1.1264251470565796, + "learning_rate": 6.532150517029478e-06, + "loss": 0.5275, + "step": 25044 + }, + { + "epoch": 1.6808496359182579, + "grad_norm": 1.1566630601882935, + "learning_rate": 6.526780594234272e-06, + "loss": 0.4925, + "step": 25046 + }, + { + "epoch": 1.6809838596020268, + "grad_norm": 1.072445034980774, + "learning_rate": 6.521412725463188e-06, + "loss": 0.4378, + "step": 25048 + }, + { + "epoch": 1.6811180832857957, + "grad_norm": 1.4630780220031738, + "learning_rate": 6.516046910969864e-06, + "loss": 0.4661, + "step": 25050 + }, + { + "epoch": 1.6812523069695646, + "grad_norm": 0.9673110842704773, + "learning_rate": 6.510683151007818e-06, + "loss": 0.4452, + "step": 25052 + }, + { + "epoch": 1.6813865306533338, + "grad_norm": 1.0803834199905396, + "learning_rate": 6.505321445830476e-06, + "loss": 0.4096, + "step": 25054 + }, + { + "epoch": 1.681520754337103, + "grad_norm": 1.049075961112976, + "learning_rate": 6.499961795691151e-06, + "loss": 0.4339, + "step": 25056 + }, + { + "epoch": 1.6816549780208718, + "grad_norm": 1.1449967622756958, + "learning_rate": 6.494604200843107e-06, + "loss": 0.4876, + "step": 25058 + }, + { + "epoch": 1.6817892017046407, + "grad_norm": 0.913049042224884, + "learning_rate": 6.489248661539454e-06, + "loss": 0.5163, + "step": 25060 + }, + { + "epoch": 1.6819234253884097, + "grad_norm": 1.0686430931091309, + "learning_rate": 6.483895178033261e-06, + "loss": 0.5323, + "step": 25062 + }, + { + "epoch": 1.6820576490721788, + "grad_norm": 1.064252257347107, + "learning_rate": 6.478543750577442e-06, + "loss": 0.478, + "step": 25064 + }, + { + "epoch": 1.682191872755948, + "grad_norm": 0.95737624168396, + "learning_rate": 6.4731943794248355e-06, + "loss": 0.4617, + "step": 25066 + }, + { + "epoch": 1.6823260964397169, + "grad_norm": 1.0545599460601807, + "learning_rate": 6.467847064828214e-06, + "loss": 0.4684, + "step": 25068 + }, + { + "epoch": 1.6824603201234858, + "grad_norm": 1.0754735469818115, + "learning_rate": 6.462501807040211e-06, + "loss": 0.484, + "step": 25070 + }, + { + "epoch": 1.6825945438072547, + "grad_norm": 1.1124367713928223, + "learning_rate": 6.457158606313402e-06, + "loss": 0.439, + "step": 25072 + }, + { + "epoch": 1.6827287674910236, + "grad_norm": 1.129748821258545, + "learning_rate": 6.451817462900206e-06, + "loss": 0.4889, + "step": 25074 + }, + { + "epoch": 1.6828629911747928, + "grad_norm": 1.002609133720398, + "learning_rate": 6.4464783770530055e-06, + "loss": 0.4862, + "step": 25076 + }, + { + "epoch": 1.682997214858562, + "grad_norm": 1.1611278057098389, + "learning_rate": 6.441141349024055e-06, + "loss": 0.5409, + "step": 25078 + }, + { + "epoch": 1.6831314385423308, + "grad_norm": 1.128123164176941, + "learning_rate": 6.435806379065529e-06, + "loss": 0.5094, + "step": 25080 + }, + { + "epoch": 1.6832656622260997, + "grad_norm": 0.9699921607971191, + "learning_rate": 6.430473467429482e-06, + "loss": 0.4155, + "step": 25082 + }, + { + "epoch": 1.6833998859098687, + "grad_norm": 1.0676101446151733, + "learning_rate": 6.4251426143678904e-06, + "loss": 0.4512, + "step": 25084 + }, + { + "epoch": 1.6835341095936378, + "grad_norm": 0.9882549047470093, + "learning_rate": 6.4198138201326145e-06, + "loss": 0.4518, + "step": 25086 + }, + { + "epoch": 1.683668333277407, + "grad_norm": 0.9625896215438843, + "learning_rate": 6.414487084975451e-06, + "loss": 0.42, + "step": 25088 + }, + { + "epoch": 1.6838025569611759, + "grad_norm": 0.9984647631645203, + "learning_rate": 6.409162409148062e-06, + "loss": 0.5404, + "step": 25090 + }, + { + "epoch": 1.6839367806449448, + "grad_norm": 0.9463441371917725, + "learning_rate": 6.403839792902033e-06, + "loss": 0.4435, + "step": 25092 + }, + { + "epoch": 1.6840710043287137, + "grad_norm": 1.0103236436843872, + "learning_rate": 6.398519236488837e-06, + "loss": 0.4823, + "step": 25094 + }, + { + "epoch": 1.6842052280124828, + "grad_norm": 0.8966814875602722, + "learning_rate": 6.393200740159877e-06, + "loss": 0.3919, + "step": 25096 + }, + { + "epoch": 1.6843394516962518, + "grad_norm": 1.2721905708312988, + "learning_rate": 6.387884304166419e-06, + "loss": 0.5342, + "step": 25098 + }, + { + "epoch": 1.684473675380021, + "grad_norm": 1.1116913557052612, + "learning_rate": 6.382569928759685e-06, + "loss": 0.4427, + "step": 25100 + }, + { + "epoch": 1.6846078990637898, + "grad_norm": 1.022383213043213, + "learning_rate": 6.3772576141907456e-06, + "loss": 0.5117, + "step": 25102 + }, + { + "epoch": 1.6847421227475587, + "grad_norm": 1.0255234241485596, + "learning_rate": 6.371947360710606e-06, + "loss": 0.4271, + "step": 25104 + }, + { + "epoch": 1.6848763464313277, + "grad_norm": 1.0771558284759521, + "learning_rate": 6.366639168570154e-06, + "loss": 0.4597, + "step": 25106 + }, + { + "epoch": 1.6850105701150968, + "grad_norm": 0.9802014827728271, + "learning_rate": 6.361333038020212e-06, + "loss": 0.4234, + "step": 25108 + }, + { + "epoch": 1.685144793798866, + "grad_norm": 1.0145399570465088, + "learning_rate": 6.356028969311467e-06, + "loss": 0.5103, + "step": 25110 + }, + { + "epoch": 1.6852790174826349, + "grad_norm": 1.097533941268921, + "learning_rate": 6.350726962694537e-06, + "loss": 0.4937, + "step": 25112 + }, + { + "epoch": 1.6854132411664038, + "grad_norm": 1.1500582695007324, + "learning_rate": 6.345427018419908e-06, + "loss": 0.4659, + "step": 25114 + }, + { + "epoch": 1.6855474648501727, + "grad_norm": 0.9849839806556702, + "learning_rate": 6.340129136738027e-06, + "loss": 0.4085, + "step": 25116 + }, + { + "epoch": 1.6856816885339418, + "grad_norm": 0.755287230014801, + "learning_rate": 6.334833317899186e-06, + "loss": 0.4087, + "step": 25118 + }, + { + "epoch": 1.6858159122177108, + "grad_norm": 1.116926670074463, + "learning_rate": 6.329539562153608e-06, + "loss": 0.4955, + "step": 25120 + }, + { + "epoch": 1.68595013590148, + "grad_norm": 0.950871467590332, + "learning_rate": 6.324247869751398e-06, + "loss": 0.4735, + "step": 25122 + }, + { + "epoch": 1.6860843595852488, + "grad_norm": 0.9359391927719116, + "learning_rate": 6.318958240942607e-06, + "loss": 0.4395, + "step": 25124 + }, + { + "epoch": 1.6862185832690177, + "grad_norm": 1.0384999513626099, + "learning_rate": 6.313670675977129e-06, + "loss": 0.4541, + "step": 25126 + }, + { + "epoch": 1.6863528069527867, + "grad_norm": 1.1261101961135864, + "learning_rate": 6.308385175104819e-06, + "loss": 0.4507, + "step": 25128 + }, + { + "epoch": 1.6864870306365558, + "grad_norm": 1.0034666061401367, + "learning_rate": 6.303101738575395e-06, + "loss": 0.4504, + "step": 25130 + }, + { + "epoch": 1.686621254320325, + "grad_norm": 1.0501008033752441, + "learning_rate": 6.297820366638485e-06, + "loss": 0.4718, + "step": 25132 + }, + { + "epoch": 1.6867554780040939, + "grad_norm": 1.0580476522445679, + "learning_rate": 6.292541059543628e-06, + "loss": 0.4951, + "step": 25134 + }, + { + "epoch": 1.6868897016878628, + "grad_norm": 0.9894029498100281, + "learning_rate": 6.2872638175402455e-06, + "loss": 0.4317, + "step": 25136 + }, + { + "epoch": 1.6870239253716317, + "grad_norm": 1.027328610420227, + "learning_rate": 6.281988640877706e-06, + "loss": 0.499, + "step": 25138 + }, + { + "epoch": 1.6871581490554008, + "grad_norm": 1.1530015468597412, + "learning_rate": 6.276715529805233e-06, + "loss": 0.4692, + "step": 25140 + }, + { + "epoch": 1.68729237273917, + "grad_norm": 0.9684929251670837, + "learning_rate": 6.27144448457197e-06, + "loss": 0.4657, + "step": 25142 + }, + { + "epoch": 1.687426596422939, + "grad_norm": 0.9814417958259583, + "learning_rate": 6.266175505426958e-06, + "loss": 0.5096, + "step": 25144 + }, + { + "epoch": 1.6875608201067078, + "grad_norm": 1.3349909782409668, + "learning_rate": 6.260908592619169e-06, + "loss": 0.4892, + "step": 25146 + }, + { + "epoch": 1.6876950437904767, + "grad_norm": 0.9152083992958069, + "learning_rate": 6.255643746397433e-06, + "loss": 0.455, + "step": 25148 + }, + { + "epoch": 1.6878292674742457, + "grad_norm": 0.9364927411079407, + "learning_rate": 6.250380967010516e-06, + "loss": 0.4415, + "step": 25150 + }, + { + "epoch": 1.6879634911580148, + "grad_norm": 1.0783220529556274, + "learning_rate": 6.2451202547070545e-06, + "loss": 0.4086, + "step": 25152 + }, + { + "epoch": 1.688097714841784, + "grad_norm": 1.1346849203109741, + "learning_rate": 6.239861609735631e-06, + "loss": 0.4136, + "step": 25154 + }, + { + "epoch": 1.6882319385255529, + "grad_norm": 1.0446324348449707, + "learning_rate": 6.2346050323446865e-06, + "loss": 0.5312, + "step": 25156 + }, + { + "epoch": 1.6883661622093218, + "grad_norm": 7.313377857208252, + "learning_rate": 6.229350522782612e-06, + "loss": 0.4643, + "step": 25158 + }, + { + "epoch": 1.6885003858930907, + "grad_norm": 1.1398403644561768, + "learning_rate": 6.224098081297636e-06, + "loss": 0.4778, + "step": 25160 + }, + { + "epoch": 1.6886346095768598, + "grad_norm": 1.1007752418518066, + "learning_rate": 6.218847708137954e-06, + "loss": 0.4406, + "step": 25162 + }, + { + "epoch": 1.688768833260629, + "grad_norm": 0.9332924485206604, + "learning_rate": 6.213599403551617e-06, + "loss": 0.434, + "step": 25164 + }, + { + "epoch": 1.688903056944398, + "grad_norm": 1.010879397392273, + "learning_rate": 6.208353167786612e-06, + "loss": 0.4313, + "step": 25166 + }, + { + "epoch": 1.6890372806281668, + "grad_norm": 0.9699853658676147, + "learning_rate": 6.203109001090812e-06, + "loss": 0.5114, + "step": 25168 + }, + { + "epoch": 1.6891715043119357, + "grad_norm": 0.9748437404632568, + "learning_rate": 6.197866903711985e-06, + "loss": 0.4725, + "step": 25170 + }, + { + "epoch": 1.6893057279957049, + "grad_norm": 0.8466166853904724, + "learning_rate": 6.192626875897806e-06, + "loss": 0.4495, + "step": 25172 + }, + { + "epoch": 1.6894399516794738, + "grad_norm": 1.0146995782852173, + "learning_rate": 6.18738891789587e-06, + "loss": 0.4434, + "step": 25174 + }, + { + "epoch": 1.689574175363243, + "grad_norm": 0.8807005286216736, + "learning_rate": 6.182153029953658e-06, + "loss": 0.4049, + "step": 25176 + }, + { + "epoch": 1.6897083990470119, + "grad_norm": 0.9264736771583557, + "learning_rate": 6.1769192123185495e-06, + "loss": 0.3903, + "step": 25178 + }, + { + "epoch": 1.6898426227307808, + "grad_norm": 1.0460312366485596, + "learning_rate": 6.171687465237824e-06, + "loss": 0.4415, + "step": 25180 + }, + { + "epoch": 1.6899768464145497, + "grad_norm": 0.9783272743225098, + "learning_rate": 6.1664577889586905e-06, + "loss": 0.4874, + "step": 25182 + }, + { + "epoch": 1.6901110700983188, + "grad_norm": 1.0166213512420654, + "learning_rate": 6.1612301837282225e-06, + "loss": 0.4665, + "step": 25184 + }, + { + "epoch": 1.690245293782088, + "grad_norm": 1.0550141334533691, + "learning_rate": 6.156004649793446e-06, + "loss": 0.4403, + "step": 25186 + }, + { + "epoch": 1.690379517465857, + "grad_norm": 1.0054047107696533, + "learning_rate": 6.150781187401211e-06, + "loss": 0.4583, + "step": 25188 + }, + { + "epoch": 1.6905137411496258, + "grad_norm": 0.877746045589447, + "learning_rate": 6.145559796798345e-06, + "loss": 0.4329, + "step": 25190 + }, + { + "epoch": 1.6906479648333947, + "grad_norm": 1.2742961645126343, + "learning_rate": 6.140340478231537e-06, + "loss": 0.5484, + "step": 25192 + }, + { + "epoch": 1.6907821885171639, + "grad_norm": 1.074097752571106, + "learning_rate": 6.135123231947404e-06, + "loss": 0.4559, + "step": 25194 + }, + { + "epoch": 1.6909164122009328, + "grad_norm": 1.1236523389816284, + "learning_rate": 6.1299080581924465e-06, + "loss": 0.5046, + "step": 25196 + }, + { + "epoch": 1.691050635884702, + "grad_norm": 1.113170862197876, + "learning_rate": 6.124694957213062e-06, + "loss": 0.4822, + "step": 25198 + }, + { + "epoch": 1.6911848595684709, + "grad_norm": 1.0551742315292358, + "learning_rate": 6.119483929255559e-06, + "loss": 0.4929, + "step": 25200 + }, + { + "epoch": 1.6913190832522398, + "grad_norm": 0.9076232314109802, + "learning_rate": 6.1142749745661536e-06, + "loss": 0.4541, + "step": 25202 + }, + { + "epoch": 1.6914533069360087, + "grad_norm": 1.0314046144485474, + "learning_rate": 6.109068093390957e-06, + "loss": 0.4655, + "step": 25204 + }, + { + "epoch": 1.6915875306197778, + "grad_norm": 1.0429438352584839, + "learning_rate": 6.103863285975992e-06, + "loss": 0.4468, + "step": 25206 + }, + { + "epoch": 1.691721754303547, + "grad_norm": 1.0242527723312378, + "learning_rate": 6.0986605525671705e-06, + "loss": 0.3975, + "step": 25208 + }, + { + "epoch": 1.691855977987316, + "grad_norm": 0.9596940875053406, + "learning_rate": 6.093459893410297e-06, + "loss": 0.523, + "step": 25210 + }, + { + "epoch": 1.6919902016710848, + "grad_norm": 1.0623986721038818, + "learning_rate": 6.0882613087511146e-06, + "loss": 0.3849, + "step": 25212 + }, + { + "epoch": 1.6921244253548537, + "grad_norm": 0.9892853498458862, + "learning_rate": 6.083064798835236e-06, + "loss": 0.4769, + "step": 25214 + }, + { + "epoch": 1.6922586490386229, + "grad_norm": 1.0125302076339722, + "learning_rate": 6.077870363908184e-06, + "loss": 0.4852, + "step": 25216 + }, + { + "epoch": 1.692392872722392, + "grad_norm": 1.0274652242660522, + "learning_rate": 6.0726780042153816e-06, + "loss": 0.5281, + "step": 25218 + }, + { + "epoch": 1.692527096406161, + "grad_norm": 1.1266363859176636, + "learning_rate": 6.067487720002174e-06, + "loss": 0.4616, + "step": 25220 + }, + { + "epoch": 1.6926613200899299, + "grad_norm": 1.0416055917739868, + "learning_rate": 6.0622995115137695e-06, + "loss": 0.4729, + "step": 25222 + }, + { + "epoch": 1.6927955437736988, + "grad_norm": 1.0067485570907593, + "learning_rate": 6.057113378995338e-06, + "loss": 0.4349, + "step": 25224 + }, + { + "epoch": 1.6929297674574677, + "grad_norm": 0.973070502281189, + "learning_rate": 6.051929322691868e-06, + "loss": 0.4329, + "step": 25226 + }, + { + "epoch": 1.6930639911412368, + "grad_norm": 0.9753096699714661, + "learning_rate": 6.046747342848325e-06, + "loss": 0.4652, + "step": 25228 + }, + { + "epoch": 1.693198214825006, + "grad_norm": 0.9495730400085449, + "learning_rate": 6.041567439709533e-06, + "loss": 0.4571, + "step": 25230 + }, + { + "epoch": 1.693332438508775, + "grad_norm": 0.9078459739685059, + "learning_rate": 6.036389613520243e-06, + "loss": 0.4677, + "step": 25232 + }, + { + "epoch": 1.6934666621925438, + "grad_norm": 1.0763219594955444, + "learning_rate": 6.031213864525098e-06, + "loss": 0.4892, + "step": 25234 + }, + { + "epoch": 1.6936008858763127, + "grad_norm": 1.1541110277175903, + "learning_rate": 6.02604019296864e-06, + "loss": 0.4815, + "step": 25236 + }, + { + "epoch": 1.6937351095600819, + "grad_norm": 0.9476569294929504, + "learning_rate": 6.0208685990952975e-06, + "loss": 0.3649, + "step": 25238 + }, + { + "epoch": 1.693869333243851, + "grad_norm": 0.9725757241249084, + "learning_rate": 6.0156990831494395e-06, + "loss": 0.4933, + "step": 25240 + }, + { + "epoch": 1.69400355692762, + "grad_norm": 0.9956469535827637, + "learning_rate": 6.010531645375312e-06, + "loss": 0.4538, + "step": 25242 + }, + { + "epoch": 1.6941377806113889, + "grad_norm": 0.9989961981773376, + "learning_rate": 6.005366286017061e-06, + "loss": 0.4813, + "step": 25244 + }, + { + "epoch": 1.6942720042951578, + "grad_norm": 1.0470447540283203, + "learning_rate": 6.000203005318733e-06, + "loss": 0.4527, + "step": 25246 + }, + { + "epoch": 1.694406227978927, + "grad_norm": 0.8920106291770935, + "learning_rate": 5.9950418035243036e-06, + "loss": 0.4608, + "step": 25248 + }, + { + "epoch": 1.6945404516626958, + "grad_norm": 1.0534292459487915, + "learning_rate": 5.9898826808776006e-06, + "loss": 0.4716, + "step": 25250 + }, + { + "epoch": 1.694674675346465, + "grad_norm": 1.0337680578231812, + "learning_rate": 5.984725637622424e-06, + "loss": 0.467, + "step": 25252 + }, + { + "epoch": 1.694808899030234, + "grad_norm": 0.9041849374771118, + "learning_rate": 5.979570674002388e-06, + "loss": 0.4682, + "step": 25254 + }, + { + "epoch": 1.6949431227140028, + "grad_norm": 1.0136576890945435, + "learning_rate": 5.974417790261083e-06, + "loss": 0.4264, + "step": 25256 + }, + { + "epoch": 1.6950773463977717, + "grad_norm": 1.0099191665649414, + "learning_rate": 5.969266986641953e-06, + "loss": 0.4767, + "step": 25258 + }, + { + "epoch": 1.6952115700815409, + "grad_norm": 1.1609923839569092, + "learning_rate": 5.964118263388391e-06, + "loss": 0.4584, + "step": 25260 + }, + { + "epoch": 1.69534579376531, + "grad_norm": 0.9966214895248413, + "learning_rate": 5.9589716207436475e-06, + "loss": 0.4714, + "step": 25262 + }, + { + "epoch": 1.695480017449079, + "grad_norm": 1.0159577131271362, + "learning_rate": 5.9538270589508895e-06, + "loss": 0.5211, + "step": 25264 + }, + { + "epoch": 1.6956142411328479, + "grad_norm": 1.0560600757598877, + "learning_rate": 5.948684578253177e-06, + "loss": 0.4752, + "step": 25266 + }, + { + "epoch": 1.6957484648166168, + "grad_norm": 1.02834153175354, + "learning_rate": 5.943544178893506e-06, + "loss": 0.4604, + "step": 25268 + }, + { + "epoch": 1.695882688500386, + "grad_norm": 1.0763250589370728, + "learning_rate": 5.938405861114743e-06, + "loss": 0.5098, + "step": 25270 + }, + { + "epoch": 1.6960169121841548, + "grad_norm": 1.0647401809692383, + "learning_rate": 5.93326962515966e-06, + "loss": 0.4427, + "step": 25272 + }, + { + "epoch": 1.696151135867924, + "grad_norm": 0.8827753067016602, + "learning_rate": 5.928135471270929e-06, + "loss": 0.4371, + "step": 25274 + }, + { + "epoch": 1.696285359551693, + "grad_norm": 1.181510329246521, + "learning_rate": 5.92300339969113e-06, + "loss": 0.4472, + "step": 25276 + }, + { + "epoch": 1.6964195832354618, + "grad_norm": 1.0063621997833252, + "learning_rate": 5.9178734106627495e-06, + "loss": 0.4782, + "step": 25278 + }, + { + "epoch": 1.6965538069192307, + "grad_norm": 0.9924083352088928, + "learning_rate": 5.9127455044281684e-06, + "loss": 0.4601, + "step": 25280 + }, + { + "epoch": 1.6966880306029999, + "grad_norm": 0.8651800751686096, + "learning_rate": 5.90761968122967e-06, + "loss": 0.4234, + "step": 25282 + }, + { + "epoch": 1.696822254286769, + "grad_norm": 0.956924319267273, + "learning_rate": 5.902495941309427e-06, + "loss": 0.4557, + "step": 25284 + }, + { + "epoch": 1.696956477970538, + "grad_norm": 1.0364547967910767, + "learning_rate": 5.897374284909546e-06, + "loss": 0.5135, + "step": 25286 + }, + { + "epoch": 1.6970907016543069, + "grad_norm": 1.1208486557006836, + "learning_rate": 5.892254712272e-06, + "loss": 0.4717, + "step": 25288 + }, + { + "epoch": 1.6972249253380758, + "grad_norm": 0.8544895648956299, + "learning_rate": 5.887137223638689e-06, + "loss": 0.3906, + "step": 25290 + }, + { + "epoch": 1.697359149021845, + "grad_norm": 1.0679324865341187, + "learning_rate": 5.882021819251405e-06, + "loss": 0.4137, + "step": 25292 + }, + { + "epoch": 1.697493372705614, + "grad_norm": 0.9690061211585999, + "learning_rate": 5.8769084993518355e-06, + "loss": 0.4583, + "step": 25294 + }, + { + "epoch": 1.697627596389383, + "grad_norm": 1.0798450708389282, + "learning_rate": 5.871797264181561e-06, + "loss": 0.4734, + "step": 25296 + }, + { + "epoch": 1.697761820073152, + "grad_norm": 1.0062329769134521, + "learning_rate": 5.866688113982105e-06, + "loss": 0.3716, + "step": 25298 + }, + { + "epoch": 1.6978960437569208, + "grad_norm": 1.0647355318069458, + "learning_rate": 5.861581048994847e-06, + "loss": 0.4397, + "step": 25300 + }, + { + "epoch": 1.6980302674406897, + "grad_norm": 1.480678915977478, + "learning_rate": 5.856476069461092e-06, + "loss": 0.5071, + "step": 25302 + }, + { + "epoch": 1.6981644911244589, + "grad_norm": 1.0006400346755981, + "learning_rate": 5.851373175622032e-06, + "loss": 0.475, + "step": 25304 + }, + { + "epoch": 1.698298714808228, + "grad_norm": 1.111854910850525, + "learning_rate": 5.84627236771878e-06, + "loss": 0.4143, + "step": 25306 + }, + { + "epoch": 1.698432938491997, + "grad_norm": 1.1463509798049927, + "learning_rate": 5.841173645992326e-06, + "loss": 0.4626, + "step": 25308 + }, + { + "epoch": 1.6985671621757659, + "grad_norm": 0.9406129717826843, + "learning_rate": 5.836077010683599e-06, + "loss": 0.3964, + "step": 25310 + }, + { + "epoch": 1.6987013858595348, + "grad_norm": 1.0047578811645508, + "learning_rate": 5.830982462033374e-06, + "loss": 0.4708, + "step": 25312 + }, + { + "epoch": 1.698835609543304, + "grad_norm": 0.8719502091407776, + "learning_rate": 5.82589000028238e-06, + "loss": 0.4133, + "step": 25314 + }, + { + "epoch": 1.698969833227073, + "grad_norm": 1.0489619970321655, + "learning_rate": 5.8207996256712084e-06, + "loss": 0.5194, + "step": 25316 + }, + { + "epoch": 1.699104056910842, + "grad_norm": 1.0129246711730957, + "learning_rate": 5.815711338440394e-06, + "loss": 0.4481, + "step": 25318 + }, + { + "epoch": 1.699238280594611, + "grad_norm": 1.257003664970398, + "learning_rate": 5.810625138830333e-06, + "loss": 0.4173, + "step": 25320 + }, + { + "epoch": 1.6993725042783798, + "grad_norm": 1.0962368249893188, + "learning_rate": 5.805541027081335e-06, + "loss": 0.5, + "step": 25322 + }, + { + "epoch": 1.699506727962149, + "grad_norm": 0.9793606400489807, + "learning_rate": 5.800459003433612e-06, + "loss": 0.4969, + "step": 25324 + }, + { + "epoch": 1.6996409516459179, + "grad_norm": 1.0478161573410034, + "learning_rate": 5.795379068127299e-06, + "loss": 0.5096, + "step": 25326 + }, + { + "epoch": 1.699775175329687, + "grad_norm": 1.0728329420089722, + "learning_rate": 5.790301221402394e-06, + "loss": 0.4675, + "step": 25328 + }, + { + "epoch": 1.699909399013456, + "grad_norm": 0.9906491041183472, + "learning_rate": 5.785225463498828e-06, + "loss": 0.4783, + "step": 25330 + }, + { + "epoch": 1.7000436226972249, + "grad_norm": 1.124284267425537, + "learning_rate": 5.780151794656397e-06, + "loss": 0.4918, + "step": 25332 + }, + { + "epoch": 1.7001778463809938, + "grad_norm": 1.0835182666778564, + "learning_rate": 5.775080215114853e-06, + "loss": 0.4822, + "step": 25334 + }, + { + "epoch": 1.700312070064763, + "grad_norm": 0.9419242739677429, + "learning_rate": 5.770010725113794e-06, + "loss": 0.4287, + "step": 25336 + }, + { + "epoch": 1.700446293748532, + "grad_norm": 0.8902066349983215, + "learning_rate": 5.76494332489278e-06, + "loss": 0.4638, + "step": 25338 + }, + { + "epoch": 1.700580517432301, + "grad_norm": 1.1273128986358643, + "learning_rate": 5.759878014691189e-06, + "loss": 0.5564, + "step": 25340 + }, + { + "epoch": 1.70071474111607, + "grad_norm": 1.0304105281829834, + "learning_rate": 5.754814794748364e-06, + "loss": 0.4371, + "step": 25342 + }, + { + "epoch": 1.7008489647998388, + "grad_norm": 1.0378000736236572, + "learning_rate": 5.749753665303542e-06, + "loss": 0.4676, + "step": 25344 + }, + { + "epoch": 1.700983188483608, + "grad_norm": 1.0077403783798218, + "learning_rate": 5.744694626595837e-06, + "loss": 0.4614, + "step": 25346 + }, + { + "epoch": 1.7011174121673769, + "grad_norm": 1.2381305694580078, + "learning_rate": 5.739637678864307e-06, + "loss": 0.5158, + "step": 25348 + }, + { + "epoch": 1.701251635851146, + "grad_norm": 0.9145381450653076, + "learning_rate": 5.734582822347839e-06, + "loss": 0.3904, + "step": 25350 + }, + { + "epoch": 1.701385859534915, + "grad_norm": 1.000479817390442, + "learning_rate": 5.729530057285304e-06, + "loss": 0.5064, + "step": 25352 + }, + { + "epoch": 1.7015200832186839, + "grad_norm": 1.0135163068771362, + "learning_rate": 5.724479383915404e-06, + "loss": 0.4627, + "step": 25354 + }, + { + "epoch": 1.7016543069024528, + "grad_norm": 0.9599625468254089, + "learning_rate": 5.719430802476805e-06, + "loss": 0.4497, + "step": 25356 + }, + { + "epoch": 1.701788530586222, + "grad_norm": 1.050414800643921, + "learning_rate": 5.714384313208021e-06, + "loss": 0.4701, + "step": 25358 + }, + { + "epoch": 1.701922754269991, + "grad_norm": 1.0206334590911865, + "learning_rate": 5.7093399163474945e-06, + "loss": 0.5321, + "step": 25360 + }, + { + "epoch": 1.70205697795376, + "grad_norm": 1.0668940544128418, + "learning_rate": 5.704297612133558e-06, + "loss": 0.4434, + "step": 25362 + }, + { + "epoch": 1.702191201637529, + "grad_norm": 0.978858470916748, + "learning_rate": 5.699257400804464e-06, + "loss": 0.4902, + "step": 25364 + }, + { + "epoch": 1.7023254253212978, + "grad_norm": 1.117708683013916, + "learning_rate": 5.69421928259834e-06, + "loss": 0.5651, + "step": 25366 + }, + { + "epoch": 1.702459649005067, + "grad_norm": 0.9856138229370117, + "learning_rate": 5.689183257753234e-06, + "loss": 0.445, + "step": 25368 + }, + { + "epoch": 1.702593872688836, + "grad_norm": 1.003671646118164, + "learning_rate": 5.684149326507077e-06, + "loss": 0.488, + "step": 25370 + }, + { + "epoch": 1.702728096372605, + "grad_norm": 1.1078087091445923, + "learning_rate": 5.67911748909773e-06, + "loss": 0.4242, + "step": 25372 + }, + { + "epoch": 1.702862320056374, + "grad_norm": 0.9358893632888794, + "learning_rate": 5.674087745762918e-06, + "loss": 0.4414, + "step": 25374 + }, + { + "epoch": 1.7029965437401429, + "grad_norm": 1.1128106117248535, + "learning_rate": 5.669060096740314e-06, + "loss": 0.459, + "step": 25376 + }, + { + "epoch": 1.7031307674239118, + "grad_norm": 1.0639312267303467, + "learning_rate": 5.6640345422674325e-06, + "loss": 0.4877, + "step": 25378 + }, + { + "epoch": 1.703264991107681, + "grad_norm": 0.9991609454154968, + "learning_rate": 5.6590110825817445e-06, + "loss": 0.4993, + "step": 25380 + }, + { + "epoch": 1.70339921479145, + "grad_norm": 0.8029139041900635, + "learning_rate": 5.653989717920577e-06, + "loss": 0.3592, + "step": 25382 + }, + { + "epoch": 1.703533438475219, + "grad_norm": 1.1822975873947144, + "learning_rate": 5.648970448521207e-06, + "loss": 0.4926, + "step": 25384 + }, + { + "epoch": 1.703667662158988, + "grad_norm": 1.2122814655303955, + "learning_rate": 5.643953274620767e-06, + "loss": 0.5176, + "step": 25386 + }, + { + "epoch": 1.7038018858427568, + "grad_norm": 1.0605030059814453, + "learning_rate": 5.6389381964563115e-06, + "loss": 0.4726, + "step": 25388 + }, + { + "epoch": 1.703936109526526, + "grad_norm": 1.047315239906311, + "learning_rate": 5.63392521426479e-06, + "loss": 0.5145, + "step": 25390 + }, + { + "epoch": 1.704070333210295, + "grad_norm": 0.8657220602035522, + "learning_rate": 5.628914328283064e-06, + "loss": 0.4303, + "step": 25392 + }, + { + "epoch": 1.704204556894064, + "grad_norm": 0.9523367881774902, + "learning_rate": 5.623905538747887e-06, + "loss": 0.4382, + "step": 25394 + }, + { + "epoch": 1.704338780577833, + "grad_norm": 0.9913052916526794, + "learning_rate": 5.618898845895909e-06, + "loss": 0.4347, + "step": 25396 + }, + { + "epoch": 1.7044730042616019, + "grad_norm": 1.1487425565719604, + "learning_rate": 5.613894249963681e-06, + "loss": 0.4715, + "step": 25398 + }, + { + "epoch": 1.704607227945371, + "grad_norm": 1.1615405082702637, + "learning_rate": 5.608891751187678e-06, + "loss": 0.4241, + "step": 25400 + }, + { + "epoch": 1.70474145162914, + "grad_norm": 1.0483362674713135, + "learning_rate": 5.603891349804241e-06, + "loss": 0.4625, + "step": 25402 + }, + { + "epoch": 1.704875675312909, + "grad_norm": 0.9386702179908752, + "learning_rate": 5.5988930460496525e-06, + "loss": 0.4794, + "step": 25404 + }, + { + "epoch": 1.705009898996678, + "grad_norm": 1.0929003953933716, + "learning_rate": 5.5938968401600455e-06, + "loss": 0.4878, + "step": 25406 + }, + { + "epoch": 1.705144122680447, + "grad_norm": 1.1455669403076172, + "learning_rate": 5.588902732371487e-06, + "loss": 0.4722, + "step": 25408 + }, + { + "epoch": 1.7052783463642158, + "grad_norm": 1.1001158952713013, + "learning_rate": 5.5839107229199495e-06, + "loss": 0.4982, + "step": 25410 + }, + { + "epoch": 1.705412570047985, + "grad_norm": 1.0804669857025146, + "learning_rate": 5.5789208120412824e-06, + "loss": 0.3945, + "step": 25412 + }, + { + "epoch": 1.705546793731754, + "grad_norm": 1.0996463298797607, + "learning_rate": 5.573932999971266e-06, + "loss": 0.4517, + "step": 25414 + }, + { + "epoch": 1.705681017415523, + "grad_norm": 0.9030675888061523, + "learning_rate": 5.568947286945553e-06, + "loss": 0.4267, + "step": 25416 + }, + { + "epoch": 1.705815241099292, + "grad_norm": 0.9615384340286255, + "learning_rate": 5.563963673199713e-06, + "loss": 0.4534, + "step": 25418 + }, + { + "epoch": 1.7059494647830609, + "grad_norm": 0.9762392640113831, + "learning_rate": 5.558982158969195e-06, + "loss": 0.4562, + "step": 25420 + }, + { + "epoch": 1.70608368846683, + "grad_norm": 0.8485289812088013, + "learning_rate": 5.554002744489395e-06, + "loss": 0.4173, + "step": 25422 + }, + { + "epoch": 1.706217912150599, + "grad_norm": 1.0777990818023682, + "learning_rate": 5.549025429995569e-06, + "loss": 0.4533, + "step": 25424 + }, + { + "epoch": 1.706352135834368, + "grad_norm": 1.0341979265213013, + "learning_rate": 5.5440502157228794e-06, + "loss": 0.4627, + "step": 25426 + }, + { + "epoch": 1.706486359518137, + "grad_norm": 1.2203598022460938, + "learning_rate": 5.539077101906387e-06, + "loss": 0.4815, + "step": 25428 + }, + { + "epoch": 1.706620583201906, + "grad_norm": 0.9701406359672546, + "learning_rate": 5.534106088781082e-06, + "loss": 0.5055, + "step": 25430 + }, + { + "epoch": 1.7067548068856748, + "grad_norm": 0.9804043769836426, + "learning_rate": 5.529137176581828e-06, + "loss": 0.4936, + "step": 25432 + }, + { + "epoch": 1.706889030569444, + "grad_norm": 0.9713636636734009, + "learning_rate": 5.5241703655433966e-06, + "loss": 0.4967, + "step": 25434 + }, + { + "epoch": 1.707023254253213, + "grad_norm": 0.9633068442344666, + "learning_rate": 5.519205655900439e-06, + "loss": 0.4242, + "step": 25436 + }, + { + "epoch": 1.707157477936982, + "grad_norm": 1.0099884271621704, + "learning_rate": 5.514243047887563e-06, + "loss": 0.5068, + "step": 25438 + }, + { + "epoch": 1.707291701620751, + "grad_norm": 1.053117036819458, + "learning_rate": 5.509282541739213e-06, + "loss": 0.4916, + "step": 25440 + }, + { + "epoch": 1.7074259253045199, + "grad_norm": 1.0143020153045654, + "learning_rate": 5.504324137689793e-06, + "loss": 0.4918, + "step": 25442 + }, + { + "epoch": 1.707560148988289, + "grad_norm": 1.090571641921997, + "learning_rate": 5.4993678359735514e-06, + "loss": 0.4509, + "step": 25444 + }, + { + "epoch": 1.7076943726720581, + "grad_norm": 1.0744800567626953, + "learning_rate": 5.494413636824675e-06, + "loss": 0.4603, + "step": 25446 + }, + { + "epoch": 1.707828596355827, + "grad_norm": 1.0884594917297363, + "learning_rate": 5.489461540477231e-06, + "loss": 0.4951, + "step": 25448 + }, + { + "epoch": 1.707962820039596, + "grad_norm": 1.0418914556503296, + "learning_rate": 5.484511547165211e-06, + "loss": 0.48, + "step": 25450 + }, + { + "epoch": 1.708097043723365, + "grad_norm": 1.0547127723693848, + "learning_rate": 5.479563657122483e-06, + "loss": 0.4791, + "step": 25452 + }, + { + "epoch": 1.7082312674071338, + "grad_norm": 1.3083374500274658, + "learning_rate": 5.474617870582826e-06, + "loss": 0.4474, + "step": 25454 + }, + { + "epoch": 1.708365491090903, + "grad_norm": 1.7928715944290161, + "learning_rate": 5.469674187779911e-06, + "loss": 0.522, + "step": 25456 + }, + { + "epoch": 1.708499714774672, + "grad_norm": 1.0545799732208252, + "learning_rate": 5.464732608947337e-06, + "loss": 0.4532, + "step": 25458 + }, + { + "epoch": 1.708633938458441, + "grad_norm": 1.0372275114059448, + "learning_rate": 5.45979313431857e-06, + "loss": 0.4123, + "step": 25460 + }, + { + "epoch": 1.70876816214221, + "grad_norm": 0.9443169236183167, + "learning_rate": 5.454855764126992e-06, + "loss": 0.4913, + "step": 25462 + }, + { + "epoch": 1.7089023858259789, + "grad_norm": 1.266764760017395, + "learning_rate": 5.4499204986058815e-06, + "loss": 0.4059, + "step": 25464 + }, + { + "epoch": 1.709036609509748, + "grad_norm": 1.0108635425567627, + "learning_rate": 5.444987337988428e-06, + "loss": 0.4567, + "step": 25466 + }, + { + "epoch": 1.7091708331935171, + "grad_norm": 1.1910806894302368, + "learning_rate": 5.440056282507699e-06, + "loss": 0.4899, + "step": 25468 + }, + { + "epoch": 1.709305056877286, + "grad_norm": 1.028946876525879, + "learning_rate": 5.435127332396695e-06, + "loss": 0.465, + "step": 25470 + }, + { + "epoch": 1.709439280561055, + "grad_norm": 0.9755556583404541, + "learning_rate": 5.4302004878882986e-06, + "loss": 0.4738, + "step": 25472 + }, + { + "epoch": 1.709573504244824, + "grad_norm": 0.9384263753890991, + "learning_rate": 5.425275749215281e-06, + "loss": 0.4892, + "step": 25474 + }, + { + "epoch": 1.709707727928593, + "grad_norm": 0.9885052442550659, + "learning_rate": 5.4203531166103325e-06, + "loss": 0.4537, + "step": 25476 + }, + { + "epoch": 1.709841951612362, + "grad_norm": 0.8746045827865601, + "learning_rate": 5.415432590306024e-06, + "loss": 0.4216, + "step": 25478 + }, + { + "epoch": 1.709976175296131, + "grad_norm": 1.0783562660217285, + "learning_rate": 5.41051417053487e-06, + "loss": 0.4666, + "step": 25480 + }, + { + "epoch": 1.7101103989799, + "grad_norm": 1.1761975288391113, + "learning_rate": 5.405597857529238e-06, + "loss": 0.4829, + "step": 25482 + }, + { + "epoch": 1.710244622663669, + "grad_norm": 1.053392767906189, + "learning_rate": 5.400683651521416e-06, + "loss": 0.4962, + "step": 25484 + }, + { + "epoch": 1.7103788463474379, + "grad_norm": 0.8936684727668762, + "learning_rate": 5.395771552743578e-06, + "loss": 0.4194, + "step": 25486 + }, + { + "epoch": 1.710513070031207, + "grad_norm": 0.8621320128440857, + "learning_rate": 5.390861561427835e-06, + "loss": 0.3909, + "step": 25488 + }, + { + "epoch": 1.7106472937149761, + "grad_norm": 0.9584031701087952, + "learning_rate": 5.3859536778061655e-06, + "loss": 0.4953, + "step": 25490 + }, + { + "epoch": 1.710781517398745, + "grad_norm": 1.0502445697784424, + "learning_rate": 5.381047902110453e-06, + "loss": 0.4979, + "step": 25492 + }, + { + "epoch": 1.710915741082514, + "grad_norm": 1.0048887729644775, + "learning_rate": 5.376144234572478e-06, + "loss": 0.3925, + "step": 25494 + }, + { + "epoch": 1.711049964766283, + "grad_norm": 0.9571940898895264, + "learning_rate": 5.371242675423949e-06, + "loss": 0.3957, + "step": 25496 + }, + { + "epoch": 1.711184188450052, + "grad_norm": 0.9335609078407288, + "learning_rate": 5.366343224896436e-06, + "loss": 0.434, + "step": 25498 + }, + { + "epoch": 1.711318412133821, + "grad_norm": 1.1270277500152588, + "learning_rate": 5.361445883221456e-06, + "loss": 0.5048, + "step": 25500 + }, + { + "epoch": 1.71145263581759, + "grad_norm": 1.101144790649414, + "learning_rate": 5.356550650630359e-06, + "loss": 0.4865, + "step": 25502 + }, + { + "epoch": 1.711586859501359, + "grad_norm": 0.9321258664131165, + "learning_rate": 5.35165752735447e-06, + "loss": 0.421, + "step": 25504 + }, + { + "epoch": 1.711721083185128, + "grad_norm": 1.087894082069397, + "learning_rate": 5.346766513624951e-06, + "loss": 0.453, + "step": 25506 + }, + { + "epoch": 1.7118553068688969, + "grad_norm": 0.9895612597465515, + "learning_rate": 5.341877609672919e-06, + "loss": 0.4835, + "step": 25508 + }, + { + "epoch": 1.711989530552666, + "grad_norm": 1.0466296672821045, + "learning_rate": 5.336990815729354e-06, + "loss": 0.4136, + "step": 25510 + }, + { + "epoch": 1.7121237542364351, + "grad_norm": 1.0212011337280273, + "learning_rate": 5.33210613202515e-06, + "loss": 0.5037, + "step": 25512 + }, + { + "epoch": 1.712257977920204, + "grad_norm": 1.071318507194519, + "learning_rate": 5.327223558791084e-06, + "loss": 0.5119, + "step": 25514 + }, + { + "epoch": 1.712392201603973, + "grad_norm": 1.0949668884277344, + "learning_rate": 5.322343096257864e-06, + "loss": 0.5161, + "step": 25516 + }, + { + "epoch": 1.712526425287742, + "grad_norm": 0.8076902627944946, + "learning_rate": 5.317464744656081e-06, + "loss": 0.4839, + "step": 25518 + }, + { + "epoch": 1.712660648971511, + "grad_norm": 1.0798898935317993, + "learning_rate": 5.312588504216226e-06, + "loss": 0.5646, + "step": 25520 + }, + { + "epoch": 1.7127948726552802, + "grad_norm": 0.9206319451332092, + "learning_rate": 5.307714375168682e-06, + "loss": 0.5107, + "step": 25522 + }, + { + "epoch": 1.712929096339049, + "grad_norm": 1.0472077131271362, + "learning_rate": 5.302842357743754e-06, + "loss": 0.5383, + "step": 25524 + }, + { + "epoch": 1.713063320022818, + "grad_norm": 1.1249159574508667, + "learning_rate": 5.297972452171629e-06, + "loss": 0.4088, + "step": 25526 + }, + { + "epoch": 1.713197543706587, + "grad_norm": 1.0618162155151367, + "learning_rate": 5.293104658682419e-06, + "loss": 0.4658, + "step": 25528 + }, + { + "epoch": 1.7133317673903559, + "grad_norm": 1.0651888847351074, + "learning_rate": 5.288238977506077e-06, + "loss": 0.5678, + "step": 25530 + }, + { + "epoch": 1.713465991074125, + "grad_norm": 1.0403186082839966, + "learning_rate": 5.283375408872537e-06, + "loss": 0.4787, + "step": 25532 + }, + { + "epoch": 1.7136002147578941, + "grad_norm": 1.0864567756652832, + "learning_rate": 5.278513953011566e-06, + "loss": 0.5136, + "step": 25534 + }, + { + "epoch": 1.713734438441663, + "grad_norm": 0.9364561438560486, + "learning_rate": 5.273654610152884e-06, + "loss": 0.4665, + "step": 25536 + }, + { + "epoch": 1.713868662125432, + "grad_norm": 1.0810825824737549, + "learning_rate": 5.268797380526064e-06, + "loss": 0.4828, + "step": 25538 + }, + { + "epoch": 1.714002885809201, + "grad_norm": 0.7876555919647217, + "learning_rate": 5.26394226436061e-06, + "loss": 0.4377, + "step": 25540 + }, + { + "epoch": 1.71413710949297, + "grad_norm": 1.094198226928711, + "learning_rate": 5.259089261885908e-06, + "loss": 0.4901, + "step": 25542 + }, + { + "epoch": 1.7142713331767392, + "grad_norm": 1.0900894403457642, + "learning_rate": 5.254238373331266e-06, + "loss": 0.4378, + "step": 25544 + }, + { + "epoch": 1.714405556860508, + "grad_norm": 0.9810633659362793, + "learning_rate": 5.249389598925869e-06, + "loss": 0.4832, + "step": 25546 + }, + { + "epoch": 1.714539780544277, + "grad_norm": 0.9120228290557861, + "learning_rate": 5.244542938898822e-06, + "loss": 0.4448, + "step": 25548 + }, + { + "epoch": 1.714674004228046, + "grad_norm": 0.9897778630256653, + "learning_rate": 5.2396983934791085e-06, + "loss": 0.4572, + "step": 25550 + }, + { + "epoch": 1.714808227911815, + "grad_norm": 0.949037492275238, + "learning_rate": 5.2348559628956185e-06, + "loss": 0.4812, + "step": 25552 + }, + { + "epoch": 1.714942451595584, + "grad_norm": 1.08169686794281, + "learning_rate": 5.230015647377168e-06, + "loss": 0.4856, + "step": 25554 + }, + { + "epoch": 1.7150766752793531, + "grad_norm": 0.8839130997657776, + "learning_rate": 5.225177447152446e-06, + "loss": 0.5099, + "step": 25556 + }, + { + "epoch": 1.715210898963122, + "grad_norm": 1.1443039178848267, + "learning_rate": 5.220341362450038e-06, + "loss": 0.4887, + "step": 25558 + }, + { + "epoch": 1.715345122646891, + "grad_norm": 1.0248621702194214, + "learning_rate": 5.215507393498437e-06, + "loss": 0.4513, + "step": 25560 + }, + { + "epoch": 1.71547934633066, + "grad_norm": 1.023232340812683, + "learning_rate": 5.2106755405260555e-06, + "loss": 0.4827, + "step": 25562 + }, + { + "epoch": 1.715613570014429, + "grad_norm": 1.0187675952911377, + "learning_rate": 5.205845803761172e-06, + "loss": 0.422, + "step": 25564 + }, + { + "epoch": 1.7157477936981982, + "grad_norm": 0.9954760670661926, + "learning_rate": 5.201018183432005e-06, + "loss": 0.4455, + "step": 25566 + }, + { + "epoch": 1.715882017381967, + "grad_norm": 0.9201191067695618, + "learning_rate": 5.196192679766626e-06, + "loss": 0.4457, + "step": 25568 + }, + { + "epoch": 1.716016241065736, + "grad_norm": 1.1385751962661743, + "learning_rate": 5.191369292993048e-06, + "loss": 0.4852, + "step": 25570 + }, + { + "epoch": 1.716150464749505, + "grad_norm": 1.0543646812438965, + "learning_rate": 5.186548023339144e-06, + "loss": 0.5159, + "step": 25572 + }, + { + "epoch": 1.716284688433274, + "grad_norm": 1.015798807144165, + "learning_rate": 5.181728871032737e-06, + "loss": 0.4604, + "step": 25574 + }, + { + "epoch": 1.716418912117043, + "grad_norm": 1.0484914779663086, + "learning_rate": 5.176911836301507e-06, + "loss": 0.4596, + "step": 25576 + }, + { + "epoch": 1.7165531358008121, + "grad_norm": 1.0813695192337036, + "learning_rate": 5.172096919373048e-06, + "loss": 0.5332, + "step": 25578 + }, + { + "epoch": 1.716687359484581, + "grad_norm": 0.9233369827270508, + "learning_rate": 5.167284120474858e-06, + "loss": 0.4424, + "step": 25580 + }, + { + "epoch": 1.71682158316835, + "grad_norm": 0.9958009719848633, + "learning_rate": 5.162473439834337e-06, + "loss": 0.4603, + "step": 25582 + }, + { + "epoch": 1.716955806852119, + "grad_norm": 1.0955454111099243, + "learning_rate": 5.157664877678781e-06, + "loss": 0.4336, + "step": 25584 + }, + { + "epoch": 1.717090030535888, + "grad_norm": 0.8009112477302551, + "learning_rate": 5.15285843423538e-06, + "loss": 0.3836, + "step": 25586 + }, + { + "epoch": 1.7172242542196572, + "grad_norm": 1.1768845319747925, + "learning_rate": 5.148054109731215e-06, + "loss": 0.4783, + "step": 25588 + }, + { + "epoch": 1.717358477903426, + "grad_norm": 1.0453424453735352, + "learning_rate": 5.143251904393309e-06, + "loss": 0.5188, + "step": 25590 + }, + { + "epoch": 1.717492701587195, + "grad_norm": 1.0332114696502686, + "learning_rate": 5.1384518184485365e-06, + "loss": 0.4412, + "step": 25592 + }, + { + "epoch": 1.717626925270964, + "grad_norm": 1.0137203931808472, + "learning_rate": 5.1336538521237045e-06, + "loss": 0.4908, + "step": 25594 + }, + { + "epoch": 1.717761148954733, + "grad_norm": 1.1090179681777954, + "learning_rate": 5.128858005645504e-06, + "loss": 0.4465, + "step": 25596 + }, + { + "epoch": 1.7178953726385022, + "grad_norm": 1.002066969871521, + "learning_rate": 5.1240642792405314e-06, + "loss": 0.4907, + "step": 25598 + }, + { + "epoch": 1.7180295963222711, + "grad_norm": 1.0695041418075562, + "learning_rate": 5.1192726731352616e-06, + "loss": 0.4986, + "step": 25600 + }, + { + "epoch": 1.71816382000604, + "grad_norm": 1.0685627460479736, + "learning_rate": 5.114483187556113e-06, + "loss": 0.4662, + "step": 25602 + }, + { + "epoch": 1.718298043689809, + "grad_norm": 1.0577702522277832, + "learning_rate": 5.1096958227293765e-06, + "loss": 0.5022, + "step": 25604 + }, + { + "epoch": 1.718432267373578, + "grad_norm": 1.0659515857696533, + "learning_rate": 5.104910578881234e-06, + "loss": 0.4163, + "step": 25606 + }, + { + "epoch": 1.718566491057347, + "grad_norm": 1.0222342014312744, + "learning_rate": 5.100127456237774e-06, + "loss": 0.5014, + "step": 25608 + }, + { + "epoch": 1.7187007147411162, + "grad_norm": 0.907884955406189, + "learning_rate": 5.095346455025013e-06, + "loss": 0.439, + "step": 25610 + }, + { + "epoch": 1.718834938424885, + "grad_norm": 0.9424964189529419, + "learning_rate": 5.09056757546883e-06, + "loss": 0.4704, + "step": 25612 + }, + { + "epoch": 1.718969162108654, + "grad_norm": 1.0116387605667114, + "learning_rate": 5.085790817795016e-06, + "loss": 0.45, + "step": 25614 + }, + { + "epoch": 1.719103385792423, + "grad_norm": 1.033856749534607, + "learning_rate": 5.08101618222927e-06, + "loss": 0.46, + "step": 25616 + }, + { + "epoch": 1.719237609476192, + "grad_norm": 1.0117535591125488, + "learning_rate": 5.076243668997166e-06, + "loss": 0.4289, + "step": 25618 + }, + { + "epoch": 1.7193718331599612, + "grad_norm": 1.018268346786499, + "learning_rate": 5.071473278324223e-06, + "loss": 0.4865, + "step": 25620 + }, + { + "epoch": 1.7195060568437301, + "grad_norm": 0.9708705544471741, + "learning_rate": 5.066705010435807e-06, + "loss": 0.4655, + "step": 25622 + }, + { + "epoch": 1.719640280527499, + "grad_norm": 1.0578694343566895, + "learning_rate": 5.061938865557242e-06, + "loss": 0.4237, + "step": 25624 + }, + { + "epoch": 1.719774504211268, + "grad_norm": 0.9732657074928284, + "learning_rate": 5.057174843913681e-06, + "loss": 0.4777, + "step": 25626 + }, + { + "epoch": 1.7199087278950371, + "grad_norm": 1.0057907104492188, + "learning_rate": 5.05241294573024e-06, + "loss": 0.4298, + "step": 25628 + }, + { + "epoch": 1.720042951578806, + "grad_norm": 0.9988004565238953, + "learning_rate": 5.047653171231892e-06, + "loss": 0.4104, + "step": 25630 + }, + { + "epoch": 1.7201771752625752, + "grad_norm": 1.0792229175567627, + "learning_rate": 5.042895520643548e-06, + "loss": 0.4588, + "step": 25632 + }, + { + "epoch": 1.720311398946344, + "grad_norm": 1.048791766166687, + "learning_rate": 5.038139994189983e-06, + "loss": 0.432, + "step": 25634 + }, + { + "epoch": 1.720445622630113, + "grad_norm": 0.9940659403800964, + "learning_rate": 5.03338659209589e-06, + "loss": 0.4821, + "step": 25636 + }, + { + "epoch": 1.720579846313882, + "grad_norm": 1.1075587272644043, + "learning_rate": 5.02863531458585e-06, + "loss": 0.4728, + "step": 25638 + }, + { + "epoch": 1.720714069997651, + "grad_norm": 0.9461389780044556, + "learning_rate": 5.023886161884372e-06, + "loss": 0.4511, + "step": 25640 + }, + { + "epoch": 1.7208482936814202, + "grad_norm": 1.0536267757415771, + "learning_rate": 5.019139134215828e-06, + "loss": 0.4944, + "step": 25642 + }, + { + "epoch": 1.7209825173651891, + "grad_norm": 1.0569151639938354, + "learning_rate": 5.014394231804503e-06, + "loss": 0.458, + "step": 25644 + }, + { + "epoch": 1.721116741048958, + "grad_norm": 1.0429084300994873, + "learning_rate": 5.009651454874587e-06, + "loss": 0.4244, + "step": 25646 + }, + { + "epoch": 1.721250964732727, + "grad_norm": 1.1130216121673584, + "learning_rate": 5.004910803650181e-06, + "loss": 0.4691, + "step": 25648 + }, + { + "epoch": 1.7213851884164961, + "grad_norm": 0.9673048257827759, + "learning_rate": 5.000172278355248e-06, + "loss": 0.4656, + "step": 25650 + }, + { + "epoch": 1.721519412100265, + "grad_norm": 1.0673524141311646, + "learning_rate": 4.9954358792137054e-06, + "loss": 0.4935, + "step": 25652 + }, + { + "epoch": 1.7216536357840342, + "grad_norm": 1.0728918313980103, + "learning_rate": 4.9907016064493e-06, + "loss": 0.4575, + "step": 25654 + }, + { + "epoch": 1.721787859467803, + "grad_norm": 1.2493226528167725, + "learning_rate": 4.98596946028575e-06, + "loss": 0.4618, + "step": 25656 + }, + { + "epoch": 1.721922083151572, + "grad_norm": 2.444298505783081, + "learning_rate": 4.981239440946612e-06, + "loss": 0.4587, + "step": 25658 + }, + { + "epoch": 1.722056306835341, + "grad_norm": 1.0520304441452026, + "learning_rate": 4.976511548655399e-06, + "loss": 0.4246, + "step": 25660 + }, + { + "epoch": 1.72219053051911, + "grad_norm": 1.1098594665527344, + "learning_rate": 4.97178578363548e-06, + "loss": 0.5034, + "step": 25662 + }, + { + "epoch": 1.7223247542028792, + "grad_norm": 1.1211425065994263, + "learning_rate": 4.967062146110135e-06, + "loss": 0.5041, + "step": 25664 + }, + { + "epoch": 1.7224589778866481, + "grad_norm": 0.9872859120368958, + "learning_rate": 4.962340636302543e-06, + "loss": 0.4887, + "step": 25666 + }, + { + "epoch": 1.722593201570417, + "grad_norm": 1.1875641345977783, + "learning_rate": 4.957621254435801e-06, + "loss": 0.5114, + "step": 25668 + }, + { + "epoch": 1.722727425254186, + "grad_norm": 0.8748855590820312, + "learning_rate": 4.952904000732883e-06, + "loss": 0.4223, + "step": 25670 + }, + { + "epoch": 1.7228616489379551, + "grad_norm": 1.0315881967544556, + "learning_rate": 4.948188875416671e-06, + "loss": 0.49, + "step": 25672 + }, + { + "epoch": 1.7229958726217243, + "grad_norm": 1.328240156173706, + "learning_rate": 4.943475878709936e-06, + "loss": 0.4946, + "step": 25674 + }, + { + "epoch": 1.7231300963054932, + "grad_norm": 1.059990644454956, + "learning_rate": 4.938765010835373e-06, + "loss": 0.5015, + "step": 25676 + }, + { + "epoch": 1.723264319989262, + "grad_norm": 1.1780343055725098, + "learning_rate": 4.934056272015541e-06, + "loss": 0.45, + "step": 25678 + }, + { + "epoch": 1.723398543673031, + "grad_norm": 1.0624507665634155, + "learning_rate": 4.92934966247296e-06, + "loss": 0.4655, + "step": 25680 + }, + { + "epoch": 1.7235327673568, + "grad_norm": 1.0479305982589722, + "learning_rate": 4.924645182429966e-06, + "loss": 0.4233, + "step": 25682 + }, + { + "epoch": 1.723666991040569, + "grad_norm": 1.0353846549987793, + "learning_rate": 4.919942832108837e-06, + "loss": 0.5271, + "step": 25684 + }, + { + "epoch": 1.7238012147243382, + "grad_norm": 1.0945695638656616, + "learning_rate": 4.915242611731774e-06, + "loss": 0.501, + "step": 25686 + }, + { + "epoch": 1.7239354384081071, + "grad_norm": 0.9750905632972717, + "learning_rate": 4.910544521520838e-06, + "loss": 0.399, + "step": 25688 + }, + { + "epoch": 1.724069662091876, + "grad_norm": 0.9547551274299622, + "learning_rate": 4.905848561698023e-06, + "loss": 0.4843, + "step": 25690 + }, + { + "epoch": 1.724203885775645, + "grad_norm": 1.0283516645431519, + "learning_rate": 4.901154732485169e-06, + "loss": 0.435, + "step": 25692 + }, + { + "epoch": 1.7243381094594141, + "grad_norm": 1.0085430145263672, + "learning_rate": 4.896463034104082e-06, + "loss": 0.4541, + "step": 25694 + }, + { + "epoch": 1.7244723331431833, + "grad_norm": 1.0000320672988892, + "learning_rate": 4.891773466776417e-06, + "loss": 0.441, + "step": 25696 + }, + { + "epoch": 1.7246065568269522, + "grad_norm": 1.0053753852844238, + "learning_rate": 4.887086030723764e-06, + "loss": 0.459, + "step": 25698 + }, + { + "epoch": 1.724740780510721, + "grad_norm": 0.9564087986946106, + "learning_rate": 4.88240072616758e-06, + "loss": 0.495, + "step": 25700 + }, + { + "epoch": 1.72487500419449, + "grad_norm": 1.08132803440094, + "learning_rate": 4.877717553329247e-06, + "loss": 0.4802, + "step": 25702 + }, + { + "epoch": 1.7250092278782592, + "grad_norm": 0.9329249262809753, + "learning_rate": 4.873036512430018e-06, + "loss": 0.4915, + "step": 25704 + }, + { + "epoch": 1.725143451562028, + "grad_norm": 0.9538440704345703, + "learning_rate": 4.868357603691087e-06, + "loss": 0.4306, + "step": 25706 + }, + { + "epoch": 1.7252776752457972, + "grad_norm": 1.119462013244629, + "learning_rate": 4.863680827333511e-06, + "loss": 0.4666, + "step": 25708 + }, + { + "epoch": 1.7254118989295661, + "grad_norm": 1.0777448415756226, + "learning_rate": 4.859006183578263e-06, + "loss": 0.4519, + "step": 25710 + }, + { + "epoch": 1.725546122613335, + "grad_norm": 1.002044677734375, + "learning_rate": 4.854333672646188e-06, + "loss": 0.3923, + "step": 25712 + }, + { + "epoch": 1.725680346297104, + "grad_norm": 1.0467420816421509, + "learning_rate": 4.849663294758089e-06, + "loss": 0.4285, + "step": 25714 + }, + { + "epoch": 1.7258145699808731, + "grad_norm": 1.0576964616775513, + "learning_rate": 4.844995050134604e-06, + "loss": 0.4917, + "step": 25716 + }, + { + "epoch": 1.7259487936646423, + "grad_norm": 1.1266741752624512, + "learning_rate": 4.840328938996325e-06, + "loss": 0.4802, + "step": 25718 + }, + { + "epoch": 1.7260830173484112, + "grad_norm": 1.1078001260757446, + "learning_rate": 4.835664961563685e-06, + "loss": 0.4634, + "step": 25720 + }, + { + "epoch": 1.72621724103218, + "grad_norm": 1.1022974252700806, + "learning_rate": 4.831003118057076e-06, + "loss": 0.4226, + "step": 25722 + }, + { + "epoch": 1.726351464715949, + "grad_norm": 1.0248111486434937, + "learning_rate": 4.826343408696732e-06, + "loss": 0.4809, + "step": 25724 + }, + { + "epoch": 1.7264856883997182, + "grad_norm": 0.9810677170753479, + "learning_rate": 4.821685833702849e-06, + "loss": 0.4922, + "step": 25726 + }, + { + "epoch": 1.726619912083487, + "grad_norm": 0.9958468675613403, + "learning_rate": 4.817030393295463e-06, + "loss": 0.4347, + "step": 25728 + }, + { + "epoch": 1.7267541357672562, + "grad_norm": 0.9662964344024658, + "learning_rate": 4.812377087694547e-06, + "loss": 0.4466, + "step": 25730 + }, + { + "epoch": 1.7268883594510251, + "grad_norm": 1.0769606828689575, + "learning_rate": 4.807725917119949e-06, + "loss": 0.4559, + "step": 25732 + }, + { + "epoch": 1.727022583134794, + "grad_norm": 1.0293160676956177, + "learning_rate": 4.803076881791441e-06, + "loss": 0.4372, + "step": 25734 + }, + { + "epoch": 1.727156806818563, + "grad_norm": 1.045389175415039, + "learning_rate": 4.798429981928676e-06, + "loss": 0.4758, + "step": 25736 + }, + { + "epoch": 1.7272910305023321, + "grad_norm": 1.0050643682479858, + "learning_rate": 4.793785217751206e-06, + "loss": 0.4356, + "step": 25738 + }, + { + "epoch": 1.7274252541861013, + "grad_norm": 1.9420435428619385, + "learning_rate": 4.7891425894784845e-06, + "loss": 0.4696, + "step": 25740 + }, + { + "epoch": 1.7275594778698702, + "grad_norm": 0.9572662115097046, + "learning_rate": 4.7845020973298795e-06, + "loss": 0.504, + "step": 25742 + }, + { + "epoch": 1.727693701553639, + "grad_norm": 0.9887237548828125, + "learning_rate": 4.779863741524626e-06, + "loss": 0.5097, + "step": 25744 + }, + { + "epoch": 1.727827925237408, + "grad_norm": 0.9231210350990295, + "learning_rate": 4.775227522281911e-06, + "loss": 0.4272, + "step": 25746 + }, + { + "epoch": 1.7279621489211772, + "grad_norm": 1.1272821426391602, + "learning_rate": 4.770593439820747e-06, + "loss": 0.4523, + "step": 25748 + }, + { + "epoch": 1.7280963726049463, + "grad_norm": 1.0660916566848755, + "learning_rate": 4.76596149436011e-06, + "loss": 0.5215, + "step": 25750 + }, + { + "epoch": 1.7282305962887152, + "grad_norm": 1.0930113792419434, + "learning_rate": 4.761331686118848e-06, + "loss": 0.4823, + "step": 25752 + }, + { + "epoch": 1.7283648199724841, + "grad_norm": 1.0196325778961182, + "learning_rate": 4.756704015315694e-06, + "loss": 0.4392, + "step": 25754 + }, + { + "epoch": 1.728499043656253, + "grad_norm": 1.0132194757461548, + "learning_rate": 4.7520784821693146e-06, + "loss": 0.441, + "step": 25756 + }, + { + "epoch": 1.728633267340022, + "grad_norm": 0.9595327377319336, + "learning_rate": 4.74745508689825e-06, + "loss": 0.3998, + "step": 25758 + }, + { + "epoch": 1.7287674910237911, + "grad_norm": 1.6109375953674316, + "learning_rate": 4.742833829720955e-06, + "loss": 0.4632, + "step": 25760 + }, + { + "epoch": 1.7289017147075603, + "grad_norm": 1.0205250978469849, + "learning_rate": 4.738214710855748e-06, + "loss": 0.4755, + "step": 25762 + }, + { + "epoch": 1.7290359383913292, + "grad_norm": 1.1388388872146606, + "learning_rate": 4.733597730520911e-06, + "loss": 0.4542, + "step": 25764 + }, + { + "epoch": 1.729170162075098, + "grad_norm": 0.931294322013855, + "learning_rate": 4.728982888934563e-06, + "loss": 0.502, + "step": 25766 + }, + { + "epoch": 1.729304385758867, + "grad_norm": 1.1117701530456543, + "learning_rate": 4.7243701863147525e-06, + "loss": 0.4964, + "step": 25768 + }, + { + "epoch": 1.7294386094426362, + "grad_norm": 1.090018630027771, + "learning_rate": 4.71975962287941e-06, + "loss": 0.5132, + "step": 25770 + }, + { + "epoch": 1.7295728331264053, + "grad_norm": 1.0194764137268066, + "learning_rate": 4.7151511988463955e-06, + "loss": 0.439, + "step": 25772 + }, + { + "epoch": 1.7297070568101742, + "grad_norm": 1.075696587562561, + "learning_rate": 4.7105449144334345e-06, + "loss": 0.5526, + "step": 25774 + }, + { + "epoch": 1.7298412804939431, + "grad_norm": 1.0920310020446777, + "learning_rate": 4.70594076985818e-06, + "loss": 0.4999, + "step": 25776 + }, + { + "epoch": 1.729975504177712, + "grad_norm": 0.9902180433273315, + "learning_rate": 4.7013387653381424e-06, + "loss": 0.4246, + "step": 25778 + }, + { + "epoch": 1.7301097278614812, + "grad_norm": 1.0894775390625, + "learning_rate": 4.696738901090781e-06, + "loss": 0.4508, + "step": 25780 + }, + { + "epoch": 1.7302439515452501, + "grad_norm": 0.9822218418121338, + "learning_rate": 4.692141177333409e-06, + "loss": 0.4512, + "step": 25782 + }, + { + "epoch": 1.7303781752290193, + "grad_norm": 0.9878543615341187, + "learning_rate": 4.687545594283282e-06, + "loss": 0.4415, + "step": 25784 + }, + { + "epoch": 1.7305123989127882, + "grad_norm": 1.0516366958618164, + "learning_rate": 4.682952152157522e-06, + "loss": 0.4311, + "step": 25786 + }, + { + "epoch": 1.730646622596557, + "grad_norm": 1.0247827768325806, + "learning_rate": 4.6783608511731635e-06, + "loss": 0.4547, + "step": 25788 + }, + { + "epoch": 1.730780846280326, + "grad_norm": 1.1121560335159302, + "learning_rate": 4.673771691547124e-06, + "loss": 0.5038, + "step": 25790 + }, + { + "epoch": 1.7309150699640952, + "grad_norm": 1.0238251686096191, + "learning_rate": 4.669184673496252e-06, + "loss": 0.3958, + "step": 25792 + }, + { + "epoch": 1.7310492936478643, + "grad_norm": 0.9520905017852783, + "learning_rate": 4.664599797237263e-06, + "loss": 0.4223, + "step": 25794 + }, + { + "epoch": 1.7311835173316332, + "grad_norm": 1.1433610916137695, + "learning_rate": 4.660017062986782e-06, + "loss": 0.4488, + "step": 25796 + }, + { + "epoch": 1.7313177410154021, + "grad_norm": 1.0909218788146973, + "learning_rate": 4.655436470961333e-06, + "loss": 0.514, + "step": 25798 + }, + { + "epoch": 1.731451964699171, + "grad_norm": 1.041691780090332, + "learning_rate": 4.650858021377352e-06, + "loss": 0.4272, + "step": 25800 + }, + { + "epoch": 1.7315861883829402, + "grad_norm": 0.889392077922821, + "learning_rate": 4.64628171445115e-06, + "loss": 0.4671, + "step": 25802 + }, + { + "epoch": 1.7317204120667091, + "grad_norm": 0.9627600908279419, + "learning_rate": 4.641707550398966e-06, + "loss": 0.4055, + "step": 25804 + }, + { + "epoch": 1.7318546357504783, + "grad_norm": 1.0190232992172241, + "learning_rate": 4.6371355294368865e-06, + "loss": 0.5, + "step": 25806 + }, + { + "epoch": 1.7319888594342472, + "grad_norm": 1.1649503707885742, + "learning_rate": 4.632565651780968e-06, + "loss": 0.4778, + "step": 25808 + }, + { + "epoch": 1.732123083118016, + "grad_norm": 1.000589370727539, + "learning_rate": 4.627997917647098e-06, + "loss": 0.428, + "step": 25810 + }, + { + "epoch": 1.732257306801785, + "grad_norm": 1.0200680494308472, + "learning_rate": 4.623432327251109e-06, + "loss": 0.5022, + "step": 25812 + }, + { + "epoch": 1.7323915304855542, + "grad_norm": 0.9837169051170349, + "learning_rate": 4.618868880808725e-06, + "loss": 0.4593, + "step": 25814 + }, + { + "epoch": 1.7325257541693233, + "grad_norm": 1.0563697814941406, + "learning_rate": 4.614307578535537e-06, + "loss": 0.3936, + "step": 25816 + }, + { + "epoch": 1.7326599778530922, + "grad_norm": 1.0165019035339355, + "learning_rate": 4.6097484206470756e-06, + "loss": 0.4896, + "step": 25818 + }, + { + "epoch": 1.7327942015368611, + "grad_norm": 1.0807873010635376, + "learning_rate": 4.605191407358733e-06, + "loss": 0.4371, + "step": 25820 + }, + { + "epoch": 1.73292842522063, + "grad_norm": 1.0654383897781372, + "learning_rate": 4.600636538885844e-06, + "loss": 0.4853, + "step": 25822 + }, + { + "epoch": 1.7330626489043992, + "grad_norm": 1.1478780508041382, + "learning_rate": 4.596083815443602e-06, + "loss": 0.5305, + "step": 25824 + }, + { + "epoch": 1.7331968725881683, + "grad_norm": 1.028700590133667, + "learning_rate": 4.5915332372471195e-06, + "loss": 0.4778, + "step": 25826 + }, + { + "epoch": 1.7333310962719373, + "grad_norm": 1.0302728414535522, + "learning_rate": 4.586984804511385e-06, + "loss": 0.4407, + "step": 25828 + }, + { + "epoch": 1.7334653199557062, + "grad_norm": 1.2134575843811035, + "learning_rate": 4.5824385174513316e-06, + "loss": 0.523, + "step": 25830 + }, + { + "epoch": 1.733599543639475, + "grad_norm": 0.983025074005127, + "learning_rate": 4.577894376281744e-06, + "loss": 0.413, + "step": 25832 + }, + { + "epoch": 1.733733767323244, + "grad_norm": 1.09456467628479, + "learning_rate": 4.573352381217333e-06, + "loss": 0.459, + "step": 25834 + }, + { + "epoch": 1.7338679910070132, + "grad_norm": 1.0942330360412598, + "learning_rate": 4.568812532472683e-06, + "loss": 0.4829, + "step": 25836 + }, + { + "epoch": 1.7340022146907823, + "grad_norm": 1.0838247537612915, + "learning_rate": 4.564274830262316e-06, + "loss": 0.4836, + "step": 25838 + }, + { + "epoch": 1.7341364383745512, + "grad_norm": 1.0898464918136597, + "learning_rate": 4.559739274800606e-06, + "loss": 0.4746, + "step": 25840 + }, + { + "epoch": 1.7342706620583201, + "grad_norm": 1.0721734762191772, + "learning_rate": 4.555205866301876e-06, + "loss": 0.416, + "step": 25842 + }, + { + "epoch": 1.734404885742089, + "grad_norm": 1.0424118041992188, + "learning_rate": 4.5506746049802925e-06, + "loss": 0.5083, + "step": 25844 + }, + { + "epoch": 1.7345391094258582, + "grad_norm": 1.02657151222229, + "learning_rate": 4.546145491049969e-06, + "loss": 0.4488, + "step": 25846 + }, + { + "epoch": 1.7346733331096273, + "grad_norm": 1.0207029581069946, + "learning_rate": 4.541618524724878e-06, + "loss": 0.4448, + "step": 25848 + }, + { + "epoch": 1.7348075567933963, + "grad_norm": 1.104272723197937, + "learning_rate": 4.537093706218937e-06, + "loss": 0.4415, + "step": 25850 + }, + { + "epoch": 1.7349417804771652, + "grad_norm": 0.89890056848526, + "learning_rate": 4.532571035745914e-06, + "loss": 0.4176, + "step": 25852 + }, + { + "epoch": 1.735076004160934, + "grad_norm": 0.9957574009895325, + "learning_rate": 4.528050513519505e-06, + "loss": 0.4607, + "step": 25854 + }, + { + "epoch": 1.7352102278447032, + "grad_norm": 0.9569705724716187, + "learning_rate": 4.523532139753278e-06, + "loss": 0.4531, + "step": 25856 + }, + { + "epoch": 1.7353444515284722, + "grad_norm": 0.9871608018875122, + "learning_rate": 4.519015914660746e-06, + "loss": 0.4515, + "step": 25858 + }, + { + "epoch": 1.7354786752122413, + "grad_norm": 1.1361002922058105, + "learning_rate": 4.514501838455276e-06, + "loss": 0.4474, + "step": 25860 + }, + { + "epoch": 1.7356128988960102, + "grad_norm": 1.1392649412155151, + "learning_rate": 4.5099899113501545e-06, + "loss": 0.4414, + "step": 25862 + }, + { + "epoch": 1.7357471225797791, + "grad_norm": 1.0350909233093262, + "learning_rate": 4.505480133558543e-06, + "loss": 0.4694, + "step": 25864 + }, + { + "epoch": 1.735881346263548, + "grad_norm": 1.1096014976501465, + "learning_rate": 4.500972505293544e-06, + "loss": 0.4315, + "step": 25866 + }, + { + "epoch": 1.7360155699473172, + "grad_norm": 0.9327784776687622, + "learning_rate": 4.4964670267681146e-06, + "loss": 0.4408, + "step": 25868 + }, + { + "epoch": 1.7361497936310863, + "grad_norm": 0.9927994608879089, + "learning_rate": 4.491963698195162e-06, + "loss": 0.4367, + "step": 25870 + }, + { + "epoch": 1.7362840173148553, + "grad_norm": 1.0583723783493042, + "learning_rate": 4.487462519787416e-06, + "loss": 0.4304, + "step": 25872 + }, + { + "epoch": 1.7364182409986242, + "grad_norm": 1.0991770029067993, + "learning_rate": 4.482963491757574e-06, + "loss": 0.5365, + "step": 25874 + }, + { + "epoch": 1.736552464682393, + "grad_norm": 1.037761926651001, + "learning_rate": 4.4784666143181996e-06, + "loss": 0.4912, + "step": 25876 + }, + { + "epoch": 1.7366866883661622, + "grad_norm": 1.0798840522766113, + "learning_rate": 4.473971887681772e-06, + "loss": 0.4917, + "step": 25878 + }, + { + "epoch": 1.7368209120499312, + "grad_norm": 1.3390872478485107, + "learning_rate": 4.469479312060643e-06, + "loss": 0.4258, + "step": 25880 + }, + { + "epoch": 1.7369551357337003, + "grad_norm": 1.0587348937988281, + "learning_rate": 4.464988887667087e-06, + "loss": 0.4164, + "step": 25882 + }, + { + "epoch": 1.7370893594174692, + "grad_norm": 1.1341004371643066, + "learning_rate": 4.460500614713259e-06, + "loss": 0.4923, + "step": 25884 + }, + { + "epoch": 1.7372235831012381, + "grad_norm": 0.9715561866760254, + "learning_rate": 4.456014493411237e-06, + "loss": 0.4884, + "step": 25886 + }, + { + "epoch": 1.737357806785007, + "grad_norm": 0.9592590928077698, + "learning_rate": 4.4515305239729685e-06, + "loss": 0.483, + "step": 25888 + }, + { + "epoch": 1.7374920304687762, + "grad_norm": 1.0191220045089722, + "learning_rate": 4.447048706610318e-06, + "loss": 0.4264, + "step": 25890 + }, + { + "epoch": 1.7376262541525453, + "grad_norm": 1.0349215269088745, + "learning_rate": 4.442569041535039e-06, + "loss": 0.4431, + "step": 25892 + }, + { + "epoch": 1.7377604778363143, + "grad_norm": 1.1754120588302612, + "learning_rate": 4.4380915289587825e-06, + "loss": 0.5029, + "step": 25894 + }, + { + "epoch": 1.7378947015200832, + "grad_norm": 1.0315639972686768, + "learning_rate": 4.433616169093113e-06, + "loss": 0.4927, + "step": 25896 + }, + { + "epoch": 1.738028925203852, + "grad_norm": 1.0900938510894775, + "learning_rate": 4.429142962149474e-06, + "loss": 0.5482, + "step": 25898 + }, + { + "epoch": 1.7381631488876212, + "grad_norm": 0.9947560429573059, + "learning_rate": 4.424671908339223e-06, + "loss": 0.4212, + "step": 25900 + }, + { + "epoch": 1.7382973725713904, + "grad_norm": 0.999907910823822, + "learning_rate": 4.420203007873597e-06, + "loss": 0.5176, + "step": 25902 + }, + { + "epoch": 1.7384315962551593, + "grad_norm": 1.1585700511932373, + "learning_rate": 4.415736260963755e-06, + "loss": 0.4646, + "step": 25904 + }, + { + "epoch": 1.7385658199389282, + "grad_norm": 1.0315663814544678, + "learning_rate": 4.411271667820727e-06, + "loss": 0.4106, + "step": 25906 + }, + { + "epoch": 1.7387000436226971, + "grad_norm": 0.9721665978431702, + "learning_rate": 4.406809228655479e-06, + "loss": 0.4395, + "step": 25908 + }, + { + "epoch": 1.738834267306466, + "grad_norm": 0.9676013588905334, + "learning_rate": 4.4023489436788355e-06, + "loss": 0.5162, + "step": 25910 + }, + { + "epoch": 1.7389684909902352, + "grad_norm": 0.998012363910675, + "learning_rate": 4.397890813101546e-06, + "loss": 0.5001, + "step": 25912 + }, + { + "epoch": 1.7391027146740043, + "grad_norm": 1.006813406944275, + "learning_rate": 4.39343483713423e-06, + "loss": 0.4572, + "step": 25914 + }, + { + "epoch": 1.7392369383577733, + "grad_norm": 0.9706727862358093, + "learning_rate": 4.388981015987448e-06, + "loss": 0.4709, + "step": 25916 + }, + { + "epoch": 1.7393711620415422, + "grad_norm": 1.0576826333999634, + "learning_rate": 4.384529349871625e-06, + "loss": 0.4457, + "step": 25918 + }, + { + "epoch": 1.739505385725311, + "grad_norm": 1.560713291168213, + "learning_rate": 4.380079838997086e-06, + "loss": 0.42, + "step": 25920 + }, + { + "epoch": 1.7396396094090802, + "grad_norm": 1.1104066371917725, + "learning_rate": 4.37563248357406e-06, + "loss": 0.5065, + "step": 25922 + }, + { + "epoch": 1.7397738330928494, + "grad_norm": 0.9899181127548218, + "learning_rate": 4.371187283812689e-06, + "loss": 0.4128, + "step": 25924 + }, + { + "epoch": 1.7399080567766183, + "grad_norm": 0.9700545072555542, + "learning_rate": 4.366744239922998e-06, + "loss": 0.4631, + "step": 25926 + }, + { + "epoch": 1.7400422804603872, + "grad_norm": 1.1491373777389526, + "learning_rate": 4.362303352114905e-06, + "loss": 0.4641, + "step": 25928 + }, + { + "epoch": 1.7401765041441561, + "grad_norm": 1.3101502656936646, + "learning_rate": 4.357864620598229e-06, + "loss": 0.4574, + "step": 25930 + }, + { + "epoch": 1.7403107278279253, + "grad_norm": 0.8173026442527771, + "learning_rate": 4.3534280455827024e-06, + "loss": 0.4014, + "step": 25932 + }, + { + "epoch": 1.7404449515116942, + "grad_norm": 1.0041718482971191, + "learning_rate": 4.348993627277936e-06, + "loss": 0.4621, + "step": 25934 + }, + { + "epoch": 1.7405791751954633, + "grad_norm": 0.9188247919082642, + "learning_rate": 4.3445613658934624e-06, + "loss": 0.4776, + "step": 25936 + }, + { + "epoch": 1.7407133988792323, + "grad_norm": 1.0293664932250977, + "learning_rate": 4.340131261638686e-06, + "loss": 0.4991, + "step": 25938 + }, + { + "epoch": 1.7408476225630012, + "grad_norm": 0.984596848487854, + "learning_rate": 4.335703314722916e-06, + "loss": 0.4099, + "step": 25940 + }, + { + "epoch": 1.74098184624677, + "grad_norm": 1.1050786972045898, + "learning_rate": 4.331277525355365e-06, + "loss": 0.441, + "step": 25942 + }, + { + "epoch": 1.7411160699305392, + "grad_norm": 1.0016874074935913, + "learning_rate": 4.326853893745153e-06, + "loss": 0.413, + "step": 25944 + }, + { + "epoch": 1.7412502936143084, + "grad_norm": 1.119035005569458, + "learning_rate": 4.3224324201012854e-06, + "loss": 0.44, + "step": 25946 + }, + { + "epoch": 1.7413845172980773, + "grad_norm": 0.9717772006988525, + "learning_rate": 4.3180131046326626e-06, + "loss": 0.4439, + "step": 25948 + }, + { + "epoch": 1.7415187409818462, + "grad_norm": 1.0132006406784058, + "learning_rate": 4.313595947548082e-06, + "loss": 0.4931, + "step": 25950 + }, + { + "epoch": 1.7416529646656151, + "grad_norm": 1.0019216537475586, + "learning_rate": 4.309180949056269e-06, + "loss": 0.4531, + "step": 25952 + }, + { + "epoch": 1.7417871883493843, + "grad_norm": 1.0533734560012817, + "learning_rate": 4.304768109365792e-06, + "loss": 0.43, + "step": 25954 + }, + { + "epoch": 1.7419214120331532, + "grad_norm": 1.076453447341919, + "learning_rate": 4.300357428685187e-06, + "loss": 0.5029, + "step": 25956 + }, + { + "epoch": 1.7420556357169223, + "grad_norm": 1.091066598892212, + "learning_rate": 4.295948907222824e-06, + "loss": 0.4205, + "step": 25958 + }, + { + "epoch": 1.7421898594006913, + "grad_norm": 1.061374545097351, + "learning_rate": 4.2915425451869884e-06, + "loss": 0.4575, + "step": 25960 + }, + { + "epoch": 1.7423240830844602, + "grad_norm": 1.1788666248321533, + "learning_rate": 4.287138342785896e-06, + "loss": 0.4443, + "step": 25962 + }, + { + "epoch": 1.742458306768229, + "grad_norm": 0.9478891491889954, + "learning_rate": 4.2827363002276135e-06, + "loss": 0.4862, + "step": 25964 + }, + { + "epoch": 1.7425925304519982, + "grad_norm": 1.106594204902649, + "learning_rate": 4.278336417720163e-06, + "loss": 0.513, + "step": 25966 + }, + { + "epoch": 1.7427267541357674, + "grad_norm": 1.0672873258590698, + "learning_rate": 4.273938695471391e-06, + "loss": 0.5416, + "step": 25968 + }, + { + "epoch": 1.7428609778195363, + "grad_norm": 0.9814993143081665, + "learning_rate": 4.269543133689108e-06, + "loss": 0.4372, + "step": 25970 + }, + { + "epoch": 1.7429952015033052, + "grad_norm": 1.0161956548690796, + "learning_rate": 4.265149732580981e-06, + "loss": 0.4437, + "step": 25972 + }, + { + "epoch": 1.7431294251870741, + "grad_norm": 1.0389659404754639, + "learning_rate": 4.2607584923546e-06, + "loss": 0.4365, + "step": 25974 + }, + { + "epoch": 1.7432636488708433, + "grad_norm": 1.1154420375823975, + "learning_rate": 4.256369413217442e-06, + "loss": 0.5105, + "step": 25976 + }, + { + "epoch": 1.7433978725546124, + "grad_norm": 1.0543137788772583, + "learning_rate": 4.251982495376872e-06, + "loss": 0.4155, + "step": 25978 + }, + { + "epoch": 1.7435320962383813, + "grad_norm": 1.1536834239959717, + "learning_rate": 4.247597739040166e-06, + "loss": 0.4518, + "step": 25980 + }, + { + "epoch": 1.7436663199221503, + "grad_norm": 1.071794867515564, + "learning_rate": 4.2432151444145085e-06, + "loss": 0.4511, + "step": 25982 + }, + { + "epoch": 1.7438005436059192, + "grad_norm": 1.0777956247329712, + "learning_rate": 4.238834711706952e-06, + "loss": 0.4817, + "step": 25984 + }, + { + "epoch": 1.743934767289688, + "grad_norm": 0.9815927147865295, + "learning_rate": 4.234456441124474e-06, + "loss": 0.4578, + "step": 25986 + }, + { + "epoch": 1.7440689909734572, + "grad_norm": 1.0425565242767334, + "learning_rate": 4.230080332873926e-06, + "loss": 0.4444, + "step": 25988 + }, + { + "epoch": 1.7442032146572264, + "grad_norm": 1.0789756774902344, + "learning_rate": 4.225706387162087e-06, + "loss": 0.4435, + "step": 25990 + }, + { + "epoch": 1.7443374383409953, + "grad_norm": 1.0805772542953491, + "learning_rate": 4.221334604195604e-06, + "loss": 0.4818, + "step": 25992 + }, + { + "epoch": 1.7444716620247642, + "grad_norm": 1.0187358856201172, + "learning_rate": 4.21696498418106e-06, + "loss": 0.399, + "step": 25994 + }, + { + "epoch": 1.7446058857085331, + "grad_norm": 0.9971460103988647, + "learning_rate": 4.212597527324869e-06, + "loss": 0.4628, + "step": 25996 + }, + { + "epoch": 1.7447401093923023, + "grad_norm": 1.0509072542190552, + "learning_rate": 4.2082322338334204e-06, + "loss": 0.4876, + "step": 25998 + }, + { + "epoch": 1.7448743330760714, + "grad_norm": 0.9012466073036194, + "learning_rate": 4.203869103912944e-06, + "loss": 0.418, + "step": 26000 + }, + { + "epoch": 1.7450085567598403, + "grad_norm": 1.104740023612976, + "learning_rate": 4.1995081377696035e-06, + "loss": 0.4798, + "step": 26002 + }, + { + "epoch": 1.7451427804436093, + "grad_norm": 1.0096579790115356, + "learning_rate": 4.195149335609444e-06, + "loss": 0.4943, + "step": 26004 + }, + { + "epoch": 1.7452770041273782, + "grad_norm": 0.9526199698448181, + "learning_rate": 4.190792697638407e-06, + "loss": 0.4269, + "step": 26006 + }, + { + "epoch": 1.7454112278111473, + "grad_norm": 1.027766466140747, + "learning_rate": 4.186438224062322e-06, + "loss": 0.4646, + "step": 26008 + }, + { + "epoch": 1.7455454514949162, + "grad_norm": 1.0538530349731445, + "learning_rate": 4.182085915086958e-06, + "loss": 0.4501, + "step": 26010 + }, + { + "epoch": 1.7456796751786854, + "grad_norm": 0.9318313002586365, + "learning_rate": 4.1777357709179345e-06, + "loss": 0.4863, + "step": 26012 + }, + { + "epoch": 1.7458138988624543, + "grad_norm": 1.1026418209075928, + "learning_rate": 4.1733877917607914e-06, + "loss": 0.5779, + "step": 26014 + }, + { + "epoch": 1.7459481225462232, + "grad_norm": 1.0496286153793335, + "learning_rate": 4.169041977820948e-06, + "loss": 0.4837, + "step": 26016 + }, + { + "epoch": 1.7460823462299921, + "grad_norm": 1.0120539665222168, + "learning_rate": 4.164698329303762e-06, + "loss": 0.469, + "step": 26018 + }, + { + "epoch": 1.7462165699137613, + "grad_norm": 0.9936702251434326, + "learning_rate": 4.1603568464144375e-06, + "loss": 0.4565, + "step": 26020 + }, + { + "epoch": 1.7463507935975304, + "grad_norm": 0.9050431847572327, + "learning_rate": 4.156017529358131e-06, + "loss": 0.4575, + "step": 26022 + }, + { + "epoch": 1.7464850172812993, + "grad_norm": 0.9547222852706909, + "learning_rate": 4.151680378339834e-06, + "loss": 0.4462, + "step": 26024 + }, + { + "epoch": 1.7466192409650683, + "grad_norm": 0.9970137476921082, + "learning_rate": 4.147345393564478e-06, + "loss": 0.459, + "step": 26026 + }, + { + "epoch": 1.7467534646488372, + "grad_norm": 1.0239909887313843, + "learning_rate": 4.1430125752368986e-06, + "loss": 0.4064, + "step": 26028 + }, + { + "epoch": 1.7468876883326063, + "grad_norm": 0.9084471464157104, + "learning_rate": 4.138681923561788e-06, + "loss": 0.4702, + "step": 26030 + }, + { + "epoch": 1.7470219120163752, + "grad_norm": 1.1111291646957397, + "learning_rate": 4.134353438743793e-06, + "loss": 0.4944, + "step": 26032 + }, + { + "epoch": 1.7471561357001444, + "grad_norm": 0.9989820718765259, + "learning_rate": 4.13002712098739e-06, + "loss": 0.4798, + "step": 26034 + }, + { + "epoch": 1.7472903593839133, + "grad_norm": 1.0311503410339355, + "learning_rate": 4.125702970497014e-06, + "loss": 0.4669, + "step": 26036 + }, + { + "epoch": 1.7474245830676822, + "grad_norm": 0.9694811105728149, + "learning_rate": 4.121380987476953e-06, + "loss": 0.4531, + "step": 26038 + }, + { + "epoch": 1.7475588067514511, + "grad_norm": 1.0406783819198608, + "learning_rate": 4.117061172131437e-06, + "loss": 0.4443, + "step": 26040 + }, + { + "epoch": 1.7476930304352203, + "grad_norm": 0.9970607161521912, + "learning_rate": 4.112743524664553e-06, + "loss": 0.3971, + "step": 26042 + }, + { + "epoch": 1.7478272541189894, + "grad_norm": 1.0036909580230713, + "learning_rate": 4.108428045280305e-06, + "loss": 0.4603, + "step": 26044 + }, + { + "epoch": 1.7479614778027583, + "grad_norm": 0.939334511756897, + "learning_rate": 4.104114734182584e-06, + "loss": 0.4816, + "step": 26046 + }, + { + "epoch": 1.7480957014865273, + "grad_norm": 1.4095072746276855, + "learning_rate": 4.0998035915751965e-06, + "loss": 0.4273, + "step": 26048 + }, + { + "epoch": 1.7482299251702962, + "grad_norm": 1.0417141914367676, + "learning_rate": 4.095494617661833e-06, + "loss": 0.4415, + "step": 26050 + }, + { + "epoch": 1.7483641488540653, + "grad_norm": 1.396823525428772, + "learning_rate": 4.0911878126460805e-06, + "loss": 0.4924, + "step": 26052 + }, + { + "epoch": 1.7484983725378345, + "grad_norm": 0.9825583100318909, + "learning_rate": 4.086883176731415e-06, + "loss": 0.439, + "step": 26054 + }, + { + "epoch": 1.7486325962216034, + "grad_norm": 1.0633548498153687, + "learning_rate": 4.082580710121248e-06, + "loss": 0.3985, + "step": 26056 + }, + { + "epoch": 1.7487668199053723, + "grad_norm": 1.0380185842514038, + "learning_rate": 4.078280413018843e-06, + "loss": 0.4545, + "step": 26058 + }, + { + "epoch": 1.7489010435891412, + "grad_norm": 0.9857343435287476, + "learning_rate": 4.0739822856273925e-06, + "loss": 0.429, + "step": 26060 + }, + { + "epoch": 1.7490352672729101, + "grad_norm": 1.0646899938583374, + "learning_rate": 4.069686328149969e-06, + "loss": 0.4804, + "step": 26062 + }, + { + "epoch": 1.7491694909566793, + "grad_norm": 1.0434508323669434, + "learning_rate": 4.065392540789553e-06, + "loss": 0.428, + "step": 26064 + }, + { + "epoch": 1.7493037146404484, + "grad_norm": 1.1091803312301636, + "learning_rate": 4.061100923749001e-06, + "loss": 0.4761, + "step": 26066 + }, + { + "epoch": 1.7494379383242173, + "grad_norm": 1.0577932596206665, + "learning_rate": 4.0568114772311035e-06, + "loss": 0.4923, + "step": 26068 + }, + { + "epoch": 1.7495721620079863, + "grad_norm": 1.0199114084243774, + "learning_rate": 4.052524201438523e-06, + "loss": 0.4548, + "step": 26070 + }, + { + "epoch": 1.7497063856917552, + "grad_norm": 1.0450282096862793, + "learning_rate": 4.048239096573819e-06, + "loss": 0.5336, + "step": 26072 + }, + { + "epoch": 1.7498406093755243, + "grad_norm": 0.8555244207382202, + "learning_rate": 4.043956162839452e-06, + "loss": 0.4252, + "step": 26074 + }, + { + "epoch": 1.7499748330592935, + "grad_norm": 1.0349127054214478, + "learning_rate": 4.0396754004377925e-06, + "loss": 0.4433, + "step": 26076 + }, + { + "epoch": 1.7501090567430624, + "grad_norm": 1.0625731945037842, + "learning_rate": 4.035396809571096e-06, + "loss": 0.4539, + "step": 26078 + }, + { + "epoch": 1.7502432804268313, + "grad_norm": 1.057253360748291, + "learning_rate": 4.0311203904415116e-06, + "loss": 0.4649, + "step": 26080 + }, + { + "epoch": 1.7503775041106002, + "grad_norm": 1.0560394525527954, + "learning_rate": 4.02684614325109e-06, + "loss": 0.4382, + "step": 26082 + }, + { + "epoch": 1.7505117277943694, + "grad_norm": 1.0100197792053223, + "learning_rate": 4.02257406820179e-06, + "loss": 0.4892, + "step": 26084 + }, + { + "epoch": 1.7506459514781383, + "grad_norm": 1.061964988708496, + "learning_rate": 4.0183041654954515e-06, + "loss": 0.4295, + "step": 26086 + }, + { + "epoch": 1.7507801751619074, + "grad_norm": 1.0337510108947754, + "learning_rate": 4.0140364353338286e-06, + "loss": 0.4432, + "step": 26088 + }, + { + "epoch": 1.7509143988456763, + "grad_norm": 0.8317381739616394, + "learning_rate": 4.009770877918556e-06, + "loss": 0.4072, + "step": 26090 + }, + { + "epoch": 1.7510486225294453, + "grad_norm": 1.0885119438171387, + "learning_rate": 4.005507493451171e-06, + "loss": 0.5211, + "step": 26092 + }, + { + "epoch": 1.7511828462132142, + "grad_norm": 1.3589690923690796, + "learning_rate": 4.0012462821331145e-06, + "loss": 0.4578, + "step": 26094 + }, + { + "epoch": 1.7513170698969833, + "grad_norm": 0.9653885364532471, + "learning_rate": 3.996987244165712e-06, + "loss": 0.4643, + "step": 26096 + }, + { + "epoch": 1.7514512935807525, + "grad_norm": 2.2596049308776855, + "learning_rate": 3.9927303797502096e-06, + "loss": 0.4178, + "step": 26098 + }, + { + "epoch": 1.7515855172645214, + "grad_norm": 1.04008150100708, + "learning_rate": 3.988475689087723e-06, + "loss": 0.4817, + "step": 26100 + }, + { + "epoch": 1.7517197409482903, + "grad_norm": 1.0107393264770508, + "learning_rate": 3.984223172379287e-06, + "loss": 0.4844, + "step": 26102 + }, + { + "epoch": 1.7518539646320592, + "grad_norm": 0.9756828546524048, + "learning_rate": 3.979972829825807e-06, + "loss": 0.4817, + "step": 26104 + }, + { + "epoch": 1.7519881883158284, + "grad_norm": 1.2110542058944702, + "learning_rate": 3.975724661628128e-06, + "loss": 0.4734, + "step": 26106 + }, + { + "epoch": 1.7521224119995973, + "grad_norm": 1.1260565519332886, + "learning_rate": 3.971478667986955e-06, + "loss": 0.4805, + "step": 26108 + }, + { + "epoch": 1.7522566356833664, + "grad_norm": 1.2594767808914185, + "learning_rate": 3.967234849102907e-06, + "loss": 0.5398, + "step": 26110 + }, + { + "epoch": 1.7523908593671353, + "grad_norm": 0.9751372337341309, + "learning_rate": 3.962993205176479e-06, + "loss": 0.4528, + "step": 26112 + }, + { + "epoch": 1.7525250830509043, + "grad_norm": 1.1253643035888672, + "learning_rate": 3.958753736408105e-06, + "loss": 0.5152, + "step": 26114 + }, + { + "epoch": 1.7526593067346732, + "grad_norm": 1.102760910987854, + "learning_rate": 3.954516442998074e-06, + "loss": 0.4306, + "step": 26116 + }, + { + "epoch": 1.7527935304184423, + "grad_norm": 1.0021374225616455, + "learning_rate": 3.9502813251466096e-06, + "loss": 0.4496, + "step": 26118 + }, + { + "epoch": 1.7529277541022115, + "grad_norm": 1.0082281827926636, + "learning_rate": 3.946048383053786e-06, + "loss": 0.5659, + "step": 26120 + }, + { + "epoch": 1.7530619777859804, + "grad_norm": 0.8066741824150085, + "learning_rate": 3.941817616919624e-06, + "loss": 0.4263, + "step": 26122 + }, + { + "epoch": 1.7531962014697493, + "grad_norm": 0.8860553503036499, + "learning_rate": 3.937589026944e-06, + "loss": 0.5009, + "step": 26124 + }, + { + "epoch": 1.7533304251535182, + "grad_norm": 1.0312938690185547, + "learning_rate": 3.9333626133267244e-06, + "loss": 0.5285, + "step": 26126 + }, + { + "epoch": 1.7534646488372874, + "grad_norm": 0.9705163836479187, + "learning_rate": 3.929138376267477e-06, + "loss": 0.5149, + "step": 26128 + }, + { + "epoch": 1.7535988725210565, + "grad_norm": 0.9016320705413818, + "learning_rate": 3.924916315965854e-06, + "loss": 0.4818, + "step": 26130 + }, + { + "epoch": 1.7537330962048254, + "grad_norm": 1.1750874519348145, + "learning_rate": 3.920696432621318e-06, + "loss": 0.4783, + "step": 26132 + }, + { + "epoch": 1.7538673198885943, + "grad_norm": 1.048884630203247, + "learning_rate": 3.916478726433276e-06, + "loss": 0.5275, + "step": 26134 + }, + { + "epoch": 1.7540015435723633, + "grad_norm": 1.061607003211975, + "learning_rate": 3.91226319760099e-06, + "loss": 0.4817, + "step": 26136 + }, + { + "epoch": 1.7541357672561322, + "grad_norm": 0.9778168201446533, + "learning_rate": 3.908049846323647e-06, + "loss": 0.5421, + "step": 26138 + }, + { + "epoch": 1.7542699909399013, + "grad_norm": 1.0087391138076782, + "learning_rate": 3.903838672800297e-06, + "loss": 0.4944, + "step": 26140 + }, + { + "epoch": 1.7544042146236705, + "grad_norm": 1.061140775680542, + "learning_rate": 3.899629677229943e-06, + "loss": 0.5601, + "step": 26142 + }, + { + "epoch": 1.7545384383074394, + "grad_norm": 1.113389015197754, + "learning_rate": 3.89542285981142e-06, + "loss": 0.4302, + "step": 26144 + }, + { + "epoch": 1.7546726619912083, + "grad_norm": 1.1354550123214722, + "learning_rate": 3.89121822074352e-06, + "loss": 0.4703, + "step": 26146 + }, + { + "epoch": 1.7548068856749772, + "grad_norm": 0.956482470035553, + "learning_rate": 3.887015760224877e-06, + "loss": 0.4745, + "step": 26148 + }, + { + "epoch": 1.7549411093587464, + "grad_norm": 1.0160974264144897, + "learning_rate": 3.8828154784540714e-06, + "loss": 0.4272, + "step": 26150 + }, + { + "epoch": 1.7550753330425155, + "grad_norm": 1.0287387371063232, + "learning_rate": 3.87861737562954e-06, + "loss": 0.5053, + "step": 26152 + }, + { + "epoch": 1.7552095567262844, + "grad_norm": 1.0463069677352905, + "learning_rate": 3.87442145194965e-06, + "loss": 0.541, + "step": 26154 + }, + { + "epoch": 1.7553437804100533, + "grad_norm": 0.9868055582046509, + "learning_rate": 3.87022770761265e-06, + "loss": 0.4387, + "step": 26156 + }, + { + "epoch": 1.7554780040938223, + "grad_norm": 1.0721690654754639, + "learning_rate": 3.866036142816676e-06, + "loss": 0.4508, + "step": 26158 + }, + { + "epoch": 1.7556122277775914, + "grad_norm": 0.9048008918762207, + "learning_rate": 3.861846757759768e-06, + "loss": 0.45, + "step": 26160 + }, + { + "epoch": 1.7557464514613603, + "grad_norm": 1.0534013509750366, + "learning_rate": 3.857659552639881e-06, + "loss": 0.4851, + "step": 26162 + }, + { + "epoch": 1.7558806751451295, + "grad_norm": 1.102059006690979, + "learning_rate": 3.853474527654849e-06, + "loss": 0.4907, + "step": 26164 + }, + { + "epoch": 1.7560148988288984, + "grad_norm": 0.8822693824768066, + "learning_rate": 3.849291683002398e-06, + "loss": 0.4201, + "step": 26166 + }, + { + "epoch": 1.7561491225126673, + "grad_norm": 0.9882935285568237, + "learning_rate": 3.845111018880165e-06, + "loss": 0.4663, + "step": 26168 + }, + { + "epoch": 1.7562833461964362, + "grad_norm": 0.9759923219680786, + "learning_rate": 3.84093253548567e-06, + "loss": 0.4165, + "step": 26170 + }, + { + "epoch": 1.7564175698802054, + "grad_norm": 0.9933663606643677, + "learning_rate": 3.836756233016348e-06, + "loss": 0.4877, + "step": 26172 + }, + { + "epoch": 1.7565517935639745, + "grad_norm": 1.0421911478042603, + "learning_rate": 3.832582111669525e-06, + "loss": 0.4357, + "step": 26174 + }, + { + "epoch": 1.7566860172477434, + "grad_norm": 0.9907096028327942, + "learning_rate": 3.828410171642404e-06, + "loss": 0.4631, + "step": 26176 + }, + { + "epoch": 1.7568202409315123, + "grad_norm": 0.8957405090332031, + "learning_rate": 3.824240413132107e-06, + "loss": 0.4537, + "step": 26178 + }, + { + "epoch": 1.7569544646152813, + "grad_norm": 1.045711874961853, + "learning_rate": 3.820072836335658e-06, + "loss": 0.4105, + "step": 26180 + }, + { + "epoch": 1.7570886882990504, + "grad_norm": 1.179186224937439, + "learning_rate": 3.815907441449951e-06, + "loss": 0.4719, + "step": 26182 + }, + { + "epoch": 1.7572229119828193, + "grad_norm": 1.2131788730621338, + "learning_rate": 3.811744228671815e-06, + "loss": 0.4861, + "step": 26184 + }, + { + "epoch": 1.7573571356665885, + "grad_norm": 1.085301160812378, + "learning_rate": 3.807583198197923e-06, + "loss": 0.4765, + "step": 26186 + }, + { + "epoch": 1.7574913593503574, + "grad_norm": 1.029398798942566, + "learning_rate": 3.8034243502249045e-06, + "loss": 0.4611, + "step": 26188 + }, + { + "epoch": 1.7576255830341263, + "grad_norm": 1.011246919631958, + "learning_rate": 3.799267684949226e-06, + "loss": 0.5085, + "step": 26190 + }, + { + "epoch": 1.7577598067178952, + "grad_norm": 1.0009100437164307, + "learning_rate": 3.795113202567313e-06, + "loss": 0.4058, + "step": 26192 + }, + { + "epoch": 1.7578940304016644, + "grad_norm": 0.7475231289863586, + "learning_rate": 3.7909609032754424e-06, + "loss": 0.4094, + "step": 26194 + }, + { + "epoch": 1.7580282540854335, + "grad_norm": 1.1544121503829956, + "learning_rate": 3.786810787269807e-06, + "loss": 0.4569, + "step": 26196 + }, + { + "epoch": 1.7581624777692024, + "grad_norm": 0.9361860752105713, + "learning_rate": 3.782662854746477e-06, + "loss": 0.5096, + "step": 26198 + }, + { + "epoch": 1.7582967014529713, + "grad_norm": 1.1230194568634033, + "learning_rate": 3.7785171059014468e-06, + "loss": 0.4443, + "step": 26200 + }, + { + "epoch": 1.7584309251367403, + "grad_norm": 0.8865429162979126, + "learning_rate": 3.7743735409305982e-06, + "loss": 0.4462, + "step": 26202 + }, + { + "epoch": 1.7585651488205094, + "grad_norm": 1.0899479389190674, + "learning_rate": 3.7702321600297017e-06, + "loss": 0.4709, + "step": 26204 + }, + { + "epoch": 1.7586993725042785, + "grad_norm": 0.9332501292228699, + "learning_rate": 3.7660929633944186e-06, + "loss": 0.4255, + "step": 26206 + }, + { + "epoch": 1.7588335961880475, + "grad_norm": 1.083202838897705, + "learning_rate": 3.7619559512203363e-06, + "loss": 0.4086, + "step": 26208 + }, + { + "epoch": 1.7589678198718164, + "grad_norm": 1.1723979711532593, + "learning_rate": 3.757821123702904e-06, + "loss": 0.5048, + "step": 26210 + }, + { + "epoch": 1.7591020435555853, + "grad_norm": 0.9791699647903442, + "learning_rate": 3.7536884810375094e-06, + "loss": 0.4624, + "step": 26212 + }, + { + "epoch": 1.7592362672393542, + "grad_norm": 1.0547990798950195, + "learning_rate": 3.7495580234193804e-06, + "loss": 0.4836, + "step": 26214 + }, + { + "epoch": 1.7593704909231234, + "grad_norm": 1.0521347522735596, + "learning_rate": 3.7454297510436885e-06, + "loss": 0.4835, + "step": 26216 + }, + { + "epoch": 1.7595047146068925, + "grad_norm": 0.9637994170188904, + "learning_rate": 3.7413036641054834e-06, + "loss": 0.4173, + "step": 26218 + }, + { + "epoch": 1.7596389382906614, + "grad_norm": 1.0964466333389282, + "learning_rate": 3.7371797627997194e-06, + "loss": 0.4228, + "step": 26220 + }, + { + "epoch": 1.7597731619744303, + "grad_norm": 0.9575738906860352, + "learning_rate": 3.7330580473212473e-06, + "loss": 0.4499, + "step": 26222 + }, + { + "epoch": 1.7599073856581993, + "grad_norm": 1.091765284538269, + "learning_rate": 3.728938517864794e-06, + "loss": 0.4908, + "step": 26224 + }, + { + "epoch": 1.7600416093419684, + "grad_norm": 1.1609419584274292, + "learning_rate": 3.724821174625004e-06, + "loss": 0.4288, + "step": 26226 + }, + { + "epoch": 1.7601758330257375, + "grad_norm": 0.9989089965820312, + "learning_rate": 3.720706017796427e-06, + "loss": 0.433, + "step": 26228 + }, + { + "epoch": 1.7603100567095065, + "grad_norm": 1.0468043088912964, + "learning_rate": 3.71659304757348e-06, + "loss": 0.4575, + "step": 26230 + }, + { + "epoch": 1.7604442803932754, + "grad_norm": 1.0158153772354126, + "learning_rate": 3.7124822641505017e-06, + "loss": 0.3912, + "step": 26232 + }, + { + "epoch": 1.7605785040770443, + "grad_norm": 1.0709056854248047, + "learning_rate": 3.7083736677217206e-06, + "loss": 0.4414, + "step": 26234 + }, + { + "epoch": 1.7607127277608134, + "grad_norm": 1.1032527685165405, + "learning_rate": 3.704267258481242e-06, + "loss": 0.4658, + "step": 26236 + }, + { + "epoch": 1.7608469514445824, + "grad_norm": 0.9846126437187195, + "learning_rate": 3.7001630366231112e-06, + "loss": 0.419, + "step": 26238 + }, + { + "epoch": 1.7609811751283515, + "grad_norm": 1.036513090133667, + "learning_rate": 3.696061002341217e-06, + "loss": 0.4931, + "step": 26240 + }, + { + "epoch": 1.7611153988121204, + "grad_norm": 1.1536115407943726, + "learning_rate": 3.6919611558294098e-06, + "loss": 0.4925, + "step": 26242 + }, + { + "epoch": 1.7612496224958893, + "grad_norm": 0.8856650590896606, + "learning_rate": 3.6878634972813576e-06, + "loss": 0.3997, + "step": 26244 + }, + { + "epoch": 1.7613838461796583, + "grad_norm": 1.0974348783493042, + "learning_rate": 3.6837680268906992e-06, + "loss": 0.4774, + "step": 26246 + }, + { + "epoch": 1.7615180698634274, + "grad_norm": 0.9839081764221191, + "learning_rate": 3.679674744850914e-06, + "loss": 0.4007, + "step": 26248 + }, + { + "epoch": 1.7616522935471965, + "grad_norm": 1.161922574043274, + "learning_rate": 3.6755836513554185e-06, + "loss": 0.4659, + "step": 26250 + }, + { + "epoch": 1.7617865172309655, + "grad_norm": 0.9850311279296875, + "learning_rate": 3.6714947465975035e-06, + "loss": 0.442, + "step": 26252 + }, + { + "epoch": 1.7619207409147344, + "grad_norm": 1.0139830112457275, + "learning_rate": 3.6674080307703586e-06, + "loss": 0.4257, + "step": 26254 + }, + { + "epoch": 1.7620549645985033, + "grad_norm": 1.1162172555923462, + "learning_rate": 3.6633235040670675e-06, + "loss": 0.4695, + "step": 26256 + }, + { + "epoch": 1.7621891882822724, + "grad_norm": 1.002410888671875, + "learning_rate": 3.6592411666806326e-06, + "loss": 0.5363, + "step": 26258 + }, + { + "epoch": 1.7623234119660414, + "grad_norm": 1.111937403678894, + "learning_rate": 3.655161018803932e-06, + "loss": 0.4447, + "step": 26260 + }, + { + "epoch": 1.7624576356498105, + "grad_norm": 1.1555240154266357, + "learning_rate": 3.6510830606297343e-06, + "loss": 0.5071, + "step": 26262 + }, + { + "epoch": 1.7625918593335794, + "grad_norm": 1.0110962390899658, + "learning_rate": 3.6470072923507125e-06, + "loss": 0.4434, + "step": 26264 + }, + { + "epoch": 1.7627260830173483, + "grad_norm": 1.3436285257339478, + "learning_rate": 3.6429337141594578e-06, + "loss": 0.5013, + "step": 26266 + }, + { + "epoch": 1.7628603067011173, + "grad_norm": 1.1013902425765991, + "learning_rate": 3.638862326248421e-06, + "loss": 0.4359, + "step": 26268 + }, + { + "epoch": 1.7629945303848864, + "grad_norm": 1.0049915313720703, + "learning_rate": 3.634793128809988e-06, + "loss": 0.4744, + "step": 26270 + }, + { + "epoch": 1.7631287540686555, + "grad_norm": 1.0206397771835327, + "learning_rate": 3.6307261220363876e-06, + "loss": 0.4962, + "step": 26272 + }, + { + "epoch": 1.7632629777524245, + "grad_norm": 1.2429076433181763, + "learning_rate": 3.626661306119805e-06, + "loss": 0.4501, + "step": 26274 + }, + { + "epoch": 1.7633972014361934, + "grad_norm": 1.0040055513381958, + "learning_rate": 3.6225986812522816e-06, + "loss": 0.4602, + "step": 26276 + }, + { + "epoch": 1.7635314251199623, + "grad_norm": 0.9989480972290039, + "learning_rate": 3.61853824762578e-06, + "loss": 0.487, + "step": 26278 + }, + { + "epoch": 1.7636656488037314, + "grad_norm": 0.9974848628044128, + "learning_rate": 3.6144800054321415e-06, + "loss": 0.4237, + "step": 26280 + }, + { + "epoch": 1.7637998724875004, + "grad_norm": 1.0944793224334717, + "learning_rate": 3.6104239548631127e-06, + "loss": 0.4444, + "step": 26282 + }, + { + "epoch": 1.7639340961712695, + "grad_norm": 1.0131146907806396, + "learning_rate": 3.6063700961103174e-06, + "loss": 0.5055, + "step": 26284 + }, + { + "epoch": 1.7640683198550384, + "grad_norm": 1.0708969831466675, + "learning_rate": 3.6023184293653143e-06, + "loss": 0.4939, + "step": 26286 + }, + { + "epoch": 1.7642025435388073, + "grad_norm": 0.9781100153923035, + "learning_rate": 3.5982689548195338e-06, + "loss": 0.432, + "step": 26288 + }, + { + "epoch": 1.7643367672225763, + "grad_norm": 1.0512197017669678, + "learning_rate": 3.594221672664294e-06, + "loss": 0.4748, + "step": 26290 + }, + { + "epoch": 1.7644709909063454, + "grad_norm": 1.1227929592132568, + "learning_rate": 3.590176583090821e-06, + "loss": 0.4332, + "step": 26292 + }, + { + "epoch": 1.7646052145901145, + "grad_norm": 1.1276625394821167, + "learning_rate": 3.586133686290255e-06, + "loss": 0.4524, + "step": 26294 + }, + { + "epoch": 1.7647394382738835, + "grad_norm": 1.1209701299667358, + "learning_rate": 3.582092982453589e-06, + "loss": 0.4821, + "step": 26296 + }, + { + "epoch": 1.7648736619576524, + "grad_norm": 1.0472911596298218, + "learning_rate": 3.5780544717717747e-06, + "loss": 0.4664, + "step": 26298 + }, + { + "epoch": 1.7650078856414213, + "grad_norm": 1.0919382572174072, + "learning_rate": 3.5740181544355875e-06, + "loss": 0.4555, + "step": 26300 + }, + { + "epoch": 1.7651421093251904, + "grad_norm": 1.0900133848190308, + "learning_rate": 3.5699840306357478e-06, + "loss": 0.5195, + "step": 26302 + }, + { + "epoch": 1.7652763330089596, + "grad_norm": 0.995741069316864, + "learning_rate": 3.565952100562869e-06, + "loss": 0.4058, + "step": 26304 + }, + { + "epoch": 1.7654105566927285, + "grad_norm": 1.073737621307373, + "learning_rate": 3.561922364407433e-06, + "loss": 0.563, + "step": 26306 + }, + { + "epoch": 1.7655447803764974, + "grad_norm": 1.0843747854232788, + "learning_rate": 3.557894822359864e-06, + "loss": 0.4034, + "step": 26308 + }, + { + "epoch": 1.7656790040602663, + "grad_norm": 1.110669732093811, + "learning_rate": 3.5538694746104274e-06, + "loss": 0.4691, + "step": 26310 + }, + { + "epoch": 1.7658132277440355, + "grad_norm": 0.9983861446380615, + "learning_rate": 3.5498463213493372e-06, + "loss": 0.459, + "step": 26312 + }, + { + "epoch": 1.7659474514278044, + "grad_norm": 0.9595610499382019, + "learning_rate": 3.545825362766653e-06, + "loss": 0.4353, + "step": 26314 + }, + { + "epoch": 1.7660816751115735, + "grad_norm": 1.0454193353652954, + "learning_rate": 3.541806599052383e-06, + "loss": 0.424, + "step": 26316 + }, + { + "epoch": 1.7662158987953425, + "grad_norm": 0.9348208904266357, + "learning_rate": 3.5377900303963986e-06, + "loss": 0.5193, + "step": 26318 + }, + { + "epoch": 1.7663501224791114, + "grad_norm": 1.0392781496047974, + "learning_rate": 3.533775656988464e-06, + "loss": 0.4424, + "step": 26320 + }, + { + "epoch": 1.7664843461628803, + "grad_norm": 1.0983844995498657, + "learning_rate": 3.5297634790182555e-06, + "loss": 0.4687, + "step": 26322 + }, + { + "epoch": 1.7666185698466494, + "grad_norm": 1.1866750717163086, + "learning_rate": 3.525753496675349e-06, + "loss": 0.5033, + "step": 26324 + }, + { + "epoch": 1.7667527935304186, + "grad_norm": 1.07992422580719, + "learning_rate": 3.5217457101492045e-06, + "loss": 0.4064, + "step": 26326 + }, + { + "epoch": 1.7668870172141875, + "grad_norm": 1.0780870914459229, + "learning_rate": 3.517740119629176e-06, + "loss": 0.4615, + "step": 26328 + }, + { + "epoch": 1.7670212408979564, + "grad_norm": 1.0299321413040161, + "learning_rate": 3.5137367253045173e-06, + "loss": 0.4651, + "step": 26330 + }, + { + "epoch": 1.7671554645817253, + "grad_norm": 0.8918436169624329, + "learning_rate": 3.509735527364394e-06, + "loss": 0.4433, + "step": 26332 + }, + { + "epoch": 1.7672896882654945, + "grad_norm": 1.1658223867416382, + "learning_rate": 3.5057365259978383e-06, + "loss": 0.5088, + "step": 26334 + }, + { + "epoch": 1.7674239119492634, + "grad_norm": 1.009101390838623, + "learning_rate": 3.501739721393826e-06, + "loss": 0.4667, + "step": 26336 + }, + { + "epoch": 1.7675581356330325, + "grad_norm": 1.0059616565704346, + "learning_rate": 3.4977451137411577e-06, + "loss": 0.4776, + "step": 26338 + }, + { + "epoch": 1.7676923593168015, + "grad_norm": 0.931182324886322, + "learning_rate": 3.4937527032285976e-06, + "loss": 0.3969, + "step": 26340 + }, + { + "epoch": 1.7678265830005704, + "grad_norm": 0.9413557052612305, + "learning_rate": 3.4897624900447624e-06, + "loss": 0.4209, + "step": 26342 + }, + { + "epoch": 1.7679608066843393, + "grad_norm": 0.9050574898719788, + "learning_rate": 3.4857744743782007e-06, + "loss": 0.4548, + "step": 26344 + }, + { + "epoch": 1.7680950303681084, + "grad_norm": 0.9665685892105103, + "learning_rate": 3.4817886564173287e-06, + "loss": 0.4743, + "step": 26346 + }, + { + "epoch": 1.7682292540518776, + "grad_norm": 1.056808590888977, + "learning_rate": 3.4778050363504675e-06, + "loss": 0.4282, + "step": 26348 + }, + { + "epoch": 1.7683634777356465, + "grad_norm": 1.005738615989685, + "learning_rate": 3.473823614365829e-06, + "loss": 0.4223, + "step": 26350 + }, + { + "epoch": 1.7684977014194154, + "grad_norm": 1.014094352722168, + "learning_rate": 3.469844390651544e-06, + "loss": 0.4461, + "step": 26352 + }, + { + "epoch": 1.7686319251031843, + "grad_norm": 1.0013539791107178, + "learning_rate": 3.465867365395614e-06, + "loss": 0.4602, + "step": 26354 + }, + { + "epoch": 1.7687661487869535, + "grad_norm": 0.9541317224502563, + "learning_rate": 3.4618925387859437e-06, + "loss": 0.478, + "step": 26356 + }, + { + "epoch": 1.7689003724707224, + "grad_norm": 1.037687063217163, + "learning_rate": 3.457919911010332e-06, + "loss": 0.486, + "step": 26358 + }, + { + "epoch": 1.7690345961544915, + "grad_norm": 0.9743841290473938, + "learning_rate": 3.453949482256491e-06, + "loss": 0.4508, + "step": 26360 + }, + { + "epoch": 1.7691688198382605, + "grad_norm": 1.5829144716262817, + "learning_rate": 3.449981252711998e-06, + "loss": 0.4721, + "step": 26362 + }, + { + "epoch": 1.7693030435220294, + "grad_norm": 0.9538119435310364, + "learning_rate": 3.4460152225643692e-06, + "loss": 0.472, + "step": 26364 + }, + { + "epoch": 1.7694372672057983, + "grad_norm": 0.9716622829437256, + "learning_rate": 3.442051392000967e-06, + "loss": 0.4669, + "step": 26366 + }, + { + "epoch": 1.7695714908895674, + "grad_norm": 0.9730298519134521, + "learning_rate": 3.4380897612090843e-06, + "loss": 0.4532, + "step": 26368 + }, + { + "epoch": 1.7697057145733366, + "grad_norm": 1.224579930305481, + "learning_rate": 3.4341303303759064e-06, + "loss": 0.4159, + "step": 26370 + }, + { + "epoch": 1.7698399382571055, + "grad_norm": 1.0344945192337036, + "learning_rate": 3.4301730996884884e-06, + "loss": 0.5461, + "step": 26372 + }, + { + "epoch": 1.7699741619408744, + "grad_norm": 1.0257664918899536, + "learning_rate": 3.4262180693338306e-06, + "loss": 0.5275, + "step": 26374 + }, + { + "epoch": 1.7701083856246433, + "grad_norm": 0.9448689818382263, + "learning_rate": 3.4222652394987843e-06, + "loss": 0.3993, + "step": 26376 + }, + { + "epoch": 1.7702426093084125, + "grad_norm": 1.0021743774414062, + "learning_rate": 3.4183146103701104e-06, + "loss": 0.4299, + "step": 26378 + }, + { + "epoch": 1.7703768329921816, + "grad_norm": 1.0137749910354614, + "learning_rate": 3.4143661821344654e-06, + "loss": 0.4464, + "step": 26380 + }, + { + "epoch": 1.7705110566759505, + "grad_norm": 1.0789955854415894, + "learning_rate": 3.4104199549784223e-06, + "loss": 0.4875, + "step": 26382 + }, + { + "epoch": 1.7706452803597195, + "grad_norm": 0.9993133544921875, + "learning_rate": 3.4064759290884207e-06, + "loss": 0.4478, + "step": 26384 + }, + { + "epoch": 1.7707795040434884, + "grad_norm": 1.0137730836868286, + "learning_rate": 3.4025341046508064e-06, + "loss": 0.5015, + "step": 26386 + }, + { + "epoch": 1.7709137277272575, + "grad_norm": 1.077858328819275, + "learning_rate": 3.398594481851819e-06, + "loss": 0.4951, + "step": 26388 + }, + { + "epoch": 1.7710479514110264, + "grad_norm": 0.948650062084198, + "learning_rate": 3.394657060877615e-06, + "loss": 0.4374, + "step": 26390 + }, + { + "epoch": 1.7711821750947956, + "grad_norm": 1.0998425483703613, + "learning_rate": 3.3907218419142182e-06, + "loss": 0.4774, + "step": 26392 + }, + { + "epoch": 1.7713163987785645, + "grad_norm": 1.0868417024612427, + "learning_rate": 3.3867888251475577e-06, + "loss": 0.4282, + "step": 26394 + }, + { + "epoch": 1.7714506224623334, + "grad_norm": 0.9734367728233337, + "learning_rate": 3.3828580107634623e-06, + "loss": 0.4114, + "step": 26396 + }, + { + "epoch": 1.7715848461461023, + "grad_norm": 1.0528538227081299, + "learning_rate": 3.378929398947661e-06, + "loss": 0.4872, + "step": 26398 + }, + { + "epoch": 1.7717190698298715, + "grad_norm": 1.1084693670272827, + "learning_rate": 3.3750029898857614e-06, + "loss": 0.4852, + "step": 26400 + }, + { + "epoch": 1.7718532935136406, + "grad_norm": 0.9850924015045166, + "learning_rate": 3.3710787837632982e-06, + "loss": 0.4211, + "step": 26402 + }, + { + "epoch": 1.7719875171974095, + "grad_norm": 0.9967346787452698, + "learning_rate": 3.367156780765668e-06, + "loss": 0.4618, + "step": 26404 + }, + { + "epoch": 1.7721217408811785, + "grad_norm": 1.0487021207809448, + "learning_rate": 3.3632369810781774e-06, + "loss": 0.4789, + "step": 26406 + }, + { + "epoch": 1.7722559645649474, + "grad_norm": 0.9948101043701172, + "learning_rate": 3.3593193848860284e-06, + "loss": 0.4703, + "step": 26408 + }, + { + "epoch": 1.7723901882487165, + "grad_norm": 1.0135157108306885, + "learning_rate": 3.3554039923743286e-06, + "loss": 0.4501, + "step": 26410 + }, + { + "epoch": 1.7725244119324854, + "grad_norm": 0.9719632863998413, + "learning_rate": 3.351490803728069e-06, + "loss": 0.4677, + "step": 26412 + }, + { + "epoch": 1.7726586356162546, + "grad_norm": 1.0641772747039795, + "learning_rate": 3.3475798191321406e-06, + "loss": 0.5544, + "step": 26414 + }, + { + "epoch": 1.7727928593000235, + "grad_norm": 0.8491989970207214, + "learning_rate": 3.3436710387713176e-06, + "loss": 0.3951, + "step": 26416 + }, + { + "epoch": 1.7729270829837924, + "grad_norm": 1.0544626712799072, + "learning_rate": 3.3397644628303026e-06, + "loss": 0.5106, + "step": 26418 + }, + { + "epoch": 1.7730613066675613, + "grad_norm": 1.0775730609893799, + "learning_rate": 3.335860091493653e-06, + "loss": 0.4315, + "step": 26420 + }, + { + "epoch": 1.7731955303513305, + "grad_norm": 1.0925928354263306, + "learning_rate": 3.3319579249458667e-06, + "loss": 0.435, + "step": 26422 + }, + { + "epoch": 1.7733297540350996, + "grad_norm": 1.0120404958724976, + "learning_rate": 3.328057963371284e-06, + "loss": 0.4855, + "step": 26424 + }, + { + "epoch": 1.7734639777188685, + "grad_norm": 1.024065375328064, + "learning_rate": 3.3241602069541967e-06, + "loss": 0.4473, + "step": 26426 + }, + { + "epoch": 1.7735982014026375, + "grad_norm": 1.0965101718902588, + "learning_rate": 3.3202646558787463e-06, + "loss": 0.5376, + "step": 26428 + }, + { + "epoch": 1.7737324250864064, + "grad_norm": 1.0976194143295288, + "learning_rate": 3.3163713103290084e-06, + "loss": 0.4475, + "step": 26430 + }, + { + "epoch": 1.7738666487701755, + "grad_norm": 0.9388175010681152, + "learning_rate": 3.3124801704889298e-06, + "loss": 0.5204, + "step": 26432 + }, + { + "epoch": 1.7740008724539444, + "grad_norm": 0.9413573741912842, + "learning_rate": 3.308591236542352e-06, + "loss": 0.4106, + "step": 26434 + }, + { + "epoch": 1.7741350961377136, + "grad_norm": 0.9693027138710022, + "learning_rate": 3.3047045086730233e-06, + "loss": 0.4768, + "step": 26436 + }, + { + "epoch": 1.7742693198214825, + "grad_norm": 1.11866295337677, + "learning_rate": 3.300819987064574e-06, + "loss": 0.4673, + "step": 26438 + }, + { + "epoch": 1.7744035435052514, + "grad_norm": 1.123133659362793, + "learning_rate": 3.296937671900563e-06, + "loss": 0.5068, + "step": 26440 + }, + { + "epoch": 1.7745377671890203, + "grad_norm": 0.9647020101547241, + "learning_rate": 3.2930575633644103e-06, + "loss": 0.4149, + "step": 26442 + }, + { + "epoch": 1.7746719908727895, + "grad_norm": 0.9570835828781128, + "learning_rate": 3.2891796616394353e-06, + "loss": 0.4342, + "step": 26444 + }, + { + "epoch": 1.7748062145565586, + "grad_norm": 0.906319797039032, + "learning_rate": 3.285303966908865e-06, + "loss": 0.4271, + "step": 26446 + }, + { + "epoch": 1.7749404382403275, + "grad_norm": 0.9696779251098633, + "learning_rate": 3.2814304793558294e-06, + "loss": 0.4509, + "step": 26448 + }, + { + "epoch": 1.7750746619240965, + "grad_norm": 1.0504570007324219, + "learning_rate": 3.277559199163338e-06, + "loss": 0.4424, + "step": 26450 + }, + { + "epoch": 1.7752088856078654, + "grad_norm": 1.1588096618652344, + "learning_rate": 3.2736901265142948e-06, + "loss": 0.4405, + "step": 26452 + }, + { + "epoch": 1.7753431092916345, + "grad_norm": 1.0212655067443848, + "learning_rate": 3.2698232615915034e-06, + "loss": 0.474, + "step": 26454 + }, + { + "epoch": 1.7754773329754037, + "grad_norm": 1.0489578247070312, + "learning_rate": 3.265958604577679e-06, + "loss": 0.4647, + "step": 26456 + }, + { + "epoch": 1.7756115566591726, + "grad_norm": 0.8789849281311035, + "learning_rate": 3.262096155655403e-06, + "loss": 0.4147, + "step": 26458 + }, + { + "epoch": 1.7757457803429415, + "grad_norm": 1.1079792976379395, + "learning_rate": 3.2582359150071963e-06, + "loss": 0.4621, + "step": 26460 + }, + { + "epoch": 1.7758800040267104, + "grad_norm": 1.1104118824005127, + "learning_rate": 3.2543778828154125e-06, + "loss": 0.4943, + "step": 26462 + }, + { + "epoch": 1.7760142277104796, + "grad_norm": 0.8925808668136597, + "learning_rate": 3.250522059262362e-06, + "loss": 0.3928, + "step": 26464 + }, + { + "epoch": 1.7761484513942485, + "grad_norm": 1.026313304901123, + "learning_rate": 3.246668444530204e-06, + "loss": 0.4922, + "step": 26466 + }, + { + "epoch": 1.7762826750780176, + "grad_norm": 1.1014049053192139, + "learning_rate": 3.2428170388010326e-06, + "loss": 0.5189, + "step": 26468 + }, + { + "epoch": 1.7764168987617865, + "grad_norm": 1.0588924884796143, + "learning_rate": 3.238967842256818e-06, + "loss": 0.4708, + "step": 26470 + }, + { + "epoch": 1.7765511224455555, + "grad_norm": 1.07292902469635, + "learning_rate": 3.2351208550794154e-06, + "loss": 0.4514, + "step": 26472 + }, + { + "epoch": 1.7766853461293244, + "grad_norm": 0.9535008072853088, + "learning_rate": 3.2312760774505845e-06, + "loss": 0.3932, + "step": 26474 + }, + { + "epoch": 1.7768195698130935, + "grad_norm": 1.1209310293197632, + "learning_rate": 3.2274335095520026e-06, + "loss": 0.4718, + "step": 26476 + }, + { + "epoch": 1.7769537934968627, + "grad_norm": 0.9832744002342224, + "learning_rate": 3.223593151565213e-06, + "loss": 0.437, + "step": 26478 + }, + { + "epoch": 1.7770880171806316, + "grad_norm": 1.0584068298339844, + "learning_rate": 3.219755003671665e-06, + "loss": 0.4624, + "step": 26480 + }, + { + "epoch": 1.7772222408644005, + "grad_norm": 1.0272608995437622, + "learning_rate": 3.2159190660526972e-06, + "loss": 0.4485, + "step": 26482 + }, + { + "epoch": 1.7773564645481694, + "grad_norm": 1.1383942365646362, + "learning_rate": 3.2120853388895643e-06, + "loss": 0.4513, + "step": 26484 + }, + { + "epoch": 1.7774906882319386, + "grad_norm": 1.079269289970398, + "learning_rate": 3.2082538223633884e-06, + "loss": 0.4312, + "step": 26486 + }, + { + "epoch": 1.7776249119157075, + "grad_norm": 1.074112892150879, + "learning_rate": 3.204424516655219e-06, + "loss": 0.4639, + "step": 26488 + }, + { + "epoch": 1.7777591355994766, + "grad_norm": 1.0962426662445068, + "learning_rate": 3.2005974219459556e-06, + "loss": 0.4773, + "step": 26490 + }, + { + "epoch": 1.7778933592832455, + "grad_norm": 0.909841775894165, + "learning_rate": 3.196772538416448e-06, + "loss": 0.4395, + "step": 26492 + }, + { + "epoch": 1.7780275829670145, + "grad_norm": 1.0619467496871948, + "learning_rate": 3.1929498662473965e-06, + "loss": 0.5161, + "step": 26494 + }, + { + "epoch": 1.7781618066507834, + "grad_norm": 1.096822738647461, + "learning_rate": 3.1891294056194233e-06, + "loss": 0.4501, + "step": 26496 + }, + { + "epoch": 1.7782960303345525, + "grad_norm": 1.1792314052581787, + "learning_rate": 3.1853111567130387e-06, + "loss": 0.477, + "step": 26498 + }, + { + "epoch": 1.7784302540183217, + "grad_norm": 0.9513255953788757, + "learning_rate": 3.1814951197086495e-06, + "loss": 0.4912, + "step": 26500 + }, + { + "epoch": 1.7785644777020906, + "grad_norm": 0.9621514081954956, + "learning_rate": 3.1776812947865385e-06, + "loss": 0.4396, + "step": 26502 + }, + { + "epoch": 1.7786987013858595, + "grad_norm": 1.0372976064682007, + "learning_rate": 3.173869682126923e-06, + "loss": 0.5395, + "step": 26504 + }, + { + "epoch": 1.7788329250696284, + "grad_norm": 0.98854660987854, + "learning_rate": 3.1700602819098868e-06, + "loss": 0.4634, + "step": 26506 + }, + { + "epoch": 1.7789671487533976, + "grad_norm": 1.076910376548767, + "learning_rate": 3.1662530943154134e-06, + "loss": 0.5016, + "step": 26508 + }, + { + "epoch": 1.7791013724371665, + "grad_norm": 1.0470190048217773, + "learning_rate": 3.162448119523387e-06, + "loss": 0.486, + "step": 26510 + }, + { + "epoch": 1.7792355961209356, + "grad_norm": 1.1797438859939575, + "learning_rate": 3.1586453577135798e-06, + "loss": 0.5459, + "step": 26512 + }, + { + "epoch": 1.7793698198047045, + "grad_norm": 1.1045492887496948, + "learning_rate": 3.154844809065677e-06, + "loss": 0.491, + "step": 26514 + }, + { + "epoch": 1.7795040434884735, + "grad_norm": 0.9104241728782654, + "learning_rate": 3.1510464737592336e-06, + "loss": 0.4016, + "step": 26516 + }, + { + "epoch": 1.7796382671722424, + "grad_norm": 1.12786865234375, + "learning_rate": 3.1472503519737295e-06, + "loss": 0.4704, + "step": 26518 + }, + { + "epoch": 1.7797724908560115, + "grad_norm": 1.0569543838500977, + "learning_rate": 3.1434564438884983e-06, + "loss": 0.4124, + "step": 26520 + }, + { + "epoch": 1.7799067145397807, + "grad_norm": 1.2627952098846436, + "learning_rate": 3.1396647496828247e-06, + "loss": 0.4108, + "step": 26522 + }, + { + "epoch": 1.7800409382235496, + "grad_norm": 0.9939997792243958, + "learning_rate": 3.1358752695358316e-06, + "loss": 0.4242, + "step": 26524 + }, + { + "epoch": 1.7801751619073185, + "grad_norm": 0.9446654915809631, + "learning_rate": 3.132088003626588e-06, + "loss": 0.4203, + "step": 26526 + }, + { + "epoch": 1.7803093855910874, + "grad_norm": 1.1433025598526, + "learning_rate": 3.1283029521340212e-06, + "loss": 0.4893, + "step": 26528 + }, + { + "epoch": 1.7804436092748566, + "grad_norm": 0.9577170014381409, + "learning_rate": 3.1245201152369783e-06, + "loss": 0.515, + "step": 26530 + }, + { + "epoch": 1.7805778329586257, + "grad_norm": 0.9782341122627258, + "learning_rate": 3.1207394931141665e-06, + "loss": 0.4448, + "step": 26532 + }, + { + "epoch": 1.7807120566423946, + "grad_norm": 0.9431700706481934, + "learning_rate": 3.116961085944242e-06, + "loss": 0.4242, + "step": 26534 + }, + { + "epoch": 1.7808462803261635, + "grad_norm": 1.0530515909194946, + "learning_rate": 3.1131848939057186e-06, + "loss": 0.4992, + "step": 26536 + }, + { + "epoch": 1.7809805040099325, + "grad_norm": 1.1985836029052734, + "learning_rate": 3.1094109171770027e-06, + "loss": 0.5346, + "step": 26538 + }, + { + "epoch": 1.7811147276937016, + "grad_norm": 0.9395465850830078, + "learning_rate": 3.105639155936413e-06, + "loss": 0.4576, + "step": 26540 + }, + { + "epoch": 1.7812489513774705, + "grad_norm": 1.0075362920761108, + "learning_rate": 3.1018696103621625e-06, + "loss": 0.4682, + "step": 26542 + }, + { + "epoch": 1.7813831750612397, + "grad_norm": 0.9399588108062744, + "learning_rate": 3.0981022806323536e-06, + "loss": 0.4313, + "step": 26544 + }, + { + "epoch": 1.7815173987450086, + "grad_norm": 1.0964415073394775, + "learning_rate": 3.094337166924982e-06, + "loss": 0.4771, + "step": 26546 + }, + { + "epoch": 1.7816516224287775, + "grad_norm": 0.9933387041091919, + "learning_rate": 3.090574269417934e-06, + "loss": 0.427, + "step": 26548 + }, + { + "epoch": 1.7817858461125464, + "grad_norm": 1.0570873022079468, + "learning_rate": 3.0868135882890224e-06, + "loss": 0.492, + "step": 26550 + }, + { + "epoch": 1.7819200697963156, + "grad_norm": 0.946513295173645, + "learning_rate": 3.0830551237159056e-06, + "loss": 0.4297, + "step": 26552 + }, + { + "epoch": 1.7820542934800847, + "grad_norm": 1.0778554677963257, + "learning_rate": 3.0792988758761854e-06, + "loss": 0.5005, + "step": 26554 + }, + { + "epoch": 1.7821885171638536, + "grad_norm": 0.9276990294456482, + "learning_rate": 3.0755448449473255e-06, + "loss": 0.4806, + "step": 26556 + }, + { + "epoch": 1.7823227408476225, + "grad_norm": 1.0314631462097168, + "learning_rate": 3.071793031106701e-06, + "loss": 0.4806, + "step": 26558 + }, + { + "epoch": 1.7824569645313915, + "grad_norm": 0.9608511924743652, + "learning_rate": 3.0680434345315643e-06, + "loss": 0.4765, + "step": 26560 + }, + { + "epoch": 1.7825911882151606, + "grad_norm": 1.042948842048645, + "learning_rate": 3.0642960553990963e-06, + "loss": 0.4899, + "step": 26562 + }, + { + "epoch": 1.7827254118989295, + "grad_norm": 0.9586542248725891, + "learning_rate": 3.0605508938863493e-06, + "loss": 0.4453, + "step": 26564 + }, + { + "epoch": 1.7828596355826987, + "grad_norm": 1.0348963737487793, + "learning_rate": 3.056807950170265e-06, + "loss": 0.484, + "step": 26566 + }, + { + "epoch": 1.7829938592664676, + "grad_norm": 1.127644658088684, + "learning_rate": 3.0530672244276912e-06, + "loss": 0.467, + "step": 26568 + }, + { + "epoch": 1.7831280829502365, + "grad_norm": 1.0233221054077148, + "learning_rate": 3.049328716835376e-06, + "loss": 0.5077, + "step": 26570 + }, + { + "epoch": 1.7832623066340054, + "grad_norm": 1.0783993005752563, + "learning_rate": 3.04559242756996e-06, + "loss": 0.4298, + "step": 26572 + }, + { + "epoch": 1.7833965303177746, + "grad_norm": 1.0966519117355347, + "learning_rate": 3.0418583568079697e-06, + "loss": 0.4818, + "step": 26574 + }, + { + "epoch": 1.7835307540015437, + "grad_norm": 1.0691412687301636, + "learning_rate": 3.03812650472583e-06, + "loss": 0.5221, + "step": 26576 + }, + { + "epoch": 1.7836649776853126, + "grad_norm": 1.1239451169967651, + "learning_rate": 3.034396871499856e-06, + "loss": 0.5107, + "step": 26578 + }, + { + "epoch": 1.7837992013690815, + "grad_norm": 1.1886765956878662, + "learning_rate": 3.0306694573062898e-06, + "loss": 0.4383, + "step": 26580 + }, + { + "epoch": 1.7839334250528505, + "grad_norm": 0.9689651131629944, + "learning_rate": 3.0269442623212186e-06, + "loss": 0.5116, + "step": 26582 + }, + { + "epoch": 1.7840676487366196, + "grad_norm": 1.0736305713653564, + "learning_rate": 3.023221286720679e-06, + "loss": 0.4788, + "step": 26584 + }, + { + "epoch": 1.7842018724203885, + "grad_norm": 1.016939401626587, + "learning_rate": 3.019500530680547e-06, + "loss": 0.462, + "step": 26586 + }, + { + "epoch": 1.7843360961041577, + "grad_norm": 1.0670020580291748, + "learning_rate": 3.015781994376632e-06, + "loss": 0.4268, + "step": 26588 + }, + { + "epoch": 1.7844703197879266, + "grad_norm": 0.9900224804878235, + "learning_rate": 3.0120656779846214e-06, + "loss": 0.5255, + "step": 26590 + }, + { + "epoch": 1.7846045434716955, + "grad_norm": 0.9999427199363708, + "learning_rate": 3.008351581680119e-06, + "loss": 0.481, + "step": 26592 + }, + { + "epoch": 1.7847387671554644, + "grad_norm": 1.0749753713607788, + "learning_rate": 3.004639705638601e-06, + "loss": 0.4785, + "step": 26594 + }, + { + "epoch": 1.7848729908392336, + "grad_norm": 0.9687768816947937, + "learning_rate": 3.0009300500354444e-06, + "loss": 0.4166, + "step": 26596 + }, + { + "epoch": 1.7850072145230027, + "grad_norm": 0.9108654856681824, + "learning_rate": 2.9972226150459137e-06, + "loss": 0.4143, + "step": 26598 + }, + { + "epoch": 1.7851414382067716, + "grad_norm": 0.9863176941871643, + "learning_rate": 2.9935174008451917e-06, + "loss": 0.4059, + "step": 26600 + }, + { + "epoch": 1.7852756618905405, + "grad_norm": 0.9309257864952087, + "learning_rate": 2.9898144076083433e-06, + "loss": 0.4821, + "step": 26602 + }, + { + "epoch": 1.7854098855743095, + "grad_norm": 1.1409417390823364, + "learning_rate": 2.9861136355103235e-06, + "loss": 0.4612, + "step": 26604 + }, + { + "epoch": 1.7855441092580786, + "grad_norm": 0.9512568116188049, + "learning_rate": 2.982415084725976e-06, + "loss": 0.4737, + "step": 26606 + }, + { + "epoch": 1.7856783329418477, + "grad_norm": 0.8105921149253845, + "learning_rate": 2.9787187554300656e-06, + "loss": 0.4305, + "step": 26608 + }, + { + "epoch": 1.7858125566256167, + "grad_norm": 1.0694352388381958, + "learning_rate": 2.97502464779722e-06, + "loss": 0.483, + "step": 26610 + }, + { + "epoch": 1.7859467803093856, + "grad_norm": 1.0770924091339111, + "learning_rate": 2.9713327620020103e-06, + "loss": 0.5138, + "step": 26612 + }, + { + "epoch": 1.7860810039931545, + "grad_norm": 1.0274219512939453, + "learning_rate": 2.967643098218831e-06, + "loss": 0.5058, + "step": 26614 + }, + { + "epoch": 1.7862152276769236, + "grad_norm": 1.0504615306854248, + "learning_rate": 2.9639556566220415e-06, + "loss": 0.4684, + "step": 26616 + }, + { + "epoch": 1.7863494513606926, + "grad_norm": 0.9464894533157349, + "learning_rate": 2.9602704373858426e-06, + "loss": 0.4762, + "step": 26618 + }, + { + "epoch": 1.7864836750444617, + "grad_norm": 1.0615830421447754, + "learning_rate": 2.9565874406843776e-06, + "loss": 0.5122, + "step": 26620 + }, + { + "epoch": 1.7866178987282306, + "grad_norm": 1.1229662895202637, + "learning_rate": 2.952906666691646e-06, + "loss": 0.4577, + "step": 26622 + }, + { + "epoch": 1.7867521224119995, + "grad_norm": 0.9584335088729858, + "learning_rate": 2.949228115581565e-06, + "loss": 0.4124, + "step": 26624 + }, + { + "epoch": 1.7868863460957685, + "grad_norm": 1.3853611946105957, + "learning_rate": 2.9455517875279225e-06, + "loss": 0.4759, + "step": 26626 + }, + { + "epoch": 1.7870205697795376, + "grad_norm": 1.0542583465576172, + "learning_rate": 2.9418776827044357e-06, + "loss": 0.503, + "step": 26628 + }, + { + "epoch": 1.7871547934633067, + "grad_norm": 1.0837205648422241, + "learning_rate": 2.9382058012846936e-06, + "loss": 0.471, + "step": 26630 + }, + { + "epoch": 1.7872890171470757, + "grad_norm": 1.0488024950027466, + "learning_rate": 2.93453614344219e-06, + "loss": 0.4463, + "step": 26632 + }, + { + "epoch": 1.7874232408308446, + "grad_norm": 1.056859016418457, + "learning_rate": 2.930868709350287e-06, + "loss": 0.4371, + "step": 26634 + }, + { + "epoch": 1.7875574645146135, + "grad_norm": 1.1180368661880493, + "learning_rate": 2.927203499182296e-06, + "loss": 0.5127, + "step": 26636 + }, + { + "epoch": 1.7876916881983826, + "grad_norm": 1.02019464969635, + "learning_rate": 2.9235405131113615e-06, + "loss": 0.4248, + "step": 26638 + }, + { + "epoch": 1.7878259118821516, + "grad_norm": 0.9511210918426514, + "learning_rate": 2.9198797513105834e-06, + "loss": 0.4121, + "step": 26640 + }, + { + "epoch": 1.7879601355659207, + "grad_norm": 0.9936085939407349, + "learning_rate": 2.9162212139528967e-06, + "loss": 0.4594, + "step": 26642 + }, + { + "epoch": 1.7880943592496896, + "grad_norm": 0.9715378880500793, + "learning_rate": 2.9125649012111678e-06, + "loss": 0.4199, + "step": 26644 + }, + { + "epoch": 1.7882285829334585, + "grad_norm": 1.149639368057251, + "learning_rate": 2.9089108132581587e-06, + "loss": 0.4676, + "step": 26646 + }, + { + "epoch": 1.7883628066172275, + "grad_norm": 0.9170129895210266, + "learning_rate": 2.9052589502665095e-06, + "loss": 0.4516, + "step": 26648 + }, + { + "epoch": 1.7884970303009966, + "grad_norm": 0.9058769345283508, + "learning_rate": 2.9016093124087817e-06, + "loss": 0.4454, + "step": 26650 + }, + { + "epoch": 1.7886312539847657, + "grad_norm": 1.0133819580078125, + "learning_rate": 2.897961899857382e-06, + "loss": 0.4251, + "step": 26652 + }, + { + "epoch": 1.7887654776685347, + "grad_norm": 1.010780692100525, + "learning_rate": 2.894316712784667e-06, + "loss": 0.4627, + "step": 26654 + }, + { + "epoch": 1.7888997013523036, + "grad_norm": 1.0257115364074707, + "learning_rate": 2.890673751362849e-06, + "loss": 0.4628, + "step": 26656 + }, + { + "epoch": 1.7890339250360725, + "grad_norm": 1.0811824798583984, + "learning_rate": 2.887033015764068e-06, + "loss": 0.4746, + "step": 26658 + }, + { + "epoch": 1.7891681487198416, + "grad_norm": 1.054804801940918, + "learning_rate": 2.883394506160336e-06, + "loss": 0.445, + "step": 26660 + }, + { + "epoch": 1.7893023724036106, + "grad_norm": 1.0269854068756104, + "learning_rate": 2.8797582227235608e-06, + "loss": 0.4701, + "step": 26662 + }, + { + "epoch": 1.7894365960873797, + "grad_norm": 1.1081819534301758, + "learning_rate": 2.876124165625549e-06, + "loss": 0.4196, + "step": 26664 + }, + { + "epoch": 1.7895708197711486, + "grad_norm": 0.932039737701416, + "learning_rate": 2.8724923350380075e-06, + "loss": 0.4012, + "step": 26666 + }, + { + "epoch": 1.7897050434549175, + "grad_norm": 1.0136710405349731, + "learning_rate": 2.868862731132532e-06, + "loss": 0.4776, + "step": 26668 + }, + { + "epoch": 1.7898392671386865, + "grad_norm": 0.9252318739891052, + "learning_rate": 2.8652353540806086e-06, + "loss": 0.452, + "step": 26670 + }, + { + "epoch": 1.7899734908224556, + "grad_norm": 1.0677413940429688, + "learning_rate": 2.8616102040536274e-06, + "loss": 0.4439, + "step": 26672 + }, + { + "epoch": 1.7901077145062247, + "grad_norm": 1.121899962425232, + "learning_rate": 2.8579872812228738e-06, + "loss": 0.4599, + "step": 26674 + }, + { + "epoch": 1.7902419381899937, + "grad_norm": 1.0476349592208862, + "learning_rate": 2.8543665857595158e-06, + "loss": 0.5122, + "step": 26676 + }, + { + "epoch": 1.7903761618737626, + "grad_norm": 1.0133148431777954, + "learning_rate": 2.850748117834645e-06, + "loss": 0.5094, + "step": 26678 + }, + { + "epoch": 1.7905103855575315, + "grad_norm": 1.0524612665176392, + "learning_rate": 2.8471318776191914e-06, + "loss": 0.4234, + "step": 26680 + }, + { + "epoch": 1.7906446092413006, + "grad_norm": 1.100936770439148, + "learning_rate": 2.8435178652840456e-06, + "loss": 0.4973, + "step": 26682 + }, + { + "epoch": 1.7907788329250698, + "grad_norm": 0.9793028235435486, + "learning_rate": 2.839906080999938e-06, + "loss": 0.4941, + "step": 26684 + }, + { + "epoch": 1.7909130566088387, + "grad_norm": 1.0425536632537842, + "learning_rate": 2.8362965249375485e-06, + "loss": 0.4424, + "step": 26686 + }, + { + "epoch": 1.7910472802926076, + "grad_norm": 1.0285500288009644, + "learning_rate": 2.8326891972673965e-06, + "loss": 0.4544, + "step": 26688 + }, + { + "epoch": 1.7911815039763765, + "grad_norm": 1.0473531484603882, + "learning_rate": 2.829084098159934e-06, + "loss": 0.4808, + "step": 26690 + }, + { + "epoch": 1.7913157276601457, + "grad_norm": 1.0424187183380127, + "learning_rate": 2.8254812277854813e-06, + "loss": 0.4182, + "step": 26692 + }, + { + "epoch": 1.7914499513439146, + "grad_norm": 1.0150495767593384, + "learning_rate": 2.8218805863142794e-06, + "loss": 0.541, + "step": 26694 + }, + { + "epoch": 1.7915841750276837, + "grad_norm": 0.9561358094215393, + "learning_rate": 2.8182821739164534e-06, + "loss": 0.414, + "step": 26696 + }, + { + "epoch": 1.7917183987114527, + "grad_norm": 1.0548745393753052, + "learning_rate": 2.8146859907620172e-06, + "loss": 0.3886, + "step": 26698 + }, + { + "epoch": 1.7918526223952216, + "grad_norm": 1.0386500358581543, + "learning_rate": 2.8110920370208682e-06, + "loss": 0.4123, + "step": 26700 + }, + { + "epoch": 1.7919868460789905, + "grad_norm": 1.0239613056182861, + "learning_rate": 2.8075003128628374e-06, + "loss": 0.5219, + "step": 26702 + }, + { + "epoch": 1.7921210697627596, + "grad_norm": 1.0330843925476074, + "learning_rate": 2.8039108184576113e-06, + "loss": 0.442, + "step": 26704 + }, + { + "epoch": 1.7922552934465288, + "grad_norm": 1.0979597568511963, + "learning_rate": 2.8003235539747984e-06, + "loss": 0.4461, + "step": 26706 + }, + { + "epoch": 1.7923895171302977, + "grad_norm": 0.9252837300300598, + "learning_rate": 2.796738519583886e-06, + "loss": 0.4517, + "step": 26708 + }, + { + "epoch": 1.7925237408140666, + "grad_norm": 1.0468565225601196, + "learning_rate": 2.7931557154542487e-06, + "loss": 0.472, + "step": 26710 + }, + { + "epoch": 1.7926579644978355, + "grad_norm": 1.0020323991775513, + "learning_rate": 2.7895751417551852e-06, + "loss": 0.5184, + "step": 26712 + }, + { + "epoch": 1.7927921881816047, + "grad_norm": 0.9411514401435852, + "learning_rate": 2.785996798655849e-06, + "loss": 0.45, + "step": 26714 + }, + { + "epoch": 1.7929264118653736, + "grad_norm": 0.7674515247344971, + "learning_rate": 2.7824206863253264e-06, + "loss": 0.4257, + "step": 26716 + }, + { + "epoch": 1.7930606355491427, + "grad_norm": 1.0416252613067627, + "learning_rate": 2.778846804932583e-06, + "loss": 0.4851, + "step": 26718 + }, + { + "epoch": 1.7931948592329117, + "grad_norm": 1.0742361545562744, + "learning_rate": 2.775275154646467e-06, + "loss": 0.4777, + "step": 26720 + }, + { + "epoch": 1.7933290829166806, + "grad_norm": 1.0092241764068604, + "learning_rate": 2.7717057356357323e-06, + "loss": 0.5018, + "step": 26722 + }, + { + "epoch": 1.7934633066004495, + "grad_norm": 0.8847936391830444, + "learning_rate": 2.768138548069038e-06, + "loss": 0.4576, + "step": 26724 + }, + { + "epoch": 1.7935975302842186, + "grad_norm": 1.039747714996338, + "learning_rate": 2.7645735921149217e-06, + "loss": 0.5439, + "step": 26726 + }, + { + "epoch": 1.7937317539679878, + "grad_norm": 0.8865048289299011, + "learning_rate": 2.7610108679418156e-06, + "loss": 0.4585, + "step": 26728 + }, + { + "epoch": 1.7938659776517567, + "grad_norm": 1.035498023033142, + "learning_rate": 2.757450375718046e-06, + "loss": 0.4834, + "step": 26730 + }, + { + "epoch": 1.7940002013355256, + "grad_norm": 1.0822051763534546, + "learning_rate": 2.753892115611856e-06, + "loss": 0.4815, + "step": 26732 + }, + { + "epoch": 1.7941344250192945, + "grad_norm": 0.8330800533294678, + "learning_rate": 2.7503360877913497e-06, + "loss": 0.4315, + "step": 26734 + }, + { + "epoch": 1.7942686487030637, + "grad_norm": 0.9999399185180664, + "learning_rate": 2.7467822924245713e-06, + "loss": 0.433, + "step": 26736 + }, + { + "epoch": 1.7944028723868326, + "grad_norm": 1.141569972038269, + "learning_rate": 2.7432307296793914e-06, + "loss": 0.4499, + "step": 26738 + }, + { + "epoch": 1.7945370960706017, + "grad_norm": 1.006667971611023, + "learning_rate": 2.739681399723637e-06, + "loss": 0.4665, + "step": 26740 + }, + { + "epoch": 1.7946713197543707, + "grad_norm": 0.9503446221351624, + "learning_rate": 2.736134302725002e-06, + "loss": 0.435, + "step": 26742 + }, + { + "epoch": 1.7948055434381396, + "grad_norm": 1.1470330953598022, + "learning_rate": 2.7325894388510855e-06, + "loss": 0.4996, + "step": 26744 + }, + { + "epoch": 1.7949397671219085, + "grad_norm": 1.1522454023361206, + "learning_rate": 2.72904680826937e-06, + "loss": 0.5004, + "step": 26746 + }, + { + "epoch": 1.7950739908056776, + "grad_norm": 0.937041699886322, + "learning_rate": 2.7255064111472385e-06, + "loss": 0.4256, + "step": 26748 + }, + { + "epoch": 1.7952082144894468, + "grad_norm": 1.0546565055847168, + "learning_rate": 2.721968247651957e-06, + "loss": 0.3999, + "step": 26750 + }, + { + "epoch": 1.7953424381732157, + "grad_norm": 0.8641650080680847, + "learning_rate": 2.71843231795072e-06, + "loss": 0.3858, + "step": 26752 + }, + { + "epoch": 1.7954766618569846, + "grad_norm": 1.1361602544784546, + "learning_rate": 2.7148986222105765e-06, + "loss": 0.5382, + "step": 26754 + }, + { + "epoch": 1.7956108855407535, + "grad_norm": 0.9405742287635803, + "learning_rate": 2.7113671605984934e-06, + "loss": 0.4319, + "step": 26756 + }, + { + "epoch": 1.7957451092245227, + "grad_norm": 1.0715886354446411, + "learning_rate": 2.707837933281321e-06, + "loss": 0.5071, + "step": 26758 + }, + { + "epoch": 1.7958793329082918, + "grad_norm": 0.885618269443512, + "learning_rate": 2.704310940425808e-06, + "loss": 0.4731, + "step": 26760 + }, + { + "epoch": 1.7960135565920607, + "grad_norm": 1.038923740386963, + "learning_rate": 2.7007861821986e-06, + "loss": 0.4156, + "step": 26762 + }, + { + "epoch": 1.7961477802758297, + "grad_norm": 1.0108329057693481, + "learning_rate": 2.697263658766247e-06, + "loss": 0.5048, + "step": 26764 + }, + { + "epoch": 1.7962820039595986, + "grad_norm": 1.0313938856124878, + "learning_rate": 2.6937433702951543e-06, + "loss": 0.4768, + "step": 26766 + }, + { + "epoch": 1.7964162276433677, + "grad_norm": 1.0988893508911133, + "learning_rate": 2.690225316951672e-06, + "loss": 0.4781, + "step": 26768 + }, + { + "epoch": 1.7965504513271366, + "grad_norm": 0.91923588514328, + "learning_rate": 2.6867094989020016e-06, + "loss": 0.4576, + "step": 26770 + }, + { + "epoch": 1.7966846750109058, + "grad_norm": 1.169795036315918, + "learning_rate": 2.683195916312281e-06, + "loss": 0.5209, + "step": 26772 + }, + { + "epoch": 1.7968188986946747, + "grad_norm": 1.0356361865997314, + "learning_rate": 2.6796845693485064e-06, + "loss": 0.4319, + "step": 26774 + }, + { + "epoch": 1.7969531223784436, + "grad_norm": 1.1351392269134521, + "learning_rate": 2.6761754581765886e-06, + "loss": 0.4949, + "step": 26776 + }, + { + "epoch": 1.7970873460622125, + "grad_norm": 1.1579749584197998, + "learning_rate": 2.672668582962312e-06, + "loss": 0.5209, + "step": 26778 + }, + { + "epoch": 1.7972215697459817, + "grad_norm": 1.3297175168991089, + "learning_rate": 2.6691639438713834e-06, + "loss": 0.5057, + "step": 26780 + }, + { + "epoch": 1.7973557934297508, + "grad_norm": 1.0900532007217407, + "learning_rate": 2.6656615410693918e-06, + "loss": 0.4534, + "step": 26782 + }, + { + "epoch": 1.7974900171135197, + "grad_norm": 1.1601152420043945, + "learning_rate": 2.662161374721811e-06, + "loss": 0.4267, + "step": 26784 + }, + { + "epoch": 1.7976242407972887, + "grad_norm": 0.9871609210968018, + "learning_rate": 2.6586634449940194e-06, + "loss": 0.5198, + "step": 26786 + }, + { + "epoch": 1.7977584644810576, + "grad_norm": 1.0594075918197632, + "learning_rate": 2.6551677520512797e-06, + "loss": 0.5015, + "step": 26788 + }, + { + "epoch": 1.7978926881648267, + "grad_norm": 0.9030765295028687, + "learning_rate": 2.651674296058776e-06, + "loss": 0.4498, + "step": 26790 + }, + { + "epoch": 1.7980269118485956, + "grad_norm": 1.0988365411758423, + "learning_rate": 2.6481830771815486e-06, + "loss": 0.4167, + "step": 26792 + }, + { + "epoch": 1.7981611355323648, + "grad_norm": 1.1187043190002441, + "learning_rate": 2.64469409558456e-06, + "loss": 0.5317, + "step": 26794 + }, + { + "epoch": 1.7982953592161337, + "grad_norm": 1.0270341634750366, + "learning_rate": 2.6412073514326508e-06, + "loss": 0.467, + "step": 26796 + }, + { + "epoch": 1.7984295828999026, + "grad_norm": 1.0437359809875488, + "learning_rate": 2.637722844890572e-06, + "loss": 0.4379, + "step": 26798 + }, + { + "epoch": 1.7985638065836715, + "grad_norm": 1.043473243713379, + "learning_rate": 2.6342405761229485e-06, + "loss": 0.4877, + "step": 26800 + }, + { + "epoch": 1.7986980302674407, + "grad_norm": 1.1618503332138062, + "learning_rate": 2.630760545294325e-06, + "loss": 0.4708, + "step": 26802 + }, + { + "epoch": 1.7988322539512098, + "grad_norm": 1.0864089727401733, + "learning_rate": 2.6272827525691104e-06, + "loss": 0.4528, + "step": 26804 + }, + { + "epoch": 1.7989664776349787, + "grad_norm": 0.9820147752761841, + "learning_rate": 2.6238071981116383e-06, + "loss": 0.4048, + "step": 26806 + }, + { + "epoch": 1.7991007013187477, + "grad_norm": 1.0986045598983765, + "learning_rate": 2.620333882086107e-06, + "loss": 0.5094, + "step": 26808 + }, + { + "epoch": 1.7992349250025166, + "grad_norm": 1.055762767791748, + "learning_rate": 2.616862804656639e-06, + "loss": 0.4684, + "step": 26810 + }, + { + "epoch": 1.7993691486862857, + "grad_norm": 1.0238934755325317, + "learning_rate": 2.6133939659872265e-06, + "loss": 0.3908, + "step": 26812 + }, + { + "epoch": 1.7995033723700546, + "grad_norm": 1.0113105773925781, + "learning_rate": 2.6099273662417713e-06, + "loss": 0.4477, + "step": 26814 + }, + { + "epoch": 1.7996375960538238, + "grad_norm": 1.0518611669540405, + "learning_rate": 2.6064630055840477e-06, + "loss": 0.4438, + "step": 26816 + }, + { + "epoch": 1.7997718197375927, + "grad_norm": 1.034557580947876, + "learning_rate": 2.6030008841777585e-06, + "loss": 0.4016, + "step": 26818 + }, + { + "epoch": 1.7999060434213616, + "grad_norm": 1.0221138000488281, + "learning_rate": 2.5995410021864787e-06, + "loss": 0.4985, + "step": 26820 + }, + { + "epoch": 1.8000402671051305, + "grad_norm": 1.2014293670654297, + "learning_rate": 2.596083359773677e-06, + "loss": 0.5002, + "step": 26822 + }, + { + "epoch": 1.8001744907888997, + "grad_norm": 1.0326858758926392, + "learning_rate": 2.5926279571027113e-06, + "loss": 0.5282, + "step": 26824 + }, + { + "epoch": 1.8003087144726688, + "grad_norm": 1.0132777690887451, + "learning_rate": 2.589174794336863e-06, + "loss": 0.3885, + "step": 26826 + }, + { + "epoch": 1.8004429381564377, + "grad_norm": 0.9559783339500427, + "learning_rate": 2.585723871639267e-06, + "loss": 0.47, + "step": 26828 + }, + { + "epoch": 1.8005771618402067, + "grad_norm": 1.0535095930099487, + "learning_rate": 2.5822751891729945e-06, + "loss": 0.4312, + "step": 26830 + }, + { + "epoch": 1.8007113855239756, + "grad_norm": 1.1592003107070923, + "learning_rate": 2.578828747100964e-06, + "loss": 0.4777, + "step": 26832 + }, + { + "epoch": 1.8008456092077447, + "grad_norm": 1.245162844657898, + "learning_rate": 2.5753845455860347e-06, + "loss": 0.5542, + "step": 26834 + }, + { + "epoch": 1.8009798328915139, + "grad_norm": 0.9433565735816956, + "learning_rate": 2.5719425847909206e-06, + "loss": 0.4472, + "step": 26836 + }, + { + "epoch": 1.8011140565752828, + "grad_norm": 1.0376611948013306, + "learning_rate": 2.5685028648782638e-06, + "loss": 0.4262, + "step": 26838 + }, + { + "epoch": 1.8012482802590517, + "grad_norm": 1.1596550941467285, + "learning_rate": 2.5650653860105733e-06, + "loss": 0.4543, + "step": 26840 + }, + { + "epoch": 1.8013825039428206, + "grad_norm": 1.075486421585083, + "learning_rate": 2.561630148350269e-06, + "loss": 0.4355, + "step": 26842 + }, + { + "epoch": 1.8015167276265898, + "grad_norm": 0.9907163977622986, + "learning_rate": 2.558197152059649e-06, + "loss": 0.4969, + "step": 26844 + }, + { + "epoch": 1.8016509513103587, + "grad_norm": 1.024023413658142, + "learning_rate": 2.5547663973009284e-06, + "loss": 0.5127, + "step": 26846 + }, + { + "epoch": 1.8017851749941278, + "grad_norm": 0.9813905954360962, + "learning_rate": 2.551337884236199e-06, + "loss": 0.4463, + "step": 26848 + }, + { + "epoch": 1.8019193986778967, + "grad_norm": 1.15963613986969, + "learning_rate": 2.547911613027454e-06, + "loss": 0.4663, + "step": 26850 + }, + { + "epoch": 1.8020536223616657, + "grad_norm": 1.0435926914215088, + "learning_rate": 2.5444875838365745e-06, + "loss": 0.4895, + "step": 26852 + }, + { + "epoch": 1.8021878460454346, + "grad_norm": 1.1256290674209595, + "learning_rate": 2.5410657968253317e-06, + "loss": 0.5043, + "step": 26854 + }, + { + "epoch": 1.8023220697292037, + "grad_norm": 1.3820325136184692, + "learning_rate": 2.5376462521554125e-06, + "loss": 0.4759, + "step": 26856 + }, + { + "epoch": 1.8024562934129729, + "grad_norm": 1.0215343236923218, + "learning_rate": 2.5342289499883766e-06, + "loss": 0.4456, + "step": 26858 + }, + { + "epoch": 1.8025905170967418, + "grad_norm": 0.9229339957237244, + "learning_rate": 2.5308138904856838e-06, + "loss": 0.4317, + "step": 26860 + }, + { + "epoch": 1.8027247407805107, + "grad_norm": 0.9657912850379944, + "learning_rate": 2.527401073808683e-06, + "loss": 0.5191, + "step": 26862 + }, + { + "epoch": 1.8028589644642796, + "grad_norm": 1.0605182647705078, + "learning_rate": 2.5239905001186394e-06, + "loss": 0.4714, + "step": 26864 + }, + { + "epoch": 1.8029931881480488, + "grad_norm": 1.1260061264038086, + "learning_rate": 2.520582169576685e-06, + "loss": 0.506, + "step": 26866 + }, + { + "epoch": 1.8031274118318177, + "grad_norm": 1.3092259168624878, + "learning_rate": 2.517176082343858e-06, + "loss": 0.4284, + "step": 26868 + }, + { + "epoch": 1.8032616355155868, + "grad_norm": 1.0388407707214355, + "learning_rate": 2.513772238581097e-06, + "loss": 0.4028, + "step": 26870 + }, + { + "epoch": 1.8033958591993557, + "grad_norm": 0.997028112411499, + "learning_rate": 2.5103706384492164e-06, + "loss": 0.4643, + "step": 26872 + }, + { + "epoch": 1.8035300828831247, + "grad_norm": 1.0246165990829468, + "learning_rate": 2.5069712821089277e-06, + "loss": 0.5137, + "step": 26874 + }, + { + "epoch": 1.8036643065668936, + "grad_norm": 0.8851560950279236, + "learning_rate": 2.503574169720868e-06, + "loss": 0.4453, + "step": 26876 + }, + { + "epoch": 1.8037985302506627, + "grad_norm": 0.9529306888580322, + "learning_rate": 2.5001793014455266e-06, + "loss": 0.4621, + "step": 26878 + }, + { + "epoch": 1.8039327539344319, + "grad_norm": 0.9867005348205566, + "learning_rate": 2.4967866774433136e-06, + "loss": 0.3798, + "step": 26880 + }, + { + "epoch": 1.8040669776182008, + "grad_norm": 0.9776849150657654, + "learning_rate": 2.4933962978745117e-06, + "loss": 0.4723, + "step": 26882 + }, + { + "epoch": 1.8042012013019697, + "grad_norm": 0.9732565879821777, + "learning_rate": 2.490008162899321e-06, + "loss": 0.4714, + "step": 26884 + }, + { + "epoch": 1.8043354249857386, + "grad_norm": 0.9912878274917603, + "learning_rate": 2.486622272677813e-06, + "loss": 0.4294, + "step": 26886 + }, + { + "epoch": 1.8044696486695078, + "grad_norm": 1.0751785039901733, + "learning_rate": 2.483238627369988e-06, + "loss": 0.4929, + "step": 26888 + }, + { + "epoch": 1.8046038723532767, + "grad_norm": 1.1649205684661865, + "learning_rate": 2.4798572271356846e-06, + "loss": 0.4399, + "step": 26890 + }, + { + "epoch": 1.8047380960370458, + "grad_norm": 0.9231991767883301, + "learning_rate": 2.4764780721346914e-06, + "loss": 0.5099, + "step": 26892 + }, + { + "epoch": 1.8048723197208147, + "grad_norm": 1.1894398927688599, + "learning_rate": 2.4731011625266477e-06, + "loss": 0.4499, + "step": 26894 + }, + { + "epoch": 1.8050065434045837, + "grad_norm": 0.9900402426719666, + "learning_rate": 2.4697264984711257e-06, + "loss": 0.4933, + "step": 26896 + }, + { + "epoch": 1.8051407670883526, + "grad_norm": 1.1936390399932861, + "learning_rate": 2.466354080127564e-06, + "loss": 0.4654, + "step": 26898 + }, + { + "epoch": 1.8052749907721217, + "grad_norm": 1.0529814958572388, + "learning_rate": 2.4629839076552974e-06, + "loss": 0.4659, + "step": 26900 + }, + { + "epoch": 1.8054092144558909, + "grad_norm": 0.9130473136901855, + "learning_rate": 2.459615981213559e-06, + "loss": 0.4362, + "step": 26902 + }, + { + "epoch": 1.8055434381396598, + "grad_norm": 1.0363892316818237, + "learning_rate": 2.456250300961488e-06, + "loss": 0.4582, + "step": 26904 + }, + { + "epoch": 1.8056776618234287, + "grad_norm": 1.144631266593933, + "learning_rate": 2.452886867058102e-06, + "loss": 0.4553, + "step": 26906 + }, + { + "epoch": 1.8058118855071976, + "grad_norm": 1.192743182182312, + "learning_rate": 2.449525679662312e-06, + "loss": 0.4986, + "step": 26908 + }, + { + "epoch": 1.8059461091909668, + "grad_norm": 0.9952451586723328, + "learning_rate": 2.44616673893292e-06, + "loss": 0.5182, + "step": 26910 + }, + { + "epoch": 1.806080332874736, + "grad_norm": 1.0130771398544312, + "learning_rate": 2.442810045028654e-06, + "loss": 0.5121, + "step": 26912 + }, + { + "epoch": 1.8062145565585048, + "grad_norm": 0.8701079487800598, + "learning_rate": 2.439455598108081e-06, + "loss": 0.4631, + "step": 26914 + }, + { + "epoch": 1.8063487802422737, + "grad_norm": 0.9158850312232971, + "learning_rate": 2.4361033983297255e-06, + "loss": 0.4168, + "step": 26916 + }, + { + "epoch": 1.8064830039260427, + "grad_norm": 1.0420955419540405, + "learning_rate": 2.432753445851943e-06, + "loss": 0.4193, + "step": 26918 + }, + { + "epoch": 1.8066172276098118, + "grad_norm": 1.009220004081726, + "learning_rate": 2.4294057408330184e-06, + "loss": 0.4435, + "step": 26920 + }, + { + "epoch": 1.8067514512935807, + "grad_norm": 1.074559211730957, + "learning_rate": 2.4260602834311364e-06, + "loss": 0.4526, + "step": 26922 + }, + { + "epoch": 1.8068856749773499, + "grad_norm": 1.014679193496704, + "learning_rate": 2.422717073804348e-06, + "loss": 0.4674, + "step": 26924 + }, + { + "epoch": 1.8070198986611188, + "grad_norm": 1.091285228729248, + "learning_rate": 2.4193761121106384e-06, + "loss": 0.4736, + "step": 26926 + }, + { + "epoch": 1.8071541223448877, + "grad_norm": 0.9874877333641052, + "learning_rate": 2.4160373985078256e-06, + "loss": 0.4707, + "step": 26928 + }, + { + "epoch": 1.8072883460286566, + "grad_norm": 1.039380669593811, + "learning_rate": 2.4127009331536832e-06, + "loss": 0.5182, + "step": 26930 + }, + { + "epoch": 1.8074225697124258, + "grad_norm": 1.1278899908065796, + "learning_rate": 2.409366716205841e-06, + "loss": 0.481, + "step": 26932 + }, + { + "epoch": 1.807556793396195, + "grad_norm": 1.0794038772583008, + "learning_rate": 2.4060347478218446e-06, + "loss": 0.5274, + "step": 26934 + }, + { + "epoch": 1.8076910170799638, + "grad_norm": 1.056154727935791, + "learning_rate": 2.402705028159119e-06, + "loss": 0.4559, + "step": 26936 + }, + { + "epoch": 1.8078252407637327, + "grad_norm": 0.8898627758026123, + "learning_rate": 2.399377557374982e-06, + "loss": 0.44, + "step": 26938 + }, + { + "epoch": 1.8079594644475017, + "grad_norm": 0.9880284667015076, + "learning_rate": 2.3960523356266475e-06, + "loss": 0.4151, + "step": 26940 + }, + { + "epoch": 1.8080936881312708, + "grad_norm": 0.878262996673584, + "learning_rate": 2.3927293630712332e-06, + "loss": 0.4972, + "step": 26942 + }, + { + "epoch": 1.8082279118150397, + "grad_norm": 0.9728204607963562, + "learning_rate": 2.3894086398657478e-06, + "loss": 0.4705, + "step": 26944 + }, + { + "epoch": 1.8083621354988089, + "grad_norm": 0.9428135752677917, + "learning_rate": 2.3860901661670764e-06, + "loss": 0.4006, + "step": 26946 + }, + { + "epoch": 1.8084963591825778, + "grad_norm": 0.9841963648796082, + "learning_rate": 2.3827739421320105e-06, + "loss": 0.4725, + "step": 26948 + }, + { + "epoch": 1.8086305828663467, + "grad_norm": 1.0408563613891602, + "learning_rate": 2.379459967917247e-06, + "loss": 0.4695, + "step": 26950 + }, + { + "epoch": 1.8087648065501156, + "grad_norm": 1.026936411857605, + "learning_rate": 2.376148243679355e-06, + "loss": 0.5047, + "step": 26952 + }, + { + "epoch": 1.8088990302338848, + "grad_norm": 1.1736093759536743, + "learning_rate": 2.37283876957482e-06, + "loss": 0.5216, + "step": 26954 + }, + { + "epoch": 1.809033253917654, + "grad_norm": 1.0411231517791748, + "learning_rate": 2.369531545759984e-06, + "loss": 0.4811, + "step": 26956 + }, + { + "epoch": 1.8091674776014228, + "grad_norm": 1.0473124980926514, + "learning_rate": 2.3662265723911337e-06, + "loss": 0.4326, + "step": 26958 + }, + { + "epoch": 1.8093017012851917, + "grad_norm": 0.8875226974487305, + "learning_rate": 2.3629238496243987e-06, + "loss": 0.4732, + "step": 26960 + }, + { + "epoch": 1.8094359249689607, + "grad_norm": 0.9540635347366333, + "learning_rate": 2.359623377615844e-06, + "loss": 0.4236, + "step": 26962 + }, + { + "epoch": 1.8095701486527298, + "grad_norm": 1.1037060022354126, + "learning_rate": 2.356325156521405e-06, + "loss": 0.4242, + "step": 26964 + }, + { + "epoch": 1.8097043723364987, + "grad_norm": 0.9551806449890137, + "learning_rate": 2.353029186496919e-06, + "loss": 0.4801, + "step": 26966 + }, + { + "epoch": 1.8098385960202679, + "grad_norm": 1.0264946222305298, + "learning_rate": 2.349735467698094e-06, + "loss": 0.4097, + "step": 26968 + }, + { + "epoch": 1.8099728197040368, + "grad_norm": 0.9813930988311768, + "learning_rate": 2.3464440002805844e-06, + "loss": 0.4443, + "step": 26970 + }, + { + "epoch": 1.8101070433878057, + "grad_norm": 1.1017805337905884, + "learning_rate": 2.3431547843998815e-06, + "loss": 0.4758, + "step": 26972 + }, + { + "epoch": 1.8102412670715746, + "grad_norm": 0.9897599816322327, + "learning_rate": 2.3398678202114054e-06, + "loss": 0.4848, + "step": 26974 + }, + { + "epoch": 1.8103754907553438, + "grad_norm": 1.0189718008041382, + "learning_rate": 2.336583107870449e-06, + "loss": 0.4375, + "step": 26976 + }, + { + "epoch": 1.810509714439113, + "grad_norm": 1.0726696252822876, + "learning_rate": 2.333300647532222e-06, + "loss": 0.4518, + "step": 26978 + }, + { + "epoch": 1.8106439381228818, + "grad_norm": 0.9022804498672485, + "learning_rate": 2.330020439351799e-06, + "loss": 0.4779, + "step": 26980 + }, + { + "epoch": 1.8107781618066507, + "grad_norm": 1.4354796409606934, + "learning_rate": 2.3267424834841845e-06, + "loss": 0.5082, + "step": 26982 + }, + { + "epoch": 1.8109123854904197, + "grad_norm": 0.8356832265853882, + "learning_rate": 2.323466780084227e-06, + "loss": 0.3892, + "step": 26984 + }, + { + "epoch": 1.8110466091741888, + "grad_norm": 1.0602096319198608, + "learning_rate": 2.3201933293067247e-06, + "loss": 0.4176, + "step": 26986 + }, + { + "epoch": 1.811180832857958, + "grad_norm": 1.0468652248382568, + "learning_rate": 2.3169221313063207e-06, + "loss": 0.468, + "step": 26988 + }, + { + "epoch": 1.8113150565417269, + "grad_norm": 1.003930687904358, + "learning_rate": 2.31365318623758e-06, + "loss": 0.4576, + "step": 26990 + }, + { + "epoch": 1.8114492802254958, + "grad_norm": 1.1603525876998901, + "learning_rate": 2.3103864942549623e-06, + "loss": 0.493, + "step": 26992 + }, + { + "epoch": 1.8115835039092647, + "grad_norm": 1.0017002820968628, + "learning_rate": 2.307122055512806e-06, + "loss": 0.4167, + "step": 26994 + }, + { + "epoch": 1.8117177275930338, + "grad_norm": 1.084191083908081, + "learning_rate": 2.3038598701653425e-06, + "loss": 0.4877, + "step": 26996 + }, + { + "epoch": 1.8118519512768028, + "grad_norm": 0.9252565503120422, + "learning_rate": 2.30059993836671e-06, + "loss": 0.415, + "step": 26998 + }, + { + "epoch": 1.811986174960572, + "grad_norm": 1.047005534172058, + "learning_rate": 2.297342260270935e-06, + "loss": 0.4355, + "step": 27000 + }, + { + "epoch": 1.8121203986443408, + "grad_norm": 1.072208285331726, + "learning_rate": 2.294086836031939e-06, + "loss": 0.4838, + "step": 27002 + }, + { + "epoch": 1.8122546223281097, + "grad_norm": 1.0251212120056152, + "learning_rate": 2.2908336658035266e-06, + "loss": 0.3947, + "step": 27004 + }, + { + "epoch": 1.8123888460118787, + "grad_norm": 1.0754388570785522, + "learning_rate": 2.287582749739403e-06, + "loss": 0.4685, + "step": 27006 + }, + { + "epoch": 1.8125230696956478, + "grad_norm": 1.0514887571334839, + "learning_rate": 2.2843340879931783e-06, + "loss": 0.4582, + "step": 27008 + }, + { + "epoch": 1.812657293379417, + "grad_norm": 1.3117904663085938, + "learning_rate": 2.2810876807183414e-06, + "loss": 0.4585, + "step": 27010 + }, + { + "epoch": 1.8127915170631859, + "grad_norm": 1.0328198671340942, + "learning_rate": 2.277843528068274e-06, + "loss": 0.4514, + "step": 27012 + }, + { + "epoch": 1.8129257407469548, + "grad_norm": 0.9663453102111816, + "learning_rate": 2.274601630196249e-06, + "loss": 0.4671, + "step": 27014 + }, + { + "epoch": 1.8130599644307237, + "grad_norm": 1.099232792854309, + "learning_rate": 2.2713619872554603e-06, + "loss": 0.4821, + "step": 27016 + }, + { + "epoch": 1.8131941881144928, + "grad_norm": 1.1244513988494873, + "learning_rate": 2.268124599398952e-06, + "loss": 0.4422, + "step": 27018 + }, + { + "epoch": 1.8133284117982618, + "grad_norm": 1.174409031867981, + "learning_rate": 2.264889466779707e-06, + "loss": 0.4618, + "step": 27020 + }, + { + "epoch": 1.813462635482031, + "grad_norm": 1.0032176971435547, + "learning_rate": 2.261656589550565e-06, + "loss": 0.4631, + "step": 27022 + }, + { + "epoch": 1.8135968591657998, + "grad_norm": 1.0601005554199219, + "learning_rate": 2.258425967864275e-06, + "loss": 0.4899, + "step": 27024 + }, + { + "epoch": 1.8137310828495687, + "grad_norm": 1.045218825340271, + "learning_rate": 2.2551976018734766e-06, + "loss": 0.5004, + "step": 27026 + }, + { + "epoch": 1.8138653065333377, + "grad_norm": 1.0657036304473877, + "learning_rate": 2.251971491730709e-06, + "loss": 0.4616, + "step": 27028 + }, + { + "epoch": 1.8139995302171068, + "grad_norm": 4.555449485778809, + "learning_rate": 2.248747637588394e-06, + "loss": 0.4756, + "step": 27030 + }, + { + "epoch": 1.814133753900876, + "grad_norm": 1.0749096870422363, + "learning_rate": 2.2455260395988553e-06, + "loss": 0.4449, + "step": 27032 + }, + { + "epoch": 1.8142679775846449, + "grad_norm": 1.120177984237671, + "learning_rate": 2.242306697914298e-06, + "loss": 0.4682, + "step": 27034 + }, + { + "epoch": 1.8144022012684138, + "grad_norm": 1.0862208604812622, + "learning_rate": 2.2390896126868453e-06, + "loss": 0.5205, + "step": 27036 + }, + { + "epoch": 1.8145364249521827, + "grad_norm": 1.0489317178726196, + "learning_rate": 2.2358747840684924e-06, + "loss": 0.4493, + "step": 27038 + }, + { + "epoch": 1.8146706486359518, + "grad_norm": 0.9261298775672913, + "learning_rate": 2.232662212211134e-06, + "loss": 0.4632, + "step": 27040 + }, + { + "epoch": 1.8148048723197208, + "grad_norm": 0.7837489247322083, + "learning_rate": 2.2294518972665434e-06, + "loss": 0.4459, + "step": 27042 + }, + { + "epoch": 1.81493909600349, + "grad_norm": 0.9737602472305298, + "learning_rate": 2.2262438393864214e-06, + "loss": 0.4512, + "step": 27044 + }, + { + "epoch": 1.8150733196872588, + "grad_norm": 1.214451551437378, + "learning_rate": 2.2230380387223305e-06, + "loss": 0.5024, + "step": 27046 + }, + { + "epoch": 1.8152075433710277, + "grad_norm": 0.860659122467041, + "learning_rate": 2.219834495425749e-06, + "loss": 0.377, + "step": 27048 + }, + { + "epoch": 1.8153417670547967, + "grad_norm": 0.9898740649223328, + "learning_rate": 2.2166332096480337e-06, + "loss": 0.4767, + "step": 27050 + }, + { + "epoch": 1.8154759907385658, + "grad_norm": 1.0652974843978882, + "learning_rate": 2.213434181540436e-06, + "loss": 0.4188, + "step": 27052 + }, + { + "epoch": 1.815610214422335, + "grad_norm": 1.0848276615142822, + "learning_rate": 2.2102374112541013e-06, + "loss": 0.4586, + "step": 27054 + }, + { + "epoch": 1.8157444381061039, + "grad_norm": 0.9306811094284058, + "learning_rate": 2.207042898940076e-06, + "loss": 0.447, + "step": 27056 + }, + { + "epoch": 1.8158786617898728, + "grad_norm": 1.0598939657211304, + "learning_rate": 2.2038506447492945e-06, + "loss": 0.4723, + "step": 27058 + }, + { + "epoch": 1.8160128854736417, + "grad_norm": 1.1055607795715332, + "learning_rate": 2.200660648832581e-06, + "loss": 0.4762, + "step": 27060 + }, + { + "epoch": 1.8161471091574108, + "grad_norm": 1.04423189163208, + "learning_rate": 2.197472911340659e-06, + "loss": 0.486, + "step": 27062 + }, + { + "epoch": 1.81628133284118, + "grad_norm": 1.1375607252120972, + "learning_rate": 2.1942874324241357e-06, + "loss": 0.5542, + "step": 27064 + }, + { + "epoch": 1.816415556524949, + "grad_norm": 1.0168747901916504, + "learning_rate": 2.1911042122335356e-06, + "loss": 0.571, + "step": 27066 + }, + { + "epoch": 1.8165497802087178, + "grad_norm": 1.0431084632873535, + "learning_rate": 2.1879232509192494e-06, + "loss": 0.4328, + "step": 27068 + }, + { + "epoch": 1.8166840038924867, + "grad_norm": 1.0373541116714478, + "learning_rate": 2.1847445486315675e-06, + "loss": 0.4037, + "step": 27070 + }, + { + "epoch": 1.8168182275762559, + "grad_norm": 1.02206289768219, + "learning_rate": 2.181568105520676e-06, + "loss": 0.4972, + "step": 27072 + }, + { + "epoch": 1.8169524512600248, + "grad_norm": 1.0911699533462524, + "learning_rate": 2.1783939217366655e-06, + "loss": 0.4703, + "step": 27074 + }, + { + "epoch": 1.817086674943794, + "grad_norm": 0.9556540846824646, + "learning_rate": 2.1752219974294995e-06, + "loss": 0.4669, + "step": 27076 + }, + { + "epoch": 1.8172208986275629, + "grad_norm": 1.0169297456741333, + "learning_rate": 2.1720523327490636e-06, + "loss": 0.4742, + "step": 27078 + }, + { + "epoch": 1.8173551223113318, + "grad_norm": 0.8982275724411011, + "learning_rate": 2.168884927845094e-06, + "loss": 0.4801, + "step": 27080 + }, + { + "epoch": 1.8174893459951007, + "grad_norm": 0.9359860420227051, + "learning_rate": 2.165719782867259e-06, + "loss": 0.488, + "step": 27082 + }, + { + "epoch": 1.8176235696788698, + "grad_norm": 0.9763116836547852, + "learning_rate": 2.1625568979651014e-06, + "loss": 0.4066, + "step": 27084 + }, + { + "epoch": 1.817757793362639, + "grad_norm": 0.8585172891616821, + "learning_rate": 2.1593962732880615e-06, + "loss": 0.4162, + "step": 27086 + }, + { + "epoch": 1.817892017046408, + "grad_norm": 1.0699492692947388, + "learning_rate": 2.1562379089854755e-06, + "loss": 0.4776, + "step": 27088 + }, + { + "epoch": 1.8180262407301768, + "grad_norm": 1.0023192167282104, + "learning_rate": 2.1530818052065695e-06, + "loss": 0.4617, + "step": 27090 + }, + { + "epoch": 1.8181604644139457, + "grad_norm": 1.0295124053955078, + "learning_rate": 2.1499279621004564e-06, + "loss": 0.484, + "step": 27092 + }, + { + "epoch": 1.8182946880977149, + "grad_norm": 1.0506134033203125, + "learning_rate": 2.146776379816157e-06, + "loss": 0.4966, + "step": 27094 + }, + { + "epoch": 1.8184289117814838, + "grad_norm": 1.12366783618927, + "learning_rate": 2.1436270585025732e-06, + "loss": 0.4895, + "step": 27096 + }, + { + "epoch": 1.818563135465253, + "grad_norm": 1.0606402158737183, + "learning_rate": 2.140479998308509e-06, + "loss": 0.546, + "step": 27098 + }, + { + "epoch": 1.8186973591490219, + "grad_norm": 1.046035885810852, + "learning_rate": 2.137335199382645e-06, + "loss": 0.4516, + "step": 27100 + }, + { + "epoch": 1.8188315828327908, + "grad_norm": 0.9801777601242065, + "learning_rate": 2.1341926618735796e-06, + "loss": 0.48, + "step": 27102 + }, + { + "epoch": 1.8189658065165597, + "grad_norm": 0.9874941110610962, + "learning_rate": 2.1310523859297828e-06, + "loss": 0.4771, + "step": 27104 + }, + { + "epoch": 1.8191000302003288, + "grad_norm": 0.9163682460784912, + "learning_rate": 2.1279143716996464e-06, + "loss": 0.4498, + "step": 27106 + }, + { + "epoch": 1.819234253884098, + "grad_norm": 0.8990226984024048, + "learning_rate": 2.1247786193314025e-06, + "loss": 0.4924, + "step": 27108 + }, + { + "epoch": 1.819368477567867, + "grad_norm": 1.0113532543182373, + "learning_rate": 2.121645128973232e-06, + "loss": 0.4578, + "step": 27110 + }, + { + "epoch": 1.8195027012516358, + "grad_norm": 1.1612234115600586, + "learning_rate": 2.118513900773178e-06, + "loss": 0.4687, + "step": 27112 + }, + { + "epoch": 1.8196369249354047, + "grad_norm": 1.055635929107666, + "learning_rate": 2.1153849348791887e-06, + "loss": 0.468, + "step": 27114 + }, + { + "epoch": 1.8197711486191739, + "grad_norm": 0.9631549119949341, + "learning_rate": 2.1122582314391013e-06, + "loss": 0.4729, + "step": 27116 + }, + { + "epoch": 1.8199053723029428, + "grad_norm": 0.9034852981567383, + "learning_rate": 2.1091337906006482e-06, + "loss": 0.4397, + "step": 27118 + }, + { + "epoch": 1.820039595986712, + "grad_norm": 1.051687240600586, + "learning_rate": 2.1060116125114436e-06, + "loss": 0.5044, + "step": 27120 + }, + { + "epoch": 1.8201738196704809, + "grad_norm": 0.9925054907798767, + "learning_rate": 2.102891697319015e-06, + "loss": 0.4694, + "step": 27122 + }, + { + "epoch": 1.8203080433542498, + "grad_norm": 1.147497296333313, + "learning_rate": 2.099774045170766e-06, + "loss": 0.4803, + "step": 27124 + }, + { + "epoch": 1.8204422670380187, + "grad_norm": 0.9984075427055359, + "learning_rate": 2.096658656214007e-06, + "loss": 0.5095, + "step": 27126 + }, + { + "epoch": 1.8205764907217878, + "grad_norm": 0.8068634271621704, + "learning_rate": 2.093545530595925e-06, + "loss": 0.4256, + "step": 27128 + }, + { + "epoch": 1.820710714405557, + "grad_norm": 1.0053539276123047, + "learning_rate": 2.090434668463609e-06, + "loss": 0.4349, + "step": 27130 + }, + { + "epoch": 1.820844938089326, + "grad_norm": 0.77165287733078, + "learning_rate": 2.0873260699640462e-06, + "loss": 0.423, + "step": 27132 + }, + { + "epoch": 1.8209791617730948, + "grad_norm": 1.0810253620147705, + "learning_rate": 2.084219735244114e-06, + "loss": 0.4758, + "step": 27134 + }, + { + "epoch": 1.8211133854568637, + "grad_norm": 1.009852409362793, + "learning_rate": 2.081115664450578e-06, + "loss": 0.4755, + "step": 27136 + }, + { + "epoch": 1.8212476091406329, + "grad_norm": 1.0758378505706787, + "learning_rate": 2.078013857730088e-06, + "loss": 0.4598, + "step": 27138 + }, + { + "epoch": 1.821381832824402, + "grad_norm": 1.057146668434143, + "learning_rate": 2.0749143152292204e-06, + "loss": 0.4423, + "step": 27140 + }, + { + "epoch": 1.821516056508171, + "grad_norm": 0.950028121471405, + "learning_rate": 2.071817037094398e-06, + "loss": 0.4749, + "step": 27142 + }, + { + "epoch": 1.8216502801919399, + "grad_norm": 0.9723566770553589, + "learning_rate": 2.068722023471992e-06, + "loss": 0.4895, + "step": 27144 + }, + { + "epoch": 1.8217845038757088, + "grad_norm": 1.006791114807129, + "learning_rate": 2.065629274508202e-06, + "loss": 0.5504, + "step": 27146 + }, + { + "epoch": 1.821918727559478, + "grad_norm": 1.1130188703536987, + "learning_rate": 2.062538790349178e-06, + "loss": 0.4283, + "step": 27148 + }, + { + "epoch": 1.8220529512432468, + "grad_norm": 0.988906741142273, + "learning_rate": 2.059450571140925e-06, + "loss": 0.4908, + "step": 27150 + }, + { + "epoch": 1.822187174927016, + "grad_norm": 1.1045299768447876, + "learning_rate": 2.056364617029366e-06, + "loss": 0.5038, + "step": 27152 + }, + { + "epoch": 1.822321398610785, + "grad_norm": 1.1224522590637207, + "learning_rate": 2.0532809281603062e-06, + "loss": 0.4549, + "step": 27154 + }, + { + "epoch": 1.8224556222945538, + "grad_norm": 1.1844680309295654, + "learning_rate": 2.050199504679434e-06, + "loss": 0.4976, + "step": 27156 + }, + { + "epoch": 1.8225898459783227, + "grad_norm": 0.9185380935668945, + "learning_rate": 2.0471203467323398e-06, + "loss": 0.4312, + "step": 27158 + }, + { + "epoch": 1.8227240696620919, + "grad_norm": 0.8855293989181519, + "learning_rate": 2.0440434544645227e-06, + "loss": 0.435, + "step": 27160 + }, + { + "epoch": 1.822858293345861, + "grad_norm": 1.1739426851272583, + "learning_rate": 2.0409688280213503e-06, + "loss": 0.4522, + "step": 27162 + }, + { + "epoch": 1.82299251702963, + "grad_norm": 1.1525919437408447, + "learning_rate": 2.037896467548095e-06, + "loss": 0.537, + "step": 27164 + }, + { + "epoch": 1.8231267407133989, + "grad_norm": 0.9660004377365112, + "learning_rate": 2.0348263731899076e-06, + "loss": 0.4421, + "step": 27166 + }, + { + "epoch": 1.8232609643971678, + "grad_norm": 1.1061416864395142, + "learning_rate": 2.0317585450918663e-06, + "loss": 0.5127, + "step": 27168 + }, + { + "epoch": 1.823395188080937, + "grad_norm": 1.0320607423782349, + "learning_rate": 2.0286929833988943e-06, + "loss": 0.4894, + "step": 27170 + }, + { + "epoch": 1.8235294117647058, + "grad_norm": 0.9689834713935852, + "learning_rate": 2.025629688255859e-06, + "loss": 0.4341, + "step": 27172 + }, + { + "epoch": 1.823663635448475, + "grad_norm": 1.2229254245758057, + "learning_rate": 2.0225686598074835e-06, + "loss": 0.4455, + "step": 27174 + }, + { + "epoch": 1.823797859132244, + "grad_norm": 0.9925333857536316, + "learning_rate": 2.0195098981983963e-06, + "loss": 0.4776, + "step": 27176 + }, + { + "epoch": 1.8239320828160128, + "grad_norm": 1.0733305215835571, + "learning_rate": 2.01645340357311e-06, + "loss": 0.4905, + "step": 27178 + }, + { + "epoch": 1.8240663064997817, + "grad_norm": 0.9591812491416931, + "learning_rate": 2.0133991760760475e-06, + "loss": 0.451, + "step": 27180 + }, + { + "epoch": 1.8242005301835509, + "grad_norm": 0.9299895763397217, + "learning_rate": 2.0103472158515158e-06, + "loss": 0.4195, + "step": 27182 + }, + { + "epoch": 1.82433475386732, + "grad_norm": 0.9533336758613586, + "learning_rate": 2.007297523043711e-06, + "loss": 0.4337, + "step": 27184 + }, + { + "epoch": 1.824468977551089, + "grad_norm": 0.9015142321586609, + "learning_rate": 2.0042500977967172e-06, + "loss": 0.4188, + "step": 27186 + }, + { + "epoch": 1.8246032012348579, + "grad_norm": 1.0596815347671509, + "learning_rate": 2.0012049402545306e-06, + "loss": 0.4273, + "step": 27188 + }, + { + "epoch": 1.8247374249186268, + "grad_norm": 1.067183256149292, + "learning_rate": 1.9981620505610255e-06, + "loss": 0.4149, + "step": 27190 + }, + { + "epoch": 1.824871648602396, + "grad_norm": 1.2220438718795776, + "learning_rate": 1.995121428859975e-06, + "loss": 0.5013, + "step": 27192 + }, + { + "epoch": 1.8250058722861648, + "grad_norm": 1.0703747272491455, + "learning_rate": 1.992083075295037e-06, + "loss": 0.4664, + "step": 27194 + }, + { + "epoch": 1.825140095969934, + "grad_norm": 1.1133596897125244, + "learning_rate": 1.989046990009763e-06, + "loss": 0.4828, + "step": 27196 + }, + { + "epoch": 1.825274319653703, + "grad_norm": 1.0013623237609863, + "learning_rate": 1.9860131731476217e-06, + "loss": 0.4463, + "step": 27198 + }, + { + "epoch": 1.8254085433374718, + "grad_norm": 1.1642730236053467, + "learning_rate": 1.982981624851932e-06, + "loss": 0.45, + "step": 27200 + }, + { + "epoch": 1.8255427670212407, + "grad_norm": 1.0820966958999634, + "learning_rate": 1.979952345265951e-06, + "loss": 0.4992, + "step": 27202 + }, + { + "epoch": 1.8256769907050099, + "grad_norm": 1.0612512826919556, + "learning_rate": 1.976925334532781e-06, + "loss": 0.4568, + "step": 27204 + }, + { + "epoch": 1.825811214388779, + "grad_norm": 1.078147292137146, + "learning_rate": 1.9739005927954633e-06, + "loss": 0.4685, + "step": 27206 + }, + { + "epoch": 1.825945438072548, + "grad_norm": 1.1312330961227417, + "learning_rate": 1.9708781201968952e-06, + "loss": 0.5071, + "step": 27208 + }, + { + "epoch": 1.8260796617563169, + "grad_norm": 1.0434460639953613, + "learning_rate": 1.9678579168799004e-06, + "loss": 0.4322, + "step": 27210 + }, + { + "epoch": 1.8262138854400858, + "grad_norm": 1.106642723083496, + "learning_rate": 1.9648399829871654e-06, + "loss": 0.453, + "step": 27212 + }, + { + "epoch": 1.826348109123855, + "grad_norm": 1.0751092433929443, + "learning_rate": 1.9618243186612815e-06, + "loss": 0.4817, + "step": 27214 + }, + { + "epoch": 1.826482332807624, + "grad_norm": 0.8817471265792847, + "learning_rate": 1.958810924044735e-06, + "loss": 0.4875, + "step": 27216 + }, + { + "epoch": 1.826616556491393, + "grad_norm": 0.8890557289123535, + "learning_rate": 1.955799799279906e-06, + "loss": 0.4036, + "step": 27218 + }, + { + "epoch": 1.826750780175162, + "grad_norm": 0.9986615180969238, + "learning_rate": 1.952790944509064e-06, + "loss": 0.458, + "step": 27220 + }, + { + "epoch": 1.8268850038589308, + "grad_norm": 0.9373031258583069, + "learning_rate": 1.949784359874368e-06, + "loss": 0.4374, + "step": 27222 + }, + { + "epoch": 1.8270192275427, + "grad_norm": 1.0496248006820679, + "learning_rate": 1.9467800455178654e-06, + "loss": 0.497, + "step": 27224 + }, + { + "epoch": 1.8271534512264689, + "grad_norm": 1.1269254684448242, + "learning_rate": 1.9437780015815254e-06, + "loss": 0.5425, + "step": 27226 + }, + { + "epoch": 1.827287674910238, + "grad_norm": 1.1114847660064697, + "learning_rate": 1.940778228207163e-06, + "loss": 0.4086, + "step": 27228 + }, + { + "epoch": 1.827421898594007, + "grad_norm": 1.167726993560791, + "learning_rate": 1.937780725536542e-06, + "loss": 0.4717, + "step": 27230 + }, + { + "epoch": 1.8275561222777759, + "grad_norm": 0.7843297123908997, + "learning_rate": 1.9347854937112606e-06, + "loss": 0.4127, + "step": 27232 + }, + { + "epoch": 1.8276903459615448, + "grad_norm": 0.9357795119285583, + "learning_rate": 1.9317925328728504e-06, + "loss": 0.466, + "step": 27234 + }, + { + "epoch": 1.827824569645314, + "grad_norm": 0.9309419393539429, + "learning_rate": 1.9288018431627143e-06, + "loss": 0.3674, + "step": 27236 + }, + { + "epoch": 1.827958793329083, + "grad_norm": 1.0073319673538208, + "learning_rate": 1.9258134247221725e-06, + "loss": 0.5228, + "step": 27238 + }, + { + "epoch": 1.828093017012852, + "grad_norm": 1.0502758026123047, + "learning_rate": 1.922827277692413e-06, + "loss": 0.4475, + "step": 27240 + }, + { + "epoch": 1.828227240696621, + "grad_norm": 1.1033629179000854, + "learning_rate": 1.9198434022145216e-06, + "loss": 0.4783, + "step": 27242 + }, + { + "epoch": 1.8283614643803898, + "grad_norm": 1.001257300376892, + "learning_rate": 1.916861798429481e-06, + "loss": 0.5018, + "step": 27244 + }, + { + "epoch": 1.828495688064159, + "grad_norm": 1.077662706375122, + "learning_rate": 1.913882466478173e-06, + "loss": 0.4509, + "step": 27246 + }, + { + "epoch": 1.8286299117479279, + "grad_norm": 1.1423128843307495, + "learning_rate": 1.910905406501362e-06, + "loss": 0.5072, + "step": 27248 + }, + { + "epoch": 1.828764135431697, + "grad_norm": 1.0833154916763306, + "learning_rate": 1.907930618639703e-06, + "loss": 0.5248, + "step": 27250 + }, + { + "epoch": 1.828898359115466, + "grad_norm": 1.0415693521499634, + "learning_rate": 1.9049581030337548e-06, + "loss": 0.4609, + "step": 27252 + }, + { + "epoch": 1.8290325827992349, + "grad_norm": 1.0075165033340454, + "learning_rate": 1.9019878598239615e-06, + "loss": 0.4583, + "step": 27254 + }, + { + "epoch": 1.8291668064830038, + "grad_norm": 0.9416204690933228, + "learning_rate": 1.8990198891506605e-06, + "loss": 0.4052, + "step": 27256 + }, + { + "epoch": 1.829301030166773, + "grad_norm": 1.0448124408721924, + "learning_rate": 1.8960541911540896e-06, + "loss": 0.4682, + "step": 27258 + }, + { + "epoch": 1.829435253850542, + "grad_norm": 1.0680210590362549, + "learning_rate": 1.8930907659743646e-06, + "loss": 0.4813, + "step": 27260 + }, + { + "epoch": 1.829569477534311, + "grad_norm": 0.9063601493835449, + "learning_rate": 1.89012961375149e-06, + "loss": 0.5119, + "step": 27262 + }, + { + "epoch": 1.82970370121808, + "grad_norm": 1.0432617664337158, + "learning_rate": 1.8871707346253986e-06, + "loss": 0.4449, + "step": 27264 + }, + { + "epoch": 1.8298379249018488, + "grad_norm": 0.9900143146514893, + "learning_rate": 1.8842141287358727e-06, + "loss": 0.4451, + "step": 27266 + }, + { + "epoch": 1.829972148585618, + "grad_norm": 1.103989839553833, + "learning_rate": 1.8812597962226286e-06, + "loss": 0.4721, + "step": 27268 + }, + { + "epoch": 1.8301063722693869, + "grad_norm": 1.019983172416687, + "learning_rate": 1.878307737225221e-06, + "loss": 0.5002, + "step": 27270 + }, + { + "epoch": 1.830240595953156, + "grad_norm": 1.1201612949371338, + "learning_rate": 1.8753579518831554e-06, + "loss": 0.5069, + "step": 27272 + }, + { + "epoch": 1.830374819636925, + "grad_norm": 1.0612053871154785, + "learning_rate": 1.8724104403357868e-06, + "loss": 0.4793, + "step": 27274 + }, + { + "epoch": 1.8305090433206939, + "grad_norm": 0.7627202272415161, + "learning_rate": 1.8694652027223925e-06, + "loss": 0.4046, + "step": 27276 + }, + { + "epoch": 1.8306432670044628, + "grad_norm": 0.9914665222167969, + "learning_rate": 1.8665222391821169e-06, + "loss": 0.4982, + "step": 27278 + }, + { + "epoch": 1.830777490688232, + "grad_norm": 0.8946560621261597, + "learning_rate": 1.863581549854021e-06, + "loss": 0.4352, + "step": 27280 + }, + { + "epoch": 1.830911714372001, + "grad_norm": 1.0527199506759644, + "learning_rate": 1.8606431348770325e-06, + "loss": 0.4364, + "step": 27282 + }, + { + "epoch": 1.83104593805577, + "grad_norm": 1.009649634361267, + "learning_rate": 1.8577069943900017e-06, + "loss": 0.4558, + "step": 27284 + }, + { + "epoch": 1.831180161739539, + "grad_norm": 1.0558478832244873, + "learning_rate": 1.854773128531645e-06, + "loss": 0.4794, + "step": 27286 + }, + { + "epoch": 1.8313143854233078, + "grad_norm": 0.9969227910041809, + "learning_rate": 1.851841537440585e-06, + "loss": 0.4659, + "step": 27288 + }, + { + "epoch": 1.831448609107077, + "grad_norm": 1.1174561977386475, + "learning_rate": 1.8489122212553279e-06, + "loss": 0.4979, + "step": 27290 + }, + { + "epoch": 1.831582832790846, + "grad_norm": 1.0655980110168457, + "learning_rate": 1.8459851801142902e-06, + "loss": 0.4702, + "step": 27292 + }, + { + "epoch": 1.831717056474615, + "grad_norm": 1.2060457468032837, + "learning_rate": 1.8430604141557562e-06, + "loss": 0.4752, + "step": 27294 + }, + { + "epoch": 1.831851280158384, + "grad_norm": 0.9956331253051758, + "learning_rate": 1.8401379235179316e-06, + "loss": 0.4351, + "step": 27296 + }, + { + "epoch": 1.8319855038421529, + "grad_norm": 1.0068447589874268, + "learning_rate": 1.8372177083388786e-06, + "loss": 0.4986, + "step": 27298 + }, + { + "epoch": 1.832119727525922, + "grad_norm": 1.085825800895691, + "learning_rate": 1.834299768756581e-06, + "loss": 0.5099, + "step": 27300 + }, + { + "epoch": 1.832253951209691, + "grad_norm": 1.0742210149765015, + "learning_rate": 1.8313841049089065e-06, + "loss": 0.4737, + "step": 27302 + }, + { + "epoch": 1.83238817489346, + "grad_norm": 1.0905416011810303, + "learning_rate": 1.8284707169336169e-06, + "loss": 0.4125, + "step": 27304 + }, + { + "epoch": 1.832522398577229, + "grad_norm": 1.0812736749649048, + "learning_rate": 1.8255596049683576e-06, + "loss": 0.4896, + "step": 27306 + }, + { + "epoch": 1.832656622260998, + "grad_norm": 1.1849606037139893, + "learning_rate": 1.82265076915068e-06, + "loss": 0.4589, + "step": 27308 + }, + { + "epoch": 1.8327908459447668, + "grad_norm": 1.093017339706421, + "learning_rate": 1.8197442096180072e-06, + "loss": 0.4888, + "step": 27310 + }, + { + "epoch": 1.832925069628536, + "grad_norm": 1.0499763488769531, + "learning_rate": 1.8168399265076852e-06, + "loss": 0.4531, + "step": 27312 + }, + { + "epoch": 1.833059293312305, + "grad_norm": 0.9453720450401306, + "learning_rate": 1.8139379199569318e-06, + "loss": 0.4253, + "step": 27314 + }, + { + "epoch": 1.833193516996074, + "grad_norm": 1.0987991094589233, + "learning_rate": 1.8110381901028539e-06, + "loss": 0.44, + "step": 27316 + }, + { + "epoch": 1.833327740679843, + "grad_norm": 1.2033222913742065, + "learning_rate": 1.8081407370824588e-06, + "loss": 0.4005, + "step": 27318 + }, + { + "epoch": 1.8334619643636119, + "grad_norm": 1.0011298656463623, + "learning_rate": 1.8052455610326534e-06, + "loss": 0.4492, + "step": 27320 + }, + { + "epoch": 1.833596188047381, + "grad_norm": 1.1182427406311035, + "learning_rate": 1.8023526620902176e-06, + "loss": 0.4544, + "step": 27322 + }, + { + "epoch": 1.83373041173115, + "grad_norm": 1.1529039144515991, + "learning_rate": 1.7994620403918527e-06, + "loss": 0.4499, + "step": 27324 + }, + { + "epoch": 1.833864635414919, + "grad_norm": 1.0249526500701904, + "learning_rate": 1.7965736960741164e-06, + "loss": 0.4595, + "step": 27326 + }, + { + "epoch": 1.833998859098688, + "grad_norm": 1.0299060344696045, + "learning_rate": 1.7936876292734883e-06, + "loss": 0.4258, + "step": 27328 + }, + { + "epoch": 1.834133082782457, + "grad_norm": 1.0333058834075928, + "learning_rate": 1.790803840126326e-06, + "loss": 0.5368, + "step": 27330 + }, + { + "epoch": 1.8342673064662258, + "grad_norm": 0.9318089485168457, + "learning_rate": 1.7879223287688761e-06, + "loss": 0.4958, + "step": 27332 + }, + { + "epoch": 1.834401530149995, + "grad_norm": 0.9352691769599915, + "learning_rate": 1.7850430953372966e-06, + "loss": 0.4329, + "step": 27334 + }, + { + "epoch": 1.834535753833764, + "grad_norm": 0.9868640899658203, + "learning_rate": 1.7821661399676226e-06, + "loss": 0.4515, + "step": 27336 + }, + { + "epoch": 1.834669977517533, + "grad_norm": 1.107014536857605, + "learning_rate": 1.779291462795779e-06, + "loss": 0.4447, + "step": 27338 + }, + { + "epoch": 1.834804201201302, + "grad_norm": 1.046768069267273, + "learning_rate": 1.7764190639575907e-06, + "loss": 0.4571, + "step": 27340 + }, + { + "epoch": 1.8349384248850709, + "grad_norm": 1.2242913246154785, + "learning_rate": 1.7735489435887764e-06, + "loss": 0.4358, + "step": 27342 + }, + { + "epoch": 1.83507264856884, + "grad_norm": 1.113647699356079, + "learning_rate": 1.7706811018249447e-06, + "loss": 0.521, + "step": 27344 + }, + { + "epoch": 1.835206872252609, + "grad_norm": 1.1053078174591064, + "learning_rate": 1.7678155388015872e-06, + "loss": 0.4808, + "step": 27346 + }, + { + "epoch": 1.835341095936378, + "grad_norm": 1.020797610282898, + "learning_rate": 1.764952254654101e-06, + "loss": 0.4903, + "step": 27348 + }, + { + "epoch": 1.835475319620147, + "grad_norm": 1.1306965351104736, + "learning_rate": 1.7620912495177721e-06, + "loss": 0.419, + "step": 27350 + }, + { + "epoch": 1.835609543303916, + "grad_norm": 1.0897547006607056, + "learning_rate": 1.7592325235277762e-06, + "loss": 0.4437, + "step": 27352 + }, + { + "epoch": 1.8357437669876848, + "grad_norm": 1.2764606475830078, + "learning_rate": 1.7563760768191828e-06, + "loss": 0.5219, + "step": 27354 + }, + { + "epoch": 1.835877990671454, + "grad_norm": 0.944107711315155, + "learning_rate": 1.7535219095269451e-06, + "loss": 0.4352, + "step": 27356 + }, + { + "epoch": 1.836012214355223, + "grad_norm": 1.0962316989898682, + "learning_rate": 1.7506700217859329e-06, + "loss": 0.4248, + "step": 27358 + }, + { + "epoch": 1.836146438038992, + "grad_norm": 1.0919740200042725, + "learning_rate": 1.7478204137308773e-06, + "loss": 0.512, + "step": 27360 + }, + { + "epoch": 1.836280661722761, + "grad_norm": 1.1897571086883545, + "learning_rate": 1.744973085496432e-06, + "loss": 0.4814, + "step": 27362 + }, + { + "epoch": 1.8364148854065299, + "grad_norm": 1.0135506391525269, + "learning_rate": 1.7421280372171167e-06, + "loss": 0.4684, + "step": 27364 + }, + { + "epoch": 1.836549109090299, + "grad_norm": 1.0432535409927368, + "learning_rate": 1.739285269027352e-06, + "loss": 0.4677, + "step": 27366 + }, + { + "epoch": 1.8366833327740681, + "grad_norm": 1.1191753149032593, + "learning_rate": 1.736444781061458e-06, + "loss": 0.4813, + "step": 27368 + }, + { + "epoch": 1.836817556457837, + "grad_norm": 1.058532476425171, + "learning_rate": 1.7336065734536444e-06, + "loss": 0.4499, + "step": 27370 + }, + { + "epoch": 1.836951780141606, + "grad_norm": 1.0622410774230957, + "learning_rate": 1.7307706463380092e-06, + "loss": 0.451, + "step": 27372 + }, + { + "epoch": 1.837086003825375, + "grad_norm": 0.9336556196212769, + "learning_rate": 1.727936999848545e-06, + "loss": 0.4338, + "step": 27374 + }, + { + "epoch": 1.837220227509144, + "grad_norm": 0.8977116346359253, + "learning_rate": 1.7251056341191285e-06, + "loss": 0.4844, + "step": 27376 + }, + { + "epoch": 1.837354451192913, + "grad_norm": 1.1026514768600464, + "learning_rate": 1.7222765492835468e-06, + "loss": 0.5071, + "step": 27378 + }, + { + "epoch": 1.837488674876682, + "grad_norm": 1.0812137126922607, + "learning_rate": 1.7194497454754599e-06, + "loss": 0.4951, + "step": 27380 + }, + { + "epoch": 1.837622898560451, + "grad_norm": 0.9849303364753723, + "learning_rate": 1.716625222828444e-06, + "loss": 0.4815, + "step": 27382 + }, + { + "epoch": 1.83775712224422, + "grad_norm": 1.177372694015503, + "learning_rate": 1.7138029814759316e-06, + "loss": 0.4735, + "step": 27384 + }, + { + "epoch": 1.8378913459279889, + "grad_norm": 1.0768537521362305, + "learning_rate": 1.7109830215512767e-06, + "loss": 0.45, + "step": 27386 + }, + { + "epoch": 1.838025569611758, + "grad_norm": 1.1244202852249146, + "learning_rate": 1.7081653431877175e-06, + "loss": 0.4956, + "step": 27388 + }, + { + "epoch": 1.8381597932955271, + "grad_norm": 1.0745270252227783, + "learning_rate": 1.7053499465183864e-06, + "loss": 0.4951, + "step": 27390 + }, + { + "epoch": 1.838294016979296, + "grad_norm": 0.9530426859855652, + "learning_rate": 1.7025368316763047e-06, + "loss": 0.4391, + "step": 27392 + }, + { + "epoch": 1.838428240663065, + "grad_norm": 0.9638584852218628, + "learning_rate": 1.699725998794388e-06, + "loss": 0.4233, + "step": 27394 + }, + { + "epoch": 1.838562464346834, + "grad_norm": 0.9115172624588013, + "learning_rate": 1.6969174480054305e-06, + "loss": 0.4646, + "step": 27396 + }, + { + "epoch": 1.838696688030603, + "grad_norm": 1.073967456817627, + "learning_rate": 1.6941111794421482e-06, + "loss": 0.4027, + "step": 27398 + }, + { + "epoch": 1.838830911714372, + "grad_norm": 1.0610753297805786, + "learning_rate": 1.6913071932371184e-06, + "loss": 0.5404, + "step": 27400 + }, + { + "epoch": 1.838965135398141, + "grad_norm": 1.1293768882751465, + "learning_rate": 1.688505489522829e-06, + "loss": 0.4975, + "step": 27402 + }, + { + "epoch": 1.83909935908191, + "grad_norm": 1.069830060005188, + "learning_rate": 1.6857060684316584e-06, + "loss": 0.442, + "step": 27404 + }, + { + "epoch": 1.839233582765679, + "grad_norm": 1.0448309183120728, + "learning_rate": 1.682908930095861e-06, + "loss": 0.4767, + "step": 27406 + }, + { + "epoch": 1.8393678064494479, + "grad_norm": 1.0019116401672363, + "learning_rate": 1.6801140746476152e-06, + "loss": 0.4503, + "step": 27408 + }, + { + "epoch": 1.839502030133217, + "grad_norm": 1.1428416967391968, + "learning_rate": 1.6773215022189592e-06, + "loss": 0.4754, + "step": 27410 + }, + { + "epoch": 1.8396362538169861, + "grad_norm": 1.1250988245010376, + "learning_rate": 1.6745312129418434e-06, + "loss": 0.4478, + "step": 27412 + }, + { + "epoch": 1.839770477500755, + "grad_norm": 1.0673909187316895, + "learning_rate": 1.6717432069480897e-06, + "loss": 0.4561, + "step": 27414 + }, + { + "epoch": 1.839904701184524, + "grad_norm": 0.9419642686843872, + "learning_rate": 1.6689574843694433e-06, + "loss": 0.442, + "step": 27416 + }, + { + "epoch": 1.840038924868293, + "grad_norm": 1.28042733669281, + "learning_rate": 1.6661740453375096e-06, + "loss": 0.5354, + "step": 27418 + }, + { + "epoch": 1.840173148552062, + "grad_norm": 1.0464138984680176, + "learning_rate": 1.6633928899838224e-06, + "loss": 0.5032, + "step": 27420 + }, + { + "epoch": 1.840307372235831, + "grad_norm": 0.9780840873718262, + "learning_rate": 1.6606140184397546e-06, + "loss": 0.4566, + "step": 27422 + }, + { + "epoch": 1.8404415959196, + "grad_norm": 1.0669537782669067, + "learning_rate": 1.6578374308366285e-06, + "loss": 0.4608, + "step": 27424 + }, + { + "epoch": 1.840575819603369, + "grad_norm": 0.8034111261367798, + "learning_rate": 1.6550631273056171e-06, + "loss": 0.4243, + "step": 27426 + }, + { + "epoch": 1.840710043287138, + "grad_norm": 1.0807321071624756, + "learning_rate": 1.6522911079778158e-06, + "loss": 0.4601, + "step": 27428 + }, + { + "epoch": 1.8408442669709069, + "grad_norm": 0.9290803670883179, + "learning_rate": 1.6495213729841808e-06, + "loss": 0.443, + "step": 27430 + }, + { + "epoch": 1.840978490654676, + "grad_norm": 0.8990726470947266, + "learning_rate": 1.6467539224555906e-06, + "loss": 0.5122, + "step": 27432 + }, + { + "epoch": 1.8411127143384451, + "grad_norm": 1.0240440368652344, + "learning_rate": 1.643988756522785e-06, + "loss": 0.4162, + "step": 27434 + }, + { + "epoch": 1.841246938022214, + "grad_norm": 1.0276577472686768, + "learning_rate": 1.6412258753164322e-06, + "loss": 0.533, + "step": 27436 + }, + { + "epoch": 1.841381161705983, + "grad_norm": 0.9784525036811829, + "learning_rate": 1.6384652789670608e-06, + "loss": 0.5061, + "step": 27438 + }, + { + "epoch": 1.841515385389752, + "grad_norm": 0.9746937155723572, + "learning_rate": 1.6357069676051051e-06, + "loss": 0.4498, + "step": 27440 + }, + { + "epoch": 1.841649609073521, + "grad_norm": 1.075529932975769, + "learning_rate": 1.6329509413608834e-06, + "loss": 0.4486, + "step": 27442 + }, + { + "epoch": 1.8417838327572902, + "grad_norm": 1.12813401222229, + "learning_rate": 1.6301972003646304e-06, + "loss": 0.4737, + "step": 27444 + }, + { + "epoch": 1.841918056441059, + "grad_norm": 0.9420235753059387, + "learning_rate": 1.6274457447464363e-06, + "loss": 0.5351, + "step": 27446 + }, + { + "epoch": 1.842052280124828, + "grad_norm": 0.9531462788581848, + "learning_rate": 1.6246965746363197e-06, + "loss": 0.4584, + "step": 27448 + }, + { + "epoch": 1.842186503808597, + "grad_norm": 0.8998656272888184, + "learning_rate": 1.6219496901641596e-06, + "loss": 0.429, + "step": 27450 + }, + { + "epoch": 1.8423207274923659, + "grad_norm": 1.0480685234069824, + "learning_rate": 1.6192050914597467e-06, + "loss": 0.4507, + "step": 27452 + }, + { + "epoch": 1.842454951176135, + "grad_norm": 1.145783543586731, + "learning_rate": 1.6164627786527498e-06, + "loss": 0.5235, + "step": 27454 + }, + { + "epoch": 1.8425891748599041, + "grad_norm": 1.1258286237716675, + "learning_rate": 1.6137227518727482e-06, + "loss": 0.4591, + "step": 27456 + }, + { + "epoch": 1.842723398543673, + "grad_norm": 0.9835421442985535, + "learning_rate": 1.6109850112492054e-06, + "loss": 0.5163, + "step": 27458 + }, + { + "epoch": 1.842857622227442, + "grad_norm": 1.0142710208892822, + "learning_rate": 1.608249556911462e-06, + "loss": 0.4324, + "step": 27460 + }, + { + "epoch": 1.842991845911211, + "grad_norm": 1.022153377532959, + "learning_rate": 1.6055163889887647e-06, + "loss": 0.5142, + "step": 27462 + }, + { + "epoch": 1.84312606959498, + "grad_norm": 1.1606864929199219, + "learning_rate": 1.6027855076102605e-06, + "loss": 0.5038, + "step": 27464 + }, + { + "epoch": 1.8432602932787492, + "grad_norm": 0.8721981048583984, + "learning_rate": 1.6000569129049735e-06, + "loss": 0.4326, + "step": 27466 + }, + { + "epoch": 1.843394516962518, + "grad_norm": 1.065577507019043, + "learning_rate": 1.5973306050018178e-06, + "loss": 0.4931, + "step": 27468 + }, + { + "epoch": 1.843528740646287, + "grad_norm": 0.9497195482254028, + "learning_rate": 1.5946065840296177e-06, + "loss": 0.4524, + "step": 27470 + }, + { + "epoch": 1.843662964330056, + "grad_norm": 0.9861379265785217, + "learning_rate": 1.5918848501170647e-06, + "loss": 0.4944, + "step": 27472 + }, + { + "epoch": 1.843797188013825, + "grad_norm": 0.9308890104293823, + "learning_rate": 1.5891654033927617e-06, + "loss": 0.4061, + "step": 27474 + }, + { + "epoch": 1.843931411697594, + "grad_norm": 1.0119004249572754, + "learning_rate": 1.5864482439852058e-06, + "loss": 0.4442, + "step": 27476 + }, + { + "epoch": 1.8440656353813631, + "grad_norm": 1.034685492515564, + "learning_rate": 1.583733372022761e-06, + "loss": 0.4174, + "step": 27478 + }, + { + "epoch": 1.844199859065132, + "grad_norm": 0.9737145900726318, + "learning_rate": 1.581020787633708e-06, + "loss": 0.4782, + "step": 27480 + }, + { + "epoch": 1.844334082748901, + "grad_norm": 0.9639759063720703, + "learning_rate": 1.5783104909462166e-06, + "loss": 0.3882, + "step": 27482 + }, + { + "epoch": 1.84446830643267, + "grad_norm": 1.7538946866989136, + "learning_rate": 1.5756024820883287e-06, + "loss": 0.4969, + "step": 27484 + }, + { + "epoch": 1.844602530116439, + "grad_norm": 1.1073065996170044, + "learning_rate": 1.5728967611880085e-06, + "loss": 0.48, + "step": 27486 + }, + { + "epoch": 1.8447367538002082, + "grad_norm": 0.9614431262016296, + "learning_rate": 1.5701933283730875e-06, + "loss": 0.453, + "step": 27488 + }, + { + "epoch": 1.844870977483977, + "grad_norm": 0.9399886727333069, + "learning_rate": 1.567492183771302e-06, + "loss": 0.5411, + "step": 27490 + }, + { + "epoch": 1.845005201167746, + "grad_norm": 1.0651518106460571, + "learning_rate": 1.564793327510261e-06, + "loss": 0.464, + "step": 27492 + }, + { + "epoch": 1.845139424851515, + "grad_norm": 0.9036216735839844, + "learning_rate": 1.5620967597175017e-06, + "loss": 0.4539, + "step": 27494 + }, + { + "epoch": 1.845273648535284, + "grad_norm": 1.1416680812835693, + "learning_rate": 1.559402480520422e-06, + "loss": 0.5153, + "step": 27496 + }, + { + "epoch": 1.845407872219053, + "grad_norm": 1.1369332075119019, + "learning_rate": 1.5567104900463147e-06, + "loss": 0.4874, + "step": 27498 + }, + { + "epoch": 1.8455420959028221, + "grad_norm": 0.9781507849693298, + "learning_rate": 1.554020788422378e-06, + "loss": 0.4902, + "step": 27500 + }, + { + "epoch": 1.845676319586591, + "grad_norm": 1.0934257507324219, + "learning_rate": 1.5513333757756987e-06, + "loss": 0.4657, + "step": 27502 + }, + { + "epoch": 1.84581054327036, + "grad_norm": 1.0390064716339111, + "learning_rate": 1.5486482522332424e-06, + "loss": 0.4627, + "step": 27504 + }, + { + "epoch": 1.845944766954129, + "grad_norm": 1.1115378141403198, + "learning_rate": 1.5459654179218852e-06, + "loss": 0.4134, + "step": 27506 + }, + { + "epoch": 1.846078990637898, + "grad_norm": 1.0079962015151978, + "learning_rate": 1.5432848729683758e-06, + "loss": 0.4602, + "step": 27508 + }, + { + "epoch": 1.8462132143216672, + "grad_norm": 0.9150170683860779, + "learning_rate": 1.5406066174993739e-06, + "loss": 0.4671, + "step": 27510 + }, + { + "epoch": 1.846347438005436, + "grad_norm": 1.2085232734680176, + "learning_rate": 1.5379306516414115e-06, + "loss": 0.5562, + "step": 27512 + }, + { + "epoch": 1.846481661689205, + "grad_norm": 1.037790298461914, + "learning_rate": 1.5352569755209378e-06, + "loss": 0.479, + "step": 27514 + }, + { + "epoch": 1.846615885372974, + "grad_norm": 0.9107427000999451, + "learning_rate": 1.5325855892642682e-06, + "loss": 0.4289, + "step": 27516 + }, + { + "epoch": 1.846750109056743, + "grad_norm": 1.0650413036346436, + "learning_rate": 1.529916492997624e-06, + "loss": 0.5249, + "step": 27518 + }, + { + "epoch": 1.8468843327405122, + "grad_norm": 1.061478853225708, + "learning_rate": 1.5272496868471043e-06, + "loss": 0.5132, + "step": 27520 + }, + { + "epoch": 1.8470185564242811, + "grad_norm": 0.9944947361946106, + "learning_rate": 1.5245851709387249e-06, + "loss": 0.4591, + "step": 27522 + }, + { + "epoch": 1.84715278010805, + "grad_norm": 0.9647601842880249, + "learning_rate": 1.5219229453983796e-06, + "loss": 0.4661, + "step": 27524 + }, + { + "epoch": 1.847287003791819, + "grad_norm": 0.9864873886108398, + "learning_rate": 1.5192630103518402e-06, + "loss": 0.4881, + "step": 27526 + }, + { + "epoch": 1.847421227475588, + "grad_norm": 1.0107039213180542, + "learning_rate": 1.5166053659247892e-06, + "loss": 0.4718, + "step": 27528 + }, + { + "epoch": 1.847555451159357, + "grad_norm": 1.0673303604125977, + "learning_rate": 1.5139500122427986e-06, + "loss": 0.4393, + "step": 27530 + }, + { + "epoch": 1.8476896748431262, + "grad_norm": 0.9937414526939392, + "learning_rate": 1.511296949431329e-06, + "loss": 0.4249, + "step": 27532 + }, + { + "epoch": 1.847823898526895, + "grad_norm": 1.0674530267715454, + "learning_rate": 1.5086461776157245e-06, + "loss": 0.5207, + "step": 27534 + }, + { + "epoch": 1.847958122210664, + "grad_norm": 0.8959336280822754, + "learning_rate": 1.5059976969212408e-06, + "loss": 0.4826, + "step": 27536 + }, + { + "epoch": 1.848092345894433, + "grad_norm": 0.9730433225631714, + "learning_rate": 1.5033515074729942e-06, + "loss": 0.4278, + "step": 27538 + }, + { + "epoch": 1.848226569578202, + "grad_norm": 1.0837477445602417, + "learning_rate": 1.5007076093960348e-06, + "loss": 0.4376, + "step": 27540 + }, + { + "epoch": 1.8483607932619712, + "grad_norm": 1.0965659618377686, + "learning_rate": 1.4980660028152627e-06, + "loss": 0.4846, + "step": 27542 + }, + { + "epoch": 1.8484950169457401, + "grad_norm": 0.9107245802879333, + "learning_rate": 1.4954266878555112e-06, + "loss": 0.4312, + "step": 27544 + }, + { + "epoch": 1.848629240629509, + "grad_norm": 1.0831120014190674, + "learning_rate": 1.4927896646414529e-06, + "loss": 0.4578, + "step": 27546 + }, + { + "epoch": 1.848763464313278, + "grad_norm": 1.0490244626998901, + "learning_rate": 1.4901549332977105e-06, + "loss": 0.4303, + "step": 27548 + }, + { + "epoch": 1.8488976879970471, + "grad_norm": 1.0324196815490723, + "learning_rate": 1.4875224939487454e-06, + "loss": 0.3801, + "step": 27550 + }, + { + "epoch": 1.849031911680816, + "grad_norm": 1.0803989171981812, + "learning_rate": 1.4848923467189524e-06, + "loss": 0.4548, + "step": 27552 + }, + { + "epoch": 1.8491661353645852, + "grad_norm": 1.1710008382797241, + "learning_rate": 1.482264491732599e-06, + "loss": 0.4808, + "step": 27554 + }, + { + "epoch": 1.849300359048354, + "grad_norm": 1.099749207496643, + "learning_rate": 1.4796389291138358e-06, + "loss": 0.4845, + "step": 27556 + }, + { + "epoch": 1.849434582732123, + "grad_norm": 0.9039605259895325, + "learning_rate": 1.4770156589867246e-06, + "loss": 0.4377, + "step": 27558 + }, + { + "epoch": 1.849568806415892, + "grad_norm": 1.1397470235824585, + "learning_rate": 1.4743946814752053e-06, + "loss": 0.5262, + "step": 27560 + }, + { + "epoch": 1.849703030099661, + "grad_norm": 0.8652995228767395, + "learning_rate": 1.471775996703123e-06, + "loss": 0.4527, + "step": 27562 + }, + { + "epoch": 1.8498372537834302, + "grad_norm": 1.0195953845977783, + "learning_rate": 1.4691596047941902e-06, + "loss": 0.4406, + "step": 27564 + }, + { + "epoch": 1.8499714774671991, + "grad_norm": 1.1417049169540405, + "learning_rate": 1.4665455058720357e-06, + "loss": 0.4805, + "step": 27566 + }, + { + "epoch": 1.850105701150968, + "grad_norm": 0.9819300770759583, + "learning_rate": 1.4639337000601717e-06, + "loss": 0.4426, + "step": 27568 + }, + { + "epoch": 1.850239924834737, + "grad_norm": 1.0256808996200562, + "learning_rate": 1.4613241874819938e-06, + "loss": 0.493, + "step": 27570 + }, + { + "epoch": 1.8503741485185061, + "grad_norm": 1.0297558307647705, + "learning_rate": 1.4587169682608093e-06, + "loss": 0.5481, + "step": 27572 + }, + { + "epoch": 1.850508372202275, + "grad_norm": 1.0053192377090454, + "learning_rate": 1.456112042519786e-06, + "loss": 0.4182, + "step": 27574 + }, + { + "epoch": 1.8506425958860442, + "grad_norm": 1.1509085893630981, + "learning_rate": 1.45350941038202e-06, + "loss": 0.4912, + "step": 27576 + }, + { + "epoch": 1.850776819569813, + "grad_norm": 1.0100876092910767, + "learning_rate": 1.4509090719704631e-06, + "loss": 0.4583, + "step": 27578 + }, + { + "epoch": 1.850911043253582, + "grad_norm": 0.8980428576469421, + "learning_rate": 1.4483110274079891e-06, + "loss": 0.415, + "step": 27580 + }, + { + "epoch": 1.851045266937351, + "grad_norm": 1.1096354722976685, + "learning_rate": 1.4457152768173444e-06, + "loss": 0.4795, + "step": 27582 + }, + { + "epoch": 1.85117949062112, + "grad_norm": 1.0847545862197876, + "learning_rate": 1.4431218203211804e-06, + "loss": 0.4038, + "step": 27584 + }, + { + "epoch": 1.8513137143048892, + "grad_norm": 1.030122995376587, + "learning_rate": 1.440530658042022e-06, + "loss": 0.4204, + "step": 27586 + }, + { + "epoch": 1.8514479379886581, + "grad_norm": 1.0669091939926147, + "learning_rate": 1.437941790102304e-06, + "loss": 0.483, + "step": 27588 + }, + { + "epoch": 1.851582161672427, + "grad_norm": 1.0459216833114624, + "learning_rate": 1.4353552166243457e-06, + "loss": 0.4687, + "step": 27590 + }, + { + "epoch": 1.851716385356196, + "grad_norm": 1.1613476276397705, + "learning_rate": 1.4327709377303544e-06, + "loss": 0.4149, + "step": 27592 + }, + { + "epoch": 1.8518506090399651, + "grad_norm": 1.0539007186889648, + "learning_rate": 1.4301889535424218e-06, + "loss": 0.4758, + "step": 27594 + }, + { + "epoch": 1.8519848327237343, + "grad_norm": 1.0329196453094482, + "learning_rate": 1.4276092641825668e-06, + "loss": 0.4155, + "step": 27596 + }, + { + "epoch": 1.8521190564075032, + "grad_norm": 1.1315053701400757, + "learning_rate": 1.4250318697726529e-06, + "loss": 0.4739, + "step": 27598 + }, + { + "epoch": 1.852253280091272, + "grad_norm": 1.0232601165771484, + "learning_rate": 1.4224567704344716e-06, + "loss": 0.4584, + "step": 27600 + }, + { + "epoch": 1.852387503775041, + "grad_norm": 1.107422947883606, + "learning_rate": 1.4198839662896758e-06, + "loss": 0.5413, + "step": 27602 + }, + { + "epoch": 1.85252172745881, + "grad_norm": 1.064710021018982, + "learning_rate": 1.4173134574598402e-06, + "loss": 0.4463, + "step": 27604 + }, + { + "epoch": 1.852655951142579, + "grad_norm": 1.0158103704452515, + "learning_rate": 1.4147452440664065e-06, + "loss": 0.4464, + "step": 27606 + }, + { + "epoch": 1.8527901748263482, + "grad_norm": 1.0038433074951172, + "learning_rate": 1.4121793262307226e-06, + "loss": 0.5025, + "step": 27608 + }, + { + "epoch": 1.8529243985101171, + "grad_norm": 1.045138955116272, + "learning_rate": 1.4096157040740244e-06, + "loss": 0.4454, + "step": 27610 + }, + { + "epoch": 1.853058622193886, + "grad_norm": 1.006141185760498, + "learning_rate": 1.4070543777174317e-06, + "loss": 0.4525, + "step": 27612 + }, + { + "epoch": 1.853192845877655, + "grad_norm": 1.0381324291229248, + "learning_rate": 1.4044953472819644e-06, + "loss": 0.4833, + "step": 27614 + }, + { + "epoch": 1.8533270695614241, + "grad_norm": 1.0100772380828857, + "learning_rate": 1.4019386128885314e-06, + "loss": 0.4649, + "step": 27616 + }, + { + "epoch": 1.8534612932451933, + "grad_norm": 1.0271670818328857, + "learning_rate": 1.3993841746579416e-06, + "loss": 0.4925, + "step": 27618 + }, + { + "epoch": 1.8535955169289622, + "grad_norm": 1.0052720308303833, + "learning_rate": 1.396832032710882e-06, + "loss": 0.4151, + "step": 27620 + }, + { + "epoch": 1.853729740612731, + "grad_norm": 0.9382982850074768, + "learning_rate": 1.3942821871679279e-06, + "loss": 0.5207, + "step": 27622 + }, + { + "epoch": 1.8538639642965, + "grad_norm": 1.1375670433044434, + "learning_rate": 1.391734638149561e-06, + "loss": 0.4553, + "step": 27624 + }, + { + "epoch": 1.8539981879802692, + "grad_norm": 1.0256612300872803, + "learning_rate": 1.3891893857761519e-06, + "loss": 0.4518, + "step": 27626 + }, + { + "epoch": 1.854132411664038, + "grad_norm": 1.0710864067077637, + "learning_rate": 1.3866464301679593e-06, + "loss": 0.5361, + "step": 27628 + }, + { + "epoch": 1.8542666353478072, + "grad_norm": 1.1199383735656738, + "learning_rate": 1.3841057714451212e-06, + "loss": 0.5286, + "step": 27630 + }, + { + "epoch": 1.8544008590315761, + "grad_norm": 1.0359209775924683, + "learning_rate": 1.3815674097276854e-06, + "loss": 0.4916, + "step": 27632 + }, + { + "epoch": 1.854535082715345, + "grad_norm": 0.8993666768074036, + "learning_rate": 1.379031345135595e-06, + "loss": 0.5123, + "step": 27634 + }, + { + "epoch": 1.854669306399114, + "grad_norm": 1.0495448112487793, + "learning_rate": 1.3764975777886547e-06, + "loss": 0.4336, + "step": 27636 + }, + { + "epoch": 1.8548035300828831, + "grad_norm": 0.7166053056716919, + "learning_rate": 1.3739661078065957e-06, + "loss": 0.4121, + "step": 27638 + }, + { + "epoch": 1.8549377537666523, + "grad_norm": 1.0230076313018799, + "learning_rate": 1.3714369353090173e-06, + "loss": 0.4041, + "step": 27640 + }, + { + "epoch": 1.8550719774504212, + "grad_norm": 1.0174587965011597, + "learning_rate": 1.3689100604154182e-06, + "loss": 0.3957, + "step": 27642 + }, + { + "epoch": 1.85520620113419, + "grad_norm": 1.0715550184249878, + "learning_rate": 1.3663854832451916e-06, + "loss": 0.4528, + "step": 27644 + }, + { + "epoch": 1.855340424817959, + "grad_norm": 0.9875494241714478, + "learning_rate": 1.3638632039176147e-06, + "loss": 0.5179, + "step": 27646 + }, + { + "epoch": 1.8554746485017282, + "grad_norm": 0.9951808452606201, + "learning_rate": 1.3613432225518696e-06, + "loss": 0.4658, + "step": 27648 + }, + { + "epoch": 1.855608872185497, + "grad_norm": 0.9630551338195801, + "learning_rate": 1.3588255392670058e-06, + "loss": 0.5425, + "step": 27650 + }, + { + "epoch": 1.8557430958692662, + "grad_norm": 1.0404685735702515, + "learning_rate": 1.3563101541819889e-06, + "loss": 0.4919, + "step": 27652 + }, + { + "epoch": 1.8558773195530351, + "grad_norm": 1.045067548751831, + "learning_rate": 1.3537970674156631e-06, + "loss": 0.5117, + "step": 27654 + }, + { + "epoch": 1.856011543236804, + "grad_norm": 1.054994821548462, + "learning_rate": 1.351286279086772e-06, + "loss": 0.5023, + "step": 27656 + }, + { + "epoch": 1.856145766920573, + "grad_norm": 1.0288803577423096, + "learning_rate": 1.3487777893139374e-06, + "loss": 0.5293, + "step": 27658 + }, + { + "epoch": 1.8562799906043421, + "grad_norm": 0.9933587908744812, + "learning_rate": 1.346271598215676e-06, + "loss": 0.5119, + "step": 27660 + }, + { + "epoch": 1.8564142142881113, + "grad_norm": 0.9720452427864075, + "learning_rate": 1.3437677059104147e-06, + "loss": 0.4397, + "step": 27662 + }, + { + "epoch": 1.8565484379718802, + "grad_norm": 1.1074039936065674, + "learning_rate": 1.3412661125164483e-06, + "loss": 0.5094, + "step": 27664 + }, + { + "epoch": 1.856682661655649, + "grad_norm": 1.0138037204742432, + "learning_rate": 1.338766818151982e-06, + "loss": 0.4095, + "step": 27666 + }, + { + "epoch": 1.856816885339418, + "grad_norm": 1.1227648258209229, + "learning_rate": 1.3362698229350935e-06, + "loss": 0.4334, + "step": 27668 + }, + { + "epoch": 1.8569511090231872, + "grad_norm": 1.0338616371154785, + "learning_rate": 1.3337751269837606e-06, + "loss": 0.5013, + "step": 27670 + }, + { + "epoch": 1.8570853327069563, + "grad_norm": 1.1134554147720337, + "learning_rate": 1.331282730415856e-06, + "loss": 0.4636, + "step": 27672 + }, + { + "epoch": 1.8572195563907252, + "grad_norm": 1.0271615982055664, + "learning_rate": 1.3287926333491353e-06, + "loss": 0.4261, + "step": 27674 + }, + { + "epoch": 1.8573537800744941, + "grad_norm": 0.9932689666748047, + "learning_rate": 1.3263048359012543e-06, + "loss": 0.501, + "step": 27676 + }, + { + "epoch": 1.857488003758263, + "grad_norm": 1.0491055250167847, + "learning_rate": 1.3238193381897635e-06, + "loss": 0.4436, + "step": 27678 + }, + { + "epoch": 1.857622227442032, + "grad_norm": 0.8929378390312195, + "learning_rate": 1.321336140332091e-06, + "loss": 0.4561, + "step": 27680 + }, + { + "epoch": 1.8577564511258011, + "grad_norm": 2.278045177459717, + "learning_rate": 1.3188552424455546e-06, + "loss": 0.5313, + "step": 27682 + }, + { + "epoch": 1.8578906748095703, + "grad_norm": 0.8573141694068909, + "learning_rate": 1.316376644647388e-06, + "loss": 0.4151, + "step": 27684 + }, + { + "epoch": 1.8580248984933392, + "grad_norm": 0.8995969891548157, + "learning_rate": 1.3139003470546918e-06, + "loss": 0.392, + "step": 27686 + }, + { + "epoch": 1.858159122177108, + "grad_norm": 1.1379754543304443, + "learning_rate": 1.3114263497844669e-06, + "loss": 0.406, + "step": 27688 + }, + { + "epoch": 1.858293345860877, + "grad_norm": 1.0205707550048828, + "learning_rate": 1.3089546529536034e-06, + "loss": 0.4731, + "step": 27690 + }, + { + "epoch": 1.8584275695446462, + "grad_norm": 1.0011777877807617, + "learning_rate": 1.306485256678891e-06, + "loss": 0.4883, + "step": 27692 + }, + { + "epoch": 1.8585617932284153, + "grad_norm": 1.0003676414489746, + "learning_rate": 1.3040181610769865e-06, + "loss": 0.4204, + "step": 27694 + }, + { + "epoch": 1.8586960169121842, + "grad_norm": 1.0474743843078613, + "learning_rate": 1.3015533662644852e-06, + "loss": 0.4178, + "step": 27696 + }, + { + "epoch": 1.8588302405959531, + "grad_norm": 1.0012686252593994, + "learning_rate": 1.299090872357811e-06, + "loss": 0.4525, + "step": 27698 + }, + { + "epoch": 1.858964464279722, + "grad_norm": 0.9653547406196594, + "learning_rate": 1.2966306794733318e-06, + "loss": 0.4891, + "step": 27700 + }, + { + "epoch": 1.8590986879634912, + "grad_norm": 0.8683306574821472, + "learning_rate": 1.2941727877272825e-06, + "loss": 0.4006, + "step": 27702 + }, + { + "epoch": 1.8592329116472601, + "grad_norm": 1.1569554805755615, + "learning_rate": 1.2917171972357922e-06, + "loss": 0.5005, + "step": 27704 + }, + { + "epoch": 1.8593671353310293, + "grad_norm": 1.2232944965362549, + "learning_rate": 1.289263908114885e-06, + "loss": 0.4743, + "step": 27706 + }, + { + "epoch": 1.8595013590147982, + "grad_norm": 0.9982020258903503, + "learning_rate": 1.2868129204804735e-06, + "loss": 0.4907, + "step": 27708 + }, + { + "epoch": 1.859635582698567, + "grad_norm": 1.1773412227630615, + "learning_rate": 1.2843642344483542e-06, + "loss": 0.4654, + "step": 27710 + }, + { + "epoch": 1.859769806382336, + "grad_norm": 1.1091125011444092, + "learning_rate": 1.2819178501342343e-06, + "loss": 0.4409, + "step": 27712 + }, + { + "epoch": 1.8599040300661052, + "grad_norm": 1.0534645318984985, + "learning_rate": 1.2794737676536994e-06, + "loss": 0.5205, + "step": 27714 + }, + { + "epoch": 1.8600382537498743, + "grad_norm": 0.9437594413757324, + "learning_rate": 1.2770319871222236e-06, + "loss": 0.4711, + "step": 27716 + }, + { + "epoch": 1.8601724774336432, + "grad_norm": 0.958147406578064, + "learning_rate": 1.2745925086551702e-06, + "loss": 0.4477, + "step": 27718 + }, + { + "epoch": 1.8603067011174121, + "grad_norm": 1.0931216478347778, + "learning_rate": 1.2721553323678137e-06, + "loss": 0.4576, + "step": 27720 + }, + { + "epoch": 1.860440924801181, + "grad_norm": 1.0033011436462402, + "learning_rate": 1.2697204583752898e-06, + "loss": 0.4928, + "step": 27722 + }, + { + "epoch": 1.8605751484849502, + "grad_norm": 1.105586290359497, + "learning_rate": 1.267287886792662e-06, + "loss": 0.4718, + "step": 27724 + }, + { + "epoch": 1.8607093721687191, + "grad_norm": 1.0070650577545166, + "learning_rate": 1.2648576177348437e-06, + "loss": 0.4843, + "step": 27726 + }, + { + "epoch": 1.8608435958524883, + "grad_norm": 1.1382700204849243, + "learning_rate": 1.2624296513166712e-06, + "loss": 0.4592, + "step": 27728 + }, + { + "epoch": 1.8609778195362572, + "grad_norm": 1.1092818975448608, + "learning_rate": 1.260003987652858e-06, + "loss": 0.43, + "step": 27730 + }, + { + "epoch": 1.861112043220026, + "grad_norm": 1.1581449508666992, + "learning_rate": 1.2575806268580182e-06, + "loss": 0.465, + "step": 27732 + }, + { + "epoch": 1.861246266903795, + "grad_norm": 0.9862481355667114, + "learning_rate": 1.2551595690466434e-06, + "loss": 0.4387, + "step": 27734 + }, + { + "epoch": 1.8613804905875642, + "grad_norm": 1.0038784742355347, + "learning_rate": 1.252740814333131e-06, + "loss": 0.4707, + "step": 27736 + }, + { + "epoch": 1.8615147142713333, + "grad_norm": 1.1910805702209473, + "learning_rate": 1.250324362831745e-06, + "loss": 0.5406, + "step": 27738 + }, + { + "epoch": 1.8616489379551022, + "grad_norm": 1.095921277999878, + "learning_rate": 1.2479102146566834e-06, + "loss": 0.4791, + "step": 27740 + }, + { + "epoch": 1.8617831616388711, + "grad_norm": 1.1198900938034058, + "learning_rate": 1.2454983699219936e-06, + "loss": 0.4982, + "step": 27742 + }, + { + "epoch": 1.86191738532264, + "grad_norm": 0.9647154211997986, + "learning_rate": 1.2430888287416342e-06, + "loss": 0.469, + "step": 27744 + }, + { + "epoch": 1.8620516090064092, + "grad_norm": 0.9845030903816223, + "learning_rate": 1.2406815912294535e-06, + "loss": 0.399, + "step": 27746 + }, + { + "epoch": 1.8621858326901783, + "grad_norm": 1.0828144550323486, + "learning_rate": 1.2382766574991766e-06, + "loss": 0.4334, + "step": 27748 + }, + { + "epoch": 1.8623200563739473, + "grad_norm": 1.013677716255188, + "learning_rate": 1.235874027664452e-06, + "loss": 0.4792, + "step": 27750 + }, + { + "epoch": 1.8624542800577162, + "grad_norm": 0.9506924748420715, + "learning_rate": 1.2334737018387887e-06, + "loss": 0.4674, + "step": 27752 + }, + { + "epoch": 1.862588503741485, + "grad_norm": 1.0657380819320679, + "learning_rate": 1.231075680135596e-06, + "loss": 0.457, + "step": 27754 + }, + { + "epoch": 1.862722727425254, + "grad_norm": 0.9487188458442688, + "learning_rate": 1.2286799626681721e-06, + "loss": 0.463, + "step": 27756 + }, + { + "epoch": 1.8628569511090232, + "grad_norm": 1.0851175785064697, + "learning_rate": 1.226286549549721e-06, + "loss": 0.4614, + "step": 27758 + }, + { + "epoch": 1.8629911747927923, + "grad_norm": 0.949742317199707, + "learning_rate": 1.2238954408933134e-06, + "loss": 0.4029, + "step": 27760 + }, + { + "epoch": 1.8631253984765612, + "grad_norm": 0.99049311876297, + "learning_rate": 1.2215066368119476e-06, + "loss": 0.4526, + "step": 27762 + }, + { + "epoch": 1.8632596221603301, + "grad_norm": 0.9609589576721191, + "learning_rate": 1.2191201374184614e-06, + "loss": 0.4424, + "step": 27764 + }, + { + "epoch": 1.863393845844099, + "grad_norm": 1.0745807886123657, + "learning_rate": 1.2167359428256253e-06, + "loss": 0.4545, + "step": 27766 + }, + { + "epoch": 1.8635280695278682, + "grad_norm": 1.0423164367675781, + "learning_rate": 1.2143540531460885e-06, + "loss": 0.4838, + "step": 27768 + }, + { + "epoch": 1.8636622932116373, + "grad_norm": 1.070115327835083, + "learning_rate": 1.211974468492394e-06, + "loss": 0.4114, + "step": 27770 + }, + { + "epoch": 1.8637965168954063, + "grad_norm": 1.0345423221588135, + "learning_rate": 1.2095971889769686e-06, + "loss": 0.4636, + "step": 27772 + }, + { + "epoch": 1.8639307405791752, + "grad_norm": 1.089133381843567, + "learning_rate": 1.2072222147121338e-06, + "loss": 0.491, + "step": 27774 + }, + { + "epoch": 1.864064964262944, + "grad_norm": 1.0854707956314087, + "learning_rate": 1.2048495458100995e-06, + "loss": 0.4679, + "step": 27776 + }, + { + "epoch": 1.8641991879467132, + "grad_norm": 0.9781652688980103, + "learning_rate": 1.2024791823829762e-06, + "loss": 0.387, + "step": 27778 + }, + { + "epoch": 1.8643334116304822, + "grad_norm": 1.0830963850021362, + "learning_rate": 1.200111124542752e-06, + "loss": 0.5194, + "step": 27780 + }, + { + "epoch": 1.8644676353142513, + "grad_norm": 1.0142366886138916, + "learning_rate": 1.1977453724013154e-06, + "loss": 0.4902, + "step": 27782 + }, + { + "epoch": 1.8646018589980202, + "grad_norm": 1.1209266185760498, + "learning_rate": 1.1953819260704436e-06, + "loss": 0.5397, + "step": 27784 + }, + { + "epoch": 1.8647360826817891, + "grad_norm": 1.0151091814041138, + "learning_rate": 1.1930207856618137e-06, + "loss": 0.4546, + "step": 27786 + }, + { + "epoch": 1.864870306365558, + "grad_norm": 1.0850834846496582, + "learning_rate": 1.1906619512869644e-06, + "loss": 0.4781, + "step": 27788 + }, + { + "epoch": 1.8650045300493272, + "grad_norm": 1.0318243503570557, + "learning_rate": 1.1883054230573731e-06, + "loss": 0.4647, + "step": 27790 + }, + { + "epoch": 1.8651387537330963, + "grad_norm": 0.9039753079414368, + "learning_rate": 1.1859512010843565e-06, + "loss": 0.4377, + "step": 27792 + }, + { + "epoch": 1.8652729774168653, + "grad_norm": 1.061113953590393, + "learning_rate": 1.183599285479159e-06, + "loss": 0.4675, + "step": 27794 + }, + { + "epoch": 1.8654072011006342, + "grad_norm": 1.1509170532226562, + "learning_rate": 1.1812496763528968e-06, + "loss": 0.5651, + "step": 27796 + }, + { + "epoch": 1.865541424784403, + "grad_norm": 1.0290863513946533, + "learning_rate": 1.1789023738165983e-06, + "loss": 0.4957, + "step": 27798 + }, + { + "epoch": 1.8656756484681722, + "grad_norm": 0.8547778725624084, + "learning_rate": 1.1765573779811578e-06, + "loss": 0.402, + "step": 27800 + }, + { + "epoch": 1.8658098721519412, + "grad_norm": 1.064781904220581, + "learning_rate": 1.1742146889573758e-06, + "loss": 0.4567, + "step": 27802 + }, + { + "epoch": 1.8659440958357103, + "grad_norm": 1.036177635192871, + "learning_rate": 1.17187430685593e-06, + "loss": 0.4933, + "step": 27804 + }, + { + "epoch": 1.8660783195194792, + "grad_norm": 1.0641292333602905, + "learning_rate": 1.1695362317874158e-06, + "loss": 0.512, + "step": 27806 + }, + { + "epoch": 1.8662125432032481, + "grad_norm": 1.0304546356201172, + "learning_rate": 1.1672004638622892e-06, + "loss": 0.4732, + "step": 27808 + }, + { + "epoch": 1.866346766887017, + "grad_norm": 1.0697804689407349, + "learning_rate": 1.164867003190917e-06, + "loss": 0.4858, + "step": 27810 + }, + { + "epoch": 1.8664809905707862, + "grad_norm": 1.2139830589294434, + "learning_rate": 1.1625358498835505e-06, + "loss": 0.5135, + "step": 27812 + }, + { + "epoch": 1.8666152142545553, + "grad_norm": 1.0166290998458862, + "learning_rate": 1.1602070040503232e-06, + "loss": 0.4203, + "step": 27814 + }, + { + "epoch": 1.8667494379383243, + "grad_norm": 0.9351518750190735, + "learning_rate": 1.157880465801281e-06, + "loss": 0.4898, + "step": 27816 + }, + { + "epoch": 1.8668836616220932, + "grad_norm": 0.9220516681671143, + "learning_rate": 1.155556235246341e-06, + "loss": 0.3922, + "step": 27818 + }, + { + "epoch": 1.867017885305862, + "grad_norm": 0.9939961433410645, + "learning_rate": 1.1532343124953216e-06, + "loss": 0.4737, + "step": 27820 + }, + { + "epoch": 1.8671521089896312, + "grad_norm": 1.5443549156188965, + "learning_rate": 1.1509146976579232e-06, + "loss": 0.4162, + "step": 27822 + }, + { + "epoch": 1.8672863326734004, + "grad_norm": 1.0167597532272339, + "learning_rate": 1.148597390843753e-06, + "loss": 0.3998, + "step": 27824 + }, + { + "epoch": 1.8674205563571693, + "grad_norm": 1.2478851079940796, + "learning_rate": 1.14628239216229e-06, + "loss": 0.5228, + "step": 27826 + }, + { + "epoch": 1.8675547800409382, + "grad_norm": 0.9998477101325989, + "learning_rate": 1.143969701722919e-06, + "loss": 0.4396, + "step": 27828 + }, + { + "epoch": 1.8676890037247071, + "grad_norm": 1.055120587348938, + "learning_rate": 1.1416593196349134e-06, + "loss": 0.4476, + "step": 27830 + }, + { + "epoch": 1.867823227408476, + "grad_norm": 1.0692424774169922, + "learning_rate": 1.139351246007425e-06, + "loss": 0.4335, + "step": 27832 + }, + { + "epoch": 1.8679574510922452, + "grad_norm": 0.9727033376693726, + "learning_rate": 1.1370454809495056e-06, + "loss": 0.4036, + "step": 27834 + }, + { + "epoch": 1.8680916747760143, + "grad_norm": 0.8969436287879944, + "learning_rate": 1.1347420245701068e-06, + "loss": 0.45, + "step": 27836 + }, + { + "epoch": 1.8682258984597833, + "grad_norm": 0.9650223255157471, + "learning_rate": 1.1324408769780636e-06, + "loss": 0.4541, + "step": 27838 + }, + { + "epoch": 1.8683601221435522, + "grad_norm": 1.0078452825546265, + "learning_rate": 1.1301420382820893e-06, + "loss": 0.5195, + "step": 27840 + }, + { + "epoch": 1.868494345827321, + "grad_norm": 1.0130294561386108, + "learning_rate": 1.1278455085908025e-06, + "loss": 0.4784, + "step": 27842 + }, + { + "epoch": 1.8686285695110902, + "grad_norm": 1.0932484865188599, + "learning_rate": 1.125551288012716e-06, + "loss": 0.4415, + "step": 27844 + }, + { + "epoch": 1.8687627931948594, + "grad_norm": 1.096427321434021, + "learning_rate": 1.123259376656216e-06, + "loss": 0.4845, + "step": 27846 + }, + { + "epoch": 1.8688970168786283, + "grad_norm": 0.9929922223091125, + "learning_rate": 1.12096977462961e-06, + "loss": 0.4352, + "step": 27848 + }, + { + "epoch": 1.8690312405623972, + "grad_norm": 1.0440667867660522, + "learning_rate": 1.1186824820410614e-06, + "loss": 0.4269, + "step": 27850 + }, + { + "epoch": 1.8691654642461661, + "grad_norm": 0.9539794325828552, + "learning_rate": 1.1163974989986447e-06, + "loss": 0.449, + "step": 27852 + }, + { + "epoch": 1.8692996879299353, + "grad_norm": 1.1850839853286743, + "learning_rate": 1.1141148256103128e-06, + "loss": 0.5031, + "step": 27854 + }, + { + "epoch": 1.8694339116137042, + "grad_norm": 1.045531988143921, + "learning_rate": 1.111834461983935e-06, + "loss": 0.5045, + "step": 27856 + }, + { + "epoch": 1.8695681352974733, + "grad_norm": 1.4767205715179443, + "learning_rate": 1.1095564082272469e-06, + "loss": 0.4553, + "step": 27858 + }, + { + "epoch": 1.8697023589812423, + "grad_norm": 1.1055418252944946, + "learning_rate": 1.1072806644478739e-06, + "loss": 0.4786, + "step": 27860 + }, + { + "epoch": 1.8698365826650112, + "grad_norm": 0.6489827036857605, + "learning_rate": 1.105007230753341e-06, + "loss": 0.4137, + "step": 27862 + }, + { + "epoch": 1.86997080634878, + "grad_norm": 1.0843795537948608, + "learning_rate": 1.1027361072510788e-06, + "loss": 0.4643, + "step": 27864 + }, + { + "epoch": 1.8701050300325492, + "grad_norm": 1.3519821166992188, + "learning_rate": 1.1004672940483796e-06, + "loss": 0.5044, + "step": 27866 + }, + { + "epoch": 1.8702392537163184, + "grad_norm": 0.9754189848899841, + "learning_rate": 1.0982007912524405e-06, + "loss": 0.3895, + "step": 27868 + }, + { + "epoch": 1.8703734774000873, + "grad_norm": 0.9900445938110352, + "learning_rate": 1.0959365989703541e-06, + "loss": 0.4306, + "step": 27870 + }, + { + "epoch": 1.8705077010838562, + "grad_norm": 1.0068804025650024, + "learning_rate": 1.093674717309101e-06, + "loss": 0.4488, + "step": 27872 + }, + { + "epoch": 1.8706419247676251, + "grad_norm": 0.9073607921600342, + "learning_rate": 1.0914151463755407e-06, + "loss": 0.4347, + "step": 27874 + }, + { + "epoch": 1.8707761484513943, + "grad_norm": 1.0566726922988892, + "learning_rate": 1.089157886276454e-06, + "loss": 0.4534, + "step": 27876 + }, + { + "epoch": 1.8709103721351632, + "grad_norm": 0.9509255886077881, + "learning_rate": 1.086902937118467e-06, + "loss": 0.4083, + "step": 27878 + }, + { + "epoch": 1.8710445958189323, + "grad_norm": 1.119420051574707, + "learning_rate": 1.084650299008133e-06, + "loss": 0.4667, + "step": 27880 + }, + { + "epoch": 1.8711788195027013, + "grad_norm": 1.0077191591262817, + "learning_rate": 1.082399972051884e-06, + "loss": 0.4978, + "step": 27882 + }, + { + "epoch": 1.8713130431864702, + "grad_norm": 0.8377633094787598, + "learning_rate": 1.080151956356046e-06, + "loss": 0.4594, + "step": 27884 + }, + { + "epoch": 1.871447266870239, + "grad_norm": 0.9869218468666077, + "learning_rate": 1.0779062520268335e-06, + "loss": 0.4655, + "step": 27886 + }, + { + "epoch": 1.8715814905540082, + "grad_norm": 0.9637986421585083, + "learning_rate": 1.0756628591703455e-06, + "loss": 0.464, + "step": 27888 + }, + { + "epoch": 1.8717157142377774, + "grad_norm": 1.01296865940094, + "learning_rate": 1.0734217778925858e-06, + "loss": 0.427, + "step": 27890 + }, + { + "epoch": 1.8718499379215463, + "grad_norm": 0.9736398458480835, + "learning_rate": 1.0711830082994312e-06, + "loss": 0.4895, + "step": 27892 + }, + { + "epoch": 1.8719841616053152, + "grad_norm": 0.9469650387763977, + "learning_rate": 1.068946550496669e-06, + "loss": 0.3917, + "step": 27894 + }, + { + "epoch": 1.8721183852890841, + "grad_norm": 1.0767292976379395, + "learning_rate": 1.0667124045899646e-06, + "loss": 0.4398, + "step": 27896 + }, + { + "epoch": 1.8722526089728533, + "grad_norm": 1.0303303003311157, + "learning_rate": 1.0644805706848782e-06, + "loss": 0.471, + "step": 27898 + }, + { + "epoch": 1.8723868326566224, + "grad_norm": 1.0257786512374878, + "learning_rate": 1.0622510488868476e-06, + "loss": 0.4258, + "step": 27900 + }, + { + "epoch": 1.8725210563403913, + "grad_norm": 1.0219508409500122, + "learning_rate": 1.0600238393012329e-06, + "loss": 0.4765, + "step": 27902 + }, + { + "epoch": 1.8726552800241603, + "grad_norm": 1.1240330934524536, + "learning_rate": 1.05779894203325e-06, + "loss": 0.5023, + "step": 27904 + }, + { + "epoch": 1.8727895037079292, + "grad_norm": 1.1029444932937622, + "learning_rate": 1.0555763571880318e-06, + "loss": 0.4776, + "step": 27906 + }, + { + "epoch": 1.872923727391698, + "grad_norm": 1.269639015197754, + "learning_rate": 1.053356084870577e-06, + "loss": 0.4947, + "step": 27908 + }, + { + "epoch": 1.8730579510754672, + "grad_norm": 0.9267687201499939, + "learning_rate": 1.051138125185802e-06, + "loss": 0.4427, + "step": 27910 + }, + { + "epoch": 1.8731921747592364, + "grad_norm": 1.052469253540039, + "learning_rate": 1.0489224782384955e-06, + "loss": 0.4915, + "step": 27912 + }, + { + "epoch": 1.8733263984430053, + "grad_norm": 0.903032660484314, + "learning_rate": 1.046709144133351e-06, + "loss": 0.4941, + "step": 27914 + }, + { + "epoch": 1.8734606221267742, + "grad_norm": 0.958709716796875, + "learning_rate": 1.0444981229749295e-06, + "loss": 0.4109, + "step": 27916 + }, + { + "epoch": 1.8735948458105431, + "grad_norm": 0.9494258761405945, + "learning_rate": 1.0422894148677087e-06, + "loss": 0.4648, + "step": 27918 + }, + { + "epoch": 1.8737290694943123, + "grad_norm": 1.0680596828460693, + "learning_rate": 1.0400830199160328e-06, + "loss": 0.625, + "step": 27920 + }, + { + "epoch": 1.8738632931780814, + "grad_norm": 1.104437232017517, + "learning_rate": 1.0378789382241682e-06, + "loss": 0.4726, + "step": 27922 + }, + { + "epoch": 1.8739975168618503, + "grad_norm": 1.0886648893356323, + "learning_rate": 1.035677169896243e-06, + "loss": 0.4234, + "step": 27924 + }, + { + "epoch": 1.8741317405456193, + "grad_norm": 1.0351425409317017, + "learning_rate": 1.033477715036285e-06, + "loss": 0.479, + "step": 27926 + }, + { + "epoch": 1.8742659642293882, + "grad_norm": 1.007570743560791, + "learning_rate": 1.0312805737482167e-06, + "loss": 0.4489, + "step": 27928 + }, + { + "epoch": 1.8744001879131573, + "grad_norm": 1.0454297065734863, + "learning_rate": 1.0290857461358495e-06, + "loss": 0.4559, + "step": 27930 + }, + { + "epoch": 1.8745344115969262, + "grad_norm": 1.1414828300476074, + "learning_rate": 1.0268932323028834e-06, + "loss": 0.4754, + "step": 27932 + }, + { + "epoch": 1.8746686352806954, + "grad_norm": 0.9029853343963623, + "learning_rate": 1.0247030323529138e-06, + "loss": 0.4742, + "step": 27934 + }, + { + "epoch": 1.8748028589644643, + "grad_norm": 1.0764508247375488, + "learning_rate": 1.0225151463894133e-06, + "loss": 0.4821, + "step": 27936 + }, + { + "epoch": 1.8749370826482332, + "grad_norm": 1.073437213897705, + "learning_rate": 1.020329574515766e-06, + "loss": 0.5068, + "step": 27938 + }, + { + "epoch": 1.8750713063320021, + "grad_norm": 1.0652241706848145, + "learning_rate": 1.018146316835228e-06, + "loss": 0.5957, + "step": 27940 + }, + { + "epoch": 1.8752055300157713, + "grad_norm": 1.028754711151123, + "learning_rate": 1.0159653734509666e-06, + "loss": 0.4757, + "step": 27942 + }, + { + "epoch": 1.8753397536995404, + "grad_norm": 1.2118598222732544, + "learning_rate": 1.0137867444660055e-06, + "loss": 0.5381, + "step": 27944 + }, + { + "epoch": 1.8754739773833093, + "grad_norm": 0.9155813455581665, + "learning_rate": 1.0116104299833062e-06, + "loss": 0.4686, + "step": 27946 + }, + { + "epoch": 1.8756082010670783, + "grad_norm": 0.9844268560409546, + "learning_rate": 1.0094364301056759e-06, + "loss": 0.4699, + "step": 27948 + }, + { + "epoch": 1.8757424247508472, + "grad_norm": 1.0277265310287476, + "learning_rate": 1.0072647449358375e-06, + "loss": 0.4672, + "step": 27950 + }, + { + "epoch": 1.8758766484346163, + "grad_norm": 0.9917311668395996, + "learning_rate": 1.0050953745764037e-06, + "loss": 0.4056, + "step": 27952 + }, + { + "epoch": 1.8760108721183852, + "grad_norm": 0.9991815090179443, + "learning_rate": 1.0029283191298644e-06, + "loss": 0.488, + "step": 27954 + }, + { + "epoch": 1.8761450958021544, + "grad_norm": 1.0271320343017578, + "learning_rate": 1.0007635786986214e-06, + "loss": 0.4535, + "step": 27956 + }, + { + "epoch": 1.8762793194859233, + "grad_norm": 0.9581825733184814, + "learning_rate": 9.98601153384937e-07, + "loss": 0.475, + "step": 27958 + }, + { + "epoch": 1.8764135431696922, + "grad_norm": 1.1085233688354492, + "learning_rate": 9.964410432909965e-07, + "loss": 0.4434, + "step": 27960 + }, + { + "epoch": 1.8765477668534611, + "grad_norm": 1.0936341285705566, + "learning_rate": 9.94283248518857e-07, + "loss": 0.5027, + "step": 27962 + }, + { + "epoch": 1.8766819905372303, + "grad_norm": 1.6282494068145752, + "learning_rate": 9.921277691704644e-07, + "loss": 0.4618, + "step": 27964 + }, + { + "epoch": 1.8768162142209994, + "grad_norm": 0.9852066040039062, + "learning_rate": 9.899746053476655e-07, + "loss": 0.4552, + "step": 27966 + }, + { + "epoch": 1.8769504379047683, + "grad_norm": 1.048144817352295, + "learning_rate": 9.878237571521898e-07, + "loss": 0.451, + "step": 27968 + }, + { + "epoch": 1.8770846615885373, + "grad_norm": 0.959335446357727, + "learning_rate": 9.85675224685667e-07, + "loss": 0.5187, + "step": 27970 + }, + { + "epoch": 1.8772188852723062, + "grad_norm": 1.0953601598739624, + "learning_rate": 9.835290080496107e-07, + "loss": 0.4798, + "step": 27972 + }, + { + "epoch": 1.8773531089560753, + "grad_norm": 1.058064579963684, + "learning_rate": 9.813851073454117e-07, + "loss": 0.4687, + "step": 27974 + }, + { + "epoch": 1.8774873326398445, + "grad_norm": 0.9155431389808655, + "learning_rate": 9.792435226743835e-07, + "loss": 0.4683, + "step": 27976 + }, + { + "epoch": 1.8776215563236134, + "grad_norm": 0.9273266196250916, + "learning_rate": 9.771042541377008e-07, + "loss": 0.4774, + "step": 27978 + }, + { + "epoch": 1.8777557800073823, + "grad_norm": 0.9717110991477966, + "learning_rate": 9.749673018364436e-07, + "loss": 0.4711, + "step": 27980 + }, + { + "epoch": 1.8778900036911512, + "grad_norm": 1.0166131258010864, + "learning_rate": 9.728326658715815e-07, + "loss": 0.5031, + "step": 27982 + }, + { + "epoch": 1.8780242273749201, + "grad_norm": 1.1387819051742554, + "learning_rate": 9.707003463439668e-07, + "loss": 0.4494, + "step": 27984 + }, + { + "epoch": 1.8781584510586893, + "grad_norm": 1.0542840957641602, + "learning_rate": 9.685703433543414e-07, + "loss": 0.4705, + "step": 27986 + }, + { + "epoch": 1.8782926747424584, + "grad_norm": 1.1977485418319702, + "learning_rate": 9.66442657003358e-07, + "loss": 0.5649, + "step": 27988 + }, + { + "epoch": 1.8784268984262273, + "grad_norm": 1.185279130935669, + "learning_rate": 9.643172873915363e-07, + "loss": 0.4437, + "step": 27990 + }, + { + "epoch": 1.8785611221099963, + "grad_norm": 0.9907231330871582, + "learning_rate": 9.621942346193013e-07, + "loss": 0.445, + "step": 27992 + }, + { + "epoch": 1.8786953457937652, + "grad_norm": 1.0656880140304565, + "learning_rate": 9.600734987869564e-07, + "loss": 0.5222, + "step": 27994 + }, + { + "epoch": 1.8788295694775343, + "grad_norm": 1.0482569932937622, + "learning_rate": 9.579550799947046e-07, + "loss": 0.433, + "step": 27996 + }, + { + "epoch": 1.8789637931613035, + "grad_norm": 1.067169189453125, + "learning_rate": 9.558389783426436e-07, + "loss": 0.4399, + "step": 27998 + }, + { + "epoch": 1.8790980168450724, + "grad_norm": 1.0769200325012207, + "learning_rate": 9.53725193930749e-07, + "loss": 0.4912, + "step": 28000 + }, + { + "epoch": 1.8792322405288413, + "grad_norm": 1.0300370454788208, + "learning_rate": 9.516137268588854e-07, + "loss": 0.4552, + "step": 28002 + }, + { + "epoch": 1.8793664642126102, + "grad_norm": 0.9600032567977905, + "learning_rate": 9.495045772268341e-07, + "loss": 0.424, + "step": 28004 + }, + { + "epoch": 1.8795006878963794, + "grad_norm": 1.1331719160079956, + "learning_rate": 9.473977451342265e-07, + "loss": 0.4525, + "step": 28006 + }, + { + "epoch": 1.8796349115801483, + "grad_norm": 1.1848702430725098, + "learning_rate": 9.452932306806273e-07, + "loss": 0.4544, + "step": 28008 + }, + { + "epoch": 1.8797691352639174, + "grad_norm": 1.0405181646347046, + "learning_rate": 9.43191033965457e-07, + "loss": 0.4596, + "step": 28010 + }, + { + "epoch": 1.8799033589476863, + "grad_norm": 1.0564866065979004, + "learning_rate": 9.410911550880475e-07, + "loss": 0.4541, + "step": 28012 + }, + { + "epoch": 1.8800375826314553, + "grad_norm": 0.9685310125350952, + "learning_rate": 9.38993594147608e-07, + "loss": 0.3954, + "step": 28014 + }, + { + "epoch": 1.8801718063152242, + "grad_norm": 1.0685757398605347, + "learning_rate": 9.368983512432483e-07, + "loss": 0.4814, + "step": 28016 + }, + { + "epoch": 1.8803060299989933, + "grad_norm": 1.1232695579528809, + "learning_rate": 9.348054264739614e-07, + "loss": 0.4593, + "step": 28018 + }, + { + "epoch": 1.8804402536827625, + "grad_norm": 1.0030947923660278, + "learning_rate": 9.327148199386404e-07, + "loss": 0.4582, + "step": 28020 + }, + { + "epoch": 1.8805744773665314, + "grad_norm": 1.0549216270446777, + "learning_rate": 9.306265317360507e-07, + "loss": 0.5093, + "step": 28022 + }, + { + "epoch": 1.8807087010503003, + "grad_norm": 0.9294421672821045, + "learning_rate": 9.28540561964869e-07, + "loss": 0.4286, + "step": 28024 + }, + { + "epoch": 1.8808429247340692, + "grad_norm": 1.1034150123596191, + "learning_rate": 9.264569107236498e-07, + "loss": 0.5005, + "step": 28026 + }, + { + "epoch": 1.8809771484178384, + "grad_norm": 1.177290678024292, + "learning_rate": 9.243755781108476e-07, + "loss": 0.4942, + "step": 28028 + }, + { + "epoch": 1.8811113721016073, + "grad_norm": 0.9686825275421143, + "learning_rate": 9.222965642247949e-07, + "loss": 0.4471, + "step": 28030 + }, + { + "epoch": 1.8812455957853764, + "grad_norm": 1.0099127292633057, + "learning_rate": 9.202198691637131e-07, + "loss": 0.4512, + "step": 28032 + }, + { + "epoch": 1.8813798194691453, + "grad_norm": 1.0454397201538086, + "learning_rate": 9.181454930257405e-07, + "loss": 0.4566, + "step": 28034 + }, + { + "epoch": 1.8815140431529143, + "grad_norm": 1.2138104438781738, + "learning_rate": 9.160734359088763e-07, + "loss": 0.4955, + "step": 28036 + }, + { + "epoch": 1.8816482668366832, + "grad_norm": 0.9711406230926514, + "learning_rate": 9.140036979110256e-07, + "loss": 0.4588, + "step": 28038 + }, + { + "epoch": 1.8817824905204523, + "grad_norm": 1.0114809274673462, + "learning_rate": 9.119362791299713e-07, + "loss": 0.427, + "step": 28040 + }, + { + "epoch": 1.8819167142042215, + "grad_norm": 1.0150818824768066, + "learning_rate": 9.098711796634018e-07, + "loss": 0.4919, + "step": 28042 + }, + { + "epoch": 1.8820509378879904, + "grad_norm": 0.9563175439834595, + "learning_rate": 9.078083996088838e-07, + "loss": 0.4625, + "step": 28044 + }, + { + "epoch": 1.8821851615717593, + "grad_norm": 0.8081610202789307, + "learning_rate": 9.05747939063889e-07, + "loss": 0.368, + "step": 28046 + }, + { + "epoch": 1.8823193852555282, + "grad_norm": 1.072668194770813, + "learning_rate": 9.036897981257675e-07, + "loss": 0.4521, + "step": 28048 + }, + { + "epoch": 1.8824536089392974, + "grad_norm": 1.107627272605896, + "learning_rate": 9.016339768917526e-07, + "loss": 0.4439, + "step": 28050 + }, + { + "epoch": 1.8825878326230665, + "grad_norm": 1.011053442955017, + "learning_rate": 8.995804754589832e-07, + "loss": 0.493, + "step": 28052 + }, + { + "epoch": 1.8827220563068354, + "grad_norm": 1.0553630590438843, + "learning_rate": 8.975292939244928e-07, + "loss": 0.4957, + "step": 28054 + }, + { + "epoch": 1.8828562799906043, + "grad_norm": 1.025673508644104, + "learning_rate": 8.954804323851818e-07, + "loss": 0.4545, + "step": 28056 + }, + { + "epoch": 1.8829905036743733, + "grad_norm": 0.9037055969238281, + "learning_rate": 8.934338909378615e-07, + "loss": 0.3896, + "step": 28058 + }, + { + "epoch": 1.8831247273581422, + "grad_norm": 1.1714078187942505, + "learning_rate": 8.913896696792212e-07, + "loss": 0.4653, + "step": 28060 + }, + { + "epoch": 1.8832589510419113, + "grad_norm": 0.8923357129096985, + "learning_rate": 8.893477687058615e-07, + "loss": 0.4543, + "step": 28062 + }, + { + "epoch": 1.8833931747256805, + "grad_norm": 0.9870643019676208, + "learning_rate": 8.873081881142386e-07, + "loss": 0.4144, + "step": 28064 + }, + { + "epoch": 1.8835273984094494, + "grad_norm": 1.1387721300125122, + "learning_rate": 8.85270928000731e-07, + "loss": 0.4651, + "step": 28066 + }, + { + "epoch": 1.8836616220932183, + "grad_norm": 1.132055401802063, + "learning_rate": 8.832359884615893e-07, + "loss": 0.4449, + "step": 28068 + }, + { + "epoch": 1.8837958457769872, + "grad_norm": 1.0421757698059082, + "learning_rate": 8.8120336959297e-07, + "loss": 0.4609, + "step": 28070 + }, + { + "epoch": 1.8839300694607564, + "grad_norm": 0.9961799383163452, + "learning_rate": 8.791730714908964e-07, + "loss": 0.4653, + "step": 28072 + }, + { + "epoch": 1.8840642931445255, + "grad_norm": 0.8882201910018921, + "learning_rate": 8.771450942513081e-07, + "loss": 0.4657, + "step": 28074 + }, + { + "epoch": 1.8841985168282944, + "grad_norm": 1.0295106172561646, + "learning_rate": 8.751194379700179e-07, + "loss": 0.4799, + "step": 28076 + }, + { + "epoch": 1.8843327405120633, + "grad_norm": 0.9758402109146118, + "learning_rate": 8.730961027427321e-07, + "loss": 0.4765, + "step": 28078 + }, + { + "epoch": 1.8844669641958323, + "grad_norm": 1.07142972946167, + "learning_rate": 8.710750886650465e-07, + "loss": 0.5259, + "step": 28080 + }, + { + "epoch": 1.8846011878796014, + "grad_norm": 1.0539863109588623, + "learning_rate": 8.690563958324627e-07, + "loss": 0.4452, + "step": 28082 + }, + { + "epoch": 1.8847354115633703, + "grad_norm": 1.1762025356292725, + "learning_rate": 8.670400243403543e-07, + "loss": 0.5145, + "step": 28084 + }, + { + "epoch": 1.8848696352471395, + "grad_norm": 0.940888524055481, + "learning_rate": 8.65025974283984e-07, + "loss": 0.505, + "step": 28086 + }, + { + "epoch": 1.8850038589309084, + "grad_norm": 0.9293429255485535, + "learning_rate": 8.6301424575852e-07, + "loss": 0.4445, + "step": 28088 + }, + { + "epoch": 1.8851380826146773, + "grad_norm": 1.121043086051941, + "learning_rate": 8.610048388590031e-07, + "loss": 0.4739, + "step": 28090 + }, + { + "epoch": 1.8852723062984462, + "grad_norm": 0.9941401481628418, + "learning_rate": 8.58997753680385e-07, + "loss": 0.4619, + "step": 28092 + }, + { + "epoch": 1.8854065299822154, + "grad_norm": 0.9810492992401123, + "learning_rate": 8.569929903174901e-07, + "loss": 0.4199, + "step": 28094 + }, + { + "epoch": 1.8855407536659845, + "grad_norm": 0.9809359312057495, + "learning_rate": 8.549905488650422e-07, + "loss": 0.417, + "step": 28096 + }, + { + "epoch": 1.8856749773497534, + "grad_norm": 1.2116690874099731, + "learning_rate": 8.529904294176494e-07, + "loss": 0.4513, + "step": 28098 + }, + { + "epoch": 1.8858092010335223, + "grad_norm": 1.0022441148757935, + "learning_rate": 8.509926320698137e-07, + "loss": 0.4976, + "step": 28100 + }, + { + "epoch": 1.8859434247172913, + "grad_norm": 1.0324827432632446, + "learning_rate": 8.489971569159261e-07, + "loss": 0.478, + "step": 28102 + }, + { + "epoch": 1.8860776484010604, + "grad_norm": 1.0444756746292114, + "learning_rate": 8.470040040502836e-07, + "loss": 0.4764, + "step": 28104 + }, + { + "epoch": 1.8862118720848293, + "grad_norm": 1.0981961488723755, + "learning_rate": 8.450131735670386e-07, + "loss": 0.5176, + "step": 28106 + }, + { + "epoch": 1.8863460957685985, + "grad_norm": 0.9942836165428162, + "learning_rate": 8.430246655602602e-07, + "loss": 0.4403, + "step": 28108 + }, + { + "epoch": 1.8864803194523674, + "grad_norm": 0.9508644342422485, + "learning_rate": 8.410384801239068e-07, + "loss": 0.4027, + "step": 28110 + }, + { + "epoch": 1.8866145431361363, + "grad_norm": 1.0890480279922485, + "learning_rate": 8.390546173518143e-07, + "loss": 0.4543, + "step": 28112 + }, + { + "epoch": 1.8867487668199052, + "grad_norm": 1.0575063228607178, + "learning_rate": 8.370730773377245e-07, + "loss": 0.5199, + "step": 28114 + }, + { + "epoch": 1.8868829905036744, + "grad_norm": 1.016358733177185, + "learning_rate": 8.350938601752567e-07, + "loss": 0.4694, + "step": 28116 + }, + { + "epoch": 1.8870172141874435, + "grad_norm": 3.7425155639648438, + "learning_rate": 8.331169659579252e-07, + "loss": 0.4485, + "step": 28118 + }, + { + "epoch": 1.8871514378712124, + "grad_norm": 0.9548282623291016, + "learning_rate": 8.31142394779133e-07, + "loss": 0.4012, + "step": 28120 + }, + { + "epoch": 1.8872856615549813, + "grad_norm": 1.1152081489562988, + "learning_rate": 8.291701467321778e-07, + "loss": 0.4721, + "step": 28122 + }, + { + "epoch": 1.8874198852387503, + "grad_norm": 1.0232319831848145, + "learning_rate": 8.272002219102459e-07, + "loss": 0.5133, + "step": 28124 + }, + { + "epoch": 1.8875541089225194, + "grad_norm": 1.034847378730774, + "learning_rate": 8.252326204064021e-07, + "loss": 0.4311, + "step": 28126 + }, + { + "epoch": 1.8876883326062885, + "grad_norm": 1.163466453552246, + "learning_rate": 8.232673423136217e-07, + "loss": 0.4596, + "step": 28128 + }, + { + "epoch": 1.8878225562900575, + "grad_norm": 1.1249845027923584, + "learning_rate": 8.213043877247528e-07, + "loss": 0.4962, + "step": 28130 + }, + { + "epoch": 1.8879567799738264, + "grad_norm": 1.1233371496200562, + "learning_rate": 8.193437567325546e-07, + "loss": 0.4717, + "step": 28132 + }, + { + "epoch": 1.8880910036575953, + "grad_norm": 1.1094117164611816, + "learning_rate": 8.173854494296529e-07, + "loss": 0.5262, + "step": 28134 + }, + { + "epoch": 1.8882252273413642, + "grad_norm": 1.107286810874939, + "learning_rate": 8.154294659085737e-07, + "loss": 0.5168, + "step": 28136 + }, + { + "epoch": 1.8883594510251334, + "grad_norm": 1.1415817737579346, + "learning_rate": 8.13475806261732e-07, + "loss": 0.5258, + "step": 28138 + }, + { + "epoch": 1.8884936747089025, + "grad_norm": 1.0087707042694092, + "learning_rate": 8.115244705814429e-07, + "loss": 0.4978, + "step": 28140 + }, + { + "epoch": 1.8886278983926714, + "grad_norm": 0.979401707649231, + "learning_rate": 8.095754589598936e-07, + "loss": 0.5136, + "step": 28142 + }, + { + "epoch": 1.8887621220764403, + "grad_norm": 1.1137760877609253, + "learning_rate": 8.076287714891773e-07, + "loss": 0.5194, + "step": 28144 + }, + { + "epoch": 1.8888963457602093, + "grad_norm": 0.983637809753418, + "learning_rate": 8.056844082612647e-07, + "loss": 0.3967, + "step": 28146 + }, + { + "epoch": 1.8890305694439784, + "grad_norm": 1.125389575958252, + "learning_rate": 8.037423693680324e-07, + "loss": 0.4757, + "step": 28148 + }, + { + "epoch": 1.8891647931277475, + "grad_norm": 1.065266728401184, + "learning_rate": 8.018026549012292e-07, + "loss": 0.4739, + "step": 28150 + }, + { + "epoch": 1.8892990168115165, + "grad_norm": 1.0932797193527222, + "learning_rate": 7.998652649525096e-07, + "loss": 0.4626, + "step": 28152 + }, + { + "epoch": 1.8894332404952854, + "grad_norm": 1.0319827795028687, + "learning_rate": 7.979301996134059e-07, + "loss": 0.4448, + "step": 28154 + }, + { + "epoch": 1.8895674641790543, + "grad_norm": 1.1214451789855957, + "learning_rate": 7.95997458975345e-07, + "loss": 0.4566, + "step": 28156 + }, + { + "epoch": 1.8897016878628234, + "grad_norm": 1.0158464908599854, + "learning_rate": 7.940670431296538e-07, + "loss": 0.4573, + "step": 28158 + }, + { + "epoch": 1.8898359115465924, + "grad_norm": 1.0854089260101318, + "learning_rate": 7.921389521675315e-07, + "loss": 0.4473, + "step": 28160 + }, + { + "epoch": 1.8899701352303615, + "grad_norm": 1.1003270149230957, + "learning_rate": 7.902131861800888e-07, + "loss": 0.4704, + "step": 28162 + }, + { + "epoch": 1.8901043589141304, + "grad_norm": 1.0552144050598145, + "learning_rate": 7.882897452583027e-07, + "loss": 0.433, + "step": 28164 + }, + { + "epoch": 1.8902385825978993, + "grad_norm": 1.0242884159088135, + "learning_rate": 7.863686294930506e-07, + "loss": 0.4525, + "step": 28166 + }, + { + "epoch": 1.8903728062816683, + "grad_norm": 1.108998417854309, + "learning_rate": 7.844498389751098e-07, + "loss": 0.4841, + "step": 28168 + }, + { + "epoch": 1.8905070299654374, + "grad_norm": 1.0566809177398682, + "learning_rate": 7.825333737951357e-07, + "loss": 0.5314, + "step": 28170 + }, + { + "epoch": 1.8906412536492065, + "grad_norm": 1.0766716003417969, + "learning_rate": 7.80619234043678e-07, + "loss": 0.4626, + "step": 28172 + }, + { + "epoch": 1.8907754773329755, + "grad_norm": 0.9279155135154724, + "learning_rate": 7.787074198111755e-07, + "loss": 0.4588, + "step": 28174 + }, + { + "epoch": 1.8909097010167444, + "grad_norm": 0.9887163043022156, + "learning_rate": 7.767979311879559e-07, + "loss": 0.4419, + "step": 28176 + }, + { + "epoch": 1.8910439247005133, + "grad_norm": 0.9838598966598511, + "learning_rate": 7.748907682642414e-07, + "loss": 0.4378, + "step": 28178 + }, + { + "epoch": 1.8911781483842824, + "grad_norm": 1.0394545793533325, + "learning_rate": 7.729859311301435e-07, + "loss": 0.5049, + "step": 28180 + }, + { + "epoch": 1.8913123720680514, + "grad_norm": 0.8739699125289917, + "learning_rate": 7.710834198756622e-07, + "loss": 0.423, + "step": 28182 + }, + { + "epoch": 1.8914465957518205, + "grad_norm": 1.0114548206329346, + "learning_rate": 7.691832345906757e-07, + "loss": 0.4613, + "step": 28184 + }, + { + "epoch": 1.8915808194355894, + "grad_norm": 1.0551714897155762, + "learning_rate": 7.672853753649789e-07, + "loss": 0.4701, + "step": 28186 + }, + { + "epoch": 1.8917150431193583, + "grad_norm": 1.108930230140686, + "learning_rate": 7.653898422882333e-07, + "loss": 0.4972, + "step": 28188 + }, + { + "epoch": 1.8918492668031273, + "grad_norm": 1.1477514505386353, + "learning_rate": 7.634966354500117e-07, + "loss": 0.4582, + "step": 28190 + }, + { + "epoch": 1.8919834904868964, + "grad_norm": 0.8802066445350647, + "learning_rate": 7.616057549397427e-07, + "loss": 0.4003, + "step": 28192 + }, + { + "epoch": 1.8921177141706655, + "grad_norm": 0.9493222236633301, + "learning_rate": 7.597172008467824e-07, + "loss": 0.444, + "step": 28194 + }, + { + "epoch": 1.8922519378544345, + "grad_norm": 1.0440152883529663, + "learning_rate": 7.578309732603539e-07, + "loss": 0.5023, + "step": 28196 + }, + { + "epoch": 1.8923861615382034, + "grad_norm": 1.0263266563415527, + "learning_rate": 7.559470722695916e-07, + "loss": 0.4919, + "step": 28198 + }, + { + "epoch": 1.8925203852219723, + "grad_norm": 1.7940959930419922, + "learning_rate": 7.540654979634909e-07, + "loss": 0.4466, + "step": 28200 + }, + { + "epoch": 1.8926546089057414, + "grad_norm": 1.1088297367095947, + "learning_rate": 7.52186250430953e-07, + "loss": 0.4774, + "step": 28202 + }, + { + "epoch": 1.8927888325895106, + "grad_norm": 0.8987371325492859, + "learning_rate": 7.503093297607732e-07, + "loss": 0.476, + "step": 28204 + }, + { + "epoch": 1.8929230562732795, + "grad_norm": 0.9417023062705994, + "learning_rate": 7.484347360416367e-07, + "loss": 0.4354, + "step": 28206 + }, + { + "epoch": 1.8930572799570484, + "grad_norm": 1.0279747247695923, + "learning_rate": 7.465624693621109e-07, + "loss": 0.4528, + "step": 28208 + }, + { + "epoch": 1.8931915036408173, + "grad_norm": 1.2020224332809448, + "learning_rate": 7.446925298106532e-07, + "loss": 0.5022, + "step": 28210 + }, + { + "epoch": 1.8933257273245863, + "grad_norm": 1.1889111995697021, + "learning_rate": 7.428249174756152e-07, + "loss": 0.541, + "step": 28212 + }, + { + "epoch": 1.8934599510083554, + "grad_norm": 1.0302174091339111, + "learning_rate": 7.409596324452428e-07, + "loss": 0.465, + "step": 28214 + }, + { + "epoch": 1.8935941746921245, + "grad_norm": 1.1223982572555542, + "learning_rate": 7.390966748076599e-07, + "loss": 0.4886, + "step": 28216 + }, + { + "epoch": 1.8937283983758935, + "grad_norm": 1.1422455310821533, + "learning_rate": 7.372360446509019e-07, + "loss": 0.4741, + "step": 28218 + }, + { + "epoch": 1.8938626220596624, + "grad_norm": 0.883145809173584, + "learning_rate": 7.353777420628594e-07, + "loss": 0.4819, + "step": 28220 + }, + { + "epoch": 1.8939968457434313, + "grad_norm": 1.1092454195022583, + "learning_rate": 7.335217671313455e-07, + "loss": 0.5085, + "step": 28222 + }, + { + "epoch": 1.8941310694272004, + "grad_norm": 1.1661089658737183, + "learning_rate": 7.316681199440568e-07, + "loss": 0.4922, + "step": 28224 + }, + { + "epoch": 1.8942652931109696, + "grad_norm": 1.0931086540222168, + "learning_rate": 7.298168005885564e-07, + "loss": 0.4556, + "step": 28226 + }, + { + "epoch": 1.8943995167947385, + "grad_norm": 1.04469895362854, + "learning_rate": 7.279678091523357e-07, + "loss": 0.4748, + "step": 28228 + }, + { + "epoch": 1.8945337404785074, + "grad_norm": 1.0515769720077515, + "learning_rate": 7.261211457227413e-07, + "loss": 0.4116, + "step": 28230 + }, + { + "epoch": 1.8946679641622763, + "grad_norm": 0.9401378035545349, + "learning_rate": 7.242768103870312e-07, + "loss": 0.4345, + "step": 28232 + }, + { + "epoch": 1.8948021878460455, + "grad_norm": 1.1771726608276367, + "learning_rate": 7.22434803232347e-07, + "loss": 0.4912, + "step": 28234 + }, + { + "epoch": 1.8949364115298144, + "grad_norm": 1.0906013250350952, + "learning_rate": 7.205951243457132e-07, + "loss": 0.4631, + "step": 28236 + }, + { + "epoch": 1.8950706352135835, + "grad_norm": 1.1231417655944824, + "learning_rate": 7.187577738140605e-07, + "loss": 0.5209, + "step": 28238 + }, + { + "epoch": 1.8952048588973525, + "grad_norm": 0.9925578236579895, + "learning_rate": 7.169227517241972e-07, + "loss": 0.4699, + "step": 28240 + }, + { + "epoch": 1.8953390825811214, + "grad_norm": 1.080661416053772, + "learning_rate": 7.150900581628206e-07, + "loss": 0.4685, + "step": 28242 + }, + { + "epoch": 1.8954733062648903, + "grad_norm": 0.9531117081642151, + "learning_rate": 7.132596932165225e-07, + "loss": 0.4411, + "step": 28244 + }, + { + "epoch": 1.8956075299486594, + "grad_norm": 0.9632938504219055, + "learning_rate": 7.114316569717894e-07, + "loss": 0.4384, + "step": 28246 + }, + { + "epoch": 1.8957417536324286, + "grad_norm": 0.9770664572715759, + "learning_rate": 7.096059495149854e-07, + "loss": 0.4697, + "step": 28248 + }, + { + "epoch": 1.8958759773161975, + "grad_norm": 1.046528935432434, + "learning_rate": 7.077825709323749e-07, + "loss": 0.455, + "step": 28250 + }, + { + "epoch": 1.8960102009999664, + "grad_norm": 1.4299813508987427, + "learning_rate": 7.059615213101112e-07, + "loss": 0.442, + "step": 28252 + }, + { + "epoch": 1.8961444246837353, + "grad_norm": 1.1879844665527344, + "learning_rate": 7.041428007342254e-07, + "loss": 0.4722, + "step": 28254 + }, + { + "epoch": 1.8962786483675045, + "grad_norm": 1.0131479501724243, + "learning_rate": 7.023264092906711e-07, + "loss": 0.5223, + "step": 28256 + }, + { + "epoch": 1.8964128720512734, + "grad_norm": 0.9336449503898621, + "learning_rate": 7.005123470652408e-07, + "loss": 0.4136, + "step": 28258 + }, + { + "epoch": 1.8965470957350425, + "grad_norm": 1.1306425333023071, + "learning_rate": 6.987006141436659e-07, + "loss": 0.4994, + "step": 28260 + }, + { + "epoch": 1.8966813194188115, + "grad_norm": 1.1440653800964355, + "learning_rate": 6.968912106115333e-07, + "loss": 0.4489, + "step": 28262 + }, + { + "epoch": 1.8968155431025804, + "grad_norm": 1.000525951385498, + "learning_rate": 6.95084136554347e-07, + "loss": 0.4658, + "step": 28264 + }, + { + "epoch": 1.8969497667863493, + "grad_norm": 1.216410756111145, + "learning_rate": 6.932793920574831e-07, + "loss": 0.4564, + "step": 28266 + }, + { + "epoch": 1.8970839904701184, + "grad_norm": 1.0653504133224487, + "learning_rate": 6.914769772062069e-07, + "loss": 0.4797, + "step": 28268 + }, + { + "epoch": 1.8972182141538876, + "grad_norm": 1.0873147249221802, + "learning_rate": 6.896768920856778e-07, + "loss": 0.4857, + "step": 28270 + }, + { + "epoch": 1.8973524378376565, + "grad_norm": 1.1958423852920532, + "learning_rate": 6.878791367809556e-07, + "loss": 0.4793, + "step": 28272 + }, + { + "epoch": 1.8974866615214254, + "grad_norm": 1.1073347330093384, + "learning_rate": 6.86083711376978e-07, + "loss": 0.4675, + "step": 28274 + }, + { + "epoch": 1.8976208852051943, + "grad_norm": 1.1181777715682983, + "learning_rate": 6.842906159585716e-07, + "loss": 0.5045, + "step": 28276 + }, + { + "epoch": 1.8977551088889635, + "grad_norm": 0.8882530331611633, + "learning_rate": 6.824998506104574e-07, + "loss": 0.4368, + "step": 28278 + }, + { + "epoch": 1.8978893325727326, + "grad_norm": 1.0279948711395264, + "learning_rate": 6.807114154172456e-07, + "loss": 0.54, + "step": 28280 + }, + { + "epoch": 1.8980235562565015, + "grad_norm": 0.9511619210243225, + "learning_rate": 6.789253104634352e-07, + "loss": 0.4265, + "step": 28282 + }, + { + "epoch": 1.8981577799402705, + "grad_norm": 1.027387261390686, + "learning_rate": 6.77141535833431e-07, + "loss": 0.4772, + "step": 28284 + }, + { + "epoch": 1.8982920036240394, + "grad_norm": 1.0526982545852661, + "learning_rate": 6.753600916114877e-07, + "loss": 0.4691, + "step": 28286 + }, + { + "epoch": 1.8984262273078083, + "grad_norm": 1.0057649612426758, + "learning_rate": 6.735809778817881e-07, + "loss": 0.4471, + "step": 28288 + }, + { + "epoch": 1.8985604509915774, + "grad_norm": 1.0767104625701904, + "learning_rate": 6.718041947283926e-07, + "loss": 0.5003, + "step": 28290 + }, + { + "epoch": 1.8986946746753466, + "grad_norm": 1.1719970703125, + "learning_rate": 6.700297422352508e-07, + "loss": 0.4701, + "step": 28292 + }, + { + "epoch": 1.8988288983591155, + "grad_norm": 0.998748242855072, + "learning_rate": 6.682576204862012e-07, + "loss": 0.4037, + "step": 28294 + }, + { + "epoch": 1.8989631220428844, + "grad_norm": 1.0387566089630127, + "learning_rate": 6.664878295649713e-07, + "loss": 0.467, + "step": 28296 + }, + { + "epoch": 1.8990973457266533, + "grad_norm": 0.9817729592323303, + "learning_rate": 6.647203695551829e-07, + "loss": 0.51, + "step": 28298 + }, + { + "epoch": 1.8992315694104225, + "grad_norm": 1.0320035219192505, + "learning_rate": 6.629552405403361e-07, + "loss": 0.4597, + "step": 28300 + }, + { + "epoch": 1.8993657930941916, + "grad_norm": 1.0402313470840454, + "learning_rate": 6.611924426038419e-07, + "loss": 0.4076, + "step": 28302 + }, + { + "epoch": 1.8995000167779605, + "grad_norm": 0.920330286026001, + "learning_rate": 6.594319758289836e-07, + "loss": 0.4829, + "step": 28304 + }, + { + "epoch": 1.8996342404617295, + "grad_norm": 0.9015540480613708, + "learning_rate": 6.576738402989447e-07, + "loss": 0.4493, + "step": 28306 + }, + { + "epoch": 1.8997684641454984, + "grad_norm": 1.0200977325439453, + "learning_rate": 6.55918036096781e-07, + "loss": 0.4504, + "step": 28308 + }, + { + "epoch": 1.8999026878292675, + "grad_norm": 0.9403783082962036, + "learning_rate": 6.54164563305465e-07, + "loss": 0.4972, + "step": 28310 + }, + { + "epoch": 1.9000369115130364, + "grad_norm": 1.0478039979934692, + "learning_rate": 6.52413422007836e-07, + "loss": 0.4678, + "step": 28312 + }, + { + "epoch": 1.9001711351968056, + "grad_norm": 1.0461442470550537, + "learning_rate": 6.506646122866445e-07, + "loss": 0.3982, + "step": 28314 + }, + { + "epoch": 1.9003053588805745, + "grad_norm": 1.138066291809082, + "learning_rate": 6.489181342244965e-07, + "loss": 0.5021, + "step": 28316 + }, + { + "epoch": 1.9004395825643434, + "grad_norm": 1.0282994508743286, + "learning_rate": 6.471739879039262e-07, + "loss": 0.4285, + "step": 28318 + }, + { + "epoch": 1.9005738062481123, + "grad_norm": 1.1668262481689453, + "learning_rate": 6.454321734073344e-07, + "loss": 0.5443, + "step": 28320 + }, + { + "epoch": 1.9007080299318815, + "grad_norm": 1.0456217527389526, + "learning_rate": 6.43692690817027e-07, + "loss": 0.5124, + "step": 28322 + }, + { + "epoch": 1.9008422536156506, + "grad_norm": 1.0460165739059448, + "learning_rate": 6.419555402151777e-07, + "loss": 0.543, + "step": 28324 + }, + { + "epoch": 1.9009764772994195, + "grad_norm": 1.1124913692474365, + "learning_rate": 6.402207216838762e-07, + "loss": 0.4607, + "step": 28326 + }, + { + "epoch": 1.9011107009831885, + "grad_norm": 0.9830761551856995, + "learning_rate": 6.384882353050791e-07, + "loss": 0.5103, + "step": 28328 + }, + { + "epoch": 1.9012449246669574, + "grad_norm": 1.0155304670333862, + "learning_rate": 6.367580811606544e-07, + "loss": 0.4781, + "step": 28330 + }, + { + "epoch": 1.9013791483507265, + "grad_norm": 1.1133677959442139, + "learning_rate": 6.350302593323365e-07, + "loss": 0.4248, + "step": 28332 + }, + { + "epoch": 1.9015133720344954, + "grad_norm": 1.009499430656433, + "learning_rate": 6.333047699017714e-07, + "loss": 0.4323, + "step": 28334 + }, + { + "epoch": 1.9016475957182646, + "grad_norm": 1.0685688257217407, + "learning_rate": 6.315816129504715e-07, + "loss": 0.4891, + "step": 28336 + }, + { + "epoch": 1.9017818194020335, + "grad_norm": 1.0232664346694946, + "learning_rate": 6.298607885598718e-07, + "loss": 0.4593, + "step": 28338 + }, + { + "epoch": 1.9019160430858024, + "grad_norm": 1.0040220022201538, + "learning_rate": 6.281422968112571e-07, + "loss": 0.469, + "step": 28340 + }, + { + "epoch": 1.9020502667695713, + "grad_norm": 0.9486731886863708, + "learning_rate": 6.26426137785846e-07, + "loss": 0.435, + "step": 28342 + }, + { + "epoch": 1.9021844904533405, + "grad_norm": 0.8470131754875183, + "learning_rate": 6.247123115647013e-07, + "loss": 0.4355, + "step": 28344 + }, + { + "epoch": 1.9023187141371096, + "grad_norm": 0.9575321078300476, + "learning_rate": 6.230008182288083e-07, + "loss": 0.4909, + "step": 28346 + }, + { + "epoch": 1.9024529378208785, + "grad_norm": 1.233555555343628, + "learning_rate": 6.212916578590355e-07, + "loss": 0.4328, + "step": 28348 + }, + { + "epoch": 1.9025871615046475, + "grad_norm": 1.0835378170013428, + "learning_rate": 6.195848305361296e-07, + "loss": 0.5304, + "step": 28350 + }, + { + "epoch": 1.9027213851884164, + "grad_norm": 1.0284099578857422, + "learning_rate": 6.178803363407371e-07, + "loss": 0.465, + "step": 28352 + }, + { + "epoch": 1.9028556088721855, + "grad_norm": 1.1114816665649414, + "learning_rate": 6.16178175353399e-07, + "loss": 0.4593, + "step": 28354 + }, + { + "epoch": 1.9029898325559547, + "grad_norm": 1.04660165309906, + "learning_rate": 6.144783476545234e-07, + "loss": 0.4788, + "step": 28356 + }, + { + "epoch": 1.9031240562397236, + "grad_norm": 0.9865087270736694, + "learning_rate": 6.127808533244406e-07, + "loss": 0.478, + "step": 28358 + }, + { + "epoch": 1.9032582799234925, + "grad_norm": 1.0369333028793335, + "learning_rate": 6.110856924433473e-07, + "loss": 0.4624, + "step": 28360 + }, + { + "epoch": 1.9033925036072614, + "grad_norm": 1.0701358318328857, + "learning_rate": 6.093928650913294e-07, + "loss": 0.4459, + "step": 28362 + }, + { + "epoch": 1.9035267272910303, + "grad_norm": 0.9676421284675598, + "learning_rate": 6.077023713483843e-07, + "loss": 0.4204, + "step": 28364 + }, + { + "epoch": 1.9036609509747995, + "grad_norm": 0.9129562377929688, + "learning_rate": 6.060142112943701e-07, + "loss": 0.4245, + "step": 28366 + }, + { + "epoch": 1.9037951746585686, + "grad_norm": 1.0327708721160889, + "learning_rate": 6.043283850090564e-07, + "loss": 0.5011, + "step": 28368 + }, + { + "epoch": 1.9039293983423375, + "grad_norm": 1.020492672920227, + "learning_rate": 6.026448925720962e-07, + "loss": 0.4124, + "step": 28370 + }, + { + "epoch": 1.9040636220261065, + "grad_norm": 1.1958445310592651, + "learning_rate": 6.009637340630258e-07, + "loss": 0.5311, + "step": 28372 + }, + { + "epoch": 1.9041978457098754, + "grad_norm": 1.1057733297348022, + "learning_rate": 5.992849095612819e-07, + "loss": 0.4411, + "step": 28374 + }, + { + "epoch": 1.9043320693936445, + "grad_norm": 1.0061222314834595, + "learning_rate": 5.97608419146184e-07, + "loss": 0.4637, + "step": 28376 + }, + { + "epoch": 1.9044662930774137, + "grad_norm": 1.0512781143188477, + "learning_rate": 5.95934262896941e-07, + "loss": 0.4505, + "step": 28378 + }, + { + "epoch": 1.9046005167611826, + "grad_norm": 1.0482077598571777, + "learning_rate": 5.942624408926623e-07, + "loss": 0.4323, + "step": 28380 + }, + { + "epoch": 1.9047347404449515, + "grad_norm": 1.0113775730133057, + "learning_rate": 5.925929532123231e-07, + "loss": 0.436, + "step": 28382 + }, + { + "epoch": 1.9048689641287204, + "grad_norm": 1.005273699760437, + "learning_rate": 5.909257999348106e-07, + "loss": 0.3966, + "step": 28384 + }, + { + "epoch": 1.9050031878124896, + "grad_norm": 1.0841667652130127, + "learning_rate": 5.892609811388949e-07, + "loss": 0.4567, + "step": 28386 + }, + { + "epoch": 1.9051374114962585, + "grad_norm": 1.0270723104476929, + "learning_rate": 5.87598496903241e-07, + "loss": 0.4843, + "step": 28388 + }, + { + "epoch": 1.9052716351800276, + "grad_norm": 1.1717411279678345, + "learning_rate": 5.859383473063918e-07, + "loss": 0.4578, + "step": 28390 + }, + { + "epoch": 1.9054058588637965, + "grad_norm": 1.0161560773849487, + "learning_rate": 5.842805324267897e-07, + "loss": 0.4392, + "step": 28392 + }, + { + "epoch": 1.9055400825475655, + "grad_norm": 0.9141351580619812, + "learning_rate": 5.826250523427557e-07, + "loss": 0.4215, + "step": 28394 + }, + { + "epoch": 1.9056743062313344, + "grad_norm": 0.9663956761360168, + "learning_rate": 5.809719071325103e-07, + "loss": 0.4045, + "step": 28396 + }, + { + "epoch": 1.9058085299151035, + "grad_norm": 1.2365949153900146, + "learning_rate": 5.793210968741691e-07, + "loss": 0.4915, + "step": 28398 + }, + { + "epoch": 1.9059427535988727, + "grad_norm": 1.044782280921936, + "learning_rate": 5.776726216457251e-07, + "loss": 0.3957, + "step": 28400 + }, + { + "epoch": 1.9060769772826416, + "grad_norm": 0.9589051604270935, + "learning_rate": 5.760264815250605e-07, + "loss": 0.4413, + "step": 28402 + }, + { + "epoch": 1.9062112009664105, + "grad_norm": 0.9788399934768677, + "learning_rate": 5.743826765899629e-07, + "loss": 0.4861, + "step": 28404 + }, + { + "epoch": 1.9063454246501794, + "grad_norm": 1.0734039545059204, + "learning_rate": 5.727412069180871e-07, + "loss": 0.4753, + "step": 28406 + }, + { + "epoch": 1.9064796483339486, + "grad_norm": 0.8587889075279236, + "learning_rate": 5.711020725869986e-07, + "loss": 0.4593, + "step": 28408 + }, + { + "epoch": 1.9066138720177175, + "grad_norm": 1.0568461418151855, + "learning_rate": 5.694652736741357e-07, + "loss": 0.5035, + "step": 28410 + }, + { + "epoch": 1.9067480957014866, + "grad_norm": 1.0681660175323486, + "learning_rate": 5.678308102568364e-07, + "loss": 0.463, + "step": 28412 + }, + { + "epoch": 1.9068823193852555, + "grad_norm": 1.0138444900512695, + "learning_rate": 5.661986824123278e-07, + "loss": 0.5085, + "step": 28414 + }, + { + "epoch": 1.9070165430690245, + "grad_norm": 1.0412541627883911, + "learning_rate": 5.645688902177315e-07, + "loss": 0.5018, + "step": 28416 + }, + { + "epoch": 1.9071507667527934, + "grad_norm": 1.213753342628479, + "learning_rate": 5.62941433750036e-07, + "loss": 0.5085, + "step": 28418 + }, + { + "epoch": 1.9072849904365625, + "grad_norm": 1.0579547882080078, + "learning_rate": 5.613163130861521e-07, + "loss": 0.4646, + "step": 28420 + }, + { + "epoch": 1.9074192141203317, + "grad_norm": 1.0913277864456177, + "learning_rate": 5.59693528302846e-07, + "loss": 0.4481, + "step": 28422 + }, + { + "epoch": 1.9075534378041006, + "grad_norm": 0.9536145329475403, + "learning_rate": 5.580730794768064e-07, + "loss": 0.4465, + "step": 28424 + }, + { + "epoch": 1.9076876614878695, + "grad_norm": 0.9059435725212097, + "learning_rate": 5.564549666845886e-07, + "loss": 0.5258, + "step": 28426 + }, + { + "epoch": 1.9078218851716384, + "grad_norm": 1.2305552959442139, + "learning_rate": 5.548391900026484e-07, + "loss": 0.5643, + "step": 28428 + }, + { + "epoch": 1.9079561088554076, + "grad_norm": 0.9224756956100464, + "learning_rate": 5.532257495073245e-07, + "loss": 0.4555, + "step": 28430 + }, + { + "epoch": 1.9080903325391767, + "grad_norm": 1.0110727548599243, + "learning_rate": 5.516146452748506e-07, + "loss": 0.4943, + "step": 28432 + }, + { + "epoch": 1.9082245562229456, + "grad_norm": 1.033556342124939, + "learning_rate": 5.500058773813543e-07, + "loss": 0.5134, + "step": 28434 + }, + { + "epoch": 1.9083587799067145, + "grad_norm": 1.0685513019561768, + "learning_rate": 5.483994459028363e-07, + "loss": 0.4744, + "step": 28436 + }, + { + "epoch": 1.9084930035904835, + "grad_norm": 1.018219232559204, + "learning_rate": 5.467953509152024e-07, + "loss": 0.5194, + "step": 28438 + }, + { + "epoch": 1.9086272272742524, + "grad_norm": 0.8616458773612976, + "learning_rate": 5.45193592494242e-07, + "loss": 0.461, + "step": 28440 + }, + { + "epoch": 1.9087614509580215, + "grad_norm": 1.039685845375061, + "learning_rate": 5.435941707156389e-07, + "loss": 0.4196, + "step": 28442 + }, + { + "epoch": 1.9088956746417907, + "grad_norm": 1.1347438097000122, + "learning_rate": 5.41997085654955e-07, + "loss": 0.4563, + "step": 28444 + }, + { + "epoch": 1.9090298983255596, + "grad_norm": 1.1098366975784302, + "learning_rate": 5.404023373876521e-07, + "loss": 0.4252, + "step": 28446 + }, + { + "epoch": 1.9091641220093285, + "grad_norm": 1.00871741771698, + "learning_rate": 5.388099259890867e-07, + "loss": 0.4486, + "step": 28448 + }, + { + "epoch": 1.9092983456930974, + "grad_norm": 0.9640150666236877, + "learning_rate": 5.372198515344929e-07, + "loss": 0.4002, + "step": 28450 + }, + { + "epoch": 1.9094325693768666, + "grad_norm": 1.0901308059692383, + "learning_rate": 5.356321140989884e-07, + "loss": 0.4761, + "step": 28452 + }, + { + "epoch": 1.9095667930606357, + "grad_norm": 1.0453557968139648, + "learning_rate": 5.340467137576022e-07, + "loss": 0.4536, + "step": 28454 + }, + { + "epoch": 1.9097010167444046, + "grad_norm": 1.0754693746566772, + "learning_rate": 5.32463650585241e-07, + "loss": 0.424, + "step": 28456 + }, + { + "epoch": 1.9098352404281735, + "grad_norm": 1.1840853691101074, + "learning_rate": 5.308829246567004e-07, + "loss": 0.4866, + "step": 28458 + }, + { + "epoch": 1.9099694641119425, + "grad_norm": 0.9412599802017212, + "learning_rate": 5.293045360466541e-07, + "loss": 0.3965, + "step": 28460 + }, + { + "epoch": 1.9101036877957116, + "grad_norm": 1.0066570043563843, + "learning_rate": 5.277284848296981e-07, + "loss": 0.4325, + "step": 28462 + }, + { + "epoch": 1.9102379114794805, + "grad_norm": 0.9843421578407288, + "learning_rate": 5.261547710802894e-07, + "loss": 0.4693, + "step": 28464 + }, + { + "epoch": 1.9103721351632497, + "grad_norm": 1.1081351041793823, + "learning_rate": 5.245833948727741e-07, + "loss": 0.4465, + "step": 28466 + }, + { + "epoch": 1.9105063588470186, + "grad_norm": 1.0450243949890137, + "learning_rate": 5.230143562814093e-07, + "loss": 0.5147, + "step": 28468 + }, + { + "epoch": 1.9106405825307875, + "grad_norm": 0.8684070706367493, + "learning_rate": 5.214476553803193e-07, + "loss": 0.4537, + "step": 28470 + }, + { + "epoch": 1.9107748062145564, + "grad_norm": 1.0927025079727173, + "learning_rate": 5.198832922435337e-07, + "loss": 0.4677, + "step": 28472 + }, + { + "epoch": 1.9109090298983256, + "grad_norm": 0.9476495385169983, + "learning_rate": 5.183212669449656e-07, + "loss": 0.3967, + "step": 28474 + }, + { + "epoch": 1.9110432535820947, + "grad_norm": 0.9752746224403381, + "learning_rate": 5.167615795584169e-07, + "loss": 0.482, + "step": 28476 + }, + { + "epoch": 1.9111774772658636, + "grad_norm": 1.0357590913772583, + "learning_rate": 5.152042301575788e-07, + "loss": 0.4535, + "step": 28478 + }, + { + "epoch": 1.9113117009496325, + "grad_norm": 1.0198721885681152, + "learning_rate": 5.136492188160313e-07, + "loss": 0.4679, + "step": 28480 + }, + { + "epoch": 1.9114459246334015, + "grad_norm": 0.9641233682632446, + "learning_rate": 5.12096545607249e-07, + "loss": 0.4682, + "step": 28482 + }, + { + "epoch": 1.9115801483171706, + "grad_norm": 0.9013912081718445, + "learning_rate": 5.105462106045955e-07, + "loss": 0.4329, + "step": 28484 + }, + { + "epoch": 1.9117143720009395, + "grad_norm": 1.004481315612793, + "learning_rate": 5.08998213881312e-07, + "loss": 0.4515, + "step": 28486 + }, + { + "epoch": 1.9118485956847087, + "grad_norm": 1.0103305578231812, + "learning_rate": 5.074525555105403e-07, + "loss": 0.526, + "step": 28488 + }, + { + "epoch": 1.9119828193684776, + "grad_norm": 1.1251330375671387, + "learning_rate": 5.059092355653161e-07, + "loss": 0.452, + "step": 28490 + }, + { + "epoch": 1.9121170430522465, + "grad_norm": 1.0311959981918335, + "learning_rate": 5.043682541185479e-07, + "loss": 0.48, + "step": 28492 + }, + { + "epoch": 1.9122512667360154, + "grad_norm": 1.04188871383667, + "learning_rate": 5.028296112430608e-07, + "loss": 0.4371, + "step": 28494 + }, + { + "epoch": 1.9123854904197846, + "grad_norm": 0.9846208095550537, + "learning_rate": 5.012933070115411e-07, + "loss": 0.4802, + "step": 28496 + }, + { + "epoch": 1.9125197141035537, + "grad_norm": 1.1321674585342407, + "learning_rate": 4.997593414965751e-07, + "loss": 0.5022, + "step": 28498 + }, + { + "epoch": 1.9126539377873226, + "grad_norm": 0.8902405500411987, + "learning_rate": 4.982277147706382e-07, + "loss": 0.426, + "step": 28500 + }, + { + "epoch": 1.9127881614710915, + "grad_norm": 1.275682806968689, + "learning_rate": 4.966984269061059e-07, + "loss": 0.5051, + "step": 28502 + }, + { + "epoch": 1.9129223851548605, + "grad_norm": 1.003994345664978, + "learning_rate": 4.951714779752314e-07, + "loss": 0.4934, + "step": 28504 + }, + { + "epoch": 1.9130566088386296, + "grad_norm": 1.0241422653198242, + "learning_rate": 4.93646868050146e-07, + "loss": 0.4803, + "step": 28506 + }, + { + "epoch": 1.9131908325223987, + "grad_norm": 0.9359074234962463, + "learning_rate": 4.921245972029087e-07, + "loss": 0.4462, + "step": 28508 + }, + { + "epoch": 1.9133250562061677, + "grad_norm": 1.139502763748169, + "learning_rate": 4.906046655054231e-07, + "loss": 0.4593, + "step": 28510 + }, + { + "epoch": 1.9134592798899366, + "grad_norm": 1.130993366241455, + "learning_rate": 4.89087073029515e-07, + "loss": 0.4675, + "step": 28512 + }, + { + "epoch": 1.9135935035737055, + "grad_norm": 1.132150411605835, + "learning_rate": 4.875718198468827e-07, + "loss": 0.4903, + "step": 28514 + }, + { + "epoch": 1.9137277272574744, + "grad_norm": 0.9900282025337219, + "learning_rate": 4.860589060291188e-07, + "loss": 0.5699, + "step": 28516 + }, + { + "epoch": 1.9138619509412436, + "grad_norm": 1.066292405128479, + "learning_rate": 4.845483316477051e-07, + "loss": 0.4572, + "step": 28518 + }, + { + "epoch": 1.9139961746250127, + "grad_norm": 0.9880896210670471, + "learning_rate": 4.830400967740178e-07, + "loss": 0.4137, + "step": 28520 + }, + { + "epoch": 1.9141303983087816, + "grad_norm": 1.0525450706481934, + "learning_rate": 4.815342014793167e-07, + "loss": 0.4943, + "step": 28522 + }, + { + "epoch": 1.9142646219925505, + "grad_norm": 1.0320472717285156, + "learning_rate": 4.800306458347448e-07, + "loss": 0.4619, + "step": 28524 + }, + { + "epoch": 1.9143988456763195, + "grad_norm": 1.0434107780456543, + "learning_rate": 4.785294299113508e-07, + "loss": 0.4554, + "step": 28526 + }, + { + "epoch": 1.9145330693600886, + "grad_norm": 0.9134626984596252, + "learning_rate": 4.770305537800613e-07, + "loss": 0.4422, + "step": 28528 + }, + { + "epoch": 1.9146672930438577, + "grad_norm": 1.0702944993972778, + "learning_rate": 4.7553401751169735e-07, + "loss": 0.4806, + "step": 28530 + }, + { + "epoch": 1.9148015167276267, + "grad_norm": 0.8887928128242493, + "learning_rate": 4.7403982117696923e-07, + "loss": 0.4599, + "step": 28532 + }, + { + "epoch": 1.9149357404113956, + "grad_norm": 0.9969332814216614, + "learning_rate": 4.7254796484645925e-07, + "loss": 0.4259, + "step": 28534 + }, + { + "epoch": 1.9150699640951645, + "grad_norm": 0.9118576645851135, + "learning_rate": 4.710584485906777e-07, + "loss": 0.4251, + "step": 28536 + }, + { + "epoch": 1.9152041877789336, + "grad_norm": 0.8886921405792236, + "learning_rate": 4.695712724799795e-07, + "loss": 0.4508, + "step": 28538 + }, + { + "epoch": 1.9153384114627026, + "grad_norm": 1.0292127132415771, + "learning_rate": 4.680864365846471e-07, + "loss": 0.4989, + "step": 28540 + }, + { + "epoch": 1.9154726351464717, + "grad_norm": 0.9974724650382996, + "learning_rate": 4.666039409748357e-07, + "loss": 0.4317, + "step": 28542 + }, + { + "epoch": 1.9156068588302406, + "grad_norm": 1.3106881380081177, + "learning_rate": 4.651237857205781e-07, + "loss": 0.4241, + "step": 28544 + }, + { + "epoch": 1.9157410825140095, + "grad_norm": 1.1389553546905518, + "learning_rate": 4.636459708918128e-07, + "loss": 0.508, + "step": 28546 + }, + { + "epoch": 1.9158753061977785, + "grad_norm": 0.8844836950302124, + "learning_rate": 4.6217049655837287e-07, + "loss": 0.4577, + "step": 28548 + }, + { + "epoch": 1.9160095298815476, + "grad_norm": 1.1401379108428955, + "learning_rate": 4.606973627899636e-07, + "loss": 0.5309, + "step": 28550 + }, + { + "epoch": 1.9161437535653167, + "grad_norm": 1.0861908197402954, + "learning_rate": 4.5922656965618484e-07, + "loss": 0.4301, + "step": 28552 + }, + { + "epoch": 1.9162779772490857, + "grad_norm": 1.0897116661071777, + "learning_rate": 4.577581172265366e-07, + "loss": 0.4486, + "step": 28554 + }, + { + "epoch": 1.9164122009328546, + "grad_norm": 1.0773807764053345, + "learning_rate": 4.5629200557039674e-07, + "loss": 0.5119, + "step": 28556 + }, + { + "epoch": 1.9165464246166235, + "grad_norm": 0.956876814365387, + "learning_rate": 4.5482823475703205e-07, + "loss": 0.5197, + "step": 28558 + }, + { + "epoch": 1.9166806483003926, + "grad_norm": 1.1136553287506104, + "learning_rate": 4.53366804855615e-07, + "loss": 0.453, + "step": 28560 + }, + { + "epoch": 1.9168148719841616, + "grad_norm": 0.9848942756652832, + "learning_rate": 4.519077159351792e-07, + "loss": 0.4496, + "step": 28562 + }, + { + "epoch": 1.9169490956679307, + "grad_norm": 1.036436915397644, + "learning_rate": 4.504509680646751e-07, + "loss": 0.4695, + "step": 28564 + }, + { + "epoch": 1.9170833193516996, + "grad_norm": 1.1097133159637451, + "learning_rate": 4.48996561312931e-07, + "loss": 0.4585, + "step": 28566 + }, + { + "epoch": 1.9172175430354685, + "grad_norm": 1.0167475938796997, + "learning_rate": 4.47544495748653e-07, + "loss": 0.4732, + "step": 28568 + }, + { + "epoch": 1.9173517667192375, + "grad_norm": 0.9866967797279358, + "learning_rate": 4.4609477144046395e-07, + "loss": 0.4768, + "step": 28570 + }, + { + "epoch": 1.9174859904030066, + "grad_norm": 5.733170032501221, + "learning_rate": 4.4464738845685363e-07, + "loss": 0.4874, + "step": 28572 + }, + { + "epoch": 1.9176202140867757, + "grad_norm": 1.0686026811599731, + "learning_rate": 4.432023468662061e-07, + "loss": 0.5011, + "step": 28574 + }, + { + "epoch": 1.9177544377705447, + "grad_norm": 0.9280849099159241, + "learning_rate": 4.417596467367946e-07, + "loss": 0.4321, + "step": 28576 + }, + { + "epoch": 1.9178886614543136, + "grad_norm": 1.0819580554962158, + "learning_rate": 4.4031928813679235e-07, + "loss": 0.4728, + "step": 28578 + }, + { + "epoch": 1.9180228851380825, + "grad_norm": 0.971687376499176, + "learning_rate": 4.3888127113424496e-07, + "loss": 0.4933, + "step": 28580 + }, + { + "epoch": 1.9181571088218516, + "grad_norm": 1.021970272064209, + "learning_rate": 4.3744559579710353e-07, + "loss": 0.4983, + "step": 28582 + }, + { + "epoch": 1.9182913325056208, + "grad_norm": 1.0264430046081543, + "learning_rate": 4.360122621931917e-07, + "loss": 0.4828, + "step": 28584 + }, + { + "epoch": 1.9184255561893897, + "grad_norm": 0.9958664178848267, + "learning_rate": 4.345812703902441e-07, + "loss": 0.4592, + "step": 28586 + }, + { + "epoch": 1.9185597798731586, + "grad_norm": 0.9951131343841553, + "learning_rate": 4.3315262045586224e-07, + "loss": 0.4928, + "step": 28588 + }, + { + "epoch": 1.9186940035569275, + "grad_norm": 0.8685447573661804, + "learning_rate": 4.317263124575477e-07, + "loss": 0.4654, + "step": 28590 + }, + { + "epoch": 1.9188282272406965, + "grad_norm": 0.9491552710533142, + "learning_rate": 4.3030234646269653e-07, + "loss": 0.4621, + "step": 28592 + }, + { + "epoch": 1.9189624509244656, + "grad_norm": 1.1487892866134644, + "learning_rate": 4.288807225385827e-07, + "loss": 0.4918, + "step": 28594 + }, + { + "epoch": 1.9190966746082347, + "grad_norm": 1.0633422136306763, + "learning_rate": 4.274614407523747e-07, + "loss": 0.4567, + "step": 28596 + }, + { + "epoch": 1.9192308982920037, + "grad_norm": 1.007346272468567, + "learning_rate": 4.2604450117114114e-07, + "loss": 0.4498, + "step": 28598 + }, + { + "epoch": 1.9193651219757726, + "grad_norm": 0.9971712231636047, + "learning_rate": 4.2462990386181735e-07, + "loss": 0.4672, + "step": 28600 + }, + { + "epoch": 1.9194993456595415, + "grad_norm": 1.153144121170044, + "learning_rate": 4.2321764889124425e-07, + "loss": 0.46, + "step": 28602 + }, + { + "epoch": 1.9196335693433106, + "grad_norm": 1.036102294921875, + "learning_rate": 4.2180773632614637e-07, + "loss": 0.4401, + "step": 28604 + }, + { + "epoch": 1.9197677930270798, + "grad_norm": 1.0317116975784302, + "learning_rate": 4.2040016623314804e-07, + "loss": 0.4512, + "step": 28606 + }, + { + "epoch": 1.9199020167108487, + "grad_norm": 0.9724912643432617, + "learning_rate": 4.189949386787462e-07, + "loss": 0.4622, + "step": 28608 + }, + { + "epoch": 1.9200362403946176, + "grad_norm": 0.9077244997024536, + "learning_rate": 4.1759205372933206e-07, + "loss": 0.4768, + "step": 28610 + }, + { + "epoch": 1.9201704640783865, + "grad_norm": 1.0253747701644897, + "learning_rate": 4.161915114511972e-07, + "loss": 0.4779, + "step": 28612 + }, + { + "epoch": 1.9203046877621557, + "grad_norm": 1.1545828580856323, + "learning_rate": 4.147933119105107e-07, + "loss": 0.4816, + "step": 28614 + }, + { + "epoch": 1.9204389114459246, + "grad_norm": 0.9470486640930176, + "learning_rate": 4.133974551733366e-07, + "loss": 0.4111, + "step": 28616 + }, + { + "epoch": 1.9205731351296937, + "grad_norm": 0.9704294800758362, + "learning_rate": 4.12003941305622e-07, + "loss": 0.5076, + "step": 28618 + }, + { + "epoch": 1.9207073588134627, + "grad_norm": 0.9441697597503662, + "learning_rate": 4.106127703732088e-07, + "loss": 0.412, + "step": 28620 + }, + { + "epoch": 1.9208415824972316, + "grad_norm": 1.0234549045562744, + "learning_rate": 4.0922394244183315e-07, + "loss": 0.4747, + "step": 28622 + }, + { + "epoch": 1.9209758061810005, + "grad_norm": 1.0847262144088745, + "learning_rate": 4.0783745757710935e-07, + "loss": 0.5063, + "step": 28624 + }, + { + "epoch": 1.9211100298647696, + "grad_norm": 1.040287733078003, + "learning_rate": 4.0645331584454606e-07, + "loss": 0.506, + "step": 28626 + }, + { + "epoch": 1.9212442535485388, + "grad_norm": 1.115692377090454, + "learning_rate": 4.0507151730954095e-07, + "loss": 0.487, + "step": 28628 + }, + { + "epoch": 1.9213784772323077, + "grad_norm": 1.7195245027542114, + "learning_rate": 4.036920620373863e-07, + "loss": 0.4531, + "step": 28630 + }, + { + "epoch": 1.9215127009160766, + "grad_norm": 0.9717667102813721, + "learning_rate": 4.0231495009325215e-07, + "loss": 0.434, + "step": 28632 + }, + { + "epoch": 1.9216469245998455, + "grad_norm": 0.8491559624671936, + "learning_rate": 4.0094018154220316e-07, + "loss": 0.4655, + "step": 28634 + }, + { + "epoch": 1.9217811482836147, + "grad_norm": 1.1206490993499756, + "learning_rate": 3.9956775644920395e-07, + "loss": 0.4388, + "step": 28636 + }, + { + "epoch": 1.9219153719673836, + "grad_norm": 0.9006564021110535, + "learning_rate": 3.9819767487909165e-07, + "loss": 0.426, + "step": 28638 + }, + { + "epoch": 1.9220495956511527, + "grad_norm": 0.9929448366165161, + "learning_rate": 3.968299368966033e-07, + "loss": 0.4574, + "step": 28640 + }, + { + "epoch": 1.9221838193349217, + "grad_norm": 1.083114743232727, + "learning_rate": 3.954645425663539e-07, + "loss": 0.4758, + "step": 28642 + }, + { + "epoch": 1.9223180430186906, + "grad_norm": 1.041306972503662, + "learning_rate": 3.9410149195286963e-07, + "loss": 0.4512, + "step": 28644 + }, + { + "epoch": 1.9224522667024595, + "grad_norm": 1.2051905393600464, + "learning_rate": 3.927407851205378e-07, + "loss": 0.4464, + "step": 28646 + }, + { + "epoch": 1.9225864903862286, + "grad_norm": 1.0940951108932495, + "learning_rate": 3.9138242213365703e-07, + "loss": 0.4367, + "step": 28648 + }, + { + "epoch": 1.9227207140699978, + "grad_norm": 1.0143513679504395, + "learning_rate": 3.900264030564038e-07, + "loss": 0.4202, + "step": 28650 + }, + { + "epoch": 1.9228549377537667, + "grad_norm": 1.453702449798584, + "learning_rate": 3.8867272795285456e-07, + "loss": 0.4457, + "step": 28652 + }, + { + "epoch": 1.9229891614375356, + "grad_norm": 1.0661327838897705, + "learning_rate": 3.8732139688695825e-07, + "loss": 0.4789, + "step": 28654 + }, + { + "epoch": 1.9231233851213045, + "grad_norm": 0.9508505463600159, + "learning_rate": 3.8597240992256924e-07, + "loss": 0.3891, + "step": 28656 + }, + { + "epoch": 1.9232576088050737, + "grad_norm": 1.0582705736160278, + "learning_rate": 3.846257671234199e-07, + "loss": 0.5021, + "step": 28658 + }, + { + "epoch": 1.9233918324888426, + "grad_norm": 0.9684615731239319, + "learning_rate": 3.8328146855314275e-07, + "loss": 0.4531, + "step": 28660 + }, + { + "epoch": 1.9235260561726117, + "grad_norm": 1.162911057472229, + "learning_rate": 3.8193951427524243e-07, + "loss": 0.4367, + "step": 28662 + }, + { + "epoch": 1.9236602798563807, + "grad_norm": 1.106896996498108, + "learning_rate": 3.8059990435313495e-07, + "loss": 0.4696, + "step": 28664 + }, + { + "epoch": 1.9237945035401496, + "grad_norm": 1.1426647901535034, + "learning_rate": 3.792626388501086e-07, + "loss": 0.5195, + "step": 28666 + }, + { + "epoch": 1.9239287272239185, + "grad_norm": 1.0694050788879395, + "learning_rate": 3.7792771782934613e-07, + "loss": 0.4896, + "step": 28668 + }, + { + "epoch": 1.9240629509076876, + "grad_norm": 0.9436160326004028, + "learning_rate": 3.7659514135391924e-07, + "loss": 0.4324, + "step": 28670 + }, + { + "epoch": 1.9241971745914568, + "grad_norm": 1.0565946102142334, + "learning_rate": 3.7526490948679995e-07, + "loss": 0.449, + "step": 28672 + }, + { + "epoch": 1.9243313982752257, + "grad_norm": 1.1161247491836548, + "learning_rate": 3.739370222908267e-07, + "loss": 0.4936, + "step": 28674 + }, + { + "epoch": 1.9244656219589946, + "grad_norm": 1.0940635204315186, + "learning_rate": 3.726114798287439e-07, + "loss": 0.4694, + "step": 28676 + }, + { + "epoch": 1.9245998456427635, + "grad_norm": 1.0296701192855835, + "learning_rate": 3.712882821631736e-07, + "loss": 0.4453, + "step": 28678 + }, + { + "epoch": 1.9247340693265327, + "grad_norm": 1.1180201768875122, + "learning_rate": 3.6996742935664906e-07, + "loss": 0.4635, + "step": 28680 + }, + { + "epoch": 1.9248682930103018, + "grad_norm": 1.0563124418258667, + "learning_rate": 3.6864892147156496e-07, + "loss": 0.3876, + "step": 28682 + }, + { + "epoch": 1.9250025166940707, + "grad_norm": 0.8866017460823059, + "learning_rate": 3.673327585702324e-07, + "loss": 0.4398, + "step": 28684 + }, + { + "epoch": 1.9251367403778397, + "grad_norm": 1.0113062858581543, + "learning_rate": 3.6601894071482403e-07, + "loss": 0.4645, + "step": 28686 + }, + { + "epoch": 1.9252709640616086, + "grad_norm": 1.0368983745574951, + "learning_rate": 3.6470746796741783e-07, + "loss": 0.5158, + "step": 28688 + }, + { + "epoch": 1.9254051877453777, + "grad_norm": 1.0530612468719482, + "learning_rate": 3.6339834038997545e-07, + "loss": 0.5536, + "step": 28690 + }, + { + "epoch": 1.9255394114291466, + "grad_norm": 1.2044345140457153, + "learning_rate": 3.62091558044364e-07, + "loss": 0.5074, + "step": 28692 + }, + { + "epoch": 1.9256736351129158, + "grad_norm": 1.0812793970108032, + "learning_rate": 3.607871209923175e-07, + "loss": 0.4488, + "step": 28694 + }, + { + "epoch": 1.9258078587966847, + "grad_norm": 0.8661904335021973, + "learning_rate": 3.5948502929546433e-07, + "loss": 0.4305, + "step": 28696 + }, + { + "epoch": 1.9259420824804536, + "grad_norm": 1.0397839546203613, + "learning_rate": 3.581852830153276e-07, + "loss": 0.4291, + "step": 28698 + }, + { + "epoch": 1.9260763061642225, + "grad_norm": 1.0755635499954224, + "learning_rate": 3.5688788221332483e-07, + "loss": 0.4847, + "step": 28700 + }, + { + "epoch": 1.9262105298479917, + "grad_norm": 1.0838686227798462, + "learning_rate": 3.555928269507458e-07, + "loss": 0.4305, + "step": 28702 + }, + { + "epoch": 1.9263447535317608, + "grad_norm": 1.0830137729644775, + "learning_rate": 3.5430011728879164e-07, + "loss": 0.4864, + "step": 28704 + }, + { + "epoch": 1.9264789772155297, + "grad_norm": 0.9551955461502075, + "learning_rate": 3.5300975328853014e-07, + "loss": 0.4718, + "step": 28706 + }, + { + "epoch": 1.9266132008992987, + "grad_norm": 1.0307772159576416, + "learning_rate": 3.517217350109236e-07, + "loss": 0.5096, + "step": 28708 + }, + { + "epoch": 1.9267474245830676, + "grad_norm": 1.031486988067627, + "learning_rate": 3.504360625168457e-07, + "loss": 0.5634, + "step": 28710 + }, + { + "epoch": 1.9268816482668367, + "grad_norm": 1.1116483211517334, + "learning_rate": 3.4915273586702546e-07, + "loss": 0.523, + "step": 28712 + }, + { + "epoch": 1.9270158719506056, + "grad_norm": 0.9586636424064636, + "learning_rate": 3.478717551221089e-07, + "loss": 0.4818, + "step": 28714 + }, + { + "epoch": 1.9271500956343748, + "grad_norm": 1.0831985473632812, + "learning_rate": 3.465931203426087e-07, + "loss": 0.4314, + "step": 28716 + }, + { + "epoch": 1.9272843193181437, + "grad_norm": 0.9724242687225342, + "learning_rate": 3.453168315889488e-07, + "loss": 0.4643, + "step": 28718 + }, + { + "epoch": 1.9274185430019126, + "grad_norm": 1.8084481954574585, + "learning_rate": 3.440428889214253e-07, + "loss": 0.4437, + "step": 28720 + }, + { + "epoch": 1.9275527666856815, + "grad_norm": 1.1648857593536377, + "learning_rate": 3.427712924002402e-07, + "loss": 0.5064, + "step": 28722 + }, + { + "epoch": 1.9276869903694507, + "grad_norm": 1.189074158668518, + "learning_rate": 3.415020420854509e-07, + "loss": 0.4949, + "step": 28724 + }, + { + "epoch": 1.9278212140532198, + "grad_norm": 1.1187207698822021, + "learning_rate": 3.402351380370483e-07, + "loss": 0.4282, + "step": 28726 + }, + { + "epoch": 1.9279554377369887, + "grad_norm": 0.9836530089378357, + "learning_rate": 3.3897058031487906e-07, + "loss": 0.4768, + "step": 28728 + }, + { + "epoch": 1.9280896614207577, + "grad_norm": 1.033627986907959, + "learning_rate": 3.3770836897870086e-07, + "loss": 0.4541, + "step": 28730 + }, + { + "epoch": 1.9282238851045266, + "grad_norm": 1.1201447248458862, + "learning_rate": 3.364485040881438e-07, + "loss": 0.4658, + "step": 28732 + }, + { + "epoch": 1.9283581087882957, + "grad_norm": 1.0621964931488037, + "learning_rate": 3.3519098570273797e-07, + "loss": 0.4132, + "step": 28734 + }, + { + "epoch": 1.9284923324720646, + "grad_norm": 1.0216847658157349, + "learning_rate": 3.3393581388189157e-07, + "loss": 0.4506, + "step": 28736 + }, + { + "epoch": 1.9286265561558338, + "grad_norm": 1.0879138708114624, + "learning_rate": 3.326829886849181e-07, + "loss": 0.5463, + "step": 28738 + }, + { + "epoch": 1.9287607798396027, + "grad_norm": 1.0055351257324219, + "learning_rate": 3.314325101710036e-07, + "loss": 0.4407, + "step": 28740 + }, + { + "epoch": 1.9288950035233716, + "grad_norm": 1.1964483261108398, + "learning_rate": 3.301843783992398e-07, + "loss": 0.4547, + "step": 28742 + }, + { + "epoch": 1.9290292272071405, + "grad_norm": 0.9628129601478577, + "learning_rate": 3.289385934285849e-07, + "loss": 0.4544, + "step": 28744 + }, + { + "epoch": 1.9291634508909097, + "grad_norm": 1.0963078737258911, + "learning_rate": 3.276951553179086e-07, + "loss": 0.4807, + "step": 28746 + }, + { + "epoch": 1.9292976745746788, + "grad_norm": 0.8719479441642761, + "learning_rate": 3.264540641259639e-07, + "loss": 0.4312, + "step": 28748 + }, + { + "epoch": 1.9294318982584477, + "grad_norm": 1.026655912399292, + "learning_rate": 3.252153199113872e-07, + "loss": 0.5329, + "step": 28750 + }, + { + "epoch": 1.9295661219422167, + "grad_norm": 0.8928234577178955, + "learning_rate": 3.2397892273269835e-07, + "loss": 0.3986, + "step": 28752 + }, + { + "epoch": 1.9297003456259856, + "grad_norm": 1.0384950637817383, + "learning_rate": 3.227448726483284e-07, + "loss": 0.4794, + "step": 28754 + }, + { + "epoch": 1.9298345693097547, + "grad_norm": 0.9321753978729248, + "learning_rate": 3.2151316971656963e-07, + "loss": 0.4443, + "step": 28756 + }, + { + "epoch": 1.9299687929935239, + "grad_norm": 1.0566009283065796, + "learning_rate": 3.2028381399563103e-07, + "loss": 0.4563, + "step": 28758 + }, + { + "epoch": 1.9301030166772928, + "grad_norm": 0.9426655173301697, + "learning_rate": 3.190568055435883e-07, + "loss": 0.4385, + "step": 28760 + }, + { + "epoch": 1.9302372403610617, + "grad_norm": 1.0611015558242798, + "learning_rate": 3.178321444184229e-07, + "loss": 0.4217, + "step": 28762 + }, + { + "epoch": 1.9303714640448306, + "grad_norm": 0.8264487981796265, + "learning_rate": 3.1660983067798856e-07, + "loss": 0.4225, + "step": 28764 + }, + { + "epoch": 1.9305056877285998, + "grad_norm": 1.009493350982666, + "learning_rate": 3.15389864380039e-07, + "loss": 0.4217, + "step": 28766 + }, + { + "epoch": 1.9306399114123687, + "grad_norm": 1.0614426136016846, + "learning_rate": 3.141722455822227e-07, + "loss": 0.4555, + "step": 28768 + }, + { + "epoch": 1.9307741350961378, + "grad_norm": 1.0587729215621948, + "learning_rate": 3.1295697434206573e-07, + "loss": 0.4292, + "step": 28770 + }, + { + "epoch": 1.9309083587799067, + "grad_norm": 1.0446990728378296, + "learning_rate": 3.117440507169833e-07, + "loss": 0.4865, + "step": 28772 + }, + { + "epoch": 1.9310425824636757, + "grad_norm": 1.14275062084198, + "learning_rate": 3.105334747642852e-07, + "loss": 0.4542, + "step": 28774 + }, + { + "epoch": 1.9311768061474446, + "grad_norm": 1.0680385828018188, + "learning_rate": 3.093252465411756e-07, + "loss": 0.4283, + "step": 28776 + }, + { + "epoch": 1.9313110298312137, + "grad_norm": 1.2552454471588135, + "learning_rate": 3.0811936610473103e-07, + "loss": 0.4425, + "step": 28778 + }, + { + "epoch": 1.9314452535149829, + "grad_norm": 0.8596842288970947, + "learning_rate": 3.0691583351193377e-07, + "loss": 0.4169, + "step": 28780 + }, + { + "epoch": 1.9315794771987518, + "grad_norm": 0.9840183258056641, + "learning_rate": 3.0571464881964385e-07, + "loss": 0.4169, + "step": 28782 + }, + { + "epoch": 1.9317137008825207, + "grad_norm": 0.9783838987350464, + "learning_rate": 3.0451581208462143e-07, + "loss": 0.4212, + "step": 28784 + }, + { + "epoch": 1.9318479245662896, + "grad_norm": 0.914157509803772, + "learning_rate": 3.0331932336349897e-07, + "loss": 0.4042, + "step": 28786 + }, + { + "epoch": 1.9319821482500588, + "grad_norm": 1.0035114288330078, + "learning_rate": 3.0212518271281466e-07, + "loss": 0.4492, + "step": 28788 + }, + { + "epoch": 1.9321163719338277, + "grad_norm": 1.0145611763000488, + "learning_rate": 3.0093339018899544e-07, + "loss": 0.4848, + "step": 28790 + }, + { + "epoch": 1.9322505956175968, + "grad_norm": 0.8387595415115356, + "learning_rate": 2.9974394584834085e-07, + "loss": 0.4025, + "step": 28792 + }, + { + "epoch": 1.9323848193013657, + "grad_norm": 0.9812015295028687, + "learning_rate": 2.9855684974705034e-07, + "loss": 0.4563, + "step": 28794 + }, + { + "epoch": 1.9325190429851347, + "grad_norm": 1.0451850891113281, + "learning_rate": 2.973721019412179e-07, + "loss": 0.4825, + "step": 28796 + }, + { + "epoch": 1.9326532666689036, + "grad_norm": 1.0891306400299072, + "learning_rate": 2.96189702486821e-07, + "loss": 0.464, + "step": 28798 + }, + { + "epoch": 1.9327874903526727, + "grad_norm": 0.9749390482902527, + "learning_rate": 2.950096514397149e-07, + "loss": 0.4394, + "step": 28800 + }, + { + "epoch": 1.9329217140364419, + "grad_norm": 0.9677867889404297, + "learning_rate": 2.9383194885566623e-07, + "loss": 0.4786, + "step": 28802 + }, + { + "epoch": 1.9330559377202108, + "grad_norm": 0.8928714394569397, + "learning_rate": 2.926565947903137e-07, + "loss": 0.4969, + "step": 28804 + }, + { + "epoch": 1.9331901614039797, + "grad_norm": 0.9890260696411133, + "learning_rate": 2.9148358929919073e-07, + "loss": 0.462, + "step": 28806 + }, + { + "epoch": 1.9333243850877486, + "grad_norm": 1.1038089990615845, + "learning_rate": 2.903129324377252e-07, + "loss": 0.4754, + "step": 28808 + }, + { + "epoch": 1.9334586087715178, + "grad_norm": 1.1482740640640259, + "learning_rate": 2.891446242612228e-07, + "loss": 0.516, + "step": 28810 + }, + { + "epoch": 1.9335928324552867, + "grad_norm": 0.973935604095459, + "learning_rate": 2.8797866482488387e-07, + "loss": 0.4626, + "step": 28812 + }, + { + "epoch": 1.9337270561390558, + "grad_norm": 0.9728585481643677, + "learning_rate": 2.868150541837922e-07, + "loss": 0.3713, + "step": 28814 + }, + { + "epoch": 1.9338612798228247, + "grad_norm": 0.8140623569488525, + "learning_rate": 2.8565379239294257e-07, + "loss": 0.463, + "step": 28816 + }, + { + "epoch": 1.9339955035065937, + "grad_norm": 0.9422245025634766, + "learning_rate": 2.844948795071856e-07, + "loss": 0.5291, + "step": 28818 + }, + { + "epoch": 1.9341297271903626, + "grad_norm": 1.8759267330169678, + "learning_rate": 2.8333831558128856e-07, + "loss": 0.3971, + "step": 28820 + }, + { + "epoch": 1.9342639508741317, + "grad_norm": 1.5151934623718262, + "learning_rate": 2.8218410066988554e-07, + "loss": 0.4601, + "step": 28822 + }, + { + "epoch": 1.9343981745579009, + "grad_norm": 1.103763461112976, + "learning_rate": 2.8103223482752183e-07, + "loss": 0.486, + "step": 28824 + }, + { + "epoch": 1.9345323982416698, + "grad_norm": 0.9701932668685913, + "learning_rate": 2.7988271810862054e-07, + "loss": 0.4239, + "step": 28826 + }, + { + "epoch": 1.9346666219254387, + "grad_norm": 0.9574766159057617, + "learning_rate": 2.787355505674882e-07, + "loss": 0.4315, + "step": 28828 + }, + { + "epoch": 1.9348008456092076, + "grad_norm": 0.9307126998901367, + "learning_rate": 2.77590732258326e-07, + "loss": 0.4835, + "step": 28830 + }, + { + "epoch": 1.9349350692929768, + "grad_norm": 0.9249545931816101, + "learning_rate": 2.7644826323522943e-07, + "loss": 0.433, + "step": 28832 + }, + { + "epoch": 1.935069292976746, + "grad_norm": 1.1061577796936035, + "learning_rate": 2.75308143552172e-07, + "loss": 0.4839, + "step": 28834 + }, + { + "epoch": 1.9352035166605148, + "grad_norm": 1.0632753372192383, + "learning_rate": 2.741703732630274e-07, + "loss": 0.4484, + "step": 28836 + }, + { + "epoch": 1.9353377403442837, + "grad_norm": 1.08527672290802, + "learning_rate": 2.7303495242155254e-07, + "loss": 0.4027, + "step": 28838 + }, + { + "epoch": 1.9354719640280527, + "grad_norm": 1.001092791557312, + "learning_rate": 2.7190188108138784e-07, + "loss": 0.4591, + "step": 28840 + }, + { + "epoch": 1.9356061877118218, + "grad_norm": 0.8313817977905273, + "learning_rate": 2.707711592960793e-07, + "loss": 0.3729, + "step": 28842 + }, + { + "epoch": 1.9357404113955907, + "grad_norm": 1.118964672088623, + "learning_rate": 2.696427871190399e-07, + "loss": 0.4507, + "step": 28844 + }, + { + "epoch": 1.9358746350793599, + "grad_norm": 1.0571320056915283, + "learning_rate": 2.6851676460359355e-07, + "loss": 0.5072, + "step": 28846 + }, + { + "epoch": 1.9360088587631288, + "grad_norm": 0.9531912803649902, + "learning_rate": 2.67393091802931e-07, + "loss": 0.469, + "step": 28848 + }, + { + "epoch": 1.9361430824468977, + "grad_norm": 1.1145519018173218, + "learning_rate": 2.6627176877015435e-07, + "loss": 0.5439, + "step": 28850 + }, + { + "epoch": 1.9362773061306666, + "grad_norm": 1.0474146604537964, + "learning_rate": 2.651527955582378e-07, + "loss": 0.5029, + "step": 28852 + }, + { + "epoch": 1.9364115298144358, + "grad_norm": 0.9402778744697571, + "learning_rate": 2.6403617222005017e-07, + "loss": 0.5068, + "step": 28854 + }, + { + "epoch": 1.936545753498205, + "grad_norm": 1.0296565294265747, + "learning_rate": 2.629218988083548e-07, + "loss": 0.4601, + "step": 28856 + }, + { + "epoch": 1.9366799771819738, + "grad_norm": 1.130409598350525, + "learning_rate": 2.6180997537579856e-07, + "loss": 0.4967, + "step": 28858 + }, + { + "epoch": 1.9368142008657427, + "grad_norm": 1.2263824939727783, + "learning_rate": 2.60700401974906e-07, + "loss": 0.5095, + "step": 28860 + }, + { + "epoch": 1.9369484245495117, + "grad_norm": 1.068340539932251, + "learning_rate": 2.595931786581185e-07, + "loss": 0.4784, + "step": 28862 + }, + { + "epoch": 1.9370826482332808, + "grad_norm": 1.0405715703964233, + "learning_rate": 2.584883054777443e-07, + "loss": 0.4375, + "step": 28864 + }, + { + "epoch": 1.9372168719170497, + "grad_norm": 1.063066005706787, + "learning_rate": 2.5738578248598044e-07, + "loss": 0.4243, + "step": 28866 + }, + { + "epoch": 1.9373510956008189, + "grad_norm": 0.9734081029891968, + "learning_rate": 2.562856097349242e-07, + "loss": 0.4596, + "step": 28868 + }, + { + "epoch": 1.9374853192845878, + "grad_norm": 1.0543913841247559, + "learning_rate": 2.551877872765562e-07, + "loss": 0.4521, + "step": 28870 + }, + { + "epoch": 1.9376195429683567, + "grad_norm": 0.9148135185241699, + "learning_rate": 2.540923151627461e-07, + "loss": 0.3845, + "step": 28872 + }, + { + "epoch": 1.9377537666521256, + "grad_norm": 0.9552580714225769, + "learning_rate": 2.52999193445258e-07, + "loss": 0.4932, + "step": 28874 + }, + { + "epoch": 1.9378879903358948, + "grad_norm": 1.0944716930389404, + "learning_rate": 2.5190842217573396e-07, + "loss": 0.4781, + "step": 28876 + }, + { + "epoch": 1.938022214019664, + "grad_norm": 0.974706768989563, + "learning_rate": 2.5082000140570493e-07, + "loss": 0.4606, + "step": 28878 + }, + { + "epoch": 1.9381564377034328, + "grad_norm": 1.061537742614746, + "learning_rate": 2.4973393118660757e-07, + "loss": 0.4437, + "step": 28880 + }, + { + "epoch": 1.9382906613872017, + "grad_norm": 1.0034071207046509, + "learning_rate": 2.4865021156975085e-07, + "loss": 0.4057, + "step": 28882 + }, + { + "epoch": 1.9384248850709707, + "grad_norm": 1.1839606761932373, + "learning_rate": 2.4756884260634384e-07, + "loss": 0.498, + "step": 28884 + }, + { + "epoch": 1.9385591087547398, + "grad_norm": 0.9004724621772766, + "learning_rate": 2.464898243474734e-07, + "loss": 0.4202, + "step": 28886 + }, + { + "epoch": 1.9386933324385087, + "grad_norm": 1.0127958059310913, + "learning_rate": 2.454131568441154e-07, + "loss": 0.4522, + "step": 28888 + }, + { + "epoch": 1.9388275561222779, + "grad_norm": 0.9200853705406189, + "learning_rate": 2.443388401471569e-07, + "loss": 0.4911, + "step": 28890 + }, + { + "epoch": 1.9389617798060468, + "grad_norm": 0.9288998246192932, + "learning_rate": 2.432668743073463e-07, + "loss": 0.4558, + "step": 28892 + }, + { + "epoch": 1.9390960034898157, + "grad_norm": 1.0898096561431885, + "learning_rate": 2.421972593753319e-07, + "loss": 0.4421, + "step": 28894 + }, + { + "epoch": 1.9392302271735846, + "grad_norm": 0.9902854561805725, + "learning_rate": 2.4112999540165103e-07, + "loss": 0.4606, + "step": 28896 + }, + { + "epoch": 1.9393644508573538, + "grad_norm": 0.9736539721488953, + "learning_rate": 2.400650824367301e-07, + "loss": 0.4236, + "step": 28898 + }, + { + "epoch": 1.939498674541123, + "grad_norm": 1.0703511238098145, + "learning_rate": 2.3900252053088435e-07, + "loss": 0.5251, + "step": 28900 + }, + { + "epoch": 1.9396328982248918, + "grad_norm": 1.0596925020217896, + "learning_rate": 2.379423097343292e-07, + "loss": 0.4527, + "step": 28902 + }, + { + "epoch": 1.9397671219086607, + "grad_norm": 0.9808062314987183, + "learning_rate": 2.3688445009713566e-07, + "loss": 0.415, + "step": 28904 + }, + { + "epoch": 1.9399013455924297, + "grad_norm": 1.0180929899215698, + "learning_rate": 2.3582894166930268e-07, + "loss": 0.4719, + "step": 28906 + }, + { + "epoch": 1.9400355692761988, + "grad_norm": 0.9373030662536621, + "learning_rate": 2.3477578450069038e-07, + "loss": 0.4148, + "step": 28908 + }, + { + "epoch": 1.940169792959968, + "grad_norm": 1.1346129179000854, + "learning_rate": 2.3372497864106445e-07, + "loss": 0.4429, + "step": 28910 + }, + { + "epoch": 1.9403040166437369, + "grad_norm": 1.9116069078445435, + "learning_rate": 2.3267652414007414e-07, + "loss": 0.446, + "step": 28912 + }, + { + "epoch": 1.9404382403275058, + "grad_norm": 1.026120901107788, + "learning_rate": 2.316304210472575e-07, + "loss": 0.4624, + "step": 28914 + }, + { + "epoch": 1.9405724640112747, + "grad_norm": 1.02186918258667, + "learning_rate": 2.3058666941203623e-07, + "loss": 0.4748, + "step": 28916 + }, + { + "epoch": 1.9407066876950438, + "grad_norm": 1.0325887203216553, + "learning_rate": 2.2954526928372632e-07, + "loss": 0.5107, + "step": 28918 + }, + { + "epoch": 1.9408409113788128, + "grad_norm": 1.205985426902771, + "learning_rate": 2.2850622071153293e-07, + "loss": 0.4616, + "step": 28920 + }, + { + "epoch": 1.940975135062582, + "grad_norm": 1.0128406286239624, + "learning_rate": 2.2746952374455011e-07, + "loss": 0.4771, + "step": 28922 + }, + { + "epoch": 1.9411093587463508, + "grad_norm": 1.089816927909851, + "learning_rate": 2.264351784317553e-07, + "loss": 0.5144, + "step": 28924 + }, + { + "epoch": 1.9412435824301197, + "grad_norm": 0.9186378717422485, + "learning_rate": 2.2540318482202615e-07, + "loss": 0.413, + "step": 28926 + }, + { + "epoch": 1.9413778061138887, + "grad_norm": 1.0192416906356812, + "learning_rate": 2.2437354296411805e-07, + "loss": 0.5828, + "step": 28928 + }, + { + "epoch": 1.9415120297976578, + "grad_norm": 1.1217014789581299, + "learning_rate": 2.23346252906681e-07, + "loss": 0.4275, + "step": 28930 + }, + { + "epoch": 1.941646253481427, + "grad_norm": 1.148184895515442, + "learning_rate": 2.223213146982539e-07, + "loss": 0.4881, + "step": 28932 + }, + { + "epoch": 1.9417804771651959, + "grad_norm": 1.0364254713058472, + "learning_rate": 2.2129872838725364e-07, + "loss": 0.4372, + "step": 28934 + }, + { + "epoch": 1.9419147008489648, + "grad_norm": 0.9857342839241028, + "learning_rate": 2.2027849402201373e-07, + "loss": 0.5483, + "step": 28936 + }, + { + "epoch": 1.9420489245327337, + "grad_norm": 1.150019645690918, + "learning_rate": 2.1926061165071788e-07, + "loss": 0.4502, + "step": 28938 + }, + { + "epoch": 1.9421831482165028, + "grad_norm": 0.9097667932510376, + "learning_rate": 2.1824508132147204e-07, + "loss": 0.4826, + "step": 28940 + }, + { + "epoch": 1.9423173719002718, + "grad_norm": 0.9855087995529175, + "learning_rate": 2.1723190308225448e-07, + "loss": 0.4767, + "step": 28942 + }, + { + "epoch": 1.942451595584041, + "grad_norm": 1.1955652236938477, + "learning_rate": 2.1622107698093808e-07, + "loss": 0.5317, + "step": 28944 + }, + { + "epoch": 1.9425858192678098, + "grad_norm": 0.9239872097969055, + "learning_rate": 2.1521260306527903e-07, + "loss": 0.4205, + "step": 28946 + }, + { + "epoch": 1.9427200429515787, + "grad_norm": 1.0808923244476318, + "learning_rate": 2.1420648138292253e-07, + "loss": 0.4204, + "step": 28948 + }, + { + "epoch": 1.9428542666353477, + "grad_norm": 1.0448721647262573, + "learning_rate": 2.1320271198141395e-07, + "loss": 0.4553, + "step": 28950 + }, + { + "epoch": 1.9429884903191168, + "grad_norm": 1.039955735206604, + "learning_rate": 2.122012949081764e-07, + "loss": 0.4373, + "step": 28952 + }, + { + "epoch": 1.943122714002886, + "grad_norm": 0.969295859336853, + "learning_rate": 2.112022302105221e-07, + "loss": 0.4894, + "step": 28954 + }, + { + "epoch": 1.9432569376866549, + "grad_norm": 1.0688284635543823, + "learning_rate": 2.1020551793565768e-07, + "loss": 0.4254, + "step": 28956 + }, + { + "epoch": 1.9433911613704238, + "grad_norm": 0.9468502998352051, + "learning_rate": 2.092111581306788e-07, + "loss": 0.449, + "step": 28958 + }, + { + "epoch": 1.9435253850541927, + "grad_norm": 0.8899710774421692, + "learning_rate": 2.0821915084255906e-07, + "loss": 0.4741, + "step": 28960 + }, + { + "epoch": 1.9436596087379618, + "grad_norm": 0.9876530766487122, + "learning_rate": 2.0722949611817198e-07, + "loss": 0.4862, + "step": 28962 + }, + { + "epoch": 1.9437938324217308, + "grad_norm": 0.9413831830024719, + "learning_rate": 2.0624219400428023e-07, + "loss": 0.4808, + "step": 28964 + }, + { + "epoch": 1.9439280561055, + "grad_norm": 1.0272241830825806, + "learning_rate": 2.052572445475298e-07, + "loss": 0.4382, + "step": 28966 + }, + { + "epoch": 1.9440622797892688, + "grad_norm": 0.9331052899360657, + "learning_rate": 2.0427464779445572e-07, + "loss": 0.4401, + "step": 28968 + }, + { + "epoch": 1.9441965034730377, + "grad_norm": 0.9220208525657654, + "learning_rate": 2.0329440379148746e-07, + "loss": 0.5323, + "step": 28970 + }, + { + "epoch": 1.9443307271568067, + "grad_norm": 1.0369528532028198, + "learning_rate": 2.02316512584938e-07, + "loss": 0.4699, + "step": 28972 + }, + { + "epoch": 1.9444649508405758, + "grad_norm": 1.038794755935669, + "learning_rate": 2.0134097422100928e-07, + "loss": 0.4573, + "step": 28974 + }, + { + "epoch": 1.944599174524345, + "grad_norm": 1.0068680047988892, + "learning_rate": 2.0036778874579775e-07, + "loss": 0.4714, + "step": 28976 + }, + { + "epoch": 1.9447333982081139, + "grad_norm": 0.9134820103645325, + "learning_rate": 1.9939695620527777e-07, + "loss": 0.4858, + "step": 28978 + }, + { + "epoch": 1.9448676218918828, + "grad_norm": 1.061395525932312, + "learning_rate": 1.984284766453237e-07, + "loss": 0.4697, + "step": 28980 + }, + { + "epoch": 1.9450018455756517, + "grad_norm": 1.0704432725906372, + "learning_rate": 1.9746235011169344e-07, + "loss": 0.4647, + "step": 28982 + }, + { + "epoch": 1.9451360692594208, + "grad_norm": 1.048775553703308, + "learning_rate": 1.964985766500338e-07, + "loss": 0.4801, + "step": 28984 + }, + { + "epoch": 1.94527029294319, + "grad_norm": 1.0661295652389526, + "learning_rate": 1.9553715630588053e-07, + "loss": 0.5136, + "step": 28986 + }, + { + "epoch": 1.945404516626959, + "grad_norm": 1.0962748527526855, + "learning_rate": 1.9457808912466402e-07, + "loss": 0.4616, + "step": 28988 + }, + { + "epoch": 1.9455387403107278, + "grad_norm": 0.9793387651443481, + "learning_rate": 1.9362137515169242e-07, + "loss": 0.421, + "step": 28990 + }, + { + "epoch": 1.9456729639944967, + "grad_norm": 0.9833454489707947, + "learning_rate": 1.9266701443217406e-07, + "loss": 0.4683, + "step": 28992 + }, + { + "epoch": 1.9458071876782659, + "grad_norm": 1.0015058517456055, + "learning_rate": 1.9171500701119504e-07, + "loss": 0.441, + "step": 28994 + }, + { + "epoch": 1.9459414113620348, + "grad_norm": 1.0324184894561768, + "learning_rate": 1.907653529337361e-07, + "loss": 0.4315, + "step": 28996 + }, + { + "epoch": 1.946075635045804, + "grad_norm": 1.0735946893692017, + "learning_rate": 1.8981805224467242e-07, + "loss": 0.4409, + "step": 28998 + }, + { + "epoch": 1.9462098587295729, + "grad_norm": 1.0525434017181396, + "learning_rate": 1.8887310498875155e-07, + "loss": 0.4902, + "step": 29000 + }, + { + "epoch": 1.9463440824133418, + "grad_norm": 1.0508967638015747, + "learning_rate": 1.879305112106322e-07, + "loss": 0.4963, + "step": 29002 + }, + { + "epoch": 1.9464783060971107, + "grad_norm": 1.0489139556884766, + "learning_rate": 1.8699027095484545e-07, + "loss": 0.4933, + "step": 29004 + }, + { + "epoch": 1.9466125297808798, + "grad_norm": 0.9729229211807251, + "learning_rate": 1.8605238426581683e-07, + "loss": 0.4763, + "step": 29006 + }, + { + "epoch": 1.946746753464649, + "grad_norm": 1.1164216995239258, + "learning_rate": 1.8511685118785538e-07, + "loss": 0.5097, + "step": 29008 + }, + { + "epoch": 1.946880977148418, + "grad_norm": 0.9576023817062378, + "learning_rate": 1.841836717651646e-07, + "loss": 0.4712, + "step": 29010 + }, + { + "epoch": 1.9470152008321868, + "grad_norm": 1.105007290840149, + "learning_rate": 1.83252846041837e-07, + "loss": 0.4731, + "step": 29012 + }, + { + "epoch": 1.9471494245159557, + "grad_norm": 1.041167974472046, + "learning_rate": 1.823243740618541e-07, + "loss": 0.4661, + "step": 29014 + }, + { + "epoch": 1.9472836481997249, + "grad_norm": 1.1564899682998657, + "learning_rate": 1.8139825586908076e-07, + "loss": 0.4755, + "step": 29016 + }, + { + "epoch": 1.9474178718834938, + "grad_norm": 1.0982869863510132, + "learning_rate": 1.8047449150727648e-07, + "loss": 0.4441, + "step": 29018 + }, + { + "epoch": 1.947552095567263, + "grad_norm": 0.9591256976127625, + "learning_rate": 1.795530810200896e-07, + "loss": 0.4082, + "step": 29020 + }, + { + "epoch": 1.9476863192510319, + "grad_norm": 1.0350632667541504, + "learning_rate": 1.786340244510465e-07, + "loss": 0.4866, + "step": 29022 + }, + { + "epoch": 1.9478205429348008, + "grad_norm": 0.9792650938034058, + "learning_rate": 1.7771732184357904e-07, + "loss": 0.4856, + "step": 29024 + }, + { + "epoch": 1.9479547666185697, + "grad_norm": 0.9688316583633423, + "learning_rate": 1.7680297324099703e-07, + "loss": 0.4627, + "step": 29026 + }, + { + "epoch": 1.9480889903023388, + "grad_norm": 0.9843719601631165, + "learning_rate": 1.758909786864993e-07, + "loss": 0.4657, + "step": 29028 + }, + { + "epoch": 1.948223213986108, + "grad_norm": 0.9596292972564697, + "learning_rate": 1.7498133822317908e-07, + "loss": 0.4706, + "step": 29030 + }, + { + "epoch": 1.948357437669877, + "grad_norm": 0.9700624942779541, + "learning_rate": 1.7407405189401315e-07, + "loss": 0.4184, + "step": 29032 + }, + { + "epoch": 1.9484916613536458, + "grad_norm": 1.0020383596420288, + "learning_rate": 1.7316911974187276e-07, + "loss": 0.4551, + "step": 29034 + }, + { + "epoch": 1.9486258850374147, + "grad_norm": 1.0333776473999023, + "learning_rate": 1.7226654180950708e-07, + "loss": 0.4732, + "step": 29036 + }, + { + "epoch": 1.9487601087211839, + "grad_norm": 1.095948338508606, + "learning_rate": 1.7136631813957082e-07, + "loss": 0.4572, + "step": 29038 + }, + { + "epoch": 1.9488943324049528, + "grad_norm": 1.1414443254470825, + "learning_rate": 1.7046844877458556e-07, + "loss": 0.4756, + "step": 29040 + }, + { + "epoch": 1.949028556088722, + "grad_norm": 1.041707158088684, + "learning_rate": 1.6957293375698403e-07, + "loss": 0.4828, + "step": 29042 + }, + { + "epoch": 1.9491627797724909, + "grad_norm": 1.217383623123169, + "learning_rate": 1.6867977312907678e-07, + "loss": 0.5353, + "step": 29044 + }, + { + "epoch": 1.9492970034562598, + "grad_norm": 1.0687098503112793, + "learning_rate": 1.677889669330579e-07, + "loss": 0.4779, + "step": 29046 + }, + { + "epoch": 1.9494312271400287, + "grad_norm": 1.08873450756073, + "learning_rate": 1.6690051521102146e-07, + "loss": 0.5443, + "step": 29048 + }, + { + "epoch": 1.9495654508237978, + "grad_norm": 1.04291570186615, + "learning_rate": 1.6601441800493945e-07, + "loss": 0.5015, + "step": 29050 + }, + { + "epoch": 1.949699674507567, + "grad_norm": 1.0000983476638794, + "learning_rate": 1.6513067535668392e-07, + "loss": 0.5141, + "step": 29052 + }, + { + "epoch": 1.949833898191336, + "grad_norm": 0.8735694885253906, + "learning_rate": 1.6424928730801036e-07, + "loss": 0.4608, + "step": 29054 + }, + { + "epoch": 1.9499681218751048, + "grad_norm": 1.115757942199707, + "learning_rate": 1.633702539005577e-07, + "loss": 0.4715, + "step": 29056 + }, + { + "epoch": 1.9501023455588737, + "grad_norm": 1.030385971069336, + "learning_rate": 1.6249357517585938e-07, + "loss": 0.4762, + "step": 29058 + }, + { + "epoch": 1.9502365692426429, + "grad_norm": 1.125352144241333, + "learning_rate": 1.6161925117533783e-07, + "loss": 0.4587, + "step": 29060 + }, + { + "epoch": 1.950370792926412, + "grad_norm": 1.1368461847305298, + "learning_rate": 1.6074728194030442e-07, + "loss": 0.4317, + "step": 29062 + }, + { + "epoch": 1.950505016610181, + "grad_norm": 1.1113425493240356, + "learning_rate": 1.5987766751195953e-07, + "loss": 0.4119, + "step": 29064 + }, + { + "epoch": 1.9506392402939499, + "grad_norm": 1.0342289209365845, + "learning_rate": 1.5901040793138699e-07, + "loss": 0.458, + "step": 29066 + }, + { + "epoch": 1.9507734639777188, + "grad_norm": 1.0388826131820679, + "learning_rate": 1.5814550323957066e-07, + "loss": 0.5091, + "step": 29068 + }, + { + "epoch": 1.950907687661488, + "grad_norm": 1.1173781156539917, + "learning_rate": 1.5728295347736123e-07, + "loss": 0.5012, + "step": 29070 + }, + { + "epoch": 1.9510419113452568, + "grad_norm": 1.036182165145874, + "learning_rate": 1.5642275868552602e-07, + "loss": 0.4631, + "step": 29072 + }, + { + "epoch": 1.951176135029026, + "grad_norm": 1.0020720958709717, + "learning_rate": 1.5556491890469927e-07, + "loss": 0.4646, + "step": 29074 + }, + { + "epoch": 1.951310358712795, + "grad_norm": 0.8500409722328186, + "learning_rate": 1.5470943417541518e-07, + "loss": 0.4186, + "step": 29076 + }, + { + "epoch": 1.9514445823965638, + "grad_norm": 0.9407816529273987, + "learning_rate": 1.538563045380914e-07, + "loss": 0.5105, + "step": 29078 + }, + { + "epoch": 1.9515788060803327, + "grad_norm": 1.0555976629257202, + "learning_rate": 1.5300553003304575e-07, + "loss": 0.4626, + "step": 29080 + }, + { + "epoch": 1.9517130297641019, + "grad_norm": 1.0615906715393066, + "learning_rate": 1.521571107004627e-07, + "loss": 0.4693, + "step": 29082 + }, + { + "epoch": 1.951847253447871, + "grad_norm": 1.036577820777893, + "learning_rate": 1.5131104658043794e-07, + "loss": 0.4407, + "step": 29084 + }, + { + "epoch": 1.95198147713164, + "grad_norm": 1.1040834188461304, + "learning_rate": 1.5046733771293953e-07, + "loss": 0.4067, + "step": 29086 + }, + { + "epoch": 1.9521157008154089, + "grad_norm": 0.9711515307426453, + "learning_rate": 1.4962598413784113e-07, + "loss": 0.4554, + "step": 29088 + }, + { + "epoch": 1.9522499244991778, + "grad_norm": 1.0760223865509033, + "learning_rate": 1.4878698589488315e-07, + "loss": 0.4432, + "step": 29090 + }, + { + "epoch": 1.952384148182947, + "grad_norm": 1.018295168876648, + "learning_rate": 1.4795034302371168e-07, + "loss": 0.5183, + "step": 29092 + }, + { + "epoch": 1.9525183718667158, + "grad_norm": 1.0077235698699951, + "learning_rate": 1.4711605556385622e-07, + "loss": 0.4489, + "step": 29094 + }, + { + "epoch": 1.952652595550485, + "grad_norm": 0.9149937629699707, + "learning_rate": 1.4628412355474076e-07, + "loss": 0.4419, + "step": 29096 + }, + { + "epoch": 1.952786819234254, + "grad_norm": 1.0449026823043823, + "learning_rate": 1.4545454703566165e-07, + "loss": 0.4531, + "step": 29098 + }, + { + "epoch": 1.9529210429180228, + "grad_norm": 1.10755455493927, + "learning_rate": 1.4462732604582086e-07, + "loss": 0.4817, + "step": 29100 + }, + { + "epoch": 1.9530552666017917, + "grad_norm": 1.0468051433563232, + "learning_rate": 1.4380246062430935e-07, + "loss": 0.4701, + "step": 29102 + }, + { + "epoch": 1.9531894902855609, + "grad_norm": 0.9423577189445496, + "learning_rate": 1.4297995081008487e-07, + "loss": 0.4525, + "step": 29104 + }, + { + "epoch": 1.95332371396933, + "grad_norm": 0.8814041614532471, + "learning_rate": 1.4215979664202183e-07, + "loss": 0.4456, + "step": 29106 + }, + { + "epoch": 1.953457937653099, + "grad_norm": 1.0376014709472656, + "learning_rate": 1.4134199815886705e-07, + "loss": 0.5057, + "step": 29108 + }, + { + "epoch": 1.9535921613368679, + "grad_norm": 1.0050175189971924, + "learning_rate": 1.4052655539926184e-07, + "loss": 0.4042, + "step": 29110 + }, + { + "epoch": 1.9537263850206368, + "grad_norm": 1.1026122570037842, + "learning_rate": 1.3971346840173095e-07, + "loss": 0.4135, + "step": 29112 + }, + { + "epoch": 1.953860608704406, + "grad_norm": 0.948855459690094, + "learning_rate": 1.389027372046936e-07, + "loss": 0.4329, + "step": 29114 + }, + { + "epoch": 1.9539948323881748, + "grad_norm": 1.0344352722167969, + "learning_rate": 1.3809436184645252e-07, + "loss": 0.4875, + "step": 29116 + }, + { + "epoch": 1.954129056071944, + "grad_norm": 1.8970121145248413, + "learning_rate": 1.3728834236520493e-07, + "loss": 0.4427, + "step": 29118 + }, + { + "epoch": 1.954263279755713, + "grad_norm": 0.8916929960250854, + "learning_rate": 1.3648467879902594e-07, + "loss": 0.4658, + "step": 29120 + }, + { + "epoch": 1.9543975034394818, + "grad_norm": 0.9808556437492371, + "learning_rate": 1.356833711859018e-07, + "loss": 0.5233, + "step": 29122 + }, + { + "epoch": 1.9545317271232507, + "grad_norm": 0.9045658111572266, + "learning_rate": 1.3488441956368003e-07, + "loss": 0.4995, + "step": 29124 + }, + { + "epoch": 1.9546659508070199, + "grad_norm": 1.0025702714920044, + "learning_rate": 1.3408782397011932e-07, + "loss": 0.5065, + "step": 29126 + }, + { + "epoch": 1.954800174490789, + "grad_norm": 0.7962448596954346, + "learning_rate": 1.3329358444284513e-07, + "loss": 0.4051, + "step": 29128 + }, + { + "epoch": 1.954934398174558, + "grad_norm": 1.0446126461029053, + "learning_rate": 1.3250170101939407e-07, + "loss": 0.4518, + "step": 29130 + }, + { + "epoch": 1.9550686218583269, + "grad_norm": 1.0310949087142944, + "learning_rate": 1.3171217373717516e-07, + "loss": 0.4761, + "step": 29132 + }, + { + "epoch": 1.9552028455420958, + "grad_norm": 1.033321499824524, + "learning_rate": 1.309250026334974e-07, + "loss": 0.4402, + "step": 29134 + }, + { + "epoch": 1.955337069225865, + "grad_norm": 0.9540125131607056, + "learning_rate": 1.301401877455477e-07, + "loss": 0.4348, + "step": 29136 + }, + { + "epoch": 1.955471292909634, + "grad_norm": 0.9811615347862244, + "learning_rate": 1.293577291104131e-07, + "loss": 0.5095, + "step": 29138 + }, + { + "epoch": 1.955605516593403, + "grad_norm": 1.055335283279419, + "learning_rate": 1.285776267650529e-07, + "loss": 0.4345, + "step": 29140 + }, + { + "epoch": 1.955739740277172, + "grad_norm": 1.139459252357483, + "learning_rate": 1.2779988074633765e-07, + "loss": 0.4297, + "step": 29142 + }, + { + "epoch": 1.9558739639609408, + "grad_norm": 0.95600426197052, + "learning_rate": 1.2702449109100455e-07, + "loss": 0.4257, + "step": 29144 + }, + { + "epoch": 1.95600818764471, + "grad_norm": 1.0222060680389404, + "learning_rate": 1.2625145783569658e-07, + "loss": 0.4449, + "step": 29146 + }, + { + "epoch": 1.9561424113284789, + "grad_norm": 1.061171293258667, + "learning_rate": 1.2548078101692894e-07, + "loss": 0.4535, + "step": 29148 + }, + { + "epoch": 1.956276635012248, + "grad_norm": 1.103115200996399, + "learning_rate": 1.2471246067112807e-07, + "loss": 0.4707, + "step": 29150 + }, + { + "epoch": 1.956410858696017, + "grad_norm": 0.9497417211532593, + "learning_rate": 1.239464968345816e-07, + "loss": 0.462, + "step": 29152 + }, + { + "epoch": 1.9565450823797859, + "grad_norm": 1.0168882608413696, + "learning_rate": 1.2318288954348833e-07, + "loss": 0.4242, + "step": 29154 + }, + { + "epoch": 1.9566793060635548, + "grad_norm": 1.0620505809783936, + "learning_rate": 1.22421638833925e-07, + "loss": 0.4896, + "step": 29156 + }, + { + "epoch": 1.956813529747324, + "grad_norm": 1.0867334604263306, + "learning_rate": 1.216627447418628e-07, + "loss": 0.4628, + "step": 29158 + }, + { + "epoch": 1.956947753431093, + "grad_norm": 0.9913773536682129, + "learning_rate": 1.2090620730315084e-07, + "loss": 0.4797, + "step": 29160 + }, + { + "epoch": 1.957081977114862, + "grad_norm": 0.995578944683075, + "learning_rate": 1.201520265535383e-07, + "loss": 0.4371, + "step": 29162 + }, + { + "epoch": 1.957216200798631, + "grad_norm": 0.9965330958366394, + "learning_rate": 1.1940020252865226e-07, + "loss": 0.5014, + "step": 29164 + }, + { + "epoch": 1.9573504244823998, + "grad_norm": 1.0482277870178223, + "learning_rate": 1.1865073526402537e-07, + "loss": 0.4753, + "step": 29166 + }, + { + "epoch": 1.957484648166169, + "grad_norm": 1.0561718940734863, + "learning_rate": 1.1790362479506822e-07, + "loss": 0.3859, + "step": 29168 + }, + { + "epoch": 1.9576188718499379, + "grad_norm": 1.0040631294250488, + "learning_rate": 1.1715887115706926e-07, + "loss": 0.5141, + "step": 29170 + }, + { + "epoch": 1.957753095533707, + "grad_norm": 1.0110114812850952, + "learning_rate": 1.1641647438522807e-07, + "loss": 0.4536, + "step": 29172 + }, + { + "epoch": 1.957887319217476, + "grad_norm": 0.9060198664665222, + "learning_rate": 1.1567643451461108e-07, + "loss": 0.4565, + "step": 29174 + }, + { + "epoch": 1.9580215429012449, + "grad_norm": 1.0186243057250977, + "learning_rate": 1.1493875158019584e-07, + "loss": 0.4437, + "step": 29176 + }, + { + "epoch": 1.9581557665850138, + "grad_norm": 1.1048253774642944, + "learning_rate": 1.1420342561682673e-07, + "loss": 0.474, + "step": 29178 + }, + { + "epoch": 1.958289990268783, + "grad_norm": 1.0118054151535034, + "learning_rate": 1.1347045665924816e-07, + "loss": 0.4664, + "step": 29180 + }, + { + "epoch": 1.958424213952552, + "grad_norm": 0.8971180319786072, + "learning_rate": 1.1273984474209354e-07, + "loss": 0.4557, + "step": 29182 + }, + { + "epoch": 1.958558437636321, + "grad_norm": 1.033743977546692, + "learning_rate": 1.1201158989988525e-07, + "loss": 0.4718, + "step": 29184 + }, + { + "epoch": 1.95869266132009, + "grad_norm": 0.9839541912078857, + "learning_rate": 1.1128569216702356e-07, + "loss": 0.4439, + "step": 29186 + }, + { + "epoch": 1.9588268850038588, + "grad_norm": 0.991456151008606, + "learning_rate": 1.1056215157781435e-07, + "loss": 0.4654, + "step": 29188 + }, + { + "epoch": 1.958961108687628, + "grad_norm": 0.9525729417800903, + "learning_rate": 1.098409681664414e-07, + "loss": 0.388, + "step": 29190 + }, + { + "epoch": 1.9590953323713969, + "grad_norm": 1.019943356513977, + "learning_rate": 1.091221419669719e-07, + "loss": 0.4282, + "step": 29192 + }, + { + "epoch": 1.959229556055166, + "grad_norm": 0.8993679881095886, + "learning_rate": 1.0840567301337868e-07, + "loss": 0.448, + "step": 29194 + }, + { + "epoch": 1.959363779738935, + "grad_norm": 1.0546905994415283, + "learning_rate": 1.0769156133951241e-07, + "loss": 0.435, + "step": 29196 + }, + { + "epoch": 1.9594980034227039, + "grad_norm": 1.1095595359802246, + "learning_rate": 1.0697980697910725e-07, + "loss": 0.4396, + "step": 29198 + }, + { + "epoch": 1.9596322271064728, + "grad_norm": 1.655231237411499, + "learning_rate": 1.0627040996579741e-07, + "loss": 0.4249, + "step": 29200 + }, + { + "epoch": 1.959766450790242, + "grad_norm": 0.9991711378097534, + "learning_rate": 1.0556337033310048e-07, + "loss": 0.4191, + "step": 29202 + }, + { + "epoch": 1.959900674474011, + "grad_norm": 1.16497004032135, + "learning_rate": 1.0485868811441757e-07, + "loss": 0.4507, + "step": 29204 + }, + { + "epoch": 1.96003489815778, + "grad_norm": 1.2120740413665771, + "learning_rate": 1.0415636334304979e-07, + "loss": 0.4976, + "step": 29206 + }, + { + "epoch": 1.960169121841549, + "grad_norm": 1.0939778089523315, + "learning_rate": 1.0345639605217616e-07, + "loss": 0.4828, + "step": 29208 + }, + { + "epoch": 1.9603033455253178, + "grad_norm": 0.9883015751838684, + "learning_rate": 1.0275878627487023e-07, + "loss": 0.4519, + "step": 29210 + }, + { + "epoch": 1.960437569209087, + "grad_norm": 0.9145817160606384, + "learning_rate": 1.0206353404409452e-07, + "loss": 0.4446, + "step": 29212 + }, + { + "epoch": 1.960571792892856, + "grad_norm": 1.0710937976837158, + "learning_rate": 1.0137063939269497e-07, + "loss": 0.4951, + "step": 29214 + }, + { + "epoch": 1.960706016576625, + "grad_norm": 1.0684460401535034, + "learning_rate": 1.0068010235341208e-07, + "loss": 0.4343, + "step": 29216 + }, + { + "epoch": 1.960840240260394, + "grad_norm": 0.9165230393409729, + "learning_rate": 9.999192295886972e-08, + "loss": 0.4662, + "step": 29218 + }, + { + "epoch": 1.9609744639441629, + "grad_norm": 0.965201199054718, + "learning_rate": 9.930610124158634e-08, + "loss": 0.4806, + "step": 29220 + }, + { + "epoch": 1.961108687627932, + "grad_norm": 1.0687464475631714, + "learning_rate": 9.862263723396382e-08, + "loss": 0.5493, + "step": 29222 + }, + { + "epoch": 1.961242911311701, + "grad_norm": 1.1389684677124023, + "learning_rate": 9.79415309682985e-08, + "loss": 0.5155, + "step": 29224 + }, + { + "epoch": 1.96137713499547, + "grad_norm": 0.9733030200004578, + "learning_rate": 9.726278247676468e-08, + "loss": 0.4105, + "step": 29226 + }, + { + "epoch": 1.961511358679239, + "grad_norm": 1.0655992031097412, + "learning_rate": 9.658639179143669e-08, + "loss": 0.4438, + "step": 29228 + }, + { + "epoch": 1.961645582363008, + "grad_norm": 1.075886845588684, + "learning_rate": 9.591235894426675e-08, + "loss": 0.4108, + "step": 29230 + }, + { + "epoch": 1.9617798060467768, + "grad_norm": 0.9965173006057739, + "learning_rate": 9.524068396710717e-08, + "loss": 0.5006, + "step": 29232 + }, + { + "epoch": 1.961914029730546, + "grad_norm": 0.8664084076881409, + "learning_rate": 9.457136689169366e-08, + "loss": 0.3999, + "step": 29234 + }, + { + "epoch": 1.962048253414315, + "grad_norm": 1.0834602117538452, + "learning_rate": 9.390440774965092e-08, + "loss": 0.516, + "step": 29236 + }, + { + "epoch": 1.962182477098084, + "grad_norm": 0.9731799960136414, + "learning_rate": 9.323980657248154e-08, + "loss": 0.4889, + "step": 29238 + }, + { + "epoch": 1.962316700781853, + "grad_norm": 0.9713391065597534, + "learning_rate": 9.257756339159929e-08, + "loss": 0.4285, + "step": 29240 + }, + { + "epoch": 1.9624509244656219, + "grad_norm": 1.1183282136917114, + "learning_rate": 9.191767823828467e-08, + "loss": 0.4732, + "step": 29242 + }, + { + "epoch": 1.962585148149391, + "grad_norm": 1.0748634338378906, + "learning_rate": 9.126015114372388e-08, + "loss": 0.4712, + "step": 29244 + }, + { + "epoch": 1.96271937183316, + "grad_norm": 0.9243441224098206, + "learning_rate": 9.060498213897539e-08, + "loss": 0.4211, + "step": 29246 + }, + { + "epoch": 1.962853595516929, + "grad_norm": 0.9169735908508301, + "learning_rate": 8.995217125500333e-08, + "loss": 0.4266, + "step": 29248 + }, + { + "epoch": 1.962987819200698, + "grad_norm": 0.9653085470199585, + "learning_rate": 8.930171852264413e-08, + "loss": 0.4021, + "step": 29250 + }, + { + "epoch": 1.963122042884467, + "grad_norm": 0.9383867383003235, + "learning_rate": 8.865362397263433e-08, + "loss": 0.433, + "step": 29252 + }, + { + "epoch": 1.9632562665682358, + "grad_norm": 1.1528648138046265, + "learning_rate": 8.800788763559386e-08, + "loss": 0.4796, + "step": 29254 + }, + { + "epoch": 1.963390490252005, + "grad_norm": 0.9118273258209229, + "learning_rate": 8.736450954203168e-08, + "loss": 0.4181, + "step": 29256 + }, + { + "epoch": 1.963524713935774, + "grad_norm": 0.8596152663230896, + "learning_rate": 8.672348972235122e-08, + "loss": 0.4154, + "step": 29258 + }, + { + "epoch": 1.963658937619543, + "grad_norm": 1.0593302249908447, + "learning_rate": 8.608482820682828e-08, + "loss": 0.4761, + "step": 29260 + }, + { + "epoch": 1.963793161303312, + "grad_norm": 1.0908609628677368, + "learning_rate": 8.544852502565537e-08, + "loss": 0.4457, + "step": 29262 + }, + { + "epoch": 1.9639273849870809, + "grad_norm": 0.8625156879425049, + "learning_rate": 8.481458020888066e-08, + "loss": 0.4576, + "step": 29264 + }, + { + "epoch": 1.96406160867085, + "grad_norm": 1.1236494779586792, + "learning_rate": 8.418299378646355e-08, + "loss": 0.5115, + "step": 29266 + }, + { + "epoch": 1.964195832354619, + "grad_norm": 1.000876545906067, + "learning_rate": 8.355376578824681e-08, + "loss": 0.4208, + "step": 29268 + }, + { + "epoch": 1.964330056038388, + "grad_norm": 1.1040709018707275, + "learning_rate": 8.292689624395666e-08, + "loss": 0.4387, + "step": 29270 + }, + { + "epoch": 1.964464279722157, + "grad_norm": 0.9892479777336121, + "learning_rate": 8.230238518321387e-08, + "loss": 0.4468, + "step": 29272 + }, + { + "epoch": 1.964598503405926, + "grad_norm": 1.0196044445037842, + "learning_rate": 8.168023263552815e-08, + "loss": 0.4451, + "step": 29274 + }, + { + "epoch": 1.9647327270896948, + "grad_norm": 0.9883550405502319, + "learning_rate": 8.106043863028157e-08, + "loss": 0.4487, + "step": 29276 + }, + { + "epoch": 1.964866950773464, + "grad_norm": 1.0349782705307007, + "learning_rate": 8.044300319677844e-08, + "loss": 0.4504, + "step": 29278 + }, + { + "epoch": 1.965001174457233, + "grad_norm": 1.0604530572891235, + "learning_rate": 7.982792636417324e-08, + "loss": 0.4185, + "step": 29280 + }, + { + "epoch": 1.965135398141002, + "grad_norm": 1.0895187854766846, + "learning_rate": 7.921520816153716e-08, + "loss": 0.4678, + "step": 29282 + }, + { + "epoch": 1.965269621824771, + "grad_norm": 0.9991888999938965, + "learning_rate": 7.860484861781925e-08, + "loss": 0.4507, + "step": 29284 + }, + { + "epoch": 1.9654038455085399, + "grad_norm": 0.9821450114250183, + "learning_rate": 7.799684776185201e-08, + "loss": 0.4439, + "step": 29286 + }, + { + "epoch": 1.965538069192309, + "grad_norm": 1.1980267763137817, + "learning_rate": 7.739120562236802e-08, + "loss": 0.5169, + "step": 29288 + }, + { + "epoch": 1.9656722928760781, + "grad_norm": 0.896367609500885, + "learning_rate": 7.678792222798325e-08, + "loss": 0.4158, + "step": 29290 + }, + { + "epoch": 1.965806516559847, + "grad_norm": 1.0855690240859985, + "learning_rate": 7.618699760719716e-08, + "loss": 0.4198, + "step": 29292 + }, + { + "epoch": 1.965940740243616, + "grad_norm": 1.1153912544250488, + "learning_rate": 7.558843178840924e-08, + "loss": 0.4523, + "step": 29294 + }, + { + "epoch": 1.966074963927385, + "grad_norm": 0.9311449527740479, + "learning_rate": 7.499222479989132e-08, + "loss": 0.39, + "step": 29296 + }, + { + "epoch": 1.966209187611154, + "grad_norm": 1.0853956937789917, + "learning_rate": 7.43983766698153e-08, + "loss": 0.474, + "step": 29298 + }, + { + "epoch": 1.966343411294923, + "grad_norm": 0.9984995126724243, + "learning_rate": 7.380688742624209e-08, + "loss": 0.4446, + "step": 29300 + }, + { + "epoch": 1.966477634978692, + "grad_norm": 1.0153001546859741, + "learning_rate": 7.321775709712153e-08, + "loss": 0.4443, + "step": 29302 + }, + { + "epoch": 1.966611858662461, + "grad_norm": 1.0898973941802979, + "learning_rate": 7.263098571028138e-08, + "loss": 0.5207, + "step": 29304 + }, + { + "epoch": 1.96674608234623, + "grad_norm": 1.0372871160507202, + "learning_rate": 7.204657329345498e-08, + "loss": 0.4754, + "step": 29306 + }, + { + "epoch": 1.9668803060299989, + "grad_norm": 1.0598440170288086, + "learning_rate": 7.146451987424252e-08, + "loss": 0.4786, + "step": 29308 + }, + { + "epoch": 1.967014529713768, + "grad_norm": 1.0670109987258911, + "learning_rate": 7.088482548015529e-08, + "loss": 0.5069, + "step": 29310 + }, + { + "epoch": 1.9671487533975371, + "grad_norm": 0.9652385115623474, + "learning_rate": 7.030749013857696e-08, + "loss": 0.4041, + "step": 29312 + }, + { + "epoch": 1.967282977081306, + "grad_norm": 0.8895744681358337, + "learning_rate": 6.97325138767857e-08, + "loss": 0.4055, + "step": 29314 + }, + { + "epoch": 1.967417200765075, + "grad_norm": 0.9850452542304993, + "learning_rate": 6.915989672195422e-08, + "loss": 0.4344, + "step": 29316 + }, + { + "epoch": 1.967551424448844, + "grad_norm": 1.1335660219192505, + "learning_rate": 6.858963870112756e-08, + "loss": 0.4607, + "step": 29318 + }, + { + "epoch": 1.967685648132613, + "grad_norm": 1.2784786224365234, + "learning_rate": 6.802173984125637e-08, + "loss": 0.4969, + "step": 29320 + }, + { + "epoch": 1.967819871816382, + "grad_norm": 0.9682573080062866, + "learning_rate": 6.745620016917476e-08, + "loss": 0.4367, + "step": 29322 + }, + { + "epoch": 1.967954095500151, + "grad_norm": 0.9108989238739014, + "learning_rate": 6.689301971159467e-08, + "loss": 0.4489, + "step": 29324 + }, + { + "epoch": 1.96808831918392, + "grad_norm": 1.087185025215149, + "learning_rate": 6.633219849513372e-08, + "loss": 0.4899, + "step": 29326 + }, + { + "epoch": 1.968222542867689, + "grad_norm": 1.101282000541687, + "learning_rate": 6.577373654628183e-08, + "loss": 0.5098, + "step": 29328 + }, + { + "epoch": 1.9683567665514579, + "grad_norm": 1.0934100151062012, + "learning_rate": 6.521763389142899e-08, + "loss": 0.4456, + "step": 29330 + }, + { + "epoch": 1.968490990235227, + "grad_norm": 1.023734211921692, + "learning_rate": 6.466389055685418e-08, + "loss": 0.5027, + "step": 29332 + }, + { + "epoch": 1.9686252139189961, + "grad_norm": 1.0267611742019653, + "learning_rate": 6.411250656871426e-08, + "loss": 0.4883, + "step": 29334 + }, + { + "epoch": 1.968759437602765, + "grad_norm": 1.0175542831420898, + "learning_rate": 6.356348195306616e-08, + "loss": 0.4625, + "step": 29336 + }, + { + "epoch": 1.968893661286534, + "grad_norm": 1.2178868055343628, + "learning_rate": 6.301681673585025e-08, + "loss": 0.5294, + "step": 29338 + }, + { + "epoch": 1.969027884970303, + "grad_norm": 1.1020519733428955, + "learning_rate": 6.24725109428903e-08, + "loss": 0.4643, + "step": 29340 + }, + { + "epoch": 1.969162108654072, + "grad_norm": 1.027024507522583, + "learning_rate": 6.193056459990465e-08, + "loss": 0.4884, + "step": 29342 + }, + { + "epoch": 1.969296332337841, + "grad_norm": 0.9886162877082825, + "learning_rate": 6.139097773250057e-08, + "loss": 0.4458, + "step": 29344 + }, + { + "epoch": 1.96943055602161, + "grad_norm": 1.0386093854904175, + "learning_rate": 6.085375036617436e-08, + "loss": 0.447, + "step": 29346 + }, + { + "epoch": 1.969564779705379, + "grad_norm": 1.0407241582870483, + "learning_rate": 6.031888252630569e-08, + "loss": 0.435, + "step": 29348 + }, + { + "epoch": 1.969699003389148, + "grad_norm": 1.0747627019882202, + "learning_rate": 5.97863742381688e-08, + "loss": 0.4715, + "step": 29350 + }, + { + "epoch": 1.9698332270729169, + "grad_norm": 1.035626769065857, + "learning_rate": 5.9256225526921336e-08, + "loss": 0.5026, + "step": 29352 + }, + { + "epoch": 1.969967450756686, + "grad_norm": 1.252779483795166, + "learning_rate": 5.8728436417615494e-08, + "loss": 0.4384, + "step": 29354 + }, + { + "epoch": 1.9701016744404551, + "grad_norm": 0.9608078002929688, + "learning_rate": 5.820300693518133e-08, + "loss": 0.4549, + "step": 29356 + }, + { + "epoch": 1.970235898124224, + "grad_norm": 1.155197024345398, + "learning_rate": 5.7679937104454516e-08, + "loss": 0.4433, + "step": 29358 + }, + { + "epoch": 1.970370121807993, + "grad_norm": 1.0186712741851807, + "learning_rate": 5.715922695013753e-08, + "loss": 0.4711, + "step": 29360 + }, + { + "epoch": 1.970504345491762, + "grad_norm": 1.0469697713851929, + "learning_rate": 5.664087649684402e-08, + "loss": 0.4341, + "step": 29362 + }, + { + "epoch": 1.970638569175531, + "grad_norm": 1.1047803163528442, + "learning_rate": 5.6124885769054394e-08, + "loss": 0.4584, + "step": 29364 + }, + { + "epoch": 1.9707727928593002, + "grad_norm": 1.0283458232879639, + "learning_rate": 5.5611254791154696e-08, + "loss": 0.4277, + "step": 29366 + }, + { + "epoch": 1.970907016543069, + "grad_norm": 1.0908386707305908, + "learning_rate": 5.509998358741442e-08, + "loss": 0.4204, + "step": 29368 + }, + { + "epoch": 1.971041240226838, + "grad_norm": 1.1800463199615479, + "learning_rate": 5.4591072181986444e-08, + "loss": 0.4752, + "step": 29370 + }, + { + "epoch": 1.971175463910607, + "grad_norm": 0.9975547194480896, + "learning_rate": 5.408452059891822e-08, + "loss": 0.4199, + "step": 29372 + }, + { + "epoch": 1.971309687594376, + "grad_norm": 1.0553650856018066, + "learning_rate": 5.358032886214059e-08, + "loss": 0.4354, + "step": 29374 + }, + { + "epoch": 1.971443911278145, + "grad_norm": 0.9321396350860596, + "learning_rate": 5.307849699547895e-08, + "loss": 0.4533, + "step": 29376 + }, + { + "epoch": 1.9715781349619141, + "grad_norm": 1.0798766613006592, + "learning_rate": 5.257902502263656e-08, + "loss": 0.4708, + "step": 29378 + }, + { + "epoch": 1.971712358645683, + "grad_norm": 1.0483887195587158, + "learning_rate": 5.208191296722231e-08, + "loss": 0.4684, + "step": 29380 + }, + { + "epoch": 1.971846582329452, + "grad_norm": 1.0603145360946655, + "learning_rate": 5.158716085271742e-08, + "loss": 0.4299, + "step": 29382 + }, + { + "epoch": 1.971980806013221, + "grad_norm": 1.0174487829208374, + "learning_rate": 5.109476870250318e-08, + "loss": 0.4365, + "step": 29384 + }, + { + "epoch": 1.97211502969699, + "grad_norm": 0.9578461647033691, + "learning_rate": 5.060473653983877e-08, + "loss": 0.4683, + "step": 29386 + }, + { + "epoch": 1.9722492533807592, + "grad_norm": 1.0425300598144531, + "learning_rate": 5.0117064387877885e-08, + "loss": 0.4852, + "step": 29388 + }, + { + "epoch": 1.972383477064528, + "grad_norm": 1.0657423734664917, + "learning_rate": 4.9631752269663213e-08, + "loss": 0.4743, + "step": 29390 + }, + { + "epoch": 1.972517700748297, + "grad_norm": 1.0595128536224365, + "learning_rate": 4.91488002081264e-08, + "loss": 0.4258, + "step": 29392 + }, + { + "epoch": 1.972651924432066, + "grad_norm": 0.9602314829826355, + "learning_rate": 4.8668208226088085e-08, + "loss": 0.3623, + "step": 29394 + }, + { + "epoch": 1.972786148115835, + "grad_norm": 0.9397258758544922, + "learning_rate": 4.818997634624678e-08, + "loss": 0.497, + "step": 29396 + }, + { + "epoch": 1.972920371799604, + "grad_norm": 1.1596510410308838, + "learning_rate": 4.771410459120662e-08, + "loss": 0.458, + "step": 29398 + }, + { + "epoch": 1.9730545954833731, + "grad_norm": 1.1102371215820312, + "learning_rate": 4.724059298344408e-08, + "loss": 0.4556, + "step": 29400 + }, + { + "epoch": 1.973188819167142, + "grad_norm": 1.0936411619186401, + "learning_rate": 4.676944154533569e-08, + "loss": 0.459, + "step": 29402 + }, + { + "epoch": 1.973323042850911, + "grad_norm": 1.1787388324737549, + "learning_rate": 4.630065029914698e-08, + "loss": 0.4643, + "step": 29404 + }, + { + "epoch": 1.97345726653468, + "grad_norm": 0.9891636371612549, + "learning_rate": 4.583421926701581e-08, + "loss": 0.4165, + "step": 29406 + }, + { + "epoch": 1.973591490218449, + "grad_norm": 1.1265305280685425, + "learning_rate": 4.537014847099119e-08, + "loss": 0.4794, + "step": 29408 + }, + { + "epoch": 1.9737257139022182, + "grad_norm": 1.053418517112732, + "learning_rate": 4.490843793300003e-08, + "loss": 0.4857, + "step": 29410 + }, + { + "epoch": 1.973859937585987, + "grad_norm": 1.0158971548080444, + "learning_rate": 4.4449087674847125e-08, + "loss": 0.4282, + "step": 29412 + }, + { + "epoch": 1.973994161269756, + "grad_norm": 0.9685587882995605, + "learning_rate": 4.399209771824287e-08, + "loss": 0.4591, + "step": 29414 + }, + { + "epoch": 1.974128384953525, + "grad_norm": 0.9691962003707886, + "learning_rate": 4.353746808477554e-08, + "loss": 0.472, + "step": 29416 + }, + { + "epoch": 1.974262608637294, + "grad_norm": 0.9528734683990479, + "learning_rate": 4.3085198795933536e-08, + "loss": 0.4626, + "step": 29418 + }, + { + "epoch": 1.974396832321063, + "grad_norm": 1.0498555898666382, + "learning_rate": 4.263528987307197e-08, + "loss": 0.4523, + "step": 29420 + }, + { + "epoch": 1.9745310560048321, + "grad_norm": 1.0218913555145264, + "learning_rate": 4.2187741337462724e-08, + "loss": 0.4251, + "step": 29422 + }, + { + "epoch": 1.974665279688601, + "grad_norm": 1.146499752998352, + "learning_rate": 4.1742553210238896e-08, + "loss": 0.5049, + "step": 29424 + }, + { + "epoch": 1.97479950337237, + "grad_norm": 1.0821541547775269, + "learning_rate": 4.129972551244476e-08, + "loss": 0.4064, + "step": 29426 + }, + { + "epoch": 1.974933727056139, + "grad_norm": 1.0310195684432983, + "learning_rate": 4.085925826499692e-08, + "loss": 0.5104, + "step": 29428 + }, + { + "epoch": 1.975067950739908, + "grad_norm": 1.0251456499099731, + "learning_rate": 4.0421151488712064e-08, + "loss": 0.4923, + "step": 29430 + }, + { + "epoch": 1.9752021744236772, + "grad_norm": 1.0707616806030273, + "learning_rate": 3.998540520428473e-08, + "loss": 0.5032, + "step": 29432 + }, + { + "epoch": 1.975336398107446, + "grad_norm": 1.109976887702942, + "learning_rate": 3.955201943230402e-08, + "loss": 0.4809, + "step": 29434 + }, + { + "epoch": 1.975470621791215, + "grad_norm": 0.980389416217804, + "learning_rate": 3.9120994193247994e-08, + "loss": 0.4546, + "step": 29436 + }, + { + "epoch": 1.975604845474984, + "grad_norm": 1.1692322492599487, + "learning_rate": 3.869232950747814e-08, + "loss": 0.4648, + "step": 29438 + }, + { + "epoch": 1.975739069158753, + "grad_norm": 1.0092170238494873, + "learning_rate": 3.826602539525603e-08, + "loss": 0.3806, + "step": 29440 + }, + { + "epoch": 1.9758732928425222, + "grad_norm": 0.8585953712463379, + "learning_rate": 3.784208187671556e-08, + "loss": 0.4346, + "step": 29442 + }, + { + "epoch": 1.9760075165262911, + "grad_norm": 1.019691824913025, + "learning_rate": 3.7420498971890706e-08, + "loss": 0.423, + "step": 29444 + }, + { + "epoch": 1.97614174021006, + "grad_norm": 1.096204400062561, + "learning_rate": 3.700127670070441e-08, + "loss": 0.4462, + "step": 29446 + }, + { + "epoch": 1.976275963893829, + "grad_norm": 0.977575957775116, + "learning_rate": 3.658441508295196e-08, + "loss": 0.4838, + "step": 29448 + }, + { + "epoch": 1.9764101875775981, + "grad_norm": 0.9653611779212952, + "learning_rate": 3.616991413834536e-08, + "loss": 0.5223, + "step": 29450 + }, + { + "epoch": 1.976544411261367, + "grad_norm": 1.0307095050811768, + "learning_rate": 3.57577738864523e-08, + "loss": 0.5271, + "step": 29452 + }, + { + "epoch": 1.9766786349451362, + "grad_norm": 1.3023886680603027, + "learning_rate": 3.534799434676273e-08, + "loss": 0.4955, + "step": 29454 + }, + { + "epoch": 1.976812858628905, + "grad_norm": 0.9763998985290527, + "learning_rate": 3.494057553862229e-08, + "loss": 0.4732, + "step": 29456 + }, + { + "epoch": 1.976947082312674, + "grad_norm": 1.1294244527816772, + "learning_rate": 3.453551748128781e-08, + "loss": 0.4562, + "step": 29458 + }, + { + "epoch": 1.977081305996443, + "grad_norm": 0.9462315440177917, + "learning_rate": 3.4132820193899514e-08, + "loss": 0.4131, + "step": 29460 + }, + { + "epoch": 1.977215529680212, + "grad_norm": 1.0207066535949707, + "learning_rate": 3.3732483695481097e-08, + "loss": 0.4648, + "step": 29462 + }, + { + "epoch": 1.9773497533639812, + "grad_norm": 0.995894193649292, + "learning_rate": 3.333450800495075e-08, + "loss": 0.4475, + "step": 29464 + }, + { + "epoch": 1.9774839770477501, + "grad_norm": 1.0989102125167847, + "learning_rate": 3.2938893141110094e-08, + "loss": 0.4478, + "step": 29466 + }, + { + "epoch": 1.977618200731519, + "grad_norm": 1.0410295724868774, + "learning_rate": 3.254563912264419e-08, + "loss": 0.5155, + "step": 29468 + }, + { + "epoch": 1.977752424415288, + "grad_norm": 0.9531170129776001, + "learning_rate": 3.215474596814372e-08, + "loss": 0.4988, + "step": 29470 + }, + { + "epoch": 1.9778866480990571, + "grad_norm": 0.8869547247886658, + "learning_rate": 3.176621369607724e-08, + "loss": 0.452, + "step": 29472 + }, + { + "epoch": 1.978020871782826, + "grad_norm": 1.021654725074768, + "learning_rate": 3.138004232479674e-08, + "loss": 0.4554, + "step": 29474 + }, + { + "epoch": 1.9781550954665952, + "grad_norm": 1.0772607326507568, + "learning_rate": 3.099623187254874e-08, + "loss": 0.4541, + "step": 29476 + }, + { + "epoch": 1.978289319150364, + "grad_norm": 1.1076545715332031, + "learning_rate": 3.061478235746873e-08, + "loss": 0.4521, + "step": 29478 + }, + { + "epoch": 1.978423542834133, + "grad_norm": 0.977824866771698, + "learning_rate": 3.023569379758118e-08, + "loss": 0.4835, + "step": 29480 + }, + { + "epoch": 1.978557766517902, + "grad_norm": 0.9521993398666382, + "learning_rate": 2.985896621079398e-08, + "loss": 0.4299, + "step": 29482 + }, + { + "epoch": 1.978691990201671, + "grad_norm": 1.0588092803955078, + "learning_rate": 2.948459961490957e-08, + "loss": 0.4646, + "step": 29484 + }, + { + "epoch": 1.9788262138854402, + "grad_norm": 1.0376704931259155, + "learning_rate": 2.9112594027619346e-08, + "loss": 0.4729, + "step": 29486 + }, + { + "epoch": 1.9789604375692091, + "grad_norm": 1.0790448188781738, + "learning_rate": 2.8742949466487036e-08, + "loss": 0.4117, + "step": 29488 + }, + { + "epoch": 1.979094661252978, + "grad_norm": 1.0049660205841064, + "learning_rate": 2.8375665948993103e-08, + "loss": 0.4509, + "step": 29490 + }, + { + "epoch": 1.979228884936747, + "grad_norm": 1.0287073850631714, + "learning_rate": 2.801074349247923e-08, + "loss": 0.446, + "step": 29492 + }, + { + "epoch": 1.9793631086205161, + "grad_norm": 0.9084295630455017, + "learning_rate": 2.7648182114198285e-08, + "loss": 0.4474, + "step": 29494 + }, + { + "epoch": 1.979497332304285, + "grad_norm": 1.1138800382614136, + "learning_rate": 2.7287981831269905e-08, + "loss": 0.4446, + "step": 29496 + }, + { + "epoch": 1.9796315559880542, + "grad_norm": 1.0034202337265015, + "learning_rate": 2.693014266071381e-08, + "loss": 0.3988, + "step": 29498 + }, + { + "epoch": 1.979765779671823, + "grad_norm": 0.9081588387489319, + "learning_rate": 2.6574664619444244e-08, + "loss": 0.4977, + "step": 29500 + }, + { + "epoch": 1.979900003355592, + "grad_norm": 1.0774911642074585, + "learning_rate": 2.6221547724253337e-08, + "loss": 0.4847, + "step": 29502 + }, + { + "epoch": 1.980034227039361, + "grad_norm": 0.9915654063224792, + "learning_rate": 2.5870791991827737e-08, + "loss": 0.529, + "step": 29504 + }, + { + "epoch": 1.98016845072313, + "grad_norm": 1.0009660720825195, + "learning_rate": 2.552239743873197e-08, + "loss": 0.4338, + "step": 29506 + }, + { + "epoch": 1.9803026744068992, + "grad_norm": 1.005157709121704, + "learning_rate": 2.51763640814362e-08, + "loss": 0.4364, + "step": 29508 + }, + { + "epoch": 1.9804368980906681, + "grad_norm": 1.1098984479904175, + "learning_rate": 2.4832691936282902e-08, + "loss": 0.4328, + "step": 29510 + }, + { + "epoch": 1.980571121774437, + "grad_norm": 0.9147427082061768, + "learning_rate": 2.4491381019520198e-08, + "loss": 0.4075, + "step": 29512 + }, + { + "epoch": 1.980705345458206, + "grad_norm": 1.0833896398544312, + "learning_rate": 2.415243134725742e-08, + "loss": 0.4841, + "step": 29514 + }, + { + "epoch": 1.9808395691419751, + "grad_norm": 0.9254489541053772, + "learning_rate": 2.381584293552619e-08, + "loss": 0.4969, + "step": 29516 + }, + { + "epoch": 1.9809737928257443, + "grad_norm": 1.0122753381729126, + "learning_rate": 2.3481615800219347e-08, + "loss": 0.4171, + "step": 29518 + }, + { + "epoch": 1.9811080165095132, + "grad_norm": 1.0806469917297363, + "learning_rate": 2.3149749957129818e-08, + "loss": 0.4285, + "step": 29520 + }, + { + "epoch": 1.981242240193282, + "grad_norm": 1.0207287073135376, + "learning_rate": 2.28202454219395e-08, + "loss": 0.4306, + "step": 29522 + }, + { + "epoch": 1.981376463877051, + "grad_norm": 1.008135199546814, + "learning_rate": 2.2493102210219275e-08, + "loss": 0.4515, + "step": 29524 + }, + { + "epoch": 1.9815106875608202, + "grad_norm": 1.120957612991333, + "learning_rate": 2.2168320337423442e-08, + "loss": 0.4642, + "step": 29526 + }, + { + "epoch": 1.981644911244589, + "grad_norm": 1.0079890489578247, + "learning_rate": 2.1845899818895287e-08, + "loss": 0.4304, + "step": 29528 + }, + { + "epoch": 1.9817791349283582, + "grad_norm": 1.0214899778366089, + "learning_rate": 2.152584066987262e-08, + "loss": 0.4601, + "step": 29530 + }, + { + "epoch": 1.9819133586121271, + "grad_norm": 0.9742206335067749, + "learning_rate": 2.120814290547668e-08, + "loss": 0.4746, + "step": 29532 + }, + { + "epoch": 1.982047582295896, + "grad_norm": 1.4604252576828003, + "learning_rate": 2.089280654071213e-08, + "loss": 0.4276, + "step": 29534 + }, + { + "epoch": 1.982181805979665, + "grad_norm": 1.1026074886322021, + "learning_rate": 2.057983159048926e-08, + "loss": 0.4672, + "step": 29536 + }, + { + "epoch": 1.9823160296634341, + "grad_norm": 1.0693613290786743, + "learning_rate": 2.026921806958515e-08, + "loss": 0.4698, + "step": 29538 + }, + { + "epoch": 1.9824502533472033, + "grad_norm": 1.0676215887069702, + "learning_rate": 1.996096599267694e-08, + "loss": 0.4791, + "step": 29540 + }, + { + "epoch": 1.9825844770309722, + "grad_norm": 0.9879465699195862, + "learning_rate": 1.965507537433631e-08, + "loss": 0.4965, + "step": 29542 + }, + { + "epoch": 1.982718700714741, + "grad_norm": 0.9821873307228088, + "learning_rate": 1.9351546229007256e-08, + "loss": 0.4378, + "step": 29544 + }, + { + "epoch": 1.98285292439851, + "grad_norm": 1.057045340538025, + "learning_rate": 1.9050378571039418e-08, + "loss": 0.4422, + "step": 29546 + }, + { + "epoch": 1.9829871480822792, + "grad_norm": 1.1257404088974, + "learning_rate": 1.875157241465475e-08, + "loss": 0.4876, + "step": 29548 + }, + { + "epoch": 1.983121371766048, + "grad_norm": 0.8850076794624329, + "learning_rate": 1.845512777397529e-08, + "loss": 0.4808, + "step": 29550 + }, + { + "epoch": 1.9832555954498172, + "grad_norm": 1.1537803411483765, + "learning_rate": 1.8161044663000948e-08, + "loss": 0.4219, + "step": 29552 + }, + { + "epoch": 1.9833898191335861, + "grad_norm": 1.0371828079223633, + "learning_rate": 1.786932309564282e-08, + "loss": 0.4715, + "step": 29554 + }, + { + "epoch": 1.983524042817355, + "grad_norm": 0.9885537028312683, + "learning_rate": 1.7579963085667672e-08, + "loss": 0.471, + "step": 29556 + }, + { + "epoch": 1.983658266501124, + "grad_norm": 0.9527798295021057, + "learning_rate": 1.7292964646753453e-08, + "loss": 0.4321, + "step": 29558 + }, + { + "epoch": 1.9837924901848931, + "grad_norm": 1.0112323760986328, + "learning_rate": 1.700832779245598e-08, + "loss": 0.4197, + "step": 29560 + }, + { + "epoch": 1.9839267138686623, + "grad_norm": 0.7867613434791565, + "learning_rate": 1.672605253623116e-08, + "loss": 0.4374, + "step": 29562 + }, + { + "epoch": 1.9840609375524312, + "grad_norm": 0.9947951436042786, + "learning_rate": 1.6446138891412777e-08, + "loss": 0.4336, + "step": 29564 + }, + { + "epoch": 1.9841951612362, + "grad_norm": 1.0333178043365479, + "learning_rate": 1.616858687122913e-08, + "loss": 0.4762, + "step": 29566 + }, + { + "epoch": 1.984329384919969, + "grad_norm": 1.0098247528076172, + "learning_rate": 1.5893396488786407e-08, + "loss": 0.4625, + "step": 29568 + }, + { + "epoch": 1.9844636086037382, + "grad_norm": 0.9545372128486633, + "learning_rate": 1.5620567757090865e-08, + "loss": 0.4488, + "step": 29570 + }, + { + "epoch": 1.984597832287507, + "grad_norm": 0.9851608276367188, + "learning_rate": 1.535010068903775e-08, + "loss": 0.4743, + "step": 29572 + }, + { + "epoch": 1.9847320559712762, + "grad_norm": 1.3361846208572388, + "learning_rate": 1.5081995297400177e-08, + "loss": 0.4902, + "step": 29574 + }, + { + "epoch": 1.9848662796550451, + "grad_norm": 1.1021267175674438, + "learning_rate": 1.4816251594845787e-08, + "loss": 0.4729, + "step": 29576 + }, + { + "epoch": 1.985000503338814, + "grad_norm": 1.0077482461929321, + "learning_rate": 1.4552869593931207e-08, + "loss": 0.4361, + "step": 29578 + }, + { + "epoch": 1.985134727022583, + "grad_norm": 1.046972632408142, + "learning_rate": 1.4291849307102034e-08, + "loss": 0.4368, + "step": 29580 + }, + { + "epoch": 1.9852689507063521, + "grad_norm": 0.9834826588630676, + "learning_rate": 1.4033190746687297e-08, + "loss": 0.5096, + "step": 29582 + }, + { + "epoch": 1.9854031743901213, + "grad_norm": 1.05916428565979, + "learning_rate": 1.37768939249161e-08, + "loss": 0.4548, + "step": 29584 + }, + { + "epoch": 1.9855373980738902, + "grad_norm": 0.9710681438446045, + "learning_rate": 1.3522958853889877e-08, + "loss": 0.5093, + "step": 29586 + }, + { + "epoch": 1.985671621757659, + "grad_norm": 0.9356110692024231, + "learning_rate": 1.3271385545610137e-08, + "loss": 0.4597, + "step": 29588 + }, + { + "epoch": 1.985805845441428, + "grad_norm": 1.1384472846984863, + "learning_rate": 1.302217401196737e-08, + "loss": 0.5089, + "step": 29590 + }, + { + "epoch": 1.9859400691251972, + "grad_norm": 0.8650492429733276, + "learning_rate": 1.2775324264724386e-08, + "loss": 0.4664, + "step": 29592 + }, + { + "epoch": 1.9860742928089663, + "grad_norm": 1.0428403615951538, + "learning_rate": 1.2530836315555183e-08, + "loss": 0.5201, + "step": 29594 + }, + { + "epoch": 1.9862085164927352, + "grad_norm": 1.044148564338684, + "learning_rate": 1.228871017601163e-08, + "loss": 0.5132, + "step": 29596 + }, + { + "epoch": 1.9863427401765041, + "grad_norm": 1.0618407726287842, + "learning_rate": 1.2048945857523475e-08, + "loss": 0.4978, + "step": 29598 + }, + { + "epoch": 1.986476963860273, + "grad_norm": 1.0523346662521362, + "learning_rate": 1.1811543371431644e-08, + "loss": 0.437, + "step": 29600 + }, + { + "epoch": 1.9866111875440422, + "grad_norm": 0.9107396006584167, + "learning_rate": 1.1576502728938287e-08, + "loss": 0.456, + "step": 29602 + }, + { + "epoch": 1.9867454112278111, + "grad_norm": 1.0899020433425903, + "learning_rate": 1.134382394116229e-08, + "loss": 0.4223, + "step": 29604 + }, + { + "epoch": 1.9868796349115803, + "grad_norm": 0.9853198528289795, + "learning_rate": 1.111350701909486e-08, + "loss": 0.4546, + "step": 29606 + }, + { + "epoch": 1.9870138585953492, + "grad_norm": 1.0362210273742676, + "learning_rate": 1.088555197361063e-08, + "loss": 0.5492, + "step": 29608 + }, + { + "epoch": 1.987148082279118, + "grad_norm": 0.9253741502761841, + "learning_rate": 1.0659958815489868e-08, + "loss": 0.4201, + "step": 29610 + }, + { + "epoch": 1.987282305962887, + "grad_norm": 1.087234377861023, + "learning_rate": 1.043672755537961e-08, + "loss": 0.4905, + "step": 29612 + }, + { + "epoch": 1.9874165296466562, + "grad_norm": 0.912765383720398, + "learning_rate": 1.021585820383808e-08, + "loss": 0.4179, + "step": 29614 + }, + { + "epoch": 1.9875507533304253, + "grad_norm": 0.8833821415901184, + "learning_rate": 9.997350771295821e-09, + "loss": 0.4319, + "step": 29616 + }, + { + "epoch": 1.9876849770141942, + "grad_norm": 1.0871175527572632, + "learning_rate": 9.781205268077908e-09, + "loss": 0.4756, + "step": 29618 + }, + { + "epoch": 1.9878192006979631, + "grad_norm": 1.0188250541687012, + "learning_rate": 9.567421704392843e-09, + "loss": 0.4835, + "step": 29620 + }, + { + "epoch": 1.987953424381732, + "grad_norm": 1.129906415939331, + "learning_rate": 9.356000090349204e-09, + "loss": 0.4576, + "step": 29622 + }, + { + "epoch": 1.9880876480655012, + "grad_norm": 0.8076863288879395, + "learning_rate": 9.146940435933449e-09, + "loss": 0.406, + "step": 29624 + }, + { + "epoch": 1.9882218717492701, + "grad_norm": 1.024215579032898, + "learning_rate": 8.94024275102101e-09, + "loss": 0.4292, + "step": 29626 + }, + { + "epoch": 1.9883560954330393, + "grad_norm": 0.9850940704345703, + "learning_rate": 8.735907045376301e-09, + "loss": 0.441, + "step": 29628 + }, + { + "epoch": 1.9884903191168082, + "grad_norm": 1.049185872077942, + "learning_rate": 8.533933328658262e-09, + "loss": 0.5439, + "step": 29630 + }, + { + "epoch": 1.988624542800577, + "grad_norm": 1.0200152397155762, + "learning_rate": 8.334321610403706e-09, + "loss": 0.4219, + "step": 29632 + }, + { + "epoch": 1.988758766484346, + "grad_norm": 1.0623215436935425, + "learning_rate": 8.137071900055082e-09, + "loss": 0.5007, + "step": 29634 + }, + { + "epoch": 1.9888929901681152, + "grad_norm": 1.196020483970642, + "learning_rate": 7.942184206921611e-09, + "loss": 0.5173, + "step": 29636 + }, + { + "epoch": 1.9890272138518843, + "grad_norm": 1.135179042816162, + "learning_rate": 7.74965854021259e-09, + "loss": 0.5123, + "step": 29638 + }, + { + "epoch": 1.9891614375356532, + "grad_norm": 1.08401358127594, + "learning_rate": 7.5594949090263e-09, + "loss": 0.5092, + "step": 29640 + }, + { + "epoch": 1.9892956612194221, + "grad_norm": 0.9641003012657166, + "learning_rate": 7.371693322349993e-09, + "loss": 0.4967, + "step": 29642 + }, + { + "epoch": 1.989429884903191, + "grad_norm": 1.1795014142990112, + "learning_rate": 7.186253789059905e-09, + "loss": 0.4712, + "step": 29644 + }, + { + "epoch": 1.9895641085869602, + "grad_norm": 1.0693312883377075, + "learning_rate": 7.003176317904591e-09, + "loss": 0.4727, + "step": 29646 + }, + { + "epoch": 1.9896983322707291, + "grad_norm": 1.0542054176330566, + "learning_rate": 6.822460917549345e-09, + "loss": 0.488, + "step": 29648 + }, + { + "epoch": 1.9898325559544983, + "grad_norm": 0.99620121717453, + "learning_rate": 6.644107596520677e-09, + "loss": 0.4666, + "step": 29650 + }, + { + "epoch": 1.9899667796382672, + "grad_norm": 1.1178016662597656, + "learning_rate": 6.4681163632507314e-09, + "loss": 0.499, + "step": 29652 + }, + { + "epoch": 1.990101003322036, + "grad_norm": 1.054388403892517, + "learning_rate": 6.294487226055079e-09, + "loss": 0.4872, + "step": 29654 + }, + { + "epoch": 1.990235227005805, + "grad_norm": 1.0646460056304932, + "learning_rate": 6.123220193132717e-09, + "loss": 0.4861, + "step": 29656 + }, + { + "epoch": 1.9903694506895742, + "grad_norm": 1.1594460010528564, + "learning_rate": 5.9543152725827226e-09, + "loss": 0.4703, + "step": 29658 + }, + { + "epoch": 1.9905036743733433, + "grad_norm": 1.0529216527938843, + "learning_rate": 5.787772472382047e-09, + "loss": 0.4481, + "step": 29660 + }, + { + "epoch": 1.9906378980571122, + "grad_norm": 0.9020050168037415, + "learning_rate": 5.623591800402173e-09, + "loss": 0.4546, + "step": 29662 + }, + { + "epoch": 1.9907721217408811, + "grad_norm": 0.9900140166282654, + "learning_rate": 5.461773264398007e-09, + "loss": 0.4591, + "step": 29664 + }, + { + "epoch": 1.99090634542465, + "grad_norm": 1.078150987625122, + "learning_rate": 5.302316872013435e-09, + "loss": 0.5246, + "step": 29666 + }, + { + "epoch": 1.9910405691084192, + "grad_norm": 0.9780765771865845, + "learning_rate": 5.145222630781321e-09, + "loss": 0.4966, + "step": 29668 + }, + { + "epoch": 1.9911747927921883, + "grad_norm": 1.0546802282333374, + "learning_rate": 4.990490548129057e-09, + "loss": 0.4696, + "step": 29670 + }, + { + "epoch": 1.9913090164759573, + "grad_norm": 0.9650538563728333, + "learning_rate": 4.838120631361909e-09, + "loss": 0.5195, + "step": 29672 + }, + { + "epoch": 1.9914432401597262, + "grad_norm": 1.0090038776397705, + "learning_rate": 4.688112887685225e-09, + "loss": 0.4585, + "step": 29674 + }, + { + "epoch": 1.991577463843495, + "grad_norm": 1.0023903846740723, + "learning_rate": 4.540467324187781e-09, + "loss": 0.4605, + "step": 29676 + }, + { + "epoch": 1.9917116875272642, + "grad_norm": 0.8548973202705383, + "learning_rate": 4.3951839478362235e-09, + "loss": 0.4265, + "step": 29678 + }, + { + "epoch": 1.9918459112110332, + "grad_norm": 1.024346947669983, + "learning_rate": 4.2522627655028346e-09, + "loss": 0.3986, + "step": 29680 + }, + { + "epoch": 1.9919801348948023, + "grad_norm": 1.0842660665512085, + "learning_rate": 4.111703783932219e-09, + "loss": 0.4498, + "step": 29682 + }, + { + "epoch": 1.9921143585785712, + "grad_norm": 0.9837660193443298, + "learning_rate": 3.973507009774613e-09, + "loss": 0.477, + "step": 29684 + }, + { + "epoch": 1.9922485822623401, + "grad_norm": 0.9357654452323914, + "learning_rate": 3.8376724495581276e-09, + "loss": 0.428, + "step": 29686 + }, + { + "epoch": 1.992382805946109, + "grad_norm": 0.9697017669677734, + "learning_rate": 3.7042001096943e-09, + "loss": 0.3818, + "step": 29688 + }, + { + "epoch": 1.9925170296298782, + "grad_norm": 1.0058552026748657, + "learning_rate": 3.5730899964947495e-09, + "loss": 0.453, + "step": 29690 + }, + { + "epoch": 1.9926512533136473, + "grad_norm": 1.1216505765914917, + "learning_rate": 3.4443421161545197e-09, + "loss": 0.4774, + "step": 29692 + }, + { + "epoch": 1.9927854769974163, + "grad_norm": 1.0027047395706177, + "learning_rate": 3.317956474757633e-09, + "loss": 0.4588, + "step": 29694 + }, + { + "epoch": 1.9929197006811852, + "grad_norm": 1.1020907163619995, + "learning_rate": 3.193933078265987e-09, + "loss": 0.4751, + "step": 29696 + }, + { + "epoch": 1.993053924364954, + "grad_norm": 1.0000810623168945, + "learning_rate": 3.0722719325526615e-09, + "loss": 0.4252, + "step": 29698 + }, + { + "epoch": 1.9931881480487232, + "grad_norm": 1.25977623462677, + "learning_rate": 2.95297304335751e-09, + "loss": 0.41, + "step": 29700 + }, + { + "epoch": 1.9933223717324922, + "grad_norm": 1.0130668878555298, + "learning_rate": 2.8360364163149135e-09, + "loss": 0.4406, + "step": 29702 + }, + { + "epoch": 1.9934565954162613, + "grad_norm": 1.1364526748657227, + "learning_rate": 2.721462056959334e-09, + "loss": 0.4426, + "step": 29704 + }, + { + "epoch": 1.9935908191000302, + "grad_norm": 0.9602410793304443, + "learning_rate": 2.609249970697558e-09, + "loss": 0.4955, + "step": 29706 + }, + { + "epoch": 1.9937250427837991, + "grad_norm": 0.9666051268577576, + "learning_rate": 2.4994001628364517e-09, + "loss": 0.5171, + "step": 29708 + }, + { + "epoch": 1.993859266467568, + "grad_norm": 1.0599029064178467, + "learning_rate": 2.391912638560756e-09, + "loss": 0.4649, + "step": 29710 + }, + { + "epoch": 1.9939934901513372, + "grad_norm": 0.8977524042129517, + "learning_rate": 2.2867874029497415e-09, + "loss": 0.4292, + "step": 29712 + }, + { + "epoch": 1.9941277138351063, + "grad_norm": 1.0639896392822266, + "learning_rate": 2.1840244609716565e-09, + "loss": 0.4996, + "step": 29714 + }, + { + "epoch": 1.9942619375188753, + "grad_norm": 1.0090793371200562, + "learning_rate": 2.0836238174837264e-09, + "loss": 0.4571, + "step": 29716 + }, + { + "epoch": 1.9943961612026442, + "grad_norm": 1.0224015712738037, + "learning_rate": 1.9855854772266037e-09, + "loss": 0.4612, + "step": 29718 + }, + { + "epoch": 1.994530384886413, + "grad_norm": 1.0634790658950806, + "learning_rate": 1.8899094448354693e-09, + "loss": 0.4556, + "step": 29720 + }, + { + "epoch": 1.9946646085701822, + "grad_norm": 1.0072276592254639, + "learning_rate": 1.796595724828931e-09, + "loss": 0.4535, + "step": 29722 + }, + { + "epoch": 1.9947988322539512, + "grad_norm": 1.0097860097885132, + "learning_rate": 1.7056443216145746e-09, + "loss": 0.458, + "step": 29724 + }, + { + "epoch": 1.9949330559377203, + "grad_norm": 1.0248351097106934, + "learning_rate": 1.6170552394889627e-09, + "loss": 0.4061, + "step": 29726 + }, + { + "epoch": 1.9950672796214892, + "grad_norm": 0.9212617874145508, + "learning_rate": 1.5308284826431874e-09, + "loss": 0.4112, + "step": 29728 + }, + { + "epoch": 1.9952015033052581, + "grad_norm": 1.416978120803833, + "learning_rate": 1.446964055146216e-09, + "loss": 0.491, + "step": 29730 + }, + { + "epoch": 1.995335726989027, + "grad_norm": 1.0837624073028564, + "learning_rate": 1.365461960961545e-09, + "loss": 0.465, + "step": 29732 + }, + { + "epoch": 1.9954699506727962, + "grad_norm": 1.1062414646148682, + "learning_rate": 1.2863222039416478e-09, + "loss": 0.4684, + "step": 29734 + }, + { + "epoch": 1.9956041743565653, + "grad_norm": 1.0122276544570923, + "learning_rate": 1.2095447878279765e-09, + "loss": 0.4579, + "step": 29736 + }, + { + "epoch": 1.9957383980403343, + "grad_norm": 1.0660096406936646, + "learning_rate": 1.1351297162398578e-09, + "loss": 0.4577, + "step": 29738 + }, + { + "epoch": 1.9958726217241032, + "grad_norm": 0.970309317111969, + "learning_rate": 1.06307699270225e-09, + "loss": 0.4796, + "step": 29740 + }, + { + "epoch": 1.996006845407872, + "grad_norm": 0.8794416785240173, + "learning_rate": 9.933866206124353e-10, + "loss": 0.4244, + "step": 29742 + }, + { + "epoch": 1.9961410690916412, + "grad_norm": 0.9972506165504456, + "learning_rate": 9.260586032677765e-10, + "loss": 0.4375, + "step": 29744 + }, + { + "epoch": 1.9962752927754104, + "grad_norm": 1.0356035232543945, + "learning_rate": 8.610929438490623e-10, + "loss": 0.4301, + "step": 29746 + }, + { + "epoch": 1.9964095164591793, + "grad_norm": 1.0438393354415894, + "learning_rate": 7.984896454260593e-10, + "loss": 0.4381, + "step": 29748 + }, + { + "epoch": 1.9965437401429482, + "grad_norm": 0.9698221683502197, + "learning_rate": 7.382487109519609e-10, + "loss": 0.4574, + "step": 29750 + }, + { + "epoch": 1.9966779638267171, + "grad_norm": 0.9586147665977478, + "learning_rate": 6.803701432744891e-10, + "loss": 0.4187, + "step": 29752 + }, + { + "epoch": 1.9968121875104863, + "grad_norm": 1.0892906188964844, + "learning_rate": 6.248539451358948e-10, + "loss": 0.5244, + "step": 29754 + }, + { + "epoch": 1.9969464111942552, + "grad_norm": 0.9858735799789429, + "learning_rate": 5.71700119145202e-10, + "loss": 0.4165, + "step": 29756 + }, + { + "epoch": 1.9970806348780243, + "grad_norm": 1.0821709632873535, + "learning_rate": 5.209086678281683e-10, + "loss": 0.4559, + "step": 29758 + }, + { + "epoch": 1.9972148585617933, + "grad_norm": 1.021770715713501, + "learning_rate": 4.724795935773241e-10, + "loss": 0.5047, + "step": 29760 + }, + { + "epoch": 1.9973490822455622, + "grad_norm": 1.1175012588500977, + "learning_rate": 4.2641289868528e-10, + "loss": 0.57, + "step": 29762 + }, + { + "epoch": 1.997483305929331, + "grad_norm": 1.196708083152771, + "learning_rate": 3.82708585316971e-10, + "loss": 0.4568, + "step": 29764 + }, + { + "epoch": 1.9976175296131002, + "grad_norm": 1.007369875907898, + "learning_rate": 3.413666555540651e-10, + "loss": 0.5123, + "step": 29766 + }, + { + "epoch": 1.9977517532968694, + "grad_norm": 1.0676368474960327, + "learning_rate": 3.023871113339016e-10, + "loss": 0.4617, + "step": 29768 + }, + { + "epoch": 1.9978859769806383, + "grad_norm": 1.0238417387008667, + "learning_rate": 2.65769954510553e-10, + "loss": 0.4486, + "step": 29770 + }, + { + "epoch": 1.9980202006644072, + "grad_norm": 1.0677101612091064, + "learning_rate": 2.3151518681041595e-10, + "loss": 0.4439, + "step": 29772 + }, + { + "epoch": 1.9981544243481761, + "grad_norm": 1.0551823377609253, + "learning_rate": 1.9962280984886507e-10, + "loss": 0.4729, + "step": 29774 + }, + { + "epoch": 1.9982886480319453, + "grad_norm": 1.0276211500167847, + "learning_rate": 1.700928251358036e-10, + "loss": 0.4413, + "step": 29776 + }, + { + "epoch": 1.9984228717157142, + "grad_norm": 1.2995710372924805, + "learning_rate": 1.429252340645615e-10, + "loss": 0.5142, + "step": 29778 + }, + { + "epoch": 1.9985570953994833, + "grad_norm": 1.126254916191101, + "learning_rate": 1.181200379174463e-10, + "loss": 0.4627, + "step": 29780 + }, + { + "epoch": 1.9986913190832523, + "grad_norm": 0.997976541519165, + "learning_rate": 9.567723787129445e-11, + "loss": 0.4312, + "step": 29782 + }, + { + "epoch": 1.9988255427670212, + "grad_norm": 0.9253402352333069, + "learning_rate": 7.559683498081782e-11, + "loss": 0.4378, + "step": 29784 + }, + { + "epoch": 1.99895976645079, + "grad_norm": 1.0526634454727173, + "learning_rate": 5.7878830200808196e-11, + "loss": 0.4286, + "step": 29786 + }, + { + "epoch": 1.9990939901345592, + "grad_norm": 1.0962398052215576, + "learning_rate": 4.2523224363932856e-11, + "loss": 0.4924, + "step": 29788 + }, + { + "epoch": 1.9992282138183284, + "grad_norm": 1.120545506477356, + "learning_rate": 2.953001819738788e-11, + "loss": 0.4837, + "step": 29790 + }, + { + "epoch": 1.9993624375020973, + "grad_norm": 1.0252364873886108, + "learning_rate": 1.8899212317347036e-11, + "loss": 0.4587, + "step": 29792 + }, + { + "epoch": 1.9994966611858662, + "grad_norm": 0.8036876320838928, + "learning_rate": 1.06308072234107e-11, + "loss": 0.4413, + "step": 29794 + }, + { + "epoch": 1.9996308848696351, + "grad_norm": 0.9679793119430542, + "learning_rate": 4.7248033041569216e-12, + "loss": 0.5315, + "step": 29796 + }, + { + "epoch": 1.9997651085534043, + "grad_norm": 1.013262391090393, + "learning_rate": 1.1812008371414608e-12, + "loss": 0.4365, + "step": 29798 + }, + { + "epoch": 1.9998993322371732, + "grad_norm": 1.0657564401626587, + "learning_rate": 0.0, + "loss": 0.4019, + "step": 29800 + } + ], + "logging_steps": 2, + "max_steps": 29800, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.9092202292301005e+18, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +}