possibly-working-glm / trainer_state.json
ToastyPigeon's picture
Upload folder using huggingface_hub
d50c568 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9950738916256157,
"eval_steps": 76,
"global_step": 608,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003284072249589491,
"grad_norm": 6271.83657859153,
"learning_rate": 0.0,
"loss": 0.7244,
"step": 1
},
{
"epoch": 0.003284072249589491,
"eval_loss": 0.5559563636779785,
"eval_runtime": 253.2891,
"eval_samples_per_second": 8.157,
"eval_steps_per_second": 0.257,
"step": 1
},
{
"epoch": 0.006568144499178982,
"grad_norm": 5855.051575469832,
"learning_rate": 3.3333333333333334e-08,
"loss": 0.7743,
"step": 2
},
{
"epoch": 0.009852216748768473,
"grad_norm": 5966.931878891552,
"learning_rate": 6.666666666666667e-08,
"loss": 0.7495,
"step": 3
},
{
"epoch": 0.013136288998357963,
"grad_norm": 2146.4401329128254,
"learning_rate": 1e-07,
"loss": 0.7853,
"step": 4
},
{
"epoch": 0.016420361247947456,
"grad_norm": 1479.4537989114276,
"learning_rate": 1.3333333333333334e-07,
"loss": 0.7486,
"step": 5
},
{
"epoch": 0.019704433497536946,
"grad_norm": 2073.875643154414,
"learning_rate": 1.6666666666666665e-07,
"loss": 0.7395,
"step": 6
},
{
"epoch": 0.022988505747126436,
"grad_norm": 3306.267015593175,
"learning_rate": 2e-07,
"loss": 0.7873,
"step": 7
},
{
"epoch": 0.026272577996715927,
"grad_norm": 5570.427974319344,
"learning_rate": 2.3333333333333333e-07,
"loss": 0.7897,
"step": 8
},
{
"epoch": 0.029556650246305417,
"grad_norm": 5192.155077763508,
"learning_rate": 2.6666666666666667e-07,
"loss": 0.7411,
"step": 9
},
{
"epoch": 0.03284072249589491,
"grad_norm": 3385.4892242590813,
"learning_rate": 3e-07,
"loss": 0.7656,
"step": 10
},
{
"epoch": 0.0361247947454844,
"grad_norm": 1680.342831450108,
"learning_rate": 3.333333333333333e-07,
"loss": 0.7801,
"step": 11
},
{
"epoch": 0.03940886699507389,
"grad_norm": 4799.608669273176,
"learning_rate": 3.666666666666666e-07,
"loss": 0.7407,
"step": 12
},
{
"epoch": 0.042692939244663386,
"grad_norm": 2371.102370606964,
"learning_rate": 4e-07,
"loss": 0.7861,
"step": 13
},
{
"epoch": 0.04597701149425287,
"grad_norm": 22928.120442114872,
"learning_rate": 4.3333333333333335e-07,
"loss": 0.7326,
"step": 14
},
{
"epoch": 0.04926108374384237,
"grad_norm": 5866.1404741418755,
"learning_rate": 4.6666666666666666e-07,
"loss": 0.7236,
"step": 15
},
{
"epoch": 0.052545155993431854,
"grad_norm": 10335.132176960362,
"learning_rate": 5e-07,
"loss": 0.7353,
"step": 16
},
{
"epoch": 0.05582922824302135,
"grad_norm": 11731.708675839953,
"learning_rate": 5.333333333333333e-07,
"loss": 0.7368,
"step": 17
},
{
"epoch": 0.059113300492610835,
"grad_norm": 8477.833222009613,
"learning_rate": 5.666666666666666e-07,
"loss": 0.7401,
"step": 18
},
{
"epoch": 0.06239737274220033,
"grad_norm": 5269.214248760142,
"learning_rate": 6e-07,
"loss": 0.7378,
"step": 19
},
{
"epoch": 0.06568144499178982,
"grad_norm": 2906.1574452333357,
"learning_rate": 6.333333333333332e-07,
"loss": 0.7661,
"step": 20
},
{
"epoch": 0.06896551724137931,
"grad_norm": 2376.80127554879,
"learning_rate": 6.666666666666666e-07,
"loss": 0.7457,
"step": 21
},
{
"epoch": 0.0722495894909688,
"grad_norm": 1796.841170343826,
"learning_rate": 7e-07,
"loss": 0.7252,
"step": 22
},
{
"epoch": 0.0755336617405583,
"grad_norm": 4583.563298783243,
"learning_rate": 7.333333333333332e-07,
"loss": 0.7417,
"step": 23
},
{
"epoch": 0.07881773399014778,
"grad_norm": 5571.34039922786,
"learning_rate": 7.666666666666667e-07,
"loss": 0.7449,
"step": 24
},
{
"epoch": 0.08210180623973727,
"grad_norm": 1788.7468612154946,
"learning_rate": 8e-07,
"loss": 0.7515,
"step": 25
},
{
"epoch": 0.08538587848932677,
"grad_norm": 3448.4947901602436,
"learning_rate": 8.333333333333333e-07,
"loss": 0.6949,
"step": 26
},
{
"epoch": 0.08866995073891626,
"grad_norm": 1036.8720881203337,
"learning_rate": 8.666666666666667e-07,
"loss": 0.7494,
"step": 27
},
{
"epoch": 0.09195402298850575,
"grad_norm": 2185.0194951434873,
"learning_rate": 9e-07,
"loss": 0.7197,
"step": 28
},
{
"epoch": 0.09523809523809523,
"grad_norm": 2269.166582616754,
"learning_rate": 9.333333333333333e-07,
"loss": 0.7033,
"step": 29
},
{
"epoch": 0.09852216748768473,
"grad_norm": 11624.24252175601,
"learning_rate": 9.666666666666666e-07,
"loss": 0.6986,
"step": 30
},
{
"epoch": 0.10180623973727422,
"grad_norm": 4520.614291277772,
"learning_rate": 1e-06,
"loss": 0.7107,
"step": 31
},
{
"epoch": 0.10509031198686371,
"grad_norm": 13701.111972610783,
"learning_rate": 1.0333333333333333e-06,
"loss": 0.7159,
"step": 32
},
{
"epoch": 0.10837438423645321,
"grad_norm": 7449.497627677214,
"learning_rate": 1.0666666666666667e-06,
"loss": 0.7135,
"step": 33
},
{
"epoch": 0.1116584564860427,
"grad_norm": 2245.9408876416637,
"learning_rate": 1.1e-06,
"loss": 0.6697,
"step": 34
},
{
"epoch": 0.11494252873563218,
"grad_norm": 926.1867127034129,
"learning_rate": 1.1333333333333332e-06,
"loss": 0.6825,
"step": 35
},
{
"epoch": 0.11822660098522167,
"grad_norm": 1329.9772406233828,
"learning_rate": 1.1666666666666668e-06,
"loss": 0.6795,
"step": 36
},
{
"epoch": 0.12151067323481117,
"grad_norm": 1000.9622533410908,
"learning_rate": 1.2e-06,
"loss": 0.7306,
"step": 37
},
{
"epoch": 0.12479474548440066,
"grad_norm": 993.8780279192466,
"learning_rate": 1.2333333333333333e-06,
"loss": 0.6348,
"step": 38
},
{
"epoch": 0.12807881773399016,
"grad_norm": 945.1494487867217,
"learning_rate": 1.2666666666666665e-06,
"loss": 0.6355,
"step": 39
},
{
"epoch": 0.13136288998357964,
"grad_norm": 179.6616198157567,
"learning_rate": 1.3e-06,
"loss": 0.65,
"step": 40
},
{
"epoch": 0.13464696223316913,
"grad_norm": 742.7710802743131,
"learning_rate": 1.3333333333333332e-06,
"loss": 0.6482,
"step": 41
},
{
"epoch": 0.13793103448275862,
"grad_norm": 587.9958891842592,
"learning_rate": 1.3666666666666666e-06,
"loss": 0.6399,
"step": 42
},
{
"epoch": 0.1412151067323481,
"grad_norm": 865.3394610899527,
"learning_rate": 1.4e-06,
"loss": 0.6455,
"step": 43
},
{
"epoch": 0.1444991789819376,
"grad_norm": 433.76765611396485,
"learning_rate": 1.4333333333333333e-06,
"loss": 0.6863,
"step": 44
},
{
"epoch": 0.1477832512315271,
"grad_norm": 112.10192933210595,
"learning_rate": 1.4666666666666665e-06,
"loss": 0.645,
"step": 45
},
{
"epoch": 0.1510673234811166,
"grad_norm": 2442.2715722240287,
"learning_rate": 1.5e-06,
"loss": 0.6667,
"step": 46
},
{
"epoch": 0.15435139573070608,
"grad_norm": 319.9299141888628,
"learning_rate": 1.5333333333333334e-06,
"loss": 0.631,
"step": 47
},
{
"epoch": 0.15763546798029557,
"grad_norm": 158.95826958739744,
"learning_rate": 1.5666666666666666e-06,
"loss": 0.6113,
"step": 48
},
{
"epoch": 0.16091954022988506,
"grad_norm": 215.59873480353468,
"learning_rate": 1.6e-06,
"loss": 0.6124,
"step": 49
},
{
"epoch": 0.16420361247947454,
"grad_norm": 126.6284211890604,
"learning_rate": 1.6333333333333333e-06,
"loss": 0.6154,
"step": 50
},
{
"epoch": 0.16748768472906403,
"grad_norm": 430.61576491190806,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.6379,
"step": 51
},
{
"epoch": 0.17077175697865354,
"grad_norm": 338.16595207485216,
"learning_rate": 1.6999999999999998e-06,
"loss": 0.6133,
"step": 52
},
{
"epoch": 0.17405582922824303,
"grad_norm": 182.98276970174174,
"learning_rate": 1.7333333333333334e-06,
"loss": 0.6288,
"step": 53
},
{
"epoch": 0.17733990147783252,
"grad_norm": 183.8220113000725,
"learning_rate": 1.7666666666666666e-06,
"loss": 0.6234,
"step": 54
},
{
"epoch": 0.180623973727422,
"grad_norm": 66.92408391602912,
"learning_rate": 1.8e-06,
"loss": 0.6182,
"step": 55
},
{
"epoch": 0.1839080459770115,
"grad_norm": 203.51503397263764,
"learning_rate": 1.833333333333333e-06,
"loss": 0.6466,
"step": 56
},
{
"epoch": 0.18719211822660098,
"grad_norm": 63.90718235472869,
"learning_rate": 1.8666666666666667e-06,
"loss": 0.6321,
"step": 57
},
{
"epoch": 0.19047619047619047,
"grad_norm": 46.03571035353064,
"learning_rate": 1.8999999999999998e-06,
"loss": 0.5958,
"step": 58
},
{
"epoch": 0.19376026272577998,
"grad_norm": 47.046353460290206,
"learning_rate": 1.933333333333333e-06,
"loss": 0.6416,
"step": 59
},
{
"epoch": 0.19704433497536947,
"grad_norm": 72.74885795414816,
"learning_rate": 1.9666666666666663e-06,
"loss": 0.6279,
"step": 60
},
{
"epoch": 0.20032840722495895,
"grad_norm": 43.09881595994001,
"learning_rate": 2e-06,
"loss": 0.6432,
"step": 61
},
{
"epoch": 0.20361247947454844,
"grad_norm": 385.8401754134751,
"learning_rate": 1.9999835673561283e-06,
"loss": 0.6153,
"step": 62
},
{
"epoch": 0.20689655172413793,
"grad_norm": 55.270632684943024,
"learning_rate": 1.999934269964577e-06,
"loss": 0.606,
"step": 63
},
{
"epoch": 0.21018062397372742,
"grad_norm": 112.35685565413114,
"learning_rate": 1.9998521094455194e-06,
"loss": 0.6397,
"step": 64
},
{
"epoch": 0.2134646962233169,
"grad_norm": 145.4743296112891,
"learning_rate": 1.999737088499184e-06,
"loss": 0.6133,
"step": 65
},
{
"epoch": 0.21674876847290642,
"grad_norm": 92.85749788738731,
"learning_rate": 1.9995892109057674e-06,
"loss": 0.59,
"step": 66
},
{
"epoch": 0.2200328407224959,
"grad_norm": 119.32804840506846,
"learning_rate": 1.9994084815253095e-06,
"loss": 0.6377,
"step": 67
},
{
"epoch": 0.2233169129720854,
"grad_norm": 337.56982786001447,
"learning_rate": 1.9991949062975332e-06,
"loss": 0.5827,
"step": 68
},
{
"epoch": 0.22660098522167488,
"grad_norm": 159.57391171072265,
"learning_rate": 1.99894849224165e-06,
"loss": 0.5926,
"step": 69
},
{
"epoch": 0.22988505747126436,
"grad_norm": 94.69383065711324,
"learning_rate": 1.998669247456129e-06,
"loss": 0.6249,
"step": 70
},
{
"epoch": 0.23316912972085385,
"grad_norm": 30.865025764067703,
"learning_rate": 1.9983571811184294e-06,
"loss": 0.615,
"step": 71
},
{
"epoch": 0.23645320197044334,
"grad_norm": 365.12385141652493,
"learning_rate": 1.9980123034847023e-06,
"loss": 0.6231,
"step": 72
},
{
"epoch": 0.23973727422003285,
"grad_norm": 24.986994644338345,
"learning_rate": 1.99763462588945e-06,
"loss": 0.6005,
"step": 73
},
{
"epoch": 0.24302134646962234,
"grad_norm": 156.97247949274475,
"learning_rate": 1.997224160745155e-06,
"loss": 0.5986,
"step": 74
},
{
"epoch": 0.24630541871921183,
"grad_norm": 104.35232755227744,
"learning_rate": 1.9967809215418726e-06,
"loss": 0.616,
"step": 75
},
{
"epoch": 0.24958949096880131,
"grad_norm": 174.47979716876907,
"learning_rate": 1.996304922846787e-06,
"loss": 0.5997,
"step": 76
},
{
"epoch": 0.24958949096880131,
"eval_loss": 0.37612438201904297,
"eval_runtime": 252.9482,
"eval_samples_per_second": 8.168,
"eval_steps_per_second": 0.257,
"step": 76
},
{
"epoch": 0.25287356321839083,
"grad_norm": 633.6612254567843,
"learning_rate": 1.9957961803037326e-06,
"loss": 0.5899,
"step": 77
},
{
"epoch": 0.2561576354679803,
"grad_norm": 138.79302081360208,
"learning_rate": 1.995254710632678e-06,
"loss": 0.638,
"step": 78
},
{
"epoch": 0.2594417077175698,
"grad_norm": 17.771737220416696,
"learning_rate": 1.994680531629181e-06,
"loss": 0.6313,
"step": 79
},
{
"epoch": 0.2627257799671593,
"grad_norm": 123.9631655047847,
"learning_rate": 1.9940736621637997e-06,
"loss": 0.6279,
"step": 80
},
{
"epoch": 0.2660098522167488,
"grad_norm": 203.7102057464755,
"learning_rate": 1.9934341221814736e-06,
"loss": 0.6131,
"step": 81
},
{
"epoch": 0.26929392446633826,
"grad_norm": 79.30066676169302,
"learning_rate": 1.992761932700868e-06,
"loss": 0.5929,
"step": 82
},
{
"epoch": 0.27257799671592775,
"grad_norm": 128.16665832374028,
"learning_rate": 1.9920571158136835e-06,
"loss": 0.6162,
"step": 83
},
{
"epoch": 0.27586206896551724,
"grad_norm": 100.90994251994853,
"learning_rate": 1.99131969468393e-06,
"loss": 0.5838,
"step": 84
},
{
"epoch": 0.2791461412151067,
"grad_norm": 68.97227461151734,
"learning_rate": 1.990549693547166e-06,
"loss": 0.5783,
"step": 85
},
{
"epoch": 0.2824302134646962,
"grad_norm": 65.63550727300574,
"learning_rate": 1.989747137709699e-06,
"loss": 0.5717,
"step": 86
},
{
"epoch": 0.2857142857142857,
"grad_norm": 105.48653502246952,
"learning_rate": 1.988912053547758e-06,
"loss": 0.6336,
"step": 87
},
{
"epoch": 0.2889983579638752,
"grad_norm": 151.45819328204334,
"learning_rate": 1.988044468506625e-06,
"loss": 0.5966,
"step": 88
},
{
"epoch": 0.2922824302134647,
"grad_norm": 73.02161989799154,
"learning_rate": 1.9871444110997308e-06,
"loss": 0.6377,
"step": 89
},
{
"epoch": 0.2955665024630542,
"grad_norm": 55.00481114192419,
"learning_rate": 1.9862119109077224e-06,
"loss": 0.6023,
"step": 90
},
{
"epoch": 0.2988505747126437,
"grad_norm": 30.849915628138252,
"learning_rate": 1.985246998577486e-06,
"loss": 0.5983,
"step": 91
},
{
"epoch": 0.3021346469622332,
"grad_norm": 105.68162830661484,
"learning_rate": 1.9842497058211427e-06,
"loss": 0.6191,
"step": 92
},
{
"epoch": 0.3054187192118227,
"grad_norm": 44.85816714268683,
"learning_rate": 1.9832200654150074e-06,
"loss": 0.6019,
"step": 93
},
{
"epoch": 0.30870279146141216,
"grad_norm": 94.09459294470936,
"learning_rate": 1.982158111198507e-06,
"loss": 0.6224,
"step": 94
},
{
"epoch": 0.31198686371100165,
"grad_norm": 72.91849721086875,
"learning_rate": 1.9810638780730727e-06,
"loss": 0.6135,
"step": 95
},
{
"epoch": 0.31527093596059114,
"grad_norm": 155.69433676198233,
"learning_rate": 1.979937402000991e-06,
"loss": 0.5619,
"step": 96
},
{
"epoch": 0.3185550082101806,
"grad_norm": 91.64919893957587,
"learning_rate": 1.978778720004222e-06,
"loss": 0.5789,
"step": 97
},
{
"epoch": 0.3218390804597701,
"grad_norm": 58.767110092919424,
"learning_rate": 1.9775878701631836e-06,
"loss": 0.6161,
"step": 98
},
{
"epoch": 0.3251231527093596,
"grad_norm": 86.39938638182066,
"learning_rate": 1.976364891615498e-06,
"loss": 0.5871,
"step": 99
},
{
"epoch": 0.3284072249589491,
"grad_norm": 30.43501640235004,
"learning_rate": 1.975109824554707e-06,
"loss": 0.6057,
"step": 100
},
{
"epoch": 0.33169129720853857,
"grad_norm": 65.6953811514428,
"learning_rate": 1.9738227102289505e-06,
"loss": 0.5836,
"step": 101
},
{
"epoch": 0.33497536945812806,
"grad_norm": 34.69550148683274,
"learning_rate": 1.972503590939612e-06,
"loss": 0.6239,
"step": 102
},
{
"epoch": 0.33825944170771755,
"grad_norm": 76.34647471342404,
"learning_rate": 1.971152510039926e-06,
"loss": 0.6006,
"step": 103
},
{
"epoch": 0.3415435139573071,
"grad_norm": 116.3926642735392,
"learning_rate": 1.9697695119335546e-06,
"loss": 0.5907,
"step": 104
},
{
"epoch": 0.3448275862068966,
"grad_norm": 26.910647996197213,
"learning_rate": 1.968354642073129e-06,
"loss": 0.5843,
"step": 105
},
{
"epoch": 0.34811165845648606,
"grad_norm": 125.40744182953463,
"learning_rate": 1.9669079469587545e-06,
"loss": 0.5907,
"step": 106
},
{
"epoch": 0.35139573070607555,
"grad_norm": 40.15061266508225,
"learning_rate": 1.965429474136482e-06,
"loss": 0.6412,
"step": 107
},
{
"epoch": 0.35467980295566504,
"grad_norm": 31.1618141731111,
"learning_rate": 1.963919272196746e-06,
"loss": 0.6126,
"step": 108
},
{
"epoch": 0.3579638752052545,
"grad_norm": 114.23556602704848,
"learning_rate": 1.962377390772768e-06,
"loss": 0.6034,
"step": 109
},
{
"epoch": 0.361247947454844,
"grad_norm": 40.58756337364396,
"learning_rate": 1.960803880538925e-06,
"loss": 0.6223,
"step": 110
},
{
"epoch": 0.3645320197044335,
"grad_norm": 30.422688962559107,
"learning_rate": 1.9591987932090833e-06,
"loss": 0.5825,
"step": 111
},
{
"epoch": 0.367816091954023,
"grad_norm": 51.96081689297722,
"learning_rate": 1.9575621815348996e-06,
"loss": 0.5922,
"step": 112
},
{
"epoch": 0.37110016420361247,
"grad_norm": 151.3652907897673,
"learning_rate": 1.9558940993040883e-06,
"loss": 0.6014,
"step": 113
},
{
"epoch": 0.37438423645320196,
"grad_norm": 179.13984258713396,
"learning_rate": 1.9541946013386506e-06,
"loss": 0.5812,
"step": 114
},
{
"epoch": 0.37766830870279144,
"grad_norm": 209.51648981274894,
"learning_rate": 1.9524637434930776e-06,
"loss": 0.5744,
"step": 115
},
{
"epoch": 0.38095238095238093,
"grad_norm": 77.78117906962474,
"learning_rate": 1.950701582652509e-06,
"loss": 0.5924,
"step": 116
},
{
"epoch": 0.3842364532019704,
"grad_norm": 56.596618911823576,
"learning_rate": 1.9489081767308697e-06,
"loss": 0.5951,
"step": 117
},
{
"epoch": 0.38752052545155996,
"grad_norm": 25.934351726934093,
"learning_rate": 1.9470835846689596e-06,
"loss": 0.6187,
"step": 118
},
{
"epoch": 0.39080459770114945,
"grad_norm": 78.54544999607046,
"learning_rate": 1.9452278664325227e-06,
"loss": 0.609,
"step": 119
},
{
"epoch": 0.39408866995073893,
"grad_norm": 79.08588241935573,
"learning_rate": 1.943341083010272e-06,
"loss": 0.56,
"step": 120
},
{
"epoch": 0.3973727422003284,
"grad_norm": 72.17985855738158,
"learning_rate": 1.9414232964118892e-06,
"loss": 0.5955,
"step": 121
},
{
"epoch": 0.4006568144499179,
"grad_norm": 56.33038272717607,
"learning_rate": 1.9394745696659807e-06,
"loss": 0.5745,
"step": 122
},
{
"epoch": 0.4039408866995074,
"grad_norm": 101.16342455706896,
"learning_rate": 1.9374949668180134e-06,
"loss": 0.5891,
"step": 123
},
{
"epoch": 0.4072249589490969,
"grad_norm": 67.57709272161114,
"learning_rate": 1.935484552928204e-06,
"loss": 0.6101,
"step": 124
},
{
"epoch": 0.41050903119868637,
"grad_norm": 22.702908704179823,
"learning_rate": 1.9334433940693826e-06,
"loss": 0.5904,
"step": 125
},
{
"epoch": 0.41379310344827586,
"grad_norm": 112.85103846233335,
"learning_rate": 1.9313715573248235e-06,
"loss": 0.5888,
"step": 126
},
{
"epoch": 0.41707717569786534,
"grad_norm": 74.44347415115806,
"learning_rate": 1.929269110786037e-06,
"loss": 0.6,
"step": 127
},
{
"epoch": 0.42036124794745483,
"grad_norm": 129.01611720733342,
"learning_rate": 1.9271361235505337e-06,
"loss": 0.5721,
"step": 128
},
{
"epoch": 0.4236453201970443,
"grad_norm": 21.624403814788998,
"learning_rate": 1.9249726657195533e-06,
"loss": 0.5868,
"step": 129
},
{
"epoch": 0.4269293924466338,
"grad_norm": 43.134924014466655,
"learning_rate": 1.9227788083957586e-06,
"loss": 0.5981,
"step": 130
},
{
"epoch": 0.4302134646962233,
"grad_norm": 102.45985784089031,
"learning_rate": 1.9205546236809032e-06,
"loss": 0.5906,
"step": 131
},
{
"epoch": 0.43349753694581283,
"grad_norm": 60.949447673351266,
"learning_rate": 1.9183001846734574e-06,
"loss": 0.6352,
"step": 132
},
{
"epoch": 0.4367816091954023,
"grad_norm": 401.41383723285713,
"learning_rate": 1.9160155654662073e-06,
"loss": 0.5842,
"step": 133
},
{
"epoch": 0.4400656814449918,
"grad_norm": 35.25770567828027,
"learning_rate": 1.913700841143821e-06,
"loss": 0.5862,
"step": 134
},
{
"epoch": 0.4433497536945813,
"grad_norm": 914.3282302161468,
"learning_rate": 1.9113560877803796e-06,
"loss": 0.5763,
"step": 135
},
{
"epoch": 0.4466338259441708,
"grad_norm": 35.54605377595188,
"learning_rate": 1.908981382436876e-06,
"loss": 0.5727,
"step": 136
},
{
"epoch": 0.44991789819376027,
"grad_norm": 43.4767061362858,
"learning_rate": 1.906576803158686e-06,
"loss": 0.5969,
"step": 137
},
{
"epoch": 0.45320197044334976,
"grad_norm": 23.151864647002014,
"learning_rate": 1.904142428972999e-06,
"loss": 0.5726,
"step": 138
},
{
"epoch": 0.45648604269293924,
"grad_norm": 19.480214528552803,
"learning_rate": 1.9016783398862226e-06,
"loss": 0.581,
"step": 139
},
{
"epoch": 0.45977011494252873,
"grad_norm": 38.790487317877094,
"learning_rate": 1.8991846168813544e-06,
"loss": 0.6188,
"step": 140
},
{
"epoch": 0.4630541871921182,
"grad_norm": 27.00623790662471,
"learning_rate": 1.8966613419153178e-06,
"loss": 0.6242,
"step": 141
},
{
"epoch": 0.4663382594417077,
"grad_norm": 22.367501221325167,
"learning_rate": 1.8941085979162713e-06,
"loss": 0.6027,
"step": 142
},
{
"epoch": 0.4696223316912972,
"grad_norm": 447.12696125749153,
"learning_rate": 1.8915264687808804e-06,
"loss": 0.5907,
"step": 143
},
{
"epoch": 0.4729064039408867,
"grad_norm": 32.44003369156479,
"learning_rate": 1.8889150393715625e-06,
"loss": 0.6062,
"step": 144
},
{
"epoch": 0.47619047619047616,
"grad_norm": 144.01477406191725,
"learning_rate": 1.8862743955136963e-06,
"loss": 0.6012,
"step": 145
},
{
"epoch": 0.4794745484400657,
"grad_norm": 75.34671278459541,
"learning_rate": 1.8836046239928022e-06,
"loss": 0.5843,
"step": 146
},
{
"epoch": 0.4827586206896552,
"grad_norm": 46.75651484429922,
"learning_rate": 1.8809058125516893e-06,
"loss": 0.5883,
"step": 147
},
{
"epoch": 0.4860426929392447,
"grad_norm": 92.53132858786016,
"learning_rate": 1.8781780498875723e-06,
"loss": 0.5989,
"step": 148
},
{
"epoch": 0.48932676518883417,
"grad_norm": 22.671976729298738,
"learning_rate": 1.875421425649156e-06,
"loss": 0.6094,
"step": 149
},
{
"epoch": 0.49261083743842365,
"grad_norm": 26.928492455992494,
"learning_rate": 1.8726360304336893e-06,
"loss": 0.5949,
"step": 150
},
{
"epoch": 0.49589490968801314,
"grad_norm": 286.98230638362656,
"learning_rate": 1.8698219557839872e-06,
"loss": 0.595,
"step": 151
},
{
"epoch": 0.49917898193760263,
"grad_norm": 79.72951321359584,
"learning_rate": 1.8669792941854229e-06,
"loss": 0.6124,
"step": 152
},
{
"epoch": 0.49917898193760263,
"eval_loss": 0.35242295265197754,
"eval_runtime": 254.4417,
"eval_samples_per_second": 8.12,
"eval_steps_per_second": 0.255,
"step": 152
},
{
"epoch": 0.5024630541871922,
"grad_norm": 64.33414761071143,
"learning_rate": 1.8641081390628876e-06,
"loss": 0.585,
"step": 153
},
{
"epoch": 0.5057471264367817,
"grad_norm": 31.517353545501283,
"learning_rate": 1.8612085847777212e-06,
"loss": 0.602,
"step": 154
},
{
"epoch": 0.5090311986863711,
"grad_norm": 45.827071354966,
"learning_rate": 1.858280726624609e-06,
"loss": 0.5806,
"step": 155
},
{
"epoch": 0.5123152709359606,
"grad_norm": 18.978499335170554,
"learning_rate": 1.855324660828452e-06,
"loss": 0.6165,
"step": 156
},
{
"epoch": 0.5155993431855501,
"grad_norm": 67.94220614733077,
"learning_rate": 1.8523404845412025e-06,
"loss": 0.6229,
"step": 157
},
{
"epoch": 0.5188834154351396,
"grad_norm": 19.582077106202128,
"learning_rate": 1.8493282958386739e-06,
"loss": 0.5532,
"step": 158
},
{
"epoch": 0.5221674876847291,
"grad_norm": 53.75335205438214,
"learning_rate": 1.846288193717314e-06,
"loss": 0.6104,
"step": 159
},
{
"epoch": 0.5254515599343186,
"grad_norm": 38.261538748017415,
"learning_rate": 1.8432202780909538e-06,
"loss": 0.5848,
"step": 160
},
{
"epoch": 0.5287356321839081,
"grad_norm": 222.09670680924359,
"learning_rate": 1.8401246497875235e-06,
"loss": 0.5867,
"step": 161
},
{
"epoch": 0.5320197044334976,
"grad_norm": 571.9505490444766,
"learning_rate": 1.8370014105457376e-06,
"loss": 0.5695,
"step": 162
},
{
"epoch": 0.535303776683087,
"grad_norm": 48.34961117571771,
"learning_rate": 1.8338506630117526e-06,
"loss": 0.5905,
"step": 163
},
{
"epoch": 0.5385878489326765,
"grad_norm": 39.05700378122741,
"learning_rate": 1.830672510735793e-06,
"loss": 0.5984,
"step": 164
},
{
"epoch": 0.541871921182266,
"grad_norm": 58.20418532997132,
"learning_rate": 1.8274670581687478e-06,
"loss": 0.587,
"step": 165
},
{
"epoch": 0.5451559934318555,
"grad_norm": 245.8655111712564,
"learning_rate": 1.8242344106587377e-06,
"loss": 0.6031,
"step": 166
},
{
"epoch": 0.548440065681445,
"grad_norm": 105.7509385578569,
"learning_rate": 1.8209746744476536e-06,
"loss": 0.5801,
"step": 167
},
{
"epoch": 0.5517241379310345,
"grad_norm": 97.03979541618602,
"learning_rate": 1.8176879566676639e-06,
"loss": 0.6077,
"step": 168
},
{
"epoch": 0.555008210180624,
"grad_norm": 14.137405437866256,
"learning_rate": 1.8143743653376943e-06,
"loss": 0.592,
"step": 169
},
{
"epoch": 0.5582922824302134,
"grad_norm": 46.24134681947936,
"learning_rate": 1.811034009359877e-06,
"loss": 0.5912,
"step": 170
},
{
"epoch": 0.5615763546798029,
"grad_norm": 101.35151290055389,
"learning_rate": 1.8076669985159725e-06,
"loss": 0.6158,
"step": 171
},
{
"epoch": 0.5648604269293924,
"grad_norm": 140.1896691880712,
"learning_rate": 1.8042734434637613e-06,
"loss": 0.5692,
"step": 172
},
{
"epoch": 0.5681444991789819,
"grad_norm": 226.4192589085135,
"learning_rate": 1.8008534557334063e-06,
"loss": 0.5754,
"step": 173
},
{
"epoch": 0.5714285714285714,
"grad_norm": 45.671389908527075,
"learning_rate": 1.7974071477237885e-06,
"loss": 0.6016,
"step": 174
},
{
"epoch": 0.5747126436781609,
"grad_norm": 19.244548559672378,
"learning_rate": 1.7939346326988125e-06,
"loss": 0.5594,
"step": 175
},
{
"epoch": 0.5779967159277504,
"grad_norm": 201.20738634541007,
"learning_rate": 1.7904360247836834e-06,
"loss": 0.5781,
"step": 176
},
{
"epoch": 0.5812807881773399,
"grad_norm": 41.13648160887617,
"learning_rate": 1.7869114389611573e-06,
"loss": 0.5769,
"step": 177
},
{
"epoch": 0.5845648604269293,
"grad_norm": 287.9411519318713,
"learning_rate": 1.7833609910677612e-06,
"loss": 0.5819,
"step": 178
},
{
"epoch": 0.5878489326765188,
"grad_norm": 24.18855361442465,
"learning_rate": 1.779784797789987e-06,
"loss": 0.5964,
"step": 179
},
{
"epoch": 0.5911330049261084,
"grad_norm": 24.012668455231225,
"learning_rate": 1.7761829766604554e-06,
"loss": 0.6134,
"step": 180
},
{
"epoch": 0.5944170771756979,
"grad_norm": 28.51880265050756,
"learning_rate": 1.772555646054055e-06,
"loss": 0.602,
"step": 181
},
{
"epoch": 0.5977011494252874,
"grad_norm": 15.794558397944432,
"learning_rate": 1.768902925184049e-06,
"loss": 0.5821,
"step": 182
},
{
"epoch": 0.6009852216748769,
"grad_norm": 58.628563783312366,
"learning_rate": 1.7652249340981604e-06,
"loss": 0.6285,
"step": 183
},
{
"epoch": 0.6042692939244664,
"grad_norm": 17.619228357058972,
"learning_rate": 1.7615217936746242e-06,
"loss": 0.596,
"step": 184
},
{
"epoch": 0.6075533661740559,
"grad_norm": 207.93744899824176,
"learning_rate": 1.7577936256182167e-06,
"loss": 0.5917,
"step": 185
},
{
"epoch": 0.6108374384236454,
"grad_norm": 41.840591224822674,
"learning_rate": 1.754040552456253e-06,
"loss": 0.5866,
"step": 186
},
{
"epoch": 0.6141215106732348,
"grad_norm": 93.8154760971863,
"learning_rate": 1.7502626975345626e-06,
"loss": 0.6073,
"step": 187
},
{
"epoch": 0.6174055829228243,
"grad_norm": 17.492683199274612,
"learning_rate": 1.746460185013435e-06,
"loss": 0.5964,
"step": 188
},
{
"epoch": 0.6206896551724138,
"grad_norm": 29.276597988664253,
"learning_rate": 1.742633139863538e-06,
"loss": 0.5754,
"step": 189
},
{
"epoch": 0.6239737274220033,
"grad_norm": 68.18460724104617,
"learning_rate": 1.7387816878618117e-06,
"loss": 0.5803,
"step": 190
},
{
"epoch": 0.6272577996715928,
"grad_norm": 56.35078039350749,
"learning_rate": 1.7349059555873343e-06,
"loss": 0.5787,
"step": 191
},
{
"epoch": 0.6305418719211823,
"grad_norm": 43.29495360037644,
"learning_rate": 1.7310060704171627e-06,
"loss": 0.5522,
"step": 192
},
{
"epoch": 0.6338259441707718,
"grad_norm": 68.5551565674711,
"learning_rate": 1.7270821605221446e-06,
"loss": 0.6055,
"step": 193
},
{
"epoch": 0.6371100164203612,
"grad_norm": 117.39744403334397,
"learning_rate": 1.723134354862708e-06,
"loss": 0.5713,
"step": 194
},
{
"epoch": 0.6403940886699507,
"grad_norm": 189.04818071139826,
"learning_rate": 1.7191627831846222e-06,
"loss": 0.5748,
"step": 195
},
{
"epoch": 0.6436781609195402,
"grad_norm": 74.96348879368328,
"learning_rate": 1.7151675760147325e-06,
"loss": 0.5767,
"step": 196
},
{
"epoch": 0.6469622331691297,
"grad_norm": 19.283409944711153,
"learning_rate": 1.7111488646566725e-06,
"loss": 0.5835,
"step": 197
},
{
"epoch": 0.6502463054187192,
"grad_norm": 28.125800950282475,
"learning_rate": 1.7071067811865474e-06,
"loss": 0.6006,
"step": 198
},
{
"epoch": 0.6535303776683087,
"grad_norm": 197.02038970341124,
"learning_rate": 1.7030414584485934e-06,
"loss": 0.6071,
"step": 199
},
{
"epoch": 0.6568144499178982,
"grad_norm": 39.027824525327794,
"learning_rate": 1.6989530300508123e-06,
"loss": 0.5967,
"step": 200
},
{
"epoch": 0.6600985221674877,
"grad_norm": 31.571940924781416,
"learning_rate": 1.6948416303605793e-06,
"loss": 0.6396,
"step": 201
},
{
"epoch": 0.6633825944170771,
"grad_norm": 129.62214181745804,
"learning_rate": 1.6907073945002288e-06,
"loss": 0.5821,
"step": 202
},
{
"epoch": 0.6666666666666666,
"grad_norm": 162.51175499429485,
"learning_rate": 1.6865504583426116e-06,
"loss": 0.6123,
"step": 203
},
{
"epoch": 0.6699507389162561,
"grad_norm": 37.68050097895022,
"learning_rate": 1.6823709585066306e-06,
"loss": 0.578,
"step": 204
},
{
"epoch": 0.6732348111658456,
"grad_norm": 20.017550125639158,
"learning_rate": 1.6781690323527509e-06,
"loss": 0.5854,
"step": 205
},
{
"epoch": 0.6765188834154351,
"grad_norm": 76.61122117606573,
"learning_rate": 1.6739448179784844e-06,
"loss": 0.589,
"step": 206
},
{
"epoch": 0.6798029556650246,
"grad_norm": 22.411521664367136,
"learning_rate": 1.6696984542138519e-06,
"loss": 0.5844,
"step": 207
},
{
"epoch": 0.6830870279146142,
"grad_norm": 98.32888397272134,
"learning_rate": 1.6654300806168206e-06,
"loss": 0.5972,
"step": 208
},
{
"epoch": 0.6863711001642037,
"grad_norm": 25.354018912353993,
"learning_rate": 1.661139837468717e-06,
"loss": 0.5749,
"step": 209
},
{
"epoch": 0.6896551724137931,
"grad_norm": 109.57794389864921,
"learning_rate": 1.6568278657696162e-06,
"loss": 0.5318,
"step": 210
},
{
"epoch": 0.6929392446633826,
"grad_norm": 184.91691946093704,
"learning_rate": 1.6524943072337092e-06,
"loss": 0.5955,
"step": 211
},
{
"epoch": 0.6962233169129721,
"grad_norm": 51.18780040494431,
"learning_rate": 1.6481393042846442e-06,
"loss": 0.5899,
"step": 212
},
{
"epoch": 0.6995073891625616,
"grad_norm": 93.66347257011115,
"learning_rate": 1.6437630000508462e-06,
"loss": 0.5723,
"step": 213
},
{
"epoch": 0.7027914614121511,
"grad_norm": 49.27370857815919,
"learning_rate": 1.6393655383608132e-06,
"loss": 0.6148,
"step": 214
},
{
"epoch": 0.7060755336617406,
"grad_norm": 25.246859061737737,
"learning_rate": 1.6349470637383888e-06,
"loss": 0.589,
"step": 215
},
{
"epoch": 0.7093596059113301,
"grad_norm": 112.54633084879946,
"learning_rate": 1.630507721398013e-06,
"loss": 0.6262,
"step": 216
},
{
"epoch": 0.7126436781609196,
"grad_norm": 59.80877460933437,
"learning_rate": 1.6260476572399493e-06,
"loss": 0.573,
"step": 217
},
{
"epoch": 0.715927750410509,
"grad_norm": 221.05125034165957,
"learning_rate": 1.6215670178454892e-06,
"loss": 0.6132,
"step": 218
},
{
"epoch": 0.7192118226600985,
"grad_norm": 32.62986909195134,
"learning_rate": 1.6170659504721363e-06,
"loss": 0.5978,
"step": 219
},
{
"epoch": 0.722495894909688,
"grad_norm": 79.18706822806207,
"learning_rate": 1.6125446030487642e-06,
"loss": 0.6197,
"step": 220
},
{
"epoch": 0.7257799671592775,
"grad_norm": 646.9711401182135,
"learning_rate": 1.6080031241707576e-06,
"loss": 0.5909,
"step": 221
},
{
"epoch": 0.729064039408867,
"grad_norm": 227.7039164426062,
"learning_rate": 1.6034416630951265e-06,
"loss": 0.5617,
"step": 222
},
{
"epoch": 0.7323481116584565,
"grad_norm": 86.24663093418314,
"learning_rate": 1.5988603697356009e-06,
"loss": 0.568,
"step": 223
},
{
"epoch": 0.735632183908046,
"grad_norm": 28.186646002874514,
"learning_rate": 1.5942593946577065e-06,
"loss": 0.5739,
"step": 224
},
{
"epoch": 0.7389162561576355,
"grad_norm": 114.91672123046412,
"learning_rate": 1.5896388890738127e-06,
"loss": 0.5854,
"step": 225
},
{
"epoch": 0.7422003284072249,
"grad_norm": 110.20008502969989,
"learning_rate": 1.5849990048381648e-06,
"loss": 0.5792,
"step": 226
},
{
"epoch": 0.7454844006568144,
"grad_norm": 118.55251219758163,
"learning_rate": 1.5803398944418933e-06,
"loss": 0.5776,
"step": 227
},
{
"epoch": 0.7487684729064039,
"grad_norm": 55.68698315902024,
"learning_rate": 1.575661711008002e-06,
"loss": 0.5644,
"step": 228
},
{
"epoch": 0.7487684729064039,
"eval_loss": 0.3468731641769409,
"eval_runtime": 254.0407,
"eval_samples_per_second": 8.133,
"eval_steps_per_second": 0.256,
"step": 228
},
{
"epoch": 0.7520525451559934,
"grad_norm": 32.913448787991214,
"learning_rate": 1.570964608286336e-06,
"loss": 0.5904,
"step": 229
},
{
"epoch": 0.7553366174055829,
"grad_norm": 48.81085602810963,
"learning_rate": 1.566248740648527e-06,
"loss": 0.585,
"step": 230
},
{
"epoch": 0.7586206896551724,
"grad_norm": 43.34405794275439,
"learning_rate": 1.5615142630829228e-06,
"loss": 0.5503,
"step": 231
},
{
"epoch": 0.7619047619047619,
"grad_norm": 24.78222801953508,
"learning_rate": 1.5567613311894907e-06,
"loss": 0.6158,
"step": 232
},
{
"epoch": 0.7651888341543513,
"grad_norm": 20.01938395622359,
"learning_rate": 1.5519901011747043e-06,
"loss": 0.5622,
"step": 233
},
{
"epoch": 0.7684729064039408,
"grad_norm": 128.40682635925788,
"learning_rate": 1.5472007298464116e-06,
"loss": 0.5866,
"step": 234
},
{
"epoch": 0.7717569786535303,
"grad_norm": 18.43305197325785,
"learning_rate": 1.5423933746086793e-06,
"loss": 0.5952,
"step": 235
},
{
"epoch": 0.7750410509031199,
"grad_norm": 100.43618051137392,
"learning_rate": 1.5375681934566202e-06,
"loss": 0.5882,
"step": 236
},
{
"epoch": 0.7783251231527094,
"grad_norm": 24.521874696776806,
"learning_rate": 1.5327253449712018e-06,
"loss": 0.5784,
"step": 237
},
{
"epoch": 0.7816091954022989,
"grad_norm": 123.57359702902865,
"learning_rate": 1.5278649883140329e-06,
"loss": 0.6413,
"step": 238
},
{
"epoch": 0.7848932676518884,
"grad_norm": 106.32565785443477,
"learning_rate": 1.5229872832221333e-06,
"loss": 0.5668,
"step": 239
},
{
"epoch": 0.7881773399014779,
"grad_norm": 99.10065342575062,
"learning_rate": 1.5180923900026845e-06,
"loss": 0.5928,
"step": 240
},
{
"epoch": 0.7914614121510674,
"grad_norm": 130.80914850175915,
"learning_rate": 1.513180469527761e-06,
"loss": 0.5926,
"step": 241
},
{
"epoch": 0.7947454844006568,
"grad_norm": 94.9924662311697,
"learning_rate": 1.5082516832290421e-06,
"loss": 0.6016,
"step": 242
},
{
"epoch": 0.7980295566502463,
"grad_norm": 197.53102698941842,
"learning_rate": 1.5033061930925078e-06,
"loss": 0.5731,
"step": 243
},
{
"epoch": 0.8013136288998358,
"grad_norm": 38.788949526243016,
"learning_rate": 1.498344161653115e-06,
"loss": 0.5963,
"step": 244
},
{
"epoch": 0.8045977011494253,
"grad_norm": 38.30742214846052,
"learning_rate": 1.493365751989454e-06,
"loss": 0.5638,
"step": 245
},
{
"epoch": 0.8078817733990148,
"grad_norm": 648.6594536144033,
"learning_rate": 1.4883711277183915e-06,
"loss": 0.5868,
"step": 246
},
{
"epoch": 0.8111658456486043,
"grad_norm": 265.95515256929093,
"learning_rate": 1.4833604529896908e-06,
"loss": 0.5512,
"step": 247
},
{
"epoch": 0.8144499178981938,
"grad_norm": 170.80635770181408,
"learning_rate": 1.478333892480619e-06,
"loss": 0.5994,
"step": 248
},
{
"epoch": 0.8177339901477833,
"grad_norm": 1398.0497669156412,
"learning_rate": 1.4732916113905333e-06,
"loss": 0.5887,
"step": 249
},
{
"epoch": 0.8210180623973727,
"grad_norm": 251.7115266399908,
"learning_rate": 1.4682337754354533e-06,
"loss": 0.5735,
"step": 250
},
{
"epoch": 0.8243021346469622,
"grad_norm": 124.88918006964796,
"learning_rate": 1.4631605508426121e-06,
"loss": 0.6044,
"step": 251
},
{
"epoch": 0.8275862068965517,
"grad_norm": 608.3413931999685,
"learning_rate": 1.4580721043449966e-06,
"loss": 0.5628,
"step": 252
},
{
"epoch": 0.8308702791461412,
"grad_norm": 1194.4815644571033,
"learning_rate": 1.4529686031758642e-06,
"loss": 0.5933,
"step": 253
},
{
"epoch": 0.8341543513957307,
"grad_norm": 113.35099537580673,
"learning_rate": 1.4478502150632501e-06,
"loss": 0.5861,
"step": 254
},
{
"epoch": 0.8374384236453202,
"grad_norm": 153.99285185333872,
"learning_rate": 1.442717108224452e-06,
"loss": 0.5812,
"step": 255
},
{
"epoch": 0.8407224958949097,
"grad_norm": 130.21821279469813,
"learning_rate": 1.4375694513605034e-06,
"loss": 0.5732,
"step": 256
},
{
"epoch": 0.8440065681444991,
"grad_norm": 35.99304972206288,
"learning_rate": 1.4324074136506282e-06,
"loss": 0.5845,
"step": 257
},
{
"epoch": 0.8472906403940886,
"grad_norm": 12.583104402647741,
"learning_rate": 1.4272311647466808e-06,
"loss": 0.6195,
"step": 258
},
{
"epoch": 0.8505747126436781,
"grad_norm": 161.8183385483987,
"learning_rate": 1.4220408747675712e-06,
"loss": 0.6064,
"step": 259
},
{
"epoch": 0.8538587848932676,
"grad_norm": 25.593864430902567,
"learning_rate": 1.4168367142936734e-06,
"loss": 0.5897,
"step": 260
},
{
"epoch": 0.8571428571428571,
"grad_norm": 60.9645827697532,
"learning_rate": 1.411618854361218e-06,
"loss": 0.6029,
"step": 261
},
{
"epoch": 0.8604269293924466,
"grad_norm": 313.86128927339445,
"learning_rate": 1.4063874664566732e-06,
"loss": 0.5766,
"step": 262
},
{
"epoch": 0.8637110016420362,
"grad_norm": 376.4774663209203,
"learning_rate": 1.401142722511109e-06,
"loss": 0.5905,
"step": 263
},
{
"epoch": 0.8669950738916257,
"grad_norm": 23.684255757274983,
"learning_rate": 1.3958847948945428e-06,
"loss": 0.5928,
"step": 264
},
{
"epoch": 0.8702791461412152,
"grad_norm": 18.327052041947013,
"learning_rate": 1.3906138564102792e-06,
"loss": 0.5659,
"step": 265
},
{
"epoch": 0.8735632183908046,
"grad_norm": 31.74783455878064,
"learning_rate": 1.3853300802892282e-06,
"loss": 0.5656,
"step": 266
},
{
"epoch": 0.8768472906403941,
"grad_norm": 419.0756338024995,
"learning_rate": 1.3800336401842127e-06,
"loss": 0.5986,
"step": 267
},
{
"epoch": 0.8801313628899836,
"grad_norm": 77.68111946128553,
"learning_rate": 1.3747247101642602e-06,
"loss": 0.6006,
"step": 268
},
{
"epoch": 0.8834154351395731,
"grad_norm": 75.41201589840144,
"learning_rate": 1.3694034647088838e-06,
"loss": 0.5778,
"step": 269
},
{
"epoch": 0.8866995073891626,
"grad_norm": 36.241115646339644,
"learning_rate": 1.3640700787023462e-06,
"loss": 0.565,
"step": 270
},
{
"epoch": 0.8899835796387521,
"grad_norm": 135.20927104693348,
"learning_rate": 1.3587247274279139e-06,
"loss": 0.5763,
"step": 271
},
{
"epoch": 0.8932676518883416,
"grad_norm": 87.93638862635923,
"learning_rate": 1.3533675865620936e-06,
"loss": 0.5963,
"step": 272
},
{
"epoch": 0.896551724137931,
"grad_norm": 29.613470120000667,
"learning_rate": 1.3479988321688618e-06,
"loss": 0.577,
"step": 273
},
{
"epoch": 0.8998357963875205,
"grad_norm": 14.319123531102576,
"learning_rate": 1.3426186406938766e-06,
"loss": 0.5961,
"step": 274
},
{
"epoch": 0.90311986863711,
"grad_norm": 169.5219058735149,
"learning_rate": 1.337227188958679e-06,
"loss": 0.5897,
"step": 275
},
{
"epoch": 0.9064039408866995,
"grad_norm": 38.00458314499751,
"learning_rate": 1.331824654154881e-06,
"loss": 0.5613,
"step": 276
},
{
"epoch": 0.909688013136289,
"grad_norm": 14.378667726194587,
"learning_rate": 1.3264112138383443e-06,
"loss": 0.6258,
"step": 277
},
{
"epoch": 0.9129720853858785,
"grad_norm": 42.72991885435684,
"learning_rate": 1.320987045923342e-06,
"loss": 0.5773,
"step": 278
},
{
"epoch": 0.916256157635468,
"grad_norm": 68.40318666339633,
"learning_rate": 1.315552328676714e-06,
"loss": 0.5839,
"step": 279
},
{
"epoch": 0.9195402298850575,
"grad_norm": 45.180537540210565,
"learning_rate": 1.3101072407120055e-06,
"loss": 0.5955,
"step": 280
},
{
"epoch": 0.922824302134647,
"grad_norm": 39.16288757906327,
"learning_rate": 1.3046519609836e-06,
"loss": 0.5927,
"step": 281
},
{
"epoch": 0.9261083743842364,
"grad_norm": 67.73988324657877,
"learning_rate": 1.2991866687808353e-06,
"loss": 0.5799,
"step": 282
},
{
"epoch": 0.9293924466338259,
"grad_norm": 353.64259849704393,
"learning_rate": 1.2937115437221117e-06,
"loss": 0.5764,
"step": 283
},
{
"epoch": 0.9326765188834154,
"grad_norm": 54.971519691287966,
"learning_rate": 1.2882267657489908e-06,
"loss": 0.5543,
"step": 284
},
{
"epoch": 0.9359605911330049,
"grad_norm": 22.759273671919335,
"learning_rate": 1.2827325151202782e-06,
"loss": 0.582,
"step": 285
},
{
"epoch": 0.9392446633825944,
"grad_norm": 14.41964639616655,
"learning_rate": 1.2772289724061014e-06,
"loss": 0.6231,
"step": 286
},
{
"epoch": 0.9425287356321839,
"grad_norm": 154.5418617403585,
"learning_rate": 1.2717163184819759e-06,
"loss": 0.6038,
"step": 287
},
{
"epoch": 0.9458128078817734,
"grad_norm": 14.212387952750024,
"learning_rate": 1.2661947345228593e-06,
"loss": 0.5684,
"step": 288
},
{
"epoch": 0.9490968801313628,
"grad_norm": 123.37078826032788,
"learning_rate": 1.2606644019971966e-06,
"loss": 0.5795,
"step": 289
},
{
"epoch": 0.9523809523809523,
"grad_norm": 17.029678211667036,
"learning_rate": 1.255125502660958e-06,
"loss": 0.6055,
"step": 290
},
{
"epoch": 0.9556650246305419,
"grad_norm": 43.985222402630065,
"learning_rate": 1.2495782185516637e-06,
"loss": 0.5785,
"step": 291
},
{
"epoch": 0.9589490968801314,
"grad_norm": 26.180114326110402,
"learning_rate": 1.2440227319824022e-06,
"loss": 0.5672,
"step": 292
},
{
"epoch": 0.9622331691297209,
"grad_norm": 58.05962982942946,
"learning_rate": 1.2384592255358384e-06,
"loss": 0.5723,
"step": 293
},
{
"epoch": 0.9655172413793104,
"grad_norm": 96.86728362154787,
"learning_rate": 1.232887882058212e-06,
"loss": 0.5667,
"step": 294
},
{
"epoch": 0.9688013136288999,
"grad_norm": 31.777624654282178,
"learning_rate": 1.2273088846533302e-06,
"loss": 0.6115,
"step": 295
},
{
"epoch": 0.9720853858784894,
"grad_norm": 83.38296129361547,
"learning_rate": 1.2217224166765475e-06,
"loss": 0.5732,
"step": 296
},
{
"epoch": 0.9753694581280788,
"grad_norm": 121.56881681339887,
"learning_rate": 1.2161286617287418e-06,
"loss": 0.5954,
"step": 297
},
{
"epoch": 0.9786535303776683,
"grad_norm": 54.8264040669832,
"learning_rate": 1.2105278036502787e-06,
"loss": 0.5931,
"step": 298
},
{
"epoch": 0.9819376026272578,
"grad_norm": 21.35246678957699,
"learning_rate": 1.2049200265149707e-06,
"loss": 0.5694,
"step": 299
},
{
"epoch": 0.9852216748768473,
"grad_norm": 1059.6690924678123,
"learning_rate": 1.1993055146240272e-06,
"loss": 0.5658,
"step": 300
},
{
"epoch": 0.9885057471264368,
"grad_norm": 26.188608952399726,
"learning_rate": 1.1936844524999966e-06,
"loss": 0.5654,
"step": 301
},
{
"epoch": 0.9917898193760263,
"grad_norm": 21.558049696210244,
"learning_rate": 1.1880570248807032e-06,
"loss": 0.5839,
"step": 302
},
{
"epoch": 0.9950738916256158,
"grad_norm": 17.803809249890314,
"learning_rate": 1.1824234167131746e-06,
"loss": 0.5563,
"step": 303
},
{
"epoch": 0.9983579638752053,
"grad_norm": 70.52662623832066,
"learning_rate": 1.1767838131475654e-06,
"loss": 0.5626,
"step": 304
},
{
"epoch": 0.9983579638752053,
"eval_loss": 0.3444424271583557,
"eval_runtime": 255.3718,
"eval_samples_per_second": 8.09,
"eval_steps_per_second": 0.255,
"step": 304
},
{
"epoch": 1.0,
"grad_norm": 70.52662623832066,
"learning_rate": 1.171138399531068e-06,
"loss": 0.6032,
"step": 305
},
{
"epoch": 1.0032840722495895,
"grad_norm": 65.13191709579829,
"learning_rate": 1.1654873614018266e-06,
"loss": 0.5511,
"step": 306
},
{
"epoch": 1.006568144499179,
"grad_norm": 285.9719298771201,
"learning_rate": 1.1598308844828345e-06,
"loss": 0.6026,
"step": 307
},
{
"epoch": 1.0098522167487685,
"grad_norm": 37.39321171527541,
"learning_rate": 1.154169154675834e-06,
"loss": 0.5785,
"step": 308
},
{
"epoch": 1.013136288998358,
"grad_norm": 41.68071927705891,
"learning_rate": 1.1485023580552039e-06,
"loss": 0.6114,
"step": 309
},
{
"epoch": 1.0164203612479474,
"grad_norm": 135.40715808644197,
"learning_rate": 1.1428306808618454e-06,
"loss": 0.5782,
"step": 310
},
{
"epoch": 1.019704433497537,
"grad_norm": 22.790140257232764,
"learning_rate": 1.137154309497062e-06,
"loss": 0.5687,
"step": 311
},
{
"epoch": 1.0229885057471264,
"grad_norm": 141.91339180106723,
"learning_rate": 1.131473430516432e-06,
"loss": 0.6137,
"step": 312
},
{
"epoch": 1.026272577996716,
"grad_norm": 72.39990304750096,
"learning_rate": 1.1257882306236775e-06,
"loss": 0.6173,
"step": 313
},
{
"epoch": 1.0295566502463054,
"grad_norm": 44.689199344971264,
"learning_rate": 1.1200988966645284e-06,
"loss": 0.5684,
"step": 314
},
{
"epoch": 1.0328407224958949,
"grad_norm": 18.212438241962214,
"learning_rate": 1.1144056156205831e-06,
"loss": 0.5885,
"step": 315
},
{
"epoch": 1.0361247947454844,
"grad_norm": 32.514249930046795,
"learning_rate": 1.108708574603161e-06,
"loss": 0.6074,
"step": 316
},
{
"epoch": 1.0394088669950738,
"grad_norm": 17.685539914578786,
"learning_rate": 1.1030079608471542e-06,
"loss": 0.5701,
"step": 317
},
{
"epoch": 1.0426929392446633,
"grad_norm": 50.50561698330298,
"learning_rate": 1.0973039617048747e-06,
"loss": 0.6148,
"step": 318
},
{
"epoch": 1.0459770114942528,
"grad_norm": 30.494423044684428,
"learning_rate": 1.0915967646398948e-06,
"loss": 0.5631,
"step": 319
},
{
"epoch": 1.0492610837438423,
"grad_norm": 43.634293011722335,
"learning_rate": 1.0858865572208891e-06,
"loss": 0.5563,
"step": 320
},
{
"epoch": 1.0525451559934318,
"grad_norm": 34.28951266031906,
"learning_rate": 1.0801735271154667e-06,
"loss": 0.5703,
"step": 321
},
{
"epoch": 1.0558292282430213,
"grad_norm": 40.35540642489708,
"learning_rate": 1.0744578620840063e-06,
"loss": 0.5682,
"step": 322
},
{
"epoch": 1.0591133004926108,
"grad_norm": 90.22818294715287,
"learning_rate": 1.068739749973484e-06,
"loss": 0.5676,
"step": 323
},
{
"epoch": 1.0623973727422003,
"grad_norm": 27.34387701783233,
"learning_rate": 1.0630193787112993e-06,
"loss": 0.5751,
"step": 324
},
{
"epoch": 1.0656814449917897,
"grad_norm": 56.638276613527445,
"learning_rate": 1.0572969362990997e-06,
"loss": 0.6014,
"step": 325
},
{
"epoch": 1.0689655172413792,
"grad_norm": 15.866790094277658,
"learning_rate": 1.0515726108066023e-06,
"loss": 0.5916,
"step": 326
},
{
"epoch": 1.0722495894909687,
"grad_norm": 12.950547272697525,
"learning_rate": 1.0458465903654105e-06,
"loss": 0.5734,
"step": 327
},
{
"epoch": 1.0755336617405582,
"grad_norm": 17.279307686255628,
"learning_rate": 1.0401190631628347e-06,
"loss": 0.5899,
"step": 328
},
{
"epoch": 1.0788177339901477,
"grad_norm": 26.36195986983623,
"learning_rate": 1.0343902174357038e-06,
"loss": 0.5936,
"step": 329
},
{
"epoch": 1.0821018062397372,
"grad_norm": 53.631919940788016,
"learning_rate": 1.0286602414641815e-06,
"loss": 0.6059,
"step": 330
},
{
"epoch": 1.0853858784893267,
"grad_norm": 37.558445588790654,
"learning_rate": 1.0229293235655768e-06,
"loss": 0.5573,
"step": 331
},
{
"epoch": 1.0886699507389164,
"grad_norm": 15.958318674221253,
"learning_rate": 1.017197652088155e-06,
"loss": 0.6102,
"step": 332
},
{
"epoch": 1.0919540229885056,
"grad_norm": 40.26474283910818,
"learning_rate": 1.0114654154049489e-06,
"loss": 0.5895,
"step": 333
},
{
"epoch": 1.0952380952380953,
"grad_norm": 146.08840587905024,
"learning_rate": 1.0057328019075668e-06,
"loss": 0.5726,
"step": 334
},
{
"epoch": 1.0985221674876848,
"grad_norm": 202.1576115663822,
"learning_rate": 1e-06,
"loss": 0.577,
"step": 335
},
{
"epoch": 1.1018062397372743,
"grad_norm": 14.523463101670913,
"learning_rate": 9.942671980924335e-07,
"loss": 0.5872,
"step": 336
},
{
"epoch": 1.1050903119868638,
"grad_norm": 32.09224127026215,
"learning_rate": 9.885345845950508e-07,
"loss": 0.6071,
"step": 337
},
{
"epoch": 1.1083743842364533,
"grad_norm": 13.896096566133576,
"learning_rate": 9.828023479118448e-07,
"loss": 0.5989,
"step": 338
},
{
"epoch": 1.1116584564860428,
"grad_norm": 58.06165751181001,
"learning_rate": 9.770706764344234e-07,
"loss": 0.5656,
"step": 339
},
{
"epoch": 1.1149425287356323,
"grad_norm": 20.036356068409564,
"learning_rate": 9.713397585358188e-07,
"loss": 0.5858,
"step": 340
},
{
"epoch": 1.1182266009852218,
"grad_norm": 10.100931882003527,
"learning_rate": 9.65609782564296e-07,
"loss": 0.5855,
"step": 341
},
{
"epoch": 1.1215106732348112,
"grad_norm": 13.587121183019379,
"learning_rate": 9.598809368371654e-07,
"loss": 0.6379,
"step": 342
},
{
"epoch": 1.1247947454844007,
"grad_norm": 126.86325897606325,
"learning_rate": 9.541534096345897e-07,
"loss": 0.5546,
"step": 343
},
{
"epoch": 1.1280788177339902,
"grad_norm": 169.89985357006515,
"learning_rate": 9.484273891933981e-07,
"loss": 0.5563,
"step": 344
},
{
"epoch": 1.1313628899835797,
"grad_norm": 49.08212390882699,
"learning_rate": 9.427030637009002e-07,
"loss": 0.5699,
"step": 345
},
{
"epoch": 1.1346469622331692,
"grad_norm": 40.18400735284968,
"learning_rate": 9.369806212887007e-07,
"loss": 0.5715,
"step": 346
},
{
"epoch": 1.1379310344827587,
"grad_norm": 18.860982747952214,
"learning_rate": 9.312602500265159e-07,
"loss": 0.5648,
"step": 347
},
{
"epoch": 1.1412151067323482,
"grad_norm": 10.902272643558419,
"learning_rate": 9.255421379159933e-07,
"loss": 0.5735,
"step": 348
},
{
"epoch": 1.1444991789819376,
"grad_norm": 66.12684936925638,
"learning_rate": 9.198264728845331e-07,
"loss": 0.6154,
"step": 349
},
{
"epoch": 1.1477832512315271,
"grad_norm": 28.02996822136127,
"learning_rate": 9.141134427791109e-07,
"loss": 0.5778,
"step": 350
},
{
"epoch": 1.1510673234811166,
"grad_norm": 30.51347632697842,
"learning_rate": 9.084032353601052e-07,
"loss": 0.6041,
"step": 351
},
{
"epoch": 1.154351395730706,
"grad_norm": 49.552466990210895,
"learning_rate": 9.026960382951252e-07,
"loss": 0.5671,
"step": 352
},
{
"epoch": 1.1576354679802956,
"grad_norm": 44.54866711025706,
"learning_rate": 8.969920391528457e-07,
"loss": 0.5506,
"step": 353
},
{
"epoch": 1.160919540229885,
"grad_norm": 52.856706858816,
"learning_rate": 8.912914253968391e-07,
"loss": 0.554,
"step": 354
},
{
"epoch": 1.1642036124794746,
"grad_norm": 47.388855420890756,
"learning_rate": 8.85594384379417e-07,
"loss": 0.5551,
"step": 355
},
{
"epoch": 1.167487684729064,
"grad_norm": 30.628301446970205,
"learning_rate": 8.799011033354715e-07,
"loss": 0.5831,
"step": 356
},
{
"epoch": 1.1707717569786535,
"grad_norm": 22.952426384099383,
"learning_rate": 8.742117693763227e-07,
"loss": 0.5597,
"step": 357
},
{
"epoch": 1.174055829228243,
"grad_norm": 17.939719949680224,
"learning_rate": 8.685265694835681e-07,
"loss": 0.577,
"step": 358
},
{
"epoch": 1.1773399014778325,
"grad_norm": 27.110777787948248,
"learning_rate": 8.628456905029382e-07,
"loss": 0.5726,
"step": 359
},
{
"epoch": 1.180623973727422,
"grad_norm": 13.896701402873619,
"learning_rate": 8.571693191381544e-07,
"loss": 0.5623,
"step": 360
},
{
"epoch": 1.1839080459770115,
"grad_norm": 15.205774656115123,
"learning_rate": 8.514976419447963e-07,
"loss": 0.5943,
"step": 361
},
{
"epoch": 1.187192118226601,
"grad_norm": 45.58828461857186,
"learning_rate": 8.458308453241663e-07,
"loss": 0.5817,
"step": 362
},
{
"epoch": 1.1904761904761905,
"grad_norm": 116.8441584869502,
"learning_rate": 8.401691155171652e-07,
"loss": 0.5444,
"step": 363
},
{
"epoch": 1.19376026272578,
"grad_norm": 170.81143104953796,
"learning_rate": 8.345126385981735e-07,
"loss": 0.5904,
"step": 364
},
{
"epoch": 1.1970443349753694,
"grad_norm": 41.71381416465335,
"learning_rate": 8.288616004689319e-07,
"loss": 0.5784,
"step": 365
},
{
"epoch": 1.200328407224959,
"grad_norm": 115.88107721819807,
"learning_rate": 8.23216186852435e-07,
"loss": 0.594,
"step": 366
},
{
"epoch": 1.2036124794745484,
"grad_norm": 17.490648876213996,
"learning_rate": 8.175765832868251e-07,
"loss": 0.576,
"step": 367
},
{
"epoch": 1.206896551724138,
"grad_norm": 39.174813390569106,
"learning_rate": 8.11942975119297e-07,
"loss": 0.5627,
"step": 368
},
{
"epoch": 1.2101806239737274,
"grad_norm": 668.8532117187957,
"learning_rate": 8.063155475000035e-07,
"loss": 0.5987,
"step": 369
},
{
"epoch": 1.2134646962233169,
"grad_norm": 29.45366016595164,
"learning_rate": 8.006944853759732e-07,
"loss": 0.5788,
"step": 370
},
{
"epoch": 1.2167487684729064,
"grad_norm": 19.85954794293891,
"learning_rate": 7.950799734850291e-07,
"loss": 0.5552,
"step": 371
},
{
"epoch": 1.2200328407224958,
"grad_norm": 11.951933824951752,
"learning_rate": 7.894721963497213e-07,
"loss": 0.5991,
"step": 372
},
{
"epoch": 1.2233169129720853,
"grad_norm": 51.890411987411,
"learning_rate": 7.838713382712583e-07,
"loss": 0.5494,
"step": 373
},
{
"epoch": 1.2266009852216748,
"grad_norm": 46.51346153636678,
"learning_rate": 7.78277583323452e-07,
"loss": 0.5601,
"step": 374
},
{
"epoch": 1.2298850574712643,
"grad_norm": 32.19714846447316,
"learning_rate": 7.726911153466697e-07,
"loss": 0.5891,
"step": 375
},
{
"epoch": 1.2331691297208538,
"grad_norm": 30.416108844072596,
"learning_rate": 7.671121179417879e-07,
"loss": 0.57,
"step": 376
},
{
"epoch": 1.2364532019704433,
"grad_norm": 9.175156240446686,
"learning_rate": 7.615407744641618e-07,
"loss": 0.592,
"step": 377
},
{
"epoch": 1.2397372742200328,
"grad_norm": 16.89408306679232,
"learning_rate": 7.559772680175978e-07,
"loss": 0.5581,
"step": 378
},
{
"epoch": 1.2430213464696223,
"grad_norm": 36.581208571354075,
"learning_rate": 7.504217814483363e-07,
"loss": 0.5684,
"step": 379
},
{
"epoch": 1.2463054187192117,
"grad_norm": 197.1580272004894,
"learning_rate": 7.448744973390422e-07,
"loss": 0.5843,
"step": 380
},
{
"epoch": 1.2463054187192117,
"eval_loss": 0.3425952196121216,
"eval_runtime": 254.289,
"eval_samples_per_second": 8.125,
"eval_steps_per_second": 0.256,
"step": 380
},
{
"epoch": 1.2495894909688012,
"grad_norm": 38.00508215291586,
"learning_rate": 7.393355980028038e-07,
"loss": 0.5705,
"step": 381
},
{
"epoch": 1.2528735632183907,
"grad_norm": 36.66893410967853,
"learning_rate": 7.338052654771407e-07,
"loss": 0.5621,
"step": 382
},
{
"epoch": 1.2561576354679804,
"grad_norm": 72.0846419386083,
"learning_rate": 7.28283681518024e-07,
"loss": 0.6053,
"step": 383
},
{
"epoch": 1.2594417077175697,
"grad_norm": 12.87863278880059,
"learning_rate": 7.227710275938987e-07,
"loss": 0.585,
"step": 384
},
{
"epoch": 1.2627257799671594,
"grad_norm": 13.782452001612908,
"learning_rate": 7.172674848797217e-07,
"loss": 0.5973,
"step": 385
},
{
"epoch": 1.2660098522167487,
"grad_norm": 208.71637259406816,
"learning_rate": 7.117732342510092e-07,
"loss": 0.586,
"step": 386
},
{
"epoch": 1.2692939244663384,
"grad_norm": 16.912304799585513,
"learning_rate": 7.062884562778882e-07,
"loss": 0.5636,
"step": 387
},
{
"epoch": 1.2725779967159276,
"grad_norm": 28.08532516731275,
"learning_rate": 7.008133312191649e-07,
"loss": 0.5897,
"step": 388
},
{
"epoch": 1.2758620689655173,
"grad_norm": 55.25592778507127,
"learning_rate": 6.953480390164e-07,
"loss": 0.5581,
"step": 389
},
{
"epoch": 1.2791461412151066,
"grad_norm": 52.471536677355836,
"learning_rate": 6.898927592879944e-07,
"loss": 0.5488,
"step": 390
},
{
"epoch": 1.2824302134646963,
"grad_norm": 33.60924814778164,
"learning_rate": 6.844476713232862e-07,
"loss": 0.5452,
"step": 391
},
{
"epoch": 1.2857142857142856,
"grad_norm": 44.60405738816861,
"learning_rate": 6.79012954076658e-07,
"loss": 0.6057,
"step": 392
},
{
"epoch": 1.2889983579638753,
"grad_norm": 52.13213673114732,
"learning_rate": 6.735887861616555e-07,
"loss": 0.5718,
"step": 393
},
{
"epoch": 1.2922824302134646,
"grad_norm": 15.614998167223849,
"learning_rate": 6.681753458451189e-07,
"loss": 0.6077,
"step": 394
},
{
"epoch": 1.2955665024630543,
"grad_norm": 21.557430927559825,
"learning_rate": 6.627728110413213e-07,
"loss": 0.5724,
"step": 395
},
{
"epoch": 1.2988505747126438,
"grad_norm": 23.88672470819002,
"learning_rate": 6.573813593061235e-07,
"loss": 0.5623,
"step": 396
},
{
"epoch": 1.3021346469622332,
"grad_norm": 95.59341827721282,
"learning_rate": 6.520011678311381e-07,
"loss": 0.5932,
"step": 397
},
{
"epoch": 1.3054187192118227,
"grad_norm": 44.267252382206344,
"learning_rate": 6.466324134379065e-07,
"loss": 0.5699,
"step": 398
},
{
"epoch": 1.3087027914614122,
"grad_norm": 20.628188906562613,
"learning_rate": 6.412752725720864e-07,
"loss": 0.5971,
"step": 399
},
{
"epoch": 1.3119868637110017,
"grad_norm": 16.82849216948797,
"learning_rate": 6.359299212976534e-07,
"loss": 0.5843,
"step": 400
},
{
"epoch": 1.3152709359605912,
"grad_norm": 18.08566660075251,
"learning_rate": 6.305965352911161e-07,
"loss": 0.5404,
"step": 401
},
{
"epoch": 1.3185550082101807,
"grad_norm": 49.91580760003982,
"learning_rate": 6.252752898357397e-07,
"loss": 0.5531,
"step": 402
},
{
"epoch": 1.3218390804597702,
"grad_norm": 15.520146318832731,
"learning_rate": 6.199663598157875e-07,
"loss": 0.5872,
"step": 403
},
{
"epoch": 1.3251231527093597,
"grad_norm": 14.249014867996666,
"learning_rate": 6.146699197107715e-07,
"loss": 0.5623,
"step": 404
},
{
"epoch": 1.3284072249589491,
"grad_norm": 13.497082574794579,
"learning_rate": 6.093861435897207e-07,
"loss": 0.5707,
"step": 405
},
{
"epoch": 1.3316912972085386,
"grad_norm": 26.344677464265622,
"learning_rate": 6.041152051054575e-07,
"loss": 0.5597,
"step": 406
},
{
"epoch": 1.3349753694581281,
"grad_norm": 14.044122777177984,
"learning_rate": 5.988572774888912e-07,
"loss": 0.5924,
"step": 407
},
{
"epoch": 1.3382594417077176,
"grad_norm": 52.51863026094184,
"learning_rate": 5.936125335433264e-07,
"loss": 0.5774,
"step": 408
},
{
"epoch": 1.341543513957307,
"grad_norm": 36.13736458430556,
"learning_rate": 5.88381145638782e-07,
"loss": 0.5676,
"step": 409
},
{
"epoch": 1.3448275862068966,
"grad_norm": 15.292686883515266,
"learning_rate": 5.83163285706327e-07,
"loss": 0.5533,
"step": 410
},
{
"epoch": 1.348111658456486,
"grad_norm": 87.55364165940539,
"learning_rate": 5.779591252324286e-07,
"loss": 0.5676,
"step": 411
},
{
"epoch": 1.3513957307060755,
"grad_norm": 90.41445749482193,
"learning_rate": 5.72768835253319e-07,
"loss": 0.6106,
"step": 412
},
{
"epoch": 1.354679802955665,
"grad_norm": 75.68880847383154,
"learning_rate": 5.67592586349372e-07,
"loss": 0.5798,
"step": 413
},
{
"epoch": 1.3579638752052545,
"grad_norm": 24.386055319065594,
"learning_rate": 5.624305486394967e-07,
"loss": 0.5831,
"step": 414
},
{
"epoch": 1.361247947454844,
"grad_norm": 218.5561635346433,
"learning_rate": 5.57282891775548e-07,
"loss": 0.5958,
"step": 415
},
{
"epoch": 1.3645320197044335,
"grad_norm": 38.47004577556696,
"learning_rate": 5.5214978493675e-07,
"loss": 0.5522,
"step": 416
},
{
"epoch": 1.367816091954023,
"grad_norm": 11.116447526825615,
"learning_rate": 5.470313968241358e-07,
"loss": 0.5655,
"step": 417
},
{
"epoch": 1.3711001642036125,
"grad_norm": 33.49087738887472,
"learning_rate": 5.419278956550036e-07,
"loss": 0.5818,
"step": 418
},
{
"epoch": 1.374384236453202,
"grad_norm": 45.327655233207565,
"learning_rate": 5.368394491573876e-07,
"loss": 0.5618,
"step": 419
},
{
"epoch": 1.3776683087027914,
"grad_norm": 266.5398840292457,
"learning_rate": 5.317662245645469e-07,
"loss": 0.5549,
"step": 420
},
{
"epoch": 1.380952380952381,
"grad_norm": 20.553987754267183,
"learning_rate": 5.267083886094668e-07,
"loss": 0.5719,
"step": 421
},
{
"epoch": 1.3842364532019704,
"grad_norm": 74.60343344764495,
"learning_rate": 5.216661075193813e-07,
"loss": 0.5709,
"step": 422
},
{
"epoch": 1.38752052545156,
"grad_norm": 81.24452905189378,
"learning_rate": 5.166395470103091e-07,
"loss": 0.5839,
"step": 423
},
{
"epoch": 1.3908045977011494,
"grad_norm": 61.57043976405265,
"learning_rate": 5.116288722816085e-07,
"loss": 0.5881,
"step": 424
},
{
"epoch": 1.3940886699507389,
"grad_norm": 66.61917309787378,
"learning_rate": 5.066342480105459e-07,
"loss": 0.5396,
"step": 425
},
{
"epoch": 1.3973727422003284,
"grad_norm": 32.18086853290272,
"learning_rate": 5.016558383468851e-07,
"loss": 0.5743,
"step": 426
},
{
"epoch": 1.4006568144499179,
"grad_norm": 20.082111947795827,
"learning_rate": 4.966938069074921e-07,
"loss": 0.5531,
"step": 427
},
{
"epoch": 1.4039408866995073,
"grad_norm": 35.324879089856665,
"learning_rate": 4.91748316770958e-07,
"loss": 0.5685,
"step": 428
},
{
"epoch": 1.4072249589490968,
"grad_norm": 52.60744635091702,
"learning_rate": 4.868195304722391e-07,
"loss": 0.5892,
"step": 429
},
{
"epoch": 1.4105090311986863,
"grad_norm": 32.96054620176144,
"learning_rate": 4.819076099973152e-07,
"loss": 0.5593,
"step": 430
},
{
"epoch": 1.4137931034482758,
"grad_norm": 203.82839428938274,
"learning_rate": 4.770127167778669e-07,
"loss": 0.5699,
"step": 431
},
{
"epoch": 1.4170771756978653,
"grad_norm": 47.57730637955671,
"learning_rate": 4.7213501168596746e-07,
"loss": 0.5789,
"step": 432
},
{
"epoch": 1.4203612479474548,
"grad_norm": 67.44290592976088,
"learning_rate": 4.6727465502879846e-07,
"loss": 0.5556,
"step": 433
},
{
"epoch": 1.4236453201970443,
"grad_norm": 44.034037450189444,
"learning_rate": 4.6243180654337966e-07,
"loss": 0.5549,
"step": 434
},
{
"epoch": 1.4269293924466337,
"grad_norm": 16.61985345751509,
"learning_rate": 4.5760662539132077e-07,
"loss": 0.5747,
"step": 435
},
{
"epoch": 1.4302134646962232,
"grad_norm": 12.882168633818488,
"learning_rate": 4.5279927015358833e-07,
"loss": 0.5717,
"step": 436
},
{
"epoch": 1.4334975369458127,
"grad_norm": 20.653148975087017,
"learning_rate": 4.480098988252957e-07,
"loss": 0.6128,
"step": 437
},
{
"epoch": 1.4367816091954024,
"grad_norm": 23.18939767829898,
"learning_rate": 4.4323866881050945e-07,
"loss": 0.5692,
"step": 438
},
{
"epoch": 1.4400656814449917,
"grad_norm": 31.061046848137774,
"learning_rate": 4.384857369170771e-07,
"loss": 0.5611,
"step": 439
},
{
"epoch": 1.4433497536945814,
"grad_norm": 62.563561066656575,
"learning_rate": 4.337512593514728e-07,
"loss": 0.56,
"step": 440
},
{
"epoch": 1.4466338259441707,
"grad_norm": 17.20298230988248,
"learning_rate": 4.290353917136639e-07,
"loss": 0.5496,
"step": 441
},
{
"epoch": 1.4499178981937604,
"grad_norm": 10.022400404488588,
"learning_rate": 4.2433828899199807e-07,
"loss": 0.5743,
"step": 442
},
{
"epoch": 1.4532019704433496,
"grad_norm": 29.312045763415277,
"learning_rate": 4.1966010555810694e-07,
"loss": 0.5445,
"step": 443
},
{
"epoch": 1.4564860426929394,
"grad_norm": 31.664457234542283,
"learning_rate": 4.150009951618355e-07,
"loss": 0.55,
"step": 444
},
{
"epoch": 1.4597701149425286,
"grad_norm": 23.68782329905495,
"learning_rate": 4.103611109261872e-07,
"loss": 0.5957,
"step": 445
},
{
"epoch": 1.4630541871921183,
"grad_norm": 15.601252467779627,
"learning_rate": 4.0574060534229327e-07,
"loss": 0.5965,
"step": 446
},
{
"epoch": 1.4663382594417076,
"grad_norm": 16.094218846568037,
"learning_rate": 4.011396302643988e-07,
"loss": 0.5731,
"step": 447
},
{
"epoch": 1.4696223316912973,
"grad_norm": 45.95169139269504,
"learning_rate": 3.965583369048737e-07,
"loss": 0.5778,
"step": 448
},
{
"epoch": 1.4729064039408866,
"grad_norm": 16.32693066785533,
"learning_rate": 3.9199687582924246e-07,
"loss": 0.582,
"step": 449
},
{
"epoch": 1.4761904761904763,
"grad_norm": 15.3081339931233,
"learning_rate": 3.8745539695123577e-07,
"loss": 0.5865,
"step": 450
},
{
"epoch": 1.4794745484400658,
"grad_norm": 41.26715984492636,
"learning_rate": 3.829340495278639e-07,
"loss": 0.5671,
"step": 451
},
{
"epoch": 1.4827586206896552,
"grad_norm": 14.84205956050775,
"learning_rate": 3.7843298215451046e-07,
"loss": 0.5676,
"step": 452
},
{
"epoch": 1.4860426929392447,
"grad_norm": 28.834583020994884,
"learning_rate": 3.739523427600508e-07,
"loss": 0.5831,
"step": 453
},
{
"epoch": 1.4893267651888342,
"grad_norm": 29.206427309967527,
"learning_rate": 3.6949227860198707e-07,
"loss": 0.5808,
"step": 454
},
{
"epoch": 1.4926108374384237,
"grad_norm": 21.208424115794667,
"learning_rate": 3.6505293626161127e-07,
"loss": 0.5701,
"step": 455
},
{
"epoch": 1.4958949096880132,
"grad_norm": 60.972107883689446,
"learning_rate": 3.6063446163918663e-07,
"loss": 0.5812,
"step": 456
},
{
"epoch": 1.4958949096880132,
"eval_loss": 0.34147635102272034,
"eval_runtime": 253.4458,
"eval_samples_per_second": 8.152,
"eval_steps_per_second": 0.256,
"step": 456
},
{
"epoch": 1.4991789819376027,
"grad_norm": 102.55339346012296,
"learning_rate": 3.5623699994915355e-07,
"loss": 0.5955,
"step": 457
},
{
"epoch": 1.5024630541871922,
"grad_norm": 76.36859833115236,
"learning_rate": 3.518606957153557e-07,
"loss": 0.5682,
"step": 458
},
{
"epoch": 1.5057471264367817,
"grad_norm": 24.75361464219726,
"learning_rate": 3.475056927662912e-07,
"loss": 0.5787,
"step": 459
},
{
"epoch": 1.5090311986863711,
"grad_norm": 27.83323345966019,
"learning_rate": 3.4317213423038384e-07,
"loss": 0.562,
"step": 460
},
{
"epoch": 1.5123152709359606,
"grad_norm": 46.505159610329876,
"learning_rate": 3.3886016253128327e-07,
"loss": 0.5851,
"step": 461
},
{
"epoch": 1.5155993431855501,
"grad_norm": 28.610228426478272,
"learning_rate": 3.345699193831795e-07,
"loss": 0.6063,
"step": 462
},
{
"epoch": 1.5188834154351396,
"grad_norm": 15.024413491910803,
"learning_rate": 3.303015457861478e-07,
"loss": 0.5277,
"step": 463
},
{
"epoch": 1.522167487684729,
"grad_norm": 12.258290893394133,
"learning_rate": 3.2605518202151574e-07,
"loss": 0.5922,
"step": 464
},
{
"epoch": 1.5254515599343186,
"grad_norm": 104.31924222325128,
"learning_rate": 3.2183096764724914e-07,
"loss": 0.5649,
"step": 465
},
{
"epoch": 1.528735632183908,
"grad_norm": 42.86381800387519,
"learning_rate": 3.1762904149336946e-07,
"loss": 0.5743,
"step": 466
},
{
"epoch": 1.5320197044334976,
"grad_norm": 18.167710389726654,
"learning_rate": 3.134495416573883e-07,
"loss": 0.5572,
"step": 467
},
{
"epoch": 1.535303776683087,
"grad_norm": 15.343486359801176,
"learning_rate": 3.092926054997711e-07,
"loss": 0.5719,
"step": 468
},
{
"epoch": 1.5385878489326765,
"grad_norm": 25.820050498948465,
"learning_rate": 3.0515836963942054e-07,
"loss": 0.5785,
"step": 469
},
{
"epoch": 1.541871921182266,
"grad_norm": 47.76654830827422,
"learning_rate": 3.01046969949188e-07,
"loss": 0.571,
"step": 470
},
{
"epoch": 1.5451559934318555,
"grad_norm": 18.685300965930125,
"learning_rate": 2.969585415514064e-07,
"loss": 0.5899,
"step": 471
},
{
"epoch": 1.548440065681445,
"grad_norm": 33.45206718151848,
"learning_rate": 2.9289321881345254e-07,
"loss": 0.5661,
"step": 472
},
{
"epoch": 1.5517241379310345,
"grad_norm": 72.16443965311746,
"learning_rate": 2.8885113534332737e-07,
"loss": 0.5931,
"step": 473
},
{
"epoch": 1.555008210180624,
"grad_norm": 42.035242935885876,
"learning_rate": 2.848324239852672e-07,
"loss": 0.5587,
"step": 474
},
{
"epoch": 1.5582922824302134,
"grad_norm": 45.15146887717905,
"learning_rate": 2.80837216815378e-07,
"loss": 0.5728,
"step": 475
},
{
"epoch": 1.561576354679803,
"grad_norm": 21.81085786153228,
"learning_rate": 2.768656451372919e-07,
"loss": 0.6011,
"step": 476
},
{
"epoch": 1.5648604269293924,
"grad_norm": 14.491067969072652,
"learning_rate": 2.729178394778554e-07,
"loss": 0.5558,
"step": 477
},
{
"epoch": 1.568144499178982,
"grad_norm": 17.912743886068196,
"learning_rate": 2.6899392958283706e-07,
"loss": 0.5631,
"step": 478
},
{
"epoch": 1.5714285714285714,
"grad_norm": 64.4857516790575,
"learning_rate": 2.6509404441266535e-07,
"loss": 0.5841,
"step": 479
},
{
"epoch": 1.5747126436781609,
"grad_norm": 40.36341697674323,
"learning_rate": 2.6121831213818826e-07,
"loss": 0.5328,
"step": 480
},
{
"epoch": 1.5779967159277504,
"grad_norm": 28.64813563751768,
"learning_rate": 2.573668601364623e-07,
"loss": 0.5665,
"step": 481
},
{
"epoch": 1.5812807881773399,
"grad_norm": 20.644096206862233,
"learning_rate": 2.5353981498656505e-07,
"loss": 0.5579,
"step": 482
},
{
"epoch": 1.5845648604269293,
"grad_norm": 42.66803007868702,
"learning_rate": 2.497373024654373e-07,
"loss": 0.5692,
"step": 483
},
{
"epoch": 1.5878489326765188,
"grad_norm": 23.741999865754043,
"learning_rate": 2.459594475437472e-07,
"loss": 0.5731,
"step": 484
},
{
"epoch": 1.5911330049261085,
"grad_norm": 25.83928960154385,
"learning_rate": 2.4220637438178313e-07,
"loss": 0.5895,
"step": 485
},
{
"epoch": 1.5944170771756978,
"grad_norm": 53.46424769894938,
"learning_rate": 2.3847820632537564e-07,
"loss": 0.5807,
"step": 486
},
{
"epoch": 1.5977011494252875,
"grad_norm": 32.58548532914844,
"learning_rate": 2.3477506590183972e-07,
"loss": 0.5525,
"step": 487
},
{
"epoch": 1.6009852216748768,
"grad_norm": 40.67612047783848,
"learning_rate": 2.310970748159511e-07,
"loss": 0.6111,
"step": 488
},
{
"epoch": 1.6042692939244665,
"grad_norm": 31.120710466672246,
"learning_rate": 2.2744435394594497e-07,
"loss": 0.569,
"step": 489
},
{
"epoch": 1.6075533661740558,
"grad_norm": 11.431716307911014,
"learning_rate": 2.2381702333954433e-07,
"loss": 0.5804,
"step": 490
},
{
"epoch": 1.6108374384236455,
"grad_norm": 13.840123564647127,
"learning_rate": 2.2021520221001299e-07,
"loss": 0.57,
"step": 491
},
{
"epoch": 1.6141215106732347,
"grad_norm": 17.928907652041175,
"learning_rate": 2.1663900893223897e-07,
"loss": 0.5922,
"step": 492
},
{
"epoch": 1.6174055829228244,
"grad_norm": 57.74383968458873,
"learning_rate": 2.1308856103884277e-07,
"loss": 0.5681,
"step": 493
},
{
"epoch": 1.6206896551724137,
"grad_norm": 49.41941546684546,
"learning_rate": 2.0956397521631662e-07,
"loss": 0.5555,
"step": 494
},
{
"epoch": 1.6239737274220034,
"grad_norm": 30.887246526627205,
"learning_rate": 2.0606536730118763e-07,
"loss": 0.5666,
"step": 495
},
{
"epoch": 1.6272577996715927,
"grad_norm": 21.706062838691153,
"learning_rate": 2.0259285227621147e-07,
"loss": 0.5633,
"step": 496
},
{
"epoch": 1.6305418719211824,
"grad_norm": 31.266696706985737,
"learning_rate": 1.991465442665937e-07,
"loss": 0.5356,
"step": 497
},
{
"epoch": 1.6338259441707716,
"grad_norm": 29.3689515647996,
"learning_rate": 1.9572655653623882e-07,
"loss": 0.5912,
"step": 498
},
{
"epoch": 1.6371100164203614,
"grad_norm": 22.322656636658127,
"learning_rate": 1.9233300148402764e-07,
"loss": 0.5595,
"step": 499
},
{
"epoch": 1.6403940886699506,
"grad_norm": 44.171884336035255,
"learning_rate": 1.8896599064012298e-07,
"loss": 0.5643,
"step": 500
},
{
"epoch": 1.6436781609195403,
"grad_norm": 17.948977196325135,
"learning_rate": 1.8562563466230575e-07,
"loss": 0.5634,
"step": 501
},
{
"epoch": 1.6469622331691296,
"grad_norm": 11.97820725055819,
"learning_rate": 1.8231204333233607e-07,
"loss": 0.5598,
"step": 502
},
{
"epoch": 1.6502463054187193,
"grad_norm": 10.612778968035474,
"learning_rate": 1.790253255523465e-07,
"loss": 0.5803,
"step": 503
},
{
"epoch": 1.6535303776683086,
"grad_norm": 18.193226588472413,
"learning_rate": 1.7576558934126217e-07,
"loss": 0.5967,
"step": 504
},
{
"epoch": 1.6568144499178983,
"grad_norm": 12.00030723839093,
"learning_rate": 1.7253294183125222e-07,
"loss": 0.5796,
"step": 505
},
{
"epoch": 1.6600985221674875,
"grad_norm": 61.889660413482744,
"learning_rate": 1.6932748926420693e-07,
"loss": 0.6197,
"step": 506
},
{
"epoch": 1.6633825944170773,
"grad_norm": 14.289243326025748,
"learning_rate": 1.6614933698824728e-07,
"loss": 0.5715,
"step": 507
},
{
"epoch": 1.6666666666666665,
"grad_norm": 83.1922121560888,
"learning_rate": 1.6299858945426248e-07,
"loss": 0.6025,
"step": 508
},
{
"epoch": 1.6699507389162562,
"grad_norm": 21.289795508971476,
"learning_rate": 1.5987535021247667e-07,
"loss": 0.5613,
"step": 509
},
{
"epoch": 1.6732348111658455,
"grad_norm": 9.12840521329095,
"learning_rate": 1.5677972190904621e-07,
"loss": 0.5611,
"step": 510
},
{
"epoch": 1.6765188834154352,
"grad_norm": 80.54244783118818,
"learning_rate": 1.5371180628268587e-07,
"loss": 0.5773,
"step": 511
},
{
"epoch": 1.6798029556650245,
"grad_norm": 18.09168094972728,
"learning_rate": 1.5067170416132603e-07,
"loss": 0.5632,
"step": 512
},
{
"epoch": 1.6830870279146142,
"grad_norm": 76.91613553570883,
"learning_rate": 1.476595154587973e-07,
"loss": 0.5859,
"step": 513
},
{
"epoch": 1.6863711001642037,
"grad_norm": 13.09746011495923,
"learning_rate": 1.446753391715484e-07,
"loss": 0.5555,
"step": 514
},
{
"epoch": 1.6896551724137931,
"grad_norm": 26.40133497602893,
"learning_rate": 1.4171927337539104e-07,
"loss": 0.5222,
"step": 515
},
{
"epoch": 1.6929392446633826,
"grad_norm": 42.003291635263835,
"learning_rate": 1.3879141522227877e-07,
"loss": 0.5857,
"step": 516
},
{
"epoch": 1.6962233169129721,
"grad_norm": 15.138211036731974,
"learning_rate": 1.3589186093711223e-07,
"loss": 0.576,
"step": 517
},
{
"epoch": 1.6995073891625616,
"grad_norm": 15.749783656567669,
"learning_rate": 1.3302070581457713e-07,
"loss": 0.5616,
"step": 518
},
{
"epoch": 1.702791461412151,
"grad_norm": 19.998784663850575,
"learning_rate": 1.3017804421601298e-07,
"loss": 0.6002,
"step": 519
},
{
"epoch": 1.7060755336617406,
"grad_norm": 12.84175073867242,
"learning_rate": 1.273639695663108e-07,
"loss": 0.5693,
"step": 520
},
{
"epoch": 1.70935960591133,
"grad_norm": 48.66516503978152,
"learning_rate": 1.2457857435084407e-07,
"loss": 0.6153,
"step": 521
},
{
"epoch": 1.7126436781609196,
"grad_norm": 16.696763387794824,
"learning_rate": 1.2182195011242747e-07,
"loss": 0.5613,
"step": 522
},
{
"epoch": 1.715927750410509,
"grad_norm": 31.172316224112276,
"learning_rate": 1.1909418744831046e-07,
"loss": 0.605,
"step": 523
},
{
"epoch": 1.7192118226600985,
"grad_norm": 19.082497779204967,
"learning_rate": 1.1639537600719761e-07,
"loss": 0.5805,
"step": 524
},
{
"epoch": 1.722495894909688,
"grad_norm": 15.143716837248418,
"learning_rate": 1.1372560448630375e-07,
"loss": 0.608,
"step": 525
},
{
"epoch": 1.7257799671592775,
"grad_norm": 37.76170562223439,
"learning_rate": 1.1108496062843741e-07,
"loss": 0.5838,
"step": 526
},
{
"epoch": 1.729064039408867,
"grad_norm": 14.565254764998345,
"learning_rate": 1.0847353121911951e-07,
"loss": 0.5539,
"step": 527
},
{
"epoch": 1.7323481116584565,
"grad_norm": 32.47683113493387,
"learning_rate": 1.0589140208372871e-07,
"loss": 0.5575,
"step": 528
},
{
"epoch": 1.735632183908046,
"grad_norm": 23.18347373384476,
"learning_rate": 1.0333865808468201e-07,
"loss": 0.5566,
"step": 529
},
{
"epoch": 1.7389162561576355,
"grad_norm": 22.4422071324967,
"learning_rate": 1.0081538311864568e-07,
"loss": 0.576,
"step": 530
},
{
"epoch": 1.742200328407225,
"grad_norm": 14.059292999171634,
"learning_rate": 9.83216601137773e-08,
"loss": 0.5677,
"step": 531
},
{
"epoch": 1.7454844006568144,
"grad_norm": 22.050768611070254,
"learning_rate": 9.58575710270011e-08,
"loss": 0.5676,
"step": 532
},
{
"epoch": 1.7454844006568144,
"eval_loss": 0.34110337495803833,
"eval_runtime": 255.1137,
"eval_samples_per_second": 8.098,
"eval_steps_per_second": 0.255,
"step": 532
},
{
"epoch": 1.748768472906404,
"grad_norm": 28.327968167510832,
"learning_rate": 9.342319684131395e-08,
"loss": 0.553,
"step": 533
},
{
"epoch": 1.7520525451559934,
"grad_norm": 23.765559507553224,
"learning_rate": 9.101861756312368e-08,
"loss": 0.5751,
"step": 534
},
{
"epoch": 1.7553366174055829,
"grad_norm": 38.32793241416281,
"learning_rate": 8.864391221962064e-08,
"loss": 0.5721,
"step": 535
},
{
"epoch": 1.7586206896551724,
"grad_norm": 23.1435189905778,
"learning_rate": 8.62991588561791e-08,
"loss": 0.5372,
"step": 536
},
{
"epoch": 1.7619047619047619,
"grad_norm": 64.41410656507954,
"learning_rate": 8.398443453379266e-08,
"loss": 0.597,
"step": 537
},
{
"epoch": 1.7651888341543513,
"grad_norm": 26.570215615082457,
"learning_rate": 8.169981532654269e-08,
"loss": 0.5425,
"step": 538
},
{
"epoch": 1.7684729064039408,
"grad_norm": 19.853728565908686,
"learning_rate": 7.944537631909664e-08,
"loss": 0.5781,
"step": 539
},
{
"epoch": 1.7717569786535303,
"grad_norm": 40.09614615765141,
"learning_rate": 7.722119160424112e-08,
"loss": 0.5738,
"step": 540
},
{
"epoch": 1.7750410509031198,
"grad_norm": 56.37553140710831,
"learning_rate": 7.502733428044683e-08,
"loss": 0.5788,
"step": 541
},
{
"epoch": 1.7783251231527095,
"grad_norm": 25.79428871095927,
"learning_rate": 7.286387644946601e-08,
"loss": 0.5614,
"step": 542
},
{
"epoch": 1.7816091954022988,
"grad_norm": 37.57206986640528,
"learning_rate": 7.073088921396286e-08,
"loss": 0.6319,
"step": 543
},
{
"epoch": 1.7848932676518885,
"grad_norm": 78.07770173709457,
"learning_rate": 6.862844267517642e-08,
"loss": 0.5585,
"step": 544
},
{
"epoch": 1.7881773399014778,
"grad_norm": 45.27512632946113,
"learning_rate": 6.655660593061718e-08,
"loss": 0.5835,
"step": 545
},
{
"epoch": 1.7914614121510675,
"grad_norm": 43.444015827492656,
"learning_rate": 6.451544707179635e-08,
"loss": 0.5839,
"step": 546
},
{
"epoch": 1.7947454844006567,
"grad_norm": 30.129718093780813,
"learning_rate": 6.250503318198663e-08,
"loss": 0.5922,
"step": 547
},
{
"epoch": 1.7980295566502464,
"grad_norm": 12.289775538527161,
"learning_rate": 6.052543033401891e-08,
"loss": 0.5659,
"step": 548
},
{
"epoch": 1.8013136288998357,
"grad_norm": 22.117467884787505,
"learning_rate": 5.8576703588110953e-08,
"loss": 0.5833,
"step": 549
},
{
"epoch": 1.8045977011494254,
"grad_norm": 18.824028673171902,
"learning_rate": 5.665891698972769e-08,
"loss": 0.5515,
"step": 550
},
{
"epoch": 1.8078817733990147,
"grad_norm": 15.156245839885349,
"learning_rate": 5.4772133567477454e-08,
"loss": 0.58,
"step": 551
},
{
"epoch": 1.8111658456486044,
"grad_norm": 17.302260811016588,
"learning_rate": 5.291641533104052e-08,
"loss": 0.5443,
"step": 552
},
{
"epoch": 1.8144499178981937,
"grad_norm": 72.23650640208795,
"learning_rate": 5.109182326913053e-08,
"loss": 0.5913,
"step": 553
},
{
"epoch": 1.8177339901477834,
"grad_norm": 41.839452624673,
"learning_rate": 4.929841734749063e-08,
"loss": 0.5826,
"step": 554
},
{
"epoch": 1.8210180623973726,
"grad_norm": 150.9975193542559,
"learning_rate": 4.75362565069225e-08,
"loss": 0.5672,
"step": 555
},
{
"epoch": 1.8243021346469623,
"grad_norm": 177.35572727759407,
"learning_rate": 4.580539866134914e-08,
"loss": 0.5962,
"step": 556
},
{
"epoch": 1.8275862068965516,
"grad_norm": 11.471174551449721,
"learning_rate": 4.410590069591191e-08,
"loss": 0.5568,
"step": 557
},
{
"epoch": 1.8308702791461413,
"grad_norm": 41.93391866191071,
"learning_rate": 4.2437818465100306e-08,
"loss": 0.5872,
"step": 558
},
{
"epoch": 1.8341543513957306,
"grad_norm": 49.54043761767362,
"learning_rate": 4.080120679091681e-08,
"loss": 0.5782,
"step": 559
},
{
"epoch": 1.8374384236453203,
"grad_norm": 43.81540298910706,
"learning_rate": 3.919611946107493e-08,
"loss": 0.5726,
"step": 560
},
{
"epoch": 1.8407224958949095,
"grad_norm": 23.38299802172784,
"learning_rate": 3.762260922723182e-08,
"loss": 0.5658,
"step": 561
},
{
"epoch": 1.8440065681444993,
"grad_norm": 24.53118761357852,
"learning_rate": 3.6080727803253996e-08,
"loss": 0.5715,
"step": 562
},
{
"epoch": 1.8472906403940885,
"grad_norm": 19.417907974207314,
"learning_rate": 3.4570525863518164e-08,
"loss": 0.591,
"step": 563
},
{
"epoch": 1.8505747126436782,
"grad_norm": 10.93698504246622,
"learning_rate": 3.309205304124552e-08,
"loss": 0.598,
"step": 564
},
{
"epoch": 1.8538587848932675,
"grad_norm": 14.732319976335372,
"learning_rate": 3.164535792687095e-08,
"loss": 0.5739,
"step": 565
},
{
"epoch": 1.8571428571428572,
"grad_norm": 17.119911911692302,
"learning_rate": 3.0230488066445457e-08,
"loss": 0.5932,
"step": 566
},
{
"epoch": 1.8604269293924465,
"grad_norm": 18.02807205348044,
"learning_rate": 2.8847489960074133e-08,
"loss": 0.5698,
"step": 567
},
{
"epoch": 1.8637110016420362,
"grad_norm": 42.32940645169373,
"learning_rate": 2.749640906038797e-08,
"loss": 0.5845,
"step": 568
},
{
"epoch": 1.8669950738916257,
"grad_norm": 21.291255161442834,
"learning_rate": 2.617728977104927e-08,
"loss": 0.5767,
"step": 569
},
{
"epoch": 1.8702791461412152,
"grad_norm": 126.41354720715246,
"learning_rate": 2.4890175445293147e-08,
"loss": 0.548,
"step": 570
},
{
"epoch": 1.8735632183908046,
"grad_norm": 89.21090716126868,
"learning_rate": 2.3635108384502e-08,
"loss": 0.5526,
"step": 571
},
{
"epoch": 1.8768472906403941,
"grad_norm": 28.90642350240842,
"learning_rate": 2.2412129836816285e-08,
"loss": 0.5923,
"step": 572
},
{
"epoch": 1.8801313628899836,
"grad_norm": 33.649895492267454,
"learning_rate": 2.122127999577783e-08,
"loss": 0.5915,
"step": 573
},
{
"epoch": 1.883415435139573,
"grad_norm": 42.351338285848115,
"learning_rate": 2.0062597999009114e-08,
"loss": 0.5692,
"step": 574
},
{
"epoch": 1.8866995073891626,
"grad_norm": 21.17999827332917,
"learning_rate": 1.8936121926927507e-08,
"loss": 0.5528,
"step": 575
},
{
"epoch": 1.889983579638752,
"grad_norm": 15.439339266974041,
"learning_rate": 1.7841888801493176e-08,
"loss": 0.5694,
"step": 576
},
{
"epoch": 1.8932676518883416,
"grad_norm": 13.470779589141687,
"learning_rate": 1.6779934584992718e-08,
"loss": 0.5879,
"step": 577
},
{
"epoch": 1.896551724137931,
"grad_norm": 27.077399389940133,
"learning_rate": 1.575029417885687e-08,
"loss": 0.5641,
"step": 578
},
{
"epoch": 1.8998357963875205,
"grad_norm": 19.647395918482665,
"learning_rate": 1.4753001422514121e-08,
"loss": 0.5752,
"step": 579
},
{
"epoch": 1.90311986863711,
"grad_norm": 33.07929236692647,
"learning_rate": 1.3788089092277688e-08,
"loss": 0.5829,
"step": 580
},
{
"epoch": 1.9064039408866995,
"grad_norm": 80.07673176809891,
"learning_rate": 1.2855588900269054e-08,
"loss": 0.5497,
"step": 581
},
{
"epoch": 1.909688013136289,
"grad_norm": 19.376860505012647,
"learning_rate": 1.1955531493375138e-08,
"loss": 0.6026,
"step": 582
},
{
"epoch": 1.9129720853858785,
"grad_norm": 35.47816644500678,
"learning_rate": 1.108794645224187e-08,
"loss": 0.5666,
"step": 583
},
{
"epoch": 1.916256157635468,
"grad_norm": 115.1418329735163,
"learning_rate": 1.0252862290301089e-08,
"loss": 0.5753,
"step": 584
},
{
"epoch": 1.9195402298850575,
"grad_norm": 30.510225961444675,
"learning_rate": 9.450306452834178e-09,
"loss": 0.5848,
"step": 585
},
{
"epoch": 1.922824302134647,
"grad_norm": 46.46127331757118,
"learning_rate": 8.68030531606967e-09,
"loss": 0.5815,
"step": 586
},
{
"epoch": 1.9261083743842364,
"grad_norm": 21.7313503869492,
"learning_rate": 7.94288418631639e-09,
"loss": 0.5717,
"step": 587
},
{
"epoch": 1.929392446633826,
"grad_norm": 127.17771447820743,
"learning_rate": 7.2380672991319e-09,
"loss": 0.571,
"step": 588
},
{
"epoch": 1.9326765188834154,
"grad_norm": 31.316585311430142,
"learning_rate": 6.565877818526244e-09,
"loss": 0.5453,
"step": 589
},
{
"epoch": 1.935960591133005,
"grad_norm": 51.54565942906159,
"learning_rate": 5.926337836199891e-09,
"loss": 0.5669,
"step": 590
},
{
"epoch": 1.9392446633825944,
"grad_norm": 17.804113285989903,
"learning_rate": 5.319468370818536e-09,
"loss": 0.601,
"step": 591
},
{
"epoch": 1.9425287356321839,
"grad_norm": 68.75929656403164,
"learning_rate": 4.745289367321658e-09,
"loss": 0.5976,
"step": 592
},
{
"epoch": 1.9458128078817734,
"grad_norm": 11.91134893164374,
"learning_rate": 4.203819696267485e-09,
"loss": 0.5482,
"step": 593
},
{
"epoch": 1.9490968801313628,
"grad_norm": 17.418945183974255,
"learning_rate": 3.6950771532126003e-09,
"loss": 0.5732,
"step": 594
},
{
"epoch": 1.9523809523809523,
"grad_norm": 12.085404257651767,
"learning_rate": 3.219078458127078e-09,
"loss": 0.5876,
"step": 595
},
{
"epoch": 1.9556650246305418,
"grad_norm": 28.855096188397088,
"learning_rate": 2.775839254844925e-09,
"loss": 0.5689,
"step": 596
},
{
"epoch": 1.9589490968801315,
"grad_norm": 40.37266990923965,
"learning_rate": 2.3653741105499336e-09,
"loss": 0.5541,
"step": 597
},
{
"epoch": 1.9622331691297208,
"grad_norm": 55.51188079956238,
"learning_rate": 1.98769651529751e-09,
"loss": 0.5642,
"step": 598
},
{
"epoch": 1.9655172413793105,
"grad_norm": 34.1692976262017,
"learning_rate": 1.6428188815703626e-09,
"loss": 0.5595,
"step": 599
},
{
"epoch": 1.9688013136288998,
"grad_norm": 14.947354957753587,
"learning_rate": 1.330752543871161e-09,
"loss": 0.6005,
"step": 600
},
{
"epoch": 1.9720853858784895,
"grad_norm": 45.62738122621848,
"learning_rate": 1.0515077583498344e-09,
"loss": 0.5665,
"step": 601
},
{
"epoch": 1.9753694581280787,
"grad_norm": 21.725111315877975,
"learning_rate": 8.050937024666193e-10,
"loss": 0.5893,
"step": 602
},
{
"epoch": 1.9786535303776684,
"grad_norm": 35.375136168343055,
"learning_rate": 5.915184746904112e-10,
"loss": 0.5841,
"step": 603
},
{
"epoch": 1.9819376026272577,
"grad_norm": 18.604441345264426,
"learning_rate": 4.107890942325332e-10,
"loss": 0.5554,
"step": 604
},
{
"epoch": 1.9852216748768474,
"grad_norm": 119.72001100708987,
"learning_rate": 2.6291150081603207e-10,
"loss": 0.5614,
"step": 605
},
{
"epoch": 1.9885057471264367,
"grad_norm": 16.41601120115481,
"learning_rate": 1.4789055448061193e-10,
"loss": 0.5533,
"step": 606
},
{
"epoch": 1.9917898193760264,
"grad_norm": 124.04621529731233,
"learning_rate": 6.57300354227619e-11,
"loss": 0.5695,
"step": 607
},
{
"epoch": 1.9950738916256157,
"grad_norm": 41.47359974511061,
"learning_rate": 1.6432643871633346e-11,
"loss": 0.5409,
"step": 608
},
{
"epoch": 1.9950738916256157,
"eval_loss": 0.34100207686424255,
"eval_runtime": 255.0349,
"eval_samples_per_second": 8.101,
"eval_steps_per_second": 0.255,
"step": 608
}
],
"logging_steps": 1,
"max_steps": 608,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 152,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.488667559261635e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}