{ "best_global_step": 606, "best_metric": 2.5098254680633545, "best_model_checkpoint": "./gpt-neo-1.3B-qlora/checkpoint-606", "epoch": 4.0, "eval_steps": 500, "global_step": 808, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004956629491945477, "grad_norm": 0.5178479552268982, "learning_rate": 0.0, "loss": 4.4997, "step": 1 }, { "epoch": 0.009913258983890954, "grad_norm": 0.5379428863525391, "learning_rate": 8.000000000000001e-06, "loss": 4.0383, "step": 2 }, { "epoch": 0.01486988847583643, "grad_norm": 0.456396222114563, "learning_rate": 1.6000000000000003e-05, "loss": 3.8483, "step": 3 }, { "epoch": 0.01982651796778191, "grad_norm": 0.4702649712562561, "learning_rate": 2.4e-05, "loss": 3.7618, "step": 4 }, { "epoch": 0.024783147459727387, "grad_norm": 0.4205624461174011, "learning_rate": 3.2000000000000005e-05, "loss": 3.7436, "step": 5 }, { "epoch": 0.02973977695167286, "grad_norm": 0.562913179397583, "learning_rate": 4e-05, "loss": 4.3193, "step": 6 }, { "epoch": 0.03469640644361834, "grad_norm": 0.4989069700241089, "learning_rate": 4.8e-05, "loss": 3.8805, "step": 7 }, { "epoch": 0.03965303593556382, "grad_norm": 0.5795283913612366, "learning_rate": 5.6000000000000006e-05, "loss": 3.751, "step": 8 }, { "epoch": 0.04460966542750929, "grad_norm": 0.4642927348613739, "learning_rate": 6.400000000000001e-05, "loss": 3.6798, "step": 9 }, { "epoch": 0.04956629491945477, "grad_norm": 0.49269798398017883, "learning_rate": 7.2e-05, "loss": 3.7732, "step": 10 }, { "epoch": 0.05452292441140025, "grad_norm": 0.7436097860336304, "learning_rate": 8e-05, "loss": 4.5229, "step": 11 }, { "epoch": 0.05947955390334572, "grad_norm": 0.6071892976760864, "learning_rate": 8.800000000000001e-05, "loss": 3.7029, "step": 12 }, { "epoch": 0.0644361833952912, "grad_norm": 0.598869800567627, "learning_rate": 9.6e-05, "loss": 4.1203, "step": 13 }, { "epoch": 0.06939281288723669, "grad_norm": 0.5983518958091736, "learning_rate": 0.00010400000000000001, "loss": 3.9319, "step": 14 }, { "epoch": 0.07434944237918216, "grad_norm": 0.5899732112884521, "learning_rate": 0.00011200000000000001, "loss": 3.7153, "step": 15 }, { "epoch": 0.07930607187112763, "grad_norm": 0.6922647356987, "learning_rate": 0.00012, "loss": 3.9409, "step": 16 }, { "epoch": 0.08426270136307311, "grad_norm": 0.6090132594108582, "learning_rate": 0.00012800000000000002, "loss": 3.8406, "step": 17 }, { "epoch": 0.08921933085501858, "grad_norm": 0.6153061389923096, "learning_rate": 0.00013600000000000003, "loss": 3.5257, "step": 18 }, { "epoch": 0.09417596034696406, "grad_norm": 0.7484270930290222, "learning_rate": 0.000144, "loss": 3.6834, "step": 19 }, { "epoch": 0.09913258983890955, "grad_norm": 0.7390730381011963, "learning_rate": 0.000152, "loss": 3.5609, "step": 20 }, { "epoch": 0.10408921933085502, "grad_norm": 0.7254196405410767, "learning_rate": 0.00016, "loss": 3.0742, "step": 21 }, { "epoch": 0.1090458488228005, "grad_norm": 0.7760971784591675, "learning_rate": 0.000168, "loss": 3.336, "step": 22 }, { "epoch": 0.11400247831474597, "grad_norm": 0.7879131436347961, "learning_rate": 0.00017600000000000002, "loss": 3.6089, "step": 23 }, { "epoch": 0.11895910780669144, "grad_norm": 0.7201817631721497, "learning_rate": 0.00018400000000000003, "loss": 3.4623, "step": 24 }, { "epoch": 0.12391573729863693, "grad_norm": 0.9186316132545471, "learning_rate": 0.000192, "loss": 3.7456, "step": 25 }, { "epoch": 0.1288723667905824, "grad_norm": 0.8382372260093689, "learning_rate": 0.0002, "loss": 3.5368, "step": 26 }, { "epoch": 0.13382899628252787, "grad_norm": 1.2407952547073364, "learning_rate": 0.00019974457215836527, "loss": 3.36, "step": 27 }, { "epoch": 0.13878562577447337, "grad_norm": 0.9732126593589783, "learning_rate": 0.00019948914431673054, "loss": 3.5255, "step": 28 }, { "epoch": 0.14374225526641884, "grad_norm": 0.971034586429596, "learning_rate": 0.0001992337164750958, "loss": 3.1694, "step": 29 }, { "epoch": 0.14869888475836432, "grad_norm": 1.368577480316162, "learning_rate": 0.00019897828863346106, "loss": 3.6517, "step": 30 }, { "epoch": 0.1536555142503098, "grad_norm": 1.1814148426055908, "learning_rate": 0.00019872286079182633, "loss": 3.3308, "step": 31 }, { "epoch": 0.15861214374225527, "grad_norm": 1.3007687330245972, "learning_rate": 0.0001984674329501916, "loss": 3.4267, "step": 32 }, { "epoch": 0.16356877323420074, "grad_norm": 1.053661823272705, "learning_rate": 0.00019821200510855685, "loss": 3.4749, "step": 33 }, { "epoch": 0.16852540272614622, "grad_norm": 1.454937219619751, "learning_rate": 0.0001979565772669221, "loss": 3.1697, "step": 34 }, { "epoch": 0.1734820322180917, "grad_norm": 1.151451587677002, "learning_rate": 0.00019770114942528738, "loss": 2.8761, "step": 35 }, { "epoch": 0.17843866171003717, "grad_norm": 1.0200417041778564, "learning_rate": 0.00019744572158365262, "loss": 3.3697, "step": 36 }, { "epoch": 0.18339529120198264, "grad_norm": 1.136304259300232, "learning_rate": 0.0001971902937420179, "loss": 3.41, "step": 37 }, { "epoch": 0.18835192069392812, "grad_norm": 1.0931888818740845, "learning_rate": 0.00019693486590038314, "loss": 3.5437, "step": 38 }, { "epoch": 0.19330855018587362, "grad_norm": 1.4094620943069458, "learning_rate": 0.00019667943805874843, "loss": 3.6909, "step": 39 }, { "epoch": 0.1982651796778191, "grad_norm": 0.9859973192214966, "learning_rate": 0.00019642401021711367, "loss": 3.1319, "step": 40 }, { "epoch": 0.20322180916976457, "grad_norm": 1.330928921699524, "learning_rate": 0.00019616858237547893, "loss": 3.671, "step": 41 }, { "epoch": 0.20817843866171004, "grad_norm": 0.8949196338653564, "learning_rate": 0.0001959131545338442, "loss": 2.7301, "step": 42 }, { "epoch": 0.21313506815365552, "grad_norm": 0.8063453435897827, "learning_rate": 0.00019565772669220946, "loss": 2.7622, "step": 43 }, { "epoch": 0.218091697645601, "grad_norm": 1.093867301940918, "learning_rate": 0.00019540229885057472, "loss": 3.0352, "step": 44 }, { "epoch": 0.22304832713754646, "grad_norm": 1.1064800024032593, "learning_rate": 0.00019514687100893999, "loss": 3.3366, "step": 45 }, { "epoch": 0.22800495662949194, "grad_norm": 0.9350125193595886, "learning_rate": 0.00019489144316730525, "loss": 2.7895, "step": 46 }, { "epoch": 0.23296158612143741, "grad_norm": 1.234007716178894, "learning_rate": 0.0001946360153256705, "loss": 3.4759, "step": 47 }, { "epoch": 0.2379182156133829, "grad_norm": 1.1140137910842896, "learning_rate": 0.00019438058748403575, "loss": 3.5556, "step": 48 }, { "epoch": 0.24287484510532836, "grad_norm": 1.0842703580856323, "learning_rate": 0.00019412515964240104, "loss": 3.7215, "step": 49 }, { "epoch": 0.24783147459727387, "grad_norm": 0.9289716482162476, "learning_rate": 0.00019386973180076628, "loss": 2.8763, "step": 50 }, { "epoch": 0.2527881040892193, "grad_norm": 0.9746398329734802, "learning_rate": 0.00019361430395913157, "loss": 2.8321, "step": 51 }, { "epoch": 0.2577447335811648, "grad_norm": 1.1200573444366455, "learning_rate": 0.0001933588761174968, "loss": 3.1944, "step": 52 }, { "epoch": 0.26270136307311026, "grad_norm": 1.002164363861084, "learning_rate": 0.0001931034482758621, "loss": 2.9787, "step": 53 }, { "epoch": 0.26765799256505574, "grad_norm": 1.002467155456543, "learning_rate": 0.00019284802043422733, "loss": 3.0303, "step": 54 }, { "epoch": 0.27261462205700127, "grad_norm": 0.8421512842178345, "learning_rate": 0.0001925925925925926, "loss": 2.7362, "step": 55 }, { "epoch": 0.27757125154894674, "grad_norm": 0.9212812185287476, "learning_rate": 0.00019233716475095786, "loss": 2.7722, "step": 56 }, { "epoch": 0.2825278810408922, "grad_norm": 0.9771202802658081, "learning_rate": 0.00019208173690932312, "loss": 2.813, "step": 57 }, { "epoch": 0.2874845105328377, "grad_norm": 0.9651036262512207, "learning_rate": 0.00019182630906768838, "loss": 3.1953, "step": 58 }, { "epoch": 0.29244114002478316, "grad_norm": 0.8853598237037659, "learning_rate": 0.00019157088122605365, "loss": 2.5714, "step": 59 }, { "epoch": 0.29739776951672864, "grad_norm": 1.213663101196289, "learning_rate": 0.0001913154533844189, "loss": 3.4754, "step": 60 }, { "epoch": 0.3023543990086741, "grad_norm": 1.020679235458374, "learning_rate": 0.00019106002554278417, "loss": 2.64, "step": 61 }, { "epoch": 0.3073110285006196, "grad_norm": 1.0737823247909546, "learning_rate": 0.00019080459770114944, "loss": 2.8907, "step": 62 }, { "epoch": 0.31226765799256506, "grad_norm": 0.8646718263626099, "learning_rate": 0.0001905491698595147, "loss": 2.7505, "step": 63 }, { "epoch": 0.31722428748451054, "grad_norm": 1.219244122505188, "learning_rate": 0.00019029374201787996, "loss": 3.4013, "step": 64 }, { "epoch": 0.322180916976456, "grad_norm": 1.0713865756988525, "learning_rate": 0.00019003831417624523, "loss": 3.3184, "step": 65 }, { "epoch": 0.3271375464684015, "grad_norm": 1.0803613662719727, "learning_rate": 0.0001897828863346105, "loss": 2.9004, "step": 66 }, { "epoch": 0.33209417596034696, "grad_norm": 0.9669873714447021, "learning_rate": 0.00018952745849297575, "loss": 2.9939, "step": 67 }, { "epoch": 0.33705080545229243, "grad_norm": 1.0816727876663208, "learning_rate": 0.00018927203065134102, "loss": 2.9023, "step": 68 }, { "epoch": 0.3420074349442379, "grad_norm": 1.0009043216705322, "learning_rate": 0.00018901660280970628, "loss": 2.8928, "step": 69 }, { "epoch": 0.3469640644361834, "grad_norm": 1.0465470552444458, "learning_rate": 0.00018876117496807154, "loss": 3.0128, "step": 70 }, { "epoch": 0.35192069392812886, "grad_norm": 1.0517007112503052, "learning_rate": 0.00018850574712643678, "loss": 3.0419, "step": 71 }, { "epoch": 0.35687732342007433, "grad_norm": 1.1139233112335205, "learning_rate": 0.00018825031928480207, "loss": 2.7776, "step": 72 }, { "epoch": 0.3618339529120198, "grad_norm": 1.0560076236724854, "learning_rate": 0.0001879948914431673, "loss": 3.276, "step": 73 }, { "epoch": 0.3667905824039653, "grad_norm": 1.054962396621704, "learning_rate": 0.0001877394636015326, "loss": 2.8159, "step": 74 }, { "epoch": 0.37174721189591076, "grad_norm": 1.1142494678497314, "learning_rate": 0.00018748403575989783, "loss": 2.829, "step": 75 }, { "epoch": 0.37670384138785623, "grad_norm": 1.9403822422027588, "learning_rate": 0.00018722860791826312, "loss": 3.2541, "step": 76 }, { "epoch": 0.38166047087980176, "grad_norm": 1.3220140933990479, "learning_rate": 0.00018697318007662836, "loss": 3.5106, "step": 77 }, { "epoch": 0.38661710037174724, "grad_norm": 1.1846554279327393, "learning_rate": 0.00018671775223499362, "loss": 3.0229, "step": 78 }, { "epoch": 0.3915737298636927, "grad_norm": 0.9640692472457886, "learning_rate": 0.00018646232439335889, "loss": 2.6364, "step": 79 }, { "epoch": 0.3965303593556382, "grad_norm": 0.9236942529678345, "learning_rate": 0.00018620689655172415, "loss": 2.8003, "step": 80 }, { "epoch": 0.40148698884758366, "grad_norm": 1.1610803604125977, "learning_rate": 0.0001859514687100894, "loss": 3.2429, "step": 81 }, { "epoch": 0.40644361833952913, "grad_norm": 1.1774544715881348, "learning_rate": 0.00018569604086845468, "loss": 3.0449, "step": 82 }, { "epoch": 0.4114002478314746, "grad_norm": 2.4880337715148926, "learning_rate": 0.00018544061302681994, "loss": 2.8562, "step": 83 }, { "epoch": 0.4163568773234201, "grad_norm": 0.991893470287323, "learning_rate": 0.0001851851851851852, "loss": 2.7547, "step": 84 }, { "epoch": 0.42131350681536556, "grad_norm": 1.6182878017425537, "learning_rate": 0.00018492975734355044, "loss": 3.5209, "step": 85 }, { "epoch": 0.42627013630731103, "grad_norm": 0.9980579018592834, "learning_rate": 0.00018467432950191573, "loss": 2.6094, "step": 86 }, { "epoch": 0.4312267657992565, "grad_norm": 0.8960520625114441, "learning_rate": 0.00018441890166028097, "loss": 2.5804, "step": 87 }, { "epoch": 0.436183395291202, "grad_norm": 1.015406847000122, "learning_rate": 0.00018416347381864626, "loss": 2.5902, "step": 88 }, { "epoch": 0.44114002478314746, "grad_norm": 0.9714847803115845, "learning_rate": 0.0001839080459770115, "loss": 2.841, "step": 89 }, { "epoch": 0.44609665427509293, "grad_norm": 1.069502592086792, "learning_rate": 0.00018365261813537678, "loss": 2.7831, "step": 90 }, { "epoch": 0.4510532837670384, "grad_norm": 0.8945897221565247, "learning_rate": 0.00018339719029374202, "loss": 2.333, "step": 91 }, { "epoch": 0.4560099132589839, "grad_norm": 1.0069011449813843, "learning_rate": 0.00018314176245210728, "loss": 2.9101, "step": 92 }, { "epoch": 0.46096654275092935, "grad_norm": 1.3870095014572144, "learning_rate": 0.00018288633461047255, "loss": 2.8714, "step": 93 }, { "epoch": 0.46592317224287483, "grad_norm": 1.0033080577850342, "learning_rate": 0.0001826309067688378, "loss": 2.7687, "step": 94 }, { "epoch": 0.4708798017348203, "grad_norm": 1.1281189918518066, "learning_rate": 0.00018237547892720307, "loss": 2.7909, "step": 95 }, { "epoch": 0.4758364312267658, "grad_norm": 1.1801944971084595, "learning_rate": 0.00018212005108556834, "loss": 2.9649, "step": 96 }, { "epoch": 0.48079306071871125, "grad_norm": 0.9499044418334961, "learning_rate": 0.0001818646232439336, "loss": 2.7359, "step": 97 }, { "epoch": 0.4857496902106567, "grad_norm": 1.0616381168365479, "learning_rate": 0.00018160919540229886, "loss": 2.6311, "step": 98 }, { "epoch": 0.49070631970260226, "grad_norm": 1.2957072257995605, "learning_rate": 0.0001813537675606641, "loss": 3.3982, "step": 99 }, { "epoch": 0.49566294919454773, "grad_norm": 1.4070500135421753, "learning_rate": 0.0001810983397190294, "loss": 3.4525, "step": 100 }, { "epoch": 0.5006195786864932, "grad_norm": 1.0367472171783447, "learning_rate": 0.00018084291187739463, "loss": 2.5367, "step": 101 }, { "epoch": 0.5055762081784386, "grad_norm": 1.1069153547286987, "learning_rate": 0.00018058748403575992, "loss": 2.8477, "step": 102 }, { "epoch": 0.5105328376703842, "grad_norm": 1.6208503246307373, "learning_rate": 0.00018033205619412515, "loss": 3.0734, "step": 103 }, { "epoch": 0.5154894671623296, "grad_norm": 1.0800952911376953, "learning_rate": 0.00018007662835249044, "loss": 3.0183, "step": 104 }, { "epoch": 0.5204460966542751, "grad_norm": 1.0784571170806885, "learning_rate": 0.00017982120051085568, "loss": 2.8266, "step": 105 }, { "epoch": 0.5254027261462205, "grad_norm": 1.3548469543457031, "learning_rate": 0.00017956577266922094, "loss": 2.8956, "step": 106 }, { "epoch": 0.530359355638166, "grad_norm": 0.967523455619812, "learning_rate": 0.0001793103448275862, "loss": 2.4301, "step": 107 }, { "epoch": 0.5353159851301115, "grad_norm": 1.3245588541030884, "learning_rate": 0.00017905491698595147, "loss": 2.6761, "step": 108 }, { "epoch": 0.540272614622057, "grad_norm": 1.3608037233352661, "learning_rate": 0.00017879948914431673, "loss": 2.5499, "step": 109 }, { "epoch": 0.5452292441140025, "grad_norm": 1.27790367603302, "learning_rate": 0.000178544061302682, "loss": 2.8868, "step": 110 }, { "epoch": 0.550185873605948, "grad_norm": 1.031919002532959, "learning_rate": 0.00017828863346104726, "loss": 2.6039, "step": 111 }, { "epoch": 0.5551425030978935, "grad_norm": 1.1233882904052734, "learning_rate": 0.00017803320561941252, "loss": 2.7838, "step": 112 }, { "epoch": 0.5600991325898389, "grad_norm": 1.2760884761810303, "learning_rate": 0.00017777777777777779, "loss": 3.0326, "step": 113 }, { "epoch": 0.5650557620817844, "grad_norm": 1.0590308904647827, "learning_rate": 0.00017752234993614305, "loss": 2.8617, "step": 114 }, { "epoch": 0.5700123915737298, "grad_norm": 1.1054567098617554, "learning_rate": 0.0001772669220945083, "loss": 2.8121, "step": 115 }, { "epoch": 0.5749690210656754, "grad_norm": 1.1285626888275146, "learning_rate": 0.00017701149425287358, "loss": 2.7076, "step": 116 }, { "epoch": 0.5799256505576208, "grad_norm": 1.1753144264221191, "learning_rate": 0.00017675606641123884, "loss": 2.8147, "step": 117 }, { "epoch": 0.5848822800495663, "grad_norm": 1.167649745941162, "learning_rate": 0.0001765006385696041, "loss": 2.8485, "step": 118 }, { "epoch": 0.5898389095415117, "grad_norm": 1.1729902029037476, "learning_rate": 0.00017624521072796937, "loss": 2.7919, "step": 119 }, { "epoch": 0.5947955390334573, "grad_norm": 1.14457368850708, "learning_rate": 0.0001759897828863346, "loss": 2.6986, "step": 120 }, { "epoch": 0.5997521685254027, "grad_norm": 1.020176649093628, "learning_rate": 0.0001757343550446999, "loss": 2.3607, "step": 121 }, { "epoch": 0.6047087980173482, "grad_norm": 1.3097666501998901, "learning_rate": 0.00017547892720306513, "loss": 2.8609, "step": 122 }, { "epoch": 0.6096654275092936, "grad_norm": 1.139198899269104, "learning_rate": 0.00017522349936143042, "loss": 2.8653, "step": 123 }, { "epoch": 0.6146220570012392, "grad_norm": 1.3521591424942017, "learning_rate": 0.00017496807151979566, "loss": 2.6918, "step": 124 }, { "epoch": 0.6195786864931846, "grad_norm": 1.0930819511413574, "learning_rate": 0.00017471264367816095, "loss": 2.7736, "step": 125 }, { "epoch": 0.6245353159851301, "grad_norm": 1.0400232076644897, "learning_rate": 0.00017445721583652618, "loss": 2.376, "step": 126 }, { "epoch": 0.6294919454770755, "grad_norm": 1.5003938674926758, "learning_rate": 0.00017420178799489145, "loss": 2.7772, "step": 127 }, { "epoch": 0.6344485749690211, "grad_norm": 1.3468868732452393, "learning_rate": 0.0001739463601532567, "loss": 2.7663, "step": 128 }, { "epoch": 0.6394052044609665, "grad_norm": 1.1979831457138062, "learning_rate": 0.00017369093231162197, "loss": 2.3247, "step": 129 }, { "epoch": 0.644361833952912, "grad_norm": 1.50636625289917, "learning_rate": 0.00017343550446998724, "loss": 3.2306, "step": 130 }, { "epoch": 0.6493184634448576, "grad_norm": 1.3549541234970093, "learning_rate": 0.0001731800766283525, "loss": 2.9172, "step": 131 }, { "epoch": 0.654275092936803, "grad_norm": 1.2008872032165527, "learning_rate": 0.00017292464878671776, "loss": 2.8561, "step": 132 }, { "epoch": 0.6592317224287485, "grad_norm": 1.166435956954956, "learning_rate": 0.00017266922094508303, "loss": 2.6782, "step": 133 }, { "epoch": 0.6641883519206939, "grad_norm": 1.8413958549499512, "learning_rate": 0.00017241379310344826, "loss": 2.932, "step": 134 }, { "epoch": 0.6691449814126395, "grad_norm": 1.2101318836212158, "learning_rate": 0.00017215836526181355, "loss": 3.0967, "step": 135 }, { "epoch": 0.6741016109045849, "grad_norm": 1.160033106803894, "learning_rate": 0.0001719029374201788, "loss": 2.79, "step": 136 }, { "epoch": 0.6790582403965304, "grad_norm": 1.367043375968933, "learning_rate": 0.00017164750957854408, "loss": 2.6743, "step": 137 }, { "epoch": 0.6840148698884758, "grad_norm": 0.9590573906898499, "learning_rate": 0.00017139208173690932, "loss": 2.6285, "step": 138 }, { "epoch": 0.6889714993804213, "grad_norm": 1.1026114225387573, "learning_rate": 0.0001711366538952746, "loss": 2.6158, "step": 139 }, { "epoch": 0.6939281288723668, "grad_norm": 1.2949557304382324, "learning_rate": 0.00017088122605363984, "loss": 2.8041, "step": 140 }, { "epoch": 0.6988847583643123, "grad_norm": 1.1252179145812988, "learning_rate": 0.0001706257982120051, "loss": 2.6227, "step": 141 }, { "epoch": 0.7038413878562577, "grad_norm": 1.2461496591567993, "learning_rate": 0.00017037037037037037, "loss": 2.9447, "step": 142 }, { "epoch": 0.7087980173482032, "grad_norm": 1.0550256967544556, "learning_rate": 0.00017011494252873563, "loss": 2.5931, "step": 143 }, { "epoch": 0.7137546468401487, "grad_norm": 1.1814954280853271, "learning_rate": 0.0001698595146871009, "loss": 2.6902, "step": 144 }, { "epoch": 0.7187112763320942, "grad_norm": 1.3461589813232422, "learning_rate": 0.00016960408684546616, "loss": 3.1877, "step": 145 }, { "epoch": 0.7236679058240396, "grad_norm": 1.5511823892593384, "learning_rate": 0.00016934865900383142, "loss": 2.8915, "step": 146 }, { "epoch": 0.7286245353159851, "grad_norm": 1.1968817710876465, "learning_rate": 0.00016909323116219669, "loss": 2.9987, "step": 147 }, { "epoch": 0.7335811648079306, "grad_norm": 1.4092822074890137, "learning_rate": 0.00016883780332056195, "loss": 2.8595, "step": 148 }, { "epoch": 0.7385377942998761, "grad_norm": 1.397154450416565, "learning_rate": 0.0001685823754789272, "loss": 2.8689, "step": 149 }, { "epoch": 0.7434944237918215, "grad_norm": 1.3408515453338623, "learning_rate": 0.00016832694763729248, "loss": 2.8842, "step": 150 }, { "epoch": 0.748451053283767, "grad_norm": 1.8174018859863281, "learning_rate": 0.00016807151979565774, "loss": 3.3432, "step": 151 }, { "epoch": 0.7534076827757125, "grad_norm": 1.179330587387085, "learning_rate": 0.000167816091954023, "loss": 2.5291, "step": 152 }, { "epoch": 0.758364312267658, "grad_norm": 1.308836579322815, "learning_rate": 0.00016756066411238827, "loss": 2.7799, "step": 153 }, { "epoch": 0.7633209417596035, "grad_norm": 1.3677353858947754, "learning_rate": 0.00016730523627075353, "loss": 2.8872, "step": 154 }, { "epoch": 0.7682775712515489, "grad_norm": 1.2296689748764038, "learning_rate": 0.0001670498084291188, "loss": 3.137, "step": 155 }, { "epoch": 0.7732342007434945, "grad_norm": 1.2084614038467407, "learning_rate": 0.00016679438058748406, "loss": 2.8232, "step": 156 }, { "epoch": 0.7781908302354399, "grad_norm": 1.1933956146240234, "learning_rate": 0.0001665389527458493, "loss": 3.0647, "step": 157 }, { "epoch": 0.7831474597273854, "grad_norm": 1.175970435142517, "learning_rate": 0.00016628352490421458, "loss": 2.3696, "step": 158 }, { "epoch": 0.7881040892193308, "grad_norm": 1.1459221839904785, "learning_rate": 0.00016602809706257982, "loss": 2.5382, "step": 159 }, { "epoch": 0.7930607187112764, "grad_norm": 1.195381760597229, "learning_rate": 0.0001657726692209451, "loss": 2.457, "step": 160 }, { "epoch": 0.7980173482032218, "grad_norm": 1.337165117263794, "learning_rate": 0.00016551724137931035, "loss": 2.5121, "step": 161 }, { "epoch": 0.8029739776951673, "grad_norm": 1.3737194538116455, "learning_rate": 0.00016526181353767564, "loss": 2.9961, "step": 162 }, { "epoch": 0.8079306071871127, "grad_norm": 1.63533353805542, "learning_rate": 0.00016500638569604087, "loss": 3.2549, "step": 163 }, { "epoch": 0.8128872366790583, "grad_norm": 1.2384507656097412, "learning_rate": 0.00016475095785440614, "loss": 2.8821, "step": 164 }, { "epoch": 0.8178438661710037, "grad_norm": 1.2741832733154297, "learning_rate": 0.0001644955300127714, "loss": 2.8996, "step": 165 }, { "epoch": 0.8228004956629492, "grad_norm": 1.559248685836792, "learning_rate": 0.00016424010217113666, "loss": 3.1911, "step": 166 }, { "epoch": 0.8277571251548946, "grad_norm": 1.3192209005355835, "learning_rate": 0.00016398467432950193, "loss": 2.6993, "step": 167 }, { "epoch": 0.8327137546468402, "grad_norm": 1.1915833950042725, "learning_rate": 0.0001637292464878672, "loss": 2.6233, "step": 168 }, { "epoch": 0.8376703841387856, "grad_norm": 1.2560778856277466, "learning_rate": 0.00016347381864623245, "loss": 2.6467, "step": 169 }, { "epoch": 0.8426270136307311, "grad_norm": 1.3174892663955688, "learning_rate": 0.00016321839080459772, "loss": 2.4764, "step": 170 }, { "epoch": 0.8475836431226765, "grad_norm": 1.3117812871932983, "learning_rate": 0.00016296296296296295, "loss": 2.5733, "step": 171 }, { "epoch": 0.8525402726146221, "grad_norm": 1.240525722503662, "learning_rate": 0.00016270753512132824, "loss": 2.6198, "step": 172 }, { "epoch": 0.8574969021065675, "grad_norm": 1.66138756275177, "learning_rate": 0.00016245210727969348, "loss": 3.3568, "step": 173 }, { "epoch": 0.862453531598513, "grad_norm": 1.5758461952209473, "learning_rate": 0.00016219667943805877, "loss": 2.7601, "step": 174 }, { "epoch": 0.8674101610904585, "grad_norm": 1.1617976427078247, "learning_rate": 0.000161941251596424, "loss": 2.5865, "step": 175 }, { "epoch": 0.872366790582404, "grad_norm": 1.435559868812561, "learning_rate": 0.0001616858237547893, "loss": 2.8164, "step": 176 }, { "epoch": 0.8773234200743495, "grad_norm": 1.208302617073059, "learning_rate": 0.00016143039591315453, "loss": 2.5785, "step": 177 }, { "epoch": 0.8822800495662949, "grad_norm": 1.4940738677978516, "learning_rate": 0.0001611749680715198, "loss": 2.7235, "step": 178 }, { "epoch": 0.8872366790582404, "grad_norm": 1.135846495628357, "learning_rate": 0.00016091954022988506, "loss": 2.3321, "step": 179 }, { "epoch": 0.8921933085501859, "grad_norm": 1.2706923484802246, "learning_rate": 0.00016066411238825032, "loss": 2.5422, "step": 180 }, { "epoch": 0.8971499380421314, "grad_norm": 1.3885927200317383, "learning_rate": 0.00016040868454661559, "loss": 2.8623, "step": 181 }, { "epoch": 0.9021065675340768, "grad_norm": 1.4354758262634277, "learning_rate": 0.00016015325670498085, "loss": 2.3397, "step": 182 }, { "epoch": 0.9070631970260223, "grad_norm": 1.3210114240646362, "learning_rate": 0.0001598978288633461, "loss": 2.5658, "step": 183 }, { "epoch": 0.9120198265179678, "grad_norm": 1.5271670818328857, "learning_rate": 0.00015964240102171138, "loss": 2.7153, "step": 184 }, { "epoch": 0.9169764560099133, "grad_norm": 1.3032946586608887, "learning_rate": 0.00015938697318007664, "loss": 2.4854, "step": 185 }, { "epoch": 0.9219330855018587, "grad_norm": 1.3772252798080444, "learning_rate": 0.0001591315453384419, "loss": 2.5493, "step": 186 }, { "epoch": 0.9268897149938042, "grad_norm": 1.344874382019043, "learning_rate": 0.00015887611749680717, "loss": 2.6904, "step": 187 }, { "epoch": 0.9318463444857497, "grad_norm": 1.4352842569351196, "learning_rate": 0.00015862068965517243, "loss": 2.7214, "step": 188 }, { "epoch": 0.9368029739776952, "grad_norm": 1.17839515209198, "learning_rate": 0.0001583652618135377, "loss": 2.1581, "step": 189 }, { "epoch": 0.9417596034696406, "grad_norm": 1.3339844942092896, "learning_rate": 0.00015810983397190296, "loss": 2.4722, "step": 190 }, { "epoch": 0.9467162329615861, "grad_norm": 1.3919591903686523, "learning_rate": 0.00015785440613026822, "loss": 2.4386, "step": 191 }, { "epoch": 0.9516728624535316, "grad_norm": 1.4500068426132202, "learning_rate": 0.00015759897828863346, "loss": 2.5789, "step": 192 }, { "epoch": 0.9566294919454771, "grad_norm": 1.3232927322387695, "learning_rate": 0.00015734355044699875, "loss": 2.5556, "step": 193 }, { "epoch": 0.9615861214374225, "grad_norm": 1.56577467918396, "learning_rate": 0.00015708812260536398, "loss": 2.8504, "step": 194 }, { "epoch": 0.966542750929368, "grad_norm": 1.3413156270980835, "learning_rate": 0.00015683269476372927, "loss": 2.2172, "step": 195 }, { "epoch": 0.9714993804213135, "grad_norm": 1.5769926309585571, "learning_rate": 0.0001565772669220945, "loss": 3.1141, "step": 196 }, { "epoch": 0.976456009913259, "grad_norm": 1.3165156841278076, "learning_rate": 0.0001563218390804598, "loss": 2.6951, "step": 197 }, { "epoch": 0.9814126394052045, "grad_norm": 1.384596347808838, "learning_rate": 0.00015606641123882504, "loss": 2.8985, "step": 198 }, { "epoch": 0.9863692688971499, "grad_norm": 1.2764040231704712, "learning_rate": 0.0001558109833971903, "loss": 2.2731, "step": 199 }, { "epoch": 0.9913258983890955, "grad_norm": 1.3206219673156738, "learning_rate": 0.00015555555555555556, "loss": 2.4783, "step": 200 }, { "epoch": 0.9962825278810409, "grad_norm": 1.1644648313522339, "learning_rate": 0.00015530012771392083, "loss": 2.3293, "step": 201 }, { "epoch": 1.0, "grad_norm": 1.5087392330169678, "learning_rate": 0.0001550446998722861, "loss": 1.9488, "step": 202 }, { "epoch": 1.0, "eval_loss": 2.729416608810425, "eval_runtime": 22.6846, "eval_samples_per_second": 17.809, "eval_steps_per_second": 2.248, "step": 202 }, { "epoch": 1.0049566294919454, "grad_norm": 1.1808276176452637, "learning_rate": 0.00015478927203065135, "loss": 2.45, "step": 203 }, { "epoch": 1.009913258983891, "grad_norm": 1.3719933032989502, "learning_rate": 0.00015453384418901662, "loss": 2.4187, "step": 204 }, { "epoch": 1.0148698884758365, "grad_norm": 1.4881116151809692, "learning_rate": 0.00015427841634738188, "loss": 2.7559, "step": 205 }, { "epoch": 1.019826517967782, "grad_norm": 1.317412257194519, "learning_rate": 0.00015402298850574712, "loss": 2.4683, "step": 206 }, { "epoch": 1.0247831474597273, "grad_norm": 1.2034761905670166, "learning_rate": 0.0001537675606641124, "loss": 2.3886, "step": 207 }, { "epoch": 1.029739776951673, "grad_norm": 1.5965017080307007, "learning_rate": 0.00015351213282247764, "loss": 2.5819, "step": 208 }, { "epoch": 1.0346964064436184, "grad_norm": 1.6514837741851807, "learning_rate": 0.00015325670498084293, "loss": 2.7572, "step": 209 }, { "epoch": 1.0396530359355638, "grad_norm": 1.4627822637557983, "learning_rate": 0.00015300127713920817, "loss": 2.8254, "step": 210 }, { "epoch": 1.0446096654275092, "grad_norm": 1.5069350004196167, "learning_rate": 0.00015274584929757346, "loss": 2.6277, "step": 211 }, { "epoch": 1.0495662949194549, "grad_norm": 1.3964656591415405, "learning_rate": 0.0001524904214559387, "loss": 2.458, "step": 212 }, { "epoch": 1.0545229244114003, "grad_norm": 1.5406875610351562, "learning_rate": 0.00015223499361430396, "loss": 2.5876, "step": 213 }, { "epoch": 1.0594795539033457, "grad_norm": 1.3873964548110962, "learning_rate": 0.00015197956577266922, "loss": 2.1244, "step": 214 }, { "epoch": 1.0644361833952911, "grad_norm": 1.5236468315124512, "learning_rate": 0.00015172413793103449, "loss": 2.4087, "step": 215 }, { "epoch": 1.0693928128872368, "grad_norm": 1.5236297845840454, "learning_rate": 0.00015146871008939975, "loss": 2.5084, "step": 216 }, { "epoch": 1.0743494423791822, "grad_norm": 1.3550326824188232, "learning_rate": 0.000151213282247765, "loss": 2.7363, "step": 217 }, { "epoch": 1.0793060718711276, "grad_norm": 1.3009722232818604, "learning_rate": 0.00015095785440613028, "loss": 2.4046, "step": 218 }, { "epoch": 1.084262701363073, "grad_norm": 1.3609213829040527, "learning_rate": 0.00015070242656449554, "loss": 2.5477, "step": 219 }, { "epoch": 1.0892193308550187, "grad_norm": 1.5530016422271729, "learning_rate": 0.00015044699872286078, "loss": 2.677, "step": 220 }, { "epoch": 1.094175960346964, "grad_norm": 1.566308856010437, "learning_rate": 0.00015019157088122607, "loss": 2.6243, "step": 221 }, { "epoch": 1.0991325898389095, "grad_norm": 1.6015573740005493, "learning_rate": 0.0001499361430395913, "loss": 2.4592, "step": 222 }, { "epoch": 1.104089219330855, "grad_norm": 2.0257885456085205, "learning_rate": 0.0001496807151979566, "loss": 3.2142, "step": 223 }, { "epoch": 1.1090458488228006, "grad_norm": 1.599144458770752, "learning_rate": 0.00014942528735632183, "loss": 2.5635, "step": 224 }, { "epoch": 1.114002478314746, "grad_norm": 1.4779815673828125, "learning_rate": 0.00014916985951468712, "loss": 2.3515, "step": 225 }, { "epoch": 1.1189591078066914, "grad_norm": 1.6670912504196167, "learning_rate": 0.00014891443167305236, "loss": 2.5193, "step": 226 }, { "epoch": 1.123915737298637, "grad_norm": 1.4645159244537354, "learning_rate": 0.00014865900383141765, "loss": 2.4852, "step": 227 }, { "epoch": 1.1288723667905824, "grad_norm": 1.39253568649292, "learning_rate": 0.00014840357598978288, "loss": 2.7357, "step": 228 }, { "epoch": 1.1338289962825279, "grad_norm": 1.5711032152175903, "learning_rate": 0.00014814814814814815, "loss": 2.6649, "step": 229 }, { "epoch": 1.1387856257744733, "grad_norm": 1.3611880540847778, "learning_rate": 0.0001478927203065134, "loss": 2.3486, "step": 230 }, { "epoch": 1.143742255266419, "grad_norm": 1.439810872077942, "learning_rate": 0.00014763729246487867, "loss": 2.1991, "step": 231 }, { "epoch": 1.1486988847583643, "grad_norm": 1.7450776100158691, "learning_rate": 0.00014738186462324394, "loss": 2.7248, "step": 232 }, { "epoch": 1.1536555142503098, "grad_norm": 1.5478737354278564, "learning_rate": 0.0001471264367816092, "loss": 2.3492, "step": 233 }, { "epoch": 1.1586121437422552, "grad_norm": 1.4742021560668945, "learning_rate": 0.00014687100893997446, "loss": 2.3586, "step": 234 }, { "epoch": 1.1635687732342008, "grad_norm": 1.6907175779342651, "learning_rate": 0.00014661558109833973, "loss": 2.8715, "step": 235 }, { "epoch": 1.1685254027261462, "grad_norm": 1.5488353967666626, "learning_rate": 0.000146360153256705, "loss": 2.7357, "step": 236 }, { "epoch": 1.1734820322180917, "grad_norm": 1.6616300344467163, "learning_rate": 0.00014610472541507025, "loss": 2.3508, "step": 237 }, { "epoch": 1.178438661710037, "grad_norm": 1.592176914215088, "learning_rate": 0.00014584929757343552, "loss": 2.432, "step": 238 }, { "epoch": 1.1833952912019827, "grad_norm": 1.5771480798721313, "learning_rate": 0.00014559386973180078, "loss": 2.8382, "step": 239 }, { "epoch": 1.1883519206939281, "grad_norm": 1.664546251296997, "learning_rate": 0.00014533844189016604, "loss": 2.5342, "step": 240 }, { "epoch": 1.1933085501858736, "grad_norm": 1.9009575843811035, "learning_rate": 0.0001450830140485313, "loss": 2.5563, "step": 241 }, { "epoch": 1.198265179677819, "grad_norm": 1.6944457292556763, "learning_rate": 0.00014482758620689657, "loss": 2.7315, "step": 242 }, { "epoch": 1.2032218091697646, "grad_norm": 1.4040586948394775, "learning_rate": 0.0001445721583652618, "loss": 2.2772, "step": 243 }, { "epoch": 1.20817843866171, "grad_norm": 1.549829363822937, "learning_rate": 0.0001443167305236271, "loss": 2.593, "step": 244 }, { "epoch": 1.2131350681536555, "grad_norm": 1.2948358058929443, "learning_rate": 0.00014406130268199233, "loss": 2.3032, "step": 245 }, { "epoch": 1.218091697645601, "grad_norm": 1.5653175115585327, "learning_rate": 0.00014380587484035762, "loss": 2.2908, "step": 246 }, { "epoch": 1.2230483271375465, "grad_norm": 1.5301649570465088, "learning_rate": 0.00014355044699872286, "loss": 2.6958, "step": 247 }, { "epoch": 1.228004956629492, "grad_norm": 1.6121726036071777, "learning_rate": 0.00014329501915708815, "loss": 2.5315, "step": 248 }, { "epoch": 1.2329615861214374, "grad_norm": 1.6542530059814453, "learning_rate": 0.00014303959131545339, "loss": 2.7279, "step": 249 }, { "epoch": 1.2379182156133828, "grad_norm": 1.5968433618545532, "learning_rate": 0.00014278416347381865, "loss": 2.1316, "step": 250 }, { "epoch": 1.2428748451053284, "grad_norm": 1.753065824508667, "learning_rate": 0.0001425287356321839, "loss": 2.7854, "step": 251 }, { "epoch": 1.2478314745972738, "grad_norm": 1.5655628442764282, "learning_rate": 0.00014227330779054918, "loss": 2.2109, "step": 252 }, { "epoch": 1.2527881040892193, "grad_norm": 1.6386921405792236, "learning_rate": 0.00014201787994891444, "loss": 2.6289, "step": 253 }, { "epoch": 1.257744733581165, "grad_norm": 1.389952301979065, "learning_rate": 0.0001417624521072797, "loss": 2.2817, "step": 254 }, { "epoch": 1.2627013630731103, "grad_norm": 1.5647423267364502, "learning_rate": 0.00014150702426564497, "loss": 2.3569, "step": 255 }, { "epoch": 1.2676579925650557, "grad_norm": 1.5217468738555908, "learning_rate": 0.00014125159642401023, "loss": 2.4843, "step": 256 }, { "epoch": 1.2726146220570014, "grad_norm": 1.5946449041366577, "learning_rate": 0.00014099616858237547, "loss": 2.482, "step": 257 }, { "epoch": 1.2775712515489468, "grad_norm": 2.6357760429382324, "learning_rate": 0.00014074074074074076, "loss": 3.0227, "step": 258 }, { "epoch": 1.2825278810408922, "grad_norm": 1.6973539590835571, "learning_rate": 0.000140485312899106, "loss": 2.3939, "step": 259 }, { "epoch": 1.2874845105328376, "grad_norm": 1.8628802299499512, "learning_rate": 0.00014022988505747128, "loss": 2.5089, "step": 260 }, { "epoch": 1.292441140024783, "grad_norm": 1.8481550216674805, "learning_rate": 0.00013997445721583652, "loss": 2.3706, "step": 261 }, { "epoch": 1.2973977695167287, "grad_norm": 1.4677447080612183, "learning_rate": 0.0001397190293742018, "loss": 2.5831, "step": 262 }, { "epoch": 1.3023543990086741, "grad_norm": 1.3649961948394775, "learning_rate": 0.00013946360153256705, "loss": 2.2881, "step": 263 }, { "epoch": 1.3073110285006195, "grad_norm": 1.6258682012557983, "learning_rate": 0.0001392081736909323, "loss": 2.326, "step": 264 }, { "epoch": 1.3122676579925652, "grad_norm": 1.869107961654663, "learning_rate": 0.00013895274584929757, "loss": 2.876, "step": 265 }, { "epoch": 1.3172242874845106, "grad_norm": 1.6073962450027466, "learning_rate": 0.00013869731800766284, "loss": 2.3673, "step": 266 }, { "epoch": 1.322180916976456, "grad_norm": 1.6778944730758667, "learning_rate": 0.0001384418901660281, "loss": 2.5112, "step": 267 }, { "epoch": 1.3271375464684014, "grad_norm": 1.7758762836456299, "learning_rate": 0.00013818646232439336, "loss": 2.5872, "step": 268 }, { "epoch": 1.3320941759603468, "grad_norm": 1.609081745147705, "learning_rate": 0.00013793103448275863, "loss": 2.4718, "step": 269 }, { "epoch": 1.3370508054522925, "grad_norm": 1.602964162826538, "learning_rate": 0.0001376756066411239, "loss": 2.2268, "step": 270 }, { "epoch": 1.342007434944238, "grad_norm": 1.532798171043396, "learning_rate": 0.00013742017879948915, "loss": 2.4378, "step": 271 }, { "epoch": 1.3469640644361833, "grad_norm": 1.4932005405426025, "learning_rate": 0.00013716475095785442, "loss": 2.6889, "step": 272 }, { "epoch": 1.351920693928129, "grad_norm": 1.6818208694458008, "learning_rate": 0.00013690932311621968, "loss": 2.6538, "step": 273 }, { "epoch": 1.3568773234200744, "grad_norm": 1.4763221740722656, "learning_rate": 0.00013665389527458494, "loss": 2.1908, "step": 274 }, { "epoch": 1.3618339529120198, "grad_norm": 1.4615814685821533, "learning_rate": 0.0001363984674329502, "loss": 2.2983, "step": 275 }, { "epoch": 1.3667905824039652, "grad_norm": 1.5107206106185913, "learning_rate": 0.00013614303959131547, "loss": 2.2305, "step": 276 }, { "epoch": 1.3717472118959106, "grad_norm": 1.4505106210708618, "learning_rate": 0.00013588761174968073, "loss": 2.1633, "step": 277 }, { "epoch": 1.3767038413878563, "grad_norm": 1.7257494926452637, "learning_rate": 0.00013563218390804597, "loss": 2.3903, "step": 278 }, { "epoch": 1.3816604708798017, "grad_norm": 1.8901737928390503, "learning_rate": 0.00013537675606641126, "loss": 2.5404, "step": 279 }, { "epoch": 1.3866171003717471, "grad_norm": 1.5865846872329712, "learning_rate": 0.0001351213282247765, "loss": 2.5774, "step": 280 }, { "epoch": 1.3915737298636928, "grad_norm": 1.631974458694458, "learning_rate": 0.0001348659003831418, "loss": 2.2341, "step": 281 }, { "epoch": 1.3965303593556382, "grad_norm": 1.7717571258544922, "learning_rate": 0.00013461047254150702, "loss": 2.1569, "step": 282 }, { "epoch": 1.4014869888475836, "grad_norm": 1.4758812189102173, "learning_rate": 0.0001343550446998723, "loss": 2.0374, "step": 283 }, { "epoch": 1.4064436183395292, "grad_norm": 1.8754217624664307, "learning_rate": 0.00013409961685823755, "loss": 2.4035, "step": 284 }, { "epoch": 1.4114002478314747, "grad_norm": 1.8739930391311646, "learning_rate": 0.0001338441890166028, "loss": 2.5747, "step": 285 }, { "epoch": 1.41635687732342, "grad_norm": 1.6370917558670044, "learning_rate": 0.00013358876117496808, "loss": 2.4051, "step": 286 }, { "epoch": 1.4213135068153655, "grad_norm": 1.6480002403259277, "learning_rate": 0.00013333333333333334, "loss": 2.4366, "step": 287 }, { "epoch": 1.426270136307311, "grad_norm": 1.7106711864471436, "learning_rate": 0.0001330779054916986, "loss": 2.4036, "step": 288 }, { "epoch": 1.4312267657992566, "grad_norm": 1.5775083303451538, "learning_rate": 0.00013282247765006387, "loss": 2.2505, "step": 289 }, { "epoch": 1.436183395291202, "grad_norm": 1.9434142112731934, "learning_rate": 0.00013256704980842913, "loss": 2.8395, "step": 290 }, { "epoch": 1.4411400247831474, "grad_norm": 1.7857332229614258, "learning_rate": 0.0001323116219667944, "loss": 2.925, "step": 291 }, { "epoch": 1.446096654275093, "grad_norm": 1.8186098337173462, "learning_rate": 0.00013205619412515963, "loss": 2.5975, "step": 292 }, { "epoch": 1.4510532837670385, "grad_norm": 1.682557225227356, "learning_rate": 0.00013180076628352492, "loss": 2.3786, "step": 293 }, { "epoch": 1.4560099132589839, "grad_norm": 1.7102502584457397, "learning_rate": 0.00013154533844189016, "loss": 2.5359, "step": 294 }, { "epoch": 1.4609665427509293, "grad_norm": 1.7254425287246704, "learning_rate": 0.00013128991060025545, "loss": 2.4446, "step": 295 }, { "epoch": 1.4659231722428747, "grad_norm": 1.6428650617599487, "learning_rate": 0.00013103448275862068, "loss": 2.7514, "step": 296 }, { "epoch": 1.4708798017348204, "grad_norm": 1.7002222537994385, "learning_rate": 0.00013077905491698597, "loss": 2.3079, "step": 297 }, { "epoch": 1.4758364312267658, "grad_norm": 2.013287305831909, "learning_rate": 0.0001305236270753512, "loss": 2.5104, "step": 298 }, { "epoch": 1.4807930607187112, "grad_norm": 1.5022965669631958, "learning_rate": 0.00013026819923371647, "loss": 2.1304, "step": 299 }, { "epoch": 1.4857496902106568, "grad_norm": 1.9274829626083374, "learning_rate": 0.00013001277139208174, "loss": 2.6626, "step": 300 }, { "epoch": 1.4907063197026023, "grad_norm": 1.9029258489608765, "learning_rate": 0.000129757343550447, "loss": 2.6778, "step": 301 }, { "epoch": 1.4956629491945477, "grad_norm": 1.4256715774536133, "learning_rate": 0.00012950191570881226, "loss": 2.4563, "step": 302 }, { "epoch": 1.5006195786864933, "grad_norm": 1.716525673866272, "learning_rate": 0.00012924648786717753, "loss": 2.5675, "step": 303 }, { "epoch": 1.5055762081784385, "grad_norm": 1.6712944507598877, "learning_rate": 0.0001289910600255428, "loss": 2.2456, "step": 304 }, { "epoch": 1.5105328376703842, "grad_norm": 1.4323303699493408, "learning_rate": 0.00012873563218390805, "loss": 2.3167, "step": 305 }, { "epoch": 1.5154894671623296, "grad_norm": 1.562535285949707, "learning_rate": 0.00012848020434227332, "loss": 2.2113, "step": 306 }, { "epoch": 1.520446096654275, "grad_norm": 1.8023464679718018, "learning_rate": 0.00012822477650063858, "loss": 2.5698, "step": 307 }, { "epoch": 1.5254027261462206, "grad_norm": 1.560618281364441, "learning_rate": 0.00012796934865900382, "loss": 2.7733, "step": 308 }, { "epoch": 1.530359355638166, "grad_norm": 1.6257753372192383, "learning_rate": 0.0001277139208173691, "loss": 2.5203, "step": 309 }, { "epoch": 1.5353159851301115, "grad_norm": 1.9268074035644531, "learning_rate": 0.00012745849297573434, "loss": 2.4339, "step": 310 }, { "epoch": 1.5402726146220571, "grad_norm": 1.728567123413086, "learning_rate": 0.00012720306513409963, "loss": 2.3867, "step": 311 }, { "epoch": 1.5452292441140025, "grad_norm": 2.204632043838501, "learning_rate": 0.00012694763729246487, "loss": 2.8223, "step": 312 }, { "epoch": 1.550185873605948, "grad_norm": 1.53752863407135, "learning_rate": 0.00012669220945083016, "loss": 2.4653, "step": 313 }, { "epoch": 1.5551425030978936, "grad_norm": 1.5143874883651733, "learning_rate": 0.0001264367816091954, "loss": 2.1158, "step": 314 }, { "epoch": 1.5600991325898388, "grad_norm": 1.635250449180603, "learning_rate": 0.00012618135376756066, "loss": 2.1365, "step": 315 }, { "epoch": 1.5650557620817844, "grad_norm": 1.4665566682815552, "learning_rate": 0.00012592592592592592, "loss": 2.3736, "step": 316 }, { "epoch": 1.5700123915737298, "grad_norm": 1.8319354057312012, "learning_rate": 0.00012567049808429119, "loss": 2.5109, "step": 317 }, { "epoch": 1.5749690210656753, "grad_norm": 1.8809919357299805, "learning_rate": 0.00012541507024265645, "loss": 2.5676, "step": 318 }, { "epoch": 1.579925650557621, "grad_norm": 1.693365216255188, "learning_rate": 0.0001251596424010217, "loss": 2.362, "step": 319 }, { "epoch": 1.5848822800495663, "grad_norm": 1.568253517150879, "learning_rate": 0.00012490421455938698, "loss": 2.221, "step": 320 }, { "epoch": 1.5898389095415117, "grad_norm": 1.5762134790420532, "learning_rate": 0.00012464878671775224, "loss": 2.5011, "step": 321 }, { "epoch": 1.5947955390334574, "grad_norm": 1.5680880546569824, "learning_rate": 0.0001243933588761175, "loss": 2.1637, "step": 322 }, { "epoch": 1.5997521685254026, "grad_norm": 1.5295878648757935, "learning_rate": 0.00012413793103448277, "loss": 2.1718, "step": 323 }, { "epoch": 1.6047087980173482, "grad_norm": 1.3603239059448242, "learning_rate": 0.00012388250319284803, "loss": 2.1767, "step": 324 }, { "epoch": 1.6096654275092936, "grad_norm": 1.6356030702590942, "learning_rate": 0.0001236270753512133, "loss": 2.8308, "step": 325 }, { "epoch": 1.614622057001239, "grad_norm": 1.5502631664276123, "learning_rate": 0.00012337164750957856, "loss": 2.0726, "step": 326 }, { "epoch": 1.6195786864931847, "grad_norm": 1.564531922340393, "learning_rate": 0.00012311621966794382, "loss": 2.283, "step": 327 }, { "epoch": 1.6245353159851301, "grad_norm": 1.775214433670044, "learning_rate": 0.00012286079182630908, "loss": 2.3739, "step": 328 }, { "epoch": 1.6294919454770755, "grad_norm": 1.7285168170928955, "learning_rate": 0.00012260536398467432, "loss": 2.6, "step": 329 }, { "epoch": 1.6344485749690212, "grad_norm": 1.9098368883132935, "learning_rate": 0.0001223499361430396, "loss": 2.6702, "step": 330 }, { "epoch": 1.6394052044609664, "grad_norm": 1.8068279027938843, "learning_rate": 0.00012209450830140485, "loss": 2.8159, "step": 331 }, { "epoch": 1.644361833952912, "grad_norm": 1.422575831413269, "learning_rate": 0.00012183908045977012, "loss": 1.9348, "step": 332 }, { "epoch": 1.6493184634448577, "grad_norm": 1.873787522315979, "learning_rate": 0.00012158365261813537, "loss": 2.6505, "step": 333 }, { "epoch": 1.6542750929368029, "grad_norm": 1.653365135192871, "learning_rate": 0.00012132822477650065, "loss": 2.4581, "step": 334 }, { "epoch": 1.6592317224287485, "grad_norm": 1.5894263982772827, "learning_rate": 0.0001210727969348659, "loss": 2.4804, "step": 335 }, { "epoch": 1.664188351920694, "grad_norm": 1.6645914316177368, "learning_rate": 0.00012081736909323116, "loss": 2.6777, "step": 336 }, { "epoch": 1.6691449814126393, "grad_norm": 1.3606253862380981, "learning_rate": 0.00012056194125159643, "loss": 2.1287, "step": 337 }, { "epoch": 1.674101610904585, "grad_norm": 1.9518952369689941, "learning_rate": 0.00012030651340996169, "loss": 2.4308, "step": 338 }, { "epoch": 1.6790582403965304, "grad_norm": 1.837501883506775, "learning_rate": 0.00012005108556832695, "loss": 2.5096, "step": 339 }, { "epoch": 1.6840148698884758, "grad_norm": 1.681298851966858, "learning_rate": 0.00011979565772669222, "loss": 1.9996, "step": 340 }, { "epoch": 1.6889714993804215, "grad_norm": 1.6063026189804077, "learning_rate": 0.00011954022988505748, "loss": 1.8223, "step": 341 }, { "epoch": 1.6939281288723667, "grad_norm": 1.7527902126312256, "learning_rate": 0.00011928480204342274, "loss": 2.3558, "step": 342 }, { "epoch": 1.6988847583643123, "grad_norm": 2.1108145713806152, "learning_rate": 0.00011902937420178799, "loss": 2.4522, "step": 343 }, { "epoch": 1.7038413878562577, "grad_norm": 1.9739495515823364, "learning_rate": 0.00011877394636015327, "loss": 2.5235, "step": 344 }, { "epoch": 1.7087980173482031, "grad_norm": NaN, "learning_rate": 0.00011851851851851852, "loss": 2.8147, "step": 345 }, { "epoch": 1.7137546468401488, "grad_norm": 1.6384708881378174, "learning_rate": 0.00011851851851851852, "loss": 2.2734, "step": 346 }, { "epoch": 1.7187112763320942, "grad_norm": 1.6661678552627563, "learning_rate": 0.0001182630906768838, "loss": 2.4438, "step": 347 }, { "epoch": 1.7236679058240396, "grad_norm": 1.987106442451477, "learning_rate": 0.00011800766283524905, "loss": 2.4275, "step": 348 }, { "epoch": 1.7286245353159853, "grad_norm": 1.9400545358657837, "learning_rate": 0.00011775223499361432, "loss": 2.446, "step": 349 }, { "epoch": 1.7335811648079305, "grad_norm": 1.7634007930755615, "learning_rate": 0.00011749680715197957, "loss": 2.2731, "step": 350 }, { "epoch": 1.738537794299876, "grad_norm": 1.5254379510879517, "learning_rate": 0.00011724137931034482, "loss": 1.8172, "step": 351 }, { "epoch": 1.7434944237918215, "grad_norm": 1.9841443300247192, "learning_rate": 0.0001169859514687101, "loss": 2.4537, "step": 352 }, { "epoch": 1.748451053283767, "grad_norm": 2.1015896797180176, "learning_rate": 0.00011673052362707535, "loss": 2.1979, "step": 353 }, { "epoch": 1.7534076827757126, "grad_norm": 1.8992552757263184, "learning_rate": 0.00011647509578544063, "loss": 2.145, "step": 354 }, { "epoch": 1.758364312267658, "grad_norm": 2.1565067768096924, "learning_rate": 0.00011621966794380588, "loss": 2.4881, "step": 355 }, { "epoch": 1.7633209417596034, "grad_norm": 1.8605690002441406, "learning_rate": 0.00011596424010217115, "loss": 2.5486, "step": 356 }, { "epoch": 1.768277571251549, "grad_norm": 1.8879384994506836, "learning_rate": 0.0001157088122605364, "loss": 2.448, "step": 357 }, { "epoch": 1.7732342007434945, "grad_norm": 2.0151591300964355, "learning_rate": 0.00011545338441890165, "loss": 2.4967, "step": 358 }, { "epoch": 1.77819083023544, "grad_norm": 1.6546062231063843, "learning_rate": 0.00011519795657726693, "loss": 2.5078, "step": 359 }, { "epoch": 1.7831474597273855, "grad_norm": 2.1501290798187256, "learning_rate": 0.00011494252873563218, "loss": 2.3383, "step": 360 }, { "epoch": 1.7881040892193307, "grad_norm": 1.8948010206222534, "learning_rate": 0.00011468710089399746, "loss": 2.3743, "step": 361 }, { "epoch": 1.7930607187112764, "grad_norm": 1.9624435901641846, "learning_rate": 0.0001144316730523627, "loss": 2.3619, "step": 362 }, { "epoch": 1.7980173482032218, "grad_norm": 2.0342910289764404, "learning_rate": 0.00011417624521072798, "loss": 2.4056, "step": 363 }, { "epoch": 1.8029739776951672, "grad_norm": 1.964487910270691, "learning_rate": 0.00011392081736909323, "loss": 2.2912, "step": 364 }, { "epoch": 1.8079306071871128, "grad_norm": 1.6878917217254639, "learning_rate": 0.00011366538952745848, "loss": 2.3647, "step": 365 }, { "epoch": 1.8128872366790583, "grad_norm": 1.959184169769287, "learning_rate": 0.00011340996168582376, "loss": 2.0794, "step": 366 }, { "epoch": 1.8178438661710037, "grad_norm": 1.7067630290985107, "learning_rate": 0.00011315453384418901, "loss": 2.5082, "step": 367 }, { "epoch": 1.8228004956629493, "grad_norm": 1.7823041677474976, "learning_rate": 0.00011289910600255429, "loss": 2.0763, "step": 368 }, { "epoch": 1.8277571251548945, "grad_norm": 1.9278924465179443, "learning_rate": 0.00011264367816091954, "loss": 2.5235, "step": 369 }, { "epoch": 1.8327137546468402, "grad_norm": 1.8295596837997437, "learning_rate": 0.00011238825031928481, "loss": 2.836, "step": 370 }, { "epoch": 1.8376703841387856, "grad_norm": 1.6599929332733154, "learning_rate": 0.00011213282247765006, "loss": 2.3678, "step": 371 }, { "epoch": 1.842627013630731, "grad_norm": 2.075873851776123, "learning_rate": 0.00011187739463601533, "loss": 2.6875, "step": 372 }, { "epoch": 1.8475836431226766, "grad_norm": 1.8870905637741089, "learning_rate": 0.00011162196679438059, "loss": 2.4988, "step": 373 }, { "epoch": 1.852540272614622, "grad_norm": 1.7830430269241333, "learning_rate": 0.00011136653895274585, "loss": 2.0613, "step": 374 }, { "epoch": 1.8574969021065675, "grad_norm": 1.8882161378860474, "learning_rate": 0.00011111111111111112, "loss": 2.627, "step": 375 }, { "epoch": 1.8624535315985131, "grad_norm": 1.9912681579589844, "learning_rate": 0.00011085568326947638, "loss": 2.7047, "step": 376 }, { "epoch": 1.8674101610904585, "grad_norm": 2.0505921840667725, "learning_rate": 0.00011060025542784164, "loss": 2.95, "step": 377 }, { "epoch": 1.872366790582404, "grad_norm": 1.9979755878448486, "learning_rate": 0.0001103448275862069, "loss": 2.5598, "step": 378 }, { "epoch": 1.8773234200743496, "grad_norm": 1.5586763620376587, "learning_rate": 0.00011008939974457216, "loss": 2.3973, "step": 379 }, { "epoch": 1.8822800495662948, "grad_norm": 1.698810338973999, "learning_rate": 0.00010983397190293743, "loss": 2.1451, "step": 380 }, { "epoch": 1.8872366790582404, "grad_norm": 2.1801044940948486, "learning_rate": 0.00010957854406130268, "loss": 2.5816, "step": 381 }, { "epoch": 1.8921933085501859, "grad_norm": 1.511141300201416, "learning_rate": 0.00010932311621966796, "loss": 2.1513, "step": 382 }, { "epoch": 1.8971499380421313, "grad_norm": 1.9838210344314575, "learning_rate": 0.00010906768837803321, "loss": 1.8924, "step": 383 }, { "epoch": 1.902106567534077, "grad_norm": 1.9898924827575684, "learning_rate": 0.00010881226053639849, "loss": 2.4117, "step": 384 }, { "epoch": 1.9070631970260223, "grad_norm": 2.090376615524292, "learning_rate": 0.00010855683269476374, "loss": 2.4818, "step": 385 }, { "epoch": 1.9120198265179678, "grad_norm": 1.8976398706436157, "learning_rate": 0.00010830140485312901, "loss": 2.5878, "step": 386 }, { "epoch": 1.9169764560099134, "grad_norm": 1.813851237297058, "learning_rate": 0.00010804597701149426, "loss": 2.1124, "step": 387 }, { "epoch": 1.9219330855018586, "grad_norm": 1.9612343311309814, "learning_rate": 0.00010779054916985951, "loss": 2.5307, "step": 388 }, { "epoch": 1.9268897149938042, "grad_norm": 2.0917444229125977, "learning_rate": 0.00010753512132822479, "loss": 2.4722, "step": 389 }, { "epoch": 1.9318463444857497, "grad_norm": 1.9183920621871948, "learning_rate": 0.00010727969348659004, "loss": 2.2143, "step": 390 }, { "epoch": 1.936802973977695, "grad_norm": 1.6060720682144165, "learning_rate": 0.00010702426564495532, "loss": 2.5327, "step": 391 }, { "epoch": 1.9417596034696407, "grad_norm": 1.6848859786987305, "learning_rate": 0.00010676883780332057, "loss": 2.3799, "step": 392 }, { "epoch": 1.9467162329615861, "grad_norm": 1.8130978345870972, "learning_rate": 0.00010651340996168584, "loss": 2.0972, "step": 393 }, { "epoch": 1.9516728624535316, "grad_norm": 2.161240816116333, "learning_rate": 0.00010625798212005109, "loss": 2.5925, "step": 394 }, { "epoch": 1.9566294919454772, "grad_norm": 2.2761476039886475, "learning_rate": 0.00010600255427841634, "loss": 2.6694, "step": 395 }, { "epoch": 1.9615861214374224, "grad_norm": 1.6192595958709717, "learning_rate": 0.00010574712643678162, "loss": 2.5228, "step": 396 }, { "epoch": 1.966542750929368, "grad_norm": 1.955870270729065, "learning_rate": 0.00010549169859514687, "loss": 2.7556, "step": 397 }, { "epoch": 1.9714993804213135, "grad_norm": 1.966259241104126, "learning_rate": 0.00010523627075351215, "loss": 2.7568, "step": 398 }, { "epoch": 1.9764560099132589, "grad_norm": 1.7899080514907837, "learning_rate": 0.0001049808429118774, "loss": 2.2289, "step": 399 }, { "epoch": 1.9814126394052045, "grad_norm": 1.9266703128814697, "learning_rate": 0.00010472541507024267, "loss": 2.8342, "step": 400 }, { "epoch": 1.98636926889715, "grad_norm": 1.9955323934555054, "learning_rate": 0.00010446998722860792, "loss": 2.5556, "step": 401 }, { "epoch": 1.9913258983890954, "grad_norm": 1.5379672050476074, "learning_rate": 0.00010421455938697317, "loss": 2.32, "step": 402 }, { "epoch": 1.996282527881041, "grad_norm": 1.9460934400558472, "learning_rate": 0.00010395913154533845, "loss": 2.4778, "step": 403 }, { "epoch": 2.0, "grad_norm": 1.82111656665802, "learning_rate": 0.0001037037037037037, "loss": 2.1362, "step": 404 }, { "epoch": 2.0, "eval_loss": 2.5610225200653076, "eval_runtime": 22.6572, "eval_samples_per_second": 17.831, "eval_steps_per_second": 2.251, "step": 404 }, { "epoch": 2.0049566294919456, "grad_norm": 1.49136221408844, "learning_rate": 0.00010344827586206898, "loss": 2.0203, "step": 405 }, { "epoch": 2.009913258983891, "grad_norm": 1.8060108423233032, "learning_rate": 0.00010319284802043423, "loss": 2.0393, "step": 406 }, { "epoch": 2.0148698884758365, "grad_norm": 1.7297966480255127, "learning_rate": 0.0001029374201787995, "loss": 2.2408, "step": 407 }, { "epoch": 2.019826517967782, "grad_norm": 1.556800365447998, "learning_rate": 0.00010268199233716475, "loss": 1.9249, "step": 408 }, { "epoch": 2.0247831474597273, "grad_norm": 1.6368837356567383, "learning_rate": 0.00010242656449553, "loss": 2.1408, "step": 409 }, { "epoch": 2.029739776951673, "grad_norm": 1.6964534521102905, "learning_rate": 0.00010217113665389528, "loss": 2.0488, "step": 410 }, { "epoch": 2.034696406443618, "grad_norm": 1.8046094179153442, "learning_rate": 0.00010191570881226053, "loss": 2.3179, "step": 411 }, { "epoch": 2.039653035935564, "grad_norm": 1.6061656475067139, "learning_rate": 0.0001016602809706258, "loss": 2.245, "step": 412 }, { "epoch": 2.0446096654275094, "grad_norm": 1.763981580734253, "learning_rate": 0.00010140485312899106, "loss": 2.1131, "step": 413 }, { "epoch": 2.0495662949194546, "grad_norm": 1.6230251789093018, "learning_rate": 0.00010114942528735633, "loss": 1.8017, "step": 414 }, { "epoch": 2.0545229244114003, "grad_norm": 2.0794503688812256, "learning_rate": 0.00010089399744572158, "loss": 2.0948, "step": 415 }, { "epoch": 2.059479553903346, "grad_norm": 1.9313017129898071, "learning_rate": 0.00010063856960408685, "loss": 2.3105, "step": 416 }, { "epoch": 2.064436183395291, "grad_norm": 2.045888662338257, "learning_rate": 0.00010038314176245211, "loss": 2.3168, "step": 417 }, { "epoch": 2.0693928128872368, "grad_norm": 1.9172661304473877, "learning_rate": 0.00010012771392081737, "loss": 2.2448, "step": 418 }, { "epoch": 2.074349442379182, "grad_norm": 1.7956594228744507, "learning_rate": 9.987228607918264e-05, "loss": 2.5406, "step": 419 }, { "epoch": 2.0793060718711276, "grad_norm": 2.210123062133789, "learning_rate": 9.96168582375479e-05, "loss": 2.3232, "step": 420 }, { "epoch": 2.0842627013630732, "grad_norm": 1.982399821281433, "learning_rate": 9.936143039591316e-05, "loss": 1.9847, "step": 421 }, { "epoch": 2.0892193308550184, "grad_norm": 1.915616750717163, "learning_rate": 9.910600255427843e-05, "loss": 2.4827, "step": 422 }, { "epoch": 2.094175960346964, "grad_norm": 2.1594433784484863, "learning_rate": 9.885057471264369e-05, "loss": 2.2737, "step": 423 }, { "epoch": 2.0991325898389097, "grad_norm": 1.9990178346633911, "learning_rate": 9.859514687100895e-05, "loss": 2.2268, "step": 424 }, { "epoch": 2.104089219330855, "grad_norm": 1.9061527252197266, "learning_rate": 9.833971902937422e-05, "loss": 2.0164, "step": 425 }, { "epoch": 2.1090458488228006, "grad_norm": 1.9608116149902344, "learning_rate": 9.808429118773947e-05, "loss": 2.2343, "step": 426 }, { "epoch": 2.114002478314746, "grad_norm": 1.7611507177352905, "learning_rate": 9.782886334610473e-05, "loss": 2.1769, "step": 427 }, { "epoch": 2.1189591078066914, "grad_norm": 1.9356578588485718, "learning_rate": 9.757343550446999e-05, "loss": 2.0807, "step": 428 }, { "epoch": 2.123915737298637, "grad_norm": 2.1335885524749756, "learning_rate": 9.731800766283526e-05, "loss": 2.2264, "step": 429 }, { "epoch": 2.1288723667905822, "grad_norm": 1.7350636720657349, "learning_rate": 9.706257982120052e-05, "loss": 2.0111, "step": 430 }, { "epoch": 2.133828996282528, "grad_norm": 2.3487319946289062, "learning_rate": 9.680715197956578e-05, "loss": 2.634, "step": 431 }, { "epoch": 2.1387856257744735, "grad_norm": 2.1219382286071777, "learning_rate": 9.655172413793105e-05, "loss": 2.2604, "step": 432 }, { "epoch": 2.1437422552664187, "grad_norm": 2.3536593914031982, "learning_rate": 9.62962962962963e-05, "loss": 2.7046, "step": 433 }, { "epoch": 2.1486988847583643, "grad_norm": 2.1545400619506836, "learning_rate": 9.604086845466156e-05, "loss": 1.8161, "step": 434 }, { "epoch": 2.15365551425031, "grad_norm": 1.805413842201233, "learning_rate": 9.578544061302682e-05, "loss": 2.0053, "step": 435 }, { "epoch": 2.158612143742255, "grad_norm": 1.9322012662887573, "learning_rate": 9.553001277139209e-05, "loss": 2.0946, "step": 436 }, { "epoch": 2.163568773234201, "grad_norm": 1.7978978157043457, "learning_rate": 9.527458492975735e-05, "loss": 2.059, "step": 437 }, { "epoch": 2.168525402726146, "grad_norm": 2.3015379905700684, "learning_rate": 9.501915708812261e-05, "loss": 2.2307, "step": 438 }, { "epoch": 2.1734820322180917, "grad_norm": 1.9533964395523071, "learning_rate": 9.476372924648788e-05, "loss": 1.7493, "step": 439 }, { "epoch": 2.1784386617100373, "grad_norm": 2.048163414001465, "learning_rate": 9.450830140485314e-05, "loss": 2.1165, "step": 440 }, { "epoch": 2.1833952912019825, "grad_norm": 1.8135693073272705, "learning_rate": 9.425287356321839e-05, "loss": 1.9652, "step": 441 }, { "epoch": 2.188351920693928, "grad_norm": 2.1022565364837646, "learning_rate": 9.399744572158365e-05, "loss": 1.8261, "step": 442 }, { "epoch": 2.193308550185874, "grad_norm": 2.4095077514648438, "learning_rate": 9.374201787994892e-05, "loss": 2.3235, "step": 443 }, { "epoch": 2.198265179677819, "grad_norm": 2.243868350982666, "learning_rate": 9.348659003831418e-05, "loss": 2.1528, "step": 444 }, { "epoch": 2.2032218091697646, "grad_norm": 2.369640350341797, "learning_rate": 9.323116219667944e-05, "loss": 2.3696, "step": 445 }, { "epoch": 2.20817843866171, "grad_norm": 1.9481455087661743, "learning_rate": 9.29757343550447e-05, "loss": 2.584, "step": 446 }, { "epoch": 2.2131350681536555, "grad_norm": 2.1010377407073975, "learning_rate": 9.272030651340997e-05, "loss": 1.97, "step": 447 }, { "epoch": 2.218091697645601, "grad_norm": 2.5184261798858643, "learning_rate": 9.246487867177522e-05, "loss": 2.1047, "step": 448 }, { "epoch": 2.2230483271375463, "grad_norm": 2.3073112964630127, "learning_rate": 9.220945083014048e-05, "loss": 2.2998, "step": 449 }, { "epoch": 2.228004956629492, "grad_norm": 2.2876088619232178, "learning_rate": 9.195402298850575e-05, "loss": 2.452, "step": 450 }, { "epoch": 2.2329615861214376, "grad_norm": 3.126934289932251, "learning_rate": 9.169859514687101e-05, "loss": 2.6555, "step": 451 }, { "epoch": 2.2379182156133828, "grad_norm": 2.048002004623413, "learning_rate": 9.144316730523627e-05, "loss": 1.8546, "step": 452 }, { "epoch": 2.2428748451053284, "grad_norm": 1.8380862474441528, "learning_rate": 9.118773946360154e-05, "loss": 2.6203, "step": 453 }, { "epoch": 2.247831474597274, "grad_norm": 1.662384033203125, "learning_rate": 9.09323116219668e-05, "loss": 1.6749, "step": 454 }, { "epoch": 2.2527881040892193, "grad_norm": 1.770814299583435, "learning_rate": 9.067688378033205e-05, "loss": 1.9807, "step": 455 }, { "epoch": 2.257744733581165, "grad_norm": 2.3417553901672363, "learning_rate": 9.042145593869731e-05, "loss": 2.1493, "step": 456 }, { "epoch": 2.26270136307311, "grad_norm": 2.1765296459198, "learning_rate": 9.016602809706258e-05, "loss": 1.9696, "step": 457 }, { "epoch": 2.2676579925650557, "grad_norm": 2.5647733211517334, "learning_rate": 8.991060025542784e-05, "loss": 2.4712, "step": 458 }, { "epoch": 2.2726146220570014, "grad_norm": 2.0637850761413574, "learning_rate": 8.96551724137931e-05, "loss": 2.4461, "step": 459 }, { "epoch": 2.2775712515489466, "grad_norm": 2.011399030685425, "learning_rate": 8.939974457215837e-05, "loss": 2.231, "step": 460 }, { "epoch": 2.282527881040892, "grad_norm": 2.521390676498413, "learning_rate": 8.914431673052363e-05, "loss": 2.0845, "step": 461 }, { "epoch": 2.287484510532838, "grad_norm": 2.049290895462036, "learning_rate": 8.888888888888889e-05, "loss": 2.1271, "step": 462 }, { "epoch": 2.292441140024783, "grad_norm": 2.2892420291900635, "learning_rate": 8.863346104725416e-05, "loss": 2.4772, "step": 463 }, { "epoch": 2.2973977695167287, "grad_norm": 2.2590816020965576, "learning_rate": 8.837803320561942e-05, "loss": 2.2858, "step": 464 }, { "epoch": 2.3023543990086743, "grad_norm": 2.2737834453582764, "learning_rate": 8.812260536398468e-05, "loss": 2.0891, "step": 465 }, { "epoch": 2.3073110285006195, "grad_norm": 2.037618398666382, "learning_rate": 8.786717752234995e-05, "loss": 2.2451, "step": 466 }, { "epoch": 2.312267657992565, "grad_norm": 2.5216243267059326, "learning_rate": 8.761174968071521e-05, "loss": 2.5849, "step": 467 }, { "epoch": 2.3172242874845104, "grad_norm": 2.036977529525757, "learning_rate": 8.735632183908047e-05, "loss": 1.9554, "step": 468 }, { "epoch": 2.322180916976456, "grad_norm": 2.277539014816284, "learning_rate": 8.710089399744572e-05, "loss": 2.3908, "step": 469 }, { "epoch": 2.3271375464684017, "grad_norm": 2.4834628105163574, "learning_rate": 8.684546615581099e-05, "loss": 2.7662, "step": 470 }, { "epoch": 2.332094175960347, "grad_norm": 2.456012487411499, "learning_rate": 8.659003831417625e-05, "loss": 2.285, "step": 471 }, { "epoch": 2.3370508054522925, "grad_norm": 2.3700196743011475, "learning_rate": 8.633461047254151e-05, "loss": 2.3568, "step": 472 }, { "epoch": 2.3420074349442377, "grad_norm": 2.248645067214966, "learning_rate": 8.607918263090678e-05, "loss": 1.9046, "step": 473 }, { "epoch": 2.3469640644361833, "grad_norm": 2.076503038406372, "learning_rate": 8.582375478927204e-05, "loss": 1.8723, "step": 474 }, { "epoch": 2.351920693928129, "grad_norm": 2.2433621883392334, "learning_rate": 8.55683269476373e-05, "loss": 2.2344, "step": 475 }, { "epoch": 2.356877323420074, "grad_norm": 1.9284616708755493, "learning_rate": 8.531289910600255e-05, "loss": 1.898, "step": 476 }, { "epoch": 2.36183395291202, "grad_norm": 1.9935495853424072, "learning_rate": 8.505747126436782e-05, "loss": 1.879, "step": 477 }, { "epoch": 2.3667905824039654, "grad_norm": 1.9034062623977661, "learning_rate": 8.480204342273308e-05, "loss": 2.1425, "step": 478 }, { "epoch": 2.3717472118959106, "grad_norm": 2.1569809913635254, "learning_rate": 8.454661558109834e-05, "loss": 1.9005, "step": 479 }, { "epoch": 2.3767038413878563, "grad_norm": 2.1620583534240723, "learning_rate": 8.42911877394636e-05, "loss": 2.4077, "step": 480 }, { "epoch": 2.381660470879802, "grad_norm": 2.290148973464966, "learning_rate": 8.403575989782887e-05, "loss": 2.3896, "step": 481 }, { "epoch": 2.386617100371747, "grad_norm": 1.9209684133529663, "learning_rate": 8.378033205619413e-05, "loss": 2.0921, "step": 482 }, { "epoch": 2.3915737298636928, "grad_norm": 2.356311798095703, "learning_rate": 8.35249042145594e-05, "loss": 2.0363, "step": 483 }, { "epoch": 2.396530359355638, "grad_norm": 2.5131113529205322, "learning_rate": 8.326947637292465e-05, "loss": 2.7636, "step": 484 }, { "epoch": 2.4014869888475836, "grad_norm": 2.132436752319336, "learning_rate": 8.301404853128991e-05, "loss": 2.2943, "step": 485 }, { "epoch": 2.4064436183395292, "grad_norm": 2.254635810852051, "learning_rate": 8.275862068965517e-05, "loss": 2.4961, "step": 486 }, { "epoch": 2.4114002478314744, "grad_norm": 2.748410701751709, "learning_rate": 8.250319284802044e-05, "loss": 2.6137, "step": 487 }, { "epoch": 2.41635687732342, "grad_norm": 2.170868158340454, "learning_rate": 8.22477650063857e-05, "loss": 2.1018, "step": 488 }, { "epoch": 2.4213135068153657, "grad_norm": 2.4472765922546387, "learning_rate": 8.199233716475096e-05, "loss": 2.5278, "step": 489 }, { "epoch": 2.426270136307311, "grad_norm": 1.8834490776062012, "learning_rate": 8.173690932311623e-05, "loss": 1.8536, "step": 490 }, { "epoch": 2.4312267657992566, "grad_norm": 1.76395583152771, "learning_rate": 8.148148148148148e-05, "loss": 1.9397, "step": 491 }, { "epoch": 2.436183395291202, "grad_norm": 2.677208662033081, "learning_rate": 8.122605363984674e-05, "loss": 2.1501, "step": 492 }, { "epoch": 2.4411400247831474, "grad_norm": 1.9943513870239258, "learning_rate": 8.0970625798212e-05, "loss": 2.3874, "step": 493 }, { "epoch": 2.446096654275093, "grad_norm": 2.273705244064331, "learning_rate": 8.071519795657727e-05, "loss": 2.0298, "step": 494 }, { "epoch": 2.4510532837670382, "grad_norm": 2.03985595703125, "learning_rate": 8.045977011494253e-05, "loss": 1.9825, "step": 495 }, { "epoch": 2.456009913258984, "grad_norm": 1.9304776191711426, "learning_rate": 8.020434227330779e-05, "loss": 1.8176, "step": 496 }, { "epoch": 2.4609665427509295, "grad_norm": 2.278892755508423, "learning_rate": 7.994891443167306e-05, "loss": 2.3092, "step": 497 }, { "epoch": 2.4659231722428747, "grad_norm": 2.163693428039551, "learning_rate": 7.969348659003832e-05, "loss": 2.195, "step": 498 }, { "epoch": 2.4708798017348204, "grad_norm": 2.5456533432006836, "learning_rate": 7.943805874840358e-05, "loss": 2.2795, "step": 499 }, { "epoch": 2.4758364312267656, "grad_norm": 2.565223455429077, "learning_rate": 7.918263090676885e-05, "loss": 2.3311, "step": 500 }, { "epoch": 2.480793060718711, "grad_norm": 2.468602180480957, "learning_rate": 7.892720306513411e-05, "loss": 2.6716, "step": 501 }, { "epoch": 2.485749690210657, "grad_norm": 2.2904815673828125, "learning_rate": 7.867177522349937e-05, "loss": 2.1953, "step": 502 }, { "epoch": 2.4907063197026025, "grad_norm": 2.0819876194000244, "learning_rate": 7.841634738186464e-05, "loss": 2.2799, "step": 503 }, { "epoch": 2.4956629491945477, "grad_norm": 2.262396812438965, "learning_rate": 7.81609195402299e-05, "loss": 2.3044, "step": 504 }, { "epoch": 2.5006195786864933, "grad_norm": 2.2627463340759277, "learning_rate": 7.790549169859515e-05, "loss": 2.3557, "step": 505 }, { "epoch": 2.5055762081784385, "grad_norm": 2.0544724464416504, "learning_rate": 7.765006385696041e-05, "loss": 1.9032, "step": 506 }, { "epoch": 2.510532837670384, "grad_norm": 2.1277432441711426, "learning_rate": 7.739463601532568e-05, "loss": 2.3961, "step": 507 }, { "epoch": 2.51548946716233, "grad_norm": 2.0343055725097656, "learning_rate": 7.713920817369094e-05, "loss": 2.037, "step": 508 }, { "epoch": 2.520446096654275, "grad_norm": 2.4001617431640625, "learning_rate": 7.68837803320562e-05, "loss": 2.4522, "step": 509 }, { "epoch": 2.5254027261462206, "grad_norm": 1.8327491283416748, "learning_rate": 7.662835249042147e-05, "loss": 1.9539, "step": 510 }, { "epoch": 2.530359355638166, "grad_norm": 1.9295401573181152, "learning_rate": 7.637292464878673e-05, "loss": 1.9554, "step": 511 }, { "epoch": 2.5353159851301115, "grad_norm": 2.409006118774414, "learning_rate": 7.611749680715198e-05, "loss": 1.8266, "step": 512 }, { "epoch": 2.540272614622057, "grad_norm": 2.3173720836639404, "learning_rate": 7.586206896551724e-05, "loss": 2.4242, "step": 513 }, { "epoch": 2.5452292441140028, "grad_norm": 2.0689756870269775, "learning_rate": 7.56066411238825e-05, "loss": 2.1029, "step": 514 }, { "epoch": 2.550185873605948, "grad_norm": 2.1280126571655273, "learning_rate": 7.535121328224777e-05, "loss": 2.1136, "step": 515 }, { "epoch": 2.5551425030978936, "grad_norm": 2.9854607582092285, "learning_rate": 7.509578544061303e-05, "loss": 2.8267, "step": 516 }, { "epoch": 2.560099132589839, "grad_norm": 2.0449581146240234, "learning_rate": 7.48403575989783e-05, "loss": 2.1204, "step": 517 }, { "epoch": 2.5650557620817844, "grad_norm": 2.152194023132324, "learning_rate": 7.458492975734356e-05, "loss": 1.6798, "step": 518 }, { "epoch": 2.57001239157373, "grad_norm": 2.301673412322998, "learning_rate": 7.432950191570882e-05, "loss": 2.2609, "step": 519 }, { "epoch": 2.5749690210656753, "grad_norm": 2.390002489089966, "learning_rate": 7.407407407407407e-05, "loss": 2.4318, "step": 520 }, { "epoch": 2.579925650557621, "grad_norm": 3.2430877685546875, "learning_rate": 7.381864623243934e-05, "loss": 2.4938, "step": 521 }, { "epoch": 2.584882280049566, "grad_norm": 2.227306842803955, "learning_rate": 7.35632183908046e-05, "loss": 2.1597, "step": 522 }, { "epoch": 2.5898389095415117, "grad_norm": 2.1724979877471924, "learning_rate": 7.330779054916986e-05, "loss": 2.1702, "step": 523 }, { "epoch": 2.5947955390334574, "grad_norm": 2.4228322505950928, "learning_rate": 7.305236270753513e-05, "loss": 2.4855, "step": 524 }, { "epoch": 2.5997521685254026, "grad_norm": 2.368286609649658, "learning_rate": 7.279693486590039e-05, "loss": 2.5217, "step": 525 }, { "epoch": 2.6047087980173482, "grad_norm": 1.8513636589050293, "learning_rate": 7.254150702426565e-05, "loss": 2.2185, "step": 526 }, { "epoch": 2.6096654275092934, "grad_norm": 2.2481653690338135, "learning_rate": 7.22860791826309e-05, "loss": 2.1565, "step": 527 }, { "epoch": 2.614622057001239, "grad_norm": 2.042464256286621, "learning_rate": 7.203065134099617e-05, "loss": 2.2602, "step": 528 }, { "epoch": 2.6195786864931847, "grad_norm": 2.1742701530456543, "learning_rate": 7.177522349936143e-05, "loss": 2.0985, "step": 529 }, { "epoch": 2.6245353159851303, "grad_norm": 2.0808088779449463, "learning_rate": 7.151979565772669e-05, "loss": 2.0583, "step": 530 }, { "epoch": 2.6294919454770755, "grad_norm": 1.9378777742385864, "learning_rate": 7.126436781609196e-05, "loss": 1.7537, "step": 531 }, { "epoch": 2.634448574969021, "grad_norm": 2.5020270347595215, "learning_rate": 7.100893997445722e-05, "loss": 2.3168, "step": 532 }, { "epoch": 2.6394052044609664, "grad_norm": 2.423002004623413, "learning_rate": 7.075351213282248e-05, "loss": 2.2957, "step": 533 }, { "epoch": 2.644361833952912, "grad_norm": 1.9235918521881104, "learning_rate": 7.049808429118773e-05, "loss": 2.1923, "step": 534 }, { "epoch": 2.6493184634448577, "grad_norm": 2.0859873294830322, "learning_rate": 7.0242656449553e-05, "loss": 2.1833, "step": 535 }, { "epoch": 2.654275092936803, "grad_norm": 2.3581674098968506, "learning_rate": 6.998722860791826e-05, "loss": 2.3217, "step": 536 }, { "epoch": 2.6592317224287485, "grad_norm": 2.184673309326172, "learning_rate": 6.973180076628352e-05, "loss": 2.2694, "step": 537 }, { "epoch": 2.6641883519206937, "grad_norm": 2.373626708984375, "learning_rate": 6.947637292464879e-05, "loss": 2.2407, "step": 538 }, { "epoch": 2.6691449814126393, "grad_norm": 2.328784704208374, "learning_rate": 6.922094508301405e-05, "loss": 2.6027, "step": 539 }, { "epoch": 2.674101610904585, "grad_norm": 2.451972007751465, "learning_rate": 6.896551724137931e-05, "loss": 2.1135, "step": 540 }, { "epoch": 2.6790582403965306, "grad_norm": 2.210793972015381, "learning_rate": 6.871008939974458e-05, "loss": 2.0632, "step": 541 }, { "epoch": 2.684014869888476, "grad_norm": 2.022038698196411, "learning_rate": 6.845466155810984e-05, "loss": 1.7921, "step": 542 }, { "epoch": 2.6889714993804215, "grad_norm": 2.6344008445739746, "learning_rate": 6.81992337164751e-05, "loss": 2.2798, "step": 543 }, { "epoch": 2.6939281288723667, "grad_norm": 2.514261484146118, "learning_rate": 6.794380587484037e-05, "loss": 2.424, "step": 544 }, { "epoch": 2.6988847583643123, "grad_norm": 2.2451043128967285, "learning_rate": 6.768837803320563e-05, "loss": 1.9677, "step": 545 }, { "epoch": 2.703841387856258, "grad_norm": 2.241933822631836, "learning_rate": 6.74329501915709e-05, "loss": 2.4206, "step": 546 }, { "epoch": 2.708798017348203, "grad_norm": 2.3536107540130615, "learning_rate": 6.717752234993616e-05, "loss": 2.5814, "step": 547 }, { "epoch": 2.7137546468401488, "grad_norm": 2.215730905532837, "learning_rate": 6.69220945083014e-05, "loss": 1.8021, "step": 548 }, { "epoch": 2.718711276332094, "grad_norm": 2.3178601264953613, "learning_rate": 6.666666666666667e-05, "loss": 2.2829, "step": 549 }, { "epoch": 2.7236679058240396, "grad_norm": 2.420584201812744, "learning_rate": 6.641123882503193e-05, "loss": 2.3131, "step": 550 }, { "epoch": 2.7286245353159853, "grad_norm": 2.242386817932129, "learning_rate": 6.61558109833972e-05, "loss": 1.9557, "step": 551 }, { "epoch": 2.7335811648079305, "grad_norm": 2.050896167755127, "learning_rate": 6.590038314176246e-05, "loss": 1.9561, "step": 552 }, { "epoch": 2.738537794299876, "grad_norm": 2.038100004196167, "learning_rate": 6.564495530012772e-05, "loss": 2.1168, "step": 553 }, { "epoch": 2.7434944237918213, "grad_norm": 2.1917381286621094, "learning_rate": 6.538952745849299e-05, "loss": 1.873, "step": 554 }, { "epoch": 2.748451053283767, "grad_norm": 2.4844274520874023, "learning_rate": 6.513409961685824e-05, "loss": 2.3741, "step": 555 }, { "epoch": 2.7534076827757126, "grad_norm": 2.1887197494506836, "learning_rate": 6.48786717752235e-05, "loss": 1.7504, "step": 556 }, { "epoch": 2.758364312267658, "grad_norm": 2.391392230987549, "learning_rate": 6.462324393358876e-05, "loss": 2.3918, "step": 557 }, { "epoch": 2.7633209417596034, "grad_norm": 2.299013376235962, "learning_rate": 6.436781609195403e-05, "loss": 2.0697, "step": 558 }, { "epoch": 2.768277571251549, "grad_norm": 2.294445276260376, "learning_rate": 6.411238825031929e-05, "loss": 2.017, "step": 559 }, { "epoch": 2.7732342007434942, "grad_norm": 2.0729377269744873, "learning_rate": 6.385696040868455e-05, "loss": 2.2024, "step": 560 }, { "epoch": 2.77819083023544, "grad_norm": 2.1257901191711426, "learning_rate": 6.360153256704982e-05, "loss": 1.9048, "step": 561 }, { "epoch": 2.7831474597273855, "grad_norm": 2.4315128326416016, "learning_rate": 6.334610472541508e-05, "loss": 2.8421, "step": 562 }, { "epoch": 2.7881040892193307, "grad_norm": 2.669069766998291, "learning_rate": 6.309067688378033e-05, "loss": 2.061, "step": 563 }, { "epoch": 2.7930607187112764, "grad_norm": 2.463329792022705, "learning_rate": 6.283524904214559e-05, "loss": 2.0992, "step": 564 }, { "epoch": 2.7980173482032216, "grad_norm": 2.218747615814209, "learning_rate": 6.257982120051086e-05, "loss": 2.0431, "step": 565 }, { "epoch": 2.802973977695167, "grad_norm": 2.394122838973999, "learning_rate": 6.232439335887612e-05, "loss": 2.1749, "step": 566 }, { "epoch": 2.807930607187113, "grad_norm": 2.188235282897949, "learning_rate": 6.206896551724138e-05, "loss": 2.2733, "step": 567 }, { "epoch": 2.8128872366790585, "grad_norm": 2.446723222732544, "learning_rate": 6.181353767560665e-05, "loss": 2.2211, "step": 568 }, { "epoch": 2.8178438661710037, "grad_norm": 2.2640678882598877, "learning_rate": 6.155810983397191e-05, "loss": 2.1568, "step": 569 }, { "epoch": 2.8228004956629493, "grad_norm": 2.8429107666015625, "learning_rate": 6.130268199233716e-05, "loss": 2.8168, "step": 570 }, { "epoch": 2.8277571251548945, "grad_norm": 2.4229469299316406, "learning_rate": 6.104725415070242e-05, "loss": 2.1375, "step": 571 }, { "epoch": 2.83271375464684, "grad_norm": 2.336423397064209, "learning_rate": 6.0791826309067686e-05, "loss": 2.7039, "step": 572 }, { "epoch": 2.837670384138786, "grad_norm": 2.691897392272949, "learning_rate": 6.053639846743295e-05, "loss": 2.4094, "step": 573 }, { "epoch": 2.842627013630731, "grad_norm": 2.258892059326172, "learning_rate": 6.028097062579821e-05, "loss": 1.7147, "step": 574 }, { "epoch": 2.8475836431226766, "grad_norm": 2.352938175201416, "learning_rate": 6.0025542784163477e-05, "loss": 2.0789, "step": 575 }, { "epoch": 2.852540272614622, "grad_norm": 2.318082094192505, "learning_rate": 5.977011494252874e-05, "loss": 1.8802, "step": 576 }, { "epoch": 2.8574969021065675, "grad_norm": 2.9496710300445557, "learning_rate": 5.9514687100893996e-05, "loss": 2.376, "step": 577 }, { "epoch": 2.862453531598513, "grad_norm": 2.5339314937591553, "learning_rate": 5.925925925925926e-05, "loss": 2.2208, "step": 578 }, { "epoch": 2.8674101610904588, "grad_norm": 2.220191717147827, "learning_rate": 5.900383141762452e-05, "loss": 1.9416, "step": 579 }, { "epoch": 2.872366790582404, "grad_norm": 2.244831085205078, "learning_rate": 5.8748403575989787e-05, "loss": 2.287, "step": 580 }, { "epoch": 2.8773234200743496, "grad_norm": 2.2947471141815186, "learning_rate": 5.849297573435505e-05, "loss": 2.233, "step": 581 }, { "epoch": 2.882280049566295, "grad_norm": 2.1228411197662354, "learning_rate": 5.823754789272031e-05, "loss": 2.012, "step": 582 }, { "epoch": 2.8872366790582404, "grad_norm": 2.3730921745300293, "learning_rate": 5.798212005108558e-05, "loss": 2.3333, "step": 583 }, { "epoch": 2.892193308550186, "grad_norm": 2.1657650470733643, "learning_rate": 5.7726692209450826e-05, "loss": 2.3758, "step": 584 }, { "epoch": 2.8971499380421313, "grad_norm": 2.4342710971832275, "learning_rate": 5.747126436781609e-05, "loss": 2.5381, "step": 585 }, { "epoch": 2.902106567534077, "grad_norm": 2.218479633331299, "learning_rate": 5.721583652618135e-05, "loss": 2.3628, "step": 586 }, { "epoch": 2.907063197026022, "grad_norm": 2.1390647888183594, "learning_rate": 5.6960408684546617e-05, "loss": 2.304, "step": 587 }, { "epoch": 2.9120198265179678, "grad_norm": 2.6552858352661133, "learning_rate": 5.670498084291188e-05, "loss": 2.2981, "step": 588 }, { "epoch": 2.9169764560099134, "grad_norm": 2.6417832374572754, "learning_rate": 5.644955300127714e-05, "loss": 2.4692, "step": 589 }, { "epoch": 2.9219330855018586, "grad_norm": 2.1957082748413086, "learning_rate": 5.6194125159642407e-05, "loss": 2.0055, "step": 590 }, { "epoch": 2.9268897149938042, "grad_norm": 2.549053192138672, "learning_rate": 5.593869731800766e-05, "loss": 1.9994, "step": 591 }, { "epoch": 2.9318463444857494, "grad_norm": 2.4547526836395264, "learning_rate": 5.5683269476372927e-05, "loss": 2.0306, "step": 592 }, { "epoch": 2.936802973977695, "grad_norm": 2.595532178878784, "learning_rate": 5.542784163473819e-05, "loss": 2.3909, "step": 593 }, { "epoch": 2.9417596034696407, "grad_norm": 2.1407456398010254, "learning_rate": 5.517241379310345e-05, "loss": 2.1741, "step": 594 }, { "epoch": 2.9467162329615864, "grad_norm": 2.4364256858825684, "learning_rate": 5.491698595146872e-05, "loss": 2.4011, "step": 595 }, { "epoch": 2.9516728624535316, "grad_norm": 2.644935369491577, "learning_rate": 5.466155810983398e-05, "loss": 2.2418, "step": 596 }, { "epoch": 2.956629491945477, "grad_norm": 2.2009565830230713, "learning_rate": 5.440613026819924e-05, "loss": 2.3574, "step": 597 }, { "epoch": 2.9615861214374224, "grad_norm": 2.40665340423584, "learning_rate": 5.415070242656451e-05, "loss": 2.2198, "step": 598 }, { "epoch": 2.966542750929368, "grad_norm": 1.834892749786377, "learning_rate": 5.3895274584929756e-05, "loss": 2.0522, "step": 599 }, { "epoch": 2.9714993804213137, "grad_norm": 2.602015256881714, "learning_rate": 5.363984674329502e-05, "loss": 2.2518, "step": 600 }, { "epoch": 2.976456009913259, "grad_norm": 2.3021697998046875, "learning_rate": 5.338441890166028e-05, "loss": 2.1467, "step": 601 }, { "epoch": 2.9814126394052045, "grad_norm": 2.2826194763183594, "learning_rate": 5.3128991060025547e-05, "loss": 2.3371, "step": 602 }, { "epoch": 2.9863692688971497, "grad_norm": 1.934054970741272, "learning_rate": 5.287356321839081e-05, "loss": 1.9699, "step": 603 }, { "epoch": 2.9913258983890954, "grad_norm": 2.2678050994873047, "learning_rate": 5.261813537675607e-05, "loss": 2.2934, "step": 604 }, { "epoch": 2.996282527881041, "grad_norm": 2.5755562782287598, "learning_rate": 5.236270753512134e-05, "loss": 2.4522, "step": 605 }, { "epoch": 3.0, "grad_norm": 3.2130398750305176, "learning_rate": 5.2107279693486586e-05, "loss": 2.4297, "step": 606 }, { "epoch": 3.0, "eval_loss": 2.5098254680633545, "eval_runtime": 22.3298, "eval_samples_per_second": 18.092, "eval_steps_per_second": 2.284, "step": 606 }, { "epoch": 3.0049566294919456, "grad_norm": 1.9976105690002441, "learning_rate": 5.185185185185185e-05, "loss": 1.9214, "step": 607 }, { "epoch": 3.009913258983891, "grad_norm": 2.0963776111602783, "learning_rate": 5.159642401021711e-05, "loss": 2.1928, "step": 608 }, { "epoch": 3.0148698884758365, "grad_norm": 1.9764257669448853, "learning_rate": 5.1340996168582377e-05, "loss": 1.8546, "step": 609 }, { "epoch": 3.019826517967782, "grad_norm": 2.3067879676818848, "learning_rate": 5.108556832694764e-05, "loss": 2.5585, "step": 610 }, { "epoch": 3.0247831474597273, "grad_norm": 2.3767294883728027, "learning_rate": 5.08301404853129e-05, "loss": 2.2411, "step": 611 }, { "epoch": 3.029739776951673, "grad_norm": 1.899346947669983, "learning_rate": 5.057471264367817e-05, "loss": 1.8045, "step": 612 }, { "epoch": 3.034696406443618, "grad_norm": 2.2742886543273926, "learning_rate": 5.031928480204342e-05, "loss": 1.9178, "step": 613 }, { "epoch": 3.039653035935564, "grad_norm": 2.208033561706543, "learning_rate": 5.0063856960408687e-05, "loss": 2.2826, "step": 614 }, { "epoch": 3.0446096654275094, "grad_norm": 2.00700306892395, "learning_rate": 4.980842911877395e-05, "loss": 1.7436, "step": 615 }, { "epoch": 3.0495662949194546, "grad_norm": 2.484027147293091, "learning_rate": 4.955300127713921e-05, "loss": 2.3736, "step": 616 }, { "epoch": 3.0545229244114003, "grad_norm": 2.4913110733032227, "learning_rate": 4.929757343550448e-05, "loss": 2.2775, "step": 617 }, { "epoch": 3.059479553903346, "grad_norm": 1.8059406280517578, "learning_rate": 4.904214559386973e-05, "loss": 1.5242, "step": 618 }, { "epoch": 3.064436183395291, "grad_norm": 2.3238720893859863, "learning_rate": 4.8786717752234997e-05, "loss": 2.1255, "step": 619 }, { "epoch": 3.0693928128872368, "grad_norm": 1.9283521175384521, "learning_rate": 4.853128991060026e-05, "loss": 1.5993, "step": 620 }, { "epoch": 3.074349442379182, "grad_norm": 2.2595303058624268, "learning_rate": 4.827586206896552e-05, "loss": 1.999, "step": 621 }, { "epoch": 3.0793060718711276, "grad_norm": 2.251521110534668, "learning_rate": 4.802043422733078e-05, "loss": 1.8928, "step": 622 }, { "epoch": 3.0842627013630732, "grad_norm": 2.7896904945373535, "learning_rate": 4.776500638569604e-05, "loss": 2.0757, "step": 623 }, { "epoch": 3.0892193308550184, "grad_norm": 2.178480863571167, "learning_rate": 4.7509578544061307e-05, "loss": 2.0942, "step": 624 }, { "epoch": 3.094175960346964, "grad_norm": 2.765531539916992, "learning_rate": 4.725415070242657e-05, "loss": 1.9807, "step": 625 }, { "epoch": 3.0991325898389097, "grad_norm": 1.9775290489196777, "learning_rate": 4.6998722860791827e-05, "loss": 1.9953, "step": 626 }, { "epoch": 3.104089219330855, "grad_norm": 2.0385220050811768, "learning_rate": 4.674329501915709e-05, "loss": 1.6417, "step": 627 }, { "epoch": 3.1090458488228006, "grad_norm": 2.535924196243286, "learning_rate": 4.648786717752235e-05, "loss": 2.3085, "step": 628 }, { "epoch": 3.114002478314746, "grad_norm": 2.8627915382385254, "learning_rate": 4.623243933588761e-05, "loss": 2.1261, "step": 629 }, { "epoch": 3.1189591078066914, "grad_norm": 2.845181941986084, "learning_rate": 4.597701149425287e-05, "loss": 2.1133, "step": 630 }, { "epoch": 3.123915737298637, "grad_norm": 2.202937602996826, "learning_rate": 4.5721583652618137e-05, "loss": 1.967, "step": 631 }, { "epoch": 3.1288723667905822, "grad_norm": 2.412930727005005, "learning_rate": 4.54661558109834e-05, "loss": 1.8034, "step": 632 }, { "epoch": 3.133828996282528, "grad_norm": 2.56500244140625, "learning_rate": 4.5210727969348656e-05, "loss": 2.2214, "step": 633 }, { "epoch": 3.1387856257744735, "grad_norm": 2.520071506500244, "learning_rate": 4.495530012771392e-05, "loss": 1.9551, "step": 634 }, { "epoch": 3.1437422552664187, "grad_norm": 2.391106605529785, "learning_rate": 4.469987228607918e-05, "loss": 1.8267, "step": 635 }, { "epoch": 3.1486988847583643, "grad_norm": 2.86560320854187, "learning_rate": 4.4444444444444447e-05, "loss": 2.0847, "step": 636 }, { "epoch": 3.15365551425031, "grad_norm": 2.5105133056640625, "learning_rate": 4.418901660280971e-05, "loss": 1.9333, "step": 637 }, { "epoch": 3.158612143742255, "grad_norm": 2.3541910648345947, "learning_rate": 4.393358876117497e-05, "loss": 2.0839, "step": 638 }, { "epoch": 3.163568773234201, "grad_norm": 2.4766783714294434, "learning_rate": 4.367816091954024e-05, "loss": 1.9689, "step": 639 }, { "epoch": 3.168525402726146, "grad_norm": 2.1712841987609863, "learning_rate": 4.342273307790549e-05, "loss": 1.7512, "step": 640 }, { "epoch": 3.1734820322180917, "grad_norm": 2.2210240364074707, "learning_rate": 4.3167305236270757e-05, "loss": 1.8299, "step": 641 }, { "epoch": 3.1784386617100373, "grad_norm": 2.5499587059020996, "learning_rate": 4.291187739463602e-05, "loss": 2.051, "step": 642 }, { "epoch": 3.1833952912019825, "grad_norm": 2.5971527099609375, "learning_rate": 4.2656449553001277e-05, "loss": 2.0613, "step": 643 }, { "epoch": 3.188351920693928, "grad_norm": 2.212960720062256, "learning_rate": 4.240102171136654e-05, "loss": 2.0225, "step": 644 }, { "epoch": 3.193308550185874, "grad_norm": 2.652787923812866, "learning_rate": 4.21455938697318e-05, "loss": 2.1012, "step": 645 }, { "epoch": 3.198265179677819, "grad_norm": 2.414275884628296, "learning_rate": 4.189016602809707e-05, "loss": 2.159, "step": 646 }, { "epoch": 3.2032218091697646, "grad_norm": 2.223020076751709, "learning_rate": 4.163473818646232e-05, "loss": 1.6145, "step": 647 }, { "epoch": 3.20817843866171, "grad_norm": 2.7731528282165527, "learning_rate": 4.1379310344827587e-05, "loss": 2.3619, "step": 648 }, { "epoch": 3.2131350681536555, "grad_norm": 2.2961618900299072, "learning_rate": 4.112388250319285e-05, "loss": 1.9229, "step": 649 }, { "epoch": 3.218091697645601, "grad_norm": 2.880171298980713, "learning_rate": 4.086845466155811e-05, "loss": 2.4915, "step": 650 }, { "epoch": 3.2230483271375463, "grad_norm": 2.2541770935058594, "learning_rate": 4.061302681992337e-05, "loss": 2.224, "step": 651 }, { "epoch": 3.228004956629492, "grad_norm": 2.63120436668396, "learning_rate": 4.035759897828863e-05, "loss": 1.9003, "step": 652 }, { "epoch": 3.2329615861214376, "grad_norm": 2.031409740447998, "learning_rate": 4.0102171136653897e-05, "loss": 1.9379, "step": 653 }, { "epoch": 3.2379182156133828, "grad_norm": 2.8178703784942627, "learning_rate": 3.984674329501916e-05, "loss": 1.8691, "step": 654 }, { "epoch": 3.2428748451053284, "grad_norm": 2.5390496253967285, "learning_rate": 3.959131545338442e-05, "loss": 1.9412, "step": 655 }, { "epoch": 3.247831474597274, "grad_norm": 2.23886775970459, "learning_rate": 3.933588761174969e-05, "loss": 1.8117, "step": 656 }, { "epoch": 3.2527881040892193, "grad_norm": 2.419747829437256, "learning_rate": 3.908045977011495e-05, "loss": 1.8492, "step": 657 }, { "epoch": 3.257744733581165, "grad_norm": 2.5116543769836426, "learning_rate": 3.8825031928480207e-05, "loss": 1.9955, "step": 658 }, { "epoch": 3.26270136307311, "grad_norm": 2.7984719276428223, "learning_rate": 3.856960408684547e-05, "loss": 2.0657, "step": 659 }, { "epoch": 3.2676579925650557, "grad_norm": 2.3362345695495605, "learning_rate": 3.831417624521073e-05, "loss": 1.9131, "step": 660 }, { "epoch": 3.2726146220570014, "grad_norm": 3.0645365715026855, "learning_rate": 3.805874840357599e-05, "loss": 2.3735, "step": 661 }, { "epoch": 3.2775712515489466, "grad_norm": 2.3989381790161133, "learning_rate": 3.780332056194125e-05, "loss": 1.4717, "step": 662 }, { "epoch": 3.282527881040892, "grad_norm": 2.7305102348327637, "learning_rate": 3.7547892720306517e-05, "loss": 2.3454, "step": 663 }, { "epoch": 3.287484510532838, "grad_norm": 2.355215311050415, "learning_rate": 3.729246487867178e-05, "loss": 1.9257, "step": 664 }, { "epoch": 3.292441140024783, "grad_norm": 2.841524600982666, "learning_rate": 3.7037037037037037e-05, "loss": 2.1013, "step": 665 }, { "epoch": 3.2973977695167287, "grad_norm": 1.97605562210083, "learning_rate": 3.67816091954023e-05, "loss": 1.415, "step": 666 }, { "epoch": 3.3023543990086743, "grad_norm": 2.803922653198242, "learning_rate": 3.652618135376756e-05, "loss": 2.3917, "step": 667 }, { "epoch": 3.3073110285006195, "grad_norm": 2.375274658203125, "learning_rate": 3.627075351213283e-05, "loss": 2.3449, "step": 668 }, { "epoch": 3.312267657992565, "grad_norm": 2.461966037750244, "learning_rate": 3.601532567049808e-05, "loss": 2.0543, "step": 669 }, { "epoch": 3.3172242874845104, "grad_norm": 2.7819225788116455, "learning_rate": 3.5759897828863347e-05, "loss": 2.0613, "step": 670 }, { "epoch": 3.322180916976456, "grad_norm": 2.084023952484131, "learning_rate": 3.550446998722861e-05, "loss": 1.5794, "step": 671 }, { "epoch": 3.3271375464684017, "grad_norm": 2.4474074840545654, "learning_rate": 3.5249042145593867e-05, "loss": 1.973, "step": 672 }, { "epoch": 3.332094175960347, "grad_norm": 2.4582390785217285, "learning_rate": 3.499361430395913e-05, "loss": 1.7705, "step": 673 }, { "epoch": 3.3370508054522925, "grad_norm": 2.56362247467041, "learning_rate": 3.473818646232439e-05, "loss": 1.9695, "step": 674 }, { "epoch": 3.3420074349442377, "grad_norm": 2.6630728244781494, "learning_rate": 3.4482758620689657e-05, "loss": 2.0741, "step": 675 }, { "epoch": 3.3469640644361833, "grad_norm": 2.5984740257263184, "learning_rate": 3.422733077905492e-05, "loss": 2.029, "step": 676 }, { "epoch": 3.351920693928129, "grad_norm": 2.3361611366271973, "learning_rate": 3.397190293742018e-05, "loss": 1.8688, "step": 677 }, { "epoch": 3.356877323420074, "grad_norm": 3.3585948944091797, "learning_rate": 3.371647509578545e-05, "loss": 2.6338, "step": 678 }, { "epoch": 3.36183395291202, "grad_norm": 2.697134494781494, "learning_rate": 3.34610472541507e-05, "loss": 1.847, "step": 679 }, { "epoch": 3.3667905824039654, "grad_norm": 2.71582293510437, "learning_rate": 3.3205619412515967e-05, "loss": 1.8604, "step": 680 }, { "epoch": 3.3717472118959106, "grad_norm": 2.484410285949707, "learning_rate": 3.295019157088123e-05, "loss": 2.3045, "step": 681 }, { "epoch": 3.3767038413878563, "grad_norm": 2.4801011085510254, "learning_rate": 3.269476372924649e-05, "loss": 1.9622, "step": 682 }, { "epoch": 3.381660470879802, "grad_norm": 2.462303638458252, "learning_rate": 3.243933588761175e-05, "loss": 2.1321, "step": 683 }, { "epoch": 3.386617100371747, "grad_norm": 2.5208513736724854, "learning_rate": 3.218390804597701e-05, "loss": 2.031, "step": 684 }, { "epoch": 3.3915737298636928, "grad_norm": 2.2665512561798096, "learning_rate": 3.192848020434228e-05, "loss": 1.9891, "step": 685 }, { "epoch": 3.396530359355638, "grad_norm": 2.9241855144500732, "learning_rate": 3.167305236270754e-05, "loss": 2.3861, "step": 686 }, { "epoch": 3.4014869888475836, "grad_norm": 2.353585720062256, "learning_rate": 3.1417624521072797e-05, "loss": 1.7977, "step": 687 }, { "epoch": 3.4064436183395292, "grad_norm": 2.5655272006988525, "learning_rate": 3.116219667943806e-05, "loss": 1.782, "step": 688 }, { "epoch": 3.4114002478314744, "grad_norm": 2.2319204807281494, "learning_rate": 3.090676883780332e-05, "loss": 1.4128, "step": 689 }, { "epoch": 3.41635687732342, "grad_norm": 2.703676462173462, "learning_rate": 3.065134099616858e-05, "loss": 2.0543, "step": 690 }, { "epoch": 3.4213135068153657, "grad_norm": 2.8589353561401367, "learning_rate": 3.0395913154533843e-05, "loss": 2.11, "step": 691 }, { "epoch": 3.426270136307311, "grad_norm": 2.237912893295288, "learning_rate": 3.0140485312899107e-05, "loss": 1.8919, "step": 692 }, { "epoch": 3.4312267657992566, "grad_norm": 2.673888683319092, "learning_rate": 2.988505747126437e-05, "loss": 2.1326, "step": 693 }, { "epoch": 3.436183395291202, "grad_norm": 2.9751524925231934, "learning_rate": 2.962962962962963e-05, "loss": 2.3678, "step": 694 }, { "epoch": 3.4411400247831474, "grad_norm": 3.3065667152404785, "learning_rate": 2.9374201787994893e-05, "loss": 2.5252, "step": 695 }, { "epoch": 3.446096654275093, "grad_norm": 2.837353229522705, "learning_rate": 2.9118773946360157e-05, "loss": 2.3186, "step": 696 }, { "epoch": 3.4510532837670382, "grad_norm": 2.4829537868499756, "learning_rate": 2.8863346104725413e-05, "loss": 2.0368, "step": 697 }, { "epoch": 3.456009913258984, "grad_norm": 3.0476629734039307, "learning_rate": 2.8607918263090677e-05, "loss": 2.124, "step": 698 }, { "epoch": 3.4609665427509295, "grad_norm": 2.930732488632202, "learning_rate": 2.835249042145594e-05, "loss": 2.2542, "step": 699 }, { "epoch": 3.4659231722428747, "grad_norm": 2.7399027347564697, "learning_rate": 2.8097062579821203e-05, "loss": 2.0625, "step": 700 }, { "epoch": 3.4708798017348204, "grad_norm": 2.625471591949463, "learning_rate": 2.7841634738186463e-05, "loss": 2.2255, "step": 701 }, { "epoch": 3.4758364312267656, "grad_norm": 2.4757065773010254, "learning_rate": 2.7586206896551727e-05, "loss": 2.0139, "step": 702 }, { "epoch": 3.480793060718711, "grad_norm": 2.4212238788604736, "learning_rate": 2.733077905491699e-05, "loss": 1.9171, "step": 703 }, { "epoch": 3.485749690210657, "grad_norm": 2.9213318824768066, "learning_rate": 2.7075351213282253e-05, "loss": 2.3692, "step": 704 }, { "epoch": 3.4907063197026025, "grad_norm": 2.563901424407959, "learning_rate": 2.681992337164751e-05, "loss": 2.2226, "step": 705 }, { "epoch": 3.4956629491945477, "grad_norm": 2.412309169769287, "learning_rate": 2.6564495530012773e-05, "loss": 1.7855, "step": 706 }, { "epoch": 3.5006195786864933, "grad_norm": 2.7574050426483154, "learning_rate": 2.6309067688378037e-05, "loss": 1.8671, "step": 707 }, { "epoch": 3.5055762081784385, "grad_norm": 2.588981866836548, "learning_rate": 2.6053639846743293e-05, "loss": 1.847, "step": 708 }, { "epoch": 3.510532837670384, "grad_norm": 2.5374417304992676, "learning_rate": 2.5798212005108557e-05, "loss": 1.8264, "step": 709 }, { "epoch": 3.51548946716233, "grad_norm": 2.9206414222717285, "learning_rate": 2.554278416347382e-05, "loss": 2.1438, "step": 710 }, { "epoch": 3.520446096654275, "grad_norm": 3.5458905696868896, "learning_rate": 2.5287356321839083e-05, "loss": 2.1679, "step": 711 }, { "epoch": 3.5254027261462206, "grad_norm": 2.4408674240112305, "learning_rate": 2.5031928480204343e-05, "loss": 1.8783, "step": 712 }, { "epoch": 3.530359355638166, "grad_norm": 2.9399070739746094, "learning_rate": 2.4776500638569607e-05, "loss": 1.7827, "step": 713 }, { "epoch": 3.5353159851301115, "grad_norm": 2.5893685817718506, "learning_rate": 2.4521072796934867e-05, "loss": 1.9307, "step": 714 }, { "epoch": 3.540272614622057, "grad_norm": 2.5607779026031494, "learning_rate": 2.426564495530013e-05, "loss": 2.1557, "step": 715 }, { "epoch": 3.5452292441140028, "grad_norm": 2.6460790634155273, "learning_rate": 2.401021711366539e-05, "loss": 1.8946, "step": 716 }, { "epoch": 3.550185873605948, "grad_norm": 2.498994827270508, "learning_rate": 2.3754789272030653e-05, "loss": 1.987, "step": 717 }, { "epoch": 3.5551425030978936, "grad_norm": 2.8072335720062256, "learning_rate": 2.3499361430395913e-05, "loss": 2.3501, "step": 718 }, { "epoch": 3.560099132589839, "grad_norm": 2.413820743560791, "learning_rate": 2.3243933588761177e-05, "loss": 2.3033, "step": 719 }, { "epoch": 3.5650557620817844, "grad_norm": 1.9895007610321045, "learning_rate": 2.2988505747126437e-05, "loss": 1.5816, "step": 720 }, { "epoch": 3.57001239157373, "grad_norm": 2.67324161529541, "learning_rate": 2.27330779054917e-05, "loss": 2.3242, "step": 721 }, { "epoch": 3.5749690210656753, "grad_norm": 2.9094364643096924, "learning_rate": 2.247765006385696e-05, "loss": 2.133, "step": 722 }, { "epoch": 3.579925650557621, "grad_norm": 2.4571454524993896, "learning_rate": 2.2222222222222223e-05, "loss": 2.089, "step": 723 }, { "epoch": 3.584882280049566, "grad_norm": 2.40213680267334, "learning_rate": 2.1966794380587487e-05, "loss": 2.013, "step": 724 }, { "epoch": 3.5898389095415117, "grad_norm": 2.4887630939483643, "learning_rate": 2.1711366538952747e-05, "loss": 2.0719, "step": 725 }, { "epoch": 3.5947955390334574, "grad_norm": 3.0302541255950928, "learning_rate": 2.145593869731801e-05, "loss": 2.2166, "step": 726 }, { "epoch": 3.5997521685254026, "grad_norm": 2.511434555053711, "learning_rate": 2.120051085568327e-05, "loss": 2.312, "step": 727 }, { "epoch": 3.6047087980173482, "grad_norm": 2.49587082862854, "learning_rate": 2.0945083014048533e-05, "loss": 2.1587, "step": 728 }, { "epoch": 3.6096654275092934, "grad_norm": 2.305344820022583, "learning_rate": 2.0689655172413793e-05, "loss": 1.9972, "step": 729 }, { "epoch": 3.614622057001239, "grad_norm": 2.829852342605591, "learning_rate": 2.0434227330779057e-05, "loss": 2.1588, "step": 730 }, { "epoch": 3.6195786864931847, "grad_norm": 2.3796768188476562, "learning_rate": 2.0178799489144317e-05, "loss": 1.7094, "step": 731 }, { "epoch": 3.6245353159851303, "grad_norm": 2.1699445247650146, "learning_rate": 1.992337164750958e-05, "loss": 1.8184, "step": 732 }, { "epoch": 3.6294919454770755, "grad_norm": 2.304624319076538, "learning_rate": 1.9667943805874843e-05, "loss": 2.1114, "step": 733 }, { "epoch": 3.634448574969021, "grad_norm": 2.507122278213501, "learning_rate": 1.9412515964240103e-05, "loss": 2.0594, "step": 734 }, { "epoch": 3.6394052044609664, "grad_norm": 2.8564951419830322, "learning_rate": 1.9157088122605367e-05, "loss": 2.1098, "step": 735 }, { "epoch": 3.644361833952912, "grad_norm": 2.4493966102600098, "learning_rate": 1.8901660280970627e-05, "loss": 2.1243, "step": 736 }, { "epoch": 3.6493184634448577, "grad_norm": 2.634030818939209, "learning_rate": 1.864623243933589e-05, "loss": 2.1568, "step": 737 }, { "epoch": 3.654275092936803, "grad_norm": 2.6991872787475586, "learning_rate": 1.839080459770115e-05, "loss": 2.094, "step": 738 }, { "epoch": 3.6592317224287485, "grad_norm": 2.587801694869995, "learning_rate": 1.8135376756066413e-05, "loss": 2.1519, "step": 739 }, { "epoch": 3.6641883519206937, "grad_norm": 2.746302366256714, "learning_rate": 1.7879948914431673e-05, "loss": 2.0224, "step": 740 }, { "epoch": 3.6691449814126393, "grad_norm": 2.5982632637023926, "learning_rate": 1.7624521072796933e-05, "loss": 2.229, "step": 741 }, { "epoch": 3.674101610904585, "grad_norm": 2.684934616088867, "learning_rate": 1.7369093231162197e-05, "loss": 2.1958, "step": 742 }, { "epoch": 3.6790582403965306, "grad_norm": 2.1835129261016846, "learning_rate": 1.711366538952746e-05, "loss": 2.0661, "step": 743 }, { "epoch": 3.684014869888476, "grad_norm": 2.6260294914245605, "learning_rate": 1.6858237547892723e-05, "loss": 2.3006, "step": 744 }, { "epoch": 3.6889714993804215, "grad_norm": 2.598024368286133, "learning_rate": 1.6602809706257983e-05, "loss": 2.184, "step": 745 }, { "epoch": 3.6939281288723667, "grad_norm": 2.5640017986297607, "learning_rate": 1.6347381864623247e-05, "loss": 1.9889, "step": 746 }, { "epoch": 3.6988847583643123, "grad_norm": 2.520355224609375, "learning_rate": 1.6091954022988507e-05, "loss": 1.8721, "step": 747 }, { "epoch": 3.703841387856258, "grad_norm": 2.578373432159424, "learning_rate": 1.583652618135377e-05, "loss": 1.6672, "step": 748 }, { "epoch": 3.708798017348203, "grad_norm": 2.8206024169921875, "learning_rate": 1.558109833971903e-05, "loss": 2.3436, "step": 749 }, { "epoch": 3.7137546468401488, "grad_norm": 2.4661734104156494, "learning_rate": 1.532567049808429e-05, "loss": 1.8102, "step": 750 }, { "epoch": 3.718711276332094, "grad_norm": 2.458994150161743, "learning_rate": 1.5070242656449553e-05, "loss": 2.2049, "step": 751 }, { "epoch": 3.7236679058240396, "grad_norm": 2.699479579925537, "learning_rate": 1.4814814814814815e-05, "loss": 1.8601, "step": 752 }, { "epoch": 3.7286245353159853, "grad_norm": 2.7305147647857666, "learning_rate": 1.4559386973180078e-05, "loss": 2.0493, "step": 753 }, { "epoch": 3.7335811648079305, "grad_norm": 2.724635601043701, "learning_rate": 1.4303959131545338e-05, "loss": 1.9252, "step": 754 }, { "epoch": 3.738537794299876, "grad_norm": 2.5254998207092285, "learning_rate": 1.4048531289910602e-05, "loss": 1.7042, "step": 755 }, { "epoch": 3.7434944237918213, "grad_norm": 2.5693461894989014, "learning_rate": 1.3793103448275863e-05, "loss": 1.9372, "step": 756 }, { "epoch": 3.748451053283767, "grad_norm": 3.0173473358154297, "learning_rate": 1.3537675606641127e-05, "loss": 2.0502, "step": 757 }, { "epoch": 3.7534076827757126, "grad_norm": 2.4707775115966797, "learning_rate": 1.3282247765006387e-05, "loss": 1.7202, "step": 758 }, { "epoch": 3.758364312267658, "grad_norm": 2.898653268814087, "learning_rate": 1.3026819923371647e-05, "loss": 1.9074, "step": 759 }, { "epoch": 3.7633209417596034, "grad_norm": 2.3936469554901123, "learning_rate": 1.277139208173691e-05, "loss": 1.7491, "step": 760 }, { "epoch": 3.768277571251549, "grad_norm": 2.245955228805542, "learning_rate": 1.2515964240102172e-05, "loss": 1.8061, "step": 761 }, { "epoch": 3.7732342007434942, "grad_norm": 2.400726079940796, "learning_rate": 1.2260536398467433e-05, "loss": 1.7594, "step": 762 }, { "epoch": 3.77819083023544, "grad_norm": 2.5081756114959717, "learning_rate": 1.2005108556832695e-05, "loss": 1.7719, "step": 763 }, { "epoch": 3.7831474597273855, "grad_norm": 2.9115960597991943, "learning_rate": 1.1749680715197957e-05, "loss": 2.307, "step": 764 }, { "epoch": 3.7881040892193307, "grad_norm": 2.9008727073669434, "learning_rate": 1.1494252873563218e-05, "loss": 2.3572, "step": 765 }, { "epoch": 3.7930607187112764, "grad_norm": 2.6959750652313232, "learning_rate": 1.123882503192848e-05, "loss": 2.3963, "step": 766 }, { "epoch": 3.7980173482032216, "grad_norm": 2.554203748703003, "learning_rate": 1.0983397190293743e-05, "loss": 2.0953, "step": 767 }, { "epoch": 3.802973977695167, "grad_norm": 3.1628901958465576, "learning_rate": 1.0727969348659005e-05, "loss": 2.322, "step": 768 }, { "epoch": 3.807930607187113, "grad_norm": 2.8941919803619385, "learning_rate": 1.0472541507024267e-05, "loss": 2.4856, "step": 769 }, { "epoch": 3.8128872366790585, "grad_norm": 2.6858787536621094, "learning_rate": 1.0217113665389528e-05, "loss": 1.7071, "step": 770 }, { "epoch": 3.8178438661710037, "grad_norm": 2.8813092708587646, "learning_rate": 9.96168582375479e-06, "loss": 2.1721, "step": 771 }, { "epoch": 3.8228004956629493, "grad_norm": 2.395799398422241, "learning_rate": 9.706257982120052e-06, "loss": 2.0502, "step": 772 }, { "epoch": 3.8277571251548945, "grad_norm": 2.8856003284454346, "learning_rate": 9.450830140485313e-06, "loss": 2.2581, "step": 773 }, { "epoch": 3.83271375464684, "grad_norm": 2.4333102703094482, "learning_rate": 9.195402298850575e-06, "loss": 2.0938, "step": 774 }, { "epoch": 3.837670384138786, "grad_norm": 2.266862392425537, "learning_rate": 8.939974457215837e-06, "loss": 2.3068, "step": 775 }, { "epoch": 3.842627013630731, "grad_norm": 2.494243860244751, "learning_rate": 8.684546615581098e-06, "loss": 1.687, "step": 776 }, { "epoch": 3.8475836431226766, "grad_norm": 2.6965909004211426, "learning_rate": 8.429118773946362e-06, "loss": 2.5372, "step": 777 }, { "epoch": 3.852540272614622, "grad_norm": 2.7889437675476074, "learning_rate": 8.173690932311623e-06, "loss": 2.1596, "step": 778 }, { "epoch": 3.8574969021065675, "grad_norm": 2.788628101348877, "learning_rate": 7.918263090676885e-06, "loss": 2.2257, "step": 779 }, { "epoch": 3.862453531598513, "grad_norm": 2.956270217895508, "learning_rate": 7.662835249042145e-06, "loss": 2.1526, "step": 780 }, { "epoch": 3.8674101610904588, "grad_norm": 2.884690523147583, "learning_rate": 7.4074074074074075e-06, "loss": 2.1096, "step": 781 }, { "epoch": 3.872366790582404, "grad_norm": 2.3916282653808594, "learning_rate": 7.151979565772669e-06, "loss": 2.0638, "step": 782 }, { "epoch": 3.8773234200743496, "grad_norm": 2.6462628841400146, "learning_rate": 6.896551724137932e-06, "loss": 2.0776, "step": 783 }, { "epoch": 3.882280049566295, "grad_norm": 2.674168348312378, "learning_rate": 6.641123882503193e-06, "loss": 1.9045, "step": 784 }, { "epoch": 3.8872366790582404, "grad_norm": 2.6898820400238037, "learning_rate": 6.385696040868455e-06, "loss": 2.2157, "step": 785 }, { "epoch": 3.892193308550186, "grad_norm": 2.8253626823425293, "learning_rate": 6.130268199233717e-06, "loss": 2.042, "step": 786 }, { "epoch": 3.8971499380421313, "grad_norm": 2.905292510986328, "learning_rate": 5.874840357598978e-06, "loss": 2.3422, "step": 787 }, { "epoch": 3.902106567534077, "grad_norm": 2.387803316116333, "learning_rate": 5.61941251596424e-06, "loss": 1.5696, "step": 788 }, { "epoch": 3.907063197026022, "grad_norm": 2.655133008956909, "learning_rate": 5.3639846743295025e-06, "loss": 1.8796, "step": 789 }, { "epoch": 3.9120198265179678, "grad_norm": 2.6278626918792725, "learning_rate": 5.108556832694764e-06, "loss": 1.8694, "step": 790 }, { "epoch": 3.9169764560099134, "grad_norm": 2.4667484760284424, "learning_rate": 4.853128991060026e-06, "loss": 2.1705, "step": 791 }, { "epoch": 3.9219330855018586, "grad_norm": 2.774759531021118, "learning_rate": 4.5977011494252875e-06, "loss": 2.1393, "step": 792 }, { "epoch": 3.9268897149938042, "grad_norm": 2.8241868019104004, "learning_rate": 4.342273307790549e-06, "loss": 2.3277, "step": 793 }, { "epoch": 3.9318463444857494, "grad_norm": 3.131408929824829, "learning_rate": 4.086845466155812e-06, "loss": 1.9978, "step": 794 }, { "epoch": 3.936802973977695, "grad_norm": 2.5323071479797363, "learning_rate": 3.8314176245210725e-06, "loss": 1.7299, "step": 795 }, { "epoch": 3.9417596034696407, "grad_norm": 2.7869760990142822, "learning_rate": 3.5759897828863346e-06, "loss": 1.9863, "step": 796 }, { "epoch": 3.9467162329615864, "grad_norm": 2.5885322093963623, "learning_rate": 3.3205619412515967e-06, "loss": 1.9325, "step": 797 }, { "epoch": 3.9516728624535316, "grad_norm": 2.5993456840515137, "learning_rate": 3.0651340996168583e-06, "loss": 2.6309, "step": 798 }, { "epoch": 3.956629491945477, "grad_norm": 2.5578904151916504, "learning_rate": 2.80970625798212e-06, "loss": 1.8496, "step": 799 }, { "epoch": 3.9615861214374224, "grad_norm": 2.696652889251709, "learning_rate": 2.554278416347382e-06, "loss": 1.9756, "step": 800 }, { "epoch": 3.966542750929368, "grad_norm": 2.8636531829833984, "learning_rate": 2.2988505747126437e-06, "loss": 2.3127, "step": 801 }, { "epoch": 3.9714993804213137, "grad_norm": 2.5958805084228516, "learning_rate": 2.043422733077906e-06, "loss": 1.5032, "step": 802 }, { "epoch": 3.976456009913259, "grad_norm": 2.6378211975097656, "learning_rate": 1.7879948914431673e-06, "loss": 1.8637, "step": 803 }, { "epoch": 3.9814126394052045, "grad_norm": 3.125774383544922, "learning_rate": 1.5325670498084292e-06, "loss": 2.3156, "step": 804 }, { "epoch": 3.9863692688971497, "grad_norm": 2.5696659088134766, "learning_rate": 1.277139208173691e-06, "loss": 2.1169, "step": 805 }, { "epoch": 3.9913258983890954, "grad_norm": 2.8002192974090576, "learning_rate": 1.021711366538953e-06, "loss": 2.2148, "step": 806 }, { "epoch": 3.996282527881041, "grad_norm": 2.850720167160034, "learning_rate": 7.662835249042146e-07, "loss": 1.8883, "step": 807 }, { "epoch": 4.0, "grad_norm": 3.3131906986236572, "learning_rate": 5.108556832694765e-07, "loss": 1.8331, "step": 808 }, { "epoch": 4.0, "eval_loss": 2.5112364292144775, "eval_runtime": 22.6481, "eval_samples_per_second": 17.838, "eval_steps_per_second": 2.252, "step": 808 } ], "logging_steps": 1, "max_steps": 808, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6015167842222080.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }