{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9999270764967549, "eval_steps": 500, "global_step": 1714, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002916940129803836, "grad_norm": 9.421533600382833, "learning_rate": 1.1627906976744187e-07, "loss": 1.4153, "num_tokens": 671048.0, "step": 5 }, { "epoch": 0.005833880259607672, "grad_norm": 9.417087087487797, "learning_rate": 2.616279069767442e-07, "loss": 1.4115, "num_tokens": 1340514.0, "step": 10 }, { "epoch": 0.008750820389411508, "grad_norm": 8.64652450056725, "learning_rate": 4.0697674418604655e-07, "loss": 1.3864, "num_tokens": 1975142.0, "step": 15 }, { "epoch": 0.011667760519215344, "grad_norm": 6.364802582777174, "learning_rate": 5.523255813953489e-07, "loss": 1.2845, "num_tokens": 2629706.0, "step": 20 }, { "epoch": 0.01458470064901918, "grad_norm": 4.036625471123866, "learning_rate": 6.976744186046513e-07, "loss": 1.1622, "num_tokens": 3282682.0, "step": 25 }, { "epoch": 0.017501640778823015, "grad_norm": 4.1984081053855205, "learning_rate": 8.430232558139535e-07, "loss": 0.9647, "num_tokens": 3956656.0, "step": 30 }, { "epoch": 0.02041858090862685, "grad_norm": 2.9803387136629893, "learning_rate": 9.88372093023256e-07, "loss": 0.7913, "num_tokens": 4609315.0, "step": 35 }, { "epoch": 0.023335521038430687, "grad_norm": 1.5758517615202334, "learning_rate": 1.1337209302325581e-06, "loss": 0.6332, "num_tokens": 5258689.0, "step": 40 }, { "epoch": 0.02625246116823452, "grad_norm": 1.5452778456677885, "learning_rate": 1.2790697674418605e-06, "loss": 0.5782, "num_tokens": 5918842.0, "step": 45 }, { "epoch": 0.02916940129803836, "grad_norm": 1.3636179516378841, "learning_rate": 1.424418604651163e-06, "loss": 0.5198, "num_tokens": 6581062.0, "step": 50 }, { "epoch": 0.0320863414278422, "grad_norm": 1.3117664668985787, "learning_rate": 1.5697674418604653e-06, "loss": 0.4967, "num_tokens": 7229502.0, "step": 55 }, { "epoch": 0.03500328155764603, "grad_norm": 1.120351932885373, "learning_rate": 1.7151162790697675e-06, "loss": 0.4632, "num_tokens": 7873010.0, "step": 60 }, { "epoch": 0.037920221687449865, "grad_norm": 1.1652086123427123, "learning_rate": 1.86046511627907e-06, "loss": 0.4468, "num_tokens": 8521924.0, "step": 65 }, { "epoch": 0.0408371618172537, "grad_norm": 1.1356964733122763, "learning_rate": 2.005813953488372e-06, "loss": 0.4216, "num_tokens": 9195311.0, "step": 70 }, { "epoch": 0.04375410194705753, "grad_norm": 1.125126715027727, "learning_rate": 2.1511627906976745e-06, "loss": 0.4149, "num_tokens": 9857902.0, "step": 75 }, { "epoch": 0.046671042076861374, "grad_norm": 1.3774618307546735, "learning_rate": 2.296511627906977e-06, "loss": 0.4073, "num_tokens": 10508649.0, "step": 80 }, { "epoch": 0.04958798220666521, "grad_norm": 0.9977532994086189, "learning_rate": 2.4418604651162793e-06, "loss": 0.4007, "num_tokens": 11165778.0, "step": 85 }, { "epoch": 0.05250492233646904, "grad_norm": 1.1279002386313963, "learning_rate": 2.5872093023255817e-06, "loss": 0.3834, "num_tokens": 11832663.0, "step": 90 }, { "epoch": 0.05542186246627288, "grad_norm": 1.1949547587228786, "learning_rate": 2.7325581395348837e-06, "loss": 0.378, "num_tokens": 12491366.0, "step": 95 }, { "epoch": 0.05833880259607672, "grad_norm": 1.2272744151775523, "learning_rate": 2.8779069767441865e-06, "loss": 0.3711, "num_tokens": 13158513.0, "step": 100 }, { "epoch": 0.06125574272588055, "grad_norm": 1.1016643876078596, "learning_rate": 3.0232558139534885e-06, "loss": 0.3673, "num_tokens": 13815015.0, "step": 105 }, { "epoch": 0.0641726828556844, "grad_norm": 1.2643000713720414, "learning_rate": 3.168604651162791e-06, "loss": 0.3597, "num_tokens": 14469541.0, "step": 110 }, { "epoch": 0.06708962298548822, "grad_norm": 1.201838639094829, "learning_rate": 3.313953488372093e-06, "loss": 0.3555, "num_tokens": 15118625.0, "step": 115 }, { "epoch": 0.07000656311529206, "grad_norm": 1.109067555947549, "learning_rate": 3.4593023255813957e-06, "loss": 0.3533, "num_tokens": 15773909.0, "step": 120 }, { "epoch": 0.07292350324509589, "grad_norm": 1.1048072692029691, "learning_rate": 3.6046511627906977e-06, "loss": 0.347, "num_tokens": 16431558.0, "step": 125 }, { "epoch": 0.07584044337489973, "grad_norm": 1.1222392707741446, "learning_rate": 3.7500000000000005e-06, "loss": 0.3561, "num_tokens": 17071991.0, "step": 130 }, { "epoch": 0.07875738350470357, "grad_norm": 1.268152454159826, "learning_rate": 3.8953488372093025e-06, "loss": 0.3534, "num_tokens": 17712413.0, "step": 135 }, { "epoch": 0.0816743236345074, "grad_norm": 1.2649313269721758, "learning_rate": 4.040697674418605e-06, "loss": 0.3299, "num_tokens": 18365792.0, "step": 140 }, { "epoch": 0.08459126376431124, "grad_norm": 1.1434149458336165, "learning_rate": 4.186046511627907e-06, "loss": 0.3306, "num_tokens": 19044956.0, "step": 145 }, { "epoch": 0.08750820389411507, "grad_norm": 1.2086728419334098, "learning_rate": 4.331395348837209e-06, "loss": 0.3374, "num_tokens": 19695160.0, "step": 150 }, { "epoch": 0.09042514402391891, "grad_norm": 1.1141028558068582, "learning_rate": 4.476744186046512e-06, "loss": 0.3329, "num_tokens": 20354718.0, "step": 155 }, { "epoch": 0.09334208415372275, "grad_norm": 1.1566468839887507, "learning_rate": 4.622093023255814e-06, "loss": 0.3341, "num_tokens": 21006198.0, "step": 160 }, { "epoch": 0.09625902428352658, "grad_norm": 1.1614415172217762, "learning_rate": 4.767441860465117e-06, "loss": 0.3443, "num_tokens": 21634895.0, "step": 165 }, { "epoch": 0.09917596441333042, "grad_norm": 1.2888710984530731, "learning_rate": 4.912790697674419e-06, "loss": 0.3332, "num_tokens": 22279296.0, "step": 170 }, { "epoch": 0.10209290454313426, "grad_norm": 1.066687159230309, "learning_rate": 4.999981321460686e-06, "loss": 0.33, "num_tokens": 22932800.0, "step": 175 }, { "epoch": 0.10500984467293809, "grad_norm": 1.0587240072052146, "learning_rate": 4.999771191454943e-06, "loss": 0.3227, "num_tokens": 23596098.0, "step": 180 }, { "epoch": 0.10792678480274193, "grad_norm": 1.0753866637705207, "learning_rate": 4.999327605146962e-06, "loss": 0.3265, "num_tokens": 24240753.0, "step": 185 }, { "epoch": 0.11084372493254575, "grad_norm": 1.1696062309729527, "learning_rate": 4.998650608567207e-06, "loss": 0.3209, "num_tokens": 24903847.0, "step": 190 }, { "epoch": 0.1137606650623496, "grad_norm": 1.0355502586963417, "learning_rate": 4.997740271966877e-06, "loss": 0.3146, "num_tokens": 25564496.0, "step": 195 }, { "epoch": 0.11667760519215344, "grad_norm": 1.119036751899036, "learning_rate": 4.996596689810608e-06, "loss": 0.315, "num_tokens": 26231532.0, "step": 200 }, { "epoch": 0.11959454532195726, "grad_norm": 1.088954101903663, "learning_rate": 4.995219980766672e-06, "loss": 0.3178, "num_tokens": 26876933.0, "step": 205 }, { "epoch": 0.1225114854517611, "grad_norm": 1.1468096627392024, "learning_rate": 4.993610287694663e-06, "loss": 0.3251, "num_tokens": 27510135.0, "step": 210 }, { "epoch": 0.12542842558156495, "grad_norm": 0.9548342307118838, "learning_rate": 4.991767777630679e-06, "loss": 0.3145, "num_tokens": 28154194.0, "step": 215 }, { "epoch": 0.1283453657113688, "grad_norm": 1.0591304201257135, "learning_rate": 4.989692641769976e-06, "loss": 0.3229, "num_tokens": 28801919.0, "step": 220 }, { "epoch": 0.1312623058411726, "grad_norm": 1.054145800912703, "learning_rate": 4.987385095447141e-06, "loss": 0.3208, "num_tokens": 29456315.0, "step": 225 }, { "epoch": 0.13417924597097644, "grad_norm": 1.1458373353014295, "learning_rate": 4.9848453781137355e-06, "loss": 0.3136, "num_tokens": 30111377.0, "step": 230 }, { "epoch": 0.13709618610078028, "grad_norm": 1.1484854864144058, "learning_rate": 4.982073753313459e-06, "loss": 0.3106, "num_tokens": 30769448.0, "step": 235 }, { "epoch": 0.14001312623058412, "grad_norm": 1.1136794674885304, "learning_rate": 4.979070508654793e-06, "loss": 0.3123, "num_tokens": 31423870.0, "step": 240 }, { "epoch": 0.14293006636038796, "grad_norm": 1.1270064595880545, "learning_rate": 4.975835955781159e-06, "loss": 0.3093, "num_tokens": 32085017.0, "step": 245 }, { "epoch": 0.14584700649019178, "grad_norm": 0.9197262026545928, "learning_rate": 4.97237043033858e-06, "loss": 0.3099, "num_tokens": 32748599.0, "step": 250 }, { "epoch": 0.14876394661999562, "grad_norm": 1.0031484672449607, "learning_rate": 4.968674291940851e-06, "loss": 0.3113, "num_tokens": 33396816.0, "step": 255 }, { "epoch": 0.15168088674979946, "grad_norm": 1.0106860095524073, "learning_rate": 4.964747924132222e-06, "loss": 0.3019, "num_tokens": 34048082.0, "step": 260 }, { "epoch": 0.1545978268796033, "grad_norm": 1.1390370333863529, "learning_rate": 4.9605917343475955e-06, "loss": 0.3094, "num_tokens": 34695245.0, "step": 265 }, { "epoch": 0.15751476700940714, "grad_norm": 0.9304690639529852, "learning_rate": 4.9562061538702525e-06, "loss": 0.3022, "num_tokens": 35357239.0, "step": 270 }, { "epoch": 0.16043170713921096, "grad_norm": 1.0043265999798545, "learning_rate": 4.9515916377870946e-06, "loss": 0.3089, "num_tokens": 36009775.0, "step": 275 }, { "epoch": 0.1633486472690148, "grad_norm": 1.2003647631325731, "learning_rate": 4.946748664941422e-06, "loss": 0.3011, "num_tokens": 36668698.0, "step": 280 }, { "epoch": 0.16626558739881864, "grad_norm": 0.8918137550325422, "learning_rate": 4.941677737883243e-06, "loss": 0.2965, "num_tokens": 37330940.0, "step": 285 }, { "epoch": 0.16918252752862248, "grad_norm": 0.9369985322231161, "learning_rate": 4.936379382817128e-06, "loss": 0.3018, "num_tokens": 37988190.0, "step": 290 }, { "epoch": 0.17209946765842632, "grad_norm": 0.9779525599227024, "learning_rate": 4.930854149547601e-06, "loss": 0.2918, "num_tokens": 38661283.0, "step": 295 }, { "epoch": 0.17501640778823013, "grad_norm": 1.1129159421490522, "learning_rate": 4.925102611422091e-06, "loss": 0.2927, "num_tokens": 39320436.0, "step": 300 }, { "epoch": 0.17793334791803397, "grad_norm": 0.9718156616589078, "learning_rate": 4.919125365271435e-06, "loss": 0.2965, "num_tokens": 39974872.0, "step": 305 }, { "epoch": 0.18085028804783781, "grad_norm": 0.997867663517212, "learning_rate": 4.912923031347944e-06, "loss": 0.2887, "num_tokens": 40624237.0, "step": 310 }, { "epoch": 0.18376722817764166, "grad_norm": 1.0604840903516273, "learning_rate": 4.9064962532610465e-06, "loss": 0.3093, "num_tokens": 41272880.0, "step": 315 }, { "epoch": 0.1866841683074455, "grad_norm": 0.9752847517901945, "learning_rate": 4.899845697910489e-06, "loss": 0.3023, "num_tokens": 41916518.0, "step": 320 }, { "epoch": 0.18960110843724934, "grad_norm": 0.9361317315869321, "learning_rate": 4.89297205541715e-06, "loss": 0.2906, "num_tokens": 42570198.0, "step": 325 }, { "epoch": 0.19251804856705315, "grad_norm": 0.924312875998321, "learning_rate": 4.885876039051408e-06, "loss": 0.2973, "num_tokens": 43235563.0, "step": 330 }, { "epoch": 0.195434988696857, "grad_norm": 0.9664895892779287, "learning_rate": 4.878558385159143e-06, "loss": 0.2932, "num_tokens": 43881944.0, "step": 335 }, { "epoch": 0.19835192882666083, "grad_norm": 1.1025545029030783, "learning_rate": 4.871019853085316e-06, "loss": 0.2969, "num_tokens": 44536573.0, "step": 340 }, { "epoch": 0.20126886895646467, "grad_norm": 1.1546365826334446, "learning_rate": 4.863261225095175e-06, "loss": 0.2975, "num_tokens": 45185298.0, "step": 345 }, { "epoch": 0.20418580908626852, "grad_norm": 0.8539653068120385, "learning_rate": 4.8552833062930845e-06, "loss": 0.2941, "num_tokens": 45833516.0, "step": 350 }, { "epoch": 0.20710274921607233, "grad_norm": 0.8887354370276694, "learning_rate": 4.847086924538975e-06, "loss": 0.2911, "num_tokens": 46492785.0, "step": 355 }, { "epoch": 0.21001968934587617, "grad_norm": 0.8965576316891322, "learning_rate": 4.838672930362438e-06, "loss": 0.2845, "num_tokens": 47147972.0, "step": 360 }, { "epoch": 0.21293662947568, "grad_norm": 1.0547471171000238, "learning_rate": 4.830042196874472e-06, "loss": 0.2889, "num_tokens": 47792460.0, "step": 365 }, { "epoch": 0.21585356960548385, "grad_norm": 0.9029069684538267, "learning_rate": 4.821195619676876e-06, "loss": 0.2935, "num_tokens": 48449608.0, "step": 370 }, { "epoch": 0.2187705097352877, "grad_norm": 1.1056029287041187, "learning_rate": 4.812134116769312e-06, "loss": 0.2927, "num_tokens": 49105080.0, "step": 375 }, { "epoch": 0.2216874498650915, "grad_norm": 0.8797309000795296, "learning_rate": 4.802858628454056e-06, "loss": 0.2895, "num_tokens": 49754281.0, "step": 380 }, { "epoch": 0.22460438999489535, "grad_norm": 1.044569804031736, "learning_rate": 4.793370117238408e-06, "loss": 0.2897, "num_tokens": 50392756.0, "step": 385 }, { "epoch": 0.2275213301246992, "grad_norm": 1.1130572813337458, "learning_rate": 4.783669567734826e-06, "loss": 0.2899, "num_tokens": 51048274.0, "step": 390 }, { "epoch": 0.23043827025450303, "grad_norm": 0.960716559832758, "learning_rate": 4.773757986558753e-06, "loss": 0.2917, "num_tokens": 51699260.0, "step": 395 }, { "epoch": 0.23335521038430687, "grad_norm": 0.9245960376672167, "learning_rate": 4.763636402224151e-06, "loss": 0.2897, "num_tokens": 52348343.0, "step": 400 }, { "epoch": 0.23627215051411068, "grad_norm": 0.8890810056058448, "learning_rate": 4.753305865036787e-06, "loss": 0.2885, "num_tokens": 53001056.0, "step": 405 }, { "epoch": 0.23918909064391453, "grad_norm": 0.8579871280345769, "learning_rate": 4.742767446985234e-06, "loss": 0.2861, "num_tokens": 53662504.0, "step": 410 }, { "epoch": 0.24210603077371837, "grad_norm": 0.8821250216443968, "learning_rate": 4.732022241629637e-06, "loss": 0.287, "num_tokens": 54318994.0, "step": 415 }, { "epoch": 0.2450229709035222, "grad_norm": 0.7974431662516955, "learning_rate": 4.721071363988235e-06, "loss": 0.2905, "num_tokens": 54969623.0, "step": 420 }, { "epoch": 0.24793991103332605, "grad_norm": 0.8353737566224922, "learning_rate": 4.709915950421653e-06, "loss": 0.2878, "num_tokens": 55620185.0, "step": 425 }, { "epoch": 0.2508568511631299, "grad_norm": 0.9111882828366351, "learning_rate": 4.698557158514988e-06, "loss": 0.2828, "num_tokens": 56274073.0, "step": 430 }, { "epoch": 0.2537737912929337, "grad_norm": 0.919475515341444, "learning_rate": 4.686996166957683e-06, "loss": 0.2822, "num_tokens": 56944526.0, "step": 435 }, { "epoch": 0.2566907314227376, "grad_norm": 0.8931932491434317, "learning_rate": 4.675234175421222e-06, "loss": 0.2892, "num_tokens": 57609076.0, "step": 440 }, { "epoch": 0.2596076715525414, "grad_norm": 0.9216610473053193, "learning_rate": 4.663272404434635e-06, "loss": 0.2774, "num_tokens": 58264579.0, "step": 445 }, { "epoch": 0.2625246116823452, "grad_norm": 0.9202742139945425, "learning_rate": 4.6511120952578484e-06, "loss": 0.2863, "num_tokens": 58902959.0, "step": 450 }, { "epoch": 0.26544155181214907, "grad_norm": 0.8943558537794027, "learning_rate": 4.638754509752878e-06, "loss": 0.2796, "num_tokens": 59564439.0, "step": 455 }, { "epoch": 0.2683584919419529, "grad_norm": 0.863810937001429, "learning_rate": 4.6262009302528915e-06, "loss": 0.2843, "num_tokens": 60214896.0, "step": 460 }, { "epoch": 0.27127543207175675, "grad_norm": 0.9983134689445968, "learning_rate": 4.61345265942914e-06, "loss": 0.2866, "num_tokens": 60876642.0, "step": 465 }, { "epoch": 0.27419237220156056, "grad_norm": 0.8983030440670102, "learning_rate": 4.600511020155778e-06, "loss": 0.2848, "num_tokens": 61518584.0, "step": 470 }, { "epoch": 0.2771093123313644, "grad_norm": 0.8031669478146738, "learning_rate": 4.587377355372596e-06, "loss": 0.2757, "num_tokens": 62183326.0, "step": 475 }, { "epoch": 0.28002625246116825, "grad_norm": 0.9265028610272964, "learning_rate": 4.57405302794566e-06, "loss": 0.2841, "num_tokens": 62825255.0, "step": 480 }, { "epoch": 0.28294319259097206, "grad_norm": 0.7881945775359938, "learning_rate": 4.5605394205258915e-06, "loss": 0.2784, "num_tokens": 63488470.0, "step": 485 }, { "epoch": 0.28586013272077593, "grad_norm": 0.890455779305603, "learning_rate": 4.546837935405591e-06, "loss": 0.2825, "num_tokens": 64144973.0, "step": 490 }, { "epoch": 0.28877707285057974, "grad_norm": 0.8245358042376737, "learning_rate": 4.532949994372922e-06, "loss": 0.2735, "num_tokens": 64800071.0, "step": 495 }, { "epoch": 0.29169401298038355, "grad_norm": 0.8022563146785544, "learning_rate": 4.518877038564374e-06, "loss": 0.2805, "step": 500 }, { "epoch": 0.29169401298038355, "eval_loss": 0.2839949429035187, "eval_num_tokens": 65444398.0, "eval_runtime": 611.5469, "eval_samples_per_second": 9.442, "eval_steps_per_second": 1.181, "step": 500 }, { "epoch": 0.2946109531101874, "grad_norm": 0.8282011678135937, "learning_rate": 4.50462052831522e-06, "loss": 0.2742, "num_tokens": 66102293.0, "step": 505 }, { "epoch": 0.29752789323999124, "grad_norm": 0.8459868862384502, "learning_rate": 4.4901819430079766e-06, "loss": 0.2797, "num_tokens": 66750025.0, "step": 510 }, { "epoch": 0.3004448333697951, "grad_norm": 1.0044989964568452, "learning_rate": 4.4755627809188885e-06, "loss": 0.2808, "num_tokens": 67398138.0, "step": 515 }, { "epoch": 0.3033617734995989, "grad_norm": 0.8645633533810008, "learning_rate": 4.460764559062458e-06, "loss": 0.2743, "num_tokens": 68064298.0, "step": 520 }, { "epoch": 0.30627871362940273, "grad_norm": 0.7902367580597776, "learning_rate": 4.445788813034024e-06, "loss": 0.2745, "num_tokens": 68730271.0, "step": 525 }, { "epoch": 0.3091956537592066, "grad_norm": 0.8074282979202131, "learning_rate": 4.430637096850415e-06, "loss": 0.278, "num_tokens": 69373992.0, "step": 530 }, { "epoch": 0.3121125938890104, "grad_norm": 0.7997689062881244, "learning_rate": 4.4153109827886894e-06, "loss": 0.2773, "num_tokens": 70013275.0, "step": 535 }, { "epoch": 0.3150295340188143, "grad_norm": 0.7763977877420887, "learning_rate": 4.399812061222983e-06, "loss": 0.2792, "num_tokens": 70667022.0, "step": 540 }, { "epoch": 0.3179464741486181, "grad_norm": 0.8027330754403003, "learning_rate": 4.384141940459482e-06, "loss": 0.2692, "num_tokens": 71329905.0, "step": 545 }, { "epoch": 0.3208634142784219, "grad_norm": 0.842566387021002, "learning_rate": 4.36830224656952e-06, "loss": 0.2777, "num_tokens": 71981698.0, "step": 550 }, { "epoch": 0.3237803544082258, "grad_norm": 0.760795844133808, "learning_rate": 4.352294623220853e-06, "loss": 0.2719, "num_tokens": 72641200.0, "step": 555 }, { "epoch": 0.3266972945380296, "grad_norm": 0.767412158526903, "learning_rate": 4.336120731507095e-06, "loss": 0.2786, "num_tokens": 73289056.0, "step": 560 }, { "epoch": 0.32961423466783346, "grad_norm": 0.839424142267615, "learning_rate": 4.319782249775343e-06, "loss": 0.2774, "num_tokens": 73940084.0, "step": 565 }, { "epoch": 0.3325311747976373, "grad_norm": 0.7953209168681115, "learning_rate": 4.303280873452022e-06, "loss": 0.2761, "num_tokens": 74602418.0, "step": 570 }, { "epoch": 0.3354481149274411, "grad_norm": 0.9703712347237416, "learning_rate": 4.286618314866953e-06, "loss": 0.2739, "num_tokens": 75260268.0, "step": 575 }, { "epoch": 0.33836505505724496, "grad_norm": 0.785005018552605, "learning_rate": 4.269796303075664e-06, "loss": 0.2719, "num_tokens": 75911746.0, "step": 580 }, { "epoch": 0.34128199518704877, "grad_norm": 0.9895796846796822, "learning_rate": 4.252816583679963e-06, "loss": 0.2703, "num_tokens": 76570208.0, "step": 585 }, { "epoch": 0.34419893531685264, "grad_norm": 0.8038675047644025, "learning_rate": 4.235680918646814e-06, "loss": 0.2733, "num_tokens": 77231968.0, "step": 590 }, { "epoch": 0.34711587544665645, "grad_norm": 0.8232561538467604, "learning_rate": 4.21839108612548e-06, "loss": 0.2753, "num_tokens": 77888752.0, "step": 595 }, { "epoch": 0.35003281557646027, "grad_norm": 0.8114352965857334, "learning_rate": 4.200948880263021e-06, "loss": 0.2764, "num_tokens": 78549259.0, "step": 600 }, { "epoch": 0.35294975570626413, "grad_norm": 0.918026001844272, "learning_rate": 4.1833561110181114e-06, "loss": 0.2696, "num_tokens": 79207039.0, "step": 605 }, { "epoch": 0.35586669583606795, "grad_norm": 0.8799204301828087, "learning_rate": 4.165614603973225e-06, "loss": 0.2663, "num_tokens": 79861860.0, "step": 610 }, { "epoch": 0.3587836359658718, "grad_norm": 0.87043880731276, "learning_rate": 4.147726200145192e-06, "loss": 0.2755, "num_tokens": 80518900.0, "step": 615 }, { "epoch": 0.36170057609567563, "grad_norm": 0.9211963447162147, "learning_rate": 4.1296927557941625e-06, "loss": 0.2767, "num_tokens": 81161582.0, "step": 620 }, { "epoch": 0.3646175162254795, "grad_norm": 0.8299533305119133, "learning_rate": 4.111516142230986e-06, "loss": 0.2696, "num_tokens": 81813290.0, "step": 625 }, { "epoch": 0.3675344563552833, "grad_norm": 0.8309388002459559, "learning_rate": 4.093198245623022e-06, "loss": 0.2718, "num_tokens": 82481134.0, "step": 630 }, { "epoch": 0.3704513964850871, "grad_norm": 0.7974940185655467, "learning_rate": 4.074740966798417e-06, "loss": 0.275, "num_tokens": 83121589.0, "step": 635 }, { "epoch": 0.373368336614891, "grad_norm": 0.7877568112781695, "learning_rate": 4.056146221048861e-06, "loss": 0.2689, "num_tokens": 83783858.0, "step": 640 }, { "epoch": 0.3762852767446948, "grad_norm": 0.7457416133818777, "learning_rate": 4.037415937930834e-06, "loss": 0.2698, "num_tokens": 84439289.0, "step": 645 }, { "epoch": 0.3792022168744987, "grad_norm": 0.804387030534503, "learning_rate": 4.018552061065385e-06, "loss": 0.2656, "num_tokens": 85096010.0, "step": 650 }, { "epoch": 0.3821191570043025, "grad_norm": 0.7866414763779639, "learning_rate": 3.999556547936433e-06, "loss": 0.2679, "num_tokens": 85763093.0, "step": 655 }, { "epoch": 0.3850360971341063, "grad_norm": 0.7975246358334895, "learning_rate": 3.980431369687657e-06, "loss": 0.2685, "num_tokens": 86414815.0, "step": 660 }, { "epoch": 0.38795303726391017, "grad_norm": 0.7541579622412656, "learning_rate": 3.961178510917938e-06, "loss": 0.2664, "num_tokens": 87072103.0, "step": 665 }, { "epoch": 0.390869977393714, "grad_norm": 0.8462615544123323, "learning_rate": 3.941799969475426e-06, "loss": 0.2697, "num_tokens": 87724238.0, "step": 670 }, { "epoch": 0.39378691752351785, "grad_norm": 0.719877168553545, "learning_rate": 3.922297756250231e-06, "loss": 0.2682, "num_tokens": 88385611.0, "step": 675 }, { "epoch": 0.39670385765332167, "grad_norm": 0.8061212707580667, "learning_rate": 3.902673894965739e-06, "loss": 0.2652, "num_tokens": 89047105.0, "step": 680 }, { "epoch": 0.3996207977831255, "grad_norm": 0.8326055553096022, "learning_rate": 3.88293042196863e-06, "loss": 0.2651, "num_tokens": 89703947.0, "step": 685 }, { "epoch": 0.40253773791292935, "grad_norm": 0.8409291362121083, "learning_rate": 3.863069386017559e-06, "loss": 0.2676, "num_tokens": 90375478.0, "step": 690 }, { "epoch": 0.40545467804273316, "grad_norm": 1.1749044956975412, "learning_rate": 3.8430928480705595e-06, "loss": 0.268, "num_tokens": 91041509.0, "step": 695 }, { "epoch": 0.40837161817253703, "grad_norm": 0.9615687348415685, "learning_rate": 3.823002881071182e-06, "loss": 0.2713, "num_tokens": 91687478.0, "step": 700 }, { "epoch": 0.41128855830234085, "grad_norm": 0.8598987130093434, "learning_rate": 3.802801569733385e-06, "loss": 0.2633, "num_tokens": 92350865.0, "step": 705 }, { "epoch": 0.41420549843214466, "grad_norm": 0.9022414410706127, "learning_rate": 3.7824910103252094e-06, "loss": 0.2705, "num_tokens": 93001864.0, "step": 710 }, { "epoch": 0.4171224385619485, "grad_norm": 0.8563602736920568, "learning_rate": 3.7620733104512457e-06, "loss": 0.2686, "num_tokens": 93646547.0, "step": 715 }, { "epoch": 0.42003937869175234, "grad_norm": 0.8403741919990089, "learning_rate": 3.741550588833938e-06, "loss": 0.261, "num_tokens": 94309272.0, "step": 720 }, { "epoch": 0.4229563188215562, "grad_norm": 0.8382205734712479, "learning_rate": 3.7209249750937194e-06, "loss": 0.2583, "num_tokens": 94974415.0, "step": 725 }, { "epoch": 0.42587325895136, "grad_norm": 0.7670934985699519, "learning_rate": 3.700198609528027e-06, "loss": 0.2621, "num_tokens": 95637473.0, "step": 730 }, { "epoch": 0.42879019908116384, "grad_norm": 0.889958478839124, "learning_rate": 3.679373642889205e-06, "loss": 0.263, "num_tokens": 96285419.0, "step": 735 }, { "epoch": 0.4317071392109677, "grad_norm": 0.8154007725747381, "learning_rate": 3.6584522361613227e-06, "loss": 0.2657, "num_tokens": 96926138.0, "step": 740 }, { "epoch": 0.4346240793407715, "grad_norm": 0.8307773360486326, "learning_rate": 3.6374365603359347e-06, "loss": 0.2662, "num_tokens": 97588450.0, "step": 745 }, { "epoch": 0.4375410194705754, "grad_norm": 0.8078529106621964, "learning_rate": 3.616328796186794e-06, "loss": 0.2681, "num_tokens": 98242023.0, "step": 750 }, { "epoch": 0.4404579596003792, "grad_norm": 0.9983558574344977, "learning_rate": 3.5951311340435597e-06, "loss": 0.2637, "num_tokens": 98902913.0, "step": 755 }, { "epoch": 0.443374899730183, "grad_norm": 0.9165975707045902, "learning_rate": 3.573845773564506e-06, "loss": 0.2727, "num_tokens": 99533885.0, "step": 760 }, { "epoch": 0.4462918398599869, "grad_norm": 0.7822942999479296, "learning_rate": 3.5524749235082728e-06, "loss": 0.2712, "num_tokens": 100188467.0, "step": 765 }, { "epoch": 0.4492087799897907, "grad_norm": 0.7478178653016506, "learning_rate": 3.5310208015046547e-06, "loss": 0.2661, "num_tokens": 100842626.0, "step": 770 }, { "epoch": 0.45212572011959457, "grad_norm": 0.8348739730561777, "learning_rate": 3.5094856338244882e-06, "loss": 0.2605, "num_tokens": 101509817.0, "step": 775 }, { "epoch": 0.4550426602493984, "grad_norm": 0.8650672549166954, "learning_rate": 3.4878716551486296e-06, "loss": 0.2626, "num_tokens": 102161268.0, "step": 780 }, { "epoch": 0.4579596003792022, "grad_norm": 0.8437538579535597, "learning_rate": 3.466181108336068e-06, "loss": 0.2631, "num_tokens": 102815935.0, "step": 785 }, { "epoch": 0.46087654050900606, "grad_norm": 0.8055131749035074, "learning_rate": 3.444416244191184e-06, "loss": 0.2548, "num_tokens": 103480501.0, "step": 790 }, { "epoch": 0.4637934806388099, "grad_norm": 0.7184736639862396, "learning_rate": 3.422579321230185e-06, "loss": 0.2615, "num_tokens": 104127953.0, "step": 795 }, { "epoch": 0.46671042076861374, "grad_norm": 0.7769160728536546, "learning_rate": 3.400672605446746e-06, "loss": 0.2649, "num_tokens": 104784029.0, "step": 800 }, { "epoch": 0.46962736089841756, "grad_norm": 0.7956791364511919, "learning_rate": 3.378698370076865e-06, "loss": 0.2585, "num_tokens": 105448623.0, "step": 805 }, { "epoch": 0.47254430102822137, "grad_norm": 0.7971292848423556, "learning_rate": 3.356658895362974e-06, "loss": 0.258, "num_tokens": 106102919.0, "step": 810 }, { "epoch": 0.47546124115802524, "grad_norm": 0.8024065230344523, "learning_rate": 3.334556468317322e-06, "loss": 0.2639, "num_tokens": 106766059.0, "step": 815 }, { "epoch": 0.47837818128782905, "grad_norm": 0.8627952778096136, "learning_rate": 3.3123933824846517e-06, "loss": 0.2637, "num_tokens": 107426299.0, "step": 820 }, { "epoch": 0.4812951214176329, "grad_norm": 0.8393622591727319, "learning_rate": 3.2901719377042003e-06, "loss": 0.2629, "num_tokens": 108065870.0, "step": 825 }, { "epoch": 0.48421206154743673, "grad_norm": 0.7495915565490366, "learning_rate": 3.2678944398710535e-06, "loss": 0.2624, "num_tokens": 108729197.0, "step": 830 }, { "epoch": 0.48712900167724055, "grad_norm": 0.7951127772717166, "learning_rate": 3.2455632006968564e-06, "loss": 0.2644, "num_tokens": 109377885.0, "step": 835 }, { "epoch": 0.4900459418070444, "grad_norm": 0.8730539703385664, "learning_rate": 3.223180537469938e-06, "loss": 0.2631, "num_tokens": 110046067.0, "step": 840 }, { "epoch": 0.49296288193684823, "grad_norm": 0.8086162516026355, "learning_rate": 3.200748772814844e-06, "loss": 0.2529, "num_tokens": 110715109.0, "step": 845 }, { "epoch": 0.4958798220666521, "grad_norm": 0.8485834526237742, "learning_rate": 3.1782702344513215e-06, "loss": 0.2607, "num_tokens": 111368579.0, "step": 850 }, { "epoch": 0.4987967621964559, "grad_norm": 0.79225012559078, "learning_rate": 3.1557472549527747e-06, "loss": 0.2618, "num_tokens": 112019671.0, "step": 855 }, { "epoch": 0.5017137023262598, "grad_norm": 0.8425810939639716, "learning_rate": 3.133182171504214e-06, "loss": 0.2627, "num_tokens": 112674101.0, "step": 860 }, { "epoch": 0.5046306424560636, "grad_norm": 0.744964617174656, "learning_rate": 3.110577325659734e-06, "loss": 0.2583, "num_tokens": 113327455.0, "step": 865 }, { "epoch": 0.5075475825858674, "grad_norm": 0.7709087869015558, "learning_rate": 3.0879350630995284e-06, "loss": 0.2631, "num_tokens": 113978391.0, "step": 870 }, { "epoch": 0.5104645227156712, "grad_norm": 0.8044798860684248, "learning_rate": 3.0652577333864812e-06, "loss": 0.2569, "num_tokens": 114620393.0, "step": 875 }, { "epoch": 0.5133814628454751, "grad_norm": 0.7842031411597096, "learning_rate": 3.0425476897223584e-06, "loss": 0.2611, "num_tokens": 115280804.0, "step": 880 }, { "epoch": 0.516298402975279, "grad_norm": 0.8035683788722667, "learning_rate": 3.019807288703615e-06, "loss": 0.2553, "num_tokens": 115935583.0, "step": 885 }, { "epoch": 0.5192153431050828, "grad_norm": 0.7544056798614338, "learning_rate": 2.99703889007686e-06, "loss": 0.2563, "num_tokens": 116605162.0, "step": 890 }, { "epoch": 0.5221322832348866, "grad_norm": 0.7470139344901265, "learning_rate": 2.9742448564939785e-06, "loss": 0.2566, "num_tokens": 117256010.0, "step": 895 }, { "epoch": 0.5250492233646904, "grad_norm": 0.7773494324524246, "learning_rate": 2.95142755326697e-06, "loss": 0.2557, "num_tokens": 117920457.0, "step": 900 }, { "epoch": 0.5279661634944943, "grad_norm": 0.908172722597612, "learning_rate": 2.9285893481224976e-06, "loss": 0.2548, "num_tokens": 118587118.0, "step": 905 }, { "epoch": 0.5308831036242981, "grad_norm": 0.7475823569135418, "learning_rate": 2.9057326109561955e-06, "loss": 0.2613, "num_tokens": 119238394.0, "step": 910 }, { "epoch": 0.533800043754102, "grad_norm": 0.7910527226931955, "learning_rate": 2.8828597135867446e-06, "loss": 0.2568, "num_tokens": 119902185.0, "step": 915 }, { "epoch": 0.5367169838839058, "grad_norm": 0.7067850895490586, "learning_rate": 2.859973029509753e-06, "loss": 0.2514, "num_tokens": 120578709.0, "step": 920 }, { "epoch": 0.5396339240137096, "grad_norm": 0.7318968008202511, "learning_rate": 2.83707493365146e-06, "loss": 0.2564, "num_tokens": 121237827.0, "step": 925 }, { "epoch": 0.5425508641435135, "grad_norm": 0.7705210064312026, "learning_rate": 2.8141678021222933e-06, "loss": 0.2525, "num_tokens": 121898853.0, "step": 930 }, { "epoch": 0.5454678042733173, "grad_norm": 0.8157907159758865, "learning_rate": 2.791254011970301e-06, "loss": 0.2558, "num_tokens": 122562375.0, "step": 935 }, { "epoch": 0.5483847444031211, "grad_norm": 0.8681587026037888, "learning_rate": 2.7683359409344905e-06, "loss": 0.2602, "num_tokens": 123211114.0, "step": 940 }, { "epoch": 0.5513016845329249, "grad_norm": 0.7556332005656767, "learning_rate": 2.745415967198093e-06, "loss": 0.254, "num_tokens": 123865098.0, "step": 945 }, { "epoch": 0.5542186246627288, "grad_norm": 0.7946261128648097, "learning_rate": 2.722496469141779e-06, "loss": 0.2586, "num_tokens": 124518258.0, "step": 950 }, { "epoch": 0.5571355647925327, "grad_norm": 0.7105424221615, "learning_rate": 2.699579825096857e-06, "loss": 0.2558, "num_tokens": 125177330.0, "step": 955 }, { "epoch": 0.5600525049223365, "grad_norm": 0.8047130073772089, "learning_rate": 2.6766684130984818e-06, "loss": 0.2518, "num_tokens": 125844180.0, "step": 960 }, { "epoch": 0.5629694450521403, "grad_norm": 0.9021937489211278, "learning_rate": 2.653764610638881e-06, "loss": 0.2622, "num_tokens": 126493064.0, "step": 965 }, { "epoch": 0.5658863851819441, "grad_norm": 0.8477417229161418, "learning_rate": 2.6308707944206487e-06, "loss": 0.256, "num_tokens": 127140885.0, "step": 970 }, { "epoch": 0.5688033253117479, "grad_norm": 0.707517618098058, "learning_rate": 2.607989340110121e-06, "loss": 0.247, "num_tokens": 127799249.0, "step": 975 }, { "epoch": 0.5717202654415519, "grad_norm": 0.7587943404137327, "learning_rate": 2.5851226220908504e-06, "loss": 0.2572, "num_tokens": 128456895.0, "step": 980 }, { "epoch": 0.5746372055713557, "grad_norm": 0.7584809790016415, "learning_rate": 2.562273013217218e-06, "loss": 0.2497, "num_tokens": 129122383.0, "step": 985 }, { "epoch": 0.5775541457011595, "grad_norm": 0.8214288920378848, "learning_rate": 2.539442884568211e-06, "loss": 0.2541, "num_tokens": 129779846.0, "step": 990 }, { "epoch": 0.5804710858309633, "grad_norm": 0.7290259201155611, "learning_rate": 2.5166346052013734e-06, "loss": 0.257, "num_tokens": 130425019.0, "step": 995 }, { "epoch": 0.5833880259607671, "grad_norm": 0.8673297271017076, "learning_rate": 2.4938505419069737e-06, "loss": 0.2524, "step": 1000 }, { "epoch": 0.5833880259607671, "eval_loss": 0.26100972294807434, "eval_num_tokens": 131094159.0, "eval_runtime": 607.9741, "eval_samples_per_second": 9.497, "eval_steps_per_second": 1.188, "step": 1000 }, { "epoch": 0.586304966090571, "grad_norm": 0.9024948847675129, "learning_rate": 2.4710930589624043e-06, "loss": 0.2543, "num_tokens": 131753695.0, "step": 1005 }, { "epoch": 0.5892219062203748, "grad_norm": 0.8681506257906868, "learning_rate": 2.4483645178868436e-06, "loss": 0.2554, "num_tokens": 132413825.0, "step": 1010 }, { "epoch": 0.5921388463501787, "grad_norm": 0.7962179648149258, "learning_rate": 2.425667277196205e-06, "loss": 0.2548, "num_tokens": 133068552.0, "step": 1015 }, { "epoch": 0.5950557864799825, "grad_norm": 0.804654580141575, "learning_rate": 2.4030036921583934e-06, "loss": 0.2564, "num_tokens": 133717600.0, "step": 1020 }, { "epoch": 0.5979727266097863, "grad_norm": 0.7346044185339496, "learning_rate": 2.380376114548905e-06, "loss": 0.2555, "num_tokens": 134372291.0, "step": 1025 }, { "epoch": 0.6008896667395902, "grad_norm": 0.8118188644094798, "learning_rate": 2.3577868924067838e-06, "loss": 0.2577, "num_tokens": 135017558.0, "step": 1030 }, { "epoch": 0.603806606869394, "grad_norm": 0.7221021652672162, "learning_rate": 2.3352383697909685e-06, "loss": 0.256, "num_tokens": 135666125.0, "step": 1035 }, { "epoch": 0.6067235469991978, "grad_norm": 0.7352033663094344, "learning_rate": 2.312732886537052e-06, "loss": 0.2575, "num_tokens": 136315008.0, "step": 1040 }, { "epoch": 0.6096404871290017, "grad_norm": 0.8607128532349083, "learning_rate": 2.29027277801448e-06, "loss": 0.2595, "num_tokens": 136959663.0, "step": 1045 }, { "epoch": 0.6125574272588055, "grad_norm": 0.7969656435537161, "learning_rate": 2.267860374884213e-06, "loss": 0.2541, "num_tokens": 137610280.0, "step": 1050 }, { "epoch": 0.6154743673886094, "grad_norm": 0.7672936399726409, "learning_rate": 2.245498002856874e-06, "loss": 0.2615, "num_tokens": 138253534.0, "step": 1055 }, { "epoch": 0.6183913075184132, "grad_norm": 0.8116002548995157, "learning_rate": 2.2231879824514114e-06, "loss": 0.2514, "num_tokens": 138918946.0, "step": 1060 }, { "epoch": 0.621308247648217, "grad_norm": 0.8343062523034186, "learning_rate": 2.2009326287543046e-06, "loss": 0.2549, "num_tokens": 139575919.0, "step": 1065 }, { "epoch": 0.6242251877780208, "grad_norm": 0.7030831591540176, "learning_rate": 2.1787342511793303e-06, "loss": 0.2543, "num_tokens": 140246394.0, "step": 1070 }, { "epoch": 0.6271421279078246, "grad_norm": 0.7676168932953417, "learning_rate": 2.156595153227911e-06, "loss": 0.2522, "num_tokens": 140908500.0, "step": 1075 }, { "epoch": 0.6300590680376286, "grad_norm": 0.7565551405042247, "learning_rate": 2.13451763225009e-06, "loss": 0.2538, "num_tokens": 141560588.0, "step": 1080 }, { "epoch": 0.6329760081674324, "grad_norm": 0.7523126517976036, "learning_rate": 2.1125039792061346e-06, "loss": 0.2507, "num_tokens": 142230913.0, "step": 1085 }, { "epoch": 0.6358929482972362, "grad_norm": 0.7547351807576873, "learning_rate": 2.0905564784288064e-06, "loss": 0.252, "num_tokens": 142880248.0, "step": 1090 }, { "epoch": 0.63880988842704, "grad_norm": 0.763981058542318, "learning_rate": 2.0686774073863183e-06, "loss": 0.2508, "num_tokens": 143538441.0, "step": 1095 }, { "epoch": 0.6417268285568438, "grad_norm": 0.7513777028630977, "learning_rate": 2.0468690364460032e-06, "loss": 0.2499, "num_tokens": 144195649.0, "step": 1100 }, { "epoch": 0.6446437686866477, "grad_norm": 0.6978388439562043, "learning_rate": 2.0251336286387246e-06, "loss": 0.2539, "num_tokens": 144840414.0, "step": 1105 }, { "epoch": 0.6475607088164516, "grad_norm": 0.7122399335583397, "learning_rate": 2.003473439424037e-06, "loss": 0.2516, "num_tokens": 145497670.0, "step": 1110 }, { "epoch": 0.6504776489462554, "grad_norm": 0.7662075027673452, "learning_rate": 1.9818907164561474e-06, "loss": 0.2508, "num_tokens": 146167309.0, "step": 1115 }, { "epoch": 0.6533945890760592, "grad_norm": 0.7079059380454418, "learning_rate": 1.960387699350673e-06, "loss": 0.2471, "num_tokens": 146836924.0, "step": 1120 }, { "epoch": 0.656311529205863, "grad_norm": 0.6995228988412967, "learning_rate": 1.9389666194522416e-06, "loss": 0.2477, "num_tokens": 147496049.0, "step": 1125 }, { "epoch": 0.6592284693356669, "grad_norm": 0.7018077763682937, "learning_rate": 1.9176296996029455e-06, "loss": 0.2505, "num_tokens": 148149073.0, "step": 1130 }, { "epoch": 0.6621454094654707, "grad_norm": 0.6955137369821246, "learning_rate": 1.8963791539116794e-06, "loss": 0.2589, "num_tokens": 148811679.0, "step": 1135 }, { "epoch": 0.6650623495952745, "grad_norm": 0.7175039810377445, "learning_rate": 1.8752171875243897e-06, "loss": 0.254, "num_tokens": 149453909.0, "step": 1140 }, { "epoch": 0.6679792897250784, "grad_norm": 0.7540266711347994, "learning_rate": 1.8541459963952401e-06, "loss": 0.2534, "num_tokens": 150113931.0, "step": 1145 }, { "epoch": 0.6708962298548822, "grad_norm": 0.6958604937899281, "learning_rate": 1.8331677670587489e-06, "loss": 0.2472, "num_tokens": 150775568.0, "step": 1150 }, { "epoch": 0.6738131699846861, "grad_norm": 0.7533171707259878, "learning_rate": 1.812284676402889e-06, "loss": 0.2532, "num_tokens": 151432991.0, "step": 1155 }, { "epoch": 0.6767301101144899, "grad_norm": 0.7169328151307629, "learning_rate": 1.7914988914431958e-06, "loss": 0.2515, "num_tokens": 152071671.0, "step": 1160 }, { "epoch": 0.6796470502442937, "grad_norm": 0.6814651782340072, "learning_rate": 1.7708125690978973e-06, "loss": 0.2577, "num_tokens": 152727004.0, "step": 1165 }, { "epoch": 0.6825639903740975, "grad_norm": 0.7553012186400901, "learning_rate": 1.7502278559641e-06, "loss": 0.246, "num_tokens": 153403038.0, "step": 1170 }, { "epoch": 0.6854809305039014, "grad_norm": 0.7916967627609816, "learning_rate": 1.7297468880950275e-06, "loss": 0.2456, "num_tokens": 154050570.0, "step": 1175 }, { "epoch": 0.6883978706337053, "grad_norm": 0.7570768314074023, "learning_rate": 1.7093717907783725e-06, "loss": 0.248, "num_tokens": 154707732.0, "step": 1180 }, { "epoch": 0.6913148107635091, "grad_norm": 0.7403254580619876, "learning_rate": 1.6891046783157577e-06, "loss": 0.2506, "num_tokens": 155357884.0, "step": 1185 }, { "epoch": 0.6942317508933129, "grad_norm": 0.7862768452334575, "learning_rate": 1.668947653803332e-06, "loss": 0.2478, "num_tokens": 156012053.0, "step": 1190 }, { "epoch": 0.6971486910231167, "grad_norm": 0.6944065718383872, "learning_rate": 1.6489028089135412e-06, "loss": 0.2403, "num_tokens": 156675319.0, "step": 1195 }, { "epoch": 0.7000656311529205, "grad_norm": 0.689978749467197, "learning_rate": 1.6289722236780708e-06, "loss": 0.2443, "num_tokens": 157341110.0, "step": 1200 }, { "epoch": 0.7029825712827245, "grad_norm": 0.7401973229886903, "learning_rate": 1.6091579662720085e-06, "loss": 0.2468, "num_tokens": 157999131.0, "step": 1205 }, { "epoch": 0.7058995114125283, "grad_norm": 0.7320601106731225, "learning_rate": 1.5894620927992305e-06, "loss": 0.2523, "num_tokens": 158639604.0, "step": 1210 }, { "epoch": 0.7088164515423321, "grad_norm": 0.7143585333776589, "learning_rate": 1.5698866470790408e-06, "loss": 0.2511, "num_tokens": 159287175.0, "step": 1215 }, { "epoch": 0.7117333916721359, "grad_norm": 0.805040758465756, "learning_rate": 1.5504336604340859e-06, "loss": 0.2527, "num_tokens": 159928090.0, "step": 1220 }, { "epoch": 0.7146503318019398, "grad_norm": 0.7525019096597073, "learning_rate": 1.5311051514795689e-06, "loss": 0.2524, "num_tokens": 160587939.0, "step": 1225 }, { "epoch": 0.7175672719317436, "grad_norm": 0.7709357498772824, "learning_rate": 1.5119031259137786e-06, "loss": 0.2495, "num_tokens": 161239965.0, "step": 1230 }, { "epoch": 0.7204842120615474, "grad_norm": 0.6930775959107991, "learning_rate": 1.4928295763099595e-06, "loss": 0.2471, "num_tokens": 161899763.0, "step": 1235 }, { "epoch": 0.7234011521913513, "grad_norm": 0.7146451766584444, "learning_rate": 1.4738864819095478e-06, "loss": 0.2468, "num_tokens": 162563978.0, "step": 1240 }, { "epoch": 0.7263180923211551, "grad_norm": 0.7233964208181292, "learning_rate": 1.4550758084167823e-06, "loss": 0.2468, "num_tokens": 163215392.0, "step": 1245 }, { "epoch": 0.729235032450959, "grad_norm": 0.7535387485017295, "learning_rate": 1.4363995077947318e-06, "loss": 0.2478, "num_tokens": 163876128.0, "step": 1250 }, { "epoch": 0.7321519725807628, "grad_norm": 0.710154633510974, "learning_rate": 1.417859518062738e-06, "loss": 0.2471, "num_tokens": 164536666.0, "step": 1255 }, { "epoch": 0.7350689127105666, "grad_norm": 0.7793325296651725, "learning_rate": 1.3994577630953085e-06, "loss": 0.2465, "num_tokens": 165186675.0, "step": 1260 }, { "epoch": 0.7379858528403704, "grad_norm": 0.7204453905441239, "learning_rate": 1.3811961524224838e-06, "loss": 0.2453, "num_tokens": 165828836.0, "step": 1265 }, { "epoch": 0.7409027929701743, "grad_norm": 0.6875863992968856, "learning_rate": 1.3630765810316799e-06, "loss": 0.2492, "num_tokens": 166480185.0, "step": 1270 }, { "epoch": 0.7438197330999782, "grad_norm": 0.7399219862362597, "learning_rate": 1.3451009291710542e-06, "loss": 0.2509, "num_tokens": 167127201.0, "step": 1275 }, { "epoch": 0.746736673229782, "grad_norm": 0.7264134555452614, "learning_rate": 1.3272710621543892e-06, "loss": 0.2488, "num_tokens": 167776261.0, "step": 1280 }, { "epoch": 0.7496536133595858, "grad_norm": 0.7732152888510079, "learning_rate": 1.309588830167536e-06, "loss": 0.2449, "num_tokens": 168426328.0, "step": 1285 }, { "epoch": 0.7525705534893896, "grad_norm": 0.7073729535659216, "learning_rate": 1.2920560680764165e-06, "loss": 0.2421, "num_tokens": 169092019.0, "step": 1290 }, { "epoch": 0.7554874936191934, "grad_norm": 0.7331560167444491, "learning_rate": 1.2746745952366275e-06, "loss": 0.2487, "num_tokens": 169741583.0, "step": 1295 }, { "epoch": 0.7584044337489974, "grad_norm": 0.6884922406400141, "learning_rate": 1.2574462153046441e-06, "loss": 0.2412, "num_tokens": 170389299.0, "step": 1300 }, { "epoch": 0.7613213738788012, "grad_norm": 0.781793314708951, "learning_rate": 1.2403727160506559e-06, "loss": 0.2444, "num_tokens": 171042211.0, "step": 1305 }, { "epoch": 0.764238314008605, "grad_norm": 0.7250494897320526, "learning_rate": 1.223455869173056e-06, "loss": 0.2438, "num_tokens": 171701780.0, "step": 1310 }, { "epoch": 0.7671552541384088, "grad_norm": 0.8740645184327206, "learning_rate": 1.2066974301145894e-06, "loss": 0.2456, "num_tokens": 172376337.0, "step": 1315 }, { "epoch": 0.7700721942682126, "grad_norm": 0.7443315304727108, "learning_rate": 1.1900991378801964e-06, "loss": 0.2439, "num_tokens": 173041371.0, "step": 1320 }, { "epoch": 0.7729891343980165, "grad_norm": 0.7116423082671238, "learning_rate": 1.1736627148565535e-06, "loss": 0.2531, "num_tokens": 173682308.0, "step": 1325 }, { "epoch": 0.7759060745278203, "grad_norm": 0.7077335089669503, "learning_rate": 1.157389866633348e-06, "loss": 0.246, "num_tokens": 174330336.0, "step": 1330 }, { "epoch": 0.7788230146576242, "grad_norm": 0.7308022672326984, "learning_rate": 1.1412822818262878e-06, "loss": 0.2524, "num_tokens": 174969619.0, "step": 1335 }, { "epoch": 0.781739954787428, "grad_norm": 0.6995063282185964, "learning_rate": 1.125341631901876e-06, "loss": 0.2406, "num_tokens": 175634291.0, "step": 1340 }, { "epoch": 0.7846568949172318, "grad_norm": 0.7514971805090548, "learning_rate": 1.1095695710039664e-06, "loss": 0.2403, "num_tokens": 176302019.0, "step": 1345 }, { "epoch": 0.7875738350470357, "grad_norm": 0.6772018040657682, "learning_rate": 1.0939677357821108e-06, "loss": 0.2445, "num_tokens": 176976275.0, "step": 1350 }, { "epoch": 0.7904907751768395, "grad_norm": 0.6817675548735839, "learning_rate": 1.0785377452217311e-06, "loss": 0.2503, "num_tokens": 177628328.0, "step": 1355 }, { "epoch": 0.7934077153066433, "grad_norm": 0.7346247121033264, "learning_rate": 1.0632812004761151e-06, "loss": 0.243, "num_tokens": 178288873.0, "step": 1360 }, { "epoch": 0.7963246554364471, "grad_norm": 0.7311842645346187, "learning_rate": 1.0481996847002676e-06, "loss": 0.2504, "num_tokens": 178942090.0, "step": 1365 }, { "epoch": 0.799241595566251, "grad_norm": 0.731394943481863, "learning_rate": 1.0332947628866273e-06, "loss": 0.2428, "num_tokens": 179600176.0, "step": 1370 }, { "epoch": 0.8021585356960549, "grad_norm": 0.7100948172511871, "learning_rate": 1.0185679817026715e-06, "loss": 0.2511, "num_tokens": 180244674.0, "step": 1375 }, { "epoch": 0.8050754758258587, "grad_norm": 0.708242217717899, "learning_rate": 1.0040208693304183e-06, "loss": 0.2456, "num_tokens": 180901761.0, "step": 1380 }, { "epoch": 0.8079924159556625, "grad_norm": 0.6803617534010951, "learning_rate": 9.89654935307848e-07, "loss": 0.2393, "num_tokens": 181564639.0, "step": 1385 }, { "epoch": 0.8109093560854663, "grad_norm": 0.6958034065313294, "learning_rate": 9.754716703722635e-07, "loss": 0.2347, "num_tokens": 182224730.0, "step": 1390 }, { "epoch": 0.8138262962152701, "grad_norm": 0.7165262095609373, "learning_rate": 9.614725463055931e-07, "loss": 0.2461, "num_tokens": 182883552.0, "step": 1395 }, { "epoch": 0.8167432363450741, "grad_norm": 0.7138480327038658, "learning_rate": 9.476590157816701e-07, "loss": 0.2411, "num_tokens": 183537453.0, "step": 1400 }, { "epoch": 0.8196601764748779, "grad_norm": 0.7133912213063753, "learning_rate": 9.340325122154878e-07, "loss": 0.2462, "num_tokens": 184199270.0, "step": 1405 }, { "epoch": 0.8225771166046817, "grad_norm": 0.6947806259213369, "learning_rate": 9.205944496144556e-07, "loss": 0.2486, "num_tokens": 184853282.0, "step": 1410 }, { "epoch": 0.8254940567344855, "grad_norm": 0.7272223644551763, "learning_rate": 9.073462224316707e-07, "loss": 0.2447, "num_tokens": 185514281.0, "step": 1415 }, { "epoch": 0.8284109968642893, "grad_norm": 0.7072312165887826, "learning_rate": 8.942892054212143e-07, "loss": 0.2397, "num_tokens": 186169820.0, "step": 1420 }, { "epoch": 0.8313279369940932, "grad_norm": 0.6903355667912614, "learning_rate": 8.814247534954983e-07, "loss": 0.2497, "num_tokens": 186817556.0, "step": 1425 }, { "epoch": 0.834244877123897, "grad_norm": 0.706857521915122, "learning_rate": 8.687542015846639e-07, "loss": 0.2438, "num_tokens": 187469631.0, "step": 1430 }, { "epoch": 0.8371618172537009, "grad_norm": 0.6988018386149026, "learning_rate": 8.562788644980624e-07, "loss": 0.2455, "num_tokens": 188113842.0, "step": 1435 }, { "epoch": 0.8400787573835047, "grad_norm": 0.673516286108323, "learning_rate": 8.440000367878115e-07, "loss": 0.2396, "num_tokens": 188782137.0, "step": 1440 }, { "epoch": 0.8429956975133085, "grad_norm": 0.699489145828766, "learning_rate": 8.319189926144688e-07, "loss": 0.2511, "num_tokens": 189419614.0, "step": 1445 }, { "epoch": 0.8459126376431124, "grad_norm": 0.7084426008074721, "learning_rate": 8.200369856148089e-07, "loss": 0.2457, "num_tokens": 190076342.0, "step": 1450 }, { "epoch": 0.8488295777729162, "grad_norm": 0.6739486565962028, "learning_rate": 8.083552487717358e-07, "loss": 0.2457, "num_tokens": 190726569.0, "step": 1455 }, { "epoch": 0.85174651790272, "grad_norm": 0.7095407182813062, "learning_rate": 7.968749942863385e-07, "loss": 0.2435, "num_tokens": 191375804.0, "step": 1460 }, { "epoch": 0.8546634580325239, "grad_norm": 0.6710001579802304, "learning_rate": 7.855974134520999e-07, "loss": 0.2405, "num_tokens": 192039149.0, "step": 1465 }, { "epoch": 0.8575803981623277, "grad_norm": 0.7079688435936735, "learning_rate": 7.745236765312819e-07, "loss": 0.2411, "num_tokens": 192695832.0, "step": 1470 }, { "epoch": 0.8604973382921316, "grad_norm": 0.6854485439890876, "learning_rate": 7.636549326334825e-07, "loss": 0.2417, "num_tokens": 193367048.0, "step": 1475 }, { "epoch": 0.8634142784219354, "grad_norm": 0.7184112786493275, "learning_rate": 7.529923095963999e-07, "loss": 0.2441, "num_tokens": 194017542.0, "step": 1480 }, { "epoch": 0.8663312185517392, "grad_norm": 0.6651542917903988, "learning_rate": 7.425369138687957e-07, "loss": 0.2422, "num_tokens": 194667338.0, "step": 1485 }, { "epoch": 0.869248158681543, "grad_norm": 0.6648797188770686, "learning_rate": 7.322898303956773e-07, "loss": 0.2416, "num_tokens": 195324822.0, "step": 1490 }, { "epoch": 0.8721650988113469, "grad_norm": 0.6743904697311516, "learning_rate": 7.222521225057187e-07, "loss": 0.2481, "num_tokens": 195972844.0, "step": 1495 }, { "epoch": 0.8750820389411508, "grad_norm": 0.6862148679988234, "learning_rate": 7.124248318009164e-07, "loss": 0.2426, "step": 1500 }, { "epoch": 0.8750820389411508, "eval_loss": 0.2491583675146103, "eval_num_tokens": 196624476.0, "eval_runtime": 607.7377, "eval_samples_per_second": 9.501, "eval_steps_per_second": 1.188, "step": 1500 }, { "epoch": 0.8779989790709546, "grad_norm": 0.7150785623554777, "learning_rate": 7.028089780485081e-07, "loss": 0.2437, "num_tokens": 197287303.0, "step": 1505 }, { "epoch": 0.8809159192007584, "grad_norm": 0.6922051228380665, "learning_rate": 6.934055590751461e-07, "loss": 0.237, "num_tokens": 197950146.0, "step": 1510 }, { "epoch": 0.8838328593305622, "grad_norm": 0.6814419896511972, "learning_rate": 6.842155506633598e-07, "loss": 0.2429, "num_tokens": 198606592.0, "step": 1515 }, { "epoch": 0.886749799460366, "grad_norm": 0.7528574216832195, "learning_rate": 6.752399064502959e-07, "loss": 0.2374, "num_tokens": 199260466.0, "step": 1520 }, { "epoch": 0.88966673959017, "grad_norm": 0.743850816504643, "learning_rate": 6.664795578287632e-07, "loss": 0.2498, "num_tokens": 199915169.0, "step": 1525 }, { "epoch": 0.8925836797199738, "grad_norm": 0.6659899742478199, "learning_rate": 6.579354138505817e-07, "loss": 0.239, "num_tokens": 200577384.0, "step": 1530 }, { "epoch": 0.8955006198497776, "grad_norm": 0.6989537096567529, "learning_rate": 6.496083611322503e-07, "loss": 0.2437, "num_tokens": 201229904.0, "step": 1535 }, { "epoch": 0.8984175599795814, "grad_norm": 0.6870906534872678, "learning_rate": 6.414992637629462e-07, "loss": 0.2459, "num_tokens": 201874398.0, "step": 1540 }, { "epoch": 0.9013345001093852, "grad_norm": 0.6851022314706521, "learning_rate": 6.336089632148566e-07, "loss": 0.2437, "num_tokens": 202514986.0, "step": 1545 }, { "epoch": 0.9042514402391891, "grad_norm": 0.708606924412846, "learning_rate": 6.259382782558623e-07, "loss": 0.2513, "num_tokens": 203157709.0, "step": 1550 }, { "epoch": 0.9071683803689929, "grad_norm": 0.6901795784562913, "learning_rate": 6.184880048645731e-07, "loss": 0.241, "num_tokens": 203820585.0, "step": 1555 }, { "epoch": 0.9100853204987968, "grad_norm": 0.6854819640738248, "learning_rate": 6.112589161477317e-07, "loss": 0.2474, "num_tokens": 204463922.0, "step": 1560 }, { "epoch": 0.9130022606286006, "grad_norm": 0.6490034167792925, "learning_rate": 6.042517622599872e-07, "loss": 0.2423, "num_tokens": 205116527.0, "step": 1565 }, { "epoch": 0.9159192007584044, "grad_norm": 0.6700066915913209, "learning_rate": 5.97467270326055e-07, "loss": 0.2399, "num_tokens": 205783661.0, "step": 1570 }, { "epoch": 0.9188361408882083, "grad_norm": 0.6666592595870292, "learning_rate": 5.909061443652619e-07, "loss": 0.2483, "num_tokens": 206431518.0, "step": 1575 }, { "epoch": 0.9217530810180121, "grad_norm": 0.6880885792235153, "learning_rate": 5.845690652184906e-07, "loss": 0.244, "num_tokens": 207082078.0, "step": 1580 }, { "epoch": 0.9246700211478159, "grad_norm": 0.7218901751907087, "learning_rate": 5.784566904775314e-07, "loss": 0.2439, "num_tokens": 207733633.0, "step": 1585 }, { "epoch": 0.9275869612776197, "grad_norm": 0.7224619275428422, "learning_rate": 5.725696544168431e-07, "loss": 0.2442, "num_tokens": 208384194.0, "step": 1590 }, { "epoch": 0.9305039014074236, "grad_norm": 0.685956055593419, "learning_rate": 5.669085679277353e-07, "loss": 0.2467, "num_tokens": 209017882.0, "step": 1595 }, { "epoch": 0.9334208415372275, "grad_norm": 0.6595594654962236, "learning_rate": 5.614740184549774e-07, "loss": 0.2385, "num_tokens": 209671100.0, "step": 1600 }, { "epoch": 0.9363377816670313, "grad_norm": 0.6635370096109607, "learning_rate": 5.562665699358395e-07, "loss": 0.2403, "num_tokens": 210332481.0, "step": 1605 }, { "epoch": 0.9392547217968351, "grad_norm": 0.6792107629780629, "learning_rate": 5.512867627415738e-07, "loss": 0.2485, "num_tokens": 210989185.0, "step": 1610 }, { "epoch": 0.9421716619266389, "grad_norm": 0.688506342901487, "learning_rate": 5.465351136213403e-07, "loss": 0.2404, "num_tokens": 211654349.0, "step": 1615 }, { "epoch": 0.9450886020564427, "grad_norm": 0.6957995520415324, "learning_rate": 5.420121156485843e-07, "loss": 0.2449, "num_tokens": 212306065.0, "step": 1620 }, { "epoch": 0.9480055421862467, "grad_norm": 0.6874524399723582, "learning_rate": 5.377182381698713e-07, "loss": 0.2461, "num_tokens": 212961559.0, "step": 1625 }, { "epoch": 0.9509224823160505, "grad_norm": 0.6531866036674674, "learning_rate": 5.336539267561834e-07, "loss": 0.2419, "num_tokens": 213623892.0, "step": 1630 }, { "epoch": 0.9538394224458543, "grad_norm": 0.6861036738287705, "learning_rate": 5.298196031566817e-07, "loss": 0.2397, "num_tokens": 214276109.0, "step": 1635 }, { "epoch": 0.9567563625756581, "grad_norm": 0.7385399272118951, "learning_rate": 5.262156652549434e-07, "loss": 0.2404, "num_tokens": 214935598.0, "step": 1640 }, { "epoch": 0.9596733027054619, "grad_norm": 0.7641216689081703, "learning_rate": 5.228424870276732e-07, "loss": 0.2413, "num_tokens": 215592154.0, "step": 1645 }, { "epoch": 0.9625902428352658, "grad_norm": 0.6626697793424543, "learning_rate": 5.197004185058957e-07, "loss": 0.2402, "num_tokens": 216248152.0, "step": 1650 }, { "epoch": 0.9655071829650697, "grad_norm": 0.6668250234146728, "learning_rate": 5.167897857386338e-07, "loss": 0.24, "num_tokens": 216908673.0, "step": 1655 }, { "epoch": 0.9684241230948735, "grad_norm": 0.6830282199194727, "learning_rate": 5.141108907590743e-07, "loss": 0.2447, "num_tokens": 217569951.0, "step": 1660 }, { "epoch": 0.9713410632246773, "grad_norm": 0.694769552967056, "learning_rate": 5.116640115532271e-07, "loss": 0.2463, "num_tokens": 218214661.0, "step": 1665 }, { "epoch": 0.9742580033544811, "grad_norm": 0.6576753862947748, "learning_rate": 5.09449402031078e-07, "loss": 0.2377, "num_tokens": 218875984.0, "step": 1670 }, { "epoch": 0.977174943484285, "grad_norm": 0.6885196158109818, "learning_rate": 5.074672920002409e-07, "loss": 0.2388, "num_tokens": 219529057.0, "step": 1675 }, { "epoch": 0.9800918836140888, "grad_norm": 0.6862150588167344, "learning_rate": 5.057178871421117e-07, "loss": 0.238, "num_tokens": 220197787.0, "step": 1680 }, { "epoch": 0.9830088237438926, "grad_norm": 0.6778601221453343, "learning_rate": 5.04201368990524e-07, "loss": 0.2513, "num_tokens": 220841743.0, "step": 1685 }, { "epoch": 0.9859257638736965, "grad_norm": 0.6899658979865586, "learning_rate": 5.029178949129118e-07, "loss": 0.243, "num_tokens": 221473805.0, "step": 1690 }, { "epoch": 0.9888427040035004, "grad_norm": 0.6528281983164207, "learning_rate": 5.018675980939805e-07, "loss": 0.2444, "num_tokens": 222120807.0, "step": 1695 }, { "epoch": 0.9917596441333042, "grad_norm": 0.6674621015647066, "learning_rate": 5.010505875218846e-07, "loss": 0.2386, "num_tokens": 222794425.0, "step": 1700 }, { "epoch": 0.994676584263108, "grad_norm": 0.7086141702806269, "learning_rate": 5.004669479769203e-07, "loss": 0.2405, "num_tokens": 223443814.0, "step": 1705 }, { "epoch": 0.9975935243929118, "grad_norm": 0.6795651043519514, "learning_rate": 5.001167400227263e-07, "loss": 0.2383, "num_tokens": 224098786.0, "step": 1710 }, { "epoch": 0.9999270764967549, "num_tokens": 224623551.0, "step": 1714, "total_flos": 9.914064352850215e+18, "train_loss": 0.2932530035076787, "train_runtime": 32864.6589, "train_samples_per_second": 3.338, "train_steps_per_second": 0.052 } ], "logging_steps": 5, "max_steps": 1714, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9.914064352850215e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }