| { | |
| "best_metric": 1.0332649946212769, | |
| "best_model_checkpoint": "/scratch/czm5kz/finetuned_Falcon-7B_16_1_0.0003_sequential/checkpoint-2700", | |
| "epoch": 0.9618810117563235, | |
| "eval_steps": 20, | |
| "global_step": 2700, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.6305112838745117, | |
| "learning_rate": 0.0002994656216601353, | |
| "loss": 2.8293, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.9544750452041626, | |
| "learning_rate": 0.0002989312433202707, | |
| "loss": 2.3887, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 1.2381925582885742, | |
| "learning_rate": 0.0002983968649804061, | |
| "loss": 2.1336, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 1.6555545330047607, | |
| "learning_rate": 0.0002978624866405415, | |
| "loss": 2.0826, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "eval_loss": 1.952557921409607, | |
| "eval_runtime": 244.0037, | |
| "eval_samples_per_second": 46.02, | |
| "eval_steps_per_second": 5.754, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.8518421053886414, | |
| "learning_rate": 0.0002973281083006768, | |
| "loss": 1.9436, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 1.0245165824890137, | |
| "learning_rate": 0.00029679372996081223, | |
| "loss": 1.8312, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 1.3011702299118042, | |
| "learning_rate": 0.0002962593516209476, | |
| "loss": 1.823, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.7834419012069702, | |
| "learning_rate": 0.000295724973281083, | |
| "loss": 1.7408, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "eval_loss": 1.688796877861023, | |
| "eval_runtime": 244.0831, | |
| "eval_samples_per_second": 46.005, | |
| "eval_steps_per_second": 5.752, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 1.2179532051086426, | |
| "learning_rate": 0.0002951905949412184, | |
| "loss": 1.7326, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.8537722229957581, | |
| "learning_rate": 0.00029465621660135373, | |
| "loss": 1.6131, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.7369029521942139, | |
| "learning_rate": 0.00029412183826148914, | |
| "loss": 1.6402, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.8261582255363464, | |
| "learning_rate": 0.0002935874599216245, | |
| "loss": 1.6312, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "eval_loss": 1.6201761960983276, | |
| "eval_runtime": 244.0446, | |
| "eval_samples_per_second": 46.012, | |
| "eval_steps_per_second": 5.753, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.6707925200462341, | |
| "learning_rate": 0.0002930530815817599, | |
| "loss": 1.688, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.5533925890922546, | |
| "learning_rate": 0.00029251870324189524, | |
| "loss": 1.6054, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 1.1391127109527588, | |
| "learning_rate": 0.0002919843249020306, | |
| "loss": 1.71, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.9591465592384338, | |
| "learning_rate": 0.000291449946562166, | |
| "loss": 1.5608, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "eval_loss": 1.5900055170059204, | |
| "eval_runtime": 242.9934, | |
| "eval_samples_per_second": 46.211, | |
| "eval_steps_per_second": 5.778, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.8105848431587219, | |
| "learning_rate": 0.00029091556822230134, | |
| "loss": 1.5863, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 1.0604382753372192, | |
| "learning_rate": 0.00029038118988243674, | |
| "loss": 1.5693, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.8133612871170044, | |
| "learning_rate": 0.0002898468115425721, | |
| "loss": 1.5386, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.8607632517814636, | |
| "learning_rate": 0.0002893124332027075, | |
| "loss": 1.5285, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "eval_loss": 1.5691229104995728, | |
| "eval_runtime": 242.5788, | |
| "eval_samples_per_second": 46.29, | |
| "eval_steps_per_second": 5.788, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.9120557904243469, | |
| "learning_rate": 0.00028877805486284284, | |
| "loss": 1.5746, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 1.2741694450378418, | |
| "learning_rate": 0.00028824367652297825, | |
| "loss": 1.6623, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.8563589453697205, | |
| "learning_rate": 0.0002877092981831136, | |
| "loss": 1.5452, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.9986954927444458, | |
| "learning_rate": 0.000287174919843249, | |
| "loss": 1.5181, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "eval_loss": 1.5524612665176392, | |
| "eval_runtime": 243.3678, | |
| "eval_samples_per_second": 46.14, | |
| "eval_steps_per_second": 5.769, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.7078434228897095, | |
| "learning_rate": 0.0002866405415033844, | |
| "loss": 1.5195, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.9647915363311768, | |
| "learning_rate": 0.00028610616316351975, | |
| "loss": 1.6456, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 1.033664345741272, | |
| "learning_rate": 0.00028557178482365515, | |
| "loss": 1.5991, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 1.3080883026123047, | |
| "learning_rate": 0.0002850374064837905, | |
| "loss": 1.5599, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "eval_loss": 1.5477995872497559, | |
| "eval_runtime": 243.1511, | |
| "eval_samples_per_second": 46.181, | |
| "eval_steps_per_second": 5.774, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.6780478954315186, | |
| "learning_rate": 0.00028450302814392585, | |
| "loss": 1.5578, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.7684407234191895, | |
| "learning_rate": 0.00028396864980406126, | |
| "loss": 1.6286, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.960607647895813, | |
| "learning_rate": 0.0002834342714641966, | |
| "loss": 1.5832, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.9325504899024963, | |
| "learning_rate": 0.000282899893124332, | |
| "loss": 1.4888, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "eval_loss": 1.5451940298080444, | |
| "eval_runtime": 243.2895, | |
| "eval_samples_per_second": 46.155, | |
| "eval_steps_per_second": 5.771, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 1.2480655908584595, | |
| "learning_rate": 0.00028236551478446736, | |
| "loss": 1.5704, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 1.0189040899276733, | |
| "learning_rate": 0.00028183113644460276, | |
| "loss": 1.5648, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.8874185085296631, | |
| "learning_rate": 0.0002812967581047381, | |
| "loss": 1.5737, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 1.0464727878570557, | |
| "learning_rate": 0.0002807623797648735, | |
| "loss": 1.4978, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "eval_loss": 1.5298898220062256, | |
| "eval_runtime": 242.8219, | |
| "eval_samples_per_second": 46.244, | |
| "eval_steps_per_second": 5.782, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.9442905187606812, | |
| "learning_rate": 0.00028022800142500886, | |
| "loss": 1.4986, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.7404810786247253, | |
| "learning_rate": 0.00027969362308514426, | |
| "loss": 1.4716, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 1.00387704372406, | |
| "learning_rate": 0.00027915924474527967, | |
| "loss": 1.5135, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 1.0239578485488892, | |
| "learning_rate": 0.000278624866405415, | |
| "loss": 1.4792, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "eval_loss": 1.5180584192276, | |
| "eval_runtime": 243.1799, | |
| "eval_samples_per_second": 46.176, | |
| "eval_steps_per_second": 5.774, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.8795331120491028, | |
| "learning_rate": 0.0002780904880655504, | |
| "loss": 1.5942, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 1.1808823347091675, | |
| "learning_rate": 0.00027755610972568577, | |
| "loss": 1.5523, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 1.10402512550354, | |
| "learning_rate": 0.0002770217313858211, | |
| "loss": 1.5887, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 1.1447386741638184, | |
| "learning_rate": 0.0002764873530459565, | |
| "loss": 1.5465, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "eval_loss": 1.514195203781128, | |
| "eval_runtime": 243.5401, | |
| "eval_samples_per_second": 46.107, | |
| "eval_steps_per_second": 5.765, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.8197668790817261, | |
| "learning_rate": 0.00027595297470609187, | |
| "loss": 1.5674, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.9403067231178284, | |
| "learning_rate": 0.0002754185963662273, | |
| "loss": 1.4565, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.8626269698143005, | |
| "learning_rate": 0.0002748842180263626, | |
| "loss": 1.5636, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.9701679348945618, | |
| "learning_rate": 0.000274349839686498, | |
| "loss": 1.4903, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "eval_loss": 1.5165505409240723, | |
| "eval_runtime": 242.914, | |
| "eval_samples_per_second": 46.226, | |
| "eval_steps_per_second": 5.78, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.8026458621025085, | |
| "learning_rate": 0.0002738154613466334, | |
| "loss": 1.5408, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.783799409866333, | |
| "learning_rate": 0.0002732810830067688, | |
| "loss": 1.5652, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.8295706510543823, | |
| "learning_rate": 0.0002727467046669041, | |
| "loss": 1.4772, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.9329605102539062, | |
| "learning_rate": 0.00027221232632703953, | |
| "loss": 1.5244, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "eval_loss": 1.505288004875183, | |
| "eval_runtime": 243.1666, | |
| "eval_samples_per_second": 46.178, | |
| "eval_steps_per_second": 5.774, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 1.1129286289215088, | |
| "learning_rate": 0.0002716779479871749, | |
| "loss": 1.4654, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 1.0153439044952393, | |
| "learning_rate": 0.0002711435696473103, | |
| "loss": 1.4938, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.7375654578208923, | |
| "learning_rate": 0.0002706091913074457, | |
| "loss": 1.454, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 1.2654181718826294, | |
| "learning_rate": 0.00027007481296758103, | |
| "loss": 1.5073, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "eval_loss": 1.4969700574874878, | |
| "eval_runtime": 243.386, | |
| "eval_samples_per_second": 46.137, | |
| "eval_steps_per_second": 5.769, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.8651628494262695, | |
| "learning_rate": 0.0002695404346277164, | |
| "loss": 1.5153, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.7497414946556091, | |
| "learning_rate": 0.0002690060562878518, | |
| "loss": 1.4811, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.9008344411849976, | |
| "learning_rate": 0.00026847167794798714, | |
| "loss": 1.5365, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 1.1004663705825806, | |
| "learning_rate": 0.00026793729960812254, | |
| "loss": 1.4646, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "eval_loss": 1.4890170097351074, | |
| "eval_runtime": 242.8989, | |
| "eval_samples_per_second": 46.229, | |
| "eval_steps_per_second": 5.78, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.9690203070640564, | |
| "learning_rate": 0.0002674029212682579, | |
| "loss": 1.5305, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 1.3153674602508545, | |
| "learning_rate": 0.0002668685429283933, | |
| "loss": 1.5001, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 1.1792179346084595, | |
| "learning_rate": 0.00026633416458852864, | |
| "loss": 1.4683, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.7645769715309143, | |
| "learning_rate": 0.000265799786248664, | |
| "loss": 1.4512, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "eval_loss": 1.4869476556777954, | |
| "eval_runtime": 243.0462, | |
| "eval_samples_per_second": 46.201, | |
| "eval_steps_per_second": 5.777, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.8460853695869446, | |
| "learning_rate": 0.0002652654079087994, | |
| "loss": 1.4417, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.8589680194854736, | |
| "learning_rate": 0.0002647310295689348, | |
| "loss": 1.5643, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.7434174418449402, | |
| "learning_rate": 0.00026419665122907014, | |
| "loss": 1.4943, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.9790274500846863, | |
| "learning_rate": 0.00026366227288920555, | |
| "loss": 1.4968, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "eval_loss": 1.4747017621994019, | |
| "eval_runtime": 243.372, | |
| "eval_samples_per_second": 46.139, | |
| "eval_steps_per_second": 5.769, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 1.0329363346099854, | |
| "learning_rate": 0.0002631278945493409, | |
| "loss": 1.4943, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 1.045272707939148, | |
| "learning_rate": 0.0002625935162094763, | |
| "loss": 1.4858, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.9788330793380737, | |
| "learning_rate": 0.00026205913786961165, | |
| "loss": 1.5004, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 1.041139006614685, | |
| "learning_rate": 0.00026152475952974705, | |
| "loss": 1.5084, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "eval_loss": 1.4680562019348145, | |
| "eval_runtime": 242.2635, | |
| "eval_samples_per_second": 46.35, | |
| "eval_steps_per_second": 5.795, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.8507019281387329, | |
| "learning_rate": 0.0002609903811898824, | |
| "loss": 1.5108, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.8787792921066284, | |
| "learning_rate": 0.0002604560028500178, | |
| "loss": 1.426, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.9176210761070251, | |
| "learning_rate": 0.00025992162451015315, | |
| "loss": 1.4643, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 1.035382866859436, | |
| "learning_rate": 0.00025938724617028856, | |
| "loss": 1.5125, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "eval_loss": 1.4658011198043823, | |
| "eval_runtime": 243.1807, | |
| "eval_samples_per_second": 46.176, | |
| "eval_steps_per_second": 5.773, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.9265539646148682, | |
| "learning_rate": 0.0002588528678304239, | |
| "loss": 1.5172, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.895794153213501, | |
| "learning_rate": 0.00025831848949055925, | |
| "loss": 1.4075, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.9875014424324036, | |
| "learning_rate": 0.00025778411115069466, | |
| "loss": 1.5134, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 1.0643417835235596, | |
| "learning_rate": 0.00025724973281083006, | |
| "loss": 1.5189, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "eval_loss": 1.4540351629257202, | |
| "eval_runtime": 243.2953, | |
| "eval_samples_per_second": 46.154, | |
| "eval_steps_per_second": 5.771, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 1.2020907402038574, | |
| "learning_rate": 0.0002567153544709654, | |
| "loss": 1.5471, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 1.2714240550994873, | |
| "learning_rate": 0.0002561809761311008, | |
| "loss": 1.4448, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.9557583928108215, | |
| "learning_rate": 0.00025564659779123616, | |
| "loss": 1.444, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.8486252427101135, | |
| "learning_rate": 0.00025511221945137156, | |
| "loss": 1.5165, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "eval_loss": 1.4487340450286865, | |
| "eval_runtime": 243.0706, | |
| "eval_samples_per_second": 46.196, | |
| "eval_steps_per_second": 5.776, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.738248884677887, | |
| "learning_rate": 0.0002545778411115069, | |
| "loss": 1.4421, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.8879930973052979, | |
| "learning_rate": 0.0002540434627716423, | |
| "loss": 1.4019, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.8274502754211426, | |
| "learning_rate": 0.00025350908443177767, | |
| "loss": 1.3916, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 1.0490825176239014, | |
| "learning_rate": 0.00025297470609191307, | |
| "loss": 1.4196, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_loss": 1.4473365545272827, | |
| "eval_runtime": 242.1251, | |
| "eval_samples_per_second": 46.377, | |
| "eval_steps_per_second": 5.799, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 1.233782410621643, | |
| "learning_rate": 0.0002524403277520484, | |
| "loss": 1.4154, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 1.2139190435409546, | |
| "learning_rate": 0.0002519059494121838, | |
| "loss": 1.4619, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.8714913725852966, | |
| "learning_rate": 0.00025137157107231917, | |
| "loss": 1.4809, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.795221745967865, | |
| "learning_rate": 0.0002508371927324545, | |
| "loss": 1.4804, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_loss": 1.4365161657333374, | |
| "eval_runtime": 242.2216, | |
| "eval_samples_per_second": 46.358, | |
| "eval_steps_per_second": 5.796, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.8788973093032837, | |
| "learning_rate": 0.0002503028143925899, | |
| "loss": 1.5129, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 1.1877810955047607, | |
| "learning_rate": 0.00024976843605272527, | |
| "loss": 1.4867, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 1.0848227739334106, | |
| "learning_rate": 0.0002492340577128607, | |
| "loss": 1.3657, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 1.0620073080062866, | |
| "learning_rate": 0.0002486996793729961, | |
| "loss": 1.4301, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "eval_loss": 1.4357354640960693, | |
| "eval_runtime": 242.0528, | |
| "eval_samples_per_second": 46.391, | |
| "eval_steps_per_second": 5.8, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 1.137902855873108, | |
| "learning_rate": 0.0002481653010331314, | |
| "loss": 1.3764, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 1.0430387258529663, | |
| "learning_rate": 0.00024763092269326683, | |
| "loss": 1.458, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.9362395405769348, | |
| "learning_rate": 0.0002470965443534022, | |
| "loss": 1.5025, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 1.0825713872909546, | |
| "learning_rate": 0.0002465621660135376, | |
| "loss": 1.4216, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "eval_loss": 1.4344379901885986, | |
| "eval_runtime": 242.4123, | |
| "eval_samples_per_second": 46.322, | |
| "eval_steps_per_second": 5.792, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 1.114099383354187, | |
| "learning_rate": 0.00024602778767367293, | |
| "loss": 1.4232, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.7564448714256287, | |
| "learning_rate": 0.00024549340933380833, | |
| "loss": 1.3576, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.9348393678665161, | |
| "learning_rate": 0.0002449590309939437, | |
| "loss": 1.4352, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 1.2036116123199463, | |
| "learning_rate": 0.0002444246526540791, | |
| "loss": 1.4637, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "eval_loss": 1.419852614402771, | |
| "eval_runtime": 243.2752, | |
| "eval_samples_per_second": 46.158, | |
| "eval_steps_per_second": 5.771, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 1.1303725242614746, | |
| "learning_rate": 0.00024389027431421443, | |
| "loss": 1.413, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.7789029479026794, | |
| "learning_rate": 0.00024335589597434984, | |
| "loss": 1.4399, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.9571925401687622, | |
| "learning_rate": 0.0002428215176344852, | |
| "loss": 1.3619, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 1.0115916728973389, | |
| "learning_rate": 0.00024228713929462056, | |
| "loss": 1.4548, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "eval_loss": 1.4131032228469849, | |
| "eval_runtime": 242.8013, | |
| "eval_samples_per_second": 46.248, | |
| "eval_steps_per_second": 5.783, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 1.2657145261764526, | |
| "learning_rate": 0.00024175276095475597, | |
| "loss": 1.4381, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.9139773845672607, | |
| "learning_rate": 0.00024121838261489132, | |
| "loss": 1.3604, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.8233699202537537, | |
| "learning_rate": 0.00024068400427502672, | |
| "loss": 1.3726, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 1.3115651607513428, | |
| "learning_rate": 0.00024014962593516207, | |
| "loss": 1.477, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "eval_loss": 1.4067269563674927, | |
| "eval_runtime": 242.5538, | |
| "eval_samples_per_second": 46.295, | |
| "eval_steps_per_second": 5.788, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 1.2231495380401611, | |
| "learning_rate": 0.00023961524759529747, | |
| "loss": 1.3265, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 1.1421154737472534, | |
| "learning_rate": 0.00023908086925543282, | |
| "loss": 1.4385, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 1.3289763927459717, | |
| "learning_rate": 0.0002385464909155682, | |
| "loss": 1.404, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.7699000835418701, | |
| "learning_rate": 0.00023801211257570357, | |
| "loss": 1.415, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "eval_loss": 1.3919585943222046, | |
| "eval_runtime": 242.7734, | |
| "eval_samples_per_second": 46.253, | |
| "eval_steps_per_second": 5.783, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.7574182152748108, | |
| "learning_rate": 0.00023747773423583895, | |
| "loss": 1.3588, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.871526837348938, | |
| "learning_rate": 0.00023694335589597435, | |
| "loss": 1.4096, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.8445005416870117, | |
| "learning_rate": 0.0002364089775561097, | |
| "loss": 1.3575, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 1.124621033668518, | |
| "learning_rate": 0.0002358745992162451, | |
| "loss": 1.4043, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "eval_loss": 1.391379475593567, | |
| "eval_runtime": 242.7975, | |
| "eval_samples_per_second": 46.248, | |
| "eval_steps_per_second": 5.783, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.824134349822998, | |
| "learning_rate": 0.00023534022087638045, | |
| "loss": 1.4347, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.7855484485626221, | |
| "learning_rate": 0.00023480584253651583, | |
| "loss": 1.4023, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.8013989329338074, | |
| "learning_rate": 0.0002342714641966512, | |
| "loss": 1.3739, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.8415730595588684, | |
| "learning_rate": 0.00023373708585678658, | |
| "loss": 1.3765, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "eval_loss": 1.3790498971939087, | |
| "eval_runtime": 242.5435, | |
| "eval_samples_per_second": 46.297, | |
| "eval_steps_per_second": 5.789, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 1.1093789339065552, | |
| "learning_rate": 0.00023320270751692198, | |
| "loss": 1.4332, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 1.311341404914856, | |
| "learning_rate": 0.00023266832917705733, | |
| "loss": 1.4063, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 1.1992008686065674, | |
| "learning_rate": 0.00023213395083719274, | |
| "loss": 1.3182, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.9611808061599731, | |
| "learning_rate": 0.00023159957249732808, | |
| "loss": 1.4105, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "eval_loss": 1.3601434230804443, | |
| "eval_runtime": 242.3398, | |
| "eval_samples_per_second": 46.336, | |
| "eval_steps_per_second": 5.794, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 1.0147098302841187, | |
| "learning_rate": 0.00023106519415746346, | |
| "loss": 1.3047, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 1.0039702653884888, | |
| "learning_rate": 0.00023053081581759884, | |
| "loss": 1.3057, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 1.056921362876892, | |
| "learning_rate": 0.0002299964374777342, | |
| "loss": 1.3699, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.8924540281295776, | |
| "learning_rate": 0.00022946205913786962, | |
| "loss": 1.3247, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "eval_loss": 1.3447396755218506, | |
| "eval_runtime": 242.972, | |
| "eval_samples_per_second": 46.215, | |
| "eval_steps_per_second": 5.778, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 1.3511056900024414, | |
| "learning_rate": 0.00022892768079800496, | |
| "loss": 1.3216, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.9699820280075073, | |
| "learning_rate": 0.00022839330245814037, | |
| "loss": 1.3659, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.8704125285148621, | |
| "learning_rate": 0.00022785892411827572, | |
| "loss": 1.3111, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 1.1639297008514404, | |
| "learning_rate": 0.0002273245457784111, | |
| "loss": 1.359, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "eval_loss": 1.3231163024902344, | |
| "eval_runtime": 242.9134, | |
| "eval_samples_per_second": 46.226, | |
| "eval_steps_per_second": 5.78, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 1.2916362285614014, | |
| "learning_rate": 0.00022679016743854647, | |
| "loss": 1.295, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.8053829669952393, | |
| "learning_rate": 0.00022625578909868185, | |
| "loss": 1.334, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 1.3208507299423218, | |
| "learning_rate": 0.00022572141075881722, | |
| "loss": 1.3426, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.9985295534133911, | |
| "learning_rate": 0.0002251870324189526, | |
| "loss": 1.2714, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "eval_loss": 1.3060131072998047, | |
| "eval_runtime": 243.2736, | |
| "eval_samples_per_second": 46.158, | |
| "eval_steps_per_second": 5.771, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.9396908283233643, | |
| "learning_rate": 0.000224652654079088, | |
| "loss": 1.3108, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 1.4417626857757568, | |
| "learning_rate": 0.00022411827573922335, | |
| "loss": 1.3042, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.9226711392402649, | |
| "learning_rate": 0.00022358389739935873, | |
| "loss": 1.2282, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 1.0715467929840088, | |
| "learning_rate": 0.0002230495190594941, | |
| "loss": 1.319, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "eval_loss": 1.2914878129959106, | |
| "eval_runtime": 243.1709, | |
| "eval_samples_per_second": 46.177, | |
| "eval_steps_per_second": 5.774, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 1.2655442953109741, | |
| "learning_rate": 0.00022251514071962948, | |
| "loss": 1.2446, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 1.3114930391311646, | |
| "learning_rate": 0.00022198076237976485, | |
| "loss": 1.2685, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 1.3273890018463135, | |
| "learning_rate": 0.00022144638403990023, | |
| "loss": 1.2024, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 1.3590787649154663, | |
| "learning_rate": 0.00022091200570003563, | |
| "loss": 1.2551, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "eval_loss": 1.2825496196746826, | |
| "eval_runtime": 242.3707, | |
| "eval_samples_per_second": 46.33, | |
| "eval_steps_per_second": 5.793, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.8397738337516785, | |
| "learning_rate": 0.00022037762736017098, | |
| "loss": 1.244, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 1.1953057050704956, | |
| "learning_rate": 0.00021984324902030636, | |
| "loss": 1.2534, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 1.427286982536316, | |
| "learning_rate": 0.00021930887068044173, | |
| "loss": 1.2897, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 2.1270551681518555, | |
| "learning_rate": 0.0002187744923405771, | |
| "loss": 1.25, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "eval_loss": 1.2916063070297241, | |
| "eval_runtime": 243.3228, | |
| "eval_samples_per_second": 46.149, | |
| "eval_steps_per_second": 5.77, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 1.1659835577011108, | |
| "learning_rate": 0.00021824011400071249, | |
| "loss": 1.3947, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 1.1815775632858276, | |
| "learning_rate": 0.00021770573566084786, | |
| "loss": 1.3375, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.9640941023826599, | |
| "learning_rate": 0.00021717135732098327, | |
| "loss": 1.2115, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 1.144339919090271, | |
| "learning_rate": 0.00021663697898111861, | |
| "loss": 1.2535, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "eval_loss": 1.2552716732025146, | |
| "eval_runtime": 243.3981, | |
| "eval_samples_per_second": 46.134, | |
| "eval_steps_per_second": 5.768, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 1.4256541728973389, | |
| "learning_rate": 0.00021610260064125396, | |
| "loss": 1.2659, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 1.6958651542663574, | |
| "learning_rate": 0.00021556822230138937, | |
| "loss": 1.2669, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 1.2060660123825073, | |
| "learning_rate": 0.00021503384396152474, | |
| "loss": 1.2205, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 1.2157624959945679, | |
| "learning_rate": 0.00021449946562166012, | |
| "loss": 1.3206, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "eval_loss": 1.2229853868484497, | |
| "eval_runtime": 242.4855, | |
| "eval_samples_per_second": 46.308, | |
| "eval_steps_per_second": 5.79, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 1.1371393203735352, | |
| "learning_rate": 0.0002139650872817955, | |
| "loss": 1.1855, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.8342422246932983, | |
| "learning_rate": 0.00021343070894193087, | |
| "loss": 1.2631, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 1.139548420906067, | |
| "learning_rate": 0.00021289633060206625, | |
| "loss": 1.2119, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 1.027574896812439, | |
| "learning_rate": 0.0002123619522622016, | |
| "loss": 1.2822, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "eval_loss": 1.2131222486495972, | |
| "eval_runtime": 242.8594, | |
| "eval_samples_per_second": 46.237, | |
| "eval_steps_per_second": 5.781, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 1.2451738119125366, | |
| "learning_rate": 0.000211827573922337, | |
| "loss": 1.1576, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 1.1453763246536255, | |
| "learning_rate": 0.00021129319558247238, | |
| "loss": 1.2733, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 1.153079867362976, | |
| "learning_rate": 0.00021075881724260775, | |
| "loss": 1.1779, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 1.0220108032226562, | |
| "learning_rate": 0.00021022443890274313, | |
| "loss": 1.294, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "eval_loss": 1.2016112804412842, | |
| "eval_runtime": 242.5462, | |
| "eval_samples_per_second": 46.296, | |
| "eval_steps_per_second": 5.789, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 1.6792449951171875, | |
| "learning_rate": 0.0002096900605628785, | |
| "loss": 1.1431, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 1.8450753688812256, | |
| "learning_rate": 0.00020915568222301388, | |
| "loss": 1.1796, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.9148487448692322, | |
| "learning_rate": 0.00020862130388314923, | |
| "loss": 1.122, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 1.242084264755249, | |
| "learning_rate": 0.00020808692554328463, | |
| "loss": 1.1864, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "eval_loss": 1.1892478466033936, | |
| "eval_runtime": 242.6928, | |
| "eval_samples_per_second": 46.268, | |
| "eval_steps_per_second": 5.785, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 1.928069829940796, | |
| "learning_rate": 0.00020755254720342, | |
| "loss": 1.2334, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 1.2880061864852905, | |
| "learning_rate": 0.00020701816886355538, | |
| "loss": 1.3125, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 1.4170634746551514, | |
| "learning_rate": 0.00020648379052369076, | |
| "loss": 1.2393, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.9899844527244568, | |
| "learning_rate": 0.00020594941218382614, | |
| "loss": 1.202, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "eval_loss": 1.1809042692184448, | |
| "eval_runtime": 242.6549, | |
| "eval_samples_per_second": 46.276, | |
| "eval_steps_per_second": 5.786, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 1.3030658960342407, | |
| "learning_rate": 0.0002054150338439615, | |
| "loss": 1.2495, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.9489758610725403, | |
| "learning_rate": 0.00020488065550409686, | |
| "loss": 1.1947, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 1.2204822301864624, | |
| "learning_rate": 0.00020434627716423226, | |
| "loss": 1.1442, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 1.2140864133834839, | |
| "learning_rate": 0.0002038118988243676, | |
| "loss": 1.2175, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "eval_loss": 1.1709703207015991, | |
| "eval_runtime": 243.5845, | |
| "eval_samples_per_second": 46.099, | |
| "eval_steps_per_second": 5.764, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 1.048947811126709, | |
| "learning_rate": 0.00020327752048450302, | |
| "loss": 1.1554, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.9663936495780945, | |
| "learning_rate": 0.0002027431421446384, | |
| "loss": 1.1373, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 1.4801690578460693, | |
| "learning_rate": 0.00020220876380477377, | |
| "loss": 1.2032, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 1.2649509906768799, | |
| "learning_rate": 0.00020167438546490914, | |
| "loss": 1.2164, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "eval_loss": 1.1549891233444214, | |
| "eval_runtime": 243.0477, | |
| "eval_samples_per_second": 46.201, | |
| "eval_steps_per_second": 5.777, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 1.5645591020584106, | |
| "learning_rate": 0.0002011400071250445, | |
| "loss": 1.2082, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 1.1867625713348389, | |
| "learning_rate": 0.0002006056287851799, | |
| "loss": 1.2105, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.8491153120994568, | |
| "learning_rate": 0.00020007125044531525, | |
| "loss": 1.1146, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 1.154305338859558, | |
| "learning_rate": 0.00019953687210545065, | |
| "loss": 1.1803, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "eval_loss": 1.1484609842300415, | |
| "eval_runtime": 242.7646, | |
| "eval_samples_per_second": 46.255, | |
| "eval_steps_per_second": 5.783, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 1.7545225620269775, | |
| "learning_rate": 0.00019900249376558603, | |
| "loss": 1.16, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 1.1874881982803345, | |
| "learning_rate": 0.0001984681154257214, | |
| "loss": 1.2113, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 1.3878768682479858, | |
| "learning_rate": 0.00019793373708585678, | |
| "loss": 1.1738, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 3.4664456844329834, | |
| "learning_rate": 0.00019739935874599213, | |
| "loss": 1.2146, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "eval_loss": 1.1772421598434448, | |
| "eval_runtime": 242.552, | |
| "eval_samples_per_second": 46.295, | |
| "eval_steps_per_second": 5.788, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 1.8794177770614624, | |
| "learning_rate": 0.00019686498040612753, | |
| "loss": 1.1482, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 1.044394850730896, | |
| "learning_rate": 0.00019633060206626288, | |
| "loss": 1.1143, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 1.1520146131515503, | |
| "learning_rate": 0.00019579622372639828, | |
| "loss": 1.2389, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 1.041759967803955, | |
| "learning_rate": 0.00019526184538653366, | |
| "loss": 1.1671, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "eval_loss": 1.1441816091537476, | |
| "eval_runtime": 243.7305, | |
| "eval_samples_per_second": 46.071, | |
| "eval_steps_per_second": 5.76, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 1.5767799615859985, | |
| "learning_rate": 0.00019472746704666903, | |
| "loss": 1.0781, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 1.2156161069869995, | |
| "learning_rate": 0.0001941930887068044, | |
| "loss": 1.1804, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 1.0692851543426514, | |
| "learning_rate": 0.00019365871036693976, | |
| "loss": 1.1296, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 1.0492011308670044, | |
| "learning_rate": 0.00019312433202707516, | |
| "loss": 1.195, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "eval_loss": 1.1373920440673828, | |
| "eval_runtime": 243.5127, | |
| "eval_samples_per_second": 46.113, | |
| "eval_steps_per_second": 5.766, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.9718597531318665, | |
| "learning_rate": 0.0001925899536872105, | |
| "loss": 1.1621, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 1.5123802423477173, | |
| "learning_rate": 0.00019205557534734591, | |
| "loss": 1.1181, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 1.209719181060791, | |
| "learning_rate": 0.00019152119700748126, | |
| "loss": 1.1386, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 2.3843159675598145, | |
| "learning_rate": 0.00019098681866761667, | |
| "loss": 1.1055, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "eval_loss": 1.1293574571609497, | |
| "eval_runtime": 243.5039, | |
| "eval_samples_per_second": 46.114, | |
| "eval_steps_per_second": 5.766, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 1.0101613998413086, | |
| "learning_rate": 0.00019045244032775204, | |
| "loss": 1.1154, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 1.160309076309204, | |
| "learning_rate": 0.0001899180619878874, | |
| "loss": 1.122, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 1.289068579673767, | |
| "learning_rate": 0.0001893836836480228, | |
| "loss": 1.1491, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 2.579099178314209, | |
| "learning_rate": 0.00018884930530815814, | |
| "loss": 1.1484, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "eval_loss": 1.1274968385696411, | |
| "eval_runtime": 243.2415, | |
| "eval_samples_per_second": 46.164, | |
| "eval_steps_per_second": 5.772, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.8187231421470642, | |
| "learning_rate": 0.00018831492696829355, | |
| "loss": 1.1322, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.8516179919242859, | |
| "learning_rate": 0.0001877805486284289, | |
| "loss": 1.0975, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.9972283244132996, | |
| "learning_rate": 0.0001872461702885643, | |
| "loss": 1.1063, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.9460062384605408, | |
| "learning_rate": 0.00018671179194869967, | |
| "loss": 1.1337, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "eval_loss": 1.124145269393921, | |
| "eval_runtime": 243.6433, | |
| "eval_samples_per_second": 46.088, | |
| "eval_steps_per_second": 5.763, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 1.1518056392669678, | |
| "learning_rate": 0.00018617741360883502, | |
| "loss": 1.139, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 1.3771902322769165, | |
| "learning_rate": 0.00018564303526897043, | |
| "loss": 1.1773, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 1.3464629650115967, | |
| "learning_rate": 0.00018510865692910578, | |
| "loss": 1.1176, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 1.1559851169586182, | |
| "learning_rate": 0.00018457427858924118, | |
| "loss": 1.1147, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "eval_loss": 1.1198575496673584, | |
| "eval_runtime": 242.729, | |
| "eval_samples_per_second": 46.261, | |
| "eval_steps_per_second": 5.784, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.8448060750961304, | |
| "learning_rate": 0.00018403990024937653, | |
| "loss": 1.1604, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 1.1841185092926025, | |
| "learning_rate": 0.00018350552190951193, | |
| "loss": 1.1384, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.97804856300354, | |
| "learning_rate": 0.0001829711435696473, | |
| "loss": 1.0934, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.7992210388183594, | |
| "learning_rate": 0.00018243676522978266, | |
| "loss": 1.1227, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "eval_loss": 1.115310549736023, | |
| "eval_runtime": 242.2075, | |
| "eval_samples_per_second": 46.361, | |
| "eval_steps_per_second": 5.797, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.9271190166473389, | |
| "learning_rate": 0.00018190238688991806, | |
| "loss": 1.1979, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 1.1744805574417114, | |
| "learning_rate": 0.0001813680085500534, | |
| "loss": 1.0411, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 1.5105717182159424, | |
| "learning_rate": 0.0001808336302101888, | |
| "loss": 1.1139, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 1.0240634679794312, | |
| "learning_rate": 0.00018029925187032416, | |
| "loss": 1.145, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "eval_loss": 1.116929054260254, | |
| "eval_runtime": 243.2586, | |
| "eval_samples_per_second": 46.161, | |
| "eval_steps_per_second": 5.772, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 1.387168049812317, | |
| "learning_rate": 0.00017976487353045956, | |
| "loss": 1.0915, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.8388767838478088, | |
| "learning_rate": 0.0001792304951905949, | |
| "loss": 1.1492, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 1.1189345121383667, | |
| "learning_rate": 0.0001786961168507303, | |
| "loss": 1.0952, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.9314697980880737, | |
| "learning_rate": 0.0001781617385108657, | |
| "loss": 1.1508, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "eval_loss": 1.1075862646102905, | |
| "eval_runtime": 242.7453, | |
| "eval_samples_per_second": 46.258, | |
| "eval_steps_per_second": 5.784, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 1.495107650756836, | |
| "learning_rate": 0.00017762736017100104, | |
| "loss": 1.1027, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 1.4054560661315918, | |
| "learning_rate": 0.00017709298183113644, | |
| "loss": 1.0901, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 1.1209503412246704, | |
| "learning_rate": 0.0001765586034912718, | |
| "loss": 1.1063, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 1.1039865016937256, | |
| "learning_rate": 0.0001760242251514072, | |
| "loss": 1.1729, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "eval_loss": 1.108239769935608, | |
| "eval_runtime": 242.5373, | |
| "eval_samples_per_second": 46.298, | |
| "eval_steps_per_second": 5.789, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 1.1784499883651733, | |
| "learning_rate": 0.00017548984681154255, | |
| "loss": 1.1479, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 1.4968849420547485, | |
| "learning_rate": 0.00017495546847167792, | |
| "loss": 1.1732, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.8712586760520935, | |
| "learning_rate": 0.00017442109013181332, | |
| "loss": 1.1105, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.8051401972770691, | |
| "learning_rate": 0.00017388671179194867, | |
| "loss": 1.1253, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "eval_loss": 1.10274338722229, | |
| "eval_runtime": 243.3942, | |
| "eval_samples_per_second": 46.135, | |
| "eval_steps_per_second": 5.768, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.945894181728363, | |
| "learning_rate": 0.00017335233345208408, | |
| "loss": 1.145, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 1.0374635457992554, | |
| "learning_rate": 0.00017281795511221943, | |
| "loss": 1.1411, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 1.0259712934494019, | |
| "learning_rate": 0.00017228357677235483, | |
| "loss": 1.0862, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 1.2664566040039062, | |
| "learning_rate": 0.00017174919843249018, | |
| "loss": 1.1152, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "eval_loss": 1.099318265914917, | |
| "eval_runtime": 243.0373, | |
| "eval_samples_per_second": 46.203, | |
| "eval_steps_per_second": 5.777, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 1.3761073350906372, | |
| "learning_rate": 0.00017121482009262555, | |
| "loss": 1.1706, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 1.5430552959442139, | |
| "learning_rate": 0.00017068044175276096, | |
| "loss": 1.0759, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 1.2533998489379883, | |
| "learning_rate": 0.0001701460634128963, | |
| "loss": 1.1051, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.9215791821479797, | |
| "learning_rate": 0.0001696116850730317, | |
| "loss": 1.1606, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "eval_loss": 1.0982745885849, | |
| "eval_runtime": 242.3387, | |
| "eval_samples_per_second": 46.336, | |
| "eval_steps_per_second": 5.794, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 1.1964141130447388, | |
| "learning_rate": 0.00016907730673316706, | |
| "loss": 1.1286, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 2.441573143005371, | |
| "learning_rate": 0.00016854292839330246, | |
| "loss": 1.0948, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 1.9406346082687378, | |
| "learning_rate": 0.0001680085500534378, | |
| "loss": 1.0833, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 1.2231512069702148, | |
| "learning_rate": 0.0001674741717135732, | |
| "loss": 1.1399, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "eval_loss": 1.097598671913147, | |
| "eval_runtime": 243.0716, | |
| "eval_samples_per_second": 46.196, | |
| "eval_steps_per_second": 5.776, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.9176976680755615, | |
| "learning_rate": 0.00016693979337370856, | |
| "loss": 1.0967, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 1.0640438795089722, | |
| "learning_rate": 0.00016640541503384394, | |
| "loss": 1.0571, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.9471141695976257, | |
| "learning_rate": 0.00016587103669397934, | |
| "loss": 1.0619, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.984430730342865, | |
| "learning_rate": 0.0001653366583541147, | |
| "loss": 1.0867, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "eval_loss": 1.0952093601226807, | |
| "eval_runtime": 243.3103, | |
| "eval_samples_per_second": 46.151, | |
| "eval_steps_per_second": 5.77, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 1.0785785913467407, | |
| "learning_rate": 0.0001648022800142501, | |
| "loss": 1.1129, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 1.0736885070800781, | |
| "learning_rate": 0.00016426790167438544, | |
| "loss": 1.165, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 2.7780587673187256, | |
| "learning_rate": 0.00016373352333452082, | |
| "loss": 1.1582, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 1.2111455202102661, | |
| "learning_rate": 0.0001631991449946562, | |
| "loss": 1.1036, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "eval_loss": 1.097571849822998, | |
| "eval_runtime": 242.4227, | |
| "eval_samples_per_second": 46.32, | |
| "eval_steps_per_second": 5.792, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 1.2604445219039917, | |
| "learning_rate": 0.00016266476665479157, | |
| "loss": 1.1973, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.853370189666748, | |
| "learning_rate": 0.00016213038831492697, | |
| "loss": 1.0787, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.808193027973175, | |
| "learning_rate": 0.00016159600997506232, | |
| "loss": 1.07, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 1.1741416454315186, | |
| "learning_rate": 0.00016106163163519773, | |
| "loss": 1.127, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "eval_loss": 1.0869979858398438, | |
| "eval_runtime": 243.2654, | |
| "eval_samples_per_second": 46.159, | |
| "eval_steps_per_second": 5.771, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 1.1077781915664673, | |
| "learning_rate": 0.00016052725329533308, | |
| "loss": 1.0511, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.9855788350105286, | |
| "learning_rate": 0.00015999287495546845, | |
| "loss": 1.0973, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 1.3673230409622192, | |
| "learning_rate": 0.00015945849661560383, | |
| "loss": 1.1164, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 1.2939677238464355, | |
| "learning_rate": 0.0001589241182757392, | |
| "loss": 1.0769, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "eval_loss": 1.0876641273498535, | |
| "eval_runtime": 242.1557, | |
| "eval_samples_per_second": 46.371, | |
| "eval_steps_per_second": 5.798, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 1.06653892993927, | |
| "learning_rate": 0.0001583897399358746, | |
| "loss": 1.0127, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.9674182534217834, | |
| "learning_rate": 0.00015785536159600996, | |
| "loss": 1.0367, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.662380576133728, | |
| "learning_rate": 0.00015732098325614536, | |
| "loss": 1.092, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 1.4071582555770874, | |
| "learning_rate": 0.0001567866049162807, | |
| "loss": 1.1572, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "eval_loss": 1.0884110927581787, | |
| "eval_runtime": 242.9901, | |
| "eval_samples_per_second": 46.212, | |
| "eval_steps_per_second": 5.778, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.9712650775909424, | |
| "learning_rate": 0.00015625222657641608, | |
| "loss": 1.0277, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 1.070185899734497, | |
| "learning_rate": 0.00015571784823655146, | |
| "loss": 1.1811, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.8417230248451233, | |
| "learning_rate": 0.00015518346989668684, | |
| "loss": 1.0426, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.8787222504615784, | |
| "learning_rate": 0.0001546490915568222, | |
| "loss": 0.9953, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "eval_loss": 1.082081913948059, | |
| "eval_runtime": 243.0414, | |
| "eval_samples_per_second": 46.202, | |
| "eval_steps_per_second": 5.777, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.9979585409164429, | |
| "learning_rate": 0.0001541147132169576, | |
| "loss": 1.0731, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 1.108880639076233, | |
| "learning_rate": 0.000153580334877093, | |
| "loss": 1.0171, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 1.2933177947998047, | |
| "learning_rate": 0.00015304595653722834, | |
| "loss": 1.0382, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 1.237389087677002, | |
| "learning_rate": 0.00015251157819736372, | |
| "loss": 1.1542, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "eval_loss": 1.0847558975219727, | |
| "eval_runtime": 243.7745, | |
| "eval_samples_per_second": 46.063, | |
| "eval_steps_per_second": 5.759, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 1.2602064609527588, | |
| "learning_rate": 0.0001519771998574991, | |
| "loss": 1.0919, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 1.1393736600875854, | |
| "learning_rate": 0.00015144282151763447, | |
| "loss": 1.0688, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 1.037117838859558, | |
| "learning_rate": 0.00015090844317776984, | |
| "loss": 1.062, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 1.169497013092041, | |
| "learning_rate": 0.00015037406483790522, | |
| "loss": 1.1268, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "eval_loss": 1.0819194316864014, | |
| "eval_runtime": 242.9727, | |
| "eval_samples_per_second": 46.215, | |
| "eval_steps_per_second": 5.778, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.9044406414031982, | |
| "learning_rate": 0.0001498396864980406, | |
| "loss": 1.0518, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.8581820726394653, | |
| "learning_rate": 0.00014930530815817597, | |
| "loss": 1.0841, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.9462387561798096, | |
| "learning_rate": 0.00014877092981831135, | |
| "loss": 1.0453, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.9421359896659851, | |
| "learning_rate": 0.00014823655147844673, | |
| "loss": 1.0587, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "eval_loss": 1.0787664651870728, | |
| "eval_runtime": 242.7641, | |
| "eval_samples_per_second": 46.255, | |
| "eval_steps_per_second": 5.783, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 1.2465983629226685, | |
| "learning_rate": 0.0001477021731385821, | |
| "loss": 1.0594, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 1.5653964281082153, | |
| "learning_rate": 0.00014716779479871748, | |
| "loss": 1.1475, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 1.5499756336212158, | |
| "learning_rate": 0.00014663341645885285, | |
| "loss": 1.0816, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 1.3674159049987793, | |
| "learning_rate": 0.00014609903811898823, | |
| "loss": 1.1724, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "eval_loss": 1.0885018110275269, | |
| "eval_runtime": 241.7321, | |
| "eval_samples_per_second": 46.452, | |
| "eval_steps_per_second": 5.808, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.9569036960601807, | |
| "learning_rate": 0.0001455646597791236, | |
| "loss": 1.1324, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.7200804352760315, | |
| "learning_rate": 0.00014503028143925898, | |
| "loss": 1.1145, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 1.5475003719329834, | |
| "learning_rate": 0.00014449590309939436, | |
| "loss": 1.1219, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.9555292725563049, | |
| "learning_rate": 0.00014396152475952973, | |
| "loss": 1.0941, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "eval_loss": 1.0766382217407227, | |
| "eval_runtime": 242.1732, | |
| "eval_samples_per_second": 46.368, | |
| "eval_steps_per_second": 5.798, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.847709596157074, | |
| "learning_rate": 0.0001434271464196651, | |
| "loss": 1.079, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.9792425632476807, | |
| "learning_rate": 0.00014289276807980049, | |
| "loss": 1.051, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 1.1272369623184204, | |
| "learning_rate": 0.00014235838973993586, | |
| "loss": 1.089, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 1.2628141641616821, | |
| "learning_rate": 0.00014182401140007124, | |
| "loss": 1.0375, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "eval_loss": 1.078391432762146, | |
| "eval_runtime": 243.4192, | |
| "eval_samples_per_second": 46.13, | |
| "eval_steps_per_second": 5.768, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 1.011176586151123, | |
| "learning_rate": 0.00014128963306020661, | |
| "loss": 1.0432, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.7509900331497192, | |
| "learning_rate": 0.000140755254720342, | |
| "loss": 1.026, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.8755093812942505, | |
| "learning_rate": 0.00014022087638047737, | |
| "loss": 1.0355, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 1.0947227478027344, | |
| "learning_rate": 0.00013968649804061274, | |
| "loss": 1.1522, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "eval_loss": 1.0746785402297974, | |
| "eval_runtime": 243.2633, | |
| "eval_samples_per_second": 46.16, | |
| "eval_steps_per_second": 5.772, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 1.576208472251892, | |
| "learning_rate": 0.00013915211970074812, | |
| "loss": 1.0697, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 1.0187541246414185, | |
| "learning_rate": 0.0001386177413608835, | |
| "loss": 1.0736, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.9235592484474182, | |
| "learning_rate": 0.00013808336302101887, | |
| "loss": 1.066, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 1.0646727085113525, | |
| "learning_rate": 0.00013754898468115425, | |
| "loss": 1.0208, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "eval_loss": 1.0796867609024048, | |
| "eval_runtime": 242.7054, | |
| "eval_samples_per_second": 46.266, | |
| "eval_steps_per_second": 5.785, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.865246057510376, | |
| "learning_rate": 0.00013701460634128962, | |
| "loss": 1.0993, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 1.3428738117218018, | |
| "learning_rate": 0.000136480228001425, | |
| "loss": 1.1218, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 1.060223937034607, | |
| "learning_rate": 0.00013594584966156037, | |
| "loss": 1.0701, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 1.3685177564620972, | |
| "learning_rate": 0.00013541147132169575, | |
| "loss": 1.1178, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "eval_loss": 1.0757360458374023, | |
| "eval_runtime": 242.3944, | |
| "eval_samples_per_second": 46.325, | |
| "eval_steps_per_second": 5.792, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 1.018256425857544, | |
| "learning_rate": 0.00013487709298183113, | |
| "loss": 1.0114, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.9656644463539124, | |
| "learning_rate": 0.0001343427146419665, | |
| "loss": 1.0382, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.9408764243125916, | |
| "learning_rate": 0.00013380833630210188, | |
| "loss": 1.0636, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 1.3006953001022339, | |
| "learning_rate": 0.00013327395796223726, | |
| "loss": 1.0406, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "eval_loss": 1.068661093711853, | |
| "eval_runtime": 243.1912, | |
| "eval_samples_per_second": 46.174, | |
| "eval_steps_per_second": 5.773, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.797767698764801, | |
| "learning_rate": 0.00013273957962237263, | |
| "loss": 1.0338, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.7967373728752136, | |
| "learning_rate": 0.000132205201282508, | |
| "loss": 1.0541, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 1.1280144453048706, | |
| "learning_rate": 0.00013167082294264338, | |
| "loss": 1.0926, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.9334732294082642, | |
| "learning_rate": 0.00013113644460277876, | |
| "loss": 1.073, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "eval_loss": 1.0698789358139038, | |
| "eval_runtime": 243.8062, | |
| "eval_samples_per_second": 46.057, | |
| "eval_steps_per_second": 5.759, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.763647198677063, | |
| "learning_rate": 0.00013060206626291414, | |
| "loss": 1.0405, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 1.2175519466400146, | |
| "learning_rate": 0.0001300676879230495, | |
| "loss": 1.0477, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.8948994278907776, | |
| "learning_rate": 0.0001295333095831849, | |
| "loss": 1.0485, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 1.2777072191238403, | |
| "learning_rate": 0.00012899893124332026, | |
| "loss": 1.1138, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "eval_loss": 1.0665650367736816, | |
| "eval_runtime": 243.1663, | |
| "eval_samples_per_second": 46.178, | |
| "eval_steps_per_second": 5.774, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 1.6561847925186157, | |
| "learning_rate": 0.00012846455290345564, | |
| "loss": 1.038, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 1.475900650024414, | |
| "learning_rate": 0.00012793017456359102, | |
| "loss": 1.0814, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 2.2426915168762207, | |
| "learning_rate": 0.0001273957962237264, | |
| "loss": 1.1084, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.9208836555480957, | |
| "learning_rate": 0.00012686141788386177, | |
| "loss": 1.0727, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "eval_loss": 1.0669995546340942, | |
| "eval_runtime": 242.7171, | |
| "eval_samples_per_second": 46.264, | |
| "eval_steps_per_second": 5.785, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 1.1481331586837769, | |
| "learning_rate": 0.00012632703954399714, | |
| "loss": 1.0752, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.9362524747848511, | |
| "learning_rate": 0.00012579266120413252, | |
| "loss": 1.0576, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 1.0292400121688843, | |
| "learning_rate": 0.0001252582828642679, | |
| "loss": 1.1245, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.7851674556732178, | |
| "learning_rate": 0.00012472390452440327, | |
| "loss": 1.0422, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "eval_loss": 1.065503716468811, | |
| "eval_runtime": 242.4737, | |
| "eval_samples_per_second": 46.31, | |
| "eval_steps_per_second": 5.79, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 1.1515681743621826, | |
| "learning_rate": 0.00012418952618453862, | |
| "loss": 1.1243, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.7999959588050842, | |
| "learning_rate": 0.00012365514784467402, | |
| "loss": 1.0751, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 1.5812344551086426, | |
| "learning_rate": 0.0001231207695048094, | |
| "loss": 1.064, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 1.4247961044311523, | |
| "learning_rate": 0.00012258639116494478, | |
| "loss": 1.0476, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "eval_loss": 1.065142035484314, | |
| "eval_runtime": 243.1043, | |
| "eval_samples_per_second": 46.19, | |
| "eval_steps_per_second": 5.775, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.6534472703933716, | |
| "learning_rate": 0.00012205201282508015, | |
| "loss": 1.0618, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.7015573382377625, | |
| "learning_rate": 0.00012151763448521553, | |
| "loss": 1.0167, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 1.0347164869308472, | |
| "learning_rate": 0.0001209832561453509, | |
| "loss": 1.1369, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.8698253035545349, | |
| "learning_rate": 0.00012044887780548627, | |
| "loss": 1.0642, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "eval_loss": 1.0631712675094604, | |
| "eval_runtime": 242.718, | |
| "eval_samples_per_second": 46.264, | |
| "eval_steps_per_second": 5.784, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 1.0342012643814087, | |
| "learning_rate": 0.00011991449946562164, | |
| "loss": 1.1106, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.7427274584770203, | |
| "learning_rate": 0.00011938012112575702, | |
| "loss": 1.0906, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.8352236151695251, | |
| "learning_rate": 0.0001188457427858924, | |
| "loss": 1.0482, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.9208966493606567, | |
| "learning_rate": 0.00011831136444602779, | |
| "loss": 1.0269, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "eval_loss": 1.061272144317627, | |
| "eval_runtime": 243.2083, | |
| "eval_samples_per_second": 46.17, | |
| "eval_steps_per_second": 5.773, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.7243204116821289, | |
| "learning_rate": 0.00011777698610616316, | |
| "loss": 1.0752, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 1.1476917266845703, | |
| "learning_rate": 0.00011724260776629854, | |
| "loss": 1.0978, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.6801986694335938, | |
| "learning_rate": 0.0001167082294264339, | |
| "loss": 1.0752, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.7917805314064026, | |
| "learning_rate": 0.00011617385108656928, | |
| "loss": 1.0756, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "eval_loss": 1.0615525245666504, | |
| "eval_runtime": 242.3231, | |
| "eval_samples_per_second": 46.339, | |
| "eval_steps_per_second": 5.794, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.9897934794425964, | |
| "learning_rate": 0.00011563947274670465, | |
| "loss": 1.0492, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.9467993378639221, | |
| "learning_rate": 0.00011510509440684003, | |
| "loss": 1.023, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.7573145627975464, | |
| "learning_rate": 0.0001145707160669754, | |
| "loss": 1.0759, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.9526081085205078, | |
| "learning_rate": 0.0001140363377271108, | |
| "loss": 1.1442, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "eval_loss": 1.0587379932403564, | |
| "eval_runtime": 241.9784, | |
| "eval_samples_per_second": 46.405, | |
| "eval_steps_per_second": 5.802, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.9118272662162781, | |
| "learning_rate": 0.00011350195938724617, | |
| "loss": 1.1047, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.951062798500061, | |
| "learning_rate": 0.00011296758104738153, | |
| "loss": 1.0981, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.9109734296798706, | |
| "learning_rate": 0.00011243320270751691, | |
| "loss": 1.0994, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.8266547918319702, | |
| "learning_rate": 0.00011189882436765228, | |
| "loss": 1.0417, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "eval_loss": 1.0577157735824585, | |
| "eval_runtime": 243.6894, | |
| "eval_samples_per_second": 46.079, | |
| "eval_steps_per_second": 5.761, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 1.2194539308547974, | |
| "learning_rate": 0.00011136444602778766, | |
| "loss": 1.1562, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.9784891605377197, | |
| "learning_rate": 0.00011083006768792304, | |
| "loss": 1.08, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.8719102740287781, | |
| "learning_rate": 0.00011029568934805841, | |
| "loss": 1.0187, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.679275631904602, | |
| "learning_rate": 0.0001097613110081938, | |
| "loss": 1.0336, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "eval_loss": 1.0579383373260498, | |
| "eval_runtime": 242.7508, | |
| "eval_samples_per_second": 46.257, | |
| "eval_steps_per_second": 5.784, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.8664696216583252, | |
| "learning_rate": 0.00010922693266832918, | |
| "loss": 1.0849, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.8156464099884033, | |
| "learning_rate": 0.00010869255432846454, | |
| "loss": 1.0582, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.7599331140518188, | |
| "learning_rate": 0.00010815817598859992, | |
| "loss": 1.0747, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 1.429771065711975, | |
| "learning_rate": 0.0001076237976487353, | |
| "loss": 1.0762, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "eval_loss": 1.0569549798965454, | |
| "eval_runtime": 241.936, | |
| "eval_samples_per_second": 46.413, | |
| "eval_steps_per_second": 5.803, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 1.29656183719635, | |
| "learning_rate": 0.00010708941930887067, | |
| "loss": 1.1023, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 1.0881860256195068, | |
| "learning_rate": 0.00010655504096900605, | |
| "loss": 1.081, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 1.3329012393951416, | |
| "learning_rate": 0.00010602066262914144, | |
| "loss": 1.1507, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 1.3197088241577148, | |
| "learning_rate": 0.00010548628428927681, | |
| "loss": 1.1399, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "eval_loss": 1.059924840927124, | |
| "eval_runtime": 243.3976, | |
| "eval_samples_per_second": 46.134, | |
| "eval_steps_per_second": 5.768, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.7774195075035095, | |
| "learning_rate": 0.00010495190594941217, | |
| "loss": 0.9988, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.8935036063194275, | |
| "learning_rate": 0.00010441752760954755, | |
| "loss": 1.0987, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 1.0009044408798218, | |
| "learning_rate": 0.00010388314926968293, | |
| "loss": 0.9937, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 1.0484709739685059, | |
| "learning_rate": 0.0001033487709298183, | |
| "loss": 1.0625, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "eval_loss": 1.0552020072937012, | |
| "eval_runtime": 243.1169, | |
| "eval_samples_per_second": 46.188, | |
| "eval_steps_per_second": 5.775, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.8426492214202881, | |
| "learning_rate": 0.00010281439258995368, | |
| "loss": 1.1666, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.8092815279960632, | |
| "learning_rate": 0.00010228001425008905, | |
| "loss": 1.033, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.9522695541381836, | |
| "learning_rate": 0.00010174563591022444, | |
| "loss": 1.0193, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.704237699508667, | |
| "learning_rate": 0.00010121125757035979, | |
| "loss": 1.0285, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "eval_loss": 1.0537731647491455, | |
| "eval_runtime": 241.9108, | |
| "eval_samples_per_second": 46.418, | |
| "eval_steps_per_second": 5.804, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 1.1165724992752075, | |
| "learning_rate": 0.00010067687923049518, | |
| "loss": 1.087, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.6024936437606812, | |
| "learning_rate": 0.00010014250089063056, | |
| "loss": 1.0875, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 1.0185704231262207, | |
| "learning_rate": 9.960812255076593e-05, | |
| "loss": 1.0953, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.7029600739479065, | |
| "learning_rate": 9.907374421090131e-05, | |
| "loss": 1.0705, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "eval_loss": 1.0541706085205078, | |
| "eval_runtime": 243.1191, | |
| "eval_samples_per_second": 46.187, | |
| "eval_steps_per_second": 5.775, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.8626222610473633, | |
| "learning_rate": 9.853936587103669e-05, | |
| "loss": 1.0528, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.8519126176834106, | |
| "learning_rate": 9.800498753117206e-05, | |
| "loss": 1.0469, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.7846513390541077, | |
| "learning_rate": 9.747060919130743e-05, | |
| "loss": 1.1003, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.7816912531852722, | |
| "learning_rate": 9.693623085144281e-05, | |
| "loss": 1.0338, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "eval_loss": 1.0513091087341309, | |
| "eval_runtime": 241.7979, | |
| "eval_samples_per_second": 46.44, | |
| "eval_steps_per_second": 5.807, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.9167221188545227, | |
| "learning_rate": 9.640185251157819e-05, | |
| "loss": 1.0941, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.6667194366455078, | |
| "learning_rate": 9.586747417171357e-05, | |
| "loss": 1.0701, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.9745165705680847, | |
| "learning_rate": 9.533309583184894e-05, | |
| "loss": 1.0187, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.6843913197517395, | |
| "learning_rate": 9.479871749198432e-05, | |
| "loss": 1.0812, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "eval_loss": 1.0521106719970703, | |
| "eval_runtime": 243.4663, | |
| "eval_samples_per_second": 46.121, | |
| "eval_steps_per_second": 5.767, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 1.2488641738891602, | |
| "learning_rate": 9.42643391521197e-05, | |
| "loss": 1.0166, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.964064359664917, | |
| "learning_rate": 9.372996081225506e-05, | |
| "loss": 1.0181, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.8608153462409973, | |
| "learning_rate": 9.319558247239043e-05, | |
| "loss": 1.0447, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.841242790222168, | |
| "learning_rate": 9.266120413252582e-05, | |
| "loss": 1.0161, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "eval_loss": 1.0501881837844849, | |
| "eval_runtime": 242.961, | |
| "eval_samples_per_second": 46.217, | |
| "eval_steps_per_second": 5.779, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 1.064781904220581, | |
| "learning_rate": 9.21268257926612e-05, | |
| "loss": 1.0696, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 1.0806126594543457, | |
| "learning_rate": 9.159244745279658e-05, | |
| "loss": 1.0982, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 1.3890222311019897, | |
| "learning_rate": 9.105806911293195e-05, | |
| "loss": 1.1408, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.9842618107795715, | |
| "learning_rate": 9.052369077306733e-05, | |
| "loss": 1.0007, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "eval_loss": 1.0495084524154663, | |
| "eval_runtime": 242.9243, | |
| "eval_samples_per_second": 46.224, | |
| "eval_steps_per_second": 5.78, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 1.2397950887680054, | |
| "learning_rate": 8.998931243320269e-05, | |
| "loss": 1.0121, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.7010890245437622, | |
| "learning_rate": 8.945493409333807e-05, | |
| "loss": 1.0179, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.7378676533699036, | |
| "learning_rate": 8.892055575347344e-05, | |
| "loss": 1.0801, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 1.0177286863327026, | |
| "learning_rate": 8.838617741360883e-05, | |
| "loss": 1.1177, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "eval_loss": 1.0490984916687012, | |
| "eval_runtime": 243.5095, | |
| "eval_samples_per_second": 46.113, | |
| "eval_steps_per_second": 5.766, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.9546262621879578, | |
| "learning_rate": 8.785179907374421e-05, | |
| "loss": 1.062, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.7818737030029297, | |
| "learning_rate": 8.731742073387958e-05, | |
| "loss": 0.9988, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 1.0717649459838867, | |
| "learning_rate": 8.678304239401496e-05, | |
| "loss": 1.0817, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.8660954236984253, | |
| "learning_rate": 8.624866405415032e-05, | |
| "loss": 1.0646, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "eval_loss": 1.050535798072815, | |
| "eval_runtime": 242.3199, | |
| "eval_samples_per_second": 46.34, | |
| "eval_steps_per_second": 5.794, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.8482767939567566, | |
| "learning_rate": 8.57142857142857e-05, | |
| "loss": 1.0838, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.7331980466842651, | |
| "learning_rate": 8.517990737442108e-05, | |
| "loss": 1.1195, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.8289487361907959, | |
| "learning_rate": 8.464552903455646e-05, | |
| "loss": 1.0255, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 1.1590027809143066, | |
| "learning_rate": 8.411115069469184e-05, | |
| "loss": 1.0598, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "eval_loss": 1.0510787963867188, | |
| "eval_runtime": 242.7439, | |
| "eval_samples_per_second": 46.259, | |
| "eval_steps_per_second": 5.784, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.8864056468009949, | |
| "learning_rate": 8.357677235482722e-05, | |
| "loss": 1.0605, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 1.1364105939865112, | |
| "learning_rate": 8.304239401496259e-05, | |
| "loss": 1.0255, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 1.044231653213501, | |
| "learning_rate": 8.250801567509796e-05, | |
| "loss": 1.1362, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 1.038897156715393, | |
| "learning_rate": 8.197363733523333e-05, | |
| "loss": 1.1022, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "eval_loss": 1.0472804307937622, | |
| "eval_runtime": 242.5807, | |
| "eval_samples_per_second": 46.29, | |
| "eval_steps_per_second": 5.788, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 1.0197745561599731, | |
| "learning_rate": 8.143925899536871e-05, | |
| "loss": 1.0834, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 1.3458046913146973, | |
| "learning_rate": 8.090488065550408e-05, | |
| "loss": 1.0143, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.6543688774108887, | |
| "learning_rate": 8.037050231563947e-05, | |
| "loss": 1.0277, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.9424459338188171, | |
| "learning_rate": 7.983612397577485e-05, | |
| "loss": 1.0171, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "eval_loss": 1.0462262630462646, | |
| "eval_runtime": 242.9409, | |
| "eval_samples_per_second": 46.221, | |
| "eval_steps_per_second": 5.779, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.9467738270759583, | |
| "learning_rate": 7.930174563591023e-05, | |
| "loss": 1.0026, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.7832115888595581, | |
| "learning_rate": 7.876736729604559e-05, | |
| "loss": 1.0163, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.8243778347969055, | |
| "learning_rate": 7.823298895618096e-05, | |
| "loss": 1.0625, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.7256205081939697, | |
| "learning_rate": 7.769861061631634e-05, | |
| "loss": 1.0061, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "eval_loss": 1.0456407070159912, | |
| "eval_runtime": 242.46, | |
| "eval_samples_per_second": 46.313, | |
| "eval_steps_per_second": 5.791, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.847520112991333, | |
| "learning_rate": 7.716423227645172e-05, | |
| "loss": 1.0741, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.7502401471138, | |
| "learning_rate": 7.662985393658709e-05, | |
| "loss": 1.0904, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.865737795829773, | |
| "learning_rate": 7.609547559672248e-05, | |
| "loss": 1.1138, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 1.011161208152771, | |
| "learning_rate": 7.556109725685786e-05, | |
| "loss": 1.137, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "eval_loss": 1.0469642877578735, | |
| "eval_runtime": 243.2507, | |
| "eval_samples_per_second": 46.162, | |
| "eval_steps_per_second": 5.772, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 1.1329538822174072, | |
| "learning_rate": 7.502671891699322e-05, | |
| "loss": 1.0289, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.9494741559028625, | |
| "learning_rate": 7.449234057712861e-05, | |
| "loss": 0.9842, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.8564668893814087, | |
| "learning_rate": 7.395796223726397e-05, | |
| "loss": 1.0844, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.7696930170059204, | |
| "learning_rate": 7.342358389739935e-05, | |
| "loss": 1.0693, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "eval_loss": 1.0445717573165894, | |
| "eval_runtime": 243.4168, | |
| "eval_samples_per_second": 46.131, | |
| "eval_steps_per_second": 5.768, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.7331804633140564, | |
| "learning_rate": 7.288920555753472e-05, | |
| "loss": 0.9968, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.696283221244812, | |
| "learning_rate": 7.23548272176701e-05, | |
| "loss": 1.0031, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.9141848683357239, | |
| "learning_rate": 7.182044887780548e-05, | |
| "loss": 1.1274, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.858689546585083, | |
| "learning_rate": 7.128607053794085e-05, | |
| "loss": 0.9971, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "eval_loss": 1.044740080833435, | |
| "eval_runtime": 243.7413, | |
| "eval_samples_per_second": 46.069, | |
| "eval_steps_per_second": 5.76, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.7686411142349243, | |
| "learning_rate": 7.075169219807623e-05, | |
| "loss": 1.035, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.8768565654754639, | |
| "learning_rate": 7.02173138582116e-05, | |
| "loss": 1.0225, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.9776067733764648, | |
| "learning_rate": 6.968293551834698e-05, | |
| "loss": 1.0651, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.8231525421142578, | |
| "learning_rate": 6.914855717848236e-05, | |
| "loss": 1.0761, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "eval_loss": 1.0438101291656494, | |
| "eval_runtime": 242.7745, | |
| "eval_samples_per_second": 46.253, | |
| "eval_steps_per_second": 5.783, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.7318928837776184, | |
| "learning_rate": 6.861417883861773e-05, | |
| "loss": 0.9767, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 1.4240113496780396, | |
| "learning_rate": 6.807980049875311e-05, | |
| "loss": 1.0337, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.8493779301643372, | |
| "learning_rate": 6.754542215888849e-05, | |
| "loss": 1.0163, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.6638810038566589, | |
| "learning_rate": 6.701104381902386e-05, | |
| "loss": 1.1004, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "eval_loss": 1.0434257984161377, | |
| "eval_runtime": 242.9626, | |
| "eval_samples_per_second": 46.217, | |
| "eval_steps_per_second": 5.779, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 1.0353915691375732, | |
| "learning_rate": 6.647666547915924e-05, | |
| "loss": 1.0303, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.8497571349143982, | |
| "learning_rate": 6.594228713929461e-05, | |
| "loss": 1.0088, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.8979742527008057, | |
| "learning_rate": 6.540790879942999e-05, | |
| "loss": 1.0702, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.791603147983551, | |
| "learning_rate": 6.487353045956537e-05, | |
| "loss": 1.0408, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "eval_loss": 1.0422961711883545, | |
| "eval_runtime": 243.1773, | |
| "eval_samples_per_second": 46.176, | |
| "eval_steps_per_second": 5.774, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.6958451271057129, | |
| "learning_rate": 6.433915211970074e-05, | |
| "loss": 1.0654, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.7097263932228088, | |
| "learning_rate": 6.380477377983612e-05, | |
| "loss": 1.0529, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.6926958560943604, | |
| "learning_rate": 6.32703954399715e-05, | |
| "loss": 1.0213, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.8106698989868164, | |
| "learning_rate": 6.273601710010687e-05, | |
| "loss": 1.0025, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "eval_loss": 1.0433177947998047, | |
| "eval_runtime": 243.1578, | |
| "eval_samples_per_second": 46.18, | |
| "eval_steps_per_second": 5.774, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.7857884168624878, | |
| "learning_rate": 6.220163876024225e-05, | |
| "loss": 0.9925, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.9492070078849792, | |
| "learning_rate": 6.166726042037762e-05, | |
| "loss": 1.0084, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.7410191297531128, | |
| "learning_rate": 6.1132882080513e-05, | |
| "loss": 1.0947, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.8114225268363953, | |
| "learning_rate": 6.0598503740648375e-05, | |
| "loss": 1.0484, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "eval_loss": 1.044475793838501, | |
| "eval_runtime": 243.0705, | |
| "eval_samples_per_second": 46.196, | |
| "eval_steps_per_second": 5.776, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.9174725413322449, | |
| "learning_rate": 6.006412540078375e-05, | |
| "loss": 1.0192, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.96259605884552, | |
| "learning_rate": 5.952974706091913e-05, | |
| "loss": 1.105, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.8854503631591797, | |
| "learning_rate": 5.8995368721054496e-05, | |
| "loss": 1.0242, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.8376172184944153, | |
| "learning_rate": 5.846099038118988e-05, | |
| "loss": 1.0875, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "eval_loss": 1.040363073348999, | |
| "eval_runtime": 243.0906, | |
| "eval_samples_per_second": 46.193, | |
| "eval_steps_per_second": 5.776, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.6750116944313049, | |
| "learning_rate": 5.7926612041325255e-05, | |
| "loss": 1.11, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 1.060460090637207, | |
| "learning_rate": 5.739223370146063e-05, | |
| "loss": 1.0577, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.9560518860816956, | |
| "learning_rate": 5.6857855361596e-05, | |
| "loss": 1.0264, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 1.3398056030273438, | |
| "learning_rate": 5.632347702173138e-05, | |
| "loss": 1.0567, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "eval_loss": 1.0413881540298462, | |
| "eval_runtime": 244.0341, | |
| "eval_samples_per_second": 46.014, | |
| "eval_steps_per_second": 5.753, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 1.0977717638015747, | |
| "learning_rate": 5.578909868186676e-05, | |
| "loss": 1.0415, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.8628262281417847, | |
| "learning_rate": 5.525472034200213e-05, | |
| "loss": 1.068, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 1.3997876644134521, | |
| "learning_rate": 5.4720342002137505e-05, | |
| "loss": 1.1642, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.9169967174530029, | |
| "learning_rate": 5.418596366227289e-05, | |
| "loss": 1.0511, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "eval_loss": 1.0395700931549072, | |
| "eval_runtime": 243.4818, | |
| "eval_samples_per_second": 46.118, | |
| "eval_steps_per_second": 5.766, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.7814826369285583, | |
| "learning_rate": 5.3651585322408264e-05, | |
| "loss": 0.9881, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.747515082359314, | |
| "learning_rate": 5.311720698254363e-05, | |
| "loss": 1.0909, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.7690112590789795, | |
| "learning_rate": 5.258282864267901e-05, | |
| "loss": 1.0002, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.9121204614639282, | |
| "learning_rate": 5.204845030281439e-05, | |
| "loss": 1.039, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "eval_loss": 1.0400930643081665, | |
| "eval_runtime": 242.4036, | |
| "eval_samples_per_second": 46.324, | |
| "eval_steps_per_second": 5.792, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.6267992854118347, | |
| "learning_rate": 5.151407196294976e-05, | |
| "loss": 1.0034, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.822167158126831, | |
| "learning_rate": 5.097969362308514e-05, | |
| "loss": 1.0909, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.8151036500930786, | |
| "learning_rate": 5.044531528322051e-05, | |
| "loss": 1.111, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.8981462717056274, | |
| "learning_rate": 4.9910936943355896e-05, | |
| "loss": 1.1096, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "eval_loss": 1.0389798879623413, | |
| "eval_runtime": 242.8318, | |
| "eval_samples_per_second": 46.242, | |
| "eval_steps_per_second": 5.782, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.8009606003761292, | |
| "learning_rate": 4.9376558603491265e-05, | |
| "loss": 1.0797, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.7975730299949646, | |
| "learning_rate": 4.884218026362664e-05, | |
| "loss": 1.1148, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.7241278290748596, | |
| "learning_rate": 4.8307801923762024e-05, | |
| "loss": 1.0344, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 1.057424545288086, | |
| "learning_rate": 4.7773423583897394e-05, | |
| "loss": 1.0552, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "eval_loss": 1.0383328199386597, | |
| "eval_runtime": 242.6213, | |
| "eval_samples_per_second": 46.282, | |
| "eval_steps_per_second": 5.787, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.554969847202301, | |
| "learning_rate": 4.723904524403277e-05, | |
| "loss": 1.0223, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.5786603093147278, | |
| "learning_rate": 4.6704666904168146e-05, | |
| "loss": 0.9964, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.9003456234931946, | |
| "learning_rate": 4.617028856430353e-05, | |
| "loss": 1.1079, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.770336925983429, | |
| "learning_rate": 4.56359102244389e-05, | |
| "loss": 1.0313, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "eval_loss": 1.038406252861023, | |
| "eval_runtime": 243.2381, | |
| "eval_samples_per_second": 46.165, | |
| "eval_steps_per_second": 5.772, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.7360727787017822, | |
| "learning_rate": 4.5101531884574274e-05, | |
| "loss": 1.0323, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.8868034482002258, | |
| "learning_rate": 4.456715354470965e-05, | |
| "loss": 1.0505, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.797825276851654, | |
| "learning_rate": 4.403277520484503e-05, | |
| "loss": 1.0687, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.6810696721076965, | |
| "learning_rate": 4.34983968649804e-05, | |
| "loss": 1.0873, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "eval_loss": 1.039219856262207, | |
| "eval_runtime": 243.163, | |
| "eval_samples_per_second": 46.179, | |
| "eval_steps_per_second": 5.774, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.7306551933288574, | |
| "learning_rate": 4.296401852511578e-05, | |
| "loss": 1.0555, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 1.3633157014846802, | |
| "learning_rate": 4.2429640185251154e-05, | |
| "loss": 1.0211, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 1.0363646745681763, | |
| "learning_rate": 4.1895261845386524e-05, | |
| "loss": 0.9987, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 1.359657645225525, | |
| "learning_rate": 4.1360883505521907e-05, | |
| "loss": 1.0497, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "eval_loss": 1.0370477437973022, | |
| "eval_runtime": 242.9392, | |
| "eval_samples_per_second": 46.221, | |
| "eval_steps_per_second": 5.779, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 1.0343430042266846, | |
| "learning_rate": 4.082650516565728e-05, | |
| "loss": 1.0131, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.6140369176864624, | |
| "learning_rate": 4.029212682579266e-05, | |
| "loss": 1.0307, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 1.0775642395019531, | |
| "learning_rate": 3.9757748485928035e-05, | |
| "loss": 1.1141, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.9400127530097961, | |
| "learning_rate": 3.922337014606341e-05, | |
| "loss": 1.0086, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "eval_loss": 1.0370286703109741, | |
| "eval_runtime": 242.1391, | |
| "eval_samples_per_second": 46.374, | |
| "eval_steps_per_second": 5.798, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.8459872603416443, | |
| "learning_rate": 3.868899180619879e-05, | |
| "loss": 1.0375, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.6824354529380798, | |
| "learning_rate": 3.8154613466334156e-05, | |
| "loss": 0.9693, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.6828789710998535, | |
| "learning_rate": 3.762023512646954e-05, | |
| "loss": 1.0693, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 1.1990960836410522, | |
| "learning_rate": 3.7085856786604915e-05, | |
| "loss": 1.0551, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "eval_loss": 1.0373618602752686, | |
| "eval_runtime": 244.1596, | |
| "eval_samples_per_second": 45.99, | |
| "eval_steps_per_second": 5.75, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.9297399520874023, | |
| "learning_rate": 3.655147844674029e-05, | |
| "loss": 1.0265, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.5350477695465088, | |
| "learning_rate": 3.601710010687567e-05, | |
| "loss": 1.0933, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.795394778251648, | |
| "learning_rate": 3.5482721767011043e-05, | |
| "loss": 1.0517, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.8578258752822876, | |
| "learning_rate": 3.494834342714641e-05, | |
| "loss": 0.9656, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "eval_loss": 1.0366795063018799, | |
| "eval_runtime": 243.2195, | |
| "eval_samples_per_second": 46.168, | |
| "eval_steps_per_second": 5.773, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.7419169545173645, | |
| "learning_rate": 3.4413965087281796e-05, | |
| "loss": 1.045, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.700157880783081, | |
| "learning_rate": 3.3879586747417165e-05, | |
| "loss": 1.005, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 1.1453505754470825, | |
| "learning_rate": 3.334520840755255e-05, | |
| "loss": 1.0606, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 1.1287420988082886, | |
| "learning_rate": 3.281083006768792e-05, | |
| "loss": 1.0068, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "eval_loss": 1.0359498262405396, | |
| "eval_runtime": 242.9315, | |
| "eval_samples_per_second": 46.223, | |
| "eval_steps_per_second": 5.779, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.7225552797317505, | |
| "learning_rate": 3.22764517278233e-05, | |
| "loss": 1.0658, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.6859227418899536, | |
| "learning_rate": 3.174207338795867e-05, | |
| "loss": 1.0376, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.9955888390541077, | |
| "learning_rate": 3.1207695048094045e-05, | |
| "loss": 1.1427, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.6565488576889038, | |
| "learning_rate": 3.067331670822942e-05, | |
| "loss": 1.0982, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "eval_loss": 1.0358622074127197, | |
| "eval_runtime": 243.3766, | |
| "eval_samples_per_second": 46.138, | |
| "eval_steps_per_second": 5.769, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.7512806057929993, | |
| "learning_rate": 3.0138938368364797e-05, | |
| "loss": 0.9968, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.8107959628105164, | |
| "learning_rate": 2.9604560028500177e-05, | |
| "loss": 1.0999, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.6686025857925415, | |
| "learning_rate": 2.907018168863555e-05, | |
| "loss": 0.9487, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 1.282087802886963, | |
| "learning_rate": 2.853580334877093e-05, | |
| "loss": 1.0322, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "eval_loss": 1.0352236032485962, | |
| "eval_runtime": 243.0812, | |
| "eval_samples_per_second": 46.194, | |
| "eval_steps_per_second": 5.776, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.9066522121429443, | |
| "learning_rate": 2.8001425008906302e-05, | |
| "loss": 1.1081, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.8009347319602966, | |
| "learning_rate": 2.7467046669041678e-05, | |
| "loss": 1.0118, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.7938979864120483, | |
| "learning_rate": 2.6932668329177054e-05, | |
| "loss": 1.0318, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.9669837355613708, | |
| "learning_rate": 2.639828998931243e-05, | |
| "loss": 1.0846, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "eval_loss": 1.034826397895813, | |
| "eval_runtime": 243.4608, | |
| "eval_samples_per_second": 46.122, | |
| "eval_steps_per_second": 5.767, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.6504008769989014, | |
| "learning_rate": 2.5863911649447806e-05, | |
| "loss": 1.0308, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.8472669124603271, | |
| "learning_rate": 2.5329533309583182e-05, | |
| "loss": 1.1081, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.6711983680725098, | |
| "learning_rate": 2.479515496971856e-05, | |
| "loss": 0.9521, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 1.2170182466506958, | |
| "learning_rate": 2.4260776629853934e-05, | |
| "loss": 1.0646, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "eval_loss": 1.034729242324829, | |
| "eval_runtime": 242.7006, | |
| "eval_samples_per_second": 46.267, | |
| "eval_steps_per_second": 5.785, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.8178135752677917, | |
| "learning_rate": 2.3726398289989314e-05, | |
| "loss": 1.1114, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.6775932908058167, | |
| "learning_rate": 2.3192019950124686e-05, | |
| "loss": 1.0224, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.634033203125, | |
| "learning_rate": 2.265764161026006e-05, | |
| "loss": 1.0107, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 1.1500298976898193, | |
| "learning_rate": 2.212326327039544e-05, | |
| "loss": 1.0405, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "eval_loss": 1.0347503423690796, | |
| "eval_runtime": 243.9831, | |
| "eval_samples_per_second": 46.024, | |
| "eval_steps_per_second": 5.754, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.7258419990539551, | |
| "learning_rate": 2.158888493053081e-05, | |
| "loss": 1.107, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.6233177185058594, | |
| "learning_rate": 2.105450659066619e-05, | |
| "loss": 1.0495, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.6832902431488037, | |
| "learning_rate": 2.0520128250801567e-05, | |
| "loss": 1.0314, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.795818030834198, | |
| "learning_rate": 1.9985749910936943e-05, | |
| "loss": 1.0524, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "eval_loss": 1.035040259361267, | |
| "eval_runtime": 243.1568, | |
| "eval_samples_per_second": 46.18, | |
| "eval_steps_per_second": 5.774, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.9443835020065308, | |
| "learning_rate": 1.945137157107232e-05, | |
| "loss": 1.0814, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.6237119436264038, | |
| "learning_rate": 1.891699323120769e-05, | |
| "loss": 1.0426, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.7754202485084534, | |
| "learning_rate": 1.838261489134307e-05, | |
| "loss": 1.0111, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.9115328788757324, | |
| "learning_rate": 1.7848236551478447e-05, | |
| "loss": 1.0586, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "eval_loss": 1.034472107887268, | |
| "eval_runtime": 242.6911, | |
| "eval_samples_per_second": 46.269, | |
| "eval_steps_per_second": 5.785, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.9519437551498413, | |
| "learning_rate": 1.731385821161382e-05, | |
| "loss": 1.0329, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 1.142217755317688, | |
| "learning_rate": 1.6779479871749196e-05, | |
| "loss": 1.0159, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.872619092464447, | |
| "learning_rate": 1.6245101531884572e-05, | |
| "loss": 1.0312, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.7944605350494385, | |
| "learning_rate": 1.5710723192019948e-05, | |
| "loss": 1.0029, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "eval_loss": 1.0337694883346558, | |
| "eval_runtime": 242.9274, | |
| "eval_samples_per_second": 46.224, | |
| "eval_steps_per_second": 5.78, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.7382352352142334, | |
| "learning_rate": 1.5176344852155324e-05, | |
| "loss": 1.05, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.706157386302948, | |
| "learning_rate": 1.46419665122907e-05, | |
| "loss": 1.0117, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.6966803669929504, | |
| "learning_rate": 1.4107588172426076e-05, | |
| "loss": 1.0328, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.8918120265007019, | |
| "learning_rate": 1.3573209832561454e-05, | |
| "loss": 1.1869, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "eval_loss": 1.034013271331787, | |
| "eval_runtime": 243.2908, | |
| "eval_samples_per_second": 46.155, | |
| "eval_steps_per_second": 5.771, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.8896795511245728, | |
| "learning_rate": 1.3038831492696827e-05, | |
| "loss": 1.0083, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.7305176258087158, | |
| "learning_rate": 1.2504453152832203e-05, | |
| "loss": 1.0409, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.9530869126319885, | |
| "learning_rate": 1.1970074812967579e-05, | |
| "loss": 1.0329, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.8421928286552429, | |
| "learning_rate": 1.1435696473102957e-05, | |
| "loss": 1.0572, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "eval_loss": 1.0332649946212769, | |
| "eval_runtime": 243.0973, | |
| "eval_samples_per_second": 46.191, | |
| "eval_steps_per_second": 5.775, | |
| "step": 2700 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 2807, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 20, | |
| "total_flos": 2.5170359635714867e+17, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |