| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 4.033884630899556, | |
| "global_step": 90000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.9998599345614274e-05, | |
| "loss": 6.6897, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.999719869122854e-05, | |
| "loss": 5.258, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.999579803684282e-05, | |
| "loss": 4.5571, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.999439738245709e-05, | |
| "loss": 4.277, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9992996728071354e-05, | |
| "loss": 4.0698, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9991596073685625e-05, | |
| "loss": 3.8909, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9990195419299904e-05, | |
| "loss": 3.7887, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.998879476491417e-05, | |
| "loss": 3.7145, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.998739411052844e-05, | |
| "loss": 3.6474, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.998599345614271e-05, | |
| "loss": 3.5479, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9984592801756984e-05, | |
| "loss": 3.4806, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.9983192147371255e-05, | |
| "loss": 3.4368, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.998179149298553e-05, | |
| "loss": 3.4167, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.998039083859979e-05, | |
| "loss": 3.3564, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.997899018421407e-05, | |
| "loss": 3.3145, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.997758952982834e-05, | |
| "loss": 3.2817, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.997618887544261e-05, | |
| "loss": 3.2437, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.997478822105688e-05, | |
| "loss": 3.1763, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.997338756667116e-05, | |
| "loss": 3.1199, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.997198691228542e-05, | |
| "loss": 3.0686, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.997058625789969e-05, | |
| "loss": 3.0862, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.9969185603513965e-05, | |
| "loss": 3.0317, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.996778494912823e-05, | |
| "loss": 2.9852, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.996638429474251e-05, | |
| "loss": 2.9593, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.996498364035678e-05, | |
| "loss": 2.9514, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9963582985971045e-05, | |
| "loss": 2.9288, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9962182331585316e-05, | |
| "loss": 2.8808, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9960781677199595e-05, | |
| "loss": 2.8553, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.995938102281386e-05, | |
| "loss": 2.849, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.995798036842813e-05, | |
| "loss": 2.8272, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.99565797140424e-05, | |
| "loss": 2.8049, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9955179059656674e-05, | |
| "loss": 2.7629, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9953778405270946e-05, | |
| "loss": 2.7371, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.995237775088522e-05, | |
| "loss": 2.728, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.995097709649948e-05, | |
| "loss": 2.6995, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.994957644211376e-05, | |
| "loss": 2.6784, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.994817578772803e-05, | |
| "loss": 2.6507, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.99467751333423e-05, | |
| "loss": 2.6582, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.994537447895657e-05, | |
| "loss": 2.629, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.994397382457085e-05, | |
| "loss": 2.5755, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.994257317018511e-05, | |
| "loss": 2.6007, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.9941172515799384e-05, | |
| "loss": 2.5898, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.9939771861413656e-05, | |
| "loss": 2.5711, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.993837120702793e-05, | |
| "loss": 2.5769, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.99369705526422e-05, | |
| "loss": 2.5123, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.993556989825647e-05, | |
| "loss": 2.471, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.9934169243870735e-05, | |
| "loss": 2.4919, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.993276858948501e-05, | |
| "loss": 2.4754, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.9931367935099285e-05, | |
| "loss": 2.4559, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.992996728071355e-05, | |
| "loss": 2.4689, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.992856662632782e-05, | |
| "loss": 2.3926, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.9927165971942094e-05, | |
| "loss": 2.4048, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.9925765317556365e-05, | |
| "loss": 2.3582, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.992436466317064e-05, | |
| "loss": 2.3653, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.992296400878491e-05, | |
| "loss": 2.3437, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.992156335439917e-05, | |
| "loss": 2.3232, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.992016270001345e-05, | |
| "loss": 2.3, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.991876204562772e-05, | |
| "loss": 2.3017, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.991736139124199e-05, | |
| "loss": 2.2912, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.991596073685626e-05, | |
| "loss": 2.2834, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.991456008247054e-05, | |
| "loss": 2.2803, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.99131594280848e-05, | |
| "loss": 2.2308, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.9911758773699075e-05, | |
| "loss": 2.2106, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.9910358119313346e-05, | |
| "loss": 2.2169, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.990895746492762e-05, | |
| "loss": 2.1805, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.990755681054189e-05, | |
| "loss": 2.1821, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.990615615615616e-05, | |
| "loss": 2.1632, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.9904755501770426e-05, | |
| "loss": 2.1328, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.99033548473847e-05, | |
| "loss": 2.1339, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.9901954192998976e-05, | |
| "loss": 2.0954, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.990055353861324e-05, | |
| "loss": 2.1149, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.989915288422751e-05, | |
| "loss": 2.0974, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.9897752229841784e-05, | |
| "loss": 2.0959, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.9896351575456056e-05, | |
| "loss": 2.0753, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.989495092107033e-05, | |
| "loss": 2.0299, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.98935502666846e-05, | |
| "loss": 2.0623, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.9892149612298864e-05, | |
| "loss": 2.0643, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.989074895791314e-05, | |
| "loss": 2.0094, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.9889348303527414e-05, | |
| "loss": 2.0117, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.988794764914168e-05, | |
| "loss": 2.0083, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.988654699475595e-05, | |
| "loss": 1.9975, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.988514634037023e-05, | |
| "loss": 1.9871, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.9883745685984494e-05, | |
| "loss": 1.9791, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.9882345031598766e-05, | |
| "loss": 1.954, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.988094437721304e-05, | |
| "loss": 1.9414, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.987954372282731e-05, | |
| "loss": 1.9618, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.987814306844158e-05, | |
| "loss": 1.9492, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.987674241405585e-05, | |
| "loss": 1.9234, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.987534175967012e-05, | |
| "loss": 1.95, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.987394110528439e-05, | |
| "loss": 1.91, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.987254045089867e-05, | |
| "loss": 1.895, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.987113979651293e-05, | |
| "loss": 1.9085, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.9869739142127204e-05, | |
| "loss": 1.9381, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.9868338487741475e-05, | |
| "loss": 1.8904, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.986693783335575e-05, | |
| "loss": 1.8767, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.986553717897002e-05, | |
| "loss": 1.89, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.986413652458429e-05, | |
| "loss": 1.8853, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.9862735870198555e-05, | |
| "loss": 1.8423, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.986133521581283e-05, | |
| "loss": 1.8952, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.9859934561427105e-05, | |
| "loss": 1.8386, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.985853390704137e-05, | |
| "loss": 1.8445, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.985713325265564e-05, | |
| "loss": 1.8287, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.985573259826991e-05, | |
| "loss": 1.8311, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.9854331943884185e-05, | |
| "loss": 1.84, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.9852931289498456e-05, | |
| "loss": 1.8407, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.985153063511273e-05, | |
| "loss": 1.8136, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.9850129980727e-05, | |
| "loss": 1.8238, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.984872932634127e-05, | |
| "loss": 1.8085, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.984732867195554e-05, | |
| "loss": 1.8123, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.984592801756981e-05, | |
| "loss": 1.7892, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.984452736318408e-05, | |
| "loss": 1.7986, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.984312670879836e-05, | |
| "loss": 1.7938, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.984172605441262e-05, | |
| "loss": 1.7872, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.9840325400026894e-05, | |
| "loss": 1.78, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.9838924745641166e-05, | |
| "loss": 1.7795, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.983752409125544e-05, | |
| "loss": 1.7643, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.983612343686971e-05, | |
| "loss": 1.7853, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.983472278248398e-05, | |
| "loss": 1.7302, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.9833322128098246e-05, | |
| "loss": 1.7454, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.9831921473712524e-05, | |
| "loss": 1.7523, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.9830520819326796e-05, | |
| "loss": 1.7306, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.982912016494106e-05, | |
| "loss": 1.7585, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.982771951055533e-05, | |
| "loss": 1.7475, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.9826318856169604e-05, | |
| "loss": 1.7738, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.9824918201783876e-05, | |
| "loss": 1.7194, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.982351754739815e-05, | |
| "loss": 1.7241, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.982211689301242e-05, | |
| "loss": 1.7459, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.982071623862669e-05, | |
| "loss": 1.745, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.981931558424096e-05, | |
| "loss": 1.709, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.9817914929855234e-05, | |
| "loss": 1.7029, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.98165142754695e-05, | |
| "loss": 1.7215, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.981511362108378e-05, | |
| "loss": 1.7268, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.981371296669804e-05, | |
| "loss": 1.7104, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.9812312312312313e-05, | |
| "loss": 1.7045, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.9810911657926585e-05, | |
| "loss": 1.7202, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.980951100354086e-05, | |
| "loss": 1.7227, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.980811034915513e-05, | |
| "loss": 1.6956, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.98067096947694e-05, | |
| "loss": 1.6987, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.980530904038367e-05, | |
| "loss": 1.7068, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.9803908385997936e-05, | |
| "loss": 1.6831, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.9802507731612215e-05, | |
| "loss": 1.6836, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.9801107077226487e-05, | |
| "loss": 1.685, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.979970642284075e-05, | |
| "loss": 1.6383, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.979830576845502e-05, | |
| "loss": 1.6637, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.9796905114069295e-05, | |
| "loss": 1.653, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.9795504459683566e-05, | |
| "loss": 1.6456, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.979410380529784e-05, | |
| "loss": 1.6906, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.979270315091211e-05, | |
| "loss": 1.681, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.979130249652638e-05, | |
| "loss": 1.6843, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.978990184214065e-05, | |
| "loss": 1.6591, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.9788501187754924e-05, | |
| "loss": 1.6505, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.978710053336919e-05, | |
| "loss": 1.6683, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.978569987898347e-05, | |
| "loss": 1.6337, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.978429922459773e-05, | |
| "loss": 1.6364, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.9782898570212004e-05, | |
| "loss": 1.6822, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.9781497915826276e-05, | |
| "loss": 1.6316, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.978009726144055e-05, | |
| "loss": 1.6467, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.977869660705482e-05, | |
| "loss": 1.6441, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.977729595266909e-05, | |
| "loss": 1.62, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.977589529828336e-05, | |
| "loss": 1.6345, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.977449464389763e-05, | |
| "loss": 1.6474, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.9773093989511906e-05, | |
| "loss": 1.6027, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.977169333512617e-05, | |
| "loss": 1.6124, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.977029268074044e-05, | |
| "loss": 1.641, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.9768892026354714e-05, | |
| "loss": 1.6292, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.9767491371968985e-05, | |
| "loss": 1.6153, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.976609071758326e-05, | |
| "loss": 1.6272, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.976469006319753e-05, | |
| "loss": 1.616, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.97632894088118e-05, | |
| "loss": 1.6062, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.976188875442607e-05, | |
| "loss": 1.5941, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.9760488100040344e-05, | |
| "loss": 1.611, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.975908744565461e-05, | |
| "loss": 1.6176, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.975768679126888e-05, | |
| "loss": 1.5742, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.975628613688316e-05, | |
| "loss": 1.5945, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.9754885482497423e-05, | |
| "loss": 1.5949, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.9753484828111695e-05, | |
| "loss": 1.6093, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.975208417372597e-05, | |
| "loss": 1.6421, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.975068351934024e-05, | |
| "loss": 1.5867, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.974928286495451e-05, | |
| "loss": 1.5893, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.974788221056878e-05, | |
| "loss": 1.6043, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.974648155618305e-05, | |
| "loss": 1.577, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.974508090179732e-05, | |
| "loss": 1.5824, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.9743680247411596e-05, | |
| "loss": 1.5603, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.974227959302586e-05, | |
| "loss": 1.5701, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.974087893864013e-05, | |
| "loss": 1.5578, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.9739478284254405e-05, | |
| "loss": 1.5759, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.9738077629868676e-05, | |
| "loss": 1.5798, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.973667697548295e-05, | |
| "loss": 1.5541, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.973527632109722e-05, | |
| "loss": 1.5648, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.973387566671149e-05, | |
| "loss": 1.5453, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.973247501232576e-05, | |
| "loss": 1.5376, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.9731074357940034e-05, | |
| "loss": 1.5535, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.97296737035543e-05, | |
| "loss": 1.5799, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.972827304916857e-05, | |
| "loss": 1.5483, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.972687239478285e-05, | |
| "loss": 1.5823, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.9725471740397114e-05, | |
| "loss": 1.5499, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.9724071086011386e-05, | |
| "loss": 1.5751, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.972267043162566e-05, | |
| "loss": 1.5588, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.972126977723993e-05, | |
| "loss": 1.5987, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.97198691228542e-05, | |
| "loss": 1.5269, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.971846846846847e-05, | |
| "loss": 1.5497, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.971706781408274e-05, | |
| "loss": 1.5329, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.971566715969701e-05, | |
| "loss": 1.5383, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.971426650531129e-05, | |
| "loss": 1.5511, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.971286585092555e-05, | |
| "loss": 1.5802, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.9711465196539824e-05, | |
| "loss": 1.5579, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.9710064542154095e-05, | |
| "loss": 1.553, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.970866388776837e-05, | |
| "loss": 1.531, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.970726323338264e-05, | |
| "loss": 1.5311, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.970586257899691e-05, | |
| "loss": 1.5493, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.970446192461118e-05, | |
| "loss": 1.5232, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.9703061270225454e-05, | |
| "loss": 1.5012, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.9701660615839725e-05, | |
| "loss": 1.5243, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.970025996145399e-05, | |
| "loss": 1.524, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.969885930706826e-05, | |
| "loss": 1.5345, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.969745865268254e-05, | |
| "loss": 1.5402, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.9696057998296805e-05, | |
| "loss": 1.5259, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.969465734391108e-05, | |
| "loss": 1.5375, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.969325668952535e-05, | |
| "loss": 1.5078, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.969185603513962e-05, | |
| "loss": 1.5454, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.969045538075389e-05, | |
| "loss": 1.5347, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.968905472636816e-05, | |
| "loss": 1.5035, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.968765407198243e-05, | |
| "loss": 1.5242, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.96862534175967e-05, | |
| "loss": 1.5151, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.968485276321098e-05, | |
| "loss": 1.5183, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.968345210882524e-05, | |
| "loss": 1.5342, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.9682051454439515e-05, | |
| "loss": 1.5023, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.9680650800053786e-05, | |
| "loss": 1.5185, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.967925014566806e-05, | |
| "loss": 1.5249, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.967784949128233e-05, | |
| "loss": 1.5236, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.96764488368966e-05, | |
| "loss": 1.5185, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.9675048182510866e-05, | |
| "loss": 1.5063, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.9673647528125144e-05, | |
| "loss": 1.5004, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.9672246873739416e-05, | |
| "loss": 1.5408, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.967084621935368e-05, | |
| "loss": 1.4791, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.966944556496795e-05, | |
| "loss": 1.5187, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.966804491058223e-05, | |
| "loss": 1.5252, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.9666644256196496e-05, | |
| "loss": 1.4862, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.966524360181077e-05, | |
| "loss": 1.4958, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.966384294742504e-05, | |
| "loss": 1.4896, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.966244229303931e-05, | |
| "loss": 1.468, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.966104163865358e-05, | |
| "loss": 1.4898, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.9659640984267854e-05, | |
| "loss": 1.5018, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.965824032988212e-05, | |
| "loss": 1.5111, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.965683967549639e-05, | |
| "loss": 1.4865, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.965543902111067e-05, | |
| "loss": 1.4759, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.9654038366724934e-05, | |
| "loss": 1.4959, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.9652637712339205e-05, | |
| "loss": 1.5015, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.965123705795348e-05, | |
| "loss": 1.4925, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.964983640356775e-05, | |
| "loss": 1.4625, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.964843574918202e-05, | |
| "loss": 1.4867, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.964703509479629e-05, | |
| "loss": 1.4986, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.964563444041056e-05, | |
| "loss": 1.4704, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.9644233786024835e-05, | |
| "loss": 1.5, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.964283313163911e-05, | |
| "loss": 1.4678, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.964143247725337e-05, | |
| "loss": 1.4832, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.964003182286764e-05, | |
| "loss": 1.4965, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.963863116848192e-05, | |
| "loss": 1.4448, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.9637230514096187e-05, | |
| "loss": 1.4972, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.963582985971046e-05, | |
| "loss": 1.4942, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.963442920532473e-05, | |
| "loss": 1.4511, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.9633028550939e-05, | |
| "loss": 1.4861, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.963162789655327e-05, | |
| "loss": 1.4808, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.9630227242167545e-05, | |
| "loss": 1.458, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.962882658778181e-05, | |
| "loss": 1.4674, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.962742593339608e-05, | |
| "loss": 1.4772, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.962602527901036e-05, | |
| "loss": 1.4613, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.9624624624624625e-05, | |
| "loss": 1.4683, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.9623223970238896e-05, | |
| "loss": 1.4877, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.962182331585317e-05, | |
| "loss": 1.448, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.962042266146744e-05, | |
| "loss": 1.471, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.961902200708171e-05, | |
| "loss": 1.4657, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.961762135269598e-05, | |
| "loss": 1.4699, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.961622069831025e-05, | |
| "loss": 1.4696, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.9614820043924526e-05, | |
| "loss": 1.4938, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.96134193895388e-05, | |
| "loss": 1.448, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.961201873515306e-05, | |
| "loss": 1.4607, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.9610618080767334e-05, | |
| "loss": 1.4393, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.960921742638161e-05, | |
| "loss": 1.4952, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.960781677199588e-05, | |
| "loss": 1.4469, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.960641611761015e-05, | |
| "loss": 1.4689, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.960501546322442e-05, | |
| "loss": 1.4632, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.960361480883869e-05, | |
| "loss": 1.4726, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.9602214154452964e-05, | |
| "loss": 1.4663, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.9600813500067236e-05, | |
| "loss": 1.4647, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.95994128456815e-05, | |
| "loss": 1.4282, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.959801219129578e-05, | |
| "loss": 1.4628, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.959661153691005e-05, | |
| "loss": 1.4443, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.9595210882524315e-05, | |
| "loss": 1.4242, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.959381022813859e-05, | |
| "loss": 1.4404, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.959240957375286e-05, | |
| "loss": 1.4191, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.959100891936713e-05, | |
| "loss": 1.4304, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.95896082649814e-05, | |
| "loss": 1.4385, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.9588207610595674e-05, | |
| "loss": 1.4845, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.958680695620994e-05, | |
| "loss": 1.4561, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.958540630182422e-05, | |
| "loss": 1.4174, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.958400564743849e-05, | |
| "loss": 1.4426, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.958260499305275e-05, | |
| "loss": 1.4449, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.9581204338667025e-05, | |
| "loss": 1.4493, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.95798036842813e-05, | |
| "loss": 1.4344, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.957840302989557e-05, | |
| "loss": 1.4331, | |
| "step": 15050 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.957700237550984e-05, | |
| "loss": 1.4231, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.957560172112411e-05, | |
| "loss": 1.4466, | |
| "step": 15150 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.957420106673838e-05, | |
| "loss": 1.4208, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.9572800412352655e-05, | |
| "loss": 1.4481, | |
| "step": 15250 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.9571399757966926e-05, | |
| "loss": 1.4266, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.956999910358119e-05, | |
| "loss": 1.4182, | |
| "step": 15350 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.956859844919547e-05, | |
| "loss": 1.439, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.956719779480974e-05, | |
| "loss": 1.4602, | |
| "step": 15450 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.9565797140424006e-05, | |
| "loss": 1.4321, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.956439648603828e-05, | |
| "loss": 1.4211, | |
| "step": 15550 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.956299583165255e-05, | |
| "loss": 1.4529, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.956159517726682e-05, | |
| "loss": 1.4812, | |
| "step": 15650 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.956019452288109e-05, | |
| "loss": 1.429, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.9558793868495364e-05, | |
| "loss": 1.4248, | |
| "step": 15750 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.955739321410963e-05, | |
| "loss": 1.4457, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.955599255972391e-05, | |
| "loss": 1.4423, | |
| "step": 15850 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.955459190533818e-05, | |
| "loss": 1.462, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.9553191250952444e-05, | |
| "loss": 1.4297, | |
| "step": 15950 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.9551790596566716e-05, | |
| "loss": 1.4194, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.9550389942180994e-05, | |
| "loss": 1.4264, | |
| "step": 16050 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.954898928779526e-05, | |
| "loss": 1.4297, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.954758863340953e-05, | |
| "loss": 1.4422, | |
| "step": 16150 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.95461879790238e-05, | |
| "loss": 1.427, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.9544787324638074e-05, | |
| "loss": 1.4065, | |
| "step": 16250 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.9543386670252346e-05, | |
| "loss": 1.4456, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.954198601586662e-05, | |
| "loss": 1.4264, | |
| "step": 16350 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.954058536148088e-05, | |
| "loss": 1.4434, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.953918470709516e-05, | |
| "loss": 1.4155, | |
| "step": 16450 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.953778405270943e-05, | |
| "loss": 1.4198, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.95363833983237e-05, | |
| "loss": 1.4149, | |
| "step": 16550 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.953498274393797e-05, | |
| "loss": 1.4168, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.953358208955224e-05, | |
| "loss": 1.4074, | |
| "step": 16650 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.953218143516651e-05, | |
| "loss": 1.4294, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.9530780780780783e-05, | |
| "loss": 1.434, | |
| "step": 16750 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.9529380126395055e-05, | |
| "loss": 1.4247, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.952797947200932e-05, | |
| "loss": 1.4213, | |
| "step": 16850 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.95265788176236e-05, | |
| "loss": 1.4084, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.952517816323787e-05, | |
| "loss": 1.4367, | |
| "step": 16950 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.9523777508852135e-05, | |
| "loss": 1.3964, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.9522376854466406e-05, | |
| "loss": 1.4384, | |
| "step": 17050 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.9520976200080685e-05, | |
| "loss": 1.3951, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.951957554569495e-05, | |
| "loss": 1.4163, | |
| "step": 17150 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.951817489130922e-05, | |
| "loss": 1.4204, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.951677423692349e-05, | |
| "loss": 1.3999, | |
| "step": 17250 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.9515373582537765e-05, | |
| "loss": 1.3874, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.9513972928152036e-05, | |
| "loss": 1.3935, | |
| "step": 17350 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.951257227376631e-05, | |
| "loss": 1.3938, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.951117161938057e-05, | |
| "loss": 1.4439, | |
| "step": 17450 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.950977096499485e-05, | |
| "loss": 1.4088, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.950837031060912e-05, | |
| "loss": 1.4006, | |
| "step": 17550 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.950696965622339e-05, | |
| "loss": 1.4287, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.950556900183766e-05, | |
| "loss": 1.3972, | |
| "step": 17650 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.950416834745194e-05, | |
| "loss": 1.3971, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.95027676930662e-05, | |
| "loss": 1.4028, | |
| "step": 17750 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.9501367038680474e-05, | |
| "loss": 1.4041, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.9499966384294746e-05, | |
| "loss": 1.3915, | |
| "step": 17850 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.949856572990901e-05, | |
| "loss": 1.4005, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.949716507552329e-05, | |
| "loss": 1.4176, | |
| "step": 17950 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.949576442113756e-05, | |
| "loss": 1.3866, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.9494363766751826e-05, | |
| "loss": 1.3905, | |
| "step": 18050 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.94929631123661e-05, | |
| "loss": 1.3843, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.9491562457980376e-05, | |
| "loss": 1.4094, | |
| "step": 18150 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.949016180359464e-05, | |
| "loss": 1.3784, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.948876114920891e-05, | |
| "loss": 1.4126, | |
| "step": 18250 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.9487360494823184e-05, | |
| "loss": 1.3982, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.9485959840437455e-05, | |
| "loss": 1.3894, | |
| "step": 18350 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.948455918605173e-05, | |
| "loss": 1.4069, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.9483158531666e-05, | |
| "loss": 1.3994, | |
| "step": 18450 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.9481757877280264e-05, | |
| "loss": 1.3741, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.948035722289454e-05, | |
| "loss": 1.3939, | |
| "step": 18550 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.9478956568508814e-05, | |
| "loss": 1.3632, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.947755591412308e-05, | |
| "loss": 1.3861, | |
| "step": 18650 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.947615525973735e-05, | |
| "loss": 1.4011, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.947475460535163e-05, | |
| "loss": 1.421, | |
| "step": 18750 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.9473353950965893e-05, | |
| "loss": 1.3762, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.9471953296580165e-05, | |
| "loss": 1.378, | |
| "step": 18850 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.947055264219444e-05, | |
| "loss": 1.3635, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.94691519878087e-05, | |
| "loss": 1.3582, | |
| "step": 18950 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.946775133342298e-05, | |
| "loss": 1.3744, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.946635067903725e-05, | |
| "loss": 1.3787, | |
| "step": 19050 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.9464950024651516e-05, | |
| "loss": 1.3899, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.946354937026579e-05, | |
| "loss": 1.3576, | |
| "step": 19150 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.9462148715880066e-05, | |
| "loss": 1.369, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.946074806149433e-05, | |
| "loss": 1.3648, | |
| "step": 19250 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.94593474071086e-05, | |
| "loss": 1.3824, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.9457946752722875e-05, | |
| "loss": 1.3722, | |
| "step": 19350 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.9456546098337146e-05, | |
| "loss": 1.3762, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.945514544395142e-05, | |
| "loss": 1.3751, | |
| "step": 19450 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.945374478956569e-05, | |
| "loss": 1.408, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.9452344135179954e-05, | |
| "loss": 1.3693, | |
| "step": 19550 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.945094348079423e-05, | |
| "loss": 1.357, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.9449542826408504e-05, | |
| "loss": 1.3611, | |
| "step": 19650 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.944814217202277e-05, | |
| "loss": 1.3758, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.944674151763704e-05, | |
| "loss": 1.3733, | |
| "step": 19750 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.944534086325132e-05, | |
| "loss": 1.3941, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.9443940208865584e-05, | |
| "loss": 1.3382, | |
| "step": 19850 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.9442539554479856e-05, | |
| "loss": 1.3743, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.944113890009413e-05, | |
| "loss": 1.4127, | |
| "step": 19950 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.943973824570839e-05, | |
| "loss": 1.3587, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.943833759132267e-05, | |
| "loss": 1.3634, | |
| "step": 20050 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.943693693693694e-05, | |
| "loss": 1.3624, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.943553628255121e-05, | |
| "loss": 1.3669, | |
| "step": 20150 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.943413562816548e-05, | |
| "loss": 1.385, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.943273497377976e-05, | |
| "loss": 1.3645, | |
| "step": 20250 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.943133431939402e-05, | |
| "loss": 1.3695, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.9429933665008294e-05, | |
| "loss": 1.339, | |
| "step": 20350 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.9428533010622565e-05, | |
| "loss": 1.3919, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.942713235623684e-05, | |
| "loss": 1.3446, | |
| "step": 20450 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.942573170185111e-05, | |
| "loss": 1.3645, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.942433104746538e-05, | |
| "loss": 1.3542, | |
| "step": 20550 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.9422930393079645e-05, | |
| "loss": 1.3525, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.9421529738693924e-05, | |
| "loss": 1.3822, | |
| "step": 20650 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.9420129084308195e-05, | |
| "loss": 1.3757, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.941872842992246e-05, | |
| "loss": 1.3809, | |
| "step": 20750 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.941732777553673e-05, | |
| "loss": 1.3601, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.9415927121151e-05, | |
| "loss": 1.3926, | |
| "step": 20850 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.9414526466765275e-05, | |
| "loss": 1.3775, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.941312581237955e-05, | |
| "loss": 1.379, | |
| "step": 20950 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.941172515799382e-05, | |
| "loss": 1.376, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.941032450360808e-05, | |
| "loss": 1.3769, | |
| "step": 21050 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.940892384922236e-05, | |
| "loss": 1.3567, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.940752319483663e-05, | |
| "loss": 1.363, | |
| "step": 21150 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.94061225404509e-05, | |
| "loss": 1.3813, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.940472188606517e-05, | |
| "loss": 1.3464, | |
| "step": 21250 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.940332123167945e-05, | |
| "loss": 1.3706, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.940192057729371e-05, | |
| "loss": 1.3981, | |
| "step": 21350 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.9400519922907985e-05, | |
| "loss": 1.3589, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.9399119268522256e-05, | |
| "loss": 1.3631, | |
| "step": 21450 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.939771861413653e-05, | |
| "loss": 1.3687, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.93963179597508e-05, | |
| "loss": 1.3863, | |
| "step": 21550 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.939491730536507e-05, | |
| "loss": 1.3843, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.9393516650979336e-05, | |
| "loss": 1.3488, | |
| "step": 21650 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.9392115996593614e-05, | |
| "loss": 1.3388, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.9390715342207886e-05, | |
| "loss": 1.3669, | |
| "step": 21750 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.938931468782215e-05, | |
| "loss": 1.3672, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.938791403343642e-05, | |
| "loss": 1.358, | |
| "step": 21850 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.9386513379050694e-05, | |
| "loss": 1.3578, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 4.9385112724664966e-05, | |
| "loss": 1.3651, | |
| "step": 21950 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.938371207027924e-05, | |
| "loss": 1.3612, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.938231141589351e-05, | |
| "loss": 1.3578, | |
| "step": 22050 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.938091076150778e-05, | |
| "loss": 1.3665, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.937951010712205e-05, | |
| "loss": 1.338, | |
| "step": 22150 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.9378109452736324e-05, | |
| "loss": 1.3644, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.937670879835059e-05, | |
| "loss": 1.3763, | |
| "step": 22250 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.937530814396486e-05, | |
| "loss": 1.367, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.937390748957913e-05, | |
| "loss": 1.3693, | |
| "step": 22350 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 4.9372506835193404e-05, | |
| "loss": 1.3245, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.9371106180807675e-05, | |
| "loss": 1.3796, | |
| "step": 22450 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.936970552642195e-05, | |
| "loss": 1.3595, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.936830487203622e-05, | |
| "loss": 1.3561, | |
| "step": 22550 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.936690421765049e-05, | |
| "loss": 1.3485, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.936550356326476e-05, | |
| "loss": 1.3683, | |
| "step": 22650 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.936410290887903e-05, | |
| "loss": 1.3529, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.9362702254493305e-05, | |
| "loss": 1.3511, | |
| "step": 22750 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.936130160010758e-05, | |
| "loss": 1.3567, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.935990094572184e-05, | |
| "loss": 1.3715, | |
| "step": 22850 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.935850029133611e-05, | |
| "loss": 1.3675, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.9357099636950385e-05, | |
| "loss": 1.3675, | |
| "step": 22950 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.9355698982564657e-05, | |
| "loss": 1.3236, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.935429832817893e-05, | |
| "loss": 1.3666, | |
| "step": 23050 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.93528976737932e-05, | |
| "loss": 1.3429, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.935149701940747e-05, | |
| "loss": 1.3192, | |
| "step": 23150 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.935009636502174e-05, | |
| "loss": 1.3289, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.9348695710636015e-05, | |
| "loss": 1.3412, | |
| "step": 23250 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.934729505625028e-05, | |
| "loss": 1.3848, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.934589440186455e-05, | |
| "loss": 1.3172, | |
| "step": 23350 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.934449374747882e-05, | |
| "loss": 1.3364, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.9343093093093095e-05, | |
| "loss": 1.3367, | |
| "step": 23450 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.9341692438707366e-05, | |
| "loss": 1.3251, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.934029178432164e-05, | |
| "loss": 1.3753, | |
| "step": 23550 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.933889112993591e-05, | |
| "loss": 1.35, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.933749047555018e-05, | |
| "loss": 1.3284, | |
| "step": 23650 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.933608982116445e-05, | |
| "loss": 1.363, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.933468916677872e-05, | |
| "loss": 1.3063, | |
| "step": 23750 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.9333288512392996e-05, | |
| "loss": 1.3346, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.933188785800726e-05, | |
| "loss": 1.3458, | |
| "step": 23850 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.933048720362153e-05, | |
| "loss": 1.3224, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.9329086549235804e-05, | |
| "loss": 1.3349, | |
| "step": 23950 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.9327685894850076e-05, | |
| "loss": 1.3494, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.932628524046435e-05, | |
| "loss": 1.3102, | |
| "step": 24050 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.932488458607862e-05, | |
| "loss": 1.2964, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.932348393169289e-05, | |
| "loss": 1.341, | |
| "step": 24150 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.932208327730716e-05, | |
| "loss": 1.3287, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.9320682622921434e-05, | |
| "loss": 1.3267, | |
| "step": 24250 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.93192819685357e-05, | |
| "loss": 1.3072, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.931788131414997e-05, | |
| "loss": 1.315, | |
| "step": 24350 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 4.931648065976424e-05, | |
| "loss": 1.3177, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.9315080005378514e-05, | |
| "loss": 1.3313, | |
| "step": 24450 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.9313679350992785e-05, | |
| "loss": 1.3679, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.931227869660706e-05, | |
| "loss": 1.3505, | |
| "step": 24550 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.931087804222133e-05, | |
| "loss": 1.324, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 4.93094773878356e-05, | |
| "loss": 1.3178, | |
| "step": 24650 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.930807673344987e-05, | |
| "loss": 1.3465, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.9306676079064143e-05, | |
| "loss": 1.3618, | |
| "step": 24750 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.930527542467841e-05, | |
| "loss": 1.3517, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.930387477029269e-05, | |
| "loss": 1.3166, | |
| "step": 24850 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.930247411590695e-05, | |
| "loss": 1.329, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.930107346152122e-05, | |
| "loss": 1.3431, | |
| "step": 24950 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.9299672807135495e-05, | |
| "loss": 1.3152, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.9298272152749767e-05, | |
| "loss": 1.3264, | |
| "step": 25050 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.929687149836404e-05, | |
| "loss": 1.3168, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.929547084397831e-05, | |
| "loss": 1.3626, | |
| "step": 25150 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.929407018959258e-05, | |
| "loss": 1.3371, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.929266953520685e-05, | |
| "loss": 1.3076, | |
| "step": 25250 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.9291268880821125e-05, | |
| "loss": 1.3047, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.928986822643539e-05, | |
| "loss": 1.3129, | |
| "step": 25350 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.928846757204966e-05, | |
| "loss": 1.3037, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.928706691766394e-05, | |
| "loss": 1.3246, | |
| "step": 25450 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 4.9285666263278204e-05, | |
| "loss": 1.3469, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.9284265608892476e-05, | |
| "loss": 1.3285, | |
| "step": 25550 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.928286495450675e-05, | |
| "loss": 1.3412, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.928146430012102e-05, | |
| "loss": 1.3245, | |
| "step": 25650 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.928006364573529e-05, | |
| "loss": 1.3652, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.927866299134956e-05, | |
| "loss": 1.3307, | |
| "step": 25750 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 4.927726233696383e-05, | |
| "loss": 1.3644, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 4.92758616825781e-05, | |
| "loss": 1.3049, | |
| "step": 25850 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 4.927446102819238e-05, | |
| "loss": 1.3461, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 4.927306037380664e-05, | |
| "loss": 1.3252, | |
| "step": 25950 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.9271659719420914e-05, | |
| "loss": 1.3157, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.9270259065035186e-05, | |
| "loss": 1.3196, | |
| "step": 26050 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.926885841064946e-05, | |
| "loss": 1.332, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.926745775626373e-05, | |
| "loss": 1.3552, | |
| "step": 26150 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 4.9266057101878e-05, | |
| "loss": 1.3087, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 4.926465644749227e-05, | |
| "loss": 1.3369, | |
| "step": 26250 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 4.9263255793106544e-05, | |
| "loss": 1.3391, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 4.9261855138720815e-05, | |
| "loss": 1.3279, | |
| "step": 26350 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 4.926045448433508e-05, | |
| "loss": 1.3224, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.925905382994935e-05, | |
| "loss": 1.3283, | |
| "step": 26450 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.925765317556363e-05, | |
| "loss": 1.3316, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.9256252521177895e-05, | |
| "loss": 1.307, | |
| "step": 26550 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.925485186679217e-05, | |
| "loss": 1.3205, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.925345121240644e-05, | |
| "loss": 1.326, | |
| "step": 26650 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 4.925205055802071e-05, | |
| "loss": 1.3374, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 4.925064990363498e-05, | |
| "loss": 1.331, | |
| "step": 26750 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 4.9249249249249253e-05, | |
| "loss": 1.3382, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 4.924784859486352e-05, | |
| "loss": 1.3381, | |
| "step": 26850 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 4.924644794047779e-05, | |
| "loss": 1.3337, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 4.924504728609207e-05, | |
| "loss": 1.321, | |
| "step": 26950 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 4.924364663170633e-05, | |
| "loss": 1.3078, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 4.9242245977320605e-05, | |
| "loss": 1.3166, | |
| "step": 27050 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 4.9240845322934876e-05, | |
| "loss": 1.3202, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 4.923944466854915e-05, | |
| "loss": 1.2926, | |
| "step": 27150 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 4.923804401416342e-05, | |
| "loss": 1.3176, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 4.923664335977769e-05, | |
| "loss": 1.2848, | |
| "step": 27250 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 4.9235242705391956e-05, | |
| "loss": 1.2999, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 4.9233842051006235e-05, | |
| "loss": 1.3205, | |
| "step": 27350 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 4.9232441396620506e-05, | |
| "loss": 1.2959, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 4.923104074223477e-05, | |
| "loss": 1.3133, | |
| "step": 27450 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 4.922964008784904e-05, | |
| "loss": 1.3072, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 4.922823943346332e-05, | |
| "loss": 1.2885, | |
| "step": 27550 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 4.9226838779077586e-05, | |
| "loss": 1.3363, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 4.922543812469186e-05, | |
| "loss": 1.3299, | |
| "step": 27650 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 4.922403747030613e-05, | |
| "loss": 1.2851, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 4.92226368159204e-05, | |
| "loss": 1.2899, | |
| "step": 27750 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 4.922123616153467e-05, | |
| "loss": 1.2967, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 4.9219835507148944e-05, | |
| "loss": 1.3213, | |
| "step": 27850 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 4.921843485276321e-05, | |
| "loss": 1.2822, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 4.921703419837748e-05, | |
| "loss": 1.2979, | |
| "step": 27950 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 4.921563354399176e-05, | |
| "loss": 1.3157, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 4.9214232889606024e-05, | |
| "loss": 1.3302, | |
| "step": 28050 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 4.9212832235220296e-05, | |
| "loss": 1.3277, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 4.921143158083457e-05, | |
| "loss": 1.3079, | |
| "step": 28150 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 4.921003092644884e-05, | |
| "loss": 1.3159, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 4.920863027206311e-05, | |
| "loss": 1.2898, | |
| "step": 28250 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 4.920722961767738e-05, | |
| "loss": 1.2955, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 4.920582896329165e-05, | |
| "loss": 1.328, | |
| "step": 28350 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 4.9204428308905925e-05, | |
| "loss": 1.3313, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 4.92030276545202e-05, | |
| "loss": 1.3039, | |
| "step": 28450 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 4.920162700013446e-05, | |
| "loss": 1.3175, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 4.9200226345748734e-05, | |
| "loss": 1.3283, | |
| "step": 28550 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 4.919882569136301e-05, | |
| "loss": 1.3091, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 4.919742503697728e-05, | |
| "loss": 1.3048, | |
| "step": 28650 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 4.919602438259155e-05, | |
| "loss": 1.2843, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 4.919462372820582e-05, | |
| "loss": 1.3218, | |
| "step": 28750 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 4.919322307382009e-05, | |
| "loss": 1.3225, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 4.9191822419434363e-05, | |
| "loss": 1.2906, | |
| "step": 28850 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 4.9190421765048635e-05, | |
| "loss": 1.3013, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 4.91890211106629e-05, | |
| "loss": 1.3285, | |
| "step": 28950 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 4.918762045627717e-05, | |
| "loss": 1.2895, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 4.918621980189145e-05, | |
| "loss": 1.2707, | |
| "step": 29050 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 4.9184819147505715e-05, | |
| "loss": 1.3069, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 4.9183418493119986e-05, | |
| "loss": 1.3097, | |
| "step": 29150 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 4.918201783873426e-05, | |
| "loss": 1.3229, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 4.918061718434853e-05, | |
| "loss": 1.3097, | |
| "step": 29250 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 4.91792165299628e-05, | |
| "loss": 1.3154, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 4.917781587557707e-05, | |
| "loss": 1.3012, | |
| "step": 29350 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 4.917641522119134e-05, | |
| "loss": 1.2962, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 4.9175014566805616e-05, | |
| "loss": 1.3123, | |
| "step": 29450 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 4.917361391241989e-05, | |
| "loss": 1.2856, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 4.917221325803415e-05, | |
| "loss": 1.3, | |
| "step": 29550 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 4.9170812603648424e-05, | |
| "loss": 1.2847, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 4.91694119492627e-05, | |
| "loss": 1.3098, | |
| "step": 29650 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 4.916801129487697e-05, | |
| "loss": 1.3073, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 4.916661064049124e-05, | |
| "loss": 1.332, | |
| "step": 29750 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 4.916520998610551e-05, | |
| "loss": 1.3081, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 4.916380933171978e-05, | |
| "loss": 1.2914, | |
| "step": 29850 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 4.9162408677334054e-05, | |
| "loss": 1.2642, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 4.9161008022948326e-05, | |
| "loss": 1.2792, | |
| "step": 29950 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 4.915960736856259e-05, | |
| "loss": 1.3048, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 4.915820671417686e-05, | |
| "loss": 1.2826, | |
| "step": 30050 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 4.915680605979114e-05, | |
| "loss": 1.2759, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 4.9155405405405406e-05, | |
| "loss": 1.2902, | |
| "step": 30150 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 4.915400475101968e-05, | |
| "loss": 1.3175, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 4.915260409663395e-05, | |
| "loss": 1.2784, | |
| "step": 30250 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 4.915120344224822e-05, | |
| "loss": 1.2925, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 4.914980278786249e-05, | |
| "loss": 1.3032, | |
| "step": 30350 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 4.9148402133476764e-05, | |
| "loss": 1.2928, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 4.914700147909103e-05, | |
| "loss": 1.3201, | |
| "step": 30450 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 4.914560082470531e-05, | |
| "loss": 1.2744, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 4.914420017031958e-05, | |
| "loss": 1.2872, | |
| "step": 30550 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 4.9142799515933844e-05, | |
| "loss": 1.2961, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 4.9141398861548115e-05, | |
| "loss": 1.3028, | |
| "step": 30650 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 4.9139998207162394e-05, | |
| "loss": 1.2775, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 4.913859755277666e-05, | |
| "loss": 1.2889, | |
| "step": 30750 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 4.913719689839093e-05, | |
| "loss": 1.2834, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 4.91357962440052e-05, | |
| "loss": 1.2941, | |
| "step": 30850 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 4.913439558961947e-05, | |
| "loss": 1.3068, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 4.9132994935233745e-05, | |
| "loss": 1.3122, | |
| "step": 30950 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 4.913159428084802e-05, | |
| "loss": 1.3007, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 4.913019362646228e-05, | |
| "loss": 1.2985, | |
| "step": 31050 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 4.912879297207655e-05, | |
| "loss": 1.296, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 4.912739231769083e-05, | |
| "loss": 1.2782, | |
| "step": 31150 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 4.9125991663305096e-05, | |
| "loss": 1.3056, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 4.912459100891937e-05, | |
| "loss": 1.2807, | |
| "step": 31250 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 4.912319035453364e-05, | |
| "loss": 1.3004, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 4.912178970014791e-05, | |
| "loss": 1.2904, | |
| "step": 31350 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 4.912038904576218e-05, | |
| "loss": 1.2898, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 4.9118988391376455e-05, | |
| "loss": 1.2742, | |
| "step": 31450 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 4.911758773699072e-05, | |
| "loss": 1.3027, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 4.9116187082605e-05, | |
| "loss": 1.3086, | |
| "step": 31550 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 4.911478642821927e-05, | |
| "loss": 1.2734, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 4.9113385773833534e-05, | |
| "loss": 1.2884, | |
| "step": 31650 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 4.9111985119447806e-05, | |
| "loss": 1.3396, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 4.9110584465062084e-05, | |
| "loss": 1.3018, | |
| "step": 31750 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 4.910918381067635e-05, | |
| "loss": 1.2828, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 4.910778315629062e-05, | |
| "loss": 1.3015, | |
| "step": 31850 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 4.910638250190489e-05, | |
| "loss": 1.3118, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 4.9104981847519164e-05, | |
| "loss": 1.288, | |
| "step": 31950 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 4.9103581193133436e-05, | |
| "loss": 1.288, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 4.910218053874771e-05, | |
| "loss": 1.2914, | |
| "step": 32050 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 4.910077988436197e-05, | |
| "loss": 1.2784, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 4.9099379229976244e-05, | |
| "loss": 1.2836, | |
| "step": 32150 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 4.909797857559052e-05, | |
| "loss": 1.2589, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 4.909657792120479e-05, | |
| "loss": 1.3241, | |
| "step": 32250 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 4.909517726681906e-05, | |
| "loss": 1.2986, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 4.909377661243333e-05, | |
| "loss": 1.2801, | |
| "step": 32350 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 4.90923759580476e-05, | |
| "loss": 1.2766, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 4.9090975303661874e-05, | |
| "loss": 1.313, | |
| "step": 32450 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 4.9089574649276145e-05, | |
| "loss": 1.2746, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 4.908817399489041e-05, | |
| "loss": 1.3009, | |
| "step": 32550 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 4.908677334050469e-05, | |
| "loss": 1.2809, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 4.908537268611896e-05, | |
| "loss": 1.2728, | |
| "step": 32650 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 4.9083972031733225e-05, | |
| "loss": 1.2999, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 4.90825713773475e-05, | |
| "loss": 1.2662, | |
| "step": 32750 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 4.9081170722961775e-05, | |
| "loss": 1.2843, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 4.907977006857604e-05, | |
| "loss": 1.2862, | |
| "step": 32850 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 4.907836941419031e-05, | |
| "loss": 1.2926, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 4.907696875980458e-05, | |
| "loss": 1.2801, | |
| "step": 32950 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 4.9075568105418855e-05, | |
| "loss": 1.3205, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 4.9074167451033127e-05, | |
| "loss": 1.2809, | |
| "step": 33050 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 4.90727667966474e-05, | |
| "loss": 1.2742, | |
| "step": 33100 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 4.907136614226166e-05, | |
| "loss": 1.2854, | |
| "step": 33150 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 4.906996548787594e-05, | |
| "loss": 1.2648, | |
| "step": 33200 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 4.906856483349021e-05, | |
| "loss": 1.3033, | |
| "step": 33250 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 4.906716417910448e-05, | |
| "loss": 1.2548, | |
| "step": 33300 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 4.906576352471875e-05, | |
| "loss": 1.3158, | |
| "step": 33350 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 4.906436287033302e-05, | |
| "loss": 1.3022, | |
| "step": 33400 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 4.906296221594729e-05, | |
| "loss": 1.2948, | |
| "step": 33450 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 4.9061561561561565e-05, | |
| "loss": 1.2916, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 4.9060160907175836e-05, | |
| "loss": 1.285, | |
| "step": 33550 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 4.90587602527901e-05, | |
| "loss": 1.2831, | |
| "step": 33600 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 4.905735959840438e-05, | |
| "loss": 1.2866, | |
| "step": 33650 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 4.905595894401865e-05, | |
| "loss": 1.266, | |
| "step": 33700 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 4.9054558289632916e-05, | |
| "loss": 1.2396, | |
| "step": 33750 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 4.905315763524719e-05, | |
| "loss": 1.2913, | |
| "step": 33800 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 4.9051756980861466e-05, | |
| "loss": 1.321, | |
| "step": 33850 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 4.905035632647573e-05, | |
| "loss": 1.2768, | |
| "step": 33900 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 4.904895567209e-05, | |
| "loss": 1.2767, | |
| "step": 33950 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 4.9047555017704274e-05, | |
| "loss": 1.2707, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 4.9046154363318546e-05, | |
| "loss": 1.2705, | |
| "step": 34050 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 4.904475370893282e-05, | |
| "loss": 1.3089, | |
| "step": 34100 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 4.904335305454709e-05, | |
| "loss": 1.2597, | |
| "step": 34150 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 4.9041952400161354e-05, | |
| "loss": 1.2656, | |
| "step": 34200 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 4.904055174577563e-05, | |
| "loss": 1.2772, | |
| "step": 34250 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 4.9039151091389904e-05, | |
| "loss": 1.2838, | |
| "step": 34300 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 4.903775043700417e-05, | |
| "loss": 1.2743, | |
| "step": 34350 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 4.903634978261844e-05, | |
| "loss": 1.2745, | |
| "step": 34400 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 4.903494912823271e-05, | |
| "loss": 1.2955, | |
| "step": 34450 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 4.9033548473846984e-05, | |
| "loss": 1.2695, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 4.9032147819461255e-05, | |
| "loss": 1.2664, | |
| "step": 34550 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 4.903074716507553e-05, | |
| "loss": 1.3064, | |
| "step": 34600 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 4.902934651068979e-05, | |
| "loss": 1.275, | |
| "step": 34650 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 4.902794585630407e-05, | |
| "loss": 1.2841, | |
| "step": 34700 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 4.902654520191834e-05, | |
| "loss": 1.2625, | |
| "step": 34750 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 4.902514454753261e-05, | |
| "loss": 1.287, | |
| "step": 34800 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 4.902374389314688e-05, | |
| "loss": 1.2852, | |
| "step": 34850 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 4.902234323876116e-05, | |
| "loss": 1.2593, | |
| "step": 34900 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 4.902094258437542e-05, | |
| "loss": 1.2636, | |
| "step": 34950 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 4.901954192998969e-05, | |
| "loss": 1.2938, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 4.9018141275603965e-05, | |
| "loss": 1.2508, | |
| "step": 35050 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 4.9016740621218237e-05, | |
| "loss": 1.3109, | |
| "step": 35100 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 4.901533996683251e-05, | |
| "loss": 1.2902, | |
| "step": 35150 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 4.901393931244678e-05, | |
| "loss": 1.2899, | |
| "step": 35200 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 4.9012538658061045e-05, | |
| "loss": 1.2717, | |
| "step": 35250 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 4.901113800367532e-05, | |
| "loss": 1.2707, | |
| "step": 35300 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 4.9009737349289595e-05, | |
| "loss": 1.3071, | |
| "step": 35350 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 4.900833669490386e-05, | |
| "loss": 1.279, | |
| "step": 35400 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 4.900693604051813e-05, | |
| "loss": 1.269, | |
| "step": 35450 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 4.90055353861324e-05, | |
| "loss": 1.2757, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 4.9004134731746674e-05, | |
| "loss": 1.2808, | |
| "step": 35550 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 4.9002734077360946e-05, | |
| "loss": 1.2825, | |
| "step": 35600 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 4.900133342297522e-05, | |
| "loss": 1.2902, | |
| "step": 35650 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 4.899993276858948e-05, | |
| "loss": 1.2926, | |
| "step": 35700 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 4.899853211420376e-05, | |
| "loss": 1.2941, | |
| "step": 35750 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 4.899713145981803e-05, | |
| "loss": 1.2875, | |
| "step": 35800 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 4.89957308054323e-05, | |
| "loss": 1.27, | |
| "step": 35850 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 4.899433015104657e-05, | |
| "loss": 1.2614, | |
| "step": 35900 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 4.899292949666085e-05, | |
| "loss": 1.2542, | |
| "step": 35950 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 4.899152884227511e-05, | |
| "loss": 1.2709, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 4.8990128187889384e-05, | |
| "loss": 1.2664, | |
| "step": 36050 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 4.8988727533503656e-05, | |
| "loss": 1.2685, | |
| "step": 36100 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 4.898732687911793e-05, | |
| "loss": 1.2621, | |
| "step": 36150 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 4.89859262247322e-05, | |
| "loss": 1.2873, | |
| "step": 36200 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 4.898452557034647e-05, | |
| "loss": 1.2842, | |
| "step": 36250 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 4.8983124915960735e-05, | |
| "loss": 1.2622, | |
| "step": 36300 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 4.8981724261575014e-05, | |
| "loss": 1.2695, | |
| "step": 36350 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 4.8980323607189285e-05, | |
| "loss": 1.2492, | |
| "step": 36400 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 4.897892295280355e-05, | |
| "loss": 1.2736, | |
| "step": 36450 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 4.897752229841782e-05, | |
| "loss": 1.2612, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 4.8976121644032094e-05, | |
| "loss": 1.2644, | |
| "step": 36550 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 4.8974720989646365e-05, | |
| "loss": 1.2708, | |
| "step": 36600 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 4.897332033526064e-05, | |
| "loss": 1.2738, | |
| "step": 36650 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 4.897191968087491e-05, | |
| "loss": 1.26, | |
| "step": 36700 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 4.897051902648917e-05, | |
| "loss": 1.2805, | |
| "step": 36750 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 4.896911837210345e-05, | |
| "loss": 1.2871, | |
| "step": 36800 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 4.8967717717717723e-05, | |
| "loss": 1.284, | |
| "step": 36850 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 4.896631706333199e-05, | |
| "loss": 1.2666, | |
| "step": 36900 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 4.896491640894626e-05, | |
| "loss": 1.261, | |
| "step": 36950 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 4.896351575456054e-05, | |
| "loss": 1.2892, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 4.89621151001748e-05, | |
| "loss": 1.2672, | |
| "step": 37050 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 4.8960714445789075e-05, | |
| "loss": 1.2791, | |
| "step": 37100 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 4.8959313791403346e-05, | |
| "loss": 1.283, | |
| "step": 37150 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 4.895791313701762e-05, | |
| "loss": 1.2304, | |
| "step": 37200 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 4.895651248263189e-05, | |
| "loss": 1.2807, | |
| "step": 37250 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 4.895511182824616e-05, | |
| "loss": 1.2712, | |
| "step": 37300 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 4.8953711173860426e-05, | |
| "loss": 1.2533, | |
| "step": 37350 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 4.8952310519474705e-05, | |
| "loss": 1.2567, | |
| "step": 37400 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 4.8950909865088976e-05, | |
| "loss": 1.259, | |
| "step": 37450 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 4.894950921070324e-05, | |
| "loss": 1.2832, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 4.894810855631751e-05, | |
| "loss": 1.2656, | |
| "step": 37550 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 4.8946707901931784e-05, | |
| "loss": 1.2681, | |
| "step": 37600 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 4.8945307247546056e-05, | |
| "loss": 1.2587, | |
| "step": 37650 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 4.894390659316033e-05, | |
| "loss": 1.263, | |
| "step": 37700 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 4.89425059387746e-05, | |
| "loss": 1.2906, | |
| "step": 37750 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 4.8941105284388864e-05, | |
| "loss": 1.241, | |
| "step": 37800 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 4.893970463000314e-05, | |
| "loss": 1.2742, | |
| "step": 37850 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 4.8938303975617414e-05, | |
| "loss": 1.2652, | |
| "step": 37900 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 4.893690332123168e-05, | |
| "loss": 1.2572, | |
| "step": 37950 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 4.893550266684595e-05, | |
| "loss": 1.2905, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 4.893410201246022e-05, | |
| "loss": 1.2526, | |
| "step": 38050 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 4.8932701358074494e-05, | |
| "loss": 1.2871, | |
| "step": 38100 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 4.8931300703688766e-05, | |
| "loss": 1.2729, | |
| "step": 38150 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 4.892990004930304e-05, | |
| "loss": 1.2516, | |
| "step": 38200 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 4.892849939491731e-05, | |
| "loss": 1.2629, | |
| "step": 38250 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 4.892709874053158e-05, | |
| "loss": 1.253, | |
| "step": 38300 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 4.892569808614585e-05, | |
| "loss": 1.2896, | |
| "step": 38350 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 4.892429743176012e-05, | |
| "loss": 1.2799, | |
| "step": 38400 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 4.8922896777374395e-05, | |
| "loss": 1.2734, | |
| "step": 38450 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 4.892149612298867e-05, | |
| "loss": 1.2748, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 4.892009546860293e-05, | |
| "loss": 1.2544, | |
| "step": 38550 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 4.8918694814217204e-05, | |
| "loss": 1.2698, | |
| "step": 38600 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 4.8917294159831475e-05, | |
| "loss": 1.2794, | |
| "step": 38650 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 4.891589350544575e-05, | |
| "loss": 1.2711, | |
| "step": 38700 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 4.891449285106002e-05, | |
| "loss": 1.2647, | |
| "step": 38750 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 4.891309219667429e-05, | |
| "loss": 1.2614, | |
| "step": 38800 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 4.8911691542288555e-05, | |
| "loss": 1.271, | |
| "step": 38850 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 4.891029088790283e-05, | |
| "loss": 1.2831, | |
| "step": 38900 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 4.8908890233517105e-05, | |
| "loss": 1.2748, | |
| "step": 38950 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 4.890748957913137e-05, | |
| "loss": 1.2796, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 4.890608892474564e-05, | |
| "loss": 1.279, | |
| "step": 39050 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 4.890468827035991e-05, | |
| "loss": 1.2432, | |
| "step": 39100 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 4.8903287615974185e-05, | |
| "loss": 1.2755, | |
| "step": 39150 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 4.8901886961588456e-05, | |
| "loss": 1.2341, | |
| "step": 39200 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 4.890048630720273e-05, | |
| "loss": 1.2592, | |
| "step": 39250 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 4.8899085652817e-05, | |
| "loss": 1.2487, | |
| "step": 39300 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 4.889768499843127e-05, | |
| "loss": 1.2737, | |
| "step": 39350 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 4.889628434404554e-05, | |
| "loss": 1.2582, | |
| "step": 39400 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 4.889488368965981e-05, | |
| "loss": 1.2543, | |
| "step": 39450 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 4.8893483035274086e-05, | |
| "loss": 1.2669, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 4.889208238088835e-05, | |
| "loss": 1.2544, | |
| "step": 39550 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 4.889068172650262e-05, | |
| "loss": 1.2764, | |
| "step": 39600 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 4.8889281072116894e-05, | |
| "loss": 1.2518, | |
| "step": 39650 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 4.8887880417731166e-05, | |
| "loss": 1.2614, | |
| "step": 39700 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 4.888647976334544e-05, | |
| "loss": 1.2271, | |
| "step": 39750 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 4.888507910895971e-05, | |
| "loss": 1.2771, | |
| "step": 39800 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 4.888367845457398e-05, | |
| "loss": 1.2457, | |
| "step": 39850 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 4.8882277800188246e-05, | |
| "loss": 1.2609, | |
| "step": 39900 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 4.8880877145802524e-05, | |
| "loss": 1.2579, | |
| "step": 39950 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 4.887947649141679e-05, | |
| "loss": 1.2853, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 4.887807583703106e-05, | |
| "loss": 1.2496, | |
| "step": 40050 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 4.887667518264533e-05, | |
| "loss": 1.2633, | |
| "step": 40100 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 4.8875274528259604e-05, | |
| "loss": 1.2668, | |
| "step": 40150 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 4.8873873873873876e-05, | |
| "loss": 1.2649, | |
| "step": 40200 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 4.887247321948815e-05, | |
| "loss": 1.2554, | |
| "step": 40250 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 4.887107256510242e-05, | |
| "loss": 1.2577, | |
| "step": 40300 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 4.886967191071669e-05, | |
| "loss": 1.2373, | |
| "step": 40350 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 4.886827125633096e-05, | |
| "loss": 1.2435, | |
| "step": 40400 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 4.8866870601945234e-05, | |
| "loss": 1.2559, | |
| "step": 40450 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.88654699475595e-05, | |
| "loss": 1.2583, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.886406929317378e-05, | |
| "loss": 1.2645, | |
| "step": 40550 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.886266863878804e-05, | |
| "loss": 1.274, | |
| "step": 40600 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.8861267984402314e-05, | |
| "loss": 1.2543, | |
| "step": 40650 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.8859867330016585e-05, | |
| "loss": 1.2648, | |
| "step": 40700 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.885846667563086e-05, | |
| "loss": 1.248, | |
| "step": 40750 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.885706602124513e-05, | |
| "loss": 1.2658, | |
| "step": 40800 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.88556653668594e-05, | |
| "loss": 1.2763, | |
| "step": 40850 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.885426471247367e-05, | |
| "loss": 1.238, | |
| "step": 40900 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.885286405808794e-05, | |
| "loss": 1.2763, | |
| "step": 40950 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.8851463403702215e-05, | |
| "loss": 1.2537, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.885006274931648e-05, | |
| "loss": 1.2257, | |
| "step": 41050 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.884866209493075e-05, | |
| "loss": 1.2472, | |
| "step": 41100 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.884726144054502e-05, | |
| "loss": 1.271, | |
| "step": 41150 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 4.8845860786159295e-05, | |
| "loss": 1.2635, | |
| "step": 41200 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 4.8844460131773566e-05, | |
| "loss": 1.233, | |
| "step": 41250 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 4.884305947738784e-05, | |
| "loss": 1.2269, | |
| "step": 41300 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 4.884165882300211e-05, | |
| "loss": 1.2529, | |
| "step": 41350 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 4.884025816861638e-05, | |
| "loss": 1.2595, | |
| "step": 41400 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 4.883885751423065e-05, | |
| "loss": 1.23, | |
| "step": 41450 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 4.883745685984492e-05, | |
| "loss": 1.2752, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 4.883605620545919e-05, | |
| "loss": 1.2482, | |
| "step": 41550 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 4.883465555107347e-05, | |
| "loss": 1.2675, | |
| "step": 41600 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 4.883325489668773e-05, | |
| "loss": 1.2196, | |
| "step": 41650 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 4.8831854242302004e-05, | |
| "loss": 1.2275, | |
| "step": 41700 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 4.8830453587916276e-05, | |
| "loss": 1.2709, | |
| "step": 41750 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 4.882905293353055e-05, | |
| "loss": 1.2194, | |
| "step": 41800 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 4.882765227914482e-05, | |
| "loss": 1.2408, | |
| "step": 41850 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 4.882625162475909e-05, | |
| "loss": 1.2282, | |
| "step": 41900 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 4.882485097037336e-05, | |
| "loss": 1.2642, | |
| "step": 41950 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 4.8823450315987634e-05, | |
| "loss": 1.225, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 4.8822049661601906e-05, | |
| "loss": 1.2541, | |
| "step": 42050 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 4.882064900721617e-05, | |
| "loss": 1.2723, | |
| "step": 42100 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 4.881924835283044e-05, | |
| "loss": 1.233, | |
| "step": 42150 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 4.8817847698444714e-05, | |
| "loss": 1.2642, | |
| "step": 42200 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 4.8816447044058986e-05, | |
| "loss": 1.2538, | |
| "step": 42250 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 4.881504638967326e-05, | |
| "loss": 1.2425, | |
| "step": 42300 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 4.881364573528753e-05, | |
| "loss": 1.2449, | |
| "step": 42350 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 4.88122450809018e-05, | |
| "loss": 1.2382, | |
| "step": 42400 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 4.881084442651607e-05, | |
| "loss": 1.1989, | |
| "step": 42450 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 4.8809443772130344e-05, | |
| "loss": 1.2473, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 4.880804311774461e-05, | |
| "loss": 1.2382, | |
| "step": 42550 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 4.880664246335888e-05, | |
| "loss": 1.2732, | |
| "step": 42600 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 4.880524180897316e-05, | |
| "loss": 1.2389, | |
| "step": 42650 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 4.8803841154587423e-05, | |
| "loss": 1.2669, | |
| "step": 42700 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 4.8802440500201695e-05, | |
| "loss": 1.2414, | |
| "step": 42750 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 4.880103984581597e-05, | |
| "loss": 1.2243, | |
| "step": 42800 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 4.879963919143024e-05, | |
| "loss": 1.2497, | |
| "step": 42850 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 4.879823853704451e-05, | |
| "loss": 1.2445, | |
| "step": 42900 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 4.879683788265878e-05, | |
| "loss": 1.2558, | |
| "step": 42950 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 4.8795437228273047e-05, | |
| "loss": 1.2713, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 4.8794036573887325e-05, | |
| "loss": 1.2751, | |
| "step": 43050 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 4.8792635919501597e-05, | |
| "loss": 1.2432, | |
| "step": 43100 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 4.879123526511586e-05, | |
| "loss": 1.2291, | |
| "step": 43150 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 4.878983461073013e-05, | |
| "loss": 1.2433, | |
| "step": 43200 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 4.8788433956344405e-05, | |
| "loss": 1.2496, | |
| "step": 43250 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 4.8787033301958676e-05, | |
| "loss": 1.256, | |
| "step": 43300 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 4.878563264757295e-05, | |
| "loss": 1.2297, | |
| "step": 43350 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 4.878423199318722e-05, | |
| "loss": 1.2287, | |
| "step": 43400 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 4.878283133880149e-05, | |
| "loss": 1.2465, | |
| "step": 43450 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 4.878143068441576e-05, | |
| "loss": 1.2325, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 4.8780030030030034e-05, | |
| "loss": 1.2047, | |
| "step": 43550 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 4.87786293756443e-05, | |
| "loss": 1.2457, | |
| "step": 43600 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 4.877722872125857e-05, | |
| "loss": 1.2407, | |
| "step": 43650 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 4.877582806687285e-05, | |
| "loss": 1.2483, | |
| "step": 43700 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 4.8774427412487114e-05, | |
| "loss": 1.2554, | |
| "step": 43750 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 4.8773026758101386e-05, | |
| "loss": 1.2362, | |
| "step": 43800 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 4.877162610371566e-05, | |
| "loss": 1.2089, | |
| "step": 43850 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 4.877022544932993e-05, | |
| "loss": 1.2505, | |
| "step": 43900 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 4.87688247949442e-05, | |
| "loss": 1.2423, | |
| "step": 43950 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 4.876742414055847e-05, | |
| "loss": 1.2723, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 4.876602348617274e-05, | |
| "loss": 1.2431, | |
| "step": 44050 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 4.8764622831787016e-05, | |
| "loss": 1.2646, | |
| "step": 44100 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 4.876322217740129e-05, | |
| "loss": 1.23, | |
| "step": 44150 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 4.876182152301555e-05, | |
| "loss": 1.2269, | |
| "step": 44200 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 4.8760420868629824e-05, | |
| "loss": 1.2469, | |
| "step": 44250 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 4.87590202142441e-05, | |
| "loss": 1.2559, | |
| "step": 44300 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 4.875761955985837e-05, | |
| "loss": 1.2738, | |
| "step": 44350 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 4.875621890547264e-05, | |
| "loss": 1.2561, | |
| "step": 44400 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 4.875481825108691e-05, | |
| "loss": 1.2353, | |
| "step": 44450 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 4.8753417596701175e-05, | |
| "loss": 1.2301, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 4.8752016942315454e-05, | |
| "loss": 1.2695, | |
| "step": 44550 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 4.8750616287929725e-05, | |
| "loss": 1.2462, | |
| "step": 44600 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 4.874921563354399e-05, | |
| "loss": 1.2627, | |
| "step": 44650 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 4.874781497915826e-05, | |
| "loss": 1.2441, | |
| "step": 44700 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.874641432477254e-05, | |
| "loss": 1.2372, | |
| "step": 44750 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.8745013670386805e-05, | |
| "loss": 1.2305, | |
| "step": 44800 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.874361301600108e-05, | |
| "loss": 1.2258, | |
| "step": 44850 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.874221236161535e-05, | |
| "loss": 1.2477, | |
| "step": 44900 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.874081170722962e-05, | |
| "loss": 1.2799, | |
| "step": 44950 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.873941105284389e-05, | |
| "loss": 1.2465, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.873801039845816e-05, | |
| "loss": 1.2296, | |
| "step": 45050 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.873660974407243e-05, | |
| "loss": 1.2494, | |
| "step": 45100 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 4.8735209089686707e-05, | |
| "loss": 1.2313, | |
| "step": 45150 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.873380843530098e-05, | |
| "loss": 1.2327, | |
| "step": 45200 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.873240778091524e-05, | |
| "loss": 1.2189, | |
| "step": 45250 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.8731007126529515e-05, | |
| "loss": 1.237, | |
| "step": 45300 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.872960647214379e-05, | |
| "loss": 1.2046, | |
| "step": 45350 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 4.872820581775806e-05, | |
| "loss": 1.2447, | |
| "step": 45400 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.872680516337233e-05, | |
| "loss": 1.2393, | |
| "step": 45450 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.87254045089866e-05, | |
| "loss": 1.2266, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.8724003854600866e-05, | |
| "loss": 1.2423, | |
| "step": 45550 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.8722603200215144e-05, | |
| "loss": 1.2575, | |
| "step": 45600 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.8721202545829416e-05, | |
| "loss": 1.2202, | |
| "step": 45650 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.871980189144368e-05, | |
| "loss": 1.2145, | |
| "step": 45700 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.871840123705795e-05, | |
| "loss": 1.2246, | |
| "step": 45750 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 4.871700058267223e-05, | |
| "loss": 1.2407, | |
| "step": 45800 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.8715599928286496e-05, | |
| "loss": 1.2289, | |
| "step": 45850 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.871419927390077e-05, | |
| "loss": 1.261, | |
| "step": 45900 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.871279861951504e-05, | |
| "loss": 1.203, | |
| "step": 45950 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.871139796512931e-05, | |
| "loss": 1.2226, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 4.870999731074358e-05, | |
| "loss": 1.2498, | |
| "step": 46050 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.8708596656357854e-05, | |
| "loss": 1.2004, | |
| "step": 46100 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.870719600197212e-05, | |
| "loss": 1.2399, | |
| "step": 46150 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.87057953475864e-05, | |
| "loss": 1.2406, | |
| "step": 46200 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.870439469320067e-05, | |
| "loss": 1.2323, | |
| "step": 46250 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.8702994038814934e-05, | |
| "loss": 1.2299, | |
| "step": 46300 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.8701593384429205e-05, | |
| "loss": 1.2521, | |
| "step": 46350 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.8700192730043484e-05, | |
| "loss": 1.2284, | |
| "step": 46400 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.869879207565775e-05, | |
| "loss": 1.2275, | |
| "step": 46450 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 4.869739142127202e-05, | |
| "loss": 1.2482, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.869599076688629e-05, | |
| "loss": 1.2348, | |
| "step": 46550 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.869459011250056e-05, | |
| "loss": 1.2323, | |
| "step": 46600 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.8693189458114835e-05, | |
| "loss": 1.2286, | |
| "step": 46650 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 4.869178880372911e-05, | |
| "loss": 1.2215, | |
| "step": 46700 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.869038814934337e-05, | |
| "loss": 1.2196, | |
| "step": 46750 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.868898749495764e-05, | |
| "loss": 1.2544, | |
| "step": 46800 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.868758684057192e-05, | |
| "loss": 1.2101, | |
| "step": 46850 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.868618618618619e-05, | |
| "loss": 1.2215, | |
| "step": 46900 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.868478553180046e-05, | |
| "loss": 1.2258, | |
| "step": 46950 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.868338487741473e-05, | |
| "loss": 1.2388, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.8681984223029e-05, | |
| "loss": 1.2367, | |
| "step": 47050 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.868058356864327e-05, | |
| "loss": 1.2129, | |
| "step": 47100 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.8679182914257545e-05, | |
| "loss": 1.2243, | |
| "step": 47150 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.867778225987181e-05, | |
| "loss": 1.229, | |
| "step": 47200 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.867638160548609e-05, | |
| "loss": 1.21, | |
| "step": 47250 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.867498095110036e-05, | |
| "loss": 1.2209, | |
| "step": 47300 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.8673580296714625e-05, | |
| "loss": 1.2399, | |
| "step": 47350 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.8672179642328896e-05, | |
| "loss": 1.2307, | |
| "step": 47400 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.8670778987943175e-05, | |
| "loss": 1.2654, | |
| "step": 47450 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.866937833355744e-05, | |
| "loss": 1.2266, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.866797767917171e-05, | |
| "loss": 1.1997, | |
| "step": 47550 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.866657702478598e-05, | |
| "loss": 1.234, | |
| "step": 47600 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.866517637040025e-05, | |
| "loss": 1.2297, | |
| "step": 47650 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.8663775716014526e-05, | |
| "loss": 1.2273, | |
| "step": 47700 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.86623750616288e-05, | |
| "loss": 1.2497, | |
| "step": 47750 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.866097440724306e-05, | |
| "loss": 1.2065, | |
| "step": 47800 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.8659573752857334e-05, | |
| "loss": 1.2506, | |
| "step": 47850 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.865817309847161e-05, | |
| "loss": 1.2257, | |
| "step": 47900 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.865677244408588e-05, | |
| "loss": 1.1959, | |
| "step": 47950 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.865537178970015e-05, | |
| "loss": 1.2571, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.865397113531442e-05, | |
| "loss": 1.2162, | |
| "step": 48050 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.865257048092869e-05, | |
| "loss": 1.2363, | |
| "step": 48100 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.8651169826542964e-05, | |
| "loss": 1.2163, | |
| "step": 48150 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.8649769172157236e-05, | |
| "loss": 1.2321, | |
| "step": 48200 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.86483685177715e-05, | |
| "loss": 1.2351, | |
| "step": 48250 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.864696786338578e-05, | |
| "loss": 1.2446, | |
| "step": 48300 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.864556720900005e-05, | |
| "loss": 1.2382, | |
| "step": 48350 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.8644166554614315e-05, | |
| "loss": 1.2366, | |
| "step": 48400 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.864276590022859e-05, | |
| "loss": 1.231, | |
| "step": 48450 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.8641365245842865e-05, | |
| "loss": 1.1926, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.863996459145713e-05, | |
| "loss": 1.2306, | |
| "step": 48550 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.86385639370714e-05, | |
| "loss": 1.2361, | |
| "step": 48600 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.8637163282685674e-05, | |
| "loss": 1.2208, | |
| "step": 48650 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.8635762628299945e-05, | |
| "loss": 1.2187, | |
| "step": 48700 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.863436197391422e-05, | |
| "loss": 1.2268, | |
| "step": 48750 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.863296131952849e-05, | |
| "loss": 1.2174, | |
| "step": 48800 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.863156066514275e-05, | |
| "loss": 1.225, | |
| "step": 48850 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.8630160010757025e-05, | |
| "loss": 1.2393, | |
| "step": 48900 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.86287593563713e-05, | |
| "loss": 1.2388, | |
| "step": 48950 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 4.862735870198557e-05, | |
| "loss": 1.2231, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 4.862595804759984e-05, | |
| "loss": 1.2132, | |
| "step": 49050 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 4.862455739321411e-05, | |
| "loss": 1.229, | |
| "step": 49100 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 4.862315673882838e-05, | |
| "loss": 1.2409, | |
| "step": 49150 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 4.8621756084442655e-05, | |
| "loss": 1.2225, | |
| "step": 49200 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 4.8620355430056926e-05, | |
| "loss": 1.2442, | |
| "step": 49250 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 4.861895477567119e-05, | |
| "loss": 1.2242, | |
| "step": 49300 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 4.861755412128547e-05, | |
| "loss": 1.2319, | |
| "step": 49350 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 4.861615346689974e-05, | |
| "loss": 1.2594, | |
| "step": 49400 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 4.8614752812514006e-05, | |
| "loss": 1.2201, | |
| "step": 49450 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 4.861335215812828e-05, | |
| "loss": 1.205, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 4.8611951503742556e-05, | |
| "loss": 1.2434, | |
| "step": 49550 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 4.861055084935682e-05, | |
| "loss": 1.2036, | |
| "step": 49600 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 4.860915019497109e-05, | |
| "loss": 1.2229, | |
| "step": 49650 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 4.8607749540585364e-05, | |
| "loss": 1.215, | |
| "step": 49700 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 4.8606348886199636e-05, | |
| "loss": 1.2174, | |
| "step": 49750 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 4.860494823181391e-05, | |
| "loss": 1.2042, | |
| "step": 49800 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 4.860354757742818e-05, | |
| "loss": 1.2096, | |
| "step": 49850 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 4.8602146923042444e-05, | |
| "loss": 1.2342, | |
| "step": 49900 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 4.8600746268656716e-05, | |
| "loss": 1.2375, | |
| "step": 49950 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 4.8599345614270994e-05, | |
| "loss": 1.2186, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 4.859794495988526e-05, | |
| "loss": 1.2533, | |
| "step": 50050 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 4.859654430549953e-05, | |
| "loss": 1.2225, | |
| "step": 50100 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 4.85951436511138e-05, | |
| "loss": 1.2329, | |
| "step": 50150 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 4.8593742996728074e-05, | |
| "loss": 1.1887, | |
| "step": 50200 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 4.8592342342342346e-05, | |
| "loss": 1.2638, | |
| "step": 50250 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 4.859094168795662e-05, | |
| "loss": 1.2038, | |
| "step": 50300 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 4.858954103357088e-05, | |
| "loss": 1.2328, | |
| "step": 50350 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 4.858814037918516e-05, | |
| "loss": 1.2415, | |
| "step": 50400 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 4.858673972479943e-05, | |
| "loss": 1.2019, | |
| "step": 50450 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 4.85853390704137e-05, | |
| "loss": 1.223, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.858393841602797e-05, | |
| "loss": 1.2338, | |
| "step": 50550 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.858253776164225e-05, | |
| "loss": 1.2546, | |
| "step": 50600 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.858113710725651e-05, | |
| "loss": 1.2223, | |
| "step": 50650 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.8579736452870784e-05, | |
| "loss": 1.2409, | |
| "step": 50700 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 4.8578335798485055e-05, | |
| "loss": 1.2193, | |
| "step": 50750 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 4.857693514409933e-05, | |
| "loss": 1.2109, | |
| "step": 50800 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 4.85755344897136e-05, | |
| "loss": 1.2168, | |
| "step": 50850 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 4.857413383532787e-05, | |
| "loss": 1.2266, | |
| "step": 50900 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 4.8572733180942135e-05, | |
| "loss": 1.2056, | |
| "step": 50950 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.8571332526556407e-05, | |
| "loss": 1.2467, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.8569931872170685e-05, | |
| "loss": 1.2152, | |
| "step": 51050 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.856853121778495e-05, | |
| "loss": 1.2269, | |
| "step": 51100 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.856713056339922e-05, | |
| "loss": 1.2012, | |
| "step": 51150 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.856572990901349e-05, | |
| "loss": 1.2035, | |
| "step": 51200 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 4.8564329254627765e-05, | |
| "loss": 1.2232, | |
| "step": 51250 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 4.8562928600242036e-05, | |
| "loss": 1.2268, | |
| "step": 51300 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 4.856152794585631e-05, | |
| "loss": 1.1912, | |
| "step": 51350 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 4.856012729147057e-05, | |
| "loss": 1.1985, | |
| "step": 51400 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.855872663708485e-05, | |
| "loss": 1.2232, | |
| "step": 51450 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.855732598269912e-05, | |
| "loss": 1.2355, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.855592532831339e-05, | |
| "loss": 1.2261, | |
| "step": 51550 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 4.855452467392766e-05, | |
| "loss": 1.2322, | |
| "step": 51600 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.855312401954194e-05, | |
| "loss": 1.2263, | |
| "step": 51650 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.85517233651562e-05, | |
| "loss": 1.1849, | |
| "step": 51700 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.8550322710770474e-05, | |
| "loss": 1.2083, | |
| "step": 51750 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.8548922056384746e-05, | |
| "loss": 1.2184, | |
| "step": 51800 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 4.854752140199902e-05, | |
| "loss": 1.2076, | |
| "step": 51850 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.854612074761329e-05, | |
| "loss": 1.2223, | |
| "step": 51900 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.854472009322756e-05, | |
| "loss": 1.2183, | |
| "step": 51950 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.8543319438841826e-05, | |
| "loss": 1.2058, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.8541918784456104e-05, | |
| "loss": 1.2209, | |
| "step": 52050 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.8540518130070376e-05, | |
| "loss": 1.1996, | |
| "step": 52100 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.853911747568464e-05, | |
| "loss": 1.216, | |
| "step": 52150 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.853771682129891e-05, | |
| "loss": 1.2479, | |
| "step": 52200 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.8536316166913184e-05, | |
| "loss": 1.2443, | |
| "step": 52250 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 4.8534915512527456e-05, | |
| "loss": 1.2272, | |
| "step": 52300 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 4.853351485814173e-05, | |
| "loss": 1.2185, | |
| "step": 52350 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 4.8532114203756e-05, | |
| "loss": 1.2367, | |
| "step": 52400 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 4.8530713549370264e-05, | |
| "loss": 1.2073, | |
| "step": 52450 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 4.852931289498454e-05, | |
| "loss": 1.2098, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.8527912240598814e-05, | |
| "loss": 1.2523, | |
| "step": 52550 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.852651158621308e-05, | |
| "loss": 1.2031, | |
| "step": 52600 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.852511093182735e-05, | |
| "loss": 1.2386, | |
| "step": 52650 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.852371027744163e-05, | |
| "loss": 1.22, | |
| "step": 52700 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.8522309623055893e-05, | |
| "loss": 1.199, | |
| "step": 52750 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.8520908968670165e-05, | |
| "loss": 1.2325, | |
| "step": 52800 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.851950831428444e-05, | |
| "loss": 1.2192, | |
| "step": 52850 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.851810765989871e-05, | |
| "loss": 1.2283, | |
| "step": 52900 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.851670700551298e-05, | |
| "loss": 1.2121, | |
| "step": 52950 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.851530635112725e-05, | |
| "loss": 1.2178, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.8513905696741516e-05, | |
| "loss": 1.2375, | |
| "step": 53050 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.8512505042355795e-05, | |
| "loss": 1.2134, | |
| "step": 53100 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.8511104387970067e-05, | |
| "loss": 1.2103, | |
| "step": 53150 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 4.850970373358433e-05, | |
| "loss": 1.2134, | |
| "step": 53200 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.85083030791986e-05, | |
| "loss": 1.2352, | |
| "step": 53250 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.8506902424812875e-05, | |
| "loss": 1.2261, | |
| "step": 53300 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.8505501770427146e-05, | |
| "loss": 1.2092, | |
| "step": 53350 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.850410111604142e-05, | |
| "loss": 1.1907, | |
| "step": 53400 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.850270046165569e-05, | |
| "loss": 1.2244, | |
| "step": 53450 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.8501299807269954e-05, | |
| "loss": 1.2199, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.849989915288423e-05, | |
| "loss": 1.2204, | |
| "step": 53550 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.8498498498498504e-05, | |
| "loss": 1.2152, | |
| "step": 53600 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 4.849709784411277e-05, | |
| "loss": 1.2411, | |
| "step": 53650 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 4.849569718972704e-05, | |
| "loss": 1.1775, | |
| "step": 53700 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 4.849429653534131e-05, | |
| "loss": 1.2329, | |
| "step": 53750 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 4.8492895880955584e-05, | |
| "loss": 1.1837, | |
| "step": 53800 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 4.8491495226569856e-05, | |
| "loss": 1.2036, | |
| "step": 53850 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 4.849009457218413e-05, | |
| "loss": 1.2453, | |
| "step": 53900 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 4.84886939177984e-05, | |
| "loss": 1.1989, | |
| "step": 53950 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 4.848729326341267e-05, | |
| "loss": 1.2009, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 4.848589260902694e-05, | |
| "loss": 1.2301, | |
| "step": 54050 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 4.848449195464121e-05, | |
| "loss": 1.1844, | |
| "step": 54100 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 4.8483091300255486e-05, | |
| "loss": 1.2068, | |
| "step": 54150 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 4.848169064586976e-05, | |
| "loss": 1.2124, | |
| "step": 54200 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 4.848028999148402e-05, | |
| "loss": 1.2168, | |
| "step": 54250 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 4.8478889337098294e-05, | |
| "loss": 1.221, | |
| "step": 54300 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 4.8477488682712565e-05, | |
| "loss": 1.2268, | |
| "step": 54350 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 4.847608802832684e-05, | |
| "loss": 1.2148, | |
| "step": 54400 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 4.847468737394111e-05, | |
| "loss": 1.229, | |
| "step": 54450 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 4.847328671955538e-05, | |
| "loss": 1.2287, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 4.8471886065169645e-05, | |
| "loss": 1.2305, | |
| "step": 54550 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 4.8470485410783924e-05, | |
| "loss": 1.2139, | |
| "step": 54600 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 4.8469084756398195e-05, | |
| "loss": 1.2446, | |
| "step": 54650 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 4.846768410201246e-05, | |
| "loss": 1.2166, | |
| "step": 54700 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 4.846628344762673e-05, | |
| "loss": 1.2111, | |
| "step": 54750 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 4.8464882793241003e-05, | |
| "loss": 1.2257, | |
| "step": 54800 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 4.8463482138855275e-05, | |
| "loss": 1.2241, | |
| "step": 54850 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 4.846208148446955e-05, | |
| "loss": 1.214, | |
| "step": 54900 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 4.846068083008382e-05, | |
| "loss": 1.2435, | |
| "step": 54950 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 4.845928017569809e-05, | |
| "loss": 1.2214, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 4.845787952131236e-05, | |
| "loss": 1.2371, | |
| "step": 55050 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 4.845647886692663e-05, | |
| "loss": 1.2145, | |
| "step": 55100 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 4.84550782125409e-05, | |
| "loss": 1.2035, | |
| "step": 55150 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 4.8453677558155176e-05, | |
| "loss": 1.2216, | |
| "step": 55200 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 4.845227690376944e-05, | |
| "loss": 1.2122, | |
| "step": 55250 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 4.845087624938371e-05, | |
| "loss": 1.235, | |
| "step": 55300 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 4.8449475594997985e-05, | |
| "loss": 1.2226, | |
| "step": 55350 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 4.8448074940612256e-05, | |
| "loss": 1.1771, | |
| "step": 55400 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 4.844667428622653e-05, | |
| "loss": 1.1974, | |
| "step": 55450 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 4.84452736318408e-05, | |
| "loss": 1.1854, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 4.844387297745507e-05, | |
| "loss": 1.1791, | |
| "step": 55550 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 4.8442472323069336e-05, | |
| "loss": 1.2525, | |
| "step": 55600 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 4.8441071668683614e-05, | |
| "loss": 1.1801, | |
| "step": 55650 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 4.843967101429788e-05, | |
| "loss": 1.2178, | |
| "step": 55700 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 4.843827035991215e-05, | |
| "loss": 1.2229, | |
| "step": 55750 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 4.843686970552642e-05, | |
| "loss": 1.2264, | |
| "step": 55800 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 4.8435469051140694e-05, | |
| "loss": 1.2389, | |
| "step": 55850 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 4.8434068396754966e-05, | |
| "loss": 1.2246, | |
| "step": 55900 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 4.843266774236924e-05, | |
| "loss": 1.2117, | |
| "step": 55950 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 4.843126708798351e-05, | |
| "loss": 1.1933, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 4.842986643359778e-05, | |
| "loss": 1.2031, | |
| "step": 56050 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 4.842846577921205e-05, | |
| "loss": 1.1818, | |
| "step": 56100 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.8427065124826324e-05, | |
| "loss": 1.2084, | |
| "step": 56150 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.842566447044059e-05, | |
| "loss": 1.2158, | |
| "step": 56200 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.842426381605487e-05, | |
| "loss": 1.2176, | |
| "step": 56250 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.842286316166913e-05, | |
| "loss": 1.2081, | |
| "step": 56300 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.8421462507283404e-05, | |
| "loss": 1.2115, | |
| "step": 56350 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.8420061852897675e-05, | |
| "loss": 1.2184, | |
| "step": 56400 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.841866119851195e-05, | |
| "loss": 1.2222, | |
| "step": 56450 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.841726054412622e-05, | |
| "loss": 1.1938, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 4.841585988974049e-05, | |
| "loss": 1.1925, | |
| "step": 56550 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.841445923535476e-05, | |
| "loss": 1.2015, | |
| "step": 56600 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.841305858096903e-05, | |
| "loss": 1.2042, | |
| "step": 56650 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.8411657926583305e-05, | |
| "loss": 1.2039, | |
| "step": 56700 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 4.841025727219757e-05, | |
| "loss": 1.2061, | |
| "step": 56750 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.840885661781184e-05, | |
| "loss": 1.1724, | |
| "step": 56800 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.840745596342611e-05, | |
| "loss": 1.2074, | |
| "step": 56850 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.8406055309040385e-05, | |
| "loss": 1.2245, | |
| "step": 56900 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.840465465465466e-05, | |
| "loss": 1.2051, | |
| "step": 56950 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.840325400026893e-05, | |
| "loss": 1.2219, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.84018533458832e-05, | |
| "loss": 1.1998, | |
| "step": 57050 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.840045269149747e-05, | |
| "loss": 1.2152, | |
| "step": 57100 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.839905203711174e-05, | |
| "loss": 1.2081, | |
| "step": 57150 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 4.839765138272601e-05, | |
| "loss": 1.2091, | |
| "step": 57200 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.839625072834028e-05, | |
| "loss": 1.2206, | |
| "step": 57250 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.839485007395456e-05, | |
| "loss": 1.2193, | |
| "step": 57300 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.839344941956882e-05, | |
| "loss": 1.2062, | |
| "step": 57350 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.8392048765183095e-05, | |
| "loss": 1.2025, | |
| "step": 57400 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 4.8390648110797366e-05, | |
| "loss": 1.2136, | |
| "step": 57450 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.838924745641164e-05, | |
| "loss": 1.227, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.838784680202591e-05, | |
| "loss": 1.1956, | |
| "step": 57550 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.838644614764018e-05, | |
| "loss": 1.2036, | |
| "step": 57600 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.838504549325445e-05, | |
| "loss": 1.2001, | |
| "step": 57650 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.838364483886872e-05, | |
| "loss": 1.1784, | |
| "step": 57700 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.8382244184482996e-05, | |
| "loss": 1.2165, | |
| "step": 57750 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.838084353009726e-05, | |
| "loss": 1.1924, | |
| "step": 57800 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.837944287571153e-05, | |
| "loss": 1.214, | |
| "step": 57850 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.8378042221325804e-05, | |
| "loss": 1.19, | |
| "step": 57900 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.8376641566940076e-05, | |
| "loss": 1.2157, | |
| "step": 57950 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.837524091255435e-05, | |
| "loss": 1.2004, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.837384025816862e-05, | |
| "loss": 1.1887, | |
| "step": 58050 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 4.837243960378289e-05, | |
| "loss": 1.1884, | |
| "step": 58100 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.837103894939716e-05, | |
| "loss": 1.1989, | |
| "step": 58150 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.8369638295011434e-05, | |
| "loss": 1.2413, | |
| "step": 58200 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.83682376406257e-05, | |
| "loss": 1.1958, | |
| "step": 58250 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.836683698623997e-05, | |
| "loss": 1.2177, | |
| "step": 58300 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.836543633185425e-05, | |
| "loss": 1.2274, | |
| "step": 58350 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.8364035677468514e-05, | |
| "loss": 1.2218, | |
| "step": 58400 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.8362635023082785e-05, | |
| "loss": 1.2179, | |
| "step": 58450 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.836123436869706e-05, | |
| "loss": 1.1932, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 4.835983371431133e-05, | |
| "loss": 1.2312, | |
| "step": 58550 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.83584330599256e-05, | |
| "loss": 1.207, | |
| "step": 58600 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.835703240553987e-05, | |
| "loss": 1.1956, | |
| "step": 58650 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.835563175115414e-05, | |
| "loss": 1.1989, | |
| "step": 58700 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.835423109676841e-05, | |
| "loss": 1.2206, | |
| "step": 58750 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.835283044238269e-05, | |
| "loss": 1.1979, | |
| "step": 58800 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.835142978799695e-05, | |
| "loss": 1.2383, | |
| "step": 58850 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.835002913361122e-05, | |
| "loss": 1.1947, | |
| "step": 58900 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.8348628479225495e-05, | |
| "loss": 1.1791, | |
| "step": 58950 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 4.8347227824839767e-05, | |
| "loss": 1.208, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.834582717045404e-05, | |
| "loss": 1.2088, | |
| "step": 59050 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.834442651606831e-05, | |
| "loss": 1.1862, | |
| "step": 59100 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.834302586168258e-05, | |
| "loss": 1.1975, | |
| "step": 59150 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.834162520729685e-05, | |
| "loss": 1.2037, | |
| "step": 59200 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.8340224552911125e-05, | |
| "loss": 1.1905, | |
| "step": 59250 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.833882389852539e-05, | |
| "loss": 1.22, | |
| "step": 59300 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.833742324413966e-05, | |
| "loss": 1.1894, | |
| "step": 59350 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.833602258975394e-05, | |
| "loss": 1.1658, | |
| "step": 59400 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 4.8334621935368205e-05, | |
| "loss": 1.2109, | |
| "step": 59450 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.8333221280982476e-05, | |
| "loss": 1.1792, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.833182062659675e-05, | |
| "loss": 1.2071, | |
| "step": 59550 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.833041997221102e-05, | |
| "loss": 1.212, | |
| "step": 59600 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 4.832901931782529e-05, | |
| "loss": 1.2113, | |
| "step": 59650 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.832761866343956e-05, | |
| "loss": 1.2427, | |
| "step": 59700 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.832621800905383e-05, | |
| "loss": 1.2051, | |
| "step": 59750 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.8324817354668106e-05, | |
| "loss": 1.2013, | |
| "step": 59800 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.832341670028238e-05, | |
| "loss": 1.1794, | |
| "step": 59850 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.832201604589664e-05, | |
| "loss": 1.1997, | |
| "step": 59900 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.8320615391510914e-05, | |
| "loss": 1.2033, | |
| "step": 59950 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.8319214737125186e-05, | |
| "loss": 1.2245, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.831781408273946e-05, | |
| "loss": 1.2218, | |
| "step": 60050 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 4.831641342835373e-05, | |
| "loss": 1.1802, | |
| "step": 60100 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.8315012773968e-05, | |
| "loss": 1.1845, | |
| "step": 60150 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.8313612119582266e-05, | |
| "loss": 1.2241, | |
| "step": 60200 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.8312211465196544e-05, | |
| "loss": 1.2201, | |
| "step": 60250 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.8310810810810816e-05, | |
| "loss": 1.1924, | |
| "step": 60300 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 4.830941015642508e-05, | |
| "loss": 1.208, | |
| "step": 60350 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.830800950203935e-05, | |
| "loss": 1.1838, | |
| "step": 60400 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.830660884765363e-05, | |
| "loss": 1.1818, | |
| "step": 60450 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.8305208193267895e-05, | |
| "loss": 1.1835, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.830380753888217e-05, | |
| "loss": 1.2058, | |
| "step": 60550 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.830240688449644e-05, | |
| "loss": 1.184, | |
| "step": 60600 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.830100623011071e-05, | |
| "loss": 1.2118, | |
| "step": 60650 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.829960557572498e-05, | |
| "loss": 1.2301, | |
| "step": 60700 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 4.8298204921339253e-05, | |
| "loss": 1.1792, | |
| "step": 60750 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.829680426695352e-05, | |
| "loss": 1.2141, | |
| "step": 60800 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.82954036125678e-05, | |
| "loss": 1.2089, | |
| "step": 60850 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.829400295818207e-05, | |
| "loss": 1.2261, | |
| "step": 60900 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.829260230379633e-05, | |
| "loss": 1.2155, | |
| "step": 60950 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.8291201649410605e-05, | |
| "loss": 1.201, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.8289800995024877e-05, | |
| "loss": 1.1915, | |
| "step": 61050 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.828840034063915e-05, | |
| "loss": 1.1811, | |
| "step": 61100 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.828699968625342e-05, | |
| "loss": 1.2219, | |
| "step": 61150 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.828559903186769e-05, | |
| "loss": 1.1838, | |
| "step": 61200 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.8284198377481956e-05, | |
| "loss": 1.1888, | |
| "step": 61250 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.8282797723096235e-05, | |
| "loss": 1.2009, | |
| "step": 61300 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.8281397068710506e-05, | |
| "loss": 1.2059, | |
| "step": 61350 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.827999641432477e-05, | |
| "loss": 1.2007, | |
| "step": 61400 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.827859575993904e-05, | |
| "loss": 1.2421, | |
| "step": 61450 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.827719510555332e-05, | |
| "loss": 1.1792, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.8275794451167586e-05, | |
| "loss": 1.2004, | |
| "step": 61550 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.827439379678186e-05, | |
| "loss": 1.2036, | |
| "step": 61600 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.827299314239613e-05, | |
| "loss": 1.2353, | |
| "step": 61650 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.82715924880104e-05, | |
| "loss": 1.1929, | |
| "step": 61700 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.827019183362467e-05, | |
| "loss": 1.1829, | |
| "step": 61750 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.8268791179238944e-05, | |
| "loss": 1.2093, | |
| "step": 61800 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.826739052485321e-05, | |
| "loss": 1.1955, | |
| "step": 61850 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.826598987046749e-05, | |
| "loss": 1.2073, | |
| "step": 61900 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.826458921608176e-05, | |
| "loss": 1.1858, | |
| "step": 61950 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.8263188561696024e-05, | |
| "loss": 1.2025, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.8261787907310296e-05, | |
| "loss": 1.2035, | |
| "step": 62050 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.826038725292457e-05, | |
| "loss": 1.1939, | |
| "step": 62100 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.825898659853884e-05, | |
| "loss": 1.187, | |
| "step": 62150 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.825758594415311e-05, | |
| "loss": 1.1798, | |
| "step": 62200 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.825618528976738e-05, | |
| "loss": 1.1662, | |
| "step": 62250 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.825478463538165e-05, | |
| "loss": 1.2151, | |
| "step": 62300 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.8253383980995926e-05, | |
| "loss": 1.1957, | |
| "step": 62350 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.82519833266102e-05, | |
| "loss": 1.1992, | |
| "step": 62400 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.825058267222446e-05, | |
| "loss": 1.184, | |
| "step": 62450 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.8249182017838734e-05, | |
| "loss": 1.2138, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.824778136345301e-05, | |
| "loss": 1.2103, | |
| "step": 62550 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.824638070906728e-05, | |
| "loss": 1.2085, | |
| "step": 62600 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.824498005468155e-05, | |
| "loss": 1.1934, | |
| "step": 62650 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.824357940029582e-05, | |
| "loss": 1.1775, | |
| "step": 62700 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.824217874591009e-05, | |
| "loss": 1.194, | |
| "step": 62750 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 4.8240778091524363e-05, | |
| "loss": 1.1881, | |
| "step": 62800 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.8239377437138635e-05, | |
| "loss": 1.2089, | |
| "step": 62850 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.82379767827529e-05, | |
| "loss": 1.2112, | |
| "step": 62900 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.823657612836718e-05, | |
| "loss": 1.1858, | |
| "step": 62950 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 4.823517547398145e-05, | |
| "loss": 1.2052, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.8233774819595715e-05, | |
| "loss": 1.1938, | |
| "step": 63050 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.8232374165209986e-05, | |
| "loss": 1.1854, | |
| "step": 63100 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.823097351082426e-05, | |
| "loss": 1.1946, | |
| "step": 63150 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.822957285643853e-05, | |
| "loss": 1.1879, | |
| "step": 63200 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 4.82281722020528e-05, | |
| "loss": 1.1887, | |
| "step": 63250 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.822677154766707e-05, | |
| "loss": 1.2325, | |
| "step": 63300 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.822537089328134e-05, | |
| "loss": 1.2207, | |
| "step": 63350 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.8223970238895616e-05, | |
| "loss": 1.1916, | |
| "step": 63400 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.822256958450989e-05, | |
| "loss": 1.1968, | |
| "step": 63450 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 4.822116893012415e-05, | |
| "loss": 1.2071, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 4.8219768275738424e-05, | |
| "loss": 1.1849, | |
| "step": 63550 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 4.82183676213527e-05, | |
| "loss": 1.1992, | |
| "step": 63600 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 4.821696696696697e-05, | |
| "loss": 1.1895, | |
| "step": 63650 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.821556631258124e-05, | |
| "loss": 1.2329, | |
| "step": 63700 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.821416565819551e-05, | |
| "loss": 1.1962, | |
| "step": 63750 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.821276500380978e-05, | |
| "loss": 1.2112, | |
| "step": 63800 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.8211364349424054e-05, | |
| "loss": 1.204, | |
| "step": 63850 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 4.8209963695038326e-05, | |
| "loss": 1.1842, | |
| "step": 63900 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.820856304065259e-05, | |
| "loss": 1.1746, | |
| "step": 63950 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.820716238626687e-05, | |
| "loss": 1.2031, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.820576173188114e-05, | |
| "loss": 1.1804, | |
| "step": 64050 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.8204361077495406e-05, | |
| "loss": 1.2164, | |
| "step": 64100 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.820296042310968e-05, | |
| "loss": 1.2014, | |
| "step": 64150 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.8201559768723956e-05, | |
| "loss": 1.1894, | |
| "step": 64200 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.820015911433822e-05, | |
| "loss": 1.1824, | |
| "step": 64250 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.819875845995249e-05, | |
| "loss": 1.1798, | |
| "step": 64300 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 4.8197357805566764e-05, | |
| "loss": 1.2037, | |
| "step": 64350 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 4.819595715118103e-05, | |
| "loss": 1.2034, | |
| "step": 64400 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 4.819455649679531e-05, | |
| "loss": 1.2143, | |
| "step": 64450 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 4.819315584240958e-05, | |
| "loss": 1.1951, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 4.8191755188023844e-05, | |
| "loss": 1.2028, | |
| "step": 64550 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.8190354533638115e-05, | |
| "loss": 1.2084, | |
| "step": 64600 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.8188953879252394e-05, | |
| "loss": 1.1917, | |
| "step": 64650 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.818755322486666e-05, | |
| "loss": 1.1943, | |
| "step": 64700 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.818615257048093e-05, | |
| "loss": 1.1768, | |
| "step": 64750 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.81847519160952e-05, | |
| "loss": 1.2175, | |
| "step": 64800 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.8183351261709473e-05, | |
| "loss": 1.2022, | |
| "step": 64850 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.8181950607323745e-05, | |
| "loss": 1.1776, | |
| "step": 64900 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.818054995293802e-05, | |
| "loss": 1.1815, | |
| "step": 64950 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 4.817914929855228e-05, | |
| "loss": 1.2024, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.817774864416656e-05, | |
| "loss": 1.19, | |
| "step": 65050 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.817634798978083e-05, | |
| "loss": 1.1711, | |
| "step": 65100 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.8174947335395096e-05, | |
| "loss": 1.205, | |
| "step": 65150 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.817354668100937e-05, | |
| "loss": 1.2095, | |
| "step": 65200 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.8172146026623646e-05, | |
| "loss": 1.2057, | |
| "step": 65250 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 4.817074537223791e-05, | |
| "loss": 1.2161, | |
| "step": 65300 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 4.816934471785218e-05, | |
| "loss": 1.2183, | |
| "step": 65350 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 4.8167944063466455e-05, | |
| "loss": 1.1944, | |
| "step": 65400 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 4.816654340908072e-05, | |
| "loss": 1.1997, | |
| "step": 65450 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 4.8165142754695e-05, | |
| "loss": 1.2016, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 4.816374210030927e-05, | |
| "loss": 1.2132, | |
| "step": 65550 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 4.8162341445923534e-05, | |
| "loss": 1.1729, | |
| "step": 65600 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 4.8160940791537806e-05, | |
| "loss": 1.1957, | |
| "step": 65650 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 4.8159540137152084e-05, | |
| "loss": 1.2012, | |
| "step": 65700 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 4.815813948276635e-05, | |
| "loss": 1.1959, | |
| "step": 65750 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 4.815673882838062e-05, | |
| "loss": 1.2101, | |
| "step": 65800 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 4.815533817399489e-05, | |
| "loss": 1.1908, | |
| "step": 65850 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 4.8153937519609164e-05, | |
| "loss": 1.1917, | |
| "step": 65900 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.8152536865223436e-05, | |
| "loss": 1.1876, | |
| "step": 65950 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.815113621083771e-05, | |
| "loss": 1.185, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.814973555645197e-05, | |
| "loss": 1.1991, | |
| "step": 66050 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.814833490206625e-05, | |
| "loss": 1.1658, | |
| "step": 66100 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.814693424768052e-05, | |
| "loss": 1.1782, | |
| "step": 66150 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.814553359329479e-05, | |
| "loss": 1.1902, | |
| "step": 66200 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.814413293890906e-05, | |
| "loss": 1.2057, | |
| "step": 66250 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.814273228452334e-05, | |
| "loss": 1.19, | |
| "step": 66300 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 4.81413316301376e-05, | |
| "loss": 1.2044, | |
| "step": 66350 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.8139930975751874e-05, | |
| "loss": 1.1954, | |
| "step": 66400 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.8138530321366145e-05, | |
| "loss": 1.1804, | |
| "step": 66450 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.813712966698041e-05, | |
| "loss": 1.2178, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.813572901259469e-05, | |
| "loss": 1.2109, | |
| "step": 66550 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.813432835820896e-05, | |
| "loss": 1.1651, | |
| "step": 66600 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.8132927703823225e-05, | |
| "loss": 1.2187, | |
| "step": 66650 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.81315270494375e-05, | |
| "loss": 1.2112, | |
| "step": 66700 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.8130126395051775e-05, | |
| "loss": 1.1817, | |
| "step": 66750 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 4.812872574066604e-05, | |
| "loss": 1.1852, | |
| "step": 66800 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.812732508628031e-05, | |
| "loss": 1.2272, | |
| "step": 66850 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.812592443189458e-05, | |
| "loss": 1.2002, | |
| "step": 66900 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.8124523777508855e-05, | |
| "loss": 1.1906, | |
| "step": 66950 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 4.812312312312313e-05, | |
| "loss": 1.1791, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4.81217224687374e-05, | |
| "loss": 1.1799, | |
| "step": 67050 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4.812032181435166e-05, | |
| "loss": 1.1557, | |
| "step": 67100 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4.811892115996594e-05, | |
| "loss": 1.1781, | |
| "step": 67150 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4.811752050558021e-05, | |
| "loss": 1.2152, | |
| "step": 67200 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 4.811611985119448e-05, | |
| "loss": 1.2041, | |
| "step": 67250 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 4.811471919680875e-05, | |
| "loss": 1.1565, | |
| "step": 67300 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 4.811331854242303e-05, | |
| "loss": 1.1857, | |
| "step": 67350 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 4.811191788803729e-05, | |
| "loss": 1.1949, | |
| "step": 67400 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 4.8110517233651565e-05, | |
| "loss": 1.1924, | |
| "step": 67450 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 4.8109116579265836e-05, | |
| "loss": 1.1577, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 4.810771592488011e-05, | |
| "loss": 1.1872, | |
| "step": 67550 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 4.810631527049438e-05, | |
| "loss": 1.1786, | |
| "step": 67600 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 4.810491461610865e-05, | |
| "loss": 1.2033, | |
| "step": 67650 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 4.8103513961722916e-05, | |
| "loss": 1.1505, | |
| "step": 67700 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 4.810211330733719e-05, | |
| "loss": 1.2049, | |
| "step": 67750 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 4.8100712652951466e-05, | |
| "loss": 1.172, | |
| "step": 67800 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 4.809931199856573e-05, | |
| "loss": 1.1887, | |
| "step": 67850 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 4.809791134418e-05, | |
| "loss": 1.1769, | |
| "step": 67900 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 4.8096510689794274e-05, | |
| "loss": 1.1956, | |
| "step": 67950 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 4.8095110035408546e-05, | |
| "loss": 1.1847, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 4.809370938102282e-05, | |
| "loss": 1.2055, | |
| "step": 68050 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 4.809230872663709e-05, | |
| "loss": 1.2131, | |
| "step": 68100 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 4.8090908072251354e-05, | |
| "loss": 1.1655, | |
| "step": 68150 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 4.808950741786563e-05, | |
| "loss": 1.2098, | |
| "step": 68200 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 4.8088106763479904e-05, | |
| "loss": 1.1924, | |
| "step": 68250 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 4.808670610909417e-05, | |
| "loss": 1.2078, | |
| "step": 68300 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 4.808530545470844e-05, | |
| "loss": 1.1757, | |
| "step": 68350 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 4.808390480032272e-05, | |
| "loss": 1.1702, | |
| "step": 68400 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 4.8082504145936984e-05, | |
| "loss": 1.1891, | |
| "step": 68450 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 4.8081103491551255e-05, | |
| "loss": 1.1555, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 4.807970283716553e-05, | |
| "loss": 1.1996, | |
| "step": 68550 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 4.80783021827798e-05, | |
| "loss": 1.1781, | |
| "step": 68600 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 4.807690152839407e-05, | |
| "loss": 1.1918, | |
| "step": 68650 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 4.807550087400834e-05, | |
| "loss": 1.1852, | |
| "step": 68700 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 4.807410021962261e-05, | |
| "loss": 1.1847, | |
| "step": 68750 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 4.807269956523688e-05, | |
| "loss": 1.1803, | |
| "step": 68800 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 4.807129891085116e-05, | |
| "loss": 1.1916, | |
| "step": 68850 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 4.806989825646542e-05, | |
| "loss": 1.1877, | |
| "step": 68900 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 4.806849760207969e-05, | |
| "loss": 1.1928, | |
| "step": 68950 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 4.8067096947693965e-05, | |
| "loss": 1.1607, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 4.8065696293308237e-05, | |
| "loss": 1.1952, | |
| "step": 69050 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 4.806429563892251e-05, | |
| "loss": 1.1779, | |
| "step": 69100 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 4.806289498453678e-05, | |
| "loss": 1.1702, | |
| "step": 69150 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 4.8061494330151045e-05, | |
| "loss": 1.1463, | |
| "step": 69200 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 4.806009367576532e-05, | |
| "loss": 1.1788, | |
| "step": 69250 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 4.8058693021379595e-05, | |
| "loss": 1.1471, | |
| "step": 69300 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 4.805729236699386e-05, | |
| "loss": 1.195, | |
| "step": 69350 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 4.805589171260813e-05, | |
| "loss": 1.1965, | |
| "step": 69400 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 4.80544910582224e-05, | |
| "loss": 1.1807, | |
| "step": 69450 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 4.8053090403836675e-05, | |
| "loss": 1.181, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 4.8051689749450946e-05, | |
| "loss": 1.1963, | |
| "step": 69550 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 4.805028909506522e-05, | |
| "loss": 1.2128, | |
| "step": 69600 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 4.804888844067949e-05, | |
| "loss": 1.204, | |
| "step": 69650 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 4.804748778629376e-05, | |
| "loss": 1.191, | |
| "step": 69700 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 4.804608713190803e-05, | |
| "loss": 1.1903, | |
| "step": 69750 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 4.80446864775223e-05, | |
| "loss": 1.1931, | |
| "step": 69800 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 4.804328582313657e-05, | |
| "loss": 1.1751, | |
| "step": 69850 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 4.804188516875085e-05, | |
| "loss": 1.1742, | |
| "step": 69900 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 4.804048451436511e-05, | |
| "loss": 1.2178, | |
| "step": 69950 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 4.8039083859979384e-05, | |
| "loss": 1.2015, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 4.8037683205593656e-05, | |
| "loss": 1.1839, | |
| "step": 70050 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 4.803628255120793e-05, | |
| "loss": 1.2029, | |
| "step": 70100 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 4.80348818968222e-05, | |
| "loss": 1.2018, | |
| "step": 70150 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 4.803348124243647e-05, | |
| "loss": 1.1907, | |
| "step": 70200 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 4.8032080588050735e-05, | |
| "loss": 1.2129, | |
| "step": 70250 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 4.8030679933665014e-05, | |
| "loss": 1.1822, | |
| "step": 70300 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 4.8029279279279286e-05, | |
| "loss": 1.1658, | |
| "step": 70350 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 4.802787862489355e-05, | |
| "loss": 1.2053, | |
| "step": 70400 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 4.802647797050782e-05, | |
| "loss": 1.1671, | |
| "step": 70450 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 4.8025077316122094e-05, | |
| "loss": 1.1622, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 4.8023676661736365e-05, | |
| "loss": 1.18, | |
| "step": 70550 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 4.802227600735064e-05, | |
| "loss": 1.1882, | |
| "step": 70600 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 4.802087535296491e-05, | |
| "loss": 1.1913, | |
| "step": 70650 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 4.801947469857918e-05, | |
| "loss": 1.1983, | |
| "step": 70700 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 4.801807404419345e-05, | |
| "loss": 1.1769, | |
| "step": 70750 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 4.8016673389807723e-05, | |
| "loss": 1.1798, | |
| "step": 70800 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 4.801527273542199e-05, | |
| "loss": 1.16, | |
| "step": 70850 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 4.801387208103626e-05, | |
| "loss": 1.1444, | |
| "step": 70900 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 4.801247142665053e-05, | |
| "loss": 1.1725, | |
| "step": 70950 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 4.80110707722648e-05, | |
| "loss": 1.1695, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 4.8009670117879075e-05, | |
| "loss": 1.1794, | |
| "step": 71050 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 4.8008269463493347e-05, | |
| "loss": 1.1908, | |
| "step": 71100 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 4.800686880910762e-05, | |
| "loss": 1.1891, | |
| "step": 71150 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 4.800546815472189e-05, | |
| "loss": 1.1919, | |
| "step": 71200 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 4.800406750033616e-05, | |
| "loss": 1.1828, | |
| "step": 71250 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 4.8002666845950426e-05, | |
| "loss": 1.1922, | |
| "step": 71300 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 4.8001266191564705e-05, | |
| "loss": 1.1989, | |
| "step": 71350 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 4.799986553717897e-05, | |
| "loss": 1.169, | |
| "step": 71400 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 4.799846488279324e-05, | |
| "loss": 1.1873, | |
| "step": 71450 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 4.799706422840751e-05, | |
| "loss": 1.2123, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 4.7995663574021784e-05, | |
| "loss": 1.1478, | |
| "step": 71550 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 4.7994262919636056e-05, | |
| "loss": 1.167, | |
| "step": 71600 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 4.799286226525033e-05, | |
| "loss": 1.1834, | |
| "step": 71650 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 4.79914616108646e-05, | |
| "loss": 1.1541, | |
| "step": 71700 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 4.799006095647887e-05, | |
| "loss": 1.1843, | |
| "step": 71750 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 4.798866030209314e-05, | |
| "loss": 1.1918, | |
| "step": 71800 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 4.7987259647707414e-05, | |
| "loss": 1.199, | |
| "step": 71850 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 4.798585899332168e-05, | |
| "loss": 1.186, | |
| "step": 71900 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 4.798445833893596e-05, | |
| "loss": 1.1835, | |
| "step": 71950 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 4.798305768455022e-05, | |
| "loss": 1.1883, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 4.7981657030164494e-05, | |
| "loss": 1.1709, | |
| "step": 72050 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 4.7980256375778766e-05, | |
| "loss": 1.1745, | |
| "step": 72100 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 4.797885572139304e-05, | |
| "loss": 1.1799, | |
| "step": 72150 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 4.797745506700731e-05, | |
| "loss": 1.1752, | |
| "step": 72200 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 4.797605441262158e-05, | |
| "loss": 1.2008, | |
| "step": 72250 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 4.797465375823585e-05, | |
| "loss": 1.1907, | |
| "step": 72300 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 4.797325310385012e-05, | |
| "loss": 1.1852, | |
| "step": 72350 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 4.7971852449464395e-05, | |
| "loss": 1.1735, | |
| "step": 72400 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 4.797045179507866e-05, | |
| "loss": 1.1996, | |
| "step": 72450 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 4.796905114069293e-05, | |
| "loss": 1.1798, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 4.7967650486307204e-05, | |
| "loss": 1.2018, | |
| "step": 72550 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 4.7966249831921475e-05, | |
| "loss": 1.1704, | |
| "step": 72600 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 4.796484917753575e-05, | |
| "loss": 1.1998, | |
| "step": 72650 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 4.796344852315002e-05, | |
| "loss": 1.1909, | |
| "step": 72700 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 4.796204786876429e-05, | |
| "loss": 1.168, | |
| "step": 72750 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 4.796064721437856e-05, | |
| "loss": 1.1671, | |
| "step": 72800 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 4.7959246559992833e-05, | |
| "loss": 1.1733, | |
| "step": 72850 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 4.79578459056071e-05, | |
| "loss": 1.1839, | |
| "step": 72900 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 4.795644525122137e-05, | |
| "loss": 1.1614, | |
| "step": 72950 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 4.795504459683565e-05, | |
| "loss": 1.1931, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 4.795364394244991e-05, | |
| "loss": 1.1983, | |
| "step": 73050 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 4.7952243288064185e-05, | |
| "loss": 1.1815, | |
| "step": 73100 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 4.7950842633678456e-05, | |
| "loss": 1.2087, | |
| "step": 73150 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 4.794944197929273e-05, | |
| "loss": 1.1793, | |
| "step": 73200 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 4.7948041324907e-05, | |
| "loss": 1.1882, | |
| "step": 73250 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 4.794664067052127e-05, | |
| "loss": 1.1631, | |
| "step": 73300 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 4.794524001613554e-05, | |
| "loss": 1.17, | |
| "step": 73350 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 4.794383936174981e-05, | |
| "loss": 1.1803, | |
| "step": 73400 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 4.7942438707364086e-05, | |
| "loss": 1.1812, | |
| "step": 73450 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 4.794103805297835e-05, | |
| "loss": 1.1729, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 4.793963739859262e-05, | |
| "loss": 1.1689, | |
| "step": 73550 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 4.7938236744206894e-05, | |
| "loss": 1.216, | |
| "step": 73600 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 4.7936836089821166e-05, | |
| "loss": 1.1708, | |
| "step": 73650 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 4.793543543543544e-05, | |
| "loss": 1.183, | |
| "step": 73700 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 4.793403478104971e-05, | |
| "loss": 1.1879, | |
| "step": 73750 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 4.793263412666398e-05, | |
| "loss": 1.1595, | |
| "step": 73800 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 4.793123347227825e-05, | |
| "loss": 1.1903, | |
| "step": 73850 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 4.7929832817892524e-05, | |
| "loss": 1.211, | |
| "step": 73900 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 4.792843216350679e-05, | |
| "loss": 1.1945, | |
| "step": 73950 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 4.792703150912106e-05, | |
| "loss": 1.1992, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 4.792563085473534e-05, | |
| "loss": 1.1741, | |
| "step": 74050 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 4.7924230200349604e-05, | |
| "loss": 1.1773, | |
| "step": 74100 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 4.7922829545963876e-05, | |
| "loss": 1.2041, | |
| "step": 74150 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 4.792142889157815e-05, | |
| "loss": 1.1684, | |
| "step": 74200 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 4.792002823719242e-05, | |
| "loss": 1.1625, | |
| "step": 74250 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 4.791862758280669e-05, | |
| "loss": 1.1374, | |
| "step": 74300 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 4.791722692842096e-05, | |
| "loss": 1.1701, | |
| "step": 74350 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 4.791582627403523e-05, | |
| "loss": 1.1549, | |
| "step": 74400 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 4.79144256196495e-05, | |
| "loss": 1.1852, | |
| "step": 74450 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 4.791302496526378e-05, | |
| "loss": 1.1694, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 4.791162431087804e-05, | |
| "loss": 1.1832, | |
| "step": 74550 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 4.7910223656492314e-05, | |
| "loss": 1.1613, | |
| "step": 74600 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 4.7908823002106585e-05, | |
| "loss": 1.1594, | |
| "step": 74650 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 4.790742234772086e-05, | |
| "loss": 1.1749, | |
| "step": 74700 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 4.790602169333513e-05, | |
| "loss": 1.1597, | |
| "step": 74750 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 4.79046210389494e-05, | |
| "loss": 1.1667, | |
| "step": 74800 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 4.790322038456367e-05, | |
| "loss": 1.1633, | |
| "step": 74850 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 4.790181973017794e-05, | |
| "loss": 1.1702, | |
| "step": 74900 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 4.7900419075792215e-05, | |
| "loss": 1.1697, | |
| "step": 74950 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 4.789901842140648e-05, | |
| "loss": 1.2037, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 4.789761776702075e-05, | |
| "loss": 1.1664, | |
| "step": 75050 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 4.789621711263503e-05, | |
| "loss": 1.1631, | |
| "step": 75100 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 4.7894816458249295e-05, | |
| "loss": 1.1686, | |
| "step": 75150 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 4.7893415803863566e-05, | |
| "loss": 1.1614, | |
| "step": 75200 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 4.789201514947784e-05, | |
| "loss": 1.1816, | |
| "step": 75250 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 4.789061449509211e-05, | |
| "loss": 1.1659, | |
| "step": 75300 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 4.788921384070638e-05, | |
| "loss": 1.1634, | |
| "step": 75350 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 4.788781318632065e-05, | |
| "loss": 1.1678, | |
| "step": 75400 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 4.788641253193492e-05, | |
| "loss": 1.1807, | |
| "step": 75450 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 4.788501187754919e-05, | |
| "loss": 1.1632, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 4.788361122316347e-05, | |
| "loss": 1.1619, | |
| "step": 75550 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 4.788221056877773e-05, | |
| "loss": 1.1506, | |
| "step": 75600 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 4.7880809914392004e-05, | |
| "loss": 1.1797, | |
| "step": 75650 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 4.7879409260006276e-05, | |
| "loss": 1.1735, | |
| "step": 75700 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 4.787800860562055e-05, | |
| "loss": 1.1532, | |
| "step": 75750 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 4.787660795123482e-05, | |
| "loss": 1.1821, | |
| "step": 75800 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 4.787520729684909e-05, | |
| "loss": 1.1809, | |
| "step": 75850 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 4.7873806642463356e-05, | |
| "loss": 1.1986, | |
| "step": 75900 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 4.7872405988077634e-05, | |
| "loss": 1.1636, | |
| "step": 75950 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 4.7871005333691906e-05, | |
| "loss": 1.1619, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 4.786960467930617e-05, | |
| "loss": 1.1842, | |
| "step": 76050 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 4.786820402492044e-05, | |
| "loss": 1.1465, | |
| "step": 76100 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 4.786680337053472e-05, | |
| "loss": 1.2114, | |
| "step": 76150 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 4.7865402716148986e-05, | |
| "loss": 1.1696, | |
| "step": 76200 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 4.786400206176326e-05, | |
| "loss": 1.1876, | |
| "step": 76250 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 4.786260140737753e-05, | |
| "loss": 1.163, | |
| "step": 76300 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 4.78612007529918e-05, | |
| "loss": 1.1856, | |
| "step": 76350 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 4.785980009860607e-05, | |
| "loss": 1.2008, | |
| "step": 76400 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 4.7858399444220344e-05, | |
| "loss": 1.1641, | |
| "step": 76450 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 4.785699878983461e-05, | |
| "loss": 1.1701, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 4.785559813544888e-05, | |
| "loss": 1.1762, | |
| "step": 76550 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 4.785419748106316e-05, | |
| "loss": 1.1818, | |
| "step": 76600 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 4.7852796826677424e-05, | |
| "loss": 1.1875, | |
| "step": 76650 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 4.7851396172291695e-05, | |
| "loss": 1.1868, | |
| "step": 76700 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 4.784999551790597e-05, | |
| "loss": 1.1566, | |
| "step": 76750 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 4.784859486352024e-05, | |
| "loss": 1.173, | |
| "step": 76800 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 4.784719420913451e-05, | |
| "loss": 1.1841, | |
| "step": 76850 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 4.784579355474878e-05, | |
| "loss": 1.1344, | |
| "step": 76900 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 4.7844392900363047e-05, | |
| "loss": 1.183, | |
| "step": 76950 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 4.7842992245977325e-05, | |
| "loss": 1.1518, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 4.7841591591591597e-05, | |
| "loss": 1.1861, | |
| "step": 77050 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 4.784019093720586e-05, | |
| "loss": 1.1881, | |
| "step": 77100 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 4.783879028282013e-05, | |
| "loss": 1.1523, | |
| "step": 77150 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 4.783738962843441e-05, | |
| "loss": 1.1709, | |
| "step": 77200 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 4.7835988974048676e-05, | |
| "loss": 1.1667, | |
| "step": 77250 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 4.783458831966295e-05, | |
| "loss": 1.2023, | |
| "step": 77300 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 4.783318766527722e-05, | |
| "loss": 1.1518, | |
| "step": 77350 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 4.783178701089149e-05, | |
| "loss": 1.1496, | |
| "step": 77400 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 4.783038635650576e-05, | |
| "loss": 1.169, | |
| "step": 77450 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 4.7828985702120035e-05, | |
| "loss": 1.1592, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 4.78275850477343e-05, | |
| "loss": 1.1706, | |
| "step": 77550 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 4.782618439334857e-05, | |
| "loss": 1.1922, | |
| "step": 77600 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 4.782478373896285e-05, | |
| "loss": 1.1403, | |
| "step": 77650 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 4.7823383084577114e-05, | |
| "loss": 1.1817, | |
| "step": 77700 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 4.7821982430191386e-05, | |
| "loss": 1.1655, | |
| "step": 77750 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 4.782058177580566e-05, | |
| "loss": 1.1826, | |
| "step": 77800 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 4.781918112141993e-05, | |
| "loss": 1.2345, | |
| "step": 77850 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 4.78177804670342e-05, | |
| "loss": 1.1927, | |
| "step": 77900 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 4.781637981264847e-05, | |
| "loss": 1.1525, | |
| "step": 77950 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 4.781497915826274e-05, | |
| "loss": 1.1508, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 4.7813578503877016e-05, | |
| "loss": 1.2115, | |
| "step": 78050 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 4.781217784949129e-05, | |
| "loss": 1.1696, | |
| "step": 78100 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 4.781077719510555e-05, | |
| "loss": 1.1726, | |
| "step": 78150 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 4.7809376540719824e-05, | |
| "loss": 1.1721, | |
| "step": 78200 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 4.78079758863341e-05, | |
| "loss": 1.1736, | |
| "step": 78250 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 4.780657523194837e-05, | |
| "loss": 1.1784, | |
| "step": 78300 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 4.780517457756264e-05, | |
| "loss": 1.14, | |
| "step": 78350 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 4.780377392317691e-05, | |
| "loss": 1.1828, | |
| "step": 78400 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 4.780237326879118e-05, | |
| "loss": 1.1638, | |
| "step": 78450 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 4.7800972614405454e-05, | |
| "loss": 1.1667, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 4.7799571960019725e-05, | |
| "loss": 1.154, | |
| "step": 78550 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 4.779817130563399e-05, | |
| "loss": 1.1937, | |
| "step": 78600 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 4.779677065124827e-05, | |
| "loss": 1.1851, | |
| "step": 78650 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 4.779536999686254e-05, | |
| "loss": 1.1986, | |
| "step": 78700 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 4.7793969342476805e-05, | |
| "loss": 1.1953, | |
| "step": 78750 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 4.779256868809108e-05, | |
| "loss": 1.1659, | |
| "step": 78800 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 4.779116803370535e-05, | |
| "loss": 1.1662, | |
| "step": 78850 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 4.778976737931962e-05, | |
| "loss": 1.1517, | |
| "step": 78900 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 4.778836672493389e-05, | |
| "loss": 1.1697, | |
| "step": 78950 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 4.778696607054816e-05, | |
| "loss": 1.1895, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 4.778556541616243e-05, | |
| "loss": 1.1425, | |
| "step": 79050 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 4.7784164761776707e-05, | |
| "loss": 1.1622, | |
| "step": 79100 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 4.778276410739098e-05, | |
| "loss": 1.1712, | |
| "step": 79150 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 4.778136345300524e-05, | |
| "loss": 1.1632, | |
| "step": 79200 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 4.7779962798619515e-05, | |
| "loss": 1.171, | |
| "step": 79250 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 4.777856214423379e-05, | |
| "loss": 1.159, | |
| "step": 79300 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 4.777716148984806e-05, | |
| "loss": 1.1655, | |
| "step": 79350 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 4.777576083546233e-05, | |
| "loss": 1.1825, | |
| "step": 79400 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 4.77743601810766e-05, | |
| "loss": 1.2058, | |
| "step": 79450 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 4.777295952669087e-05, | |
| "loss": 1.1672, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 4.7771558872305145e-05, | |
| "loss": 1.1774, | |
| "step": 79550 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 4.7770158217919416e-05, | |
| "loss": 1.1497, | |
| "step": 79600 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 4.776875756353368e-05, | |
| "loss": 1.1707, | |
| "step": 79650 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 4.776735690914796e-05, | |
| "loss": 1.2059, | |
| "step": 79700 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 4.776595625476223e-05, | |
| "loss": 1.1527, | |
| "step": 79750 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 4.7764555600376496e-05, | |
| "loss": 1.1573, | |
| "step": 79800 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 4.776315494599077e-05, | |
| "loss": 1.1846, | |
| "step": 79850 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 4.776175429160504e-05, | |
| "loss": 1.2171, | |
| "step": 79900 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 4.776035363721931e-05, | |
| "loss": 1.1666, | |
| "step": 79950 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 4.775895298283358e-05, | |
| "loss": 1.1736, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 4.7757552328447854e-05, | |
| "loss": 1.157, | |
| "step": 80050 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 4.775615167406212e-05, | |
| "loss": 1.1828, | |
| "step": 80100 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 4.77547510196764e-05, | |
| "loss": 1.1898, | |
| "step": 80150 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 4.775335036529067e-05, | |
| "loss": 1.1761, | |
| "step": 80200 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 4.7751949710904934e-05, | |
| "loss": 1.1554, | |
| "step": 80250 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 4.7750549056519205e-05, | |
| "loss": 1.1623, | |
| "step": 80300 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 4.7749148402133484e-05, | |
| "loss": 1.1596, | |
| "step": 80350 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 4.774774774774775e-05, | |
| "loss": 1.1777, | |
| "step": 80400 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 4.774634709336202e-05, | |
| "loss": 1.1608, | |
| "step": 80450 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 4.774494643897629e-05, | |
| "loss": 1.1305, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 4.7743545784590564e-05, | |
| "loss": 1.1636, | |
| "step": 80550 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 4.7742145130204835e-05, | |
| "loss": 1.1405, | |
| "step": 80600 | |
| }, | |
| { | |
| "epoch": 3.61, | |
| "learning_rate": 4.774074447581911e-05, | |
| "loss": 1.164, | |
| "step": 80650 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 4.773934382143337e-05, | |
| "loss": 1.1696, | |
| "step": 80700 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 4.773794316704765e-05, | |
| "loss": 1.1624, | |
| "step": 80750 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 4.773654251266192e-05, | |
| "loss": 1.1748, | |
| "step": 80800 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 4.773514185827619e-05, | |
| "loss": 1.1486, | |
| "step": 80850 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 4.773374120389046e-05, | |
| "loss": 1.1813, | |
| "step": 80900 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 4.773234054950473e-05, | |
| "loss": 1.1725, | |
| "step": 80950 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 4.7730939895119e-05, | |
| "loss": 1.1872, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 4.772953924073327e-05, | |
| "loss": 1.1534, | |
| "step": 81050 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 4.7728138586347545e-05, | |
| "loss": 1.1725, | |
| "step": 81100 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 4.772673793196181e-05, | |
| "loss": 1.1668, | |
| "step": 81150 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 4.772533727757609e-05, | |
| "loss": 1.1781, | |
| "step": 81200 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 4.772393662319036e-05, | |
| "loss": 1.1684, | |
| "step": 81250 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 4.7722535968804625e-05, | |
| "loss": 1.1879, | |
| "step": 81300 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 4.7721135314418896e-05, | |
| "loss": 1.1728, | |
| "step": 81350 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 4.7719734660033175e-05, | |
| "loss": 1.1831, | |
| "step": 81400 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 4.771833400564744e-05, | |
| "loss": 1.1714, | |
| "step": 81450 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 4.771693335126171e-05, | |
| "loss": 1.1622, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 4.771553269687598e-05, | |
| "loss": 1.1586, | |
| "step": 81550 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 4.7714132042490254e-05, | |
| "loss": 1.1451, | |
| "step": 81600 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 4.7712731388104526e-05, | |
| "loss": 1.1707, | |
| "step": 81650 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 4.77113307337188e-05, | |
| "loss": 1.1656, | |
| "step": 81700 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 4.770993007933306e-05, | |
| "loss": 1.1915, | |
| "step": 81750 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 4.770852942494734e-05, | |
| "loss": 1.1767, | |
| "step": 81800 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 4.770712877056161e-05, | |
| "loss": 1.1862, | |
| "step": 81850 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 4.770572811617588e-05, | |
| "loss": 1.1925, | |
| "step": 81900 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 4.770432746179015e-05, | |
| "loss": 1.1957, | |
| "step": 81950 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 4.770292680740442e-05, | |
| "loss": 1.1652, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 4.770152615301869e-05, | |
| "loss": 1.1723, | |
| "step": 82050 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 4.7700125498632964e-05, | |
| "loss": 1.183, | |
| "step": 82100 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 4.7698724844247236e-05, | |
| "loss": 1.1691, | |
| "step": 82150 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 4.76973241898615e-05, | |
| "loss": 1.1415, | |
| "step": 82200 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 4.769592353547578e-05, | |
| "loss": 1.1928, | |
| "step": 82250 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 4.769452288109005e-05, | |
| "loss": 1.1779, | |
| "step": 82300 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 4.7693122226704315e-05, | |
| "loss": 1.155, | |
| "step": 82350 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 4.769172157231859e-05, | |
| "loss": 1.1755, | |
| "step": 82400 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 4.7690320917932865e-05, | |
| "loss": 1.1804, | |
| "step": 82450 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 4.768892026354713e-05, | |
| "loss": 1.1769, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 4.76875196091614e-05, | |
| "loss": 1.1553, | |
| "step": 82550 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 4.7686118954775674e-05, | |
| "loss": 1.1611, | |
| "step": 82600 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 4.7684718300389945e-05, | |
| "loss": 1.1786, | |
| "step": 82650 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 4.768331764600422e-05, | |
| "loss": 1.1768, | |
| "step": 82700 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 4.768191699161849e-05, | |
| "loss": 1.1694, | |
| "step": 82750 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 4.768051633723275e-05, | |
| "loss": 1.214, | |
| "step": 82800 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 4.767911568284703e-05, | |
| "loss": 1.1711, | |
| "step": 82850 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 4.7677715028461303e-05, | |
| "loss": 1.1475, | |
| "step": 82900 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 4.767631437407557e-05, | |
| "loss": 1.1716, | |
| "step": 82950 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 4.767491371968984e-05, | |
| "loss": 1.184, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 4.767351306530412e-05, | |
| "loss": 1.1876, | |
| "step": 83050 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 4.767211241091838e-05, | |
| "loss": 1.2014, | |
| "step": 83100 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 4.7670711756532655e-05, | |
| "loss": 1.1782, | |
| "step": 83150 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 4.7669311102146926e-05, | |
| "loss": 1.1751, | |
| "step": 83200 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 4.766791044776119e-05, | |
| "loss": 1.1593, | |
| "step": 83250 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 4.766650979337547e-05, | |
| "loss": 1.1758, | |
| "step": 83300 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 4.766510913898974e-05, | |
| "loss": 1.157, | |
| "step": 83350 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 4.7663708484604006e-05, | |
| "loss": 1.1296, | |
| "step": 83400 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 4.766230783021828e-05, | |
| "loss": 1.1392, | |
| "step": 83450 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 4.7660907175832556e-05, | |
| "loss": 1.145, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 4.765950652144682e-05, | |
| "loss": 1.1592, | |
| "step": 83550 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 4.765810586706109e-05, | |
| "loss": 1.1619, | |
| "step": 83600 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 4.7656705212675364e-05, | |
| "loss": 1.1685, | |
| "step": 83650 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 4.7655304558289636e-05, | |
| "loss": 1.1478, | |
| "step": 83700 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 4.765390390390391e-05, | |
| "loss": 1.1554, | |
| "step": 83750 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 4.765250324951818e-05, | |
| "loss": 1.1587, | |
| "step": 83800 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 4.7651102595132444e-05, | |
| "loss": 1.1696, | |
| "step": 83850 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 4.764970194074672e-05, | |
| "loss": 1.1849, | |
| "step": 83900 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 4.7648301286360994e-05, | |
| "loss": 1.1463, | |
| "step": 83950 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 4.764690063197526e-05, | |
| "loss": 1.1289, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 4.764549997758953e-05, | |
| "loss": 1.1503, | |
| "step": 84050 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 4.764409932320381e-05, | |
| "loss": 1.1848, | |
| "step": 84100 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 4.7642698668818074e-05, | |
| "loss": 1.167, | |
| "step": 84150 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 4.7641298014432346e-05, | |
| "loss": 1.1467, | |
| "step": 84200 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 4.763989736004662e-05, | |
| "loss": 1.1529, | |
| "step": 84250 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 4.763849670566088e-05, | |
| "loss": 1.1826, | |
| "step": 84300 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 4.763709605127516e-05, | |
| "loss": 1.1499, | |
| "step": 84350 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 4.763569539688943e-05, | |
| "loss": 1.18, | |
| "step": 84400 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 4.76342947425037e-05, | |
| "loss": 1.1758, | |
| "step": 84450 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 4.763289408811797e-05, | |
| "loss": 1.1627, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 4.763149343373225e-05, | |
| "loss": 1.1634, | |
| "step": 84550 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 4.763009277934651e-05, | |
| "loss": 1.1744, | |
| "step": 84600 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 4.7628692124960784e-05, | |
| "loss": 1.1388, | |
| "step": 84650 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 4.7627291470575055e-05, | |
| "loss": 1.1546, | |
| "step": 84700 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 4.762589081618933e-05, | |
| "loss": 1.1508, | |
| "step": 84750 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 4.76244901618036e-05, | |
| "loss": 1.1785, | |
| "step": 84800 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 4.762308950741787e-05, | |
| "loss": 1.184, | |
| "step": 84850 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 4.7621688853032135e-05, | |
| "loss": 1.1683, | |
| "step": 84900 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 4.762028819864641e-05, | |
| "loss": 1.155, | |
| "step": 84950 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 4.7618887544260685e-05, | |
| "loss": 1.1493, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 4.761748688987495e-05, | |
| "loss": 1.1736, | |
| "step": 85050 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 4.761608623548922e-05, | |
| "loss": 1.1701, | |
| "step": 85100 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 4.761468558110349e-05, | |
| "loss": 1.1473, | |
| "step": 85150 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 4.7613284926717765e-05, | |
| "loss": 1.1543, | |
| "step": 85200 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 4.7611884272332036e-05, | |
| "loss": 1.16, | |
| "step": 85250 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 4.761048361794631e-05, | |
| "loss": 1.1549, | |
| "step": 85300 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 4.760908296356057e-05, | |
| "loss": 1.1629, | |
| "step": 85350 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 4.760768230917485e-05, | |
| "loss": 1.1526, | |
| "step": 85400 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 4.760628165478912e-05, | |
| "loss": 1.1535, | |
| "step": 85450 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 4.760488100040339e-05, | |
| "loss": 1.1512, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 3.83, | |
| "learning_rate": 4.760348034601766e-05, | |
| "loss": 1.1895, | |
| "step": 85550 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 4.760207969163194e-05, | |
| "loss": 1.1491, | |
| "step": 85600 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 4.76006790372462e-05, | |
| "loss": 1.1818, | |
| "step": 85650 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 4.7599278382860474e-05, | |
| "loss": 1.1703, | |
| "step": 85700 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 4.7597877728474746e-05, | |
| "loss": 1.1637, | |
| "step": 85750 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 4.759647707408902e-05, | |
| "loss": 1.1561, | |
| "step": 85800 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 4.759507641970329e-05, | |
| "loss": 1.179, | |
| "step": 85850 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 4.759367576531756e-05, | |
| "loss": 1.1416, | |
| "step": 85900 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 4.7592275110931826e-05, | |
| "loss": 1.1354, | |
| "step": 85950 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 4.7590874456546104e-05, | |
| "loss": 1.1555, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 4.7589473802160376e-05, | |
| "loss": 1.1897, | |
| "step": 86050 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 4.758807314777464e-05, | |
| "loss": 1.159, | |
| "step": 86100 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 4.758667249338891e-05, | |
| "loss": 1.1614, | |
| "step": 86150 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 4.7585271839003184e-05, | |
| "loss": 1.177, | |
| "step": 86200 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 4.7583871184617456e-05, | |
| "loss": 1.1443, | |
| "step": 86250 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 4.758247053023173e-05, | |
| "loss": 1.1603, | |
| "step": 86300 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 4.7581069875846e-05, | |
| "loss": 1.198, | |
| "step": 86350 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 4.757966922146027e-05, | |
| "loss": 1.1655, | |
| "step": 86400 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 4.757826856707454e-05, | |
| "loss": 1.2059, | |
| "step": 86450 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 4.7576867912688814e-05, | |
| "loss": 1.1499, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 4.757546725830308e-05, | |
| "loss": 1.159, | |
| "step": 86550 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 4.757406660391735e-05, | |
| "loss": 1.1589, | |
| "step": 86600 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 4.757266594953162e-05, | |
| "loss": 1.1699, | |
| "step": 86650 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 4.7571265295145894e-05, | |
| "loss": 1.1401, | |
| "step": 86700 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 4.7569864640760165e-05, | |
| "loss": 1.1742, | |
| "step": 86750 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 4.756846398637444e-05, | |
| "loss": 1.1591, | |
| "step": 86800 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 4.756706333198871e-05, | |
| "loss": 1.1519, | |
| "step": 86850 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 4.756566267760298e-05, | |
| "loss": 1.1516, | |
| "step": 86900 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 4.756426202321725e-05, | |
| "loss": 1.1428, | |
| "step": 86950 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 4.7562861368831517e-05, | |
| "loss": 1.1925, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 4.7561460714445795e-05, | |
| "loss": 1.1404, | |
| "step": 87050 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 4.756006006006006e-05, | |
| "loss": 1.1496, | |
| "step": 87100 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 4.755865940567433e-05, | |
| "loss": 1.1784, | |
| "step": 87150 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 4.75572587512886e-05, | |
| "loss": 1.1627, | |
| "step": 87200 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 4.7555858096902875e-05, | |
| "loss": 1.1426, | |
| "step": 87250 | |
| }, | |
| { | |
| "epoch": 3.91, | |
| "learning_rate": 4.7554457442517146e-05, | |
| "loss": 1.1499, | |
| "step": 87300 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 4.755305678813142e-05, | |
| "loss": 1.1624, | |
| "step": 87350 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 4.755165613374569e-05, | |
| "loss": 1.1788, | |
| "step": 87400 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 4.755025547935996e-05, | |
| "loss": 1.1411, | |
| "step": 87450 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 4.754885482497423e-05, | |
| "loss": 1.1643, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 4.7547454170588505e-05, | |
| "loss": 1.1557, | |
| "step": 87550 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 4.754605351620277e-05, | |
| "loss": 1.1751, | |
| "step": 87600 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 4.754465286181704e-05, | |
| "loss": 1.1482, | |
| "step": 87650 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 4.754325220743131e-05, | |
| "loss": 1.1461, | |
| "step": 87700 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 4.7541851553045584e-05, | |
| "loss": 1.175, | |
| "step": 87750 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 4.7540450898659856e-05, | |
| "loss": 1.1366, | |
| "step": 87800 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 4.753905024427413e-05, | |
| "loss": 1.1696, | |
| "step": 87850 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 4.75376495898884e-05, | |
| "loss": 1.1603, | |
| "step": 87900 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 4.753624893550267e-05, | |
| "loss": 1.1581, | |
| "step": 87950 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 4.753484828111694e-05, | |
| "loss": 1.1782, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 4.753344762673121e-05, | |
| "loss": 1.1512, | |
| "step": 88050 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 4.7532046972345486e-05, | |
| "loss": 1.1958, | |
| "step": 88100 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 4.753064631795975e-05, | |
| "loss": 1.1588, | |
| "step": 88150 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 4.752924566357402e-05, | |
| "loss": 1.1503, | |
| "step": 88200 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 4.7527845009188294e-05, | |
| "loss": 1.1552, | |
| "step": 88250 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 4.7526444354802566e-05, | |
| "loss": 1.178, | |
| "step": 88300 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 4.752504370041684e-05, | |
| "loss": 1.1439, | |
| "step": 88350 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 4.752364304603111e-05, | |
| "loss": 1.1596, | |
| "step": 88400 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 4.752224239164538e-05, | |
| "loss": 1.1738, | |
| "step": 88450 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 4.752084173725965e-05, | |
| "loss": 1.1629, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 4.7519441082873924e-05, | |
| "loss": 1.1586, | |
| "step": 88550 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 4.751804042848819e-05, | |
| "loss": 1.1825, | |
| "step": 88600 | |
| }, | |
| { | |
| "epoch": 3.97, | |
| "learning_rate": 4.751663977410246e-05, | |
| "loss": 1.1345, | |
| "step": 88650 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 4.751523911971673e-05, | |
| "loss": 1.1739, | |
| "step": 88700 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 4.7513838465331003e-05, | |
| "loss": 1.1766, | |
| "step": 88750 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 4.7512437810945275e-05, | |
| "loss": 1.1577, | |
| "step": 88800 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 4.751103715655955e-05, | |
| "loss": 1.1732, | |
| "step": 88850 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 4.750963650217382e-05, | |
| "loss": 1.1485, | |
| "step": 88900 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 4.750823584778809e-05, | |
| "loss": 1.1586, | |
| "step": 88950 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 4.750683519340236e-05, | |
| "loss": 1.1265, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 4.750543453901663e-05, | |
| "loss": 1.1527, | |
| "step": 89050 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 4.75040338846309e-05, | |
| "loss": 1.1542, | |
| "step": 89100 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 4.7502633230245177e-05, | |
| "loss": 1.1522, | |
| "step": 89150 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 4.750123257585944e-05, | |
| "loss": 1.1272, | |
| "step": 89200 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 4.749983192147371e-05, | |
| "loss": 1.1496, | |
| "step": 89250 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 4.7498431267087985e-05, | |
| "loss": 1.1714, | |
| "step": 89300 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 4.7497030612702256e-05, | |
| "loss": 1.1457, | |
| "step": 89350 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 4.749562995831653e-05, | |
| "loss": 1.149, | |
| "step": 89400 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 4.74942293039308e-05, | |
| "loss": 1.1828, | |
| "step": 89450 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 4.749282864954507e-05, | |
| "loss": 1.1107, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 4.749142799515934e-05, | |
| "loss": 1.1306, | |
| "step": 89550 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 4.7490027340773614e-05, | |
| "loss": 1.1476, | |
| "step": 89600 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 4.748862668638788e-05, | |
| "loss": 1.1422, | |
| "step": 89650 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 4.748722603200215e-05, | |
| "loss": 1.1734, | |
| "step": 89700 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 4.748582537761642e-05, | |
| "loss": 1.1281, | |
| "step": 89750 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 4.7484424723230694e-05, | |
| "loss": 1.1658, | |
| "step": 89800 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 4.7483024068844966e-05, | |
| "loss": 1.1493, | |
| "step": 89850 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 4.748162341445924e-05, | |
| "loss": 1.1633, | |
| "step": 89900 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 4.748022276007351e-05, | |
| "loss": 1.1526, | |
| "step": 89950 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 4.747882210568778e-05, | |
| "loss": 1.1744, | |
| "step": 90000 | |
| } | |
| ], | |
| "max_steps": 1784880, | |
| "num_train_epochs": 80, | |
| "total_flos": 1.1156563427328e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |