dialect-debiasing-gpt2-medium-translated-pnlogmse-e1-r50_eval-n10.0
/
checkpoint-3125
/trainer_state.json
| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 3125, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0032, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 6.369426751592357e-06, | |
| "loss": 22.4391, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0064, | |
| "grad_norm": 10.0, | |
| "learning_rate": 1.2738853503184714e-05, | |
| "loss": 18.5401, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0096, | |
| "grad_norm": 10.0, | |
| "learning_rate": 1.910828025477707e-05, | |
| "loss": 18.8722, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.0128, | |
| "grad_norm": 10.0, | |
| "learning_rate": 2.5477707006369428e-05, | |
| "loss": 17.9956, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.016, | |
| "grad_norm": 10.000000953674316, | |
| "learning_rate": 3.184713375796178e-05, | |
| "loss": 15.0615, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.0192, | |
| "grad_norm": 9.999998092651367, | |
| "learning_rate": 3.821656050955414e-05, | |
| "loss": 14.3948, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.0224, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 4.45859872611465e-05, | |
| "loss": 12.1943, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.0256, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 5.0955414012738855e-05, | |
| "loss": 10.114, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.0288, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 5.732484076433121e-05, | |
| "loss": 8.9744, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.032, | |
| "grad_norm": 10.0, | |
| "learning_rate": 6.369426751592356e-05, | |
| "loss": 8.1171, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0352, | |
| "grad_norm": 10.0, | |
| "learning_rate": 7.006369426751592e-05, | |
| "loss": 6.1562, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.0384, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 7.643312101910829e-05, | |
| "loss": 6.1523, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.0416, | |
| "grad_norm": 10.0, | |
| "learning_rate": 8.280254777070065e-05, | |
| "loss": 5.1105, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.0448, | |
| "grad_norm": 10.0, | |
| "learning_rate": 8.9171974522293e-05, | |
| "loss": 4.6447, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.048, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 9.554140127388536e-05, | |
| "loss": 4.441, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.0512, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 9.989892183288411e-05, | |
| "loss": 3.9956, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.0544, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 9.956199460916442e-05, | |
| "loss": 3.6844, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.0576, | |
| "grad_norm": 9.74394702911377, | |
| "learning_rate": 9.922506738544474e-05, | |
| "loss": 3.6594, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.0608, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 9.888814016172507e-05, | |
| "loss": 3.5963, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.064, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 9.85512129380054e-05, | |
| "loss": 3.0941, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.0672, | |
| "grad_norm": 10.0, | |
| "learning_rate": 9.821428571428572e-05, | |
| "loss": 3.0641, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.0704, | |
| "grad_norm": 10.0, | |
| "learning_rate": 9.787735849056603e-05, | |
| "loss": 2.8093, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.0736, | |
| "grad_norm": 9.999998092651367, | |
| "learning_rate": 9.754043126684636e-05, | |
| "loss": 2.9413, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.0768, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 9.720350404312669e-05, | |
| "loss": 2.7816, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 7.968898773193359, | |
| "learning_rate": 9.686657681940702e-05, | |
| "loss": 2.6371, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.0832, | |
| "grad_norm": 9.999999046325684, | |
| "learning_rate": 9.652964959568734e-05, | |
| "loss": 2.5965, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.0864, | |
| "grad_norm": 9.171067237854004, | |
| "learning_rate": 9.619272237196765e-05, | |
| "loss": 2.7478, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.0896, | |
| "grad_norm": 9.04393196105957, | |
| "learning_rate": 9.585579514824798e-05, | |
| "loss": 2.4879, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.0928, | |
| "grad_norm": 8.444461822509766, | |
| "learning_rate": 9.551886792452831e-05, | |
| "loss": 2.2215, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.096, | |
| "grad_norm": 8.200078964233398, | |
| "learning_rate": 9.518194070080863e-05, | |
| "loss": 2.6473, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.0992, | |
| "grad_norm": 9.681734085083008, | |
| "learning_rate": 9.484501347708896e-05, | |
| "loss": 2.3853, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.1024, | |
| "grad_norm": 9.111586570739746, | |
| "learning_rate": 9.450808625336927e-05, | |
| "loss": 2.2498, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.1056, | |
| "grad_norm": 9.999998092651367, | |
| "learning_rate": 9.41711590296496e-05, | |
| "loss": 2.0398, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.1088, | |
| "grad_norm": 9.552145957946777, | |
| "learning_rate": 9.383423180592993e-05, | |
| "loss": 1.9816, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.112, | |
| "grad_norm": 8.0576171875, | |
| "learning_rate": 9.349730458221025e-05, | |
| "loss": 2.0454, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.1152, | |
| "grad_norm": 6.90339994430542, | |
| "learning_rate": 9.316037735849057e-05, | |
| "loss": 1.8505, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.1184, | |
| "grad_norm": 8.755167007446289, | |
| "learning_rate": 9.282345013477089e-05, | |
| "loss": 1.8279, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.1216, | |
| "grad_norm": 8.947823524475098, | |
| "learning_rate": 9.248652291105122e-05, | |
| "loss": 1.8657, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.1248, | |
| "grad_norm": 9.317538261413574, | |
| "learning_rate": 9.214959568733154e-05, | |
| "loss": 1.6115, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.128, | |
| "grad_norm": 9.361852645874023, | |
| "learning_rate": 9.181266846361186e-05, | |
| "loss": 1.6446, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.1312, | |
| "grad_norm": 7.044137001037598, | |
| "learning_rate": 9.14757412398922e-05, | |
| "loss": 1.5443, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.1344, | |
| "grad_norm": 6.381120681762695, | |
| "learning_rate": 9.113881401617251e-05, | |
| "loss": 1.4368, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.1376, | |
| "grad_norm": 7.995136260986328, | |
| "learning_rate": 9.080188679245284e-05, | |
| "loss": 1.4774, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.1408, | |
| "grad_norm": 6.258317470550537, | |
| "learning_rate": 9.046495956873315e-05, | |
| "loss": 1.3529, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.144, | |
| "grad_norm": 5.482784748077393, | |
| "learning_rate": 9.012803234501348e-05, | |
| "loss": 1.3406, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.1472, | |
| "grad_norm": 5.9611921310424805, | |
| "learning_rate": 8.979110512129381e-05, | |
| "loss": 1.3311, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.1504, | |
| "grad_norm": 5.832077503204346, | |
| "learning_rate": 8.945417789757413e-05, | |
| "loss": 1.2302, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.1536, | |
| "grad_norm": 5.394923686981201, | |
| "learning_rate": 8.911725067385444e-05, | |
| "loss": 1.1862, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.1568, | |
| "grad_norm": 4.73380708694458, | |
| "learning_rate": 8.878032345013477e-05, | |
| "loss": 1.1394, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 5.2955427169799805, | |
| "learning_rate": 8.84433962264151e-05, | |
| "loss": 1.2688, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.1632, | |
| "grad_norm": 6.693230628967285, | |
| "learning_rate": 8.810646900269543e-05, | |
| "loss": 1.1342, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.1664, | |
| "grad_norm": 6.776840686798096, | |
| "learning_rate": 8.776954177897575e-05, | |
| "loss": 1.1557, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.1696, | |
| "grad_norm": 5.950035572052002, | |
| "learning_rate": 8.743261455525606e-05, | |
| "loss": 1.0529, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.1728, | |
| "grad_norm": 6.108712673187256, | |
| "learning_rate": 8.709568733153639e-05, | |
| "loss": 1.1195, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.176, | |
| "grad_norm": 7.9674787521362305, | |
| "learning_rate": 8.675876010781672e-05, | |
| "loss": 1.1074, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.1792, | |
| "grad_norm": 5.42255163192749, | |
| "learning_rate": 8.642183288409704e-05, | |
| "loss": 1.0794, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.1824, | |
| "grad_norm": 8.32884693145752, | |
| "learning_rate": 8.608490566037735e-05, | |
| "loss": 1.0796, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.1856, | |
| "grad_norm": 5.415431022644043, | |
| "learning_rate": 8.574797843665768e-05, | |
| "loss": 1.0595, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.1888, | |
| "grad_norm": 5.638887405395508, | |
| "learning_rate": 8.541105121293801e-05, | |
| "loss": 1.0503, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.192, | |
| "grad_norm": 6.9637908935546875, | |
| "learning_rate": 8.507412398921834e-05, | |
| "loss": 1.037, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.1952, | |
| "grad_norm": 4.847858905792236, | |
| "learning_rate": 8.473719676549866e-05, | |
| "loss": 0.915, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.1984, | |
| "grad_norm": 4.449883460998535, | |
| "learning_rate": 8.440026954177897e-05, | |
| "loss": 0.9626, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.2016, | |
| "grad_norm": 6.299098968505859, | |
| "learning_rate": 8.40633423180593e-05, | |
| "loss": 0.9122, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.2048, | |
| "grad_norm": 9.391380310058594, | |
| "learning_rate": 8.372641509433963e-05, | |
| "loss": 0.9696, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.208, | |
| "grad_norm": 7.544974327087402, | |
| "learning_rate": 8.338948787061996e-05, | |
| "loss": 0.8972, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.2112, | |
| "grad_norm": 5.396461009979248, | |
| "learning_rate": 8.305256064690027e-05, | |
| "loss": 0.9055, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.2144, | |
| "grad_norm": 6.584786891937256, | |
| "learning_rate": 8.271563342318059e-05, | |
| "loss": 0.9223, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.2176, | |
| "grad_norm": 5.7641072273254395, | |
| "learning_rate": 8.237870619946092e-05, | |
| "loss": 0.8602, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.2208, | |
| "grad_norm": 5.566444396972656, | |
| "learning_rate": 8.204177897574125e-05, | |
| "loss": 0.8629, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.224, | |
| "grad_norm": 5.248049736022949, | |
| "learning_rate": 8.170485175202158e-05, | |
| "loss": 0.8402, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.2272, | |
| "grad_norm": 4.847179889678955, | |
| "learning_rate": 8.136792452830189e-05, | |
| "loss": 0.8218, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.2304, | |
| "grad_norm": 5.082803726196289, | |
| "learning_rate": 8.103099730458221e-05, | |
| "loss": 0.8042, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.2336, | |
| "grad_norm": 3.536363124847412, | |
| "learning_rate": 8.069407008086254e-05, | |
| "loss": 0.8003, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.2368, | |
| "grad_norm": 7.2759857177734375, | |
| "learning_rate": 8.035714285714287e-05, | |
| "loss": 0.8405, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 5.047908306121826, | |
| "learning_rate": 8.002021563342318e-05, | |
| "loss": 0.7769, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.2432, | |
| "grad_norm": 4.044074535369873, | |
| "learning_rate": 7.968328840970351e-05, | |
| "loss": 0.7445, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.2464, | |
| "grad_norm": 4.787796497344971, | |
| "learning_rate": 7.934636118598383e-05, | |
| "loss": 0.7424, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.2496, | |
| "grad_norm": 5.314538955688477, | |
| "learning_rate": 7.900943396226416e-05, | |
| "loss": 0.7705, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.2528, | |
| "grad_norm": 7.510671138763428, | |
| "learning_rate": 7.867250673854449e-05, | |
| "loss": 0.8262, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.256, | |
| "grad_norm": 6.341707706451416, | |
| "learning_rate": 7.83355795148248e-05, | |
| "loss": 0.8339, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.2592, | |
| "grad_norm": 5.165682792663574, | |
| "learning_rate": 7.799865229110512e-05, | |
| "loss": 0.6955, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.2624, | |
| "grad_norm": 6.235734939575195, | |
| "learning_rate": 7.766172506738545e-05, | |
| "loss": 0.7221, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.2656, | |
| "grad_norm": 6.6283488273620605, | |
| "learning_rate": 7.732479784366577e-05, | |
| "loss": 0.6867, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.2688, | |
| "grad_norm": 8.073585510253906, | |
| "learning_rate": 7.69878706199461e-05, | |
| "loss": 0.733, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.272, | |
| "grad_norm": 8.499881744384766, | |
| "learning_rate": 7.665094339622642e-05, | |
| "loss": 0.7132, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.2752, | |
| "grad_norm": 6.54105806350708, | |
| "learning_rate": 7.631401617250674e-05, | |
| "loss": 0.7183, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.2784, | |
| "grad_norm": 5.989783763885498, | |
| "learning_rate": 7.597708894878706e-05, | |
| "loss": 0.6529, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.2816, | |
| "grad_norm": 4.016666412353516, | |
| "learning_rate": 7.56401617250674e-05, | |
| "loss": 0.6531, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.2848, | |
| "grad_norm": 6.746622085571289, | |
| "learning_rate": 7.530323450134771e-05, | |
| "loss": 0.6357, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.288, | |
| "grad_norm": 5.905118942260742, | |
| "learning_rate": 7.496630727762804e-05, | |
| "loss": 0.6638, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.2912, | |
| "grad_norm": 5.090343952178955, | |
| "learning_rate": 7.462938005390835e-05, | |
| "loss": 0.6784, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.2944, | |
| "grad_norm": 6.740263938903809, | |
| "learning_rate": 7.429245283018868e-05, | |
| "loss": 0.635, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.2976, | |
| "grad_norm": 6.317070484161377, | |
| "learning_rate": 7.395552560646901e-05, | |
| "loss": 0.6421, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.3008, | |
| "grad_norm": 6.079349994659424, | |
| "learning_rate": 7.361859838274933e-05, | |
| "loss": 0.5963, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.304, | |
| "grad_norm": 4.388522624969482, | |
| "learning_rate": 7.328167115902966e-05, | |
| "loss": 0.6379, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.3072, | |
| "grad_norm": 3.835901975631714, | |
| "learning_rate": 7.294474393530997e-05, | |
| "loss": 0.5692, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.3104, | |
| "grad_norm": 3.3163001537323, | |
| "learning_rate": 7.26078167115903e-05, | |
| "loss": 0.6069, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.3136, | |
| "grad_norm": 5.156872749328613, | |
| "learning_rate": 7.227088948787062e-05, | |
| "loss": 0.5933, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.3168, | |
| "grad_norm": 3.8711464405059814, | |
| "learning_rate": 7.193396226415095e-05, | |
| "loss": 0.544, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 6.842670917510986, | |
| "learning_rate": 7.159703504043128e-05, | |
| "loss": 0.6124, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.3232, | |
| "grad_norm": 5.548664569854736, | |
| "learning_rate": 7.126010781671159e-05, | |
| "loss": 0.5312, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.3264, | |
| "grad_norm": 6.509451389312744, | |
| "learning_rate": 7.092318059299192e-05, | |
| "loss": 0.6036, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.3296, | |
| "grad_norm": 3.848813533782959, | |
| "learning_rate": 7.058625336927224e-05, | |
| "loss": 0.5867, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.3328, | |
| "grad_norm": 4.100753307342529, | |
| "learning_rate": 7.024932614555257e-05, | |
| "loss": 0.5403, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.336, | |
| "grad_norm": 4.498255252838135, | |
| "learning_rate": 6.99123989218329e-05, | |
| "loss": 0.5903, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.3392, | |
| "grad_norm": 5.71539831161499, | |
| "learning_rate": 6.957547169811321e-05, | |
| "loss": 0.5171, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.3424, | |
| "grad_norm": 6.740539073944092, | |
| "learning_rate": 6.923854447439353e-05, | |
| "loss": 0.5315, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.3456, | |
| "grad_norm": 5.979411602020264, | |
| "learning_rate": 6.890161725067386e-05, | |
| "loss": 0.6017, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.3488, | |
| "grad_norm": 5.852504253387451, | |
| "learning_rate": 6.856469002695418e-05, | |
| "loss": 0.5629, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.352, | |
| "grad_norm": 3.0850918292999268, | |
| "learning_rate": 6.822776280323451e-05, | |
| "loss": 0.5144, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.3552, | |
| "grad_norm": 4.634555339813232, | |
| "learning_rate": 6.789083557951483e-05, | |
| "loss": 0.4906, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.3584, | |
| "grad_norm": 6.15895414352417, | |
| "learning_rate": 6.755390835579514e-05, | |
| "loss": 0.5649, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.3616, | |
| "grad_norm": 4.997654914855957, | |
| "learning_rate": 6.721698113207547e-05, | |
| "loss": 0.4627, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.3648, | |
| "grad_norm": 4.061686038970947, | |
| "learning_rate": 6.68800539083558e-05, | |
| "loss": 0.4985, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.368, | |
| "grad_norm": 4.477031707763672, | |
| "learning_rate": 6.654312668463612e-05, | |
| "loss": 0.4774, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.3712, | |
| "grad_norm": 5.712886810302734, | |
| "learning_rate": 6.620619946091643e-05, | |
| "loss": 0.4635, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.3744, | |
| "grad_norm": 7.275344371795654, | |
| "learning_rate": 6.586927223719676e-05, | |
| "loss": 0.4568, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.3776, | |
| "grad_norm": 4.3618364334106445, | |
| "learning_rate": 6.553234501347709e-05, | |
| "loss": 0.4543, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.3808, | |
| "grad_norm": 6.843659400939941, | |
| "learning_rate": 6.519541778975742e-05, | |
| "loss": 0.4615, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.384, | |
| "grad_norm": 4.021126747131348, | |
| "learning_rate": 6.485849056603774e-05, | |
| "loss": 0.4053, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.3872, | |
| "grad_norm": 4.5041069984436035, | |
| "learning_rate": 6.452156334231805e-05, | |
| "loss": 0.4241, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.3904, | |
| "grad_norm": 4.459962844848633, | |
| "learning_rate": 6.418463611859838e-05, | |
| "loss": 0.4412, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.3936, | |
| "grad_norm": 3.881894588470459, | |
| "learning_rate": 6.384770889487871e-05, | |
| "loss": 0.4122, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.3968, | |
| "grad_norm": 4.284637451171875, | |
| "learning_rate": 6.351078167115904e-05, | |
| "loss": 0.4358, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 4.454357624053955, | |
| "learning_rate": 6.317385444743936e-05, | |
| "loss": 0.407, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.4032, | |
| "grad_norm": 3.3918070793151855, | |
| "learning_rate": 6.283692722371967e-05, | |
| "loss": 0.3787, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.4064, | |
| "grad_norm": 4.96063756942749, | |
| "learning_rate": 6.25e-05, | |
| "loss": 0.4132, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.4096, | |
| "grad_norm": 5.919780254364014, | |
| "learning_rate": 6.216307277628033e-05, | |
| "loss": 0.409, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.4128, | |
| "grad_norm": 5.709549427032471, | |
| "learning_rate": 6.182614555256066e-05, | |
| "loss": 0.4547, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.416, | |
| "grad_norm": 4.5489349365234375, | |
| "learning_rate": 6.148921832884098e-05, | |
| "loss": 0.4131, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.4192, | |
| "grad_norm": 3.3304429054260254, | |
| "learning_rate": 6.115229110512129e-05, | |
| "loss": 0.4027, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.4224, | |
| "grad_norm": 2.7704129219055176, | |
| "learning_rate": 6.081536388140162e-05, | |
| "loss": 0.3894, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.4256, | |
| "grad_norm": 6.625534534454346, | |
| "learning_rate": 6.047843665768195e-05, | |
| "loss": 0.3766, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.4288, | |
| "grad_norm": 4.533291339874268, | |
| "learning_rate": 6.0141509433962265e-05, | |
| "loss": 0.4129, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.432, | |
| "grad_norm": 6.264132976531982, | |
| "learning_rate": 5.980458221024259e-05, | |
| "loss": 0.366, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.4352, | |
| "grad_norm": 4.117740631103516, | |
| "learning_rate": 5.9467654986522916e-05, | |
| "loss": 0.3728, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.4384, | |
| "grad_norm": 3.296051025390625, | |
| "learning_rate": 5.913072776280324e-05, | |
| "loss": 0.3514, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.4416, | |
| "grad_norm": 4.544425964355469, | |
| "learning_rate": 5.879380053908357e-05, | |
| "loss": 0.3392, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.4448, | |
| "grad_norm": 6.508903980255127, | |
| "learning_rate": 5.8456873315363884e-05, | |
| "loss": 0.3427, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.448, | |
| "grad_norm": 3.8851804733276367, | |
| "learning_rate": 5.8119946091644206e-05, | |
| "loss": 0.3154, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.4512, | |
| "grad_norm": 4.247873783111572, | |
| "learning_rate": 5.7783018867924535e-05, | |
| "loss": 0.3421, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.4544, | |
| "grad_norm": 5.067176818847656, | |
| "learning_rate": 5.744609164420486e-05, | |
| "loss": 0.3086, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.4576, | |
| "grad_norm": 3.9404332637786865, | |
| "learning_rate": 5.710916442048517e-05, | |
| "loss": 0.3494, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.4608, | |
| "grad_norm": 4.0142822265625, | |
| "learning_rate": 5.6772237196765496e-05, | |
| "loss": 0.3144, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.464, | |
| "grad_norm": 2.31613826751709, | |
| "learning_rate": 5.6435309973045825e-05, | |
| "loss": 0.3061, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.4672, | |
| "grad_norm": 3.2096686363220215, | |
| "learning_rate": 5.609838274932615e-05, | |
| "loss": 0.2921, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.4704, | |
| "grad_norm": 5.634882926940918, | |
| "learning_rate": 5.5761455525606476e-05, | |
| "loss": 0.3227, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.4736, | |
| "grad_norm": 4.442248344421387, | |
| "learning_rate": 5.542452830188679e-05, | |
| "loss": 0.3521, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.4768, | |
| "grad_norm": 3.5769202709198, | |
| "learning_rate": 5.5087601078167114e-05, | |
| "loss": 0.3034, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 4.759718894958496, | |
| "learning_rate": 5.4750673854447444e-05, | |
| "loss": 0.2806, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.4832, | |
| "grad_norm": 5.413734436035156, | |
| "learning_rate": 5.4413746630727766e-05, | |
| "loss": 0.2903, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.4864, | |
| "grad_norm": 5.242001533508301, | |
| "learning_rate": 5.407681940700808e-05, | |
| "loss": 0.3088, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.4896, | |
| "grad_norm": 3.810133218765259, | |
| "learning_rate": 5.373989218328841e-05, | |
| "loss": 0.292, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.4928, | |
| "grad_norm": 5.0193023681640625, | |
| "learning_rate": 5.340296495956873e-05, | |
| "loss": 0.2982, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.496, | |
| "grad_norm": 3.1278297901153564, | |
| "learning_rate": 5.306603773584906e-05, | |
| "loss": 0.2494, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.4992, | |
| "grad_norm": 3.620786190032959, | |
| "learning_rate": 5.2729110512129385e-05, | |
| "loss": 0.2765, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.5024, | |
| "grad_norm": 3.282259464263916, | |
| "learning_rate": 5.23921832884097e-05, | |
| "loss": 0.273, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.5056, | |
| "grad_norm": 2.5962984561920166, | |
| "learning_rate": 5.205525606469003e-05, | |
| "loss": 0.2675, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.5088, | |
| "grad_norm": 5.7002787590026855, | |
| "learning_rate": 5.171832884097035e-05, | |
| "loss": 0.2561, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.512, | |
| "grad_norm": 3.456076145172119, | |
| "learning_rate": 5.138140161725068e-05, | |
| "loss": 0.2642, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.5152, | |
| "grad_norm": 3.859783411026001, | |
| "learning_rate": 5.1044474393531e-05, | |
| "loss": 0.289, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.5184, | |
| "grad_norm": 2.2919702529907227, | |
| "learning_rate": 5.070754716981132e-05, | |
| "loss": 0.2722, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.5216, | |
| "grad_norm": 2.624751567840576, | |
| "learning_rate": 5.037061994609165e-05, | |
| "loss": 0.2393, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.5248, | |
| "grad_norm": 3.1490752696990967, | |
| "learning_rate": 5.003369272237197e-05, | |
| "loss": 0.2401, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.528, | |
| "grad_norm": 3.9848804473876953, | |
| "learning_rate": 4.969676549865229e-05, | |
| "loss": 0.2362, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.5312, | |
| "grad_norm": 5.311415672302246, | |
| "learning_rate": 4.9359838274932616e-05, | |
| "loss": 0.2363, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.5344, | |
| "grad_norm": 2.246136426925659, | |
| "learning_rate": 4.902291105121294e-05, | |
| "loss": 0.244, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.5376, | |
| "grad_norm": 2.677457571029663, | |
| "learning_rate": 4.868598382749327e-05, | |
| "loss": 0.2428, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.5408, | |
| "grad_norm": 3.197678327560425, | |
| "learning_rate": 4.834905660377358e-05, | |
| "loss": 0.2147, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.544, | |
| "grad_norm": 4.309765815734863, | |
| "learning_rate": 4.801212938005391e-05, | |
| "loss": 0.225, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.5472, | |
| "grad_norm": 3.1181559562683105, | |
| "learning_rate": 4.7675202156334234e-05, | |
| "loss": 0.2173, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.5504, | |
| "grad_norm": 2.730708360671997, | |
| "learning_rate": 4.733827493261456e-05, | |
| "loss": 0.2174, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.5536, | |
| "grad_norm": 2.331942081451416, | |
| "learning_rate": 4.7001347708894886e-05, | |
| "loss": 0.2065, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.5568, | |
| "grad_norm": 4.71842622756958, | |
| "learning_rate": 4.66644204851752e-05, | |
| "loss": 0.2154, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 2.290827989578247, | |
| "learning_rate": 4.632749326145553e-05, | |
| "loss": 0.225, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.5632, | |
| "grad_norm": 1.7604676485061646, | |
| "learning_rate": 4.5990566037735846e-05, | |
| "loss": 0.2234, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.5664, | |
| "grad_norm": 2.8914430141448975, | |
| "learning_rate": 4.5653638814016176e-05, | |
| "loss": 0.2069, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.5696, | |
| "grad_norm": 2.695112466812134, | |
| "learning_rate": 4.53167115902965e-05, | |
| "loss": 0.2137, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.5728, | |
| "grad_norm": 3.153864860534668, | |
| "learning_rate": 4.497978436657682e-05, | |
| "loss": 0.2265, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.576, | |
| "grad_norm": 2.9276316165924072, | |
| "learning_rate": 4.464285714285715e-05, | |
| "loss": 0.2184, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.5792, | |
| "grad_norm": 2.0935676097869873, | |
| "learning_rate": 4.4305929919137465e-05, | |
| "loss": 0.2047, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.5824, | |
| "grad_norm": 2.0270116329193115, | |
| "learning_rate": 4.3969002695417794e-05, | |
| "loss": 0.196, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.5856, | |
| "grad_norm": 3.581533670425415, | |
| "learning_rate": 4.363207547169812e-05, | |
| "loss": 0.2023, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.5888, | |
| "grad_norm": 2.189016342163086, | |
| "learning_rate": 4.329514824797844e-05, | |
| "loss": 0.1947, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.592, | |
| "grad_norm": 2.613018274307251, | |
| "learning_rate": 4.295822102425876e-05, | |
| "loss": 0.1968, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.5952, | |
| "grad_norm": 1.6381279230117798, | |
| "learning_rate": 4.2621293800539084e-05, | |
| "loss": 0.1795, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.5984, | |
| "grad_norm": 2.9138712882995605, | |
| "learning_rate": 4.2284366576819406e-05, | |
| "loss": 0.185, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.6016, | |
| "grad_norm": 3.8309102058410645, | |
| "learning_rate": 4.1947439353099736e-05, | |
| "loss": 0.2061, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.6048, | |
| "grad_norm": 2.3997795581817627, | |
| "learning_rate": 4.161051212938006e-05, | |
| "loss": 0.1847, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.608, | |
| "grad_norm": 2.6803364753723145, | |
| "learning_rate": 4.127358490566038e-05, | |
| "loss": 0.1898, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.6112, | |
| "grad_norm": 2.676018238067627, | |
| "learning_rate": 4.09366576819407e-05, | |
| "loss": 0.1935, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.6144, | |
| "grad_norm": 4.005608558654785, | |
| "learning_rate": 4.0599730458221025e-05, | |
| "loss": 0.194, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.6176, | |
| "grad_norm": 2.0579423904418945, | |
| "learning_rate": 4.026280323450135e-05, | |
| "loss": 0.178, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.6208, | |
| "grad_norm": 3.387993574142456, | |
| "learning_rate": 3.992587601078167e-05, | |
| "loss": 0.1769, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.624, | |
| "grad_norm": 2.394421339035034, | |
| "learning_rate": 3.9588948787062e-05, | |
| "loss": 0.172, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.6272, | |
| "grad_norm": 3.314922571182251, | |
| "learning_rate": 3.9252021563342315e-05, | |
| "loss": 0.1776, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.6304, | |
| "grad_norm": 3.107218027114868, | |
| "learning_rate": 3.8915094339622644e-05, | |
| "loss": 0.1765, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.6336, | |
| "grad_norm": 4.710299015045166, | |
| "learning_rate": 3.8578167115902966e-05, | |
| "loss": 0.1879, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.6368, | |
| "grad_norm": 2.651448965072632, | |
| "learning_rate": 3.824123989218329e-05, | |
| "loss": 0.1689, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 2.7613463401794434, | |
| "learning_rate": 3.790431266846362e-05, | |
| "loss": 0.1854, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.6432, | |
| "grad_norm": 2.5603368282318115, | |
| "learning_rate": 3.7567385444743934e-05, | |
| "loss": 0.201, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.6464, | |
| "grad_norm": 2.4434683322906494, | |
| "learning_rate": 3.723045822102426e-05, | |
| "loss": 0.1606, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.6496, | |
| "grad_norm": 3.0501456260681152, | |
| "learning_rate": 3.689353099730458e-05, | |
| "loss": 0.1695, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.6528, | |
| "grad_norm": 1.5982695817947388, | |
| "learning_rate": 3.655660377358491e-05, | |
| "loss": 0.177, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.656, | |
| "grad_norm": 2.7528181076049805, | |
| "learning_rate": 3.621967654986524e-05, | |
| "loss": 0.1763, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.6592, | |
| "grad_norm": 2.7957491874694824, | |
| "learning_rate": 3.588274932614555e-05, | |
| "loss": 0.1702, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.6624, | |
| "grad_norm": 1.6270503997802734, | |
| "learning_rate": 3.554582210242588e-05, | |
| "loss": 0.164, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.6656, | |
| "grad_norm": 2.3618762493133545, | |
| "learning_rate": 3.52088948787062e-05, | |
| "loss": 0.1748, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.6688, | |
| "grad_norm": 3.2674052715301514, | |
| "learning_rate": 3.4871967654986526e-05, | |
| "loss": 0.1541, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.672, | |
| "grad_norm": 1.9046889543533325, | |
| "learning_rate": 3.453504043126685e-05, | |
| "loss": 0.1517, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.6752, | |
| "grad_norm": 1.8505709171295166, | |
| "learning_rate": 3.419811320754717e-05, | |
| "loss": 0.1861, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.6784, | |
| "grad_norm": 3.82865047454834, | |
| "learning_rate": 3.3861185983827494e-05, | |
| "loss": 0.1548, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.6816, | |
| "grad_norm": 3.949946641921997, | |
| "learning_rate": 3.3524258760107816e-05, | |
| "loss": 0.1797, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.6848, | |
| "grad_norm": 2.983675956726074, | |
| "learning_rate": 3.3187331536388145e-05, | |
| "loss": 0.1639, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.688, | |
| "grad_norm": 1.9660401344299316, | |
| "learning_rate": 3.285040431266847e-05, | |
| "loss": 0.1588, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.6912, | |
| "grad_norm": 2.7449395656585693, | |
| "learning_rate": 3.251347708894879e-05, | |
| "loss": 0.1553, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.6944, | |
| "grad_norm": 1.9276952743530273, | |
| "learning_rate": 3.217654986522911e-05, | |
| "loss": 0.1603, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.6976, | |
| "grad_norm": 2.2562904357910156, | |
| "learning_rate": 3.1839622641509435e-05, | |
| "loss": 0.1598, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.7008, | |
| "grad_norm": 2.2897818088531494, | |
| "learning_rate": 3.150269541778976e-05, | |
| "loss": 0.1601, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.704, | |
| "grad_norm": 1.7737034559249878, | |
| "learning_rate": 3.1165768194070086e-05, | |
| "loss": 0.1417, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.7072, | |
| "grad_norm": 2.640472173690796, | |
| "learning_rate": 3.08288409703504e-05, | |
| "loss": 0.1794, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.7104, | |
| "grad_norm": 2.8277909755706787, | |
| "learning_rate": 3.0491913746630728e-05, | |
| "loss": 0.1445, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.7136, | |
| "grad_norm": 2.287991523742676, | |
| "learning_rate": 3.0154986522911054e-05, | |
| "loss": 0.1528, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.7168, | |
| "grad_norm": 2.436365842819214, | |
| "learning_rate": 2.9818059299191376e-05, | |
| "loss": 0.1393, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 2.4227402210235596, | |
| "learning_rate": 2.9481132075471702e-05, | |
| "loss": 0.1391, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.7232, | |
| "grad_norm": 1.8195456266403198, | |
| "learning_rate": 2.914420485175202e-05, | |
| "loss": 0.1423, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.7264, | |
| "grad_norm": 1.9462803602218628, | |
| "learning_rate": 2.8807277628032347e-05, | |
| "loss": 0.1465, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.7296, | |
| "grad_norm": 1.5456150770187378, | |
| "learning_rate": 2.847035040431267e-05, | |
| "loss": 0.1217, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.7328, | |
| "grad_norm": 3.13179612159729, | |
| "learning_rate": 2.8133423180592995e-05, | |
| "loss": 0.1316, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.736, | |
| "grad_norm": 2.0412580966949463, | |
| "learning_rate": 2.7796495956873314e-05, | |
| "loss": 0.1432, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.7392, | |
| "grad_norm": 2.3401479721069336, | |
| "learning_rate": 2.745956873315364e-05, | |
| "loss": 0.1325, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.7424, | |
| "grad_norm": 1.8098496198654175, | |
| "learning_rate": 2.7122641509433965e-05, | |
| "loss": 0.1503, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.7456, | |
| "grad_norm": 1.7403185367584229, | |
| "learning_rate": 2.6785714285714288e-05, | |
| "loss": 0.1452, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.7488, | |
| "grad_norm": 2.0069713592529297, | |
| "learning_rate": 2.6448787061994614e-05, | |
| "loss": 0.1484, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.752, | |
| "grad_norm": 2.249183416366577, | |
| "learning_rate": 2.6111859838274933e-05, | |
| "loss": 0.138, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.7552, | |
| "grad_norm": 1.7765642404556274, | |
| "learning_rate": 2.577493261455526e-05, | |
| "loss": 0.1389, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.7584, | |
| "grad_norm": 1.5941318273544312, | |
| "learning_rate": 2.5438005390835577e-05, | |
| "loss": 0.1552, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.7616, | |
| "grad_norm": 1.5062505006790161, | |
| "learning_rate": 2.5101078167115903e-05, | |
| "loss": 0.1375, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.7648, | |
| "grad_norm": 2.394498586654663, | |
| "learning_rate": 2.476415094339623e-05, | |
| "loss": 0.1368, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.768, | |
| "grad_norm": 2.680593967437744, | |
| "learning_rate": 2.442722371967655e-05, | |
| "loss": 0.1367, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.7712, | |
| "grad_norm": 1.6163920164108276, | |
| "learning_rate": 2.4090296495956874e-05, | |
| "loss": 0.1297, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.7744, | |
| "grad_norm": 2.071157455444336, | |
| "learning_rate": 2.3753369272237196e-05, | |
| "loss": 0.131, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.7776, | |
| "grad_norm": 3.0522313117980957, | |
| "learning_rate": 2.341644204851752e-05, | |
| "loss": 0.1304, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.7808, | |
| "grad_norm": 2.017282009124756, | |
| "learning_rate": 2.3079514824797844e-05, | |
| "loss": 0.127, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.784, | |
| "grad_norm": 1.9524980783462524, | |
| "learning_rate": 2.274258760107817e-05, | |
| "loss": 0.1244, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.7872, | |
| "grad_norm": 2.556171417236328, | |
| "learning_rate": 2.2405660377358493e-05, | |
| "loss": 0.1409, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.7904, | |
| "grad_norm": 1.6622446775436401, | |
| "learning_rate": 2.2068733153638815e-05, | |
| "loss": 0.1274, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.7936, | |
| "grad_norm": 1.9248820543289185, | |
| "learning_rate": 2.1731805929919137e-05, | |
| "loss": 0.1323, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.7968, | |
| "grad_norm": 2.1036221981048584, | |
| "learning_rate": 2.1394878706199463e-05, | |
| "loss": 0.1301, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 1.6761208772659302, | |
| "learning_rate": 2.1057951482479785e-05, | |
| "loss": 0.1216, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.8032, | |
| "grad_norm": 2.89397931098938, | |
| "learning_rate": 2.0721024258760108e-05, | |
| "loss": 0.1338, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.8064, | |
| "grad_norm": 2.2912402153015137, | |
| "learning_rate": 2.038409703504043e-05, | |
| "loss": 0.1264, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.8096, | |
| "grad_norm": 2.0714144706726074, | |
| "learning_rate": 2.0047169811320756e-05, | |
| "loss": 0.1321, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.8128, | |
| "grad_norm": 3.272960901260376, | |
| "learning_rate": 1.971024258760108e-05, | |
| "loss": 0.1277, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.816, | |
| "grad_norm": 1.8504010438919067, | |
| "learning_rate": 1.9373315363881404e-05, | |
| "loss": 0.1301, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.8192, | |
| "grad_norm": 2.447061061859131, | |
| "learning_rate": 1.9036388140161727e-05, | |
| "loss": 0.1225, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.8224, | |
| "grad_norm": 2.020902395248413, | |
| "learning_rate": 1.869946091644205e-05, | |
| "loss": 0.1377, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.8256, | |
| "grad_norm": 1.4962462186813354, | |
| "learning_rate": 1.836253369272237e-05, | |
| "loss": 0.1238, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.8288, | |
| "grad_norm": 2.4895267486572266, | |
| "learning_rate": 1.8025606469002694e-05, | |
| "loss": 0.1254, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.832, | |
| "grad_norm": 1.4031554460525513, | |
| "learning_rate": 1.768867924528302e-05, | |
| "loss": 0.1391, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.8352, | |
| "grad_norm": 1.472172498703003, | |
| "learning_rate": 1.7351752021563345e-05, | |
| "loss": 0.1228, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.8384, | |
| "grad_norm": 1.7857037782669067, | |
| "learning_rate": 1.7014824797843668e-05, | |
| "loss": 0.1216, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.8416, | |
| "grad_norm": 2.2501163482666016, | |
| "learning_rate": 1.667789757412399e-05, | |
| "loss": 0.1271, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.8448, | |
| "grad_norm": 2.6882874965667725, | |
| "learning_rate": 1.6340970350404313e-05, | |
| "loss": 0.1394, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.848, | |
| "grad_norm": 1.6687155961990356, | |
| "learning_rate": 1.600404312668464e-05, | |
| "loss": 0.1163, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.8512, | |
| "grad_norm": 2.6670000553131104, | |
| "learning_rate": 1.566711590296496e-05, | |
| "loss": 0.1257, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.8544, | |
| "grad_norm": 1.77886164188385, | |
| "learning_rate": 1.5330188679245283e-05, | |
| "loss": 0.1333, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.8576, | |
| "grad_norm": 1.5165330171585083, | |
| "learning_rate": 1.4993261455525606e-05, | |
| "loss": 0.1151, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.8608, | |
| "grad_norm": 1.5670900344848633, | |
| "learning_rate": 1.465633423180593e-05, | |
| "loss": 0.1213, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.864, | |
| "grad_norm": 1.878943920135498, | |
| "learning_rate": 1.4319407008086256e-05, | |
| "loss": 0.1232, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.8672, | |
| "grad_norm": 1.5453932285308838, | |
| "learning_rate": 1.3982479784366578e-05, | |
| "loss": 0.1244, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.8704, | |
| "grad_norm": 1.8712074756622314, | |
| "learning_rate": 1.3645552560646902e-05, | |
| "loss": 0.1189, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.8736, | |
| "grad_norm": 1.4645411968231201, | |
| "learning_rate": 1.3308625336927224e-05, | |
| "loss": 0.1202, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.8768, | |
| "grad_norm": 1.605932354927063, | |
| "learning_rate": 1.2971698113207547e-05, | |
| "loss": 0.1162, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 1.7104394435882568, | |
| "learning_rate": 1.2634770889487871e-05, | |
| "loss": 0.1287, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.8832, | |
| "grad_norm": 1.9555670022964478, | |
| "learning_rate": 1.2297843665768195e-05, | |
| "loss": 0.2169, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.8864, | |
| "grad_norm": 2.0241458415985107, | |
| "learning_rate": 1.1960916442048519e-05, | |
| "loss": 0.1144, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.8896, | |
| "grad_norm": 1.8091477155685425, | |
| "learning_rate": 1.1623989218328842e-05, | |
| "loss": 0.1134, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.8928, | |
| "grad_norm": 2.328983783721924, | |
| "learning_rate": 1.1287061994609164e-05, | |
| "loss": 0.122, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.896, | |
| "grad_norm": 2.8249378204345703, | |
| "learning_rate": 1.0950134770889488e-05, | |
| "loss": 0.1177, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.8992, | |
| "grad_norm": 1.338085651397705, | |
| "learning_rate": 1.0613207547169812e-05, | |
| "loss": 0.128, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.9024, | |
| "grad_norm": 1.635053038597107, | |
| "learning_rate": 1.0276280323450135e-05, | |
| "loss": 0.1057, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.9056, | |
| "grad_norm": 1.7977182865142822, | |
| "learning_rate": 9.939353099730459e-06, | |
| "loss": 0.1097, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.9088, | |
| "grad_norm": 1.781990647315979, | |
| "learning_rate": 9.602425876010781e-06, | |
| "loss": 0.1237, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.912, | |
| "grad_norm": 2.48148775100708, | |
| "learning_rate": 9.265498652291107e-06, | |
| "loss": 0.1171, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.9152, | |
| "grad_norm": 1.3490359783172607, | |
| "learning_rate": 8.92857142857143e-06, | |
| "loss": 0.1077, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.9184, | |
| "grad_norm": 1.5256094932556152, | |
| "learning_rate": 8.591644204851752e-06, | |
| "loss": 0.1066, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.9216, | |
| "grad_norm": 1.5943819284439087, | |
| "learning_rate": 8.254716981132076e-06, | |
| "loss": 0.125, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.9248, | |
| "grad_norm": 1.8614747524261475, | |
| "learning_rate": 7.9177897574124e-06, | |
| "loss": 0.1112, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.928, | |
| "grad_norm": 1.2486650943756104, | |
| "learning_rate": 7.580862533692723e-06, | |
| "loss": 0.1246, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.9312, | |
| "grad_norm": 2.3569819927215576, | |
| "learning_rate": 7.243935309973046e-06, | |
| "loss": 0.1231, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.9344, | |
| "grad_norm": 2.198814868927002, | |
| "learning_rate": 6.90700808625337e-06, | |
| "loss": 0.114, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.9376, | |
| "grad_norm": 1.7546306848526, | |
| "learning_rate": 6.570080862533692e-06, | |
| "loss": 0.1127, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.9408, | |
| "grad_norm": 1.148824691772461, | |
| "learning_rate": 6.233153638814016e-06, | |
| "loss": 0.1128, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.944, | |
| "grad_norm": 1.61479914188385, | |
| "learning_rate": 5.89622641509434e-06, | |
| "loss": 0.1126, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.9472, | |
| "grad_norm": 2.365548849105835, | |
| "learning_rate": 5.5592991913746634e-06, | |
| "loss": 0.1133, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.9504, | |
| "grad_norm": 1.6197785139083862, | |
| "learning_rate": 5.222371967654987e-06, | |
| "loss": 0.1285, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.9536, | |
| "grad_norm": 1.727784276008606, | |
| "learning_rate": 4.88544474393531e-06, | |
| "loss": 0.1108, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.9568, | |
| "grad_norm": 1.6709702014923096, | |
| "learning_rate": 4.548517520215634e-06, | |
| "loss": 0.1145, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 2.0024709701538086, | |
| "learning_rate": 4.211590296495957e-06, | |
| "loss": 0.1135, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.9632, | |
| "grad_norm": 1.8355417251586914, | |
| "learning_rate": 3.8746630727762805e-06, | |
| "loss": 0.1063, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.9664, | |
| "grad_norm": 1.5643547773361206, | |
| "learning_rate": 3.5377358490566038e-06, | |
| "loss": 0.1203, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.9696, | |
| "grad_norm": 2.120378255844116, | |
| "learning_rate": 3.200808625336928e-06, | |
| "loss": 0.1158, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.9728, | |
| "grad_norm": 2.3552708625793457, | |
| "learning_rate": 2.8638814016172507e-06, | |
| "loss": 0.12, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.976, | |
| "grad_norm": 1.9691171646118164, | |
| "learning_rate": 2.5269541778975744e-06, | |
| "loss": 0.1143, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.9792, | |
| "grad_norm": 2.707933187484741, | |
| "learning_rate": 2.1900269541778976e-06, | |
| "loss": 0.1239, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.9824, | |
| "grad_norm": 1.4246569871902466, | |
| "learning_rate": 1.853099730458221e-06, | |
| "loss": 0.1111, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.9856, | |
| "grad_norm": 1.4995771646499634, | |
| "learning_rate": 1.5161725067385445e-06, | |
| "loss": 0.1151, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.9888, | |
| "grad_norm": 2.139070749282837, | |
| "learning_rate": 1.179245283018868e-06, | |
| "loss": 0.1178, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.992, | |
| "grad_norm": 1.5565009117126465, | |
| "learning_rate": 8.423180592991913e-07, | |
| "loss": 0.109, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.9952, | |
| "grad_norm": 1.771331787109375, | |
| "learning_rate": 5.053908355795148e-07, | |
| "loss": 0.1226, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.9984, | |
| "grad_norm": 2.2940099239349365, | |
| "learning_rate": 1.684636118598383e-07, | |
| "loss": 0.1086, | |
| "step": 3120 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 3125, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 32, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |