| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 6.0, | |
| "eval_steps": 500, | |
| "global_step": 16506, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.09087604507451835, | |
| "grad_norm": 0.8166529536247253, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8718, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.1817520901490367, | |
| "grad_norm": 0.8011488914489746, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5089, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.27262813522355506, | |
| "grad_norm": 0.8325187563896179, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3852, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.3635041802980734, | |
| "grad_norm": 1.6889874935150146, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3127, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.45438022537259176, | |
| "grad_norm": 0.6524374485015869, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2749, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.5452562704471101, | |
| "grad_norm": 0.8681249022483826, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2379, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.6361323155216285, | |
| "grad_norm": 0.9492128491401672, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2204, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.7270083605961468, | |
| "grad_norm": 0.35517942905426025, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2017, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.8178844056706652, | |
| "grad_norm": 0.5566073060035706, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1898, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.9087604507451835, | |
| "grad_norm": 0.23202739655971527, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1788, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.9996364958197019, | |
| "grad_norm": 0.4493115246295929, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1736, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 1.0905125408942202, | |
| "grad_norm": 0.41077283024787903, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1647, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.1813885859687385, | |
| "grad_norm": 0.24447296559810638, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1598, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.272264631043257, | |
| "grad_norm": 0.3004068434238434, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1594, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.3631406761177753, | |
| "grad_norm": 0.20352505147457123, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1554, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 1.4540167211922936, | |
| "grad_norm": 0.2659279704093933, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1502, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.5448927662668122, | |
| "grad_norm": 0.616990327835083, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1485, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 1.6357688113413305, | |
| "grad_norm": 0.1787092685699463, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1465, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.7266448564158487, | |
| "grad_norm": 0.26382288336753845, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1452, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 1.8175209014903673, | |
| "grad_norm": 0.22121331095695496, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1429, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.9083969465648853, | |
| "grad_norm": 0.23968210816383362, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1415, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 1.9992729916394039, | |
| "grad_norm": 0.20464320480823517, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1392, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.0901490367139224, | |
| "grad_norm": 0.23473434150218964, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1363, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 2.1810250817884405, | |
| "grad_norm": 0.3769285976886749, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1373, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.271901126862959, | |
| "grad_norm": 0.11411458998918533, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1355, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 2.362777171937477, | |
| "grad_norm": 0.14283251762390137, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1351, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 2.4536532170119956, | |
| "grad_norm": 0.1278432458639145, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1344, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 2.544529262086514, | |
| "grad_norm": 0.1538127213716507, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1351, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 2.635405307161032, | |
| "grad_norm": 0.1636444330215454, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1341, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 2.7262813522355507, | |
| "grad_norm": 0.23468904197216034, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1343, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 2.817157397310069, | |
| "grad_norm": 0.34467971324920654, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1342, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 2.9080334423845873, | |
| "grad_norm": 0.15800316631793976, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1331, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 2.998909487459106, | |
| "grad_norm": 0.18056762218475342, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1324, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 3.0897855325336243, | |
| "grad_norm": 0.12928904592990875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1301, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 3.1806615776081424, | |
| "grad_norm": 0.10186909139156342, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1299, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 3.271537622682661, | |
| "grad_norm": 0.1294797658920288, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1311, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 3.3624136677571794, | |
| "grad_norm": 0.96567702293396, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1292, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 3.4532897128316975, | |
| "grad_norm": 0.7819355130195618, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1289, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 3.544165757906216, | |
| "grad_norm": 0.19112642109394073, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1305, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 3.6350418029807345, | |
| "grad_norm": 0.09955710917711258, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1298, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 3.7259178480552526, | |
| "grad_norm": 0.13611435890197754, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1286, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 3.816793893129771, | |
| "grad_norm": 0.11884398013353348, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1307, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 3.907669938204289, | |
| "grad_norm": 0.1615493893623352, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1284, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 3.9985459832788077, | |
| "grad_norm": 0.31517332792282104, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1291, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 4.089422028353326, | |
| "grad_norm": 0.6462264060974121, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1276, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 4.180298073427845, | |
| "grad_norm": 0.09960414469242096, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1265, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 4.271174118502363, | |
| "grad_norm": 0.09911854565143585, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1286, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 4.362050163576881, | |
| "grad_norm": 0.12108103185892105, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1282, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 4.4529262086514, | |
| "grad_norm": 0.09403906762599945, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1277, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 4.543802253725918, | |
| "grad_norm": 0.12478747963905334, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1294, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 4.634678298800436, | |
| "grad_norm": 0.5839309692382812, | |
| "learning_rate": 0.0002, | |
| "loss": 0.128, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 4.725554343874954, | |
| "grad_norm": 0.0925154760479927, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1278, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 4.816430388949473, | |
| "grad_norm": 0.09721515327692032, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1243, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 4.907306434023991, | |
| "grad_norm": 0.11027718335390091, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1264, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 4.99818247909851, | |
| "grad_norm": 0.24596811830997467, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1247, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 5.089058524173028, | |
| "grad_norm": 0.3751857876777649, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1244, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 5.179934569247546, | |
| "grad_norm": 0.0758715569972992, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1235, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 5.270810614322064, | |
| "grad_norm": 0.10280770063400269, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1263, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 5.361686659396583, | |
| "grad_norm": 0.10937105864286423, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1247, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 5.452562704471101, | |
| "grad_norm": 0.10805700719356537, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1257, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 5.543438749545619, | |
| "grad_norm": 0.1117258295416832, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1289, | |
| "step": 15250 | |
| }, | |
| { | |
| "epoch": 5.634314794620138, | |
| "grad_norm": 0.08629261702299118, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1258, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 5.7251908396946565, | |
| "grad_norm": 0.10845168679952621, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1259, | |
| "step": 15750 | |
| }, | |
| { | |
| "epoch": 5.8160668847691745, | |
| "grad_norm": 0.08773403614759445, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1255, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 5.9069429298436935, | |
| "grad_norm": 0.08745443820953369, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1268, | |
| "step": 16250 | |
| }, | |
| { | |
| "epoch": 5.997818974918212, | |
| "grad_norm": 0.29043182730674744, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1253, | |
| "step": 16500 | |
| } | |
| ], | |
| "logging_steps": 250, | |
| "max_steps": 16506, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 6, | |
| "save_steps": 250, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.32857340165632e+17, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |