| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 6.0, | |
| "global_step": 73008, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.965757177295639e-05, | |
| "loss": 3.431, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.931514354591278e-05, | |
| "loss": 3.3117, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.8972715318869165e-05, | |
| "loss": 3.2778, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.863028709182555e-05, | |
| "loss": 3.2331, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.828785886478194e-05, | |
| "loss": 3.2122, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.794543063773833e-05, | |
| "loss": 3.1961, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.7603002410694716e-05, | |
| "loss": 3.1842, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.726057418365111e-05, | |
| "loss": 3.1603, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.69181459566075e-05, | |
| "loss": 3.147, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.6575717729563886e-05, | |
| "loss": 3.1572, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.623328950252027e-05, | |
| "loss": 3.1329, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.589086127547666e-05, | |
| "loss": 3.1132, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.554843304843305e-05, | |
| "loss": 3.1188, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.520600482138944e-05, | |
| "loss": 3.1141, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.486357659434583e-05, | |
| "loss": 3.1044, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.452114836730222e-05, | |
| "loss": 3.0998, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.4178720140258606e-05, | |
| "loss": 3.0963, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.3836291913214994e-05, | |
| "loss": 3.0978, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.349386368617138e-05, | |
| "loss": 3.0674, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.315143545912777e-05, | |
| "loss": 3.0855, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.2809007232084157e-05, | |
| "loss": 3.0664, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.2466579005040544e-05, | |
| "loss": 3.0678, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.212415077799693e-05, | |
| "loss": 3.0681, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.178172255095332e-05, | |
| "loss": 3.0565, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 4.1439294323909714e-05, | |
| "loss": 3.0114, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.10968660968661e-05, | |
| "loss": 2.9869, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.075443786982249e-05, | |
| "loss": 2.9743, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.041200964277888e-05, | |
| "loss": 2.9838, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.0069581415735264e-05, | |
| "loss": 2.9806, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.972715318869165e-05, | |
| "loss": 2.9715, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.938472496164804e-05, | |
| "loss": 2.9702, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.904229673460443e-05, | |
| "loss": 2.9667, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.8699868507560815e-05, | |
| "loss": 2.9737, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.83574402805172e-05, | |
| "loss": 2.9569, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.801501205347359e-05, | |
| "loss": 2.9759, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 3.767258382642998e-05, | |
| "loss": 2.9749, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.7330155599386366e-05, | |
| "loss": 2.9742, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 3.698772737234275e-05, | |
| "loss": 2.97, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 3.664529914529915e-05, | |
| "loss": 2.9551, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 3.6302870918255535e-05, | |
| "loss": 2.9464, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 3.596044269121192e-05, | |
| "loss": 2.9549, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 3.561801446416832e-05, | |
| "loss": 2.969, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 3.5275586237124705e-05, | |
| "loss": 2.9478, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 3.493315801008109e-05, | |
| "loss": 2.9568, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 3.459072978303748e-05, | |
| "loss": 2.9509, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 3.424830155599387e-05, | |
| "loss": 2.9484, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 3.3905873328950256e-05, | |
| "loss": 2.9646, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 3.356344510190664e-05, | |
| "loss": 2.9476, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 3.322101687486303e-05, | |
| "loss": 2.9165, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 3.287858864781942e-05, | |
| "loss": 2.8741, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 3.2536160420775806e-05, | |
| "loss": 2.8902, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 3.2193732193732194e-05, | |
| "loss": 2.8788, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.185130396668858e-05, | |
| "loss": 2.8775, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.150887573964497e-05, | |
| "loss": 2.885, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 3.116644751260136e-05, | |
| "loss": 2.8678, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 3.082401928555775e-05, | |
| "loss": 2.8771, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 3.0481591058514136e-05, | |
| "loss": 2.8873, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 3.0139162831470523e-05, | |
| "loss": 2.8768, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 2.979673460442691e-05, | |
| "loss": 2.8938, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 2.9454306377383302e-05, | |
| "loss": 2.8774, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 2.911187815033969e-05, | |
| "loss": 2.8761, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 2.8769449923296077e-05, | |
| "loss": 2.8824, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 2.8427021696252465e-05, | |
| "loss": 2.8737, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 2.8084593469208852e-05, | |
| "loss": 2.8802, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 2.774216524216524e-05, | |
| "loss": 2.8744, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 2.7399737015121628e-05, | |
| "loss": 2.8713, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 2.705730878807802e-05, | |
| "loss": 2.8822, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.671488056103441e-05, | |
| "loss": 2.8809, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 2.63724523339908e-05, | |
| "loss": 2.8694, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.603002410694719e-05, | |
| "loss": 2.8723, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 2.5687595879903576e-05, | |
| "loss": 2.8734, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 2.5345167652859964e-05, | |
| "loss": 2.8757, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 2.500273942581635e-05, | |
| "loss": 2.8737, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 2.466031119877274e-05, | |
| "loss": 2.8095, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 2.4317882971729127e-05, | |
| "loss": 2.8181, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 2.3975454744685514e-05, | |
| "loss": 2.8102, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 2.3633026517641905e-05, | |
| "loss": 2.8261, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 2.3290598290598293e-05, | |
| "loss": 2.8121, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 2.294817006355468e-05, | |
| "loss": 2.8127, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 2.260574183651107e-05, | |
| "loss": 2.8246, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 2.2263313609467456e-05, | |
| "loss": 2.8188, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 3.37, | |
| "learning_rate": 2.1920885382423844e-05, | |
| "loss": 2.8247, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 2.157845715538023e-05, | |
| "loss": 2.8087, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 2.1236028928336622e-05, | |
| "loss": 2.8135, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 2.089360070129301e-05, | |
| "loss": 2.8208, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "learning_rate": 2.0551172474249398e-05, | |
| "loss": 2.8227, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 2.0208744247205785e-05, | |
| "loss": 2.8272, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 1.9866316020162176e-05, | |
| "loss": 2.8168, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 1.9523887793118564e-05, | |
| "loss": 2.8304, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 1.918145956607495e-05, | |
| "loss": 2.8102, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 1.883903133903134e-05, | |
| "loss": 2.8139, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 1.849660311198773e-05, | |
| "loss": 2.806, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 1.8154174884944118e-05, | |
| "loss": 2.8253, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 1.7811746657900506e-05, | |
| "loss": 2.8237, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 1.7469318430856893e-05, | |
| "loss": 2.8118, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 1.712689020381328e-05, | |
| "loss": 2.8203, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 1.678446197676967e-05, | |
| "loss": 2.8142, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 1.6442033749726056e-05, | |
| "loss": 2.7918, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 1.6099605522682447e-05, | |
| "loss": 2.7683, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 1.5757177295638835e-05, | |
| "loss": 2.7645, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 1.5414749068595226e-05, | |
| "loss": 2.7562, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 1.5072320841551612e-05, | |
| "loss": 2.7706, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 1.4729892614508001e-05, | |
| "loss": 2.779, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 1.4387464387464389e-05, | |
| "loss": 2.7665, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 1.4045036160420776e-05, | |
| "loss": 2.7696, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 1.3702607933377166e-05, | |
| "loss": 2.7725, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 1.3360179706333553e-05, | |
| "loss": 2.7724, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 1.3017751479289941e-05, | |
| "loss": 2.7725, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 1.2675323252246329e-05, | |
| "loss": 2.7759, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 1.2332895025202718e-05, | |
| "loss": 2.7655, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 1.1990466798159107e-05, | |
| "loss": 2.7836, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 1.1648038571115495e-05, | |
| "loss": 2.7709, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 1.1305610344071883e-05, | |
| "loss": 2.7758, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 1.0963182117028272e-05, | |
| "loss": 2.7825, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 1.062075388998466e-05, | |
| "loss": 2.7641, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 4.77, | |
| "learning_rate": 1.0278325662941047e-05, | |
| "loss": 2.7886, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 9.935897435897435e-06, | |
| "loss": 2.7784, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 9.593469208853826e-06, | |
| "loss": 2.7692, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 9.251040981810214e-06, | |
| "loss": 2.7633, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 8.908612754766601e-06, | |
| "loss": 2.7832, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 8.566184527722989e-06, | |
| "loss": 2.7778, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 5.01, | |
| "learning_rate": 8.223756300679378e-06, | |
| "loss": 2.7643, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 5.05, | |
| "learning_rate": 7.881328073635766e-06, | |
| "loss": 2.7539, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 5.1, | |
| "learning_rate": 7.538899846592154e-06, | |
| "loss": 2.7482, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 5.14, | |
| "learning_rate": 7.196471619548544e-06, | |
| "loss": 2.7536, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 5.18, | |
| "learning_rate": 6.854043392504931e-06, | |
| "loss": 2.7419, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 5.22, | |
| "learning_rate": 6.51161516546132e-06, | |
| "loss": 2.731, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 5.26, | |
| "learning_rate": 6.1691869384177075e-06, | |
| "loss": 2.7355, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 5.3, | |
| "learning_rate": 5.826758711374096e-06, | |
| "loss": 2.7378, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 5.34, | |
| "learning_rate": 5.4843304843304845e-06, | |
| "loss": 2.7447, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 5.38, | |
| "learning_rate": 5.141902257286873e-06, | |
| "loss": 2.7548, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 5.42, | |
| "learning_rate": 4.799474030243261e-06, | |
| "loss": 2.7362, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 5.47, | |
| "learning_rate": 4.45704580319965e-06, | |
| "loss": 2.7434, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 4.114617576156038e-06, | |
| "loss": 2.7424, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 5.55, | |
| "learning_rate": 3.772189349112426e-06, | |
| "loss": 2.7468, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 5.59, | |
| "learning_rate": 3.429761122068815e-06, | |
| "loss": 2.7451, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 5.63, | |
| "learning_rate": 3.087332895025203e-06, | |
| "loss": 2.7504, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 5.67, | |
| "learning_rate": 2.744904667981591e-06, | |
| "loss": 2.7372, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 5.71, | |
| "learning_rate": 2.4024764409379797e-06, | |
| "loss": 2.7349, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 5.75, | |
| "learning_rate": 2.0600482138943677e-06, | |
| "loss": 2.7477, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 5.79, | |
| "learning_rate": 1.7176199868507562e-06, | |
| "loss": 2.7394, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 5.83, | |
| "learning_rate": 1.3751917598071445e-06, | |
| "loss": 2.7303, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 5.88, | |
| "learning_rate": 1.0327635327635328e-06, | |
| "loss": 2.7408, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 5.92, | |
| "learning_rate": 6.903353057199211e-07, | |
| "loss": 2.7455, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 5.96, | |
| "learning_rate": 3.479070786763094e-07, | |
| "loss": 2.7378, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "learning_rate": 5.478851632697787e-09, | |
| "loss": 2.7414, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "step": 73008, | |
| "total_flos": 3.8152817344512e+16, | |
| "train_loss": 2.887446778400044, | |
| "train_runtime": 18728.1376, | |
| "train_samples_per_second": 3.898, | |
| "train_steps_per_second": 3.898 | |
| } | |
| ], | |
| "max_steps": 73008, | |
| "num_train_epochs": 6, | |
| "total_flos": 3.8152817344512e+16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |