| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.21606648199446, | |
| "eval_steps": 500, | |
| "global_step": 200, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0110803324099723, | |
| "grad_norm": 0.9041995406150818, | |
| "learning_rate": 0.0001, | |
| "loss": 2.5533, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0221606648199446, | |
| "grad_norm": 0.9356410503387451, | |
| "learning_rate": 9.949748743718594e-05, | |
| "loss": 2.5542, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0332409972299169, | |
| "grad_norm": 0.9131423830986023, | |
| "learning_rate": 9.899497487437186e-05, | |
| "loss": 2.5028, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.0443213296398892, | |
| "grad_norm": 0.9732369780540466, | |
| "learning_rate": 9.84924623115578e-05, | |
| "loss": 2.393, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.055401662049861494, | |
| "grad_norm": 0.9332369565963745, | |
| "learning_rate": 9.798994974874372e-05, | |
| "loss": 2.2448, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0664819944598338, | |
| "grad_norm": 1.0083566904067993, | |
| "learning_rate": 9.748743718592965e-05, | |
| "loss": 2.1345, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.07756232686980609, | |
| "grad_norm": 0.8932923078536987, | |
| "learning_rate": 9.698492462311559e-05, | |
| "loss": 1.9912, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0886426592797784, | |
| "grad_norm": 1.8232415914535522, | |
| "learning_rate": 9.64824120603015e-05, | |
| "loss": 1.844, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0997229916897507, | |
| "grad_norm": 0.8558672070503235, | |
| "learning_rate": 9.597989949748745e-05, | |
| "loss": 1.7406, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.11080332409972299, | |
| "grad_norm": 0.7986319661140442, | |
| "learning_rate": 9.547738693467337e-05, | |
| "loss": 1.636, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.12188365650969529, | |
| "grad_norm": 0.8156765699386597, | |
| "learning_rate": 9.49748743718593e-05, | |
| "loss": 1.6183, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.1329639889196676, | |
| "grad_norm": 0.7248062491416931, | |
| "learning_rate": 9.447236180904523e-05, | |
| "loss": 1.57, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.1440443213296399, | |
| "grad_norm": 0.6793098449707031, | |
| "learning_rate": 9.396984924623115e-05, | |
| "loss": 1.4801, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.15512465373961218, | |
| "grad_norm": 0.566728949546814, | |
| "learning_rate": 9.34673366834171e-05, | |
| "loss": 1.4867, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.16620498614958448, | |
| "grad_norm": 0.5523749589920044, | |
| "learning_rate": 9.296482412060302e-05, | |
| "loss": 1.36, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.1772853185595568, | |
| "grad_norm": 0.5163611173629761, | |
| "learning_rate": 9.246231155778895e-05, | |
| "loss": 1.4472, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.1883656509695291, | |
| "grad_norm": 0.5090933442115784, | |
| "learning_rate": 9.195979899497488e-05, | |
| "loss": 1.4181, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.1994459833795014, | |
| "grad_norm": 0.5989904999732971, | |
| "learning_rate": 9.14572864321608e-05, | |
| "loss": 1.4414, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.21052631578947367, | |
| "grad_norm": 0.5392615795135498, | |
| "learning_rate": 9.095477386934675e-05, | |
| "loss": 1.3723, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.22160664819944598, | |
| "grad_norm": 0.6069510579109192, | |
| "learning_rate": 9.045226130653267e-05, | |
| "loss": 1.5178, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.23268698060941828, | |
| "grad_norm": 0.5653948187828064, | |
| "learning_rate": 8.99497487437186e-05, | |
| "loss": 1.2593, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.24376731301939059, | |
| "grad_norm": 0.5368112325668335, | |
| "learning_rate": 8.944723618090453e-05, | |
| "loss": 1.3471, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.2548476454293629, | |
| "grad_norm": 0.5639390349388123, | |
| "learning_rate": 8.894472361809045e-05, | |
| "loss": 1.4432, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.2659279778393352, | |
| "grad_norm": 0.5520769953727722, | |
| "learning_rate": 8.84422110552764e-05, | |
| "loss": 1.3426, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.2770083102493075, | |
| "grad_norm": 0.6374968886375427, | |
| "learning_rate": 8.793969849246232e-05, | |
| "loss": 1.5108, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.2880886426592798, | |
| "grad_norm": 0.6635875701904297, | |
| "learning_rate": 8.743718592964825e-05, | |
| "loss": 1.4302, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.29916897506925205, | |
| "grad_norm": 0.6544961929321289, | |
| "learning_rate": 8.693467336683418e-05, | |
| "loss": 1.4542, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.31024930747922436, | |
| "grad_norm": 0.5790326595306396, | |
| "learning_rate": 8.64321608040201e-05, | |
| "loss": 1.3176, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.32132963988919666, | |
| "grad_norm": 0.6066296696662903, | |
| "learning_rate": 8.592964824120603e-05, | |
| "loss": 1.355, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.33240997229916897, | |
| "grad_norm": 0.6547830700874329, | |
| "learning_rate": 8.542713567839196e-05, | |
| "loss": 1.3518, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.34349030470914127, | |
| "grad_norm": 0.6181479096412659, | |
| "learning_rate": 8.49246231155779e-05, | |
| "loss": 1.2498, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.3545706371191136, | |
| "grad_norm": 0.7298603057861328, | |
| "learning_rate": 8.442211055276383e-05, | |
| "loss": 1.2897, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.3656509695290859, | |
| "grad_norm": 0.7295474410057068, | |
| "learning_rate": 8.391959798994975e-05, | |
| "loss": 1.3034, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.3767313019390582, | |
| "grad_norm": 0.7686471343040466, | |
| "learning_rate": 8.341708542713568e-05, | |
| "loss": 1.3327, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.3878116343490305, | |
| "grad_norm": 0.7613719701766968, | |
| "learning_rate": 8.291457286432161e-05, | |
| "loss": 1.2834, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.3988919667590028, | |
| "grad_norm": 0.8543422222137451, | |
| "learning_rate": 8.241206030150754e-05, | |
| "loss": 1.3813, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.4099722991689751, | |
| "grad_norm": 0.9008685946464539, | |
| "learning_rate": 8.190954773869348e-05, | |
| "loss": 1.3528, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.42105263157894735, | |
| "grad_norm": 0.8236178159713745, | |
| "learning_rate": 8.14070351758794e-05, | |
| "loss": 1.3074, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.43213296398891965, | |
| "grad_norm": 0.8271133899688721, | |
| "learning_rate": 8.090452261306533e-05, | |
| "loss": 1.264, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.44321329639889195, | |
| "grad_norm": 0.8218770623207092, | |
| "learning_rate": 8.040201005025126e-05, | |
| "loss": 1.2871, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.45429362880886426, | |
| "grad_norm": 0.7466350197792053, | |
| "learning_rate": 7.989949748743719e-05, | |
| "loss": 1.2691, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.46537396121883656, | |
| "grad_norm": 0.7745970487594604, | |
| "learning_rate": 7.939698492462313e-05, | |
| "loss": 1.2766, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.47645429362880887, | |
| "grad_norm": 0.7701446413993835, | |
| "learning_rate": 7.889447236180904e-05, | |
| "loss": 1.294, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.48753462603878117, | |
| "grad_norm": 0.6183106899261475, | |
| "learning_rate": 7.839195979899498e-05, | |
| "loss": 1.3002, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.4986149584487535, | |
| "grad_norm": 0.5864247679710388, | |
| "learning_rate": 7.788944723618091e-05, | |
| "loss": 1.2178, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.5096952908587258, | |
| "grad_norm": 0.5792540907859802, | |
| "learning_rate": 7.738693467336684e-05, | |
| "loss": 1.2126, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.5207756232686981, | |
| "grad_norm": 0.6436092853546143, | |
| "learning_rate": 7.688442211055277e-05, | |
| "loss": 1.317, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.5318559556786704, | |
| "grad_norm": 0.5778934359550476, | |
| "learning_rate": 7.638190954773869e-05, | |
| "loss": 1.2303, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.5429362880886427, | |
| "grad_norm": 0.5667629837989807, | |
| "learning_rate": 7.587939698492463e-05, | |
| "loss": 1.239, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.554016620498615, | |
| "grad_norm": 0.579045832157135, | |
| "learning_rate": 7.537688442211056e-05, | |
| "loss": 1.1976, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.5650969529085873, | |
| "grad_norm": 0.5645351409912109, | |
| "learning_rate": 7.487437185929649e-05, | |
| "loss": 1.2011, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.5761772853185596, | |
| "grad_norm": 0.6186327934265137, | |
| "learning_rate": 7.437185929648241e-05, | |
| "loss": 1.26, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.5872576177285319, | |
| "grad_norm": 0.6174798011779785, | |
| "learning_rate": 7.386934673366834e-05, | |
| "loss": 1.2812, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.5983379501385041, | |
| "grad_norm": 0.6200773119926453, | |
| "learning_rate": 7.336683417085427e-05, | |
| "loss": 1.2493, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.6094182825484764, | |
| "grad_norm": 0.5842644572257996, | |
| "learning_rate": 7.28643216080402e-05, | |
| "loss": 1.2587, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.6204986149584487, | |
| "grad_norm": 0.5945526957511902, | |
| "learning_rate": 7.236180904522614e-05, | |
| "loss": 1.2105, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.631578947368421, | |
| "grad_norm": 0.6002059578895569, | |
| "learning_rate": 7.185929648241206e-05, | |
| "loss": 1.2404, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.6426592797783933, | |
| "grad_norm": 0.5362327098846436, | |
| "learning_rate": 7.135678391959799e-05, | |
| "loss": 1.3024, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.6537396121883656, | |
| "grad_norm": 0.5753970146179199, | |
| "learning_rate": 7.085427135678392e-05, | |
| "loss": 1.2076, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.6648199445983379, | |
| "grad_norm": 0.6161749958992004, | |
| "learning_rate": 7.035175879396985e-05, | |
| "loss": 1.1563, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.6759002770083102, | |
| "grad_norm": 0.578284502029419, | |
| "learning_rate": 6.984924623115579e-05, | |
| "loss": 1.2165, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.6869806094182825, | |
| "grad_norm": 0.6425468325614929, | |
| "learning_rate": 6.93467336683417e-05, | |
| "loss": 1.2972, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.6980609418282548, | |
| "grad_norm": 0.5888572335243225, | |
| "learning_rate": 6.884422110552764e-05, | |
| "loss": 1.1187, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.7091412742382271, | |
| "grad_norm": 0.6597657203674316, | |
| "learning_rate": 6.834170854271357e-05, | |
| "loss": 1.2613, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.7202216066481995, | |
| "grad_norm": 0.6418899893760681, | |
| "learning_rate": 6.78391959798995e-05, | |
| "loss": 1.1614, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.7313019390581718, | |
| "grad_norm": 0.5714394450187683, | |
| "learning_rate": 6.733668341708544e-05, | |
| "loss": 1.1293, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.7423822714681441, | |
| "grad_norm": 0.6186115145683289, | |
| "learning_rate": 6.683417085427135e-05, | |
| "loss": 1.1836, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.7534626038781164, | |
| "grad_norm": 0.5801815390586853, | |
| "learning_rate": 6.633165829145729e-05, | |
| "loss": 1.2035, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.7645429362880887, | |
| "grad_norm": 0.6500537991523743, | |
| "learning_rate": 6.582914572864322e-05, | |
| "loss": 1.2234, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.775623268698061, | |
| "grad_norm": 0.5818614363670349, | |
| "learning_rate": 6.532663316582915e-05, | |
| "loss": 1.2251, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.7867036011080333, | |
| "grad_norm": 0.6213693022727966, | |
| "learning_rate": 6.482412060301508e-05, | |
| "loss": 1.2281, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.7977839335180056, | |
| "grad_norm": 0.5809823870658875, | |
| "learning_rate": 6.4321608040201e-05, | |
| "loss": 1.18, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.8088642659279779, | |
| "grad_norm": 0.5340937376022339, | |
| "learning_rate": 6.381909547738694e-05, | |
| "loss": 1.1571, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.8199445983379502, | |
| "grad_norm": 0.5844171047210693, | |
| "learning_rate": 6.331658291457287e-05, | |
| "loss": 1.2212, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.8310249307479224, | |
| "grad_norm": 0.5754027366638184, | |
| "learning_rate": 6.28140703517588e-05, | |
| "loss": 1.2023, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.8421052631578947, | |
| "grad_norm": 0.6343287229537964, | |
| "learning_rate": 6.231155778894473e-05, | |
| "loss": 1.2817, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.853185595567867, | |
| "grad_norm": 0.6712394952774048, | |
| "learning_rate": 6.180904522613065e-05, | |
| "loss": 1.2516, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.8642659279778393, | |
| "grad_norm": 0.5741068720817566, | |
| "learning_rate": 6.130653266331658e-05, | |
| "loss": 1.2114, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.8753462603878116, | |
| "grad_norm": 0.6043746471405029, | |
| "learning_rate": 6.080402010050251e-05, | |
| "loss": 1.1762, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.8864265927977839, | |
| "grad_norm": 0.6717391610145569, | |
| "learning_rate": 6.030150753768844e-05, | |
| "loss": 1.228, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.8975069252077562, | |
| "grad_norm": 0.6319631338119507, | |
| "learning_rate": 5.979899497487438e-05, | |
| "loss": 1.2198, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.9085872576177285, | |
| "grad_norm": 0.6088309288024902, | |
| "learning_rate": 5.929648241206031e-05, | |
| "loss": 1.211, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.9196675900277008, | |
| "grad_norm": 0.6102275252342224, | |
| "learning_rate": 5.879396984924623e-05, | |
| "loss": 1.187, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.9307479224376731, | |
| "grad_norm": 0.5819908976554871, | |
| "learning_rate": 5.829145728643216e-05, | |
| "loss": 1.2206, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.9418282548476454, | |
| "grad_norm": 0.601245641708374, | |
| "learning_rate": 5.778894472361809e-05, | |
| "loss": 1.2269, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.9529085872576177, | |
| "grad_norm": 0.6378527283668518, | |
| "learning_rate": 5.728643216080403e-05, | |
| "loss": 1.2919, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.96398891966759, | |
| "grad_norm": 0.6004720330238342, | |
| "learning_rate": 5.6783919597989955e-05, | |
| "loss": 1.204, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.9750692520775623, | |
| "grad_norm": 0.6289650797843933, | |
| "learning_rate": 5.628140703517588e-05, | |
| "loss": 1.1608, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.9861495844875346, | |
| "grad_norm": 0.6542637944221497, | |
| "learning_rate": 5.577889447236181e-05, | |
| "loss": 1.1592, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.997229916897507, | |
| "grad_norm": 0.6260697245597839, | |
| "learning_rate": 5.527638190954774e-05, | |
| "loss": 1.2151, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.0083102493074791, | |
| "grad_norm": 1.5121455192565918, | |
| "learning_rate": 5.477386934673368e-05, | |
| "loss": 2.0839, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 1.0193905817174516, | |
| "grad_norm": 0.574207067489624, | |
| "learning_rate": 5.4271356783919604e-05, | |
| "loss": 1.1613, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.0304709141274238, | |
| "grad_norm": 0.6616407036781311, | |
| "learning_rate": 5.376884422110553e-05, | |
| "loss": 1.3486, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 1.0415512465373962, | |
| "grad_norm": 0.5284487009048462, | |
| "learning_rate": 5.3266331658291455e-05, | |
| "loss": 1.041, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 1.0526315789473684, | |
| "grad_norm": 0.5977286696434021, | |
| "learning_rate": 5.276381909547739e-05, | |
| "loss": 1.1503, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 1.0637119113573408, | |
| "grad_norm": 0.5718993544578552, | |
| "learning_rate": 5.226130653266332e-05, | |
| "loss": 1.1083, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 1.074792243767313, | |
| "grad_norm": 0.6474930644035339, | |
| "learning_rate": 5.175879396984925e-05, | |
| "loss": 1.2006, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 1.0858725761772854, | |
| "grad_norm": 0.5949592590332031, | |
| "learning_rate": 5.125628140703518e-05, | |
| "loss": 1.1012, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 1.0969529085872576, | |
| "grad_norm": 0.5743479132652283, | |
| "learning_rate": 5.0753768844221104e-05, | |
| "loss": 1.094, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 1.10803324099723, | |
| "grad_norm": 0.7052620053291321, | |
| "learning_rate": 5.0251256281407036e-05, | |
| "loss": 1.2209, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.1191135734072022, | |
| "grad_norm": 0.6266711354255676, | |
| "learning_rate": 4.974874371859297e-05, | |
| "loss": 1.0795, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 1.1301939058171746, | |
| "grad_norm": 0.6345641016960144, | |
| "learning_rate": 4.92462311557789e-05, | |
| "loss": 1.1237, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 1.1412742382271468, | |
| "grad_norm": 0.5873332023620605, | |
| "learning_rate": 4.874371859296483e-05, | |
| "loss": 1.062, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 1.1523545706371192, | |
| "grad_norm": 0.6368873119354248, | |
| "learning_rate": 4.824120603015075e-05, | |
| "loss": 0.9626, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 1.1634349030470914, | |
| "grad_norm": 0.6816115975379944, | |
| "learning_rate": 4.7738693467336685e-05, | |
| "loss": 1.0869, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 1.1745152354570636, | |
| "grad_norm": 0.7037101984024048, | |
| "learning_rate": 4.723618090452262e-05, | |
| "loss": 1.2178, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 1.185595567867036, | |
| "grad_norm": 0.6814127564430237, | |
| "learning_rate": 4.673366834170855e-05, | |
| "loss": 1.1456, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 1.1966759002770084, | |
| "grad_norm": 0.6351640224456787, | |
| "learning_rate": 4.6231155778894475e-05, | |
| "loss": 1.0543, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 1.2077562326869806, | |
| "grad_norm": 0.729073703289032, | |
| "learning_rate": 4.57286432160804e-05, | |
| "loss": 1.1579, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 1.2188365650969528, | |
| "grad_norm": 0.7375718355178833, | |
| "learning_rate": 4.522613065326633e-05, | |
| "loss": 1.178, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.2299168975069252, | |
| "grad_norm": 0.686299204826355, | |
| "learning_rate": 4.4723618090452266e-05, | |
| "loss": 1.1791, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 1.2409972299168974, | |
| "grad_norm": 0.6791194081306458, | |
| "learning_rate": 4.42211055276382e-05, | |
| "loss": 1.1277, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 1.2520775623268698, | |
| "grad_norm": 0.7525886297225952, | |
| "learning_rate": 4.3718592964824124e-05, | |
| "loss": 1.1455, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 1.263157894736842, | |
| "grad_norm": 0.7229343056678772, | |
| "learning_rate": 4.321608040201005e-05, | |
| "loss": 1.2113, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 1.2742382271468145, | |
| "grad_norm": 0.7222431302070618, | |
| "learning_rate": 4.271356783919598e-05, | |
| "loss": 1.1342, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 1.2853185595567866, | |
| "grad_norm": 0.677331805229187, | |
| "learning_rate": 4.2211055276381914e-05, | |
| "loss": 1.1411, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 1.296398891966759, | |
| "grad_norm": 0.6559180617332458, | |
| "learning_rate": 4.170854271356784e-05, | |
| "loss": 1.1108, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 1.3074792243767313, | |
| "grad_norm": 0.6477547287940979, | |
| "learning_rate": 4.120603015075377e-05, | |
| "loss": 1.079, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 1.3185595567867037, | |
| "grad_norm": 0.710292398929596, | |
| "learning_rate": 4.07035175879397e-05, | |
| "loss": 1.1465, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 1.3296398891966759, | |
| "grad_norm": 0.7174103260040283, | |
| "learning_rate": 4.020100502512563e-05, | |
| "loss": 1.2304, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.3407202216066483, | |
| "grad_norm": 0.6680272221565247, | |
| "learning_rate": 3.969849246231156e-05, | |
| "loss": 1.0935, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 1.3518005540166205, | |
| "grad_norm": 0.6414808630943298, | |
| "learning_rate": 3.919597989949749e-05, | |
| "loss": 1.0343, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 1.3628808864265927, | |
| "grad_norm": 0.7363560795783997, | |
| "learning_rate": 3.869346733668342e-05, | |
| "loss": 1.248, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 1.373961218836565, | |
| "grad_norm": 0.6751046776771545, | |
| "learning_rate": 3.8190954773869346e-05, | |
| "loss": 1.109, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 1.3850415512465375, | |
| "grad_norm": 0.6871734261512756, | |
| "learning_rate": 3.768844221105528e-05, | |
| "loss": 1.1761, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.3961218836565097, | |
| "grad_norm": 0.7280701398849487, | |
| "learning_rate": 3.7185929648241204e-05, | |
| "loss": 1.1457, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 1.4072022160664819, | |
| "grad_norm": 0.6830523014068604, | |
| "learning_rate": 3.668341708542714e-05, | |
| "loss": 0.985, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 1.4182825484764543, | |
| "grad_norm": 0.732204794883728, | |
| "learning_rate": 3.618090452261307e-05, | |
| "loss": 1.1521, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 1.4293628808864267, | |
| "grad_norm": 0.7047545909881592, | |
| "learning_rate": 3.5678391959798995e-05, | |
| "loss": 1.0978, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 1.440443213296399, | |
| "grad_norm": 0.7437470555305481, | |
| "learning_rate": 3.517587939698493e-05, | |
| "loss": 1.1808, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.451523545706371, | |
| "grad_norm": 0.6570298671722412, | |
| "learning_rate": 3.467336683417085e-05, | |
| "loss": 0.9725, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 1.4626038781163435, | |
| "grad_norm": 0.681265115737915, | |
| "learning_rate": 3.4170854271356785e-05, | |
| "loss": 1.1196, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 1.4736842105263157, | |
| "grad_norm": 0.7734697461128235, | |
| "learning_rate": 3.366834170854272e-05, | |
| "loss": 1.0839, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 1.4847645429362881, | |
| "grad_norm": 0.6945009231567383, | |
| "learning_rate": 3.3165829145728643e-05, | |
| "loss": 1.0336, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 1.4958448753462603, | |
| "grad_norm": 0.7448641657829285, | |
| "learning_rate": 3.2663316582914576e-05, | |
| "loss": 1.0924, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.5069252077562327, | |
| "grad_norm": 0.7725421190261841, | |
| "learning_rate": 3.21608040201005e-05, | |
| "loss": 1.1409, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 1.5180055401662051, | |
| "grad_norm": 0.8401060700416565, | |
| "learning_rate": 3.1658291457286434e-05, | |
| "loss": 1.1962, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 1.5290858725761773, | |
| "grad_norm": 0.708717405796051, | |
| "learning_rate": 3.1155778894472366e-05, | |
| "loss": 1.1007, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 1.5401662049861495, | |
| "grad_norm": 0.7954943776130676, | |
| "learning_rate": 3.065326633165829e-05, | |
| "loss": 1.1094, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 1.5512465373961217, | |
| "grad_norm": 0.7701108455657959, | |
| "learning_rate": 3.015075376884422e-05, | |
| "loss": 1.152, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.5623268698060941, | |
| "grad_norm": 0.7365975379943848, | |
| "learning_rate": 2.9648241206030153e-05, | |
| "loss": 1.0735, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 1.5734072022160666, | |
| "grad_norm": 0.6943490505218506, | |
| "learning_rate": 2.914572864321608e-05, | |
| "loss": 1.0759, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 1.5844875346260388, | |
| "grad_norm": 0.7694918513298035, | |
| "learning_rate": 2.8643216080402015e-05, | |
| "loss": 1.1438, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 1.595567867036011, | |
| "grad_norm": 0.6781268119812012, | |
| "learning_rate": 2.814070351758794e-05, | |
| "loss": 0.9776, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 1.6066481994459834, | |
| "grad_norm": 0.6973868012428284, | |
| "learning_rate": 2.763819095477387e-05, | |
| "loss": 1.0492, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.6177285318559558, | |
| "grad_norm": 0.770706295967102, | |
| "learning_rate": 2.7135678391959802e-05, | |
| "loss": 1.0905, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 1.628808864265928, | |
| "grad_norm": 0.7229887247085571, | |
| "learning_rate": 2.6633165829145728e-05, | |
| "loss": 1.1664, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 1.6398891966759002, | |
| "grad_norm": 0.762946367263794, | |
| "learning_rate": 2.613065326633166e-05, | |
| "loss": 1.1373, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 1.6509695290858726, | |
| "grad_norm": 0.7556053996086121, | |
| "learning_rate": 2.562814070351759e-05, | |
| "loss": 1.0613, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 1.662049861495845, | |
| "grad_norm": 0.7181993722915649, | |
| "learning_rate": 2.5125628140703518e-05, | |
| "loss": 1.0686, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.6731301939058172, | |
| "grad_norm": 0.7401473522186279, | |
| "learning_rate": 2.462311557788945e-05, | |
| "loss": 1.0952, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 1.6842105263157894, | |
| "grad_norm": 0.7067743539810181, | |
| "learning_rate": 2.4120603015075376e-05, | |
| "loss": 0.9935, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 1.6952908587257618, | |
| "grad_norm": 0.7622341513633728, | |
| "learning_rate": 2.361809045226131e-05, | |
| "loss": 1.1452, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 1.7063711911357342, | |
| "grad_norm": 0.7768684029579163, | |
| "learning_rate": 2.3115577889447238e-05, | |
| "loss": 1.1314, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 1.7174515235457064, | |
| "grad_norm": 0.7184272408485413, | |
| "learning_rate": 2.2613065326633167e-05, | |
| "loss": 1.049, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 1.7285318559556786, | |
| "grad_norm": 0.7589651942253113, | |
| "learning_rate": 2.21105527638191e-05, | |
| "loss": 1.0791, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 1.739612188365651, | |
| "grad_norm": 0.7551384568214417, | |
| "learning_rate": 2.1608040201005025e-05, | |
| "loss": 1.1264, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 1.7506925207756234, | |
| "grad_norm": 0.7738797664642334, | |
| "learning_rate": 2.1105527638190957e-05, | |
| "loss": 1.0658, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 1.7617728531855956, | |
| "grad_norm": 0.8027400374412537, | |
| "learning_rate": 2.0603015075376886e-05, | |
| "loss": 1.0526, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 1.7728531855955678, | |
| "grad_norm": 0.693533718585968, | |
| "learning_rate": 2.0100502512562815e-05, | |
| "loss": 1.0348, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.78393351800554, | |
| "grad_norm": 0.7291271686553955, | |
| "learning_rate": 1.9597989949748744e-05, | |
| "loss": 1.0686, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 1.7950138504155124, | |
| "grad_norm": 0.7514436841011047, | |
| "learning_rate": 1.9095477386934673e-05, | |
| "loss": 1.1501, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 1.8060941828254848, | |
| "grad_norm": 0.7121478915214539, | |
| "learning_rate": 1.8592964824120602e-05, | |
| "loss": 1.0153, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 1.817174515235457, | |
| "grad_norm": 0.7232415676116943, | |
| "learning_rate": 1.8090452261306535e-05, | |
| "loss": 1.034, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 1.8282548476454292, | |
| "grad_norm": 0.8162721991539001, | |
| "learning_rate": 1.7587939698492464e-05, | |
| "loss": 1.1254, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 1.8393351800554016, | |
| "grad_norm": 0.7215307354927063, | |
| "learning_rate": 1.7085427135678393e-05, | |
| "loss": 1.0638, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 1.850415512465374, | |
| "grad_norm": 0.7444539666175842, | |
| "learning_rate": 1.6582914572864322e-05, | |
| "loss": 1.0895, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 1.8614958448753463, | |
| "grad_norm": 0.787027895450592, | |
| "learning_rate": 1.608040201005025e-05, | |
| "loss": 0.9595, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 1.8725761772853184, | |
| "grad_norm": 0.7815292477607727, | |
| "learning_rate": 1.5577889447236183e-05, | |
| "loss": 1.1505, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 1.8836565096952909, | |
| "grad_norm": 0.7654871344566345, | |
| "learning_rate": 1.507537688442211e-05, | |
| "loss": 1.1074, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.8947368421052633, | |
| "grad_norm": 0.7809323072433472, | |
| "learning_rate": 1.457286432160804e-05, | |
| "loss": 1.0832, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 1.9058171745152355, | |
| "grad_norm": 0.8374095559120178, | |
| "learning_rate": 1.407035175879397e-05, | |
| "loss": 1.1028, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 1.9168975069252077, | |
| "grad_norm": 0.7976056933403015, | |
| "learning_rate": 1.3567839195979901e-05, | |
| "loss": 1.0124, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 1.92797783933518, | |
| "grad_norm": 0.8494656682014465, | |
| "learning_rate": 1.306532663316583e-05, | |
| "loss": 1.1487, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 1.9390581717451525, | |
| "grad_norm": 0.7171152234077454, | |
| "learning_rate": 1.2562814070351759e-05, | |
| "loss": 1.0215, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 1.9501385041551247, | |
| "grad_norm": 0.7844187617301941, | |
| "learning_rate": 1.2060301507537688e-05, | |
| "loss": 1.1526, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 1.9612188365650969, | |
| "grad_norm": 0.7948229908943176, | |
| "learning_rate": 1.1557788944723619e-05, | |
| "loss": 1.1165, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 1.9722991689750693, | |
| "grad_norm": 0.6978301405906677, | |
| "learning_rate": 1.105527638190955e-05, | |
| "loss": 0.9733, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 1.9833795013850417, | |
| "grad_norm": 0.8851218819618225, | |
| "learning_rate": 1.0552763819095479e-05, | |
| "loss": 1.167, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 1.994459833795014, | |
| "grad_norm": 0.7181246876716614, | |
| "learning_rate": 1.0050251256281408e-05, | |
| "loss": 1.0187, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 2.005540166204986, | |
| "grad_norm": 1.8050236701965332, | |
| "learning_rate": 9.547738693467337e-06, | |
| "loss": 1.97, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 2.0166204986149583, | |
| "grad_norm": 0.7432425618171692, | |
| "learning_rate": 9.045226130653267e-06, | |
| "loss": 1.0561, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 2.027700831024931, | |
| "grad_norm": 0.7358068227767944, | |
| "learning_rate": 8.542713567839196e-06, | |
| "loss": 0.9627, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 2.038781163434903, | |
| "grad_norm": 0.7742936015129089, | |
| "learning_rate": 8.040201005025125e-06, | |
| "loss": 1.143, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 2.0498614958448753, | |
| "grad_norm": 0.7715827226638794, | |
| "learning_rate": 7.537688442211055e-06, | |
| "loss": 1.0638, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 2.0609418282548475, | |
| "grad_norm": 0.7287455797195435, | |
| "learning_rate": 7.035175879396985e-06, | |
| "loss": 1.0142, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 2.07202216066482, | |
| "grad_norm": 0.7636083960533142, | |
| "learning_rate": 6.532663316582915e-06, | |
| "loss": 1.0196, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 2.0831024930747923, | |
| "grad_norm": 0.7313657999038696, | |
| "learning_rate": 6.030150753768844e-06, | |
| "loss": 1.0466, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 2.0941828254847645, | |
| "grad_norm": 0.7313005924224854, | |
| "learning_rate": 5.527638190954775e-06, | |
| "loss": 1.0009, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 2.1052631578947367, | |
| "grad_norm": 0.729709267616272, | |
| "learning_rate": 5.025125628140704e-06, | |
| "loss": 1.0601, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.1163434903047094, | |
| "grad_norm": 0.7446528077125549, | |
| "learning_rate": 4.522613065326634e-06, | |
| "loss": 0.9965, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 2.1274238227146816, | |
| "grad_norm": 0.7408772110939026, | |
| "learning_rate": 4.020100502512563e-06, | |
| "loss": 0.9824, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 2.1385041551246537, | |
| "grad_norm": 0.7374000549316406, | |
| "learning_rate": 3.5175879396984926e-06, | |
| "loss": 1.13, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 2.149584487534626, | |
| "grad_norm": 0.7011594772338867, | |
| "learning_rate": 3.015075376884422e-06, | |
| "loss": 0.9952, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 2.160664819944598, | |
| "grad_norm": 0.6886879801750183, | |
| "learning_rate": 2.512562814070352e-06, | |
| "loss": 0.9617, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 2.1717451523545708, | |
| "grad_norm": 0.8077765703201294, | |
| "learning_rate": 2.0100502512562813e-06, | |
| "loss": 1.0507, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 2.182825484764543, | |
| "grad_norm": 0.7857282757759094, | |
| "learning_rate": 1.507537688442211e-06, | |
| "loss": 1.1114, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 2.193905817174515, | |
| "grad_norm": 0.7135636806488037, | |
| "learning_rate": 1.0050251256281407e-06, | |
| "loss": 1.0691, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 2.2049861495844874, | |
| "grad_norm": 0.720593273639679, | |
| "learning_rate": 5.025125628140703e-07, | |
| "loss": 0.9977, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 2.21606648199446, | |
| "grad_norm": 0.7399870157241821, | |
| "learning_rate": 0.0, | |
| "loss": 1.0131, | |
| "step": 200 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 200, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.7777393737908224e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |