| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.236421725239617, | |
| "eval_steps": 500, | |
| "global_step": 1400, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.001597444089456869, | |
| "grad_norm": 128.6180419921875, | |
| "learning_rate": 0.0, | |
| "loss": 1.9747, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.003194888178913738, | |
| "grad_norm": 127.44181060791016, | |
| "learning_rate": 7.628243530057963e-08, | |
| "loss": 1.9176, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.004792332268370607, | |
| "grad_norm": 130.29476928710938, | |
| "learning_rate": 1.2090479941510652e-07, | |
| "loss": 1.9397, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.006389776357827476, | |
| "grad_norm": 128.3740234375, | |
| "learning_rate": 1.5256487060115926e-07, | |
| "loss": 1.9046, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.007987220447284345, | |
| "grad_norm": 136.06153869628906, | |
| "learning_rate": 1.7712232967084336e-07, | |
| "loss": 1.947, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.009584664536741214, | |
| "grad_norm": 134.7310333251953, | |
| "learning_rate": 1.9718723471568615e-07, | |
| "loss": 1.9161, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.011182108626198083, | |
| "grad_norm": 136.66058349609375, | |
| "learning_rate": 2.14151870207623e-07, | |
| "loss": 1.779, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.012779552715654952, | |
| "grad_norm": 142.6785888671875, | |
| "learning_rate": 2.288473059017389e-07, | |
| "loss": 1.7793, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.01437699680511182, | |
| "grad_norm": 139.76708984375, | |
| "learning_rate": 2.4180959883021303e-07, | |
| "loss": 1.7003, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.01597444089456869, | |
| "grad_norm": 147.7816162109375, | |
| "learning_rate": 2.53404764971423e-07, | |
| "loss": 1.742, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01757188498402556, | |
| "grad_norm": 159.0968780517578, | |
| "learning_rate": 2.638938686254791e-07, | |
| "loss": 1.6079, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.019169329073482427, | |
| "grad_norm": 138.03765869140625, | |
| "learning_rate": 2.7346967001626583e-07, | |
| "loss": 1.1859, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.020766773162939296, | |
| "grad_norm": 130.79318237304688, | |
| "learning_rate": 2.8227855338279304e-07, | |
| "loss": 1.1544, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.022364217252396165, | |
| "grad_norm": 128.29359436035156, | |
| "learning_rate": 2.904343055082026e-07, | |
| "loss": 1.0924, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.023961661341853034, | |
| "grad_norm": 118.21534729003906, | |
| "learning_rate": 2.9802712908594985e-07, | |
| "loss": 1.0143, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.025559105431309903, | |
| "grad_norm": 110.80689239501953, | |
| "learning_rate": 3.051297412023185e-07, | |
| "loss": 0.941, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.027156549520766772, | |
| "grad_norm": 112.3552474975586, | |
| "learning_rate": 3.1180161973120243e-07, | |
| "loss": 0.9826, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.02875399361022364, | |
| "grad_norm": 59.84565353393555, | |
| "learning_rate": 3.1809203413079266e-07, | |
| "loss": 0.6646, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.03035143769968051, | |
| "grad_norm": 61.63838195800781, | |
| "learning_rate": 3.2404225570581245e-07, | |
| "loss": 0.535, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.03194888178913738, | |
| "grad_norm": 57.68476486206055, | |
| "learning_rate": 3.2968720027200265e-07, | |
| "loss": 0.451, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.03354632587859425, | |
| "grad_norm": 58.70195388793945, | |
| "learning_rate": 3.350566696227295e-07, | |
| "loss": 0.4398, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.03514376996805112, | |
| "grad_norm": 60.67943572998047, | |
| "learning_rate": 3.401763039260588e-07, | |
| "loss": 0.3542, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.036741214057507986, | |
| "grad_norm": 54.30119705200195, | |
| "learning_rate": 3.450683222410826e-07, | |
| "loss": 0.3362, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.038338658146964855, | |
| "grad_norm": 27.211286544799805, | |
| "learning_rate": 3.4975210531684546e-07, | |
| "loss": 0.3206, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.039936102236421724, | |
| "grad_norm": 20.16046905517578, | |
| "learning_rate": 3.542446593416867e-07, | |
| "loss": 0.2747, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.04153354632587859, | |
| "grad_norm": 18.191621780395508, | |
| "learning_rate": 3.585609886833727e-07, | |
| "loss": 0.2485, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.04313099041533546, | |
| "grad_norm": 17.691118240356445, | |
| "learning_rate": 3.627143982453195e-07, | |
| "loss": 0.2113, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.04472843450479233, | |
| "grad_norm": 16.775856018066406, | |
| "learning_rate": 3.6671674080878223e-07, | |
| "loss": 0.2395, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.0463258785942492, | |
| "grad_norm": 11.414810180664062, | |
| "learning_rate": 3.7057862095226456e-07, | |
| "loss": 0.1754, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.04792332268370607, | |
| "grad_norm": 11.763614654541016, | |
| "learning_rate": 3.7430956438652954e-07, | |
| "loss": 0.171, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.04952076677316294, | |
| "grad_norm": 11.317436218261719, | |
| "learning_rate": 3.7791815951345716e-07, | |
| "loss": 0.1304, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.051118210862619806, | |
| "grad_norm": 11.866049766540527, | |
| "learning_rate": 3.814121765028982e-07, | |
| "loss": 0.1429, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.052715654952076675, | |
| "grad_norm": 7.949328899383545, | |
| "learning_rate": 3.8479866804058563e-07, | |
| "loss": 0.1008, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.054313099041533544, | |
| "grad_norm": 6.994161605834961, | |
| "learning_rate": 3.880840550317821e-07, | |
| "loss": 0.0818, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.05591054313099041, | |
| "grad_norm": 8.409013748168945, | |
| "learning_rate": 3.912741998784663e-07, | |
| "loss": 0.1117, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.05750798722044728, | |
| "grad_norm": 5.039752006530762, | |
| "learning_rate": 3.943744694313723e-07, | |
| "loss": 0.0769, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.05910543130990415, | |
| "grad_norm": 5.305444717407227, | |
| "learning_rate": 3.973897893149772e-07, | |
| "loss": 0.0547, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.06070287539936102, | |
| "grad_norm": 6.901062488555908, | |
| "learning_rate": 4.003246910063921e-07, | |
| "loss": 0.0888, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.06230031948881789, | |
| "grad_norm": 5.630770683288574, | |
| "learning_rate": 4.031833527978995e-07, | |
| "loss": 0.0664, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.06389776357827476, | |
| "grad_norm": 5.812962532043457, | |
| "learning_rate": 4.059696355725823e-07, | |
| "loss": 0.0643, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.06549520766773163, | |
| "grad_norm": 4.684011936187744, | |
| "learning_rate": 4.0868711416176974e-07, | |
| "loss": 0.0563, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.0670926517571885, | |
| "grad_norm": 5.2407050132751465, | |
| "learning_rate": 4.113391049233091e-07, | |
| "loss": 0.0546, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.06869009584664537, | |
| "grad_norm": 5.062054634094238, | |
| "learning_rate": 4.139286900743784e-07, | |
| "loss": 0.0488, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.07028753993610223, | |
| "grad_norm": 4.678827285766602, | |
| "learning_rate": 4.1645873922663837e-07, | |
| "loss": 0.039, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.07188498402555911, | |
| "grad_norm": 4.474348068237305, | |
| "learning_rate": 4.1893192850105637e-07, | |
| "loss": 0.0327, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.07348242811501597, | |
| "grad_norm": 5.580406665802002, | |
| "learning_rate": 4.213507575416622e-07, | |
| "loss": 0.0265, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.07507987220447285, | |
| "grad_norm": 4.145895957946777, | |
| "learning_rate": 4.2371756469942035e-07, | |
| "loss": 0.034, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.07667731629392971, | |
| "grad_norm": 2.864013671875, | |
| "learning_rate": 4.260345406174251e-07, | |
| "loss": 0.0195, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.07827476038338659, | |
| "grad_norm": 3.931077718734741, | |
| "learning_rate": 4.28303740415246e-07, | |
| "loss": 0.028, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.07987220447284345, | |
| "grad_norm": 3.2536861896514893, | |
| "learning_rate": 4.305270946422664e-07, | |
| "loss": 0.0167, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.08146964856230032, | |
| "grad_norm": 2.361112356185913, | |
| "learning_rate": 4.3270641914630895e-07, | |
| "loss": 0.0137, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.08306709265175719, | |
| "grad_norm": 3.177391529083252, | |
| "learning_rate": 4.3484342398395235e-07, | |
| "loss": 0.0168, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.08466453674121406, | |
| "grad_norm": 3.647589921951294, | |
| "learning_rate": 4.3693972148208397e-07, | |
| "loss": 0.0171, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.08626198083067092, | |
| "grad_norm": 3.1123926639556885, | |
| "learning_rate": 4.389968335458992e-07, | |
| "loss": 0.0096, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.0878594249201278, | |
| "grad_norm": 3.218924045562744, | |
| "learning_rate": 4.4101619829632255e-07, | |
| "loss": 0.0128, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.08945686900958466, | |
| "grad_norm": 3.422558307647705, | |
| "learning_rate": 4.429991761093619e-07, | |
| "loss": 0.0075, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.09105431309904154, | |
| "grad_norm": 4.4138288497924805, | |
| "learning_rate": 4.44947055120919e-07, | |
| "loss": 0.0106, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.0926517571884984, | |
| "grad_norm": 4.1317219734191895, | |
| "learning_rate": 4.4686105625284414e-07, | |
| "loss": 0.0147, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.09424920127795527, | |
| "grad_norm": 1.8419156074523926, | |
| "learning_rate": 4.4874233780934923e-07, | |
| "loss": 0.002, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.09584664536741214, | |
| "grad_norm": 5.771890640258789, | |
| "learning_rate": 4.5059199968710917e-07, | |
| "loss": 0.0132, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.09744408945686901, | |
| "grad_norm": 3.2066822052001953, | |
| "learning_rate": 4.5241108723737287e-07, | |
| "loss": 0.0121, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.09904153354632587, | |
| "grad_norm": 2.5449273586273193, | |
| "learning_rate": 4.5420059481403685e-07, | |
| "loss": 0.0112, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.10063897763578275, | |
| "grad_norm": 1.833320140838623, | |
| "learning_rate": 4.55961469037836e-07, | |
| "loss": 0.0057, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.10223642172523961, | |
| "grad_norm": 0.9052779674530029, | |
| "learning_rate": 4.576946118034778e-07, | |
| "loss": 0.0029, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.10383386581469649, | |
| "grad_norm": 1.7721703052520752, | |
| "learning_rate": 4.594008830536363e-07, | |
| "loss": 0.0073, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.10543130990415335, | |
| "grad_norm": 1.3743584156036377, | |
| "learning_rate": 4.6108110334116526e-07, | |
| "loss": 0.0042, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.10702875399361023, | |
| "grad_norm": 5.354936599731445, | |
| "learning_rate": 4.627360561986405e-07, | |
| "loss": 0.0067, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.10862619808306709, | |
| "grad_norm": 0.8433341979980469, | |
| "learning_rate": 4.6436649033236175e-07, | |
| "loss": 0.0036, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.11022364217252396, | |
| "grad_norm": 1.8066205978393555, | |
| "learning_rate": 4.6597312165618915e-07, | |
| "loss": 0.0065, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.11182108626198083, | |
| "grad_norm": 2.4936585426330566, | |
| "learning_rate": 4.6755663517904604e-07, | |
| "loss": 0.013, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.1134185303514377, | |
| "grad_norm": 1.3142043352127075, | |
| "learning_rate": 4.691176867585419e-07, | |
| "loss": 0.0106, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.11501597444089456, | |
| "grad_norm": 1.3923885822296143, | |
| "learning_rate": 4.70656904731952e-07, | |
| "loss": 0.0033, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.11661341853035144, | |
| "grad_norm": 1.4490530490875244, | |
| "learning_rate": 4.721748914347038e-07, | |
| "loss": 0.0079, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.1182108626198083, | |
| "grad_norm": 0.8172028660774231, | |
| "learning_rate": 4.736722246155569e-07, | |
| "loss": 0.0039, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.11980830670926518, | |
| "grad_norm": 1.627395510673523, | |
| "learning_rate": 4.751494587567932e-07, | |
| "loss": 0.0038, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.12140575079872204, | |
| "grad_norm": 1.3568391799926758, | |
| "learning_rate": 4.7660712630697176e-07, | |
| "loss": 0.0067, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.12300319488817892, | |
| "grad_norm": 1.5196893215179443, | |
| "learning_rate": 4.780457388331021e-07, | |
| "loss": 0.0064, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.12460063897763578, | |
| "grad_norm": 1.7342876195907593, | |
| "learning_rate": 4.794657880984791e-07, | |
| "loss": 0.0112, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.12619808306709265, | |
| "grad_norm": 1.1238106489181519, | |
| "learning_rate": 4.808677470718594e-07, | |
| "loss": 0.008, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.12779552715654952, | |
| "grad_norm": 0.6927244663238525, | |
| "learning_rate": 4.822520708731619e-07, | |
| "loss": 0.0028, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.12939297124600638, | |
| "grad_norm": 0.710464596748352, | |
| "learning_rate": 4.836191976604261e-07, | |
| "loss": 0.0046, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.13099041533546327, | |
| "grad_norm": 1.29885995388031, | |
| "learning_rate": 4.849695494623494e-07, | |
| "loss": 0.0028, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.13258785942492013, | |
| "grad_norm": 0.7128888964653015, | |
| "learning_rate": 4.863035329603659e-07, | |
| "loss": 0.0017, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.134185303514377, | |
| "grad_norm": 1.3970448970794678, | |
| "learning_rate": 4.876215402238887e-07, | |
| "loss": 0.007, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.13578274760383385, | |
| "grad_norm": 1.8840289115905762, | |
| "learning_rate": 4.889239494020458e-07, | |
| "loss": 0.0073, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.13738019169329074, | |
| "grad_norm": 0.9983999729156494, | |
| "learning_rate": 4.902111253749581e-07, | |
| "loss": 0.0008, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.1389776357827476, | |
| "grad_norm": 1.001712441444397, | |
| "learning_rate": 4.91483420367371e-07, | |
| "loss": 0.0045, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.14057507987220447, | |
| "grad_norm": 0.8022336959838867, | |
| "learning_rate": 4.927411745272181e-07, | |
| "loss": 0.005, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.14217252396166133, | |
| "grad_norm": 1.9497357606887817, | |
| "learning_rate": 4.939847164714948e-07, | |
| "loss": 0.0132, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.14376996805111822, | |
| "grad_norm": 0.4873954951763153, | |
| "learning_rate": 4.952143638016361e-07, | |
| "loss": 0.0036, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.14536741214057508, | |
| "grad_norm": 1.6238665580749512, | |
| "learning_rate": 4.96430423590416e-07, | |
| "loss": 0.0079, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.14696485623003194, | |
| "grad_norm": 2.2722995281219482, | |
| "learning_rate": 4.976331928422418e-07, | |
| "loss": 0.0112, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.1485623003194888, | |
| "grad_norm": 1.2474795579910278, | |
| "learning_rate": 4.988229589285637e-07, | |
| "loss": 0.0063, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.1501597444089457, | |
| "grad_norm": 1.0009785890579224, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0085, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.15175718849840256, | |
| "grad_norm": 0.6343518495559692, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.15335463258785942, | |
| "grad_norm": 0.4515174329280853, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.15495207667731628, | |
| "grad_norm": 0.7897706627845764, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0057, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.15654952076677317, | |
| "grad_norm": 1.2340161800384521, | |
| "learning_rate": 5e-07, | |
| "loss": 0.006, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.15814696485623003, | |
| "grad_norm": 0.5955281257629395, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.1597444089456869, | |
| "grad_norm": 1.140373706817627, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0094, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.16134185303514376, | |
| "grad_norm": 0.5656710267066956, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0049, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.16293929712460065, | |
| "grad_norm": 1.3329702615737915, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0102, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.1645367412140575, | |
| "grad_norm": 0.42556625604629517, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.16613418530351437, | |
| "grad_norm": 0.6244504451751709, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.16773162939297126, | |
| "grad_norm": 0.6920517086982727, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.16932907348242812, | |
| "grad_norm": 0.660217821598053, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.17092651757188498, | |
| "grad_norm": 1.0536694526672363, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.17252396166134185, | |
| "grad_norm": 0.862185537815094, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.17412140575079874, | |
| "grad_norm": 0.33221954107284546, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.1757188498402556, | |
| "grad_norm": 3.744462013244629, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0047, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.17731629392971246, | |
| "grad_norm": 1.6666382551193237, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0067, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.17891373801916932, | |
| "grad_norm": 0.3559662103652954, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.1805111821086262, | |
| "grad_norm": 0.3662293255329132, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.18210862619808307, | |
| "grad_norm": 0.49407243728637695, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.18370607028753994, | |
| "grad_norm": 1.0715949535369873, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0054, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.1853035143769968, | |
| "grad_norm": 0.618974506855011, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.1869009584664537, | |
| "grad_norm": 1.6139345169067383, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0089, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.18849840255591055, | |
| "grad_norm": 0.9241402745246887, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.1900958466453674, | |
| "grad_norm": 0.7571372389793396, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.19169329073482427, | |
| "grad_norm": 0.5674700736999512, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.19329073482428116, | |
| "grad_norm": 2.3434317111968994, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0113, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.19488817891373802, | |
| "grad_norm": 0.5285120010375977, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.1964856230031949, | |
| "grad_norm": 0.3353559076786041, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.19808306709265175, | |
| "grad_norm": 0.6090297698974609, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0052, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.19968051118210864, | |
| "grad_norm": 0.6662172079086304, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.2012779552715655, | |
| "grad_norm": 0.6109351515769958, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.20287539936102236, | |
| "grad_norm": 1.2161667346954346, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0077, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.20447284345047922, | |
| "grad_norm": 0.5375782251358032, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.20607028753993611, | |
| "grad_norm": 0.8255560398101807, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.20766773162939298, | |
| "grad_norm": 1.2899589538574219, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0073, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.20926517571884984, | |
| "grad_norm": 0.4597526490688324, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.2108626198083067, | |
| "grad_norm": 0.2904457151889801, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.2124600638977636, | |
| "grad_norm": 0.2778950035572052, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.21405750798722045, | |
| "grad_norm": 0.45779383182525635, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.21565495207667731, | |
| "grad_norm": 1.7134373188018799, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0073, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.21725239616613418, | |
| "grad_norm": 0.8295688033103943, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.21884984025559107, | |
| "grad_norm": 1.5231693983078003, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0087, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.22044728434504793, | |
| "grad_norm": 0.6921976804733276, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.2220447284345048, | |
| "grad_norm": 0.613254964351654, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.22364217252396165, | |
| "grad_norm": 0.9046704769134521, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.22523961661341854, | |
| "grad_norm": 0.827433705329895, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.2268370607028754, | |
| "grad_norm": 0.42438846826553345, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.22843450479233227, | |
| "grad_norm": 0.6912664771080017, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.23003194888178913, | |
| "grad_norm": 0.6359020471572876, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.23162939297124602, | |
| "grad_norm": 0.45392975211143494, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.23322683706070288, | |
| "grad_norm": 0.2925805151462555, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.23482428115015974, | |
| "grad_norm": 0.6216511130332947, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.2364217252396166, | |
| "grad_norm": 1.4257566928863525, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0049, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.2380191693290735, | |
| "grad_norm": 0.5238153338432312, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.23961661341853036, | |
| "grad_norm": 0.4455612301826477, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.24121405750798722, | |
| "grad_norm": 0.23266546428203583, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.24281150159744408, | |
| "grad_norm": 0.3799627125263214, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.24440894568690097, | |
| "grad_norm": 0.48497480154037476, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.24600638977635783, | |
| "grad_norm": 0.6147518157958984, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.2476038338658147, | |
| "grad_norm": 0.6456011533737183, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.24920127795527156, | |
| "grad_norm": 0.5384286642074585, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.2507987220447284, | |
| "grad_norm": 0.8868528008460999, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.2523961661341853, | |
| "grad_norm": 0.25091347098350525, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.2539936102236422, | |
| "grad_norm": 0.5789719820022583, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.25559105431309903, | |
| "grad_norm": 1.5586203336715698, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.2571884984025559, | |
| "grad_norm": 0.705531895160675, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.25878594249201275, | |
| "grad_norm": 0.7628393173217773, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.26038338658146964, | |
| "grad_norm": 0.8703659176826477, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.26198083067092653, | |
| "grad_norm": 0.3005010783672333, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.26357827476038337, | |
| "grad_norm": 0.19402694702148438, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.26517571884984026, | |
| "grad_norm": 2.271894931793213, | |
| "learning_rate": 5e-07, | |
| "loss": 0.006, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.26677316293929715, | |
| "grad_norm": 0.3555186092853546, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.268370607028754, | |
| "grad_norm": 1.9556201696395874, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.26996805111821087, | |
| "grad_norm": 1.7738531827926636, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0072, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.2715654952076677, | |
| "grad_norm": 2.2895989418029785, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0049, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.2731629392971246, | |
| "grad_norm": 1.7725859880447388, | |
| "learning_rate": 5e-07, | |
| "loss": 0.009, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.2747603833865815, | |
| "grad_norm": 0.5722166895866394, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.2763578274760383, | |
| "grad_norm": 1.063871145248413, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.2779552715654952, | |
| "grad_norm": 0.3389774560928345, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.2795527156549521, | |
| "grad_norm": 0.45647579431533813, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.28115015974440893, | |
| "grad_norm": 0.7398928999900818, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.2827476038338658, | |
| "grad_norm": 0.7850615382194519, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.28434504792332266, | |
| "grad_norm": 3.63692045211792, | |
| "learning_rate": 5e-07, | |
| "loss": 0.006, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.28594249201277955, | |
| "grad_norm": 0.44707363843917847, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.28753993610223644, | |
| "grad_norm": 0.9111068248748779, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.28913738019169327, | |
| "grad_norm": 0.6590073704719543, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.29073482428115016, | |
| "grad_norm": 0.5553420782089233, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.29233226837060705, | |
| "grad_norm": 0.47674331068992615, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.2939297124600639, | |
| "grad_norm": 0.29876044392585754, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.2955271565495208, | |
| "grad_norm": 0.6842983365058899, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.2971246006389776, | |
| "grad_norm": 0.4935712516307831, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.2987220447284345, | |
| "grad_norm": 1.2657763957977295, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0076, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.3003194888178914, | |
| "grad_norm": 0.48839613795280457, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.3019169329073482, | |
| "grad_norm": 1.063694715499878, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.3035143769968051, | |
| "grad_norm": 0.7980586290359497, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.305111821086262, | |
| "grad_norm": 0.45484256744384766, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.30670926517571884, | |
| "grad_norm": 1.688433051109314, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0056, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.3083067092651757, | |
| "grad_norm": 0.7312079071998596, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.30990415335463256, | |
| "grad_norm": 0.4801974594593048, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.31150159744408945, | |
| "grad_norm": 0.4754345118999481, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0058, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.31309904153354634, | |
| "grad_norm": 2.1157238483428955, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0091, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.3146964856230032, | |
| "grad_norm": 0.6343661546707153, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.31629392971246006, | |
| "grad_norm": 0.5099918842315674, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.31789137380191695, | |
| "grad_norm": 0.333951860666275, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.3194888178913738, | |
| "grad_norm": 0.7018607258796692, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0044, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.3210862619808307, | |
| "grad_norm": 0.6244291067123413, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.3226837060702875, | |
| "grad_norm": 0.8400171399116516, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.3242811501597444, | |
| "grad_norm": 1.1659806966781616, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0047, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.3258785942492013, | |
| "grad_norm": 0.9981520175933838, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0056, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.3274760383386581, | |
| "grad_norm": 0.32153695821762085, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.329073482428115, | |
| "grad_norm": 3.8695359230041504, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.3306709265175719, | |
| "grad_norm": 0.6529371738433838, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.33226837060702874, | |
| "grad_norm": 0.7385216951370239, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0048, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.33386581469648563, | |
| "grad_norm": 0.2696060836315155, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.3354632587859425, | |
| "grad_norm": 2.1431872844696045, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0104, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.33706070287539935, | |
| "grad_norm": 0.45158419013023376, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.33865814696485624, | |
| "grad_norm": 0.28861358761787415, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.3402555910543131, | |
| "grad_norm": 0.8536441326141357, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.34185303514376997, | |
| "grad_norm": 0.3299103081226349, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.34345047923322686, | |
| "grad_norm": 0.6404473185539246, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.3450479233226837, | |
| "grad_norm": 0.2772029638290405, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.3466453674121406, | |
| "grad_norm": 3.0106074810028076, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0051, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.34824281150159747, | |
| "grad_norm": 1.7244184017181396, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0066, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.3498402555910543, | |
| "grad_norm": 0.25184711813926697, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.3514376996805112, | |
| "grad_norm": 0.6831691861152649, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.35303514376996803, | |
| "grad_norm": 0.309230774641037, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.3546325878594249, | |
| "grad_norm": 0.5623671412467957, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.3562300319488818, | |
| "grad_norm": 0.31381672620773315, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.35782747603833864, | |
| "grad_norm": 0.6246241927146912, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.35942492012779553, | |
| "grad_norm": 0.9443689584732056, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0056, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.3610223642172524, | |
| "grad_norm": 0.8880484104156494, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.36261980830670926, | |
| "grad_norm": 0.9324328899383545, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.36421725239616615, | |
| "grad_norm": 0.7231195569038391, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.365814696485623, | |
| "grad_norm": 0.5298541188240051, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0047, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.36741214057507987, | |
| "grad_norm": 1.5357270240783691, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0054, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.36900958466453676, | |
| "grad_norm": 0.4374171495437622, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.3706070287539936, | |
| "grad_norm": 0.2250860631465912, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.3722044728434505, | |
| "grad_norm": 1.8229856491088867, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.3738019169329074, | |
| "grad_norm": 1.0502337217330933, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.3753993610223642, | |
| "grad_norm": 1.1811085939407349, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.3769968051118211, | |
| "grad_norm": 0.5202884674072266, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.37859424920127793, | |
| "grad_norm": 0.29556524753570557, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.3801916932907348, | |
| "grad_norm": 0.7253068685531616, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.3817891373801917, | |
| "grad_norm": 0.9315080642700195, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0044, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.38338658146964855, | |
| "grad_norm": 0.5284143090248108, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.38498402555910544, | |
| "grad_norm": 0.15872539579868317, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.3865814696485623, | |
| "grad_norm": 0.3570381999015808, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.38817891373801916, | |
| "grad_norm": 1.1983716487884521, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.38977635782747605, | |
| "grad_norm": 0.4511491656303406, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.3913738019169329, | |
| "grad_norm": 0.21462838351726532, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.3929712460063898, | |
| "grad_norm": 3.4422619342803955, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.39456869009584666, | |
| "grad_norm": 2.029860496520996, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.3961661341853035, | |
| "grad_norm": 0.4887297749519348, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.3977635782747604, | |
| "grad_norm": 3.5348706245422363, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.3993610223642173, | |
| "grad_norm": 0.12261940538883209, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0006, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.4009584664536741, | |
| "grad_norm": 1.2506657838821411, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0044, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.402555910543131, | |
| "grad_norm": 1.385932445526123, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0071, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.40415335463258784, | |
| "grad_norm": 0.6065950393676758, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.4057507987220447, | |
| "grad_norm": 0.3198873698711395, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.4073482428115016, | |
| "grad_norm": 0.21003501117229462, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.40894568690095845, | |
| "grad_norm": 0.70599764585495, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.41054313099041534, | |
| "grad_norm": 0.27071261405944824, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.41214057507987223, | |
| "grad_norm": 0.4544181525707245, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.41373801916932906, | |
| "grad_norm": 0.3489130437374115, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.41533546325878595, | |
| "grad_norm": 2.232909679412842, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.4169329073482428, | |
| "grad_norm": 0.24653400480747223, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.4185303514376997, | |
| "grad_norm": 0.70169597864151, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.42012779552715657, | |
| "grad_norm": 0.26590874791145325, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.4217252396166134, | |
| "grad_norm": 1.3357137441635132, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.4233226837060703, | |
| "grad_norm": 0.17348721623420715, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.4249201277955272, | |
| "grad_norm": 0.21971674263477325, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.426517571884984, | |
| "grad_norm": 0.3568759262561798, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.4281150159744409, | |
| "grad_norm": 0.22301608324050903, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.42971246006389774, | |
| "grad_norm": 0.2780276834964752, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.43130990415335463, | |
| "grad_norm": 0.399155855178833, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.4329073482428115, | |
| "grad_norm": 0.35904622077941895, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.43450479233226835, | |
| "grad_norm": 0.14762169122695923, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.43610223642172524, | |
| "grad_norm": 0.26977628469467163, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.43769968051118213, | |
| "grad_norm": 0.9291149377822876, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.43929712460063897, | |
| "grad_norm": 0.6625785827636719, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.44089456869009586, | |
| "grad_norm": 0.8231006860733032, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.4424920127795527, | |
| "grad_norm": 0.44783130288124084, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.4440894568690096, | |
| "grad_norm": 0.8134899139404297, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.44568690095846647, | |
| "grad_norm": 0.18782544136047363, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.4472843450479233, | |
| "grad_norm": 2.3220651149749756, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0057, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.4488817891373802, | |
| "grad_norm": 3.4821548461914062, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.4504792332268371, | |
| "grad_norm": 1.2853020429611206, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.4520766773162939, | |
| "grad_norm": 2.967637777328491, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.4536741214057508, | |
| "grad_norm": 1.1847063302993774, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0066, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.45527156549520764, | |
| "grad_norm": 0.6564436554908752, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.45686900958466453, | |
| "grad_norm": 1.7592343091964722, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.4584664536741214, | |
| "grad_norm": 0.21900290250778198, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.46006389776357826, | |
| "grad_norm": 0.31253165006637573, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.46166134185303515, | |
| "grad_norm": 0.7273770570755005, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.46325878594249204, | |
| "grad_norm": 0.9572885036468506, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.46485623003194887, | |
| "grad_norm": 4.609229564666748, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0105, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.46645367412140576, | |
| "grad_norm": 0.7994480729103088, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.4680511182108626, | |
| "grad_norm": 0.512992262840271, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.4696485623003195, | |
| "grad_norm": 0.7132886052131653, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.4712460063897764, | |
| "grad_norm": 0.4453645348548889, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.4728434504792332, | |
| "grad_norm": 0.32878679037094116, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.4744408945686901, | |
| "grad_norm": 0.4557294249534607, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.476038338658147, | |
| "grad_norm": 0.7930683493614197, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.4776357827476038, | |
| "grad_norm": 0.16142797470092773, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.4792332268370607, | |
| "grad_norm": 0.16838717460632324, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.48083067092651754, | |
| "grad_norm": 1.0744272470474243, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.48242811501597443, | |
| "grad_norm": 1.177935004234314, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.4840255591054313, | |
| "grad_norm": 0.33150434494018555, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.48562300319488816, | |
| "grad_norm": 0.14007702469825745, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.48722044728434505, | |
| "grad_norm": 0.23931963741779327, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.48881789137380194, | |
| "grad_norm": 1.721514344215393, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.4904153354632588, | |
| "grad_norm": 1.3192071914672852, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0095, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.49201277955271566, | |
| "grad_norm": 6.429077625274658, | |
| "learning_rate": 5e-07, | |
| "loss": 0.011, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.4936102236421725, | |
| "grad_norm": 0.919262707233429, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0066, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.4952076677316294, | |
| "grad_norm": 0.6547032594680786, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.4968051118210863, | |
| "grad_norm": 0.487663209438324, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.4984025559105431, | |
| "grad_norm": 0.33844229578971863, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.686505138874054, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.5015974440894568, | |
| "grad_norm": 0.2250794619321823, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.5031948881789138, | |
| "grad_norm": 0.40359318256378174, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.5047923322683706, | |
| "grad_norm": 0.3062373697757721, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.5063897763578274, | |
| "grad_norm": 0.3083527386188507, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.5079872204472844, | |
| "grad_norm": 0.4790252447128296, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.5095846645367412, | |
| "grad_norm": 0.4106828570365906, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.5111821086261981, | |
| "grad_norm": 0.5916593670845032, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.512779552715655, | |
| "grad_norm": 0.29794517159461975, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.5143769968051118, | |
| "grad_norm": 0.3094945251941681, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.5159744408945687, | |
| "grad_norm": 0.8037744760513306, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.5175718849840255, | |
| "grad_norm": 1.0918713808059692, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.5191693290734825, | |
| "grad_norm": 0.1457972228527069, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.5207667731629393, | |
| "grad_norm": 0.9666080474853516, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.5223642172523961, | |
| "grad_norm": 0.5684059262275696, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.5239616613418531, | |
| "grad_norm": 0.5098831057548523, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.5255591054313099, | |
| "grad_norm": 0.23925842344760895, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.5271565495207667, | |
| "grad_norm": 0.17939390242099762, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.5287539936102237, | |
| "grad_norm": 0.16178075969219208, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.5303514376996805, | |
| "grad_norm": 0.671220064163208, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.5319488817891374, | |
| "grad_norm": 0.816387414932251, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.5335463258785943, | |
| "grad_norm": 0.24421027302742004, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.5351437699680511, | |
| "grad_norm": 0.557515025138855, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.536741214057508, | |
| "grad_norm": 1.0557501316070557, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.5383386581469649, | |
| "grad_norm": 0.959538459777832, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.5399361022364217, | |
| "grad_norm": 0.3274182677268982, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.5415335463258786, | |
| "grad_norm": 0.6471118927001953, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.5431309904153354, | |
| "grad_norm": 0.38304996490478516, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.5447284345047924, | |
| "grad_norm": 1.0141305923461914, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.5463258785942492, | |
| "grad_norm": 0.9558109641075134, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.547923322683706, | |
| "grad_norm": 0.20122268795967102, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.549520766773163, | |
| "grad_norm": 0.4421844780445099, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.5511182108626198, | |
| "grad_norm": 0.4365698993206024, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.5527156549520766, | |
| "grad_norm": 1.578316330909729, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0053, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.5543130990415336, | |
| "grad_norm": 0.21290868520736694, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.5559105431309904, | |
| "grad_norm": 0.6181979179382324, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.5575079872204473, | |
| "grad_norm": 0.20121750235557556, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.5591054313099042, | |
| "grad_norm": 0.5325523018836975, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.560702875399361, | |
| "grad_norm": 0.6079647541046143, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0052, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.5623003194888179, | |
| "grad_norm": 0.2826935946941376, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.5638977635782748, | |
| "grad_norm": 0.15892595052719116, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.5654952076677316, | |
| "grad_norm": 0.3881969749927521, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.5670926517571885, | |
| "grad_norm": 0.3390671908855438, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.5686900958466453, | |
| "grad_norm": 0.2904258668422699, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.5702875399361023, | |
| "grad_norm": 0.46645137667655945, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.5718849840255591, | |
| "grad_norm": 0.6814522743225098, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.5734824281150159, | |
| "grad_norm": 0.9146270751953125, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0044, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.5750798722044729, | |
| "grad_norm": 0.21922031044960022, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.5766773162939297, | |
| "grad_norm": 0.939836859703064, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.5782747603833865, | |
| "grad_norm": 0.6049763560295105, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.5798722044728435, | |
| "grad_norm": 0.25285854935646057, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.5814696485623003, | |
| "grad_norm": 0.3022539019584656, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.5830670926517572, | |
| "grad_norm": 0.25078529119491577, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.5846645367412141, | |
| "grad_norm": 0.3507131040096283, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.5862619808306709, | |
| "grad_norm": 0.6294477581977844, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0058, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.5878594249201278, | |
| "grad_norm": 0.48274868726730347, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.5894568690095847, | |
| "grad_norm": 0.39043542742729187, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.5910543130990416, | |
| "grad_norm": 0.19649553298950195, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.5926517571884984, | |
| "grad_norm": 0.12885907292366028, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.5942492012779552, | |
| "grad_norm": 0.1488872468471527, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.5958466453674122, | |
| "grad_norm": 0.3566787540912628, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.597444089456869, | |
| "grad_norm": 0.986865758895874, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0047, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.5990415335463258, | |
| "grad_norm": 0.1579206883907318, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.6006389776357828, | |
| "grad_norm": 0.13746462762355804, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.6022364217252396, | |
| "grad_norm": 0.25995907187461853, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.6038338658146964, | |
| "grad_norm": 0.38791167736053467, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.6054313099041534, | |
| "grad_norm": 1.6591804027557373, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.6070287539936102, | |
| "grad_norm": 0.11244110763072968, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.6086261980830671, | |
| "grad_norm": 0.28090763092041016, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.610223642172524, | |
| "grad_norm": 0.8654282689094543, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.6118210862619808, | |
| "grad_norm": 0.3197280764579773, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.6134185303514377, | |
| "grad_norm": 0.13376931846141815, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.6150159744408946, | |
| "grad_norm": 0.3771628439426422, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.6166134185303515, | |
| "grad_norm": 3.0465352535247803, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0057, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.6182108626198083, | |
| "grad_norm": 0.6134064793586731, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.6198083067092651, | |
| "grad_norm": 1.984861969947815, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.6214057507987221, | |
| "grad_norm": 0.7715848684310913, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.6230031948881789, | |
| "grad_norm": 1.4465832710266113, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.6246006389776357, | |
| "grad_norm": 0.4437814950942993, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.6261980830670927, | |
| "grad_norm": 0.7803134322166443, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.6277955271565495, | |
| "grad_norm": 0.4495057761669159, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.6293929712460063, | |
| "grad_norm": 0.9235338568687439, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0062, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.6309904153354633, | |
| "grad_norm": 1.1879374980926514, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.6325878594249201, | |
| "grad_norm": 0.22155530750751495, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.634185303514377, | |
| "grad_norm": 0.6639124155044556, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.6357827476038339, | |
| "grad_norm": 0.24812071025371552, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.6373801916932907, | |
| "grad_norm": 0.24182091653347015, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.6389776357827476, | |
| "grad_norm": 0.25194039940834045, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.6405750798722045, | |
| "grad_norm": 1.6056197881698608, | |
| "learning_rate": 5e-07, | |
| "loss": 0.006, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.6421725239616614, | |
| "grad_norm": 3.634866952896118, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0048, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.6437699680511182, | |
| "grad_norm": 1.0231901407241821, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.645367412140575, | |
| "grad_norm": 0.27629023790359497, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.646964856230032, | |
| "grad_norm": 0.1949998289346695, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.6485623003194888, | |
| "grad_norm": 0.34071701765060425, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.6501597444089456, | |
| "grad_norm": 0.21671850979328156, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.6517571884984026, | |
| "grad_norm": 0.27686911821365356, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.6533546325878594, | |
| "grad_norm": 0.3393467366695404, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.6549520766773163, | |
| "grad_norm": 0.5887823700904846, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.6565495207667732, | |
| "grad_norm": 0.22454427182674408, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.65814696485623, | |
| "grad_norm": 0.13433396816253662, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.6597444089456869, | |
| "grad_norm": 0.14173242449760437, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.6613418530351438, | |
| "grad_norm": 0.30982744693756104, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.6629392971246006, | |
| "grad_norm": 0.47411805391311646, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.6645367412140575, | |
| "grad_norm": 0.19577734172344208, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.6661341853035144, | |
| "grad_norm": 0.6344949007034302, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.6677316293929713, | |
| "grad_norm": 0.45841068029403687, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.6693290734824281, | |
| "grad_norm": 0.4367104768753052, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.670926517571885, | |
| "grad_norm": 0.363915354013443, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.6725239616613419, | |
| "grad_norm": 0.18759390711784363, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.6741214057507987, | |
| "grad_norm": 0.15988245606422424, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.6757188498402555, | |
| "grad_norm": 0.1806800663471222, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.6773162939297125, | |
| "grad_norm": 0.21876202523708344, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.6789137380191693, | |
| "grad_norm": 0.13135991990566254, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0005, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.6805111821086262, | |
| "grad_norm": 1.3671129941940308, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.6821086261980831, | |
| "grad_norm": 0.1489122062921524, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.6837060702875399, | |
| "grad_norm": 0.17866690456867218, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.6853035143769968, | |
| "grad_norm": 0.655277669429779, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.6869009584664537, | |
| "grad_norm": 0.17315036058425903, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.6884984025559105, | |
| "grad_norm": 0.20632481575012207, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.6900958466453674, | |
| "grad_norm": 0.2345188409090042, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.6916932907348243, | |
| "grad_norm": 3.234980583190918, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.6932907348242812, | |
| "grad_norm": 0.5852277874946594, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.694888178913738, | |
| "grad_norm": 2.7778074741363525, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0047, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.6964856230031949, | |
| "grad_norm": 0.13274647295475006, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.6980830670926518, | |
| "grad_norm": 0.3051118850708008, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.6996805111821086, | |
| "grad_norm": 0.1823217123746872, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.7012779552715654, | |
| "grad_norm": 0.24809595942497253, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.7028753993610224, | |
| "grad_norm": 0.191731795668602, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.7044728434504792, | |
| "grad_norm": 0.32459452748298645, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.7060702875399361, | |
| "grad_norm": 0.46944916248321533, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.707667731629393, | |
| "grad_norm": 0.28335535526275635, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.7092651757188498, | |
| "grad_norm": 0.4613920748233795, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.7108626198083067, | |
| "grad_norm": 0.25062379240989685, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.7124600638977636, | |
| "grad_norm": 0.18467797338962555, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.7140575079872205, | |
| "grad_norm": 0.11972685158252716, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.7156549520766773, | |
| "grad_norm": 0.42692673206329346, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.7172523961661342, | |
| "grad_norm": 0.24261529743671417, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.7188498402555911, | |
| "grad_norm": 0.5654745101928711, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.7204472843450479, | |
| "grad_norm": 1.1742912530899048, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.7220447284345048, | |
| "grad_norm": 0.4380282163619995, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.7236421725239617, | |
| "grad_norm": 0.10869846493005753, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.7252396166134185, | |
| "grad_norm": 0.24093355238437653, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.7268370607028753, | |
| "grad_norm": 0.10973268747329712, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.7284345047923323, | |
| "grad_norm": 0.3785342276096344, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.7300319488817891, | |
| "grad_norm": 0.9914066791534424, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.731629392971246, | |
| "grad_norm": 0.4453500211238861, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.7332268370607029, | |
| "grad_norm": 0.2674676477909088, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.7348242811501597, | |
| "grad_norm": 0.9725813269615173, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.7364217252396166, | |
| "grad_norm": 0.14569059014320374, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.7380191693290735, | |
| "grad_norm": 0.25622859597206116, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.7396166134185304, | |
| "grad_norm": 0.33373990654945374, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.7412140575079872, | |
| "grad_norm": 0.28780269622802734, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.7428115015974441, | |
| "grad_norm": 0.3068907856941223, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.744408945686901, | |
| "grad_norm": 0.1721941977739334, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.7460063897763578, | |
| "grad_norm": 0.08544081449508667, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.7476038338658147, | |
| "grad_norm": 0.5271486043930054, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.7492012779552716, | |
| "grad_norm": 1.1334803104400635, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.7507987220447284, | |
| "grad_norm": 0.7498302459716797, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.7523961661341853, | |
| "grad_norm": 0.2573578357696533, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.7539936102236422, | |
| "grad_norm": 0.658050537109375, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.755591054313099, | |
| "grad_norm": 0.9222865104675293, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.7571884984025559, | |
| "grad_norm": 0.15267683565616608, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.7587859424920128, | |
| "grad_norm": 1.6186271905899048, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0042, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.7603833865814696, | |
| "grad_norm": 0.27657464146614075, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.7619808306709265, | |
| "grad_norm": 1.7262461185455322, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.7635782747603834, | |
| "grad_norm": 0.2438318431377411, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.7651757188498403, | |
| "grad_norm": 1.9847776889801025, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.7667731629392971, | |
| "grad_norm": 1.081231713294983, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.768370607028754, | |
| "grad_norm": 0.23922041058540344, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.7699680511182109, | |
| "grad_norm": 0.9737673401832581, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0065, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.7715654952076677, | |
| "grad_norm": 0.14736409485340118, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.7731629392971247, | |
| "grad_norm": 0.2849075198173523, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.7747603833865815, | |
| "grad_norm": 0.7391708493232727, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.7763578274760383, | |
| "grad_norm": 0.24760426580905914, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.7779552715654952, | |
| "grad_norm": 0.1888248175382614, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.7795527156549521, | |
| "grad_norm": 0.27008160948753357, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.7811501597444089, | |
| "grad_norm": 0.17157332599163055, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.7827476038338658, | |
| "grad_norm": 0.2907833755016327, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.7843450479233227, | |
| "grad_norm": 0.4900013506412506, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.7859424920127795, | |
| "grad_norm": 0.1433720886707306, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.7875399361022364, | |
| "grad_norm": 0.09725961834192276, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.7891373801916933, | |
| "grad_norm": 0.8984055519104004, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0048, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.7907348242811502, | |
| "grad_norm": 0.22446835041046143, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.792332268370607, | |
| "grad_norm": 0.2916003167629242, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.7939297124600639, | |
| "grad_norm": 0.42406606674194336, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.7955271565495208, | |
| "grad_norm": 0.8635863661766052, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.7971246006389776, | |
| "grad_norm": 0.4940405488014221, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.7987220447284346, | |
| "grad_norm": 0.6001198291778564, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.8003194888178914, | |
| "grad_norm": 0.4194888174533844, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.8019169329073482, | |
| "grad_norm": 0.1558961719274521, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.8035143769968051, | |
| "grad_norm": 0.1700737476348877, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.805111821086262, | |
| "grad_norm": 0.15276680886745453, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.8067092651757188, | |
| "grad_norm": 0.21490108966827393, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.8083067092651757, | |
| "grad_norm": 0.48414719104766846, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.8099041533546326, | |
| "grad_norm": 0.14232124388217926, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.8115015974440895, | |
| "grad_norm": 0.13643090426921844, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.8130990415335463, | |
| "grad_norm": 0.4378473162651062, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.8146964856230032, | |
| "grad_norm": 0.5300367474555969, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.8162939297124601, | |
| "grad_norm": 0.3394503593444824, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.8178913738019169, | |
| "grad_norm": 0.15926502645015717, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.8194888178913738, | |
| "grad_norm": 0.14627420902252197, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.8210862619808307, | |
| "grad_norm": 0.2294456660747528, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.8226837060702875, | |
| "grad_norm": 0.2830236554145813, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.8242811501597445, | |
| "grad_norm": 0.49337977170944214, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.8258785942492013, | |
| "grad_norm": 0.12382300198078156, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.8274760383386581, | |
| "grad_norm": 0.5072153210639954, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.829073482428115, | |
| "grad_norm": 0.6549058556556702, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0044, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.8306709265175719, | |
| "grad_norm": 0.21236664056777954, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.8322683706070287, | |
| "grad_norm": 0.6704789996147156, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.8338658146964856, | |
| "grad_norm": 0.1254369020462036, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.8354632587859425, | |
| "grad_norm": 2.3178699016571045, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.8370607028753994, | |
| "grad_norm": 3.5792298316955566, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.8386581469648562, | |
| "grad_norm": 0.7073982357978821, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.8402555910543131, | |
| "grad_norm": 0.5759369134902954, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.84185303514377, | |
| "grad_norm": 1.6669633388519287, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.8434504792332268, | |
| "grad_norm": 0.6380763649940491, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.8450479233226837, | |
| "grad_norm": 0.19805264472961426, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.8466453674121406, | |
| "grad_norm": 0.6069849133491516, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.8482428115015974, | |
| "grad_norm": 0.1687164306640625, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.8498402555910544, | |
| "grad_norm": 8.362351417541504, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0054, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.8514376996805112, | |
| "grad_norm": 0.2093094438314438, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.853035143769968, | |
| "grad_norm": 0.34365883469581604, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.854632587859425, | |
| "grad_norm": 0.12474914640188217, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.8562300319488818, | |
| "grad_norm": 0.510906994342804, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.8578274760383386, | |
| "grad_norm": 0.587397038936615, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.8594249201277955, | |
| "grad_norm": 0.844524621963501, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0052, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.8610223642172524, | |
| "grad_norm": 0.13176505267620087, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.8626198083067093, | |
| "grad_norm": 0.16751867532730103, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.8642172523961661, | |
| "grad_norm": 0.41158822178840637, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.865814696485623, | |
| "grad_norm": 0.12536601722240448, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.8674121405750799, | |
| "grad_norm": 0.08342156559228897, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.8690095846645367, | |
| "grad_norm": 1.5413963794708252, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.8706070287539937, | |
| "grad_norm": 0.24465155601501465, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.8722044728434505, | |
| "grad_norm": 0.1531795710325241, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.8738019169329073, | |
| "grad_norm": 0.1711510270833969, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.8753993610223643, | |
| "grad_norm": 0.10270589590072632, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.8769968051118211, | |
| "grad_norm": 0.6564199328422546, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.8785942492012779, | |
| "grad_norm": 0.15586939454078674, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.8801916932907349, | |
| "grad_norm": 0.11600361764431, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.8817891373801917, | |
| "grad_norm": 0.10974587500095367, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.8833865814696485, | |
| "grad_norm": 0.2567409574985504, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.8849840255591054, | |
| "grad_norm": 0.27306219935417175, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.8865814696485623, | |
| "grad_norm": 0.3767613470554352, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.8881789137380192, | |
| "grad_norm": 1.1817584037780762, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0057, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.889776357827476, | |
| "grad_norm": 0.3905037045478821, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.8913738019169329, | |
| "grad_norm": 0.23763906955718994, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0006, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.8929712460063898, | |
| "grad_norm": 0.3837161362171173, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.8945686900958466, | |
| "grad_norm": 0.6008383631706238, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.8961661341853036, | |
| "grad_norm": 0.1975802332162857, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.8977635782747604, | |
| "grad_norm": 0.15433986485004425, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.8993610223642172, | |
| "grad_norm": 0.4065248966217041, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.9009584664536742, | |
| "grad_norm": 0.25533992052078247, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.902555910543131, | |
| "grad_norm": 0.46665945649147034, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.9041533546325878, | |
| "grad_norm": 0.16390737891197205, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.9057507987220448, | |
| "grad_norm": 0.06191791966557503, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0005, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.9073482428115016, | |
| "grad_norm": 1.692974328994751, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.9089456869009584, | |
| "grad_norm": 0.21359291672706604, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.9105431309904153, | |
| "grad_norm": 0.09417448937892914, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.9121405750798722, | |
| "grad_norm": 0.22416774928569794, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.9137380191693291, | |
| "grad_norm": 0.15862461924552917, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.9153354632587859, | |
| "grad_norm": 0.325225293636322, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.9169329073482428, | |
| "grad_norm": 0.4235207140445709, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.9185303514376997, | |
| "grad_norm": 0.17048902809619904, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.9201277955271565, | |
| "grad_norm": 0.28779876232147217, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.9217252396166135, | |
| "grad_norm": 0.16058605909347534, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.9233226837060703, | |
| "grad_norm": 0.08717779070138931, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.9249201277955271, | |
| "grad_norm": 0.852260172367096, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.9265175718849841, | |
| "grad_norm": 0.34097403287887573, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.9281150159744409, | |
| "grad_norm": 0.3408906161785126, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.9297124600638977, | |
| "grad_norm": 0.4731082618236542, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.9313099041533547, | |
| "grad_norm": 0.49347537755966187, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.9329073482428115, | |
| "grad_norm": 0.17723320424556732, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.9345047923322684, | |
| "grad_norm": 0.16061758995056152, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.9361022364217252, | |
| "grad_norm": 0.11341333389282227, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.9376996805111821, | |
| "grad_norm": 0.12614861130714417, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.939297124600639, | |
| "grad_norm": 0.1919727921485901, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.9408945686900958, | |
| "grad_norm": 0.5518249273300171, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.9424920127795527, | |
| "grad_norm": 0.11890781670808792, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.9440894568690096, | |
| "grad_norm": 0.5765843391418457, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.9456869009584664, | |
| "grad_norm": 0.17196467518806458, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.9472843450479234, | |
| "grad_norm": 0.2513171136379242, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.9488817891373802, | |
| "grad_norm": 1.7397063970565796, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0108, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.950479233226837, | |
| "grad_norm": 0.2406977415084839, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.952076677316294, | |
| "grad_norm": 0.23791703581809998, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.9536741214057508, | |
| "grad_norm": 0.22479400038719177, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.9552715654952076, | |
| "grad_norm": 0.23164299130439758, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.9568690095846646, | |
| "grad_norm": 0.1600014865398407, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.9584664536741214, | |
| "grad_norm": 0.2001635879278183, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.9600638977635783, | |
| "grad_norm": 0.2946987450122833, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.9616613418530351, | |
| "grad_norm": 0.2080981582403183, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.963258785942492, | |
| "grad_norm": 0.6526065468788147, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.9648562300319489, | |
| "grad_norm": 0.3620392382144928, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.9664536741214057, | |
| "grad_norm": 0.21057778596878052, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.9680511182108626, | |
| "grad_norm": 0.11293834447860718, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.9696485623003195, | |
| "grad_norm": 1.2960716485977173, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.9712460063897763, | |
| "grad_norm": 0.1437765210866928, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.9728434504792333, | |
| "grad_norm": 0.6358578205108643, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0049, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.9744408945686901, | |
| "grad_norm": 0.14403465390205383, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.9760383386581469, | |
| "grad_norm": 0.4725677967071533, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.9776357827476039, | |
| "grad_norm": 0.8543768525123596, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.9792332268370607, | |
| "grad_norm": 0.27371206879615784, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.9808306709265175, | |
| "grad_norm": 0.23279932141304016, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.9824281150159745, | |
| "grad_norm": 0.8952144384384155, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.9840255591054313, | |
| "grad_norm": 0.11720918118953705, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.9856230031948882, | |
| "grad_norm": 0.9963851571083069, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.987220447284345, | |
| "grad_norm": 0.08971522003412247, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.9888178913738019, | |
| "grad_norm": 0.1214229092001915, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.9904153354632588, | |
| "grad_norm": 0.2924940288066864, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.9920127795527156, | |
| "grad_norm": 1.189788579940796, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.9936102236421726, | |
| "grad_norm": 0.5802843570709229, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0044, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.9952076677316294, | |
| "grad_norm": 0.11823207885026932, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.9968051118210862, | |
| "grad_norm": 0.16288286447525024, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.9984025559105432, | |
| "grad_norm": 0.13317349553108215, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.1848553717136383, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.0015974440894568, | |
| "grad_norm": 0.5851624608039856, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.0031948881789137, | |
| "grad_norm": 0.09685046225786209, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.0047923322683705, | |
| "grad_norm": 0.11342930048704147, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.0063897763578276, | |
| "grad_norm": 0.46440932154655457, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.0079872204472844, | |
| "grad_norm": 0.19794267416000366, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.0095846645367412, | |
| "grad_norm": 0.3643769919872284, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.011182108626198, | |
| "grad_norm": 0.08710943907499313, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.012779552715655, | |
| "grad_norm": 0.4915333688259125, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.0143769968051117, | |
| "grad_norm": 0.23040416836738586, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.0159744408945688, | |
| "grad_norm": 0.24800904095172882, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.0175718849840256, | |
| "grad_norm": 0.29919004440307617, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.0191693290734825, | |
| "grad_norm": 0.11212212592363358, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 1.0207667731629393, | |
| "grad_norm": 0.16430723667144775, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 1.0223642172523961, | |
| "grad_norm": 0.32202014327049255, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.023961661341853, | |
| "grad_norm": 0.3237399458885193, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 1.0255591054313098, | |
| "grad_norm": 0.1953740268945694, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 1.0271565495207668, | |
| "grad_norm": 0.12679028511047363, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 1.0287539936102237, | |
| "grad_norm": 0.10133963078260422, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 1.0303514376996805, | |
| "grad_norm": 0.0973275899887085, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.0319488817891374, | |
| "grad_norm": 0.5217294096946716, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 1.0335463258785942, | |
| "grad_norm": 0.15706409513950348, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 1.035143769968051, | |
| "grad_norm": 0.10915224999189377, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 1.036741214057508, | |
| "grad_norm": 0.4759727716445923, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 1.038338658146965, | |
| "grad_norm": 0.11390835046768188, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.0399361022364217, | |
| "grad_norm": 1.093390703201294, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 1.0415335463258786, | |
| "grad_norm": 0.07437872141599655, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 1.0431309904153354, | |
| "grad_norm": 0.21833601593971252, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 1.0447284345047922, | |
| "grad_norm": 0.21499225497245789, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 1.0463258785942493, | |
| "grad_norm": 0.12159962207078934, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.0479233226837061, | |
| "grad_norm": 0.3396832346916199, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.049520766773163, | |
| "grad_norm": 0.21546514332294464, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.0511182108626198, | |
| "grad_norm": 0.19969788193702698, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 1.0527156549520766, | |
| "grad_norm": 0.08988461643457413, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 1.0543130990415335, | |
| "grad_norm": 0.26418858766555786, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.0559105431309903, | |
| "grad_norm": 0.2093944102525711, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 1.0575079872204474, | |
| "grad_norm": 0.36498188972473145, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 1.0591054313099042, | |
| "grad_norm": 0.07945135235786438, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 1.060702875399361, | |
| "grad_norm": 0.12991008162498474, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 1.0623003194888179, | |
| "grad_norm": 0.2589464783668518, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.0638977635782747, | |
| "grad_norm": 0.10802029818296432, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 1.0654952076677315, | |
| "grad_norm": 0.4823663532733917, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.0670926517571886, | |
| "grad_norm": 0.44493567943573, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 1.0686900958466454, | |
| "grad_norm": 0.23464570939540863, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 1.0702875399361023, | |
| "grad_norm": 0.28489264845848083, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.071884984025559, | |
| "grad_norm": 0.2952616214752197, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 1.073482428115016, | |
| "grad_norm": 0.20882096886634827, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.0750798722044728, | |
| "grad_norm": 0.12549445033073425, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 1.0766773162939298, | |
| "grad_norm": 0.22115099430084229, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.0782747603833867, | |
| "grad_norm": 0.15710076689720154, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.0798722044728435, | |
| "grad_norm": 0.15250109136104584, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 1.0814696485623003, | |
| "grad_norm": 0.1936073750257492, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 1.0830670926517572, | |
| "grad_norm": 0.08779247850179672, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0006, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 1.084664536741214, | |
| "grad_norm": 0.1437022089958191, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 1.0862619808306708, | |
| "grad_norm": 0.2850930392742157, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.0878594249201279, | |
| "grad_norm": 0.3779478967189789, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0051, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 1.0894568690095847, | |
| "grad_norm": 0.20855233073234558, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 1.0910543130990416, | |
| "grad_norm": 0.11478544771671295, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 1.0926517571884984, | |
| "grad_norm": 0.3283560276031494, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 1.0942492012779552, | |
| "grad_norm": 0.35279855132102966, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.095846645367412, | |
| "grad_norm": 0.166551411151886, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 1.097444089456869, | |
| "grad_norm": 0.10785111039876938, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 1.099041533546326, | |
| "grad_norm": 0.1150643453001976, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.1006389776357828, | |
| "grad_norm": 0.10543517768383026, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 1.1022364217252396, | |
| "grad_norm": 0.12300563603639603, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.1038338658146964, | |
| "grad_norm": 0.15775950253009796, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 1.1054313099041533, | |
| "grad_norm": 0.052068062126636505, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 1.1070287539936103, | |
| "grad_norm": 0.1371004432439804, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 1.1086261980830672, | |
| "grad_norm": 0.11871246248483658, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 1.110223642172524, | |
| "grad_norm": 0.106499083340168, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.1118210862619808, | |
| "grad_norm": 0.05005495250225067, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 1.1134185303514377, | |
| "grad_norm": 0.6656812429428101, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 1.1150159744408945, | |
| "grad_norm": 0.14445379376411438, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 1.1166134185303513, | |
| "grad_norm": 0.268867164850235, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 1.1182108626198084, | |
| "grad_norm": 0.16358579695224762, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.1198083067092652, | |
| "grad_norm": 0.11881609261035919, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 1.121405750798722, | |
| "grad_norm": 0.10697460919618607, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 1.123003194888179, | |
| "grad_norm": 0.1314931958913803, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 1.1246006389776357, | |
| "grad_norm": 0.13953609764575958, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 1.1261980830670926, | |
| "grad_norm": 0.30124762654304504, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.1277955271565494, | |
| "grad_norm": 0.2047244906425476, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 1.1293929712460065, | |
| "grad_norm": 0.28825926780700684, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 1.1309904153354633, | |
| "grad_norm": 0.1030370369553566, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 1.1325878594249201, | |
| "grad_norm": 0.08416575193405151, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 1.134185303514377, | |
| "grad_norm": 0.6813434958457947, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.1357827476038338, | |
| "grad_norm": 0.7300125360488892, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 1.1373801916932909, | |
| "grad_norm": 0.21185138821601868, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 1.1389776357827477, | |
| "grad_norm": 0.3095090389251709, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 1.1405750798722045, | |
| "grad_norm": 0.13913819193840027, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 1.1421725239616614, | |
| "grad_norm": 0.2016637921333313, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.1437699680511182, | |
| "grad_norm": 0.12468226253986359, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 1.145367412140575, | |
| "grad_norm": 0.1510785073041916, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 1.1469648562300319, | |
| "grad_norm": 0.11004221439361572, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 1.148562300319489, | |
| "grad_norm": 0.12669166922569275, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 1.1501597444089458, | |
| "grad_norm": 0.11249173432588577, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.1517571884984026, | |
| "grad_norm": 0.21114954352378845, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 1.1533546325878594, | |
| "grad_norm": 0.3067387342453003, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 1.1549520766773163, | |
| "grad_norm": 0.11942733824253082, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 1.156549520766773, | |
| "grad_norm": 0.2498818337917328, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 1.15814696485623, | |
| "grad_norm": 0.589302122592926, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.159744408945687, | |
| "grad_norm": 0.11040360480546951, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 1.1613418530351438, | |
| "grad_norm": 0.1865539252758026, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 1.1629392971246006, | |
| "grad_norm": 0.09886899590492249, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 1.1645367412140575, | |
| "grad_norm": 0.10603392124176025, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 1.1661341853035143, | |
| "grad_norm": 0.17114412784576416, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.1677316293929714, | |
| "grad_norm": 0.11507996916770935, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 1.1693290734824282, | |
| "grad_norm": 0.10117539763450623, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 1.170926517571885, | |
| "grad_norm": 0.18606479465961456, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 1.1725239616613419, | |
| "grad_norm": 0.1562090963125229, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 1.1741214057507987, | |
| "grad_norm": 0.2511482834815979, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.1757188498402555, | |
| "grad_norm": 0.131776362657547, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 1.1773162939297124, | |
| "grad_norm": 0.2350272536277771, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 1.1789137380191694, | |
| "grad_norm": 0.5341865420341492, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 1.1805111821086263, | |
| "grad_norm": 0.18602542579174042, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 1.182108626198083, | |
| "grad_norm": 0.08645334094762802, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.18370607028754, | |
| "grad_norm": 0.1708468645811081, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 1.1853035143769968, | |
| "grad_norm": 0.11233728379011154, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 1.1869009584664536, | |
| "grad_norm": 0.20240649580955505, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 1.1884984025559104, | |
| "grad_norm": 0.14982427656650543, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 1.1900958466453675, | |
| "grad_norm": 0.13709725439548492, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.1916932907348243, | |
| "grad_norm": 0.18215787410736084, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 1.1932907348242812, | |
| "grad_norm": 0.12986968457698822, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 1.194888178913738, | |
| "grad_norm": 0.10665430873632431, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 1.1964856230031948, | |
| "grad_norm": 0.11768273264169693, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 1.1980830670926517, | |
| "grad_norm": 0.6945746541023254, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.1996805111821087, | |
| "grad_norm": 0.25220727920532227, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 1.2012779552715656, | |
| "grad_norm": 0.07079087197780609, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0005, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 1.2028753993610224, | |
| "grad_norm": 0.16933268308639526, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 1.2044728434504792, | |
| "grad_norm": 0.15669192373752594, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 1.206070287539936, | |
| "grad_norm": 6.026957035064697, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.207667731629393, | |
| "grad_norm": 0.5998969674110413, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 1.20926517571885, | |
| "grad_norm": 0.06143517419695854, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 1.2108626198083068, | |
| "grad_norm": 0.2319423258304596, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 1.2124600638977636, | |
| "grad_norm": 0.701130211353302, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0056, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 1.2140575079872205, | |
| "grad_norm": 0.3482913076877594, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.2156549520766773, | |
| "grad_norm": 0.21112686395645142, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 1.2172523961661341, | |
| "grad_norm": 5.14704704284668, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 1.218849840255591, | |
| "grad_norm": 0.41259634494781494, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 1.220447284345048, | |
| "grad_norm": 0.2882460951805115, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 1.2220447284345048, | |
| "grad_norm": 0.22736166417598724, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.2236421725239617, | |
| "grad_norm": 0.2859097719192505, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 1.2252396166134185, | |
| "grad_norm": 1.8469651937484741, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 1.2268370607028753, | |
| "grad_norm": 0.15771770477294922, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 1.2284345047923322, | |
| "grad_norm": 0.10112886875867844, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 1.230031948881789, | |
| "grad_norm": 0.7125353813171387, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.231629392971246, | |
| "grad_norm": 0.10188297927379608, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 1.233226837060703, | |
| "grad_norm": 0.9515169262886047, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 1.2348242811501597, | |
| "grad_norm": 0.06836796551942825, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 1.2364217252396166, | |
| "grad_norm": 0.13632221519947052, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 1.2380191693290734, | |
| "grad_norm": 0.572127103805542, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.2396166134185305, | |
| "grad_norm": 0.11243536323308945, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 1.2412140575079873, | |
| "grad_norm": 0.47509658336639404, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 1.2428115015974441, | |
| "grad_norm": 3.768028974533081, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 1.244408945686901, | |
| "grad_norm": 0.346608966588974, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 1.2460063897763578, | |
| "grad_norm": 0.23147884011268616, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.2476038338658146, | |
| "grad_norm": 0.14679104089736938, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 1.2492012779552715, | |
| "grad_norm": 0.23181918263435364, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 1.2507987220447285, | |
| "grad_norm": 0.16475149989128113, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 1.2523961661341854, | |
| "grad_norm": 0.08807860314846039, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 1.2539936102236422, | |
| "grad_norm": 0.16376349329948425, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.255591054313099, | |
| "grad_norm": 0.19524359703063965, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 1.2571884984025559, | |
| "grad_norm": 0.06623630225658417, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 1.2587859424920127, | |
| "grad_norm": 0.03929225727915764, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0003, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 1.2603833865814695, | |
| "grad_norm": 0.11669357866048813, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 1.2619808306709266, | |
| "grad_norm": 0.00831072498112917, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.2635782747603834, | |
| "grad_norm": 0.13926006853580475, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 1.2651757188498403, | |
| "grad_norm": 4.567328929901123, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 1.266773162939297, | |
| "grad_norm": 1.018916368484497, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0067, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 1.268370607028754, | |
| "grad_norm": 0.2978529632091522, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 1.269968051118211, | |
| "grad_norm": 0.5247751474380493, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.2715654952076676, | |
| "grad_norm": 0.06592018157243729, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0006, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 1.2731629392971247, | |
| "grad_norm": 0.1426411271095276, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 1.2747603833865815, | |
| "grad_norm": 0.14098455011844635, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 1.2763578274760383, | |
| "grad_norm": 0.11731888353824615, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 1.2779552715654952, | |
| "grad_norm": 0.8000310063362122, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0044, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.279552715654952, | |
| "grad_norm": 0.8283488154411316, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 1.281150159744409, | |
| "grad_norm": 0.6855292916297913, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 1.2827476038338659, | |
| "grad_norm": 0.12922891974449158, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 1.2843450479233227, | |
| "grad_norm": 0.8001663088798523, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0055, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 1.2859424920127795, | |
| "grad_norm": 0.09436812996864319, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.2875399361022364, | |
| "grad_norm": 0.3059661388397217, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 1.2891373801916932, | |
| "grad_norm": 0.22149628400802612, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 1.29073482428115, | |
| "grad_norm": 0.1476268321275711, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 1.292332268370607, | |
| "grad_norm": 0.11144911497831345, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 1.293929712460064, | |
| "grad_norm": 0.2204524427652359, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.2955271565495208, | |
| "grad_norm": 0.07029737532138824, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 1.2971246006389776, | |
| "grad_norm": 0.09939467906951904, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 1.2987220447284344, | |
| "grad_norm": 0.1769058108329773, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 1.3003194888178915, | |
| "grad_norm": 0.3592926561832428, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 1.3019169329073481, | |
| "grad_norm": 0.2275543361902237, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.3035143769968052, | |
| "grad_norm": 0.07945606857538223, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 1.305111821086262, | |
| "grad_norm": 0.08711861073970795, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 1.3067092651757188, | |
| "grad_norm": 0.28836479783058167, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 1.3083067092651757, | |
| "grad_norm": 0.08159520477056503, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 1.3099041533546325, | |
| "grad_norm": 0.07466080784797668, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.3115015974440896, | |
| "grad_norm": 0.5451674461364746, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 1.3130990415335464, | |
| "grad_norm": 0.14322105050086975, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 1.3146964856230032, | |
| "grad_norm": 0.12718692421913147, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 1.31629392971246, | |
| "grad_norm": 0.13329675793647766, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 1.317891373801917, | |
| "grad_norm": 0.22774100303649902, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.3194888178913737, | |
| "grad_norm": 0.8411527276039124, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 1.3210862619808306, | |
| "grad_norm": 0.18316972255706787, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 1.3226837060702876, | |
| "grad_norm": 0.07081547379493713, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 1.3242811501597445, | |
| "grad_norm": 0.3037576973438263, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 1.3258785942492013, | |
| "grad_norm": 0.10193005204200745, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.3274760383386581, | |
| "grad_norm": 0.10602962970733643, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 1.329073482428115, | |
| "grad_norm": 0.2908097505569458, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 1.330670926517572, | |
| "grad_norm": 0.3724125027656555, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 1.3322683706070286, | |
| "grad_norm": 0.11064999550580978, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 1.3338658146964857, | |
| "grad_norm": 0.228192538022995, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.3354632587859425, | |
| "grad_norm": 0.7090662717819214, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 1.3370607028753994, | |
| "grad_norm": 0.14330582320690155, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 1.3386581469648562, | |
| "grad_norm": 0.22641916573047638, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 1.340255591054313, | |
| "grad_norm": 0.21619798243045807, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 1.34185303514377, | |
| "grad_norm": 0.10568831115961075, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.343450479233227, | |
| "grad_norm": 0.1081472709774971, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 1.3450479233226837, | |
| "grad_norm": 0.10540477186441422, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 1.3466453674121406, | |
| "grad_norm": 0.11737114936113358, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 1.3482428115015974, | |
| "grad_norm": 0.17713451385498047, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 1.3498402555910542, | |
| "grad_norm": 0.2664765417575836, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.351437699680511, | |
| "grad_norm": 0.11203097552061081, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 1.3530351437699681, | |
| "grad_norm": 0.24115419387817383, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 1.354632587859425, | |
| "grad_norm": 0.13034223020076752, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 1.3562300319488818, | |
| "grad_norm": 1.2168869972229004, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 1.3578274760383386, | |
| "grad_norm": 0.04777536913752556, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0005, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.3594249201277955, | |
| "grad_norm": 0.14492247998714447, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 1.3610223642172525, | |
| "grad_norm": 0.11650210618972778, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 1.3626198083067091, | |
| "grad_norm": 0.13418716192245483, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 1.3642172523961662, | |
| "grad_norm": 1.2964794635772705, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 1.365814696485623, | |
| "grad_norm": 0.17427192628383636, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.3674121405750799, | |
| "grad_norm": 0.8252497911453247, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0049, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 1.3690095846645367, | |
| "grad_norm": 0.3598852753639221, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 1.3706070287539935, | |
| "grad_norm": 0.23161448538303375, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 1.3722044728434506, | |
| "grad_norm": 0.17750093340873718, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 1.3738019169329074, | |
| "grad_norm": 0.07043975591659546, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.3753993610223643, | |
| "grad_norm": 0.06552740186452866, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 1.376996805111821, | |
| "grad_norm": 0.167598158121109, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 1.378594249201278, | |
| "grad_norm": 0.8627744913101196, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 1.3801916932907348, | |
| "grad_norm": 0.23403562605381012, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 1.3817891373801916, | |
| "grad_norm": 1.3149168491363525, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.3833865814696487, | |
| "grad_norm": 0.16277176141738892, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 1.3849840255591055, | |
| "grad_norm": 0.1406593918800354, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 1.3865814696485623, | |
| "grad_norm": 0.24820701777935028, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 1.3881789137380192, | |
| "grad_norm": 0.10332539677619934, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 1.389776357827476, | |
| "grad_norm": 0.16070209443569183, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.3913738019169328, | |
| "grad_norm": 0.16537906229496002, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 1.3929712460063897, | |
| "grad_norm": 0.2067955881357193, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 1.3945686900958467, | |
| "grad_norm": 0.1595505326986313, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 1.3961661341853036, | |
| "grad_norm": 0.7071959972381592, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 1.3977635782747604, | |
| "grad_norm": 0.08548534661531448, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0005, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.3993610223642172, | |
| "grad_norm": 1.5511912107467651, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 1.400958466453674, | |
| "grad_norm": 0.622124969959259, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 1.4025559105431311, | |
| "grad_norm": 3.853699207305908, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0054, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 1.4041533546325877, | |
| "grad_norm": 0.07736141234636307, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 1.4057507987220448, | |
| "grad_norm": 0.2597537338733673, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.4073482428115016, | |
| "grad_norm": 0.09090615063905716, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 1.4089456869009584, | |
| "grad_norm": 0.4754495620727539, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 1.4105431309904153, | |
| "grad_norm": 0.4713440537452698, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 1.4121405750798721, | |
| "grad_norm": 0.3018096387386322, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 1.4137380191693292, | |
| "grad_norm": 0.21582446992397308, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 1.415335463258786, | |
| "grad_norm": 0.28057220578193665, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 1.4169329073482428, | |
| "grad_norm": 0.25230610370635986, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 1.4185303514376997, | |
| "grad_norm": 0.3251979947090149, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 1.4201277955271565, | |
| "grad_norm": 1.2839736938476562, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0045, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 1.4217252396166133, | |
| "grad_norm": 0.6521425247192383, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.4233226837060702, | |
| "grad_norm": 0.19354148209095, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 1.4249201277955272, | |
| "grad_norm": 0.26791778206825256, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 1.426517571884984, | |
| "grad_norm": 0.9978561997413635, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 1.428115015974441, | |
| "grad_norm": 1.0886672735214233, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 1.4297124600638977, | |
| "grad_norm": 0.1294880360364914, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 1.4313099041533546, | |
| "grad_norm": 0.1127467229962349, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 1.4329073482428116, | |
| "grad_norm": 0.14215292036533356, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 1.4345047923322682, | |
| "grad_norm": 0.11055054515600204, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 1.4361022364217253, | |
| "grad_norm": 0.8130955100059509, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 1.4376996805111821, | |
| "grad_norm": 0.6208978295326233, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.439297124600639, | |
| "grad_norm": 0.10077293962240219, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 1.4408945686900958, | |
| "grad_norm": 0.15302683413028717, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 1.4424920127795526, | |
| "grad_norm": 0.2711629867553711, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 1.4440894568690097, | |
| "grad_norm": 0.6740185022354126, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 1.4456869009584665, | |
| "grad_norm": 0.09207923710346222, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 1.4472843450479234, | |
| "grad_norm": 0.13100707530975342, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 1.4488817891373802, | |
| "grad_norm": 0.22404153645038605, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 1.450479233226837, | |
| "grad_norm": 0.19576634466648102, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 1.4520766773162939, | |
| "grad_norm": 0.3745575547218323, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 1.4536741214057507, | |
| "grad_norm": 0.3619185984134674, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0049, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.4552715654952078, | |
| "grad_norm": 0.11963102966547012, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 1.4568690095846646, | |
| "grad_norm": 0.6312240958213806, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 1.4584664536741214, | |
| "grad_norm": 0.44361862540245056, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 1.4600638977635783, | |
| "grad_norm": 0.40353521704673767, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 1.461661341853035, | |
| "grad_norm": 0.14119647443294525, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 1.4632587859424921, | |
| "grad_norm": 0.07256528735160828, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 1.4648562300319488, | |
| "grad_norm": 0.4794658422470093, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 1.4664536741214058, | |
| "grad_norm": 0.2629548907279968, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 1.4680511182108626, | |
| "grad_norm": 0.6776370406150818, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 1.4696485623003195, | |
| "grad_norm": 0.14644776284694672, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.4712460063897763, | |
| "grad_norm": 0.09278970211744308, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 1.4728434504792332, | |
| "grad_norm": 0.08292179554700851, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 1.4744408945686902, | |
| "grad_norm": 0.9622533321380615, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 1.476038338658147, | |
| "grad_norm": 0.17527416348457336, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 1.4776357827476039, | |
| "grad_norm": 0.07049129158258438, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 1.4792332268370607, | |
| "grad_norm": 0.07232940942049026, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 1.4808306709265175, | |
| "grad_norm": 0.06436185538768768, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 1.4824281150159744, | |
| "grad_norm": 0.24670840799808502, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 1.4840255591054312, | |
| "grad_norm": 0.7605392932891846, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 1.4856230031948883, | |
| "grad_norm": 0.06602557003498077, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.487220447284345, | |
| "grad_norm": 0.15429018437862396, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 1.488817891373802, | |
| "grad_norm": 0.155389204621315, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 1.4904153354632588, | |
| "grad_norm": 0.3137715458869934, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 1.4920127795527156, | |
| "grad_norm": 0.13918346166610718, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 1.4936102236421724, | |
| "grad_norm": 0.161741241812706, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 1.4952076677316293, | |
| "grad_norm": 0.2871551513671875, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 1.4968051118210863, | |
| "grad_norm": 0.3078411817550659, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 1.4984025559105432, | |
| "grad_norm": 0.1792358011007309, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 0.08885855227708817, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 1.5015974440894568, | |
| "grad_norm": 0.10388941317796707, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.5031948881789137, | |
| "grad_norm": 0.1879977583885193, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 1.5047923322683707, | |
| "grad_norm": 0.08109968155622482, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 1.5063897763578273, | |
| "grad_norm": 0.14684544503688812, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 1.5079872204472844, | |
| "grad_norm": 0.250420480966568, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 1.5095846645367412, | |
| "grad_norm": 0.0925571396946907, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 1.511182108626198, | |
| "grad_norm": 0.1837424635887146, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 1.5127795527156551, | |
| "grad_norm": 0.17538028955459595, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 1.5143769968051117, | |
| "grad_norm": 0.08512210845947266, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 1.5159744408945688, | |
| "grad_norm": 0.08973109722137451, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 1.5175718849840254, | |
| "grad_norm": 0.1241975948214531, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.5191693290734825, | |
| "grad_norm": 3.0702695846557617, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 1.5207667731629393, | |
| "grad_norm": 0.07310276478528976, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0006, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 1.5223642172523961, | |
| "grad_norm": 0.2920989692211151, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 1.5239616613418532, | |
| "grad_norm": 0.5082775950431824, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 1.5255591054313098, | |
| "grad_norm": 0.1549331098794937, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 1.5271565495207668, | |
| "grad_norm": 0.1703399121761322, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 1.5287539936102237, | |
| "grad_norm": 0.21957652270793915, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 1.5303514376996805, | |
| "grad_norm": 0.3131657838821411, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 1.5319488817891374, | |
| "grad_norm": 0.3629818260669708, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 1.5335463258785942, | |
| "grad_norm": 0.2978671193122864, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.5351437699680512, | |
| "grad_norm": 0.26085761189460754, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 1.5367412140575079, | |
| "grad_norm": 0.13067355751991272, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 1.538338658146965, | |
| "grad_norm": 1.0939770936965942, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0075, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 1.5399361022364217, | |
| "grad_norm": 0.4009106755256653, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 1.5415335463258786, | |
| "grad_norm": 0.69194495677948, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0071, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 1.5431309904153354, | |
| "grad_norm": 0.15541456639766693, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 1.5447284345047922, | |
| "grad_norm": 0.1687278002500534, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 1.5463258785942493, | |
| "grad_norm": 0.08637956529855728, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0004, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 1.547923322683706, | |
| "grad_norm": 0.49107223749160767, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 1.549520766773163, | |
| "grad_norm": 0.40780824422836304, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.5511182108626198, | |
| "grad_norm": 0.14083029329776764, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 1.5527156549520766, | |
| "grad_norm": 0.1509598046541214, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 1.5543130990415337, | |
| "grad_norm": 0.22124463319778442, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 1.5559105431309903, | |
| "grad_norm": 0.29571980237960815, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 1.5575079872204474, | |
| "grad_norm": 0.12444217503070831, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 1.5591054313099042, | |
| "grad_norm": 0.19605819880962372, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 1.560702875399361, | |
| "grad_norm": 0.36710816621780396, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 1.5623003194888179, | |
| "grad_norm": 0.4394298493862152, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 1.5638977635782747, | |
| "grad_norm": 3.3451671600341797, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 1.5654952076677318, | |
| "grad_norm": 0.0950264111161232, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0004, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.5670926517571884, | |
| "grad_norm": 0.5253363847732544, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 1.5686900958466454, | |
| "grad_norm": 0.08567991107702255, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 1.5702875399361023, | |
| "grad_norm": 0.21886466443538666, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 1.571884984025559, | |
| "grad_norm": 0.12182791531085968, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 1.573482428115016, | |
| "grad_norm": 0.10883834213018417, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 1.5750798722044728, | |
| "grad_norm": 0.2287680059671402, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 1.5766773162939298, | |
| "grad_norm": 0.7266914248466492, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 1.5782747603833864, | |
| "grad_norm": 0.11474244296550751, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 1.5798722044728435, | |
| "grad_norm": 0.23904937505722046, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 1.5814696485623003, | |
| "grad_norm": 0.055184707045555115, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.5830670926517572, | |
| "grad_norm": 0.22353950142860413, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 1.5846645367412142, | |
| "grad_norm": 0.19132664799690247, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 1.5862619808306708, | |
| "grad_norm": 0.2778601348400116, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 1.5878594249201279, | |
| "grad_norm": 0.05275161191821098, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 1.5894568690095847, | |
| "grad_norm": 0.2460019290447235, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 1.5910543130990416, | |
| "grad_norm": 0.3010371923446655, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 1.5926517571884984, | |
| "grad_norm": 0.10904088616371155, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 1.5942492012779552, | |
| "grad_norm": 0.05686507746577263, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 1.5958466453674123, | |
| "grad_norm": 0.20314623415470123, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 1.5974440894568689, | |
| "grad_norm": 1.588057279586792, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.599041533546326, | |
| "grad_norm": 0.48463931679725647, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 1.6006389776357828, | |
| "grad_norm": 0.6624436974525452, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 1.6022364217252396, | |
| "grad_norm": 0.14726407825946808, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 1.6038338658146964, | |
| "grad_norm": 0.7131458520889282, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 1.6054313099041533, | |
| "grad_norm": 0.46038198471069336, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 1.6070287539936103, | |
| "grad_norm": 0.2227628529071808, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 1.608626198083067, | |
| "grad_norm": 0.13341805338859558, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 1.610223642172524, | |
| "grad_norm": 0.07892493903636932, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 1.6118210862619808, | |
| "grad_norm": 0.12278909236192703, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 1.6134185303514377, | |
| "grad_norm": 0.5151563882827759, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.6150159744408947, | |
| "grad_norm": 0.23579691350460052, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 1.6166134185303513, | |
| "grad_norm": 0.05294935405254364, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 1.6182108626198084, | |
| "grad_norm": 0.12360315769910812, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 1.619808306709265, | |
| "grad_norm": 0.35527291893959045, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 1.621405750798722, | |
| "grad_norm": 0.25084754824638367, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 1.623003194888179, | |
| "grad_norm": 0.13500788807868958, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 1.6246006389776357, | |
| "grad_norm": 0.17942559719085693, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 1.6261980830670928, | |
| "grad_norm": 0.30760741233825684, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 1.6277955271565494, | |
| "grad_norm": 0.12108216434717178, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 1.6293929712460065, | |
| "grad_norm": 0.36486342549324036, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.6309904153354633, | |
| "grad_norm": 0.21493099629878998, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 1.6325878594249201, | |
| "grad_norm": 0.2136039435863495, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 1.634185303514377, | |
| "grad_norm": 0.19852106273174286, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 1.6357827476038338, | |
| "grad_norm": 0.07942666113376617, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 1.6373801916932909, | |
| "grad_norm": 0.14488880336284637, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 1.6389776357827475, | |
| "grad_norm": 0.07575450092554092, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 1.6405750798722045, | |
| "grad_norm": 0.18136127293109894, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 1.6421725239616614, | |
| "grad_norm": 0.09243427962064743, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 1.6437699680511182, | |
| "grad_norm": 0.15622855722904205, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 1.645367412140575, | |
| "grad_norm": 0.20699208974838257, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.6469648562300319, | |
| "grad_norm": 0.17613987624645233, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 1.648562300319489, | |
| "grad_norm": 0.10338564217090607, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 1.6501597444089455, | |
| "grad_norm": 0.07882916927337646, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 1.6517571884984026, | |
| "grad_norm": 0.14468686282634735, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 1.6533546325878594, | |
| "grad_norm": 0.33097293972969055, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0037, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 1.6549520766773163, | |
| "grad_norm": 0.46003833413124084, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 1.6565495207667733, | |
| "grad_norm": 0.38343682885169983, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 1.65814696485623, | |
| "grad_norm": 0.22435209155082703, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 1.659744408945687, | |
| "grad_norm": 0.10138783603906631, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 1.6613418530351438, | |
| "grad_norm": 0.42318132519721985, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.6629392971246006, | |
| "grad_norm": 0.10508925467729568, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 1.6645367412140575, | |
| "grad_norm": 0.3548614978790283, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 1.6661341853035143, | |
| "grad_norm": 0.22214291989803314, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 1.6677316293929714, | |
| "grad_norm": 0.23011279106140137, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 1.669329073482428, | |
| "grad_norm": 0.08691424876451492, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 1.670926517571885, | |
| "grad_norm": 0.6495136618614197, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 1.6725239616613419, | |
| "grad_norm": 0.20477008819580078, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 1.6741214057507987, | |
| "grad_norm": 0.3025433123111725, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 1.6757188498402555, | |
| "grad_norm": 0.1731894165277481, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 1.6773162939297124, | |
| "grad_norm": 0.06371040642261505, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.6789137380191694, | |
| "grad_norm": 0.14911451935768127, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 1.680511182108626, | |
| "grad_norm": 0.09349862486124039, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 1.682108626198083, | |
| "grad_norm": 0.21330159902572632, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 1.68370607028754, | |
| "grad_norm": 0.13860629498958588, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 1.6853035143769968, | |
| "grad_norm": 0.1365477740764618, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 1.6869009584664538, | |
| "grad_norm": 0.17503461241722107, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 1.6884984025559104, | |
| "grad_norm": 0.111385278403759, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 1.6900958466453675, | |
| "grad_norm": 0.06377507001161575, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 1.6916932907348243, | |
| "grad_norm": 0.0641062781214714, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 1.6932907348242812, | |
| "grad_norm": 0.09682592004537582, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.694888178913738, | |
| "grad_norm": 0.1440698206424713, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 1.6964856230031948, | |
| "grad_norm": 0.04743053764104843, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 1.6980830670926519, | |
| "grad_norm": 0.18199731409549713, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 1.6996805111821085, | |
| "grad_norm": 0.053704094141721725, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 1.7012779552715656, | |
| "grad_norm": 0.10925064235925674, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 1.7028753993610224, | |
| "grad_norm": 0.09475322812795639, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0005, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 1.7044728434504792, | |
| "grad_norm": 0.14424358308315277, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 1.706070287539936, | |
| "grad_norm": 0.07309069484472275, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 1.707667731629393, | |
| "grad_norm": 0.17733772099018097, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 1.70926517571885, | |
| "grad_norm": 0.9875695705413818, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0086, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.7108626198083066, | |
| "grad_norm": 0.408796101808548, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 1.7124600638977636, | |
| "grad_norm": 0.17582127451896667, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 1.7140575079872205, | |
| "grad_norm": 0.1303548812866211, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 1.7156549520766773, | |
| "grad_norm": 0.10945620387792587, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 1.7172523961661343, | |
| "grad_norm": 0.07630528509616852, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 1.718849840255591, | |
| "grad_norm": 0.07517647743225098, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 1.720447284345048, | |
| "grad_norm": 0.0669432058930397, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 1.7220447284345048, | |
| "grad_norm": 0.1877792626619339, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 1.7236421725239617, | |
| "grad_norm": 0.04336528107523918, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 1.7252396166134185, | |
| "grad_norm": 0.06632347404956818, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.7268370607028753, | |
| "grad_norm": 0.09459614008665085, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 1.7284345047923324, | |
| "grad_norm": 0.5509624481201172, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 1.730031948881789, | |
| "grad_norm": 0.2771284878253937, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 1.731629392971246, | |
| "grad_norm": 0.05486772954463959, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 1.733226837060703, | |
| "grad_norm": 0.2757831811904907, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 1.7348242811501597, | |
| "grad_norm": 0.04002672806382179, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 1.7364217252396166, | |
| "grad_norm": 0.06649244576692581, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0003, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 1.7380191693290734, | |
| "grad_norm": 0.11669522523880005, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 1.7396166134185305, | |
| "grad_norm": 0.12358154356479645, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1089 | |
| }, | |
| { | |
| "epoch": 1.741214057507987, | |
| "grad_norm": 0.0829407200217247, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.7428115015974441, | |
| "grad_norm": 0.09355901926755905, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 1.744408945686901, | |
| "grad_norm": 0.0956120640039444, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 1.7460063897763578, | |
| "grad_norm": 0.4703699052333832, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0055, | |
| "step": 1093 | |
| }, | |
| { | |
| "epoch": 1.7476038338658149, | |
| "grad_norm": 0.3924386203289032, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 1.7492012779552715, | |
| "grad_norm": 0.16191145777702332, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 1.7507987220447285, | |
| "grad_norm": 0.06510366499423981, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 1.7523961661341851, | |
| "grad_norm": 0.19621287286281586, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1097 | |
| }, | |
| { | |
| "epoch": 1.7539936102236422, | |
| "grad_norm": 0.1870349496603012, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 1.755591054313099, | |
| "grad_norm": 0.14898639917373657, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1099 | |
| }, | |
| { | |
| "epoch": 1.7571884984025559, | |
| "grad_norm": 0.15484674274921417, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.758785942492013, | |
| "grad_norm": 0.1420629471540451, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1101 | |
| }, | |
| { | |
| "epoch": 1.7603833865814695, | |
| "grad_norm": 0.12516824901103973, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 1.7619808306709266, | |
| "grad_norm": 0.09688828140497208, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1103 | |
| }, | |
| { | |
| "epoch": 1.7635782747603834, | |
| "grad_norm": 0.09412126988172531, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 1.7651757188498403, | |
| "grad_norm": 1.9070310592651367, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 1.766773162939297, | |
| "grad_norm": 0.16800744831562042, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 1.768370607028754, | |
| "grad_norm": 0.10442492365837097, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1107 | |
| }, | |
| { | |
| "epoch": 1.769968051118211, | |
| "grad_norm": 0.16451716423034668, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 1.7715654952076676, | |
| "grad_norm": 0.8484131097793579, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 1109 | |
| }, | |
| { | |
| "epoch": 1.7731629392971247, | |
| "grad_norm": 0.3258044719696045, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.7747603833865815, | |
| "grad_norm": 0.17745210230350494, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1111 | |
| }, | |
| { | |
| "epoch": 1.7763578274760383, | |
| "grad_norm": 0.07777510583400726, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 1.7779552715654952, | |
| "grad_norm": 0.08588631451129913, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1113 | |
| }, | |
| { | |
| "epoch": 1.779552715654952, | |
| "grad_norm": 0.6624420881271362, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 1.781150159744409, | |
| "grad_norm": 0.11576593667268753, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 1.7827476038338657, | |
| "grad_norm": 0.859341561794281, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 1.7843450479233227, | |
| "grad_norm": 0.10600177943706512, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1117 | |
| }, | |
| { | |
| "epoch": 1.7859424920127795, | |
| "grad_norm": 0.14881321787834167, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 1.7875399361022364, | |
| "grad_norm": 0.05584081634879112, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1119 | |
| }, | |
| { | |
| "epoch": 1.7891373801916934, | |
| "grad_norm": 0.06437966227531433, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.79073482428115, | |
| "grad_norm": 0.2528286874294281, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 1.792332268370607, | |
| "grad_norm": 0.18848662078380585, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 1.793929712460064, | |
| "grad_norm": 0.14006567001342773, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 1123 | |
| }, | |
| { | |
| "epoch": 1.7955271565495208, | |
| "grad_norm": 0.07045773416757584, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 1.7971246006389776, | |
| "grad_norm": 0.0998934954404831, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 1.7987220447284344, | |
| "grad_norm": 0.09063876420259476, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 1.8003194888178915, | |
| "grad_norm": 0.2747619152069092, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1127 | |
| }, | |
| { | |
| "epoch": 1.8019169329073481, | |
| "grad_norm": 0.0671829804778099, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 1.8035143769968052, | |
| "grad_norm": 0.23768068850040436, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1129 | |
| }, | |
| { | |
| "epoch": 1.805111821086262, | |
| "grad_norm": 0.14285339415073395, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.8067092651757188, | |
| "grad_norm": 3.77717661857605, | |
| "learning_rate": 5e-07, | |
| "loss": 0.005, | |
| "step": 1131 | |
| }, | |
| { | |
| "epoch": 1.8083067092651757, | |
| "grad_norm": 0.24507032334804535, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 1.8099041533546325, | |
| "grad_norm": 0.15166565775871277, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1133 | |
| }, | |
| { | |
| "epoch": 1.8115015974440896, | |
| "grad_norm": 0.1067359670996666, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 1.8130990415335462, | |
| "grad_norm": 0.18612955510616302, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 1.8146964856230032, | |
| "grad_norm": 0.17414437234401703, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 1.81629392971246, | |
| "grad_norm": 0.3064412474632263, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1137 | |
| }, | |
| { | |
| "epoch": 1.817891373801917, | |
| "grad_norm": 0.071932353079319, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 1.819488817891374, | |
| "grad_norm": 0.1119665578007698, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1139 | |
| }, | |
| { | |
| "epoch": 1.8210862619808306, | |
| "grad_norm": 0.20450662076473236, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.8226837060702876, | |
| "grad_norm": 0.13636766374111176, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1141 | |
| }, | |
| { | |
| "epoch": 1.8242811501597445, | |
| "grad_norm": 0.07295648753643036, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 1.8258785942492013, | |
| "grad_norm": 0.17274846136569977, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 1143 | |
| }, | |
| { | |
| "epoch": 1.8274760383386581, | |
| "grad_norm": 0.14178737998008728, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 1.829073482428115, | |
| "grad_norm": 0.2096203863620758, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 1.830670926517572, | |
| "grad_norm": 0.09562284499406815, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 1.8322683706070286, | |
| "grad_norm": 0.10589710623025894, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1147 | |
| }, | |
| { | |
| "epoch": 1.8338658146964857, | |
| "grad_norm": 0.2623519003391266, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 1.8354632587859425, | |
| "grad_norm": 0.06434428691864014, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1149 | |
| }, | |
| { | |
| "epoch": 1.8370607028753994, | |
| "grad_norm": 0.31623294949531555, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.8386581469648562, | |
| "grad_norm": 0.11535608768463135, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1151 | |
| }, | |
| { | |
| "epoch": 1.840255591054313, | |
| "grad_norm": 0.25482088327407837, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0006, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 1.84185303514377, | |
| "grad_norm": 0.1523568332195282, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1153 | |
| }, | |
| { | |
| "epoch": 1.8434504792332267, | |
| "grad_norm": 0.09644993394613266, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 1.8450479233226837, | |
| "grad_norm": 0.30026814341545105, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 1.8466453674121406, | |
| "grad_norm": 5.034646511077881, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 1.8482428115015974, | |
| "grad_norm": 0.1875925064086914, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1157 | |
| }, | |
| { | |
| "epoch": 1.8498402555910545, | |
| "grad_norm": 0.15650227665901184, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 1.851437699680511, | |
| "grad_norm": 0.17901848256587982, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1159 | |
| }, | |
| { | |
| "epoch": 1.8530351437699681, | |
| "grad_norm": 0.27782970666885376, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.854632587859425, | |
| "grad_norm": 0.11091984808444977, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1161 | |
| }, | |
| { | |
| "epoch": 1.8562300319488818, | |
| "grad_norm": 0.3836122453212738, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 1.8578274760383386, | |
| "grad_norm": 2.26542067527771, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 1163 | |
| }, | |
| { | |
| "epoch": 1.8594249201277955, | |
| "grad_norm": 0.8616245985031128, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0051, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 1.8610223642172525, | |
| "grad_norm": 0.1418013572692871, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 1.8626198083067091, | |
| "grad_norm": 0.20582883059978485, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 1.8642172523961662, | |
| "grad_norm": 0.05279695987701416, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0006, | |
| "step": 1167 | |
| }, | |
| { | |
| "epoch": 1.865814696485623, | |
| "grad_norm": 0.30152246356010437, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 1.8674121405750799, | |
| "grad_norm": 1.215949296951294, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0042, | |
| "step": 1169 | |
| }, | |
| { | |
| "epoch": 1.8690095846645367, | |
| "grad_norm": 0.1583871841430664, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.8706070287539935, | |
| "grad_norm": 0.18552114069461823, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1171 | |
| }, | |
| { | |
| "epoch": 1.8722044728434506, | |
| "grad_norm": 0.25958728790283203, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 1.8738019169329072, | |
| "grad_norm": 0.3198787271976471, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1173 | |
| }, | |
| { | |
| "epoch": 1.8753993610223643, | |
| "grad_norm": 0.4293941557407379, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 1.876996805111821, | |
| "grad_norm": 0.12936322391033173, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 1.878594249201278, | |
| "grad_norm": 0.26641571521759033, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 1.880191693290735, | |
| "grad_norm": 1.098074197769165, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0036, | |
| "step": 1177 | |
| }, | |
| { | |
| "epoch": 1.8817891373801916, | |
| "grad_norm": 0.11431043595075607, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 1.8833865814696487, | |
| "grad_norm": 0.37071993947029114, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1179 | |
| }, | |
| { | |
| "epoch": 1.8849840255591053, | |
| "grad_norm": 0.08628468960523605, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.8865814696485623, | |
| "grad_norm": 0.14398378133773804, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1181 | |
| }, | |
| { | |
| "epoch": 1.8881789137380192, | |
| "grad_norm": 0.35993340611457825, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 1.889776357827476, | |
| "grad_norm": 0.1559559851884842, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1183 | |
| }, | |
| { | |
| "epoch": 1.891373801916933, | |
| "grad_norm": 0.2564754784107208, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 1.8929712460063897, | |
| "grad_norm": 0.11279148608446121, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0005, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 1.8945686900958467, | |
| "grad_norm": 0.08808083087205887, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 1.8961661341853036, | |
| "grad_norm": 0.10382703691720963, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0004, | |
| "step": 1187 | |
| }, | |
| { | |
| "epoch": 1.8977635782747604, | |
| "grad_norm": 0.15366709232330322, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 1.8993610223642172, | |
| "grad_norm": 0.10865357518196106, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1189 | |
| }, | |
| { | |
| "epoch": 1.900958466453674, | |
| "grad_norm": 0.24182426929473877, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.9025559105431311, | |
| "grad_norm": 0.4804140627384186, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0048, | |
| "step": 1191 | |
| }, | |
| { | |
| "epoch": 1.9041533546325877, | |
| "grad_norm": 0.23404286801815033, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 1.9057507987220448, | |
| "grad_norm": 0.36086928844451904, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1193 | |
| }, | |
| { | |
| "epoch": 1.9073482428115016, | |
| "grad_norm": 0.5354902744293213, | |
| "learning_rate": 5e-07, | |
| "loss": 0.004, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 1.9089456869009584, | |
| "grad_norm": 0.11621209979057312, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 1.9105431309904153, | |
| "grad_norm": 0.1943303495645523, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 1.9121405750798721, | |
| "grad_norm": 0.2628275454044342, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1197 | |
| }, | |
| { | |
| "epoch": 1.9137380191693292, | |
| "grad_norm": 0.20775623619556427, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 1.9153354632587858, | |
| "grad_norm": 0.12529809772968292, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1199 | |
| }, | |
| { | |
| "epoch": 1.9169329073482428, | |
| "grad_norm": 0.09771095961332321, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.9185303514376997, | |
| "grad_norm": 0.10013385117053986, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1201 | |
| }, | |
| { | |
| "epoch": 1.9201277955271565, | |
| "grad_norm": 0.29045844078063965, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1202 | |
| }, | |
| { | |
| "epoch": 1.9217252396166136, | |
| "grad_norm": 0.15383721888065338, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1203 | |
| }, | |
| { | |
| "epoch": 1.9233226837060702, | |
| "grad_norm": 0.6897266507148743, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 1.9249201277955272, | |
| "grad_norm": 0.15652583539485931, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 1.926517571884984, | |
| "grad_norm": 0.0853152722120285, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1206 | |
| }, | |
| { | |
| "epoch": 1.928115015974441, | |
| "grad_norm": 0.11881982535123825, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1207 | |
| }, | |
| { | |
| "epoch": 1.9297124600638977, | |
| "grad_norm": 0.10427961498498917, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 1.9313099041533546, | |
| "grad_norm": 0.1068776324391365, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1209 | |
| }, | |
| { | |
| "epoch": 1.9329073482428116, | |
| "grad_norm": 0.13248351216316223, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.9345047923322682, | |
| "grad_norm": 0.25468289852142334, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1211 | |
| }, | |
| { | |
| "epoch": 1.9361022364217253, | |
| "grad_norm": 0.16791756451129913, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 1.9376996805111821, | |
| "grad_norm": 0.13099579513072968, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1213 | |
| }, | |
| { | |
| "epoch": 1.939297124600639, | |
| "grad_norm": 0.26930350065231323, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 1214 | |
| }, | |
| { | |
| "epoch": 1.9408945686900958, | |
| "grad_norm": 2.077728271484375, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 1.9424920127795526, | |
| "grad_norm": 0.31897902488708496, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 1.9440894568690097, | |
| "grad_norm": 0.07247374951839447, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1217 | |
| }, | |
| { | |
| "epoch": 1.9456869009584663, | |
| "grad_norm": 0.11995717883110046, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 1.9472843450479234, | |
| "grad_norm": 0.2621629536151886, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1219 | |
| }, | |
| { | |
| "epoch": 1.9488817891373802, | |
| "grad_norm": 0.1726490557193756, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.950479233226837, | |
| "grad_norm": 0.16272880136966705, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1221 | |
| }, | |
| { | |
| "epoch": 1.952076677316294, | |
| "grad_norm": 0.10036841779947281, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 1.9536741214057507, | |
| "grad_norm": 0.7988063097000122, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0046, | |
| "step": 1223 | |
| }, | |
| { | |
| "epoch": 1.9552715654952078, | |
| "grad_norm": 0.2782542407512665, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 1.9568690095846646, | |
| "grad_norm": 0.2055627703666687, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 1.9584664536741214, | |
| "grad_norm": 0.19908927381038666, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1226 | |
| }, | |
| { | |
| "epoch": 1.9600638977635783, | |
| "grad_norm": 0.21738676726818085, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1227 | |
| }, | |
| { | |
| "epoch": 1.961661341853035, | |
| "grad_norm": 0.16997866332530975, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 1.9632587859424921, | |
| "grad_norm": 0.10853756219148636, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1229 | |
| }, | |
| { | |
| "epoch": 1.9648562300319488, | |
| "grad_norm": 0.2967356741428375, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.9664536741214058, | |
| "grad_norm": 0.24150174856185913, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 1231 | |
| }, | |
| { | |
| "epoch": 1.9680511182108626, | |
| "grad_norm": 0.09346190094947815, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 1.9696485623003195, | |
| "grad_norm": 0.16779378056526184, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1233 | |
| }, | |
| { | |
| "epoch": 1.9712460063897763, | |
| "grad_norm": 0.1114976555109024, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1234 | |
| }, | |
| { | |
| "epoch": 1.9728434504792332, | |
| "grad_norm": 0.45826414227485657, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 1.9744408945686902, | |
| "grad_norm": 0.4743911921977997, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0034, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 1.9760383386581468, | |
| "grad_norm": 0.144659623503685, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1237 | |
| }, | |
| { | |
| "epoch": 1.9776357827476039, | |
| "grad_norm": 0.9392029643058777, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1238 | |
| }, | |
| { | |
| "epoch": 1.9792332268370607, | |
| "grad_norm": 0.5866590738296509, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1239 | |
| }, | |
| { | |
| "epoch": 1.9808306709265175, | |
| "grad_norm": 0.2669076919555664, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.9824281150159746, | |
| "grad_norm": 0.34604692459106445, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1241 | |
| }, | |
| { | |
| "epoch": 1.9840255591054312, | |
| "grad_norm": 0.27668237686157227, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 1.9856230031948883, | |
| "grad_norm": 0.1560208648443222, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1243 | |
| }, | |
| { | |
| "epoch": 1.9872204472843449, | |
| "grad_norm": 0.35375985503196716, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 1.988817891373802, | |
| "grad_norm": 0.134053036570549, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 1.9904153354632588, | |
| "grad_norm": 0.10250476002693176, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1246 | |
| }, | |
| { | |
| "epoch": 1.9920127795527156, | |
| "grad_norm": 0.13041843473911285, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1247 | |
| }, | |
| { | |
| "epoch": 1.9936102236421727, | |
| "grad_norm": 0.062465403228998184, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 1.9952076677316293, | |
| "grad_norm": 0.10093759000301361, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1249 | |
| }, | |
| { | |
| "epoch": 1.9968051118210863, | |
| "grad_norm": 0.08476297557353973, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.9984025559105432, | |
| "grad_norm": 0.22871960699558258, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1251 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.19058428704738617, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 2.001597444089457, | |
| "grad_norm": 0.2544306218624115, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1253 | |
| }, | |
| { | |
| "epoch": 2.0031948881789137, | |
| "grad_norm": 0.14264468848705292, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 1254 | |
| }, | |
| { | |
| "epoch": 2.0047923322683707, | |
| "grad_norm": 0.18681064248085022, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 2.0063897763578273, | |
| "grad_norm": 0.17501874268054962, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 2.0079872204472844, | |
| "grad_norm": 0.1236346885561943, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1257 | |
| }, | |
| { | |
| "epoch": 2.009584664536741, | |
| "grad_norm": 0.08623294532299042, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 1258 | |
| }, | |
| { | |
| "epoch": 2.011182108626198, | |
| "grad_norm": 0.12931625545024872, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1259 | |
| }, | |
| { | |
| "epoch": 2.012779552715655, | |
| "grad_norm": 0.10413195192813873, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.0143769968051117, | |
| "grad_norm": 0.23485882580280304, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1261 | |
| }, | |
| { | |
| "epoch": 2.015974440894569, | |
| "grad_norm": 0.07863974571228027, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0005, | |
| "step": 1262 | |
| }, | |
| { | |
| "epoch": 2.0175718849840254, | |
| "grad_norm": 0.33484408259391785, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1263 | |
| }, | |
| { | |
| "epoch": 2.0191693290734825, | |
| "grad_norm": 0.31939253211021423, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 2.0207667731629395, | |
| "grad_norm": 0.10941127687692642, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 2.022364217252396, | |
| "grad_norm": 0.10784590244293213, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1266 | |
| }, | |
| { | |
| "epoch": 2.023961661341853, | |
| "grad_norm": 0.16436795890331268, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1267 | |
| }, | |
| { | |
| "epoch": 2.02555910543131, | |
| "grad_norm": 0.09606944024562836, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 2.027156549520767, | |
| "grad_norm": 0.17230142652988434, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1269 | |
| }, | |
| { | |
| "epoch": 2.0287539936102235, | |
| "grad_norm": 0.5120985507965088, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.0303514376996805, | |
| "grad_norm": 0.06792975217103958, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 1271 | |
| }, | |
| { | |
| "epoch": 2.0319488817891376, | |
| "grad_norm": 0.12373680621385574, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 2.033546325878594, | |
| "grad_norm": 0.10998155921697617, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1273 | |
| }, | |
| { | |
| "epoch": 2.0351437699680512, | |
| "grad_norm": 0.15938100218772888, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1274 | |
| }, | |
| { | |
| "epoch": 2.036741214057508, | |
| "grad_norm": 0.07384390383958817, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 2.038338658146965, | |
| "grad_norm": 0.23505450785160065, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 2.0399361022364215, | |
| "grad_norm": 1.6043593883514404, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0044, | |
| "step": 1277 | |
| }, | |
| { | |
| "epoch": 2.0415335463258786, | |
| "grad_norm": 0.08369333297014236, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1278 | |
| }, | |
| { | |
| "epoch": 2.0431309904153356, | |
| "grad_norm": 0.15266691148281097, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1279 | |
| }, | |
| { | |
| "epoch": 2.0447284345047922, | |
| "grad_norm": 0.1948157548904419, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.0463258785942493, | |
| "grad_norm": 0.33584949374198914, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1281 | |
| }, | |
| { | |
| "epoch": 2.047923322683706, | |
| "grad_norm": 0.6691318154335022, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1282 | |
| }, | |
| { | |
| "epoch": 2.049520766773163, | |
| "grad_norm": 0.1382388472557068, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1283 | |
| }, | |
| { | |
| "epoch": 2.0511182108626196, | |
| "grad_norm": 0.12497832626104355, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 2.0527156549520766, | |
| "grad_norm": 0.27157703042030334, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 2.0543130990415337, | |
| "grad_norm": 0.2626846134662628, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 1286 | |
| }, | |
| { | |
| "epoch": 2.0559105431309903, | |
| "grad_norm": 0.16861018538475037, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1287 | |
| }, | |
| { | |
| "epoch": 2.0575079872204474, | |
| "grad_norm": 0.049949146807193756, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0004, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 2.059105431309904, | |
| "grad_norm": 0.0886000543832779, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1289 | |
| }, | |
| { | |
| "epoch": 2.060702875399361, | |
| "grad_norm": 0.20434832572937012, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.062300319488818, | |
| "grad_norm": 0.11731091886758804, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1291 | |
| }, | |
| { | |
| "epoch": 2.0638977635782747, | |
| "grad_norm": 0.20026597380638123, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 2.0654952076677318, | |
| "grad_norm": 0.27669885754585266, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 1293 | |
| }, | |
| { | |
| "epoch": 2.0670926517571884, | |
| "grad_norm": 0.2785263657569885, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 1294 | |
| }, | |
| { | |
| "epoch": 2.0686900958466454, | |
| "grad_norm": 0.15540768206119537, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 2.070287539936102, | |
| "grad_norm": 0.09532786905765533, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 2.071884984025559, | |
| "grad_norm": 0.1427212655544281, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1297 | |
| }, | |
| { | |
| "epoch": 2.073482428115016, | |
| "grad_norm": 0.3582955300807953, | |
| "learning_rate": 5e-07, | |
| "loss": 0.003, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 2.0750798722044728, | |
| "grad_norm": 0.3676643669605255, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1299 | |
| }, | |
| { | |
| "epoch": 2.07667731629393, | |
| "grad_norm": 0.15875279903411865, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.0782747603833864, | |
| "grad_norm": 0.17880000174045563, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1301 | |
| }, | |
| { | |
| "epoch": 2.0798722044728435, | |
| "grad_norm": 0.09481951594352722, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 2.0814696485623, | |
| "grad_norm": 0.0720115378499031, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1303 | |
| }, | |
| { | |
| "epoch": 2.083067092651757, | |
| "grad_norm": 0.1885913908481598, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 2.084664536741214, | |
| "grad_norm": 0.07471724599599838, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 2.086261980830671, | |
| "grad_norm": 0.10295652598142624, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1306 | |
| }, | |
| { | |
| "epoch": 2.087859424920128, | |
| "grad_norm": 0.48676663637161255, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1307 | |
| }, | |
| { | |
| "epoch": 2.0894568690095845, | |
| "grad_norm": 0.06689424067735672, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1308 | |
| }, | |
| { | |
| "epoch": 2.0910543130990416, | |
| "grad_norm": 0.3771530091762543, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 1309 | |
| }, | |
| { | |
| "epoch": 2.0926517571884986, | |
| "grad_norm": 0.37652865052223206, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.094249201277955, | |
| "grad_norm": 0.0743250697851181, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1311 | |
| }, | |
| { | |
| "epoch": 2.0958466453674123, | |
| "grad_norm": 0.1055615022778511, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 2.097444089456869, | |
| "grad_norm": 0.25487279891967773, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1313 | |
| }, | |
| { | |
| "epoch": 2.099041533546326, | |
| "grad_norm": 0.13495591282844543, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 2.1006389776357826, | |
| "grad_norm": 0.17297884821891785, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 2.1022364217252396, | |
| "grad_norm": 0.14259670674800873, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 2.1038338658146967, | |
| "grad_norm": 0.06441140174865723, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1317 | |
| }, | |
| { | |
| "epoch": 2.1054313099041533, | |
| "grad_norm": 0.0832604393362999, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1318 | |
| }, | |
| { | |
| "epoch": 2.1070287539936103, | |
| "grad_norm": 0.6112750172615051, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1319 | |
| }, | |
| { | |
| "epoch": 2.108626198083067, | |
| "grad_norm": 0.14800317585468292, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.110223642172524, | |
| "grad_norm": 0.389436811208725, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1321 | |
| }, | |
| { | |
| "epoch": 2.1118210862619806, | |
| "grad_norm": 0.1899157464504242, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1322 | |
| }, | |
| { | |
| "epoch": 2.1134185303514377, | |
| "grad_norm": 0.08389966934919357, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1323 | |
| }, | |
| { | |
| "epoch": 2.1150159744408947, | |
| "grad_norm": 0.10899113118648529, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1324 | |
| }, | |
| { | |
| "epoch": 2.1166134185303513, | |
| "grad_norm": 0.10662740468978882, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 2.1182108626198084, | |
| "grad_norm": 0.11692536622285843, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1326 | |
| }, | |
| { | |
| "epoch": 2.119808306709265, | |
| "grad_norm": 0.09900128096342087, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1327 | |
| }, | |
| { | |
| "epoch": 2.121405750798722, | |
| "grad_norm": 0.08334943652153015, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 1328 | |
| }, | |
| { | |
| "epoch": 2.123003194888179, | |
| "grad_norm": 0.10552877187728882, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1329 | |
| }, | |
| { | |
| "epoch": 2.1246006389776357, | |
| "grad_norm": 0.8137688040733337, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0014, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.126198083067093, | |
| "grad_norm": 0.06990483403205872, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1331 | |
| }, | |
| { | |
| "epoch": 2.1277955271565494, | |
| "grad_norm": 0.1675483137369156, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1332 | |
| }, | |
| { | |
| "epoch": 2.1293929712460065, | |
| "grad_norm": 0.2427297979593277, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0024, | |
| "step": 1333 | |
| }, | |
| { | |
| "epoch": 2.130990415335463, | |
| "grad_norm": 0.1828804910182953, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 2.13258785942492, | |
| "grad_norm": 0.35362717509269714, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0029, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 2.134185303514377, | |
| "grad_norm": 0.21310366690158844, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1336 | |
| }, | |
| { | |
| "epoch": 2.135782747603834, | |
| "grad_norm": 0.14100836217403412, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1337 | |
| }, | |
| { | |
| "epoch": 2.137380191693291, | |
| "grad_norm": 0.10196174681186676, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1338 | |
| }, | |
| { | |
| "epoch": 2.1389776357827475, | |
| "grad_norm": 0.2148902863264084, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1339 | |
| }, | |
| { | |
| "epoch": 2.1405750798722045, | |
| "grad_norm": 0.05017423257231712, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.142172523961661, | |
| "grad_norm": 2.62032151222229, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0035, | |
| "step": 1341 | |
| }, | |
| { | |
| "epoch": 2.143769968051118, | |
| "grad_norm": 0.23405258357524872, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 2.1453674121405752, | |
| "grad_norm": 0.05571659281849861, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1343 | |
| }, | |
| { | |
| "epoch": 2.146964856230032, | |
| "grad_norm": 0.17637008428573608, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 2.148562300319489, | |
| "grad_norm": 0.10391382873058319, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 2.1501597444089455, | |
| "grad_norm": 0.5443282723426819, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0027, | |
| "step": 1346 | |
| }, | |
| { | |
| "epoch": 2.1517571884984026, | |
| "grad_norm": 0.09875518828630447, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1347 | |
| }, | |
| { | |
| "epoch": 2.1533546325878596, | |
| "grad_norm": 0.3345401883125305, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1348 | |
| }, | |
| { | |
| "epoch": 2.1549520766773163, | |
| "grad_norm": 0.14103494584560394, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1349 | |
| }, | |
| { | |
| "epoch": 2.1565495207667733, | |
| "grad_norm": 0.1581650972366333, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.15814696485623, | |
| "grad_norm": 0.1288004070520401, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1351 | |
| }, | |
| { | |
| "epoch": 2.159744408945687, | |
| "grad_norm": 0.19039283692836761, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0026, | |
| "step": 1352 | |
| }, | |
| { | |
| "epoch": 2.1613418530351436, | |
| "grad_norm": 0.15048496425151825, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0009, | |
| "step": 1353 | |
| }, | |
| { | |
| "epoch": 2.1629392971246006, | |
| "grad_norm": 0.14754633605480194, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1354 | |
| }, | |
| { | |
| "epoch": 2.1645367412140577, | |
| "grad_norm": 0.20571894943714142, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0033, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 2.1661341853035143, | |
| "grad_norm": 0.12298233062028885, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1356 | |
| }, | |
| { | |
| "epoch": 2.1677316293929714, | |
| "grad_norm": 0.060658592730760574, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0007, | |
| "step": 1357 | |
| }, | |
| { | |
| "epoch": 2.169329073482428, | |
| "grad_norm": 0.2829779088497162, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1358 | |
| }, | |
| { | |
| "epoch": 2.170926517571885, | |
| "grad_norm": 0.1882810741662979, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1359 | |
| }, | |
| { | |
| "epoch": 2.1725239616613417, | |
| "grad_norm": 0.5703380703926086, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0039, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.1741214057507987, | |
| "grad_norm": 0.7452375292778015, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0052, | |
| "step": 1361 | |
| }, | |
| { | |
| "epoch": 2.1757188498402558, | |
| "grad_norm": 0.593784749507904, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 1362 | |
| }, | |
| { | |
| "epoch": 2.1773162939297124, | |
| "grad_norm": 0.472791850566864, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0041, | |
| "step": 1363 | |
| }, | |
| { | |
| "epoch": 2.1789137380191694, | |
| "grad_norm": 0.2769353687763214, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1364 | |
| }, | |
| { | |
| "epoch": 2.180511182108626, | |
| "grad_norm": 0.5703862309455872, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0031, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 2.182108626198083, | |
| "grad_norm": 0.19744770228862762, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1366 | |
| }, | |
| { | |
| "epoch": 2.18370607028754, | |
| "grad_norm": 0.7886673212051392, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0043, | |
| "step": 1367 | |
| }, | |
| { | |
| "epoch": 2.1853035143769968, | |
| "grad_norm": 0.13506360352039337, | |
| "learning_rate": 5e-07, | |
| "loss": 0.001, | |
| "step": 1368 | |
| }, | |
| { | |
| "epoch": 2.186900958466454, | |
| "grad_norm": 0.2344270497560501, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1369 | |
| }, | |
| { | |
| "epoch": 2.1884984025559104, | |
| "grad_norm": 0.19468432664871216, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.1900958466453675, | |
| "grad_norm": 0.1934957057237625, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1371 | |
| }, | |
| { | |
| "epoch": 2.191693290734824, | |
| "grad_norm": 0.2196214497089386, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1372 | |
| }, | |
| { | |
| "epoch": 2.193290734824281, | |
| "grad_norm": 0.13796208798885345, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 1373 | |
| }, | |
| { | |
| "epoch": 2.194888178913738, | |
| "grad_norm": 0.27561473846435547, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 1374 | |
| }, | |
| { | |
| "epoch": 2.196485623003195, | |
| "grad_norm": 0.12160798162221909, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0013, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 2.198083067092652, | |
| "grad_norm": 0.10912802815437317, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0022, | |
| "step": 1376 | |
| }, | |
| { | |
| "epoch": 2.1996805111821085, | |
| "grad_norm": 0.09718818962574005, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 1377 | |
| }, | |
| { | |
| "epoch": 2.2012779552715656, | |
| "grad_norm": 0.11051057279109955, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0006, | |
| "step": 1378 | |
| }, | |
| { | |
| "epoch": 2.202875399361022, | |
| "grad_norm": 0.12758323550224304, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1379 | |
| }, | |
| { | |
| "epoch": 2.2044728434504792, | |
| "grad_norm": 0.08249010145664215, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0012, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.2060702875399363, | |
| "grad_norm": 0.1392274647951126, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1381 | |
| }, | |
| { | |
| "epoch": 2.207667731629393, | |
| "grad_norm": 0.10520713776350021, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0023, | |
| "step": 1382 | |
| }, | |
| { | |
| "epoch": 2.20926517571885, | |
| "grad_norm": 0.10856632888317108, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0038, | |
| "step": 1383 | |
| }, | |
| { | |
| "epoch": 2.2108626198083066, | |
| "grad_norm": 0.08818315714597702, | |
| "learning_rate": 5e-07, | |
| "loss": 0.002, | |
| "step": 1384 | |
| }, | |
| { | |
| "epoch": 2.2124600638977636, | |
| "grad_norm": 0.13421356678009033, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0016, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 2.2140575079872207, | |
| "grad_norm": 0.32427483797073364, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1386 | |
| }, | |
| { | |
| "epoch": 2.2156549520766773, | |
| "grad_norm": 0.2926160991191864, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1387 | |
| }, | |
| { | |
| "epoch": 2.2172523961661343, | |
| "grad_norm": 0.2298290878534317, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0032, | |
| "step": 1388 | |
| }, | |
| { | |
| "epoch": 2.218849840255591, | |
| "grad_norm": 0.09151104092597961, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0015, | |
| "step": 1389 | |
| }, | |
| { | |
| "epoch": 2.220447284345048, | |
| "grad_norm": 0.10674551874399185, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0021, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.2220447284345046, | |
| "grad_norm": 0.2997196614742279, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0025, | |
| "step": 1391 | |
| }, | |
| { | |
| "epoch": 2.2236421725239617, | |
| "grad_norm": 0.11480361223220825, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0028, | |
| "step": 1392 | |
| }, | |
| { | |
| "epoch": 2.2252396166134187, | |
| "grad_norm": 0.07671017944812775, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1393 | |
| }, | |
| { | |
| "epoch": 2.2268370607028753, | |
| "grad_norm": 0.1539941430091858, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1394 | |
| }, | |
| { | |
| "epoch": 2.2284345047923324, | |
| "grad_norm": 0.12852609157562256, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0008, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 2.230031948881789, | |
| "grad_norm": 0.13677293062210083, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1396 | |
| }, | |
| { | |
| "epoch": 2.231629392971246, | |
| "grad_norm": 0.05409352481365204, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0011, | |
| "step": 1397 | |
| }, | |
| { | |
| "epoch": 2.2332268370607027, | |
| "grad_norm": 0.07923945039510727, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0018, | |
| "step": 1398 | |
| }, | |
| { | |
| "epoch": 2.2348242811501597, | |
| "grad_norm": 0.10353945195674896, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0017, | |
| "step": 1399 | |
| }, | |
| { | |
| "epoch": 2.236421725239617, | |
| "grad_norm": 0.24928894639015198, | |
| "learning_rate": 5e-07, | |
| "loss": 0.0019, | |
| "step": 1400 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 3130, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 387658364567552.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |