| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.006, | |
| "eval_steps": 100, | |
| "global_step": 300, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 2e-05, | |
| "grad_norm": 1.8784615993499756, | |
| "learning_rate": 4e-10, | |
| "loss": 1.247075080871582, | |
| "memory(GiB)": 45.58, | |
| "step": 1, | |
| "token_acc": 0.7174163783160323, | |
| "train_speed(iter/s)": 0.013609 | |
| }, | |
| { | |
| "epoch": 0.0001, | |
| "grad_norm": 1.7829604148864746, | |
| "learning_rate": 2e-09, | |
| "loss": 1.2329106330871582, | |
| "memory(GiB)": 47.68, | |
| "step": 5, | |
| "token_acc": 0.7190702324418895, | |
| "train_speed(iter/s)": 0.056616 | |
| }, | |
| { | |
| "epoch": 0.0002, | |
| "grad_norm": 1.8510538339614868, | |
| "learning_rate": 4e-09, | |
| "loss": 1.2224847793579101, | |
| "memory(GiB)": 47.68, | |
| "step": 10, | |
| "token_acc": 0.7056490003173596, | |
| "train_speed(iter/s)": 0.093003 | |
| }, | |
| { | |
| "epoch": 0.0003, | |
| "grad_norm": 1.9363449811935425, | |
| "learning_rate": 5.6e-09, | |
| "loss": 1.2302563667297364, | |
| "memory(GiB)": 48.7, | |
| "step": 15, | |
| "token_acc": 0.7107342772472836, | |
| "train_speed(iter/s)": 0.119381 | |
| }, | |
| { | |
| "epoch": 0.0004, | |
| "grad_norm": 1.935133695602417, | |
| "learning_rate": 7.6e-09, | |
| "loss": 1.2335556983947753, | |
| "memory(GiB)": 48.7, | |
| "step": 20, | |
| "token_acc": 0.7361481714016638, | |
| "train_speed(iter/s)": 0.137465 | |
| }, | |
| { | |
| "epoch": 0.0005, | |
| "grad_norm": 1.7830060720443726, | |
| "learning_rate": 9.599999999999998e-09, | |
| "loss": 1.2144426345825194, | |
| "memory(GiB)": 48.7, | |
| "step": 25, | |
| "token_acc": 0.7571428571428571, | |
| "train_speed(iter/s)": 0.150572 | |
| }, | |
| { | |
| "epoch": 0.0006, | |
| "grad_norm": 1.7309236526489258, | |
| "learning_rate": 1.1599999999999998e-08, | |
| "loss": 1.2357244491577148, | |
| "memory(GiB)": 48.7, | |
| "step": 30, | |
| "token_acc": 0.7153189910979229, | |
| "train_speed(iter/s)": 0.162512 | |
| }, | |
| { | |
| "epoch": 0.0007, | |
| "grad_norm": 1.7474730014801025, | |
| "learning_rate": 1.36e-08, | |
| "loss": 1.231928825378418, | |
| "memory(GiB)": 48.7, | |
| "step": 35, | |
| "token_acc": 0.7078455355582737, | |
| "train_speed(iter/s)": 0.171663 | |
| }, | |
| { | |
| "epoch": 0.0008, | |
| "grad_norm": 1.780007004737854, | |
| "learning_rate": 1.5599999999999997e-08, | |
| "loss": 1.2193016052246093, | |
| "memory(GiB)": 48.7, | |
| "step": 40, | |
| "token_acc": 0.6932599724896836, | |
| "train_speed(iter/s)": 0.178935 | |
| }, | |
| { | |
| "epoch": 0.0009, | |
| "grad_norm": 1.723268747329712, | |
| "learning_rate": 1.72e-08, | |
| "loss": 1.204256820678711, | |
| "memory(GiB)": 48.7, | |
| "step": 45, | |
| "token_acc": 0.6967172137977004, | |
| "train_speed(iter/s)": 0.186547 | |
| }, | |
| { | |
| "epoch": 0.001, | |
| "grad_norm": 1.7750556468963623, | |
| "learning_rate": 1.9199999999999997e-08, | |
| "loss": 1.2007877349853515, | |
| "memory(GiB)": 48.7, | |
| "step": 50, | |
| "token_acc": 0.6921988682295878, | |
| "train_speed(iter/s)": 0.19125 | |
| }, | |
| { | |
| "epoch": 0.0011, | |
| "grad_norm": 1.299729585647583, | |
| "learning_rate": 2.1199999999999998e-08, | |
| "loss": 0.8069381713867188, | |
| "memory(GiB)": 48.7, | |
| "step": 55, | |
| "token_acc": 0.7952855847688123, | |
| "train_speed(iter/s)": 0.195908 | |
| }, | |
| { | |
| "epoch": 0.0012, | |
| "grad_norm": 1.3525291681289673, | |
| "learning_rate": 2.3199999999999996e-08, | |
| "loss": 0.42714319229125974, | |
| "memory(GiB)": 48.7, | |
| "step": 60, | |
| "token_acc": 0.8841448189762796, | |
| "train_speed(iter/s)": 0.201 | |
| }, | |
| { | |
| "epoch": 0.0013, | |
| "grad_norm": 1.080306053161621, | |
| "learning_rate": 2.52e-08, | |
| "loss": 0.4192944526672363, | |
| "memory(GiB)": 48.7, | |
| "step": 65, | |
| "token_acc": 0.8929637526652452, | |
| "train_speed(iter/s)": 0.204749 | |
| }, | |
| { | |
| "epoch": 0.0014, | |
| "grad_norm": 1.177667498588562, | |
| "learning_rate": 2.72e-08, | |
| "loss": 0.4133622169494629, | |
| "memory(GiB)": 48.7, | |
| "step": 70, | |
| "token_acc": 0.8836150845253576, | |
| "train_speed(iter/s)": 0.20853 | |
| }, | |
| { | |
| "epoch": 0.0015, | |
| "grad_norm": 1.4200434684753418, | |
| "learning_rate": 2.92e-08, | |
| "loss": 0.4155101776123047, | |
| "memory(GiB)": 48.7, | |
| "step": 75, | |
| "token_acc": 0.8933333333333333, | |
| "train_speed(iter/s)": 0.211872 | |
| }, | |
| { | |
| "epoch": 0.0016, | |
| "grad_norm": 1.1931238174438477, | |
| "learning_rate": 3.1199999999999995e-08, | |
| "loss": 0.40213637351989745, | |
| "memory(GiB)": 48.7, | |
| "step": 80, | |
| "token_acc": 0.9073569482288828, | |
| "train_speed(iter/s)": 0.214198 | |
| }, | |
| { | |
| "epoch": 0.0017, | |
| "grad_norm": 1.2268942594528198, | |
| "learning_rate": 3.32e-08, | |
| "loss": 0.4193448543548584, | |
| "memory(GiB)": 48.7, | |
| "step": 85, | |
| "token_acc": 0.890360895386021, | |
| "train_speed(iter/s)": 0.217149 | |
| }, | |
| { | |
| "epoch": 0.0018, | |
| "grad_norm": 1.168769121170044, | |
| "learning_rate": 3.52e-08, | |
| "loss": 0.4069235324859619, | |
| "memory(GiB)": 48.7, | |
| "step": 90, | |
| "token_acc": 0.8803038239916187, | |
| "train_speed(iter/s)": 0.219373 | |
| }, | |
| { | |
| "epoch": 0.0019, | |
| "grad_norm": 1.1890403032302856, | |
| "learning_rate": 3.7199999999999996e-08, | |
| "loss": 0.4017478942871094, | |
| "memory(GiB)": 48.7, | |
| "step": 95, | |
| "token_acc": 0.8852933438568797, | |
| "train_speed(iter/s)": 0.221331 | |
| }, | |
| { | |
| "epoch": 0.002, | |
| "grad_norm": 1.0330846309661865, | |
| "learning_rate": 3.9199999999999994e-08, | |
| "loss": 0.3853747844696045, | |
| "memory(GiB)": 48.7, | |
| "step": 100, | |
| "token_acc": 0.8942406230784997, | |
| "train_speed(iter/s)": 0.223547 | |
| }, | |
| { | |
| "epoch": 0.0021, | |
| "grad_norm": 0.8882772922515869, | |
| "learning_rate": 4.12e-08, | |
| "loss": 0.38479089736938477, | |
| "memory(GiB)": 48.7, | |
| "step": 105, | |
| "token_acc": 0.888404744613895, | |
| "train_speed(iter/s)": 0.165159 | |
| }, | |
| { | |
| "epoch": 0.0022, | |
| "grad_norm": 0.9768219590187073, | |
| "learning_rate": 4.32e-08, | |
| "loss": 0.3818492889404297, | |
| "memory(GiB)": 48.7, | |
| "step": 110, | |
| "token_acc": 0.8976019604410993, | |
| "train_speed(iter/s)": 0.168006 | |
| }, | |
| { | |
| "epoch": 0.0023, | |
| "grad_norm": 1.0652796030044556, | |
| "learning_rate": 4.5199999999999994e-08, | |
| "loss": 0.37880630493164064, | |
| "memory(GiB)": 48.7, | |
| "step": 115, | |
| "token_acc": 0.8944169293111212, | |
| "train_speed(iter/s)": 0.170961 | |
| }, | |
| { | |
| "epoch": 0.0024, | |
| "grad_norm": 0.9667234420776367, | |
| "learning_rate": 4.72e-08, | |
| "loss": 0.3773456573486328, | |
| "memory(GiB)": 48.7, | |
| "step": 120, | |
| "token_acc": 0.9041720990873533, | |
| "train_speed(iter/s)": 0.17359 | |
| }, | |
| { | |
| "epoch": 0.0025, | |
| "grad_norm": 1.1004765033721924, | |
| "learning_rate": 4.92e-08, | |
| "loss": 0.37454307079315186, | |
| "memory(GiB)": 48.7, | |
| "step": 125, | |
| "token_acc": 0.8971794358871774, | |
| "train_speed(iter/s)": 0.17613 | |
| }, | |
| { | |
| "epoch": 0.0026, | |
| "grad_norm": 0.9110294580459595, | |
| "learning_rate": 5.12e-08, | |
| "loss": 0.35935871601104735, | |
| "memory(GiB)": 48.7, | |
| "step": 130, | |
| "token_acc": 0.9002339927674963, | |
| "train_speed(iter/s)": 0.178791 | |
| }, | |
| { | |
| "epoch": 0.0027, | |
| "grad_norm": 0.8307713270187378, | |
| "learning_rate": 5.319999999999999e-08, | |
| "loss": 0.3583024501800537, | |
| "memory(GiB)": 48.7, | |
| "step": 135, | |
| "token_acc": 0.9001166861143524, | |
| "train_speed(iter/s)": 0.181064 | |
| }, | |
| { | |
| "epoch": 0.0028, | |
| "grad_norm": 0.8742683529853821, | |
| "learning_rate": 5.52e-08, | |
| "loss": 0.34030709266662595, | |
| "memory(GiB)": 48.7, | |
| "step": 140, | |
| "token_acc": 0.8936170212765957, | |
| "train_speed(iter/s)": 0.183337 | |
| }, | |
| { | |
| "epoch": 0.0029, | |
| "grad_norm": 0.8469884395599365, | |
| "learning_rate": 5.7199999999999996e-08, | |
| "loss": 0.3286393404006958, | |
| "memory(GiB)": 48.7, | |
| "step": 145, | |
| "token_acc": 0.9112193294624461, | |
| "train_speed(iter/s)": 0.185295 | |
| }, | |
| { | |
| "epoch": 0.003, | |
| "grad_norm": 0.7903546094894409, | |
| "learning_rate": 5.92e-08, | |
| "loss": 0.34013702869415285, | |
| "memory(GiB)": 48.7, | |
| "step": 150, | |
| "token_acc": 0.9123177283192633, | |
| "train_speed(iter/s)": 0.187277 | |
| }, | |
| { | |
| "epoch": 0.0031, | |
| "grad_norm": 0.7647843956947327, | |
| "learning_rate": 6.119999999999999e-08, | |
| "loss": 0.32967448234558105, | |
| "memory(GiB)": 48.7, | |
| "step": 155, | |
| "token_acc": 0.9061872909698997, | |
| "train_speed(iter/s)": 0.189242 | |
| }, | |
| { | |
| "epoch": 0.0032, | |
| "grad_norm": 0.7588908076286316, | |
| "learning_rate": 6.32e-08, | |
| "loss": 0.31885499954223634, | |
| "memory(GiB)": 48.7, | |
| "step": 160, | |
| "token_acc": 0.9148387096774193, | |
| "train_speed(iter/s)": 0.190988 | |
| }, | |
| { | |
| "epoch": 0.0033, | |
| "grad_norm": 0.7278595566749573, | |
| "learning_rate": 6.519999999999999e-08, | |
| "loss": 0.3225527048110962, | |
| "memory(GiB)": 48.7, | |
| "step": 165, | |
| "token_acc": 0.9134559535333979, | |
| "train_speed(iter/s)": 0.192579 | |
| }, | |
| { | |
| "epoch": 0.0034, | |
| "grad_norm": 0.7548233270645142, | |
| "learning_rate": 6.719999999999999e-08, | |
| "loss": 0.3111454725265503, | |
| "memory(GiB)": 48.7, | |
| "step": 170, | |
| "token_acc": 0.9035423647678315, | |
| "train_speed(iter/s)": 0.194372 | |
| }, | |
| { | |
| "epoch": 0.0035, | |
| "grad_norm": 0.7382190823554993, | |
| "learning_rate": 6.92e-08, | |
| "loss": 0.30791687965393066, | |
| "memory(GiB)": 48.7, | |
| "step": 175, | |
| "token_acc": 0.9182072829131652, | |
| "train_speed(iter/s)": 0.195886 | |
| }, | |
| { | |
| "epoch": 0.0036, | |
| "grad_norm": 0.6201812028884888, | |
| "learning_rate": 7.12e-08, | |
| "loss": 0.30755660533905027, | |
| "memory(GiB)": 48.7, | |
| "step": 180, | |
| "token_acc": 0.9121319199057715, | |
| "train_speed(iter/s)": 0.197365 | |
| }, | |
| { | |
| "epoch": 0.0037, | |
| "grad_norm": 0.680054247379303, | |
| "learning_rate": 7.32e-08, | |
| "loss": 0.3023200511932373, | |
| "memory(GiB)": 48.7, | |
| "step": 185, | |
| "token_acc": 0.9189243427654024, | |
| "train_speed(iter/s)": 0.198991 | |
| }, | |
| { | |
| "epoch": 0.0038, | |
| "grad_norm": 0.7230331301689148, | |
| "learning_rate": 7.52e-08, | |
| "loss": 0.29792306423187254, | |
| "memory(GiB)": 48.7, | |
| "step": 190, | |
| "token_acc": 0.9166977960403436, | |
| "train_speed(iter/s)": 0.200242 | |
| }, | |
| { | |
| "epoch": 0.0039, | |
| "grad_norm": 0.6341392397880554, | |
| "learning_rate": 7.72e-08, | |
| "loss": 0.2936398983001709, | |
| "memory(GiB)": 48.7, | |
| "step": 195, | |
| "token_acc": 0.916235294117647, | |
| "train_speed(iter/s)": 0.20159 | |
| }, | |
| { | |
| "epoch": 0.004, | |
| "grad_norm": 0.6513913869857788, | |
| "learning_rate": 7.920000000000001e-08, | |
| "loss": 0.29084038734436035, | |
| "memory(GiB)": 48.7, | |
| "step": 200, | |
| "token_acc": 0.915651358950328, | |
| "train_speed(iter/s)": 0.202853 | |
| }, | |
| { | |
| "epoch": 0.0041, | |
| "grad_norm": 0.5849307775497437, | |
| "learning_rate": 8.119999999999999e-08, | |
| "loss": 0.28313345909118653, | |
| "memory(GiB)": 48.7, | |
| "step": 205, | |
| "token_acc": 0.9201498422712934, | |
| "train_speed(iter/s)": 0.174461 | |
| }, | |
| { | |
| "epoch": 0.0042, | |
| "grad_norm": 0.5864230990409851, | |
| "learning_rate": 8.319999999999999e-08, | |
| "loss": 0.27917160987854006, | |
| "memory(GiB)": 48.7, | |
| "step": 210, | |
| "token_acc": 0.9227850926132559, | |
| "train_speed(iter/s)": 0.176086 | |
| }, | |
| { | |
| "epoch": 0.0043, | |
| "grad_norm": 0.5705774426460266, | |
| "learning_rate": 8.52e-08, | |
| "loss": 0.27348690032958983, | |
| "memory(GiB)": 48.7, | |
| "step": 215, | |
| "token_acc": 0.9209313047487321, | |
| "train_speed(iter/s)": 0.177517 | |
| }, | |
| { | |
| "epoch": 0.0044, | |
| "grad_norm": 0.5566577911376953, | |
| "learning_rate": 8.72e-08, | |
| "loss": 0.2738009214401245, | |
| "memory(GiB)": 48.7, | |
| "step": 220, | |
| "token_acc": 0.9212003454231433, | |
| "train_speed(iter/s)": 0.178985 | |
| }, | |
| { | |
| "epoch": 0.0045, | |
| "grad_norm": 0.5548610091209412, | |
| "learning_rate": 8.919999999999999e-08, | |
| "loss": 0.270387601852417, | |
| "memory(GiB)": 48.7, | |
| "step": 225, | |
| "token_acc": 0.9163515016685205, | |
| "train_speed(iter/s)": 0.180404 | |
| }, | |
| { | |
| "epoch": 0.0046, | |
| "grad_norm": 0.5132172107696533, | |
| "learning_rate": 9.12e-08, | |
| "loss": 0.2689182758331299, | |
| "memory(GiB)": 48.7, | |
| "step": 230, | |
| "token_acc": 0.9161837748344371, | |
| "train_speed(iter/s)": 0.181679 | |
| }, | |
| { | |
| "epoch": 0.0047, | |
| "grad_norm": 0.5392292141914368, | |
| "learning_rate": 9.32e-08, | |
| "loss": 0.2692979335784912, | |
| "memory(GiB)": 48.7, | |
| "step": 235, | |
| "token_acc": 0.9189686924493554, | |
| "train_speed(iter/s)": 0.18288 | |
| }, | |
| { | |
| "epoch": 0.0048, | |
| "grad_norm": 0.5800856947898865, | |
| "learning_rate": 9.52e-08, | |
| "loss": 0.2671097755432129, | |
| "memory(GiB)": 48.7, | |
| "step": 240, | |
| "token_acc": 0.9276693455797933, | |
| "train_speed(iter/s)": 0.184197 | |
| }, | |
| { | |
| "epoch": 0.0049, | |
| "grad_norm": 0.5248638391494751, | |
| "learning_rate": 9.72e-08, | |
| "loss": 0.2651660442352295, | |
| "memory(GiB)": 48.7, | |
| "step": 245, | |
| "token_acc": 0.9340950826361967, | |
| "train_speed(iter/s)": 0.185361 | |
| }, | |
| { | |
| "epoch": 0.005, | |
| "grad_norm": 0.4779960811138153, | |
| "learning_rate": 9.919999999999999e-08, | |
| "loss": 0.2598712921142578, | |
| "memory(GiB)": 48.7, | |
| "step": 250, | |
| "token_acc": 0.9171786120591582, | |
| "train_speed(iter/s)": 0.186527 | |
| }, | |
| { | |
| "epoch": 0.0051, | |
| "grad_norm": 0.5446517467498779, | |
| "learning_rate": 1.0119999999999999e-07, | |
| "loss": 0.2589933633804321, | |
| "memory(GiB)": 48.7, | |
| "step": 255, | |
| "token_acc": 0.919714165615805, | |
| "train_speed(iter/s)": 0.187704 | |
| }, | |
| { | |
| "epoch": 0.0052, | |
| "grad_norm": 0.5014457702636719, | |
| "learning_rate": 1.032e-07, | |
| "loss": 0.2622577667236328, | |
| "memory(GiB)": 48.7, | |
| "step": 260, | |
| "token_acc": 0.9149246475449684, | |
| "train_speed(iter/s)": 0.188846 | |
| }, | |
| { | |
| "epoch": 0.0053, | |
| "grad_norm": 0.486289918422699, | |
| "learning_rate": 1.052e-07, | |
| "loss": 0.2525398969650269, | |
| "memory(GiB)": 48.7, | |
| "step": 265, | |
| "token_acc": 0.9272396543883583, | |
| "train_speed(iter/s)": 0.189871 | |
| }, | |
| { | |
| "epoch": 0.0054, | |
| "grad_norm": 0.5184631943702698, | |
| "learning_rate": 1.072e-07, | |
| "loss": 0.2508587121963501, | |
| "memory(GiB)": 48.7, | |
| "step": 270, | |
| "token_acc": 0.9343525179856115, | |
| "train_speed(iter/s)": 0.190895 | |
| }, | |
| { | |
| "epoch": 0.0055, | |
| "grad_norm": 0.7472315430641174, | |
| "learning_rate": 1.092e-07, | |
| "loss": 0.3759446620941162, | |
| "memory(GiB)": 48.7, | |
| "step": 275, | |
| "token_acc": 0.8695306284805091, | |
| "train_speed(iter/s)": 0.191441 | |
| }, | |
| { | |
| "epoch": 0.0056, | |
| "grad_norm": 0.6762179136276245, | |
| "learning_rate": 1.1119999999999999e-07, | |
| "loss": 0.5864349365234375, | |
| "memory(GiB)": 48.7, | |
| "step": 280, | |
| "token_acc": 0.8324331623008372, | |
| "train_speed(iter/s)": 0.191626 | |
| }, | |
| { | |
| "epoch": 0.0057, | |
| "grad_norm": 0.5527864694595337, | |
| "learning_rate": 1.132e-07, | |
| "loss": 0.5735197067260742, | |
| "memory(GiB)": 48.7, | |
| "step": 285, | |
| "token_acc": 0.8262280306444344, | |
| "train_speed(iter/s)": 0.19159 | |
| }, | |
| { | |
| "epoch": 0.0058, | |
| "grad_norm": 0.46731290221214294, | |
| "learning_rate": 1.1519999999999999e-07, | |
| "loss": 0.5600314140319824, | |
| "memory(GiB)": 48.7, | |
| "step": 290, | |
| "token_acc": 0.8467968392868804, | |
| "train_speed(iter/s)": 0.191504 | |
| }, | |
| { | |
| "epoch": 0.0059, | |
| "grad_norm": 0.42628228664398193, | |
| "learning_rate": 1.1719999999999999e-07, | |
| "loss": 0.5390614986419677, | |
| "memory(GiB)": 50.81, | |
| "step": 295, | |
| "token_acc": 0.8327679356703104, | |
| "train_speed(iter/s)": 0.191542 | |
| }, | |
| { | |
| "epoch": 0.006, | |
| "grad_norm": 0.3921428322792053, | |
| "learning_rate": 1.192e-07, | |
| "loss": 0.5223519325256347, | |
| "memory(GiB)": 50.81, | |
| "step": 300, | |
| "token_acc": 0.8428060564559081, | |
| "train_speed(iter/s)": 0.191529 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 50000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 9223372036854775807, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.573770266594509e+17, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |