| { | |
| "best_metric": 35.8859514966135, | |
| "best_model_checkpoint": "whisper-small-mn-6/checkpoint-5000", | |
| "epoch": 12.057877813504824, | |
| "global_step": 15000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.2000000000000006e-07, | |
| "loss": 5.7495, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 9.200000000000001e-07, | |
| "loss": 4.3046, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.42e-06, | |
| "loss": 3.3553, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9200000000000003e-06, | |
| "loss": 2.6498, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 2.42e-06, | |
| "loss": 2.1301, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 2.92e-06, | |
| "loss": 1.7195, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 3.4200000000000007e-06, | |
| "loss": 1.4733, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 3.920000000000001e-06, | |
| "loss": 1.2885, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.42e-06, | |
| "loss": 1.1661, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.92e-06, | |
| "loss": 1.0562, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 5.420000000000001e-06, | |
| "loss": 0.9722, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 5.92e-06, | |
| "loss": 0.9045, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 6.42e-06, | |
| "loss": 0.8404, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 6.92e-06, | |
| "loss": 0.7772, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 7.420000000000001e-06, | |
| "loss": 0.756, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 7.92e-06, | |
| "loss": 0.6819, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 8.42e-06, | |
| "loss": 0.6408, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 8.920000000000001e-06, | |
| "loss": 0.6201, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 9.42e-06, | |
| "loss": 0.5853, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 9.920000000000002e-06, | |
| "loss": 0.5725, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 9.98551724137931e-06, | |
| "loss": 0.5482, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 9.968275862068967e-06, | |
| "loss": 0.5424, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 9.951034482758622e-06, | |
| "loss": 0.5108, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 9.933793103448277e-06, | |
| "loss": 0.5109, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 9.916551724137932e-06, | |
| "loss": 0.5008, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.899310344827588e-06, | |
| "loss": 0.4925, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.882068965517241e-06, | |
| "loss": 0.4937, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 9.864827586206898e-06, | |
| "loss": 0.4793, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.847586206896553e-06, | |
| "loss": 0.4551, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 9.830344827586208e-06, | |
| "loss": 0.4473, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 9.813103448275862e-06, | |
| "loss": 0.438, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 9.795862068965517e-06, | |
| "loss": 0.4431, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 9.778620689655172e-06, | |
| "loss": 0.4341, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 9.761379310344829e-06, | |
| "loss": 0.4247, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 9.744137931034484e-06, | |
| "loss": 0.4065, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 9.726896551724139e-06, | |
| "loss": 0.4051, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 9.709655172413795e-06, | |
| "loss": 0.4003, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 9.692413793103448e-06, | |
| "loss": 0.4015, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.675172413793105e-06, | |
| "loss": 0.407, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 9.65793103448276e-06, | |
| "loss": 0.3774, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "eval_cer": 19.662699434970033, | |
| "eval_loss": 0.43194717168807983, | |
| "eval_runtime": 826.084, | |
| "eval_samples_per_second": 2.283, | |
| "eval_steps_per_second": 0.071, | |
| "eval_wer": 53.27725584443959, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 9.640689655172415e-06, | |
| "loss": 0.3923, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.62344827586207e-06, | |
| "loss": 0.3848, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.606206896551726e-06, | |
| "loss": 0.3854, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 9.58896551724138e-06, | |
| "loss": 0.3842, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 9.571724137931036e-06, | |
| "loss": 0.378, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 9.55448275862069e-06, | |
| "loss": 0.3658, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 9.537241379310345e-06, | |
| "loss": 0.3796, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.52e-06, | |
| "loss": 0.3611, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 9.502758620689655e-06, | |
| "loss": 0.3669, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 9.48551724137931e-06, | |
| "loss": 0.3458, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 9.468275862068967e-06, | |
| "loss": 0.3253, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.451034482758622e-06, | |
| "loss": 0.3084, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 9.433793103448276e-06, | |
| "loss": 0.3157, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 9.416551724137933e-06, | |
| "loss": 0.3149, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 9.399310344827586e-06, | |
| "loss": 0.3088, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 9.382068965517243e-06, | |
| "loss": 0.3094, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 9.364827586206898e-06, | |
| "loss": 0.296, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 9.347586206896552e-06, | |
| "loss": 0.3077, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 9.330344827586207e-06, | |
| "loss": 0.3119, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 9.313103448275864e-06, | |
| "loss": 0.3019, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 9.295862068965517e-06, | |
| "loss": 0.3028, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 9.278620689655174e-06, | |
| "loss": 0.3009, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 9.261379310344828e-06, | |
| "loss": 0.3041, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 9.244137931034483e-06, | |
| "loss": 0.2979, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 9.226896551724138e-06, | |
| "loss": 0.2961, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 9.209655172413793e-06, | |
| "loss": 0.2946, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 9.192413793103448e-06, | |
| "loss": 0.2975, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 9.175172413793105e-06, | |
| "loss": 0.3021, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 9.15793103448276e-06, | |
| "loss": 0.304, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 9.140689655172414e-06, | |
| "loss": 0.2808, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 9.12344827586207e-06, | |
| "loss": 0.286, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 9.106206896551724e-06, | |
| "loss": 0.3008, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 9.08896551724138e-06, | |
| "loss": 0.2982, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 9.071724137931035e-06, | |
| "loss": 0.2874, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 9.05448275862069e-06, | |
| "loss": 0.2895, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 9.037241379310345e-06, | |
| "loss": 0.287, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 9.020000000000002e-06, | |
| "loss": 0.2771, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 9.002758620689655e-06, | |
| "loss": 0.2903, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 8.985517241379311e-06, | |
| "loss": 0.2989, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 8.968275862068966e-06, | |
| "loss": 0.2926, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "eval_cer": 15.021381833170178, | |
| "eval_loss": 0.34934598207473755, | |
| "eval_runtime": 733.1505, | |
| "eval_samples_per_second": 2.572, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 40.49595805112519, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 8.951034482758621e-06, | |
| "loss": 0.2904, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.933793103448276e-06, | |
| "loss": 0.2825, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.916551724137931e-06, | |
| "loss": 0.2875, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 8.899310344827588e-06, | |
| "loss": 0.2747, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 8.882068965517242e-06, | |
| "loss": 0.2865, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 8.864827586206897e-06, | |
| "loss": 0.2767, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 8.847586206896552e-06, | |
| "loss": 0.285, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 8.830344827586209e-06, | |
| "loss": 0.2751, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 8.813103448275862e-06, | |
| "loss": 0.2746, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 8.795862068965518e-06, | |
| "loss": 0.2787, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 8.778620689655173e-06, | |
| "loss": 0.2783, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 8.761379310344828e-06, | |
| "loss": 0.272, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 8.744137931034483e-06, | |
| "loss": 0.2697, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 8.72689655172414e-06, | |
| "loss": 0.2741, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 8.709655172413793e-06, | |
| "loss": 0.2723, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 8.69241379310345e-06, | |
| "loss": 0.2641, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 8.675172413793104e-06, | |
| "loss": 0.273, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.657931034482759e-06, | |
| "loss": 0.2661, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 8.640689655172414e-06, | |
| "loss": 0.2738, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 8.623448275862069e-06, | |
| "loss": 0.2526, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 8.606206896551725e-06, | |
| "loss": 0.2275, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 8.58896551724138e-06, | |
| "loss": 0.2086, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 8.571724137931035e-06, | |
| "loss": 0.2074, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 8.55448275862069e-06, | |
| "loss": 0.2135, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 8.537241379310347e-06, | |
| "loss": 0.2135, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 8.52e-06, | |
| "loss": 0.2161, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 8.502758620689656e-06, | |
| "loss": 0.2214, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 8.485517241379311e-06, | |
| "loss": 0.214, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 8.468275862068966e-06, | |
| "loss": 0.2225, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 8.451034482758621e-06, | |
| "loss": 0.2206, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 8.433793103448277e-06, | |
| "loss": 0.2152, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 8.41655172413793e-06, | |
| "loss": 0.2197, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 8.399310344827587e-06, | |
| "loss": 0.217, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 8.382068965517242e-06, | |
| "loss": 0.2184, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 8.364827586206897e-06, | |
| "loss": 0.2313, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 8.347586206896552e-06, | |
| "loss": 0.2108, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 8.330344827586208e-06, | |
| "loss": 0.224, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 8.313103448275863e-06, | |
| "loss": 0.2182, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 8.295862068965518e-06, | |
| "loss": 0.2154, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 8.278620689655173e-06, | |
| "loss": 0.2331, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "eval_cer": 14.768921634293369, | |
| "eval_loss": 0.3346174359321594, | |
| "eval_runtime": 736.358, | |
| "eval_samples_per_second": 2.561, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 39.174131527201226, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 8.261379310344828e-06, | |
| "loss": 0.2181, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 8.244137931034484e-06, | |
| "loss": 0.2237, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 8.226896551724138e-06, | |
| "loss": 0.2072, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 8.209655172413794e-06, | |
| "loss": 0.2184, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 8.192413793103449e-06, | |
| "loss": 0.2147, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 8.175172413793104e-06, | |
| "loss": 0.2062, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 8.157931034482759e-06, | |
| "loss": 0.2077, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.140689655172415e-06, | |
| "loss": 0.2136, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.123448275862069e-06, | |
| "loss": 0.2136, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 8.106206896551725e-06, | |
| "loss": 0.2135, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 8.08896551724138e-06, | |
| "loss": 0.2167, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 8.071724137931035e-06, | |
| "loss": 0.2096, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 8.054482758620691e-06, | |
| "loss": 0.2167, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 8.037241379310346e-06, | |
| "loss": 0.2033, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 8.020000000000001e-06, | |
| "loss": 0.2097, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 8.002758620689656e-06, | |
| "loss": 0.2202, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 7.985517241379311e-06, | |
| "loss": 0.2108, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 7.968275862068966e-06, | |
| "loss": 0.2043, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 7.951034482758622e-06, | |
| "loss": 0.2126, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 7.933793103448275e-06, | |
| "loss": 0.2157, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 7.916551724137932e-06, | |
| "loss": 0.2125, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 7.899310344827587e-06, | |
| "loss": 0.2255, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 7.882068965517242e-06, | |
| "loss": 0.2096, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 7.864827586206897e-06, | |
| "loss": 0.2101, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 7.847586206896553e-06, | |
| "loss": 0.2104, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 7.830344827586206e-06, | |
| "loss": 0.2102, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 7.813103448275863e-06, | |
| "loss": 0.2228, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 7.795862068965518e-06, | |
| "loss": 0.2104, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 7.778620689655173e-06, | |
| "loss": 0.2014, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 7.76137931034483e-06, | |
| "loss": 0.1691, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 7.744137931034484e-06, | |
| "loss": 0.1584, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 7.726896551724139e-06, | |
| "loss": 0.1638, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 7.709655172413794e-06, | |
| "loss": 0.1601, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 7.692413793103449e-06, | |
| "loss": 0.1668, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 7.675172413793104e-06, | |
| "loss": 0.1608, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 7.65793103448276e-06, | |
| "loss": 0.159, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 7.640689655172413e-06, | |
| "loss": 0.1653, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 7.62344827586207e-06, | |
| "loss": 0.1594, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 7.606206896551725e-06, | |
| "loss": 0.1593, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 7.58896551724138e-06, | |
| "loss": 0.1636, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "eval_cer": 13.794287873323372, | |
| "eval_loss": 0.3287133276462555, | |
| "eval_runtime": 732.0901, | |
| "eval_samples_per_second": 2.576, | |
| "eval_steps_per_second": 0.081, | |
| "eval_wer": 36.92374918068604, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 7.5717241379310345e-06, | |
| "loss": 0.1628, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 3.26, | |
| "learning_rate": 7.55448275862069e-06, | |
| "loss": 0.1543, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 7.537241379310345e-06, | |
| "loss": 0.1745, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 7.520000000000001e-06, | |
| "loss": 0.1591, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 7.5027586206896566e-06, | |
| "loss": 0.1662, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 7.485517241379311e-06, | |
| "loss": 0.1681, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 7.468275862068966e-06, | |
| "loss": 0.164, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 7.451034482758621e-06, | |
| "loss": 0.1645, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 7.433793103448277e-06, | |
| "loss": 0.1634, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 3.42, | |
| "learning_rate": 7.416551724137932e-06, | |
| "loss": 0.165, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 7.3993103448275875e-06, | |
| "loss": 0.166, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 7.3820689655172415e-06, | |
| "loss": 0.1696, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "learning_rate": 7.364827586206897e-06, | |
| "loss": 0.1582, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "learning_rate": 7.347586206896552e-06, | |
| "loss": 0.1681, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 7.330344827586208e-06, | |
| "loss": 0.1691, | |
| "step": 4375 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 7.313103448275863e-06, | |
| "loss": 0.1626, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 7.2958620689655175e-06, | |
| "loss": 0.1577, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 7.278620689655172e-06, | |
| "loss": 0.165, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 7.261379310344828e-06, | |
| "loss": 0.1631, | |
| "step": 4475 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 7.244137931034483e-06, | |
| "loss": 0.1676, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 7.226896551724139e-06, | |
| "loss": 0.1701, | |
| "step": 4525 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 7.2096551724137944e-06, | |
| "loss": 0.1701, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 7.1924137931034485e-06, | |
| "loss": 0.1593, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 7.175172413793104e-06, | |
| "loss": 0.1653, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 3.72, | |
| "learning_rate": 7.157931034482759e-06, | |
| "loss": 0.1682, | |
| "step": 4625 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 7.140689655172415e-06, | |
| "loss": 0.1644, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 7.12344827586207e-06, | |
| "loss": 0.1662, | |
| "step": 4675 | |
| }, | |
| { | |
| "epoch": 3.78, | |
| "learning_rate": 7.106206896551725e-06, | |
| "loss": 0.1594, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 7.088965517241379e-06, | |
| "loss": 0.1589, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 7.071724137931035e-06, | |
| "loss": 0.1714, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 7.05448275862069e-06, | |
| "loss": 0.1742, | |
| "step": 4775 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 7.037241379310346e-06, | |
| "loss": 0.1661, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 7.0200000000000006e-06, | |
| "loss": 0.1597, | |
| "step": 4825 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 7.002758620689655e-06, | |
| "loss": 0.1632, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 6.98551724137931e-06, | |
| "loss": 0.1655, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 6.968275862068966e-06, | |
| "loss": 0.1613, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 6.951034482758622e-06, | |
| "loss": 0.1703, | |
| "step": 4925 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 6.933793103448277e-06, | |
| "loss": 0.1651, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 6.916551724137932e-06, | |
| "loss": 0.1625, | |
| "step": 4975 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 6.899310344827586e-06, | |
| "loss": 0.1157, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "eval_cer": 13.310835179555877, | |
| "eval_loss": 0.3296343684196472, | |
| "eval_runtime": 742.4097, | |
| "eval_samples_per_second": 2.54, | |
| "eval_steps_per_second": 0.079, | |
| "eval_wer": 35.8859514966135, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 6.882068965517242e-06, | |
| "loss": 0.1221, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 6.864827586206897e-06, | |
| "loss": 0.1173, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "learning_rate": 6.847586206896553e-06, | |
| "loss": 0.116, | |
| "step": 5075 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 6.8303448275862075e-06, | |
| "loss": 0.1226, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "learning_rate": 6.813103448275863e-06, | |
| "loss": 0.1194, | |
| "step": 5125 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 6.795862068965517e-06, | |
| "loss": 0.1239, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 6.778620689655173e-06, | |
| "loss": 0.1195, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 6.761379310344828e-06, | |
| "loss": 0.1213, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 6.7441379310344836e-06, | |
| "loss": 0.1244, | |
| "step": 5225 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 6.7268965517241384e-06, | |
| "loss": 0.1245, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 6.709655172413793e-06, | |
| "loss": 0.124, | |
| "step": 5275 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 6.692413793103448e-06, | |
| "loss": 0.1226, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 6.675172413793104e-06, | |
| "loss": 0.1223, | |
| "step": 5325 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 6.65793103448276e-06, | |
| "loss": 0.1199, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 6.6406896551724145e-06, | |
| "loss": 0.1163, | |
| "step": 5375 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 6.62344827586207e-06, | |
| "loss": 0.1169, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 6.606206896551724e-06, | |
| "loss": 0.1289, | |
| "step": 5425 | |
| }, | |
| { | |
| "epoch": 4.38, | |
| "learning_rate": 6.58896551724138e-06, | |
| "loss": 0.121, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 6.571724137931035e-06, | |
| "loss": 0.1194, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 6.5544827586206905e-06, | |
| "loss": 0.1248, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 6.537241379310345e-06, | |
| "loss": 0.1238, | |
| "step": 5525 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 6.520000000000001e-06, | |
| "loss": 0.1212, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 6.502758620689655e-06, | |
| "loss": 0.1255, | |
| "step": 5575 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 6.485517241379311e-06, | |
| "loss": 0.1267, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 6.468275862068966e-06, | |
| "loss": 0.121, | |
| "step": 5625 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 6.4510344827586214e-06, | |
| "loss": 0.1198, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 6.433793103448276e-06, | |
| "loss": 0.1324, | |
| "step": 5675 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 6.416551724137931e-06, | |
| "loss": 0.1242, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 6.399310344827587e-06, | |
| "loss": 0.1223, | |
| "step": 5725 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 6.382068965517242e-06, | |
| "loss": 0.1187, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 6.3648275862068975e-06, | |
| "loss": 0.1224, | |
| "step": 5775 | |
| }, | |
| { | |
| "epoch": 4.66, | |
| "learning_rate": 6.347586206896552e-06, | |
| "loss": 0.1243, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 4.68, | |
| "learning_rate": 6.330344827586208e-06, | |
| "loss": 0.1255, | |
| "step": 5825 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 6.313103448275862e-06, | |
| "loss": 0.1199, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 6.295862068965518e-06, | |
| "loss": 0.1215, | |
| "step": 5875 | |
| }, | |
| { | |
| "epoch": 4.74, | |
| "learning_rate": 6.278620689655173e-06, | |
| "loss": 0.1253, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 6.261379310344828e-06, | |
| "loss": 0.1282, | |
| "step": 5925 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 6.244137931034483e-06, | |
| "loss": 0.1316, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 6.226896551724139e-06, | |
| "loss": 0.126, | |
| "step": 5975 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 6.209655172413793e-06, | |
| "loss": 0.1271, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "eval_cer": 13.570165043708245, | |
| "eval_loss": 0.34215304255485535, | |
| "eval_runtime": 739.2004, | |
| "eval_samples_per_second": 2.551, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 36.07166266113175, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 6.192413793103449e-06, | |
| "loss": 0.1185, | |
| "step": 6025 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "learning_rate": 6.175172413793104e-06, | |
| "loss": 0.1304, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 4.88, | |
| "learning_rate": 6.157931034482759e-06, | |
| "loss": 0.1242, | |
| "step": 6075 | |
| }, | |
| { | |
| "epoch": 4.9, | |
| "learning_rate": 6.140689655172414e-06, | |
| "loss": 0.1211, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 6.123448275862069e-06, | |
| "loss": 0.1248, | |
| "step": 6125 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 6.106206896551725e-06, | |
| "loss": 0.1212, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 6.08896551724138e-06, | |
| "loss": 0.1186, | |
| "step": 6175 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 6.071724137931035e-06, | |
| "loss": 0.1262, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "learning_rate": 6.05448275862069e-06, | |
| "loss": 0.1178, | |
| "step": 6225 | |
| }, | |
| { | |
| "epoch": 5.02, | |
| "learning_rate": 6.037241379310346e-06, | |
| "loss": 0.0885, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "learning_rate": 6.02e-06, | |
| "loss": 0.0871, | |
| "step": 6275 | |
| }, | |
| { | |
| "epoch": 5.06, | |
| "learning_rate": 6.002758620689656e-06, | |
| "loss": 0.0839, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 5.08, | |
| "learning_rate": 5.9855172413793105e-06, | |
| "loss": 0.0832, | |
| "step": 6325 | |
| }, | |
| { | |
| "epoch": 5.1, | |
| "learning_rate": 5.968275862068966e-06, | |
| "loss": 0.0845, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 5.12, | |
| "learning_rate": 5.951034482758621e-06, | |
| "loss": 0.0853, | |
| "step": 6375 | |
| }, | |
| { | |
| "epoch": 5.14, | |
| "learning_rate": 5.933793103448277e-06, | |
| "loss": 0.0823, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 5.16, | |
| "learning_rate": 5.916551724137931e-06, | |
| "loss": 0.0845, | |
| "step": 6425 | |
| }, | |
| { | |
| "epoch": 5.18, | |
| "learning_rate": 5.899310344827587e-06, | |
| "loss": 0.0838, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 5.2, | |
| "learning_rate": 5.8820689655172415e-06, | |
| "loss": 0.0855, | |
| "step": 6475 | |
| }, | |
| { | |
| "epoch": 5.23, | |
| "learning_rate": 5.864827586206897e-06, | |
| "loss": 0.0855, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 5.25, | |
| "learning_rate": 5.847586206896552e-06, | |
| "loss": 0.0829, | |
| "step": 6525 | |
| }, | |
| { | |
| "epoch": 5.27, | |
| "learning_rate": 5.830344827586208e-06, | |
| "loss": 0.0895, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 5.29, | |
| "learning_rate": 5.813103448275863e-06, | |
| "loss": 0.0875, | |
| "step": 6575 | |
| }, | |
| { | |
| "epoch": 5.31, | |
| "learning_rate": 5.7958620689655175e-06, | |
| "loss": 0.0848, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 5.33, | |
| "learning_rate": 5.778620689655173e-06, | |
| "loss": 0.084, | |
| "step": 6625 | |
| }, | |
| { | |
| "epoch": 5.35, | |
| "learning_rate": 5.761379310344828e-06, | |
| "loss": 0.0844, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 5.37, | |
| "learning_rate": 5.744137931034484e-06, | |
| "loss": 0.0878, | |
| "step": 6675 | |
| }, | |
| { | |
| "epoch": 5.39, | |
| "learning_rate": 5.726896551724138e-06, | |
| "loss": 0.0893, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 5.41, | |
| "learning_rate": 5.7096551724137936e-06, | |
| "loss": 0.0883, | |
| "step": 6725 | |
| }, | |
| { | |
| "epoch": 5.43, | |
| "learning_rate": 5.692413793103448e-06, | |
| "loss": 0.0841, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 5.45, | |
| "learning_rate": 5.675172413793104e-06, | |
| "loss": 0.0912, | |
| "step": 6775 | |
| }, | |
| { | |
| "epoch": 5.47, | |
| "learning_rate": 5.657931034482759e-06, | |
| "loss": 0.0892, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 5.49, | |
| "learning_rate": 5.640689655172415e-06, | |
| "loss": 0.0853, | |
| "step": 6825 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 5.623448275862069e-06, | |
| "loss": 0.0888, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 5.53, | |
| "learning_rate": 5.6062068965517245e-06, | |
| "loss": 0.0934, | |
| "step": 6875 | |
| }, | |
| { | |
| "epoch": 5.55, | |
| "learning_rate": 5.588965517241379e-06, | |
| "loss": 0.0887, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 5.57, | |
| "learning_rate": 5.571724137931035e-06, | |
| "loss": 0.0902, | |
| "step": 6925 | |
| }, | |
| { | |
| "epoch": 5.59, | |
| "learning_rate": 5.554482758620691e-06, | |
| "loss": 0.0895, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 5.61, | |
| "learning_rate": 5.537241379310346e-06, | |
| "loss": 0.092, | |
| "step": 6975 | |
| }, | |
| { | |
| "epoch": 5.63, | |
| "learning_rate": 5.5200000000000005e-06, | |
| "loss": 0.0879, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 5.63, | |
| "eval_cer": 13.777972418293919, | |
| "eval_loss": 0.3660544753074646, | |
| "eval_runtime": 738.6606, | |
| "eval_samples_per_second": 2.553, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 36.69434127157527, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 5.65, | |
| "learning_rate": 5.502758620689655e-06, | |
| "loss": 0.0883, | |
| "step": 7025 | |
| }, | |
| { | |
| "epoch": 5.67, | |
| "learning_rate": 5.485517241379311e-06, | |
| "loss": 0.0813, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 5.69, | |
| "learning_rate": 5.468275862068966e-06, | |
| "loss": 0.0939, | |
| "step": 7075 | |
| }, | |
| { | |
| "epoch": 5.71, | |
| "learning_rate": 5.451034482758622e-06, | |
| "loss": 0.0883, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 5.73, | |
| "learning_rate": 5.433793103448276e-06, | |
| "loss": 0.0938, | |
| "step": 7125 | |
| }, | |
| { | |
| "epoch": 5.75, | |
| "learning_rate": 5.4165517241379314e-06, | |
| "loss": 0.0915, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 5.399310344827586e-06, | |
| "loss": 0.0899, | |
| "step": 7175 | |
| }, | |
| { | |
| "epoch": 5.79, | |
| "learning_rate": 5.382068965517242e-06, | |
| "loss": 0.0879, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 5.81, | |
| "learning_rate": 5.364827586206897e-06, | |
| "loss": 0.0861, | |
| "step": 7225 | |
| }, | |
| { | |
| "epoch": 5.83, | |
| "learning_rate": 5.347586206896553e-06, | |
| "loss": 0.0948, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 5.85, | |
| "learning_rate": 5.330344827586207e-06, | |
| "loss": 0.0923, | |
| "step": 7275 | |
| }, | |
| { | |
| "epoch": 5.87, | |
| "learning_rate": 5.313103448275862e-06, | |
| "loss": 0.091, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 5.89, | |
| "learning_rate": 5.295862068965517e-06, | |
| "loss": 0.0899, | |
| "step": 7325 | |
| }, | |
| { | |
| "epoch": 5.91, | |
| "learning_rate": 5.278620689655173e-06, | |
| "loss": 0.0862, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 5.93, | |
| "learning_rate": 5.261379310344829e-06, | |
| "loss": 0.0874, | |
| "step": 7375 | |
| }, | |
| { | |
| "epoch": 5.95, | |
| "learning_rate": 5.2441379310344835e-06, | |
| "loss": 0.0924, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "learning_rate": 5.226896551724138e-06, | |
| "loss": 0.0914, | |
| "step": 7425 | |
| }, | |
| { | |
| "epoch": 5.99, | |
| "learning_rate": 5.209655172413793e-06, | |
| "loss": 0.0948, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 6.01, | |
| "learning_rate": 5.192413793103449e-06, | |
| "loss": 0.0757, | |
| "step": 7475 | |
| }, | |
| { | |
| "epoch": 6.03, | |
| "learning_rate": 5.175172413793104e-06, | |
| "loss": 0.0593, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 6.05, | |
| "learning_rate": 5.1579310344827596e-06, | |
| "loss": 0.0544, | |
| "step": 7525 | |
| }, | |
| { | |
| "epoch": 6.07, | |
| "learning_rate": 5.140689655172414e-06, | |
| "loss": 0.0586, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 5.123448275862069e-06, | |
| "loss": 0.0599, | |
| "step": 7575 | |
| }, | |
| { | |
| "epoch": 6.11, | |
| "learning_rate": 5.106206896551724e-06, | |
| "loss": 0.0564, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 6.13, | |
| "learning_rate": 5.08896551724138e-06, | |
| "loss": 0.0606, | |
| "step": 7625 | |
| }, | |
| { | |
| "epoch": 6.15, | |
| "learning_rate": 5.071724137931035e-06, | |
| "loss": 0.0574, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 6.17, | |
| "learning_rate": 5.0544827586206905e-06, | |
| "loss": 0.0562, | |
| "step": 7675 | |
| }, | |
| { | |
| "epoch": 6.19, | |
| "learning_rate": 5.0372413793103445e-06, | |
| "loss": 0.0591, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 6.21, | |
| "learning_rate": 5.02e-06, | |
| "loss": 0.0545, | |
| "step": 7725 | |
| }, | |
| { | |
| "epoch": 6.23, | |
| "learning_rate": 5.002758620689656e-06, | |
| "loss": 0.0597, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 6.25, | |
| "learning_rate": 4.985517241379311e-06, | |
| "loss": 0.0626, | |
| "step": 7775 | |
| }, | |
| { | |
| "epoch": 6.27, | |
| "learning_rate": 4.968275862068966e-06, | |
| "loss": 0.0604, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 6.29, | |
| "learning_rate": 4.951034482758621e-06, | |
| "loss": 0.0575, | |
| "step": 7825 | |
| }, | |
| { | |
| "epoch": 6.31, | |
| "learning_rate": 4.933793103448276e-06, | |
| "loss": 0.062, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 6.33, | |
| "learning_rate": 4.916551724137931e-06, | |
| "loss": 0.0573, | |
| "step": 7875 | |
| }, | |
| { | |
| "epoch": 6.35, | |
| "learning_rate": 4.899310344827586e-06, | |
| "loss": 0.063, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 6.37, | |
| "learning_rate": 4.882068965517242e-06, | |
| "loss": 0.0569, | |
| "step": 7925 | |
| }, | |
| { | |
| "epoch": 6.39, | |
| "learning_rate": 4.8648275862068974e-06, | |
| "loss": 0.0638, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 6.41, | |
| "learning_rate": 4.847586206896552e-06, | |
| "loss": 0.0606, | |
| "step": 7975 | |
| }, | |
| { | |
| "epoch": 6.43, | |
| "learning_rate": 4.830344827586207e-06, | |
| "loss": 0.0574, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 6.43, | |
| "eval_cer": 13.50146839095265, | |
| "eval_loss": 0.3883569538593292, | |
| "eval_runtime": 731.3755, | |
| "eval_samples_per_second": 2.579, | |
| "eval_steps_per_second": 0.081, | |
| "eval_wer": 36.459471269390434, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 6.45, | |
| "learning_rate": 4.813103448275863e-06, | |
| "loss": 0.0566, | |
| "step": 8025 | |
| }, | |
| { | |
| "epoch": 6.47, | |
| "learning_rate": 4.795862068965518e-06, | |
| "loss": 0.0604, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 6.49, | |
| "learning_rate": 4.778620689655173e-06, | |
| "loss": 0.0592, | |
| "step": 8075 | |
| }, | |
| { | |
| "epoch": 6.51, | |
| "learning_rate": 4.761379310344828e-06, | |
| "loss": 0.06, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 4.744137931034483e-06, | |
| "loss": 0.058, | |
| "step": 8125 | |
| }, | |
| { | |
| "epoch": 6.55, | |
| "learning_rate": 4.726896551724138e-06, | |
| "loss": 0.0582, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 6.57, | |
| "learning_rate": 4.709655172413794e-06, | |
| "loss": 0.0568, | |
| "step": 8175 | |
| }, | |
| { | |
| "epoch": 6.59, | |
| "learning_rate": 4.692413793103449e-06, | |
| "loss": 0.0563, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 6.61, | |
| "learning_rate": 4.6751724137931035e-06, | |
| "loss": 0.0611, | |
| "step": 8225 | |
| }, | |
| { | |
| "epoch": 6.63, | |
| "learning_rate": 4.657931034482759e-06, | |
| "loss": 0.0621, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 6.65, | |
| "learning_rate": 4.640689655172414e-06, | |
| "loss": 0.0605, | |
| "step": 8275 | |
| }, | |
| { | |
| "epoch": 6.67, | |
| "learning_rate": 4.623448275862069e-06, | |
| "loss": 0.06, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 6.69, | |
| "learning_rate": 4.606206896551724e-06, | |
| "loss": 0.057, | |
| "step": 8325 | |
| }, | |
| { | |
| "epoch": 6.71, | |
| "learning_rate": 4.58896551724138e-06, | |
| "loss": 0.0629, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 6.73, | |
| "learning_rate": 4.571724137931035e-06, | |
| "loss": 0.0594, | |
| "step": 8375 | |
| }, | |
| { | |
| "epoch": 6.75, | |
| "learning_rate": 4.55448275862069e-06, | |
| "loss": 0.0634, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 6.77, | |
| "learning_rate": 4.537241379310345e-06, | |
| "loss": 0.0594, | |
| "step": 8425 | |
| }, | |
| { | |
| "epoch": 6.79, | |
| "learning_rate": 4.520000000000001e-06, | |
| "loss": 0.0591, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 6.81, | |
| "learning_rate": 4.502758620689656e-06, | |
| "loss": 0.0602, | |
| "step": 8475 | |
| }, | |
| { | |
| "epoch": 6.83, | |
| "learning_rate": 4.4855172413793105e-06, | |
| "loss": 0.0626, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 6.85, | |
| "learning_rate": 4.468275862068966e-06, | |
| "loss": 0.0611, | |
| "step": 8525 | |
| }, | |
| { | |
| "epoch": 6.87, | |
| "learning_rate": 4.451034482758621e-06, | |
| "loss": 0.0608, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 4.433793103448276e-06, | |
| "loss": 0.0632, | |
| "step": 8575 | |
| }, | |
| { | |
| "epoch": 6.91, | |
| "learning_rate": 4.416551724137932e-06, | |
| "loss": 0.059, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 6.93, | |
| "learning_rate": 4.3993103448275866e-06, | |
| "loss": 0.0618, | |
| "step": 8625 | |
| }, | |
| { | |
| "epoch": 6.95, | |
| "learning_rate": 4.3820689655172414e-06, | |
| "loss": 0.0599, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 6.97, | |
| "learning_rate": 4.364827586206897e-06, | |
| "loss": 0.0582, | |
| "step": 8675 | |
| }, | |
| { | |
| "epoch": 6.99, | |
| "learning_rate": 4.347586206896552e-06, | |
| "loss": 0.0608, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 7.01, | |
| "learning_rate": 4.330344827586207e-06, | |
| "loss": 0.0451, | |
| "step": 8725 | |
| }, | |
| { | |
| "epoch": 7.03, | |
| "learning_rate": 4.313103448275863e-06, | |
| "loss": 0.0363, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 7.05, | |
| "learning_rate": 4.2958620689655175e-06, | |
| "loss": 0.0367, | |
| "step": 8775 | |
| }, | |
| { | |
| "epoch": 7.07, | |
| "learning_rate": 4.278620689655173e-06, | |
| "loss": 0.0357, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 7.09, | |
| "learning_rate": 4.261379310344828e-06, | |
| "loss": 0.0365, | |
| "step": 8825 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 4.244137931034483e-06, | |
| "loss": 0.0371, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 7.13, | |
| "learning_rate": 4.226896551724139e-06, | |
| "loss": 0.0353, | |
| "step": 8875 | |
| }, | |
| { | |
| "epoch": 7.15, | |
| "learning_rate": 4.210344827586207e-06, | |
| "loss": 0.036, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 7.17, | |
| "learning_rate": 4.193103448275863e-06, | |
| "loss": 0.0377, | |
| "step": 8925 | |
| }, | |
| { | |
| "epoch": 7.19, | |
| "learning_rate": 4.175862068965518e-06, | |
| "loss": 0.0387, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 7.21, | |
| "learning_rate": 4.158620689655173e-06, | |
| "loss": 0.0374, | |
| "step": 8975 | |
| }, | |
| { | |
| "epoch": 7.23, | |
| "learning_rate": 4.141379310344828e-06, | |
| "loss": 0.036, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 7.23, | |
| "eval_cer": 13.842375530252287, | |
| "eval_loss": 0.41278308629989624, | |
| "eval_runtime": 740.4031, | |
| "eval_samples_per_second": 2.547, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 37.14223290364868, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 7.25, | |
| "learning_rate": 4.124137931034483e-06, | |
| "loss": 0.039, | |
| "step": 9025 | |
| }, | |
| { | |
| "epoch": 7.27, | |
| "learning_rate": 4.106896551724138e-06, | |
| "loss": 0.0388, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 7.3, | |
| "learning_rate": 4.089655172413794e-06, | |
| "loss": 0.0376, | |
| "step": 9075 | |
| }, | |
| { | |
| "epoch": 7.32, | |
| "learning_rate": 4.072413793103449e-06, | |
| "loss": 0.036, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 7.34, | |
| "learning_rate": 4.0551724137931036e-06, | |
| "loss": 0.0345, | |
| "step": 9125 | |
| }, | |
| { | |
| "epoch": 7.36, | |
| "learning_rate": 4.0379310344827584e-06, | |
| "loss": 0.0407, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 7.38, | |
| "learning_rate": 4.020689655172414e-06, | |
| "loss": 0.0391, | |
| "step": 9175 | |
| }, | |
| { | |
| "epoch": 7.4, | |
| "learning_rate": 4.003448275862069e-06, | |
| "loss": 0.0367, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 7.42, | |
| "learning_rate": 3.986206896551724e-06, | |
| "loss": 0.0378, | |
| "step": 9225 | |
| }, | |
| { | |
| "epoch": 7.44, | |
| "learning_rate": 3.96896551724138e-06, | |
| "loss": 0.0356, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 7.46, | |
| "learning_rate": 3.951724137931035e-06, | |
| "loss": 0.0392, | |
| "step": 9275 | |
| }, | |
| { | |
| "epoch": 7.48, | |
| "learning_rate": 3.93448275862069e-06, | |
| "loss": 0.0371, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "learning_rate": 3.917241379310345e-06, | |
| "loss": 0.039, | |
| "step": 9325 | |
| }, | |
| { | |
| "epoch": 7.52, | |
| "learning_rate": 3.900000000000001e-06, | |
| "loss": 0.0386, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 7.54, | |
| "learning_rate": 3.882758620689656e-06, | |
| "loss": 0.0377, | |
| "step": 9375 | |
| }, | |
| { | |
| "epoch": 7.56, | |
| "learning_rate": 3.8655172413793105e-06, | |
| "loss": 0.0377, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 7.58, | |
| "learning_rate": 3.848275862068966e-06, | |
| "loss": 0.0397, | |
| "step": 9425 | |
| }, | |
| { | |
| "epoch": 7.6, | |
| "learning_rate": 3.831034482758621e-06, | |
| "loss": 0.0355, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "learning_rate": 3.813793103448276e-06, | |
| "loss": 0.0374, | |
| "step": 9475 | |
| }, | |
| { | |
| "epoch": 7.64, | |
| "learning_rate": 3.7965517241379313e-06, | |
| "loss": 0.0362, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 7.66, | |
| "learning_rate": 3.7793103448275866e-06, | |
| "loss": 0.0402, | |
| "step": 9525 | |
| }, | |
| { | |
| "epoch": 7.68, | |
| "learning_rate": 3.7620689655172414e-06, | |
| "loss": 0.0375, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 7.7, | |
| "learning_rate": 3.7448275862068967e-06, | |
| "loss": 0.0377, | |
| "step": 9575 | |
| }, | |
| { | |
| "epoch": 7.72, | |
| "learning_rate": 3.727586206896552e-06, | |
| "loss": 0.0371, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 7.74, | |
| "learning_rate": 3.710344827586207e-06, | |
| "loss": 0.038, | |
| "step": 9625 | |
| }, | |
| { | |
| "epoch": 7.76, | |
| "learning_rate": 3.6931034482758626e-06, | |
| "loss": 0.0361, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 7.78, | |
| "learning_rate": 3.675862068965518e-06, | |
| "loss": 0.0381, | |
| "step": 9675 | |
| }, | |
| { | |
| "epoch": 7.8, | |
| "learning_rate": 3.6586206896551728e-06, | |
| "loss": 0.0394, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 7.82, | |
| "learning_rate": 3.641379310344828e-06, | |
| "loss": 0.0396, | |
| "step": 9725 | |
| }, | |
| { | |
| "epoch": 7.84, | |
| "learning_rate": 3.6241379310344834e-06, | |
| "loss": 0.0388, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 7.86, | |
| "learning_rate": 3.6068965517241382e-06, | |
| "loss": 0.0343, | |
| "step": 9775 | |
| }, | |
| { | |
| "epoch": 7.88, | |
| "learning_rate": 3.5896551724137935e-06, | |
| "loss": 0.039, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 7.9, | |
| "learning_rate": 3.5724137931034484e-06, | |
| "loss": 0.0362, | |
| "step": 9825 | |
| }, | |
| { | |
| "epoch": 7.92, | |
| "learning_rate": 3.5551724137931037e-06, | |
| "loss": 0.0387, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 7.94, | |
| "learning_rate": 3.537931034482759e-06, | |
| "loss": 0.0397, | |
| "step": 9875 | |
| }, | |
| { | |
| "epoch": 7.96, | |
| "learning_rate": 3.520689655172414e-06, | |
| "loss": 0.0376, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 7.98, | |
| "learning_rate": 3.503448275862069e-06, | |
| "loss": 0.0389, | |
| "step": 9925 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 3.4862068965517244e-06, | |
| "loss": 0.0361, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 8.02, | |
| "learning_rate": 3.4689655172413793e-06, | |
| "loss": 0.0238, | |
| "step": 9975 | |
| }, | |
| { | |
| "epoch": 8.04, | |
| "learning_rate": 3.4517241379310346e-06, | |
| "loss": 0.0229, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 8.04, | |
| "eval_cer": 13.847527779208956, | |
| "eval_loss": 0.4321258068084717, | |
| "eval_runtime": 737.7687, | |
| "eval_samples_per_second": 2.556, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 36.858204063797245, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 8.06, | |
| "learning_rate": 3.43448275862069e-06, | |
| "loss": 0.0222, | |
| "step": 10025 | |
| }, | |
| { | |
| "epoch": 8.08, | |
| "learning_rate": 3.4172413793103448e-06, | |
| "loss": 0.022, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 8.1, | |
| "learning_rate": 3.4000000000000005e-06, | |
| "loss": 0.0213, | |
| "step": 10075 | |
| }, | |
| { | |
| "epoch": 8.12, | |
| "learning_rate": 3.3827586206896558e-06, | |
| "loss": 0.022, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 8.14, | |
| "learning_rate": 3.3655172413793106e-06, | |
| "loss": 0.0225, | |
| "step": 10125 | |
| }, | |
| { | |
| "epoch": 8.16, | |
| "learning_rate": 3.348275862068966e-06, | |
| "loss": 0.0229, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 8.18, | |
| "learning_rate": 3.3310344827586212e-06, | |
| "loss": 0.0225, | |
| "step": 10175 | |
| }, | |
| { | |
| "epoch": 8.2, | |
| "learning_rate": 3.313793103448276e-06, | |
| "loss": 0.0231, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 8.22, | |
| "learning_rate": 3.2965517241379314e-06, | |
| "loss": 0.0232, | |
| "step": 10225 | |
| }, | |
| { | |
| "epoch": 8.24, | |
| "learning_rate": 3.2793103448275863e-06, | |
| "loss": 0.0208, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 8.26, | |
| "learning_rate": 3.2620689655172416e-06, | |
| "loss": 0.023, | |
| "step": 10275 | |
| }, | |
| { | |
| "epoch": 8.28, | |
| "learning_rate": 3.244827586206897e-06, | |
| "loss": 0.0225, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 8.3, | |
| "learning_rate": 3.2275862068965517e-06, | |
| "loss": 0.0216, | |
| "step": 10325 | |
| }, | |
| { | |
| "epoch": 8.32, | |
| "learning_rate": 3.210344827586207e-06, | |
| "loss": 0.0225, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 8.34, | |
| "learning_rate": 3.1931034482758623e-06, | |
| "loss": 0.0228, | |
| "step": 10375 | |
| }, | |
| { | |
| "epoch": 8.36, | |
| "learning_rate": 3.175862068965517e-06, | |
| "loss": 0.0221, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "learning_rate": 3.1586206896551725e-06, | |
| "loss": 0.0213, | |
| "step": 10425 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 3.1413793103448278e-06, | |
| "loss": 0.0227, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 8.42, | |
| "learning_rate": 3.124137931034483e-06, | |
| "loss": 0.0232, | |
| "step": 10475 | |
| }, | |
| { | |
| "epoch": 8.44, | |
| "learning_rate": 3.1068965517241384e-06, | |
| "loss": 0.0207, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 8.46, | |
| "learning_rate": 3.0896551724137937e-06, | |
| "loss": 0.0205, | |
| "step": 10525 | |
| }, | |
| { | |
| "epoch": 8.48, | |
| "learning_rate": 3.0724137931034485e-06, | |
| "loss": 0.0216, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 8.5, | |
| "learning_rate": 3.055172413793104e-06, | |
| "loss": 0.0219, | |
| "step": 10575 | |
| }, | |
| { | |
| "epoch": 8.52, | |
| "learning_rate": 3.037931034482759e-06, | |
| "loss": 0.0215, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 8.54, | |
| "learning_rate": 3.020689655172414e-06, | |
| "loss": 0.022, | |
| "step": 10625 | |
| }, | |
| { | |
| "epoch": 8.56, | |
| "learning_rate": 3.0034482758620693e-06, | |
| "loss": 0.0221, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 8.58, | |
| "learning_rate": 2.9862068965517246e-06, | |
| "loss": 0.0218, | |
| "step": 10675 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 2.9689655172413794e-06, | |
| "loss": 0.0223, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 8.62, | |
| "learning_rate": 2.9517241379310347e-06, | |
| "loss": 0.0225, | |
| "step": 10725 | |
| }, | |
| { | |
| "epoch": 8.64, | |
| "learning_rate": 2.9344827586206896e-06, | |
| "loss": 0.022, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 8.66, | |
| "learning_rate": 2.917241379310345e-06, | |
| "loss": 0.0228, | |
| "step": 10775 | |
| }, | |
| { | |
| "epoch": 8.68, | |
| "learning_rate": 2.9e-06, | |
| "loss": 0.0214, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 8.7, | |
| "learning_rate": 2.882758620689655e-06, | |
| "loss": 0.0215, | |
| "step": 10825 | |
| }, | |
| { | |
| "epoch": 8.72, | |
| "learning_rate": 2.8655172413793104e-06, | |
| "loss": 0.023, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 8.74, | |
| "learning_rate": 2.848275862068966e-06, | |
| "loss": 0.022, | |
| "step": 10875 | |
| }, | |
| { | |
| "epoch": 8.76, | |
| "learning_rate": 2.831034482758621e-06, | |
| "loss": 0.0237, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 8.78, | |
| "learning_rate": 2.8137931034482762e-06, | |
| "loss": 0.0204, | |
| "step": 10925 | |
| }, | |
| { | |
| "epoch": 8.8, | |
| "learning_rate": 2.7965517241379315e-06, | |
| "loss": 0.0218, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 8.82, | |
| "learning_rate": 2.7793103448275864e-06, | |
| "loss": 0.0208, | |
| "step": 10975 | |
| }, | |
| { | |
| "epoch": 8.84, | |
| "learning_rate": 2.7620689655172417e-06, | |
| "loss": 0.0241, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 8.84, | |
| "eval_cer": 13.86727806687619, | |
| "eval_loss": 0.45295876264572144, | |
| "eval_runtime": 736.8724, | |
| "eval_samples_per_second": 2.559, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 37.10946034520428, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 8.86, | |
| "learning_rate": 2.744827586206897e-06, | |
| "loss": 0.0211, | |
| "step": 11025 | |
| }, | |
| { | |
| "epoch": 8.88, | |
| "learning_rate": 2.727586206896552e-06, | |
| "loss": 0.0222, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 8.9, | |
| "learning_rate": 2.710344827586207e-06, | |
| "loss": 0.0223, | |
| "step": 11075 | |
| }, | |
| { | |
| "epoch": 8.92, | |
| "learning_rate": 2.6931034482758624e-06, | |
| "loss": 0.023, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 8.94, | |
| "learning_rate": 2.6758620689655173e-06, | |
| "loss": 0.0219, | |
| "step": 11125 | |
| }, | |
| { | |
| "epoch": 8.96, | |
| "learning_rate": 2.6586206896551726e-06, | |
| "loss": 0.022, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 8.98, | |
| "learning_rate": 2.6413793103448275e-06, | |
| "loss": 0.0217, | |
| "step": 11175 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "learning_rate": 2.6241379310344828e-06, | |
| "loss": 0.02, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 9.02, | |
| "learning_rate": 2.606896551724138e-06, | |
| "loss": 0.0126, | |
| "step": 11225 | |
| }, | |
| { | |
| "epoch": 9.04, | |
| "learning_rate": 2.589655172413793e-06, | |
| "loss": 0.0116, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 9.06, | |
| "learning_rate": 2.5724137931034486e-06, | |
| "loss": 0.0113, | |
| "step": 11275 | |
| }, | |
| { | |
| "epoch": 9.08, | |
| "learning_rate": 2.555172413793104e-06, | |
| "loss": 0.0118, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 9.1, | |
| "learning_rate": 2.537931034482759e-06, | |
| "loss": 0.0122, | |
| "step": 11325 | |
| }, | |
| { | |
| "epoch": 9.12, | |
| "learning_rate": 2.520689655172414e-06, | |
| "loss": 0.0123, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 9.14, | |
| "learning_rate": 2.5034482758620694e-06, | |
| "loss": 0.0126, | |
| "step": 11375 | |
| }, | |
| { | |
| "epoch": 9.16, | |
| "learning_rate": 2.4862068965517243e-06, | |
| "loss": 0.0125, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 2.4689655172413796e-06, | |
| "loss": 0.0122, | |
| "step": 11425 | |
| }, | |
| { | |
| "epoch": 9.2, | |
| "learning_rate": 2.451724137931035e-06, | |
| "loss": 0.012, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 9.22, | |
| "learning_rate": 2.4344827586206897e-06, | |
| "loss": 0.0122, | |
| "step": 11475 | |
| }, | |
| { | |
| "epoch": 9.24, | |
| "learning_rate": 2.417241379310345e-06, | |
| "loss": 0.0125, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 9.26, | |
| "learning_rate": 2.4000000000000003e-06, | |
| "loss": 0.013, | |
| "step": 11525 | |
| }, | |
| { | |
| "epoch": 9.28, | |
| "learning_rate": 2.382758620689655e-06, | |
| "loss": 0.0122, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 9.3, | |
| "learning_rate": 2.3655172413793105e-06, | |
| "loss": 0.0119, | |
| "step": 11575 | |
| }, | |
| { | |
| "epoch": 9.32, | |
| "learning_rate": 2.3482758620689658e-06, | |
| "loss": 0.0122, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 9.34, | |
| "learning_rate": 2.331034482758621e-06, | |
| "loss": 0.0125, | |
| "step": 11625 | |
| }, | |
| { | |
| "epoch": 9.36, | |
| "learning_rate": 2.313793103448276e-06, | |
| "loss": 0.0119, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 9.39, | |
| "learning_rate": 2.2965517241379312e-06, | |
| "loss": 0.013, | |
| "step": 11675 | |
| }, | |
| { | |
| "epoch": 9.41, | |
| "learning_rate": 2.2793103448275865e-06, | |
| "loss": 0.0121, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 9.43, | |
| "learning_rate": 2.2620689655172414e-06, | |
| "loss": 0.0123, | |
| "step": 11725 | |
| }, | |
| { | |
| "epoch": 9.45, | |
| "learning_rate": 2.2448275862068967e-06, | |
| "loss": 0.0118, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 9.47, | |
| "learning_rate": 2.227586206896552e-06, | |
| "loss": 0.0123, | |
| "step": 11775 | |
| }, | |
| { | |
| "epoch": 9.49, | |
| "learning_rate": 2.2103448275862073e-06, | |
| "loss": 0.0111, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 9.51, | |
| "learning_rate": 2.193103448275862e-06, | |
| "loss": 0.0121, | |
| "step": 11825 | |
| }, | |
| { | |
| "epoch": 9.53, | |
| "learning_rate": 2.1758620689655174e-06, | |
| "loss": 0.0134, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 9.55, | |
| "learning_rate": 2.1586206896551727e-06, | |
| "loss": 0.0124, | |
| "step": 11875 | |
| }, | |
| { | |
| "epoch": 9.57, | |
| "learning_rate": 2.1413793103448276e-06, | |
| "loss": 0.0116, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 9.59, | |
| "learning_rate": 2.124137931034483e-06, | |
| "loss": 0.0129, | |
| "step": 11925 | |
| }, | |
| { | |
| "epoch": 9.61, | |
| "learning_rate": 2.106896551724138e-06, | |
| "loss": 0.0122, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 9.63, | |
| "learning_rate": 2.0896551724137935e-06, | |
| "loss": 0.0126, | |
| "step": 11975 | |
| }, | |
| { | |
| "epoch": 9.65, | |
| "learning_rate": 2.0724137931034484e-06, | |
| "loss": 0.0123, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 9.65, | |
| "eval_cer": 13.958301131777354, | |
| "eval_loss": 0.47625598311424255, | |
| "eval_runtime": 740.1003, | |
| "eval_samples_per_second": 2.548, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 37.595586628796156, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 9.67, | |
| "learning_rate": 2.0551724137931036e-06, | |
| "loss": 0.0117, | |
| "step": 12025 | |
| }, | |
| { | |
| "epoch": 9.69, | |
| "learning_rate": 2.037931034482759e-06, | |
| "loss": 0.0123, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 9.71, | |
| "learning_rate": 2.020689655172414e-06, | |
| "loss": 0.0125, | |
| "step": 12075 | |
| }, | |
| { | |
| "epoch": 9.73, | |
| "learning_rate": 2.003448275862069e-06, | |
| "loss": 0.0124, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 9.75, | |
| "learning_rate": 1.9862068965517244e-06, | |
| "loss": 0.0122, | |
| "step": 12125 | |
| }, | |
| { | |
| "epoch": 9.77, | |
| "learning_rate": 1.9689655172413793e-06, | |
| "loss": 0.0117, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 9.79, | |
| "learning_rate": 1.9517241379310346e-06, | |
| "loss": 0.0117, | |
| "step": 12175 | |
| }, | |
| { | |
| "epoch": 9.81, | |
| "learning_rate": 1.93448275862069e-06, | |
| "loss": 0.0126, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 9.83, | |
| "learning_rate": 1.917241379310345e-06, | |
| "loss": 0.0125, | |
| "step": 12225 | |
| }, | |
| { | |
| "epoch": 9.85, | |
| "learning_rate": 1.9000000000000002e-06, | |
| "loss": 0.0125, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 9.87, | |
| "learning_rate": 1.8827586206896553e-06, | |
| "loss": 0.0133, | |
| "step": 12275 | |
| }, | |
| { | |
| "epoch": 9.89, | |
| "learning_rate": 1.8655172413793104e-06, | |
| "loss": 0.0121, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 9.91, | |
| "learning_rate": 1.8482758620689657e-06, | |
| "loss": 0.0123, | |
| "step": 12325 | |
| }, | |
| { | |
| "epoch": 9.93, | |
| "learning_rate": 1.8310344827586208e-06, | |
| "loss": 0.013, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 9.95, | |
| "learning_rate": 1.813793103448276e-06, | |
| "loss": 0.0117, | |
| "step": 12375 | |
| }, | |
| { | |
| "epoch": 9.97, | |
| "learning_rate": 1.7965517241379314e-06, | |
| "loss": 0.012, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 9.99, | |
| "learning_rate": 1.7793103448275864e-06, | |
| "loss": 0.0122, | |
| "step": 12425 | |
| }, | |
| { | |
| "epoch": 10.01, | |
| "learning_rate": 1.7620689655172415e-06, | |
| "loss": 0.01, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 10.03, | |
| "learning_rate": 1.7448275862068966e-06, | |
| "loss": 0.0072, | |
| "step": 12475 | |
| }, | |
| { | |
| "epoch": 10.05, | |
| "learning_rate": 1.727586206896552e-06, | |
| "loss": 0.0074, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 10.07, | |
| "learning_rate": 1.710344827586207e-06, | |
| "loss": 0.0074, | |
| "step": 12525 | |
| }, | |
| { | |
| "epoch": 10.09, | |
| "learning_rate": 1.693103448275862e-06, | |
| "loss": 0.0069, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 10.11, | |
| "learning_rate": 1.6758620689655176e-06, | |
| "loss": 0.007, | |
| "step": 12575 | |
| }, | |
| { | |
| "epoch": 10.13, | |
| "learning_rate": 1.6586206896551726e-06, | |
| "loss": 0.0064, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 10.15, | |
| "learning_rate": 1.6413793103448277e-06, | |
| "loss": 0.0069, | |
| "step": 12625 | |
| }, | |
| { | |
| "epoch": 10.17, | |
| "learning_rate": 1.6241379310344828e-06, | |
| "loss": 0.0069, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 10.19, | |
| "learning_rate": 1.6068965517241381e-06, | |
| "loss": 0.0073, | |
| "step": 12675 | |
| }, | |
| { | |
| "epoch": 10.21, | |
| "learning_rate": 1.5896551724137932e-06, | |
| "loss": 0.0076, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 10.23, | |
| "learning_rate": 1.5724137931034483e-06, | |
| "loss": 0.0068, | |
| "step": 12725 | |
| }, | |
| { | |
| "epoch": 10.25, | |
| "learning_rate": 1.5551724137931036e-06, | |
| "loss": 0.0062, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 10.27, | |
| "learning_rate": 1.5379310344827589e-06, | |
| "loss": 0.007, | |
| "step": 12775 | |
| }, | |
| { | |
| "epoch": 10.29, | |
| "learning_rate": 1.520689655172414e-06, | |
| "loss": 0.0069, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 10.31, | |
| "learning_rate": 1.5034482758620692e-06, | |
| "loss": 0.0076, | |
| "step": 12825 | |
| }, | |
| { | |
| "epoch": 10.33, | |
| "learning_rate": 1.4862068965517243e-06, | |
| "loss": 0.0073, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 10.35, | |
| "learning_rate": 1.4689655172413794e-06, | |
| "loss": 0.0069, | |
| "step": 12875 | |
| }, | |
| { | |
| "epoch": 10.37, | |
| "learning_rate": 1.4517241379310345e-06, | |
| "loss": 0.0069, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 10.39, | |
| "learning_rate": 1.4344827586206898e-06, | |
| "loss": 0.0066, | |
| "step": 12925 | |
| }, | |
| { | |
| "epoch": 10.41, | |
| "learning_rate": 1.4172413793103449e-06, | |
| "loss": 0.0067, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 10.43, | |
| "learning_rate": 1.4000000000000001e-06, | |
| "loss": 0.0068, | |
| "step": 12975 | |
| }, | |
| { | |
| "epoch": 10.45, | |
| "learning_rate": 1.3827586206896554e-06, | |
| "loss": 0.007, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 10.45, | |
| "eval_cer": 13.935974719631785, | |
| "eval_loss": 0.49392759799957275, | |
| "eval_runtime": 733.9554, | |
| "eval_samples_per_second": 2.57, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 37.31155778894472, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 10.47, | |
| "learning_rate": 1.3655172413793105e-06, | |
| "loss": 0.0067, | |
| "step": 13025 | |
| }, | |
| { | |
| "epoch": 10.49, | |
| "learning_rate": 1.3482758620689656e-06, | |
| "loss": 0.0066, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 10.51, | |
| "learning_rate": 1.331034482758621e-06, | |
| "loss": 0.0069, | |
| "step": 13075 | |
| }, | |
| { | |
| "epoch": 10.53, | |
| "learning_rate": 1.313793103448276e-06, | |
| "loss": 0.0074, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 10.55, | |
| "learning_rate": 1.296551724137931e-06, | |
| "loss": 0.0066, | |
| "step": 13125 | |
| }, | |
| { | |
| "epoch": 10.57, | |
| "learning_rate": 1.2793103448275861e-06, | |
| "loss": 0.0068, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 10.59, | |
| "learning_rate": 1.2620689655172417e-06, | |
| "loss": 0.0074, | |
| "step": 13175 | |
| }, | |
| { | |
| "epoch": 10.61, | |
| "learning_rate": 1.2448275862068965e-06, | |
| "loss": 0.0072, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 10.63, | |
| "learning_rate": 1.2275862068965518e-06, | |
| "loss": 0.0062, | |
| "step": 13225 | |
| }, | |
| { | |
| "epoch": 10.65, | |
| "learning_rate": 1.2103448275862071e-06, | |
| "loss": 0.007, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 10.67, | |
| "learning_rate": 1.1931034482758622e-06, | |
| "loss": 0.0074, | |
| "step": 13275 | |
| }, | |
| { | |
| "epoch": 10.69, | |
| "learning_rate": 1.1758620689655173e-06, | |
| "loss": 0.0073, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 10.71, | |
| "learning_rate": 1.1586206896551726e-06, | |
| "loss": 0.0066, | |
| "step": 13325 | |
| }, | |
| { | |
| "epoch": 10.73, | |
| "learning_rate": 1.1413793103448276e-06, | |
| "loss": 0.007, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 10.75, | |
| "learning_rate": 1.124137931034483e-06, | |
| "loss": 0.0072, | |
| "step": 13375 | |
| }, | |
| { | |
| "epoch": 10.77, | |
| "learning_rate": 1.106896551724138e-06, | |
| "loss": 0.0067, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 10.79, | |
| "learning_rate": 1.0896551724137933e-06, | |
| "loss": 0.0071, | |
| "step": 13425 | |
| }, | |
| { | |
| "epoch": 10.81, | |
| "learning_rate": 1.0724137931034484e-06, | |
| "loss": 0.0072, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 10.83, | |
| "learning_rate": 1.0551724137931035e-06, | |
| "loss": 0.0065, | |
| "step": 13475 | |
| }, | |
| { | |
| "epoch": 10.85, | |
| "learning_rate": 1.0379310344827588e-06, | |
| "loss": 0.0067, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 10.87, | |
| "learning_rate": 1.0206896551724139e-06, | |
| "loss": 0.0067, | |
| "step": 13525 | |
| }, | |
| { | |
| "epoch": 10.89, | |
| "learning_rate": 1.0034482758620692e-06, | |
| "loss": 0.0066, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 10.91, | |
| "learning_rate": 9.862068965517242e-07, | |
| "loss": 0.0068, | |
| "step": 13575 | |
| }, | |
| { | |
| "epoch": 10.93, | |
| "learning_rate": 9.689655172413793e-07, | |
| "loss": 0.0063, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 10.95, | |
| "learning_rate": 9.517241379310345e-07, | |
| "loss": 0.0069, | |
| "step": 13625 | |
| }, | |
| { | |
| "epoch": 10.97, | |
| "learning_rate": 9.344827586206898e-07, | |
| "loss": 0.0065, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 10.99, | |
| "learning_rate": 9.172413793103449e-07, | |
| "loss": 0.0061, | |
| "step": 13675 | |
| }, | |
| { | |
| "epoch": 11.01, | |
| "learning_rate": 9.000000000000001e-07, | |
| "loss": 0.0054, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 11.03, | |
| "learning_rate": 8.827586206896551e-07, | |
| "loss": 0.0049, | |
| "step": 13725 | |
| }, | |
| { | |
| "epoch": 11.05, | |
| "learning_rate": 8.655172413793104e-07, | |
| "loss": 0.005, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 11.07, | |
| "learning_rate": 8.482758620689656e-07, | |
| "loss": 0.0046, | |
| "step": 13775 | |
| }, | |
| { | |
| "epoch": 11.09, | |
| "learning_rate": 8.310344827586207e-07, | |
| "loss": 0.0051, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 11.11, | |
| "learning_rate": 8.137931034482759e-07, | |
| "loss": 0.0046, | |
| "step": 13825 | |
| }, | |
| { | |
| "epoch": 11.13, | |
| "learning_rate": 7.965517241379311e-07, | |
| "loss": 0.0045, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 11.15, | |
| "learning_rate": 7.793103448275863e-07, | |
| "loss": 0.0048, | |
| "step": 13875 | |
| }, | |
| { | |
| "epoch": 11.17, | |
| "learning_rate": 7.620689655172415e-07, | |
| "loss": 0.0047, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 11.19, | |
| "learning_rate": 7.448275862068965e-07, | |
| "loss": 0.0047, | |
| "step": 13925 | |
| }, | |
| { | |
| "epoch": 11.21, | |
| "learning_rate": 7.275862068965518e-07, | |
| "loss": 0.0047, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 11.23, | |
| "learning_rate": 7.103448275862069e-07, | |
| "loss": 0.0045, | |
| "step": 13975 | |
| }, | |
| { | |
| "epoch": 11.25, | |
| "learning_rate": 6.931034482758621e-07, | |
| "loss": 0.0047, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 11.25, | |
| "eval_cer": 13.810603328352828, | |
| "eval_loss": 0.5054397583007812, | |
| "eval_runtime": 736.0874, | |
| "eval_samples_per_second": 2.562, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 37.175005462093075, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 11.27, | |
| "learning_rate": 6.758620689655172e-07, | |
| "loss": 0.0046, | |
| "step": 14025 | |
| }, | |
| { | |
| "epoch": 11.29, | |
| "learning_rate": 6.586206896551725e-07, | |
| "loss": 0.0046, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 11.31, | |
| "learning_rate": 6.413793103448277e-07, | |
| "loss": 0.0043, | |
| "step": 14075 | |
| }, | |
| { | |
| "epoch": 11.33, | |
| "learning_rate": 6.241379310344828e-07, | |
| "loss": 0.0046, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 11.35, | |
| "learning_rate": 6.068965517241379e-07, | |
| "loss": 0.0049, | |
| "step": 14125 | |
| }, | |
| { | |
| "epoch": 11.37, | |
| "learning_rate": 5.896551724137931e-07, | |
| "loss": 0.0048, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 11.39, | |
| "learning_rate": 5.724137931034483e-07, | |
| "loss": 0.0047, | |
| "step": 14175 | |
| }, | |
| { | |
| "epoch": 11.41, | |
| "learning_rate": 5.551724137931035e-07, | |
| "loss": 0.0043, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 11.43, | |
| "learning_rate": 5.379310344827587e-07, | |
| "loss": 0.0044, | |
| "step": 14225 | |
| }, | |
| { | |
| "epoch": 11.45, | |
| "learning_rate": 5.206896551724138e-07, | |
| "loss": 0.0045, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 11.48, | |
| "learning_rate": 5.03448275862069e-07, | |
| "loss": 0.0045, | |
| "step": 14275 | |
| }, | |
| { | |
| "epoch": 11.5, | |
| "learning_rate": 4.862068965517241e-07, | |
| "loss": 0.0049, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 11.52, | |
| "learning_rate": 4.6896551724137934e-07, | |
| "loss": 0.0046, | |
| "step": 14325 | |
| }, | |
| { | |
| "epoch": 11.54, | |
| "learning_rate": 4.5172413793103447e-07, | |
| "loss": 0.0047, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 11.56, | |
| "learning_rate": 4.344827586206897e-07, | |
| "loss": 0.0045, | |
| "step": 14375 | |
| }, | |
| { | |
| "epoch": 11.58, | |
| "learning_rate": 4.1724137931034485e-07, | |
| "loss": 0.0045, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 11.6, | |
| "learning_rate": 4.0000000000000003e-07, | |
| "loss": 0.0043, | |
| "step": 14425 | |
| }, | |
| { | |
| "epoch": 11.62, | |
| "learning_rate": 3.8275862068965517e-07, | |
| "loss": 0.0045, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 11.64, | |
| "learning_rate": 3.6551724137931036e-07, | |
| "loss": 0.0045, | |
| "step": 14475 | |
| }, | |
| { | |
| "epoch": 11.66, | |
| "learning_rate": 3.4827586206896555e-07, | |
| "loss": 0.005, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 11.68, | |
| "learning_rate": 3.3103448275862073e-07, | |
| "loss": 0.0046, | |
| "step": 14525 | |
| }, | |
| { | |
| "epoch": 11.7, | |
| "learning_rate": 3.1379310344827587e-07, | |
| "loss": 0.0048, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 11.72, | |
| "learning_rate": 2.9655172413793106e-07, | |
| "loss": 0.005, | |
| "step": 14575 | |
| }, | |
| { | |
| "epoch": 11.74, | |
| "learning_rate": 2.7931034482758624e-07, | |
| "loss": 0.0046, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 11.76, | |
| "learning_rate": 2.6206896551724143e-07, | |
| "loss": 0.0046, | |
| "step": 14625 | |
| }, | |
| { | |
| "epoch": 11.78, | |
| "learning_rate": 2.4482758620689657e-07, | |
| "loss": 0.0045, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 11.8, | |
| "learning_rate": 2.2758620689655175e-07, | |
| "loss": 0.0045, | |
| "step": 14675 | |
| }, | |
| { | |
| "epoch": 11.82, | |
| "learning_rate": 2.1034482758620692e-07, | |
| "loss": 0.0045, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 11.84, | |
| "learning_rate": 1.931034482758621e-07, | |
| "loss": 0.0043, | |
| "step": 14725 | |
| }, | |
| { | |
| "epoch": 11.86, | |
| "learning_rate": 1.7586206896551726e-07, | |
| "loss": 0.0044, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 11.88, | |
| "learning_rate": 1.5862068965517243e-07, | |
| "loss": 0.0048, | |
| "step": 14775 | |
| }, | |
| { | |
| "epoch": 11.9, | |
| "learning_rate": 1.413793103448276e-07, | |
| "loss": 0.0049, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 11.92, | |
| "learning_rate": 1.2413793103448277e-07, | |
| "loss": 0.0042, | |
| "step": 14825 | |
| }, | |
| { | |
| "epoch": 11.94, | |
| "learning_rate": 1.0689655172413794e-07, | |
| "loss": 0.0045, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 11.96, | |
| "learning_rate": 8.965517241379311e-08, | |
| "loss": 0.0046, | |
| "step": 14875 | |
| }, | |
| { | |
| "epoch": 11.98, | |
| "learning_rate": 7.241379310344829e-08, | |
| "loss": 0.0048, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "learning_rate": 5.517241379310345e-08, | |
| "loss": 0.0042, | |
| "step": 14925 | |
| }, | |
| { | |
| "epoch": 12.02, | |
| "learning_rate": 3.793103448275862e-08, | |
| "loss": 0.0039, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 12.04, | |
| "learning_rate": 2.0689655172413796e-08, | |
| "loss": 0.0038, | |
| "step": 14975 | |
| }, | |
| { | |
| "epoch": 12.06, | |
| "learning_rate": 3.4482758620689654e-09, | |
| "loss": 0.0036, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 12.06, | |
| "eval_cer": 13.893039311659539, | |
| "eval_loss": 0.5093265175819397, | |
| "eval_runtime": 735.2398, | |
| "eval_samples_per_second": 2.565, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 37.5081931396111, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 12.06, | |
| "step": 15000, | |
| "total_flos": 1.3843787955830784e+20, | |
| "train_loss": 0.159961280053854, | |
| "train_runtime": 44667.1197, | |
| "train_samples_per_second": 10.746, | |
| "train_steps_per_second": 0.336 | |
| }, | |
| { | |
| "epoch": 12.06, | |
| "eval_cer": 13.310835179555877, | |
| "eval_loss": 0.3296343684196472, | |
| "eval_runtime": 742.0911, | |
| "eval_samples_per_second": 2.541, | |
| "eval_steps_per_second": 0.08, | |
| "eval_wer": 35.8859514966135, | |
| "step": 15000 | |
| } | |
| ], | |
| "max_steps": 15000, | |
| "num_train_epochs": 13, | |
| "total_flos": 1.3843787955830784e+20, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |