{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.4149274849663955, "eval_steps": 1000, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.017686593562079942, "grad_norm": 28.968021392822266, "learning_rate": 4.6000000000000004e-07, "loss": 3.698, "step": 25 }, { "epoch": 0.035373187124159884, "grad_norm": 21.002214431762695, "learning_rate": 9.600000000000001e-07, "loss": 3.5124, "step": 50 }, { "epoch": 0.05305978068623983, "grad_norm": 21.64297866821289, "learning_rate": 1.46e-06, "loss": 3.2335, "step": 75 }, { "epoch": 0.07074637424831977, "grad_norm": 19.800973892211914, "learning_rate": 1.9600000000000003e-06, "loss": 2.9933, "step": 100 }, { "epoch": 0.08843296781039972, "grad_norm": 19.281307220458984, "learning_rate": 2.46e-06, "loss": 2.5097, "step": 125 }, { "epoch": 0.10611956137247966, "grad_norm": 19.187774658203125, "learning_rate": 2.96e-06, "loss": 2.3572, "step": 150 }, { "epoch": 0.1238061549345596, "grad_norm": 17.483562469482422, "learning_rate": 3.46e-06, "loss": 2.1479, "step": 175 }, { "epoch": 0.14149274849663954, "grad_norm": 17.524105072021484, "learning_rate": 3.96e-06, "loss": 2.2237, "step": 200 }, { "epoch": 0.1591793420587195, "grad_norm": 17.354618072509766, "learning_rate": 4.4600000000000005e-06, "loss": 2.0436, "step": 225 }, { "epoch": 0.17686593562079944, "grad_norm": 18.539562225341797, "learning_rate": 4.960000000000001e-06, "loss": 2.0212, "step": 250 }, { "epoch": 0.19455252918287938, "grad_norm": 18.09463882446289, "learning_rate": 5.460000000000001e-06, "loss": 1.9828, "step": 275 }, { "epoch": 0.21223912274495932, "grad_norm": 16.73234748840332, "learning_rate": 5.9600000000000005e-06, "loss": 1.9199, "step": 300 }, { "epoch": 0.22992571630703926, "grad_norm": 17.19611930847168, "learning_rate": 6.460000000000001e-06, "loss": 1.8943, "step": 325 }, { "epoch": 0.2476123098691192, "grad_norm": 18.489797592163086, "learning_rate": 6.96e-06, "loss": 1.7673, "step": 350 }, { "epoch": 0.26529890343119916, "grad_norm": 17.019506454467773, "learning_rate": 7.4600000000000006e-06, "loss": 1.6931, "step": 375 }, { "epoch": 0.2829854969932791, "grad_norm": 16.345083236694336, "learning_rate": 7.960000000000002e-06, "loss": 1.6946, "step": 400 }, { "epoch": 0.30067209055535904, "grad_norm": 16.138513565063477, "learning_rate": 8.46e-06, "loss": 1.6849, "step": 425 }, { "epoch": 0.318358684117439, "grad_norm": 17.647178649902344, "learning_rate": 8.96e-06, "loss": 1.677, "step": 450 }, { "epoch": 0.3360452776795189, "grad_norm": 17.51073455810547, "learning_rate": 9.460000000000001e-06, "loss": 1.608, "step": 475 }, { "epoch": 0.3537318712415989, "grad_norm": 15.542696952819824, "learning_rate": 9.960000000000001e-06, "loss": 1.5617, "step": 500 }, { "epoch": 0.3714184648036788, "grad_norm": 14.3616361618042, "learning_rate": 9.846666666666668e-06, "loss": 1.4832, "step": 525 }, { "epoch": 0.38910505836575876, "grad_norm": 14.655681610107422, "learning_rate": 9.68e-06, "loss": 1.521, "step": 550 }, { "epoch": 0.40679165192783867, "grad_norm": 15.943355560302734, "learning_rate": 9.513333333333334e-06, "loss": 1.4946, "step": 575 }, { "epoch": 0.42447824548991864, "grad_norm": 17.380294799804688, "learning_rate": 9.346666666666666e-06, "loss": 1.4358, "step": 600 }, { "epoch": 0.4421648390519986, "grad_norm": 16.160900115966797, "learning_rate": 9.180000000000002e-06, "loss": 1.4093, "step": 625 }, { "epoch": 0.4598514326140785, "grad_norm": 14.763204574584961, "learning_rate": 9.013333333333334e-06, "loss": 1.4101, "step": 650 }, { "epoch": 0.4775380261761585, "grad_norm": 15.164654731750488, "learning_rate": 8.846666666666668e-06, "loss": 1.3897, "step": 675 }, { "epoch": 0.4952246197382384, "grad_norm": 15.769131660461426, "learning_rate": 8.68e-06, "loss": 1.3422, "step": 700 }, { "epoch": 0.5129112133003184, "grad_norm": 16.859813690185547, "learning_rate": 8.513333333333335e-06, "loss": 1.3451, "step": 725 }, { "epoch": 0.5305978068623983, "grad_norm": 14.078543663024902, "learning_rate": 8.346666666666668e-06, "loss": 1.3333, "step": 750 }, { "epoch": 0.5482844004244782, "grad_norm": 13.418024063110352, "learning_rate": 8.18e-06, "loss": 1.3618, "step": 775 }, { "epoch": 0.5659709939865581, "grad_norm": 14.887019157409668, "learning_rate": 8.013333333333333e-06, "loss": 1.2983, "step": 800 }, { "epoch": 0.5836575875486382, "grad_norm": 13.075471878051758, "learning_rate": 7.846666666666667e-06, "loss": 1.3021, "step": 825 }, { "epoch": 0.6013441811107181, "grad_norm": 14.660274505615234, "learning_rate": 7.680000000000001e-06, "loss": 1.2375, "step": 850 }, { "epoch": 0.619030774672798, "grad_norm": 16.32628631591797, "learning_rate": 7.513333333333334e-06, "loss": 1.3038, "step": 875 }, { "epoch": 0.636717368234878, "grad_norm": 17.701143264770508, "learning_rate": 7.346666666666668e-06, "loss": 1.2082, "step": 900 }, { "epoch": 0.6544039617969579, "grad_norm": 14.465027809143066, "learning_rate": 7.180000000000001e-06, "loss": 1.2543, "step": 925 }, { "epoch": 0.6720905553590378, "grad_norm": 12.923417091369629, "learning_rate": 7.0133333333333345e-06, "loss": 1.2544, "step": 950 }, { "epoch": 0.6897771489211177, "grad_norm": 15.675958633422852, "learning_rate": 6.846666666666667e-06, "loss": 1.2279, "step": 975 }, { "epoch": 0.7074637424831978, "grad_norm": 15.041681289672852, "learning_rate": 6.680000000000001e-06, "loss": 1.2368, "step": 1000 }, { "epoch": 0.7074637424831978, "eval_loss": 0.6446413397789001, "eval_runtime": 2060.0729, "eval_samples_per_second": 5.225, "eval_steps_per_second": 0.327, "eval_wer": 0.965980208088868, "step": 1000 }, { "epoch": 0.7251503360452777, "grad_norm": 15.67192554473877, "learning_rate": 6.513333333333333e-06, "loss": 1.1917, "step": 1025 }, { "epoch": 0.7428369296073576, "grad_norm": 14.699666023254395, "learning_rate": 6.346666666666668e-06, "loss": 1.1812, "step": 1050 }, { "epoch": 0.7605235231694376, "grad_norm": 14.321735382080078, "learning_rate": 6.18e-06, "loss": 1.2046, "step": 1075 }, { "epoch": 0.7782101167315175, "grad_norm": 14.692147254943848, "learning_rate": 6.013333333333335e-06, "loss": 1.2071, "step": 1100 }, { "epoch": 0.7958967102935974, "grad_norm": 13.778362274169922, "learning_rate": 5.846666666666667e-06, "loss": 1.1476, "step": 1125 }, { "epoch": 0.8135833038556773, "grad_norm": 13.074742317199707, "learning_rate": 5.68e-06, "loss": 1.1543, "step": 1150 }, { "epoch": 0.8312698974177574, "grad_norm": 14.026896476745605, "learning_rate": 5.513333333333334e-06, "loss": 1.1841, "step": 1175 }, { "epoch": 0.8489564909798373, "grad_norm": 13.244227409362793, "learning_rate": 5.346666666666667e-06, "loss": 1.1698, "step": 1200 }, { "epoch": 0.8666430845419172, "grad_norm": 15.449938774108887, "learning_rate": 5.18e-06, "loss": 1.1213, "step": 1225 }, { "epoch": 0.8843296781039972, "grad_norm": 13.706199645996094, "learning_rate": 5.013333333333333e-06, "loss": 1.1935, "step": 1250 }, { "epoch": 0.9020162716660771, "grad_norm": 14.426807403564453, "learning_rate": 4.846666666666667e-06, "loss": 1.0659, "step": 1275 }, { "epoch": 0.919702865228157, "grad_norm": 14.945075035095215, "learning_rate": 4.680000000000001e-06, "loss": 1.1169, "step": 1300 }, { "epoch": 0.937389458790237, "grad_norm": 13.335543632507324, "learning_rate": 4.513333333333333e-06, "loss": 1.0949, "step": 1325 }, { "epoch": 0.955076052352317, "grad_norm": 15.83086109161377, "learning_rate": 4.346666666666667e-06, "loss": 1.0845, "step": 1350 }, { "epoch": 0.9727626459143969, "grad_norm": 12.246453285217285, "learning_rate": 4.18e-06, "loss": 1.148, "step": 1375 }, { "epoch": 0.9904492394764768, "grad_norm": 16.202903747558594, "learning_rate": 4.013333333333334e-06, "loss": 1.1118, "step": 1400 }, { "epoch": 1.0081358330385568, "grad_norm": 12.500658988952637, "learning_rate": 3.8466666666666665e-06, "loss": 1.0352, "step": 1425 }, { "epoch": 1.0258224266006368, "grad_norm": 11.863042831420898, "learning_rate": 3.6800000000000003e-06, "loss": 1.0224, "step": 1450 }, { "epoch": 1.0435090201627166, "grad_norm": 13.093315124511719, "learning_rate": 3.5133333333333337e-06, "loss": 0.966, "step": 1475 }, { "epoch": 1.0611956137247966, "grad_norm": 11.886918067932129, "learning_rate": 3.346666666666667e-06, "loss": 0.9643, "step": 1500 }, { "epoch": 1.0788822072868764, "grad_norm": 15.925506591796875, "learning_rate": 3.1800000000000005e-06, "loss": 0.8792, "step": 1525 }, { "epoch": 1.0965688008489565, "grad_norm": 12.430258750915527, "learning_rate": 3.013333333333334e-06, "loss": 0.9462, "step": 1550 }, { "epoch": 1.1142553944110365, "grad_norm": 13.365354537963867, "learning_rate": 2.8466666666666672e-06, "loss": 0.9246, "step": 1575 }, { "epoch": 1.1319419879731163, "grad_norm": 13.153193473815918, "learning_rate": 2.68e-06, "loss": 0.8996, "step": 1600 }, { "epoch": 1.1496285815351963, "grad_norm": 13.20574951171875, "learning_rate": 2.5133333333333336e-06, "loss": 0.943, "step": 1625 }, { "epoch": 1.1673151750972763, "grad_norm": 10.716314315795898, "learning_rate": 2.346666666666667e-06, "loss": 0.9071, "step": 1650 }, { "epoch": 1.1850017686593561, "grad_norm": 15.43217658996582, "learning_rate": 2.1800000000000003e-06, "loss": 0.9232, "step": 1675 }, { "epoch": 1.2026883622214362, "grad_norm": 14.106558799743652, "learning_rate": 2.0133333333333337e-06, "loss": 0.8998, "step": 1700 }, { "epoch": 1.2203749557835162, "grad_norm": 12.580400466918945, "learning_rate": 1.8466666666666668e-06, "loss": 0.928, "step": 1725 }, { "epoch": 1.238061549345596, "grad_norm": 12.888578414916992, "learning_rate": 1.6800000000000002e-06, "loss": 0.9354, "step": 1750 }, { "epoch": 1.255748142907676, "grad_norm": 11.692194938659668, "learning_rate": 1.5133333333333334e-06, "loss": 0.8441, "step": 1775 }, { "epoch": 1.2734347364697558, "grad_norm": 14.173469543457031, "learning_rate": 1.3466666666666668e-06, "loss": 0.925, "step": 1800 }, { "epoch": 1.2911213300318358, "grad_norm": 12.65889835357666, "learning_rate": 1.1800000000000001e-06, "loss": 0.905, "step": 1825 }, { "epoch": 1.3088079235939158, "grad_norm": 12.559049606323242, "learning_rate": 1.0133333333333333e-06, "loss": 0.8771, "step": 1850 }, { "epoch": 1.3264945171559956, "grad_norm": 11.348992347717285, "learning_rate": 8.466666666666668e-07, "loss": 0.9033, "step": 1875 }, { "epoch": 1.3441811107180757, "grad_norm": 11.88202953338623, "learning_rate": 6.800000000000001e-07, "loss": 0.8451, "step": 1900 }, { "epoch": 1.3618677042801557, "grad_norm": 11.244508743286133, "learning_rate": 5.133333333333334e-07, "loss": 0.8839, "step": 1925 }, { "epoch": 1.3795542978422355, "grad_norm": 12.49364185333252, "learning_rate": 3.466666666666667e-07, "loss": 0.8928, "step": 1950 }, { "epoch": 1.3972408914043155, "grad_norm": 11.834848403930664, "learning_rate": 1.8e-07, "loss": 0.9105, "step": 1975 }, { "epoch": 1.4149274849663955, "grad_norm": 12.054742813110352, "learning_rate": 1.3333333333333334e-08, "loss": 0.8994, "step": 2000 }, { "epoch": 1.4149274849663955, "eval_loss": 0.5812551379203796, "eval_runtime": 1719.3277, "eval_samples_per_second": 6.26, "eval_steps_per_second": 0.391, "eval_wer": 1.254253696315419, "step": 2000 }, { "epoch": 1.4149274849663955, "step": 2000, "total_flos": 4.15045601427456e+18, "train_loss": 1.393887762069702, "train_runtime": 12787.317, "train_samples_per_second": 5.005, "train_steps_per_second": 0.156 } ], "logging_steps": 25, "max_steps": 2000, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.15045601427456e+18, "train_batch_size": 16, "trial_name": null, "trial_params": null }