{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.0045333998231974065, "eval_steps": 4, "global_step": 50, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 9.066799646394814e-05, "grad_norm": 1.305655837059021, "learning_rate": 2e-05, "loss": 0.9466, "step": 1 }, { "epoch": 9.066799646394814e-05, "eval_loss": 1.1514371633529663, "eval_runtime": 281.757, "eval_samples_per_second": 8.241, "eval_steps_per_second": 8.241, "step": 1 }, { "epoch": 0.0001813359929278963, "grad_norm": 1.7540494203567505, "learning_rate": 4e-05, "loss": 1.2689, "step": 2 }, { "epoch": 0.00027200398939184443, "grad_norm": 1.3806679248809814, "learning_rate": 6e-05, "loss": 0.9866, "step": 3 }, { "epoch": 0.0003626719858557926, "grad_norm": 1.6540546417236328, "learning_rate": 8e-05, "loss": 0.886, "step": 4 }, { "epoch": 0.0003626719858557926, "eval_loss": 1.063585877418518, "eval_runtime": 281.615, "eval_samples_per_second": 8.245, "eval_steps_per_second": 8.245, "step": 4 }, { "epoch": 0.00045333998231974066, "grad_norm": 1.351820945739746, "learning_rate": 0.0001, "loss": 1.0148, "step": 5 }, { "epoch": 0.0005440079787836889, "grad_norm": 1.6633617877960205, "learning_rate": 0.00012, "loss": 1.1922, "step": 6 }, { "epoch": 0.000634675975247637, "grad_norm": 1.216033697128296, "learning_rate": 0.00014, "loss": 0.7838, "step": 7 }, { "epoch": 0.0007253439717115851, "grad_norm": 1.0886664390563965, "learning_rate": 0.00016, "loss": 0.6905, "step": 8 }, { "epoch": 0.0007253439717115851, "eval_loss": 0.6156781911849976, "eval_runtime": 282.2824, "eval_samples_per_second": 8.226, "eval_steps_per_second": 8.226, "step": 8 }, { "epoch": 0.0008160119681755333, "grad_norm": 1.2936314344406128, "learning_rate": 0.00018, "loss": 0.6603, "step": 9 }, { "epoch": 0.0009066799646394813, "grad_norm": 1.429738163948059, "learning_rate": 0.0002, "loss": 0.518, "step": 10 }, { "epoch": 0.0009973479611034295, "grad_norm": 1.501217007637024, "learning_rate": 0.0001996917333733128, "loss": 0.475, "step": 11 }, { "epoch": 0.0010880159575673777, "grad_norm": 1.0113286972045898, "learning_rate": 0.00019876883405951377, "loss": 0.3233, "step": 12 }, { "epoch": 0.0010880159575673777, "eval_loss": 0.29599475860595703, "eval_runtime": 282.4723, "eval_samples_per_second": 8.22, "eval_steps_per_second": 8.22, "step": 12 }, { "epoch": 0.0011786839540313258, "grad_norm": 1.045129656791687, "learning_rate": 0.00019723699203976766, "loss": 0.305, "step": 13 }, { "epoch": 0.001269351950495274, "grad_norm": 0.7782332897186279, "learning_rate": 0.00019510565162951537, "loss": 0.2367, "step": 14 }, { "epoch": 0.001360019946959222, "grad_norm": 1.1300699710845947, "learning_rate": 0.0001923879532511287, "loss": 0.1923, "step": 15 }, { "epoch": 0.0014506879434231703, "grad_norm": 0.6594939231872559, "learning_rate": 0.0001891006524188368, "loss": 0.1823, "step": 16 }, { "epoch": 0.0014506879434231703, "eval_loss": 0.2610986828804016, "eval_runtime": 282.2669, "eval_samples_per_second": 8.226, "eval_steps_per_second": 8.226, "step": 16 }, { "epoch": 0.0015413559398871183, "grad_norm": 1.0713460445404053, "learning_rate": 0.00018526401643540922, "loss": 0.2457, "step": 17 }, { "epoch": 0.0016320239363510666, "grad_norm": 1.2344204187393188, "learning_rate": 0.00018090169943749476, "loss": 0.272, "step": 18 }, { "epoch": 0.0017226919328150146, "grad_norm": 0.7098881006240845, "learning_rate": 0.0001760405965600031, "loss": 0.225, "step": 19 }, { "epoch": 0.0018133599292789627, "grad_norm": 0.8920848965644836, "learning_rate": 0.00017071067811865476, "loss": 0.1479, "step": 20 }, { "epoch": 0.0018133599292789627, "eval_loss": 0.23455704748630524, "eval_runtime": 282.1688, "eval_samples_per_second": 8.229, "eval_steps_per_second": 8.229, "step": 20 }, { "epoch": 0.001904027925742911, "grad_norm": 0.5211899876594543, "learning_rate": 0.00016494480483301836, "loss": 0.1664, "step": 21 }, { "epoch": 0.001994695922206859, "grad_norm": 0.6904815435409546, "learning_rate": 0.00015877852522924732, "loss": 0.2361, "step": 22 }, { "epoch": 0.002085363918670807, "grad_norm": 0.6779294013977051, "learning_rate": 0.0001522498564715949, "loss": 0.1939, "step": 23 }, { "epoch": 0.0021760319151347554, "grad_norm": 0.8565766215324402, "learning_rate": 0.00014539904997395468, "loss": 0.2695, "step": 24 }, { "epoch": 0.0021760319151347554, "eval_loss": 0.22548052668571472, "eval_runtime": 282.2744, "eval_samples_per_second": 8.226, "eval_steps_per_second": 8.226, "step": 24 }, { "epoch": 0.0022666999115987033, "grad_norm": 0.7228334546089172, "learning_rate": 0.000138268343236509, "loss": 0.2259, "step": 25 }, { "epoch": 0.0023573679080626515, "grad_norm": 0.9707162976264954, "learning_rate": 0.00013090169943749476, "loss": 0.2173, "step": 26 }, { "epoch": 0.0024480359045265998, "grad_norm": 0.5939818024635315, "learning_rate": 0.00012334453638559057, "loss": 0.2012, "step": 27 }, { "epoch": 0.002538703900990548, "grad_norm": 0.3913775682449341, "learning_rate": 0.0001156434465040231, "loss": 0.1645, "step": 28 }, { "epoch": 0.002538703900990548, "eval_loss": 0.21409040689468384, "eval_runtime": 282.2725, "eval_samples_per_second": 8.226, "eval_steps_per_second": 8.226, "step": 28 }, { "epoch": 0.002629371897454496, "grad_norm": 0.652338981628418, "learning_rate": 0.0001078459095727845, "loss": 0.1954, "step": 29 }, { "epoch": 0.002720039893918444, "grad_norm": 0.5130963325500488, "learning_rate": 0.0001, "loss": 0.1994, "step": 30 }, { "epoch": 0.0028107078903823923, "grad_norm": 0.38596802949905396, "learning_rate": 9.215409042721552e-05, "loss": 0.1804, "step": 31 }, { "epoch": 0.0029013758868463406, "grad_norm": 0.481663316488266, "learning_rate": 8.435655349597689e-05, "loss": 0.1804, "step": 32 }, { "epoch": 0.0029013758868463406, "eval_loss": 0.20881743729114532, "eval_runtime": 281.3146, "eval_samples_per_second": 8.254, "eval_steps_per_second": 8.254, "step": 32 }, { "epoch": 0.0029920438833102884, "grad_norm": 0.3412702679634094, "learning_rate": 7.66554636144095e-05, "loss": 0.2066, "step": 33 }, { "epoch": 0.0030827118797742367, "grad_norm": 0.5542587041854858, "learning_rate": 6.909830056250527e-05, "loss": 0.1895, "step": 34 }, { "epoch": 0.003173379876238185, "grad_norm": 0.5084295868873596, "learning_rate": 6.173165676349103e-05, "loss": 0.2111, "step": 35 }, { "epoch": 0.003264047872702133, "grad_norm": 0.6083145141601562, "learning_rate": 5.4600950026045326e-05, "loss": 0.2318, "step": 36 }, { "epoch": 0.003264047872702133, "eval_loss": 0.20627427101135254, "eval_runtime": 282.4785, "eval_samples_per_second": 8.22, "eval_steps_per_second": 8.22, "step": 36 }, { "epoch": 0.003354715869166081, "grad_norm": 0.4692092835903168, "learning_rate": 4.7750143528405126e-05, "loss": 0.1626, "step": 37 }, { "epoch": 0.0034453838656300292, "grad_norm": 0.4230363965034485, "learning_rate": 4.12214747707527e-05, "loss": 0.1752, "step": 38 }, { "epoch": 0.0035360518620939775, "grad_norm": 0.5133092403411865, "learning_rate": 3.5055195166981645e-05, "loss": 0.209, "step": 39 }, { "epoch": 0.0036267198585579253, "grad_norm": 0.3762070834636688, "learning_rate": 2.9289321881345254e-05, "loss": 0.1563, "step": 40 }, { "epoch": 0.0036267198585579253, "eval_loss": 0.20195676386356354, "eval_runtime": 282.3618, "eval_samples_per_second": 8.223, "eval_steps_per_second": 8.223, "step": 40 }, { "epoch": 0.0037173878550218736, "grad_norm": 0.3519875407218933, "learning_rate": 2.3959403439996907e-05, "loss": 0.1516, "step": 41 }, { "epoch": 0.003808055851485822, "grad_norm": 0.41183072328567505, "learning_rate": 1.9098300562505266e-05, "loss": 0.1585, "step": 42 }, { "epoch": 0.00389872384794977, "grad_norm": 0.613614022731781, "learning_rate": 1.4735983564590783e-05, "loss": 0.1563, "step": 43 }, { "epoch": 0.003989391844413718, "grad_norm": 0.4138711988925934, "learning_rate": 1.0899347581163221e-05, "loss": 0.1512, "step": 44 }, { "epoch": 0.003989391844413718, "eval_loss": 0.20044909417629242, "eval_runtime": 282.2051, "eval_samples_per_second": 8.228, "eval_steps_per_second": 8.228, "step": 44 }, { "epoch": 0.004080059840877667, "grad_norm": 0.4004136025905609, "learning_rate": 7.612046748871327e-06, "loss": 0.1641, "step": 45 }, { "epoch": 0.004170727837341614, "grad_norm": 0.4379856586456299, "learning_rate": 4.8943483704846475e-06, "loss": 0.1492, "step": 46 }, { "epoch": 0.004261395833805562, "grad_norm": 0.32673540711402893, "learning_rate": 2.7630079602323442e-06, "loss": 0.1826, "step": 47 }, { "epoch": 0.004352063830269511, "grad_norm": 0.2924938499927521, "learning_rate": 1.231165940486234e-06, "loss": 0.1529, "step": 48 }, { "epoch": 0.004352063830269511, "eval_loss": 0.20006079971790314, "eval_runtime": 281.9182, "eval_samples_per_second": 8.236, "eval_steps_per_second": 8.236, "step": 48 }, { "epoch": 0.004442731826733459, "grad_norm": 0.44641199707984924, "learning_rate": 3.0826662668720364e-07, "loss": 0.1428, "step": 49 }, { "epoch": 0.0045333998231974065, "grad_norm": 0.2606041431427002, "learning_rate": 0.0, "loss": 0.1358, "step": 50 } ], "logging_steps": 1, "max_steps": 50, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 4, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.8825716774731776e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }