{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.963109354413702, "eval_steps": 500, "global_step": 376, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01, "grad_norm": 476.0, "learning_rate": 5.263157894736842e-06, "loss": 13.5076, "step": 1 }, { "epoch": 0.05, "grad_norm": 27.125, "learning_rate": 2.6315789473684212e-05, "loss": 10.3382, "step": 5 }, { "epoch": 0.11, "grad_norm": 7.09375, "learning_rate": 5.2631578947368424e-05, "loss": 6.5621, "step": 10 }, { "epoch": 0.16, "grad_norm": 5.09375, "learning_rate": 7.894736842105263e-05, "loss": 5.4048, "step": 15 }, { "epoch": 0.21, "grad_norm": 13.625, "learning_rate": 0.00010526315789473685, "loss": 3.8714, "step": 20 }, { "epoch": 0.26, "grad_norm": 4.3125, "learning_rate": 0.00013157894736842108, "loss": 1.907, "step": 25 }, { "epoch": 0.32, "grad_norm": 2.484375, "learning_rate": 0.00015789473684210527, "loss": 1.7156, "step": 30 }, { "epoch": 0.37, "grad_norm": 1.71875, "learning_rate": 0.00018421052631578948, "loss": 1.5905, "step": 35 }, { "epoch": 0.42, "grad_norm": 1.4375, "learning_rate": 0.00019998272238333606, "loss": 1.4996, "step": 40 }, { "epoch": 0.47, "grad_norm": 1.2109375, "learning_rate": 0.00019978841775475367, "loss": 1.424, "step": 45 }, { "epoch": 0.53, "grad_norm": 2.515625, "learning_rate": 0.00019937863245275304, "loss": 1.3526, "step": 50 }, { "epoch": 0.58, "grad_norm": 1.6953125, "learning_rate": 0.0001987542513577122, "loss": 1.3334, "step": 55 }, { "epoch": 0.63, "grad_norm": 2.3125, "learning_rate": 0.00019791662274292637, "loss": 1.2863, "step": 60 }, { "epoch": 0.69, "grad_norm": 0.91015625, "learning_rate": 0.00019686755536317945, "loss": 1.2384, "step": 65 }, { "epoch": 0.74, "grad_norm": 0.99609375, "learning_rate": 0.00019560931454896298, "loss": 1.2248, "step": 70 }, { "epoch": 0.79, "grad_norm": 0.73046875, "learning_rate": 0.000194144617314777, "loss": 1.1772, "step": 75 }, { "epoch": 0.84, "grad_norm": 0.58203125, "learning_rate": 0.0001924766264920751, "loss": 1.1654, "step": 80 }, { "epoch": 0.9, "grad_norm": 0.74609375, "learning_rate": 0.00019060894389952328, "loss": 1.1382, "step": 85 }, { "epoch": 0.95, "grad_norm": 0.78515625, "learning_rate": 0.000188545602565321, "loss": 1.1537, "step": 90 }, { "epoch": 0.99, "eval_loss": 1.0988351106643677, "eval_runtime": 16.9422, "eval_samples_per_second": 16.822, "eval_steps_per_second": 2.125, "step": 94 }, { "epoch": 1.0, "grad_norm": 0.65625, "learning_rate": 0.00018629105801837818, "loss": 1.1162, "step": 95 }, { "epoch": 1.05, "grad_norm": 0.84375, "learning_rate": 0.00018385017866715507, "loss": 0.9838, "step": 100 }, { "epoch": 1.11, "grad_norm": 0.61328125, "learning_rate": 0.00018122823528693966, "loss": 0.9895, "step": 105 }, { "epoch": 1.16, "grad_norm": 0.8359375, "learning_rate": 0.00017843088963826435, "loss": 0.9588, "step": 110 }, { "epoch": 1.21, "grad_norm": 0.74609375, "learning_rate": 0.00017546418224103838, "loss": 0.9304, "step": 115 }, { "epoch": 1.26, "grad_norm": 0.63671875, "learning_rate": 0.00017233451933079664, "loss": 0.9315, "step": 120 }, { "epoch": 1.32, "grad_norm": 0.53125, "learning_rate": 0.00016904865902523096, "loss": 0.9869, "step": 125 }, { "epoch": 1.37, "grad_norm": 0.796875, "learning_rate": 0.00016561369673087588, "loss": 0.9399, "step": 130 }, { "epoch": 1.42, "grad_norm": 0.75390625, "learning_rate": 0.00016203704982146073, "loss": 0.9061, "step": 135 }, { "epoch": 1.48, "grad_norm": 2.296875, "learning_rate": 0.00015832644162101417, "loss": 0.9537, "step": 140 }, { "epoch": 1.53, "grad_norm": 0.6875, "learning_rate": 0.00015448988472630654, "loss": 0.9107, "step": 145 }, { "epoch": 1.58, "grad_norm": 0.61328125, "learning_rate": 0.00015053566370464415, "loss": 0.9483, "step": 150 }, { "epoch": 1.63, "grad_norm": 0.7265625, "learning_rate": 0.00014647231720437686, "loss": 0.9075, "step": 155 }, { "epoch": 1.69, "grad_norm": 0.53515625, "learning_rate": 0.00014230861951674913, "loss": 0.9148, "step": 160 }, { "epoch": 1.74, "grad_norm": 0.51171875, "learning_rate": 0.0001380535616289099, "loss": 0.9142, "step": 165 }, { "epoch": 1.79, "grad_norm": 0.5703125, "learning_rate": 0.00013371633180899416, "loss": 0.8966, "step": 170 }, { "epoch": 1.84, "grad_norm": 0.53515625, "learning_rate": 0.00012930629576520132, "loss": 0.8795, "step": 175 }, { "epoch": 1.9, "grad_norm": 0.609375, "learning_rate": 0.00012483297642171333, "loss": 0.8548, "step": 180 }, { "epoch": 1.95, "grad_norm": 0.53515625, "learning_rate": 0.00012030603335512468, "loss": 0.9028, "step": 185 }, { "epoch": 1.99, "eval_loss": 0.8055516481399536, "eval_runtime": 16.4906, "eval_samples_per_second": 17.283, "eval_steps_per_second": 2.183, "step": 189 }, { "epoch": 2.0, "grad_norm": 0.58203125, "learning_rate": 0.00011573524193578863, "loss": 0.8422, "step": 190 }, { "epoch": 2.06, "grad_norm": 0.57421875, "learning_rate": 0.00011113047221912096, "loss": 0.7173, "step": 195 }, { "epoch": 2.11, "grad_norm": 0.52734375, "learning_rate": 0.0001065016676324433, "loss": 0.6956, "step": 200 }, { "epoch": 2.16, "grad_norm": 0.53125, "learning_rate": 0.0001018588235033888, "loss": 0.6894, "step": 205 }, { "epoch": 2.21, "grad_norm": 0.54296875, "learning_rate": 9.721196547623584e-05, "loss": 0.6804, "step": 210 }, { "epoch": 2.27, "grad_norm": 0.4765625, "learning_rate": 9.257112786277631e-05, "loss": 0.6327, "step": 215 }, { "epoch": 2.32, "grad_norm": 0.51953125, "learning_rate": 8.79463319744677e-05, "loss": 0.637, "step": 220 }, { "epoch": 2.37, "grad_norm": 0.53125, "learning_rate": 8.334756448265781e-05, "loss": 0.6765, "step": 225 }, { "epoch": 2.42, "grad_norm": 0.51953125, "learning_rate": 7.878475585361045e-05, "loss": 0.6518, "step": 230 }, { "epoch": 2.48, "grad_norm": 0.5390625, "learning_rate": 7.42677589048989e-05, "loss": 0.6535, "step": 235 }, { "epoch": 2.53, "grad_norm": 0.5078125, "learning_rate": 6.98063275294722e-05, "loss": 0.6311, "step": 240 }, { "epoch": 2.58, "grad_norm": 0.52734375, "learning_rate": 6.54100956333369e-05, "loss": 0.6438, "step": 245 }, { "epoch": 2.64, "grad_norm": 0.55859375, "learning_rate": 6.108855633233546e-05, "loss": 0.6366, "step": 250 }, { "epoch": 2.69, "grad_norm": 0.51171875, "learning_rate": 5.6851041452943646e-05, "loss": 0.627, "step": 255 }, { "epoch": 2.74, "grad_norm": 0.498046875, "learning_rate": 5.270670138135234e-05, "loss": 0.6359, "step": 260 }, { "epoch": 2.79, "grad_norm": 0.49609375, "learning_rate": 4.866448530434692e-05, "loss": 0.6727, "step": 265 }, { "epoch": 2.85, "grad_norm": 0.49609375, "learning_rate": 4.4733121884651664e-05, "loss": 0.6406, "step": 270 }, { "epoch": 2.9, "grad_norm": 0.54296875, "learning_rate": 4.092110041246865e-05, "loss": 0.6535, "step": 275 }, { "epoch": 2.95, "grad_norm": 0.48046875, "learning_rate": 3.7236652473911814e-05, "loss": 0.6553, "step": 280 }, { "epoch": 2.99, "eval_loss": 0.6577403545379639, "eval_runtime": 16.6273, "eval_samples_per_second": 17.141, "eval_steps_per_second": 2.165, "step": 284 }, { "epoch": 3.0, "grad_norm": 0.578125, "learning_rate": 3.36877341759205e-05, "loss": 0.6292, "step": 285 }, { "epoch": 3.06, "grad_norm": 0.7421875, "learning_rate": 3.0282008966036646e-05, "loss": 0.4973, "step": 290 }, { "epoch": 3.11, "grad_norm": 0.54296875, "learning_rate": 2.7026831084143255e-05, "loss": 0.4967, "step": 295 }, { "epoch": 3.16, "grad_norm": 0.5234375, "learning_rate": 2.3929229681898003e-05, "loss": 0.5102, "step": 300 }, { "epoch": 3.21, "grad_norm": 0.50390625, "learning_rate": 2.0995893644155008e-05, "loss": 0.4894, "step": 305 }, { "epoch": 3.27, "grad_norm": 0.515625, "learning_rate": 1.823315714515018e-05, "loss": 0.5026, "step": 310 }, { "epoch": 3.32, "grad_norm": 0.49609375, "learning_rate": 1.5646985970639717e-05, "loss": 0.4853, "step": 315 }, { "epoch": 3.37, "grad_norm": 0.50390625, "learning_rate": 1.324296463552821e-05, "loss": 0.483, "step": 320 }, { "epoch": 3.43, "grad_norm": 0.4921875, "learning_rate": 1.1026284324803494e-05, "loss": 0.5077, "step": 325 }, { "epoch": 3.48, "grad_norm": 0.51953125, "learning_rate": 9.001731683818337e-06, "loss": 0.5099, "step": 330 }, { "epoch": 3.53, "grad_norm": 0.50390625, "learning_rate": 7.17367848212539e-06, "loss": 0.4787, "step": 335 }, { "epoch": 3.58, "grad_norm": 0.5078125, "learning_rate": 5.546072173184791e-06, "loss": 0.4859, "step": 340 }, { "epoch": 3.64, "grad_norm": 0.5, "learning_rate": 4.1224273703294514e-06, "loss": 0.5011, "step": 345 }, { "epoch": 3.69, "grad_norm": 0.5078125, "learning_rate": 2.905818257394799e-06, "loss": 0.5044, "step": 350 }, { "epoch": 3.74, "grad_norm": 0.5078125, "learning_rate": 1.8988719504013374e-06, "loss": 0.4728, "step": 355 }, { "epoch": 3.79, "grad_norm": 0.51171875, "learning_rate": 1.103762824624377e-06, "loss": 0.5035, "step": 360 }, { "epoch": 3.85, "grad_norm": 0.498046875, "learning_rate": 5.222078193011126e-07, "loss": 0.4899, "step": 365 }, { "epoch": 3.9, "grad_norm": 0.50390625, "learning_rate": 1.554627301140199e-07, "loss": 0.4906, "step": 370 }, { "epoch": 3.95, "grad_norm": 0.4921875, "learning_rate": 4.319497456273247e-09, "loss": 0.4936, "step": 375 }, { "epoch": 3.96, "eval_loss": 0.6385390162467957, "eval_runtime": 16.4691, "eval_samples_per_second": 17.305, "eval_steps_per_second": 2.186, "step": 376 }, { "epoch": 3.96, "step": 376, "total_flos": 2.3511765822965023e+18, "train_loss": 1.1451922758938151, "train_runtime": 4455.6354, "train_samples_per_second": 5.447, "train_steps_per_second": 0.084 } ], "logging_steps": 5, "max_steps": 376, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 500, "total_flos": 2.3511765822965023e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }