|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.11199211575505084, |
|
"eval_steps": 1000, |
|
"global_step": 5000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.002239842315101017, |
|
"grad_norm": 71.10852813720703, |
|
"learning_rate": 0.0001, |
|
"loss": 3.5057, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.004479684630202034, |
|
"grad_norm": 57.756534576416016, |
|
"learning_rate": 9.9998756572327e-05, |
|
"loss": 0.8254, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.006719526945303051, |
|
"grad_norm": 44.61080551147461, |
|
"learning_rate": 9.999502635115246e-05, |
|
"loss": 0.6935, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.008959369260404068, |
|
"grad_norm": 50.82619094848633, |
|
"learning_rate": 9.998880952200681e-05, |
|
"loss": 0.5972, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.011199211575505085, |
|
"grad_norm": 44.26677703857422, |
|
"learning_rate": 9.998010639409713e-05, |
|
"loss": 0.5408, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.013439053890606102, |
|
"grad_norm": 42.8045654296875, |
|
"learning_rate": 9.996891740029186e-05, |
|
"loss": 0.5509, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.01567889620570712, |
|
"grad_norm": 30.989139556884766, |
|
"learning_rate": 9.995524309709913e-05, |
|
"loss": 0.4823, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.017918738520808136, |
|
"grad_norm": 34.01952362060547, |
|
"learning_rate": 9.993908416463927e-05, |
|
"loss": 0.5111, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.020158580835909153, |
|
"grad_norm": 34.079307556152344, |
|
"learning_rate": 9.992044140661079e-05, |
|
"loss": 0.4635, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.02239842315101017, |
|
"grad_norm": 26.16071128845215, |
|
"learning_rate": 9.989931575025056e-05, |
|
"loss": 0.4883, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.02239842315101017, |
|
"eval_avg_non_pair_similarity": 0.0020343252948339737, |
|
"eval_avg_pair_similarity": 0.008852629057131708, |
|
"eval_loss": 0.5431402921676636, |
|
"eval_runtime": 19.6178, |
|
"eval_samples_per_second": 25.487, |
|
"eval_similarity_ratio": 4.35162905343228, |
|
"eval_steps_per_second": 0.816, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.024638265466111187, |
|
"grad_norm": 35.38695526123047, |
|
"learning_rate": 9.987570824628759e-05, |
|
"loss": 0.4655, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.026878107781212204, |
|
"grad_norm": 32.071346282958984, |
|
"learning_rate": 9.984962006889084e-05, |
|
"loss": 0.4342, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.029117950096313218, |
|
"grad_norm": 39.6610221862793, |
|
"learning_rate": 9.982105251561082e-05, |
|
"loss": 0.458, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.03135779241141424, |
|
"grad_norm": 31.493322372436523, |
|
"learning_rate": 9.979000700731491e-05, |
|
"loss": 0.4525, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.03359763472651525, |
|
"grad_norm": 34.453399658203125, |
|
"learning_rate": 9.975648508811693e-05, |
|
"loss": 0.41, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.03583747704161627, |
|
"grad_norm": 33.990074157714844, |
|
"learning_rate": 9.972048842530012e-05, |
|
"loss": 0.4097, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.038077319356717286, |
|
"grad_norm": 30.44228172302246, |
|
"learning_rate": 9.968201880923439e-05, |
|
"loss": 0.4257, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.040317161671818307, |
|
"grad_norm": 31.427162170410156, |
|
"learning_rate": 9.964107815328711e-05, |
|
"loss": 0.3821, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.04255700398691932, |
|
"grad_norm": 25.877887725830078, |
|
"learning_rate": 9.959766849372808e-05, |
|
"loss": 0.3788, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.04479684630202034, |
|
"grad_norm": 25.36798095703125, |
|
"learning_rate": 9.955179198962817e-05, |
|
"loss": 0.3854, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.04479684630202034, |
|
"eval_avg_non_pair_similarity": 0.0016289287904792565, |
|
"eval_avg_pair_similarity": 0.0032495629731565715, |
|
"eval_loss": 0.4394480586051941, |
|
"eval_runtime": 19.5064, |
|
"eval_samples_per_second": 25.633, |
|
"eval_similarity_ratio": 1.9949079371360972, |
|
"eval_steps_per_second": 0.82, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.047036688617121354, |
|
"grad_norm": 24.537384033203125, |
|
"learning_rate": 9.950345092275198e-05, |
|
"loss": 0.3886, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.049276530932222375, |
|
"grad_norm": 26.4981632232666, |
|
"learning_rate": 9.945264769744431e-05, |
|
"loss": 0.3926, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.05151637324732339, |
|
"grad_norm": 40.585941314697266, |
|
"learning_rate": 9.939938484051063e-05, |
|
"loss": 0.3796, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.05375621556242441, |
|
"grad_norm": 26.374149322509766, |
|
"learning_rate": 9.934366500109132e-05, |
|
"loss": 0.3723, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.05599605787752542, |
|
"grad_norm": 31.54728126525879, |
|
"learning_rate": 9.928549095053001e-05, |
|
"loss": 0.3639, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.058235900192626436, |
|
"grad_norm": 26.00455093383789, |
|
"learning_rate": 9.922486558223567e-05, |
|
"loss": 0.3387, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.060475742507727456, |
|
"grad_norm": 23.684057235717773, |
|
"learning_rate": 9.916179191153873e-05, |
|
"loss": 0.3494, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.06271558482282848, |
|
"grad_norm": 24.034656524658203, |
|
"learning_rate": 9.909627307554108e-05, |
|
"loss": 0.362, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.06495542713792948, |
|
"grad_norm": 31.07891082763672, |
|
"learning_rate": 9.902831233296009e-05, |
|
"loss": 0.3481, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.0671952694530305, |
|
"grad_norm": 24.183626174926758, |
|
"learning_rate": 9.895791306396644e-05, |
|
"loss": 0.3535, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.0671952694530305, |
|
"eval_avg_non_pair_similarity": -0.0007878901720614375, |
|
"eval_avg_pair_similarity": -0.008789425778668375, |
|
"eval_loss": 0.4187028110027313, |
|
"eval_runtime": 19.4839, |
|
"eval_samples_per_second": 25.662, |
|
"eval_similarity_ratio": 11.155648452463497, |
|
"eval_steps_per_second": 0.821, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.06943511176813152, |
|
"grad_norm": 20.178213119506836, |
|
"learning_rate": 9.888507877001616e-05, |
|
"loss": 0.3375, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.07167495408323254, |
|
"grad_norm": 31.757244110107422, |
|
"learning_rate": 9.880981307367627e-05, |
|
"loss": 0.3489, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.07391479639833355, |
|
"grad_norm": 27.889982223510742, |
|
"learning_rate": 9.873211971844477e-05, |
|
"loss": 0.3463, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.07615463871343457, |
|
"grad_norm": 23.282093048095703, |
|
"learning_rate": 9.865200256856437e-05, |
|
"loss": 0.3235, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.07839448102853559, |
|
"grad_norm": 24.573863983154297, |
|
"learning_rate": 9.856946560883034e-05, |
|
"loss": 0.3474, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.08063432334363661, |
|
"grad_norm": 20.7421932220459, |
|
"learning_rate": 9.848451294439224e-05, |
|
"loss": 0.3533, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.08287416565873762, |
|
"grad_norm": 22.051513671875, |
|
"learning_rate": 9.839714880054987e-05, |
|
"loss": 0.3277, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.08511400797383864, |
|
"grad_norm": 22.470027923583984, |
|
"learning_rate": 9.830737752254294e-05, |
|
"loss": 0.3261, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.08735385028893966, |
|
"grad_norm": 20.80890464782715, |
|
"learning_rate": 9.821520357533513e-05, |
|
"loss": 0.3474, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.08959369260404068, |
|
"grad_norm": 20.187280654907227, |
|
"learning_rate": 9.812063154339191e-05, |
|
"loss": 0.2984, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.08959369260404068, |
|
"eval_avg_non_pair_similarity": -0.0008396646829899625, |
|
"eval_avg_pair_similarity": -0.011315496074035763, |
|
"eval_loss": 0.32538020610809326, |
|
"eval_runtime": 19.4916, |
|
"eval_samples_per_second": 25.652, |
|
"eval_similarity_ratio": 13.476208185560939, |
|
"eval_steps_per_second": 0.821, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.09183353491914169, |
|
"grad_norm": 26.496036529541016, |
|
"learning_rate": 9.802366613045254e-05, |
|
"loss": 0.3326, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.09407337723424271, |
|
"grad_norm": 23.102359771728516, |
|
"learning_rate": 9.792431215929613e-05, |
|
"loss": 0.3341, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.09631321954934373, |
|
"grad_norm": 21.27369499206543, |
|
"learning_rate": 9.782257457150177e-05, |
|
"loss": 0.3254, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.09855306186444475, |
|
"grad_norm": 19.25406265258789, |
|
"learning_rate": 9.771845842720274e-05, |
|
"loss": 0.3334, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.10079290417954576, |
|
"grad_norm": 15.666335105895996, |
|
"learning_rate": 9.761196890483482e-05, |
|
"loss": 0.3064, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.10303274649464678, |
|
"grad_norm": 24.592592239379883, |
|
"learning_rate": 9.75031113008788e-05, |
|
"loss": 0.2902, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.1052725888097478, |
|
"grad_norm": 20.0572566986084, |
|
"learning_rate": 9.739189102959696e-05, |
|
"loss": 0.3121, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.10751243112484882, |
|
"grad_norm": 19.887725830078125, |
|
"learning_rate": 9.727831362276381e-05, |
|
"loss": 0.3014, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.10975227343994982, |
|
"grad_norm": 15.561097145080566, |
|
"learning_rate": 9.716238472939101e-05, |
|
"loss": 0.316, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.11199211575505084, |
|
"grad_norm": 17.96786880493164, |
|
"learning_rate": 9.704411011544629e-05, |
|
"loss": 0.3215, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.11199211575505084, |
|
"eval_avg_non_pair_similarity": 0.00024146916974524413, |
|
"eval_avg_pair_similarity": -0.009633154251612723, |
|
"eval_loss": 0.2951599061489105, |
|
"eval_runtime": 19.5025, |
|
"eval_samples_per_second": 25.638, |
|
"eval_similarity_ratio": -39.893930400207765, |
|
"eval_steps_per_second": 0.82, |
|
"step": 5000 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 44646, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|