|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.7242524916943522, |
|
"eval_steps": 25, |
|
"global_step": 50, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.053156146179401995, |
|
"grad_norm": 16.070100784301758, |
|
"learning_rate": 5e-05, |
|
"loss": 4.8852, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.053156146179401995, |
|
"eval_loss": 4.792490005493164, |
|
"eval_runtime": 0.7317, |
|
"eval_samples_per_second": 68.332, |
|
"eval_steps_per_second": 17.766, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.10631229235880399, |
|
"grad_norm": 15.72305965423584, |
|
"learning_rate": 0.0001, |
|
"loss": 4.6739, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.15946843853820597, |
|
"grad_norm": 14.131994247436523, |
|
"learning_rate": 9.990365154573717e-05, |
|
"loss": 4.3832, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.21262458471760798, |
|
"grad_norm": 16.100086212158203, |
|
"learning_rate": 9.961501876182148e-05, |
|
"loss": 3.2328, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.26578073089701, |
|
"grad_norm": 9.526208877563477, |
|
"learning_rate": 9.913533761814537e-05, |
|
"loss": 2.0447, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.31893687707641194, |
|
"grad_norm": 6.778939723968506, |
|
"learning_rate": 9.846666218300807e-05, |
|
"loss": 1.0156, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.37209302325581395, |
|
"grad_norm": 3.811636209487915, |
|
"learning_rate": 9.761185582727977e-05, |
|
"loss": 0.6336, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.42524916943521596, |
|
"grad_norm": 4.02534294128418, |
|
"learning_rate": 9.657457896300791e-05, |
|
"loss": 0.4985, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.47840531561461797, |
|
"grad_norm": 3.899996280670166, |
|
"learning_rate": 9.535927336897098e-05, |
|
"loss": 0.4257, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.53156146179402, |
|
"grad_norm": 2.4773781299591064, |
|
"learning_rate": 9.397114317029975e-05, |
|
"loss": 0.2646, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.584717607973422, |
|
"grad_norm": 1.6356959342956543, |
|
"learning_rate": 9.241613255361455e-05, |
|
"loss": 0.2353, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.6378737541528239, |
|
"grad_norm": 1.855863332748413, |
|
"learning_rate": 9.070090031310558e-05, |
|
"loss": 0.1652, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.6910299003322259, |
|
"grad_norm": 1.7059516906738281, |
|
"learning_rate": 8.883279133655399e-05, |
|
"loss": 0.1623, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.7441860465116279, |
|
"grad_norm": 3.5583112239837646, |
|
"learning_rate": 8.681980515339464e-05, |
|
"loss": 0.2193, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.7973421926910299, |
|
"grad_norm": 2.0394420623779297, |
|
"learning_rate": 8.467056167950311e-05, |
|
"loss": 0.1814, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.8504983388704319, |
|
"grad_norm": 2.178466558456421, |
|
"learning_rate": 8.239426430539243e-05, |
|
"loss": 0.1836, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.9036544850498339, |
|
"grad_norm": 2.2939162254333496, |
|
"learning_rate": 8.000066048588211e-05, |
|
"loss": 0.2283, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.9568106312292359, |
|
"grad_norm": 1.9878804683685303, |
|
"learning_rate": 7.75e-05, |
|
"loss": 0.1923, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 1.043189368770764, |
|
"grad_norm": 8.555853843688965, |
|
"learning_rate": 7.490299105985507e-05, |
|
"loss": 0.5503, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 1.0963455149501662, |
|
"grad_norm": 1.7653311491012573, |
|
"learning_rate": 7.222075445642904e-05, |
|
"loss": 0.0921, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 1.149501661129568, |
|
"grad_norm": 0.7832552790641785, |
|
"learning_rate": 6.946477593864228e-05, |
|
"loss": 0.0621, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 1.2026578073089702, |
|
"grad_norm": 0.47030219435691833, |
|
"learning_rate": 6.664685702961344e-05, |
|
"loss": 0.0903, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 1.255813953488372, |
|
"grad_norm": 1.1789742708206177, |
|
"learning_rate": 6.377906449072578e-05, |
|
"loss": 0.1346, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 1.308970099667774, |
|
"grad_norm": 1.045068621635437, |
|
"learning_rate": 6.087367864990233e-05, |
|
"loss": 0.126, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 1.3621262458471761, |
|
"grad_norm": 0.8990850448608398, |
|
"learning_rate": 5.794314081535644e-05, |
|
"loss": 0.1019, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 1.3621262458471761, |
|
"eval_loss": 0.12839345633983612, |
|
"eval_runtime": 0.7402, |
|
"eval_samples_per_second": 67.551, |
|
"eval_steps_per_second": 17.563, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 1.415282392026578, |
|
"grad_norm": 0.7274023294448853, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 0.0638, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 1.4684385382059801, |
|
"grad_norm": 1.0477628707885742, |
|
"learning_rate": 5.205685918464356e-05, |
|
"loss": 0.1032, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 1.521594684385382, |
|
"grad_norm": 1.7702083587646484, |
|
"learning_rate": 4.912632135009769e-05, |
|
"loss": 0.0773, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 1.574750830564784, |
|
"grad_norm": 0.9557569026947021, |
|
"learning_rate": 4.6220935509274235e-05, |
|
"loss": 0.0616, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 1.627906976744186, |
|
"grad_norm": 1.0757166147232056, |
|
"learning_rate": 4.3353142970386564e-05, |
|
"loss": 0.1396, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 1.6810631229235882, |
|
"grad_norm": 0.8739386796951294, |
|
"learning_rate": 4.053522406135775e-05, |
|
"loss": 0.0655, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 1.73421926910299, |
|
"grad_norm": 0.9457898736000061, |
|
"learning_rate": 3.777924554357096e-05, |
|
"loss": 0.0842, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 1.787375415282392, |
|
"grad_norm": 0.5630450248718262, |
|
"learning_rate": 3.509700894014496e-05, |
|
"loss": 0.0476, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 1.8405315614617939, |
|
"grad_norm": 0.7644055485725403, |
|
"learning_rate": 3.250000000000001e-05, |
|
"loss": 0.1071, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 1.893687707641196, |
|
"grad_norm": 1.1439839601516724, |
|
"learning_rate": 2.9999339514117912e-05, |
|
"loss": 0.1219, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 1.946843853820598, |
|
"grad_norm": 0.68359375, |
|
"learning_rate": 2.760573569460757e-05, |
|
"loss": 0.1116, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 2.0332225913621262, |
|
"grad_norm": 1.7485463619232178, |
|
"learning_rate": 2.53294383204969e-05, |
|
"loss": 0.1243, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 2.086378737541528, |
|
"grad_norm": 0.47398442029953003, |
|
"learning_rate": 2.3180194846605367e-05, |
|
"loss": 0.0464, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 2.13953488372093, |
|
"grad_norm": 0.9993898272514343, |
|
"learning_rate": 2.1167208663446025e-05, |
|
"loss": 0.082, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 2.1926910299003324, |
|
"grad_norm": 0.46211889386177063, |
|
"learning_rate": 1.9299099686894423e-05, |
|
"loss": 0.0404, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 2.2458471760797343, |
|
"grad_norm": 0.8578104972839355, |
|
"learning_rate": 1.758386744638546e-05, |
|
"loss": 0.1049, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 2.299003322259136, |
|
"grad_norm": 0.7607599496841431, |
|
"learning_rate": 1.602885682970026e-05, |
|
"loss": 0.086, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 2.352159468438538, |
|
"grad_norm": 0.46289491653442383, |
|
"learning_rate": 1.464072663102903e-05, |
|
"loss": 0.0477, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 2.4053156146179404, |
|
"grad_norm": 0.8016815185546875, |
|
"learning_rate": 1.3425421036992098e-05, |
|
"loss": 0.0811, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 2.4584717607973423, |
|
"grad_norm": 0.43612438440322876, |
|
"learning_rate": 1.2388144172720251e-05, |
|
"loss": 0.0495, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 2.511627906976744, |
|
"grad_norm": 0.7244629263877869, |
|
"learning_rate": 1.1533337816991932e-05, |
|
"loss": 0.0904, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 2.564784053156146, |
|
"grad_norm": 0.526421844959259, |
|
"learning_rate": 1.0864662381854632e-05, |
|
"loss": 0.0517, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 2.617940199335548, |
|
"grad_norm": 0.7599980235099792, |
|
"learning_rate": 1.0384981238178534e-05, |
|
"loss": 0.0553, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 2.6710963455149503, |
|
"grad_norm": 0.7634601593017578, |
|
"learning_rate": 1.0096348454262845e-05, |
|
"loss": 0.0647, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 2.7242524916943522, |
|
"grad_norm": 0.7769803404808044, |
|
"learning_rate": 1e-05, |
|
"loss": 0.1012, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 2.7242524916943522, |
|
"eval_loss": 0.14976823329925537, |
|
"eval_runtime": 0.725, |
|
"eval_samples_per_second": 68.97, |
|
"eval_steps_per_second": 17.932, |
|
"step": 50 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 50, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.54445405339648e+16, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|