|
{ |
|
"best_metric": 0.42311346530914307, |
|
"best_model_checkpoint": "speecht5_finetuned_bphc/checkpoint-500", |
|
"epoch": 2.6595744680851063, |
|
"eval_steps": 100, |
|
"global_step": 500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.13297872340425532, |
|
"grad_norm": 3.937445640563965, |
|
"learning_rate": 2.4e-05, |
|
"loss": 1.0732, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.26595744680851063, |
|
"grad_norm": 3.3802008628845215, |
|
"learning_rate": 4.9e-05, |
|
"loss": 0.7567, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.39893617021276595, |
|
"grad_norm": 7.172617435455322, |
|
"learning_rate": 7.3e-05, |
|
"loss": 0.6624, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.5319148936170213, |
|
"grad_norm": 5.82205867767334, |
|
"learning_rate": 9.8e-05, |
|
"loss": 0.6294, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.5319148936170213, |
|
"eval_loss": 0.5470437407493591, |
|
"eval_runtime": 19.8146, |
|
"eval_samples_per_second": 33.763, |
|
"eval_steps_per_second": 16.907, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.6648936170212766, |
|
"grad_norm": 3.176703929901123, |
|
"learning_rate": 9.425e-05, |
|
"loss": 0.6045, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.7978723404255319, |
|
"grad_norm": 2.801500082015991, |
|
"learning_rate": 8.800000000000001e-05, |
|
"loss": 0.5722, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.9308510638297872, |
|
"grad_norm": 7.730113506317139, |
|
"learning_rate": 8.175000000000001e-05, |
|
"loss": 0.5563, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.0638297872340425, |
|
"grad_norm": 11.911551475524902, |
|
"learning_rate": 7.55e-05, |
|
"loss": 0.5486, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.0638297872340425, |
|
"eval_loss": 0.4846374988555908, |
|
"eval_runtime": 18.6315, |
|
"eval_samples_per_second": 35.907, |
|
"eval_steps_per_second": 17.98, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.196808510638298, |
|
"grad_norm": 4.457734107971191, |
|
"learning_rate": 6.925e-05, |
|
"loss": 0.5378, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.3297872340425532, |
|
"grad_norm": 2.7005956172943115, |
|
"learning_rate": 6.3e-05, |
|
"loss": 0.5355, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.4627659574468086, |
|
"grad_norm": 2.9917397499084473, |
|
"learning_rate": 5.6750000000000004e-05, |
|
"loss": 0.5138, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.5957446808510638, |
|
"grad_norm": 3.650026321411133, |
|
"learning_rate": 5.05e-05, |
|
"loss": 0.5046, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.5957446808510638, |
|
"eval_loss": 0.45161741971969604, |
|
"eval_runtime": 18.3387, |
|
"eval_samples_per_second": 36.48, |
|
"eval_steps_per_second": 18.267, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.728723404255319, |
|
"grad_norm": 3.8126158714294434, |
|
"learning_rate": 4.4250000000000005e-05, |
|
"loss": 0.4897, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.8617021276595744, |
|
"grad_norm": 4.279622554779053, |
|
"learning_rate": 3.8e-05, |
|
"loss": 0.4865, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.9946808510638299, |
|
"grad_norm": 4.044638633728027, |
|
"learning_rate": 3.175e-05, |
|
"loss": 0.4823, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 2.127659574468085, |
|
"grad_norm": 4.797245979309082, |
|
"learning_rate": 2.5500000000000003e-05, |
|
"loss": 0.4732, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.127659574468085, |
|
"eval_loss": 0.43503323197364807, |
|
"eval_runtime": 18.8477, |
|
"eval_samples_per_second": 35.495, |
|
"eval_steps_per_second": 17.774, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.2606382978723403, |
|
"grad_norm": 2.5267035961151123, |
|
"learning_rate": 1.925e-05, |
|
"loss": 0.4774, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 2.393617021276596, |
|
"grad_norm": 3.2455756664276123, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"loss": 0.4646, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.526595744680851, |
|
"grad_norm": 3.491698980331421, |
|
"learning_rate": 6.750000000000001e-06, |
|
"loss": 0.4601, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.6595744680851063, |
|
"grad_norm": 3.7166852951049805, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 0.4601, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.6595744680851063, |
|
"eval_loss": 0.42311346530914307, |
|
"eval_runtime": 18.8917, |
|
"eval_samples_per_second": 35.412, |
|
"eval_steps_per_second": 17.733, |
|
"step": 500 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 500, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1481834827297440.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|