|
{ |
|
"best_metric": 0.9787709497206704, |
|
"best_model_checkpoint": "resnet-50-LongSleeveCleanedData/checkpoint-2733", |
|
"epoch": 19.88083416087388, |
|
"global_step": 2860, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.7241379310344828e-05, |
|
"loss": 1.1031, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 3.4482758620689657e-05, |
|
"loss": 1.102, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.998233839632639e-05, |
|
"loss": 1.0994, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.980572235959025e-05, |
|
"loss": 1.0905, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.962910632285412e-05, |
|
"loss": 1.0828, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.945249028611798e-05, |
|
"loss": 1.0761, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.927587424938185e-05, |
|
"loss": 1.0646, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.909925821264571e-05, |
|
"loss": 1.0606, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.892264217590958e-05, |
|
"loss": 1.0502, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.874602613917344e-05, |
|
"loss": 1.0411, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.85694101024373e-05, |
|
"loss": 1.0358, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.8392794065701166e-05, |
|
"loss": 1.023, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.821617802896503e-05, |
|
"loss": 1.0164, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.8039561992228896e-05, |
|
"loss": 0.9906, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"eval_accuracy": 0.6134078212290502, |
|
"eval_loss": 1.0393954515457153, |
|
"eval_runtime": 725.2732, |
|
"eval_samples_per_second": 1.234, |
|
"eval_steps_per_second": 0.154, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.786294595549276e-05, |
|
"loss": 0.9853, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.768632991875663e-05, |
|
"loss": 0.9606, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 4.750971388202049e-05, |
|
"loss": 0.9562, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 4.7333097845284354e-05, |
|
"loss": 0.9219, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 4.715648180854822e-05, |
|
"loss": 0.9033, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.697986577181208e-05, |
|
"loss": 0.8942, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 4.680324973507595e-05, |
|
"loss": 0.8555, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.662663369833981e-05, |
|
"loss": 0.8475, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.645001766160367e-05, |
|
"loss": 0.8118, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.627340162486754e-05, |
|
"loss": 0.82, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 4.6096785588131406e-05, |
|
"loss": 0.8028, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.592016955139527e-05, |
|
"loss": 0.7645, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.5743553514659135e-05, |
|
"loss": 0.7472, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.5566937477923e-05, |
|
"loss": 0.7315, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.7631284916201118, |
|
"eval_loss": 0.6789541244506836, |
|
"eval_runtime": 20.8953, |
|
"eval_samples_per_second": 42.833, |
|
"eval_steps_per_second": 5.36, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 4.5390321441186864e-05, |
|
"loss": 0.757, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.521370540445072e-05, |
|
"loss": 0.7235, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.503708936771459e-05, |
|
"loss": 0.6899, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.486047333097845e-05, |
|
"loss": 0.6831, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.468385729424232e-05, |
|
"loss": 0.6265, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.450724125750618e-05, |
|
"loss": 0.6394, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 4.433062522077005e-05, |
|
"loss": 0.64, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.415400918403391e-05, |
|
"loss": 0.6319, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.3977393147297774e-05, |
|
"loss": 0.5885, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 4.3800777110561645e-05, |
|
"loss": 0.6172, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.36241610738255e-05, |
|
"loss": 0.6075, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 4.3447545037089374e-05, |
|
"loss": 0.5672, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 4.327092900035323e-05, |
|
"loss": 0.5717, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 4.30943129636171e-05, |
|
"loss": 0.5241, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 4.291769692688096e-05, |
|
"loss": 0.559, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.8547486033519553, |
|
"eval_loss": 0.47351497411727905, |
|
"eval_runtime": 19.826, |
|
"eval_samples_per_second": 45.143, |
|
"eval_steps_per_second": 5.649, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 4.2741080890144825e-05, |
|
"loss": 0.5018, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 4.256446485340869e-05, |
|
"loss": 0.5239, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 4.2387848816672555e-05, |
|
"loss": 0.5113, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 4.221123277993642e-05, |
|
"loss": 0.5464, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 4.2034616743200284e-05, |
|
"loss": 0.5247, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 4.185800070646415e-05, |
|
"loss": 0.5309, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 4.168138466972801e-05, |
|
"loss": 0.4472, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 4.150476863299188e-05, |
|
"loss": 0.4852, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 4.132815259625574e-05, |
|
"loss": 0.4296, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 4.1151536559519606e-05, |
|
"loss": 0.4878, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 4.097492052278347e-05, |
|
"loss": 0.4741, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 4.0798304486047335e-05, |
|
"loss": 0.4571, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 4.06216884493112e-05, |
|
"loss": 0.4514, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 4.0445072412575065e-05, |
|
"loss": 0.4905, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.8983240223463688, |
|
"eval_loss": 0.3148014545440674, |
|
"eval_runtime": 21.1724, |
|
"eval_samples_per_second": 42.272, |
|
"eval_steps_per_second": 5.29, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 4.026845637583892e-05, |
|
"loss": 0.4468, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 4.0091840339102794e-05, |
|
"loss": 0.4162, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 3.991522430236666e-05, |
|
"loss": 0.4565, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 3.973860826563052e-05, |
|
"loss": 0.4272, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 3.956199222889439e-05, |
|
"loss": 0.3927, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 3.938537619215825e-05, |
|
"loss": 0.4419, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 3.9208760155422116e-05, |
|
"loss": 0.4796, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 3.9032144118685974e-05, |
|
"loss": 0.3772, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 3.8855528081949846e-05, |
|
"loss": 0.4285, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 3.86789120452137e-05, |
|
"loss": 0.4611, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 3.8502296008477575e-05, |
|
"loss": 0.4045, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 3.832567997174143e-05, |
|
"loss": 0.4066, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 3.8149063935005304e-05, |
|
"loss": 0.4457, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 3.797244789826916e-05, |
|
"loss": 0.3465, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.9363128491620112, |
|
"eval_loss": 0.22248676419258118, |
|
"eval_runtime": 20.5776, |
|
"eval_samples_per_second": 43.494, |
|
"eval_steps_per_second": 5.443, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 3.7795831861533026e-05, |
|
"loss": 0.3614, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 3.761921582479689e-05, |
|
"loss": 0.3649, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 3.7442599788060755e-05, |
|
"loss": 0.3456, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 3.7265983751324627e-05, |
|
"loss": 0.3546, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 3.7089367714588484e-05, |
|
"loss": 0.3802, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 3.6912751677852356e-05, |
|
"loss": 0.3694, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 3.6736135641116213e-05, |
|
"loss": 0.3551, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 3.655951960438008e-05, |
|
"loss": 0.3756, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 3.638290356764394e-05, |
|
"loss": 0.4003, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 3.620628753090781e-05, |
|
"loss": 0.4294, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 3.602967149417167e-05, |
|
"loss": 0.3199, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 3.5853055457435536e-05, |
|
"loss": 0.3732, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 3.56764394206994e-05, |
|
"loss": 0.3084, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 3.5499823383963265e-05, |
|
"loss": 0.3275, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 3.532320734722713e-05, |
|
"loss": 0.3372, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.9486033519553073, |
|
"eval_loss": 0.18386144936084747, |
|
"eval_runtime": 19.9501, |
|
"eval_samples_per_second": 44.862, |
|
"eval_steps_per_second": 5.614, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"learning_rate": 3.5146591310490994e-05, |
|
"loss": 0.3084, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 3.496997527375486e-05, |
|
"loss": 0.3171, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 3.4793359237018724e-05, |
|
"loss": 0.3379, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 3.461674320028259e-05, |
|
"loss": 0.3426, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 3.444012716354645e-05, |
|
"loss": 0.3396, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 3.426351112681032e-05, |
|
"loss": 0.3184, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 3.4086895090074175e-05, |
|
"loss": 0.3158, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 3.3910279053338046e-05, |
|
"loss": 0.3197, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 3.3733663016601904e-05, |
|
"loss": 0.3312, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 3.3557046979865775e-05, |
|
"loss": 0.3256, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 3.338043094312964e-05, |
|
"loss": 0.3237, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 3.3203814906393505e-05, |
|
"loss": 0.3162, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 3.302719886965737e-05, |
|
"loss": 0.2939, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 6.95, |
|
"learning_rate": 3.285058283292123e-05, |
|
"loss": 0.3349, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.958659217877095, |
|
"eval_loss": 0.16170552372932434, |
|
"eval_runtime": 21.263, |
|
"eval_samples_per_second": 42.092, |
|
"eval_steps_per_second": 5.267, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 3.26739667961851e-05, |
|
"loss": 0.3104, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 3.2497350759448956e-05, |
|
"loss": 0.3108, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 3.232073472271283e-05, |
|
"loss": 0.2673, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 3.2144118685976685e-05, |
|
"loss": 0.3208, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 3.1967502649240556e-05, |
|
"loss": 0.2951, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 3.1790886612504414e-05, |
|
"loss": 0.2816, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 7.44, |
|
"learning_rate": 3.161427057576828e-05, |
|
"loss": 0.2879, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 7.51, |
|
"learning_rate": 3.143765453903214e-05, |
|
"loss": 0.2891, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 3.126103850229601e-05, |
|
"loss": 0.2988, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 3.108442246555988e-05, |
|
"loss": 0.3067, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 3.090780642882374e-05, |
|
"loss": 0.3065, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 3.073119039208761e-05, |
|
"loss": 0.3315, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 3.0554574355351466e-05, |
|
"loss": 0.2949, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 3.037795831861533e-05, |
|
"loss": 0.3493, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 3.02013422818792e-05, |
|
"loss": 0.3159, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"eval_accuracy": 0.9620111731843576, |
|
"eval_loss": 0.13231398165225983, |
|
"eval_runtime": 19.7019, |
|
"eval_samples_per_second": 45.427, |
|
"eval_steps_per_second": 5.685, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 3.002472624514306e-05, |
|
"loss": 0.3193, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 2.9848110208406928e-05, |
|
"loss": 0.3245, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 2.967149417167079e-05, |
|
"loss": 0.2705, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 2.9494878134934657e-05, |
|
"loss": 0.2941, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 2.9318262098198518e-05, |
|
"loss": 0.2936, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 2.914164606146238e-05, |
|
"loss": 0.3415, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 2.8965030024726247e-05, |
|
"loss": 0.2976, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 8.55, |
|
"learning_rate": 2.8788413987990108e-05, |
|
"loss": 0.2979, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 2.8611797951253976e-05, |
|
"loss": 0.3391, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 2.8435181914517837e-05, |
|
"loss": 0.3329, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 8.76, |
|
"learning_rate": 2.8258565877781705e-05, |
|
"loss": 0.2764, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 2.808194984104557e-05, |
|
"loss": 0.2915, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"learning_rate": 2.790533380430943e-05, |
|
"loss": 0.2607, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 2.77287177675733e-05, |
|
"loss": 0.2805, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_accuracy": 0.958659217877095, |
|
"eval_loss": 0.16597720980644226, |
|
"eval_runtime": 21.3333, |
|
"eval_samples_per_second": 41.953, |
|
"eval_steps_per_second": 5.25, |
|
"step": 1294 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 2.755210173083716e-05, |
|
"loss": 0.2688, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 2.7375485694101028e-05, |
|
"loss": 0.2965, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 2.719886965736489e-05, |
|
"loss": 0.2919, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 2.7022253620628757e-05, |
|
"loss": 0.2912, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 2.6845637583892618e-05, |
|
"loss": 0.26, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 2.666902154715648e-05, |
|
"loss": 0.2639, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 9.45, |
|
"learning_rate": 2.6492405510420347e-05, |
|
"loss": 0.3168, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 2.6315789473684212e-05, |
|
"loss": 0.2797, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"learning_rate": 2.6139173436948077e-05, |
|
"loss": 0.2514, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 2.596255740021194e-05, |
|
"loss": 0.2826, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 2.578594136347581e-05, |
|
"loss": 0.3124, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 2.560932532673967e-05, |
|
"loss": 0.27, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 2.543270929000353e-05, |
|
"loss": 0.271, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 2.52560932532674e-05, |
|
"loss": 0.2657, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_accuracy": 0.9530726256983241, |
|
"eval_loss": 0.14559155702590942, |
|
"eval_runtime": 19.8491, |
|
"eval_samples_per_second": 45.09, |
|
"eval_steps_per_second": 5.643, |
|
"step": 1438 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 2.507947721653126e-05, |
|
"loss": 0.2865, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 10.08, |
|
"learning_rate": 2.4902861179795125e-05, |
|
"loss": 0.2725, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 10.15, |
|
"learning_rate": 2.472624514305899e-05, |
|
"loss": 0.2701, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 10.22, |
|
"learning_rate": 2.4549629106322854e-05, |
|
"loss": 0.2808, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 2.437301306958672e-05, |
|
"loss": 0.2918, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 10.36, |
|
"learning_rate": 2.4196397032850583e-05, |
|
"loss": 0.291, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 10.43, |
|
"learning_rate": 2.4019780996114448e-05, |
|
"loss": 0.2738, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 2.3843164959378316e-05, |
|
"loss": 0.254, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"learning_rate": 2.3666548922642177e-05, |
|
"loss": 0.245, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 10.64, |
|
"learning_rate": 2.348993288590604e-05, |
|
"loss": 0.265, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"learning_rate": 2.3313316849169906e-05, |
|
"loss": 0.267, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"learning_rate": 2.313670081243377e-05, |
|
"loss": 0.2262, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 10.84, |
|
"learning_rate": 2.2960084775697635e-05, |
|
"loss": 0.2396, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 10.91, |
|
"learning_rate": 2.27834687389615e-05, |
|
"loss": 0.2945, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 10.98, |
|
"learning_rate": 2.260685270222536e-05, |
|
"loss": 0.2929, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_accuracy": 0.9698324022346368, |
|
"eval_loss": 0.10857631266117096, |
|
"eval_runtime": 20.1943, |
|
"eval_samples_per_second": 44.319, |
|
"eval_steps_per_second": 5.546, |
|
"step": 1582 |
|
}, |
|
{ |
|
"epoch": 11.05, |
|
"learning_rate": 2.2430236665489225e-05, |
|
"loss": 0.2728, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 11.12, |
|
"learning_rate": 2.225362062875309e-05, |
|
"loss": 0.2714, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 11.19, |
|
"learning_rate": 2.2077004592016954e-05, |
|
"loss": 0.2536, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 11.26, |
|
"learning_rate": 2.1900388555280822e-05, |
|
"loss": 0.2504, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 11.33, |
|
"learning_rate": 2.1723772518544687e-05, |
|
"loss": 0.2501, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 11.4, |
|
"learning_rate": 2.154715648180855e-05, |
|
"loss": 0.2787, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 11.47, |
|
"learning_rate": 2.1370540445072413e-05, |
|
"loss": 0.2295, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 11.54, |
|
"learning_rate": 2.1193924408336277e-05, |
|
"loss": 0.2616, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 11.61, |
|
"learning_rate": 2.1017308371600142e-05, |
|
"loss": 0.2796, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 11.68, |
|
"learning_rate": 2.0840692334864006e-05, |
|
"loss": 0.2889, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 11.75, |
|
"learning_rate": 2.066407629812787e-05, |
|
"loss": 0.3062, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 11.82, |
|
"learning_rate": 2.0487460261391735e-05, |
|
"loss": 0.2692, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 11.89, |
|
"learning_rate": 2.03108442246556e-05, |
|
"loss": 0.2466, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 11.96, |
|
"learning_rate": 2.013422818791946e-05, |
|
"loss": 0.2763, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.976536312849162, |
|
"eval_loss": 0.088605597615242, |
|
"eval_runtime": 21.0879, |
|
"eval_samples_per_second": 42.441, |
|
"eval_steps_per_second": 5.311, |
|
"step": 1726 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 1.995761215118333e-05, |
|
"loss": 0.2598, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 12.1, |
|
"learning_rate": 1.9780996114447194e-05, |
|
"loss": 0.2472, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 12.16, |
|
"learning_rate": 1.9604380077711058e-05, |
|
"loss": 0.3078, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 12.23, |
|
"learning_rate": 1.9427764040974923e-05, |
|
"loss": 0.2277, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 12.3, |
|
"learning_rate": 1.9251148004238787e-05, |
|
"loss": 0.2445, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 12.37, |
|
"learning_rate": 1.9074531967502652e-05, |
|
"loss": 0.2041, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 12.44, |
|
"learning_rate": 1.8897915930766513e-05, |
|
"loss": 0.2447, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 12.51, |
|
"learning_rate": 1.8721299894030378e-05, |
|
"loss": 0.2281, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 12.58, |
|
"learning_rate": 1.8544683857294242e-05, |
|
"loss": 0.234, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 12.65, |
|
"learning_rate": 1.8368067820558107e-05, |
|
"loss": 0.2434, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 12.72, |
|
"learning_rate": 1.819145178382197e-05, |
|
"loss": 0.2309, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 12.79, |
|
"learning_rate": 1.8014835747085836e-05, |
|
"loss": 0.2143, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 12.86, |
|
"learning_rate": 1.78382197103497e-05, |
|
"loss": 0.2691, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 12.93, |
|
"learning_rate": 1.7661603673613565e-05, |
|
"loss": 0.2371, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 1.748498763687743e-05, |
|
"loss": 0.2475, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_accuracy": 0.9731843575418995, |
|
"eval_loss": 0.1041172444820404, |
|
"eval_runtime": 19.6519, |
|
"eval_samples_per_second": 45.543, |
|
"eval_steps_per_second": 5.699, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 13.07, |
|
"learning_rate": 1.7308371600141294e-05, |
|
"loss": 0.2237, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 13.14, |
|
"learning_rate": 1.713175556340516e-05, |
|
"loss": 0.2541, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 13.21, |
|
"learning_rate": 1.6955139526669023e-05, |
|
"loss": 0.223, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 13.28, |
|
"learning_rate": 1.6778523489932888e-05, |
|
"loss": 0.2379, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 13.35, |
|
"learning_rate": 1.6601907453196752e-05, |
|
"loss": 0.2184, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 13.42, |
|
"learning_rate": 1.6425291416460613e-05, |
|
"loss": 0.1985, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 13.49, |
|
"learning_rate": 1.6248675379724478e-05, |
|
"loss": 0.2464, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 13.56, |
|
"learning_rate": 1.6072059342988343e-05, |
|
"loss": 0.2067, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 13.62, |
|
"learning_rate": 1.5895443306252207e-05, |
|
"loss": 0.2622, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 13.69, |
|
"learning_rate": 1.571882726951607e-05, |
|
"loss": 0.2322, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 13.76, |
|
"learning_rate": 1.554221123277994e-05, |
|
"loss": 0.2618, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 13.83, |
|
"learning_rate": 1.5365595196043804e-05, |
|
"loss": 0.2699, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 13.9, |
|
"learning_rate": 1.5188979159307665e-05, |
|
"loss": 0.2608, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 13.97, |
|
"learning_rate": 1.501236312257153e-05, |
|
"loss": 0.2148, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_accuracy": 0.9776536312849162, |
|
"eval_loss": 0.09550000727176666, |
|
"eval_runtime": 20.0949, |
|
"eval_samples_per_second": 44.539, |
|
"eval_steps_per_second": 5.574, |
|
"step": 2014 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 1.4835747085835394e-05, |
|
"loss": 0.2454, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 14.11, |
|
"learning_rate": 1.4659131049099259e-05, |
|
"loss": 0.2304, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 14.18, |
|
"learning_rate": 1.4482515012363124e-05, |
|
"loss": 0.254, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 14.25, |
|
"learning_rate": 1.4305898975626988e-05, |
|
"loss": 0.2417, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 14.32, |
|
"learning_rate": 1.4129282938890853e-05, |
|
"loss": 0.235, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 14.39, |
|
"learning_rate": 1.3952666902154715e-05, |
|
"loss": 0.2349, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 14.46, |
|
"learning_rate": 1.377605086541858e-05, |
|
"loss": 0.244, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 14.53, |
|
"learning_rate": 1.3599434828682445e-05, |
|
"loss": 0.2315, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 14.6, |
|
"learning_rate": 1.3422818791946309e-05, |
|
"loss": 0.219, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 14.67, |
|
"learning_rate": 1.3246202755210174e-05, |
|
"loss": 0.2722, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 14.74, |
|
"learning_rate": 1.3069586718474038e-05, |
|
"loss": 0.2546, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 14.81, |
|
"learning_rate": 1.2892970681737905e-05, |
|
"loss": 0.2179, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 14.88, |
|
"learning_rate": 1.2716354645001766e-05, |
|
"loss": 0.2407, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 14.95, |
|
"learning_rate": 1.253973860826563e-05, |
|
"loss": 0.209, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 14.99, |
|
"eval_accuracy": 0.9709497206703911, |
|
"eval_loss": 0.10605238378047943, |
|
"eval_runtime": 21.3223, |
|
"eval_samples_per_second": 41.975, |
|
"eval_steps_per_second": 5.253, |
|
"step": 2157 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 1.2363122571529495e-05, |
|
"loss": 0.2344, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 15.08, |
|
"learning_rate": 1.218650653479336e-05, |
|
"loss": 0.2087, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 15.15, |
|
"learning_rate": 1.2009890498057224e-05, |
|
"loss": 0.1903, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 15.22, |
|
"learning_rate": 1.1833274461321088e-05, |
|
"loss": 0.2297, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 15.29, |
|
"learning_rate": 1.1656658424584953e-05, |
|
"loss": 0.2814, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 15.36, |
|
"learning_rate": 1.1480042387848818e-05, |
|
"loss": 0.2455, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 15.43, |
|
"learning_rate": 1.130342635111268e-05, |
|
"loss": 0.2502, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 15.5, |
|
"learning_rate": 1.1126810314376545e-05, |
|
"loss": 0.2388, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 15.57, |
|
"learning_rate": 1.0950194277640411e-05, |
|
"loss": 0.23, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 15.64, |
|
"learning_rate": 1.0773578240904276e-05, |
|
"loss": 0.2555, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 15.71, |
|
"learning_rate": 1.0596962204168139e-05, |
|
"loss": 0.2256, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 15.78, |
|
"learning_rate": 1.0420346167432003e-05, |
|
"loss": 0.2611, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 15.85, |
|
"learning_rate": 1.0243730130695868e-05, |
|
"loss": 0.232, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 15.92, |
|
"learning_rate": 1.006711409395973e-05, |
|
"loss": 0.2425, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 15.99, |
|
"learning_rate": 9.890498057223597e-06, |
|
"loss": 0.2408, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_accuracy": 0.9743016759776536, |
|
"eval_loss": 0.07841791957616806, |
|
"eval_runtime": 19.8182, |
|
"eval_samples_per_second": 45.161, |
|
"eval_steps_per_second": 5.651, |
|
"step": 2301 |
|
}, |
|
{ |
|
"epoch": 16.06, |
|
"learning_rate": 9.713882020487461e-06, |
|
"loss": 0.2557, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 16.13, |
|
"learning_rate": 9.537265983751326e-06, |
|
"loss": 0.2286, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 16.2, |
|
"learning_rate": 9.360649947015189e-06, |
|
"loss": 0.2412, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 16.27, |
|
"learning_rate": 9.184033910279053e-06, |
|
"loss": 0.2506, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 16.34, |
|
"learning_rate": 9.007417873542918e-06, |
|
"loss": 0.261, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 16.41, |
|
"learning_rate": 8.830801836806782e-06, |
|
"loss": 0.2101, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 16.47, |
|
"learning_rate": 8.654185800070647e-06, |
|
"loss": 0.2341, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 16.54, |
|
"learning_rate": 8.477569763334512e-06, |
|
"loss": 0.1933, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 16.61, |
|
"learning_rate": 8.300953726598376e-06, |
|
"loss": 0.2145, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 16.68, |
|
"learning_rate": 8.124337689862239e-06, |
|
"loss": 0.2417, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 16.75, |
|
"learning_rate": 7.947721653126104e-06, |
|
"loss": 0.2545, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 16.82, |
|
"learning_rate": 7.77110561638997e-06, |
|
"loss": 0.2472, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 16.89, |
|
"learning_rate": 7.594489579653833e-06, |
|
"loss": 0.2041, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 16.96, |
|
"learning_rate": 7.417873542917697e-06, |
|
"loss": 0.222, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_accuracy": 0.9698324022346368, |
|
"eval_loss": 0.08389975875616074, |
|
"eval_runtime": 20.7685, |
|
"eval_samples_per_second": 43.094, |
|
"eval_steps_per_second": 5.393, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 7.241257506181562e-06, |
|
"loss": 0.2161, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 17.1, |
|
"learning_rate": 7.064641469445426e-06, |
|
"loss": 0.2392, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 17.17, |
|
"learning_rate": 6.88802543270929e-06, |
|
"loss": 0.2356, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 17.24, |
|
"learning_rate": 6.7114093959731546e-06, |
|
"loss": 0.2305, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 17.31, |
|
"learning_rate": 6.534793359237019e-06, |
|
"loss": 0.2393, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 17.38, |
|
"learning_rate": 6.358177322500883e-06, |
|
"loss": 0.1977, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 17.45, |
|
"learning_rate": 6.181561285764747e-06, |
|
"loss": 0.1951, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 17.52, |
|
"learning_rate": 6.004945249028612e-06, |
|
"loss": 0.2562, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 17.59, |
|
"learning_rate": 5.8283292122924765e-06, |
|
"loss": 0.2741, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 17.66, |
|
"learning_rate": 5.65171317555634e-06, |
|
"loss": 0.2385, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 17.73, |
|
"learning_rate": 5.475097138820206e-06, |
|
"loss": 0.2471, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 17.8, |
|
"learning_rate": 5.298481102084069e-06, |
|
"loss": 0.2479, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 17.86, |
|
"learning_rate": 5.121865065347934e-06, |
|
"loss": 0.2439, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 17.93, |
|
"learning_rate": 4.945249028611798e-06, |
|
"loss": 0.208, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_accuracy": 0.9731843575418995, |
|
"eval_loss": 0.08734852075576782, |
|
"eval_runtime": 20.3969, |
|
"eval_samples_per_second": 43.879, |
|
"eval_steps_per_second": 5.491, |
|
"step": 2589 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 4.768632991875663e-06, |
|
"loss": 0.2174, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 18.07, |
|
"learning_rate": 4.592016955139527e-06, |
|
"loss": 0.1979, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 18.14, |
|
"learning_rate": 4.415400918403391e-06, |
|
"loss": 0.2435, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 18.21, |
|
"learning_rate": 4.238784881667256e-06, |
|
"loss": 0.2102, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 18.28, |
|
"learning_rate": 4.0621688449311195e-06, |
|
"loss": 0.2216, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 18.35, |
|
"learning_rate": 3.885552808194985e-06, |
|
"loss": 0.233, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 18.42, |
|
"learning_rate": 3.7089367714588486e-06, |
|
"loss": 0.2159, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 18.49, |
|
"learning_rate": 3.532320734722713e-06, |
|
"loss": 0.2606, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 18.56, |
|
"learning_rate": 3.3557046979865773e-06, |
|
"loss": 0.1915, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 18.63, |
|
"learning_rate": 3.1790886612504414e-06, |
|
"loss": 0.2747, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 18.7, |
|
"learning_rate": 3.002472624514306e-06, |
|
"loss": 0.2493, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 18.77, |
|
"learning_rate": 2.82585658777817e-06, |
|
"loss": 0.1871, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 18.84, |
|
"learning_rate": 2.6492405510420347e-06, |
|
"loss": 0.209, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 18.91, |
|
"learning_rate": 2.472624514305899e-06, |
|
"loss": 0.2328, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 18.98, |
|
"learning_rate": 2.2960084775697633e-06, |
|
"loss": 0.2214, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_accuracy": 0.9787709497206704, |
|
"eval_loss": 0.08888570219278336, |
|
"eval_runtime": 19.8638, |
|
"eval_samples_per_second": 45.057, |
|
"eval_steps_per_second": 5.638, |
|
"step": 2733 |
|
}, |
|
{ |
|
"epoch": 19.05, |
|
"learning_rate": 2.119392440833628e-06, |
|
"loss": 0.2477, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 19.12, |
|
"learning_rate": 1.9427764040974924e-06, |
|
"loss": 0.2084, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 19.19, |
|
"learning_rate": 1.7661603673613566e-06, |
|
"loss": 0.2407, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 19.26, |
|
"learning_rate": 1.5895443306252207e-06, |
|
"loss": 0.205, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 19.32, |
|
"learning_rate": 1.412928293889085e-06, |
|
"loss": 0.2047, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 19.39, |
|
"learning_rate": 1.2363122571529496e-06, |
|
"loss": 0.2102, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 19.46, |
|
"learning_rate": 1.059696220416814e-06, |
|
"loss": 0.2407, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 19.53, |
|
"learning_rate": 8.830801836806783e-07, |
|
"loss": 0.2137, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 19.6, |
|
"learning_rate": 7.064641469445425e-07, |
|
"loss": 0.2519, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 19.67, |
|
"learning_rate": 5.29848110208407e-07, |
|
"loss": 0.2114, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 19.74, |
|
"learning_rate": 3.5323207347227126e-07, |
|
"loss": 0.2116, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 19.81, |
|
"learning_rate": 1.7661603673613563e-07, |
|
"loss": 0.2044, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 19.88, |
|
"learning_rate": 0.0, |
|
"loss": 0.2375, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 19.88, |
|
"eval_accuracy": 0.9743016759776536, |
|
"eval_loss": 0.0864485576748848, |
|
"eval_runtime": 21.3872, |
|
"eval_samples_per_second": 41.847, |
|
"eval_steps_per_second": 5.237, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 19.88, |
|
"step": 2860, |
|
"total_flos": 3.399744572389122e+18, |
|
"train_loss": 0.37100943968846245, |
|
"train_runtime": 12472.4089, |
|
"train_samples_per_second": 12.912, |
|
"train_steps_per_second": 0.229 |
|
} |
|
], |
|
"max_steps": 2860, |
|
"num_train_epochs": 20, |
|
"total_flos": 3.399744572389122e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|