|
{ |
|
"best_metric": 0.07669652253389359, |
|
"best_model_checkpoint": "/teamspace/studios/this_studio/output/Mistral_End/checkpoint-4000", |
|
"epoch": 1.9997660545092995, |
|
"eval_steps": 500, |
|
"global_step": 4274, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0023394549070066676, |
|
"grad_norm": 1.2956504821777344, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 0.7204, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.004678909814013335, |
|
"grad_norm": 1.2448140382766724, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 0.7411, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.007018364721020002, |
|
"grad_norm": 1.3956960439682007, |
|
"learning_rate": 1.5e-06, |
|
"loss": 0.7235, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.00935781962802667, |
|
"grad_norm": 1.10906982421875, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.7176, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.011697274535033338, |
|
"grad_norm": 1.1599094867706299, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.7439, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.014036729442040005, |
|
"grad_norm": 1.085955023765564, |
|
"learning_rate": 3e-06, |
|
"loss": 0.6894, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.016376184349046672, |
|
"grad_norm": 0.9708353877067566, |
|
"learning_rate": 3.5000000000000004e-06, |
|
"loss": 0.6607, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.01871563925605334, |
|
"grad_norm": 0.8504016399383545, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.6335, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.021055094163060006, |
|
"grad_norm": 0.7278764843940735, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.5814, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.023394549070066675, |
|
"grad_norm": 0.5420047044754028, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5636, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02573400397707334, |
|
"grad_norm": 0.5314227938652039, |
|
"learning_rate": 5.500000000000001e-06, |
|
"loss": 0.5845, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.02807345888408001, |
|
"grad_norm": 0.48584049940109253, |
|
"learning_rate": 6e-06, |
|
"loss": 0.494, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.030412913791086675, |
|
"grad_norm": 0.49247828125953674, |
|
"learning_rate": 6.5000000000000004e-06, |
|
"loss": 0.4809, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.032752368698093344, |
|
"grad_norm": 0.470062255859375, |
|
"learning_rate": 7.000000000000001e-06, |
|
"loss": 0.446, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.03509182360510001, |
|
"grad_norm": 0.4149915874004364, |
|
"learning_rate": 7.5e-06, |
|
"loss": 0.3915, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.03743127851210668, |
|
"grad_norm": 0.41831153631210327, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.3696, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.039770733419113344, |
|
"grad_norm": 0.40496742725372314, |
|
"learning_rate": 8.500000000000002e-06, |
|
"loss": 0.3464, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.04211018832612001, |
|
"grad_norm": 0.41090327501296997, |
|
"learning_rate": 9e-06, |
|
"loss": 0.3097, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.04444964323312668, |
|
"grad_norm": 0.41252198815345764, |
|
"learning_rate": 9.5e-06, |
|
"loss": 0.3111, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.04678909814013335, |
|
"grad_norm": 0.44755125045776367, |
|
"learning_rate": 1e-05, |
|
"loss": 0.2696, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.04912855304714002, |
|
"grad_norm": 0.4423600137233734, |
|
"learning_rate": 1.05e-05, |
|
"loss": 0.2461, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.05146800795414668, |
|
"grad_norm": 0.44379884004592896, |
|
"learning_rate": 1.1000000000000001e-05, |
|
"loss": 0.2139, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.05380746286115335, |
|
"grad_norm": 0.43690061569213867, |
|
"learning_rate": 1.1500000000000002e-05, |
|
"loss": 0.2045, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.05614691776816002, |
|
"grad_norm": 0.4275108277797699, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.1987, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.05848637267516669, |
|
"grad_norm": 0.5754674077033997, |
|
"learning_rate": 1.25e-05, |
|
"loss": 0.1899, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.06082582758217335, |
|
"grad_norm": 0.4689083397388458, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"loss": 0.171, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.06316528248918002, |
|
"grad_norm": 0.42702773213386536, |
|
"learning_rate": 1.3500000000000001e-05, |
|
"loss": 0.1636, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.06550473739618669, |
|
"grad_norm": 0.4773492217063904, |
|
"learning_rate": 1.4000000000000001e-05, |
|
"loss": 0.1579, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.06784419230319336, |
|
"grad_norm": 0.5190818309783936, |
|
"learning_rate": 1.45e-05, |
|
"loss": 0.1377, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.07018364721020003, |
|
"grad_norm": 0.4696808159351349, |
|
"learning_rate": 1.5e-05, |
|
"loss": 0.1325, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.0725231021172067, |
|
"grad_norm": 0.4878705143928528, |
|
"learning_rate": 1.55e-05, |
|
"loss": 0.1384, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.07486255702421336, |
|
"grad_norm": 0.48388373851776123, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 0.1198, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.07720201193122003, |
|
"grad_norm": 0.47326740622520447, |
|
"learning_rate": 1.65e-05, |
|
"loss": 0.1192, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.07954146683822669, |
|
"grad_norm": 0.4732869565486908, |
|
"learning_rate": 1.7000000000000003e-05, |
|
"loss": 0.1208, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.08188092174523336, |
|
"grad_norm": 0.45708340406417847, |
|
"learning_rate": 1.75e-05, |
|
"loss": 0.1099, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.08422037665224003, |
|
"grad_norm": 0.5060539245605469, |
|
"learning_rate": 1.8e-05, |
|
"loss": 0.104, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.0865598315592467, |
|
"grad_norm": 0.45235252380371094, |
|
"learning_rate": 1.85e-05, |
|
"loss": 0.1489, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.08889928646625336, |
|
"grad_norm": 0.48685696721076965, |
|
"learning_rate": 1.9e-05, |
|
"loss": 0.1102, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.09123874137326003, |
|
"grad_norm": 0.45249509811401367, |
|
"learning_rate": 1.9500000000000003e-05, |
|
"loss": 0.1172, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.0935781962802667, |
|
"grad_norm": 0.456840455532074, |
|
"learning_rate": 2e-05, |
|
"loss": 0.1046, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.09591765118727337, |
|
"grad_norm": 0.4239897131919861, |
|
"learning_rate": 2.05e-05, |
|
"loss": 0.1024, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.09825710609428004, |
|
"grad_norm": 0.3757131099700928, |
|
"learning_rate": 2.1e-05, |
|
"loss": 0.1117, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.1005965610012867, |
|
"grad_norm": 0.43743741512298584, |
|
"learning_rate": 2.15e-05, |
|
"loss": 0.1177, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.10293601590829336, |
|
"grad_norm": 0.39099183678627014, |
|
"learning_rate": 2.2000000000000003e-05, |
|
"loss": 0.0988, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.10527547081530003, |
|
"grad_norm": 0.44669702649116516, |
|
"learning_rate": 2.25e-05, |
|
"loss": 0.0943, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.1076149257223067, |
|
"grad_norm": 0.4071020185947418, |
|
"learning_rate": 2.3000000000000003e-05, |
|
"loss": 0.0989, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.10995438062931337, |
|
"grad_norm": 0.3775785565376282, |
|
"learning_rate": 2.35e-05, |
|
"loss": 0.0953, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.11229383553632004, |
|
"grad_norm": 0.417341947555542, |
|
"learning_rate": 2.4e-05, |
|
"loss": 0.0989, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.11463329044332671, |
|
"grad_norm": 0.43018296360969543, |
|
"learning_rate": 2.45e-05, |
|
"loss": 0.0933, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.11697274535033338, |
|
"grad_norm": 0.4740806818008423, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.1041, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.11931220025734005, |
|
"grad_norm": 0.3639010488986969, |
|
"learning_rate": 2.5500000000000003e-05, |
|
"loss": 0.0983, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.1216516551643467, |
|
"grad_norm": 0.46424400806427, |
|
"learning_rate": 2.6000000000000002e-05, |
|
"loss": 0.1017, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.12399111007135337, |
|
"grad_norm": 0.5055582523345947, |
|
"learning_rate": 2.6500000000000004e-05, |
|
"loss": 0.1486, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.12633056497836004, |
|
"grad_norm": 0.39390507340431213, |
|
"learning_rate": 2.7000000000000002e-05, |
|
"loss": 0.0973, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.12867001988536672, |
|
"grad_norm": 0.35929396748542786, |
|
"learning_rate": 2.7500000000000004e-05, |
|
"loss": 0.0957, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.13100947479237338, |
|
"grad_norm": 0.4053295850753784, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 0.1, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.13334892969938003, |
|
"grad_norm": 0.45273882150650024, |
|
"learning_rate": 2.8499999999999998e-05, |
|
"loss": 0.0878, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.1356883846063867, |
|
"grad_norm": 0.3821883797645569, |
|
"learning_rate": 2.9e-05, |
|
"loss": 0.0881, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.13802783951339337, |
|
"grad_norm": 0.47926971316337585, |
|
"learning_rate": 2.95e-05, |
|
"loss": 0.1115, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.14036729442040005, |
|
"grad_norm": 0.37854063510894775, |
|
"learning_rate": 3e-05, |
|
"loss": 0.1025, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.1427067493274067, |
|
"grad_norm": 0.49879732728004456, |
|
"learning_rate": 3.05e-05, |
|
"loss": 0.0978, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.1450462042344134, |
|
"grad_norm": 0.372403085231781, |
|
"learning_rate": 3.1e-05, |
|
"loss": 0.0931, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.14738565914142004, |
|
"grad_norm": 0.3892665505409241, |
|
"learning_rate": 3.15e-05, |
|
"loss": 0.0947, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.14972511404842673, |
|
"grad_norm": 0.39579498767852783, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 0.0876, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.15206456895543338, |
|
"grad_norm": 0.33920520544052124, |
|
"learning_rate": 3.2500000000000004e-05, |
|
"loss": 0.0941, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.15440402386244007, |
|
"grad_norm": 0.32931971549987793, |
|
"learning_rate": 3.3e-05, |
|
"loss": 0.0825, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.15674347876944672, |
|
"grad_norm": 0.34293580055236816, |
|
"learning_rate": 3.35e-05, |
|
"loss": 0.0803, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.15908293367645338, |
|
"grad_norm": 0.2816343903541565, |
|
"learning_rate": 3.4000000000000007e-05, |
|
"loss": 0.081, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.16142238858346006, |
|
"grad_norm": 0.4197327792644501, |
|
"learning_rate": 3.45e-05, |
|
"loss": 0.0895, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.1637618434904667, |
|
"grad_norm": 0.33747074007987976, |
|
"learning_rate": 3.5e-05, |
|
"loss": 0.0882, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.1661012983974734, |
|
"grad_norm": 0.31165778636932373, |
|
"learning_rate": 3.55e-05, |
|
"loss": 0.0928, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.16844075330448005, |
|
"grad_norm": 0.3359187841415405, |
|
"learning_rate": 3.6e-05, |
|
"loss": 0.0939, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.17078020821148673, |
|
"grad_norm": 0.30003583431243896, |
|
"learning_rate": 3.65e-05, |
|
"loss": 0.0842, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.1731196631184934, |
|
"grad_norm": 0.35046064853668213, |
|
"learning_rate": 3.7e-05, |
|
"loss": 0.0954, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.17545911802550007, |
|
"grad_norm": 0.37043917179107666, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 0.0838, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.17779857293250673, |
|
"grad_norm": 0.30147868394851685, |
|
"learning_rate": 3.8e-05, |
|
"loss": 0.101, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.18013802783951338, |
|
"grad_norm": 0.3372386395931244, |
|
"learning_rate": 3.85e-05, |
|
"loss": 0.0873, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.18247748274652006, |
|
"grad_norm": 0.32464468479156494, |
|
"learning_rate": 3.9000000000000006e-05, |
|
"loss": 0.09, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.18481693765352672, |
|
"grad_norm": 0.3182346224784851, |
|
"learning_rate": 3.9500000000000005e-05, |
|
"loss": 0.0892, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.1871563925605334, |
|
"grad_norm": 0.2824024558067322, |
|
"learning_rate": 4e-05, |
|
"loss": 0.0832, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.18949584746754006, |
|
"grad_norm": 0.30845776200294495, |
|
"learning_rate": 4.05e-05, |
|
"loss": 0.093, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.19183530237454674, |
|
"grad_norm": 0.3025493323802948, |
|
"learning_rate": 4.1e-05, |
|
"loss": 0.0877, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.1941747572815534, |
|
"grad_norm": 0.35149937868118286, |
|
"learning_rate": 4.15e-05, |
|
"loss": 0.0841, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.19651421218856008, |
|
"grad_norm": 0.34748056530952454, |
|
"learning_rate": 4.2e-05, |
|
"loss": 0.0923, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.19885366709556673, |
|
"grad_norm": 0.3216325640678406, |
|
"learning_rate": 4.25e-05, |
|
"loss": 0.081, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.2011931220025734, |
|
"grad_norm": 0.2797127366065979, |
|
"learning_rate": 4.3e-05, |
|
"loss": 0.0808, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.20353257690958007, |
|
"grad_norm": 0.2927537262439728, |
|
"learning_rate": 4.35e-05, |
|
"loss": 0.1383, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.20587203181658673, |
|
"grad_norm": 0.2892308533191681, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"loss": 0.0836, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.2082114867235934, |
|
"grad_norm": 0.35608819127082825, |
|
"learning_rate": 4.4500000000000004e-05, |
|
"loss": 0.0796, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.21055094163060006, |
|
"grad_norm": 0.29528915882110596, |
|
"learning_rate": 4.5e-05, |
|
"loss": 0.09, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.21289039653760675, |
|
"grad_norm": 0.3688681721687317, |
|
"learning_rate": 4.55e-05, |
|
"loss": 0.0928, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.2152298514446134, |
|
"grad_norm": 0.3129563629627228, |
|
"learning_rate": 4.600000000000001e-05, |
|
"loss": 0.0914, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.21756930635162008, |
|
"grad_norm": 0.34032294154167175, |
|
"learning_rate": 4.6500000000000005e-05, |
|
"loss": 0.0978, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.21990876125862674, |
|
"grad_norm": 0.283371239900589, |
|
"learning_rate": 4.7e-05, |
|
"loss": 0.1109, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.2222482161656334, |
|
"grad_norm": 0.25409555435180664, |
|
"learning_rate": 4.75e-05, |
|
"loss": 0.0804, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.22458767107264008, |
|
"grad_norm": 0.24365410208702087, |
|
"learning_rate": 4.8e-05, |
|
"loss": 0.0737, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.22692712597964673, |
|
"grad_norm": 0.28039586544036865, |
|
"learning_rate": 4.85e-05, |
|
"loss": 0.0801, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.22926658088665341, |
|
"grad_norm": 0.28205862641334534, |
|
"learning_rate": 4.9e-05, |
|
"loss": 0.0784, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.23160603579366007, |
|
"grad_norm": 0.2688550353050232, |
|
"learning_rate": 4.9500000000000004e-05, |
|
"loss": 0.0838, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.23394549070066675, |
|
"grad_norm": 0.31778618693351746, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0893, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.23394549070066675, |
|
"eval_loss": 0.09720832854509354, |
|
"eval_runtime": 235.4562, |
|
"eval_samples_per_second": 2.964, |
|
"eval_steps_per_second": 0.743, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.2362849456076734, |
|
"grad_norm": 0.27649974822998047, |
|
"learning_rate": 4.999978345640206e-05, |
|
"loss": 0.0818, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.2386244005146801, |
|
"grad_norm": 0.24827256798744202, |
|
"learning_rate": 4.9999133829359514e-05, |
|
"loss": 0.0812, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.24096385542168675, |
|
"grad_norm": 0.24930593371391296, |
|
"learning_rate": 4.999805113012618e-05, |
|
"loss": 0.0732, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.2433033103286934, |
|
"grad_norm": 0.2697354555130005, |
|
"learning_rate": 4.999653537745817e-05, |
|
"loss": 0.0843, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.24564276523570008, |
|
"grad_norm": 0.258434534072876, |
|
"learning_rate": 4.999458659761363e-05, |
|
"loss": 0.07, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.24798222014270674, |
|
"grad_norm": 0.2943266034126282, |
|
"learning_rate": 4.9992204824352213e-05, |
|
"loss": 0.0852, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.2503216750497134, |
|
"grad_norm": 0.23471561074256897, |
|
"learning_rate": 4.998939009893453e-05, |
|
"loss": 0.0847, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.2526611299567201, |
|
"grad_norm": 0.27082759141921997, |
|
"learning_rate": 4.9986142470121453e-05, |
|
"loss": 0.0727, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.25500058486372673, |
|
"grad_norm": 0.26077497005462646, |
|
"learning_rate": 4.9982461994173234e-05, |
|
"loss": 0.0893, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.25734003977073344, |
|
"grad_norm": 0.2686660587787628, |
|
"learning_rate": 4.9978348734848556e-05, |
|
"loss": 0.0764, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.2596794946777401, |
|
"grad_norm": 0.2592894434928894, |
|
"learning_rate": 4.9973802763403424e-05, |
|
"loss": 0.1747, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.26201894958474675, |
|
"grad_norm": 0.26700106263160706, |
|
"learning_rate": 4.996882415858991e-05, |
|
"loss": 0.0808, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.2643584044917534, |
|
"grad_norm": 0.29679951071739197, |
|
"learning_rate": 4.996341300665481e-05, |
|
"loss": 0.0847, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.26669785939876006, |
|
"grad_norm": 0.29322996735572815, |
|
"learning_rate": 4.995756940133817e-05, |
|
"loss": 0.0884, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.2690373143057668, |
|
"grad_norm": 0.24166248738765717, |
|
"learning_rate": 4.995129344387158e-05, |
|
"loss": 0.0784, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.2713767692127734, |
|
"grad_norm": 0.23448926210403442, |
|
"learning_rate": 4.9944585242976546e-05, |
|
"loss": 0.146, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.2737162241197801, |
|
"grad_norm": 0.2622728645801544, |
|
"learning_rate": 4.99374449148625e-05, |
|
"loss": 0.0894, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.27605567902678674, |
|
"grad_norm": 0.23316816985607147, |
|
"learning_rate": 4.992987258322481e-05, |
|
"loss": 0.0821, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.27839513393379345, |
|
"grad_norm": 0.29072973132133484, |
|
"learning_rate": 4.992186837924268e-05, |
|
"loss": 0.098, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.2807345888408001, |
|
"grad_norm": 0.20675267279148102, |
|
"learning_rate": 4.9913432441576845e-05, |
|
"loss": 0.0776, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.28307404374780676, |
|
"grad_norm": 0.26899468898773193, |
|
"learning_rate": 4.9904564916367166e-05, |
|
"loss": 0.0862, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.2854134986548134, |
|
"grad_norm": 0.22569362819194794, |
|
"learning_rate": 4.989526595723012e-05, |
|
"loss": 0.0921, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.28775295356182007, |
|
"grad_norm": 0.2874622046947479, |
|
"learning_rate": 4.988553572525609e-05, |
|
"loss": 0.0811, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.2900924084688268, |
|
"grad_norm": 0.2119138389825821, |
|
"learning_rate": 4.987537438900664e-05, |
|
"loss": 0.0711, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.29243186337583343, |
|
"grad_norm": 0.2240922749042511, |
|
"learning_rate": 4.986478212451157e-05, |
|
"loss": 0.0802, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.2947713182828401, |
|
"grad_norm": 0.24686302244663239, |
|
"learning_rate": 4.985375911526582e-05, |
|
"loss": 0.0825, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.29711077318984674, |
|
"grad_norm": 0.22553415596485138, |
|
"learning_rate": 4.9842305552226365e-05, |
|
"loss": 0.0676, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.29945022809685345, |
|
"grad_norm": 0.24262481927871704, |
|
"learning_rate": 4.9830421633808876e-05, |
|
"loss": 0.0752, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.3017896830038601, |
|
"grad_norm": 0.24724553525447845, |
|
"learning_rate": 4.981810756588426e-05, |
|
"loss": 0.0823, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.30412913791086676, |
|
"grad_norm": 0.2152285873889923, |
|
"learning_rate": 4.980536356177513e-05, |
|
"loss": 0.0795, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.3064685928178734, |
|
"grad_norm": 0.2218402624130249, |
|
"learning_rate": 4.9792189842252074e-05, |
|
"loss": 0.0711, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.30880804772488013, |
|
"grad_norm": 0.24227651953697205, |
|
"learning_rate": 4.977858663552987e-05, |
|
"loss": 0.0756, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.3111475026318868, |
|
"grad_norm": 0.22776497900485992, |
|
"learning_rate": 4.9764554177263507e-05, |
|
"loss": 0.0742, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.31348695753889344, |
|
"grad_norm": 0.20029255747795105, |
|
"learning_rate": 4.975009271054409e-05, |
|
"loss": 0.0698, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.3158264124459001, |
|
"grad_norm": 0.21585966646671295, |
|
"learning_rate": 4.973520248589469e-05, |
|
"loss": 0.0858, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.31816586735290675, |
|
"grad_norm": 0.23440401256084442, |
|
"learning_rate": 4.9719883761265906e-05, |
|
"loss": 0.0818, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.32050532225991346, |
|
"grad_norm": 0.22107180953025818, |
|
"learning_rate": 4.9704136802031485e-05, |
|
"loss": 0.0671, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.3228447771669201, |
|
"grad_norm": 0.19328860938549042, |
|
"learning_rate": 4.968796188098369e-05, |
|
"loss": 0.0725, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.32518423207392677, |
|
"grad_norm": 0.2299336940050125, |
|
"learning_rate": 4.967135927832856e-05, |
|
"loss": 0.0852, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.3275236869809334, |
|
"grad_norm": 0.24605049192905426, |
|
"learning_rate": 4.9654329281681094e-05, |
|
"loss": 0.0754, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.32986314188794014, |
|
"grad_norm": 0.23175948858261108, |
|
"learning_rate": 4.9636872186060215e-05, |
|
"loss": 0.0961, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.3322025967949468, |
|
"grad_norm": 0.2416304647922516, |
|
"learning_rate": 4.961898829388372e-05, |
|
"loss": 0.1093, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.33454205170195345, |
|
"grad_norm": 0.20699331164360046, |
|
"learning_rate": 4.960067791496299e-05, |
|
"loss": 0.0793, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.3368815066089601, |
|
"grad_norm": 0.22395586967468262, |
|
"learning_rate": 4.958194136649765e-05, |
|
"loss": 0.0805, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.33922096151596676, |
|
"grad_norm": 0.21491140127182007, |
|
"learning_rate": 4.956277897307008e-05, |
|
"loss": 0.073, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.34156041642297347, |
|
"grad_norm": 0.23053660988807678, |
|
"learning_rate": 4.954319106663976e-05, |
|
"loss": 0.0769, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.3438998713299801, |
|
"grad_norm": 0.21306245028972626, |
|
"learning_rate": 4.952317798653755e-05, |
|
"loss": 0.075, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.3462393262369868, |
|
"grad_norm": 0.1973639279603958, |
|
"learning_rate": 4.95027400794598e-05, |
|
"loss": 0.0697, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.34857878114399343, |
|
"grad_norm": 0.2036924660205841, |
|
"learning_rate": 4.948187769946234e-05, |
|
"loss": 0.08, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.35091823605100014, |
|
"grad_norm": 0.2153988480567932, |
|
"learning_rate": 4.946059120795437e-05, |
|
"loss": 0.0963, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.3532576909580068, |
|
"grad_norm": 0.20555338263511658, |
|
"learning_rate": 4.943888097369216e-05, |
|
"loss": 0.0913, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.35559714586501345, |
|
"grad_norm": 0.26464715600013733, |
|
"learning_rate": 4.941674737277268e-05, |
|
"loss": 0.0845, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.3579366007720201, |
|
"grad_norm": 0.19657152891159058, |
|
"learning_rate": 4.9394190788627106e-05, |
|
"loss": 0.0756, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.36027605567902676, |
|
"grad_norm": 0.21668066084384918, |
|
"learning_rate": 4.937121161201415e-05, |
|
"loss": 0.0803, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.3626155105860335, |
|
"grad_norm": 0.24815738201141357, |
|
"learning_rate": 4.9347810241013294e-05, |
|
"loss": 0.0823, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.36495496549304013, |
|
"grad_norm": 0.20182116329669952, |
|
"learning_rate": 4.932398708101791e-05, |
|
"loss": 0.0703, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.3672944204000468, |
|
"grad_norm": 0.2374674379825592, |
|
"learning_rate": 4.92997425447282e-05, |
|
"loss": 0.084, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.36963387530705344, |
|
"grad_norm": 0.2246353179216385, |
|
"learning_rate": 4.927507705214412e-05, |
|
"loss": 0.0868, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.37197333021406015, |
|
"grad_norm": 0.21759703755378723, |
|
"learning_rate": 4.924999103055802e-05, |
|
"loss": 0.0754, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.3743127851210668, |
|
"grad_norm": 0.20431989431381226, |
|
"learning_rate": 4.922448491454729e-05, |
|
"loss": 0.0803, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.37665224002807346, |
|
"grad_norm": 0.20232437551021576, |
|
"learning_rate": 4.9198559145966824e-05, |
|
"loss": 0.0751, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.3789916949350801, |
|
"grad_norm": 0.1968168169260025, |
|
"learning_rate": 4.917221417394134e-05, |
|
"loss": 0.0798, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.38133114984208677, |
|
"grad_norm": 0.2001713663339615, |
|
"learning_rate": 4.914545045485767e-05, |
|
"loss": 0.0776, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.3836706047490935, |
|
"grad_norm": 0.19992566108703613, |
|
"learning_rate": 4.911826845235676e-05, |
|
"loss": 0.0767, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.38601005965610014, |
|
"grad_norm": 0.21567636728286743, |
|
"learning_rate": 4.9090668637325696e-05, |
|
"loss": 0.0752, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.3883495145631068, |
|
"grad_norm": 0.19452187418937683, |
|
"learning_rate": 4.906265148788954e-05, |
|
"loss": 0.0722, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.39068896947011345, |
|
"grad_norm": 0.23549924790859222, |
|
"learning_rate": 4.9034217489403045e-05, |
|
"loss": 0.0764, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.39302842437712016, |
|
"grad_norm": 0.2083604633808136, |
|
"learning_rate": 4.9005367134442235e-05, |
|
"loss": 0.0767, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.3953678792841268, |
|
"grad_norm": 0.2386019378900528, |
|
"learning_rate": 4.8976100922795884e-05, |
|
"loss": 0.0815, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.39770733419113347, |
|
"grad_norm": 0.19933900237083435, |
|
"learning_rate": 4.894641936145686e-05, |
|
"loss": 0.0935, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.4000467890981401, |
|
"grad_norm": 0.16899636387825012, |
|
"learning_rate": 4.8916322964613316e-05, |
|
"loss": 0.0767, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.4023862440051468, |
|
"grad_norm": 0.20377202332019806, |
|
"learning_rate": 4.888581225363982e-05, |
|
"loss": 0.0771, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.4047256989121535, |
|
"grad_norm": 0.2286689430475235, |
|
"learning_rate": 4.885488775708831e-05, |
|
"loss": 0.0798, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.40706515381916014, |
|
"grad_norm": 0.2132563591003418, |
|
"learning_rate": 4.882355001067892e-05, |
|
"loss": 0.0723, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.4094046087261668, |
|
"grad_norm": 0.21681547164916992, |
|
"learning_rate": 4.879179955729072e-05, |
|
"loss": 0.0777, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.41174406363317345, |
|
"grad_norm": 0.20257344841957092, |
|
"learning_rate": 4.87596369469523e-05, |
|
"loss": 0.0777, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.41408351854018016, |
|
"grad_norm": 0.18595652282238007, |
|
"learning_rate": 4.872706273683225e-05, |
|
"loss": 0.0648, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.4164229734471868, |
|
"grad_norm": 0.1963788866996765, |
|
"learning_rate": 4.869407749122951e-05, |
|
"loss": 0.0688, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.41876242835419347, |
|
"grad_norm": 0.22948507964611053, |
|
"learning_rate": 4.8660681781563576e-05, |
|
"loss": 0.0715, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.4211018832612001, |
|
"grad_norm": 0.16155906021595, |
|
"learning_rate": 4.8626876186364624e-05, |
|
"loss": 0.0651, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.4234413381682068, |
|
"grad_norm": 0.193013533949852, |
|
"learning_rate": 4.859266129126345e-05, |
|
"loss": 0.0754, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.4257807930752135, |
|
"grad_norm": 0.21157501637935638, |
|
"learning_rate": 4.85580376889814e-05, |
|
"loss": 0.1229, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.42812024798222015, |
|
"grad_norm": 0.2193852812051773, |
|
"learning_rate": 4.8523005979320015e-05, |
|
"loss": 0.0755, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.4304597028892268, |
|
"grad_norm": 0.20411770045757294, |
|
"learning_rate": 4.848756676915069e-05, |
|
"loss": 0.0848, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.43279915779623346, |
|
"grad_norm": 0.23284126818180084, |
|
"learning_rate": 4.845172067240415e-05, |
|
"loss": 0.0764, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.43513861270324017, |
|
"grad_norm": 0.21067671477794647, |
|
"learning_rate": 4.841546831005982e-05, |
|
"loss": 0.0821, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.4374780676102468, |
|
"grad_norm": 0.1912618726491928, |
|
"learning_rate": 4.837881031013506e-05, |
|
"loss": 0.0755, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.4398175225172535, |
|
"grad_norm": 0.1677297055721283, |
|
"learning_rate": 4.8341747307674276e-05, |
|
"loss": 0.0695, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.44215697742426013, |
|
"grad_norm": 0.22734542191028595, |
|
"learning_rate": 4.8304279944737954e-05, |
|
"loss": 0.086, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.4444964323312668, |
|
"grad_norm": 0.1839355230331421, |
|
"learning_rate": 4.8266408870391484e-05, |
|
"loss": 0.0797, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.4468358872382735, |
|
"grad_norm": 0.20233304798603058, |
|
"learning_rate": 4.822813474069398e-05, |
|
"loss": 0.084, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.44917534214528015, |
|
"grad_norm": 0.20918181538581848, |
|
"learning_rate": 4.8189458218686846e-05, |
|
"loss": 0.0754, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.4515147970522868, |
|
"grad_norm": 0.2046424299478531, |
|
"learning_rate": 4.815037997438234e-05, |
|
"loss": 0.0762, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.45385425195929346, |
|
"grad_norm": 0.20791682600975037, |
|
"learning_rate": 4.811090068475197e-05, |
|
"loss": 0.0686, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.4561937068663002, |
|
"grad_norm": 0.1891777068376541, |
|
"learning_rate": 4.807102103371472e-05, |
|
"loss": 0.0818, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.45853316177330683, |
|
"grad_norm": 0.1632051318883896, |
|
"learning_rate": 4.803074171212524e-05, |
|
"loss": 0.0693, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.4608726166803135, |
|
"grad_norm": 0.18187515437602997, |
|
"learning_rate": 4.799006341776185e-05, |
|
"loss": 0.0642, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.46321207158732014, |
|
"grad_norm": 0.20138341188430786, |
|
"learning_rate": 4.7948986855314523e-05, |
|
"loss": 0.0759, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.4655515264943268, |
|
"grad_norm": 0.19479338824748993, |
|
"learning_rate": 4.790751273637256e-05, |
|
"loss": 0.0761, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.4678909814013335, |
|
"grad_norm": 0.18999288976192474, |
|
"learning_rate": 4.7865641779412366e-05, |
|
"loss": 0.0753, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.4678909814013335, |
|
"eval_loss": 0.08787659555673599, |
|
"eval_runtime": 236.9363, |
|
"eval_samples_per_second": 2.946, |
|
"eval_steps_per_second": 0.739, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.47023043630834016, |
|
"grad_norm": 0.22196418046951294, |
|
"learning_rate": 4.7823374709784955e-05, |
|
"loss": 0.0731, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.4725698912153468, |
|
"grad_norm": 0.17047221958637238, |
|
"learning_rate": 4.77807122597034e-05, |
|
"loss": 0.082, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.47490934612235347, |
|
"grad_norm": 0.23016561567783356, |
|
"learning_rate": 4.7737655168230114e-05, |
|
"loss": 0.0775, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.4772488010293602, |
|
"grad_norm": 0.21696537733078003, |
|
"learning_rate": 4.769420418126412e-05, |
|
"loss": 0.0805, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.47958825593636684, |
|
"grad_norm": 0.19352155923843384, |
|
"learning_rate": 4.7650360051528054e-05, |
|
"loss": 0.0754, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.4819277108433735, |
|
"grad_norm": 0.1966516673564911, |
|
"learning_rate": 4.760612353855517e-05, |
|
"loss": 0.068, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.48426716575038015, |
|
"grad_norm": 0.22759053111076355, |
|
"learning_rate": 4.7561495408676146e-05, |
|
"loss": 0.073, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.4866066206573868, |
|
"grad_norm": 0.17850154638290405, |
|
"learning_rate": 4.7516476435005865e-05, |
|
"loss": 0.074, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.4889460755643935, |
|
"grad_norm": 0.19016915559768677, |
|
"learning_rate": 4.7471067397429956e-05, |
|
"loss": 0.0734, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.49128553047140017, |
|
"grad_norm": 0.19601960480213165, |
|
"learning_rate": 4.742526908259134e-05, |
|
"loss": 0.0763, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.4936249853784068, |
|
"grad_norm": 0.17694705724716187, |
|
"learning_rate": 4.7379082283876566e-05, |
|
"loss": 0.0703, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.4959644402854135, |
|
"grad_norm": 0.18579581379890442, |
|
"learning_rate": 4.733250780140206e-05, |
|
"loss": 0.0863, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.4983038951924202, |
|
"grad_norm": 0.18680576980113983, |
|
"learning_rate": 4.728554644200034e-05, |
|
"loss": 0.1042, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.5006433500994268, |
|
"grad_norm": 0.1797022670507431, |
|
"learning_rate": 4.723819901920591e-05, |
|
"loss": 0.0743, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.5029828050064336, |
|
"grad_norm": 0.20039916038513184, |
|
"learning_rate": 4.719046635324129e-05, |
|
"loss": 0.0707, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.5053222599134402, |
|
"grad_norm": 0.19813776016235352, |
|
"learning_rate": 4.7142349271002735e-05, |
|
"loss": 0.0738, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.5076617148204469, |
|
"grad_norm": 0.18406710028648376, |
|
"learning_rate": 4.709384860604593e-05, |
|
"loss": 0.0724, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.5100011697274535, |
|
"grad_norm": 0.20730999112129211, |
|
"learning_rate": 4.704496519857156e-05, |
|
"loss": 0.0738, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.5123406246344602, |
|
"grad_norm": 0.2113994061946869, |
|
"learning_rate": 4.699569989541074e-05, |
|
"loss": 0.0699, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.5146800795414669, |
|
"grad_norm": 0.1848883181810379, |
|
"learning_rate": 4.694605355001034e-05, |
|
"loss": 0.072, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.5170195344484735, |
|
"grad_norm": 0.19623607397079468, |
|
"learning_rate": 4.689602702241823e-05, |
|
"loss": 0.0761, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.5193589893554802, |
|
"grad_norm": 0.18376454710960388, |
|
"learning_rate": 4.684562117926835e-05, |
|
"loss": 0.0793, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.5216984442624868, |
|
"grad_norm": 0.18506892025470734, |
|
"learning_rate": 4.679483689376571e-05, |
|
"loss": 0.0731, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.5240378991694935, |
|
"grad_norm": 0.1819106787443161, |
|
"learning_rate": 4.674367504567127e-05, |
|
"loss": 0.0795, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.5263773540765002, |
|
"grad_norm": 0.19489003717899323, |
|
"learning_rate": 4.669213652128667e-05, |
|
"loss": 0.0716, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.5287168089835068, |
|
"grad_norm": 0.16565723717212677, |
|
"learning_rate": 4.664022221343892e-05, |
|
"loss": 0.0717, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.5310562638905135, |
|
"grad_norm": 0.18749143183231354, |
|
"learning_rate": 4.658793302146489e-05, |
|
"loss": 0.0756, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.5333957187975201, |
|
"grad_norm": 0.2000792771577835, |
|
"learning_rate": 4.653526985119577e-05, |
|
"loss": 0.0789, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.5357351737045268, |
|
"grad_norm": 0.17934930324554443, |
|
"learning_rate": 4.648223361494135e-05, |
|
"loss": 0.0823, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.5380746286115335, |
|
"grad_norm": 0.204596146941185, |
|
"learning_rate": 4.642882523147422e-05, |
|
"loss": 0.0804, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.5404140835185401, |
|
"grad_norm": 0.17674653232097626, |
|
"learning_rate": 4.637504562601386e-05, |
|
"loss": 0.0699, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.5427535384255469, |
|
"grad_norm": 0.17949679493904114, |
|
"learning_rate": 4.6320895730210616e-05, |
|
"loss": 0.0717, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.5450929933325536, |
|
"grad_norm": 0.1828782856464386, |
|
"learning_rate": 4.626637648212955e-05, |
|
"loss": 0.0816, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.5474324482395602, |
|
"grad_norm": 0.1803893744945526, |
|
"learning_rate": 4.6211488826234186e-05, |
|
"loss": 0.0766, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.5497719031465669, |
|
"grad_norm": 0.1853674352169037, |
|
"learning_rate": 4.615623371337016e-05, |
|
"loss": 0.0728, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.5521113580535735, |
|
"grad_norm": 0.20650415122509003, |
|
"learning_rate": 4.6100612100748765e-05, |
|
"loss": 0.073, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.5544508129605802, |
|
"grad_norm": 0.18589161336421967, |
|
"learning_rate": 4.604462495193031e-05, |
|
"loss": 0.0803, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.5567902678675869, |
|
"grad_norm": 0.16948671638965607, |
|
"learning_rate": 4.59882732368075e-05, |
|
"loss": 0.0728, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.5591297227745935, |
|
"grad_norm": 0.17474383115768433, |
|
"learning_rate": 4.593155793158859e-05, |
|
"loss": 0.0697, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.5614691776816002, |
|
"grad_norm": 0.1744118481874466, |
|
"learning_rate": 4.5874480018780456e-05, |
|
"loss": 0.0771, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.5638086325886068, |
|
"grad_norm": 0.17954926192760468, |
|
"learning_rate": 4.581704048717166e-05, |
|
"loss": 0.0722, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.5661480874956135, |
|
"grad_norm": 0.19267572462558746, |
|
"learning_rate": 4.57592403318152e-05, |
|
"loss": 0.0739, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.5684875424026202, |
|
"grad_norm": 0.1731371432542801, |
|
"learning_rate": 4.570108055401138e-05, |
|
"loss": 0.076, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.5708269973096268, |
|
"grad_norm": 0.18417227268218994, |
|
"learning_rate": 4.5642562161290406e-05, |
|
"loss": 0.0708, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.5731664522166335, |
|
"grad_norm": 0.19248449802398682, |
|
"learning_rate": 4.558368616739493e-05, |
|
"loss": 0.0775, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.5755059071236401, |
|
"grad_norm": 0.20043963193893433, |
|
"learning_rate": 4.552445359226252e-05, |
|
"loss": 0.0758, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.5778453620306468, |
|
"grad_norm": 0.20923319458961487, |
|
"learning_rate": 4.546486546200798e-05, |
|
"loss": 0.0648, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.5801848169376536, |
|
"grad_norm": 0.19847044348716736, |
|
"learning_rate": 4.540492280890555e-05, |
|
"loss": 0.0684, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.5825242718446602, |
|
"grad_norm": 0.17543412744998932, |
|
"learning_rate": 4.534462667137105e-05, |
|
"loss": 0.0746, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.5848637267516669, |
|
"grad_norm": 0.1842990517616272, |
|
"learning_rate": 4.528397809394389e-05, |
|
"loss": 0.0771, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.5872031816586736, |
|
"grad_norm": 0.15888415277004242, |
|
"learning_rate": 4.522297812726897e-05, |
|
"loss": 0.0654, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.5895426365656802, |
|
"grad_norm": 0.20020891726016998, |
|
"learning_rate": 4.516162782807845e-05, |
|
"loss": 0.0759, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.5918820914726869, |
|
"grad_norm": 0.164358988404274, |
|
"learning_rate": 4.5099928259173516e-05, |
|
"loss": 0.0666, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.5942215463796935, |
|
"grad_norm": 0.16114068031311035, |
|
"learning_rate": 4.503788048940589e-05, |
|
"loss": 0.0677, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.5965610012867002, |
|
"grad_norm": 0.18885543942451477, |
|
"learning_rate": 4.497548559365935e-05, |
|
"loss": 0.0697, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.5989004561937069, |
|
"grad_norm": 0.16143983602523804, |
|
"learning_rate": 4.4912744652831116e-05, |
|
"loss": 0.072, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.6012399111007135, |
|
"grad_norm": 0.16090819239616394, |
|
"learning_rate": 4.4849658753813126e-05, |
|
"loss": 0.068, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.6035793660077202, |
|
"grad_norm": 0.19243429601192474, |
|
"learning_rate": 4.4786228989473164e-05, |
|
"loss": 0.0744, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.6059188209147268, |
|
"grad_norm": 0.17546018958091736, |
|
"learning_rate": 4.4722456458635995e-05, |
|
"loss": 0.0745, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.6082582758217335, |
|
"grad_norm": 0.1936197578907013, |
|
"learning_rate": 4.465834226606428e-05, |
|
"loss": 0.0769, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.6105977307287402, |
|
"grad_norm": 0.16884423792362213, |
|
"learning_rate": 4.459388752243945e-05, |
|
"loss": 0.0809, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.6129371856357468, |
|
"grad_norm": 0.20882828533649445, |
|
"learning_rate": 4.452909334434247e-05, |
|
"loss": 0.08, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.6152766405427536, |
|
"grad_norm": 0.1857665777206421, |
|
"learning_rate": 4.4463960854234506e-05, |
|
"loss": 0.0668, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.6176160954497603, |
|
"grad_norm": 0.17508915066719055, |
|
"learning_rate": 4.4398491180437446e-05, |
|
"loss": 0.0775, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.6199555503567669, |
|
"grad_norm": 0.16894185543060303, |
|
"learning_rate": 4.43326854571144e-05, |
|
"loss": 0.0714, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.6222950052637736, |
|
"grad_norm": 0.19014866650104523, |
|
"learning_rate": 4.426654482425e-05, |
|
"loss": 0.0789, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.6246344601707802, |
|
"grad_norm": 0.15668885409832, |
|
"learning_rate": 4.420007042763071e-05, |
|
"loss": 0.0641, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.6269739150777869, |
|
"grad_norm": 0.15578554570674896, |
|
"learning_rate": 4.413326341882492e-05, |
|
"loss": 0.0665, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.6293133699847936, |
|
"grad_norm": 0.20437653362751007, |
|
"learning_rate": 4.4066124955163046e-05, |
|
"loss": 0.0784, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.6316528248918002, |
|
"grad_norm": 0.16404947638511658, |
|
"learning_rate": 4.3998656199717435e-05, |
|
"loss": 0.0675, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.6339922797988069, |
|
"grad_norm": 0.147497296333313, |
|
"learning_rate": 4.393085832128226e-05, |
|
"loss": 0.0711, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.6363317347058135, |
|
"grad_norm": 0.17336557805538177, |
|
"learning_rate": 4.386273249435324e-05, |
|
"loss": 0.0766, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.6386711896128202, |
|
"grad_norm": 0.18283897638320923, |
|
"learning_rate": 4.3794279899107304e-05, |
|
"loss": 0.0746, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.6410106445198269, |
|
"grad_norm": 0.16502517461776733, |
|
"learning_rate": 4.3725501721382165e-05, |
|
"loss": 0.0638, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.6433500994268335, |
|
"grad_norm": 0.16824820637702942, |
|
"learning_rate": 4.3656399152655735e-05, |
|
"loss": 0.0698, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.6456895543338402, |
|
"grad_norm": 0.17634496092796326, |
|
"learning_rate": 4.358697339002553e-05, |
|
"loss": 0.0952, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.6480290092408468, |
|
"grad_norm": 0.1962941437959671, |
|
"learning_rate": 4.3517225636187906e-05, |
|
"loss": 0.0728, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.6503684641478535, |
|
"grad_norm": 0.1725791096687317, |
|
"learning_rate": 4.344715709941722e-05, |
|
"loss": 0.0721, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.6527079190548603, |
|
"grad_norm": 0.1703636348247528, |
|
"learning_rate": 4.337676899354493e-05, |
|
"loss": 0.072, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.6550473739618669, |
|
"grad_norm": 0.16488994657993317, |
|
"learning_rate": 4.330606253793851e-05, |
|
"loss": 0.072, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.6573868288688736, |
|
"grad_norm": 0.16995395720005035, |
|
"learning_rate": 4.32350389574804e-05, |
|
"loss": 0.0747, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.6597262837758803, |
|
"grad_norm": 0.17242836952209473, |
|
"learning_rate": 4.316369948254674e-05, |
|
"loss": 0.0698, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.6620657386828869, |
|
"grad_norm": 0.16680659353733063, |
|
"learning_rate": 4.3092045348986034e-05, |
|
"loss": 0.0683, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.6644051935898936, |
|
"grad_norm": 0.18627899885177612, |
|
"learning_rate": 4.302007779809781e-05, |
|
"loss": 0.0672, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.6667446484969002, |
|
"grad_norm": 0.14819131791591644, |
|
"learning_rate": 4.294779807661105e-05, |
|
"loss": 0.0683, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.6690841034039069, |
|
"grad_norm": 0.18117517232894897, |
|
"learning_rate": 4.287520743666263e-05, |
|
"loss": 0.0728, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.6714235583109136, |
|
"grad_norm": 0.17058628797531128, |
|
"learning_rate": 4.280230713577564e-05, |
|
"loss": 0.077, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.6737630132179202, |
|
"grad_norm": 0.15503695607185364, |
|
"learning_rate": 4.2729098436837536e-05, |
|
"loss": 0.0739, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.6761024681249269, |
|
"grad_norm": 0.1836390346288681, |
|
"learning_rate": 4.2655582608078315e-05, |
|
"loss": 0.079, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.6784419230319335, |
|
"grad_norm": 0.15766695141792297, |
|
"learning_rate": 4.2581760923048554e-05, |
|
"loss": 0.0653, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.6807813779389402, |
|
"grad_norm": 0.18035438656806946, |
|
"learning_rate": 4.2507634660597315e-05, |
|
"loss": 0.0793, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.6831208328459469, |
|
"grad_norm": 0.1778097152709961, |
|
"learning_rate": 4.243320510485001e-05, |
|
"loss": 0.0689, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.6854602877529535, |
|
"grad_norm": 0.1773742139339447, |
|
"learning_rate": 4.235847354518614e-05, |
|
"loss": 0.0726, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.6877997426599602, |
|
"grad_norm": 0.16845591366291046, |
|
"learning_rate": 4.228344127621695e-05, |
|
"loss": 0.0691, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.6901391975669668, |
|
"grad_norm": 0.14897610247135162, |
|
"learning_rate": 4.220810959776307e-05, |
|
"loss": 0.0684, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.6924786524739736, |
|
"grad_norm": 0.17343005537986755, |
|
"learning_rate": 4.213247981483189e-05, |
|
"loss": 0.0624, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.6948181073809803, |
|
"grad_norm": 0.19513456523418427, |
|
"learning_rate": 4.205655323759505e-05, |
|
"loss": 0.0751, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.6971575622879869, |
|
"grad_norm": 0.17424218356609344, |
|
"learning_rate": 4.1980331181365685e-05, |
|
"loss": 0.0704, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.6994970171949936, |
|
"grad_norm": 0.1491064429283142, |
|
"learning_rate": 4.190381496657565e-05, |
|
"loss": 0.0717, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.7018364721020003, |
|
"grad_norm": 0.1655229777097702, |
|
"learning_rate": 4.182700591875267e-05, |
|
"loss": 0.071, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.7018364721020003, |
|
"eval_loss": 0.0841016098856926, |
|
"eval_runtime": 237.2254, |
|
"eval_samples_per_second": 2.942, |
|
"eval_steps_per_second": 0.738, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.7041759270090069, |
|
"grad_norm": 0.18065688014030457, |
|
"learning_rate": 4.1749905368497345e-05, |
|
"loss": 0.0733, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.7065153819160136, |
|
"grad_norm": 0.16063852608203888, |
|
"learning_rate": 4.1672514651460124e-05, |
|
"loss": 0.0646, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.7088548368230202, |
|
"grad_norm": 0.18853044509887695, |
|
"learning_rate": 4.159483510831814e-05, |
|
"loss": 0.0623, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.7111942917300269, |
|
"grad_norm": 0.15105347335338593, |
|
"learning_rate": 4.151686808475204e-05, |
|
"loss": 0.0638, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.7135337466370336, |
|
"grad_norm": 0.17402097582817078, |
|
"learning_rate": 4.143861493142258e-05, |
|
"loss": 0.0714, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.7158732015440402, |
|
"grad_norm": 0.17532867193222046, |
|
"learning_rate": 4.136007700394733e-05, |
|
"loss": 0.0732, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.7182126564510469, |
|
"grad_norm": 0.16897249221801758, |
|
"learning_rate": 4.128125566287711e-05, |
|
"loss": 0.0705, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.7205521113580535, |
|
"grad_norm": 0.19206973910331726, |
|
"learning_rate": 4.120215227367247e-05, |
|
"loss": 0.0722, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.7228915662650602, |
|
"grad_norm": 0.1461183875799179, |
|
"learning_rate": 4.112276820668001e-05, |
|
"loss": 0.0654, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.725231021172067, |
|
"grad_norm": 0.14818954467773438, |
|
"learning_rate": 4.104310483710864e-05, |
|
"loss": 0.0664, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.7275704760790735, |
|
"grad_norm": 0.1520642787218094, |
|
"learning_rate": 4.096316354500578e-05, |
|
"loss": 0.0647, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.7299099309860803, |
|
"grad_norm": 0.18049833178520203, |
|
"learning_rate": 4.0882945715233426e-05, |
|
"loss": 0.0647, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.7322493858930869, |
|
"grad_norm": 0.14968477189540863, |
|
"learning_rate": 4.080245273744419e-05, |
|
"loss": 0.0681, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.7345888408000936, |
|
"grad_norm": 0.17562703788280487, |
|
"learning_rate": 4.072168600605718e-05, |
|
"loss": 0.0682, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.7369282957071003, |
|
"grad_norm": 0.14080215990543365, |
|
"learning_rate": 4.064064692023389e-05, |
|
"loss": 0.0626, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.7392677506141069, |
|
"grad_norm": 0.16380788385868073, |
|
"learning_rate": 4.055933688385394e-05, |
|
"loss": 0.1316, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.7416072055211136, |
|
"grad_norm": 0.15547692775726318, |
|
"learning_rate": 4.047775730549075e-05, |
|
"loss": 0.08, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.7439466604281203, |
|
"grad_norm": 0.15663617849349976, |
|
"learning_rate": 4.039590959838715e-05, |
|
"loss": 0.0722, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.7462861153351269, |
|
"grad_norm": 0.16741621494293213, |
|
"learning_rate": 4.031379518043091e-05, |
|
"loss": 0.0708, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.7486255702421336, |
|
"grad_norm": 0.15329977869987488, |
|
"learning_rate": 4.023141547413015e-05, |
|
"loss": 0.063, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.7509650251491402, |
|
"grad_norm": 0.16843506693840027, |
|
"learning_rate": 4.0148771906588706e-05, |
|
"loss": 0.0766, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.7533044800561469, |
|
"grad_norm": 0.17443476617336273, |
|
"learning_rate": 4.0065865909481417e-05, |
|
"loss": 0.0663, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.7556439349631536, |
|
"grad_norm": 0.14947757124900818, |
|
"learning_rate": 3.9982698919029305e-05, |
|
"loss": 0.0804, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.7579833898701602, |
|
"grad_norm": 0.16239669919013977, |
|
"learning_rate": 3.9899272375974726e-05, |
|
"loss": 0.076, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.7603228447771669, |
|
"grad_norm": 0.15543022751808167, |
|
"learning_rate": 3.981558772555638e-05, |
|
"loss": 0.0697, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.7626622996841735, |
|
"grad_norm": 0.16913045942783356, |
|
"learning_rate": 3.97316464174843e-05, |
|
"loss": 0.0752, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.7650017545911802, |
|
"grad_norm": 0.16426704823970795, |
|
"learning_rate": 3.964744990591471e-05, |
|
"loss": 0.0827, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.767341209498187, |
|
"grad_norm": 0.1630501002073288, |
|
"learning_rate": 3.956299964942485e-05, |
|
"loss": 0.0716, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.7696806644051936, |
|
"grad_norm": 0.15198907256126404, |
|
"learning_rate": 3.947829711098772e-05, |
|
"loss": 0.0763, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.7720201193122003, |
|
"grad_norm": 0.17114469408988953, |
|
"learning_rate": 3.9393343757946696e-05, |
|
"loss": 0.0662, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.7743595742192069, |
|
"grad_norm": 0.14625594019889832, |
|
"learning_rate": 3.930814106199018e-05, |
|
"loss": 0.0721, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.7766990291262136, |
|
"grad_norm": 0.16365419328212738, |
|
"learning_rate": 3.922269049912602e-05, |
|
"loss": 0.0755, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.7790384840332203, |
|
"grad_norm": 0.19672849774360657, |
|
"learning_rate": 3.9136993549656006e-05, |
|
"loss": 0.0721, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.7813779389402269, |
|
"grad_norm": 0.17917795479297638, |
|
"learning_rate": 3.905105169815021e-05, |
|
"loss": 0.0638, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.7837173938472336, |
|
"grad_norm": 0.15261490643024445, |
|
"learning_rate": 3.896486643342124e-05, |
|
"loss": 0.0779, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.7860568487542403, |
|
"grad_norm": 0.1785430610179901, |
|
"learning_rate": 3.887843924849849e-05, |
|
"loss": 0.0715, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.7883963036612469, |
|
"grad_norm": 0.15738226473331451, |
|
"learning_rate": 3.879177164060225e-05, |
|
"loss": 0.0719, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.7907357585682536, |
|
"grad_norm": 0.17169606685638428, |
|
"learning_rate": 3.8704865111117746e-05, |
|
"loss": 0.0661, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.7930752134752602, |
|
"grad_norm": 0.15554682910442352, |
|
"learning_rate": 3.861772116556921e-05, |
|
"loss": 0.0715, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.7954146683822669, |
|
"grad_norm": 0.2050098329782486, |
|
"learning_rate": 3.853034131359371e-05, |
|
"loss": 0.0726, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.7977541232892736, |
|
"grad_norm": 0.14537398517131805, |
|
"learning_rate": 3.8442727068915066e-05, |
|
"loss": 0.0736, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.8000935781962802, |
|
"grad_norm": 0.16626664996147156, |
|
"learning_rate": 3.8354879949317546e-05, |
|
"loss": 0.0641, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.802433033103287, |
|
"grad_norm": 0.17925380170345306, |
|
"learning_rate": 3.8266801476619694e-05, |
|
"loss": 0.07, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.8047724880102936, |
|
"grad_norm": 0.1543145477771759, |
|
"learning_rate": 3.817849317664784e-05, |
|
"loss": 0.069, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.8071119429173003, |
|
"grad_norm": 0.15055689215660095, |
|
"learning_rate": 3.808995657920975e-05, |
|
"loss": 0.0686, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.809451397824307, |
|
"grad_norm": 0.13166533410549164, |
|
"learning_rate": 3.800119321806809e-05, |
|
"loss": 0.0643, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.8117908527313136, |
|
"grad_norm": 0.20518459379673004, |
|
"learning_rate": 3.791220463091387e-05, |
|
"loss": 0.0846, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.8141303076383203, |
|
"grad_norm": 0.13176603615283966, |
|
"learning_rate": 3.7822992359339794e-05, |
|
"loss": 0.0697, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.816469762545327, |
|
"grad_norm": 0.16138117015361786, |
|
"learning_rate": 3.7733557948813575e-05, |
|
"loss": 0.0679, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.8188092174523336, |
|
"grad_norm": 0.20597168803215027, |
|
"learning_rate": 3.764390294865112e-05, |
|
"loss": 0.0779, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.8211486723593403, |
|
"grad_norm": 0.18723969161510468, |
|
"learning_rate": 3.755402891198974e-05, |
|
"loss": 0.141, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.8234881272663469, |
|
"grad_norm": 0.16847847402095795, |
|
"learning_rate": 3.746393739576121e-05, |
|
"loss": 0.0675, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.8258275821733536, |
|
"grad_norm": 0.1512673944234848, |
|
"learning_rate": 3.737362996066483e-05, |
|
"loss": 0.0708, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.8281670370803603, |
|
"grad_norm": 0.17795756459236145, |
|
"learning_rate": 3.728310817114034e-05, |
|
"loss": 0.0675, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.8305064919873669, |
|
"grad_norm": 0.18552307784557343, |
|
"learning_rate": 3.719237359534087e-05, |
|
"loss": 0.0758, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.8328459468943736, |
|
"grad_norm": 0.16442035138607025, |
|
"learning_rate": 3.710142780510573e-05, |
|
"loss": 0.0728, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.8351854018013802, |
|
"grad_norm": 0.16192203760147095, |
|
"learning_rate": 3.7010272375933216e-05, |
|
"loss": 0.0651, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.8375248567083869, |
|
"grad_norm": 0.15759891271591187, |
|
"learning_rate": 3.691890888695329e-05, |
|
"loss": 0.0659, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.8398643116153937, |
|
"grad_norm": 0.16758672893047333, |
|
"learning_rate": 3.6827338920900254e-05, |
|
"loss": 0.065, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.8422037665224003, |
|
"grad_norm": 0.1629868447780609, |
|
"learning_rate": 3.6735564064085296e-05, |
|
"loss": 0.0748, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.844543221429407, |
|
"grad_norm": 0.1754574477672577, |
|
"learning_rate": 3.664358590636903e-05, |
|
"loss": 0.0857, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.8468826763364136, |
|
"grad_norm": 0.16642498970031738, |
|
"learning_rate": 3.655140604113395e-05, |
|
"loss": 0.0721, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.8492221312434203, |
|
"grad_norm": 0.15343014895915985, |
|
"learning_rate": 3.645902606525683e-05, |
|
"loss": 0.0653, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.851561586150427, |
|
"grad_norm": 0.15997187793254852, |
|
"learning_rate": 3.636644757908107e-05, |
|
"loss": 0.0752, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.8539010410574336, |
|
"grad_norm": 0.1555909514427185, |
|
"learning_rate": 3.627367218638893e-05, |
|
"loss": 0.0692, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.8562404959644403, |
|
"grad_norm": 0.1649237424135208, |
|
"learning_rate": 3.618070149437381e-05, |
|
"loss": 0.0706, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.858579950871447, |
|
"grad_norm": 0.14074915647506714, |
|
"learning_rate": 3.6087537113612364e-05, |
|
"loss": 0.0612, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.8609194057784536, |
|
"grad_norm": 0.15589754283428192, |
|
"learning_rate": 3.59941806580366e-05, |
|
"loss": 0.0724, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.8632588606854603, |
|
"grad_norm": 0.1727776825428009, |
|
"learning_rate": 3.590063374490595e-05, |
|
"loss": 0.0851, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.8655983155924669, |
|
"grad_norm": 0.14394600689411163, |
|
"learning_rate": 3.580689799477921e-05, |
|
"loss": 0.0669, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.8679377704994736, |
|
"grad_norm": 0.15605291724205017, |
|
"learning_rate": 3.5712975031486525e-05, |
|
"loss": 0.0716, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.8702772254064803, |
|
"grad_norm": 0.19002071022987366, |
|
"learning_rate": 3.56188664821012e-05, |
|
"loss": 0.0748, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.8726166803134869, |
|
"grad_norm": 0.16365236043930054, |
|
"learning_rate": 3.5524573976911546e-05, |
|
"loss": 0.079, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 0.8749561352204936, |
|
"grad_norm": 0.19456696510314941, |
|
"learning_rate": 3.543009914939261e-05, |
|
"loss": 0.0747, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.8772955901275002, |
|
"grad_norm": 0.15589672327041626, |
|
"learning_rate": 3.533544363617796e-05, |
|
"loss": 0.0657, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.879635045034507, |
|
"grad_norm": 0.14598876237869263, |
|
"learning_rate": 3.5240609077031196e-05, |
|
"loss": 0.0642, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.8819744999415137, |
|
"grad_norm": 0.15636853873729706, |
|
"learning_rate": 3.514559711481765e-05, |
|
"loss": 0.0608, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 0.8843139548485203, |
|
"grad_norm": 0.1746116727590561, |
|
"learning_rate": 3.505040939547591e-05, |
|
"loss": 0.077, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.886653409755527, |
|
"grad_norm": 0.15822593867778778, |
|
"learning_rate": 3.495504756798926e-05, |
|
"loss": 0.0788, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 0.8889928646625336, |
|
"grad_norm": 0.16369085013866425, |
|
"learning_rate": 3.485951328435716e-05, |
|
"loss": 0.0764, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.8913323195695403, |
|
"grad_norm": 0.1527598798274994, |
|
"learning_rate": 3.476380819956662e-05, |
|
"loss": 0.0722, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 0.893671774476547, |
|
"grad_norm": 0.13147583603858948, |
|
"learning_rate": 3.466793397156351e-05, |
|
"loss": 0.0694, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.8960112293835536, |
|
"grad_norm": 0.15432552993297577, |
|
"learning_rate": 3.457189226122384e-05, |
|
"loss": 0.0652, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 0.8983506842905603, |
|
"grad_norm": 0.15792711079120636, |
|
"learning_rate": 3.4475684732325024e-05, |
|
"loss": 0.0763, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.900690139197567, |
|
"grad_norm": 0.14269225299358368, |
|
"learning_rate": 3.437931305151701e-05, |
|
"loss": 0.0739, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.9030295941045736, |
|
"grad_norm": 0.18912151455879211, |
|
"learning_rate": 3.428277888829344e-05, |
|
"loss": 0.0839, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.9053690490115803, |
|
"grad_norm": 0.16357168555259705, |
|
"learning_rate": 3.418608391496273e-05, |
|
"loss": 0.0695, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 0.9077085039185869, |
|
"grad_norm": 0.1561209261417389, |
|
"learning_rate": 3.408922980661905e-05, |
|
"loss": 0.0681, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.9100479588255936, |
|
"grad_norm": 0.17604918777942657, |
|
"learning_rate": 3.39922182411134e-05, |
|
"loss": 0.0694, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 0.9123874137326003, |
|
"grad_norm": 0.16690443456172943, |
|
"learning_rate": 3.3895050899024416e-05, |
|
"loss": 0.1077, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.914726868639607, |
|
"grad_norm": 0.15058398246765137, |
|
"learning_rate": 3.379772946362939e-05, |
|
"loss": 0.0604, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 0.9170663235466137, |
|
"grad_norm": 0.12406554818153381, |
|
"learning_rate": 3.3700255620875025e-05, |
|
"loss": 0.061, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.9194057784536203, |
|
"grad_norm": 0.15860305726528168, |
|
"learning_rate": 3.3602631059348245e-05, |
|
"loss": 0.0714, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 0.921745233360627, |
|
"grad_norm": 0.15389299392700195, |
|
"learning_rate": 3.350485747024695e-05, |
|
"loss": 0.0724, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.9240846882676337, |
|
"grad_norm": 0.16053031384944916, |
|
"learning_rate": 3.340693654735073e-05, |
|
"loss": 0.064, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.9264241431746403, |
|
"grad_norm": 0.14700767397880554, |
|
"learning_rate": 3.330886998699149e-05, |
|
"loss": 0.0639, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.928763598081647, |
|
"grad_norm": 0.15613354742527008, |
|
"learning_rate": 3.321065948802411e-05, |
|
"loss": 0.0667, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 0.9311030529886536, |
|
"grad_norm": 0.15523682534694672, |
|
"learning_rate": 3.311230675179697e-05, |
|
"loss": 0.0634, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.9334425078956603, |
|
"grad_norm": 0.14495091140270233, |
|
"learning_rate": 3.301381348212249e-05, |
|
"loss": 0.0632, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 0.935781962802667, |
|
"grad_norm": 0.17717325687408447, |
|
"learning_rate": 3.291518138524764e-05, |
|
"loss": 0.0678, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.935781962802667, |
|
"eval_loss": 0.08131425082683563, |
|
"eval_runtime": 232.0604, |
|
"eval_samples_per_second": 3.008, |
|
"eval_steps_per_second": 0.754, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.9381214177096736, |
|
"grad_norm": 0.15504471957683563, |
|
"learning_rate": 3.281641216982435e-05, |
|
"loss": 0.1141, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 0.9404608726166803, |
|
"grad_norm": 0.15814577043056488, |
|
"learning_rate": 3.271750754687991e-05, |
|
"loss": 0.0658, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.942800327523687, |
|
"grad_norm": 0.18318745493888855, |
|
"learning_rate": 3.261846922978736e-05, |
|
"loss": 0.0727, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 0.9451397824306936, |
|
"grad_norm": 0.16373957693576813, |
|
"learning_rate": 3.2519298934235796e-05, |
|
"loss": 0.0683, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.9474792373377003, |
|
"grad_norm": 0.1619662046432495, |
|
"learning_rate": 3.24199983782006e-05, |
|
"loss": 0.0663, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.9498186922447069, |
|
"grad_norm": 0.15787597000598907, |
|
"learning_rate": 3.232056928191376e-05, |
|
"loss": 0.0743, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.9521581471517137, |
|
"grad_norm": 0.18520085513591766, |
|
"learning_rate": 3.2221013367834016e-05, |
|
"loss": 0.0994, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 0.9544976020587204, |
|
"grad_norm": 0.1397067755460739, |
|
"learning_rate": 3.212133236061704e-05, |
|
"loss": 0.0835, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.956837056965727, |
|
"grad_norm": 0.1584145724773407, |
|
"learning_rate": 3.202152798708554e-05, |
|
"loss": 0.0843, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 0.9591765118727337, |
|
"grad_norm": 0.14914660155773163, |
|
"learning_rate": 3.1921601976199365e-05, |
|
"loss": 0.0744, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.9615159667797403, |
|
"grad_norm": 0.14191728830337524, |
|
"learning_rate": 3.182155605902556e-05, |
|
"loss": 0.0654, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 0.963855421686747, |
|
"grad_norm": 0.15701289474964142, |
|
"learning_rate": 3.1721391968708335e-05, |
|
"loss": 0.0665, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.9661948765937537, |
|
"grad_norm": 0.14742569625377655, |
|
"learning_rate": 3.162111144043911e-05, |
|
"loss": 0.072, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 0.9685343315007603, |
|
"grad_norm": 0.15007531642913818, |
|
"learning_rate": 3.1520716211426385e-05, |
|
"loss": 0.0754, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.970873786407767, |
|
"grad_norm": 0.1743524670600891, |
|
"learning_rate": 3.142020802086569e-05, |
|
"loss": 0.0739, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.9732132413147736, |
|
"grad_norm": 0.12957176566123962, |
|
"learning_rate": 3.131958860990945e-05, |
|
"loss": 0.0617, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.9755526962217803, |
|
"grad_norm": 0.15021301805973053, |
|
"learning_rate": 3.121885972163681e-05, |
|
"loss": 0.0785, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 0.977892151128787, |
|
"grad_norm": 0.18265359103679657, |
|
"learning_rate": 3.111802310102341e-05, |
|
"loss": 0.0675, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.9802316060357936, |
|
"grad_norm": 0.14601291716098785, |
|
"learning_rate": 3.101708049491125e-05, |
|
"loss": 0.0694, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 0.9825710609428003, |
|
"grad_norm": 0.1612151712179184, |
|
"learning_rate": 3.091603365197833e-05, |
|
"loss": 0.0743, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.984910515849807, |
|
"grad_norm": 0.14992724359035492, |
|
"learning_rate": 3.081488432270841e-05, |
|
"loss": 0.0664, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 0.9872499707568136, |
|
"grad_norm": 0.14685116708278656, |
|
"learning_rate": 3.071363425936066e-05, |
|
"loss": 0.0762, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.9895894256638204, |
|
"grad_norm": 0.14769916236400604, |
|
"learning_rate": 3.061228521593931e-05, |
|
"loss": 0.0636, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 0.991928880570827, |
|
"grad_norm": 0.14650078117847443, |
|
"learning_rate": 3.0510838948163308e-05, |
|
"loss": 0.0836, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.9942683354778337, |
|
"grad_norm": 0.1642710417509079, |
|
"learning_rate": 3.0409297213435818e-05, |
|
"loss": 0.0723, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.9966077903848404, |
|
"grad_norm": 0.13404929637908936, |
|
"learning_rate": 3.030766177081385e-05, |
|
"loss": 0.0766, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.998947245291847, |
|
"grad_norm": 0.1478378027677536, |
|
"learning_rate": 3.020593438097776e-05, |
|
"loss": 0.0619, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 1.0012867001988537, |
|
"grad_norm": 0.1335192173719406, |
|
"learning_rate": 3.010411680620075e-05, |
|
"loss": 0.0641, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.0036261551058603, |
|
"grad_norm": 0.1439552754163742, |
|
"learning_rate": 3.0002210810318338e-05, |
|
"loss": 0.0735, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 1.005965610012867, |
|
"grad_norm": 0.14612635970115662, |
|
"learning_rate": 2.9900218158697797e-05, |
|
"loss": 0.0576, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.0083050649198737, |
|
"grad_norm": 0.17644114792346954, |
|
"learning_rate": 2.9798140618207598e-05, |
|
"loss": 0.0721, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 1.0106445198268803, |
|
"grad_norm": 0.15315324068069458, |
|
"learning_rate": 2.9695979957186764e-05, |
|
"loss": 0.0642, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.012983974733887, |
|
"grad_norm": 0.15840911865234375, |
|
"learning_rate": 2.9593737945414264e-05, |
|
"loss": 0.0751, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 1.0153234296408937, |
|
"grad_norm": 0.14203301072120667, |
|
"learning_rate": 2.9491416354078343e-05, |
|
"loss": 0.0619, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 1.0176628845479003, |
|
"grad_norm": 0.17222335934638977, |
|
"learning_rate": 2.938901695574585e-05, |
|
"loss": 0.0719, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 1.020002339454907, |
|
"grad_norm": 0.16653592884540558, |
|
"learning_rate": 2.9286541524331525e-05, |
|
"loss": 0.0597, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 1.0223417943619137, |
|
"grad_norm": 0.12632368505001068, |
|
"learning_rate": 2.9183991835067237e-05, |
|
"loss": 0.0585, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 1.0246812492689203, |
|
"grad_norm": 0.15063339471817017, |
|
"learning_rate": 2.9081369664471293e-05, |
|
"loss": 0.0712, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 1.027020704175927, |
|
"grad_norm": 0.15402260422706604, |
|
"learning_rate": 2.8978676790317605e-05, |
|
"loss": 0.0631, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 1.0293601590829338, |
|
"grad_norm": 0.17674040794372559, |
|
"learning_rate": 2.8875914991604948e-05, |
|
"loss": 0.0611, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.0316996139899404, |
|
"grad_norm": 0.15764425694942474, |
|
"learning_rate": 2.8773086048526076e-05, |
|
"loss": 0.1481, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 1.034039068896947, |
|
"grad_norm": 0.17565783858299255, |
|
"learning_rate": 2.8670191742436946e-05, |
|
"loss": 0.0623, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 1.0363785238039536, |
|
"grad_norm": 0.1751687377691269, |
|
"learning_rate": 2.8567233855825813e-05, |
|
"loss": 0.0599, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 1.0387179787109604, |
|
"grad_norm": 0.15617790818214417, |
|
"learning_rate": 2.8464214172282367e-05, |
|
"loss": 0.0593, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.041057433617967, |
|
"grad_norm": 0.1573198437690735, |
|
"learning_rate": 2.8361134476466843e-05, |
|
"loss": 0.0643, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 1.0433968885249736, |
|
"grad_norm": 0.15318118035793304, |
|
"learning_rate": 2.8257996554079113e-05, |
|
"loss": 0.0582, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 1.0457363434319804, |
|
"grad_norm": 0.14526399970054626, |
|
"learning_rate": 2.8154802191827706e-05, |
|
"loss": 0.0577, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 1.048075798338987, |
|
"grad_norm": 0.1734457015991211, |
|
"learning_rate": 2.805155317739891e-05, |
|
"loss": 0.0618, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 1.0504152532459936, |
|
"grad_norm": 0.16416044533252716, |
|
"learning_rate": 2.7948251299425758e-05, |
|
"loss": 0.067, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 1.0527547081530004, |
|
"grad_norm": 0.14763504266738892, |
|
"learning_rate": 2.784489834745709e-05, |
|
"loss": 0.0626, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.055094163060007, |
|
"grad_norm": 0.15287157893180847, |
|
"learning_rate": 2.77414961119265e-05, |
|
"loss": 0.0627, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 1.0574336179670136, |
|
"grad_norm": 0.17158065736293793, |
|
"learning_rate": 2.763804638412137e-05, |
|
"loss": 0.0653, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 1.0597730728740204, |
|
"grad_norm": 0.1766432225704193, |
|
"learning_rate": 2.753455095615179e-05, |
|
"loss": 0.0667, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 1.062112527781027, |
|
"grad_norm": 0.16371361911296844, |
|
"learning_rate": 2.7431011620919554e-05, |
|
"loss": 0.0653, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 1.0644519826880336, |
|
"grad_norm": 0.26466837525367737, |
|
"learning_rate": 2.7327430172087077e-05, |
|
"loss": 0.0605, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 1.0667914375950405, |
|
"grad_norm": 0.1517406851053238, |
|
"learning_rate": 2.722380840404632e-05, |
|
"loss": 0.0595, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.069130892502047, |
|
"grad_norm": 0.16470152139663696, |
|
"learning_rate": 2.7120148111887732e-05, |
|
"loss": 0.0706, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 1.0714703474090537, |
|
"grad_norm": 0.15149088203907013, |
|
"learning_rate": 2.7016451091369116e-05, |
|
"loss": 0.0646, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 1.0738098023160603, |
|
"grad_norm": 0.14142438769340515, |
|
"learning_rate": 2.6912719138884553e-05, |
|
"loss": 0.0612, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 1.076149257223067, |
|
"grad_norm": 0.17522796988487244, |
|
"learning_rate": 2.6808954051433248e-05, |
|
"loss": 0.0657, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.0784887121300737, |
|
"grad_norm": 0.1683465838432312, |
|
"learning_rate": 2.670515762658844e-05, |
|
"loss": 0.0649, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 1.0808281670370803, |
|
"grad_norm": 0.1598363220691681, |
|
"learning_rate": 2.6601331662466227e-05, |
|
"loss": 0.0523, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 1.0831676219440871, |
|
"grad_norm": 0.14117397367954254, |
|
"learning_rate": 2.6497477957694443e-05, |
|
"loss": 0.0565, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 1.0855070768510937, |
|
"grad_norm": 0.16452661156654358, |
|
"learning_rate": 2.6393598311381473e-05, |
|
"loss": 0.0629, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 1.0878465317581003, |
|
"grad_norm": 0.14130878448486328, |
|
"learning_rate": 2.6289694523085112e-05, |
|
"loss": 0.0685, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 1.0901859866651071, |
|
"grad_norm": 0.1380264163017273, |
|
"learning_rate": 2.6185768392781366e-05, |
|
"loss": 0.056, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 1.0925254415721137, |
|
"grad_norm": 0.16482488811016083, |
|
"learning_rate": 2.6081821720833287e-05, |
|
"loss": 0.0569, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 1.0948648964791203, |
|
"grad_norm": 0.1513761430978775, |
|
"learning_rate": 2.5977856307959798e-05, |
|
"loss": 0.0622, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.097204351386127, |
|
"grad_norm": 0.15945203602313995, |
|
"learning_rate": 2.5873873955204448e-05, |
|
"loss": 0.0574, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 1.0995438062931338, |
|
"grad_norm": 0.16861629486083984, |
|
"learning_rate": 2.5769876463904265e-05, |
|
"loss": 0.0575, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.1018832612001404, |
|
"grad_norm": 0.15520039200782776, |
|
"learning_rate": 2.5665865635658527e-05, |
|
"loss": 0.063, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 1.104222716107147, |
|
"grad_norm": 0.1708049774169922, |
|
"learning_rate": 2.5561843272297536e-05, |
|
"loss": 0.0625, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 1.1065621710141538, |
|
"grad_norm": 0.1442188173532486, |
|
"learning_rate": 2.5457811175851465e-05, |
|
"loss": 0.0539, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 1.1089016259211604, |
|
"grad_norm": 0.15366148948669434, |
|
"learning_rate": 2.5353771148519057e-05, |
|
"loss": 0.0639, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 1.111241080828167, |
|
"grad_norm": 0.1677359789609909, |
|
"learning_rate": 2.524972499263646e-05, |
|
"loss": 0.0749, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 1.1135805357351738, |
|
"grad_norm": 0.16887301206588745, |
|
"learning_rate": 2.514567451064599e-05, |
|
"loss": 0.0632, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 1.1159199906421804, |
|
"grad_norm": 0.16333596408367157, |
|
"learning_rate": 2.50416215050649e-05, |
|
"loss": 0.0578, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 1.118259445549187, |
|
"grad_norm": 0.1616666167974472, |
|
"learning_rate": 2.4937567778454188e-05, |
|
"loss": 0.0603, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 1.1205989004561938, |
|
"grad_norm": 0.15231609344482422, |
|
"learning_rate": 2.4833515133387296e-05, |
|
"loss": 0.0577, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 1.1229383553632004, |
|
"grad_norm": 0.16632795333862305, |
|
"learning_rate": 2.4729465372418972e-05, |
|
"loss": 0.0715, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.125277810270207, |
|
"grad_norm": 0.16579392552375793, |
|
"learning_rate": 2.4625420298053968e-05, |
|
"loss": 0.0696, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 1.1276172651772136, |
|
"grad_norm": 0.1642322987318039, |
|
"learning_rate": 2.4521381712715884e-05, |
|
"loss": 0.0631, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 1.1299567200842204, |
|
"grad_norm": 0.1781720668077469, |
|
"learning_rate": 2.441735141871587e-05, |
|
"loss": 0.0584, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 1.132296174991227, |
|
"grad_norm": 0.16532465815544128, |
|
"learning_rate": 2.4313331218221476e-05, |
|
"loss": 0.0658, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 1.1346356298982336, |
|
"grad_norm": 0.1574028581380844, |
|
"learning_rate": 2.420932291322536e-05, |
|
"loss": 0.0636, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 1.1369750848052405, |
|
"grad_norm": 0.1727520376443863, |
|
"learning_rate": 2.410532830551415e-05, |
|
"loss": 0.0687, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 1.139314539712247, |
|
"grad_norm": 0.17281392216682434, |
|
"learning_rate": 2.4001349196637144e-05, |
|
"loss": 0.065, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 1.1416539946192537, |
|
"grad_norm": 0.16799040138721466, |
|
"learning_rate": 2.3897387387875188e-05, |
|
"loss": 0.0597, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 1.1439934495262605, |
|
"grad_norm": 0.1646750122308731, |
|
"learning_rate": 2.3793444680209397e-05, |
|
"loss": 0.0617, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 1.146332904433267, |
|
"grad_norm": 0.15122398734092712, |
|
"learning_rate": 2.3689522874290028e-05, |
|
"loss": 0.0646, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.1486723593402737, |
|
"grad_norm": 0.15441857278347015, |
|
"learning_rate": 2.358562377040519e-05, |
|
"loss": 0.0734, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 1.1510118142472803, |
|
"grad_norm": 0.1397976279258728, |
|
"learning_rate": 2.3481749168449774e-05, |
|
"loss": 0.0599, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.153351269154287, |
|
"grad_norm": 0.1505647897720337, |
|
"learning_rate": 2.3377900867894158e-05, |
|
"loss": 0.0587, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 1.1556907240612937, |
|
"grad_norm": 0.17228984832763672, |
|
"learning_rate": 2.3274080667753128e-05, |
|
"loss": 0.0713, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 1.1580301789683003, |
|
"grad_norm": 0.1795542687177658, |
|
"learning_rate": 2.3170290366554653e-05, |
|
"loss": 0.0681, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 1.1603696338753071, |
|
"grad_norm": 0.16465342044830322, |
|
"learning_rate": 2.3066531762308766e-05, |
|
"loss": 0.0598, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 1.1627090887823137, |
|
"grad_norm": 0.18022869527339935, |
|
"learning_rate": 2.2962806652476363e-05, |
|
"loss": 0.0553, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 1.1650485436893203, |
|
"grad_norm": 0.17633071541786194, |
|
"learning_rate": 2.2859116833938146e-05, |
|
"loss": 0.0694, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 1.1673879985963271, |
|
"grad_norm": 0.1658298820257187, |
|
"learning_rate": 2.2755464102963408e-05, |
|
"loss": 0.0576, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 1.1697274535033337, |
|
"grad_norm": 0.1415170431137085, |
|
"learning_rate": 2.2651850255178974e-05, |
|
"loss": 0.0697, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.1697274535033337, |
|
"eval_loss": 0.07991591095924377, |
|
"eval_runtime": 236.928, |
|
"eval_samples_per_second": 2.946, |
|
"eval_steps_per_second": 0.739, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.1720669084103403, |
|
"grad_norm": 0.17509862780570984, |
|
"learning_rate": 2.2548277085538092e-05, |
|
"loss": 0.0669, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 1.1744063633173472, |
|
"grad_norm": 0.17153185606002808, |
|
"learning_rate": 2.2444746388289283e-05, |
|
"loss": 0.0853, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 1.1767458182243538, |
|
"grad_norm": 0.1431114226579666, |
|
"learning_rate": 2.2341259956945342e-05, |
|
"loss": 0.0595, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 1.1790852731313604, |
|
"grad_norm": 0.14989079535007477, |
|
"learning_rate": 2.2237819584252188e-05, |
|
"loss": 0.0544, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 1.1814247280383672, |
|
"grad_norm": 0.13587923347949982, |
|
"learning_rate": 2.2134427062157866e-05, |
|
"loss": 0.0652, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 1.1837641829453738, |
|
"grad_norm": 0.1634293496608734, |
|
"learning_rate": 2.2031084181781476e-05, |
|
"loss": 0.0637, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 1.1861036378523804, |
|
"grad_norm": 0.16133365035057068, |
|
"learning_rate": 2.192779273338215e-05, |
|
"loss": 0.079, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 1.188443092759387, |
|
"grad_norm": 0.12266691774129868, |
|
"learning_rate": 2.182455450632803e-05, |
|
"loss": 0.0569, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 1.1907825476663938, |
|
"grad_norm": 0.15535689890384674, |
|
"learning_rate": 2.17213712890653e-05, |
|
"loss": 0.0569, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 1.1931220025734004, |
|
"grad_norm": 0.16880477964878082, |
|
"learning_rate": 2.1618244869087157e-05, |
|
"loss": 0.0651, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.195461457480407, |
|
"grad_norm": 0.13848161697387695, |
|
"learning_rate": 2.151517703290289e-05, |
|
"loss": 0.0693, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 1.1978009123874138, |
|
"grad_norm": 0.16127823293209076, |
|
"learning_rate": 2.1412169566006897e-05, |
|
"loss": 0.0598, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 1.2001403672944204, |
|
"grad_norm": 0.11767289787530899, |
|
"learning_rate": 2.1309224252847793e-05, |
|
"loss": 0.0593, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 1.202479822201427, |
|
"grad_norm": 0.17509087920188904, |
|
"learning_rate": 2.1206342876797436e-05, |
|
"loss": 0.0604, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 1.2048192771084336, |
|
"grad_norm": 0.16853219270706177, |
|
"learning_rate": 2.110352722012011e-05, |
|
"loss": 0.063, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 1.2071587320154404, |
|
"grad_norm": 0.13158155977725983, |
|
"learning_rate": 2.1000779063941585e-05, |
|
"loss": 0.0568, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 1.209498186922447, |
|
"grad_norm": 0.15704578161239624, |
|
"learning_rate": 2.0898100188218313e-05, |
|
"loss": 0.0634, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 1.2118376418294536, |
|
"grad_norm": 0.157634437084198, |
|
"learning_rate": 2.0795492371706527e-05, |
|
"loss": 0.0597, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 1.2141770967364605, |
|
"grad_norm": 0.1686418503522873, |
|
"learning_rate": 2.0692957391931485e-05, |
|
"loss": 0.0654, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 1.216516551643467, |
|
"grad_norm": 0.15651817619800568, |
|
"learning_rate": 2.0590497025156676e-05, |
|
"loss": 0.0591, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.2188560065504737, |
|
"grad_norm": 0.16979578137397766, |
|
"learning_rate": 2.0488113046353004e-05, |
|
"loss": 0.0619, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 1.2211954614574805, |
|
"grad_norm": 0.16162872314453125, |
|
"learning_rate": 2.0385807229168104e-05, |
|
"loss": 0.0653, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 1.223534916364487, |
|
"grad_norm": 0.17247694730758667, |
|
"learning_rate": 2.028358134589553e-05, |
|
"loss": 0.0647, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 1.2258743712714937, |
|
"grad_norm": 0.16605614125728607, |
|
"learning_rate": 2.0181437167444144e-05, |
|
"loss": 0.0632, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 1.2282138261785005, |
|
"grad_norm": 0.14373953640460968, |
|
"learning_rate": 2.0079376463307368e-05, |
|
"loss": 0.0619, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 1.230553281085507, |
|
"grad_norm": 0.1507708579301834, |
|
"learning_rate": 1.9977401001532576e-05, |
|
"loss": 0.0588, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 1.2328927359925137, |
|
"grad_norm": 0.16803239285945892, |
|
"learning_rate": 1.9875512548690433e-05, |
|
"loss": 0.0591, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 1.2352321908995205, |
|
"grad_norm": 0.16970671713352203, |
|
"learning_rate": 1.9773712869844323e-05, |
|
"loss": 0.0708, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 1.2375716458065271, |
|
"grad_norm": 0.14553260803222656, |
|
"learning_rate": 1.9672003728519728e-05, |
|
"loss": 0.0609, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 1.2399111007135337, |
|
"grad_norm": 0.1582130342721939, |
|
"learning_rate": 1.957038688667374e-05, |
|
"loss": 0.0642, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 1.2422505556205403, |
|
"grad_norm": 0.1633974015712738, |
|
"learning_rate": 1.9468864104664464e-05, |
|
"loss": 0.0638, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 1.2445900105275471, |
|
"grad_norm": 0.15454533696174622, |
|
"learning_rate": 1.9367437141220594e-05, |
|
"loss": 0.0643, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 1.2469294654345537, |
|
"grad_norm": 0.14684420824050903, |
|
"learning_rate": 1.9266107753410884e-05, |
|
"loss": 0.0611, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 1.2492689203415603, |
|
"grad_norm": 0.13646100461483002, |
|
"learning_rate": 1.916487769661377e-05, |
|
"loss": 0.055, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 1.2516083752485672, |
|
"grad_norm": 0.15355093777179718, |
|
"learning_rate": 1.9063748724486887e-05, |
|
"loss": 0.0518, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 1.2539478301555738, |
|
"grad_norm": 0.16275903582572937, |
|
"learning_rate": 1.896272258893677e-05, |
|
"loss": 0.0618, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 1.2562872850625804, |
|
"grad_norm": 0.14904741942882538, |
|
"learning_rate": 1.8861801040088445e-05, |
|
"loss": 0.0553, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 1.258626739969587, |
|
"grad_norm": 0.16873393952846527, |
|
"learning_rate": 1.8760985826255125e-05, |
|
"loss": 0.063, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 1.2609661948765938, |
|
"grad_norm": 0.1706465184688568, |
|
"learning_rate": 1.8660278693907962e-05, |
|
"loss": 0.0645, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 1.2633056497836004, |
|
"grad_norm": 0.14628972113132477, |
|
"learning_rate": 1.8559681387645715e-05, |
|
"loss": 0.0579, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.265645104690607, |
|
"grad_norm": 0.1530769318342209, |
|
"learning_rate": 1.8459195650164612e-05, |
|
"loss": 0.0629, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 1.2679845595976138, |
|
"grad_norm": 0.16960960626602173, |
|
"learning_rate": 1.8358823222228097e-05, |
|
"loss": 0.0627, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 1.2703240145046204, |
|
"grad_norm": 0.1773470938205719, |
|
"learning_rate": 1.8258565842636703e-05, |
|
"loss": 0.0582, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 1.272663469411627, |
|
"grad_norm": 0.162547767162323, |
|
"learning_rate": 1.815842524819793e-05, |
|
"loss": 0.0606, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 1.2750029243186338, |
|
"grad_norm": 0.15616440773010254, |
|
"learning_rate": 1.8058403173696152e-05, |
|
"loss": 0.0967, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 1.2773423792256404, |
|
"grad_norm": 0.1639740914106369, |
|
"learning_rate": 1.7958501351862545e-05, |
|
"loss": 0.0585, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 1.279681834132647, |
|
"grad_norm": 0.16226714849472046, |
|
"learning_rate": 1.7858721513345117e-05, |
|
"loss": 0.0664, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 1.2820212890396538, |
|
"grad_norm": 0.1616683155298233, |
|
"learning_rate": 1.7759065386678674e-05, |
|
"loss": 0.0548, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 1.2843607439466604, |
|
"grad_norm": 0.15211808681488037, |
|
"learning_rate": 1.765953469825494e-05, |
|
"loss": 0.0577, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 1.286700198853667, |
|
"grad_norm": 0.15566258132457733, |
|
"learning_rate": 1.7560131172292556e-05, |
|
"loss": 0.0632, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.2890396537606739, |
|
"grad_norm": 0.14558292925357819, |
|
"learning_rate": 1.7460856530807315e-05, |
|
"loss": 0.0592, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 1.2913791086676805, |
|
"grad_norm": 0.16073749959468842, |
|
"learning_rate": 1.7361712493582242e-05, |
|
"loss": 0.057, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 1.293718563574687, |
|
"grad_norm": 0.14757321774959564, |
|
"learning_rate": 1.726270077813788e-05, |
|
"loss": 0.0603, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 1.2960580184816939, |
|
"grad_norm": 0.17697452008724213, |
|
"learning_rate": 1.7163823099702466e-05, |
|
"loss": 0.0625, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 1.2983974733887005, |
|
"grad_norm": 0.17188061773777008, |
|
"learning_rate": 1.7065081171182264e-05, |
|
"loss": 0.0585, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 1.300736928295707, |
|
"grad_norm": 0.1670866757631302, |
|
"learning_rate": 1.696647670313186e-05, |
|
"loss": 0.0569, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 1.303076383202714, |
|
"grad_norm": 0.15840460360050201, |
|
"learning_rate": 1.6868011403724582e-05, |
|
"loss": 0.0616, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 1.3054158381097205, |
|
"grad_norm": 0.16217677295207977, |
|
"learning_rate": 1.676968697872282e-05, |
|
"loss": 0.063, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 1.307755293016727, |
|
"grad_norm": 0.15185709297657013, |
|
"learning_rate": 1.667150513144856e-05, |
|
"loss": 0.0517, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 1.3100947479237337, |
|
"grad_norm": 0.1776837259531021, |
|
"learning_rate": 1.657346756275386e-05, |
|
"loss": 0.0614, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.3124342028307403, |
|
"grad_norm": 0.17348352074623108, |
|
"learning_rate": 1.6475575970991313e-05, |
|
"loss": 0.0593, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 1.3147736577377471, |
|
"grad_norm": 0.1484275907278061, |
|
"learning_rate": 1.6377832051984755e-05, |
|
"loss": 0.0558, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 1.3171131126447537, |
|
"grad_norm": 0.15603071451187134, |
|
"learning_rate": 1.628023749899975e-05, |
|
"loss": 0.0544, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 1.3194525675517603, |
|
"grad_norm": 0.16645818948745728, |
|
"learning_rate": 1.618279400271436e-05, |
|
"loss": 0.0635, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 1.3217920224587671, |
|
"grad_norm": 0.1469196379184723, |
|
"learning_rate": 1.608550325118981e-05, |
|
"loss": 0.0559, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 1.3241314773657737, |
|
"grad_norm": 0.19546544551849365, |
|
"learning_rate": 1.5988366929841254e-05, |
|
"loss": 0.07, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 1.3264709322727803, |
|
"grad_norm": 0.1820397824048996, |
|
"learning_rate": 1.589138672140856e-05, |
|
"loss": 0.0718, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 1.3288103871797872, |
|
"grad_norm": 0.14199690520763397, |
|
"learning_rate": 1.57945643059272e-05, |
|
"loss": 0.0566, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 1.3311498420867938, |
|
"grad_norm": 0.1681860238313675, |
|
"learning_rate": 1.56979013606991e-05, |
|
"loss": 0.0628, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 1.3334892969938004, |
|
"grad_norm": 0.16973961889743805, |
|
"learning_rate": 1.560139956026363e-05, |
|
"loss": 0.058, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.3358287519008072, |
|
"grad_norm": 0.15116538107395172, |
|
"learning_rate": 1.5505060576368545e-05, |
|
"loss": 0.059, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 1.3381682068078138, |
|
"grad_norm": 0.1909974217414856, |
|
"learning_rate": 1.5408886077941074e-05, |
|
"loss": 0.066, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 1.3405076617148204, |
|
"grad_norm": 0.17362594604492188, |
|
"learning_rate": 1.531287773105895e-05, |
|
"loss": 0.0602, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 1.3428471166218272, |
|
"grad_norm": 0.16387321054935455, |
|
"learning_rate": 1.5217037198921625e-05, |
|
"loss": 0.0633, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 1.3451865715288338, |
|
"grad_norm": 0.15900340676307678, |
|
"learning_rate": 1.512136614182137e-05, |
|
"loss": 0.0694, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 1.3475260264358404, |
|
"grad_norm": 0.15678687393665314, |
|
"learning_rate": 1.5025866217114592e-05, |
|
"loss": 0.0531, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 1.3498654813428472, |
|
"grad_norm": 0.17534701526165009, |
|
"learning_rate": 1.4930539079193078e-05, |
|
"loss": 0.0648, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 1.3522049362498538, |
|
"grad_norm": 0.15762171149253845, |
|
"learning_rate": 1.483538637945533e-05, |
|
"loss": 0.0684, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 1.3545443911568604, |
|
"grad_norm": 0.13863767683506012, |
|
"learning_rate": 1.474040976627799e-05, |
|
"loss": 0.055, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 1.3568838460638672, |
|
"grad_norm": 0.1766417771577835, |
|
"learning_rate": 1.4645610884987265e-05, |
|
"loss": 0.1258, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.3592233009708738, |
|
"grad_norm": 0.17303304374217987, |
|
"learning_rate": 1.4550991377830426e-05, |
|
"loss": 0.0607, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 1.3615627558778804, |
|
"grad_norm": 0.17528830468654633, |
|
"learning_rate": 1.4456552883947333e-05, |
|
"loss": 0.0618, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 1.363902210784887, |
|
"grad_norm": 0.16436804831027985, |
|
"learning_rate": 1.4362297039342098e-05, |
|
"loss": 0.0648, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 1.3662416656918939, |
|
"grad_norm": 0.1755647212266922, |
|
"learning_rate": 1.4268225476854704e-05, |
|
"loss": 0.0681, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 1.3685811205989005, |
|
"grad_norm": 0.15491195023059845, |
|
"learning_rate": 1.4174339826132727e-05, |
|
"loss": 0.0587, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 1.370920575505907, |
|
"grad_norm": 0.1469022035598755, |
|
"learning_rate": 1.4080641713603072e-05, |
|
"loss": 0.0669, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 1.3732600304129137, |
|
"grad_norm": 0.18021321296691895, |
|
"learning_rate": 1.3987132762443873e-05, |
|
"loss": 0.0601, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 1.3755994853199205, |
|
"grad_norm": 0.1528056561946869, |
|
"learning_rate": 1.3893814592556287e-05, |
|
"loss": 0.0599, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 1.377938940226927, |
|
"grad_norm": 0.15603193640708923, |
|
"learning_rate": 1.3800688820536517e-05, |
|
"loss": 0.0539, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 1.3802783951339337, |
|
"grad_norm": 0.15558800101280212, |
|
"learning_rate": 1.3707757059647725e-05, |
|
"loss": 0.0671, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.3826178500409405, |
|
"grad_norm": 0.156876340508461, |
|
"learning_rate": 1.3615020919792151e-05, |
|
"loss": 0.0586, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 1.384957304947947, |
|
"grad_norm": 0.1524912714958191, |
|
"learning_rate": 1.3522482007483172e-05, |
|
"loss": 0.0662, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 1.3872967598549537, |
|
"grad_norm": 0.15926173329353333, |
|
"learning_rate": 1.3430141925817532e-05, |
|
"loss": 0.0637, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 1.3896362147619605, |
|
"grad_norm": 0.16642306745052338, |
|
"learning_rate": 1.333800227444749e-05, |
|
"loss": 0.067, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 1.3919756696689671, |
|
"grad_norm": 0.13672704994678497, |
|
"learning_rate": 1.324606464955318e-05, |
|
"loss": 0.0618, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 1.3943151245759737, |
|
"grad_norm": 0.16178594529628754, |
|
"learning_rate": 1.3154330643814938e-05, |
|
"loss": 0.0564, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 1.3966545794829806, |
|
"grad_norm": 0.16218972206115723, |
|
"learning_rate": 1.30628018463857e-05, |
|
"loss": 0.0592, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 1.3989940343899872, |
|
"grad_norm": 0.16459548473358154, |
|
"learning_rate": 1.2971479842863465e-05, |
|
"loss": 0.0533, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 1.4013334892969938, |
|
"grad_norm": 0.14604820311069489, |
|
"learning_rate": 1.2880366215263845e-05, |
|
"loss": 0.0629, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 1.4036729442040006, |
|
"grad_norm": 0.16797035932540894, |
|
"learning_rate": 1.2789462541992676e-05, |
|
"loss": 0.0553, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.4036729442040006, |
|
"eval_loss": 0.07826597988605499, |
|
"eval_runtime": 234.1597, |
|
"eval_samples_per_second": 2.981, |
|
"eval_steps_per_second": 0.747, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.4060123991110072, |
|
"grad_norm": 0.1841832548379898, |
|
"learning_rate": 1.2698770397818616e-05, |
|
"loss": 0.13, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 1.4083518540180138, |
|
"grad_norm": 0.1567966490983963, |
|
"learning_rate": 1.2608291353845941e-05, |
|
"loss": 0.063, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 1.4106913089250206, |
|
"grad_norm": 0.14550812542438507, |
|
"learning_rate": 1.2518026977487252e-05, |
|
"loss": 0.0598, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 1.4130307638320272, |
|
"grad_norm": 0.15681859850883484, |
|
"learning_rate": 1.242797883243638e-05, |
|
"loss": 0.063, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 1.4153702187390338, |
|
"grad_norm": 0.16205298900604248, |
|
"learning_rate": 1.233814847864127e-05, |
|
"loss": 0.0691, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 1.4177096736460404, |
|
"grad_norm": 0.16490066051483154, |
|
"learning_rate": 1.224853747227698e-05, |
|
"loss": 0.0628, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 1.4200491285530472, |
|
"grad_norm": 0.19161145389080048, |
|
"learning_rate": 1.2159147365718664e-05, |
|
"loss": 0.0646, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 1.4223885834600538, |
|
"grad_norm": 0.16144663095474243, |
|
"learning_rate": 1.2069979707514766e-05, |
|
"loss": 0.0542, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 1.4247280383670604, |
|
"grad_norm": 0.15800189971923828, |
|
"learning_rate": 1.1981036042360109e-05, |
|
"loss": 0.059, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 1.427067493274067, |
|
"grad_norm": 0.15965986251831055, |
|
"learning_rate": 1.1892317911069212e-05, |
|
"loss": 0.0575, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.4294069481810738, |
|
"grad_norm": 0.15217727422714233, |
|
"learning_rate": 1.1803826850549526e-05, |
|
"loss": 0.0604, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 1.4317464030880804, |
|
"grad_norm": 0.15865880250930786, |
|
"learning_rate": 1.171556439377488e-05, |
|
"loss": 0.0659, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 1.434085857995087, |
|
"grad_norm": 0.18922170996665955, |
|
"learning_rate": 1.1627532069758851e-05, |
|
"loss": 0.0605, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 1.4364253129020939, |
|
"grad_norm": 0.17972765862941742, |
|
"learning_rate": 1.1539731403528353e-05, |
|
"loss": 0.0754, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 1.4387647678091005, |
|
"grad_norm": 0.18478457629680634, |
|
"learning_rate": 1.1452163916097141e-05, |
|
"loss": 0.0689, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 1.441104222716107, |
|
"grad_norm": 0.16333162784576416, |
|
"learning_rate": 1.1364831124439526e-05, |
|
"loss": 0.0669, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 1.4434436776231139, |
|
"grad_norm": 0.1580970138311386, |
|
"learning_rate": 1.1277734541464072e-05, |
|
"loss": 0.0596, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 1.4457831325301205, |
|
"grad_norm": 0.17736057937145233, |
|
"learning_rate": 1.1190875675987356e-05, |
|
"loss": 0.057, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 1.448122587437127, |
|
"grad_norm": 0.17096631228923798, |
|
"learning_rate": 1.1104256032707894e-05, |
|
"loss": 0.0671, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 1.450462042344134, |
|
"grad_norm": 0.17101797461509705, |
|
"learning_rate": 1.1017877112180009e-05, |
|
"loss": 0.0583, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.4528014972511405, |
|
"grad_norm": 0.1477900892496109, |
|
"learning_rate": 1.0931740410787895e-05, |
|
"loss": 0.0617, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 1.455140952158147, |
|
"grad_norm": 0.16209881007671356, |
|
"learning_rate": 1.0845847420719632e-05, |
|
"loss": 0.0619, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 1.457480407065154, |
|
"grad_norm": 0.16828244924545288, |
|
"learning_rate": 1.076019962994141e-05, |
|
"loss": 0.0572, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 1.4598198619721605, |
|
"grad_norm": 0.1566639393568039, |
|
"learning_rate": 1.0674798522171668e-05, |
|
"loss": 0.0632, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 1.4621593168791671, |
|
"grad_norm": 0.1641826629638672, |
|
"learning_rate": 1.0589645576855477e-05, |
|
"loss": 0.055, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 1.464498771786174, |
|
"grad_norm": 0.1548743098974228, |
|
"learning_rate": 1.0504742269138835e-05, |
|
"loss": 0.0631, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 1.4668382266931805, |
|
"grad_norm": 0.16996270418167114, |
|
"learning_rate": 1.0420090069843167e-05, |
|
"loss": 0.0557, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 1.4691776816001871, |
|
"grad_norm": 0.18014395236968994, |
|
"learning_rate": 1.0335690445439817e-05, |
|
"loss": 0.0623, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 1.471517136507194, |
|
"grad_norm": 0.17947927117347717, |
|
"learning_rate": 1.0251544858024662e-05, |
|
"loss": 0.056, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 1.4738565914142006, |
|
"grad_norm": 0.19152309000492096, |
|
"learning_rate": 1.0167654765292742e-05, |
|
"loss": 0.0581, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.4761960463212072, |
|
"grad_norm": 0.16634751856327057, |
|
"learning_rate": 1.0084021620513079e-05, |
|
"loss": 0.0619, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 1.4785355012282138, |
|
"grad_norm": 0.19897738099098206, |
|
"learning_rate": 1.0000646872503419e-05, |
|
"loss": 0.0784, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 1.4808749561352204, |
|
"grad_norm": 0.1585209220647812, |
|
"learning_rate": 9.917531965605211e-06, |
|
"loss": 0.062, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 1.4832144110422272, |
|
"grad_norm": 0.18893523514270782, |
|
"learning_rate": 9.83467833965852e-06, |
|
"loss": 0.057, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 1.4855538659492338, |
|
"grad_norm": 0.17129787802696228, |
|
"learning_rate": 9.752087429977143e-06, |
|
"loss": 0.0674, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 1.4878933208562404, |
|
"grad_norm": 0.1496947705745697, |
|
"learning_rate": 9.669760667323685e-06, |
|
"loss": 0.0842, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 1.4902327757632472, |
|
"grad_norm": 0.17729486525058746, |
|
"learning_rate": 9.58769947788482e-06, |
|
"loss": 0.0736, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 1.4925722306702538, |
|
"grad_norm": 0.17603404819965363, |
|
"learning_rate": 9.505905283246578e-06, |
|
"loss": 0.0638, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 1.4949116855772604, |
|
"grad_norm": 0.17529185116291046, |
|
"learning_rate": 9.424379500369674e-06, |
|
"loss": 0.0672, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 1.4972511404842672, |
|
"grad_norm": 0.1493815779685974, |
|
"learning_rate": 9.343123541565035e-06, |
|
"loss": 0.058, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.4995905953912738, |
|
"grad_norm": 0.15651191771030426, |
|
"learning_rate": 9.26213881446926e-06, |
|
"loss": 0.0588, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 1.5019300502982804, |
|
"grad_norm": 0.15472684800624847, |
|
"learning_rate": 9.181426722020297e-06, |
|
"loss": 0.0626, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 1.5042695052052872, |
|
"grad_norm": 0.1653946340084076, |
|
"learning_rate": 9.100988662433082e-06, |
|
"loss": 0.0824, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 1.5066089601122938, |
|
"grad_norm": 0.1525607407093048, |
|
"learning_rate": 9.020826029175384e-06, |
|
"loss": 0.0621, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 1.5089484150193004, |
|
"grad_norm": 0.18066176772117615, |
|
"learning_rate": 8.94094021094358e-06, |
|
"loss": 0.0632, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 1.5112878699263073, |
|
"grad_norm": 0.13805319368839264, |
|
"learning_rate": 8.86133259163869e-06, |
|
"loss": 0.0638, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 1.5136273248333139, |
|
"grad_norm": 0.1525852382183075, |
|
"learning_rate": 8.782004550342323e-06, |
|
"loss": 0.0586, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 1.5159667797403205, |
|
"grad_norm": 0.16709508001804352, |
|
"learning_rate": 8.702957461292846e-06, |
|
"loss": 0.0632, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 1.5183062346473273, |
|
"grad_norm": 0.1565026491880417, |
|
"learning_rate": 8.624192693861544e-06, |
|
"loss": 0.0578, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 1.5206456895543339, |
|
"grad_norm": 0.16678424179553986, |
|
"learning_rate": 8.545711612528915e-06, |
|
"loss": 0.063, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.5229851444613405, |
|
"grad_norm": 0.19322334229946136, |
|
"learning_rate": 8.467515576861002e-06, |
|
"loss": 0.0657, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 1.5253245993683473, |
|
"grad_norm": 0.1959463208913803, |
|
"learning_rate": 8.389605941485893e-06, |
|
"loss": 0.1184, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 1.5276640542753537, |
|
"grad_norm": 0.1491149216890335, |
|
"learning_rate": 8.31198405607019e-06, |
|
"loss": 0.055, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 1.5300035091823605, |
|
"grad_norm": 0.1650906205177307, |
|
"learning_rate": 8.234651265295702e-06, |
|
"loss": 0.0669, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 1.5323429640893673, |
|
"grad_norm": 0.13482898473739624, |
|
"learning_rate": 8.15760890883607e-06, |
|
"loss": 0.0577, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 1.5346824189963737, |
|
"grad_norm": 0.15607015788555145, |
|
"learning_rate": 8.080858321333633e-06, |
|
"loss": 0.063, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 1.5370218739033805, |
|
"grad_norm": 0.15770550072193146, |
|
"learning_rate": 8.004400832376258e-06, |
|
"loss": 0.0617, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 1.5393613288103873, |
|
"grad_norm": 0.15652437508106232, |
|
"learning_rate": 7.928237766474314e-06, |
|
"loss": 0.0668, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 1.5417007837173937, |
|
"grad_norm": 0.174765482544899, |
|
"learning_rate": 7.852370443037751e-06, |
|
"loss": 0.0604, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 1.5440402386244005, |
|
"grad_norm": 0.14293302595615387, |
|
"learning_rate": 7.776800176353227e-06, |
|
"loss": 0.0852, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.5463796935314071, |
|
"grad_norm": 0.16163261234760284, |
|
"learning_rate": 7.701528275561348e-06, |
|
"loss": 0.0571, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 1.5487191484384137, |
|
"grad_norm": 0.13872607052326202, |
|
"learning_rate": 7.626556044633951e-06, |
|
"loss": 0.0606, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 1.5510586033454206, |
|
"grad_norm": 0.16931146383285522, |
|
"learning_rate": 7.551884782351587e-06, |
|
"loss": 0.0582, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 1.5533980582524272, |
|
"grad_norm": 0.16759054362773895, |
|
"learning_rate": 7.477515782280939e-06, |
|
"loss": 0.0656, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 1.5557375131594338, |
|
"grad_norm": 0.15919411182403564, |
|
"learning_rate": 7.403450332752493e-06, |
|
"loss": 0.0685, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 1.5580769680664406, |
|
"grad_norm": 0.1733466535806656, |
|
"learning_rate": 7.329689716838148e-06, |
|
"loss": 0.06, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 1.5604164229734472, |
|
"grad_norm": 0.15883946418762207, |
|
"learning_rate": 7.256235212329055e-06, |
|
"loss": 0.0537, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 1.5627558778804538, |
|
"grad_norm": 0.20026150345802307, |
|
"learning_rate": 7.183088091713405e-06, |
|
"loss": 0.0676, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 1.5650953327874606, |
|
"grad_norm": 0.14357200264930725, |
|
"learning_rate": 7.110249622154469e-06, |
|
"loss": 0.0596, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 1.5674347876944672, |
|
"grad_norm": 0.15919791162014008, |
|
"learning_rate": 7.037721065468591e-06, |
|
"loss": 0.0504, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.5697742426014738, |
|
"grad_norm": 0.16171149909496307, |
|
"learning_rate": 6.96550367810333e-06, |
|
"loss": 0.0688, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 1.5721136975084806, |
|
"grad_norm": 0.17881818115711212, |
|
"learning_rate": 6.89359871111572e-06, |
|
"loss": 0.0545, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 1.5744531524154872, |
|
"grad_norm": 0.1703099012374878, |
|
"learning_rate": 6.822007410150591e-06, |
|
"loss": 0.0624, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 1.5767926073224938, |
|
"grad_norm": 0.17320412397384644, |
|
"learning_rate": 6.750731015418959e-06, |
|
"loss": 0.0659, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 1.5791320622295006, |
|
"grad_norm": 0.15360009670257568, |
|
"learning_rate": 6.679770761676596e-06, |
|
"loss": 0.062, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 1.5814715171365072, |
|
"grad_norm": 0.15514342486858368, |
|
"learning_rate": 6.609127878202581e-06, |
|
"loss": 0.0522, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 1.5838109720435138, |
|
"grad_norm": 0.14531250298023224, |
|
"learning_rate": 6.538803588778064e-06, |
|
"loss": 0.0588, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 1.5861504269505207, |
|
"grad_norm": 0.16243524849414825, |
|
"learning_rate": 6.468799111665003e-06, |
|
"loss": 0.0641, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 1.588489881857527, |
|
"grad_norm": 0.19768266379833221, |
|
"learning_rate": 6.399115659585103e-06, |
|
"loss": 0.0668, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 1.5908293367645339, |
|
"grad_norm": 0.1622917205095291, |
|
"learning_rate": 6.329754439698804e-06, |
|
"loss": 0.0531, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.5931687916715407, |
|
"grad_norm": 0.18452420830726624, |
|
"learning_rate": 6.260716653584359e-06, |
|
"loss": 0.0992, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 1.595508246578547, |
|
"grad_norm": 0.16582529246807098, |
|
"learning_rate": 6.192003497217016e-06, |
|
"loss": 0.0574, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 1.5978477014855539, |
|
"grad_norm": 0.1616390198469162, |
|
"learning_rate": 6.123616160948295e-06, |
|
"loss": 0.0616, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 1.6001871563925605, |
|
"grad_norm": 0.16133980453014374, |
|
"learning_rate": 6.055555829485396e-06, |
|
"loss": 0.0584, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 1.602526611299567, |
|
"grad_norm": 0.14773264527320862, |
|
"learning_rate": 5.98782368187063e-06, |
|
"loss": 0.0557, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 1.604866066206574, |
|
"grad_norm": 0.1554953008890152, |
|
"learning_rate": 5.920420891461046e-06, |
|
"loss": 0.0618, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 1.6072055211135805, |
|
"grad_norm": 0.1828761100769043, |
|
"learning_rate": 5.853348625908048e-06, |
|
"loss": 0.0623, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 1.609544976020587, |
|
"grad_norm": 0.1556580513715744, |
|
"learning_rate": 5.786608047137229e-06, |
|
"loss": 0.0564, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 1.611884430927594, |
|
"grad_norm": 0.1514003425836563, |
|
"learning_rate": 5.720200311328178e-06, |
|
"loss": 0.0595, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 1.6142238858346005, |
|
"grad_norm": 0.14597107470035553, |
|
"learning_rate": 5.654126568894513e-06, |
|
"loss": 0.0595, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.6165633407416071, |
|
"grad_norm": 0.17368368804454803, |
|
"learning_rate": 5.5883879644638925e-06, |
|
"loss": 0.0606, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 1.618902795648614, |
|
"grad_norm": 0.16808243095874786, |
|
"learning_rate": 5.522985636858239e-06, |
|
"loss": 0.058, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 1.6212422505556205, |
|
"grad_norm": 0.1682329773902893, |
|
"learning_rate": 5.4579207190739775e-06, |
|
"loss": 0.0578, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 1.6235817054626271, |
|
"grad_norm": 0.1642475128173828, |
|
"learning_rate": 5.393194338262428e-06, |
|
"loss": 0.0651, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 1.625921160369634, |
|
"grad_norm": 0.18125009536743164, |
|
"learning_rate": 5.328807615710246e-06, |
|
"loss": 0.0586, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 1.6282606152766406, |
|
"grad_norm": 0.14442673325538635, |
|
"learning_rate": 5.264761666820054e-06, |
|
"loss": 0.059, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 1.6306000701836472, |
|
"grad_norm": 0.1675989329814911, |
|
"learning_rate": 5.201057601091056e-06, |
|
"loss": 0.0574, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 1.632939525090654, |
|
"grad_norm": 0.17742133140563965, |
|
"learning_rate": 5.137696522099858e-06, |
|
"loss": 0.0577, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 1.6352789799976606, |
|
"grad_norm": 0.15731069445610046, |
|
"learning_rate": 5.074679527481349e-06, |
|
"loss": 0.055, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 1.6376184349046672, |
|
"grad_norm": 0.17156293988227844, |
|
"learning_rate": 5.012007708909661e-06, |
|
"loss": 0.057, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.6376184349046672, |
|
"eval_loss": 0.07722621411085129, |
|
"eval_runtime": 235.8253, |
|
"eval_samples_per_second": 2.96, |
|
"eval_steps_per_second": 0.742, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.639957889811674, |
|
"grad_norm": 0.1923573613166809, |
|
"learning_rate": 4.949682152079291e-06, |
|
"loss": 0.0658, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 1.6422973447186804, |
|
"grad_norm": 0.15531805157661438, |
|
"learning_rate": 4.887703936686252e-06, |
|
"loss": 0.0549, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 1.6446367996256872, |
|
"grad_norm": 0.18201015889644623, |
|
"learning_rate": 4.826074136409411e-06, |
|
"loss": 0.0665, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 1.646976254532694, |
|
"grad_norm": 0.14315198361873627, |
|
"learning_rate": 4.764793818891866e-06, |
|
"loss": 0.0557, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 1.6493157094397004, |
|
"grad_norm": 0.16865558922290802, |
|
"learning_rate": 4.703864045722453e-06, |
|
"loss": 0.0609, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 1.6516551643467072, |
|
"grad_norm": 0.16090945899486542, |
|
"learning_rate": 4.643285872417346e-06, |
|
"loss": 0.0602, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 1.653994619253714, |
|
"grad_norm": 0.16474011540412903, |
|
"learning_rate": 4.583060348401808e-06, |
|
"loss": 0.0529, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 1.6563340741607204, |
|
"grad_norm": 0.18071883916854858, |
|
"learning_rate": 4.523188516991958e-06, |
|
"loss": 0.0598, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 1.6586735290677272, |
|
"grad_norm": 0.14429545402526855, |
|
"learning_rate": 4.4636714153767496e-06, |
|
"loss": 0.058, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 1.6610129839747338, |
|
"grad_norm": 0.17122824490070343, |
|
"learning_rate": 4.40451007459996e-06, |
|
"loss": 0.061, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.6633524388817404, |
|
"grad_norm": 0.17276746034622192, |
|
"learning_rate": 4.3457055195423645e-06, |
|
"loss": 0.0557, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 1.6656918937887473, |
|
"grad_norm": 0.1375202089548111, |
|
"learning_rate": 4.2872587689039484e-06, |
|
"loss": 0.055, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 1.6680313486957539, |
|
"grad_norm": 0.14362306892871857, |
|
"learning_rate": 4.229170835186292e-06, |
|
"loss": 0.0573, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 1.6703708036027605, |
|
"grad_norm": 0.16020837426185608, |
|
"learning_rate": 4.171442724675004e-06, |
|
"loss": 0.0599, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 1.6727102585097673, |
|
"grad_norm": 0.1534327119588852, |
|
"learning_rate": 4.114075437422313e-06, |
|
"loss": 0.0663, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 1.6750497134167739, |
|
"grad_norm": 0.15872815251350403, |
|
"learning_rate": 4.057069967229718e-06, |
|
"loss": 0.0759, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 1.6773891683237805, |
|
"grad_norm": 0.17739912867546082, |
|
"learning_rate": 4.00042730163078e-06, |
|
"loss": 0.0576, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 1.6797286232307873, |
|
"grad_norm": 0.14820364117622375, |
|
"learning_rate": 3.944148421874042e-06, |
|
"loss": 0.0579, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 1.682068078137794, |
|
"grad_norm": 0.17816150188446045, |
|
"learning_rate": 3.888234302905977e-06, |
|
"loss": 0.0625, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 1.6844075330448005, |
|
"grad_norm": 0.1766812652349472, |
|
"learning_rate": 3.832685913354159e-06, |
|
"loss": 0.0551, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.6867469879518073, |
|
"grad_norm": 0.1790032684803009, |
|
"learning_rate": 3.777504215510427e-06, |
|
"loss": 0.0643, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 1.689086442858814, |
|
"grad_norm": 0.16145819425582886, |
|
"learning_rate": 3.7226901653142634e-06, |
|
"loss": 0.0616, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 1.6914258977658205, |
|
"grad_norm": 0.18062575161457062, |
|
"learning_rate": 3.6682447123361885e-06, |
|
"loss": 0.0688, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 1.6937653526728274, |
|
"grad_norm": 0.15464866161346436, |
|
"learning_rate": 3.61416879976135e-06, |
|
"loss": 0.062, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 1.696104807579834, |
|
"grad_norm": 0.1730775237083435, |
|
"learning_rate": 3.560463364373165e-06, |
|
"loss": 0.0552, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 1.6984442624868406, |
|
"grad_norm": 0.17476975917816162, |
|
"learning_rate": 3.5071293365370904e-06, |
|
"loss": 0.0694, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 1.7007837173938474, |
|
"grad_norm": 0.1686089038848877, |
|
"learning_rate": 3.4541676401845027e-06, |
|
"loss": 0.0562, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 1.7031231723008537, |
|
"grad_norm": 0.1799984872341156, |
|
"learning_rate": 3.4015791927967083e-06, |
|
"loss": 0.0596, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 1.7054626272078606, |
|
"grad_norm": 0.15171782672405243, |
|
"learning_rate": 3.3493649053890326e-06, |
|
"loss": 0.0797, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 1.7078020821148674, |
|
"grad_norm": 0.17303191125392914, |
|
"learning_rate": 3.2975256824950564e-06, |
|
"loss": 0.0594, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.7101415370218738, |
|
"grad_norm": 0.17267484962940216, |
|
"learning_rate": 3.2460624221509137e-06, |
|
"loss": 0.0563, |
|
"step": 3655 |
|
}, |
|
{ |
|
"epoch": 1.7124809919288806, |
|
"grad_norm": 0.16894906759262085, |
|
"learning_rate": 3.1949760158797833e-06, |
|
"loss": 0.0614, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.7148204468358872, |
|
"grad_norm": 0.15203548967838287, |
|
"learning_rate": 3.1442673486763912e-06, |
|
"loss": 0.0648, |
|
"step": 3665 |
|
}, |
|
{ |
|
"epoch": 1.7171599017428938, |
|
"grad_norm": 0.16010317206382751, |
|
"learning_rate": 3.0939372989917254e-06, |
|
"loss": 0.0593, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 1.7194993566499006, |
|
"grad_norm": 0.14804290235042572, |
|
"learning_rate": 3.04398673871778e-06, |
|
"loss": 0.0588, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 1.7218388115569072, |
|
"grad_norm": 0.17553718388080597, |
|
"learning_rate": 2.9944165331724866e-06, |
|
"loss": 0.0567, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 1.7241782664639138, |
|
"grad_norm": 0.16042271256446838, |
|
"learning_rate": 2.945227541084697e-06, |
|
"loss": 0.0548, |
|
"step": 3685 |
|
}, |
|
{ |
|
"epoch": 1.7265177213709206, |
|
"grad_norm": 0.16253893077373505, |
|
"learning_rate": 2.896420614579312e-06, |
|
"loss": 0.0515, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 1.7288571762779272, |
|
"grad_norm": 0.16040562093257904, |
|
"learning_rate": 2.847996599162539e-06, |
|
"loss": 0.0575, |
|
"step": 3695 |
|
}, |
|
{ |
|
"epoch": 1.7311966311849338, |
|
"grad_norm": 0.1720447987318039, |
|
"learning_rate": 2.799956333707207e-06, |
|
"loss": 0.064, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.7335360860919407, |
|
"grad_norm": 0.15470455586910248, |
|
"learning_rate": 2.752300650438283e-06, |
|
"loss": 0.0572, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 1.7358755409989473, |
|
"grad_norm": 0.15257127583026886, |
|
"learning_rate": 2.7050303749184046e-06, |
|
"loss": 0.0676, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 1.7382149959059539, |
|
"grad_norm": 0.17101050913333893, |
|
"learning_rate": 2.6581463260336252e-06, |
|
"loss": 0.0646, |
|
"step": 3715 |
|
}, |
|
{ |
|
"epoch": 1.7405544508129607, |
|
"grad_norm": 0.16087070107460022, |
|
"learning_rate": 2.6116493159791866e-06, |
|
"loss": 0.0499, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.7428939057199673, |
|
"grad_norm": 0.16433356702327728, |
|
"learning_rate": 2.5655401502454795e-06, |
|
"loss": 0.0668, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 1.7452333606269739, |
|
"grad_norm": 0.19104856252670288, |
|
"learning_rate": 2.5198196276040782e-06, |
|
"loss": 0.0579, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 1.7475728155339807, |
|
"grad_norm": 0.16840414702892303, |
|
"learning_rate": 2.4744885400939026e-06, |
|
"loss": 0.0642, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 1.7499122704409873, |
|
"grad_norm": 0.16252835094928741, |
|
"learning_rate": 2.42954767300749e-06, |
|
"loss": 0.0561, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 1.752251725347994, |
|
"grad_norm": 0.16803041100502014, |
|
"learning_rate": 2.384997804877412e-06, |
|
"loss": 0.0545, |
|
"step": 3745 |
|
}, |
|
{ |
|
"epoch": 1.7545911802550007, |
|
"grad_norm": 0.1815616637468338, |
|
"learning_rate": 2.340839707462755e-06, |
|
"loss": 0.0556, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.756930635162007, |
|
"grad_norm": 0.16640643775463104, |
|
"learning_rate": 2.2970741457357973e-06, |
|
"loss": 0.0587, |
|
"step": 3755 |
|
}, |
|
{ |
|
"epoch": 1.759270090069014, |
|
"grad_norm": 0.1885158121585846, |
|
"learning_rate": 2.2537018778687026e-06, |
|
"loss": 0.0729, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 1.7616095449760207, |
|
"grad_norm": 0.12472023814916611, |
|
"learning_rate": 2.210723655220434e-06, |
|
"loss": 0.0619, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 1.7639489998830271, |
|
"grad_norm": 0.17403416335582733, |
|
"learning_rate": 2.168140222323703e-06, |
|
"loss": 0.0619, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 1.766288454790034, |
|
"grad_norm": 0.16542308032512665, |
|
"learning_rate": 2.1259523168720957e-06, |
|
"loss": 0.0672, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 1.7686279096970405, |
|
"grad_norm": 0.1458907127380371, |
|
"learning_rate": 2.0841606697072745e-06, |
|
"loss": 0.056, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.7709673646040471, |
|
"grad_norm": 0.1915542483329773, |
|
"learning_rate": 2.042766004806332e-06, |
|
"loss": 0.1111, |
|
"step": 3785 |
|
}, |
|
{ |
|
"epoch": 1.773306819511054, |
|
"grad_norm": 0.1618765890598297, |
|
"learning_rate": 2.001769039269247e-06, |
|
"loss": 0.0606, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 1.7756462744180606, |
|
"grad_norm": 0.17050661146640778, |
|
"learning_rate": 1.9611704833064486e-06, |
|
"loss": 0.0702, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 1.7779857293250672, |
|
"grad_norm": 0.1698133945465088, |
|
"learning_rate": 1.9209710402265295e-06, |
|
"loss": 0.0586, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.780325184232074, |
|
"grad_norm": 0.16448405385017395, |
|
"learning_rate": 1.8811714064240481e-06, |
|
"loss": 0.059, |
|
"step": 3805 |
|
}, |
|
{ |
|
"epoch": 1.7826646391390806, |
|
"grad_norm": 0.14546535909175873, |
|
"learning_rate": 1.841772271367484e-06, |
|
"loss": 0.0543, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 1.7850040940460872, |
|
"grad_norm": 0.1465018391609192, |
|
"learning_rate": 1.8027743175872664e-06, |
|
"loss": 0.0507, |
|
"step": 3815 |
|
}, |
|
{ |
|
"epoch": 1.787343548953094, |
|
"grad_norm": 0.15931351482868195, |
|
"learning_rate": 1.7641782206639813e-06, |
|
"loss": 0.0601, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 1.7896830038601006, |
|
"grad_norm": 0.18377573788166046, |
|
"learning_rate": 1.7259846492166359e-06, |
|
"loss": 0.0604, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 1.7920224587671072, |
|
"grad_norm": 0.1727667599916458, |
|
"learning_rate": 1.6881942648911076e-06, |
|
"loss": 0.0547, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 1.794361913674114, |
|
"grad_norm": 0.1737833321094513, |
|
"learning_rate": 1.6508077223486546e-06, |
|
"loss": 0.0605, |
|
"step": 3835 |
|
}, |
|
{ |
|
"epoch": 1.7967013685811206, |
|
"grad_norm": 0.18085430562496185, |
|
"learning_rate": 1.61382566925459e-06, |
|
"loss": 0.0553, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.7990408234881272, |
|
"grad_norm": 0.1712871491909027, |
|
"learning_rate": 1.5772487462670681e-06, |
|
"loss": 0.0654, |
|
"step": 3845 |
|
}, |
|
{ |
|
"epoch": 1.801380278395134, |
|
"grad_norm": 0.16300112009048462, |
|
"learning_rate": 1.5410775870259686e-06, |
|
"loss": 0.0558, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.8037197333021406, |
|
"grad_norm": 0.17991948127746582, |
|
"learning_rate": 1.5053128181419184e-06, |
|
"loss": 0.06, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 1.8060591882091472, |
|
"grad_norm": 0.15448680520057678, |
|
"learning_rate": 1.469955059185471e-06, |
|
"loss": 0.0645, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 1.808398643116154, |
|
"grad_norm": 0.17701593041419983, |
|
"learning_rate": 1.4350049226763224e-06, |
|
"loss": 0.0631, |
|
"step": 3865 |
|
}, |
|
{ |
|
"epoch": 1.8107380980231607, |
|
"grad_norm": 0.185321643948555, |
|
"learning_rate": 1.400463014072742e-06, |
|
"loss": 0.0639, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 1.8130775529301673, |
|
"grad_norm": 0.18532027304172516, |
|
"learning_rate": 1.3663299317610595e-06, |
|
"loss": 0.0631, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 1.815417007837174, |
|
"grad_norm": 0.16126079857349396, |
|
"learning_rate": 1.3326062670453194e-06, |
|
"loss": 0.0693, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 1.8177564627441805, |
|
"grad_norm": 0.15626025199890137, |
|
"learning_rate": 1.2992926041370064e-06, |
|
"loss": 0.059, |
|
"step": 3885 |
|
}, |
|
{ |
|
"epoch": 1.8200959176511873, |
|
"grad_norm": 0.1502443104982376, |
|
"learning_rate": 1.2663895201449588e-06, |
|
"loss": 0.058, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 1.822435372558194, |
|
"grad_norm": 0.16258084774017334, |
|
"learning_rate": 1.2338975850653579e-06, |
|
"loss": 0.0562, |
|
"step": 3895 |
|
}, |
|
{ |
|
"epoch": 1.8247748274652005, |
|
"grad_norm": 0.1836717575788498, |
|
"learning_rate": 1.201817361771837e-06, |
|
"loss": 0.0555, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.8271142823722073, |
|
"grad_norm": 0.17480716109275818, |
|
"learning_rate": 1.1701494060057606e-06, |
|
"loss": 0.0635, |
|
"step": 3905 |
|
}, |
|
{ |
|
"epoch": 1.829453737279214, |
|
"grad_norm": 0.17331267893314362, |
|
"learning_rate": 1.138894266366572e-06, |
|
"loss": 0.0628, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 1.8317931921862205, |
|
"grad_norm": 0.15012997388839722, |
|
"learning_rate": 1.108052484302302e-06, |
|
"loss": 0.0569, |
|
"step": 3915 |
|
}, |
|
{ |
|
"epoch": 1.8341326470932273, |
|
"grad_norm": 0.16882111132144928, |
|
"learning_rate": 1.0776245941001878e-06, |
|
"loss": 0.0558, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 1.836472102000234, |
|
"grad_norm": 0.17416958510875702, |
|
"learning_rate": 1.0476111228774178e-06, |
|
"loss": 0.0571, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 1.8388115569072405, |
|
"grad_norm": 0.1751696616411209, |
|
"learning_rate": 1.01801259057199e-06, |
|
"loss": 0.054, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 1.8411510118142473, |
|
"grad_norm": 0.17922084033489227, |
|
"learning_rate": 9.888295099337252e-07, |
|
"loss": 0.0611, |
|
"step": 3935 |
|
}, |
|
{ |
|
"epoch": 1.843490466721254, |
|
"grad_norm": 0.20534135401248932, |
|
"learning_rate": 9.60062386515359e-07, |
|
"loss": 0.065, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 1.8458299216282605, |
|
"grad_norm": 0.17067112028598785, |
|
"learning_rate": 9.317117186638108e-07, |
|
"loss": 0.0591, |
|
"step": 3945 |
|
}, |
|
{ |
|
"epoch": 1.8481693765352674, |
|
"grad_norm": 0.17611610889434814, |
|
"learning_rate": 9.037779975115235e-07, |
|
"loss": 0.0567, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.850508831442274, |
|
"grad_norm": 0.1562628149986267, |
|
"learning_rate": 8.762617069679846e-07, |
|
"loss": 0.0613, |
|
"step": 3955 |
|
}, |
|
{ |
|
"epoch": 1.8528482863492806, |
|
"grad_norm": 0.1408929079771042, |
|
"learning_rate": 8.491633237113078e-07, |
|
"loss": 0.0593, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.8551877412562874, |
|
"grad_norm": 0.18245179951190948, |
|
"learning_rate": 8.224833171800173e-07, |
|
"loss": 0.0607, |
|
"step": 3965 |
|
}, |
|
{ |
|
"epoch": 1.857527196163294, |
|
"grad_norm": 0.17572057247161865, |
|
"learning_rate": 7.962221495648708e-07, |
|
"loss": 0.0625, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 1.8598666510703006, |
|
"grad_norm": 0.1581767499446869, |
|
"learning_rate": 7.703802758008943e-07, |
|
"loss": 0.0629, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 1.8622061059773074, |
|
"grad_norm": 0.1815463751554489, |
|
"learning_rate": 7.44958143559471e-07, |
|
"loss": 0.0592, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 1.864545560884314, |
|
"grad_norm": 0.17877933382987976, |
|
"learning_rate": 7.199561932405952e-07, |
|
"loss": 0.0606, |
|
"step": 3985 |
|
}, |
|
{ |
|
"epoch": 1.8668850157913206, |
|
"grad_norm": 0.17959001660346985, |
|
"learning_rate": 6.953748579652558e-07, |
|
"loss": 0.0687, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 1.8692244706983274, |
|
"grad_norm": 0.19308309257030487, |
|
"learning_rate": 6.712145635679096e-07, |
|
"loss": 0.0681, |
|
"step": 3995 |
|
}, |
|
{ |
|
"epoch": 1.8715639256053338, |
|
"grad_norm": 0.1864127218723297, |
|
"learning_rate": 6.474757285891253e-07, |
|
"loss": 0.0599, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.8715639256053338, |
|
"eval_loss": 0.07669652253389359, |
|
"eval_runtime": 233.842, |
|
"eval_samples_per_second": 2.985, |
|
"eval_steps_per_second": 0.748, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.8739033805123406, |
|
"grad_norm": 0.17673756182193756, |
|
"learning_rate": 6.241587642683206e-07, |
|
"loss": 0.0624, |
|
"step": 4005 |
|
}, |
|
{ |
|
"epoch": 1.8762428354193474, |
|
"grad_norm": 0.16083566844463348, |
|
"learning_rate": 6.012640745366477e-07, |
|
"loss": 0.0526, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 1.8785822903263538, |
|
"grad_norm": 0.1742735356092453, |
|
"learning_rate": 5.78792056009983e-07, |
|
"loss": 0.058, |
|
"step": 4015 |
|
}, |
|
{ |
|
"epoch": 1.8809217452333606, |
|
"grad_norm": 0.1606372892856598, |
|
"learning_rate": 5.567430979820654e-07, |
|
"loss": 0.1072, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.8832612001403672, |
|
"grad_norm": 0.14640302956104279, |
|
"learning_rate": 5.35117582417749e-07, |
|
"loss": 0.0527, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 1.8856006550473738, |
|
"grad_norm": 0.17613400518894196, |
|
"learning_rate": 5.139158839463948e-07, |
|
"loss": 0.1046, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 1.8879401099543807, |
|
"grad_norm": 0.1592911034822464, |
|
"learning_rate": 4.931383698553643e-07, |
|
"loss": 0.0654, |
|
"step": 4035 |
|
}, |
|
{ |
|
"epoch": 1.8902795648613873, |
|
"grad_norm": 0.16795676946640015, |
|
"learning_rate": 4.727854000836723e-07, |
|
"loss": 0.0818, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 1.8926190197683939, |
|
"grad_norm": 0.16940264403820038, |
|
"learning_rate": 4.528573272157416e-07, |
|
"loss": 0.0562, |
|
"step": 4045 |
|
}, |
|
{ |
|
"epoch": 1.8949584746754007, |
|
"grad_norm": 0.17346608638763428, |
|
"learning_rate": 4.333544964753022e-07, |
|
"loss": 0.0563, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.8972979295824073, |
|
"grad_norm": 0.16219955682754517, |
|
"learning_rate": 4.1427724571940217e-07, |
|
"loss": 0.0677, |
|
"step": 4055 |
|
}, |
|
{ |
|
"epoch": 1.8996373844894139, |
|
"grad_norm": 0.20030000805854797, |
|
"learning_rate": 3.9562590543256175e-07, |
|
"loss": 0.0651, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 1.9019768393964207, |
|
"grad_norm": 0.14878813922405243, |
|
"learning_rate": 3.7740079872105317e-07, |
|
"loss": 0.0555, |
|
"step": 4065 |
|
}, |
|
{ |
|
"epoch": 1.9043162943034273, |
|
"grad_norm": 0.1636265218257904, |
|
"learning_rate": 3.5960224130728857e-07, |
|
"loss": 0.0634, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 1.906655749210434, |
|
"grad_norm": 0.1574837565422058, |
|
"learning_rate": 3.422305415243576e-07, |
|
"loss": 0.0564, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 1.9089952041174407, |
|
"grad_norm": 0.16645601391792297, |
|
"learning_rate": 3.2528600031069557e-07, |
|
"loss": 0.0586, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.9113346590244473, |
|
"grad_norm": 0.16494524478912354, |
|
"learning_rate": 3.087689112048542e-07, |
|
"loss": 0.0564, |
|
"step": 4085 |
|
}, |
|
{ |
|
"epoch": 1.913674113931454, |
|
"grad_norm": 0.14947259426116943, |
|
"learning_rate": 2.926795603404198e-07, |
|
"loss": 0.0609, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 1.9160135688384607, |
|
"grad_norm": 0.1580345779657364, |
|
"learning_rate": 2.7701822644107536e-07, |
|
"loss": 0.0557, |
|
"step": 4095 |
|
}, |
|
{ |
|
"epoch": 1.9183530237454673, |
|
"grad_norm": 0.15407022833824158, |
|
"learning_rate": 2.6178518081574064e-07, |
|
"loss": 0.0565, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.920692478652474, |
|
"grad_norm": 0.1476811021566391, |
|
"learning_rate": 2.46980687353901e-07, |
|
"loss": 0.0609, |
|
"step": 4105 |
|
}, |
|
{ |
|
"epoch": 1.9230319335594808, |
|
"grad_norm": 0.18727290630340576, |
|
"learning_rate": 2.3260500252101636e-07, |
|
"loss": 0.0618, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 1.9253713884664871, |
|
"grad_norm": 0.18050184845924377, |
|
"learning_rate": 2.1865837535409162e-07, |
|
"loss": 0.0516, |
|
"step": 4115 |
|
}, |
|
{ |
|
"epoch": 1.927710843373494, |
|
"grad_norm": 0.19610415399074554, |
|
"learning_rate": 2.0514104745734674e-07, |
|
"loss": 0.0646, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 1.9300502982805008, |
|
"grad_norm": 0.16899576783180237, |
|
"learning_rate": 1.9205325299805622e-07, |
|
"loss": 0.072, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 1.9323897531875072, |
|
"grad_norm": 0.17279615998268127, |
|
"learning_rate": 1.7939521870245779e-07, |
|
"loss": 0.0584, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 1.934729208094514, |
|
"grad_norm": 0.1838502138853073, |
|
"learning_rate": 1.6716716385186126e-07, |
|
"loss": 0.0576, |
|
"step": 4135 |
|
}, |
|
{ |
|
"epoch": 1.9370686630015208, |
|
"grad_norm": 0.1634179651737213, |
|
"learning_rate": 1.5536930027882368e-07, |
|
"loss": 0.1097, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.9394081179085272, |
|
"grad_norm": 0.17711836099624634, |
|
"learning_rate": 1.4400183236349397e-07, |
|
"loss": 0.0637, |
|
"step": 4145 |
|
}, |
|
{ |
|
"epoch": 1.941747572815534, |
|
"grad_norm": 0.1731368601322174, |
|
"learning_rate": 1.330649570300574e-07, |
|
"loss": 0.0683, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.9440870277225406, |
|
"grad_norm": 0.16300569474697113, |
|
"learning_rate": 1.2255886374334946e-07, |
|
"loss": 0.0539, |
|
"step": 4155 |
|
}, |
|
{ |
|
"epoch": 1.9464264826295472, |
|
"grad_norm": 0.21170642971992493, |
|
"learning_rate": 1.1248373450554462e-07, |
|
"loss": 0.0618, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 1.948765937536554, |
|
"grad_norm": 0.14262203872203827, |
|
"learning_rate": 1.0283974385301986e-07, |
|
"loss": 0.06, |
|
"step": 4165 |
|
}, |
|
{ |
|
"epoch": 1.9511053924435606, |
|
"grad_norm": 0.17411337792873383, |
|
"learning_rate": 9.3627058853335e-08, |
|
"loss": 0.0574, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 1.9534448473505672, |
|
"grad_norm": 0.15016813576221466, |
|
"learning_rate": 8.484583910232379e-08, |
|
"loss": 0.0595, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 1.955784302257574, |
|
"grad_norm": 0.16608966886997223, |
|
"learning_rate": 7.649623672134065e-08, |
|
"loss": 0.0605, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 1.9581237571645806, |
|
"grad_norm": 0.1695844531059265, |
|
"learning_rate": 6.857839635462104e-08, |
|
"loss": 0.0619, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 1.9604632120715872, |
|
"grad_norm": 0.1505594253540039, |
|
"learning_rate": 6.109245516677242e-08, |
|
"loss": 0.0623, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 1.962802666978594, |
|
"grad_norm": 0.17143121361732483, |
|
"learning_rate": 5.403854284040943e-08, |
|
"loss": 0.0573, |
|
"step": 4195 |
|
}, |
|
{ |
|
"epoch": 1.9651421218856007, |
|
"grad_norm": 0.18096469342708588, |
|
"learning_rate": 4.741678157389739e-08, |
|
"loss": 0.0646, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.9674815767926073, |
|
"grad_norm": 0.17704518139362335, |
|
"learning_rate": 4.122728607923731e-08, |
|
"loss": 0.0602, |
|
"step": 4205 |
|
}, |
|
{ |
|
"epoch": 1.969821031699614, |
|
"grad_norm": 0.16505348682403564, |
|
"learning_rate": 3.5470163580073046e-08, |
|
"loss": 0.0947, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 1.9721604866066207, |
|
"grad_norm": 0.13319675624370575, |
|
"learning_rate": 3.014551380985386e-08, |
|
"loss": 0.0566, |
|
"step": 4215 |
|
}, |
|
{ |
|
"epoch": 1.9744999415136273, |
|
"grad_norm": 0.17260295152664185, |
|
"learning_rate": 2.525342901008032e-08, |
|
"loss": 0.0612, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 1.976839396420634, |
|
"grad_norm": 0.19151194393634796, |
|
"learning_rate": 2.0793993928724942e-08, |
|
"loss": 0.0652, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 1.9791788513276407, |
|
"grad_norm": 0.1425478458404541, |
|
"learning_rate": 1.676728581876119e-08, |
|
"loss": 0.0576, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 1.9815183062346473, |
|
"grad_norm": 0.14790157973766327, |
|
"learning_rate": 1.3173374436811769e-08, |
|
"loss": 0.0639, |
|
"step": 4235 |
|
}, |
|
{ |
|
"epoch": 1.9838577611416541, |
|
"grad_norm": 0.17651847004890442, |
|
"learning_rate": 1.0012322041960676e-08, |
|
"loss": 0.0503, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 1.9861972160486605, |
|
"grad_norm": 0.16545896232128143, |
|
"learning_rate": 7.284183394656863e-09, |
|
"loss": 0.0557, |
|
"step": 4245 |
|
}, |
|
{ |
|
"epoch": 1.9885366709556673, |
|
"grad_norm": 0.15864944458007812, |
|
"learning_rate": 4.98900575578165e-09, |
|
"loss": 0.0591, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.9908761258626742, |
|
"grad_norm": 0.1790982186794281, |
|
"learning_rate": 3.126828885816058e-09, |
|
"loss": 0.0607, |
|
"step": 4255 |
|
}, |
|
{ |
|
"epoch": 1.9932155807696805, |
|
"grad_norm": 0.20370884239673615, |
|
"learning_rate": 1.6976850441552394e-09, |
|
"loss": 0.059, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.9955550356766873, |
|
"grad_norm": 0.16718322038650513, |
|
"learning_rate": 7.015989885589269e-10, |
|
"loss": 0.0575, |
|
"step": 4265 |
|
}, |
|
{ |
|
"epoch": 1.997894490583694, |
|
"grad_norm": 0.17118395864963531, |
|
"learning_rate": 1.3858797471011376e-10, |
|
"loss": 0.0704, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 1.9997660545092995, |
|
"step": 4274, |
|
"total_flos": 3.0051315317703967e+18, |
|
"train_loss": 0.08305150749153994, |
|
"train_runtime": 80721.3637, |
|
"train_samples_per_second": 0.847, |
|
"train_steps_per_second": 0.053 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 4274, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.0051315317703967e+18, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|