{ "best_metric": 0.07669652253389359, "best_model_checkpoint": "/teamspace/studios/this_studio/output/Mistral_End/checkpoint-4000", "epoch": 1.9997660545092995, "eval_steps": 500, "global_step": 4274, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0023394549070066676, "grad_norm": 1.2956504821777344, "learning_rate": 5.000000000000001e-07, "loss": 0.7204, "step": 5 }, { "epoch": 0.004678909814013335, "grad_norm": 1.2448140382766724, "learning_rate": 1.0000000000000002e-06, "loss": 0.7411, "step": 10 }, { "epoch": 0.007018364721020002, "grad_norm": 1.3956960439682007, "learning_rate": 1.5e-06, "loss": 0.7235, "step": 15 }, { "epoch": 0.00935781962802667, "grad_norm": 1.10906982421875, "learning_rate": 2.0000000000000003e-06, "loss": 0.7176, "step": 20 }, { "epoch": 0.011697274535033338, "grad_norm": 1.1599094867706299, "learning_rate": 2.5e-06, "loss": 0.7439, "step": 25 }, { "epoch": 0.014036729442040005, "grad_norm": 1.085955023765564, "learning_rate": 3e-06, "loss": 0.6894, "step": 30 }, { "epoch": 0.016376184349046672, "grad_norm": 0.9708353877067566, "learning_rate": 3.5000000000000004e-06, "loss": 0.6607, "step": 35 }, { "epoch": 0.01871563925605334, "grad_norm": 0.8504016399383545, "learning_rate": 4.000000000000001e-06, "loss": 0.6335, "step": 40 }, { "epoch": 0.021055094163060006, "grad_norm": 0.7278764843940735, "learning_rate": 4.5e-06, "loss": 0.5814, "step": 45 }, { "epoch": 0.023394549070066675, "grad_norm": 0.5420047044754028, "learning_rate": 5e-06, "loss": 0.5636, "step": 50 }, { "epoch": 0.02573400397707334, "grad_norm": 0.5314227938652039, "learning_rate": 5.500000000000001e-06, "loss": 0.5845, "step": 55 }, { "epoch": 0.02807345888408001, "grad_norm": 0.48584049940109253, "learning_rate": 6e-06, "loss": 0.494, "step": 60 }, { "epoch": 0.030412913791086675, "grad_norm": 0.49247828125953674, "learning_rate": 6.5000000000000004e-06, "loss": 0.4809, "step": 65 }, { "epoch": 0.032752368698093344, "grad_norm": 0.470062255859375, "learning_rate": 7.000000000000001e-06, "loss": 0.446, "step": 70 }, { "epoch": 0.03509182360510001, "grad_norm": 0.4149915874004364, "learning_rate": 7.5e-06, "loss": 0.3915, "step": 75 }, { "epoch": 0.03743127851210668, "grad_norm": 0.41831153631210327, "learning_rate": 8.000000000000001e-06, "loss": 0.3696, "step": 80 }, { "epoch": 0.039770733419113344, "grad_norm": 0.40496742725372314, "learning_rate": 8.500000000000002e-06, "loss": 0.3464, "step": 85 }, { "epoch": 0.04211018832612001, "grad_norm": 0.41090327501296997, "learning_rate": 9e-06, "loss": 0.3097, "step": 90 }, { "epoch": 0.04444964323312668, "grad_norm": 0.41252198815345764, "learning_rate": 9.5e-06, "loss": 0.3111, "step": 95 }, { "epoch": 0.04678909814013335, "grad_norm": 0.44755125045776367, "learning_rate": 1e-05, "loss": 0.2696, "step": 100 }, { "epoch": 0.04912855304714002, "grad_norm": 0.4423600137233734, "learning_rate": 1.05e-05, "loss": 0.2461, "step": 105 }, { "epoch": 0.05146800795414668, "grad_norm": 0.44379884004592896, "learning_rate": 1.1000000000000001e-05, "loss": 0.2139, "step": 110 }, { "epoch": 0.05380746286115335, "grad_norm": 0.43690061569213867, "learning_rate": 1.1500000000000002e-05, "loss": 0.2045, "step": 115 }, { "epoch": 0.05614691776816002, "grad_norm": 0.4275108277797699, "learning_rate": 1.2e-05, "loss": 0.1987, "step": 120 }, { "epoch": 0.05848637267516669, "grad_norm": 0.5754674077033997, "learning_rate": 1.25e-05, "loss": 0.1899, "step": 125 }, { "epoch": 0.06082582758217335, "grad_norm": 0.4689083397388458, "learning_rate": 1.3000000000000001e-05, "loss": 0.171, "step": 130 }, { "epoch": 0.06316528248918002, "grad_norm": 0.42702773213386536, "learning_rate": 1.3500000000000001e-05, "loss": 0.1636, "step": 135 }, { "epoch": 0.06550473739618669, "grad_norm": 0.4773492217063904, "learning_rate": 1.4000000000000001e-05, "loss": 0.1579, "step": 140 }, { "epoch": 0.06784419230319336, "grad_norm": 0.5190818309783936, "learning_rate": 1.45e-05, "loss": 0.1377, "step": 145 }, { "epoch": 0.07018364721020003, "grad_norm": 0.4696808159351349, "learning_rate": 1.5e-05, "loss": 0.1325, "step": 150 }, { "epoch": 0.0725231021172067, "grad_norm": 0.4878705143928528, "learning_rate": 1.55e-05, "loss": 0.1384, "step": 155 }, { "epoch": 0.07486255702421336, "grad_norm": 0.48388373851776123, "learning_rate": 1.6000000000000003e-05, "loss": 0.1198, "step": 160 }, { "epoch": 0.07720201193122003, "grad_norm": 0.47326740622520447, "learning_rate": 1.65e-05, "loss": 0.1192, "step": 165 }, { "epoch": 0.07954146683822669, "grad_norm": 0.4732869565486908, "learning_rate": 1.7000000000000003e-05, "loss": 0.1208, "step": 170 }, { "epoch": 0.08188092174523336, "grad_norm": 0.45708340406417847, "learning_rate": 1.75e-05, "loss": 0.1099, "step": 175 }, { "epoch": 0.08422037665224003, "grad_norm": 0.5060539245605469, "learning_rate": 1.8e-05, "loss": 0.104, "step": 180 }, { "epoch": 0.0865598315592467, "grad_norm": 0.45235252380371094, "learning_rate": 1.85e-05, "loss": 0.1489, "step": 185 }, { "epoch": 0.08889928646625336, "grad_norm": 0.48685696721076965, "learning_rate": 1.9e-05, "loss": 0.1102, "step": 190 }, { "epoch": 0.09123874137326003, "grad_norm": 0.45249509811401367, "learning_rate": 1.9500000000000003e-05, "loss": 0.1172, "step": 195 }, { "epoch": 0.0935781962802667, "grad_norm": 0.456840455532074, "learning_rate": 2e-05, "loss": 0.1046, "step": 200 }, { "epoch": 0.09591765118727337, "grad_norm": 0.4239897131919861, "learning_rate": 2.05e-05, "loss": 0.1024, "step": 205 }, { "epoch": 0.09825710609428004, "grad_norm": 0.3757131099700928, "learning_rate": 2.1e-05, "loss": 0.1117, "step": 210 }, { "epoch": 0.1005965610012867, "grad_norm": 0.43743741512298584, "learning_rate": 2.15e-05, "loss": 0.1177, "step": 215 }, { "epoch": 0.10293601590829336, "grad_norm": 0.39099183678627014, "learning_rate": 2.2000000000000003e-05, "loss": 0.0988, "step": 220 }, { "epoch": 0.10527547081530003, "grad_norm": 0.44669702649116516, "learning_rate": 2.25e-05, "loss": 0.0943, "step": 225 }, { "epoch": 0.1076149257223067, "grad_norm": 0.4071020185947418, "learning_rate": 2.3000000000000003e-05, "loss": 0.0989, "step": 230 }, { "epoch": 0.10995438062931337, "grad_norm": 0.3775785565376282, "learning_rate": 2.35e-05, "loss": 0.0953, "step": 235 }, { "epoch": 0.11229383553632004, "grad_norm": 0.417341947555542, "learning_rate": 2.4e-05, "loss": 0.0989, "step": 240 }, { "epoch": 0.11463329044332671, "grad_norm": 0.43018296360969543, "learning_rate": 2.45e-05, "loss": 0.0933, "step": 245 }, { "epoch": 0.11697274535033338, "grad_norm": 0.4740806818008423, "learning_rate": 2.5e-05, "loss": 0.1041, "step": 250 }, { "epoch": 0.11931220025734005, "grad_norm": 0.3639010488986969, "learning_rate": 2.5500000000000003e-05, "loss": 0.0983, "step": 255 }, { "epoch": 0.1216516551643467, "grad_norm": 0.46424400806427, "learning_rate": 2.6000000000000002e-05, "loss": 0.1017, "step": 260 }, { "epoch": 0.12399111007135337, "grad_norm": 0.5055582523345947, "learning_rate": 2.6500000000000004e-05, "loss": 0.1486, "step": 265 }, { "epoch": 0.12633056497836004, "grad_norm": 0.39390507340431213, "learning_rate": 2.7000000000000002e-05, "loss": 0.0973, "step": 270 }, { "epoch": 0.12867001988536672, "grad_norm": 0.35929396748542786, "learning_rate": 2.7500000000000004e-05, "loss": 0.0957, "step": 275 }, { "epoch": 0.13100947479237338, "grad_norm": 0.4053295850753784, "learning_rate": 2.8000000000000003e-05, "loss": 0.1, "step": 280 }, { "epoch": 0.13334892969938003, "grad_norm": 0.45273882150650024, "learning_rate": 2.8499999999999998e-05, "loss": 0.0878, "step": 285 }, { "epoch": 0.1356883846063867, "grad_norm": 0.3821883797645569, "learning_rate": 2.9e-05, "loss": 0.0881, "step": 290 }, { "epoch": 0.13802783951339337, "grad_norm": 0.47926971316337585, "learning_rate": 2.95e-05, "loss": 0.1115, "step": 295 }, { "epoch": 0.14036729442040005, "grad_norm": 0.37854063510894775, "learning_rate": 3e-05, "loss": 0.1025, "step": 300 }, { "epoch": 0.1427067493274067, "grad_norm": 0.49879732728004456, "learning_rate": 3.05e-05, "loss": 0.0978, "step": 305 }, { "epoch": 0.1450462042344134, "grad_norm": 0.372403085231781, "learning_rate": 3.1e-05, "loss": 0.0931, "step": 310 }, { "epoch": 0.14738565914142004, "grad_norm": 0.3892665505409241, "learning_rate": 3.15e-05, "loss": 0.0947, "step": 315 }, { "epoch": 0.14972511404842673, "grad_norm": 0.39579498767852783, "learning_rate": 3.2000000000000005e-05, "loss": 0.0876, "step": 320 }, { "epoch": 0.15206456895543338, "grad_norm": 0.33920520544052124, "learning_rate": 3.2500000000000004e-05, "loss": 0.0941, "step": 325 }, { "epoch": 0.15440402386244007, "grad_norm": 0.32931971549987793, "learning_rate": 3.3e-05, "loss": 0.0825, "step": 330 }, { "epoch": 0.15674347876944672, "grad_norm": 0.34293580055236816, "learning_rate": 3.35e-05, "loss": 0.0803, "step": 335 }, { "epoch": 0.15908293367645338, "grad_norm": 0.2816343903541565, "learning_rate": 3.4000000000000007e-05, "loss": 0.081, "step": 340 }, { "epoch": 0.16142238858346006, "grad_norm": 0.4197327792644501, "learning_rate": 3.45e-05, "loss": 0.0895, "step": 345 }, { "epoch": 0.1637618434904667, "grad_norm": 0.33747074007987976, "learning_rate": 3.5e-05, "loss": 0.0882, "step": 350 }, { "epoch": 0.1661012983974734, "grad_norm": 0.31165778636932373, "learning_rate": 3.55e-05, "loss": 0.0928, "step": 355 }, { "epoch": 0.16844075330448005, "grad_norm": 0.3359187841415405, "learning_rate": 3.6e-05, "loss": 0.0939, "step": 360 }, { "epoch": 0.17078020821148673, "grad_norm": 0.30003583431243896, "learning_rate": 3.65e-05, "loss": 0.0842, "step": 365 }, { "epoch": 0.1731196631184934, "grad_norm": 0.35046064853668213, "learning_rate": 3.7e-05, "loss": 0.0954, "step": 370 }, { "epoch": 0.17545911802550007, "grad_norm": 0.37043917179107666, "learning_rate": 3.7500000000000003e-05, "loss": 0.0838, "step": 375 }, { "epoch": 0.17779857293250673, "grad_norm": 0.30147868394851685, "learning_rate": 3.8e-05, "loss": 0.101, "step": 380 }, { "epoch": 0.18013802783951338, "grad_norm": 0.3372386395931244, "learning_rate": 3.85e-05, "loss": 0.0873, "step": 385 }, { "epoch": 0.18247748274652006, "grad_norm": 0.32464468479156494, "learning_rate": 3.9000000000000006e-05, "loss": 0.09, "step": 390 }, { "epoch": 0.18481693765352672, "grad_norm": 0.3182346224784851, "learning_rate": 3.9500000000000005e-05, "loss": 0.0892, "step": 395 }, { "epoch": 0.1871563925605334, "grad_norm": 0.2824024558067322, "learning_rate": 4e-05, "loss": 0.0832, "step": 400 }, { "epoch": 0.18949584746754006, "grad_norm": 0.30845776200294495, "learning_rate": 4.05e-05, "loss": 0.093, "step": 405 }, { "epoch": 0.19183530237454674, "grad_norm": 0.3025493323802948, "learning_rate": 4.1e-05, "loss": 0.0877, "step": 410 }, { "epoch": 0.1941747572815534, "grad_norm": 0.35149937868118286, "learning_rate": 4.15e-05, "loss": 0.0841, "step": 415 }, { "epoch": 0.19651421218856008, "grad_norm": 0.34748056530952454, "learning_rate": 4.2e-05, "loss": 0.0923, "step": 420 }, { "epoch": 0.19885366709556673, "grad_norm": 0.3216325640678406, "learning_rate": 4.25e-05, "loss": 0.081, "step": 425 }, { "epoch": 0.2011931220025734, "grad_norm": 0.2797127366065979, "learning_rate": 4.3e-05, "loss": 0.0808, "step": 430 }, { "epoch": 0.20353257690958007, "grad_norm": 0.2927537262439728, "learning_rate": 4.35e-05, "loss": 0.1383, "step": 435 }, { "epoch": 0.20587203181658673, "grad_norm": 0.2892308533191681, "learning_rate": 4.4000000000000006e-05, "loss": 0.0836, "step": 440 }, { "epoch": 0.2082114867235934, "grad_norm": 0.35608819127082825, "learning_rate": 4.4500000000000004e-05, "loss": 0.0796, "step": 445 }, { "epoch": 0.21055094163060006, "grad_norm": 0.29528915882110596, "learning_rate": 4.5e-05, "loss": 0.09, "step": 450 }, { "epoch": 0.21289039653760675, "grad_norm": 0.3688681721687317, "learning_rate": 4.55e-05, "loss": 0.0928, "step": 455 }, { "epoch": 0.2152298514446134, "grad_norm": 0.3129563629627228, "learning_rate": 4.600000000000001e-05, "loss": 0.0914, "step": 460 }, { "epoch": 0.21756930635162008, "grad_norm": 0.34032294154167175, "learning_rate": 4.6500000000000005e-05, "loss": 0.0978, "step": 465 }, { "epoch": 0.21990876125862674, "grad_norm": 0.283371239900589, "learning_rate": 4.7e-05, "loss": 0.1109, "step": 470 }, { "epoch": 0.2222482161656334, "grad_norm": 0.25409555435180664, "learning_rate": 4.75e-05, "loss": 0.0804, "step": 475 }, { "epoch": 0.22458767107264008, "grad_norm": 0.24365410208702087, "learning_rate": 4.8e-05, "loss": 0.0737, "step": 480 }, { "epoch": 0.22692712597964673, "grad_norm": 0.28039586544036865, "learning_rate": 4.85e-05, "loss": 0.0801, "step": 485 }, { "epoch": 0.22926658088665341, "grad_norm": 0.28205862641334534, "learning_rate": 4.9e-05, "loss": 0.0784, "step": 490 }, { "epoch": 0.23160603579366007, "grad_norm": 0.2688550353050232, "learning_rate": 4.9500000000000004e-05, "loss": 0.0838, "step": 495 }, { "epoch": 0.23394549070066675, "grad_norm": 0.31778618693351746, "learning_rate": 5e-05, "loss": 0.0893, "step": 500 }, { "epoch": 0.23394549070066675, "eval_loss": 0.09720832854509354, "eval_runtime": 235.4562, "eval_samples_per_second": 2.964, "eval_steps_per_second": 0.743, "step": 500 }, { "epoch": 0.2362849456076734, "grad_norm": 0.27649974822998047, "learning_rate": 4.999978345640206e-05, "loss": 0.0818, "step": 505 }, { "epoch": 0.2386244005146801, "grad_norm": 0.24827256798744202, "learning_rate": 4.9999133829359514e-05, "loss": 0.0812, "step": 510 }, { "epoch": 0.24096385542168675, "grad_norm": 0.24930593371391296, "learning_rate": 4.999805113012618e-05, "loss": 0.0732, "step": 515 }, { "epoch": 0.2433033103286934, "grad_norm": 0.2697354555130005, "learning_rate": 4.999653537745817e-05, "loss": 0.0843, "step": 520 }, { "epoch": 0.24564276523570008, "grad_norm": 0.258434534072876, "learning_rate": 4.999458659761363e-05, "loss": 0.07, "step": 525 }, { "epoch": 0.24798222014270674, "grad_norm": 0.2943266034126282, "learning_rate": 4.9992204824352213e-05, "loss": 0.0852, "step": 530 }, { "epoch": 0.2503216750497134, "grad_norm": 0.23471561074256897, "learning_rate": 4.998939009893453e-05, "loss": 0.0847, "step": 535 }, { "epoch": 0.2526611299567201, "grad_norm": 0.27082759141921997, "learning_rate": 4.9986142470121453e-05, "loss": 0.0727, "step": 540 }, { "epoch": 0.25500058486372673, "grad_norm": 0.26077497005462646, "learning_rate": 4.9982461994173234e-05, "loss": 0.0893, "step": 545 }, { "epoch": 0.25734003977073344, "grad_norm": 0.2686660587787628, "learning_rate": 4.9978348734848556e-05, "loss": 0.0764, "step": 550 }, { "epoch": 0.2596794946777401, "grad_norm": 0.2592894434928894, "learning_rate": 4.9973802763403424e-05, "loss": 0.1747, "step": 555 }, { "epoch": 0.26201894958474675, "grad_norm": 0.26700106263160706, "learning_rate": 4.996882415858991e-05, "loss": 0.0808, "step": 560 }, { "epoch": 0.2643584044917534, "grad_norm": 0.29679951071739197, "learning_rate": 4.996341300665481e-05, "loss": 0.0847, "step": 565 }, { "epoch": 0.26669785939876006, "grad_norm": 0.29322996735572815, "learning_rate": 4.995756940133817e-05, "loss": 0.0884, "step": 570 }, { "epoch": 0.2690373143057668, "grad_norm": 0.24166248738765717, "learning_rate": 4.995129344387158e-05, "loss": 0.0784, "step": 575 }, { "epoch": 0.2713767692127734, "grad_norm": 0.23448926210403442, "learning_rate": 4.9944585242976546e-05, "loss": 0.146, "step": 580 }, { "epoch": 0.2737162241197801, "grad_norm": 0.2622728645801544, "learning_rate": 4.99374449148625e-05, "loss": 0.0894, "step": 585 }, { "epoch": 0.27605567902678674, "grad_norm": 0.23316816985607147, "learning_rate": 4.992987258322481e-05, "loss": 0.0821, "step": 590 }, { "epoch": 0.27839513393379345, "grad_norm": 0.29072973132133484, "learning_rate": 4.992186837924268e-05, "loss": 0.098, "step": 595 }, { "epoch": 0.2807345888408001, "grad_norm": 0.20675267279148102, "learning_rate": 4.9913432441576845e-05, "loss": 0.0776, "step": 600 }, { "epoch": 0.28307404374780676, "grad_norm": 0.26899468898773193, "learning_rate": 4.9904564916367166e-05, "loss": 0.0862, "step": 605 }, { "epoch": 0.2854134986548134, "grad_norm": 0.22569362819194794, "learning_rate": 4.989526595723012e-05, "loss": 0.0921, "step": 610 }, { "epoch": 0.28775295356182007, "grad_norm": 0.2874622046947479, "learning_rate": 4.988553572525609e-05, "loss": 0.0811, "step": 615 }, { "epoch": 0.2900924084688268, "grad_norm": 0.2119138389825821, "learning_rate": 4.987537438900664e-05, "loss": 0.0711, "step": 620 }, { "epoch": 0.29243186337583343, "grad_norm": 0.2240922749042511, "learning_rate": 4.986478212451157e-05, "loss": 0.0802, "step": 625 }, { "epoch": 0.2947713182828401, "grad_norm": 0.24686302244663239, "learning_rate": 4.985375911526582e-05, "loss": 0.0825, "step": 630 }, { "epoch": 0.29711077318984674, "grad_norm": 0.22553415596485138, "learning_rate": 4.9842305552226365e-05, "loss": 0.0676, "step": 635 }, { "epoch": 0.29945022809685345, "grad_norm": 0.24262481927871704, "learning_rate": 4.9830421633808876e-05, "loss": 0.0752, "step": 640 }, { "epoch": 0.3017896830038601, "grad_norm": 0.24724553525447845, "learning_rate": 4.981810756588426e-05, "loss": 0.0823, "step": 645 }, { "epoch": 0.30412913791086676, "grad_norm": 0.2152285873889923, "learning_rate": 4.980536356177513e-05, "loss": 0.0795, "step": 650 }, { "epoch": 0.3064685928178734, "grad_norm": 0.2218402624130249, "learning_rate": 4.9792189842252074e-05, "loss": 0.0711, "step": 655 }, { "epoch": 0.30880804772488013, "grad_norm": 0.24227651953697205, "learning_rate": 4.977858663552987e-05, "loss": 0.0756, "step": 660 }, { "epoch": 0.3111475026318868, "grad_norm": 0.22776497900485992, "learning_rate": 4.9764554177263507e-05, "loss": 0.0742, "step": 665 }, { "epoch": 0.31348695753889344, "grad_norm": 0.20029255747795105, "learning_rate": 4.975009271054409e-05, "loss": 0.0698, "step": 670 }, { "epoch": 0.3158264124459001, "grad_norm": 0.21585966646671295, "learning_rate": 4.973520248589469e-05, "loss": 0.0858, "step": 675 }, { "epoch": 0.31816586735290675, "grad_norm": 0.23440401256084442, "learning_rate": 4.9719883761265906e-05, "loss": 0.0818, "step": 680 }, { "epoch": 0.32050532225991346, "grad_norm": 0.22107180953025818, "learning_rate": 4.9704136802031485e-05, "loss": 0.0671, "step": 685 }, { "epoch": 0.3228447771669201, "grad_norm": 0.19328860938549042, "learning_rate": 4.968796188098369e-05, "loss": 0.0725, "step": 690 }, { "epoch": 0.32518423207392677, "grad_norm": 0.2299336940050125, "learning_rate": 4.967135927832856e-05, "loss": 0.0852, "step": 695 }, { "epoch": 0.3275236869809334, "grad_norm": 0.24605049192905426, "learning_rate": 4.9654329281681094e-05, "loss": 0.0754, "step": 700 }, { "epoch": 0.32986314188794014, "grad_norm": 0.23175948858261108, "learning_rate": 4.9636872186060215e-05, "loss": 0.0961, "step": 705 }, { "epoch": 0.3322025967949468, "grad_norm": 0.2416304647922516, "learning_rate": 4.961898829388372e-05, "loss": 0.1093, "step": 710 }, { "epoch": 0.33454205170195345, "grad_norm": 0.20699331164360046, "learning_rate": 4.960067791496299e-05, "loss": 0.0793, "step": 715 }, { "epoch": 0.3368815066089601, "grad_norm": 0.22395586967468262, "learning_rate": 4.958194136649765e-05, "loss": 0.0805, "step": 720 }, { "epoch": 0.33922096151596676, "grad_norm": 0.21491140127182007, "learning_rate": 4.956277897307008e-05, "loss": 0.073, "step": 725 }, { "epoch": 0.34156041642297347, "grad_norm": 0.23053660988807678, "learning_rate": 4.954319106663976e-05, "loss": 0.0769, "step": 730 }, { "epoch": 0.3438998713299801, "grad_norm": 0.21306245028972626, "learning_rate": 4.952317798653755e-05, "loss": 0.075, "step": 735 }, { "epoch": 0.3462393262369868, "grad_norm": 0.1973639279603958, "learning_rate": 4.95027400794598e-05, "loss": 0.0697, "step": 740 }, { "epoch": 0.34857878114399343, "grad_norm": 0.2036924660205841, "learning_rate": 4.948187769946234e-05, "loss": 0.08, "step": 745 }, { "epoch": 0.35091823605100014, "grad_norm": 0.2153988480567932, "learning_rate": 4.946059120795437e-05, "loss": 0.0963, "step": 750 }, { "epoch": 0.3532576909580068, "grad_norm": 0.20555338263511658, "learning_rate": 4.943888097369216e-05, "loss": 0.0913, "step": 755 }, { "epoch": 0.35559714586501345, "grad_norm": 0.26464715600013733, "learning_rate": 4.941674737277268e-05, "loss": 0.0845, "step": 760 }, { "epoch": 0.3579366007720201, "grad_norm": 0.19657152891159058, "learning_rate": 4.9394190788627106e-05, "loss": 0.0756, "step": 765 }, { "epoch": 0.36027605567902676, "grad_norm": 0.21668066084384918, "learning_rate": 4.937121161201415e-05, "loss": 0.0803, "step": 770 }, { "epoch": 0.3626155105860335, "grad_norm": 0.24815738201141357, "learning_rate": 4.9347810241013294e-05, "loss": 0.0823, "step": 775 }, { "epoch": 0.36495496549304013, "grad_norm": 0.20182116329669952, "learning_rate": 4.932398708101791e-05, "loss": 0.0703, "step": 780 }, { "epoch": 0.3672944204000468, "grad_norm": 0.2374674379825592, "learning_rate": 4.92997425447282e-05, "loss": 0.084, "step": 785 }, { "epoch": 0.36963387530705344, "grad_norm": 0.2246353179216385, "learning_rate": 4.927507705214412e-05, "loss": 0.0868, "step": 790 }, { "epoch": 0.37197333021406015, "grad_norm": 0.21759703755378723, "learning_rate": 4.924999103055802e-05, "loss": 0.0754, "step": 795 }, { "epoch": 0.3743127851210668, "grad_norm": 0.20431989431381226, "learning_rate": 4.922448491454729e-05, "loss": 0.0803, "step": 800 }, { "epoch": 0.37665224002807346, "grad_norm": 0.20232437551021576, "learning_rate": 4.9198559145966824e-05, "loss": 0.0751, "step": 805 }, { "epoch": 0.3789916949350801, "grad_norm": 0.1968168169260025, "learning_rate": 4.917221417394134e-05, "loss": 0.0798, "step": 810 }, { "epoch": 0.38133114984208677, "grad_norm": 0.2001713663339615, "learning_rate": 4.914545045485767e-05, "loss": 0.0776, "step": 815 }, { "epoch": 0.3836706047490935, "grad_norm": 0.19992566108703613, "learning_rate": 4.911826845235676e-05, "loss": 0.0767, "step": 820 }, { "epoch": 0.38601005965610014, "grad_norm": 0.21567636728286743, "learning_rate": 4.9090668637325696e-05, "loss": 0.0752, "step": 825 }, { "epoch": 0.3883495145631068, "grad_norm": 0.19452187418937683, "learning_rate": 4.906265148788954e-05, "loss": 0.0722, "step": 830 }, { "epoch": 0.39068896947011345, "grad_norm": 0.23549924790859222, "learning_rate": 4.9034217489403045e-05, "loss": 0.0764, "step": 835 }, { "epoch": 0.39302842437712016, "grad_norm": 0.2083604633808136, "learning_rate": 4.9005367134442235e-05, "loss": 0.0767, "step": 840 }, { "epoch": 0.3953678792841268, "grad_norm": 0.2386019378900528, "learning_rate": 4.8976100922795884e-05, "loss": 0.0815, "step": 845 }, { "epoch": 0.39770733419113347, "grad_norm": 0.19933900237083435, "learning_rate": 4.894641936145686e-05, "loss": 0.0935, "step": 850 }, { "epoch": 0.4000467890981401, "grad_norm": 0.16899636387825012, "learning_rate": 4.8916322964613316e-05, "loss": 0.0767, "step": 855 }, { "epoch": 0.4023862440051468, "grad_norm": 0.20377202332019806, "learning_rate": 4.888581225363982e-05, "loss": 0.0771, "step": 860 }, { "epoch": 0.4047256989121535, "grad_norm": 0.2286689430475235, "learning_rate": 4.885488775708831e-05, "loss": 0.0798, "step": 865 }, { "epoch": 0.40706515381916014, "grad_norm": 0.2132563591003418, "learning_rate": 4.882355001067892e-05, "loss": 0.0723, "step": 870 }, { "epoch": 0.4094046087261668, "grad_norm": 0.21681547164916992, "learning_rate": 4.879179955729072e-05, "loss": 0.0777, "step": 875 }, { "epoch": 0.41174406363317345, "grad_norm": 0.20257344841957092, "learning_rate": 4.87596369469523e-05, "loss": 0.0777, "step": 880 }, { "epoch": 0.41408351854018016, "grad_norm": 0.18595652282238007, "learning_rate": 4.872706273683225e-05, "loss": 0.0648, "step": 885 }, { "epoch": 0.4164229734471868, "grad_norm": 0.1963788866996765, "learning_rate": 4.869407749122951e-05, "loss": 0.0688, "step": 890 }, { "epoch": 0.41876242835419347, "grad_norm": 0.22948507964611053, "learning_rate": 4.8660681781563576e-05, "loss": 0.0715, "step": 895 }, { "epoch": 0.4211018832612001, "grad_norm": 0.16155906021595, "learning_rate": 4.8626876186364624e-05, "loss": 0.0651, "step": 900 }, { "epoch": 0.4234413381682068, "grad_norm": 0.193013533949852, "learning_rate": 4.859266129126345e-05, "loss": 0.0754, "step": 905 }, { "epoch": 0.4257807930752135, "grad_norm": 0.21157501637935638, "learning_rate": 4.85580376889814e-05, "loss": 0.1229, "step": 910 }, { "epoch": 0.42812024798222015, "grad_norm": 0.2193852812051773, "learning_rate": 4.8523005979320015e-05, "loss": 0.0755, "step": 915 }, { "epoch": 0.4304597028892268, "grad_norm": 0.20411770045757294, "learning_rate": 4.848756676915069e-05, "loss": 0.0848, "step": 920 }, { "epoch": 0.43279915779623346, "grad_norm": 0.23284126818180084, "learning_rate": 4.845172067240415e-05, "loss": 0.0764, "step": 925 }, { "epoch": 0.43513861270324017, "grad_norm": 0.21067671477794647, "learning_rate": 4.841546831005982e-05, "loss": 0.0821, "step": 930 }, { "epoch": 0.4374780676102468, "grad_norm": 0.1912618726491928, "learning_rate": 4.837881031013506e-05, "loss": 0.0755, "step": 935 }, { "epoch": 0.4398175225172535, "grad_norm": 0.1677297055721283, "learning_rate": 4.8341747307674276e-05, "loss": 0.0695, "step": 940 }, { "epoch": 0.44215697742426013, "grad_norm": 0.22734542191028595, "learning_rate": 4.8304279944737954e-05, "loss": 0.086, "step": 945 }, { "epoch": 0.4444964323312668, "grad_norm": 0.1839355230331421, "learning_rate": 4.8266408870391484e-05, "loss": 0.0797, "step": 950 }, { "epoch": 0.4468358872382735, "grad_norm": 0.20233304798603058, "learning_rate": 4.822813474069398e-05, "loss": 0.084, "step": 955 }, { "epoch": 0.44917534214528015, "grad_norm": 0.20918181538581848, "learning_rate": 4.8189458218686846e-05, "loss": 0.0754, "step": 960 }, { "epoch": 0.4515147970522868, "grad_norm": 0.2046424299478531, "learning_rate": 4.815037997438234e-05, "loss": 0.0762, "step": 965 }, { "epoch": 0.45385425195929346, "grad_norm": 0.20791682600975037, "learning_rate": 4.811090068475197e-05, "loss": 0.0686, "step": 970 }, { "epoch": 0.4561937068663002, "grad_norm": 0.1891777068376541, "learning_rate": 4.807102103371472e-05, "loss": 0.0818, "step": 975 }, { "epoch": 0.45853316177330683, "grad_norm": 0.1632051318883896, "learning_rate": 4.803074171212524e-05, "loss": 0.0693, "step": 980 }, { "epoch": 0.4608726166803135, "grad_norm": 0.18187515437602997, "learning_rate": 4.799006341776185e-05, "loss": 0.0642, "step": 985 }, { "epoch": 0.46321207158732014, "grad_norm": 0.20138341188430786, "learning_rate": 4.7948986855314523e-05, "loss": 0.0759, "step": 990 }, { "epoch": 0.4655515264943268, "grad_norm": 0.19479338824748993, "learning_rate": 4.790751273637256e-05, "loss": 0.0761, "step": 995 }, { "epoch": 0.4678909814013335, "grad_norm": 0.18999288976192474, "learning_rate": 4.7865641779412366e-05, "loss": 0.0753, "step": 1000 }, { "epoch": 0.4678909814013335, "eval_loss": 0.08787659555673599, "eval_runtime": 236.9363, "eval_samples_per_second": 2.946, "eval_steps_per_second": 0.739, "step": 1000 }, { "epoch": 0.47023043630834016, "grad_norm": 0.22196418046951294, "learning_rate": 4.7823374709784955e-05, "loss": 0.0731, "step": 1005 }, { "epoch": 0.4725698912153468, "grad_norm": 0.17047221958637238, "learning_rate": 4.77807122597034e-05, "loss": 0.082, "step": 1010 }, { "epoch": 0.47490934612235347, "grad_norm": 0.23016561567783356, "learning_rate": 4.7737655168230114e-05, "loss": 0.0775, "step": 1015 }, { "epoch": 0.4772488010293602, "grad_norm": 0.21696537733078003, "learning_rate": 4.769420418126412e-05, "loss": 0.0805, "step": 1020 }, { "epoch": 0.47958825593636684, "grad_norm": 0.19352155923843384, "learning_rate": 4.7650360051528054e-05, "loss": 0.0754, "step": 1025 }, { "epoch": 0.4819277108433735, "grad_norm": 0.1966516673564911, "learning_rate": 4.760612353855517e-05, "loss": 0.068, "step": 1030 }, { "epoch": 0.48426716575038015, "grad_norm": 0.22759053111076355, "learning_rate": 4.7561495408676146e-05, "loss": 0.073, "step": 1035 }, { "epoch": 0.4866066206573868, "grad_norm": 0.17850154638290405, "learning_rate": 4.7516476435005865e-05, "loss": 0.074, "step": 1040 }, { "epoch": 0.4889460755643935, "grad_norm": 0.19016915559768677, "learning_rate": 4.7471067397429956e-05, "loss": 0.0734, "step": 1045 }, { "epoch": 0.49128553047140017, "grad_norm": 0.19601960480213165, "learning_rate": 4.742526908259134e-05, "loss": 0.0763, "step": 1050 }, { "epoch": 0.4936249853784068, "grad_norm": 0.17694705724716187, "learning_rate": 4.7379082283876566e-05, "loss": 0.0703, "step": 1055 }, { "epoch": 0.4959644402854135, "grad_norm": 0.18579581379890442, "learning_rate": 4.733250780140206e-05, "loss": 0.0863, "step": 1060 }, { "epoch": 0.4983038951924202, "grad_norm": 0.18680576980113983, "learning_rate": 4.728554644200034e-05, "loss": 0.1042, "step": 1065 }, { "epoch": 0.5006433500994268, "grad_norm": 0.1797022670507431, "learning_rate": 4.723819901920591e-05, "loss": 0.0743, "step": 1070 }, { "epoch": 0.5029828050064336, "grad_norm": 0.20039916038513184, "learning_rate": 4.719046635324129e-05, "loss": 0.0707, "step": 1075 }, { "epoch": 0.5053222599134402, "grad_norm": 0.19813776016235352, "learning_rate": 4.7142349271002735e-05, "loss": 0.0738, "step": 1080 }, { "epoch": 0.5076617148204469, "grad_norm": 0.18406710028648376, "learning_rate": 4.709384860604593e-05, "loss": 0.0724, "step": 1085 }, { "epoch": 0.5100011697274535, "grad_norm": 0.20730999112129211, "learning_rate": 4.704496519857156e-05, "loss": 0.0738, "step": 1090 }, { "epoch": 0.5123406246344602, "grad_norm": 0.2113994061946869, "learning_rate": 4.699569989541074e-05, "loss": 0.0699, "step": 1095 }, { "epoch": 0.5146800795414669, "grad_norm": 0.1848883181810379, "learning_rate": 4.694605355001034e-05, "loss": 0.072, "step": 1100 }, { "epoch": 0.5170195344484735, "grad_norm": 0.19623607397079468, "learning_rate": 4.689602702241823e-05, "loss": 0.0761, "step": 1105 }, { "epoch": 0.5193589893554802, "grad_norm": 0.18376454710960388, "learning_rate": 4.684562117926835e-05, "loss": 0.0793, "step": 1110 }, { "epoch": 0.5216984442624868, "grad_norm": 0.18506892025470734, "learning_rate": 4.679483689376571e-05, "loss": 0.0731, "step": 1115 }, { "epoch": 0.5240378991694935, "grad_norm": 0.1819106787443161, "learning_rate": 4.674367504567127e-05, "loss": 0.0795, "step": 1120 }, { "epoch": 0.5263773540765002, "grad_norm": 0.19489003717899323, "learning_rate": 4.669213652128667e-05, "loss": 0.0716, "step": 1125 }, { "epoch": 0.5287168089835068, "grad_norm": 0.16565723717212677, "learning_rate": 4.664022221343892e-05, "loss": 0.0717, "step": 1130 }, { "epoch": 0.5310562638905135, "grad_norm": 0.18749143183231354, "learning_rate": 4.658793302146489e-05, "loss": 0.0756, "step": 1135 }, { "epoch": 0.5333957187975201, "grad_norm": 0.2000792771577835, "learning_rate": 4.653526985119577e-05, "loss": 0.0789, "step": 1140 }, { "epoch": 0.5357351737045268, "grad_norm": 0.17934930324554443, "learning_rate": 4.648223361494135e-05, "loss": 0.0823, "step": 1145 }, { "epoch": 0.5380746286115335, "grad_norm": 0.204596146941185, "learning_rate": 4.642882523147422e-05, "loss": 0.0804, "step": 1150 }, { "epoch": 0.5404140835185401, "grad_norm": 0.17674653232097626, "learning_rate": 4.637504562601386e-05, "loss": 0.0699, "step": 1155 }, { "epoch": 0.5427535384255469, "grad_norm": 0.17949679493904114, "learning_rate": 4.6320895730210616e-05, "loss": 0.0717, "step": 1160 }, { "epoch": 0.5450929933325536, "grad_norm": 0.1828782856464386, "learning_rate": 4.626637648212955e-05, "loss": 0.0816, "step": 1165 }, { "epoch": 0.5474324482395602, "grad_norm": 0.1803893744945526, "learning_rate": 4.6211488826234186e-05, "loss": 0.0766, "step": 1170 }, { "epoch": 0.5497719031465669, "grad_norm": 0.1853674352169037, "learning_rate": 4.615623371337016e-05, "loss": 0.0728, "step": 1175 }, { "epoch": 0.5521113580535735, "grad_norm": 0.20650415122509003, "learning_rate": 4.6100612100748765e-05, "loss": 0.073, "step": 1180 }, { "epoch": 0.5544508129605802, "grad_norm": 0.18589161336421967, "learning_rate": 4.604462495193031e-05, "loss": 0.0803, "step": 1185 }, { "epoch": 0.5567902678675869, "grad_norm": 0.16948671638965607, "learning_rate": 4.59882732368075e-05, "loss": 0.0728, "step": 1190 }, { "epoch": 0.5591297227745935, "grad_norm": 0.17474383115768433, "learning_rate": 4.593155793158859e-05, "loss": 0.0697, "step": 1195 }, { "epoch": 0.5614691776816002, "grad_norm": 0.1744118481874466, "learning_rate": 4.5874480018780456e-05, "loss": 0.0771, "step": 1200 }, { "epoch": 0.5638086325886068, "grad_norm": 0.17954926192760468, "learning_rate": 4.581704048717166e-05, "loss": 0.0722, "step": 1205 }, { "epoch": 0.5661480874956135, "grad_norm": 0.19267572462558746, "learning_rate": 4.57592403318152e-05, "loss": 0.0739, "step": 1210 }, { "epoch": 0.5684875424026202, "grad_norm": 0.1731371432542801, "learning_rate": 4.570108055401138e-05, "loss": 0.076, "step": 1215 }, { "epoch": 0.5708269973096268, "grad_norm": 0.18417227268218994, "learning_rate": 4.5642562161290406e-05, "loss": 0.0708, "step": 1220 }, { "epoch": 0.5731664522166335, "grad_norm": 0.19248449802398682, "learning_rate": 4.558368616739493e-05, "loss": 0.0775, "step": 1225 }, { "epoch": 0.5755059071236401, "grad_norm": 0.20043963193893433, "learning_rate": 4.552445359226252e-05, "loss": 0.0758, "step": 1230 }, { "epoch": 0.5778453620306468, "grad_norm": 0.20923319458961487, "learning_rate": 4.546486546200798e-05, "loss": 0.0648, "step": 1235 }, { "epoch": 0.5801848169376536, "grad_norm": 0.19847044348716736, "learning_rate": 4.540492280890555e-05, "loss": 0.0684, "step": 1240 }, { "epoch": 0.5825242718446602, "grad_norm": 0.17543412744998932, "learning_rate": 4.534462667137105e-05, "loss": 0.0746, "step": 1245 }, { "epoch": 0.5848637267516669, "grad_norm": 0.1842990517616272, "learning_rate": 4.528397809394389e-05, "loss": 0.0771, "step": 1250 }, { "epoch": 0.5872031816586736, "grad_norm": 0.15888415277004242, "learning_rate": 4.522297812726897e-05, "loss": 0.0654, "step": 1255 }, { "epoch": 0.5895426365656802, "grad_norm": 0.20020891726016998, "learning_rate": 4.516162782807845e-05, "loss": 0.0759, "step": 1260 }, { "epoch": 0.5918820914726869, "grad_norm": 0.164358988404274, "learning_rate": 4.5099928259173516e-05, "loss": 0.0666, "step": 1265 }, { "epoch": 0.5942215463796935, "grad_norm": 0.16114068031311035, "learning_rate": 4.503788048940589e-05, "loss": 0.0677, "step": 1270 }, { "epoch": 0.5965610012867002, "grad_norm": 0.18885543942451477, "learning_rate": 4.497548559365935e-05, "loss": 0.0697, "step": 1275 }, { "epoch": 0.5989004561937069, "grad_norm": 0.16143983602523804, "learning_rate": 4.4912744652831116e-05, "loss": 0.072, "step": 1280 }, { "epoch": 0.6012399111007135, "grad_norm": 0.16090819239616394, "learning_rate": 4.4849658753813126e-05, "loss": 0.068, "step": 1285 }, { "epoch": 0.6035793660077202, "grad_norm": 0.19243429601192474, "learning_rate": 4.4786228989473164e-05, "loss": 0.0744, "step": 1290 }, { "epoch": 0.6059188209147268, "grad_norm": 0.17546018958091736, "learning_rate": 4.4722456458635995e-05, "loss": 0.0745, "step": 1295 }, { "epoch": 0.6082582758217335, "grad_norm": 0.1936197578907013, "learning_rate": 4.465834226606428e-05, "loss": 0.0769, "step": 1300 }, { "epoch": 0.6105977307287402, "grad_norm": 0.16884423792362213, "learning_rate": 4.459388752243945e-05, "loss": 0.0809, "step": 1305 }, { "epoch": 0.6129371856357468, "grad_norm": 0.20882828533649445, "learning_rate": 4.452909334434247e-05, "loss": 0.08, "step": 1310 }, { "epoch": 0.6152766405427536, "grad_norm": 0.1857665777206421, "learning_rate": 4.4463960854234506e-05, "loss": 0.0668, "step": 1315 }, { "epoch": 0.6176160954497603, "grad_norm": 0.17508915066719055, "learning_rate": 4.4398491180437446e-05, "loss": 0.0775, "step": 1320 }, { "epoch": 0.6199555503567669, "grad_norm": 0.16894185543060303, "learning_rate": 4.43326854571144e-05, "loss": 0.0714, "step": 1325 }, { "epoch": 0.6222950052637736, "grad_norm": 0.19014866650104523, "learning_rate": 4.426654482425e-05, "loss": 0.0789, "step": 1330 }, { "epoch": 0.6246344601707802, "grad_norm": 0.15668885409832, "learning_rate": 4.420007042763071e-05, "loss": 0.0641, "step": 1335 }, { "epoch": 0.6269739150777869, "grad_norm": 0.15578554570674896, "learning_rate": 4.413326341882492e-05, "loss": 0.0665, "step": 1340 }, { "epoch": 0.6293133699847936, "grad_norm": 0.20437653362751007, "learning_rate": 4.4066124955163046e-05, "loss": 0.0784, "step": 1345 }, { "epoch": 0.6316528248918002, "grad_norm": 0.16404947638511658, "learning_rate": 4.3998656199717435e-05, "loss": 0.0675, "step": 1350 }, { "epoch": 0.6339922797988069, "grad_norm": 0.147497296333313, "learning_rate": 4.393085832128226e-05, "loss": 0.0711, "step": 1355 }, { "epoch": 0.6363317347058135, "grad_norm": 0.17336557805538177, "learning_rate": 4.386273249435324e-05, "loss": 0.0766, "step": 1360 }, { "epoch": 0.6386711896128202, "grad_norm": 0.18283897638320923, "learning_rate": 4.3794279899107304e-05, "loss": 0.0746, "step": 1365 }, { "epoch": 0.6410106445198269, "grad_norm": 0.16502517461776733, "learning_rate": 4.3725501721382165e-05, "loss": 0.0638, "step": 1370 }, { "epoch": 0.6433500994268335, "grad_norm": 0.16824820637702942, "learning_rate": 4.3656399152655735e-05, "loss": 0.0698, "step": 1375 }, { "epoch": 0.6456895543338402, "grad_norm": 0.17634496092796326, "learning_rate": 4.358697339002553e-05, "loss": 0.0952, "step": 1380 }, { "epoch": 0.6480290092408468, "grad_norm": 0.1962941437959671, "learning_rate": 4.3517225636187906e-05, "loss": 0.0728, "step": 1385 }, { "epoch": 0.6503684641478535, "grad_norm": 0.1725791096687317, "learning_rate": 4.344715709941722e-05, "loss": 0.0721, "step": 1390 }, { "epoch": 0.6527079190548603, "grad_norm": 0.1703636348247528, "learning_rate": 4.337676899354493e-05, "loss": 0.072, "step": 1395 }, { "epoch": 0.6550473739618669, "grad_norm": 0.16488994657993317, "learning_rate": 4.330606253793851e-05, "loss": 0.072, "step": 1400 }, { "epoch": 0.6573868288688736, "grad_norm": 0.16995395720005035, "learning_rate": 4.32350389574804e-05, "loss": 0.0747, "step": 1405 }, { "epoch": 0.6597262837758803, "grad_norm": 0.17242836952209473, "learning_rate": 4.316369948254674e-05, "loss": 0.0698, "step": 1410 }, { "epoch": 0.6620657386828869, "grad_norm": 0.16680659353733063, "learning_rate": 4.3092045348986034e-05, "loss": 0.0683, "step": 1415 }, { "epoch": 0.6644051935898936, "grad_norm": 0.18627899885177612, "learning_rate": 4.302007779809781e-05, "loss": 0.0672, "step": 1420 }, { "epoch": 0.6667446484969002, "grad_norm": 0.14819131791591644, "learning_rate": 4.294779807661105e-05, "loss": 0.0683, "step": 1425 }, { "epoch": 0.6690841034039069, "grad_norm": 0.18117517232894897, "learning_rate": 4.287520743666263e-05, "loss": 0.0728, "step": 1430 }, { "epoch": 0.6714235583109136, "grad_norm": 0.17058628797531128, "learning_rate": 4.280230713577564e-05, "loss": 0.077, "step": 1435 }, { "epoch": 0.6737630132179202, "grad_norm": 0.15503695607185364, "learning_rate": 4.2729098436837536e-05, "loss": 0.0739, "step": 1440 }, { "epoch": 0.6761024681249269, "grad_norm": 0.1836390346288681, "learning_rate": 4.2655582608078315e-05, "loss": 0.079, "step": 1445 }, { "epoch": 0.6784419230319335, "grad_norm": 0.15766695141792297, "learning_rate": 4.2581760923048554e-05, "loss": 0.0653, "step": 1450 }, { "epoch": 0.6807813779389402, "grad_norm": 0.18035438656806946, "learning_rate": 4.2507634660597315e-05, "loss": 0.0793, "step": 1455 }, { "epoch": 0.6831208328459469, "grad_norm": 0.1778097152709961, "learning_rate": 4.243320510485001e-05, "loss": 0.0689, "step": 1460 }, { "epoch": 0.6854602877529535, "grad_norm": 0.1773742139339447, "learning_rate": 4.235847354518614e-05, "loss": 0.0726, "step": 1465 }, { "epoch": 0.6877997426599602, "grad_norm": 0.16845591366291046, "learning_rate": 4.228344127621695e-05, "loss": 0.0691, "step": 1470 }, { "epoch": 0.6901391975669668, "grad_norm": 0.14897610247135162, "learning_rate": 4.220810959776307e-05, "loss": 0.0684, "step": 1475 }, { "epoch": 0.6924786524739736, "grad_norm": 0.17343005537986755, "learning_rate": 4.213247981483189e-05, "loss": 0.0624, "step": 1480 }, { "epoch": 0.6948181073809803, "grad_norm": 0.19513456523418427, "learning_rate": 4.205655323759505e-05, "loss": 0.0751, "step": 1485 }, { "epoch": 0.6971575622879869, "grad_norm": 0.17424218356609344, "learning_rate": 4.1980331181365685e-05, "loss": 0.0704, "step": 1490 }, { "epoch": 0.6994970171949936, "grad_norm": 0.1491064429283142, "learning_rate": 4.190381496657565e-05, "loss": 0.0717, "step": 1495 }, { "epoch": 0.7018364721020003, "grad_norm": 0.1655229777097702, "learning_rate": 4.182700591875267e-05, "loss": 0.071, "step": 1500 }, { "epoch": 0.7018364721020003, "eval_loss": 0.0841016098856926, "eval_runtime": 237.2254, "eval_samples_per_second": 2.942, "eval_steps_per_second": 0.738, "step": 1500 }, { "epoch": 0.7041759270090069, "grad_norm": 0.18065688014030457, "learning_rate": 4.1749905368497345e-05, "loss": 0.0733, "step": 1505 }, { "epoch": 0.7065153819160136, "grad_norm": 0.16063852608203888, "learning_rate": 4.1672514651460124e-05, "loss": 0.0646, "step": 1510 }, { "epoch": 0.7088548368230202, "grad_norm": 0.18853044509887695, "learning_rate": 4.159483510831814e-05, "loss": 0.0623, "step": 1515 }, { "epoch": 0.7111942917300269, "grad_norm": 0.15105347335338593, "learning_rate": 4.151686808475204e-05, "loss": 0.0638, "step": 1520 }, { "epoch": 0.7135337466370336, "grad_norm": 0.17402097582817078, "learning_rate": 4.143861493142258e-05, "loss": 0.0714, "step": 1525 }, { "epoch": 0.7158732015440402, "grad_norm": 0.17532867193222046, "learning_rate": 4.136007700394733e-05, "loss": 0.0732, "step": 1530 }, { "epoch": 0.7182126564510469, "grad_norm": 0.16897249221801758, "learning_rate": 4.128125566287711e-05, "loss": 0.0705, "step": 1535 }, { "epoch": 0.7205521113580535, "grad_norm": 0.19206973910331726, "learning_rate": 4.120215227367247e-05, "loss": 0.0722, "step": 1540 }, { "epoch": 0.7228915662650602, "grad_norm": 0.1461183875799179, "learning_rate": 4.112276820668001e-05, "loss": 0.0654, "step": 1545 }, { "epoch": 0.725231021172067, "grad_norm": 0.14818954467773438, "learning_rate": 4.104310483710864e-05, "loss": 0.0664, "step": 1550 }, { "epoch": 0.7275704760790735, "grad_norm": 0.1520642787218094, "learning_rate": 4.096316354500578e-05, "loss": 0.0647, "step": 1555 }, { "epoch": 0.7299099309860803, "grad_norm": 0.18049833178520203, "learning_rate": 4.0882945715233426e-05, "loss": 0.0647, "step": 1560 }, { "epoch": 0.7322493858930869, "grad_norm": 0.14968477189540863, "learning_rate": 4.080245273744419e-05, "loss": 0.0681, "step": 1565 }, { "epoch": 0.7345888408000936, "grad_norm": 0.17562703788280487, "learning_rate": 4.072168600605718e-05, "loss": 0.0682, "step": 1570 }, { "epoch": 0.7369282957071003, "grad_norm": 0.14080215990543365, "learning_rate": 4.064064692023389e-05, "loss": 0.0626, "step": 1575 }, { "epoch": 0.7392677506141069, "grad_norm": 0.16380788385868073, "learning_rate": 4.055933688385394e-05, "loss": 0.1316, "step": 1580 }, { "epoch": 0.7416072055211136, "grad_norm": 0.15547692775726318, "learning_rate": 4.047775730549075e-05, "loss": 0.08, "step": 1585 }, { "epoch": 0.7439466604281203, "grad_norm": 0.15663617849349976, "learning_rate": 4.039590959838715e-05, "loss": 0.0722, "step": 1590 }, { "epoch": 0.7462861153351269, "grad_norm": 0.16741621494293213, "learning_rate": 4.031379518043091e-05, "loss": 0.0708, "step": 1595 }, { "epoch": 0.7486255702421336, "grad_norm": 0.15329977869987488, "learning_rate": 4.023141547413015e-05, "loss": 0.063, "step": 1600 }, { "epoch": 0.7509650251491402, "grad_norm": 0.16843506693840027, "learning_rate": 4.0148771906588706e-05, "loss": 0.0766, "step": 1605 }, { "epoch": 0.7533044800561469, "grad_norm": 0.17443476617336273, "learning_rate": 4.0065865909481417e-05, "loss": 0.0663, "step": 1610 }, { "epoch": 0.7556439349631536, "grad_norm": 0.14947757124900818, "learning_rate": 3.9982698919029305e-05, "loss": 0.0804, "step": 1615 }, { "epoch": 0.7579833898701602, "grad_norm": 0.16239669919013977, "learning_rate": 3.9899272375974726e-05, "loss": 0.076, "step": 1620 }, { "epoch": 0.7603228447771669, "grad_norm": 0.15543022751808167, "learning_rate": 3.981558772555638e-05, "loss": 0.0697, "step": 1625 }, { "epoch": 0.7626622996841735, "grad_norm": 0.16913045942783356, "learning_rate": 3.97316464174843e-05, "loss": 0.0752, "step": 1630 }, { "epoch": 0.7650017545911802, "grad_norm": 0.16426704823970795, "learning_rate": 3.964744990591471e-05, "loss": 0.0827, "step": 1635 }, { "epoch": 0.767341209498187, "grad_norm": 0.1630501002073288, "learning_rate": 3.956299964942485e-05, "loss": 0.0716, "step": 1640 }, { "epoch": 0.7696806644051936, "grad_norm": 0.15198907256126404, "learning_rate": 3.947829711098772e-05, "loss": 0.0763, "step": 1645 }, { "epoch": 0.7720201193122003, "grad_norm": 0.17114469408988953, "learning_rate": 3.9393343757946696e-05, "loss": 0.0662, "step": 1650 }, { "epoch": 0.7743595742192069, "grad_norm": 0.14625594019889832, "learning_rate": 3.930814106199018e-05, "loss": 0.0721, "step": 1655 }, { "epoch": 0.7766990291262136, "grad_norm": 0.16365419328212738, "learning_rate": 3.922269049912602e-05, "loss": 0.0755, "step": 1660 }, { "epoch": 0.7790384840332203, "grad_norm": 0.19672849774360657, "learning_rate": 3.9136993549656006e-05, "loss": 0.0721, "step": 1665 }, { "epoch": 0.7813779389402269, "grad_norm": 0.17917795479297638, "learning_rate": 3.905105169815021e-05, "loss": 0.0638, "step": 1670 }, { "epoch": 0.7837173938472336, "grad_norm": 0.15261490643024445, "learning_rate": 3.896486643342124e-05, "loss": 0.0779, "step": 1675 }, { "epoch": 0.7860568487542403, "grad_norm": 0.1785430610179901, "learning_rate": 3.887843924849849e-05, "loss": 0.0715, "step": 1680 }, { "epoch": 0.7883963036612469, "grad_norm": 0.15738226473331451, "learning_rate": 3.879177164060225e-05, "loss": 0.0719, "step": 1685 }, { "epoch": 0.7907357585682536, "grad_norm": 0.17169606685638428, "learning_rate": 3.8704865111117746e-05, "loss": 0.0661, "step": 1690 }, { "epoch": 0.7930752134752602, "grad_norm": 0.15554682910442352, "learning_rate": 3.861772116556921e-05, "loss": 0.0715, "step": 1695 }, { "epoch": 0.7954146683822669, "grad_norm": 0.2050098329782486, "learning_rate": 3.853034131359371e-05, "loss": 0.0726, "step": 1700 }, { "epoch": 0.7977541232892736, "grad_norm": 0.14537398517131805, "learning_rate": 3.8442727068915066e-05, "loss": 0.0736, "step": 1705 }, { "epoch": 0.8000935781962802, "grad_norm": 0.16626664996147156, "learning_rate": 3.8354879949317546e-05, "loss": 0.0641, "step": 1710 }, { "epoch": 0.802433033103287, "grad_norm": 0.17925380170345306, "learning_rate": 3.8266801476619694e-05, "loss": 0.07, "step": 1715 }, { "epoch": 0.8047724880102936, "grad_norm": 0.1543145477771759, "learning_rate": 3.817849317664784e-05, "loss": 0.069, "step": 1720 }, { "epoch": 0.8071119429173003, "grad_norm": 0.15055689215660095, "learning_rate": 3.808995657920975e-05, "loss": 0.0686, "step": 1725 }, { "epoch": 0.809451397824307, "grad_norm": 0.13166533410549164, "learning_rate": 3.800119321806809e-05, "loss": 0.0643, "step": 1730 }, { "epoch": 0.8117908527313136, "grad_norm": 0.20518459379673004, "learning_rate": 3.791220463091387e-05, "loss": 0.0846, "step": 1735 }, { "epoch": 0.8141303076383203, "grad_norm": 0.13176603615283966, "learning_rate": 3.7822992359339794e-05, "loss": 0.0697, "step": 1740 }, { "epoch": 0.816469762545327, "grad_norm": 0.16138117015361786, "learning_rate": 3.7733557948813575e-05, "loss": 0.0679, "step": 1745 }, { "epoch": 0.8188092174523336, "grad_norm": 0.20597168803215027, "learning_rate": 3.764390294865112e-05, "loss": 0.0779, "step": 1750 }, { "epoch": 0.8211486723593403, "grad_norm": 0.18723969161510468, "learning_rate": 3.755402891198974e-05, "loss": 0.141, "step": 1755 }, { "epoch": 0.8234881272663469, "grad_norm": 0.16847847402095795, "learning_rate": 3.746393739576121e-05, "loss": 0.0675, "step": 1760 }, { "epoch": 0.8258275821733536, "grad_norm": 0.1512673944234848, "learning_rate": 3.737362996066483e-05, "loss": 0.0708, "step": 1765 }, { "epoch": 0.8281670370803603, "grad_norm": 0.17795756459236145, "learning_rate": 3.728310817114034e-05, "loss": 0.0675, "step": 1770 }, { "epoch": 0.8305064919873669, "grad_norm": 0.18552307784557343, "learning_rate": 3.719237359534087e-05, "loss": 0.0758, "step": 1775 }, { "epoch": 0.8328459468943736, "grad_norm": 0.16442035138607025, "learning_rate": 3.710142780510573e-05, "loss": 0.0728, "step": 1780 }, { "epoch": 0.8351854018013802, "grad_norm": 0.16192203760147095, "learning_rate": 3.7010272375933216e-05, "loss": 0.0651, "step": 1785 }, { "epoch": 0.8375248567083869, "grad_norm": 0.15759891271591187, "learning_rate": 3.691890888695329e-05, "loss": 0.0659, "step": 1790 }, { "epoch": 0.8398643116153937, "grad_norm": 0.16758672893047333, "learning_rate": 3.6827338920900254e-05, "loss": 0.065, "step": 1795 }, { "epoch": 0.8422037665224003, "grad_norm": 0.1629868447780609, "learning_rate": 3.6735564064085296e-05, "loss": 0.0748, "step": 1800 }, { "epoch": 0.844543221429407, "grad_norm": 0.1754574477672577, "learning_rate": 3.664358590636903e-05, "loss": 0.0857, "step": 1805 }, { "epoch": 0.8468826763364136, "grad_norm": 0.16642498970031738, "learning_rate": 3.655140604113395e-05, "loss": 0.0721, "step": 1810 }, { "epoch": 0.8492221312434203, "grad_norm": 0.15343014895915985, "learning_rate": 3.645902606525683e-05, "loss": 0.0653, "step": 1815 }, { "epoch": 0.851561586150427, "grad_norm": 0.15997187793254852, "learning_rate": 3.636644757908107e-05, "loss": 0.0752, "step": 1820 }, { "epoch": 0.8539010410574336, "grad_norm": 0.1555909514427185, "learning_rate": 3.627367218638893e-05, "loss": 0.0692, "step": 1825 }, { "epoch": 0.8562404959644403, "grad_norm": 0.1649237424135208, "learning_rate": 3.618070149437381e-05, "loss": 0.0706, "step": 1830 }, { "epoch": 0.858579950871447, "grad_norm": 0.14074915647506714, "learning_rate": 3.6087537113612364e-05, "loss": 0.0612, "step": 1835 }, { "epoch": 0.8609194057784536, "grad_norm": 0.15589754283428192, "learning_rate": 3.59941806580366e-05, "loss": 0.0724, "step": 1840 }, { "epoch": 0.8632588606854603, "grad_norm": 0.1727776825428009, "learning_rate": 3.590063374490595e-05, "loss": 0.0851, "step": 1845 }, { "epoch": 0.8655983155924669, "grad_norm": 0.14394600689411163, "learning_rate": 3.580689799477921e-05, "loss": 0.0669, "step": 1850 }, { "epoch": 0.8679377704994736, "grad_norm": 0.15605291724205017, "learning_rate": 3.5712975031486525e-05, "loss": 0.0716, "step": 1855 }, { "epoch": 0.8702772254064803, "grad_norm": 0.19002071022987366, "learning_rate": 3.56188664821012e-05, "loss": 0.0748, "step": 1860 }, { "epoch": 0.8726166803134869, "grad_norm": 0.16365236043930054, "learning_rate": 3.5524573976911546e-05, "loss": 0.079, "step": 1865 }, { "epoch": 0.8749561352204936, "grad_norm": 0.19456696510314941, "learning_rate": 3.543009914939261e-05, "loss": 0.0747, "step": 1870 }, { "epoch": 0.8772955901275002, "grad_norm": 0.15589672327041626, "learning_rate": 3.533544363617796e-05, "loss": 0.0657, "step": 1875 }, { "epoch": 0.879635045034507, "grad_norm": 0.14598876237869263, "learning_rate": 3.5240609077031196e-05, "loss": 0.0642, "step": 1880 }, { "epoch": 0.8819744999415137, "grad_norm": 0.15636853873729706, "learning_rate": 3.514559711481765e-05, "loss": 0.0608, "step": 1885 }, { "epoch": 0.8843139548485203, "grad_norm": 0.1746116727590561, "learning_rate": 3.505040939547591e-05, "loss": 0.077, "step": 1890 }, { "epoch": 0.886653409755527, "grad_norm": 0.15822593867778778, "learning_rate": 3.495504756798926e-05, "loss": 0.0788, "step": 1895 }, { "epoch": 0.8889928646625336, "grad_norm": 0.16369085013866425, "learning_rate": 3.485951328435716e-05, "loss": 0.0764, "step": 1900 }, { "epoch": 0.8913323195695403, "grad_norm": 0.1527598798274994, "learning_rate": 3.476380819956662e-05, "loss": 0.0722, "step": 1905 }, { "epoch": 0.893671774476547, "grad_norm": 0.13147583603858948, "learning_rate": 3.466793397156351e-05, "loss": 0.0694, "step": 1910 }, { "epoch": 0.8960112293835536, "grad_norm": 0.15432552993297577, "learning_rate": 3.457189226122384e-05, "loss": 0.0652, "step": 1915 }, { "epoch": 0.8983506842905603, "grad_norm": 0.15792711079120636, "learning_rate": 3.4475684732325024e-05, "loss": 0.0763, "step": 1920 }, { "epoch": 0.900690139197567, "grad_norm": 0.14269225299358368, "learning_rate": 3.437931305151701e-05, "loss": 0.0739, "step": 1925 }, { "epoch": 0.9030295941045736, "grad_norm": 0.18912151455879211, "learning_rate": 3.428277888829344e-05, "loss": 0.0839, "step": 1930 }, { "epoch": 0.9053690490115803, "grad_norm": 0.16357168555259705, "learning_rate": 3.418608391496273e-05, "loss": 0.0695, "step": 1935 }, { "epoch": 0.9077085039185869, "grad_norm": 0.1561209261417389, "learning_rate": 3.408922980661905e-05, "loss": 0.0681, "step": 1940 }, { "epoch": 0.9100479588255936, "grad_norm": 0.17604918777942657, "learning_rate": 3.39922182411134e-05, "loss": 0.0694, "step": 1945 }, { "epoch": 0.9123874137326003, "grad_norm": 0.16690443456172943, "learning_rate": 3.3895050899024416e-05, "loss": 0.1077, "step": 1950 }, { "epoch": 0.914726868639607, "grad_norm": 0.15058398246765137, "learning_rate": 3.379772946362939e-05, "loss": 0.0604, "step": 1955 }, { "epoch": 0.9170663235466137, "grad_norm": 0.12406554818153381, "learning_rate": 3.3700255620875025e-05, "loss": 0.061, "step": 1960 }, { "epoch": 0.9194057784536203, "grad_norm": 0.15860305726528168, "learning_rate": 3.3602631059348245e-05, "loss": 0.0714, "step": 1965 }, { "epoch": 0.921745233360627, "grad_norm": 0.15389299392700195, "learning_rate": 3.350485747024695e-05, "loss": 0.0724, "step": 1970 }, { "epoch": 0.9240846882676337, "grad_norm": 0.16053031384944916, "learning_rate": 3.340693654735073e-05, "loss": 0.064, "step": 1975 }, { "epoch": 0.9264241431746403, "grad_norm": 0.14700767397880554, "learning_rate": 3.330886998699149e-05, "loss": 0.0639, "step": 1980 }, { "epoch": 0.928763598081647, "grad_norm": 0.15613354742527008, "learning_rate": 3.321065948802411e-05, "loss": 0.0667, "step": 1985 }, { "epoch": 0.9311030529886536, "grad_norm": 0.15523682534694672, "learning_rate": 3.311230675179697e-05, "loss": 0.0634, "step": 1990 }, { "epoch": 0.9334425078956603, "grad_norm": 0.14495091140270233, "learning_rate": 3.301381348212249e-05, "loss": 0.0632, "step": 1995 }, { "epoch": 0.935781962802667, "grad_norm": 0.17717325687408447, "learning_rate": 3.291518138524764e-05, "loss": 0.0678, "step": 2000 }, { "epoch": 0.935781962802667, "eval_loss": 0.08131425082683563, "eval_runtime": 232.0604, "eval_samples_per_second": 3.008, "eval_steps_per_second": 0.754, "step": 2000 }, { "epoch": 0.9381214177096736, "grad_norm": 0.15504471957683563, "learning_rate": 3.281641216982435e-05, "loss": 0.1141, "step": 2005 }, { "epoch": 0.9404608726166803, "grad_norm": 0.15814577043056488, "learning_rate": 3.271750754687991e-05, "loss": 0.0658, "step": 2010 }, { "epoch": 0.942800327523687, "grad_norm": 0.18318745493888855, "learning_rate": 3.261846922978736e-05, "loss": 0.0727, "step": 2015 }, { "epoch": 0.9451397824306936, "grad_norm": 0.16373957693576813, "learning_rate": 3.2519298934235796e-05, "loss": 0.0683, "step": 2020 }, { "epoch": 0.9474792373377003, "grad_norm": 0.1619662046432495, "learning_rate": 3.24199983782006e-05, "loss": 0.0663, "step": 2025 }, { "epoch": 0.9498186922447069, "grad_norm": 0.15787597000598907, "learning_rate": 3.232056928191376e-05, "loss": 0.0743, "step": 2030 }, { "epoch": 0.9521581471517137, "grad_norm": 0.18520085513591766, "learning_rate": 3.2221013367834016e-05, "loss": 0.0994, "step": 2035 }, { "epoch": 0.9544976020587204, "grad_norm": 0.1397067755460739, "learning_rate": 3.212133236061704e-05, "loss": 0.0835, "step": 2040 }, { "epoch": 0.956837056965727, "grad_norm": 0.1584145724773407, "learning_rate": 3.202152798708554e-05, "loss": 0.0843, "step": 2045 }, { "epoch": 0.9591765118727337, "grad_norm": 0.14914660155773163, "learning_rate": 3.1921601976199365e-05, "loss": 0.0744, "step": 2050 }, { "epoch": 0.9615159667797403, "grad_norm": 0.14191728830337524, "learning_rate": 3.182155605902556e-05, "loss": 0.0654, "step": 2055 }, { "epoch": 0.963855421686747, "grad_norm": 0.15701289474964142, "learning_rate": 3.1721391968708335e-05, "loss": 0.0665, "step": 2060 }, { "epoch": 0.9661948765937537, "grad_norm": 0.14742569625377655, "learning_rate": 3.162111144043911e-05, "loss": 0.072, "step": 2065 }, { "epoch": 0.9685343315007603, "grad_norm": 0.15007531642913818, "learning_rate": 3.1520716211426385e-05, "loss": 0.0754, "step": 2070 }, { "epoch": 0.970873786407767, "grad_norm": 0.1743524670600891, "learning_rate": 3.142020802086569e-05, "loss": 0.0739, "step": 2075 }, { "epoch": 0.9732132413147736, "grad_norm": 0.12957176566123962, "learning_rate": 3.131958860990945e-05, "loss": 0.0617, "step": 2080 }, { "epoch": 0.9755526962217803, "grad_norm": 0.15021301805973053, "learning_rate": 3.121885972163681e-05, "loss": 0.0785, "step": 2085 }, { "epoch": 0.977892151128787, "grad_norm": 0.18265359103679657, "learning_rate": 3.111802310102341e-05, "loss": 0.0675, "step": 2090 }, { "epoch": 0.9802316060357936, "grad_norm": 0.14601291716098785, "learning_rate": 3.101708049491125e-05, "loss": 0.0694, "step": 2095 }, { "epoch": 0.9825710609428003, "grad_norm": 0.1612151712179184, "learning_rate": 3.091603365197833e-05, "loss": 0.0743, "step": 2100 }, { "epoch": 0.984910515849807, "grad_norm": 0.14992724359035492, "learning_rate": 3.081488432270841e-05, "loss": 0.0664, "step": 2105 }, { "epoch": 0.9872499707568136, "grad_norm": 0.14685116708278656, "learning_rate": 3.071363425936066e-05, "loss": 0.0762, "step": 2110 }, { "epoch": 0.9895894256638204, "grad_norm": 0.14769916236400604, "learning_rate": 3.061228521593931e-05, "loss": 0.0636, "step": 2115 }, { "epoch": 0.991928880570827, "grad_norm": 0.14650078117847443, "learning_rate": 3.0510838948163308e-05, "loss": 0.0836, "step": 2120 }, { "epoch": 0.9942683354778337, "grad_norm": 0.1642710417509079, "learning_rate": 3.0409297213435818e-05, "loss": 0.0723, "step": 2125 }, { "epoch": 0.9966077903848404, "grad_norm": 0.13404929637908936, "learning_rate": 3.030766177081385e-05, "loss": 0.0766, "step": 2130 }, { "epoch": 0.998947245291847, "grad_norm": 0.1478378027677536, "learning_rate": 3.020593438097776e-05, "loss": 0.0619, "step": 2135 }, { "epoch": 1.0012867001988537, "grad_norm": 0.1335192173719406, "learning_rate": 3.010411680620075e-05, "loss": 0.0641, "step": 2140 }, { "epoch": 1.0036261551058603, "grad_norm": 0.1439552754163742, "learning_rate": 3.0002210810318338e-05, "loss": 0.0735, "step": 2145 }, { "epoch": 1.005965610012867, "grad_norm": 0.14612635970115662, "learning_rate": 2.9900218158697797e-05, "loss": 0.0576, "step": 2150 }, { "epoch": 1.0083050649198737, "grad_norm": 0.17644114792346954, "learning_rate": 2.9798140618207598e-05, "loss": 0.0721, "step": 2155 }, { "epoch": 1.0106445198268803, "grad_norm": 0.15315324068069458, "learning_rate": 2.9695979957186764e-05, "loss": 0.0642, "step": 2160 }, { "epoch": 1.012983974733887, "grad_norm": 0.15840911865234375, "learning_rate": 2.9593737945414264e-05, "loss": 0.0751, "step": 2165 }, { "epoch": 1.0153234296408937, "grad_norm": 0.14203301072120667, "learning_rate": 2.9491416354078343e-05, "loss": 0.0619, "step": 2170 }, { "epoch": 1.0176628845479003, "grad_norm": 0.17222335934638977, "learning_rate": 2.938901695574585e-05, "loss": 0.0719, "step": 2175 }, { "epoch": 1.020002339454907, "grad_norm": 0.16653592884540558, "learning_rate": 2.9286541524331525e-05, "loss": 0.0597, "step": 2180 }, { "epoch": 1.0223417943619137, "grad_norm": 0.12632368505001068, "learning_rate": 2.9183991835067237e-05, "loss": 0.0585, "step": 2185 }, { "epoch": 1.0246812492689203, "grad_norm": 0.15063339471817017, "learning_rate": 2.9081369664471293e-05, "loss": 0.0712, "step": 2190 }, { "epoch": 1.027020704175927, "grad_norm": 0.15402260422706604, "learning_rate": 2.8978676790317605e-05, "loss": 0.0631, "step": 2195 }, { "epoch": 1.0293601590829338, "grad_norm": 0.17674040794372559, "learning_rate": 2.8875914991604948e-05, "loss": 0.0611, "step": 2200 }, { "epoch": 1.0316996139899404, "grad_norm": 0.15764425694942474, "learning_rate": 2.8773086048526076e-05, "loss": 0.1481, "step": 2205 }, { "epoch": 1.034039068896947, "grad_norm": 0.17565783858299255, "learning_rate": 2.8670191742436946e-05, "loss": 0.0623, "step": 2210 }, { "epoch": 1.0363785238039536, "grad_norm": 0.1751687377691269, "learning_rate": 2.8567233855825813e-05, "loss": 0.0599, "step": 2215 }, { "epoch": 1.0387179787109604, "grad_norm": 0.15617790818214417, "learning_rate": 2.8464214172282367e-05, "loss": 0.0593, "step": 2220 }, { "epoch": 1.041057433617967, "grad_norm": 0.1573198437690735, "learning_rate": 2.8361134476466843e-05, "loss": 0.0643, "step": 2225 }, { "epoch": 1.0433968885249736, "grad_norm": 0.15318118035793304, "learning_rate": 2.8257996554079113e-05, "loss": 0.0582, "step": 2230 }, { "epoch": 1.0457363434319804, "grad_norm": 0.14526399970054626, "learning_rate": 2.8154802191827706e-05, "loss": 0.0577, "step": 2235 }, { "epoch": 1.048075798338987, "grad_norm": 0.1734457015991211, "learning_rate": 2.805155317739891e-05, "loss": 0.0618, "step": 2240 }, { "epoch": 1.0504152532459936, "grad_norm": 0.16416044533252716, "learning_rate": 2.7948251299425758e-05, "loss": 0.067, "step": 2245 }, { "epoch": 1.0527547081530004, "grad_norm": 0.14763504266738892, "learning_rate": 2.784489834745709e-05, "loss": 0.0626, "step": 2250 }, { "epoch": 1.055094163060007, "grad_norm": 0.15287157893180847, "learning_rate": 2.77414961119265e-05, "loss": 0.0627, "step": 2255 }, { "epoch": 1.0574336179670136, "grad_norm": 0.17158065736293793, "learning_rate": 2.763804638412137e-05, "loss": 0.0653, "step": 2260 }, { "epoch": 1.0597730728740204, "grad_norm": 0.1766432225704193, "learning_rate": 2.753455095615179e-05, "loss": 0.0667, "step": 2265 }, { "epoch": 1.062112527781027, "grad_norm": 0.16371361911296844, "learning_rate": 2.7431011620919554e-05, "loss": 0.0653, "step": 2270 }, { "epoch": 1.0644519826880336, "grad_norm": 0.26466837525367737, "learning_rate": 2.7327430172087077e-05, "loss": 0.0605, "step": 2275 }, { "epoch": 1.0667914375950405, "grad_norm": 0.1517406851053238, "learning_rate": 2.722380840404632e-05, "loss": 0.0595, "step": 2280 }, { "epoch": 1.069130892502047, "grad_norm": 0.16470152139663696, "learning_rate": 2.7120148111887732e-05, "loss": 0.0706, "step": 2285 }, { "epoch": 1.0714703474090537, "grad_norm": 0.15149088203907013, "learning_rate": 2.7016451091369116e-05, "loss": 0.0646, "step": 2290 }, { "epoch": 1.0738098023160603, "grad_norm": 0.14142438769340515, "learning_rate": 2.6912719138884553e-05, "loss": 0.0612, "step": 2295 }, { "epoch": 1.076149257223067, "grad_norm": 0.17522796988487244, "learning_rate": 2.6808954051433248e-05, "loss": 0.0657, "step": 2300 }, { "epoch": 1.0784887121300737, "grad_norm": 0.1683465838432312, "learning_rate": 2.670515762658844e-05, "loss": 0.0649, "step": 2305 }, { "epoch": 1.0808281670370803, "grad_norm": 0.1598363220691681, "learning_rate": 2.6601331662466227e-05, "loss": 0.0523, "step": 2310 }, { "epoch": 1.0831676219440871, "grad_norm": 0.14117397367954254, "learning_rate": 2.6497477957694443e-05, "loss": 0.0565, "step": 2315 }, { "epoch": 1.0855070768510937, "grad_norm": 0.16452661156654358, "learning_rate": 2.6393598311381473e-05, "loss": 0.0629, "step": 2320 }, { "epoch": 1.0878465317581003, "grad_norm": 0.14130878448486328, "learning_rate": 2.6289694523085112e-05, "loss": 0.0685, "step": 2325 }, { "epoch": 1.0901859866651071, "grad_norm": 0.1380264163017273, "learning_rate": 2.6185768392781366e-05, "loss": 0.056, "step": 2330 }, { "epoch": 1.0925254415721137, "grad_norm": 0.16482488811016083, "learning_rate": 2.6081821720833287e-05, "loss": 0.0569, "step": 2335 }, { "epoch": 1.0948648964791203, "grad_norm": 0.1513761430978775, "learning_rate": 2.5977856307959798e-05, "loss": 0.0622, "step": 2340 }, { "epoch": 1.097204351386127, "grad_norm": 0.15945203602313995, "learning_rate": 2.5873873955204448e-05, "loss": 0.0574, "step": 2345 }, { "epoch": 1.0995438062931338, "grad_norm": 0.16861629486083984, "learning_rate": 2.5769876463904265e-05, "loss": 0.0575, "step": 2350 }, { "epoch": 1.1018832612001404, "grad_norm": 0.15520039200782776, "learning_rate": 2.5665865635658527e-05, "loss": 0.063, "step": 2355 }, { "epoch": 1.104222716107147, "grad_norm": 0.1708049774169922, "learning_rate": 2.5561843272297536e-05, "loss": 0.0625, "step": 2360 }, { "epoch": 1.1065621710141538, "grad_norm": 0.1442188173532486, "learning_rate": 2.5457811175851465e-05, "loss": 0.0539, "step": 2365 }, { "epoch": 1.1089016259211604, "grad_norm": 0.15366148948669434, "learning_rate": 2.5353771148519057e-05, "loss": 0.0639, "step": 2370 }, { "epoch": 1.111241080828167, "grad_norm": 0.1677359789609909, "learning_rate": 2.524972499263646e-05, "loss": 0.0749, "step": 2375 }, { "epoch": 1.1135805357351738, "grad_norm": 0.16887301206588745, "learning_rate": 2.514567451064599e-05, "loss": 0.0632, "step": 2380 }, { "epoch": 1.1159199906421804, "grad_norm": 0.16333596408367157, "learning_rate": 2.50416215050649e-05, "loss": 0.0578, "step": 2385 }, { "epoch": 1.118259445549187, "grad_norm": 0.1616666167974472, "learning_rate": 2.4937567778454188e-05, "loss": 0.0603, "step": 2390 }, { "epoch": 1.1205989004561938, "grad_norm": 0.15231609344482422, "learning_rate": 2.4833515133387296e-05, "loss": 0.0577, "step": 2395 }, { "epoch": 1.1229383553632004, "grad_norm": 0.16632795333862305, "learning_rate": 2.4729465372418972e-05, "loss": 0.0715, "step": 2400 }, { "epoch": 1.125277810270207, "grad_norm": 0.16579392552375793, "learning_rate": 2.4625420298053968e-05, "loss": 0.0696, "step": 2405 }, { "epoch": 1.1276172651772136, "grad_norm": 0.1642322987318039, "learning_rate": 2.4521381712715884e-05, "loss": 0.0631, "step": 2410 }, { "epoch": 1.1299567200842204, "grad_norm": 0.1781720668077469, "learning_rate": 2.441735141871587e-05, "loss": 0.0584, "step": 2415 }, { "epoch": 1.132296174991227, "grad_norm": 0.16532465815544128, "learning_rate": 2.4313331218221476e-05, "loss": 0.0658, "step": 2420 }, { "epoch": 1.1346356298982336, "grad_norm": 0.1574028581380844, "learning_rate": 2.420932291322536e-05, "loss": 0.0636, "step": 2425 }, { "epoch": 1.1369750848052405, "grad_norm": 0.1727520376443863, "learning_rate": 2.410532830551415e-05, "loss": 0.0687, "step": 2430 }, { "epoch": 1.139314539712247, "grad_norm": 0.17281392216682434, "learning_rate": 2.4001349196637144e-05, "loss": 0.065, "step": 2435 }, { "epoch": 1.1416539946192537, "grad_norm": 0.16799040138721466, "learning_rate": 2.3897387387875188e-05, "loss": 0.0597, "step": 2440 }, { "epoch": 1.1439934495262605, "grad_norm": 0.1646750122308731, "learning_rate": 2.3793444680209397e-05, "loss": 0.0617, "step": 2445 }, { "epoch": 1.146332904433267, "grad_norm": 0.15122398734092712, "learning_rate": 2.3689522874290028e-05, "loss": 0.0646, "step": 2450 }, { "epoch": 1.1486723593402737, "grad_norm": 0.15441857278347015, "learning_rate": 2.358562377040519e-05, "loss": 0.0734, "step": 2455 }, { "epoch": 1.1510118142472803, "grad_norm": 0.1397976279258728, "learning_rate": 2.3481749168449774e-05, "loss": 0.0599, "step": 2460 }, { "epoch": 1.153351269154287, "grad_norm": 0.1505647897720337, "learning_rate": 2.3377900867894158e-05, "loss": 0.0587, "step": 2465 }, { "epoch": 1.1556907240612937, "grad_norm": 0.17228984832763672, "learning_rate": 2.3274080667753128e-05, "loss": 0.0713, "step": 2470 }, { "epoch": 1.1580301789683003, "grad_norm": 0.1795542687177658, "learning_rate": 2.3170290366554653e-05, "loss": 0.0681, "step": 2475 }, { "epoch": 1.1603696338753071, "grad_norm": 0.16465342044830322, "learning_rate": 2.3066531762308766e-05, "loss": 0.0598, "step": 2480 }, { "epoch": 1.1627090887823137, "grad_norm": 0.18022869527339935, "learning_rate": 2.2962806652476363e-05, "loss": 0.0553, "step": 2485 }, { "epoch": 1.1650485436893203, "grad_norm": 0.17633071541786194, "learning_rate": 2.2859116833938146e-05, "loss": 0.0694, "step": 2490 }, { "epoch": 1.1673879985963271, "grad_norm": 0.1658298820257187, "learning_rate": 2.2755464102963408e-05, "loss": 0.0576, "step": 2495 }, { "epoch": 1.1697274535033337, "grad_norm": 0.1415170431137085, "learning_rate": 2.2651850255178974e-05, "loss": 0.0697, "step": 2500 }, { "epoch": 1.1697274535033337, "eval_loss": 0.07991591095924377, "eval_runtime": 236.928, "eval_samples_per_second": 2.946, "eval_steps_per_second": 0.739, "step": 2500 }, { "epoch": 1.1720669084103403, "grad_norm": 0.17509862780570984, "learning_rate": 2.2548277085538092e-05, "loss": 0.0669, "step": 2505 }, { "epoch": 1.1744063633173472, "grad_norm": 0.17153185606002808, "learning_rate": 2.2444746388289283e-05, "loss": 0.0853, "step": 2510 }, { "epoch": 1.1767458182243538, "grad_norm": 0.1431114226579666, "learning_rate": 2.2341259956945342e-05, "loss": 0.0595, "step": 2515 }, { "epoch": 1.1790852731313604, "grad_norm": 0.14989079535007477, "learning_rate": 2.2237819584252188e-05, "loss": 0.0544, "step": 2520 }, { "epoch": 1.1814247280383672, "grad_norm": 0.13587923347949982, "learning_rate": 2.2134427062157866e-05, "loss": 0.0652, "step": 2525 }, { "epoch": 1.1837641829453738, "grad_norm": 0.1634293496608734, "learning_rate": 2.2031084181781476e-05, "loss": 0.0637, "step": 2530 }, { "epoch": 1.1861036378523804, "grad_norm": 0.16133365035057068, "learning_rate": 2.192779273338215e-05, "loss": 0.079, "step": 2535 }, { "epoch": 1.188443092759387, "grad_norm": 0.12266691774129868, "learning_rate": 2.182455450632803e-05, "loss": 0.0569, "step": 2540 }, { "epoch": 1.1907825476663938, "grad_norm": 0.15535689890384674, "learning_rate": 2.17213712890653e-05, "loss": 0.0569, "step": 2545 }, { "epoch": 1.1931220025734004, "grad_norm": 0.16880477964878082, "learning_rate": 2.1618244869087157e-05, "loss": 0.0651, "step": 2550 }, { "epoch": 1.195461457480407, "grad_norm": 0.13848161697387695, "learning_rate": 2.151517703290289e-05, "loss": 0.0693, "step": 2555 }, { "epoch": 1.1978009123874138, "grad_norm": 0.16127823293209076, "learning_rate": 2.1412169566006897e-05, "loss": 0.0598, "step": 2560 }, { "epoch": 1.2001403672944204, "grad_norm": 0.11767289787530899, "learning_rate": 2.1309224252847793e-05, "loss": 0.0593, "step": 2565 }, { "epoch": 1.202479822201427, "grad_norm": 0.17509087920188904, "learning_rate": 2.1206342876797436e-05, "loss": 0.0604, "step": 2570 }, { "epoch": 1.2048192771084336, "grad_norm": 0.16853219270706177, "learning_rate": 2.110352722012011e-05, "loss": 0.063, "step": 2575 }, { "epoch": 1.2071587320154404, "grad_norm": 0.13158155977725983, "learning_rate": 2.1000779063941585e-05, "loss": 0.0568, "step": 2580 }, { "epoch": 1.209498186922447, "grad_norm": 0.15704578161239624, "learning_rate": 2.0898100188218313e-05, "loss": 0.0634, "step": 2585 }, { "epoch": 1.2118376418294536, "grad_norm": 0.157634437084198, "learning_rate": 2.0795492371706527e-05, "loss": 0.0597, "step": 2590 }, { "epoch": 1.2141770967364605, "grad_norm": 0.1686418503522873, "learning_rate": 2.0692957391931485e-05, "loss": 0.0654, "step": 2595 }, { "epoch": 1.216516551643467, "grad_norm": 0.15651817619800568, "learning_rate": 2.0590497025156676e-05, "loss": 0.0591, "step": 2600 }, { "epoch": 1.2188560065504737, "grad_norm": 0.16979578137397766, "learning_rate": 2.0488113046353004e-05, "loss": 0.0619, "step": 2605 }, { "epoch": 1.2211954614574805, "grad_norm": 0.16162872314453125, "learning_rate": 2.0385807229168104e-05, "loss": 0.0653, "step": 2610 }, { "epoch": 1.223534916364487, "grad_norm": 0.17247694730758667, "learning_rate": 2.028358134589553e-05, "loss": 0.0647, "step": 2615 }, { "epoch": 1.2258743712714937, "grad_norm": 0.16605614125728607, "learning_rate": 2.0181437167444144e-05, "loss": 0.0632, "step": 2620 }, { "epoch": 1.2282138261785005, "grad_norm": 0.14373953640460968, "learning_rate": 2.0079376463307368e-05, "loss": 0.0619, "step": 2625 }, { "epoch": 1.230553281085507, "grad_norm": 0.1507708579301834, "learning_rate": 1.9977401001532576e-05, "loss": 0.0588, "step": 2630 }, { "epoch": 1.2328927359925137, "grad_norm": 0.16803239285945892, "learning_rate": 1.9875512548690433e-05, "loss": 0.0591, "step": 2635 }, { "epoch": 1.2352321908995205, "grad_norm": 0.16970671713352203, "learning_rate": 1.9773712869844323e-05, "loss": 0.0708, "step": 2640 }, { "epoch": 1.2375716458065271, "grad_norm": 0.14553260803222656, "learning_rate": 1.9672003728519728e-05, "loss": 0.0609, "step": 2645 }, { "epoch": 1.2399111007135337, "grad_norm": 0.1582130342721939, "learning_rate": 1.957038688667374e-05, "loss": 0.0642, "step": 2650 }, { "epoch": 1.2422505556205403, "grad_norm": 0.1633974015712738, "learning_rate": 1.9468864104664464e-05, "loss": 0.0638, "step": 2655 }, { "epoch": 1.2445900105275471, "grad_norm": 0.15454533696174622, "learning_rate": 1.9367437141220594e-05, "loss": 0.0643, "step": 2660 }, { "epoch": 1.2469294654345537, "grad_norm": 0.14684420824050903, "learning_rate": 1.9266107753410884e-05, "loss": 0.0611, "step": 2665 }, { "epoch": 1.2492689203415603, "grad_norm": 0.13646100461483002, "learning_rate": 1.916487769661377e-05, "loss": 0.055, "step": 2670 }, { "epoch": 1.2516083752485672, "grad_norm": 0.15355093777179718, "learning_rate": 1.9063748724486887e-05, "loss": 0.0518, "step": 2675 }, { "epoch": 1.2539478301555738, "grad_norm": 0.16275903582572937, "learning_rate": 1.896272258893677e-05, "loss": 0.0618, "step": 2680 }, { "epoch": 1.2562872850625804, "grad_norm": 0.14904741942882538, "learning_rate": 1.8861801040088445e-05, "loss": 0.0553, "step": 2685 }, { "epoch": 1.258626739969587, "grad_norm": 0.16873393952846527, "learning_rate": 1.8760985826255125e-05, "loss": 0.063, "step": 2690 }, { "epoch": 1.2609661948765938, "grad_norm": 0.1706465184688568, "learning_rate": 1.8660278693907962e-05, "loss": 0.0645, "step": 2695 }, { "epoch": 1.2633056497836004, "grad_norm": 0.14628972113132477, "learning_rate": 1.8559681387645715e-05, "loss": 0.0579, "step": 2700 }, { "epoch": 1.265645104690607, "grad_norm": 0.1530769318342209, "learning_rate": 1.8459195650164612e-05, "loss": 0.0629, "step": 2705 }, { "epoch": 1.2679845595976138, "grad_norm": 0.16960960626602173, "learning_rate": 1.8358823222228097e-05, "loss": 0.0627, "step": 2710 }, { "epoch": 1.2703240145046204, "grad_norm": 0.1773470938205719, "learning_rate": 1.8258565842636703e-05, "loss": 0.0582, "step": 2715 }, { "epoch": 1.272663469411627, "grad_norm": 0.162547767162323, "learning_rate": 1.815842524819793e-05, "loss": 0.0606, "step": 2720 }, { "epoch": 1.2750029243186338, "grad_norm": 0.15616440773010254, "learning_rate": 1.8058403173696152e-05, "loss": 0.0967, "step": 2725 }, { "epoch": 1.2773423792256404, "grad_norm": 0.1639740914106369, "learning_rate": 1.7958501351862545e-05, "loss": 0.0585, "step": 2730 }, { "epoch": 1.279681834132647, "grad_norm": 0.16226714849472046, "learning_rate": 1.7858721513345117e-05, "loss": 0.0664, "step": 2735 }, { "epoch": 1.2820212890396538, "grad_norm": 0.1616683155298233, "learning_rate": 1.7759065386678674e-05, "loss": 0.0548, "step": 2740 }, { "epoch": 1.2843607439466604, "grad_norm": 0.15211808681488037, "learning_rate": 1.765953469825494e-05, "loss": 0.0577, "step": 2745 }, { "epoch": 1.286700198853667, "grad_norm": 0.15566258132457733, "learning_rate": 1.7560131172292556e-05, "loss": 0.0632, "step": 2750 }, { "epoch": 1.2890396537606739, "grad_norm": 0.14558292925357819, "learning_rate": 1.7460856530807315e-05, "loss": 0.0592, "step": 2755 }, { "epoch": 1.2913791086676805, "grad_norm": 0.16073749959468842, "learning_rate": 1.7361712493582242e-05, "loss": 0.057, "step": 2760 }, { "epoch": 1.293718563574687, "grad_norm": 0.14757321774959564, "learning_rate": 1.726270077813788e-05, "loss": 0.0603, "step": 2765 }, { "epoch": 1.2960580184816939, "grad_norm": 0.17697452008724213, "learning_rate": 1.7163823099702466e-05, "loss": 0.0625, "step": 2770 }, { "epoch": 1.2983974733887005, "grad_norm": 0.17188061773777008, "learning_rate": 1.7065081171182264e-05, "loss": 0.0585, "step": 2775 }, { "epoch": 1.300736928295707, "grad_norm": 0.1670866757631302, "learning_rate": 1.696647670313186e-05, "loss": 0.0569, "step": 2780 }, { "epoch": 1.303076383202714, "grad_norm": 0.15840460360050201, "learning_rate": 1.6868011403724582e-05, "loss": 0.0616, "step": 2785 }, { "epoch": 1.3054158381097205, "grad_norm": 0.16217677295207977, "learning_rate": 1.676968697872282e-05, "loss": 0.063, "step": 2790 }, { "epoch": 1.307755293016727, "grad_norm": 0.15185709297657013, "learning_rate": 1.667150513144856e-05, "loss": 0.0517, "step": 2795 }, { "epoch": 1.3100947479237337, "grad_norm": 0.1776837259531021, "learning_rate": 1.657346756275386e-05, "loss": 0.0614, "step": 2800 }, { "epoch": 1.3124342028307403, "grad_norm": 0.17348352074623108, "learning_rate": 1.6475575970991313e-05, "loss": 0.0593, "step": 2805 }, { "epoch": 1.3147736577377471, "grad_norm": 0.1484275907278061, "learning_rate": 1.6377832051984755e-05, "loss": 0.0558, "step": 2810 }, { "epoch": 1.3171131126447537, "grad_norm": 0.15603071451187134, "learning_rate": 1.628023749899975e-05, "loss": 0.0544, "step": 2815 }, { "epoch": 1.3194525675517603, "grad_norm": 0.16645818948745728, "learning_rate": 1.618279400271436e-05, "loss": 0.0635, "step": 2820 }, { "epoch": 1.3217920224587671, "grad_norm": 0.1469196379184723, "learning_rate": 1.608550325118981e-05, "loss": 0.0559, "step": 2825 }, { "epoch": 1.3241314773657737, "grad_norm": 0.19546544551849365, "learning_rate": 1.5988366929841254e-05, "loss": 0.07, "step": 2830 }, { "epoch": 1.3264709322727803, "grad_norm": 0.1820397824048996, "learning_rate": 1.589138672140856e-05, "loss": 0.0718, "step": 2835 }, { "epoch": 1.3288103871797872, "grad_norm": 0.14199690520763397, "learning_rate": 1.57945643059272e-05, "loss": 0.0566, "step": 2840 }, { "epoch": 1.3311498420867938, "grad_norm": 0.1681860238313675, "learning_rate": 1.56979013606991e-05, "loss": 0.0628, "step": 2845 }, { "epoch": 1.3334892969938004, "grad_norm": 0.16973961889743805, "learning_rate": 1.560139956026363e-05, "loss": 0.058, "step": 2850 }, { "epoch": 1.3358287519008072, "grad_norm": 0.15116538107395172, "learning_rate": 1.5505060576368545e-05, "loss": 0.059, "step": 2855 }, { "epoch": 1.3381682068078138, "grad_norm": 0.1909974217414856, "learning_rate": 1.5408886077941074e-05, "loss": 0.066, "step": 2860 }, { "epoch": 1.3405076617148204, "grad_norm": 0.17362594604492188, "learning_rate": 1.531287773105895e-05, "loss": 0.0602, "step": 2865 }, { "epoch": 1.3428471166218272, "grad_norm": 0.16387321054935455, "learning_rate": 1.5217037198921625e-05, "loss": 0.0633, "step": 2870 }, { "epoch": 1.3451865715288338, "grad_norm": 0.15900340676307678, "learning_rate": 1.512136614182137e-05, "loss": 0.0694, "step": 2875 }, { "epoch": 1.3475260264358404, "grad_norm": 0.15678687393665314, "learning_rate": 1.5025866217114592e-05, "loss": 0.0531, "step": 2880 }, { "epoch": 1.3498654813428472, "grad_norm": 0.17534701526165009, "learning_rate": 1.4930539079193078e-05, "loss": 0.0648, "step": 2885 }, { "epoch": 1.3522049362498538, "grad_norm": 0.15762171149253845, "learning_rate": 1.483538637945533e-05, "loss": 0.0684, "step": 2890 }, { "epoch": 1.3545443911568604, "grad_norm": 0.13863767683506012, "learning_rate": 1.474040976627799e-05, "loss": 0.055, "step": 2895 }, { "epoch": 1.3568838460638672, "grad_norm": 0.1766417771577835, "learning_rate": 1.4645610884987265e-05, "loss": 0.1258, "step": 2900 }, { "epoch": 1.3592233009708738, "grad_norm": 0.17303304374217987, "learning_rate": 1.4550991377830426e-05, "loss": 0.0607, "step": 2905 }, { "epoch": 1.3615627558778804, "grad_norm": 0.17528830468654633, "learning_rate": 1.4456552883947333e-05, "loss": 0.0618, "step": 2910 }, { "epoch": 1.363902210784887, "grad_norm": 0.16436804831027985, "learning_rate": 1.4362297039342098e-05, "loss": 0.0648, "step": 2915 }, { "epoch": 1.3662416656918939, "grad_norm": 0.1755647212266922, "learning_rate": 1.4268225476854704e-05, "loss": 0.0681, "step": 2920 }, { "epoch": 1.3685811205989005, "grad_norm": 0.15491195023059845, "learning_rate": 1.4174339826132727e-05, "loss": 0.0587, "step": 2925 }, { "epoch": 1.370920575505907, "grad_norm": 0.1469022035598755, "learning_rate": 1.4080641713603072e-05, "loss": 0.0669, "step": 2930 }, { "epoch": 1.3732600304129137, "grad_norm": 0.18021321296691895, "learning_rate": 1.3987132762443873e-05, "loss": 0.0601, "step": 2935 }, { "epoch": 1.3755994853199205, "grad_norm": 0.1528056561946869, "learning_rate": 1.3893814592556287e-05, "loss": 0.0599, "step": 2940 }, { "epoch": 1.377938940226927, "grad_norm": 0.15603193640708923, "learning_rate": 1.3800688820536517e-05, "loss": 0.0539, "step": 2945 }, { "epoch": 1.3802783951339337, "grad_norm": 0.15558800101280212, "learning_rate": 1.3707757059647725e-05, "loss": 0.0671, "step": 2950 }, { "epoch": 1.3826178500409405, "grad_norm": 0.156876340508461, "learning_rate": 1.3615020919792151e-05, "loss": 0.0586, "step": 2955 }, { "epoch": 1.384957304947947, "grad_norm": 0.1524912714958191, "learning_rate": 1.3522482007483172e-05, "loss": 0.0662, "step": 2960 }, { "epoch": 1.3872967598549537, "grad_norm": 0.15926173329353333, "learning_rate": 1.3430141925817532e-05, "loss": 0.0637, "step": 2965 }, { "epoch": 1.3896362147619605, "grad_norm": 0.16642306745052338, "learning_rate": 1.333800227444749e-05, "loss": 0.067, "step": 2970 }, { "epoch": 1.3919756696689671, "grad_norm": 0.13672704994678497, "learning_rate": 1.324606464955318e-05, "loss": 0.0618, "step": 2975 }, { "epoch": 1.3943151245759737, "grad_norm": 0.16178594529628754, "learning_rate": 1.3154330643814938e-05, "loss": 0.0564, "step": 2980 }, { "epoch": 1.3966545794829806, "grad_norm": 0.16218972206115723, "learning_rate": 1.30628018463857e-05, "loss": 0.0592, "step": 2985 }, { "epoch": 1.3989940343899872, "grad_norm": 0.16459548473358154, "learning_rate": 1.2971479842863465e-05, "loss": 0.0533, "step": 2990 }, { "epoch": 1.4013334892969938, "grad_norm": 0.14604820311069489, "learning_rate": 1.2880366215263845e-05, "loss": 0.0629, "step": 2995 }, { "epoch": 1.4036729442040006, "grad_norm": 0.16797035932540894, "learning_rate": 1.2789462541992676e-05, "loss": 0.0553, "step": 3000 }, { "epoch": 1.4036729442040006, "eval_loss": 0.07826597988605499, "eval_runtime": 234.1597, "eval_samples_per_second": 2.981, "eval_steps_per_second": 0.747, "step": 3000 }, { "epoch": 1.4060123991110072, "grad_norm": 0.1841832548379898, "learning_rate": 1.2698770397818616e-05, "loss": 0.13, "step": 3005 }, { "epoch": 1.4083518540180138, "grad_norm": 0.1567966490983963, "learning_rate": 1.2608291353845941e-05, "loss": 0.063, "step": 3010 }, { "epoch": 1.4106913089250206, "grad_norm": 0.14550812542438507, "learning_rate": 1.2518026977487252e-05, "loss": 0.0598, "step": 3015 }, { "epoch": 1.4130307638320272, "grad_norm": 0.15681859850883484, "learning_rate": 1.242797883243638e-05, "loss": 0.063, "step": 3020 }, { "epoch": 1.4153702187390338, "grad_norm": 0.16205298900604248, "learning_rate": 1.233814847864127e-05, "loss": 0.0691, "step": 3025 }, { "epoch": 1.4177096736460404, "grad_norm": 0.16490066051483154, "learning_rate": 1.224853747227698e-05, "loss": 0.0628, "step": 3030 }, { "epoch": 1.4200491285530472, "grad_norm": 0.19161145389080048, "learning_rate": 1.2159147365718664e-05, "loss": 0.0646, "step": 3035 }, { "epoch": 1.4223885834600538, "grad_norm": 0.16144663095474243, "learning_rate": 1.2069979707514766e-05, "loss": 0.0542, "step": 3040 }, { "epoch": 1.4247280383670604, "grad_norm": 0.15800189971923828, "learning_rate": 1.1981036042360109e-05, "loss": 0.059, "step": 3045 }, { "epoch": 1.427067493274067, "grad_norm": 0.15965986251831055, "learning_rate": 1.1892317911069212e-05, "loss": 0.0575, "step": 3050 }, { "epoch": 1.4294069481810738, "grad_norm": 0.15217727422714233, "learning_rate": 1.1803826850549526e-05, "loss": 0.0604, "step": 3055 }, { "epoch": 1.4317464030880804, "grad_norm": 0.15865880250930786, "learning_rate": 1.171556439377488e-05, "loss": 0.0659, "step": 3060 }, { "epoch": 1.434085857995087, "grad_norm": 0.18922170996665955, "learning_rate": 1.1627532069758851e-05, "loss": 0.0605, "step": 3065 }, { "epoch": 1.4364253129020939, "grad_norm": 0.17972765862941742, "learning_rate": 1.1539731403528353e-05, "loss": 0.0754, "step": 3070 }, { "epoch": 1.4387647678091005, "grad_norm": 0.18478457629680634, "learning_rate": 1.1452163916097141e-05, "loss": 0.0689, "step": 3075 }, { "epoch": 1.441104222716107, "grad_norm": 0.16333162784576416, "learning_rate": 1.1364831124439526e-05, "loss": 0.0669, "step": 3080 }, { "epoch": 1.4434436776231139, "grad_norm": 0.1580970138311386, "learning_rate": 1.1277734541464072e-05, "loss": 0.0596, "step": 3085 }, { "epoch": 1.4457831325301205, "grad_norm": 0.17736057937145233, "learning_rate": 1.1190875675987356e-05, "loss": 0.057, "step": 3090 }, { "epoch": 1.448122587437127, "grad_norm": 0.17096631228923798, "learning_rate": 1.1104256032707894e-05, "loss": 0.0671, "step": 3095 }, { "epoch": 1.450462042344134, "grad_norm": 0.17101797461509705, "learning_rate": 1.1017877112180009e-05, "loss": 0.0583, "step": 3100 }, { "epoch": 1.4528014972511405, "grad_norm": 0.1477900892496109, "learning_rate": 1.0931740410787895e-05, "loss": 0.0617, "step": 3105 }, { "epoch": 1.455140952158147, "grad_norm": 0.16209881007671356, "learning_rate": 1.0845847420719632e-05, "loss": 0.0619, "step": 3110 }, { "epoch": 1.457480407065154, "grad_norm": 0.16828244924545288, "learning_rate": 1.076019962994141e-05, "loss": 0.0572, "step": 3115 }, { "epoch": 1.4598198619721605, "grad_norm": 0.1566639393568039, "learning_rate": 1.0674798522171668e-05, "loss": 0.0632, "step": 3120 }, { "epoch": 1.4621593168791671, "grad_norm": 0.1641826629638672, "learning_rate": 1.0589645576855477e-05, "loss": 0.055, "step": 3125 }, { "epoch": 1.464498771786174, "grad_norm": 0.1548743098974228, "learning_rate": 1.0504742269138835e-05, "loss": 0.0631, "step": 3130 }, { "epoch": 1.4668382266931805, "grad_norm": 0.16996270418167114, "learning_rate": 1.0420090069843167e-05, "loss": 0.0557, "step": 3135 }, { "epoch": 1.4691776816001871, "grad_norm": 0.18014395236968994, "learning_rate": 1.0335690445439817e-05, "loss": 0.0623, "step": 3140 }, { "epoch": 1.471517136507194, "grad_norm": 0.17947927117347717, "learning_rate": 1.0251544858024662e-05, "loss": 0.056, "step": 3145 }, { "epoch": 1.4738565914142006, "grad_norm": 0.19152309000492096, "learning_rate": 1.0167654765292742e-05, "loss": 0.0581, "step": 3150 }, { "epoch": 1.4761960463212072, "grad_norm": 0.16634751856327057, "learning_rate": 1.0084021620513079e-05, "loss": 0.0619, "step": 3155 }, { "epoch": 1.4785355012282138, "grad_norm": 0.19897738099098206, "learning_rate": 1.0000646872503419e-05, "loss": 0.0784, "step": 3160 }, { "epoch": 1.4808749561352204, "grad_norm": 0.1585209220647812, "learning_rate": 9.917531965605211e-06, "loss": 0.062, "step": 3165 }, { "epoch": 1.4832144110422272, "grad_norm": 0.18893523514270782, "learning_rate": 9.83467833965852e-06, "loss": 0.057, "step": 3170 }, { "epoch": 1.4855538659492338, "grad_norm": 0.17129787802696228, "learning_rate": 9.752087429977143e-06, "loss": 0.0674, "step": 3175 }, { "epoch": 1.4878933208562404, "grad_norm": 0.1496947705745697, "learning_rate": 9.669760667323685e-06, "loss": 0.0842, "step": 3180 }, { "epoch": 1.4902327757632472, "grad_norm": 0.17729486525058746, "learning_rate": 9.58769947788482e-06, "loss": 0.0736, "step": 3185 }, { "epoch": 1.4925722306702538, "grad_norm": 0.17603404819965363, "learning_rate": 9.505905283246578e-06, "loss": 0.0638, "step": 3190 }, { "epoch": 1.4949116855772604, "grad_norm": 0.17529185116291046, "learning_rate": 9.424379500369674e-06, "loss": 0.0672, "step": 3195 }, { "epoch": 1.4972511404842672, "grad_norm": 0.1493815779685974, "learning_rate": 9.343123541565035e-06, "loss": 0.058, "step": 3200 }, { "epoch": 1.4995905953912738, "grad_norm": 0.15651191771030426, "learning_rate": 9.26213881446926e-06, "loss": 0.0588, "step": 3205 }, { "epoch": 1.5019300502982804, "grad_norm": 0.15472684800624847, "learning_rate": 9.181426722020297e-06, "loss": 0.0626, "step": 3210 }, { "epoch": 1.5042695052052872, "grad_norm": 0.1653946340084076, "learning_rate": 9.100988662433082e-06, "loss": 0.0824, "step": 3215 }, { "epoch": 1.5066089601122938, "grad_norm": 0.1525607407093048, "learning_rate": 9.020826029175384e-06, "loss": 0.0621, "step": 3220 }, { "epoch": 1.5089484150193004, "grad_norm": 0.18066176772117615, "learning_rate": 8.94094021094358e-06, "loss": 0.0632, "step": 3225 }, { "epoch": 1.5112878699263073, "grad_norm": 0.13805319368839264, "learning_rate": 8.86133259163869e-06, "loss": 0.0638, "step": 3230 }, { "epoch": 1.5136273248333139, "grad_norm": 0.1525852382183075, "learning_rate": 8.782004550342323e-06, "loss": 0.0586, "step": 3235 }, { "epoch": 1.5159667797403205, "grad_norm": 0.16709508001804352, "learning_rate": 8.702957461292846e-06, "loss": 0.0632, "step": 3240 }, { "epoch": 1.5183062346473273, "grad_norm": 0.1565026491880417, "learning_rate": 8.624192693861544e-06, "loss": 0.0578, "step": 3245 }, { "epoch": 1.5206456895543339, "grad_norm": 0.16678424179553986, "learning_rate": 8.545711612528915e-06, "loss": 0.063, "step": 3250 }, { "epoch": 1.5229851444613405, "grad_norm": 0.19322334229946136, "learning_rate": 8.467515576861002e-06, "loss": 0.0657, "step": 3255 }, { "epoch": 1.5253245993683473, "grad_norm": 0.1959463208913803, "learning_rate": 8.389605941485893e-06, "loss": 0.1184, "step": 3260 }, { "epoch": 1.5276640542753537, "grad_norm": 0.1491149216890335, "learning_rate": 8.31198405607019e-06, "loss": 0.055, "step": 3265 }, { "epoch": 1.5300035091823605, "grad_norm": 0.1650906205177307, "learning_rate": 8.234651265295702e-06, "loss": 0.0669, "step": 3270 }, { "epoch": 1.5323429640893673, "grad_norm": 0.13482898473739624, "learning_rate": 8.15760890883607e-06, "loss": 0.0577, "step": 3275 }, { "epoch": 1.5346824189963737, "grad_norm": 0.15607015788555145, "learning_rate": 8.080858321333633e-06, "loss": 0.063, "step": 3280 }, { "epoch": 1.5370218739033805, "grad_norm": 0.15770550072193146, "learning_rate": 8.004400832376258e-06, "loss": 0.0617, "step": 3285 }, { "epoch": 1.5393613288103873, "grad_norm": 0.15652437508106232, "learning_rate": 7.928237766474314e-06, "loss": 0.0668, "step": 3290 }, { "epoch": 1.5417007837173937, "grad_norm": 0.174765482544899, "learning_rate": 7.852370443037751e-06, "loss": 0.0604, "step": 3295 }, { "epoch": 1.5440402386244005, "grad_norm": 0.14293302595615387, "learning_rate": 7.776800176353227e-06, "loss": 0.0852, "step": 3300 }, { "epoch": 1.5463796935314071, "grad_norm": 0.16163261234760284, "learning_rate": 7.701528275561348e-06, "loss": 0.0571, "step": 3305 }, { "epoch": 1.5487191484384137, "grad_norm": 0.13872607052326202, "learning_rate": 7.626556044633951e-06, "loss": 0.0606, "step": 3310 }, { "epoch": 1.5510586033454206, "grad_norm": 0.16931146383285522, "learning_rate": 7.551884782351587e-06, "loss": 0.0582, "step": 3315 }, { "epoch": 1.5533980582524272, "grad_norm": 0.16759054362773895, "learning_rate": 7.477515782280939e-06, "loss": 0.0656, "step": 3320 }, { "epoch": 1.5557375131594338, "grad_norm": 0.15919411182403564, "learning_rate": 7.403450332752493e-06, "loss": 0.0685, "step": 3325 }, { "epoch": 1.5580769680664406, "grad_norm": 0.1733466535806656, "learning_rate": 7.329689716838148e-06, "loss": 0.06, "step": 3330 }, { "epoch": 1.5604164229734472, "grad_norm": 0.15883946418762207, "learning_rate": 7.256235212329055e-06, "loss": 0.0537, "step": 3335 }, { "epoch": 1.5627558778804538, "grad_norm": 0.20026150345802307, "learning_rate": 7.183088091713405e-06, "loss": 0.0676, "step": 3340 }, { "epoch": 1.5650953327874606, "grad_norm": 0.14357200264930725, "learning_rate": 7.110249622154469e-06, "loss": 0.0596, "step": 3345 }, { "epoch": 1.5674347876944672, "grad_norm": 0.15919791162014008, "learning_rate": 7.037721065468591e-06, "loss": 0.0504, "step": 3350 }, { "epoch": 1.5697742426014738, "grad_norm": 0.16171149909496307, "learning_rate": 6.96550367810333e-06, "loss": 0.0688, "step": 3355 }, { "epoch": 1.5721136975084806, "grad_norm": 0.17881818115711212, "learning_rate": 6.89359871111572e-06, "loss": 0.0545, "step": 3360 }, { "epoch": 1.5744531524154872, "grad_norm": 0.1703099012374878, "learning_rate": 6.822007410150591e-06, "loss": 0.0624, "step": 3365 }, { "epoch": 1.5767926073224938, "grad_norm": 0.17320412397384644, "learning_rate": 6.750731015418959e-06, "loss": 0.0659, "step": 3370 }, { "epoch": 1.5791320622295006, "grad_norm": 0.15360009670257568, "learning_rate": 6.679770761676596e-06, "loss": 0.062, "step": 3375 }, { "epoch": 1.5814715171365072, "grad_norm": 0.15514342486858368, "learning_rate": 6.609127878202581e-06, "loss": 0.0522, "step": 3380 }, { "epoch": 1.5838109720435138, "grad_norm": 0.14531250298023224, "learning_rate": 6.538803588778064e-06, "loss": 0.0588, "step": 3385 }, { "epoch": 1.5861504269505207, "grad_norm": 0.16243524849414825, "learning_rate": 6.468799111665003e-06, "loss": 0.0641, "step": 3390 }, { "epoch": 1.588489881857527, "grad_norm": 0.19768266379833221, "learning_rate": 6.399115659585103e-06, "loss": 0.0668, "step": 3395 }, { "epoch": 1.5908293367645339, "grad_norm": 0.1622917205095291, "learning_rate": 6.329754439698804e-06, "loss": 0.0531, "step": 3400 }, { "epoch": 1.5931687916715407, "grad_norm": 0.18452420830726624, "learning_rate": 6.260716653584359e-06, "loss": 0.0992, "step": 3405 }, { "epoch": 1.595508246578547, "grad_norm": 0.16582529246807098, "learning_rate": 6.192003497217016e-06, "loss": 0.0574, "step": 3410 }, { "epoch": 1.5978477014855539, "grad_norm": 0.1616390198469162, "learning_rate": 6.123616160948295e-06, "loss": 0.0616, "step": 3415 }, { "epoch": 1.6001871563925605, "grad_norm": 0.16133980453014374, "learning_rate": 6.055555829485396e-06, "loss": 0.0584, "step": 3420 }, { "epoch": 1.602526611299567, "grad_norm": 0.14773264527320862, "learning_rate": 5.98782368187063e-06, "loss": 0.0557, "step": 3425 }, { "epoch": 1.604866066206574, "grad_norm": 0.1554953008890152, "learning_rate": 5.920420891461046e-06, "loss": 0.0618, "step": 3430 }, { "epoch": 1.6072055211135805, "grad_norm": 0.1828761100769043, "learning_rate": 5.853348625908048e-06, "loss": 0.0623, "step": 3435 }, { "epoch": 1.609544976020587, "grad_norm": 0.1556580513715744, "learning_rate": 5.786608047137229e-06, "loss": 0.0564, "step": 3440 }, { "epoch": 1.611884430927594, "grad_norm": 0.1514003425836563, "learning_rate": 5.720200311328178e-06, "loss": 0.0595, "step": 3445 }, { "epoch": 1.6142238858346005, "grad_norm": 0.14597107470035553, "learning_rate": 5.654126568894513e-06, "loss": 0.0595, "step": 3450 }, { "epoch": 1.6165633407416071, "grad_norm": 0.17368368804454803, "learning_rate": 5.5883879644638925e-06, "loss": 0.0606, "step": 3455 }, { "epoch": 1.618902795648614, "grad_norm": 0.16808243095874786, "learning_rate": 5.522985636858239e-06, "loss": 0.058, "step": 3460 }, { "epoch": 1.6212422505556205, "grad_norm": 0.1682329773902893, "learning_rate": 5.4579207190739775e-06, "loss": 0.0578, "step": 3465 }, { "epoch": 1.6235817054626271, "grad_norm": 0.1642475128173828, "learning_rate": 5.393194338262428e-06, "loss": 0.0651, "step": 3470 }, { "epoch": 1.625921160369634, "grad_norm": 0.18125009536743164, "learning_rate": 5.328807615710246e-06, "loss": 0.0586, "step": 3475 }, { "epoch": 1.6282606152766406, "grad_norm": 0.14442673325538635, "learning_rate": 5.264761666820054e-06, "loss": 0.059, "step": 3480 }, { "epoch": 1.6306000701836472, "grad_norm": 0.1675989329814911, "learning_rate": 5.201057601091056e-06, "loss": 0.0574, "step": 3485 }, { "epoch": 1.632939525090654, "grad_norm": 0.17742133140563965, "learning_rate": 5.137696522099858e-06, "loss": 0.0577, "step": 3490 }, { "epoch": 1.6352789799976606, "grad_norm": 0.15731069445610046, "learning_rate": 5.074679527481349e-06, "loss": 0.055, "step": 3495 }, { "epoch": 1.6376184349046672, "grad_norm": 0.17156293988227844, "learning_rate": 5.012007708909661e-06, "loss": 0.057, "step": 3500 }, { "epoch": 1.6376184349046672, "eval_loss": 0.07722621411085129, "eval_runtime": 235.8253, "eval_samples_per_second": 2.96, "eval_steps_per_second": 0.742, "step": 3500 }, { "epoch": 1.639957889811674, "grad_norm": 0.1923573613166809, "learning_rate": 4.949682152079291e-06, "loss": 0.0658, "step": 3505 }, { "epoch": 1.6422973447186804, "grad_norm": 0.15531805157661438, "learning_rate": 4.887703936686252e-06, "loss": 0.0549, "step": 3510 }, { "epoch": 1.6446367996256872, "grad_norm": 0.18201015889644623, "learning_rate": 4.826074136409411e-06, "loss": 0.0665, "step": 3515 }, { "epoch": 1.646976254532694, "grad_norm": 0.14315198361873627, "learning_rate": 4.764793818891866e-06, "loss": 0.0557, "step": 3520 }, { "epoch": 1.6493157094397004, "grad_norm": 0.16865558922290802, "learning_rate": 4.703864045722453e-06, "loss": 0.0609, "step": 3525 }, { "epoch": 1.6516551643467072, "grad_norm": 0.16090945899486542, "learning_rate": 4.643285872417346e-06, "loss": 0.0602, "step": 3530 }, { "epoch": 1.653994619253714, "grad_norm": 0.16474011540412903, "learning_rate": 4.583060348401808e-06, "loss": 0.0529, "step": 3535 }, { "epoch": 1.6563340741607204, "grad_norm": 0.18071883916854858, "learning_rate": 4.523188516991958e-06, "loss": 0.0598, "step": 3540 }, { "epoch": 1.6586735290677272, "grad_norm": 0.14429545402526855, "learning_rate": 4.4636714153767496e-06, "loss": 0.058, "step": 3545 }, { "epoch": 1.6610129839747338, "grad_norm": 0.17122824490070343, "learning_rate": 4.40451007459996e-06, "loss": 0.061, "step": 3550 }, { "epoch": 1.6633524388817404, "grad_norm": 0.17276746034622192, "learning_rate": 4.3457055195423645e-06, "loss": 0.0557, "step": 3555 }, { "epoch": 1.6656918937887473, "grad_norm": 0.1375202089548111, "learning_rate": 4.2872587689039484e-06, "loss": 0.055, "step": 3560 }, { "epoch": 1.6680313486957539, "grad_norm": 0.14362306892871857, "learning_rate": 4.229170835186292e-06, "loss": 0.0573, "step": 3565 }, { "epoch": 1.6703708036027605, "grad_norm": 0.16020837426185608, "learning_rate": 4.171442724675004e-06, "loss": 0.0599, "step": 3570 }, { "epoch": 1.6727102585097673, "grad_norm": 0.1534327119588852, "learning_rate": 4.114075437422313e-06, "loss": 0.0663, "step": 3575 }, { "epoch": 1.6750497134167739, "grad_norm": 0.15872815251350403, "learning_rate": 4.057069967229718e-06, "loss": 0.0759, "step": 3580 }, { "epoch": 1.6773891683237805, "grad_norm": 0.17739912867546082, "learning_rate": 4.00042730163078e-06, "loss": 0.0576, "step": 3585 }, { "epoch": 1.6797286232307873, "grad_norm": 0.14820364117622375, "learning_rate": 3.944148421874042e-06, "loss": 0.0579, "step": 3590 }, { "epoch": 1.682068078137794, "grad_norm": 0.17816150188446045, "learning_rate": 3.888234302905977e-06, "loss": 0.0625, "step": 3595 }, { "epoch": 1.6844075330448005, "grad_norm": 0.1766812652349472, "learning_rate": 3.832685913354159e-06, "loss": 0.0551, "step": 3600 }, { "epoch": 1.6867469879518073, "grad_norm": 0.1790032684803009, "learning_rate": 3.777504215510427e-06, "loss": 0.0643, "step": 3605 }, { "epoch": 1.689086442858814, "grad_norm": 0.16145819425582886, "learning_rate": 3.7226901653142634e-06, "loss": 0.0616, "step": 3610 }, { "epoch": 1.6914258977658205, "grad_norm": 0.18062575161457062, "learning_rate": 3.6682447123361885e-06, "loss": 0.0688, "step": 3615 }, { "epoch": 1.6937653526728274, "grad_norm": 0.15464866161346436, "learning_rate": 3.61416879976135e-06, "loss": 0.062, "step": 3620 }, { "epoch": 1.696104807579834, "grad_norm": 0.1730775237083435, "learning_rate": 3.560463364373165e-06, "loss": 0.0552, "step": 3625 }, { "epoch": 1.6984442624868406, "grad_norm": 0.17476975917816162, "learning_rate": 3.5071293365370904e-06, "loss": 0.0694, "step": 3630 }, { "epoch": 1.7007837173938474, "grad_norm": 0.1686089038848877, "learning_rate": 3.4541676401845027e-06, "loss": 0.0562, "step": 3635 }, { "epoch": 1.7031231723008537, "grad_norm": 0.1799984872341156, "learning_rate": 3.4015791927967083e-06, "loss": 0.0596, "step": 3640 }, { "epoch": 1.7054626272078606, "grad_norm": 0.15171782672405243, "learning_rate": 3.3493649053890326e-06, "loss": 0.0797, "step": 3645 }, { "epoch": 1.7078020821148674, "grad_norm": 0.17303191125392914, "learning_rate": 3.2975256824950564e-06, "loss": 0.0594, "step": 3650 }, { "epoch": 1.7101415370218738, "grad_norm": 0.17267484962940216, "learning_rate": 3.2460624221509137e-06, "loss": 0.0563, "step": 3655 }, { "epoch": 1.7124809919288806, "grad_norm": 0.16894906759262085, "learning_rate": 3.1949760158797833e-06, "loss": 0.0614, "step": 3660 }, { "epoch": 1.7148204468358872, "grad_norm": 0.15203548967838287, "learning_rate": 3.1442673486763912e-06, "loss": 0.0648, "step": 3665 }, { "epoch": 1.7171599017428938, "grad_norm": 0.16010317206382751, "learning_rate": 3.0939372989917254e-06, "loss": 0.0593, "step": 3670 }, { "epoch": 1.7194993566499006, "grad_norm": 0.14804290235042572, "learning_rate": 3.04398673871778e-06, "loss": 0.0588, "step": 3675 }, { "epoch": 1.7218388115569072, "grad_norm": 0.17553718388080597, "learning_rate": 2.9944165331724866e-06, "loss": 0.0567, "step": 3680 }, { "epoch": 1.7241782664639138, "grad_norm": 0.16042271256446838, "learning_rate": 2.945227541084697e-06, "loss": 0.0548, "step": 3685 }, { "epoch": 1.7265177213709206, "grad_norm": 0.16253893077373505, "learning_rate": 2.896420614579312e-06, "loss": 0.0515, "step": 3690 }, { "epoch": 1.7288571762779272, "grad_norm": 0.16040562093257904, "learning_rate": 2.847996599162539e-06, "loss": 0.0575, "step": 3695 }, { "epoch": 1.7311966311849338, "grad_norm": 0.1720447987318039, "learning_rate": 2.799956333707207e-06, "loss": 0.064, "step": 3700 }, { "epoch": 1.7335360860919407, "grad_norm": 0.15470455586910248, "learning_rate": 2.752300650438283e-06, "loss": 0.0572, "step": 3705 }, { "epoch": 1.7358755409989473, "grad_norm": 0.15257127583026886, "learning_rate": 2.7050303749184046e-06, "loss": 0.0676, "step": 3710 }, { "epoch": 1.7382149959059539, "grad_norm": 0.17101050913333893, "learning_rate": 2.6581463260336252e-06, "loss": 0.0646, "step": 3715 }, { "epoch": 1.7405544508129607, "grad_norm": 0.16087070107460022, "learning_rate": 2.6116493159791866e-06, "loss": 0.0499, "step": 3720 }, { "epoch": 1.7428939057199673, "grad_norm": 0.16433356702327728, "learning_rate": 2.5655401502454795e-06, "loss": 0.0668, "step": 3725 }, { "epoch": 1.7452333606269739, "grad_norm": 0.19104856252670288, "learning_rate": 2.5198196276040782e-06, "loss": 0.0579, "step": 3730 }, { "epoch": 1.7475728155339807, "grad_norm": 0.16840414702892303, "learning_rate": 2.4744885400939026e-06, "loss": 0.0642, "step": 3735 }, { "epoch": 1.7499122704409873, "grad_norm": 0.16252835094928741, "learning_rate": 2.42954767300749e-06, "loss": 0.0561, "step": 3740 }, { "epoch": 1.752251725347994, "grad_norm": 0.16803041100502014, "learning_rate": 2.384997804877412e-06, "loss": 0.0545, "step": 3745 }, { "epoch": 1.7545911802550007, "grad_norm": 0.1815616637468338, "learning_rate": 2.340839707462755e-06, "loss": 0.0556, "step": 3750 }, { "epoch": 1.756930635162007, "grad_norm": 0.16640643775463104, "learning_rate": 2.2970741457357973e-06, "loss": 0.0587, "step": 3755 }, { "epoch": 1.759270090069014, "grad_norm": 0.1885158121585846, "learning_rate": 2.2537018778687026e-06, "loss": 0.0729, "step": 3760 }, { "epoch": 1.7616095449760207, "grad_norm": 0.12472023814916611, "learning_rate": 2.210723655220434e-06, "loss": 0.0619, "step": 3765 }, { "epoch": 1.7639489998830271, "grad_norm": 0.17403416335582733, "learning_rate": 2.168140222323703e-06, "loss": 0.0619, "step": 3770 }, { "epoch": 1.766288454790034, "grad_norm": 0.16542308032512665, "learning_rate": 2.1259523168720957e-06, "loss": 0.0672, "step": 3775 }, { "epoch": 1.7686279096970405, "grad_norm": 0.1458907127380371, "learning_rate": 2.0841606697072745e-06, "loss": 0.056, "step": 3780 }, { "epoch": 1.7709673646040471, "grad_norm": 0.1915542483329773, "learning_rate": 2.042766004806332e-06, "loss": 0.1111, "step": 3785 }, { "epoch": 1.773306819511054, "grad_norm": 0.1618765890598297, "learning_rate": 2.001769039269247e-06, "loss": 0.0606, "step": 3790 }, { "epoch": 1.7756462744180606, "grad_norm": 0.17050661146640778, "learning_rate": 1.9611704833064486e-06, "loss": 0.0702, "step": 3795 }, { "epoch": 1.7779857293250672, "grad_norm": 0.1698133945465088, "learning_rate": 1.9209710402265295e-06, "loss": 0.0586, "step": 3800 }, { "epoch": 1.780325184232074, "grad_norm": 0.16448405385017395, "learning_rate": 1.8811714064240481e-06, "loss": 0.059, "step": 3805 }, { "epoch": 1.7826646391390806, "grad_norm": 0.14546535909175873, "learning_rate": 1.841772271367484e-06, "loss": 0.0543, "step": 3810 }, { "epoch": 1.7850040940460872, "grad_norm": 0.1465018391609192, "learning_rate": 1.8027743175872664e-06, "loss": 0.0507, "step": 3815 }, { "epoch": 1.787343548953094, "grad_norm": 0.15931351482868195, "learning_rate": 1.7641782206639813e-06, "loss": 0.0601, "step": 3820 }, { "epoch": 1.7896830038601006, "grad_norm": 0.18377573788166046, "learning_rate": 1.7259846492166359e-06, "loss": 0.0604, "step": 3825 }, { "epoch": 1.7920224587671072, "grad_norm": 0.1727667599916458, "learning_rate": 1.6881942648911076e-06, "loss": 0.0547, "step": 3830 }, { "epoch": 1.794361913674114, "grad_norm": 0.1737833321094513, "learning_rate": 1.6508077223486546e-06, "loss": 0.0605, "step": 3835 }, { "epoch": 1.7967013685811206, "grad_norm": 0.18085430562496185, "learning_rate": 1.61382566925459e-06, "loss": 0.0553, "step": 3840 }, { "epoch": 1.7990408234881272, "grad_norm": 0.1712871491909027, "learning_rate": 1.5772487462670681e-06, "loss": 0.0654, "step": 3845 }, { "epoch": 1.801380278395134, "grad_norm": 0.16300112009048462, "learning_rate": 1.5410775870259686e-06, "loss": 0.0558, "step": 3850 }, { "epoch": 1.8037197333021406, "grad_norm": 0.17991948127746582, "learning_rate": 1.5053128181419184e-06, "loss": 0.06, "step": 3855 }, { "epoch": 1.8060591882091472, "grad_norm": 0.15448680520057678, "learning_rate": 1.469955059185471e-06, "loss": 0.0645, "step": 3860 }, { "epoch": 1.808398643116154, "grad_norm": 0.17701593041419983, "learning_rate": 1.4350049226763224e-06, "loss": 0.0631, "step": 3865 }, { "epoch": 1.8107380980231607, "grad_norm": 0.185321643948555, "learning_rate": 1.400463014072742e-06, "loss": 0.0639, "step": 3870 }, { "epoch": 1.8130775529301673, "grad_norm": 0.18532027304172516, "learning_rate": 1.3663299317610595e-06, "loss": 0.0631, "step": 3875 }, { "epoch": 1.815417007837174, "grad_norm": 0.16126079857349396, "learning_rate": 1.3326062670453194e-06, "loss": 0.0693, "step": 3880 }, { "epoch": 1.8177564627441805, "grad_norm": 0.15626025199890137, "learning_rate": 1.2992926041370064e-06, "loss": 0.059, "step": 3885 }, { "epoch": 1.8200959176511873, "grad_norm": 0.1502443104982376, "learning_rate": 1.2663895201449588e-06, "loss": 0.058, "step": 3890 }, { "epoch": 1.822435372558194, "grad_norm": 0.16258084774017334, "learning_rate": 1.2338975850653579e-06, "loss": 0.0562, "step": 3895 }, { "epoch": 1.8247748274652005, "grad_norm": 0.1836717575788498, "learning_rate": 1.201817361771837e-06, "loss": 0.0555, "step": 3900 }, { "epoch": 1.8271142823722073, "grad_norm": 0.17480716109275818, "learning_rate": 1.1701494060057606e-06, "loss": 0.0635, "step": 3905 }, { "epoch": 1.829453737279214, "grad_norm": 0.17331267893314362, "learning_rate": 1.138894266366572e-06, "loss": 0.0628, "step": 3910 }, { "epoch": 1.8317931921862205, "grad_norm": 0.15012997388839722, "learning_rate": 1.108052484302302e-06, "loss": 0.0569, "step": 3915 }, { "epoch": 1.8341326470932273, "grad_norm": 0.16882111132144928, "learning_rate": 1.0776245941001878e-06, "loss": 0.0558, "step": 3920 }, { "epoch": 1.836472102000234, "grad_norm": 0.17416958510875702, "learning_rate": 1.0476111228774178e-06, "loss": 0.0571, "step": 3925 }, { "epoch": 1.8388115569072405, "grad_norm": 0.1751696616411209, "learning_rate": 1.01801259057199e-06, "loss": 0.054, "step": 3930 }, { "epoch": 1.8411510118142473, "grad_norm": 0.17922084033489227, "learning_rate": 9.888295099337252e-07, "loss": 0.0611, "step": 3935 }, { "epoch": 1.843490466721254, "grad_norm": 0.20534135401248932, "learning_rate": 9.60062386515359e-07, "loss": 0.065, "step": 3940 }, { "epoch": 1.8458299216282605, "grad_norm": 0.17067112028598785, "learning_rate": 9.317117186638108e-07, "loss": 0.0591, "step": 3945 }, { "epoch": 1.8481693765352674, "grad_norm": 0.17611610889434814, "learning_rate": 9.037779975115235e-07, "loss": 0.0567, "step": 3950 }, { "epoch": 1.850508831442274, "grad_norm": 0.1562628149986267, "learning_rate": 8.762617069679846e-07, "loss": 0.0613, "step": 3955 }, { "epoch": 1.8528482863492806, "grad_norm": 0.1408929079771042, "learning_rate": 8.491633237113078e-07, "loss": 0.0593, "step": 3960 }, { "epoch": 1.8551877412562874, "grad_norm": 0.18245179951190948, "learning_rate": 8.224833171800173e-07, "loss": 0.0607, "step": 3965 }, { "epoch": 1.857527196163294, "grad_norm": 0.17572057247161865, "learning_rate": 7.962221495648708e-07, "loss": 0.0625, "step": 3970 }, { "epoch": 1.8598666510703006, "grad_norm": 0.1581767499446869, "learning_rate": 7.703802758008943e-07, "loss": 0.0629, "step": 3975 }, { "epoch": 1.8622061059773074, "grad_norm": 0.1815463751554489, "learning_rate": 7.44958143559471e-07, "loss": 0.0592, "step": 3980 }, { "epoch": 1.864545560884314, "grad_norm": 0.17877933382987976, "learning_rate": 7.199561932405952e-07, "loss": 0.0606, "step": 3985 }, { "epoch": 1.8668850157913206, "grad_norm": 0.17959001660346985, "learning_rate": 6.953748579652558e-07, "loss": 0.0687, "step": 3990 }, { "epoch": 1.8692244706983274, "grad_norm": 0.19308309257030487, "learning_rate": 6.712145635679096e-07, "loss": 0.0681, "step": 3995 }, { "epoch": 1.8715639256053338, "grad_norm": 0.1864127218723297, "learning_rate": 6.474757285891253e-07, "loss": 0.0599, "step": 4000 }, { "epoch": 1.8715639256053338, "eval_loss": 0.07669652253389359, "eval_runtime": 233.842, "eval_samples_per_second": 2.985, "eval_steps_per_second": 0.748, "step": 4000 }, { "epoch": 1.8739033805123406, "grad_norm": 0.17673756182193756, "learning_rate": 6.241587642683206e-07, "loss": 0.0624, "step": 4005 }, { "epoch": 1.8762428354193474, "grad_norm": 0.16083566844463348, "learning_rate": 6.012640745366477e-07, "loss": 0.0526, "step": 4010 }, { "epoch": 1.8785822903263538, "grad_norm": 0.1742735356092453, "learning_rate": 5.78792056009983e-07, "loss": 0.058, "step": 4015 }, { "epoch": 1.8809217452333606, "grad_norm": 0.1606372892856598, "learning_rate": 5.567430979820654e-07, "loss": 0.1072, "step": 4020 }, { "epoch": 1.8832612001403672, "grad_norm": 0.14640302956104279, "learning_rate": 5.35117582417749e-07, "loss": 0.0527, "step": 4025 }, { "epoch": 1.8856006550473738, "grad_norm": 0.17613400518894196, "learning_rate": 5.139158839463948e-07, "loss": 0.1046, "step": 4030 }, { "epoch": 1.8879401099543807, "grad_norm": 0.1592911034822464, "learning_rate": 4.931383698553643e-07, "loss": 0.0654, "step": 4035 }, { "epoch": 1.8902795648613873, "grad_norm": 0.16795676946640015, "learning_rate": 4.727854000836723e-07, "loss": 0.0818, "step": 4040 }, { "epoch": 1.8926190197683939, "grad_norm": 0.16940264403820038, "learning_rate": 4.528573272157416e-07, "loss": 0.0562, "step": 4045 }, { "epoch": 1.8949584746754007, "grad_norm": 0.17346608638763428, "learning_rate": 4.333544964753022e-07, "loss": 0.0563, "step": 4050 }, { "epoch": 1.8972979295824073, "grad_norm": 0.16219955682754517, "learning_rate": 4.1427724571940217e-07, "loss": 0.0677, "step": 4055 }, { "epoch": 1.8996373844894139, "grad_norm": 0.20030000805854797, "learning_rate": 3.9562590543256175e-07, "loss": 0.0651, "step": 4060 }, { "epoch": 1.9019768393964207, "grad_norm": 0.14878813922405243, "learning_rate": 3.7740079872105317e-07, "loss": 0.0555, "step": 4065 }, { "epoch": 1.9043162943034273, "grad_norm": 0.1636265218257904, "learning_rate": 3.5960224130728857e-07, "loss": 0.0634, "step": 4070 }, { "epoch": 1.906655749210434, "grad_norm": 0.1574837565422058, "learning_rate": 3.422305415243576e-07, "loss": 0.0564, "step": 4075 }, { "epoch": 1.9089952041174407, "grad_norm": 0.16645601391792297, "learning_rate": 3.2528600031069557e-07, "loss": 0.0586, "step": 4080 }, { "epoch": 1.9113346590244473, "grad_norm": 0.16494524478912354, "learning_rate": 3.087689112048542e-07, "loss": 0.0564, "step": 4085 }, { "epoch": 1.913674113931454, "grad_norm": 0.14947259426116943, "learning_rate": 2.926795603404198e-07, "loss": 0.0609, "step": 4090 }, { "epoch": 1.9160135688384607, "grad_norm": 0.1580345779657364, "learning_rate": 2.7701822644107536e-07, "loss": 0.0557, "step": 4095 }, { "epoch": 1.9183530237454673, "grad_norm": 0.15407022833824158, "learning_rate": 2.6178518081574064e-07, "loss": 0.0565, "step": 4100 }, { "epoch": 1.920692478652474, "grad_norm": 0.1476811021566391, "learning_rate": 2.46980687353901e-07, "loss": 0.0609, "step": 4105 }, { "epoch": 1.9230319335594808, "grad_norm": 0.18727290630340576, "learning_rate": 2.3260500252101636e-07, "loss": 0.0618, "step": 4110 }, { "epoch": 1.9253713884664871, "grad_norm": 0.18050184845924377, "learning_rate": 2.1865837535409162e-07, "loss": 0.0516, "step": 4115 }, { "epoch": 1.927710843373494, "grad_norm": 0.19610415399074554, "learning_rate": 2.0514104745734674e-07, "loss": 0.0646, "step": 4120 }, { "epoch": 1.9300502982805008, "grad_norm": 0.16899576783180237, "learning_rate": 1.9205325299805622e-07, "loss": 0.072, "step": 4125 }, { "epoch": 1.9323897531875072, "grad_norm": 0.17279615998268127, "learning_rate": 1.7939521870245779e-07, "loss": 0.0584, "step": 4130 }, { "epoch": 1.934729208094514, "grad_norm": 0.1838502138853073, "learning_rate": 1.6716716385186126e-07, "loss": 0.0576, "step": 4135 }, { "epoch": 1.9370686630015208, "grad_norm": 0.1634179651737213, "learning_rate": 1.5536930027882368e-07, "loss": 0.1097, "step": 4140 }, { "epoch": 1.9394081179085272, "grad_norm": 0.17711836099624634, "learning_rate": 1.4400183236349397e-07, "loss": 0.0637, "step": 4145 }, { "epoch": 1.941747572815534, "grad_norm": 0.1731368601322174, "learning_rate": 1.330649570300574e-07, "loss": 0.0683, "step": 4150 }, { "epoch": 1.9440870277225406, "grad_norm": 0.16300569474697113, "learning_rate": 1.2255886374334946e-07, "loss": 0.0539, "step": 4155 }, { "epoch": 1.9464264826295472, "grad_norm": 0.21170642971992493, "learning_rate": 1.1248373450554462e-07, "loss": 0.0618, "step": 4160 }, { "epoch": 1.948765937536554, "grad_norm": 0.14262203872203827, "learning_rate": 1.0283974385301986e-07, "loss": 0.06, "step": 4165 }, { "epoch": 1.9511053924435606, "grad_norm": 0.17411337792873383, "learning_rate": 9.3627058853335e-08, "loss": 0.0574, "step": 4170 }, { "epoch": 1.9534448473505672, "grad_norm": 0.15016813576221466, "learning_rate": 8.484583910232379e-08, "loss": 0.0595, "step": 4175 }, { "epoch": 1.955784302257574, "grad_norm": 0.16608966886997223, "learning_rate": 7.649623672134065e-08, "loss": 0.0605, "step": 4180 }, { "epoch": 1.9581237571645806, "grad_norm": 0.1695844531059265, "learning_rate": 6.857839635462104e-08, "loss": 0.0619, "step": 4185 }, { "epoch": 1.9604632120715872, "grad_norm": 0.1505594253540039, "learning_rate": 6.109245516677242e-08, "loss": 0.0623, "step": 4190 }, { "epoch": 1.962802666978594, "grad_norm": 0.17143121361732483, "learning_rate": 5.403854284040943e-08, "loss": 0.0573, "step": 4195 }, { "epoch": 1.9651421218856007, "grad_norm": 0.18096469342708588, "learning_rate": 4.741678157389739e-08, "loss": 0.0646, "step": 4200 }, { "epoch": 1.9674815767926073, "grad_norm": 0.17704518139362335, "learning_rate": 4.122728607923731e-08, "loss": 0.0602, "step": 4205 }, { "epoch": 1.969821031699614, "grad_norm": 0.16505348682403564, "learning_rate": 3.5470163580073046e-08, "loss": 0.0947, "step": 4210 }, { "epoch": 1.9721604866066207, "grad_norm": 0.13319675624370575, "learning_rate": 3.014551380985386e-08, "loss": 0.0566, "step": 4215 }, { "epoch": 1.9744999415136273, "grad_norm": 0.17260295152664185, "learning_rate": 2.525342901008032e-08, "loss": 0.0612, "step": 4220 }, { "epoch": 1.976839396420634, "grad_norm": 0.19151194393634796, "learning_rate": 2.0793993928724942e-08, "loss": 0.0652, "step": 4225 }, { "epoch": 1.9791788513276407, "grad_norm": 0.1425478458404541, "learning_rate": 1.676728581876119e-08, "loss": 0.0576, "step": 4230 }, { "epoch": 1.9815183062346473, "grad_norm": 0.14790157973766327, "learning_rate": 1.3173374436811769e-08, "loss": 0.0639, "step": 4235 }, { "epoch": 1.9838577611416541, "grad_norm": 0.17651847004890442, "learning_rate": 1.0012322041960676e-08, "loss": 0.0503, "step": 4240 }, { "epoch": 1.9861972160486605, "grad_norm": 0.16545896232128143, "learning_rate": 7.284183394656863e-09, "loss": 0.0557, "step": 4245 }, { "epoch": 1.9885366709556673, "grad_norm": 0.15864944458007812, "learning_rate": 4.98900575578165e-09, "loss": 0.0591, "step": 4250 }, { "epoch": 1.9908761258626742, "grad_norm": 0.1790982186794281, "learning_rate": 3.126828885816058e-09, "loss": 0.0607, "step": 4255 }, { "epoch": 1.9932155807696805, "grad_norm": 0.20370884239673615, "learning_rate": 1.6976850441552394e-09, "loss": 0.059, "step": 4260 }, { "epoch": 1.9955550356766873, "grad_norm": 0.16718322038650513, "learning_rate": 7.015989885589269e-10, "loss": 0.0575, "step": 4265 }, { "epoch": 1.997894490583694, "grad_norm": 0.17118395864963531, "learning_rate": 1.3858797471011376e-10, "loss": 0.0704, "step": 4270 }, { "epoch": 1.9997660545092995, "step": 4274, "total_flos": 3.0051315317703967e+18, "train_loss": 0.08305150749153994, "train_runtime": 80721.3637, "train_samples_per_second": 0.847, "train_steps_per_second": 0.053 } ], "logging_steps": 5, "max_steps": 4274, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.0051315317703967e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }