{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.989373007438894, "eval_steps": 500, "global_step": 4700, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0021253985122210413, "grad_norm": 512.0, "learning_rate": 4.2553191489361704e-07, "loss": 23.7791, "step": 1 }, { "epoch": 0.010626992561105207, "grad_norm": 488.0, "learning_rate": 2.1276595744680853e-06, "loss": 23.5159, "step": 5 }, { "epoch": 0.021253985122210415, "grad_norm": 244.0, "learning_rate": 4.255319148936171e-06, "loss": 20.5987, "step": 10 }, { "epoch": 0.031880977683315624, "grad_norm": 83.5, "learning_rate": 6.3829787234042555e-06, "loss": 17.7875, "step": 15 }, { "epoch": 0.04250797024442083, "grad_norm": 45.25, "learning_rate": 8.510638297872341e-06, "loss": 15.0704, "step": 20 }, { "epoch": 0.053134962805526036, "grad_norm": 19.125, "learning_rate": 1.0638297872340426e-05, "loss": 14.1038, "step": 25 }, { "epoch": 0.06376195536663125, "grad_norm": 9.5, "learning_rate": 1.2765957446808511e-05, "loss": 13.5893, "step": 30 }, { "epoch": 0.07438894792773645, "grad_norm": 5.96875, "learning_rate": 1.4893617021276596e-05, "loss": 12.8509, "step": 35 }, { "epoch": 0.08501594048884166, "grad_norm": 6.34375, "learning_rate": 1.7021276595744682e-05, "loss": 12.4395, "step": 40 }, { "epoch": 0.09564293304994687, "grad_norm": 8.875, "learning_rate": 1.9148936170212766e-05, "loss": 12.1862, "step": 45 }, { "epoch": 0.10626992561105207, "grad_norm": 14.5625, "learning_rate": 2.1276595744680852e-05, "loss": 11.1895, "step": 50 }, { "epoch": 0.11689691817215728, "grad_norm": 29.5, "learning_rate": 2.340425531914894e-05, "loss": 9.9832, "step": 55 }, { "epoch": 0.1275239107332625, "grad_norm": 46.75, "learning_rate": 2.5531914893617022e-05, "loss": 6.8574, "step": 60 }, { "epoch": 0.1381509032943677, "grad_norm": 9.8125, "learning_rate": 2.765957446808511e-05, "loss": 2.6513, "step": 65 }, { "epoch": 0.1487778958554729, "grad_norm": 2.9375, "learning_rate": 2.9787234042553192e-05, "loss": 1.4605, "step": 70 }, { "epoch": 0.1594048884165781, "grad_norm": 3.453125, "learning_rate": 3.191489361702128e-05, "loss": 1.1919, "step": 75 }, { "epoch": 0.17003188097768332, "grad_norm": 2.015625, "learning_rate": 3.4042553191489365e-05, "loss": 1.0938, "step": 80 }, { "epoch": 0.1806588735387885, "grad_norm": 2.265625, "learning_rate": 3.617021276595745e-05, "loss": 0.9763, "step": 85 }, { "epoch": 0.19128586609989373, "grad_norm": 4.96875, "learning_rate": 3.829787234042553e-05, "loss": 0.9307, "step": 90 }, { "epoch": 0.20191285866099895, "grad_norm": 1.5390625, "learning_rate": 4.0425531914893614e-05, "loss": 0.866, "step": 95 }, { "epoch": 0.21253985122210414, "grad_norm": 4.40625, "learning_rate": 4.2553191489361704e-05, "loss": 0.8456, "step": 100 }, { "epoch": 0.22316684378320936, "grad_norm": 2.6875, "learning_rate": 4.468085106382979e-05, "loss": 0.821, "step": 105 }, { "epoch": 0.23379383634431455, "grad_norm": 2.75, "learning_rate": 4.680851063829788e-05, "loss": 0.7823, "step": 110 }, { "epoch": 0.24442082890541977, "grad_norm": 2.953125, "learning_rate": 4.893617021276596e-05, "loss": 0.7831, "step": 115 }, { "epoch": 0.255047821466525, "grad_norm": 3.90625, "learning_rate": 5.1063829787234044e-05, "loss": 0.7535, "step": 120 }, { "epoch": 0.26567481402763016, "grad_norm": 4.3125, "learning_rate": 5.319148936170213e-05, "loss": 0.714, "step": 125 }, { "epoch": 0.2763018065887354, "grad_norm": 7.09375, "learning_rate": 5.531914893617022e-05, "loss": 0.7084, "step": 130 }, { "epoch": 0.2869287991498406, "grad_norm": 4.15625, "learning_rate": 5.744680851063831e-05, "loss": 0.7074, "step": 135 }, { "epoch": 0.2975557917109458, "grad_norm": 3.203125, "learning_rate": 5.9574468085106384e-05, "loss": 0.6913, "step": 140 }, { "epoch": 0.30818278427205104, "grad_norm": 1.9296875, "learning_rate": 6.170212765957447e-05, "loss": 0.6714, "step": 145 }, { "epoch": 0.3188097768331562, "grad_norm": 29.25, "learning_rate": 6.382978723404256e-05, "loss": 0.6761, "step": 150 }, { "epoch": 0.3294367693942614, "grad_norm": 2.921875, "learning_rate": 6.595744680851063e-05, "loss": 0.6739, "step": 155 }, { "epoch": 0.34006376195536664, "grad_norm": 3.921875, "learning_rate": 6.808510638297873e-05, "loss": 0.6572, "step": 160 }, { "epoch": 0.35069075451647186, "grad_norm": 8.25, "learning_rate": 7.021276595744681e-05, "loss": 0.644, "step": 165 }, { "epoch": 0.361317747077577, "grad_norm": 11.125, "learning_rate": 7.23404255319149e-05, "loss": 0.6699, "step": 170 }, { "epoch": 0.37194473963868224, "grad_norm": 2.828125, "learning_rate": 7.446808510638298e-05, "loss": 0.6613, "step": 175 }, { "epoch": 0.38257173219978746, "grad_norm": 4.84375, "learning_rate": 7.659574468085106e-05, "loss": 0.6376, "step": 180 }, { "epoch": 0.3931987247608927, "grad_norm": 3.796875, "learning_rate": 7.872340425531916e-05, "loss": 0.6444, "step": 185 }, { "epoch": 0.4038257173219979, "grad_norm": 2.078125, "learning_rate": 8.085106382978723e-05, "loss": 0.6349, "step": 190 }, { "epoch": 0.41445270988310307, "grad_norm": 3.53125, "learning_rate": 8.297872340425533e-05, "loss": 0.644, "step": 195 }, { "epoch": 0.4250797024442083, "grad_norm": 1.9921875, "learning_rate": 8.510638297872341e-05, "loss": 0.6224, "step": 200 }, { "epoch": 0.4357066950053135, "grad_norm": 3.1875, "learning_rate": 8.723404255319149e-05, "loss": 0.6213, "step": 205 }, { "epoch": 0.4463336875664187, "grad_norm": 6.375, "learning_rate": 8.936170212765958e-05, "loss": 0.6095, "step": 210 }, { "epoch": 0.4569606801275239, "grad_norm": 4.28125, "learning_rate": 9.148936170212766e-05, "loss": 0.6226, "step": 215 }, { "epoch": 0.4675876726886291, "grad_norm": 2.671875, "learning_rate": 9.361702127659576e-05, "loss": 0.6042, "step": 220 }, { "epoch": 0.4782146652497343, "grad_norm": 0.88671875, "learning_rate": 9.574468085106384e-05, "loss": 0.6009, "step": 225 }, { "epoch": 0.48884165781083955, "grad_norm": 1.2265625, "learning_rate": 9.787234042553192e-05, "loss": 0.5855, "step": 230 }, { "epoch": 0.49946865037194477, "grad_norm": 1.8671875, "learning_rate": 0.0001, "loss": 0.5729, "step": 235 }, { "epoch": 0.51009564293305, "grad_norm": 2.171875, "learning_rate": 0.00010212765957446809, "loss": 0.5751, "step": 240 }, { "epoch": 0.5207226354941552, "grad_norm": 13.5625, "learning_rate": 0.00010425531914893618, "loss": 0.5832, "step": 245 }, { "epoch": 0.5313496280552603, "grad_norm": 1.0859375, "learning_rate": 0.00010638297872340425, "loss": 0.5858, "step": 250 }, { "epoch": 0.5419766206163655, "grad_norm": 14.5, "learning_rate": 0.00010851063829787234, "loss": 0.5908, "step": 255 }, { "epoch": 0.5526036131774708, "grad_norm": 1.5546875, "learning_rate": 0.00011063829787234043, "loss": 0.5924, "step": 260 }, { "epoch": 0.563230605738576, "grad_norm": 1.7578125, "learning_rate": 0.00011276595744680852, "loss": 0.5704, "step": 265 }, { "epoch": 0.5738575982996812, "grad_norm": 0.7734375, "learning_rate": 0.00011489361702127661, "loss": 0.5695, "step": 270 }, { "epoch": 0.5844845908607864, "grad_norm": 2.25, "learning_rate": 0.00011702127659574468, "loss": 0.5654, "step": 275 }, { "epoch": 0.5951115834218916, "grad_norm": 2.703125, "learning_rate": 0.00011914893617021277, "loss": 0.5816, "step": 280 }, { "epoch": 0.6057385759829969, "grad_norm": 0.95703125, "learning_rate": 0.00012127659574468086, "loss": 0.5627, "step": 285 }, { "epoch": 0.6163655685441021, "grad_norm": 2.703125, "learning_rate": 0.00012340425531914893, "loss": 0.5557, "step": 290 }, { "epoch": 0.6269925611052072, "grad_norm": 1.8046875, "learning_rate": 0.00012553191489361702, "loss": 0.5571, "step": 295 }, { "epoch": 0.6376195536663124, "grad_norm": 4.46875, "learning_rate": 0.00012765957446808513, "loss": 0.544, "step": 300 }, { "epoch": 0.6482465462274176, "grad_norm": 3.421875, "learning_rate": 0.00012978723404255318, "loss": 0.5593, "step": 305 }, { "epoch": 0.6588735387885228, "grad_norm": 3.5625, "learning_rate": 0.00013191489361702127, "loss": 0.5513, "step": 310 }, { "epoch": 0.6695005313496281, "grad_norm": 9.0, "learning_rate": 0.00013404255319148938, "loss": 0.5681, "step": 315 }, { "epoch": 0.6801275239107333, "grad_norm": 1.03125, "learning_rate": 0.00013617021276595746, "loss": 0.554, "step": 320 }, { "epoch": 0.6907545164718385, "grad_norm": 1.484375, "learning_rate": 0.00013829787234042554, "loss": 0.5383, "step": 325 }, { "epoch": 0.7013815090329437, "grad_norm": 1.5234375, "learning_rate": 0.00014042553191489363, "loss": 0.5381, "step": 330 }, { "epoch": 0.7120085015940489, "grad_norm": 1.125, "learning_rate": 0.0001425531914893617, "loss": 0.5544, "step": 335 }, { "epoch": 0.722635494155154, "grad_norm": 0.96484375, "learning_rate": 0.0001446808510638298, "loss": 0.539, "step": 340 }, { "epoch": 0.7332624867162593, "grad_norm": 2.890625, "learning_rate": 0.00014680851063829788, "loss": 0.542, "step": 345 }, { "epoch": 0.7438894792773645, "grad_norm": 0.83984375, "learning_rate": 0.00014893617021276596, "loss": 0.531, "step": 350 }, { "epoch": 0.7545164718384697, "grad_norm": 3.015625, "learning_rate": 0.00015106382978723407, "loss": 0.5358, "step": 355 }, { "epoch": 0.7651434643995749, "grad_norm": 0.9296875, "learning_rate": 0.00015319148936170213, "loss": 0.5366, "step": 360 }, { "epoch": 0.7757704569606801, "grad_norm": 2.71875, "learning_rate": 0.0001553191489361702, "loss": 0.5359, "step": 365 }, { "epoch": 0.7863974495217854, "grad_norm": 1.8984375, "learning_rate": 0.00015744680851063832, "loss": 0.5506, "step": 370 }, { "epoch": 0.7970244420828906, "grad_norm": 2.484375, "learning_rate": 0.00015957446808510637, "loss": 0.5286, "step": 375 }, { "epoch": 0.8076514346439958, "grad_norm": 0.72265625, "learning_rate": 0.00016170212765957446, "loss": 0.5319, "step": 380 }, { "epoch": 0.8182784272051009, "grad_norm": 0.5859375, "learning_rate": 0.00016382978723404257, "loss": 0.5312, "step": 385 }, { "epoch": 0.8289054197662061, "grad_norm": 1.234375, "learning_rate": 0.00016595744680851065, "loss": 0.5284, "step": 390 }, { "epoch": 0.8395324123273114, "grad_norm": 2.859375, "learning_rate": 0.00016808510638297873, "loss": 0.5448, "step": 395 }, { "epoch": 0.8501594048884166, "grad_norm": 1.90625, "learning_rate": 0.00017021276595744682, "loss": 0.5292, "step": 400 }, { "epoch": 0.8607863974495218, "grad_norm": 1.0234375, "learning_rate": 0.0001723404255319149, "loss": 0.5246, "step": 405 }, { "epoch": 0.871413390010627, "grad_norm": 0.87109375, "learning_rate": 0.00017446808510638298, "loss": 0.5353, "step": 410 }, { "epoch": 0.8820403825717322, "grad_norm": 21.0, "learning_rate": 0.00017659574468085107, "loss": 0.5399, "step": 415 }, { "epoch": 0.8926673751328374, "grad_norm": 1.296875, "learning_rate": 0.00017872340425531915, "loss": 0.5405, "step": 420 }, { "epoch": 0.9032943676939427, "grad_norm": 0.58984375, "learning_rate": 0.00018085106382978726, "loss": 0.5233, "step": 425 }, { "epoch": 0.9139213602550478, "grad_norm": 2.21875, "learning_rate": 0.00018297872340425532, "loss": 0.5225, "step": 430 }, { "epoch": 0.924548352816153, "grad_norm": 0.6328125, "learning_rate": 0.0001851063829787234, "loss": 0.532, "step": 435 }, { "epoch": 0.9351753453772582, "grad_norm": 0.87109375, "learning_rate": 0.0001872340425531915, "loss": 0.5353, "step": 440 }, { "epoch": 0.9458023379383634, "grad_norm": 1.8671875, "learning_rate": 0.00018936170212765957, "loss": 0.5365, "step": 445 }, { "epoch": 0.9564293304994687, "grad_norm": 0.74609375, "learning_rate": 0.00019148936170212768, "loss": 0.5395, "step": 450 }, { "epoch": 0.9670563230605739, "grad_norm": 2.421875, "learning_rate": 0.00019361702127659576, "loss": 0.5309, "step": 455 }, { "epoch": 0.9776833156216791, "grad_norm": 4.96875, "learning_rate": 0.00019574468085106384, "loss": 0.5281, "step": 460 }, { "epoch": 0.9883103081827843, "grad_norm": 1.0234375, "learning_rate": 0.00019787234042553193, "loss": 0.5182, "step": 465 }, { "epoch": 0.9989373007438895, "grad_norm": 1.7109375, "learning_rate": 0.0002, "loss": 0.5262, "step": 470 }, { "epoch": 0.9989373007438895, "eval_loss": 1.3223822116851807, "eval_runtime": 0.8107, "eval_samples_per_second": 6.167, "eval_steps_per_second": 1.233, "step": 470 }, { "epoch": 1.0095642933049946, "grad_norm": 1.484375, "learning_rate": 0.00019999931050912237, "loss": 0.5078, "step": 475 }, { "epoch": 1.0201912858661, "grad_norm": 1.125, "learning_rate": 0.00019999724204599747, "loss": 0.4809, "step": 480 }, { "epoch": 1.030818278427205, "grad_norm": 2.703125, "learning_rate": 0.00019999379463914898, "loss": 0.4983, "step": 485 }, { "epoch": 1.0414452709883104, "grad_norm": 1.90625, "learning_rate": 0.00019998896833611603, "loss": 0.4957, "step": 490 }, { "epoch": 1.0520722635494155, "grad_norm": 0.6640625, "learning_rate": 0.00019998276320345247, "loss": 0.4901, "step": 495 }, { "epoch": 1.0626992561105206, "grad_norm": 2.171875, "learning_rate": 0.0001999751793267259, "loss": 0.4931, "step": 500 }, { "epoch": 1.073326248671626, "grad_norm": 1.40625, "learning_rate": 0.00019996621681051669, "loss": 0.4916, "step": 505 }, { "epoch": 1.083953241232731, "grad_norm": 2.75, "learning_rate": 0.0001999558757784162, "loss": 0.501, "step": 510 }, { "epoch": 1.0945802337938364, "grad_norm": 1.390625, "learning_rate": 0.00019994415637302547, "loss": 0.5011, "step": 515 }, { "epoch": 1.1052072263549415, "grad_norm": 2.625, "learning_rate": 0.0001999310587559529, "loss": 0.4966, "step": 520 }, { "epoch": 1.1158342189160468, "grad_norm": 2.703125, "learning_rate": 0.00019991658310781224, "loss": 0.5037, "step": 525 }, { "epoch": 1.126461211477152, "grad_norm": 1.0078125, "learning_rate": 0.00019990072962822007, "loss": 0.5134, "step": 530 }, { "epoch": 1.1370882040382573, "grad_norm": 0.5703125, "learning_rate": 0.00019988349853579295, "loss": 0.4997, "step": 535 }, { "epoch": 1.1477151965993624, "grad_norm": 1.0390625, "learning_rate": 0.00019986489006814452, "loss": 0.4965, "step": 540 }, { "epoch": 1.1583421891604675, "grad_norm": 7.1875, "learning_rate": 0.00019984490448188218, "loss": 0.4843, "step": 545 }, { "epoch": 1.1689691817215728, "grad_norm": 2.546875, "learning_rate": 0.00019982354205260347, "loss": 0.5033, "step": 550 }, { "epoch": 1.179596174282678, "grad_norm": 0.443359375, "learning_rate": 0.0001998008030748924, "loss": 0.4976, "step": 555 }, { "epoch": 1.1902231668437833, "grad_norm": 0.58984375, "learning_rate": 0.00019977668786231534, "loss": 0.497, "step": 560 }, { "epoch": 1.2008501594048884, "grad_norm": 0.79296875, "learning_rate": 0.00019975119674741664, "loss": 0.4973, "step": 565 }, { "epoch": 1.2114771519659937, "grad_norm": 1.171875, "learning_rate": 0.00019972433008171416, "loss": 0.5, "step": 570 }, { "epoch": 1.2221041445270988, "grad_norm": 1.609375, "learning_rate": 0.00019969608823569433, "loss": 0.4925, "step": 575 }, { "epoch": 1.2327311370882041, "grad_norm": 1.6875, "learning_rate": 0.00019966647159880703, "loss": 0.4848, "step": 580 }, { "epoch": 1.2433581296493093, "grad_norm": 2.734375, "learning_rate": 0.00019963548057946024, "loss": 0.5033, "step": 585 }, { "epoch": 1.2539851222104144, "grad_norm": 0.5390625, "learning_rate": 0.00019960311560501454, "loss": 0.4925, "step": 590 }, { "epoch": 1.2646121147715197, "grad_norm": 1.3125, "learning_rate": 0.000199569377121777, "loss": 0.487, "step": 595 }, { "epoch": 1.2752391073326248, "grad_norm": 0.44921875, "learning_rate": 0.0001995342655949951, "loss": 0.497, "step": 600 }, { "epoch": 1.2858660998937301, "grad_norm": 0.57421875, "learning_rate": 0.00019949778150885042, "loss": 0.4877, "step": 605 }, { "epoch": 1.2964930924548352, "grad_norm": 0.56640625, "learning_rate": 0.00019945992536645187, "loss": 0.4942, "step": 610 }, { "epoch": 1.3071200850159406, "grad_norm": 0.55859375, "learning_rate": 0.00019942069768982872, "loss": 0.4881, "step": 615 }, { "epoch": 1.3177470775770457, "grad_norm": 0.5390625, "learning_rate": 0.0001993800990199235, "loss": 0.4959, "step": 620 }, { "epoch": 1.328374070138151, "grad_norm": 0.61328125, "learning_rate": 0.0001993381299165844, "loss": 0.4883, "step": 625 }, { "epoch": 1.3390010626992561, "grad_norm": 0.546875, "learning_rate": 0.0001992947909585578, "loss": 0.4861, "step": 630 }, { "epoch": 1.3496280552603612, "grad_norm": 0.74609375, "learning_rate": 0.00019925008274347995, "loss": 0.4868, "step": 635 }, { "epoch": 1.3602550478214666, "grad_norm": 0.66796875, "learning_rate": 0.000199204005887869, "loss": 0.4979, "step": 640 }, { "epoch": 1.3708820403825717, "grad_norm": 1.0703125, "learning_rate": 0.00019915656102711634, "loss": 0.4885, "step": 645 }, { "epoch": 1.381509032943677, "grad_norm": 0.55078125, "learning_rate": 0.000199107748815478, "loss": 0.4928, "step": 650 }, { "epoch": 1.392136025504782, "grad_norm": 1.46875, "learning_rate": 0.00019905756992606548, "loss": 0.4879, "step": 655 }, { "epoch": 1.4027630180658874, "grad_norm": 0.70703125, "learning_rate": 0.00019900602505083648, "loss": 0.4854, "step": 660 }, { "epoch": 1.4133900106269925, "grad_norm": 5.0, "learning_rate": 0.00019895311490058542, "loss": 0.4932, "step": 665 }, { "epoch": 1.4240170031880979, "grad_norm": 0.5390625, "learning_rate": 0.0001988988402049336, "loss": 0.4911, "step": 670 }, { "epoch": 1.434643995749203, "grad_norm": 0.490234375, "learning_rate": 0.00019884320171231925, "loss": 0.4967, "step": 675 }, { "epoch": 1.445270988310308, "grad_norm": 0.859375, "learning_rate": 0.00019878620018998696, "loss": 0.4893, "step": 680 }, { "epoch": 1.4558979808714134, "grad_norm": 1.09375, "learning_rate": 0.00019872783642397733, "loss": 0.4849, "step": 685 }, { "epoch": 1.4665249734325185, "grad_norm": 0.431640625, "learning_rate": 0.00019866811121911607, "loss": 0.485, "step": 690 }, { "epoch": 1.4771519659936239, "grad_norm": 0.984375, "learning_rate": 0.00019860702539900287, "loss": 0.4993, "step": 695 }, { "epoch": 1.487778958554729, "grad_norm": 3.546875, "learning_rate": 0.000198544579806, "loss": 0.4858, "step": 700 }, { "epoch": 1.4984059511158343, "grad_norm": 0.8359375, "learning_rate": 0.00019848077530122083, "loss": 0.4781, "step": 705 }, { "epoch": 1.5090329436769394, "grad_norm": 1.109375, "learning_rate": 0.0001984156127645178, "loss": 0.4771, "step": 710 }, { "epoch": 1.5196599362380447, "grad_norm": 1.734375, "learning_rate": 0.00019834909309447045, "loss": 0.4913, "step": 715 }, { "epoch": 1.5302869287991498, "grad_norm": 0.62890625, "learning_rate": 0.00019828121720837286, "loss": 0.4957, "step": 720 }, { "epoch": 1.540913921360255, "grad_norm": 0.63671875, "learning_rate": 0.00019821198604222113, "loss": 0.4829, "step": 725 }, { "epoch": 1.5515409139213603, "grad_norm": 1.9140625, "learning_rate": 0.00019814140055070042, "loss": 0.4901, "step": 730 }, { "epoch": 1.5621679064824656, "grad_norm": 0.6640625, "learning_rate": 0.00019806946170717175, "loss": 0.4813, "step": 735 }, { "epoch": 1.5727948990435707, "grad_norm": 0.408203125, "learning_rate": 0.0001979961705036587, "loss": 0.472, "step": 740 }, { "epoch": 1.5834218916046758, "grad_norm": 0.484375, "learning_rate": 0.00019792152795083351, "loss": 0.4895, "step": 745 }, { "epoch": 1.594048884165781, "grad_norm": 1.21875, "learning_rate": 0.00019784553507800349, "loss": 0.4911, "step": 750 }, { "epoch": 1.6046758767268863, "grad_norm": 0.7421875, "learning_rate": 0.00019776819293309633, "loss": 0.4801, "step": 755 }, { "epoch": 1.6153028692879916, "grad_norm": 0.98046875, "learning_rate": 0.00019768950258264623, "loss": 0.488, "step": 760 }, { "epoch": 1.6259298618490967, "grad_norm": 0.63671875, "learning_rate": 0.00019760946511177872, "loss": 0.4795, "step": 765 }, { "epoch": 1.6365568544102018, "grad_norm": 0.94140625, "learning_rate": 0.0001975280816241959, "loss": 0.4797, "step": 770 }, { "epoch": 1.6471838469713072, "grad_norm": 5.125, "learning_rate": 0.00019744535324216127, "loss": 0.4913, "step": 775 }, { "epoch": 1.6578108395324125, "grad_norm": 0.7109375, "learning_rate": 0.00019736128110648407, "loss": 0.4751, "step": 780 }, { "epoch": 1.6684378320935176, "grad_norm": 6.5, "learning_rate": 0.00019727586637650373, "loss": 0.4956, "step": 785 }, { "epoch": 1.6790648246546227, "grad_norm": 1.1640625, "learning_rate": 0.0001971891102300738, "loss": 0.4918, "step": 790 }, { "epoch": 1.6896918172157278, "grad_norm": 0.53125, "learning_rate": 0.0001971010138635457, "loss": 0.4869, "step": 795 }, { "epoch": 1.7003188097768331, "grad_norm": 0.451171875, "learning_rate": 0.00019701157849175228, "loss": 0.4823, "step": 800 }, { "epoch": 1.7109458023379385, "grad_norm": 0.6484375, "learning_rate": 0.00019692080534799096, "loss": 0.4922, "step": 805 }, { "epoch": 1.7215727948990436, "grad_norm": 0.80078125, "learning_rate": 0.00019682869568400684, "loss": 0.4918, "step": 810 }, { "epoch": 1.7321997874601487, "grad_norm": 0.5, "learning_rate": 0.0001967352507699754, "loss": 0.4834, "step": 815 }, { "epoch": 1.742826780021254, "grad_norm": 6.21875, "learning_rate": 0.00019664047189448493, "loss": 0.4807, "step": 820 }, { "epoch": 1.7534537725823593, "grad_norm": 0.78515625, "learning_rate": 0.0001965443603645189, "loss": 0.486, "step": 825 }, { "epoch": 1.7640807651434645, "grad_norm": 1.3046875, "learning_rate": 0.00019644691750543767, "loss": 0.5035, "step": 830 }, { "epoch": 1.7747077577045696, "grad_norm": 1.40625, "learning_rate": 0.00019634814466096056, "loss": 0.4976, "step": 835 }, { "epoch": 1.7853347502656747, "grad_norm": 0.78515625, "learning_rate": 0.00019624804319314705, "loss": 0.5016, "step": 840 }, { "epoch": 1.79596174282678, "grad_norm": 1.125, "learning_rate": 0.0001961466144823781, "loss": 0.4952, "step": 845 }, { "epoch": 1.8065887353878853, "grad_norm": 0.5546875, "learning_rate": 0.00019604385992733715, "loss": 0.4849, "step": 850 }, { "epoch": 1.8172157279489904, "grad_norm": 0.61328125, "learning_rate": 0.00019593978094499076, "loss": 0.4792, "step": 855 }, { "epoch": 1.8278427205100956, "grad_norm": 0.494140625, "learning_rate": 0.00019583437897056915, "loss": 0.4858, "step": 860 }, { "epoch": 1.8384697130712009, "grad_norm": 0.85546875, "learning_rate": 0.00019572765545754626, "loss": 0.4905, "step": 865 }, { "epoch": 1.8490967056323062, "grad_norm": 1.3359375, "learning_rate": 0.00019561961187761985, "loss": 0.4786, "step": 870 }, { "epoch": 1.8597236981934113, "grad_norm": 0.5078125, "learning_rate": 0.00019551024972069126, "loss": 0.4694, "step": 875 }, { "epoch": 1.8703506907545164, "grad_norm": 0.91796875, "learning_rate": 0.00019539957049484458, "loss": 0.474, "step": 880 }, { "epoch": 1.8809776833156215, "grad_norm": 1.5, "learning_rate": 0.00019528757572632622, "loss": 0.4803, "step": 885 }, { "epoch": 1.8916046758767269, "grad_norm": 0.890625, "learning_rate": 0.00019517426695952358, "loss": 0.4798, "step": 890 }, { "epoch": 1.9022316684378322, "grad_norm": 0.57421875, "learning_rate": 0.00019505964575694385, "loss": 0.4779, "step": 895 }, { "epoch": 1.9128586609989373, "grad_norm": 0.9296875, "learning_rate": 0.0001949437136991925, "loss": 0.4796, "step": 900 }, { "epoch": 1.9234856535600424, "grad_norm": 0.6484375, "learning_rate": 0.00019482647238495152, "loss": 0.4808, "step": 905 }, { "epoch": 1.9341126461211477, "grad_norm": 0.69140625, "learning_rate": 0.00019470792343095718, "loss": 0.4837, "step": 910 }, { "epoch": 1.944739638682253, "grad_norm": 0.4921875, "learning_rate": 0.000194588068471978, "loss": 0.4861, "step": 915 }, { "epoch": 1.9553666312433582, "grad_norm": 0.5859375, "learning_rate": 0.0001944669091607919, "loss": 0.4632, "step": 920 }, { "epoch": 1.9659936238044633, "grad_norm": 0.76171875, "learning_rate": 0.00019434444716816374, "loss": 0.4767, "step": 925 }, { "epoch": 1.9766206163655684, "grad_norm": 0.47265625, "learning_rate": 0.00019422068418282202, "loss": 0.465, "step": 930 }, { "epoch": 1.9872476089266737, "grad_norm": 0.63671875, "learning_rate": 0.00019409562191143577, "loss": 0.4613, "step": 935 }, { "epoch": 1.997874601487779, "grad_norm": 0.484375, "learning_rate": 0.00019396926207859084, "loss": 0.4826, "step": 940 }, { "epoch": 2.0, "eval_loss": 1.3435465097427368, "eval_runtime": 0.7971, "eval_samples_per_second": 6.273, "eval_steps_per_second": 1.255, "step": 941 }, { "epoch": 2.008501594048884, "grad_norm": 0.734375, "learning_rate": 0.00019384160642676635, "loss": 0.4356, "step": 945 }, { "epoch": 2.0191285866099893, "grad_norm": 0.5625, "learning_rate": 0.00019371265671631037, "loss": 0.4104, "step": 950 }, { "epoch": 2.0297555791710944, "grad_norm": 0.484375, "learning_rate": 0.0001935824147254159, "loss": 0.4108, "step": 955 }, { "epoch": 2.0403825717322, "grad_norm": 0.59765625, "learning_rate": 0.00019345088225009626, "loss": 0.4187, "step": 960 }, { "epoch": 2.051009564293305, "grad_norm": 0.470703125, "learning_rate": 0.00019331806110416027, "loss": 0.4235, "step": 965 }, { "epoch": 2.06163655685441, "grad_norm": 0.42578125, "learning_rate": 0.0001931839531191873, "loss": 0.4266, "step": 970 }, { "epoch": 2.0722635494155153, "grad_norm": 0.7421875, "learning_rate": 0.000193048560144502, "loss": 0.4157, "step": 975 }, { "epoch": 2.082890541976621, "grad_norm": 0.84765625, "learning_rate": 0.00019291188404714878, "loss": 0.4255, "step": 980 }, { "epoch": 2.093517534537726, "grad_norm": 0.7421875, "learning_rate": 0.00019277392671186608, "loss": 0.4309, "step": 985 }, { "epoch": 2.104144527098831, "grad_norm": 1.0234375, "learning_rate": 0.0001926346900410604, "loss": 0.4288, "step": 990 }, { "epoch": 2.114771519659936, "grad_norm": 1.8359375, "learning_rate": 0.00019249417595478002, "loss": 0.4208, "step": 995 }, { "epoch": 2.1253985122210413, "grad_norm": 0.427734375, "learning_rate": 0.00019235238639068856, "loss": 0.4249, "step": 1000 }, { "epoch": 2.136025504782147, "grad_norm": 0.61328125, "learning_rate": 0.00019220932330403823, "loss": 0.427, "step": 1005 }, { "epoch": 2.146652497343252, "grad_norm": 0.6875, "learning_rate": 0.00019206498866764288, "loss": 0.4276, "step": 1010 }, { "epoch": 2.157279489904357, "grad_norm": 0.50390625, "learning_rate": 0.00019191938447185084, "loss": 0.4201, "step": 1015 }, { "epoch": 2.167906482465462, "grad_norm": 1.6328125, "learning_rate": 0.0001917725127245174, "loss": 0.4299, "step": 1020 }, { "epoch": 2.1785334750265677, "grad_norm": 0.5078125, "learning_rate": 0.00019162437545097719, "loss": 0.4268, "step": 1025 }, { "epoch": 2.189160467587673, "grad_norm": 1.2578125, "learning_rate": 0.0001914749746940161, "loss": 0.4241, "step": 1030 }, { "epoch": 2.199787460148778, "grad_norm": 0.69140625, "learning_rate": 0.00019132431251384335, "loss": 0.4297, "step": 1035 }, { "epoch": 2.210414452709883, "grad_norm": 0.423828125, "learning_rate": 0.00019117239098806295, "loss": 0.4275, "step": 1040 }, { "epoch": 2.221041445270988, "grad_norm": 0.55859375, "learning_rate": 0.000191019212211645, "loss": 0.4293, "step": 1045 }, { "epoch": 2.2316684378320937, "grad_norm": 0.427734375, "learning_rate": 0.00019086477829689685, "loss": 0.421, "step": 1050 }, { "epoch": 2.242295430393199, "grad_norm": 0.419921875, "learning_rate": 0.00019070909137343408, "loss": 0.4312, "step": 1055 }, { "epoch": 2.252922422954304, "grad_norm": 0.42578125, "learning_rate": 0.0001905521535881509, "loss": 0.4299, "step": 1060 }, { "epoch": 2.263549415515409, "grad_norm": 0.5859375, "learning_rate": 0.00019039396710519076, "loss": 0.4291, "step": 1065 }, { "epoch": 2.2741764080765146, "grad_norm": 0.404296875, "learning_rate": 0.00019023453410591635, "loss": 0.4236, "step": 1070 }, { "epoch": 2.2848034006376197, "grad_norm": 0.51171875, "learning_rate": 0.00019007385678887975, "loss": 0.4201, "step": 1075 }, { "epoch": 2.2954303931987248, "grad_norm": 0.47265625, "learning_rate": 0.00018991193736979175, "loss": 0.4257, "step": 1080 }, { "epoch": 2.30605738575983, "grad_norm": 0.439453125, "learning_rate": 0.0001897487780814916, "loss": 0.4219, "step": 1085 }, { "epoch": 2.316684378320935, "grad_norm": 0.50390625, "learning_rate": 0.00018958438117391618, "loss": 0.4212, "step": 1090 }, { "epoch": 2.3273113708820405, "grad_norm": 0.4765625, "learning_rate": 0.00018941874891406882, "loss": 0.4351, "step": 1095 }, { "epoch": 2.3379383634431457, "grad_norm": 0.474609375, "learning_rate": 0.00018925188358598813, "loss": 0.4141, "step": 1100 }, { "epoch": 2.3485653560042508, "grad_norm": 1.8515625, "learning_rate": 0.0001890837874907166, "loss": 0.4243, "step": 1105 }, { "epoch": 2.359192348565356, "grad_norm": 0.66015625, "learning_rate": 0.00018891446294626866, "loss": 0.4408, "step": 1110 }, { "epoch": 2.369819341126461, "grad_norm": 0.53515625, "learning_rate": 0.00018874391228759893, "loss": 0.4277, "step": 1115 }, { "epoch": 2.3804463336875665, "grad_norm": 0.546875, "learning_rate": 0.00018857213786656985, "loss": 0.4279, "step": 1120 }, { "epoch": 2.3910733262486716, "grad_norm": 0.77734375, "learning_rate": 0.00018839914205191936, "loss": 0.4259, "step": 1125 }, { "epoch": 2.4017003188097767, "grad_norm": 0.79296875, "learning_rate": 0.0001882249272292282, "loss": 0.4287, "step": 1130 }, { "epoch": 2.412327311370882, "grad_norm": 0.76953125, "learning_rate": 0.00018804949580088692, "loss": 0.4394, "step": 1135 }, { "epoch": 2.4229543039319874, "grad_norm": 0.435546875, "learning_rate": 0.00018787285018606297, "loss": 0.4268, "step": 1140 }, { "epoch": 2.4335812964930925, "grad_norm": 0.462890625, "learning_rate": 0.00018769499282066717, "loss": 0.4271, "step": 1145 }, { "epoch": 2.4442082890541976, "grad_norm": 0.6484375, "learning_rate": 0.00018751592615732005, "loss": 0.4338, "step": 1150 }, { "epoch": 2.4548352816153027, "grad_norm": 0.89453125, "learning_rate": 0.0001873356526653183, "loss": 0.4306, "step": 1155 }, { "epoch": 2.4654622741764083, "grad_norm": 1.09375, "learning_rate": 0.0001871541748306005, "loss": 0.4374, "step": 1160 }, { "epoch": 2.4760892667375134, "grad_norm": 0.44921875, "learning_rate": 0.00018697149515571284, "loss": 0.429, "step": 1165 }, { "epoch": 2.4867162592986185, "grad_norm": 1.3203125, "learning_rate": 0.00018678761615977468, "loss": 0.4351, "step": 1170 }, { "epoch": 2.4973432518597236, "grad_norm": 0.47265625, "learning_rate": 0.00018660254037844388, "loss": 0.4264, "step": 1175 }, { "epoch": 2.5079702444208287, "grad_norm": 0.58203125, "learning_rate": 0.00018641627036388169, "loss": 0.431, "step": 1180 }, { "epoch": 2.5185972369819343, "grad_norm": 0.41015625, "learning_rate": 0.00018622880868471756, "loss": 0.4265, "step": 1185 }, { "epoch": 2.5292242295430394, "grad_norm": 0.400390625, "learning_rate": 0.00018604015792601396, "loss": 0.4259, "step": 1190 }, { "epoch": 2.5398512221041445, "grad_norm": 7.6875, "learning_rate": 0.00018585032068923032, "loss": 0.4309, "step": 1195 }, { "epoch": 2.5504782146652496, "grad_norm": 0.4296875, "learning_rate": 0.00018565929959218758, "loss": 0.4254, "step": 1200 }, { "epoch": 2.5611052072263547, "grad_norm": 0.42578125, "learning_rate": 0.00018546709726903178, "loss": 0.4253, "step": 1205 }, { "epoch": 2.5717321997874603, "grad_norm": 1.0625, "learning_rate": 0.0001852737163701979, "loss": 0.4403, "step": 1210 }, { "epoch": 2.5823591923485654, "grad_norm": 0.640625, "learning_rate": 0.00018507915956237326, "loss": 0.4326, "step": 1215 }, { "epoch": 2.5929861849096705, "grad_norm": 2.25, "learning_rate": 0.00018488342952846073, "loss": 0.437, "step": 1220 }, { "epoch": 2.603613177470776, "grad_norm": 0.90625, "learning_rate": 0.00018468652896754177, "loss": 0.4344, "step": 1225 }, { "epoch": 2.614240170031881, "grad_norm": 0.87890625, "learning_rate": 0.0001844884605948392, "loss": 0.4387, "step": 1230 }, { "epoch": 2.6248671625929862, "grad_norm": 0.66796875, "learning_rate": 0.0001842892271416797, "loss": 0.4371, "step": 1235 }, { "epoch": 2.6354941551540914, "grad_norm": 0.515625, "learning_rate": 0.00018408883135545632, "loss": 0.4345, "step": 1240 }, { "epoch": 2.6461211477151965, "grad_norm": 0.8515625, "learning_rate": 0.00018388727599959033, "loss": 0.4405, "step": 1245 }, { "epoch": 2.656748140276302, "grad_norm": 1.078125, "learning_rate": 0.00018368456385349334, "loss": 0.4369, "step": 1250 }, { "epoch": 2.667375132837407, "grad_norm": 0.50390625, "learning_rate": 0.0001834806977125288, "loss": 0.4354, "step": 1255 }, { "epoch": 2.6780021253985122, "grad_norm": 0.41796875, "learning_rate": 0.0001832756803879737, "loss": 0.4297, "step": 1260 }, { "epoch": 2.6886291179596173, "grad_norm": 0.578125, "learning_rate": 0.00018306951470697946, "loss": 0.4268, "step": 1265 }, { "epoch": 2.6992561105207225, "grad_norm": 0.44140625, "learning_rate": 0.0001828622035125332, "loss": 0.4299, "step": 1270 }, { "epoch": 2.709883103081828, "grad_norm": 0.8046875, "learning_rate": 0.0001826537496634186, "loss": 0.4362, "step": 1275 }, { "epoch": 2.720510095642933, "grad_norm": 1.484375, "learning_rate": 0.00018244415603417603, "loss": 0.4364, "step": 1280 }, { "epoch": 2.731137088204038, "grad_norm": 0.953125, "learning_rate": 0.0001822334255150635, "loss": 0.439, "step": 1285 }, { "epoch": 2.7417640807651433, "grad_norm": 0.921875, "learning_rate": 0.00018202156101201645, "loss": 0.4342, "step": 1290 }, { "epoch": 2.7523910733262484, "grad_norm": 0.55859375, "learning_rate": 0.0001818085654466076, "loss": 0.4308, "step": 1295 }, { "epoch": 2.763018065887354, "grad_norm": 0.494140625, "learning_rate": 0.00018159444175600703, "loss": 0.4306, "step": 1300 }, { "epoch": 2.773645058448459, "grad_norm": 1.125, "learning_rate": 0.00018137919289294135, "loss": 0.4342, "step": 1305 }, { "epoch": 2.784272051009564, "grad_norm": 0.53125, "learning_rate": 0.00018116282182565311, "loss": 0.428, "step": 1310 }, { "epoch": 2.7948990435706698, "grad_norm": 0.50390625, "learning_rate": 0.00018094533153785984, "loss": 0.4221, "step": 1315 }, { "epoch": 2.805526036131775, "grad_norm": 1.5859375, "learning_rate": 0.00018072672502871296, "loss": 0.433, "step": 1320 }, { "epoch": 2.81615302869288, "grad_norm": 0.44140625, "learning_rate": 0.0001805070053127563, "loss": 0.4306, "step": 1325 }, { "epoch": 2.826780021253985, "grad_norm": 0.6171875, "learning_rate": 0.00018028617541988472, "loss": 0.4217, "step": 1330 }, { "epoch": 2.83740701381509, "grad_norm": 0.49609375, "learning_rate": 0.00018006423839530205, "loss": 0.426, "step": 1335 }, { "epoch": 2.8480340063761957, "grad_norm": 0.6484375, "learning_rate": 0.00017984119729947944, "loss": 0.4402, "step": 1340 }, { "epoch": 2.858660998937301, "grad_norm": 0.4609375, "learning_rate": 0.00017961705520811293, "loss": 0.4306, "step": 1345 }, { "epoch": 2.869287991498406, "grad_norm": 0.375, "learning_rate": 0.000179391815212081, "loss": 0.434, "step": 1350 }, { "epoch": 2.879914984059511, "grad_norm": 0.70703125, "learning_rate": 0.00017916548041740213, "loss": 0.4295, "step": 1355 }, { "epoch": 2.890541976620616, "grad_norm": 0.72265625, "learning_rate": 0.0001789380539451919, "loss": 0.4373, "step": 1360 }, { "epoch": 2.9011689691817217, "grad_norm": 0.4296875, "learning_rate": 0.00017870953893161975, "loss": 0.4287, "step": 1365 }, { "epoch": 2.911795961742827, "grad_norm": 0.5625, "learning_rate": 0.0001784799385278661, "loss": 0.4358, "step": 1370 }, { "epoch": 2.922422954303932, "grad_norm": 0.419921875, "learning_rate": 0.00017824925590007864, "loss": 0.4304, "step": 1375 }, { "epoch": 2.933049946865037, "grad_norm": 0.98046875, "learning_rate": 0.0001780174942293287, "loss": 0.4296, "step": 1380 }, { "epoch": 2.943676939426142, "grad_norm": 0.6796875, "learning_rate": 0.00017778465671156743, "loss": 0.4326, "step": 1385 }, { "epoch": 2.9543039319872477, "grad_norm": 0.486328125, "learning_rate": 0.00017755074655758174, "loss": 0.4348, "step": 1390 }, { "epoch": 2.964930924548353, "grad_norm": 0.61328125, "learning_rate": 0.0001773157669929499, "loss": 0.4345, "step": 1395 }, { "epoch": 2.975557917109458, "grad_norm": 0.51171875, "learning_rate": 0.00017707972125799735, "loss": 0.4259, "step": 1400 }, { "epoch": 2.9861849096705635, "grad_norm": 0.5859375, "learning_rate": 0.00017684261260775163, "loss": 0.4306, "step": 1405 }, { "epoch": 2.9968119022316686, "grad_norm": 0.60546875, "learning_rate": 0.0001766044443118978, "loss": 0.4369, "step": 1410 }, { "epoch": 2.9989373007438895, "eval_loss": 1.4787013530731201, "eval_runtime": 0.8123, "eval_samples_per_second": 6.155, "eval_steps_per_second": 1.231, "step": 1411 }, { "epoch": 3.0074388947927737, "grad_norm": 0.54296875, "learning_rate": 0.00017636521965473323, "loss": 0.3835, "step": 1415 }, { "epoch": 3.018065887353879, "grad_norm": 0.455078125, "learning_rate": 0.0001761249419351222, "loss": 0.3594, "step": 1420 }, { "epoch": 3.028692879914984, "grad_norm": 0.478515625, "learning_rate": 0.00017588361446645073, "loss": 0.3596, "step": 1425 }, { "epoch": 3.0393198724760895, "grad_norm": 0.51171875, "learning_rate": 0.00017564124057658056, "loss": 0.3504, "step": 1430 }, { "epoch": 3.0499468650371946, "grad_norm": 0.4375, "learning_rate": 0.00017539782360780334, "loss": 0.3575, "step": 1435 }, { "epoch": 3.0605738575982997, "grad_norm": 0.419921875, "learning_rate": 0.00017515336691679477, "loss": 0.3619, "step": 1440 }, { "epoch": 3.071200850159405, "grad_norm": 0.41015625, "learning_rate": 0.0001749078738745679, "loss": 0.3678, "step": 1445 }, { "epoch": 3.08182784272051, "grad_norm": 0.57421875, "learning_rate": 0.0001746613478664271, "loss": 0.3551, "step": 1450 }, { "epoch": 3.0924548352816155, "grad_norm": 0.75, "learning_rate": 0.00017441379229192098, "loss": 0.3668, "step": 1455 }, { "epoch": 3.1030818278427206, "grad_norm": 0.498046875, "learning_rate": 0.00017416521056479577, "loss": 0.3664, "step": 1460 }, { "epoch": 3.1137088204038257, "grad_norm": 1.6640625, "learning_rate": 0.0001739156061129481, "loss": 0.3709, "step": 1465 }, { "epoch": 3.124335812964931, "grad_norm": 0.8125, "learning_rate": 0.0001736649823783779, "loss": 0.3612, "step": 1470 }, { "epoch": 3.134962805526036, "grad_norm": 0.64453125, "learning_rate": 0.00017341334281714064, "loss": 0.3619, "step": 1475 }, { "epoch": 3.1455897980871415, "grad_norm": 1.0859375, "learning_rate": 0.00017316069089930007, "loss": 0.37, "step": 1480 }, { "epoch": 3.1562167906482466, "grad_norm": 0.5859375, "learning_rate": 0.00017290703010887994, "loss": 0.3717, "step": 1485 }, { "epoch": 3.1668437832093517, "grad_norm": 0.53515625, "learning_rate": 0.00017265236394381633, "loss": 0.3749, "step": 1490 }, { "epoch": 3.177470775770457, "grad_norm": 0.5, "learning_rate": 0.00017239669591590916, "loss": 0.3754, "step": 1495 }, { "epoch": 3.1880977683315623, "grad_norm": 0.55859375, "learning_rate": 0.00017214002955077393, "loss": 0.3726, "step": 1500 }, { "epoch": 3.1987247608926674, "grad_norm": 0.53125, "learning_rate": 0.00017188236838779295, "loss": 0.3726, "step": 1505 }, { "epoch": 3.2093517534537725, "grad_norm": 0.7265625, "learning_rate": 0.00017162371598006666, "loss": 0.3798, "step": 1510 }, { "epoch": 3.2199787460148777, "grad_norm": 0.94140625, "learning_rate": 0.00017136407589436457, "loss": 0.3767, "step": 1515 }, { "epoch": 3.230605738575983, "grad_norm": 0.5703125, "learning_rate": 0.0001711034517110761, "loss": 0.3763, "step": 1520 }, { "epoch": 3.2412327311370883, "grad_norm": 0.53125, "learning_rate": 0.00017084184702416115, "loss": 0.3703, "step": 1525 }, { "epoch": 3.2518597236981934, "grad_norm": 1.0234375, "learning_rate": 0.0001705792654411007, "loss": 0.378, "step": 1530 }, { "epoch": 3.2624867162592985, "grad_norm": 0.5234375, "learning_rate": 0.00017031571058284678, "loss": 0.3818, "step": 1535 }, { "epoch": 3.2731137088204036, "grad_norm": 0.55859375, "learning_rate": 0.00017005118608377288, "loss": 0.3709, "step": 1540 }, { "epoch": 3.283740701381509, "grad_norm": 0.59375, "learning_rate": 0.00016978569559162357, "loss": 0.3761, "step": 1545 }, { "epoch": 3.2943676939426143, "grad_norm": 0.44921875, "learning_rate": 0.00016951924276746425, "loss": 0.3786, "step": 1550 }, { "epoch": 3.3049946865037194, "grad_norm": 0.451171875, "learning_rate": 0.00016925183128563078, "loss": 0.3661, "step": 1555 }, { "epoch": 3.3156216790648245, "grad_norm": 0.490234375, "learning_rate": 0.00016898346483367867, "loss": 0.3721, "step": 1560 }, { "epoch": 3.3262486716259296, "grad_norm": 0.4375, "learning_rate": 0.0001687141471123324, "loss": 0.3752, "step": 1565 }, { "epoch": 3.336875664187035, "grad_norm": 0.4921875, "learning_rate": 0.00016844388183543418, "loss": 0.3784, "step": 1570 }, { "epoch": 3.3475026567481403, "grad_norm": 0.44921875, "learning_rate": 0.00016817267272989286, "loss": 0.3807, "step": 1575 }, { "epoch": 3.3581296493092454, "grad_norm": 0.55859375, "learning_rate": 0.00016790052353563253, "loss": 0.3887, "step": 1580 }, { "epoch": 3.3687566418703505, "grad_norm": 0.6484375, "learning_rate": 0.0001676274380055409, "loss": 0.3831, "step": 1585 }, { "epoch": 3.379383634431456, "grad_norm": 0.45703125, "learning_rate": 0.00016735341990541764, "loss": 0.3772, "step": 1590 }, { "epoch": 3.390010626992561, "grad_norm": 0.56640625, "learning_rate": 0.00016707847301392236, "loss": 0.3799, "step": 1595 }, { "epoch": 3.4006376195536663, "grad_norm": 0.431640625, "learning_rate": 0.0001668026011225225, "loss": 0.3854, "step": 1600 }, { "epoch": 3.4112646121147714, "grad_norm": 0.439453125, "learning_rate": 0.00016652580803544112, "loss": 0.3725, "step": 1605 }, { "epoch": 3.421891604675877, "grad_norm": 0.439453125, "learning_rate": 0.00016624809756960444, "loss": 0.3785, "step": 1610 }, { "epoch": 3.432518597236982, "grad_norm": 0.490234375, "learning_rate": 0.00016596947355458904, "loss": 0.3803, "step": 1615 }, { "epoch": 3.443145589798087, "grad_norm": 0.640625, "learning_rate": 0.0001656899398325693, "loss": 0.3893, "step": 1620 }, { "epoch": 3.4537725823591923, "grad_norm": 0.53515625, "learning_rate": 0.00016540950025826422, "loss": 0.3776, "step": 1625 }, { "epoch": 3.4643995749202974, "grad_norm": 0.494140625, "learning_rate": 0.0001651281586988844, "loss": 0.3802, "step": 1630 }, { "epoch": 3.475026567481403, "grad_norm": 0.73046875, "learning_rate": 0.00016484591903407857, "loss": 0.3765, "step": 1635 }, { "epoch": 3.485653560042508, "grad_norm": 0.57421875, "learning_rate": 0.00016456278515588024, "loss": 0.3843, "step": 1640 }, { "epoch": 3.496280552603613, "grad_norm": 0.59375, "learning_rate": 0.00016427876096865394, "loss": 0.3748, "step": 1645 }, { "epoch": 3.5069075451647183, "grad_norm": 0.4609375, "learning_rate": 0.00016399385038904138, "loss": 0.3785, "step": 1650 }, { "epoch": 3.5175345377258234, "grad_norm": 0.466796875, "learning_rate": 0.00016370805734590747, "loss": 0.3879, "step": 1655 }, { "epoch": 3.528161530286929, "grad_norm": 0.4609375, "learning_rate": 0.00016342138578028613, "loss": 0.382, "step": 1660 }, { "epoch": 3.538788522848034, "grad_norm": 0.423828125, "learning_rate": 0.00016313383964532596, "loss": 0.3915, "step": 1665 }, { "epoch": 3.549415515409139, "grad_norm": 0.42578125, "learning_rate": 0.00016284542290623567, "loss": 0.3767, "step": 1670 }, { "epoch": 3.5600425079702447, "grad_norm": 0.443359375, "learning_rate": 0.0001625561395402295, "loss": 0.373, "step": 1675 }, { "epoch": 3.57066950053135, "grad_norm": 0.4453125, "learning_rate": 0.00016226599353647228, "loss": 0.3811, "step": 1680 }, { "epoch": 3.581296493092455, "grad_norm": 0.43359375, "learning_rate": 0.00016197498889602448, "loss": 0.3842, "step": 1685 }, { "epoch": 3.59192348565356, "grad_norm": 0.59765625, "learning_rate": 0.00016168312963178697, "loss": 0.3864, "step": 1690 }, { "epoch": 3.602550478214665, "grad_norm": 1.2578125, "learning_rate": 0.00016139041976844583, "loss": 0.3835, "step": 1695 }, { "epoch": 3.6131774707757707, "grad_norm": 0.578125, "learning_rate": 0.00016109686334241655, "loss": 0.3801, "step": 1700 }, { "epoch": 3.623804463336876, "grad_norm": 0.5078125, "learning_rate": 0.00016080246440178874, "loss": 0.382, "step": 1705 }, { "epoch": 3.634431455897981, "grad_norm": 0.66796875, "learning_rate": 0.00016050722700627012, "loss": 0.3923, "step": 1710 }, { "epoch": 3.645058448459086, "grad_norm": 0.75390625, "learning_rate": 0.00016021115522713047, "loss": 0.3836, "step": 1715 }, { "epoch": 3.655685441020191, "grad_norm": 0.77734375, "learning_rate": 0.0001599142531471456, "loss": 0.3801, "step": 1720 }, { "epoch": 3.6663124335812967, "grad_norm": 0.61328125, "learning_rate": 0.00015961652486054103, "loss": 0.3871, "step": 1725 }, { "epoch": 3.6769394261424018, "grad_norm": 0.51171875, "learning_rate": 0.00015931797447293552, "loss": 0.3843, "step": 1730 }, { "epoch": 3.687566418703507, "grad_norm": 0.4765625, "learning_rate": 0.00015901860610128448, "loss": 0.3846, "step": 1735 }, { "epoch": 3.698193411264612, "grad_norm": 0.50390625, "learning_rate": 0.00015871842387382305, "loss": 0.3788, "step": 1740 }, { "epoch": 3.708820403825717, "grad_norm": 0.66015625, "learning_rate": 0.00015841743193000944, "loss": 0.3899, "step": 1745 }, { "epoch": 3.7194473963868226, "grad_norm": 0.50390625, "learning_rate": 0.00015811563442046767, "loss": 0.3863, "step": 1750 }, { "epoch": 3.7300743889479278, "grad_norm": 0.396484375, "learning_rate": 0.00015781303550693024, "loss": 0.3822, "step": 1755 }, { "epoch": 3.740701381509033, "grad_norm": 0.5859375, "learning_rate": 0.00015750963936218105, "loss": 0.3821, "step": 1760 }, { "epoch": 3.7513283740701384, "grad_norm": 0.70703125, "learning_rate": 0.00015720545016999752, "loss": 0.3809, "step": 1765 }, { "epoch": 3.761955366631243, "grad_norm": 0.515625, "learning_rate": 0.00015690047212509316, "loss": 0.3809, "step": 1770 }, { "epoch": 3.7725823591923486, "grad_norm": 0.5859375, "learning_rate": 0.00015659470943305955, "loss": 0.3815, "step": 1775 }, { "epoch": 3.7832093517534537, "grad_norm": 0.443359375, "learning_rate": 0.00015628816631030836, "loss": 0.3808, "step": 1780 }, { "epoch": 3.793836344314559, "grad_norm": 0.44140625, "learning_rate": 0.00015598084698401342, "loss": 0.38, "step": 1785 }, { "epoch": 3.8044633368756644, "grad_norm": 0.53125, "learning_rate": 0.00015567275569205218, "loss": 0.3851, "step": 1790 }, { "epoch": 3.8150903294367695, "grad_norm": 0.44140625, "learning_rate": 0.00015536389668294724, "loss": 0.3917, "step": 1795 }, { "epoch": 3.8257173219978746, "grad_norm": 0.40625, "learning_rate": 0.00015505427421580808, "loss": 0.3799, "step": 1800 }, { "epoch": 3.8363443145589797, "grad_norm": 0.45703125, "learning_rate": 0.000154743892560272, "loss": 0.3846, "step": 1805 }, { "epoch": 3.846971307120085, "grad_norm": 0.46484375, "learning_rate": 0.00015443275599644538, "loss": 0.3865, "step": 1810 }, { "epoch": 3.8575982996811904, "grad_norm": 0.486328125, "learning_rate": 0.0001541208688148447, "loss": 0.3802, "step": 1815 }, { "epoch": 3.8682252922422955, "grad_norm": 0.462890625, "learning_rate": 0.00015380823531633729, "loss": 0.3823, "step": 1820 }, { "epoch": 3.8788522848034006, "grad_norm": 0.61328125, "learning_rate": 0.00015349485981208202, "loss": 0.3866, "step": 1825 }, { "epoch": 3.8894792773645057, "grad_norm": 0.490234375, "learning_rate": 0.00015318074662346994, "loss": 0.3758, "step": 1830 }, { "epoch": 3.900106269925611, "grad_norm": 0.5625, "learning_rate": 0.00015286590008206465, "loss": 0.3858, "step": 1835 }, { "epoch": 3.9107332624867164, "grad_norm": 0.65625, "learning_rate": 0.00015255032452954245, "loss": 0.3867, "step": 1840 }, { "epoch": 3.9213602550478215, "grad_norm": 0.42578125, "learning_rate": 0.00015223402431763269, "loss": 0.3893, "step": 1845 }, { "epoch": 3.9319872476089266, "grad_norm": 0.58984375, "learning_rate": 0.00015191700380805752, "loss": 0.3791, "step": 1850 }, { "epoch": 3.942614240170032, "grad_norm": 0.51953125, "learning_rate": 0.00015159926737247202, "loss": 0.3932, "step": 1855 }, { "epoch": 3.953241232731137, "grad_norm": 0.69921875, "learning_rate": 0.00015128081939240357, "loss": 0.3794, "step": 1860 }, { "epoch": 3.9638682252922424, "grad_norm": 0.54296875, "learning_rate": 0.00015096166425919175, "loss": 0.3857, "step": 1865 }, { "epoch": 3.9744952178533475, "grad_norm": 0.412109375, "learning_rate": 0.00015064180637392764, "loss": 0.3887, "step": 1870 }, { "epoch": 3.9851222104144526, "grad_norm": 0.51171875, "learning_rate": 0.0001503212501473931, "loss": 0.3886, "step": 1875 }, { "epoch": 3.995749202975558, "grad_norm": 0.5078125, "learning_rate": 0.00015000000000000001, "loss": 0.3819, "step": 1880 }, { "epoch": 4.0, "eval_loss": 1.7431671619415283, "eval_runtime": 0.7941, "eval_samples_per_second": 6.296, "eval_steps_per_second": 1.259, "step": 1882 }, { "epoch": 4.006376195536663, "grad_norm": 0.47265625, "learning_rate": 0.0001496780603617293, "loss": 0.3464, "step": 1885 }, { "epoch": 4.017003188097768, "grad_norm": 0.46484375, "learning_rate": 0.00014935543567206984, "loss": 0.3125, "step": 1890 }, { "epoch": 4.027630180658874, "grad_norm": 0.5546875, "learning_rate": 0.00014903213037995724, "loss": 0.3159, "step": 1895 }, { "epoch": 4.038257173219979, "grad_norm": 0.478515625, "learning_rate": 0.00014870814894371245, "loss": 0.3227, "step": 1900 }, { "epoch": 4.048884165781084, "grad_norm": 0.470703125, "learning_rate": 0.00014838349583098045, "loss": 0.3122, "step": 1905 }, { "epoch": 4.059511158342189, "grad_norm": 0.43359375, "learning_rate": 0.00014805817551866838, "loss": 0.3194, "step": 1910 }, { "epoch": 4.070138150903294, "grad_norm": 0.458984375, "learning_rate": 0.00014773219249288402, "loss": 0.3123, "step": 1915 }, { "epoch": 4.0807651434644, "grad_norm": 0.4765625, "learning_rate": 0.00014740555124887375, "loss": 0.3254, "step": 1920 }, { "epoch": 4.0913921360255046, "grad_norm": 0.478515625, "learning_rate": 0.00014707825629096084, "loss": 0.3178, "step": 1925 }, { "epoch": 4.10201912858661, "grad_norm": 0.609375, "learning_rate": 0.00014675031213248296, "loss": 0.3214, "step": 1930 }, { "epoch": 4.112646121147715, "grad_norm": 0.470703125, "learning_rate": 0.00014642172329573026, "loss": 0.321, "step": 1935 }, { "epoch": 4.12327311370882, "grad_norm": 0.474609375, "learning_rate": 0.00014609249431188278, "loss": 0.3227, "step": 1940 }, { "epoch": 4.133900106269926, "grad_norm": 0.470703125, "learning_rate": 0.00014576262972094828, "loss": 0.3277, "step": 1945 }, { "epoch": 4.1445270988310305, "grad_norm": 0.578125, "learning_rate": 0.0001454321340716992, "loss": 0.3216, "step": 1950 }, { "epoch": 4.155154091392136, "grad_norm": 0.4453125, "learning_rate": 0.00014510101192161018, "loss": 0.3196, "step": 1955 }, { "epoch": 4.165781083953242, "grad_norm": 0.57421875, "learning_rate": 0.00014476926783679538, "loss": 0.3284, "step": 1960 }, { "epoch": 4.176408076514346, "grad_norm": 0.46875, "learning_rate": 0.00014443690639194515, "loss": 0.3244, "step": 1965 }, { "epoch": 4.187035069075452, "grad_norm": 0.80078125, "learning_rate": 0.00014410393217026318, "loss": 0.3205, "step": 1970 }, { "epoch": 4.1976620616365565, "grad_norm": 0.71875, "learning_rate": 0.0001437703497634032, "loss": 0.3216, "step": 1975 }, { "epoch": 4.208289054197662, "grad_norm": 0.76171875, "learning_rate": 0.00014343616377140582, "loss": 0.3178, "step": 1980 }, { "epoch": 4.218916046758768, "grad_norm": 0.86328125, "learning_rate": 0.00014310137880263482, "loss": 0.3219, "step": 1985 }, { "epoch": 4.229543039319872, "grad_norm": 0.5546875, "learning_rate": 0.00014276599947371388, "loss": 0.3262, "step": 1990 }, { "epoch": 4.240170031880978, "grad_norm": 0.51171875, "learning_rate": 0.00014243003040946274, "loss": 0.3287, "step": 1995 }, { "epoch": 4.2507970244420825, "grad_norm": 0.57421875, "learning_rate": 0.0001420934762428335, "loss": 0.328, "step": 2000 }, { "epoch": 4.261424017003188, "grad_norm": 0.64453125, "learning_rate": 0.00014175634161484675, "loss": 0.3334, "step": 2005 }, { "epoch": 4.272051009564294, "grad_norm": 0.515625, "learning_rate": 0.00014141863117452745, "loss": 0.3255, "step": 2010 }, { "epoch": 4.282678002125398, "grad_norm": 0.419921875, "learning_rate": 0.00014108034957884094, "loss": 0.3226, "step": 2015 }, { "epoch": 4.293304994686504, "grad_norm": 0.48046875, "learning_rate": 0.0001407415014926288, "loss": 0.3314, "step": 2020 }, { "epoch": 4.3039319872476085, "grad_norm": 0.67578125, "learning_rate": 0.00014040209158854426, "loss": 0.3334, "step": 2025 }, { "epoch": 4.314558979808714, "grad_norm": 0.48828125, "learning_rate": 0.00014006212454698797, "loss": 0.3349, "step": 2030 }, { "epoch": 4.32518597236982, "grad_norm": 0.455078125, "learning_rate": 0.00013972160505604342, "loss": 0.3301, "step": 2035 }, { "epoch": 4.335812964930924, "grad_norm": 0.6171875, "learning_rate": 0.00013938053781141222, "loss": 0.325, "step": 2040 }, { "epoch": 4.34643995749203, "grad_norm": 0.5, "learning_rate": 0.00013903892751634947, "loss": 0.3293, "step": 2045 }, { "epoch": 4.357066950053135, "grad_norm": 0.6484375, "learning_rate": 0.00013869677888159887, "loss": 0.3285, "step": 2050 }, { "epoch": 4.36769394261424, "grad_norm": 0.71875, "learning_rate": 0.00013835409662532762, "loss": 0.3271, "step": 2055 }, { "epoch": 4.378320935175346, "grad_norm": 0.515625, "learning_rate": 0.00013801088547306148, "loss": 0.3296, "step": 2060 }, { "epoch": 4.38894792773645, "grad_norm": 0.6015625, "learning_rate": 0.0001376671501576197, "loss": 0.3292, "step": 2065 }, { "epoch": 4.399574920297556, "grad_norm": 0.53515625, "learning_rate": 0.00013732289541904948, "loss": 0.3324, "step": 2070 }, { "epoch": 4.410201912858661, "grad_norm": 0.55078125, "learning_rate": 0.00013697812600456093, "loss": 0.3326, "step": 2075 }, { "epoch": 4.420828905419766, "grad_norm": 0.462890625, "learning_rate": 0.00013663284666846134, "loss": 0.3329, "step": 2080 }, { "epoch": 4.431455897980872, "grad_norm": 0.65234375, "learning_rate": 0.00013628706217208976, "loss": 0.3305, "step": 2085 }, { "epoch": 4.442082890541976, "grad_norm": 0.4765625, "learning_rate": 0.00013594077728375128, "loss": 0.3305, "step": 2090 }, { "epoch": 4.452709883103082, "grad_norm": 0.451171875, "learning_rate": 0.0001355939967786514, "loss": 0.3327, "step": 2095 }, { "epoch": 4.463336875664187, "grad_norm": 0.427734375, "learning_rate": 0.00013524672543882996, "loss": 0.3329, "step": 2100 }, { "epoch": 4.473963868225292, "grad_norm": 0.63671875, "learning_rate": 0.00013489896805309542, "loss": 0.3376, "step": 2105 }, { "epoch": 4.484590860786398, "grad_norm": 0.63671875, "learning_rate": 0.00013455072941695863, "loss": 0.329, "step": 2110 }, { "epoch": 4.495217853347503, "grad_norm": 0.53125, "learning_rate": 0.00013420201433256689, "loss": 0.333, "step": 2115 }, { "epoch": 4.505844845908608, "grad_norm": 0.470703125, "learning_rate": 0.00013385282760863758, "loss": 0.328, "step": 2120 }, { "epoch": 4.516471838469713, "grad_norm": 0.439453125, "learning_rate": 0.00013350317406039187, "loss": 0.3321, "step": 2125 }, { "epoch": 4.527098831030818, "grad_norm": 0.48046875, "learning_rate": 0.00013315305850948846, "loss": 0.3325, "step": 2130 }, { "epoch": 4.537725823591924, "grad_norm": 0.48046875, "learning_rate": 0.0001328024857839569, "loss": 0.3324, "step": 2135 }, { "epoch": 4.548352816153029, "grad_norm": 0.447265625, "learning_rate": 0.00013245146071813114, "loss": 0.3351, "step": 2140 }, { "epoch": 4.558979808714134, "grad_norm": 0.41796875, "learning_rate": 0.00013209998815258273, "loss": 0.3264, "step": 2145 }, { "epoch": 4.569606801275239, "grad_norm": 0.6640625, "learning_rate": 0.00013174807293405428, "loss": 0.3303, "step": 2150 }, { "epoch": 4.580233793836344, "grad_norm": 0.5625, "learning_rate": 0.00013139571991539238, "loss": 0.3308, "step": 2155 }, { "epoch": 4.5908607863974495, "grad_norm": 0.494140625, "learning_rate": 0.00013104293395548098, "loss": 0.3314, "step": 2160 }, { "epoch": 4.601487778958555, "grad_norm": 0.45703125, "learning_rate": 0.000130689719919174, "loss": 0.3328, "step": 2165 }, { "epoch": 4.61211477151966, "grad_norm": 0.453125, "learning_rate": 0.00013033608267722858, "loss": 0.3337, "step": 2170 }, { "epoch": 4.622741764080765, "grad_norm": 0.54296875, "learning_rate": 0.0001299820271062378, "loss": 0.3343, "step": 2175 }, { "epoch": 4.63336875664187, "grad_norm": 0.63671875, "learning_rate": 0.00012962755808856342, "loss": 0.3317, "step": 2180 }, { "epoch": 4.6439957492029755, "grad_norm": 0.478515625, "learning_rate": 0.00012927268051226844, "loss": 0.3311, "step": 2185 }, { "epoch": 4.654622741764081, "grad_norm": 0.55078125, "learning_rate": 0.0001289173992710499, "loss": 0.3347, "step": 2190 }, { "epoch": 4.665249734325186, "grad_norm": 0.5859375, "learning_rate": 0.00012856171926417133, "loss": 0.3353, "step": 2195 }, { "epoch": 4.675876726886291, "grad_norm": 0.59375, "learning_rate": 0.00012820564539639512, "loss": 0.3393, "step": 2200 }, { "epoch": 4.686503719447396, "grad_norm": 0.73046875, "learning_rate": 0.00012784918257791495, "loss": 0.3373, "step": 2205 }, { "epoch": 4.6971307120085015, "grad_norm": 0.7265625, "learning_rate": 0.00012749233572428804, "loss": 0.3271, "step": 2210 }, { "epoch": 4.707757704569607, "grad_norm": 0.77734375, "learning_rate": 0.00012713510975636741, "loss": 0.3363, "step": 2215 }, { "epoch": 4.718384697130712, "grad_norm": 0.67578125, "learning_rate": 0.00012677750960023396, "loss": 0.3355, "step": 2220 }, { "epoch": 4.729011689691817, "grad_norm": 0.5703125, "learning_rate": 0.00012641954018712863, "loss": 0.3378, "step": 2225 }, { "epoch": 4.739638682252922, "grad_norm": 0.484375, "learning_rate": 0.0001260612064533843, "loss": 0.3354, "step": 2230 }, { "epoch": 4.7502656748140275, "grad_norm": 0.5, "learning_rate": 0.0001257025133403577, "loss": 0.3363, "step": 2235 }, { "epoch": 4.760892667375133, "grad_norm": 0.57421875, "learning_rate": 0.0001253434657943616, "loss": 0.3296, "step": 2240 }, { "epoch": 4.771519659936238, "grad_norm": 0.5546875, "learning_rate": 0.00012498406876659598, "loss": 0.3381, "step": 2245 }, { "epoch": 4.782146652497343, "grad_norm": 0.5234375, "learning_rate": 0.0001246243272130804, "loss": 0.3222, "step": 2250 }, { "epoch": 4.792773645058449, "grad_norm": 0.474609375, "learning_rate": 0.00012426424609458518, "loss": 0.3348, "step": 2255 }, { "epoch": 4.8034006376195535, "grad_norm": 0.431640625, "learning_rate": 0.00012390383037656327, "loss": 0.3321, "step": 2260 }, { "epoch": 4.814027630180659, "grad_norm": 0.43359375, "learning_rate": 0.00012354308502908164, "loss": 0.3349, "step": 2265 }, { "epoch": 4.824654622741764, "grad_norm": 0.453125, "learning_rate": 0.00012318201502675285, "loss": 0.3269, "step": 2270 }, { "epoch": 4.835281615302869, "grad_norm": 0.4453125, "learning_rate": 0.00012282062534866632, "loss": 0.3379, "step": 2275 }, { "epoch": 4.845908607863975, "grad_norm": 0.423828125, "learning_rate": 0.00012245892097831982, "loss": 0.3316, "step": 2280 }, { "epoch": 4.8565356004250795, "grad_norm": 0.4609375, "learning_rate": 0.0001220969069035506, "loss": 0.3411, "step": 2285 }, { "epoch": 4.867162592986185, "grad_norm": 0.5546875, "learning_rate": 0.0001217345881164667, "loss": 0.3362, "step": 2290 }, { "epoch": 4.877789585547291, "grad_norm": 0.46484375, "learning_rate": 0.00012137196961337811, "loss": 0.3316, "step": 2295 }, { "epoch": 4.888416578108395, "grad_norm": 0.546875, "learning_rate": 0.00012100905639472779, "loss": 0.3309, "step": 2300 }, { "epoch": 4.899043570669501, "grad_norm": 0.50390625, "learning_rate": 0.00012064585346502286, "loss": 0.3341, "step": 2305 }, { "epoch": 4.9096705632306055, "grad_norm": 0.53515625, "learning_rate": 0.00012028236583276542, "loss": 0.3362, "step": 2310 }, { "epoch": 4.920297555791711, "grad_norm": 0.734375, "learning_rate": 0.0001199185985103836, "loss": 0.3397, "step": 2315 }, { "epoch": 4.930924548352817, "grad_norm": 0.49609375, "learning_rate": 0.00011955455651416246, "loss": 0.3363, "step": 2320 }, { "epoch": 4.941551540913921, "grad_norm": 0.48828125, "learning_rate": 0.00011919024486417463, "loss": 0.3335, "step": 2325 }, { "epoch": 4.952178533475027, "grad_norm": 0.5078125, "learning_rate": 0.00011882566858421135, "loss": 0.3311, "step": 2330 }, { "epoch": 4.9628055260361315, "grad_norm": 0.54296875, "learning_rate": 0.000118460832701713, "loss": 0.3346, "step": 2335 }, { "epoch": 4.973432518597237, "grad_norm": 0.54296875, "learning_rate": 0.00011809574224769981, "loss": 0.3334, "step": 2340 }, { "epoch": 4.984059511158343, "grad_norm": 0.462890625, "learning_rate": 0.00011773040225670256, "loss": 0.3309, "step": 2345 }, { "epoch": 4.994686503719447, "grad_norm": 0.490234375, "learning_rate": 0.00011736481776669306, "loss": 0.3345, "step": 2350 }, { "epoch": 4.99893730074389, "eval_loss": 2.1233744621276855, "eval_runtime": 0.8188, "eval_samples_per_second": 6.106, "eval_steps_per_second": 1.221, "step": 2352 }, { "epoch": 5.005313496280553, "grad_norm": 0.435546875, "learning_rate": 0.0001169989938190147, "loss": 0.3151, "step": 2355 }, { "epoch": 5.015940488841657, "grad_norm": 0.55859375, "learning_rate": 0.00011663293545831302, "loss": 0.2706, "step": 2360 }, { "epoch": 5.026567481402763, "grad_norm": 0.5390625, "learning_rate": 0.000116266647732466, "loss": 0.276, "step": 2365 }, { "epoch": 5.0371944739638685, "grad_norm": 0.55078125, "learning_rate": 0.00011590013569251457, "loss": 0.2787, "step": 2370 }, { "epoch": 5.047821466524973, "grad_norm": 0.56640625, "learning_rate": 0.00011553340439259286, "loss": 0.2753, "step": 2375 }, { "epoch": 5.058448459086079, "grad_norm": 0.466796875, "learning_rate": 0.0001151664588898586, "loss": 0.2707, "step": 2380 }, { "epoch": 5.069075451647183, "grad_norm": 0.52734375, "learning_rate": 0.00011479930424442335, "loss": 0.2703, "step": 2385 }, { "epoch": 5.079702444208289, "grad_norm": 0.53125, "learning_rate": 0.00011443194551928266, "loss": 0.2715, "step": 2390 }, { "epoch": 5.0903294367693945, "grad_norm": 0.6328125, "learning_rate": 0.00011406438778024635, "loss": 0.2739, "step": 2395 }, { "epoch": 5.100956429330499, "grad_norm": 0.5625, "learning_rate": 0.00011369663609586854, "loss": 0.278, "step": 2400 }, { "epoch": 5.111583421891605, "grad_norm": 0.52734375, "learning_rate": 0.0001133286955373779, "loss": 0.2809, "step": 2405 }, { "epoch": 5.12221041445271, "grad_norm": 0.57421875, "learning_rate": 0.00011296057117860759, "loss": 0.2706, "step": 2410 }, { "epoch": 5.132837407013815, "grad_norm": 0.54296875, "learning_rate": 0.00011259226809592534, "loss": 0.2765, "step": 2415 }, { "epoch": 5.1434643995749205, "grad_norm": 0.46484375, "learning_rate": 0.00011222379136816345, "loss": 0.2717, "step": 2420 }, { "epoch": 5.154091392136025, "grad_norm": 0.515625, "learning_rate": 0.00011185514607654881, "loss": 0.2796, "step": 2425 }, { "epoch": 5.164718384697131, "grad_norm": 0.51953125, "learning_rate": 0.00011148633730463273, "loss": 0.2801, "step": 2430 }, { "epoch": 5.175345377258236, "grad_norm": 0.5078125, "learning_rate": 0.00011111737013822088, "loss": 0.2745, "step": 2435 }, { "epoch": 5.185972369819341, "grad_norm": 0.51171875, "learning_rate": 0.00011074824966530312, "loss": 0.2786, "step": 2440 }, { "epoch": 5.1965993623804465, "grad_norm": 0.55078125, "learning_rate": 0.00011037898097598352, "loss": 0.2728, "step": 2445 }, { "epoch": 5.207226354941551, "grad_norm": 0.5625, "learning_rate": 0.00011000956916240985, "loss": 0.2827, "step": 2450 }, { "epoch": 5.217853347502657, "grad_norm": 0.59765625, "learning_rate": 0.00010964001931870365, "loss": 0.2783, "step": 2455 }, { "epoch": 5.228480340063762, "grad_norm": 0.51171875, "learning_rate": 0.00010927033654088983, "loss": 0.2847, "step": 2460 }, { "epoch": 5.239107332624867, "grad_norm": 0.78125, "learning_rate": 0.0001089005259268265, "loss": 0.2833, "step": 2465 }, { "epoch": 5.2497343251859725, "grad_norm": 0.6953125, "learning_rate": 0.00010853059257613448, "loss": 0.2819, "step": 2470 }, { "epoch": 5.260361317747078, "grad_norm": 0.5859375, "learning_rate": 0.00010816054159012723, "loss": 0.2766, "step": 2475 }, { "epoch": 5.270988310308183, "grad_norm": 0.53125, "learning_rate": 0.00010779037807174033, "loss": 0.2767, "step": 2480 }, { "epoch": 5.281615302869288, "grad_norm": 0.48828125, "learning_rate": 0.00010742010712546116, "loss": 0.2737, "step": 2485 }, { "epoch": 5.292242295430393, "grad_norm": 0.47265625, "learning_rate": 0.00010704973385725851, "loss": 0.2799, "step": 2490 }, { "epoch": 5.3028692879914985, "grad_norm": 0.58984375, "learning_rate": 0.00010667926337451217, "loss": 0.2815, "step": 2495 }, { "epoch": 5.313496280552604, "grad_norm": 0.48046875, "learning_rate": 0.00010630870078594249, "loss": 0.2796, "step": 2500 }, { "epoch": 5.324123273113709, "grad_norm": 0.5703125, "learning_rate": 0.00010593805120154001, "loss": 0.2764, "step": 2505 }, { "epoch": 5.334750265674814, "grad_norm": 0.48828125, "learning_rate": 0.00010556731973249485, "loss": 0.279, "step": 2510 }, { "epoch": 5.345377258235919, "grad_norm": 0.5625, "learning_rate": 0.00010519651149112631, "loss": 0.2817, "step": 2515 }, { "epoch": 5.3560042507970245, "grad_norm": 0.578125, "learning_rate": 0.00010482563159081238, "loss": 0.283, "step": 2520 }, { "epoch": 5.36663124335813, "grad_norm": 0.5390625, "learning_rate": 0.00010445468514591925, "loss": 0.2815, "step": 2525 }, { "epoch": 5.377258235919235, "grad_norm": 0.50390625, "learning_rate": 0.00010408367727173067, "loss": 0.2793, "step": 2530 }, { "epoch": 5.38788522848034, "grad_norm": 0.5078125, "learning_rate": 0.0001037126130843776, "loss": 0.2839, "step": 2535 }, { "epoch": 5.398512221041445, "grad_norm": 0.46875, "learning_rate": 0.00010334149770076747, "loss": 0.2843, "step": 2540 }, { "epoch": 5.4091392136025505, "grad_norm": 0.494140625, "learning_rate": 0.00010297033623851369, "loss": 0.2864, "step": 2545 }, { "epoch": 5.419766206163656, "grad_norm": 0.609375, "learning_rate": 0.0001025991338158651, "loss": 0.2883, "step": 2550 }, { "epoch": 5.430393198724761, "grad_norm": 0.609375, "learning_rate": 0.0001022278955516354, "loss": 0.2919, "step": 2555 }, { "epoch": 5.441020191285866, "grad_norm": 0.55859375, "learning_rate": 0.00010185662656513251, "loss": 0.28, "step": 2560 }, { "epoch": 5.451647183846971, "grad_norm": 0.51171875, "learning_rate": 0.00010148533197608803, "loss": 0.2815, "step": 2565 }, { "epoch": 5.462274176408076, "grad_norm": 0.5546875, "learning_rate": 0.00010111401690458654, "loss": 0.2858, "step": 2570 }, { "epoch": 5.472901168969182, "grad_norm": 0.474609375, "learning_rate": 0.00010074268647099525, "loss": 0.2824, "step": 2575 }, { "epoch": 5.483528161530287, "grad_norm": 0.58203125, "learning_rate": 0.00010037134579589302, "loss": 0.2863, "step": 2580 }, { "epoch": 5.494155154091392, "grad_norm": 0.58203125, "learning_rate": 0.0001, "loss": 0.2755, "step": 2585 }, { "epoch": 5.504782146652497, "grad_norm": 0.59375, "learning_rate": 9.962865420410701e-05, "loss": 0.2812, "step": 2590 }, { "epoch": 5.515409139213602, "grad_norm": 0.72265625, "learning_rate": 9.925731352900478e-05, "loss": 0.2781, "step": 2595 }, { "epoch": 5.526036131774708, "grad_norm": 0.73828125, "learning_rate": 9.888598309541347e-05, "loss": 0.2864, "step": 2600 }, { "epoch": 5.536663124335813, "grad_norm": 0.5234375, "learning_rate": 9.851466802391201e-05, "loss": 0.2813, "step": 2605 }, { "epoch": 5.547290116896918, "grad_norm": 0.60546875, "learning_rate": 9.814337343486754e-05, "loss": 0.2801, "step": 2610 }, { "epoch": 5.557917109458024, "grad_norm": 0.546875, "learning_rate": 9.777210444836463e-05, "loss": 0.281, "step": 2615 }, { "epoch": 5.568544102019128, "grad_norm": 0.51953125, "learning_rate": 9.740086618413495e-05, "loss": 0.2817, "step": 2620 }, { "epoch": 5.579171094580234, "grad_norm": 0.578125, "learning_rate": 9.702966376148635e-05, "loss": 0.2859, "step": 2625 }, { "epoch": 5.589798087141339, "grad_norm": 0.52734375, "learning_rate": 9.665850229923258e-05, "loss": 0.2814, "step": 2630 }, { "epoch": 5.600425079702444, "grad_norm": 0.609375, "learning_rate": 9.62873869156224e-05, "loss": 0.2827, "step": 2635 }, { "epoch": 5.61105207226355, "grad_norm": 0.48046875, "learning_rate": 9.591632272826934e-05, "loss": 0.2831, "step": 2640 }, { "epoch": 5.621679064824654, "grad_norm": 0.609375, "learning_rate": 9.554531485408078e-05, "loss": 0.2791, "step": 2645 }, { "epoch": 5.63230605738576, "grad_norm": 0.51953125, "learning_rate": 9.517436840918766e-05, "loss": 0.2829, "step": 2650 }, { "epoch": 5.6429330499468655, "grad_norm": 0.5390625, "learning_rate": 9.480348850887373e-05, "loss": 0.2826, "step": 2655 }, { "epoch": 5.65356004250797, "grad_norm": 0.490234375, "learning_rate": 9.44326802675052e-05, "loss": 0.2827, "step": 2660 }, { "epoch": 5.664187035069076, "grad_norm": 0.46875, "learning_rate": 9.406194879846e-05, "loss": 0.2857, "step": 2665 }, { "epoch": 5.67481402763018, "grad_norm": 0.484375, "learning_rate": 9.369129921405754e-05, "loss": 0.2818, "step": 2670 }, { "epoch": 5.685441020191286, "grad_norm": 0.45703125, "learning_rate": 9.332073662548784e-05, "loss": 0.278, "step": 2675 }, { "epoch": 5.6960680127523915, "grad_norm": 0.5234375, "learning_rate": 9.295026614274152e-05, "loss": 0.2815, "step": 2680 }, { "epoch": 5.706695005313496, "grad_norm": 0.5546875, "learning_rate": 9.257989287453883e-05, "loss": 0.2832, "step": 2685 }, { "epoch": 5.717321997874602, "grad_norm": 0.59375, "learning_rate": 9.220962192825968e-05, "loss": 0.285, "step": 2690 }, { "epoch": 5.727948990435706, "grad_norm": 0.45703125, "learning_rate": 9.183945840987276e-05, "loss": 0.283, "step": 2695 }, { "epoch": 5.738575982996812, "grad_norm": 0.4921875, "learning_rate": 9.146940742386553e-05, "loss": 0.2869, "step": 2700 }, { "epoch": 5.7492029755579175, "grad_norm": 0.466796875, "learning_rate": 9.109947407317352e-05, "loss": 0.2876, "step": 2705 }, { "epoch": 5.759829968119022, "grad_norm": 0.51953125, "learning_rate": 9.072966345911019e-05, "loss": 0.2879, "step": 2710 }, { "epoch": 5.770456960680128, "grad_norm": 0.5, "learning_rate": 9.035998068129636e-05, "loss": 0.2773, "step": 2715 }, { "epoch": 5.781083953241232, "grad_norm": 0.5390625, "learning_rate": 8.999043083759017e-05, "loss": 0.2863, "step": 2720 }, { "epoch": 5.791710945802338, "grad_norm": 0.47265625, "learning_rate": 8.962101902401648e-05, "loss": 0.2813, "step": 2725 }, { "epoch": 5.8023379383634435, "grad_norm": 0.54296875, "learning_rate": 8.925175033469688e-05, "loss": 0.2772, "step": 2730 }, { "epoch": 5.812964930924548, "grad_norm": 0.5234375, "learning_rate": 8.888262986177913e-05, "loss": 0.2842, "step": 2735 }, { "epoch": 5.823591923485654, "grad_norm": 0.53125, "learning_rate": 8.851366269536729e-05, "loss": 0.2806, "step": 2740 }, { "epoch": 5.834218916046758, "grad_norm": 0.51953125, "learning_rate": 8.814485392345118e-05, "loss": 0.2754, "step": 2745 }, { "epoch": 5.844845908607864, "grad_norm": 0.482421875, "learning_rate": 8.777620863183657e-05, "loss": 0.2852, "step": 2750 }, { "epoch": 5.8554729011689695, "grad_norm": 0.50390625, "learning_rate": 8.740773190407471e-05, "loss": 0.2782, "step": 2755 }, { "epoch": 5.866099893730074, "grad_norm": 0.486328125, "learning_rate": 8.703942882139245e-05, "loss": 0.2887, "step": 2760 }, { "epoch": 5.87672688629118, "grad_norm": 0.46875, "learning_rate": 8.667130446262214e-05, "loss": 0.2812, "step": 2765 }, { "epoch": 5.887353878852284, "grad_norm": 0.53515625, "learning_rate": 8.630336390413147e-05, "loss": 0.2809, "step": 2770 }, { "epoch": 5.89798087141339, "grad_norm": 0.498046875, "learning_rate": 8.59356122197537e-05, "loss": 0.2816, "step": 2775 }, { "epoch": 5.9086078639744954, "grad_norm": 0.59375, "learning_rate": 8.556805448071735e-05, "loss": 0.2788, "step": 2780 }, { "epoch": 5.9192348565356, "grad_norm": 0.515625, "learning_rate": 8.520069575557667e-05, "loss": 0.2773, "step": 2785 }, { "epoch": 5.929861849096706, "grad_norm": 0.53125, "learning_rate": 8.483354111014141e-05, "loss": 0.2882, "step": 2790 }, { "epoch": 5.940488841657811, "grad_norm": 0.50390625, "learning_rate": 8.446659560740717e-05, "loss": 0.285, "step": 2795 }, { "epoch": 5.951115834218916, "grad_norm": 0.515625, "learning_rate": 8.409986430748545e-05, "loss": 0.2788, "step": 2800 }, { "epoch": 5.961742826780021, "grad_norm": 0.50390625, "learning_rate": 8.373335226753404e-05, "loss": 0.2799, "step": 2805 }, { "epoch": 5.972369819341126, "grad_norm": 0.478515625, "learning_rate": 8.336706454168701e-05, "loss": 0.2835, "step": 2810 }, { "epoch": 5.982996811902232, "grad_norm": 0.5234375, "learning_rate": 8.300100618098534e-05, "loss": 0.2865, "step": 2815 }, { "epoch": 5.993623804463337, "grad_norm": 0.53515625, "learning_rate": 8.263518223330697e-05, "loss": 0.2875, "step": 2820 }, { "epoch": 6.0, "eval_loss": 2.5845518112182617, "eval_runtime": 0.7933, "eval_samples_per_second": 6.303, "eval_steps_per_second": 1.261, "step": 2823 }, { "epoch": 6.004250797024442, "grad_norm": 0.3984375, "learning_rate": 8.226959774329747e-05, "loss": 0.2603, "step": 2825 }, { "epoch": 6.014877789585547, "grad_norm": 0.578125, "learning_rate": 8.190425775230021e-05, "loss": 0.2316, "step": 2830 }, { "epoch": 6.025504782146652, "grad_norm": 0.470703125, "learning_rate": 8.153916729828703e-05, "loss": 0.2244, "step": 2835 }, { "epoch": 6.036131774707758, "grad_norm": 0.470703125, "learning_rate": 8.117433141578866e-05, "loss": 0.2276, "step": 2840 }, { "epoch": 6.046758767268863, "grad_norm": 0.484375, "learning_rate": 8.080975513582539e-05, "loss": 0.2338, "step": 2845 }, { "epoch": 6.057385759829968, "grad_norm": 0.51171875, "learning_rate": 8.044544348583755e-05, "loss": 0.2334, "step": 2850 }, { "epoch": 6.068012752391073, "grad_norm": 0.51953125, "learning_rate": 8.008140148961641e-05, "loss": 0.2319, "step": 2855 }, { "epoch": 6.078639744952179, "grad_norm": 0.53125, "learning_rate": 7.971763416723459e-05, "loss": 0.2326, "step": 2860 }, { "epoch": 6.089266737513284, "grad_norm": 0.494140625, "learning_rate": 7.935414653497715e-05, "loss": 0.2305, "step": 2865 }, { "epoch": 6.099893730074389, "grad_norm": 0.46484375, "learning_rate": 7.89909436052722e-05, "loss": 0.2297, "step": 2870 }, { "epoch": 6.110520722635494, "grad_norm": 0.482421875, "learning_rate": 7.862803038662191e-05, "loss": 0.2251, "step": 2875 }, { "epoch": 6.121147715196599, "grad_norm": 0.515625, "learning_rate": 7.826541188353329e-05, "loss": 0.2321, "step": 2880 }, { "epoch": 6.131774707757705, "grad_norm": 0.52734375, "learning_rate": 7.790309309644942e-05, "loss": 0.2324, "step": 2885 }, { "epoch": 6.14240170031881, "grad_norm": 0.5078125, "learning_rate": 7.754107902168019e-05, "loss": 0.2305, "step": 2890 }, { "epoch": 6.153028692879915, "grad_norm": 0.52734375, "learning_rate": 7.717937465133371e-05, "loss": 0.2258, "step": 2895 }, { "epoch": 6.16365568544102, "grad_norm": 0.5078125, "learning_rate": 7.681798497324716e-05, "loss": 0.2375, "step": 2900 }, { "epoch": 6.174282678002125, "grad_norm": 0.50390625, "learning_rate": 7.645691497091838e-05, "loss": 0.2351, "step": 2905 }, { "epoch": 6.184909670563231, "grad_norm": 0.52734375, "learning_rate": 7.609616962343675e-05, "loss": 0.2359, "step": 2910 }, { "epoch": 6.195536663124336, "grad_norm": 0.5234375, "learning_rate": 7.573575390541485e-05, "loss": 0.2396, "step": 2915 }, { "epoch": 6.206163655685441, "grad_norm": 0.5390625, "learning_rate": 7.537567278691964e-05, "loss": 0.2316, "step": 2920 }, { "epoch": 6.216790648246546, "grad_norm": 0.486328125, "learning_rate": 7.501593123340403e-05, "loss": 0.2356, "step": 2925 }, { "epoch": 6.227417640807651, "grad_norm": 0.52734375, "learning_rate": 7.465653420563845e-05, "loss": 0.2373, "step": 2930 }, { "epoch": 6.238044633368757, "grad_norm": 0.53515625, "learning_rate": 7.42974866596423e-05, "loss": 0.2349, "step": 2935 }, { "epoch": 6.248671625929862, "grad_norm": 0.5625, "learning_rate": 7.393879354661577e-05, "loss": 0.2326, "step": 2940 }, { "epoch": 6.259298618490967, "grad_norm": 0.60546875, "learning_rate": 7.358045981287141e-05, "loss": 0.2366, "step": 2945 }, { "epoch": 6.269925611052072, "grad_norm": 0.5390625, "learning_rate": 7.322249039976608e-05, "loss": 0.2372, "step": 2950 }, { "epoch": 6.280552603613177, "grad_norm": 0.4765625, "learning_rate": 7.286489024363261e-05, "loss": 0.2319, "step": 2955 }, { "epoch": 6.291179596174283, "grad_norm": 0.546875, "learning_rate": 7.2507664275712e-05, "loss": 0.2341, "step": 2960 }, { "epoch": 6.301806588735388, "grad_norm": 0.5703125, "learning_rate": 7.215081742208508e-05, "loss": 0.2334, "step": 2965 }, { "epoch": 6.312433581296493, "grad_norm": 0.515625, "learning_rate": 7.179435460360491e-05, "loss": 0.2305, "step": 2970 }, { "epoch": 6.323060573857599, "grad_norm": 0.61328125, "learning_rate": 7.14382807358287e-05, "loss": 0.2351, "step": 2975 }, { "epoch": 6.333687566418703, "grad_norm": 0.51953125, "learning_rate": 7.108260072895013e-05, "loss": 0.2312, "step": 2980 }, { "epoch": 6.344314558979809, "grad_norm": 0.52734375, "learning_rate": 7.072731948773159e-05, "loss": 0.2309, "step": 2985 }, { "epoch": 6.354941551540914, "grad_norm": 0.54296875, "learning_rate": 7.037244191143661e-05, "loss": 0.2392, "step": 2990 }, { "epoch": 6.365568544102019, "grad_norm": 0.56640625, "learning_rate": 7.001797289376218e-05, "loss": 0.2363, "step": 2995 }, { "epoch": 6.376195536663125, "grad_norm": 0.55859375, "learning_rate": 6.966391732277143e-05, "loss": 0.2342, "step": 3000 }, { "epoch": 6.386822529224229, "grad_norm": 0.55859375, "learning_rate": 6.931028008082602e-05, "loss": 0.2329, "step": 3005 }, { "epoch": 6.397449521785335, "grad_norm": 0.55078125, "learning_rate": 6.895706604451905e-05, "loss": 0.2377, "step": 3010 }, { "epoch": 6.4080765143464395, "grad_norm": 0.53515625, "learning_rate": 6.860428008460762e-05, "loss": 0.2402, "step": 3015 }, { "epoch": 6.418703506907545, "grad_norm": 0.5703125, "learning_rate": 6.825192706594575e-05, "loss": 0.2366, "step": 3020 }, { "epoch": 6.429330499468651, "grad_norm": 0.6015625, "learning_rate": 6.790001184741728e-05, "loss": 0.2331, "step": 3025 }, { "epoch": 6.439957492029755, "grad_norm": 0.56640625, "learning_rate": 6.75485392818689e-05, "loss": 0.2337, "step": 3030 }, { "epoch": 6.450584484590861, "grad_norm": 0.53515625, "learning_rate": 6.719751421604309e-05, "loss": 0.2357, "step": 3035 }, { "epoch": 6.461211477151966, "grad_norm": 0.5, "learning_rate": 6.684694149051156e-05, "loss": 0.2395, "step": 3040 }, { "epoch": 6.471838469713071, "grad_norm": 0.54296875, "learning_rate": 6.649682593960814e-05, "loss": 0.2258, "step": 3045 }, { "epoch": 6.482465462274177, "grad_norm": 0.46484375, "learning_rate": 6.614717239136246e-05, "loss": 0.234, "step": 3050 }, { "epoch": 6.493092454835281, "grad_norm": 0.5234375, "learning_rate": 6.579798566743314e-05, "loss": 0.2417, "step": 3055 }, { "epoch": 6.503719447396387, "grad_norm": 0.66796875, "learning_rate": 6.54492705830414e-05, "loss": 0.2319, "step": 3060 }, { "epoch": 6.514346439957492, "grad_norm": 0.58203125, "learning_rate": 6.51010319469046e-05, "loss": 0.2325, "step": 3065 }, { "epoch": 6.524973432518597, "grad_norm": 0.578125, "learning_rate": 6.475327456117005e-05, "loss": 0.2353, "step": 3070 }, { "epoch": 6.535600425079703, "grad_norm": 0.490234375, "learning_rate": 6.440600322134864e-05, "loss": 0.2345, "step": 3075 }, { "epoch": 6.546227417640807, "grad_norm": 0.52734375, "learning_rate": 6.405922271624874e-05, "loss": 0.2327, "step": 3080 }, { "epoch": 6.556854410201913, "grad_norm": 0.54296875, "learning_rate": 6.371293782791028e-05, "loss": 0.2356, "step": 3085 }, { "epoch": 6.567481402763018, "grad_norm": 0.49609375, "learning_rate": 6.336715333153869e-05, "loss": 0.2336, "step": 3090 }, { "epoch": 6.578108395324123, "grad_norm": 0.51953125, "learning_rate": 6.302187399543911e-05, "loss": 0.2365, "step": 3095 }, { "epoch": 6.588735387885229, "grad_norm": 0.46875, "learning_rate": 6.267710458095053e-05, "loss": 0.2377, "step": 3100 }, { "epoch": 6.599362380446333, "grad_norm": 0.53515625, "learning_rate": 6.233284984238035e-05, "loss": 0.2297, "step": 3105 }, { "epoch": 6.609989373007439, "grad_norm": 0.53515625, "learning_rate": 6.198911452693853e-05, "loss": 0.2347, "step": 3110 }, { "epoch": 6.620616365568544, "grad_norm": 0.51953125, "learning_rate": 6.164590337467243e-05, "loss": 0.2363, "step": 3115 }, { "epoch": 6.631243358129649, "grad_norm": 0.5078125, "learning_rate": 6.130322111840114e-05, "loss": 0.2339, "step": 3120 }, { "epoch": 6.641870350690755, "grad_norm": 0.57421875, "learning_rate": 6.0961072483650526e-05, "loss": 0.2328, "step": 3125 }, { "epoch": 6.652497343251859, "grad_norm": 0.578125, "learning_rate": 6.0619462188587793e-05, "loss": 0.234, "step": 3130 }, { "epoch": 6.663124335812965, "grad_norm": 0.5390625, "learning_rate": 6.027839494395664e-05, "loss": 0.2397, "step": 3135 }, { "epoch": 6.67375132837407, "grad_norm": 0.50390625, "learning_rate": 5.993787545301204e-05, "loss": 0.2379, "step": 3140 }, { "epoch": 6.684378320935175, "grad_norm": 0.50390625, "learning_rate": 5.959790841145577e-05, "loss": 0.236, "step": 3145 }, { "epoch": 6.695005313496281, "grad_norm": 0.53125, "learning_rate": 5.9258498507371194e-05, "loss": 0.2393, "step": 3150 }, { "epoch": 6.705632306057386, "grad_norm": 0.50390625, "learning_rate": 5.891965042115907e-05, "loss": 0.2399, "step": 3155 }, { "epoch": 6.716259298618491, "grad_norm": 0.59765625, "learning_rate": 5.8581368825472585e-05, "loss": 0.2261, "step": 3160 }, { "epoch": 6.726886291179596, "grad_norm": 0.51953125, "learning_rate": 5.8243658385153285e-05, "loss": 0.2379, "step": 3165 }, { "epoch": 6.737513283740701, "grad_norm": 0.55859375, "learning_rate": 5.790652375716652e-05, "loss": 0.2327, "step": 3170 }, { "epoch": 6.748140276301807, "grad_norm": 0.58984375, "learning_rate": 5.7569969590537284e-05, "loss": 0.2299, "step": 3175 }, { "epoch": 6.758767268862912, "grad_norm": 0.490234375, "learning_rate": 5.7234000526286156e-05, "loss": 0.2373, "step": 3180 }, { "epoch": 6.769394261424017, "grad_norm": 0.55859375, "learning_rate": 5.689862119736522e-05, "loss": 0.2406, "step": 3185 }, { "epoch": 6.780021253985122, "grad_norm": 0.55078125, "learning_rate": 5.656383622859418e-05, "loss": 0.2394, "step": 3190 }, { "epoch": 6.790648246546228, "grad_norm": 0.5625, "learning_rate": 5.622965023659683e-05, "loss": 0.2378, "step": 3195 }, { "epoch": 6.801275239107333, "grad_norm": 0.5703125, "learning_rate": 5.589606782973683e-05, "loss": 0.2342, "step": 3200 }, { "epoch": 6.811902231668438, "grad_norm": 0.51171875, "learning_rate": 5.55630936080549e-05, "loss": 0.2344, "step": 3205 }, { "epoch": 6.822529224229543, "grad_norm": 0.53125, "learning_rate": 5.5230732163204615e-05, "loss": 0.2344, "step": 3210 }, { "epoch": 6.833156216790648, "grad_norm": 0.5390625, "learning_rate": 5.48989880783898e-05, "loss": 0.2352, "step": 3215 }, { "epoch": 6.843783209351754, "grad_norm": 0.51953125, "learning_rate": 5.456786592830083e-05, "loss": 0.2337, "step": 3220 }, { "epoch": 6.8544102019128585, "grad_norm": 0.50390625, "learning_rate": 5.423737027905173e-05, "loss": 0.2326, "step": 3225 }, { "epoch": 6.865037194473964, "grad_norm": 0.50390625, "learning_rate": 5.39075056881172e-05, "loss": 0.2348, "step": 3230 }, { "epoch": 6.875664187035069, "grad_norm": 0.54296875, "learning_rate": 5.357827670426977e-05, "loss": 0.2361, "step": 3235 }, { "epoch": 6.886291179596174, "grad_norm": 0.4921875, "learning_rate": 5.3249687867517095e-05, "loss": 0.2298, "step": 3240 }, { "epoch": 6.89691817215728, "grad_norm": 0.53515625, "learning_rate": 5.292174370903919e-05, "loss": 0.2344, "step": 3245 }, { "epoch": 6.9075451647183845, "grad_norm": 0.51953125, "learning_rate": 5.259444875112624e-05, "loss": 0.2366, "step": 3250 }, { "epoch": 6.91817215727949, "grad_norm": 0.8203125, "learning_rate": 5.226780750711602e-05, "loss": 0.2343, "step": 3255 }, { "epoch": 6.928799149840595, "grad_norm": 0.53125, "learning_rate": 5.1941824481331626e-05, "loss": 0.236, "step": 3260 }, { "epoch": 6.9394261424017, "grad_norm": 0.498046875, "learning_rate": 5.1616504169019564e-05, "loss": 0.2271, "step": 3265 }, { "epoch": 6.950053134962806, "grad_norm": 0.50390625, "learning_rate": 5.129185105628756e-05, "loss": 0.2338, "step": 3270 }, { "epoch": 6.9606801275239105, "grad_norm": 0.515625, "learning_rate": 5.0967869620042794e-05, "loss": 0.2311, "step": 3275 }, { "epoch": 6.971307120085016, "grad_norm": 0.51171875, "learning_rate": 5.064456432793019e-05, "loss": 0.23, "step": 3280 }, { "epoch": 6.981934112646121, "grad_norm": 0.59765625, "learning_rate": 5.032193963827073e-05, "loss": 0.2347, "step": 3285 }, { "epoch": 6.992561105207226, "grad_norm": 0.56640625, "learning_rate": 5.000000000000002e-05, "loss": 0.2319, "step": 3290 }, { "epoch": 6.99893730074389, "eval_loss": 3.1057090759277344, "eval_runtime": 0.816, "eval_samples_per_second": 6.127, "eval_steps_per_second": 1.225, "step": 3293 }, { "epoch": 7.003188097768332, "grad_norm": 0.40234375, "learning_rate": 4.96787498526069e-05, "loss": 0.2177, "step": 3295 }, { "epoch": 7.0138150903294365, "grad_norm": 0.458984375, "learning_rate": 4.93581936260724e-05, "loss": 0.1956, "step": 3300 }, { "epoch": 7.024442082890542, "grad_norm": 0.56640625, "learning_rate": 4.903833574080825e-05, "loss": 0.1954, "step": 3305 }, { "epoch": 7.035069075451648, "grad_norm": 0.494140625, "learning_rate": 4.8719180607596484e-05, "loss": 0.1939, "step": 3310 }, { "epoch": 7.045696068012752, "grad_norm": 0.4453125, "learning_rate": 4.8400732627528e-05, "loss": 0.1996, "step": 3315 }, { "epoch": 7.056323060573858, "grad_norm": 0.55859375, "learning_rate": 4.808299619194251e-05, "loss": 0.1946, "step": 3320 }, { "epoch": 7.0669500531349625, "grad_norm": 0.54296875, "learning_rate": 4.776597568236731e-05, "loss": 0.1914, "step": 3325 }, { "epoch": 7.077577045696068, "grad_norm": 0.4609375, "learning_rate": 4.744967547045754e-05, "loss": 0.1963, "step": 3330 }, { "epoch": 7.088204038257174, "grad_norm": 0.4921875, "learning_rate": 4.713409991793536e-05, "loss": 0.2024, "step": 3335 }, { "epoch": 7.098831030818278, "grad_norm": 0.5625, "learning_rate": 4.681925337653006e-05, "loss": 0.1974, "step": 3340 }, { "epoch": 7.109458023379384, "grad_norm": 0.53515625, "learning_rate": 4.650514018791799e-05, "loss": 0.1954, "step": 3345 }, { "epoch": 7.1200850159404885, "grad_norm": 0.53125, "learning_rate": 4.6191764683662744e-05, "loss": 0.2007, "step": 3350 }, { "epoch": 7.130712008501594, "grad_norm": 0.478515625, "learning_rate": 4.587913118515532e-05, "loss": 0.1959, "step": 3355 }, { "epoch": 7.1413390010627, "grad_norm": 0.5078125, "learning_rate": 4.5567244003554645e-05, "loss": 0.1994, "step": 3360 }, { "epoch": 7.151965993623804, "grad_norm": 0.51953125, "learning_rate": 4.5256107439728035e-05, "loss": 0.1979, "step": 3365 }, { "epoch": 7.16259298618491, "grad_norm": 0.53125, "learning_rate": 4.494572578419194e-05, "loss": 0.1951, "step": 3370 }, { "epoch": 7.1732199787460145, "grad_norm": 0.474609375, "learning_rate": 4.463610331705273e-05, "loss": 0.1996, "step": 3375 }, { "epoch": 7.18384697130712, "grad_norm": 0.474609375, "learning_rate": 4.432724430794786e-05, "loss": 0.1989, "step": 3380 }, { "epoch": 7.194473963868226, "grad_norm": 0.51953125, "learning_rate": 4.4019153015986546e-05, "loss": 0.2009, "step": 3385 }, { "epoch": 7.20510095642933, "grad_norm": 0.53125, "learning_rate": 4.371183368969165e-05, "loss": 0.1926, "step": 3390 }, { "epoch": 7.215727948990436, "grad_norm": 0.51171875, "learning_rate": 4.340529056694047e-05, "loss": 0.1997, "step": 3395 }, { "epoch": 7.226354941551541, "grad_norm": 0.4765625, "learning_rate": 4.309952787490689e-05, "loss": 0.1969, "step": 3400 }, { "epoch": 7.236981934112646, "grad_norm": 0.48046875, "learning_rate": 4.279454983000251e-05, "loss": 0.1974, "step": 3405 }, { "epoch": 7.247608926673752, "grad_norm": 0.63671875, "learning_rate": 4.249036063781896e-05, "loss": 0.1969, "step": 3410 }, { "epoch": 7.258235919234856, "grad_norm": 0.5234375, "learning_rate": 4.2186964493069794e-05, "loss": 0.1953, "step": 3415 }, { "epoch": 7.268862911795962, "grad_norm": 0.54296875, "learning_rate": 4.1884365579532346e-05, "loss": 0.1973, "step": 3420 }, { "epoch": 7.279489904357067, "grad_norm": 0.62890625, "learning_rate": 4.158256806999059e-05, "loss": 0.1938, "step": 3425 }, { "epoch": 7.290116896918172, "grad_norm": 0.51953125, "learning_rate": 4.128157612617696e-05, "loss": 0.1964, "step": 3430 }, { "epoch": 7.3007438894792775, "grad_norm": 0.53125, "learning_rate": 4.0981393898715545e-05, "loss": 0.202, "step": 3435 }, { "epoch": 7.311370882040382, "grad_norm": 0.494140625, "learning_rate": 4.0682025527064486e-05, "loss": 0.1964, "step": 3440 }, { "epoch": 7.321997874601488, "grad_norm": 0.53515625, "learning_rate": 4.038347513945898e-05, "loss": 0.1999, "step": 3445 }, { "epoch": 7.332624867162593, "grad_norm": 0.4921875, "learning_rate": 4.008574685285442e-05, "loss": 0.1938, "step": 3450 }, { "epoch": 7.343251859723698, "grad_norm": 0.94921875, "learning_rate": 3.978884477286956e-05, "loss": 0.1941, "step": 3455 }, { "epoch": 7.3538788522848035, "grad_norm": 0.5859375, "learning_rate": 3.94927729937299e-05, "loss": 0.1996, "step": 3460 }, { "epoch": 7.364505844845908, "grad_norm": 0.578125, "learning_rate": 3.9197535598211274e-05, "loss": 0.2, "step": 3465 }, { "epoch": 7.375132837407014, "grad_norm": 0.50390625, "learning_rate": 3.890313665758348e-05, "loss": 0.1937, "step": 3470 }, { "epoch": 7.385759829968119, "grad_norm": 0.5390625, "learning_rate": 3.860958023155422e-05, "loss": 0.1991, "step": 3475 }, { "epoch": 7.396386822529224, "grad_norm": 0.5078125, "learning_rate": 3.8316870368213e-05, "loss": 0.1979, "step": 3480 }, { "epoch": 7.4070138150903295, "grad_norm": 0.546875, "learning_rate": 3.802501110397553e-05, "loss": 0.2021, "step": 3485 }, { "epoch": 7.417640807651434, "grad_norm": 0.54296875, "learning_rate": 3.773400646352769e-05, "loss": 0.2034, "step": 3490 }, { "epoch": 7.42826780021254, "grad_norm": 0.50390625, "learning_rate": 3.744386045977052e-05, "loss": 0.2033, "step": 3495 }, { "epoch": 7.438894792773645, "grad_norm": 0.5, "learning_rate": 3.7154577093764334e-05, "loss": 0.2012, "step": 3500 }, { "epoch": 7.44952178533475, "grad_norm": 0.53515625, "learning_rate": 3.686616035467408e-05, "loss": 0.1952, "step": 3505 }, { "epoch": 7.4601487778958555, "grad_norm": 0.57421875, "learning_rate": 3.657861421971388e-05, "loss": 0.2011, "step": 3510 }, { "epoch": 7.470775770456961, "grad_norm": 0.490234375, "learning_rate": 3.629194265409255e-05, "loss": 0.2033, "step": 3515 }, { "epoch": 7.481402763018066, "grad_norm": 0.546875, "learning_rate": 3.6006149610958625e-05, "loss": 0.2003, "step": 3520 }, { "epoch": 7.492029755579171, "grad_norm": 0.56640625, "learning_rate": 3.5721239031346066e-05, "loss": 0.1988, "step": 3525 }, { "epoch": 7.502656748140276, "grad_norm": 0.484375, "learning_rate": 3.543721484411976e-05, "loss": 0.1903, "step": 3530 }, { "epoch": 7.5132837407013815, "grad_norm": 0.53125, "learning_rate": 3.515408096592144e-05, "loss": 0.1991, "step": 3535 }, { "epoch": 7.523910733262487, "grad_norm": 0.49609375, "learning_rate": 3.487184130111562e-05, "loss": 0.1967, "step": 3540 }, { "epoch": 7.534537725823592, "grad_norm": 0.54296875, "learning_rate": 3.459049974173579e-05, "loss": 0.1983, "step": 3545 }, { "epoch": 7.545164718384697, "grad_norm": 0.53125, "learning_rate": 3.4310060167430725e-05, "loss": 0.1998, "step": 3550 }, { "epoch": 7.555791710945803, "grad_norm": 0.5390625, "learning_rate": 3.4030526445410984e-05, "loss": 0.198, "step": 3555 }, { "epoch": 7.5664187035069075, "grad_norm": 0.49609375, "learning_rate": 3.375190243039556e-05, "loss": 0.1961, "step": 3560 }, { "epoch": 7.577045696068013, "grad_norm": 0.478515625, "learning_rate": 3.3474191964558885e-05, "loss": 0.1974, "step": 3565 }, { "epoch": 7.587672688629118, "grad_norm": 0.5234375, "learning_rate": 3.319739887747752e-05, "loss": 0.195, "step": 3570 }, { "epoch": 7.598299681190223, "grad_norm": 0.51171875, "learning_rate": 3.292152698607768e-05, "loss": 0.1992, "step": 3575 }, { "epoch": 7.608926673751329, "grad_norm": 0.50390625, "learning_rate": 3.264658009458239e-05, "loss": 0.1996, "step": 3580 }, { "epoch": 7.6195536663124335, "grad_norm": 0.5, "learning_rate": 3.2372561994459136e-05, "loss": 0.1966, "step": 3585 }, { "epoch": 7.630180658873539, "grad_norm": 0.5078125, "learning_rate": 3.209947646436752e-05, "loss": 0.1941, "step": 3590 }, { "epoch": 7.640807651434644, "grad_norm": 0.54296875, "learning_rate": 3.182732727010715e-05, "loss": 0.1959, "step": 3595 }, { "epoch": 7.651434643995749, "grad_norm": 0.53515625, "learning_rate": 3.155611816456586e-05, "loss": 0.2002, "step": 3600 }, { "epoch": 7.662061636556855, "grad_norm": 0.51171875, "learning_rate": 3.12858528876676e-05, "loss": 0.196, "step": 3605 }, { "epoch": 7.6726886291179595, "grad_norm": 0.53125, "learning_rate": 3.1016535166321356e-05, "loss": 0.1988, "step": 3610 }, { "epoch": 7.683315621679065, "grad_norm": 0.51171875, "learning_rate": 3.074816871436924e-05, "loss": 0.2016, "step": 3615 }, { "epoch": 7.69394261424017, "grad_norm": 0.54296875, "learning_rate": 3.0480757232535772e-05, "loss": 0.1973, "step": 3620 }, { "epoch": 7.704569606801275, "grad_norm": 0.5234375, "learning_rate": 3.021430440837646e-05, "loss": 0.1945, "step": 3625 }, { "epoch": 7.715196599362381, "grad_norm": 0.5234375, "learning_rate": 2.9948813916227115e-05, "loss": 0.1985, "step": 3630 }, { "epoch": 7.7258235919234854, "grad_norm": 0.5859375, "learning_rate": 2.968428941715321e-05, "loss": 0.1963, "step": 3635 }, { "epoch": 7.736450584484591, "grad_norm": 0.50390625, "learning_rate": 2.9420734558899322e-05, "loss": 0.2008, "step": 3640 }, { "epoch": 7.747077577045696, "grad_norm": 0.5234375, "learning_rate": 2.915815297583886e-05, "loss": 0.1983, "step": 3645 }, { "epoch": 7.757704569606801, "grad_norm": 0.51953125, "learning_rate": 2.889654828892393e-05, "loss": 0.2, "step": 3650 }, { "epoch": 7.768331562167907, "grad_norm": 0.515625, "learning_rate": 2.8635924105635458e-05, "loss": 0.1978, "step": 3655 }, { "epoch": 7.778958554729011, "grad_norm": 0.546875, "learning_rate": 2.8376284019933373e-05, "loss": 0.1944, "step": 3660 }, { "epoch": 7.789585547290117, "grad_norm": 0.474609375, "learning_rate": 2.8117631612207084e-05, "loss": 0.1963, "step": 3665 }, { "epoch": 7.800212539851222, "grad_norm": 0.53515625, "learning_rate": 2.7859970449226104e-05, "loss": 0.1961, "step": 3670 }, { "epoch": 7.810839532412327, "grad_norm": 0.54296875, "learning_rate": 2.760330408409083e-05, "loss": 0.1974, "step": 3675 }, { "epoch": 7.821466524973433, "grad_norm": 0.51171875, "learning_rate": 2.73476360561837e-05, "loss": 0.1977, "step": 3680 }, { "epoch": 7.832093517534537, "grad_norm": 0.55078125, "learning_rate": 2.7092969891120058e-05, "loss": 0.1956, "step": 3685 }, { "epoch": 7.842720510095643, "grad_norm": 0.51171875, "learning_rate": 2.6839309100699973e-05, "loss": 0.1905, "step": 3690 }, { "epoch": 7.8533475026567485, "grad_norm": 0.5234375, "learning_rate": 2.6586657182859343e-05, "loss": 0.1984, "step": 3695 }, { "epoch": 7.863974495217853, "grad_norm": 0.5625, "learning_rate": 2.6335017621622116e-05, "loss": 0.1949, "step": 3700 }, { "epoch": 7.874601487778959, "grad_norm": 0.5703125, "learning_rate": 2.6084393887051884e-05, "loss": 0.196, "step": 3705 }, { "epoch": 7.885228480340063, "grad_norm": 0.5234375, "learning_rate": 2.5834789435204243e-05, "loss": 0.1922, "step": 3710 }, { "epoch": 7.895855472901169, "grad_norm": 0.478515625, "learning_rate": 2.5586207708079034e-05, "loss": 0.1921, "step": 3715 }, { "epoch": 7.9064824654622745, "grad_norm": 0.55078125, "learning_rate": 2.5338652133572915e-05, "loss": 0.1969, "step": 3720 }, { "epoch": 7.917109458023379, "grad_norm": 0.54296875, "learning_rate": 2.5092126125432136e-05, "loss": 0.2008, "step": 3725 }, { "epoch": 7.927736450584485, "grad_norm": 0.54296875, "learning_rate": 2.4846633083205263e-05, "loss": 0.1963, "step": 3730 }, { "epoch": 7.93836344314559, "grad_norm": 0.49609375, "learning_rate": 2.4602176392196662e-05, "loss": 0.1944, "step": 3735 }, { "epoch": 7.948990435706695, "grad_norm": 0.51953125, "learning_rate": 2.4358759423419474e-05, "loss": 0.1954, "step": 3740 }, { "epoch": 7.9596174282678005, "grad_norm": 0.5, "learning_rate": 2.411638553354928e-05, "loss": 0.1979, "step": 3745 }, { "epoch": 7.970244420828905, "grad_norm": 0.50390625, "learning_rate": 2.3875058064877807e-05, "loss": 0.195, "step": 3750 }, { "epoch": 7.980871413390011, "grad_norm": 0.515625, "learning_rate": 2.3634780345266806e-05, "loss": 0.1936, "step": 3755 }, { "epoch": 7.991498405951116, "grad_norm": 0.5, "learning_rate": 2.339555568810221e-05, "loss": 0.1968, "step": 3760 }, { "epoch": 8.0, "eval_loss": 3.6609108448028564, "eval_runtime": 0.7997, "eval_samples_per_second": 6.252, "eval_steps_per_second": 1.25, "step": 3764 }, { "epoch": 8.002125398512222, "grad_norm": 0.431640625, "learning_rate": 2.3157387392248385e-05, "loss": 0.1921, "step": 3765 }, { "epoch": 8.012752391073326, "grad_norm": 0.41015625, "learning_rate": 2.2920278742002676e-05, "loss": 0.1828, "step": 3770 }, { "epoch": 8.023379383634431, "grad_norm": 0.453125, "learning_rate": 2.2684233007050115e-05, "loss": 0.1811, "step": 3775 }, { "epoch": 8.034006376195537, "grad_norm": 0.515625, "learning_rate": 2.244925344241828e-05, "loss": 0.1809, "step": 3780 }, { "epoch": 8.044633368756642, "grad_norm": 0.4765625, "learning_rate": 2.221534328843261e-05, "loss": 0.1758, "step": 3785 }, { "epoch": 8.055260361317748, "grad_norm": 0.51171875, "learning_rate": 2.1982505770671303e-05, "loss": 0.1764, "step": 3790 }, { "epoch": 8.065887353878852, "grad_norm": 0.51953125, "learning_rate": 2.1750744099921396e-05, "loss": 0.1779, "step": 3795 }, { "epoch": 8.076514346439957, "grad_norm": 0.478515625, "learning_rate": 2.1520061472133902e-05, "loss": 0.1798, "step": 3800 }, { "epoch": 8.087141339001063, "grad_norm": 0.51953125, "learning_rate": 2.1290461068380298e-05, "loss": 0.1756, "step": 3805 }, { "epoch": 8.097768331562168, "grad_norm": 0.4921875, "learning_rate": 2.1061946054808146e-05, "loss": 0.1804, "step": 3810 }, { "epoch": 8.108395324123274, "grad_norm": 0.51953125, "learning_rate": 2.0834519582597878e-05, "loss": 0.1782, "step": 3815 }, { "epoch": 8.119022316684378, "grad_norm": 0.490234375, "learning_rate": 2.0608184787919026e-05, "loss": 0.1804, "step": 3820 }, { "epoch": 8.129649309245483, "grad_norm": 0.51171875, "learning_rate": 2.0382944791887106e-05, "loss": 0.1795, "step": 3825 }, { "epoch": 8.140276301806589, "grad_norm": 0.490234375, "learning_rate": 2.0158802700520574e-05, "loss": 0.1815, "step": 3830 }, { "epoch": 8.150903294367694, "grad_norm": 0.4921875, "learning_rate": 1.9935761604697976e-05, "loss": 0.1793, "step": 3835 }, { "epoch": 8.1615302869288, "grad_norm": 0.4765625, "learning_rate": 1.9713824580115335e-05, "loss": 0.1785, "step": 3840 }, { "epoch": 8.172157279489904, "grad_norm": 0.486328125, "learning_rate": 1.9492994687243714e-05, "loss": 0.1784, "step": 3845 }, { "epoch": 8.182784272051009, "grad_norm": 0.48046875, "learning_rate": 1.927327497128706e-05, "loss": 0.177, "step": 3850 }, { "epoch": 8.193411264612115, "grad_norm": 0.4921875, "learning_rate": 1.9054668462140168e-05, "loss": 0.1804, "step": 3855 }, { "epoch": 8.20403825717322, "grad_norm": 0.515625, "learning_rate": 1.883717817434688e-05, "loss": 0.1753, "step": 3860 }, { "epoch": 8.214665249734326, "grad_norm": 0.484375, "learning_rate": 1.8620807107058668e-05, "loss": 0.1785, "step": 3865 }, { "epoch": 8.22529224229543, "grad_norm": 0.46875, "learning_rate": 1.840555824399296e-05, "loss": 0.1775, "step": 3870 }, { "epoch": 8.235919234856535, "grad_norm": 0.578125, "learning_rate": 1.8191434553392428e-05, "loss": 0.18, "step": 3875 }, { "epoch": 8.24654622741764, "grad_norm": 0.515625, "learning_rate": 1.797843898798358e-05, "loss": 0.1806, "step": 3880 }, { "epoch": 8.257173219978746, "grad_norm": 0.51171875, "learning_rate": 1.7766574484936482e-05, "loss": 0.1837, "step": 3885 }, { "epoch": 8.267800212539852, "grad_norm": 0.494140625, "learning_rate": 1.7555843965823992e-05, "loss": 0.1817, "step": 3890 }, { "epoch": 8.278427205100957, "grad_norm": 0.5390625, "learning_rate": 1.7346250336581427e-05, "loss": 0.1806, "step": 3895 }, { "epoch": 8.289054197662061, "grad_norm": 0.486328125, "learning_rate": 1.7137796487466797e-05, "loss": 0.1791, "step": 3900 }, { "epoch": 8.299681190223167, "grad_norm": 0.498046875, "learning_rate": 1.693048529302056e-05, "loss": 0.1788, "step": 3905 }, { "epoch": 8.310308182784272, "grad_norm": 0.46484375, "learning_rate": 1.672431961202635e-05, "loss": 0.1809, "step": 3910 }, { "epoch": 8.320935175345378, "grad_norm": 0.484375, "learning_rate": 1.6519302287471207e-05, "loss": 0.1774, "step": 3915 }, { "epoch": 8.331562167906483, "grad_norm": 0.5, "learning_rate": 1.6315436146506703e-05, "loss": 0.1737, "step": 3920 }, { "epoch": 8.342189160467587, "grad_norm": 0.50390625, "learning_rate": 1.6112724000409694e-05, "loss": 0.1813, "step": 3925 }, { "epoch": 8.352816153028693, "grad_norm": 0.4921875, "learning_rate": 1.5911168644543707e-05, "loss": 0.175, "step": 3930 }, { "epoch": 8.363443145589798, "grad_norm": 0.5546875, "learning_rate": 1.57107728583203e-05, "loss": 0.1838, "step": 3935 }, { "epoch": 8.374070138150904, "grad_norm": 0.51953125, "learning_rate": 1.5511539405160825e-05, "loss": 0.1783, "step": 3940 }, { "epoch": 8.38469713071201, "grad_norm": 0.5234375, "learning_rate": 1.5313471032458247e-05, "loss": 0.1763, "step": 3945 }, { "epoch": 8.395324123273113, "grad_norm": 0.51953125, "learning_rate": 1.5116570471539293e-05, "loss": 0.1758, "step": 3950 }, { "epoch": 8.405951115834219, "grad_norm": 0.50390625, "learning_rate": 1.4920840437626759e-05, "loss": 0.1829, "step": 3955 }, { "epoch": 8.416578108395324, "grad_norm": 0.53125, "learning_rate": 1.4726283629802107e-05, "loss": 0.1779, "step": 3960 }, { "epoch": 8.42720510095643, "grad_norm": 0.53125, "learning_rate": 1.4532902730968212e-05, "loss": 0.1792, "step": 3965 }, { "epoch": 8.437832093517535, "grad_norm": 0.48828125, "learning_rate": 1.4340700407812435e-05, "loss": 0.1813, "step": 3970 }, { "epoch": 8.448459086078639, "grad_norm": 0.50390625, "learning_rate": 1.414967931076967e-05, "loss": 0.1785, "step": 3975 }, { "epoch": 8.459086078639745, "grad_norm": 0.5625, "learning_rate": 1.3959842073986085e-05, "loss": 0.1746, "step": 3980 }, { "epoch": 8.46971307120085, "grad_norm": 0.62109375, "learning_rate": 1.3771191315282427e-05, "loss": 0.1832, "step": 3985 }, { "epoch": 8.480340063761956, "grad_norm": 0.490234375, "learning_rate": 1.3583729636118358e-05, "loss": 0.1843, "step": 3990 }, { "epoch": 8.490967056323061, "grad_norm": 0.515625, "learning_rate": 1.339745962155613e-05, "loss": 0.1801, "step": 3995 }, { "epoch": 8.501594048884165, "grad_norm": 0.54296875, "learning_rate": 1.3212383840225329e-05, "loss": 0.18, "step": 4000 }, { "epoch": 8.51222104144527, "grad_norm": 0.55078125, "learning_rate": 1.3028504844287182e-05, "loss": 0.18, "step": 4005 }, { "epoch": 8.522848034006376, "grad_norm": 0.51171875, "learning_rate": 1.2845825169399507e-05, "loss": 0.1778, "step": 4010 }, { "epoch": 8.533475026567482, "grad_norm": 0.46875, "learning_rate": 1.2664347334681681e-05, "loss": 0.1849, "step": 4015 }, { "epoch": 8.544102019128587, "grad_norm": 0.46875, "learning_rate": 1.2484073842679944e-05, "loss": 0.1799, "step": 4020 }, { "epoch": 8.554729011689691, "grad_norm": 0.478515625, "learning_rate": 1.230500717933285e-05, "loss": 0.1802, "step": 4025 }, { "epoch": 8.565356004250797, "grad_norm": 0.51953125, "learning_rate": 1.2127149813937022e-05, "loss": 0.1776, "step": 4030 }, { "epoch": 8.575982996811902, "grad_norm": 0.52734375, "learning_rate": 1.1950504199113088e-05, "loss": 0.1794, "step": 4035 }, { "epoch": 8.586609989373008, "grad_norm": 0.478515625, "learning_rate": 1.1775072770771834e-05, "loss": 0.1757, "step": 4040 }, { "epoch": 8.597236981934113, "grad_norm": 0.5078125, "learning_rate": 1.1600857948080624e-05, "loss": 0.1786, "step": 4045 }, { "epoch": 8.607863974495217, "grad_norm": 0.50390625, "learning_rate": 1.1427862133430156e-05, "loss": 0.1838, "step": 4050 }, { "epoch": 8.618490967056323, "grad_norm": 0.5234375, "learning_rate": 1.1256087712401087e-05, "loss": 0.1755, "step": 4055 }, { "epoch": 8.629117959617428, "grad_norm": 0.478515625, "learning_rate": 1.1085537053731354e-05, "loss": 0.1777, "step": 4060 }, { "epoch": 8.639744952178534, "grad_norm": 0.5234375, "learning_rate": 1.091621250928343e-05, "loss": 0.1799, "step": 4065 }, { "epoch": 8.65037194473964, "grad_norm": 0.51953125, "learning_rate": 1.0748116414011888e-05, "loss": 0.1774, "step": 4070 }, { "epoch": 8.660998937300743, "grad_norm": 0.5, "learning_rate": 1.058125108593122e-05, "loss": 0.1844, "step": 4075 }, { "epoch": 8.671625929861849, "grad_norm": 0.50390625, "learning_rate": 1.0415618826083828e-05, "loss": 0.1803, "step": 4080 }, { "epoch": 8.682252922422954, "grad_norm": 0.546875, "learning_rate": 1.0251221918508425e-05, "loss": 0.1787, "step": 4085 }, { "epoch": 8.69287991498406, "grad_norm": 0.5546875, "learning_rate": 1.0088062630208273e-05, "loss": 0.1804, "step": 4090 }, { "epoch": 8.703506907545165, "grad_norm": 0.50390625, "learning_rate": 9.926143211120286e-06, "loss": 0.1868, "step": 4095 }, { "epoch": 8.71413390010627, "grad_norm": 0.5, "learning_rate": 9.765465894083636e-06, "loss": 0.1831, "step": 4100 }, { "epoch": 8.724760892667375, "grad_norm": 0.458984375, "learning_rate": 9.606032894809258e-06, "loss": 0.1774, "step": 4105 }, { "epoch": 8.73538788522848, "grad_norm": 0.515625, "learning_rate": 9.447846411849115e-06, "loss": 0.1775, "step": 4110 }, { "epoch": 8.746014877789586, "grad_norm": 0.51953125, "learning_rate": 9.29090862656593e-06, "loss": 0.1796, "step": 4115 }, { "epoch": 8.756641870350691, "grad_norm": 0.5, "learning_rate": 9.135221703103136e-06, "loss": 0.1798, "step": 4120 }, { "epoch": 8.767268862911797, "grad_norm": 0.5, "learning_rate": 8.980787788355016e-06, "loss": 0.179, "step": 4125 }, { "epoch": 8.7778958554729, "grad_norm": 0.46484375, "learning_rate": 8.827609011937066e-06, "loss": 0.1802, "step": 4130 }, { "epoch": 8.788522848034006, "grad_norm": 0.51171875, "learning_rate": 8.675687486156659e-06, "loss": 0.184, "step": 4135 }, { "epoch": 8.799149840595112, "grad_norm": 0.60546875, "learning_rate": 8.525025305983936e-06, "loss": 0.173, "step": 4140 }, { "epoch": 8.809776833156217, "grad_norm": 0.53125, "learning_rate": 8.375624549022854e-06, "loss": 0.179, "step": 4145 }, { "epoch": 8.820403825717323, "grad_norm": 0.56640625, "learning_rate": 8.227487275482592e-06, "loss": 0.1812, "step": 4150 }, { "epoch": 8.831030818278427, "grad_norm": 0.494140625, "learning_rate": 8.08061552814917e-06, "loss": 0.183, "step": 4155 }, { "epoch": 8.841657810839532, "grad_norm": 0.51953125, "learning_rate": 7.935011332357112e-06, "loss": 0.1786, "step": 4160 }, { "epoch": 8.852284803400638, "grad_norm": 0.474609375, "learning_rate": 7.790676695961796e-06, "loss": 0.178, "step": 4165 }, { "epoch": 8.862911795961743, "grad_norm": 0.5, "learning_rate": 7.647613609311455e-06, "loss": 0.1809, "step": 4170 }, { "epoch": 8.873538788522849, "grad_norm": 0.55078125, "learning_rate": 7.505824045220011e-06, "loss": 0.1756, "step": 4175 }, { "epoch": 8.884165781083952, "grad_norm": 0.5390625, "learning_rate": 7.365309958939615e-06, "loss": 0.175, "step": 4180 }, { "epoch": 8.894792773645058, "grad_norm": 0.5078125, "learning_rate": 7.226073288133939e-06, "loss": 0.1792, "step": 4185 }, { "epoch": 8.905419766206164, "grad_norm": 0.486328125, "learning_rate": 7.088115952851238e-06, "loss": 0.1733, "step": 4190 }, { "epoch": 8.91604675876727, "grad_norm": 0.49609375, "learning_rate": 6.951439855498021e-06, "loss": 0.1792, "step": 4195 }, { "epoch": 8.926673751328375, "grad_norm": 0.498046875, "learning_rate": 6.81604688081271e-06, "loss": 0.1803, "step": 4200 }, { "epoch": 8.937300743889478, "grad_norm": 0.45703125, "learning_rate": 6.681938895839746e-06, "loss": 0.177, "step": 4205 }, { "epoch": 8.947927736450584, "grad_norm": 0.486328125, "learning_rate": 6.549117749903755e-06, "loss": 0.1791, "step": 4210 }, { "epoch": 8.95855472901169, "grad_norm": 0.498046875, "learning_rate": 6.417585274584115e-06, "loss": 0.1778, "step": 4215 }, { "epoch": 8.969181721572795, "grad_norm": 0.494140625, "learning_rate": 6.287343283689661e-06, "loss": 0.176, "step": 4220 }, { "epoch": 8.9798087141339, "grad_norm": 0.5234375, "learning_rate": 6.158393573233678e-06, "loss": 0.1759, "step": 4225 }, { "epoch": 8.990435706695006, "grad_norm": 0.5234375, "learning_rate": 6.030737921409169e-06, "loss": 0.1809, "step": 4230 }, { "epoch": 8.99893730074389, "eval_loss": 3.939990520477295, "eval_runtime": 0.8191, "eval_samples_per_second": 6.105, "eval_steps_per_second": 1.221, "step": 4234 }, { "epoch": 9.00106269925611, "grad_norm": 0.47265625, "learning_rate": 5.904378088564255e-06, "loss": 0.1819, "step": 4235 }, { "epoch": 9.011689691817216, "grad_norm": 0.5, "learning_rate": 5.779315817178e-06, "loss": 0.1763, "step": 4240 }, { "epoch": 9.022316684378321, "grad_norm": 0.4921875, "learning_rate": 5.655552831836297e-06, "loss": 0.1767, "step": 4245 }, { "epoch": 9.032943676939427, "grad_norm": 0.46875, "learning_rate": 5.533090839208133e-06, "loss": 0.1744, "step": 4250 }, { "epoch": 9.043570669500532, "grad_norm": 0.490234375, "learning_rate": 5.4119315280220405e-06, "loss": 0.1777, "step": 4255 }, { "epoch": 9.054197662061636, "grad_norm": 0.44921875, "learning_rate": 5.292076569042826e-06, "loss": 0.1744, "step": 4260 }, { "epoch": 9.064824654622742, "grad_norm": 0.451171875, "learning_rate": 5.1735276150484905e-06, "loss": 0.1792, "step": 4265 }, { "epoch": 9.075451647183847, "grad_norm": 0.50390625, "learning_rate": 5.056286300807511e-06, "loss": 0.1751, "step": 4270 }, { "epoch": 9.086078639744953, "grad_norm": 0.47265625, "learning_rate": 4.940354243056178e-06, "loss": 0.174, "step": 4275 }, { "epoch": 9.096705632306058, "grad_norm": 0.458984375, "learning_rate": 4.825733040476465e-06, "loss": 0.1732, "step": 4280 }, { "epoch": 9.107332624867162, "grad_norm": 0.51953125, "learning_rate": 4.712424273673788e-06, "loss": 0.1767, "step": 4285 }, { "epoch": 9.117959617428268, "grad_norm": 0.44921875, "learning_rate": 4.600429505155424e-06, "loss": 0.1726, "step": 4290 }, { "epoch": 9.128586609989373, "grad_norm": 0.4765625, "learning_rate": 4.489750279308757e-06, "loss": 0.174, "step": 4295 }, { "epoch": 9.139213602550479, "grad_norm": 0.484375, "learning_rate": 4.380388122380141e-06, "loss": 0.1787, "step": 4300 }, { "epoch": 9.149840595111584, "grad_norm": 0.515625, "learning_rate": 4.2723445424537544e-06, "loss": 0.1802, "step": 4305 }, { "epoch": 9.160467587672688, "grad_norm": 0.5, "learning_rate": 4.165621029430855e-06, "loss": 0.1786, "step": 4310 }, { "epoch": 9.171094580233794, "grad_norm": 0.462890625, "learning_rate": 4.060219055009229e-06, "loss": 0.1752, "step": 4315 }, { "epoch": 9.181721572794899, "grad_norm": 0.48828125, "learning_rate": 3.9561400726628505e-06, "loss": 0.1793, "step": 4320 }, { "epoch": 9.192348565356005, "grad_norm": 0.56640625, "learning_rate": 3.85338551762191e-06, "loss": 0.175, "step": 4325 }, { "epoch": 9.20297555791711, "grad_norm": 0.462890625, "learning_rate": 3.7519568068529855e-06, "loss": 0.1728, "step": 4330 }, { "epoch": 9.213602550478214, "grad_norm": 0.478515625, "learning_rate": 3.651855339039456e-06, "loss": 0.1698, "step": 4335 }, { "epoch": 9.22422954303932, "grad_norm": 0.69921875, "learning_rate": 3.5530824945623542e-06, "loss": 0.1765, "step": 4340 }, { "epoch": 9.234856535600425, "grad_norm": 0.4765625, "learning_rate": 3.4556396354811226e-06, "loss": 0.1762, "step": 4345 }, { "epoch": 9.24548352816153, "grad_norm": 0.490234375, "learning_rate": 3.359528105515064e-06, "loss": 0.1781, "step": 4350 }, { "epoch": 9.256110520722636, "grad_norm": 0.49609375, "learning_rate": 3.2647492300245994e-06, "loss": 0.1754, "step": 4355 }, { "epoch": 9.26673751328374, "grad_norm": 0.50390625, "learning_rate": 3.1713043159931734e-06, "loss": 0.1767, "step": 4360 }, { "epoch": 9.277364505844846, "grad_norm": 0.515625, "learning_rate": 3.079194652009054e-06, "loss": 0.1708, "step": 4365 }, { "epoch": 9.287991498405951, "grad_norm": 0.48046875, "learning_rate": 2.9884215082477408e-06, "loss": 0.1723, "step": 4370 }, { "epoch": 9.298618490967057, "grad_norm": 0.51171875, "learning_rate": 2.898986136454318e-06, "loss": 0.1801, "step": 4375 }, { "epoch": 9.309245483528162, "grad_norm": 0.494140625, "learning_rate": 2.810889769926217e-06, "loss": 0.1774, "step": 4380 }, { "epoch": 9.319872476089266, "grad_norm": 0.447265625, "learning_rate": 2.7241336234962944e-06, "loss": 0.1722, "step": 4385 }, { "epoch": 9.330499468650371, "grad_norm": 0.515625, "learning_rate": 2.6387188935159456e-06, "loss": 0.1779, "step": 4390 }, { "epoch": 9.341126461211477, "grad_norm": 0.46875, "learning_rate": 2.5546467578387635e-06, "loss": 0.1743, "step": 4395 }, { "epoch": 9.351753453772583, "grad_norm": 0.4921875, "learning_rate": 2.471918375804105e-06, "loss": 0.1733, "step": 4400 }, { "epoch": 9.362380446333688, "grad_norm": 0.4921875, "learning_rate": 2.3905348882212964e-06, "loss": 0.179, "step": 4405 }, { "epoch": 9.373007438894792, "grad_norm": 0.49609375, "learning_rate": 2.3104974173537743e-06, "loss": 0.1742, "step": 4410 }, { "epoch": 9.383634431455897, "grad_norm": 0.484375, "learning_rate": 2.2318070669036685e-06, "loss": 0.1743, "step": 4415 }, { "epoch": 9.394261424017003, "grad_norm": 0.4921875, "learning_rate": 2.1544649219965575e-06, "loss": 0.1734, "step": 4420 }, { "epoch": 9.404888416578109, "grad_norm": 0.46875, "learning_rate": 2.07847204916648e-06, "loss": 0.1729, "step": 4425 }, { "epoch": 9.415515409139214, "grad_norm": 0.546875, "learning_rate": 2.003829496341325e-06, "loss": 0.1762, "step": 4430 }, { "epoch": 9.426142401700318, "grad_norm": 0.51953125, "learning_rate": 1.9305382928282546e-06, "loss": 0.1798, "step": 4435 }, { "epoch": 9.436769394261423, "grad_norm": 0.51953125, "learning_rate": 1.8585994492995916e-06, "loss": 0.1757, "step": 4440 }, { "epoch": 9.447396386822529, "grad_norm": 0.5234375, "learning_rate": 1.788013957778878e-06, "loss": 0.1765, "step": 4445 }, { "epoch": 9.458023379383635, "grad_norm": 0.5, "learning_rate": 1.7187827916271382e-06, "loss": 0.1739, "step": 4450 }, { "epoch": 9.46865037194474, "grad_norm": 0.45703125, "learning_rate": 1.65090690552957e-06, "loss": 0.1732, "step": 4455 }, { "epoch": 9.479277364505846, "grad_norm": 0.498046875, "learning_rate": 1.5843872354822097e-06, "loss": 0.173, "step": 4460 }, { "epoch": 9.48990435706695, "grad_norm": 0.5078125, "learning_rate": 1.5192246987791981e-06, "loss": 0.177, "step": 4465 }, { "epoch": 9.500531349628055, "grad_norm": 0.49609375, "learning_rate": 1.4554201940000123e-06, "loss": 0.177, "step": 4470 }, { "epoch": 9.51115834218916, "grad_norm": 0.5, "learning_rate": 1.3929746009971433e-06, "loss": 0.179, "step": 4475 }, { "epoch": 9.521785334750266, "grad_norm": 0.5234375, "learning_rate": 1.3318887808839274e-06, "loss": 0.1714, "step": 4480 }, { "epoch": 9.532412327311372, "grad_norm": 0.484375, "learning_rate": 1.272163576022667e-06, "loss": 0.1729, "step": 4485 }, { "epoch": 9.543039319872475, "grad_norm": 0.46484375, "learning_rate": 1.21379981001305e-06, "loss": 0.1733, "step": 4490 }, { "epoch": 9.553666312433581, "grad_norm": 0.5078125, "learning_rate": 1.1567982876807603e-06, "loss": 0.175, "step": 4495 }, { "epoch": 9.564293304994687, "grad_norm": 0.474609375, "learning_rate": 1.1011597950663865e-06, "loss": 0.1816, "step": 4500 }, { "epoch": 9.574920297555792, "grad_norm": 0.490234375, "learning_rate": 1.0468850994145963e-06, "loss": 0.1777, "step": 4505 }, { "epoch": 9.585547290116898, "grad_norm": 0.5, "learning_rate": 9.939749491635341e-07, "loss": 0.1776, "step": 4510 }, { "epoch": 9.596174282678001, "grad_norm": 0.5, "learning_rate": 9.424300739345304e-07, "loss": 0.1727, "step": 4515 }, { "epoch": 9.606801275239107, "grad_norm": 0.478515625, "learning_rate": 8.922511845219971e-07, "loss": 0.1734, "step": 4520 }, { "epoch": 9.617428267800213, "grad_norm": 0.49609375, "learning_rate": 8.434389728836589e-07, "loss": 0.1806, "step": 4525 }, { "epoch": 9.628055260361318, "grad_norm": 0.51953125, "learning_rate": 7.959941121310266e-07, "loss": 0.1729, "step": 4530 }, { "epoch": 9.638682252922424, "grad_norm": 0.486328125, "learning_rate": 7.499172565200718e-07, "loss": 0.1729, "step": 4535 }, { "epoch": 9.649309245483527, "grad_norm": 0.455078125, "learning_rate": 7.052090414422119e-07, "loss": 0.1821, "step": 4540 }, { "epoch": 9.659936238044633, "grad_norm": 0.47265625, "learning_rate": 6.618700834155945e-07, "loss": 0.1825, "step": 4545 }, { "epoch": 9.670563230605739, "grad_norm": 0.51171875, "learning_rate": 6.199009800765265e-07, "loss": 0.1709, "step": 4550 }, { "epoch": 9.681190223166844, "grad_norm": 0.5, "learning_rate": 5.793023101712924e-07, "loss": 0.1771, "step": 4555 }, { "epoch": 9.69181721572795, "grad_norm": 0.5390625, "learning_rate": 5.400746335481488e-07, "loss": 0.1771, "step": 4560 }, { "epoch": 9.702444208289053, "grad_norm": 0.52734375, "learning_rate": 5.022184911495864e-07, "loss": 0.1807, "step": 4565 }, { "epoch": 9.713071200850159, "grad_norm": 0.474609375, "learning_rate": 4.6573440500492504e-07, "loss": 0.1677, "step": 4570 }, { "epoch": 9.723698193411265, "grad_norm": 0.5, "learning_rate": 4.306228782230304e-07, "loss": 0.1755, "step": 4575 }, { "epoch": 9.73432518597237, "grad_norm": 0.494140625, "learning_rate": 3.96884394985475e-07, "loss": 0.1741, "step": 4580 }, { "epoch": 9.744952178533476, "grad_norm": 0.466796875, "learning_rate": 3.6451942053975507e-07, "loss": 0.1708, "step": 4585 }, { "epoch": 9.755579171094581, "grad_norm": 0.490234375, "learning_rate": 3.335284011929951e-07, "loss": 0.1732, "step": 4590 }, { "epoch": 9.766206163655685, "grad_norm": 0.48828125, "learning_rate": 3.0391176430567547e-07, "loss": 0.1757, "step": 4595 }, { "epoch": 9.77683315621679, "grad_norm": 0.50390625, "learning_rate": 2.756699182858369e-07, "loss": 0.1786, "step": 4600 }, { "epoch": 9.787460148777896, "grad_norm": 0.48828125, "learning_rate": 2.488032525833628e-07, "loss": 0.1746, "step": 4605 }, { "epoch": 9.798087141339002, "grad_norm": 0.494140625, "learning_rate": 2.2331213768468363e-07, "loss": 0.1767, "step": 4610 }, { "epoch": 9.808714133900107, "grad_norm": 0.484375, "learning_rate": 1.9919692510762533e-07, "loss": 0.1795, "step": 4615 }, { "epoch": 9.819341126461211, "grad_norm": 0.5078125, "learning_rate": 1.7645794739654665e-07, "loss": 0.1764, "step": 4620 }, { "epoch": 9.829968119022316, "grad_norm": 0.5234375, "learning_rate": 1.5509551811783152e-07, "loss": 0.1781, "step": 4625 }, { "epoch": 9.840595111583422, "grad_norm": 0.52734375, "learning_rate": 1.351099318554705e-07, "loss": 0.1755, "step": 4630 }, { "epoch": 9.851222104144528, "grad_norm": 0.498046875, "learning_rate": 1.1650146420704166e-07, "loss": 0.1758, "step": 4635 }, { "epoch": 9.861849096705633, "grad_norm": 0.515625, "learning_rate": 9.927037177993592e-08, "loss": 0.1774, "step": 4640 }, { "epoch": 9.872476089266737, "grad_norm": 0.5234375, "learning_rate": 8.341689218775984e-08, "loss": 0.1827, "step": 4645 }, { "epoch": 9.883103081827842, "grad_norm": 0.49609375, "learning_rate": 6.894124404711599e-08, "loss": 0.1717, "step": 4650 }, { "epoch": 9.893730074388948, "grad_norm": 0.4609375, "learning_rate": 5.584362697453882e-08, "loss": 0.1773, "step": 4655 }, { "epoch": 9.904357066950054, "grad_norm": 0.494140625, "learning_rate": 4.4124221583785595e-08, "loss": 0.1738, "step": 4660 }, { "epoch": 9.91498405951116, "grad_norm": 0.49609375, "learning_rate": 3.378318948332737e-08, "loss": 0.1792, "step": 4665 }, { "epoch": 9.925611052072263, "grad_norm": 0.5859375, "learning_rate": 2.482067327409521e-08, "loss": 0.173, "step": 4670 }, { "epoch": 9.936238044633368, "grad_norm": 0.455078125, "learning_rate": 1.7236796547559497e-08, "loss": 0.1719, "step": 4675 }, { "epoch": 9.946865037194474, "grad_norm": 0.451171875, "learning_rate": 1.103166388398691e-08, "loss": 0.1741, "step": 4680 }, { "epoch": 9.95749202975558, "grad_norm": 0.486328125, "learning_rate": 6.205360851041508e-09, "loss": 0.1773, "step": 4685 }, { "epoch": 9.968119022316685, "grad_norm": 0.490234375, "learning_rate": 2.7579540025524097e-09, "loss": 0.1762, "step": 4690 }, { "epoch": 9.978746014877789, "grad_norm": 0.44140625, "learning_rate": 6.894908776255982e-10, "loss": 0.1745, "step": 4695 }, { "epoch": 9.989373007438894, "grad_norm": 0.51953125, "learning_rate": 0.0, "loss": 0.1757, "step": 4700 }, { "epoch": 9.989373007438894, "eval_loss": 3.9657981395721436, "eval_runtime": 0.7855, "eval_samples_per_second": 6.366, "eval_steps_per_second": 1.273, "step": 4700 }, { "epoch": 9.989373007438894, "step": 4700, "total_flos": 7.255013733871649e+18, "train_loss": 0.5076825852977469, "train_runtime": 24662.4426, "train_samples_per_second": 6.102, "train_steps_per_second": 0.191 } ], "logging_steps": 5, "max_steps": 4700, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 100, "total_flos": 7.255013733871649e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }