diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7087 @@ +{ + "best_metric": 0.10381071222883471, + "best_model_checkpoint": "d:\\\\whisper-medium-pt-cv19-fleurs2-lr\\checkpoint-10000", + "epoch": 11.441647597254004, + "eval_steps": 5000, + "global_step": 25000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.011441647597254004, + "grad_norm": 17.1010799407959, + "learning_rate": 2.875e-08, + "loss": 0.7462, + "step": 25 + }, + { + "epoch": 0.02288329519450801, + "grad_norm": 36.57398986816406, + "learning_rate": 5.8750000000000007e-08, + "loss": 1.2736, + "step": 50 + }, + { + "epoch": 0.034324942791762014, + "grad_norm": 14.23887825012207, + "learning_rate": 9e-08, + "loss": 0.7475, + "step": 75 + }, + { + "epoch": 0.04576659038901602, + "grad_norm": 37.83841323852539, + "learning_rate": 1.2125e-07, + "loss": 1.2338, + "step": 100 + }, + { + "epoch": 0.057208237986270026, + "grad_norm": 13.831230163574219, + "learning_rate": 1.5250000000000002e-07, + "loss": 0.6932, + "step": 125 + }, + { + "epoch": 0.06864988558352403, + "grad_norm": 33.25807189941406, + "learning_rate": 1.8375000000000001e-07, + "loss": 1.1706, + "step": 150 + }, + { + "epoch": 0.08009153318077804, + "grad_norm": 15.954742431640625, + "learning_rate": 2.15e-07, + "loss": 0.5765, + "step": 175 + }, + { + "epoch": 0.09153318077803203, + "grad_norm": 28.214441299438477, + "learning_rate": 2.4624999999999997e-07, + "loss": 0.7797, + "step": 200 + }, + { + "epoch": 0.10297482837528604, + "grad_norm": 9.24880313873291, + "learning_rate": 2.7750000000000004e-07, + "loss": 0.317, + "step": 225 + }, + { + "epoch": 0.11441647597254005, + "grad_norm": 15.88664722442627, + "learning_rate": 3.0875e-07, + "loss": 0.3906, + "step": 250 + }, + { + "epoch": 0.12585812356979406, + "grad_norm": 7.954398155212402, + "learning_rate": 3.4e-07, + "loss": 0.2576, + "step": 275 + }, + { + "epoch": 0.13729977116704806, + "grad_norm": 20.36141586303711, + "learning_rate": 3.7125000000000005e-07, + "loss": 0.3185, + "step": 300 + }, + { + "epoch": 0.14874141876430205, + "grad_norm": 5.76043176651001, + "learning_rate": 4.025e-07, + "loss": 0.1886, + "step": 325 + }, + { + "epoch": 0.16018306636155608, + "grad_norm": 24.122772216796875, + "learning_rate": 4.3375000000000003e-07, + "loss": 0.2623, + "step": 350 + }, + { + "epoch": 0.17162471395881007, + "grad_norm": 8.021683692932129, + "learning_rate": 4.65e-07, + "loss": 0.2107, + "step": 375 + }, + { + "epoch": 0.18306636155606407, + "grad_norm": 20.607276916503906, + "learning_rate": 4.9625e-07, + "loss": 0.2821, + "step": 400 + }, + { + "epoch": 0.1945080091533181, + "grad_norm": 9.00802230834961, + "learning_rate": 5.275e-07, + "loss": 0.1848, + "step": 425 + }, + { + "epoch": 0.20594965675057209, + "grad_norm": 16.48225975036621, + "learning_rate": 5.587499999999999e-07, + "loss": 0.2534, + "step": 450 + }, + { + "epoch": 0.21739130434782608, + "grad_norm": 7.142576217651367, + "learning_rate": 5.9e-07, + "loss": 0.1887, + "step": 475 + }, + { + "epoch": 0.2288329519450801, + "grad_norm": 13.774003982543945, + "learning_rate": 6.212500000000001e-07, + "loss": 0.2385, + "step": 500 + }, + { + "epoch": 0.2402745995423341, + "grad_norm": 9.393098831176758, + "learning_rate": 6.525000000000001e-07, + "loss": 0.1894, + "step": 525 + }, + { + "epoch": 0.2517162471395881, + "grad_norm": 17.331106185913086, + "learning_rate": 6.8375e-07, + "loss": 0.2414, + "step": 550 + }, + { + "epoch": 0.2631578947368421, + "grad_norm": 6.325589656829834, + "learning_rate": 7.15e-07, + "loss": 0.1827, + "step": 575 + }, + { + "epoch": 0.2745995423340961, + "grad_norm": 19.494165420532227, + "learning_rate": 7.462500000000001e-07, + "loss": 0.2265, + "step": 600 + }, + { + "epoch": 0.28604118993135014, + "grad_norm": 8.660221099853516, + "learning_rate": 7.775e-07, + "loss": 0.2014, + "step": 625 + }, + { + "epoch": 0.2974828375286041, + "grad_norm": 16.032672882080078, + "learning_rate": 8.0875e-07, + "loss": 0.2182, + "step": 650 + }, + { + "epoch": 0.30892448512585813, + "grad_norm": 9.232745170593262, + "learning_rate": 8.4e-07, + "loss": 0.1863, + "step": 675 + }, + { + "epoch": 0.32036613272311215, + "grad_norm": 17.713306427001953, + "learning_rate": 8.7125e-07, + "loss": 0.2344, + "step": 700 + }, + { + "epoch": 0.3318077803203661, + "grad_norm": 6.892812728881836, + "learning_rate": 9.025e-07, + "loss": 0.1698, + "step": 725 + }, + { + "epoch": 0.34324942791762014, + "grad_norm": 14.548686981201172, + "learning_rate": 9.337500000000001e-07, + "loss": 0.2197, + "step": 750 + }, + { + "epoch": 0.35469107551487417, + "grad_norm": 6.604827404022217, + "learning_rate": 9.65e-07, + "loss": 0.1939, + "step": 775 + }, + { + "epoch": 0.36613272311212813, + "grad_norm": 12.459800720214844, + "learning_rate": 9.9625e-07, + "loss": 0.2181, + "step": 800 + }, + { + "epoch": 0.37757437070938216, + "grad_norm": 6.9460248947143555, + "learning_rate": 1.0275e-06, + "loss": 0.1656, + "step": 825 + }, + { + "epoch": 0.3890160183066362, + "grad_norm": 14.406033515930176, + "learning_rate": 1.05875e-06, + "loss": 0.2054, + "step": 850 + }, + { + "epoch": 0.40045766590389015, + "grad_norm": 4.875965118408203, + "learning_rate": 1.0900000000000002e-06, + "loss": 0.1801, + "step": 875 + }, + { + "epoch": 0.41189931350114417, + "grad_norm": 10.513298988342285, + "learning_rate": 1.12125e-06, + "loss": 0.2108, + "step": 900 + }, + { + "epoch": 0.4233409610983982, + "grad_norm": 6.491100311279297, + "learning_rate": 1.1525000000000002e-06, + "loss": 0.1719, + "step": 925 + }, + { + "epoch": 0.43478260869565216, + "grad_norm": 17.76156997680664, + "learning_rate": 1.18375e-06, + "loss": 0.1924, + "step": 950 + }, + { + "epoch": 0.4462242562929062, + "grad_norm": 4.623416900634766, + "learning_rate": 1.215e-06, + "loss": 0.1684, + "step": 975 + }, + { + "epoch": 0.4576659038901602, + "grad_norm": 15.875606536865234, + "learning_rate": 1.24625e-06, + "loss": 0.1971, + "step": 1000 + }, + { + "epoch": 0.4691075514874142, + "grad_norm": 5.360893726348877, + "learning_rate": 1.2775e-06, + "loss": 0.1661, + "step": 1025 + }, + { + "epoch": 0.4805491990846682, + "grad_norm": 11.622519493103027, + "learning_rate": 1.3087500000000002e-06, + "loss": 0.1868, + "step": 1050 + }, + { + "epoch": 0.4919908466819222, + "grad_norm": 8.163434982299805, + "learning_rate": 1.34e-06, + "loss": 0.1643, + "step": 1075 + }, + { + "epoch": 0.5034324942791762, + "grad_norm": 14.534429550170898, + "learning_rate": 1.3712500000000002e-06, + "loss": 0.1727, + "step": 1100 + }, + { + "epoch": 0.5148741418764302, + "grad_norm": 9.445674896240234, + "learning_rate": 1.4025e-06, + "loss": 0.1659, + "step": 1125 + }, + { + "epoch": 0.5263157894736842, + "grad_norm": 17.26341438293457, + "learning_rate": 1.43375e-06, + "loss": 0.1929, + "step": 1150 + }, + { + "epoch": 0.5377574370709383, + "grad_norm": 4.513171672821045, + "learning_rate": 1.465e-06, + "loss": 0.169, + "step": 1175 + }, + { + "epoch": 0.5491990846681922, + "grad_norm": 10.165334701538086, + "learning_rate": 1.49625e-06, + "loss": 0.1726, + "step": 1200 + }, + { + "epoch": 0.5606407322654462, + "grad_norm": 4.283674716949463, + "learning_rate": 1.5275000000000002e-06, + "loss": 0.1672, + "step": 1225 + }, + { + "epoch": 0.5720823798627003, + "grad_norm": 14.090361595153809, + "learning_rate": 1.5587500000000001e-06, + "loss": 0.181, + "step": 1250 + }, + { + "epoch": 0.5835240274599542, + "grad_norm": 6.832610130310059, + "learning_rate": 1.5900000000000002e-06, + "loss": 0.1542, + "step": 1275 + }, + { + "epoch": 0.5949656750572082, + "grad_norm": 14.610993385314941, + "learning_rate": 1.6212500000000001e-06, + "loss": 0.2017, + "step": 1300 + }, + { + "epoch": 0.6064073226544623, + "grad_norm": 6.391531467437744, + "learning_rate": 1.6525000000000003e-06, + "loss": 0.1752, + "step": 1325 + }, + { + "epoch": 0.6178489702517163, + "grad_norm": 15.447869300842285, + "learning_rate": 1.68375e-06, + "loss": 0.1862, + "step": 1350 + }, + { + "epoch": 0.6292906178489702, + "grad_norm": 10.263774871826172, + "learning_rate": 1.7149999999999999e-06, + "loss": 0.1728, + "step": 1375 + }, + { + "epoch": 0.6407322654462243, + "grad_norm": 9.934491157531738, + "learning_rate": 1.74625e-06, + "loss": 0.162, + "step": 1400 + }, + { + "epoch": 0.6521739130434783, + "grad_norm": 10.328812599182129, + "learning_rate": 1.7775e-06, + "loss": 0.1635, + "step": 1425 + }, + { + "epoch": 0.6636155606407322, + "grad_norm": 16.462087631225586, + "learning_rate": 1.80875e-06, + "loss": 0.1785, + "step": 1450 + }, + { + "epoch": 0.6750572082379863, + "grad_norm": 6.4470534324646, + "learning_rate": 1.84e-06, + "loss": 0.1706, + "step": 1475 + }, + { + "epoch": 0.6864988558352403, + "grad_norm": 14.632269859313965, + "learning_rate": 1.87125e-06, + "loss": 0.1897, + "step": 1500 + }, + { + "epoch": 0.6979405034324943, + "grad_norm": 5.021676063537598, + "learning_rate": 1.9025000000000002e-06, + "loss": 0.1711, + "step": 1525 + }, + { + "epoch": 0.7093821510297483, + "grad_norm": 14.754373550415039, + "learning_rate": 1.9337500000000003e-06, + "loss": 0.2165, + "step": 1550 + }, + { + "epoch": 0.7208237986270023, + "grad_norm": 8.087540626525879, + "learning_rate": 1.9650000000000002e-06, + "loss": 0.1793, + "step": 1575 + }, + { + "epoch": 0.7322654462242563, + "grad_norm": 13.461847305297852, + "learning_rate": 1.99625e-06, + "loss": 0.176, + "step": 1600 + }, + { + "epoch": 0.7437070938215103, + "grad_norm": 6.565804958343506, + "learning_rate": 2.0275e-06, + "loss": 0.1696, + "step": 1625 + }, + { + "epoch": 0.7551487414187643, + "grad_norm": 16.43035888671875, + "learning_rate": 2.0587500000000004e-06, + "loss": 0.1796, + "step": 1650 + }, + { + "epoch": 0.7665903890160183, + "grad_norm": 7.316359996795654, + "learning_rate": 2.09e-06, + "loss": 0.1565, + "step": 1675 + }, + { + "epoch": 0.7780320366132724, + "grad_norm": 14.20641803741455, + "learning_rate": 2.12125e-06, + "loss": 0.1793, + "step": 1700 + }, + { + "epoch": 0.7894736842105263, + "grad_norm": 5.601787090301514, + "learning_rate": 2.1525e-06, + "loss": 0.1518, + "step": 1725 + }, + { + "epoch": 0.8009153318077803, + "grad_norm": 13.47598648071289, + "learning_rate": 2.18375e-06, + "loss": 0.1601, + "step": 1750 + }, + { + "epoch": 0.8123569794050344, + "grad_norm": 5.27864408493042, + "learning_rate": 2.215e-06, + "loss": 0.1656, + "step": 1775 + }, + { + "epoch": 0.8237986270022883, + "grad_norm": 9.679104804992676, + "learning_rate": 2.24625e-06, + "loss": 0.1842, + "step": 1800 + }, + { + "epoch": 0.8352402745995423, + "grad_norm": 4.940381050109863, + "learning_rate": 2.2775000000000002e-06, + "loss": 0.1455, + "step": 1825 + }, + { + "epoch": 0.8466819221967964, + "grad_norm": 15.962414741516113, + "learning_rate": 2.30875e-06, + "loss": 0.1936, + "step": 1850 + }, + { + "epoch": 0.8581235697940504, + "grad_norm": 6.16251802444458, + "learning_rate": 2.34e-06, + "loss": 0.1623, + "step": 1875 + }, + { + "epoch": 0.8695652173913043, + "grad_norm": 16.07195281982422, + "learning_rate": 2.3712500000000004e-06, + "loss": 0.1932, + "step": 1900 + }, + { + "epoch": 0.8810068649885584, + "grad_norm": 5.088130950927734, + "learning_rate": 2.4025000000000003e-06, + "loss": 0.1568, + "step": 1925 + }, + { + "epoch": 0.8924485125858124, + "grad_norm": 10.709908485412598, + "learning_rate": 2.43375e-06, + "loss": 0.165, + "step": 1950 + }, + { + "epoch": 0.9038901601830663, + "grad_norm": 6.499959945678711, + "learning_rate": 2.465e-06, + "loss": 0.1385, + "step": 1975 + }, + { + "epoch": 0.9153318077803204, + "grad_norm": 13.85750675201416, + "learning_rate": 2.49625e-06, + "loss": 0.1911, + "step": 2000 + }, + { + "epoch": 0.9267734553775744, + "grad_norm": 5.7312822341918945, + "learning_rate": 2.5275e-06, + "loss": 0.1525, + "step": 2025 + }, + { + "epoch": 0.9382151029748284, + "grad_norm": 12.252706527709961, + "learning_rate": 2.55875e-06, + "loss": 0.2005, + "step": 2050 + }, + { + "epoch": 0.9496567505720824, + "grad_norm": 11.799100875854492, + "learning_rate": 2.59e-06, + "loss": 0.1695, + "step": 2075 + }, + { + "epoch": 0.9610983981693364, + "grad_norm": 16.282901763916016, + "learning_rate": 2.62125e-06, + "loss": 0.2019, + "step": 2100 + }, + { + "epoch": 0.9725400457665904, + "grad_norm": 4.300443172454834, + "learning_rate": 2.6525e-06, + "loss": 0.1514, + "step": 2125 + }, + { + "epoch": 0.9839816933638444, + "grad_norm": 17.060945510864258, + "learning_rate": 2.6837500000000004e-06, + "loss": 0.1909, + "step": 2150 + }, + { + "epoch": 0.9954233409610984, + "grad_norm": 8.843390464782715, + "learning_rate": 2.7150000000000003e-06, + "loss": 0.1605, + "step": 2175 + }, + { + "epoch": 1.0068649885583525, + "grad_norm": 3.554812431335449, + "learning_rate": 2.74625e-06, + "loss": 0.1245, + "step": 2200 + }, + { + "epoch": 1.0183066361556063, + "grad_norm": 5.7890305519104, + "learning_rate": 2.7775e-06, + "loss": 0.0978, + "step": 2225 + }, + { + "epoch": 1.0297482837528604, + "grad_norm": 2.891451358795166, + "learning_rate": 2.8087500000000004e-06, + "loss": 0.1469, + "step": 2250 + }, + { + "epoch": 1.0411899313501145, + "grad_norm": 5.289301872253418, + "learning_rate": 2.8400000000000003e-06, + "loss": 0.1022, + "step": 2275 + }, + { + "epoch": 1.0526315789473684, + "grad_norm": 6.965038299560547, + "learning_rate": 2.87125e-06, + "loss": 0.134, + "step": 2300 + }, + { + "epoch": 1.0640732265446224, + "grad_norm": 18.425371170043945, + "learning_rate": 2.9025e-06, + "loss": 0.0921, + "step": 2325 + }, + { + "epoch": 1.0755148741418765, + "grad_norm": 4.118972301483154, + "learning_rate": 2.93375e-06, + "loss": 0.1319, + "step": 2350 + }, + { + "epoch": 1.0869565217391304, + "grad_norm": 5.5514702796936035, + "learning_rate": 2.965e-06, + "loss": 0.0917, + "step": 2375 + }, + { + "epoch": 1.0983981693363845, + "grad_norm": 3.7737677097320557, + "learning_rate": 2.99625e-06, + "loss": 0.1433, + "step": 2400 + }, + { + "epoch": 1.1098398169336385, + "grad_norm": 5.6356987953186035, + "learning_rate": 3.0275000000000002e-06, + "loss": 0.1072, + "step": 2425 + }, + { + "epoch": 1.1212814645308924, + "grad_norm": 3.293067693710327, + "learning_rate": 3.05875e-06, + "loss": 0.1306, + "step": 2450 + }, + { + "epoch": 1.1327231121281465, + "grad_norm": 8.140530586242676, + "learning_rate": 3.09e-06, + "loss": 0.0957, + "step": 2475 + }, + { + "epoch": 1.1441647597254005, + "grad_norm": 12.179439544677734, + "learning_rate": 3.1212500000000004e-06, + "loss": 0.1193, + "step": 2500 + }, + { + "epoch": 1.1556064073226544, + "grad_norm": 5.836298942565918, + "learning_rate": 3.1525e-06, + "loss": 0.1076, + "step": 2525 + }, + { + "epoch": 1.1670480549199085, + "grad_norm": 5.364569664001465, + "learning_rate": 3.18375e-06, + "loss": 0.1626, + "step": 2550 + }, + { + "epoch": 1.1784897025171626, + "grad_norm": 5.4423723220825195, + "learning_rate": 3.215e-06, + "loss": 0.1094, + "step": 2575 + }, + { + "epoch": 1.1899313501144164, + "grad_norm": 3.647977590560913, + "learning_rate": 3.24625e-06, + "loss": 0.129, + "step": 2600 + }, + { + "epoch": 1.2013729977116705, + "grad_norm": 6.781228065490723, + "learning_rate": 3.2775e-06, + "loss": 0.1072, + "step": 2625 + }, + { + "epoch": 1.2128146453089246, + "grad_norm": 3.458967924118042, + "learning_rate": 3.30875e-06, + "loss": 0.1365, + "step": 2650 + }, + { + "epoch": 1.2242562929061784, + "grad_norm": 7.483636856079102, + "learning_rate": 3.34e-06, + "loss": 0.1026, + "step": 2675 + }, + { + "epoch": 1.2356979405034325, + "grad_norm": 5.426140308380127, + "learning_rate": 3.37125e-06, + "loss": 0.1561, + "step": 2700 + }, + { + "epoch": 1.2471395881006866, + "grad_norm": 6.9238104820251465, + "learning_rate": 3.4025e-06, + "loss": 0.1244, + "step": 2725 + }, + { + "epoch": 1.2585812356979404, + "grad_norm": 3.6151397228240967, + "learning_rate": 3.4337500000000004e-06, + "loss": 0.1412, + "step": 2750 + }, + { + "epoch": 1.2700228832951945, + "grad_norm": 8.311691284179688, + "learning_rate": 3.4650000000000003e-06, + "loss": 0.108, + "step": 2775 + }, + { + "epoch": 1.2814645308924484, + "grad_norm": 3.9945647716522217, + "learning_rate": 3.49625e-06, + "loss": 0.1473, + "step": 2800 + }, + { + "epoch": 1.2929061784897025, + "grad_norm": 8.727825164794922, + "learning_rate": 3.5275e-06, + "loss": 0.1101, + "step": 2825 + }, + { + "epoch": 1.3043478260869565, + "grad_norm": 6.813472270965576, + "learning_rate": 3.5587500000000004e-06, + "loss": 0.1693, + "step": 2850 + }, + { + "epoch": 1.3157894736842106, + "grad_norm": 7.848366737365723, + "learning_rate": 3.5900000000000004e-06, + "loss": 0.1067, + "step": 2875 + }, + { + "epoch": 1.3272311212814645, + "grad_norm": 7.600481986999512, + "learning_rate": 3.6212500000000003e-06, + "loss": 0.1386, + "step": 2900 + }, + { + "epoch": 1.3386727688787186, + "grad_norm": 6.826284885406494, + "learning_rate": 3.6525e-06, + "loss": 0.1102, + "step": 2925 + }, + { + "epoch": 1.3501144164759724, + "grad_norm": 6.185801029205322, + "learning_rate": 3.6837500000000005e-06, + "loss": 0.1441, + "step": 2950 + }, + { + "epoch": 1.3615560640732265, + "grad_norm": 7.203296661376953, + "learning_rate": 3.7150000000000004e-06, + "loss": 0.1033, + "step": 2975 + }, + { + "epoch": 1.3729977116704806, + "grad_norm": 4.766386032104492, + "learning_rate": 3.7462500000000003e-06, + "loss": 0.152, + "step": 3000 + }, + { + "epoch": 1.3844393592677346, + "grad_norm": 5.796128273010254, + "learning_rate": 3.7775000000000007e-06, + "loss": 0.0942, + "step": 3025 + }, + { + "epoch": 1.3958810068649885, + "grad_norm": 3.6194448471069336, + "learning_rate": 3.8087500000000006e-06, + "loss": 0.1314, + "step": 3050 + }, + { + "epoch": 1.4073226544622426, + "grad_norm": 1.8742519617080688, + "learning_rate": 3.84e-06, + "loss": 0.0959, + "step": 3075 + }, + { + "epoch": 1.4187643020594964, + "grad_norm": 5.452184200286865, + "learning_rate": 3.8712499999999996e-06, + "loss": 0.1477, + "step": 3100 + }, + { + "epoch": 1.4302059496567505, + "grad_norm": 9.028647422790527, + "learning_rate": 3.9025e-06, + "loss": 0.0983, + "step": 3125 + }, + { + "epoch": 1.4416475972540046, + "grad_norm": 5.759121417999268, + "learning_rate": 3.93375e-06, + "loss": 0.1264, + "step": 3150 + }, + { + "epoch": 1.4530892448512587, + "grad_norm": 4.013973236083984, + "learning_rate": 3.965e-06, + "loss": 0.1103, + "step": 3175 + }, + { + "epoch": 1.4645308924485125, + "grad_norm": 4.706231594085693, + "learning_rate": 3.99625e-06, + "loss": 0.1425, + "step": 3200 + }, + { + "epoch": 1.4759725400457666, + "grad_norm": 7.723937511444092, + "learning_rate": 4.0275e-06, + "loss": 0.0989, + "step": 3225 + }, + { + "epoch": 1.4874141876430205, + "grad_norm": 4.614337921142578, + "learning_rate": 4.05875e-06, + "loss": 0.1456, + "step": 3250 + }, + { + "epoch": 1.4988558352402745, + "grad_norm": 6.104960918426514, + "learning_rate": 4.09e-06, + "loss": 0.1257, + "step": 3275 + }, + { + "epoch": 1.5102974828375286, + "grad_norm": 3.505063772201538, + "learning_rate": 4.12125e-06, + "loss": 0.1506, + "step": 3300 + }, + { + "epoch": 1.5217391304347827, + "grad_norm": 11.545547485351562, + "learning_rate": 4.1525000000000005e-06, + "loss": 0.1014, + "step": 3325 + }, + { + "epoch": 1.5331807780320366, + "grad_norm": 5.136053085327148, + "learning_rate": 4.18375e-06, + "loss": 0.1383, + "step": 3350 + }, + { + "epoch": 1.5446224256292906, + "grad_norm": 7.054224491119385, + "learning_rate": 4.215e-06, + "loss": 0.1002, + "step": 3375 + }, + { + "epoch": 1.5560640732265445, + "grad_norm": 6.568813800811768, + "learning_rate": 4.24625e-06, + "loss": 0.1384, + "step": 3400 + }, + { + "epoch": 1.5675057208237986, + "grad_norm": 5.4689202308654785, + "learning_rate": 4.2775e-06, + "loss": 0.1155, + "step": 3425 + }, + { + "epoch": 1.5789473684210527, + "grad_norm": 3.3778083324432373, + "learning_rate": 4.30875e-06, + "loss": 0.1466, + "step": 3450 + }, + { + "epoch": 1.5903890160183067, + "grad_norm": 6.343507766723633, + "learning_rate": 4.34e-06, + "loss": 0.1067, + "step": 3475 + }, + { + "epoch": 1.6018306636155606, + "grad_norm": 4.513115882873535, + "learning_rate": 4.371250000000001e-06, + "loss": 0.1327, + "step": 3500 + }, + { + "epoch": 1.6132723112128147, + "grad_norm": 7.246913909912109, + "learning_rate": 4.402500000000001e-06, + "loss": 0.1074, + "step": 3525 + }, + { + "epoch": 1.6247139588100685, + "grad_norm": 6.542562961578369, + "learning_rate": 4.4337500000000005e-06, + "loss": 0.1428, + "step": 3550 + }, + { + "epoch": 1.6361556064073226, + "grad_norm": 5.578348159790039, + "learning_rate": 4.4650000000000004e-06, + "loss": 0.1148, + "step": 3575 + }, + { + "epoch": 1.6475972540045767, + "grad_norm": 5.8001885414123535, + "learning_rate": 4.49625e-06, + "loss": 0.1406, + "step": 3600 + }, + { + "epoch": 1.6590389016018308, + "grad_norm": 9.58809757232666, + "learning_rate": 4.5275e-06, + "loss": 0.1139, + "step": 3625 + }, + { + "epoch": 1.6704805491990846, + "grad_norm": 5.428323268890381, + "learning_rate": 4.55875e-06, + "loss": 0.1471, + "step": 3650 + }, + { + "epoch": 1.6819221967963387, + "grad_norm": 8.763864517211914, + "learning_rate": 4.590000000000001e-06, + "loss": 0.1075, + "step": 3675 + }, + { + "epoch": 1.6933638443935926, + "grad_norm": 3.4743332862854004, + "learning_rate": 4.62125e-06, + "loss": 0.1537, + "step": 3700 + }, + { + "epoch": 1.7048054919908466, + "grad_norm": 6.93314790725708, + "learning_rate": 4.6525e-06, + "loss": 0.1257, + "step": 3725 + }, + { + "epoch": 1.7162471395881007, + "grad_norm": 3.2530903816223145, + "learning_rate": 4.68375e-06, + "loss": 0.1429, + "step": 3750 + }, + { + "epoch": 1.7276887871853548, + "grad_norm": 5.646566390991211, + "learning_rate": 4.715e-06, + "loss": 0.1167, + "step": 3775 + }, + { + "epoch": 1.7391304347826086, + "grad_norm": 4.806797981262207, + "learning_rate": 4.74625e-06, + "loss": 0.1387, + "step": 3800 + }, + { + "epoch": 1.7505720823798627, + "grad_norm": 8.664536476135254, + "learning_rate": 4.7775e-06, + "loss": 0.1173, + "step": 3825 + }, + { + "epoch": 1.7620137299771166, + "grad_norm": 2.8637678623199463, + "learning_rate": 4.80875e-06, + "loss": 0.1539, + "step": 3850 + }, + { + "epoch": 1.7734553775743707, + "grad_norm": 7.306221961975098, + "learning_rate": 4.84e-06, + "loss": 0.1342, + "step": 3875 + }, + { + "epoch": 1.7848970251716247, + "grad_norm": 3.557720184326172, + "learning_rate": 4.87125e-06, + "loss": 0.1611, + "step": 3900 + }, + { + "epoch": 1.7963386727688788, + "grad_norm": 10.298198699951172, + "learning_rate": 4.9025e-06, + "loss": 0.1093, + "step": 3925 + }, + { + "epoch": 1.8077803203661327, + "grad_norm": 3.6880478858947754, + "learning_rate": 4.93375e-06, + "loss": 0.1332, + "step": 3950 + }, + { + "epoch": 1.8192219679633868, + "grad_norm": 5.447417259216309, + "learning_rate": 4.965e-06, + "loss": 0.1232, + "step": 3975 + }, + { + "epoch": 1.8306636155606406, + "grad_norm": 7.719545364379883, + "learning_rate": 4.996250000000001e-06, + "loss": 0.1445, + "step": 4000 + }, + { + "epoch": 1.8421052631578947, + "grad_norm": 5.616191387176514, + "learning_rate": 5.0275000000000006e-06, + "loss": 0.1335, + "step": 4025 + }, + { + "epoch": 1.8535469107551488, + "grad_norm": 4.927710056304932, + "learning_rate": 5.0587500000000005e-06, + "loss": 0.139, + "step": 4050 + }, + { + "epoch": 1.8649885583524028, + "grad_norm": 9.265386581420898, + "learning_rate": 5.09e-06, + "loss": 0.1206, + "step": 4075 + }, + { + "epoch": 1.8764302059496567, + "grad_norm": 6.632187366485596, + "learning_rate": 5.12125e-06, + "loss": 0.1312, + "step": 4100 + }, + { + "epoch": 1.8878718535469108, + "grad_norm": 11.779533386230469, + "learning_rate": 5.1525e-06, + "loss": 0.1116, + "step": 4125 + }, + { + "epoch": 1.8993135011441646, + "grad_norm": 3.810981035232544, + "learning_rate": 5.182500000000001e-06, + "loss": 0.1401, + "step": 4150 + }, + { + "epoch": 1.9107551487414187, + "grad_norm": 4.736062526702881, + "learning_rate": 5.213750000000001e-06, + "loss": 0.1171, + "step": 4175 + }, + { + "epoch": 1.9221967963386728, + "grad_norm": 5.427768707275391, + "learning_rate": 5.245e-06, + "loss": 0.1392, + "step": 4200 + }, + { + "epoch": 1.9336384439359269, + "grad_norm": 10.339317321777344, + "learning_rate": 5.27625e-06, + "loss": 0.1249, + "step": 4225 + }, + { + "epoch": 1.9450800915331807, + "grad_norm": 5.156798362731934, + "learning_rate": 5.3075e-06, + "loss": 0.1413, + "step": 4250 + }, + { + "epoch": 1.9565217391304348, + "grad_norm": 6.890321731567383, + "learning_rate": 5.33875e-06, + "loss": 0.1059, + "step": 4275 + }, + { + "epoch": 1.9679633867276887, + "grad_norm": 4.059100151062012, + "learning_rate": 5.37e-06, + "loss": 0.1443, + "step": 4300 + }, + { + "epoch": 1.9794050343249427, + "grad_norm": 12.56505012512207, + "learning_rate": 5.40125e-06, + "loss": 0.1082, + "step": 4325 + }, + { + "epoch": 1.9908466819221968, + "grad_norm": 5.877668857574463, + "learning_rate": 5.4325e-06, + "loss": 0.1328, + "step": 4350 + }, + { + "epoch": 2.002288329519451, + "grad_norm": 3.3472976684570312, + "learning_rate": 5.46375e-06, + "loss": 0.1337, + "step": 4375 + }, + { + "epoch": 2.013729977116705, + "grad_norm": 1.5074313879013062, + "learning_rate": 5.495e-06, + "loss": 0.0594, + "step": 4400 + }, + { + "epoch": 2.0251716247139586, + "grad_norm": 4.803617000579834, + "learning_rate": 5.52625e-06, + "loss": 0.0803, + "step": 4425 + }, + { + "epoch": 2.0366132723112127, + "grad_norm": 3.4739410877227783, + "learning_rate": 5.557500000000001e-06, + "loss": 0.0581, + "step": 4450 + }, + { + "epoch": 2.0480549199084668, + "grad_norm": 2.3615224361419678, + "learning_rate": 5.5887500000000005e-06, + "loss": 0.0815, + "step": 4475 + }, + { + "epoch": 2.059496567505721, + "grad_norm": 2.8116395473480225, + "learning_rate": 5.62e-06, + "loss": 0.066, + "step": 4500 + }, + { + "epoch": 2.070938215102975, + "grad_norm": 3.5427844524383545, + "learning_rate": 5.65125e-06, + "loss": 0.0756, + "step": 4525 + }, + { + "epoch": 2.082379862700229, + "grad_norm": 10.534377098083496, + "learning_rate": 5.6825e-06, + "loss": 0.0608, + "step": 4550 + }, + { + "epoch": 2.0938215102974826, + "grad_norm": 3.33672833442688, + "learning_rate": 5.71375e-06, + "loss": 0.0937, + "step": 4575 + }, + { + "epoch": 2.1052631578947367, + "grad_norm": 3.8001768589019775, + "learning_rate": 5.745e-06, + "loss": 0.058, + "step": 4600 + }, + { + "epoch": 2.116704805491991, + "grad_norm": 4.036559104919434, + "learning_rate": 5.776250000000001e-06, + "loss": 0.0789, + "step": 4625 + }, + { + "epoch": 2.128146453089245, + "grad_norm": 2.8355629444122314, + "learning_rate": 5.807500000000001e-06, + "loss": 0.0503, + "step": 4650 + }, + { + "epoch": 2.139588100686499, + "grad_norm": 6.225577354431152, + "learning_rate": 5.838750000000001e-06, + "loss": 0.0859, + "step": 4675 + }, + { + "epoch": 2.151029748283753, + "grad_norm": 3.4654979705810547, + "learning_rate": 5.8700000000000005e-06, + "loss": 0.0635, + "step": 4700 + }, + { + "epoch": 2.1624713958810067, + "grad_norm": 3.9897263050079346, + "learning_rate": 5.9012500000000005e-06, + "loss": 0.078, + "step": 4725 + }, + { + "epoch": 2.1739130434782608, + "grad_norm": 3.4768152236938477, + "learning_rate": 5.9325e-06, + "loss": 0.0707, + "step": 4750 + }, + { + "epoch": 2.185354691075515, + "grad_norm": 4.6715192794799805, + "learning_rate": 5.96375e-06, + "loss": 0.0937, + "step": 4775 + }, + { + "epoch": 2.196796338672769, + "grad_norm": 3.6018433570861816, + "learning_rate": 5.995e-06, + "loss": 0.0616, + "step": 4800 + }, + { + "epoch": 2.208237986270023, + "grad_norm": 4.695530891418457, + "learning_rate": 6.02625e-06, + "loss": 0.0853, + "step": 4825 + }, + { + "epoch": 2.219679633867277, + "grad_norm": 1.7175928354263306, + "learning_rate": 6.0575e-06, + "loss": 0.0527, + "step": 4850 + }, + { + "epoch": 2.2311212814645307, + "grad_norm": 2.272045850753784, + "learning_rate": 6.08875e-06, + "loss": 0.1014, + "step": 4875 + }, + { + "epoch": 2.242562929061785, + "grad_norm": 9.221810340881348, + "learning_rate": 6.12e-06, + "loss": 0.0698, + "step": 4900 + }, + { + "epoch": 2.254004576659039, + "grad_norm": 3.7419273853302, + "learning_rate": 6.15125e-06, + "loss": 0.1009, + "step": 4925 + }, + { + "epoch": 2.265446224256293, + "grad_norm": 4.7529802322387695, + "learning_rate": 6.1825e-06, + "loss": 0.0715, + "step": 4950 + }, + { + "epoch": 2.276887871853547, + "grad_norm": 3.084136486053467, + "learning_rate": 6.2137500000000004e-06, + "loss": 0.1042, + "step": 4975 + }, + { + "epoch": 2.288329519450801, + "grad_norm": 6.452476501464844, + "learning_rate": 6.245e-06, + "loss": 0.0697, + "step": 5000 + }, + { + "epoch": 2.288329519450801, + "eval_loss": 0.16196583211421967, + "eval_runtime": 5941.1089, + "eval_samples_per_second": 1.603, + "eval_steps_per_second": 0.2, + "eval_wer": 0.10273884942727331, + "step": 5000 + }, + { + "epoch": 2.2997711670480547, + "grad_norm": 3.5335402488708496, + "learning_rate": 6.2434375e-06, + "loss": 0.0812, + "step": 5025 + }, + { + "epoch": 2.311212814645309, + "grad_norm": 7.161555290222168, + "learning_rate": 6.235625e-06, + "loss": 0.0688, + "step": 5050 + }, + { + "epoch": 2.322654462242563, + "grad_norm": 2.389561653137207, + "learning_rate": 6.2278125e-06, + "loss": 0.0872, + "step": 5075 + }, + { + "epoch": 2.334096109839817, + "grad_norm": 6.461544513702393, + "learning_rate": 6.22e-06, + "loss": 0.0598, + "step": 5100 + }, + { + "epoch": 2.345537757437071, + "grad_norm": 4.625803470611572, + "learning_rate": 6.2121875e-06, + "loss": 0.0825, + "step": 5125 + }, + { + "epoch": 2.356979405034325, + "grad_norm": 3.9252727031707764, + "learning_rate": 6.204375e-06, + "loss": 0.0631, + "step": 5150 + }, + { + "epoch": 2.3684210526315788, + "grad_norm": 2.284882068634033, + "learning_rate": 6.196562500000001e-06, + "loss": 0.085, + "step": 5175 + }, + { + "epoch": 2.379862700228833, + "grad_norm": 5.8602399826049805, + "learning_rate": 6.18875e-06, + "loss": 0.0697, + "step": 5200 + }, + { + "epoch": 2.391304347826087, + "grad_norm": 5.568091869354248, + "learning_rate": 6.1809375000000005e-06, + "loss": 0.0872, + "step": 5225 + }, + { + "epoch": 2.402745995423341, + "grad_norm": 7.008312225341797, + "learning_rate": 6.173125e-06, + "loss": 0.0661, + "step": 5250 + }, + { + "epoch": 2.414187643020595, + "grad_norm": 3.3845436573028564, + "learning_rate": 6.165312500000001e-06, + "loss": 0.0912, + "step": 5275 + }, + { + "epoch": 2.425629290617849, + "grad_norm": 3.0679447650909424, + "learning_rate": 6.1575e-06, + "loss": 0.0723, + "step": 5300 + }, + { + "epoch": 2.437070938215103, + "grad_norm": 5.007796287536621, + "learning_rate": 6.1496875000000006e-06, + "loss": 0.0937, + "step": 5325 + }, + { + "epoch": 2.448512585812357, + "grad_norm": 4.2327046394348145, + "learning_rate": 6.141875e-06, + "loss": 0.073, + "step": 5350 + }, + { + "epoch": 2.459954233409611, + "grad_norm": 4.146753311157227, + "learning_rate": 6.1340625e-06, + "loss": 0.0894, + "step": 5375 + }, + { + "epoch": 2.471395881006865, + "grad_norm": 1.6775671243667603, + "learning_rate": 6.12625e-06, + "loss": 0.0866, + "step": 5400 + }, + { + "epoch": 2.482837528604119, + "grad_norm": 3.461623191833496, + "learning_rate": 6.1184375e-06, + "loss": 0.0859, + "step": 5425 + }, + { + "epoch": 2.494279176201373, + "grad_norm": 3.211845636367798, + "learning_rate": 6.1106250000000005e-06, + "loss": 0.0732, + "step": 5450 + }, + { + "epoch": 2.505720823798627, + "grad_norm": 4.490880489349365, + "learning_rate": 6.1028125e-06, + "loss": 0.0871, + "step": 5475 + }, + { + "epoch": 2.517162471395881, + "grad_norm": 4.470268249511719, + "learning_rate": 6.095e-06, + "loss": 0.0734, + "step": 5500 + }, + { + "epoch": 2.528604118993135, + "grad_norm": 5.099210262298584, + "learning_rate": 6.0871875e-06, + "loss": 0.0961, + "step": 5525 + }, + { + "epoch": 2.540045766590389, + "grad_norm": 2.920400381088257, + "learning_rate": 6.0793750000000006e-06, + "loss": 0.0712, + "step": 5550 + }, + { + "epoch": 2.551487414187643, + "grad_norm": 4.1212592124938965, + "learning_rate": 6.0715625e-06, + "loss": 0.0911, + "step": 5575 + }, + { + "epoch": 2.5629290617848968, + "grad_norm": 1.774526596069336, + "learning_rate": 6.06375e-06, + "loss": 0.0664, + "step": 5600 + }, + { + "epoch": 2.5743707093821513, + "grad_norm": 3.186657428741455, + "learning_rate": 6.0559375e-06, + "loss": 0.0807, + "step": 5625 + }, + { + "epoch": 2.585812356979405, + "grad_norm": 1.9952179193496704, + "learning_rate": 6.048125000000001e-06, + "loss": 0.0776, + "step": 5650 + }, + { + "epoch": 2.597254004576659, + "grad_norm": 2.9018378257751465, + "learning_rate": 6.0403125000000005e-06, + "loss": 0.092, + "step": 5675 + }, + { + "epoch": 2.608695652173913, + "grad_norm": 3.827449083328247, + "learning_rate": 6.0325e-06, + "loss": 0.0645, + "step": 5700 + }, + { + "epoch": 2.620137299771167, + "grad_norm": 3.021289825439453, + "learning_rate": 6.0246875e-06, + "loss": 0.0791, + "step": 5725 + }, + { + "epoch": 2.6315789473684212, + "grad_norm": 3.3749895095825195, + "learning_rate": 6.016875e-06, + "loss": 0.075, + "step": 5750 + }, + { + "epoch": 2.643020594965675, + "grad_norm": 4.381983757019043, + "learning_rate": 6.0090625000000005e-06, + "loss": 0.0977, + "step": 5775 + }, + { + "epoch": 2.654462242562929, + "grad_norm": 8.246472358703613, + "learning_rate": 6.00125e-06, + "loss": 0.0669, + "step": 5800 + }, + { + "epoch": 2.665903890160183, + "grad_norm": 2.754366874694824, + "learning_rate": 5.9934375e-06, + "loss": 0.0845, + "step": 5825 + }, + { + "epoch": 2.677345537757437, + "grad_norm": 3.511594295501709, + "learning_rate": 5.985625e-06, + "loss": 0.0663, + "step": 5850 + }, + { + "epoch": 2.688787185354691, + "grad_norm": 3.16924786567688, + "learning_rate": 5.977812500000001e-06, + "loss": 0.0868, + "step": 5875 + }, + { + "epoch": 2.700228832951945, + "grad_norm": 3.8349554538726807, + "learning_rate": 5.9700000000000004e-06, + "loss": 0.0644, + "step": 5900 + }, + { + "epoch": 2.7116704805491993, + "grad_norm": 2.3711776733398438, + "learning_rate": 5.9621875e-06, + "loss": 0.0786, + "step": 5925 + }, + { + "epoch": 2.723112128146453, + "grad_norm": 2.4516854286193848, + "learning_rate": 5.954375e-06, + "loss": 0.0774, + "step": 5950 + }, + { + "epoch": 2.734553775743707, + "grad_norm": 2.746248483657837, + "learning_rate": 5.946562500000001e-06, + "loss": 0.1053, + "step": 5975 + }, + { + "epoch": 2.745995423340961, + "grad_norm": 5.512765407562256, + "learning_rate": 5.9387500000000005e-06, + "loss": 0.0624, + "step": 6000 + }, + { + "epoch": 2.757437070938215, + "grad_norm": 3.6978447437286377, + "learning_rate": 5.9309375e-06, + "loss": 0.0832, + "step": 6025 + }, + { + "epoch": 2.7688787185354693, + "grad_norm": 3.7485339641571045, + "learning_rate": 5.923125e-06, + "loss": 0.0646, + "step": 6050 + }, + { + "epoch": 2.780320366132723, + "grad_norm": 2.9584758281707764, + "learning_rate": 5.9153125e-06, + "loss": 0.0797, + "step": 6075 + }, + { + "epoch": 2.791762013729977, + "grad_norm": 3.063089609146118, + "learning_rate": 5.907500000000001e-06, + "loss": 0.0594, + "step": 6100 + }, + { + "epoch": 2.803203661327231, + "grad_norm": 3.4016189575195312, + "learning_rate": 5.8996875000000004e-06, + "loss": 0.0841, + "step": 6125 + }, + { + "epoch": 2.814645308924485, + "grad_norm": 5.774538993835449, + "learning_rate": 5.8921875e-06, + "loss": 0.0851, + "step": 6150 + }, + { + "epoch": 2.8260869565217392, + "grad_norm": 2.9045443534851074, + "learning_rate": 5.884375e-06, + "loss": 0.0849, + "step": 6175 + }, + { + "epoch": 2.837528604118993, + "grad_norm": 9.496201515197754, + "learning_rate": 5.8765625000000005e-06, + "loss": 0.08, + "step": 6200 + }, + { + "epoch": 2.8489702517162474, + "grad_norm": 4.791667938232422, + "learning_rate": 5.86875e-06, + "loss": 0.0912, + "step": 6225 + }, + { + "epoch": 2.860411899313501, + "grad_norm": 3.803701400756836, + "learning_rate": 5.8609375e-06, + "loss": 0.0662, + "step": 6250 + }, + { + "epoch": 2.871853546910755, + "grad_norm": 7.12540864944458, + "learning_rate": 5.853125e-06, + "loss": 0.0981, + "step": 6275 + }, + { + "epoch": 2.883295194508009, + "grad_norm": 6.205677032470703, + "learning_rate": 5.845312500000001e-06, + "loss": 0.0758, + "step": 6300 + }, + { + "epoch": 2.8947368421052633, + "grad_norm": 4.83319616317749, + "learning_rate": 5.8375000000000004e-06, + "loss": 0.0805, + "step": 6325 + }, + { + "epoch": 2.9061784897025174, + "grad_norm": 6.701275825500488, + "learning_rate": 5.8296875e-06, + "loss": 0.0796, + "step": 6350 + }, + { + "epoch": 2.917620137299771, + "grad_norm": 4.209991455078125, + "learning_rate": 5.821875e-06, + "loss": 0.0876, + "step": 6375 + }, + { + "epoch": 2.929061784897025, + "grad_norm": 3.636922597885132, + "learning_rate": 5.814062500000001e-06, + "loss": 0.0719, + "step": 6400 + }, + { + "epoch": 2.940503432494279, + "grad_norm": 4.24993896484375, + "learning_rate": 5.8062500000000005e-06, + "loss": 0.0848, + "step": 6425 + }, + { + "epoch": 2.9519450800915332, + "grad_norm": 4.109525203704834, + "learning_rate": 5.7984375e-06, + "loss": 0.0775, + "step": 6450 + }, + { + "epoch": 2.9633867276887873, + "grad_norm": 3.515380382537842, + "learning_rate": 5.790625e-06, + "loss": 0.0949, + "step": 6475 + }, + { + "epoch": 2.974828375286041, + "grad_norm": 2.294419765472412, + "learning_rate": 5.782812500000001e-06, + "loss": 0.0721, + "step": 6500 + }, + { + "epoch": 2.9862700228832955, + "grad_norm": 3.1478168964385986, + "learning_rate": 5.775000000000001e-06, + "loss": 0.0784, + "step": 6525 + }, + { + "epoch": 2.997711670480549, + "grad_norm": 6.680340766906738, + "learning_rate": 5.7671875e-06, + "loss": 0.0817, + "step": 6550 + }, + { + "epoch": 3.009153318077803, + "grad_norm": 5.719060897827148, + "learning_rate": 5.759375e-06, + "loss": 0.0532, + "step": 6575 + }, + { + "epoch": 3.0205949656750573, + "grad_norm": 12.910862922668457, + "learning_rate": 5.7515625e-06, + "loss": 0.039, + "step": 6600 + }, + { + "epoch": 3.0320366132723113, + "grad_norm": 2.577425003051758, + "learning_rate": 5.743750000000001e-06, + "loss": 0.0385, + "step": 6625 + }, + { + "epoch": 3.0434782608695654, + "grad_norm": 6.705196380615234, + "learning_rate": 5.7359375e-06, + "loss": 0.0373, + "step": 6650 + }, + { + "epoch": 3.054919908466819, + "grad_norm": 3.4962539672851562, + "learning_rate": 5.728125e-06, + "loss": 0.0542, + "step": 6675 + }, + { + "epoch": 3.066361556064073, + "grad_norm": 7.721837043762207, + "learning_rate": 5.7203125e-06, + "loss": 0.0384, + "step": 6700 + }, + { + "epoch": 3.077803203661327, + "grad_norm": 2.549494504928589, + "learning_rate": 5.712500000000001e-06, + "loss": 0.0443, + "step": 6725 + }, + { + "epoch": 3.0892448512585813, + "grad_norm": 7.875229358673096, + "learning_rate": 5.7046875e-06, + "loss": 0.0546, + "step": 6750 + }, + { + "epoch": 3.1006864988558354, + "grad_norm": 6.519637584686279, + "learning_rate": 5.696875e-06, + "loss": 0.0572, + "step": 6775 + }, + { + "epoch": 3.1121281464530894, + "grad_norm": 3.3160250186920166, + "learning_rate": 5.6890625e-06, + "loss": 0.0268, + "step": 6800 + }, + { + "epoch": 3.123569794050343, + "grad_norm": 1.0826241970062256, + "learning_rate": 5.681250000000001e-06, + "loss": 0.0455, + "step": 6825 + }, + { + "epoch": 3.135011441647597, + "grad_norm": 2.4593420028686523, + "learning_rate": 5.6734375e-06, + "loss": 0.0374, + "step": 6850 + }, + { + "epoch": 3.1464530892448512, + "grad_norm": 5.654216289520264, + "learning_rate": 5.6656250000000005e-06, + "loss": 0.0528, + "step": 6875 + }, + { + "epoch": 3.1578947368421053, + "grad_norm": 4.648204803466797, + "learning_rate": 5.6578125e-06, + "loss": 0.0345, + "step": 6900 + }, + { + "epoch": 3.1693363844393594, + "grad_norm": 4.878942489624023, + "learning_rate": 5.65e-06, + "loss": 0.0397, + "step": 6925 + }, + { + "epoch": 3.1807780320366135, + "grad_norm": 5.794234752655029, + "learning_rate": 5.642187500000001e-06, + "loss": 0.0457, + "step": 6950 + }, + { + "epoch": 3.192219679633867, + "grad_norm": 3.9493894577026367, + "learning_rate": 5.634375e-06, + "loss": 0.0547, + "step": 6975 + }, + { + "epoch": 3.203661327231121, + "grad_norm": 4.193353652954102, + "learning_rate": 5.6265625e-06, + "loss": 0.0383, + "step": 7000 + }, + { + "epoch": 3.2151029748283753, + "grad_norm": 3.423971176147461, + "learning_rate": 5.61875e-06, + "loss": 0.0462, + "step": 7025 + }, + { + "epoch": 3.2265446224256293, + "grad_norm": 6.13396692276001, + "learning_rate": 5.610937500000001e-06, + "loss": 0.0473, + "step": 7050 + }, + { + "epoch": 3.2379862700228834, + "grad_norm": 4.359652996063232, + "learning_rate": 5.603125e-06, + "loss": 0.0511, + "step": 7075 + }, + { + "epoch": 3.2494279176201375, + "grad_norm": 1.4082869291305542, + "learning_rate": 5.5953125000000005e-06, + "loss": 0.0423, + "step": 7100 + }, + { + "epoch": 3.260869565217391, + "grad_norm": 4.264839172363281, + "learning_rate": 5.5875e-06, + "loss": 0.0462, + "step": 7125 + }, + { + "epoch": 3.272311212814645, + "grad_norm": 7.262903213500977, + "learning_rate": 5.579687500000001e-06, + "loss": 0.0478, + "step": 7150 + }, + { + "epoch": 3.2837528604118993, + "grad_norm": 3.770082950592041, + "learning_rate": 5.571875e-06, + "loss": 0.0361, + "step": 7175 + }, + { + "epoch": 3.2951945080091534, + "grad_norm": 7.296878814697266, + "learning_rate": 5.5640625000000006e-06, + "loss": 0.0432, + "step": 7200 + }, + { + "epoch": 3.3066361556064074, + "grad_norm": 2.4708614349365234, + "learning_rate": 5.55625e-06, + "loss": 0.0599, + "step": 7225 + }, + { + "epoch": 3.3180778032036615, + "grad_norm": 4.431317329406738, + "learning_rate": 5.5484375e-06, + "loss": 0.0424, + "step": 7250 + }, + { + "epoch": 3.329519450800915, + "grad_norm": 4.9188127517700195, + "learning_rate": 5.540625e-06, + "loss": 0.0514, + "step": 7275 + }, + { + "epoch": 3.3409610983981692, + "grad_norm": 5.590330123901367, + "learning_rate": 5.5328125e-06, + "loss": 0.0502, + "step": 7300 + }, + { + "epoch": 3.3524027459954233, + "grad_norm": 1.5474858283996582, + "learning_rate": 5.5250000000000005e-06, + "loss": 0.0506, + "step": 7325 + }, + { + "epoch": 3.3638443935926774, + "grad_norm": 6.954276084899902, + "learning_rate": 5.5171875e-06, + "loss": 0.0432, + "step": 7350 + }, + { + "epoch": 3.3752860411899315, + "grad_norm": 2.652070999145508, + "learning_rate": 5.509375e-06, + "loss": 0.0516, + "step": 7375 + }, + { + "epoch": 3.386727688787185, + "grad_norm": 1.9480232000350952, + "learning_rate": 5.5015625e-06, + "loss": 0.0408, + "step": 7400 + }, + { + "epoch": 3.398169336384439, + "grad_norm": 1.5268117189407349, + "learning_rate": 5.4937500000000006e-06, + "loss": 0.0468, + "step": 7425 + }, + { + "epoch": 3.4096109839816933, + "grad_norm": 3.8624870777130127, + "learning_rate": 5.4859375e-06, + "loss": 0.0401, + "step": 7450 + }, + { + "epoch": 3.4210526315789473, + "grad_norm": 3.136197805404663, + "learning_rate": 5.478125e-06, + "loss": 0.0485, + "step": 7475 + }, + { + "epoch": 3.4324942791762014, + "grad_norm": 4.95338249206543, + "learning_rate": 5.4703125e-06, + "loss": 0.0381, + "step": 7500 + }, + { + "epoch": 3.4439359267734555, + "grad_norm": 4.111715793609619, + "learning_rate": 5.462500000000001e-06, + "loss": 0.0484, + "step": 7525 + }, + { + "epoch": 3.4553775743707096, + "grad_norm": 8.481527328491211, + "learning_rate": 5.4546875000000004e-06, + "loss": 0.0573, + "step": 7550 + }, + { + "epoch": 3.466819221967963, + "grad_norm": 2.4452638626098633, + "learning_rate": 5.446875e-06, + "loss": 0.0511, + "step": 7575 + }, + { + "epoch": 3.4782608695652173, + "grad_norm": 4.839463710784912, + "learning_rate": 5.4390625e-06, + "loss": 0.0454, + "step": 7600 + }, + { + "epoch": 3.4897025171624714, + "grad_norm": 3.3421833515167236, + "learning_rate": 5.43125e-06, + "loss": 0.0531, + "step": 7625 + }, + { + "epoch": 3.5011441647597255, + "grad_norm": 5.8438334465026855, + "learning_rate": 5.4234375000000005e-06, + "loss": 0.0464, + "step": 7650 + }, + { + "epoch": 3.5125858123569795, + "grad_norm": 3.1992125511169434, + "learning_rate": 5.415625e-06, + "loss": 0.0558, + "step": 7675 + }, + { + "epoch": 3.524027459954233, + "grad_norm": 2.4741952419281006, + "learning_rate": 5.4078125e-06, + "loss": 0.0496, + "step": 7700 + }, + { + "epoch": 3.5354691075514877, + "grad_norm": 2.3519113063812256, + "learning_rate": 5.4e-06, + "loss": 0.0517, + "step": 7725 + }, + { + "epoch": 3.5469107551487413, + "grad_norm": 3.754331111907959, + "learning_rate": 5.392187500000001e-06, + "loss": 0.0417, + "step": 7750 + }, + { + "epoch": 3.5583524027459954, + "grad_norm": 2.477574586868286, + "learning_rate": 5.3843750000000004e-06, + "loss": 0.048, + "step": 7775 + }, + { + "epoch": 3.5697940503432495, + "grad_norm": 4.506994724273682, + "learning_rate": 5.3765625e-06, + "loss": 0.0393, + "step": 7800 + }, + { + "epoch": 3.5812356979405036, + "grad_norm": 2.5902769565582275, + "learning_rate": 5.36875e-06, + "loss": 0.0493, + "step": 7825 + }, + { + "epoch": 3.5926773455377576, + "grad_norm": 9.635528564453125, + "learning_rate": 5.360937500000001e-06, + "loss": 0.0456, + "step": 7850 + }, + { + "epoch": 3.6041189931350113, + "grad_norm": 5.572010040283203, + "learning_rate": 5.3531250000000005e-06, + "loss": 0.0451, + "step": 7875 + }, + { + "epoch": 3.6155606407322654, + "grad_norm": 8.494470596313477, + "learning_rate": 5.3453125e-06, + "loss": 0.0388, + "step": 7900 + }, + { + "epoch": 3.6270022883295194, + "grad_norm": 3.9414212703704834, + "learning_rate": 5.3375e-06, + "loss": 0.0555, + "step": 7925 + }, + { + "epoch": 3.6384439359267735, + "grad_norm": 4.169886112213135, + "learning_rate": 5.3296875e-06, + "loss": 0.0468, + "step": 7950 + }, + { + "epoch": 3.6498855835240276, + "grad_norm": 3.8904569149017334, + "learning_rate": 5.321875000000001e-06, + "loss": 0.0566, + "step": 7975 + }, + { + "epoch": 3.6613272311212812, + "grad_norm": 4.280102729797363, + "learning_rate": 5.3140624999999996e-06, + "loss": 0.0416, + "step": 8000 + }, + { + "epoch": 3.6727688787185357, + "grad_norm": 3.302794933319092, + "learning_rate": 5.30625e-06, + "loss": 0.0451, + "step": 8025 + }, + { + "epoch": 3.6842105263157894, + "grad_norm": 9.184480667114258, + "learning_rate": 5.2984375e-06, + "loss": 0.0502, + "step": 8050 + }, + { + "epoch": 3.6956521739130435, + "grad_norm": 4.913455009460449, + "learning_rate": 5.290625000000001e-06, + "loss": 0.0452, + "step": 8075 + }, + { + "epoch": 3.7070938215102975, + "grad_norm": 3.561953544616699, + "learning_rate": 5.2828125e-06, + "loss": 0.0485, + "step": 8100 + }, + { + "epoch": 3.7185354691075516, + "grad_norm": 3.3016109466552734, + "learning_rate": 5.275e-06, + "loss": 0.0471, + "step": 8125 + }, + { + "epoch": 3.7299771167048057, + "grad_norm": 4.43864107131958, + "learning_rate": 5.2671875e-06, + "loss": 0.0499, + "step": 8150 + }, + { + "epoch": 3.7414187643020593, + "grad_norm": 6.333988666534424, + "learning_rate": 5.259687500000001e-06, + "loss": 0.0554, + "step": 8175 + }, + { + "epoch": 3.7528604118993134, + "grad_norm": 11.240910530090332, + "learning_rate": 5.2518750000000004e-06, + "loss": 0.034, + "step": 8200 + }, + { + "epoch": 3.7643020594965675, + "grad_norm": 3.428675889968872, + "learning_rate": 5.2440625e-06, + "loss": 0.0474, + "step": 8225 + }, + { + "epoch": 3.7757437070938216, + "grad_norm": 2.0469939708709717, + "learning_rate": 5.23625e-06, + "loss": 0.0366, + "step": 8250 + }, + { + "epoch": 3.7871853546910756, + "grad_norm": 2.763183116912842, + "learning_rate": 5.228437500000001e-06, + "loss": 0.0451, + "step": 8275 + }, + { + "epoch": 3.7986270022883293, + "grad_norm": 5.365248203277588, + "learning_rate": 5.2206250000000005e-06, + "loss": 0.0409, + "step": 8300 + }, + { + "epoch": 3.8100686498855834, + "grad_norm": 3.84425687789917, + "learning_rate": 5.2128125e-06, + "loss": 0.0493, + "step": 8325 + }, + { + "epoch": 3.8215102974828374, + "grad_norm": 8.628647804260254, + "learning_rate": 5.205e-06, + "loss": 0.0468, + "step": 8350 + }, + { + "epoch": 3.8329519450800915, + "grad_norm": 6.355470657348633, + "learning_rate": 5.1971875e-06, + "loss": 0.0532, + "step": 8375 + }, + { + "epoch": 3.8443935926773456, + "grad_norm": 5.792243003845215, + "learning_rate": 5.189375000000001e-06, + "loss": 0.0389, + "step": 8400 + }, + { + "epoch": 3.8558352402745997, + "grad_norm": 5.9728312492370605, + "learning_rate": 5.1815624999999996e-06, + "loss": 0.0495, + "step": 8425 + }, + { + "epoch": 3.8672768878718538, + "grad_norm": 3.019531488418579, + "learning_rate": 5.17375e-06, + "loss": 0.0474, + "step": 8450 + }, + { + "epoch": 3.8787185354691074, + "grad_norm": 4.252594947814941, + "learning_rate": 5.1659375e-06, + "loss": 0.0571, + "step": 8475 + }, + { + "epoch": 3.8901601830663615, + "grad_norm": 2.1574008464813232, + "learning_rate": 5.158125000000001e-06, + "loss": 0.0353, + "step": 8500 + }, + { + "epoch": 3.9016018306636155, + "grad_norm": 1.5579230785369873, + "learning_rate": 5.1503125e-06, + "loss": 0.04, + "step": 8525 + }, + { + "epoch": 3.9130434782608696, + "grad_norm": 9.67751693725586, + "learning_rate": 5.1425e-06, + "loss": 0.0446, + "step": 8550 + }, + { + "epoch": 3.9244851258581237, + "grad_norm": 3.2914412021636963, + "learning_rate": 5.1346875e-06, + "loss": 0.0491, + "step": 8575 + }, + { + "epoch": 3.9359267734553773, + "grad_norm": 5.403482437133789, + "learning_rate": 5.126875000000001e-06, + "loss": 0.0501, + "step": 8600 + }, + { + "epoch": 3.9473684210526314, + "grad_norm": 4.028888702392578, + "learning_rate": 5.1190625e-06, + "loss": 0.0554, + "step": 8625 + }, + { + "epoch": 3.9588100686498855, + "grad_norm": 4.119118690490723, + "learning_rate": 5.11125e-06, + "loss": 0.0411, + "step": 8650 + }, + { + "epoch": 3.9702517162471396, + "grad_norm": 3.5604629516601562, + "learning_rate": 5.1034375e-06, + "loss": 0.0595, + "step": 8675 + }, + { + "epoch": 3.9816933638443937, + "grad_norm": 4.4073686599731445, + "learning_rate": 5.095625e-06, + "loss": 0.0362, + "step": 8700 + }, + { + "epoch": 3.9931350114416477, + "grad_norm": 3.6186680793762207, + "learning_rate": 5.087812500000001e-06, + "loss": 0.05, + "step": 8725 + }, + { + "epoch": 4.004576659038902, + "grad_norm": 2.3267414569854736, + "learning_rate": 5.08e-06, + "loss": 0.0403, + "step": 8750 + }, + { + "epoch": 4.016018306636155, + "grad_norm": 3.6123080253601074, + "learning_rate": 5.0721875e-06, + "loss": 0.0205, + "step": 8775 + }, + { + "epoch": 4.02745995423341, + "grad_norm": 4.297911643981934, + "learning_rate": 5.064375e-06, + "loss": 0.0269, + "step": 8800 + }, + { + "epoch": 4.038901601830664, + "grad_norm": 3.034499168395996, + "learning_rate": 5.056562500000001e-06, + "loss": 0.0195, + "step": 8825 + }, + { + "epoch": 4.050343249427917, + "grad_norm": 2.369044780731201, + "learning_rate": 5.04875e-06, + "loss": 0.0293, + "step": 8850 + }, + { + "epoch": 4.061784897025172, + "grad_norm": 2.833590269088745, + "learning_rate": 5.0409375e-06, + "loss": 0.0187, + "step": 8875 + }, + { + "epoch": 4.073226544622425, + "grad_norm": 2.4066593647003174, + "learning_rate": 5.033125e-06, + "loss": 0.0298, + "step": 8900 + }, + { + "epoch": 4.08466819221968, + "grad_norm": 2.521658420562744, + "learning_rate": 5.025312500000001e-06, + "loss": 0.0219, + "step": 8925 + }, + { + "epoch": 4.0961098398169336, + "grad_norm": 2.493262529373169, + "learning_rate": 5.0175e-06, + "loss": 0.0319, + "step": 8950 + }, + { + "epoch": 4.107551487414188, + "grad_norm": 4.92982816696167, + "learning_rate": 5.0096875000000005e-06, + "loss": 0.0229, + "step": 8975 + }, + { + "epoch": 4.118993135011442, + "grad_norm": 5.053268909454346, + "learning_rate": 5.001875e-06, + "loss": 0.0348, + "step": 9000 + }, + { + "epoch": 4.130434782608695, + "grad_norm": 1.4297044277191162, + "learning_rate": 4.9940625e-06, + "loss": 0.0204, + "step": 9025 + }, + { + "epoch": 4.14187643020595, + "grad_norm": 1.9617934226989746, + "learning_rate": 4.98625e-06, + "loss": 0.0369, + "step": 9050 + }, + { + "epoch": 4.1533180778032035, + "grad_norm": 3.376234531402588, + "learning_rate": 4.9784375e-06, + "loss": 0.0286, + "step": 9075 + }, + { + "epoch": 4.164759725400458, + "grad_norm": 1.698594570159912, + "learning_rate": 4.970625e-06, + "loss": 0.035, + "step": 9100 + }, + { + "epoch": 4.176201372997712, + "grad_norm": 1.5813435316085815, + "learning_rate": 4.9628125e-06, + "loss": 0.0215, + "step": 9125 + }, + { + "epoch": 4.187643020594965, + "grad_norm": 1.6306517124176025, + "learning_rate": 4.955e-06, + "loss": 0.0401, + "step": 9150 + }, + { + "epoch": 4.19908466819222, + "grad_norm": 2.4657962322235107, + "learning_rate": 4.9471875e-06, + "loss": 0.0211, + "step": 9175 + }, + { + "epoch": 4.2105263157894735, + "grad_norm": 0.9647684693336487, + "learning_rate": 4.9393750000000005e-06, + "loss": 0.0292, + "step": 9200 + }, + { + "epoch": 4.221967963386728, + "grad_norm": 3.233245849609375, + "learning_rate": 4.9315625e-06, + "loss": 0.0218, + "step": 9225 + }, + { + "epoch": 4.233409610983982, + "grad_norm": 3.3073842525482178, + "learning_rate": 4.92375e-06, + "loss": 0.0301, + "step": 9250 + }, + { + "epoch": 4.244851258581235, + "grad_norm": 3.375771999359131, + "learning_rate": 4.9159375e-06, + "loss": 0.0243, + "step": 9275 + }, + { + "epoch": 4.25629290617849, + "grad_norm": 2.0122270584106445, + "learning_rate": 4.9081250000000005e-06, + "loss": 0.0274, + "step": 9300 + }, + { + "epoch": 4.267734553775743, + "grad_norm": 4.077823638916016, + "learning_rate": 4.9003125e-06, + "loss": 0.0186, + "step": 9325 + }, + { + "epoch": 4.279176201372998, + "grad_norm": 2.2521536350250244, + "learning_rate": 4.8925e-06, + "loss": 0.0305, + "step": 9350 + }, + { + "epoch": 4.290617848970252, + "grad_norm": 2.4649903774261475, + "learning_rate": 4.8846875e-06, + "loss": 0.0218, + "step": 9375 + }, + { + "epoch": 4.302059496567506, + "grad_norm": 2.196124792098999, + "learning_rate": 4.876875e-06, + "loss": 0.0342, + "step": 9400 + }, + { + "epoch": 4.31350114416476, + "grad_norm": 2.979475975036621, + "learning_rate": 4.8690625000000004e-06, + "loss": 0.0242, + "step": 9425 + }, + { + "epoch": 4.324942791762013, + "grad_norm": 7.853615760803223, + "learning_rate": 4.86125e-06, + "loss": 0.03, + "step": 9450 + }, + { + "epoch": 4.336384439359268, + "grad_norm": 1.174206256866455, + "learning_rate": 4.8534375e-06, + "loss": 0.0195, + "step": 9475 + }, + { + "epoch": 4.3478260869565215, + "grad_norm": 3.248457193374634, + "learning_rate": 4.845625e-06, + "loss": 0.0377, + "step": 9500 + }, + { + "epoch": 4.359267734553776, + "grad_norm": 1.723983645439148, + "learning_rate": 4.8378125000000005e-06, + "loss": 0.0266, + "step": 9525 + }, + { + "epoch": 4.37070938215103, + "grad_norm": 8.764972686767578, + "learning_rate": 4.83e-06, + "loss": 0.0431, + "step": 9550 + }, + { + "epoch": 4.382151029748284, + "grad_norm": 1.3390229940414429, + "learning_rate": 4.8221875e-06, + "loss": 0.0176, + "step": 9575 + }, + { + "epoch": 4.393592677345538, + "grad_norm": 4.078359603881836, + "learning_rate": 4.814375e-06, + "loss": 0.0338, + "step": 9600 + }, + { + "epoch": 4.4050343249427915, + "grad_norm": 2.31992244720459, + "learning_rate": 4.806562500000001e-06, + "loss": 0.0202, + "step": 9625 + }, + { + "epoch": 4.416475972540046, + "grad_norm": 4.521693229675293, + "learning_rate": 4.7987500000000004e-06, + "loss": 0.0302, + "step": 9650 + }, + { + "epoch": 4.4279176201373, + "grad_norm": 5.699862480163574, + "learning_rate": 4.7909375e-06, + "loss": 0.0262, + "step": 9675 + }, + { + "epoch": 4.439359267734554, + "grad_norm": 2.659482955932617, + "learning_rate": 4.783125e-06, + "loss": 0.0252, + "step": 9700 + }, + { + "epoch": 4.450800915331808, + "grad_norm": 9.114242553710938, + "learning_rate": 4.7753125e-06, + "loss": 0.0292, + "step": 9725 + }, + { + "epoch": 4.462242562929061, + "grad_norm": 2.6791884899139404, + "learning_rate": 4.7675000000000005e-06, + "loss": 0.0304, + "step": 9750 + }, + { + "epoch": 4.473684210526316, + "grad_norm": 3.7395987510681152, + "learning_rate": 4.7596875e-06, + "loss": 0.0266, + "step": 9775 + }, + { + "epoch": 4.48512585812357, + "grad_norm": 1.6912297010421753, + "learning_rate": 4.751875e-06, + "loss": 0.0446, + "step": 9800 + }, + { + "epoch": 4.496567505720824, + "grad_norm": 3.4774134159088135, + "learning_rate": 4.7440625e-06, + "loss": 0.0266, + "step": 9825 + }, + { + "epoch": 4.508009153318078, + "grad_norm": 3.602811098098755, + "learning_rate": 4.736250000000001e-06, + "loss": 0.0343, + "step": 9850 + }, + { + "epoch": 4.519450800915331, + "grad_norm": 3.7448768615722656, + "learning_rate": 4.7284374999999996e-06, + "loss": 0.0241, + "step": 9875 + }, + { + "epoch": 4.530892448512586, + "grad_norm": 2.9017863273620605, + "learning_rate": 4.720625e-06, + "loss": 0.0313, + "step": 9900 + }, + { + "epoch": 4.5423340961098395, + "grad_norm": 2.8648183345794678, + "learning_rate": 4.7128125e-06, + "loss": 0.0224, + "step": 9925 + }, + { + "epoch": 4.553775743707094, + "grad_norm": 2.749983549118042, + "learning_rate": 4.705000000000001e-06, + "loss": 0.0326, + "step": 9950 + }, + { + "epoch": 4.565217391304348, + "grad_norm": 0.6364901661872864, + "learning_rate": 4.6971875000000005e-06, + "loss": 0.0192, + "step": 9975 + }, + { + "epoch": 4.576659038901602, + "grad_norm": 0.9560081362724304, + "learning_rate": 4.689375e-06, + "loss": 0.0272, + "step": 10000 + }, + { + "epoch": 4.576659038901602, + "eval_loss": 0.1832965463399887, + "eval_runtime": 5483.9142, + "eval_samples_per_second": 1.736, + "eval_steps_per_second": 0.217, + "eval_wer": 0.10381071222883471, + "step": 10000 + }, + { + "epoch": 4.588100686498856, + "grad_norm": 4.101663589477539, + "learning_rate": 4.6815625e-06, + "loss": 0.0217, + "step": 10025 + }, + { + "epoch": 4.5995423340961095, + "grad_norm": 2.0789246559143066, + "learning_rate": 4.67375e-06, + "loss": 0.0289, + "step": 10050 + }, + { + "epoch": 4.610983981693364, + "grad_norm": 4.012986660003662, + "learning_rate": 4.665937500000001e-06, + "loss": 0.0272, + "step": 10075 + }, + { + "epoch": 4.622425629290618, + "grad_norm": 2.1427500247955322, + "learning_rate": 4.658125e-06, + "loss": 0.0299, + "step": 10100 + }, + { + "epoch": 4.633867276887872, + "grad_norm": 6.746981620788574, + "learning_rate": 4.6503125e-06, + "loss": 0.0222, + "step": 10125 + }, + { + "epoch": 4.645308924485126, + "grad_norm": 0.5219627022743225, + "learning_rate": 4.6425e-06, + "loss": 0.0311, + "step": 10150 + }, + { + "epoch": 4.65675057208238, + "grad_norm": 2.274179697036743, + "learning_rate": 4.634687500000001e-06, + "loss": 0.0189, + "step": 10175 + }, + { + "epoch": 4.668192219679634, + "grad_norm": 1.2450759410858154, + "learning_rate": 4.6271875e-06, + "loss": 0.0319, + "step": 10200 + }, + { + "epoch": 4.679633867276888, + "grad_norm": 7.748188018798828, + "learning_rate": 4.619375e-06, + "loss": 0.0249, + "step": 10225 + }, + { + "epoch": 4.691075514874142, + "grad_norm": 2.7961182594299316, + "learning_rate": 4.6115625e-06, + "loss": 0.0367, + "step": 10250 + }, + { + "epoch": 4.702517162471396, + "grad_norm": 1.6634931564331055, + "learning_rate": 4.603750000000001e-06, + "loss": 0.0193, + "step": 10275 + }, + { + "epoch": 4.71395881006865, + "grad_norm": 3.7419142723083496, + "learning_rate": 4.5959374999999996e-06, + "loss": 0.0346, + "step": 10300 + }, + { + "epoch": 4.725400457665904, + "grad_norm": 11.037392616271973, + "learning_rate": 4.588125e-06, + "loss": 0.0251, + "step": 10325 + }, + { + "epoch": 4.7368421052631575, + "grad_norm": 2.88437819480896, + "learning_rate": 4.5803125e-06, + "loss": 0.0386, + "step": 10350 + }, + { + "epoch": 4.748283752860412, + "grad_norm": 3.546595335006714, + "learning_rate": 4.572500000000001e-06, + "loss": 0.0184, + "step": 10375 + }, + { + "epoch": 4.759725400457666, + "grad_norm": 1.2331124544143677, + "learning_rate": 4.5646875000000005e-06, + "loss": 0.0322, + "step": 10400 + }, + { + "epoch": 4.77116704805492, + "grad_norm": 1.0554383993148804, + "learning_rate": 4.556875e-06, + "loss": 0.02, + "step": 10425 + }, + { + "epoch": 4.782608695652174, + "grad_norm": 0.897574245929718, + "learning_rate": 4.5490625e-06, + "loss": 0.028, + "step": 10450 + }, + { + "epoch": 4.7940503432494275, + "grad_norm": 3.45660138130188, + "learning_rate": 4.541250000000001e-06, + "loss": 0.0196, + "step": 10475 + }, + { + "epoch": 4.805491990846682, + "grad_norm": 4.324442386627197, + "learning_rate": 4.533437500000001e-06, + "loss": 0.0329, + "step": 10500 + }, + { + "epoch": 4.816933638443936, + "grad_norm": 7.328847885131836, + "learning_rate": 4.525625e-06, + "loss": 0.0286, + "step": 10525 + }, + { + "epoch": 4.82837528604119, + "grad_norm": 0.6618658900260925, + "learning_rate": 4.5178125e-06, + "loss": 0.0257, + "step": 10550 + }, + { + "epoch": 4.839816933638444, + "grad_norm": 3.4053359031677246, + "learning_rate": 4.51e-06, + "loss": 0.0188, + "step": 10575 + }, + { + "epoch": 4.851258581235698, + "grad_norm": 1.5626916885375977, + "learning_rate": 4.502187500000001e-06, + "loss": 0.0388, + "step": 10600 + }, + { + "epoch": 4.862700228832952, + "grad_norm": 2.1022493839263916, + "learning_rate": 4.494375e-06, + "loss": 0.0228, + "step": 10625 + }, + { + "epoch": 4.874141876430206, + "grad_norm": 1.5179095268249512, + "learning_rate": 4.4865625e-06, + "loss": 0.0446, + "step": 10650 + }, + { + "epoch": 4.88558352402746, + "grad_norm": 3.457355499267578, + "learning_rate": 4.47875e-06, + "loss": 0.0213, + "step": 10675 + }, + { + "epoch": 4.897025171624714, + "grad_norm": 1.2291810512542725, + "learning_rate": 4.470937500000001e-06, + "loss": 0.0308, + "step": 10700 + }, + { + "epoch": 4.908466819221968, + "grad_norm": 4.664734363555908, + "learning_rate": 4.463125e-06, + "loss": 0.0187, + "step": 10725 + }, + { + "epoch": 4.919908466819222, + "grad_norm": 2.5962226390838623, + "learning_rate": 4.4553125e-06, + "loss": 0.0348, + "step": 10750 + }, + { + "epoch": 4.931350114416476, + "grad_norm": 1.759839415550232, + "learning_rate": 4.4475e-06, + "loss": 0.0223, + "step": 10775 + }, + { + "epoch": 4.94279176201373, + "grad_norm": 3.38629412651062, + "learning_rate": 4.439687500000001e-06, + "loss": 0.0439, + "step": 10800 + }, + { + "epoch": 4.954233409610984, + "grad_norm": 0.6326417326927185, + "learning_rate": 4.431875e-06, + "loss": 0.0204, + "step": 10825 + }, + { + "epoch": 4.965675057208238, + "grad_norm": 3.3679585456848145, + "learning_rate": 4.4240625000000005e-06, + "loss": 0.0322, + "step": 10850 + }, + { + "epoch": 4.977116704805492, + "grad_norm": 3.8049206733703613, + "learning_rate": 4.41625e-06, + "loss": 0.0243, + "step": 10875 + }, + { + "epoch": 4.988558352402746, + "grad_norm": 2.581594944000244, + "learning_rate": 4.4084375e-06, + "loss": 0.0332, + "step": 10900 + }, + { + "epoch": 5.0, + "grad_norm": 3.7988698482513428, + "learning_rate": 4.400625e-06, + "loss": 0.0275, + "step": 10925 + }, + { + "epoch": 5.011441647597254, + "grad_norm": 2.600945234298706, + "learning_rate": 4.3928125e-06, + "loss": 0.0131, + "step": 10950 + }, + { + "epoch": 5.022883295194508, + "grad_norm": 8.007580757141113, + "learning_rate": 4.385e-06, + "loss": 0.0133, + "step": 10975 + }, + { + "epoch": 5.034324942791762, + "grad_norm": 3.332880735397339, + "learning_rate": 4.3771875e-06, + "loss": 0.0149, + "step": 11000 + }, + { + "epoch": 5.045766590389016, + "grad_norm": 15.387001037597656, + "learning_rate": 4.369375000000001e-06, + "loss": 0.0207, + "step": 11025 + }, + { + "epoch": 5.05720823798627, + "grad_norm": 4.535585403442383, + "learning_rate": 4.3615625e-06, + "loss": 0.0199, + "step": 11050 + }, + { + "epoch": 5.068649885583524, + "grad_norm": 4.385061740875244, + "learning_rate": 4.3537500000000005e-06, + "loss": 0.0248, + "step": 11075 + }, + { + "epoch": 5.080091533180778, + "grad_norm": 2.6416256427764893, + "learning_rate": 4.3459375e-06, + "loss": 0.0129, + "step": 11100 + }, + { + "epoch": 5.091533180778032, + "grad_norm": 2.3068976402282715, + "learning_rate": 4.338125000000001e-06, + "loss": 0.0271, + "step": 11125 + }, + { + "epoch": 5.102974828375286, + "grad_norm": 5.006783485412598, + "learning_rate": 4.3303125e-06, + "loss": 0.0156, + "step": 11150 + }, + { + "epoch": 5.11441647597254, + "grad_norm": 8.350238800048828, + "learning_rate": 4.3225000000000005e-06, + "loss": 0.0152, + "step": 11175 + }, + { + "epoch": 5.125858123569794, + "grad_norm": 3.0252091884613037, + "learning_rate": 4.3146875e-06, + "loss": 0.0142, + "step": 11200 + }, + { + "epoch": 5.137299771167048, + "grad_norm": 17.71987533569336, + "learning_rate": 4.306875e-06, + "loss": 0.0261, + "step": 11225 + }, + { + "epoch": 5.148741418764302, + "grad_norm": 0.3432025611400604, + "learning_rate": 4.2990625e-06, + "loss": 0.0141, + "step": 11250 + }, + { + "epoch": 5.160183066361556, + "grad_norm": 3.693594455718994, + "learning_rate": 4.29125e-06, + "loss": 0.0284, + "step": 11275 + }, + { + "epoch": 5.17162471395881, + "grad_norm": 2.1907877922058105, + "learning_rate": 4.2834375000000004e-06, + "loss": 0.0103, + "step": 11300 + }, + { + "epoch": 5.183066361556064, + "grad_norm": 5.526587963104248, + "learning_rate": 4.275625e-06, + "loss": 0.0254, + "step": 11325 + }, + { + "epoch": 5.194508009153318, + "grad_norm": 5.500503063201904, + "learning_rate": 4.2678125e-06, + "loss": 0.0108, + "step": 11350 + }, + { + "epoch": 5.2059496567505725, + "grad_norm": 7.589873790740967, + "learning_rate": 4.26e-06, + "loss": 0.0237, + "step": 11375 + }, + { + "epoch": 5.217391304347826, + "grad_norm": 2.3860023021698, + "learning_rate": 4.2521875000000005e-06, + "loss": 0.0186, + "step": 11400 + }, + { + "epoch": 5.22883295194508, + "grad_norm": 4.4563307762146, + "learning_rate": 4.244375e-06, + "loss": 0.0217, + "step": 11425 + }, + { + "epoch": 5.240274599542334, + "grad_norm": 3.0135042667388916, + "learning_rate": 4.2365625e-06, + "loss": 0.0155, + "step": 11450 + }, + { + "epoch": 5.251716247139588, + "grad_norm": 1.9074753522872925, + "learning_rate": 4.22875e-06, + "loss": 0.0251, + "step": 11475 + }, + { + "epoch": 5.2631578947368425, + "grad_norm": 3.4067037105560303, + "learning_rate": 4.220937500000001e-06, + "loss": 0.0119, + "step": 11500 + }, + { + "epoch": 5.274599542334096, + "grad_norm": 2.8989033699035645, + "learning_rate": 4.2131250000000004e-06, + "loss": 0.0188, + "step": 11525 + }, + { + "epoch": 5.28604118993135, + "grad_norm": 2.8292016983032227, + "learning_rate": 4.2053125e-06, + "loss": 0.0159, + "step": 11550 + }, + { + "epoch": 5.297482837528604, + "grad_norm": 6.212322235107422, + "learning_rate": 4.1975e-06, + "loss": 0.0233, + "step": 11575 + }, + { + "epoch": 5.308924485125858, + "grad_norm": 0.9031001925468445, + "learning_rate": 4.1896875e-06, + "loss": 0.0146, + "step": 11600 + }, + { + "epoch": 5.320366132723112, + "grad_norm": 7.238254070281982, + "learning_rate": 4.1818750000000005e-06, + "loss": 0.0158, + "step": 11625 + }, + { + "epoch": 5.331807780320366, + "grad_norm": 2.4615652561187744, + "learning_rate": 4.1740625e-06, + "loss": 0.0122, + "step": 11650 + }, + { + "epoch": 5.34324942791762, + "grad_norm": 3.7764835357666016, + "learning_rate": 4.16625e-06, + "loss": 0.0221, + "step": 11675 + }, + { + "epoch": 5.354691075514874, + "grad_norm": 1.053728699684143, + "learning_rate": 4.1584375e-06, + "loss": 0.0186, + "step": 11700 + }, + { + "epoch": 5.366132723112128, + "grad_norm": 7.169757843017578, + "learning_rate": 4.150625000000001e-06, + "loss": 0.0198, + "step": 11725 + }, + { + "epoch": 5.377574370709382, + "grad_norm": 0.351454496383667, + "learning_rate": 4.1428125e-06, + "loss": 0.0121, + "step": 11750 + }, + { + "epoch": 5.389016018306636, + "grad_norm": 4.189688682556152, + "learning_rate": 4.135e-06, + "loss": 0.0245, + "step": 11775 + }, + { + "epoch": 5.4004576659038905, + "grad_norm": 3.2234408855438232, + "learning_rate": 4.1271875e-06, + "loss": 0.0149, + "step": 11800 + }, + { + "epoch": 5.411899313501144, + "grad_norm": 7.368442058563232, + "learning_rate": 4.119375000000001e-06, + "loss": 0.0185, + "step": 11825 + }, + { + "epoch": 5.423340961098398, + "grad_norm": 4.849565505981445, + "learning_rate": 4.1115625000000005e-06, + "loss": 0.0155, + "step": 11850 + }, + { + "epoch": 5.434782608695652, + "grad_norm": 8.018900871276855, + "learning_rate": 4.10375e-06, + "loss": 0.0272, + "step": 11875 + }, + { + "epoch": 5.446224256292906, + "grad_norm": 3.36999773979187, + "learning_rate": 4.0959375e-06, + "loss": 0.0168, + "step": 11900 + }, + { + "epoch": 5.4576659038901605, + "grad_norm": 8.504731178283691, + "learning_rate": 4.088125e-06, + "loss": 0.0187, + "step": 11925 + }, + { + "epoch": 5.469107551487414, + "grad_norm": 1.719467282295227, + "learning_rate": 4.080312500000001e-06, + "loss": 0.0213, + "step": 11950 + }, + { + "epoch": 5.480549199084669, + "grad_norm": 7.953880310058594, + "learning_rate": 4.0724999999999995e-06, + "loss": 0.0209, + "step": 11975 + }, + { + "epoch": 5.491990846681922, + "grad_norm": 3.389706611633301, + "learning_rate": 4.0646875e-06, + "loss": 0.0151, + "step": 12000 + }, + { + "epoch": 5.503432494279176, + "grad_norm": 6.694132328033447, + "learning_rate": 4.056875e-06, + "loss": 0.0202, + "step": 12025 + }, + { + "epoch": 5.51487414187643, + "grad_norm": 2.6016061305999756, + "learning_rate": 4.049062500000001e-06, + "loss": 0.0172, + "step": 12050 + }, + { + "epoch": 5.526315789473684, + "grad_norm": 5.176774024963379, + "learning_rate": 4.04125e-06, + "loss": 0.0281, + "step": 12075 + }, + { + "epoch": 5.537757437070939, + "grad_norm": 1.5615991353988647, + "learning_rate": 4.0334375e-06, + "loss": 0.0128, + "step": 12100 + }, + { + "epoch": 5.549199084668192, + "grad_norm": 4.4115376472473145, + "learning_rate": 4.025625e-06, + "loss": 0.0278, + "step": 12125 + }, + { + "epoch": 5.560640732265446, + "grad_norm": 0.18079940974712372, + "learning_rate": 4.017812500000001e-06, + "loss": 0.0129, + "step": 12150 + }, + { + "epoch": 5.5720823798627, + "grad_norm": 1.855873942375183, + "learning_rate": 4.01e-06, + "loss": 0.024, + "step": 12175 + }, + { + "epoch": 5.583524027459954, + "grad_norm": 3.256206512451172, + "learning_rate": 4.0021875e-06, + "loss": 0.0152, + "step": 12200 + }, + { + "epoch": 5.5949656750572085, + "grad_norm": 6.555622100830078, + "learning_rate": 3.994375e-06, + "loss": 0.0261, + "step": 12225 + }, + { + "epoch": 5.606407322654462, + "grad_norm": 3.024768352508545, + "learning_rate": 3.9865625e-06, + "loss": 0.0107, + "step": 12250 + }, + { + "epoch": 5.617848970251716, + "grad_norm": 2.868229389190674, + "learning_rate": 3.978750000000001e-06, + "loss": 0.0255, + "step": 12275 + }, + { + "epoch": 5.62929061784897, + "grad_norm": 5.408130168914795, + "learning_rate": 3.9709375e-06, + "loss": 0.0121, + "step": 12300 + }, + { + "epoch": 5.640732265446224, + "grad_norm": 8.534693717956543, + "learning_rate": 3.963125e-06, + "loss": 0.0245, + "step": 12325 + }, + { + "epoch": 5.6521739130434785, + "grad_norm": 0.8499470353126526, + "learning_rate": 3.9553125e-06, + "loss": 0.0114, + "step": 12350 + }, + { + "epoch": 5.663615560640732, + "grad_norm": 1.4523133039474487, + "learning_rate": 3.947500000000001e-06, + "loss": 0.0251, + "step": 12375 + }, + { + "epoch": 5.675057208237987, + "grad_norm": 1.7775627374649048, + "learning_rate": 3.9396875e-06, + "loss": 0.0191, + "step": 12400 + }, + { + "epoch": 5.68649885583524, + "grad_norm": 4.454090118408203, + "learning_rate": 3.931875e-06, + "loss": 0.0195, + "step": 12425 + }, + { + "epoch": 5.697940503432494, + "grad_norm": 1.4271085262298584, + "learning_rate": 3.9240625e-06, + "loss": 0.0138, + "step": 12450 + }, + { + "epoch": 5.709382151029748, + "grad_norm": 9.393786430358887, + "learning_rate": 3.916250000000001e-06, + "loss": 0.0273, + "step": 12475 + }, + { + "epoch": 5.720823798627002, + "grad_norm": 3.8522861003875732, + "learning_rate": 3.9084375e-06, + "loss": 0.0197, + "step": 12500 + }, + { + "epoch": 5.732265446224257, + "grad_norm": 5.545007705688477, + "learning_rate": 3.9006250000000005e-06, + "loss": 0.0206, + "step": 12525 + }, + { + "epoch": 5.74370709382151, + "grad_norm": 2.3138723373413086, + "learning_rate": 3.8928125e-06, + "loss": 0.0135, + "step": 12550 + }, + { + "epoch": 5.755148741418765, + "grad_norm": 6.381573677062988, + "learning_rate": 3.885e-06, + "loss": 0.0264, + "step": 12575 + }, + { + "epoch": 5.766590389016018, + "grad_norm": 2.9235124588012695, + "learning_rate": 3.8771875e-06, + "loss": 0.014, + "step": 12600 + }, + { + "epoch": 5.778032036613272, + "grad_norm": 9.768881797790527, + "learning_rate": 3.869375e-06, + "loss": 0.0237, + "step": 12625 + }, + { + "epoch": 5.7894736842105265, + "grad_norm": 2.5591719150543213, + "learning_rate": 3.8615625e-06, + "loss": 0.0129, + "step": 12650 + }, + { + "epoch": 5.80091533180778, + "grad_norm": 8.308956146240234, + "learning_rate": 3.85375e-06, + "loss": 0.0174, + "step": 12675 + }, + { + "epoch": 5.812356979405035, + "grad_norm": 1.5682135820388794, + "learning_rate": 3.8459375e-06, + "loss": 0.0134, + "step": 12700 + }, + { + "epoch": 5.823798627002288, + "grad_norm": 5.886026859283447, + "learning_rate": 3.838125e-06, + "loss": 0.0221, + "step": 12725 + }, + { + "epoch": 5.835240274599542, + "grad_norm": 5.910558223724365, + "learning_rate": 3.8303125000000004e-06, + "loss": 0.0163, + "step": 12750 + }, + { + "epoch": 5.8466819221967965, + "grad_norm": 6.313669204711914, + "learning_rate": 3.8225e-06, + "loss": 0.0222, + "step": 12775 + }, + { + "epoch": 5.85812356979405, + "grad_norm": 4.852139949798584, + "learning_rate": 3.8146875e-06, + "loss": 0.013, + "step": 12800 + }, + { + "epoch": 5.869565217391305, + "grad_norm": 6.2581963539123535, + "learning_rate": 3.806875e-06, + "loss": 0.0169, + "step": 12825 + }, + { + "epoch": 5.881006864988558, + "grad_norm": 5.26676607131958, + "learning_rate": 3.7990625e-06, + "loss": 0.0159, + "step": 12850 + }, + { + "epoch": 5.892448512585812, + "grad_norm": 0.4348423182964325, + "learning_rate": 3.7912500000000003e-06, + "loss": 0.0276, + "step": 12875 + }, + { + "epoch": 5.9038901601830664, + "grad_norm": 0.7607429623603821, + "learning_rate": 3.7834375000000006e-06, + "loss": 0.017, + "step": 12900 + }, + { + "epoch": 5.91533180778032, + "grad_norm": 4.235646724700928, + "learning_rate": 3.775625e-06, + "loss": 0.0277, + "step": 12925 + }, + { + "epoch": 5.926773455377575, + "grad_norm": 0.33068224787712097, + "learning_rate": 3.7678125e-06, + "loss": 0.0177, + "step": 12950 + }, + { + "epoch": 5.938215102974828, + "grad_norm": 8.94812297821045, + "learning_rate": 3.7600000000000004e-06, + "loss": 0.0189, + "step": 12975 + }, + { + "epoch": 5.949656750572083, + "grad_norm": 3.01238751411438, + "learning_rate": 3.7521875000000007e-06, + "loss": 0.0135, + "step": 13000 + }, + { + "epoch": 5.961098398169336, + "grad_norm": 6.622538089752197, + "learning_rate": 3.744375e-06, + "loss": 0.021, + "step": 13025 + }, + { + "epoch": 5.97254004576659, + "grad_norm": 5.920276641845703, + "learning_rate": 3.7365625000000003e-06, + "loss": 0.0208, + "step": 13050 + }, + { + "epoch": 5.983981693363845, + "grad_norm": 1.632116436958313, + "learning_rate": 3.7287500000000005e-06, + "loss": 0.0206, + "step": 13075 + }, + { + "epoch": 5.995423340961098, + "grad_norm": 6.903181076049805, + "learning_rate": 3.7209375000000003e-06, + "loss": 0.0155, + "step": 13100 + }, + { + "epoch": 6.006864988558353, + "grad_norm": 0.6914149522781372, + "learning_rate": 3.7134375e-06, + "loss": 0.0194, + "step": 13125 + }, + { + "epoch": 6.018306636155606, + "grad_norm": 6.280193328857422, + "learning_rate": 3.705625e-06, + "loss": 0.0168, + "step": 13150 + }, + { + "epoch": 6.02974828375286, + "grad_norm": 0.3250044286251068, + "learning_rate": 3.6978125000000004e-06, + "loss": 0.0273, + "step": 13175 + }, + { + "epoch": 6.0411899313501145, + "grad_norm": 2.4546151161193848, + "learning_rate": 3.6900000000000002e-06, + "loss": 0.0144, + "step": 13200 + }, + { + "epoch": 6.052631578947368, + "grad_norm": 1.1390433311462402, + "learning_rate": 3.6821875e-06, + "loss": 0.0184, + "step": 13225 + }, + { + "epoch": 6.064073226544623, + "grad_norm": 1.7204899787902832, + "learning_rate": 3.674375e-06, + "loss": 0.0092, + "step": 13250 + }, + { + "epoch": 6.075514874141876, + "grad_norm": 1.4025479555130005, + "learning_rate": 3.6665625e-06, + "loss": 0.024, + "step": 13275 + }, + { + "epoch": 6.086956521739131, + "grad_norm": 0.4468977153301239, + "learning_rate": 3.6587500000000003e-06, + "loss": 0.0139, + "step": 13300 + }, + { + "epoch": 6.0983981693363845, + "grad_norm": 0.18809981644153595, + "learning_rate": 3.6509374999999997e-06, + "loss": 0.011, + "step": 13325 + }, + { + "epoch": 6.109839816933638, + "grad_norm": 3.639801502227783, + "learning_rate": 3.643125e-06, + "loss": 0.0116, + "step": 13350 + }, + { + "epoch": 6.121281464530893, + "grad_norm": 0.617262601852417, + "learning_rate": 3.6353125e-06, + "loss": 0.0127, + "step": 13375 + }, + { + "epoch": 6.132723112128146, + "grad_norm": 3.9678919315338135, + "learning_rate": 3.6275000000000004e-06, + "loss": 0.0131, + "step": 13400 + }, + { + "epoch": 6.144164759725401, + "grad_norm": 1.4072645902633667, + "learning_rate": 3.6196875000000007e-06, + "loss": 0.0144, + "step": 13425 + }, + { + "epoch": 6.155606407322654, + "grad_norm": 6.396816253662109, + "learning_rate": 3.611875e-06, + "loss": 0.0094, + "step": 13450 + }, + { + "epoch": 6.167048054919908, + "grad_norm": 2.001157760620117, + "learning_rate": 3.6040625000000003e-06, + "loss": 0.0189, + "step": 13475 + }, + { + "epoch": 6.178489702517163, + "grad_norm": 4.379182815551758, + "learning_rate": 3.5962500000000005e-06, + "loss": 0.0135, + "step": 13500 + }, + { + "epoch": 6.189931350114416, + "grad_norm": 1.7968692779541016, + "learning_rate": 3.5884375000000003e-06, + "loss": 0.0197, + "step": 13525 + }, + { + "epoch": 6.201372997711671, + "grad_norm": 0.30796533823013306, + "learning_rate": 3.580625e-06, + "loss": 0.0091, + "step": 13550 + }, + { + "epoch": 6.212814645308924, + "grad_norm": 0.5206483006477356, + "learning_rate": 3.5728125e-06, + "loss": 0.0193, + "step": 13575 + }, + { + "epoch": 6.224256292906179, + "grad_norm": 0.16793321073055267, + "learning_rate": 3.565e-06, + "loss": 0.011, + "step": 13600 + }, + { + "epoch": 6.2356979405034325, + "grad_norm": 3.4714410305023193, + "learning_rate": 3.5571875000000004e-06, + "loss": 0.0163, + "step": 13625 + }, + { + "epoch": 6.247139588100686, + "grad_norm": 2.5397183895111084, + "learning_rate": 3.549375e-06, + "loss": 0.0098, + "step": 13650 + }, + { + "epoch": 6.258581235697941, + "grad_norm": 0.8021149039268494, + "learning_rate": 3.5415625e-06, + "loss": 0.0261, + "step": 13675 + }, + { + "epoch": 6.270022883295194, + "grad_norm": 2.6485466957092285, + "learning_rate": 3.5337500000000003e-06, + "loss": 0.0116, + "step": 13700 + }, + { + "epoch": 6.281464530892449, + "grad_norm": 0.5701060891151428, + "learning_rate": 3.5259375000000005e-06, + "loss": 0.0144, + "step": 13725 + }, + { + "epoch": 6.2929061784897025, + "grad_norm": 0.46138399839401245, + "learning_rate": 3.518125e-06, + "loss": 0.0074, + "step": 13750 + }, + { + "epoch": 6.304347826086957, + "grad_norm": 2.842442512512207, + "learning_rate": 3.5103125e-06, + "loss": 0.0194, + "step": 13775 + }, + { + "epoch": 6.315789473684211, + "grad_norm": 0.676575243473053, + "learning_rate": 3.5025000000000003e-06, + "loss": 0.0085, + "step": 13800 + }, + { + "epoch": 6.327231121281464, + "grad_norm": 1.1983352899551392, + "learning_rate": 3.4946875000000006e-06, + "loss": 0.0118, + "step": 13825 + }, + { + "epoch": 6.338672768878719, + "grad_norm": 3.670888900756836, + "learning_rate": 3.486875e-06, + "loss": 0.0126, + "step": 13850 + }, + { + "epoch": 6.350114416475972, + "grad_norm": 0.25366395711898804, + "learning_rate": 3.4790625e-06, + "loss": 0.0154, + "step": 13875 + }, + { + "epoch": 6.361556064073227, + "grad_norm": 3.8486428260803223, + "learning_rate": 3.47125e-06, + "loss": 0.0117, + "step": 13900 + }, + { + "epoch": 6.372997711670481, + "grad_norm": 1.135246753692627, + "learning_rate": 3.4634375000000002e-06, + "loss": 0.0156, + "step": 13925 + }, + { + "epoch": 6.384439359267734, + "grad_norm": 1.3957738876342773, + "learning_rate": 3.4556249999999996e-06, + "loss": 0.0191, + "step": 13950 + }, + { + "epoch": 6.395881006864989, + "grad_norm": 0.1521887332201004, + "learning_rate": 3.4478125e-06, + "loss": 0.0161, + "step": 13975 + }, + { + "epoch": 6.407322654462242, + "grad_norm": 0.3171142041683197, + "learning_rate": 3.44e-06, + "loss": 0.0106, + "step": 14000 + }, + { + "epoch": 6.418764302059497, + "grad_norm": 1.6162185668945312, + "learning_rate": 3.4321875000000003e-06, + "loss": 0.016, + "step": 14025 + }, + { + "epoch": 6.4302059496567505, + "grad_norm": 3.2044668197631836, + "learning_rate": 3.4243750000000006e-06, + "loss": 0.0088, + "step": 14050 + }, + { + "epoch": 6.441647597254004, + "grad_norm": 16.820634841918945, + "learning_rate": 3.4165625e-06, + "loss": 0.0094, + "step": 14075 + }, + { + "epoch": 6.453089244851259, + "grad_norm": 2.0577335357666016, + "learning_rate": 3.40875e-06, + "loss": 0.0113, + "step": 14100 + }, + { + "epoch": 6.464530892448512, + "grad_norm": 2.2943193912506104, + "learning_rate": 3.4009375000000004e-06, + "loss": 0.0136, + "step": 14125 + }, + { + "epoch": 6.475972540045767, + "grad_norm": 3.4963417053222656, + "learning_rate": 3.3931250000000007e-06, + "loss": 0.0138, + "step": 14150 + }, + { + "epoch": 6.4874141876430205, + "grad_norm": 0.4365079402923584, + "learning_rate": 3.3853125e-06, + "loss": 0.0131, + "step": 14175 + }, + { + "epoch": 6.498855835240275, + "grad_norm": 3.457284450531006, + "learning_rate": 3.3775000000000003e-06, + "loss": 0.0115, + "step": 14200 + }, + { + "epoch": 6.510297482837529, + "grad_norm": 0.40817612409591675, + "learning_rate": 3.3696875e-06, + "loss": 0.0181, + "step": 14225 + }, + { + "epoch": 6.521739130434782, + "grad_norm": 1.6476938724517822, + "learning_rate": 3.3618750000000003e-06, + "loss": 0.0097, + "step": 14250 + }, + { + "epoch": 6.533180778032037, + "grad_norm": 0.25183960795402527, + "learning_rate": 3.3540624999999997e-06, + "loss": 0.0217, + "step": 14275 + }, + { + "epoch": 6.54462242562929, + "grad_norm": 1.9451416730880737, + "learning_rate": 3.34625e-06, + "loss": 0.0097, + "step": 14300 + }, + { + "epoch": 6.556064073226545, + "grad_norm": 4.131077289581299, + "learning_rate": 3.3384375e-06, + "loss": 0.0183, + "step": 14325 + }, + { + "epoch": 6.567505720823799, + "grad_norm": 0.5438878536224365, + "learning_rate": 3.3306250000000004e-06, + "loss": 0.0145, + "step": 14350 + }, + { + "epoch": 6.578947368421053, + "grad_norm": 0.5045881867408752, + "learning_rate": 3.3228125e-06, + "loss": 0.0172, + "step": 14375 + }, + { + "epoch": 6.590389016018307, + "grad_norm": 0.4585050344467163, + "learning_rate": 3.315e-06, + "loss": 0.0126, + "step": 14400 + }, + { + "epoch": 6.60183066361556, + "grad_norm": 0.8127508759498596, + "learning_rate": 3.3071875000000003e-06, + "loss": 0.0151, + "step": 14425 + }, + { + "epoch": 6.613272311212815, + "grad_norm": 1.7050418853759766, + "learning_rate": 3.2993750000000005e-06, + "loss": 0.009, + "step": 14450 + }, + { + "epoch": 6.6247139588100685, + "grad_norm": 2.609612464904785, + "learning_rate": 3.2915625e-06, + "loss": 0.014, + "step": 14475 + }, + { + "epoch": 6.636155606407323, + "grad_norm": 0.10997484624385834, + "learning_rate": 3.28375e-06, + "loss": 0.0094, + "step": 14500 + }, + { + "epoch": 6.647597254004577, + "grad_norm": 5.7333173751831055, + "learning_rate": 3.2759375000000003e-06, + "loss": 0.0154, + "step": 14525 + }, + { + "epoch": 6.65903890160183, + "grad_norm": 2.564332962036133, + "learning_rate": 3.268125e-06, + "loss": 0.0155, + "step": 14550 + }, + { + "epoch": 6.670480549199085, + "grad_norm": 4.814024448394775, + "learning_rate": 3.2603125e-06, + "loss": 0.0179, + "step": 14575 + }, + { + "epoch": 6.6819221967963385, + "grad_norm": 1.1423102617263794, + "learning_rate": 3.2525e-06, + "loss": 0.0096, + "step": 14600 + }, + { + "epoch": 6.693363844393593, + "grad_norm": 3.1103646755218506, + "learning_rate": 3.2446875e-06, + "loss": 0.0164, + "step": 14625 + }, + { + "epoch": 6.704805491990847, + "grad_norm": 1.021026372909546, + "learning_rate": 3.2368750000000002e-06, + "loss": 0.0075, + "step": 14650 + }, + { + "epoch": 6.7162471395881, + "grad_norm": 0.4085843861103058, + "learning_rate": 3.2290625000000005e-06, + "loss": 0.0167, + "step": 14675 + }, + { + "epoch": 6.727688787185355, + "grad_norm": 0.8315253257751465, + "learning_rate": 3.22125e-06, + "loss": 0.0085, + "step": 14700 + }, + { + "epoch": 6.739130434782608, + "grad_norm": 2.2641472816467285, + "learning_rate": 3.2134375e-06, + "loss": 0.0209, + "step": 14725 + }, + { + "epoch": 6.750572082379863, + "grad_norm": 6.209875583648682, + "learning_rate": 3.2056250000000003e-06, + "loss": 0.0088, + "step": 14750 + }, + { + "epoch": 6.762013729977117, + "grad_norm": 0.5979451537132263, + "learning_rate": 3.1978125000000006e-06, + "loss": 0.0179, + "step": 14775 + }, + { + "epoch": 6.77345537757437, + "grad_norm": 2.0241873264312744, + "learning_rate": 3.19e-06, + "loss": 0.0103, + "step": 14800 + }, + { + "epoch": 6.784897025171625, + "grad_norm": 0.17280341684818268, + "learning_rate": 3.1821875e-06, + "loss": 0.0227, + "step": 14825 + }, + { + "epoch": 6.796338672768878, + "grad_norm": 0.12923689186573029, + "learning_rate": 3.1743750000000004e-06, + "loss": 0.0089, + "step": 14850 + }, + { + "epoch": 6.807780320366133, + "grad_norm": 0.43938687443733215, + "learning_rate": 3.1665625000000002e-06, + "loss": 0.0209, + "step": 14875 + }, + { + "epoch": 6.8192219679633865, + "grad_norm": 0.2618752419948578, + "learning_rate": 3.15875e-06, + "loss": 0.0109, + "step": 14900 + }, + { + "epoch": 6.830663615560641, + "grad_norm": 3.6829495429992676, + "learning_rate": 3.1509375000000003e-06, + "loss": 0.0188, + "step": 14925 + }, + { + "epoch": 6.842105263157895, + "grad_norm": 4.100037574768066, + "learning_rate": 3.143125e-06, + "loss": 0.0103, + "step": 14950 + }, + { + "epoch": 6.853546910755149, + "grad_norm": 3.131757974624634, + "learning_rate": 3.1353125000000003e-06, + "loss": 0.0199, + "step": 14975 + }, + { + "epoch": 6.864988558352403, + "grad_norm": 2.9150915145874023, + "learning_rate": 3.1274999999999997e-06, + "loss": 0.0125, + "step": 15000 + }, + { + "epoch": 6.864988558352403, + "eval_loss": 0.20364722609519958, + "eval_runtime": 5377.8653, + "eval_samples_per_second": 1.771, + "eval_steps_per_second": 0.221, + "eval_wer": 0.09913931016829845, + "step": 15000 + }, + { + "epoch": 6.8764302059496565, + "grad_norm": 3.365856885910034, + "learning_rate": 3.1196875e-06, + "loss": 0.0218, + "step": 15025 + }, + { + "epoch": 6.887871853546911, + "grad_norm": 0.5725418329238892, + "learning_rate": 3.111875e-06, + "loss": 0.0073, + "step": 15050 + }, + { + "epoch": 6.899313501144165, + "grad_norm": 1.2597202062606812, + "learning_rate": 3.1040625e-06, + "loss": 0.0155, + "step": 15075 + }, + { + "epoch": 6.910755148741419, + "grad_norm": 4.598627090454102, + "learning_rate": 3.0962500000000002e-06, + "loss": 0.0098, + "step": 15100 + }, + { + "epoch": 6.922196796338673, + "grad_norm": 0.7468834519386292, + "learning_rate": 3.0884375e-06, + "loss": 0.0193, + "step": 15125 + }, + { + "epoch": 6.933638443935926, + "grad_norm": 4.2396464347839355, + "learning_rate": 3.0806250000000003e-06, + "loss": 0.0102, + "step": 15150 + }, + { + "epoch": 6.945080091533181, + "grad_norm": 2.1572372913360596, + "learning_rate": 3.0728125e-06, + "loss": 0.02, + "step": 15175 + }, + { + "epoch": 6.956521739130435, + "grad_norm": 2.8561244010925293, + "learning_rate": 3.0650000000000003e-06, + "loss": 0.0096, + "step": 15200 + }, + { + "epoch": 6.967963386727689, + "grad_norm": 0.29279011487960815, + "learning_rate": 3.0571875e-06, + "loss": 0.0175, + "step": 15225 + }, + { + "epoch": 6.979405034324943, + "grad_norm": 1.120410442352295, + "learning_rate": 3.0493750000000003e-06, + "loss": 0.0117, + "step": 15250 + }, + { + "epoch": 6.990846681922196, + "grad_norm": 2.2122271060943604, + "learning_rate": 3.0415625e-06, + "loss": 0.0133, + "step": 15275 + }, + { + "epoch": 7.002288329519451, + "grad_norm": 1.584817886352539, + "learning_rate": 3.03375e-06, + "loss": 0.0088, + "step": 15300 + }, + { + "epoch": 7.0137299771167045, + "grad_norm": 0.8186086416244507, + "learning_rate": 3.0259375e-06, + "loss": 0.0064, + "step": 15325 + }, + { + "epoch": 7.025171624713959, + "grad_norm": 1.3982340097427368, + "learning_rate": 3.018125e-06, + "loss": 0.0173, + "step": 15350 + }, + { + "epoch": 7.036613272311213, + "grad_norm": 0.623369574546814, + "learning_rate": 3.0103125000000002e-06, + "loss": 0.0072, + "step": 15375 + }, + { + "epoch": 7.048054919908467, + "grad_norm": 0.9551053643226624, + "learning_rate": 3.0025e-06, + "loss": 0.0104, + "step": 15400 + }, + { + "epoch": 7.059496567505721, + "grad_norm": 0.4638427495956421, + "learning_rate": 2.9946875000000003e-06, + "loss": 0.0074, + "step": 15425 + }, + { + "epoch": 7.0709382151029745, + "grad_norm": 0.7113415598869324, + "learning_rate": 2.986875e-06, + "loss": 0.0121, + "step": 15450 + }, + { + "epoch": 7.082379862700229, + "grad_norm": 1.5616718530654907, + "learning_rate": 2.9790625000000003e-06, + "loss": 0.0084, + "step": 15475 + }, + { + "epoch": 7.093821510297483, + "grad_norm": 5.3068528175354, + "learning_rate": 2.97125e-06, + "loss": 0.0115, + "step": 15500 + }, + { + "epoch": 7.105263157894737, + "grad_norm": 6.045175075531006, + "learning_rate": 2.9634375000000004e-06, + "loss": 0.007, + "step": 15525 + }, + { + "epoch": 7.116704805491991, + "grad_norm": 0.12478996068239212, + "learning_rate": 2.955625e-06, + "loss": 0.0122, + "step": 15550 + }, + { + "epoch": 7.128146453089244, + "grad_norm": 0.3576812744140625, + "learning_rate": 2.9478125000000004e-06, + "loss": 0.007, + "step": 15575 + }, + { + "epoch": 7.139588100686499, + "grad_norm": 1.844315767288208, + "learning_rate": 2.9400000000000002e-06, + "loss": 0.0115, + "step": 15600 + }, + { + "epoch": 7.151029748283753, + "grad_norm": 9.679522514343262, + "learning_rate": 2.9321875e-06, + "loss": 0.0087, + "step": 15625 + }, + { + "epoch": 7.162471395881007, + "grad_norm": 0.7042823433876038, + "learning_rate": 2.924375e-06, + "loss": 0.0125, + "step": 15650 + }, + { + "epoch": 7.173913043478261, + "grad_norm": 0.31098079681396484, + "learning_rate": 2.9165625e-06, + "loss": 0.0099, + "step": 15675 + }, + { + "epoch": 7.185354691075515, + "grad_norm": 2.236577033996582, + "learning_rate": 2.90875e-06, + "loss": 0.0152, + "step": 15700 + }, + { + "epoch": 7.196796338672769, + "grad_norm": 0.2922148108482361, + "learning_rate": 2.9009375e-06, + "loss": 0.0066, + "step": 15725 + }, + { + "epoch": 7.2082379862700225, + "grad_norm": 3.0574800968170166, + "learning_rate": 2.893125e-06, + "loss": 0.0142, + "step": 15750 + }, + { + "epoch": 7.219679633867277, + "grad_norm": 0.9282683730125427, + "learning_rate": 2.8853125e-06, + "loss": 0.007, + "step": 15775 + }, + { + "epoch": 7.231121281464531, + "grad_norm": 1.8068912029266357, + "learning_rate": 2.8775e-06, + "loss": 0.0065, + "step": 15800 + }, + { + "epoch": 7.242562929061785, + "grad_norm": 2.347912311553955, + "learning_rate": 2.8696875000000002e-06, + "loss": 0.0098, + "step": 15825 + }, + { + "epoch": 7.254004576659039, + "grad_norm": 1.0899382829666138, + "learning_rate": 2.861875e-06, + "loss": 0.0143, + "step": 15850 + }, + { + "epoch": 7.2654462242562925, + "grad_norm": 0.7211443185806274, + "learning_rate": 2.8540625000000003e-06, + "loss": 0.0057, + "step": 15875 + }, + { + "epoch": 7.276887871853547, + "grad_norm": 1.6489779949188232, + "learning_rate": 2.8462500000000005e-06, + "loss": 0.0175, + "step": 15900 + }, + { + "epoch": 7.288329519450801, + "grad_norm": 1.7807652950286865, + "learning_rate": 2.8384375000000003e-06, + "loss": 0.0074, + "step": 15925 + }, + { + "epoch": 7.299771167048055, + "grad_norm": 2.1116390228271484, + "learning_rate": 2.830625e-06, + "loss": 0.0111, + "step": 15950 + }, + { + "epoch": 7.311212814645309, + "grad_norm": 0.1760861575603485, + "learning_rate": 2.8228125e-06, + "loss": 0.0098, + "step": 15975 + }, + { + "epoch": 7.322654462242563, + "grad_norm": 1.8475443124771118, + "learning_rate": 2.815e-06, + "loss": 0.0093, + "step": 16000 + }, + { + "epoch": 7.334096109839817, + "grad_norm": 0.14219874143600464, + "learning_rate": 2.8071875e-06, + "loss": 0.0057, + "step": 16025 + }, + { + "epoch": 7.345537757437071, + "grad_norm": 0.7175412774085999, + "learning_rate": 2.799375e-06, + "loss": 0.0195, + "step": 16050 + }, + { + "epoch": 7.356979405034325, + "grad_norm": 0.1249295026063919, + "learning_rate": 2.7915625e-06, + "loss": 0.0085, + "step": 16075 + }, + { + "epoch": 7.368421052631579, + "grad_norm": 1.440131664276123, + "learning_rate": 2.7837500000000002e-06, + "loss": 0.0176, + "step": 16100 + }, + { + "epoch": 7.379862700228833, + "grad_norm": 0.3971177339553833, + "learning_rate": 2.7759375e-06, + "loss": 0.0073, + "step": 16125 + }, + { + "epoch": 7.391304347826087, + "grad_norm": 3.1487958431243896, + "learning_rate": 2.7681250000000003e-06, + "loss": 0.0099, + "step": 16150 + }, + { + "epoch": 7.4027459954233406, + "grad_norm": 4.346505165100098, + "learning_rate": 2.7603125e-06, + "loss": 0.0061, + "step": 16175 + }, + { + "epoch": 7.414187643020595, + "grad_norm": 1.5192569494247437, + "learning_rate": 2.7525000000000003e-06, + "loss": 0.0195, + "step": 16200 + }, + { + "epoch": 7.425629290617849, + "grad_norm": 2.2581429481506348, + "learning_rate": 2.7446875e-06, + "loss": 0.0071, + "step": 16225 + }, + { + "epoch": 7.437070938215103, + "grad_norm": 1.6350927352905273, + "learning_rate": 2.7368750000000004e-06, + "loss": 0.0115, + "step": 16250 + }, + { + "epoch": 7.448512585812357, + "grad_norm": 0.6200711727142334, + "learning_rate": 2.7290625e-06, + "loss": 0.0083, + "step": 16275 + }, + { + "epoch": 7.459954233409611, + "grad_norm": 0.9812812805175781, + "learning_rate": 2.72125e-06, + "loss": 0.013, + "step": 16300 + }, + { + "epoch": 7.471395881006865, + "grad_norm": 0.33793002367019653, + "learning_rate": 2.7134375e-06, + "loss": 0.0074, + "step": 16325 + }, + { + "epoch": 7.482837528604119, + "grad_norm": 1.2503453493118286, + "learning_rate": 2.705625e-06, + "loss": 0.011, + "step": 16350 + }, + { + "epoch": 7.494279176201373, + "grad_norm": 0.3572548031806946, + "learning_rate": 2.6978125e-06, + "loss": 0.0058, + "step": 16375 + }, + { + "epoch": 7.505720823798627, + "grad_norm": 1.6965501308441162, + "learning_rate": 2.69e-06, + "loss": 0.0101, + "step": 16400 + }, + { + "epoch": 7.517162471395881, + "grad_norm": 1.6538255214691162, + "learning_rate": 2.6821875e-06, + "loss": 0.0079, + "step": 16425 + }, + { + "epoch": 7.528604118993135, + "grad_norm": 1.0653047561645508, + "learning_rate": 2.674375e-06, + "loss": 0.0247, + "step": 16450 + }, + { + "epoch": 7.540045766590389, + "grad_norm": 0.19629091024398804, + "learning_rate": 2.6665625e-06, + "loss": 0.0093, + "step": 16475 + }, + { + "epoch": 7.551487414187643, + "grad_norm": 0.48006531596183777, + "learning_rate": 2.65875e-06, + "loss": 0.0133, + "step": 16500 + }, + { + "epoch": 7.562929061784897, + "grad_norm": 0.11753380298614502, + "learning_rate": 2.6509375000000004e-06, + "loss": 0.0052, + "step": 16525 + }, + { + "epoch": 7.574370709382151, + "grad_norm": 0.975831151008606, + "learning_rate": 2.6434375e-06, + "loss": 0.0115, + "step": 16550 + }, + { + "epoch": 7.585812356979405, + "grad_norm": 3.755007028579712, + "learning_rate": 2.6356250000000003e-06, + "loss": 0.0086, + "step": 16575 + }, + { + "epoch": 7.597254004576659, + "grad_norm": 0.4555506408214569, + "learning_rate": 2.6278125e-06, + "loss": 0.0178, + "step": 16600 + }, + { + "epoch": 7.608695652173913, + "grad_norm": 0.7911310791969299, + "learning_rate": 2.6200000000000003e-06, + "loss": 0.0098, + "step": 16625 + }, + { + "epoch": 7.620137299771167, + "grad_norm": 1.7671183347702026, + "learning_rate": 2.6121875e-06, + "loss": 0.0178, + "step": 16650 + }, + { + "epoch": 7.631578947368421, + "grad_norm": 3.4077210426330566, + "learning_rate": 2.6043750000000004e-06, + "loss": 0.0107, + "step": 16675 + }, + { + "epoch": 7.643020594965675, + "grad_norm": 0.5805812478065491, + "learning_rate": 2.5965625e-06, + "loss": 0.014, + "step": 16700 + }, + { + "epoch": 7.654462242562929, + "grad_norm": 0.3688525855541229, + "learning_rate": 2.5887500000000004e-06, + "loss": 0.0069, + "step": 16725 + }, + { + "epoch": 7.665903890160183, + "grad_norm": 0.8044073581695557, + "learning_rate": 2.5809375000000002e-06, + "loss": 0.0158, + "step": 16750 + }, + { + "epoch": 7.6773455377574376, + "grad_norm": 0.10179438441991806, + "learning_rate": 2.573125e-06, + "loss": 0.0083, + "step": 16775 + }, + { + "epoch": 7.688787185354691, + "grad_norm": 1.3863494396209717, + "learning_rate": 2.5653125e-06, + "loss": 0.015, + "step": 16800 + }, + { + "epoch": 7.700228832951945, + "grad_norm": 0.7638989090919495, + "learning_rate": 2.5575e-06, + "loss": 0.0102, + "step": 16825 + }, + { + "epoch": 7.711670480549199, + "grad_norm": 1.0550347566604614, + "learning_rate": 2.5496875e-06, + "loss": 0.0142, + "step": 16850 + }, + { + "epoch": 7.723112128146453, + "grad_norm": 6.330323219299316, + "learning_rate": 2.541875e-06, + "loss": 0.0099, + "step": 16875 + }, + { + "epoch": 7.7345537757437075, + "grad_norm": 3.124359369277954, + "learning_rate": 2.5340625e-06, + "loss": 0.008, + "step": 16900 + }, + { + "epoch": 7.745995423340961, + "grad_norm": 0.30625733733177185, + "learning_rate": 2.52625e-06, + "loss": 0.006, + "step": 16925 + }, + { + "epoch": 7.757437070938215, + "grad_norm": 1.6125200986862183, + "learning_rate": 2.5184375e-06, + "loss": 0.0142, + "step": 16950 + }, + { + "epoch": 7.768878718535469, + "grad_norm": 0.9505332708358765, + "learning_rate": 2.510625e-06, + "loss": 0.0062, + "step": 16975 + }, + { + "epoch": 7.780320366132723, + "grad_norm": 3.020979404449463, + "learning_rate": 2.5028125e-06, + "loss": 0.0169, + "step": 17000 + }, + { + "epoch": 7.7917620137299775, + "grad_norm": 0.600764274597168, + "learning_rate": 2.4950000000000003e-06, + "loss": 0.0064, + "step": 17025 + }, + { + "epoch": 7.803203661327231, + "grad_norm": 1.8066636323928833, + "learning_rate": 2.4871875000000005e-06, + "loss": 0.0189, + "step": 17050 + }, + { + "epoch": 7.814645308924485, + "grad_norm": 9.108110427856445, + "learning_rate": 2.4793750000000003e-06, + "loss": 0.0071, + "step": 17075 + }, + { + "epoch": 7.826086956521739, + "grad_norm": 2.945603132247925, + "learning_rate": 2.4715625e-06, + "loss": 0.0089, + "step": 17100 + }, + { + "epoch": 7.837528604118993, + "grad_norm": 0.26592642068862915, + "learning_rate": 2.46375e-06, + "loss": 0.0071, + "step": 17125 + }, + { + "epoch": 7.848970251716247, + "grad_norm": 0.4964580833911896, + "learning_rate": 2.4559375e-06, + "loss": 0.0089, + "step": 17150 + }, + { + "epoch": 7.860411899313501, + "grad_norm": 3.0700931549072266, + "learning_rate": 2.448125e-06, + "loss": 0.0129, + "step": 17175 + }, + { + "epoch": 7.871853546910755, + "grad_norm": 1.6877917051315308, + "learning_rate": 2.4403125e-06, + "loss": 0.0187, + "step": 17200 + }, + { + "epoch": 7.883295194508009, + "grad_norm": 0.2074529379606247, + "learning_rate": 2.4325e-06, + "loss": 0.0101, + "step": 17225 + }, + { + "epoch": 7.894736842105263, + "grad_norm": 0.1611488312482834, + "learning_rate": 2.4246875000000002e-06, + "loss": 0.0163, + "step": 17250 + }, + { + "epoch": 7.906178489702517, + "grad_norm": 21.526241302490234, + "learning_rate": 2.416875e-06, + "loss": 0.0096, + "step": 17275 + }, + { + "epoch": 7.917620137299771, + "grad_norm": 1.5529296398162842, + "learning_rate": 2.4090625000000003e-06, + "loss": 0.0142, + "step": 17300 + }, + { + "epoch": 7.9290617848970255, + "grad_norm": 0.151444673538208, + "learning_rate": 2.40125e-06, + "loss": 0.0066, + "step": 17325 + }, + { + "epoch": 7.940503432494279, + "grad_norm": 1.1893980503082275, + "learning_rate": 2.3934375000000003e-06, + "loss": 0.0076, + "step": 17350 + }, + { + "epoch": 7.951945080091534, + "grad_norm": 3.2917895317077637, + "learning_rate": 2.385625e-06, + "loss": 0.0082, + "step": 17375 + }, + { + "epoch": 7.963386727688787, + "grad_norm": 1.4442905187606812, + "learning_rate": 2.3778125000000004e-06, + "loss": 0.0079, + "step": 17400 + }, + { + "epoch": 7.974828375286041, + "grad_norm": 3.447230577468872, + "learning_rate": 2.37e-06, + "loss": 0.0117, + "step": 17425 + }, + { + "epoch": 7.9862700228832955, + "grad_norm": 0.0969911515712738, + "learning_rate": 2.3621875e-06, + "loss": 0.0186, + "step": 17450 + }, + { + "epoch": 7.997711670480549, + "grad_norm": 3.7749619483947754, + "learning_rate": 2.354375e-06, + "loss": 0.0064, + "step": 17475 + }, + { + "epoch": 8.009153318077804, + "grad_norm": 0.26138579845428467, + "learning_rate": 2.3465625e-06, + "loss": 0.0155, + "step": 17500 + }, + { + "epoch": 8.020594965675057, + "grad_norm": 1.6767950057983398, + "learning_rate": 2.33875e-06, + "loss": 0.0067, + "step": 17525 + }, + { + "epoch": 8.03203661327231, + "grad_norm": 2.5147764682769775, + "learning_rate": 2.3309375e-06, + "loss": 0.0128, + "step": 17550 + }, + { + "epoch": 8.043478260869565, + "grad_norm": 0.12537100911140442, + "learning_rate": 2.323125e-06, + "loss": 0.0071, + "step": 17575 + }, + { + "epoch": 8.05491990846682, + "grad_norm": 7.097085952758789, + "learning_rate": 2.3153125e-06, + "loss": 0.0069, + "step": 17600 + }, + { + "epoch": 8.066361556064074, + "grad_norm": 0.1975085735321045, + "learning_rate": 2.3075e-06, + "loss": 0.0063, + "step": 17625 + }, + { + "epoch": 8.077803203661327, + "grad_norm": 2.5416934490203857, + "learning_rate": 2.2996875e-06, + "loss": 0.0067, + "step": 17650 + }, + { + "epoch": 8.08924485125858, + "grad_norm": 0.9218634963035583, + "learning_rate": 2.2918750000000004e-06, + "loss": 0.0081, + "step": 17675 + }, + { + "epoch": 8.100686498855834, + "grad_norm": 1.0671167373657227, + "learning_rate": 2.2840625e-06, + "loss": 0.0153, + "step": 17700 + }, + { + "epoch": 8.11212814645309, + "grad_norm": 0.07980553060770035, + "learning_rate": 2.2762500000000004e-06, + "loss": 0.0121, + "step": 17725 + }, + { + "epoch": 8.123569794050344, + "grad_norm": 0.6690077781677246, + "learning_rate": 2.2684375000000003e-06, + "loss": 0.0061, + "step": 17750 + }, + { + "epoch": 8.135011441647597, + "grad_norm": 14.465129852294922, + "learning_rate": 2.260625e-06, + "loss": 0.0083, + "step": 17775 + }, + { + "epoch": 8.14645308924485, + "grad_norm": 1.2041326761245728, + "learning_rate": 2.2528125e-06, + "loss": 0.0093, + "step": 17800 + }, + { + "epoch": 8.157894736842104, + "grad_norm": 6.081242561340332, + "learning_rate": 2.245e-06, + "loss": 0.0068, + "step": 17825 + }, + { + "epoch": 8.16933638443936, + "grad_norm": 1.041059970855713, + "learning_rate": 2.2371875e-06, + "loss": 0.012, + "step": 17850 + }, + { + "epoch": 8.180778032036613, + "grad_norm": 2.5304393768310547, + "learning_rate": 2.229375e-06, + "loss": 0.0145, + "step": 17875 + }, + { + "epoch": 8.192219679633867, + "grad_norm": 0.06209346652030945, + "learning_rate": 2.2215625e-06, + "loss": 0.0061, + "step": 17900 + }, + { + "epoch": 8.20366132723112, + "grad_norm": 0.9206859469413757, + "learning_rate": 2.21375e-06, + "loss": 0.0116, + "step": 17925 + }, + { + "epoch": 8.215102974828376, + "grad_norm": 1.2147653102874756, + "learning_rate": 2.2059375e-06, + "loss": 0.0067, + "step": 17950 + }, + { + "epoch": 8.22654462242563, + "grad_norm": 0.1464168131351471, + "learning_rate": 2.1981250000000002e-06, + "loss": 0.0078, + "step": 17975 + }, + { + "epoch": 8.237986270022883, + "grad_norm": 1.9483212232589722, + "learning_rate": 2.1903125e-06, + "loss": 0.0112, + "step": 18000 + }, + { + "epoch": 8.249427917620137, + "grad_norm": 3.755913734436035, + "learning_rate": 2.1825000000000003e-06, + "loss": 0.0093, + "step": 18025 + }, + { + "epoch": 8.26086956521739, + "grad_norm": 1.1962236166000366, + "learning_rate": 2.1746875e-06, + "loss": 0.007, + "step": 18050 + }, + { + "epoch": 8.272311212814646, + "grad_norm": 1.5700790882110596, + "learning_rate": 2.1668750000000003e-06, + "loss": 0.0115, + "step": 18075 + }, + { + "epoch": 8.2837528604119, + "grad_norm": 0.4117244780063629, + "learning_rate": 2.1590625e-06, + "loss": 0.0089, + "step": 18100 + }, + { + "epoch": 8.295194508009153, + "grad_norm": 2.4972431659698486, + "learning_rate": 2.15125e-06, + "loss": 0.0051, + "step": 18125 + }, + { + "epoch": 8.306636155606407, + "grad_norm": 0.6796992421150208, + "learning_rate": 2.1434374999999998e-06, + "loss": 0.01, + "step": 18150 + }, + { + "epoch": 8.31807780320366, + "grad_norm": 1.471119999885559, + "learning_rate": 2.135625e-06, + "loss": 0.0051, + "step": 18175 + }, + { + "epoch": 8.329519450800916, + "grad_norm": 0.32256969809532166, + "learning_rate": 2.1278125e-06, + "loss": 0.01, + "step": 18200 + }, + { + "epoch": 8.34096109839817, + "grad_norm": 1.6328589916229248, + "learning_rate": 2.12e-06, + "loss": 0.0045, + "step": 18225 + }, + { + "epoch": 8.352402745995423, + "grad_norm": 0.09561590105295181, + "learning_rate": 2.1121875e-06, + "loss": 0.0106, + "step": 18250 + }, + { + "epoch": 8.363844393592677, + "grad_norm": 4.785649299621582, + "learning_rate": 2.104375e-06, + "loss": 0.0094, + "step": 18275 + }, + { + "epoch": 8.37528604118993, + "grad_norm": 1.282939076423645, + "learning_rate": 2.0965625000000003e-06, + "loss": 0.0095, + "step": 18300 + }, + { + "epoch": 8.386727688787186, + "grad_norm": 0.11949565261602402, + "learning_rate": 2.08875e-06, + "loss": 0.0082, + "step": 18325 + }, + { + "epoch": 8.39816933638444, + "grad_norm": 0.13847044110298157, + "learning_rate": 2.0809375000000004e-06, + "loss": 0.01, + "step": 18350 + }, + { + "epoch": 8.409610983981693, + "grad_norm": 0.5209012031555176, + "learning_rate": 2.073125e-06, + "loss": 0.0073, + "step": 18375 + }, + { + "epoch": 8.421052631578947, + "grad_norm": 0.20265115797519684, + "learning_rate": 2.0653125000000004e-06, + "loss": 0.0115, + "step": 18400 + }, + { + "epoch": 8.4324942791762, + "grad_norm": 0.9013919830322266, + "learning_rate": 2.0575e-06, + "loss": 0.0099, + "step": 18425 + }, + { + "epoch": 8.443935926773456, + "grad_norm": 1.9062713384628296, + "learning_rate": 2.0496875e-06, + "loss": 0.0112, + "step": 18450 + }, + { + "epoch": 8.45537757437071, + "grad_norm": 3.9823992252349854, + "learning_rate": 2.041875e-06, + "loss": 0.0066, + "step": 18475 + }, + { + "epoch": 8.466819221967963, + "grad_norm": 0.3470716178417206, + "learning_rate": 2.0340625e-06, + "loss": 0.0049, + "step": 18500 + }, + { + "epoch": 8.478260869565217, + "grad_norm": 1.912090539932251, + "learning_rate": 2.02625e-06, + "loss": 0.0072, + "step": 18525 + }, + { + "epoch": 8.48970251716247, + "grad_norm": 2.102421522140503, + "learning_rate": 2.0184375e-06, + "loss": 0.0081, + "step": 18550 + }, + { + "epoch": 8.501144164759726, + "grad_norm": 0.11117032915353775, + "learning_rate": 2.010625e-06, + "loss": 0.0057, + "step": 18575 + }, + { + "epoch": 8.51258581235698, + "grad_norm": 0.21074584126472473, + "learning_rate": 2.0028125e-06, + "loss": 0.0103, + "step": 18600 + }, + { + "epoch": 8.524027459954233, + "grad_norm": 0.12100081890821457, + "learning_rate": 1.995e-06, + "loss": 0.0038, + "step": 18625 + }, + { + "epoch": 8.535469107551487, + "grad_norm": 0.08020362257957458, + "learning_rate": 1.9871875e-06, + "loss": 0.0077, + "step": 18650 + }, + { + "epoch": 8.546910755148742, + "grad_norm": 2.3164095878601074, + "learning_rate": 1.979375e-06, + "loss": 0.0087, + "step": 18675 + }, + { + "epoch": 8.558352402745996, + "grad_norm": 0.48693156242370605, + "learning_rate": 1.9715625000000002e-06, + "loss": 0.0079, + "step": 18700 + }, + { + "epoch": 8.56979405034325, + "grad_norm": 0.06513327360153198, + "learning_rate": 1.96375e-06, + "loss": 0.0093, + "step": 18725 + }, + { + "epoch": 8.581235697940503, + "grad_norm": 0.049644988030195236, + "learning_rate": 1.9559375000000003e-06, + "loss": 0.0122, + "step": 18750 + }, + { + "epoch": 8.592677345537757, + "grad_norm": 0.46888506412506104, + "learning_rate": 1.948125e-06, + "loss": 0.0066, + "step": 18775 + }, + { + "epoch": 8.604118993135012, + "grad_norm": 1.531025767326355, + "learning_rate": 1.9403125000000003e-06, + "loss": 0.0089, + "step": 18800 + }, + { + "epoch": 8.615560640732266, + "grad_norm": 3.112746000289917, + "learning_rate": 1.9325e-06, + "loss": 0.0101, + "step": 18825 + }, + { + "epoch": 8.62700228832952, + "grad_norm": 0.7876713871955872, + "learning_rate": 1.9246875e-06, + "loss": 0.0138, + "step": 18850 + }, + { + "epoch": 8.638443935926773, + "grad_norm": 16.939899444580078, + "learning_rate": 1.9168749999999998e-06, + "loss": 0.0098, + "step": 18875 + }, + { + "epoch": 8.649885583524027, + "grad_norm": 0.5488317608833313, + "learning_rate": 1.9090625e-06, + "loss": 0.0098, + "step": 18900 + }, + { + "epoch": 8.661327231121282, + "grad_norm": 9.495604515075684, + "learning_rate": 1.9012500000000002e-06, + "loss": 0.0125, + "step": 18925 + }, + { + "epoch": 8.672768878718536, + "grad_norm": 0.20728307962417603, + "learning_rate": 1.8934375e-06, + "loss": 0.0084, + "step": 18950 + }, + { + "epoch": 8.68421052631579, + "grad_norm": 13.799328804016113, + "learning_rate": 1.8856250000000003e-06, + "loss": 0.0169, + "step": 18975 + }, + { + "epoch": 8.695652173913043, + "grad_norm": 0.3443595767021179, + "learning_rate": 1.8778125e-06, + "loss": 0.0129, + "step": 19000 + }, + { + "epoch": 8.707093821510298, + "grad_norm": 0.05261196941137314, + "learning_rate": 1.8700000000000003e-06, + "loss": 0.0081, + "step": 19025 + }, + { + "epoch": 8.718535469107552, + "grad_norm": 1.2424705028533936, + "learning_rate": 1.8621875000000001e-06, + "loss": 0.0138, + "step": 19050 + }, + { + "epoch": 8.729977116704806, + "grad_norm": 2.1272494792938232, + "learning_rate": 1.8543750000000001e-06, + "loss": 0.0087, + "step": 19075 + }, + { + "epoch": 8.74141876430206, + "grad_norm": 0.19991889595985413, + "learning_rate": 1.8465625e-06, + "loss": 0.011, + "step": 19100 + }, + { + "epoch": 8.752860411899313, + "grad_norm": 2.920642375946045, + "learning_rate": 1.8387500000000002e-06, + "loss": 0.0072, + "step": 19125 + }, + { + "epoch": 8.764302059496568, + "grad_norm": 2.8176496028900146, + "learning_rate": 1.8309375e-06, + "loss": 0.0093, + "step": 19150 + }, + { + "epoch": 8.775743707093822, + "grad_norm": 2.0975587368011475, + "learning_rate": 1.8231250000000002e-06, + "loss": 0.0071, + "step": 19175 + }, + { + "epoch": 8.787185354691076, + "grad_norm": 0.34839051961898804, + "learning_rate": 1.8153125e-06, + "loss": 0.0061, + "step": 19200 + }, + { + "epoch": 8.79862700228833, + "grad_norm": 1.3293085098266602, + "learning_rate": 1.8075000000000003e-06, + "loss": 0.0088, + "step": 19225 + }, + { + "epoch": 8.810068649885583, + "grad_norm": 0.2044825553894043, + "learning_rate": 1.7996875e-06, + "loss": 0.0085, + "step": 19250 + }, + { + "epoch": 8.821510297482838, + "grad_norm": 0.24040932953357697, + "learning_rate": 1.791875e-06, + "loss": 0.0073, + "step": 19275 + }, + { + "epoch": 8.832951945080092, + "grad_norm": 4.296323299407959, + "learning_rate": 1.784375e-06, + "loss": 0.0103, + "step": 19300 + }, + { + "epoch": 8.844393592677346, + "grad_norm": 0.16996651887893677, + "learning_rate": 1.7765625000000002e-06, + "loss": 0.0046, + "step": 19325 + }, + { + "epoch": 8.8558352402746, + "grad_norm": 7.846490859985352, + "learning_rate": 1.76875e-06, + "loss": 0.0091, + "step": 19350 + }, + { + "epoch": 8.867276887871853, + "grad_norm": 6.62661600112915, + "learning_rate": 1.7609375e-06, + "loss": 0.0038, + "step": 19375 + }, + { + "epoch": 8.878718535469108, + "grad_norm": 0.48901718854904175, + "learning_rate": 1.7531250000000003e-06, + "loss": 0.0084, + "step": 19400 + }, + { + "epoch": 8.890160183066362, + "grad_norm": 0.669843852519989, + "learning_rate": 1.7453125e-06, + "loss": 0.0073, + "step": 19425 + }, + { + "epoch": 8.901601830663616, + "grad_norm": 5.433796405792236, + "learning_rate": 1.7375000000000003e-06, + "loss": 0.0065, + "step": 19450 + }, + { + "epoch": 8.91304347826087, + "grad_norm": 0.07918695360422134, + "learning_rate": 1.7296875000000001e-06, + "loss": 0.0056, + "step": 19475 + }, + { + "epoch": 8.924485125858123, + "grad_norm": 0.07582997530698776, + "learning_rate": 1.7218750000000001e-06, + "loss": 0.0115, + "step": 19500 + }, + { + "epoch": 8.935926773455378, + "grad_norm": 0.5449735522270203, + "learning_rate": 1.7140625e-06, + "loss": 0.0106, + "step": 19525 + }, + { + "epoch": 8.947368421052632, + "grad_norm": 0.7232295274734497, + "learning_rate": 1.7062500000000002e-06, + "loss": 0.0059, + "step": 19550 + }, + { + "epoch": 8.958810068649885, + "grad_norm": 4.808300495147705, + "learning_rate": 1.6984375e-06, + "loss": 0.0103, + "step": 19575 + }, + { + "epoch": 8.97025171624714, + "grad_norm": 5.207526683807373, + "learning_rate": 1.6906250000000002e-06, + "loss": 0.0104, + "step": 19600 + }, + { + "epoch": 8.981693363844393, + "grad_norm": 0.05005680024623871, + "learning_rate": 1.6828125e-06, + "loss": 0.0056, + "step": 19625 + }, + { + "epoch": 8.993135011441648, + "grad_norm": 0.113949254155159, + "learning_rate": 1.6750000000000003e-06, + "loss": 0.0061, + "step": 19650 + }, + { + "epoch": 9.004576659038902, + "grad_norm": 0.4994412064552307, + "learning_rate": 1.6671875e-06, + "loss": 0.0114, + "step": 19675 + }, + { + "epoch": 9.016018306636155, + "grad_norm": 1.9782668352127075, + "learning_rate": 1.659375e-06, + "loss": 0.005, + "step": 19700 + }, + { + "epoch": 9.027459954233409, + "grad_norm": 0.5642163157463074, + "learning_rate": 1.6515625e-06, + "loss": 0.0128, + "step": 19725 + }, + { + "epoch": 9.038901601830664, + "grad_norm": 0.13898131251335144, + "learning_rate": 1.6437500000000001e-06, + "loss": 0.0057, + "step": 19750 + }, + { + "epoch": 9.050343249427918, + "grad_norm": 0.932721734046936, + "learning_rate": 1.6359375e-06, + "loss": 0.0059, + "step": 19775 + }, + { + "epoch": 9.061784897025172, + "grad_norm": 3.428349494934082, + "learning_rate": 1.6281250000000002e-06, + "loss": 0.0084, + "step": 19800 + }, + { + "epoch": 9.073226544622425, + "grad_norm": 3.3652799129486084, + "learning_rate": 1.6203125e-06, + "loss": 0.0117, + "step": 19825 + }, + { + "epoch": 9.084668192219679, + "grad_norm": 0.9847972393035889, + "learning_rate": 1.6125e-06, + "loss": 0.0039, + "step": 19850 + }, + { + "epoch": 9.096109839816934, + "grad_norm": 0.21777117252349854, + "learning_rate": 1.6046875e-06, + "loss": 0.0077, + "step": 19875 + }, + { + "epoch": 9.107551487414188, + "grad_norm": 0.054129473865032196, + "learning_rate": 1.596875e-06, + "loss": 0.0053, + "step": 19900 + }, + { + "epoch": 9.118993135011442, + "grad_norm": 0.08982221782207489, + "learning_rate": 1.5890624999999999e-06, + "loss": 0.0065, + "step": 19925 + }, + { + "epoch": 9.130434782608695, + "grad_norm": 0.7745457291603088, + "learning_rate": 1.5812500000000001e-06, + "loss": 0.0055, + "step": 19950 + }, + { + "epoch": 9.141876430205949, + "grad_norm": 1.2945647239685059, + "learning_rate": 1.5734375e-06, + "loss": 0.0082, + "step": 19975 + }, + { + "epoch": 9.153318077803204, + "grad_norm": 1.5046677589416504, + "learning_rate": 1.5656250000000002e-06, + "loss": 0.0057, + "step": 20000 + }, + { + "epoch": 9.153318077803204, + "eval_loss": 0.2091646045446396, + "eval_runtime": 5410.4377, + "eval_samples_per_second": 1.76, + "eval_steps_per_second": 0.22, + "eval_wer": 0.0982754207461445, + "step": 20000 + }, + { + "epoch": 9.164759725400458, + "grad_norm": 0.4961409866809845, + "learning_rate": 1.5578125000000002e-06, + "loss": 0.0101, + "step": 20025 + }, + { + "epoch": 9.176201372997712, + "grad_norm": 0.1165071576833725, + "learning_rate": 1.55e-06, + "loss": 0.0049, + "step": 20050 + }, + { + "epoch": 9.187643020594965, + "grad_norm": 4.693184852600098, + "learning_rate": 1.5421875e-06, + "loss": 0.0071, + "step": 20075 + }, + { + "epoch": 9.199084668192219, + "grad_norm": 2.6000266075134277, + "learning_rate": 1.534375e-06, + "loss": 0.0042, + "step": 20100 + }, + { + "epoch": 9.210526315789474, + "grad_norm": 0.3854849338531494, + "learning_rate": 1.5265625e-06, + "loss": 0.0047, + "step": 20125 + }, + { + "epoch": 9.221967963386728, + "grad_norm": 0.0833137035369873, + "learning_rate": 1.51875e-06, + "loss": 0.0045, + "step": 20150 + }, + { + "epoch": 9.233409610983982, + "grad_norm": 0.8595607876777649, + "learning_rate": 1.5109375e-06, + "loss": 0.0059, + "step": 20175 + }, + { + "epoch": 9.244851258581235, + "grad_norm": 0.15006107091903687, + "learning_rate": 1.5031250000000001e-06, + "loss": 0.0035, + "step": 20200 + }, + { + "epoch": 9.256292906178489, + "grad_norm": 0.9896960258483887, + "learning_rate": 1.4953125e-06, + "loss": 0.0064, + "step": 20225 + }, + { + "epoch": 9.267734553775744, + "grad_norm": 6.129720211029053, + "learning_rate": 1.4875e-06, + "loss": 0.0043, + "step": 20250 + }, + { + "epoch": 9.279176201372998, + "grad_norm": 0.558292031288147, + "learning_rate": 1.4796875e-06, + "loss": 0.0134, + "step": 20275 + }, + { + "epoch": 9.290617848970252, + "grad_norm": 0.06486662477254868, + "learning_rate": 1.471875e-06, + "loss": 0.0048, + "step": 20300 + }, + { + "epoch": 9.302059496567505, + "grad_norm": 0.475968599319458, + "learning_rate": 1.4640625000000002e-06, + "loss": 0.0102, + "step": 20325 + }, + { + "epoch": 9.31350114416476, + "grad_norm": 1.7022502422332764, + "learning_rate": 1.4562500000000002e-06, + "loss": 0.0038, + "step": 20350 + }, + { + "epoch": 9.324942791762014, + "grad_norm": 3.257913589477539, + "learning_rate": 1.4484375e-06, + "loss": 0.0123, + "step": 20375 + }, + { + "epoch": 9.336384439359268, + "grad_norm": 0.739531397819519, + "learning_rate": 1.440625e-06, + "loss": 0.0066, + "step": 20400 + }, + { + "epoch": 9.347826086956522, + "grad_norm": 0.6568574905395508, + "learning_rate": 1.4328125e-06, + "loss": 0.0081, + "step": 20425 + }, + { + "epoch": 9.359267734553775, + "grad_norm": 0.15944476425647736, + "learning_rate": 1.4250000000000001e-06, + "loss": 0.0041, + "step": 20450 + }, + { + "epoch": 9.37070938215103, + "grad_norm": 2.0112485885620117, + "learning_rate": 1.4171875000000001e-06, + "loss": 0.0114, + "step": 20475 + }, + { + "epoch": 9.382151029748284, + "grad_norm": 6.05720853805542, + "learning_rate": 1.4093750000000002e-06, + "loss": 0.0074, + "step": 20500 + }, + { + "epoch": 9.393592677345538, + "grad_norm": 5.496974468231201, + "learning_rate": 1.4015625000000002e-06, + "loss": 0.0071, + "step": 20525 + }, + { + "epoch": 9.405034324942791, + "grad_norm": 0.04030141234397888, + "learning_rate": 1.39375e-06, + "loss": 0.003, + "step": 20550 + }, + { + "epoch": 9.416475972540045, + "grad_norm": 2.8988194465637207, + "learning_rate": 1.3859375e-06, + "loss": 0.0073, + "step": 20575 + }, + { + "epoch": 9.4279176201373, + "grad_norm": 3.288280963897705, + "learning_rate": 1.378125e-06, + "loss": 0.0038, + "step": 20600 + }, + { + "epoch": 9.439359267734554, + "grad_norm": 0.5999969244003296, + "learning_rate": 1.3703125e-06, + "loss": 0.0072, + "step": 20625 + }, + { + "epoch": 9.450800915331808, + "grad_norm": 0.05087543651461601, + "learning_rate": 1.3625e-06, + "loss": 0.0047, + "step": 20650 + }, + { + "epoch": 9.462242562929061, + "grad_norm": 0.31114375591278076, + "learning_rate": 1.3546875e-06, + "loss": 0.0118, + "step": 20675 + }, + { + "epoch": 9.473684210526315, + "grad_norm": 0.05578223988413811, + "learning_rate": 1.3468750000000001e-06, + "loss": 0.0068, + "step": 20700 + }, + { + "epoch": 9.48512585812357, + "grad_norm": 0.2985018789768219, + "learning_rate": 1.3390625e-06, + "loss": 0.0135, + "step": 20725 + }, + { + "epoch": 9.496567505720824, + "grad_norm": 8.602644920349121, + "learning_rate": 1.33125e-06, + "loss": 0.0064, + "step": 20750 + }, + { + "epoch": 9.508009153318078, + "grad_norm": 0.8129026293754578, + "learning_rate": 1.3234375e-06, + "loss": 0.0074, + "step": 20775 + }, + { + "epoch": 9.519450800915331, + "grad_norm": 13.631547927856445, + "learning_rate": 1.315625e-06, + "loss": 0.0054, + "step": 20800 + }, + { + "epoch": 9.530892448512585, + "grad_norm": 1.1775178909301758, + "learning_rate": 1.3078125e-06, + "loss": 0.0147, + "step": 20825 + }, + { + "epoch": 9.54233409610984, + "grad_norm": 3.9796142578125, + "learning_rate": 1.3e-06, + "loss": 0.0072, + "step": 20850 + }, + { + "epoch": 9.553775743707094, + "grad_norm": 0.5741226077079773, + "learning_rate": 1.2921875e-06, + "loss": 0.0093, + "step": 20875 + }, + { + "epoch": 9.565217391304348, + "grad_norm": 0.49431973695755005, + "learning_rate": 1.284375e-06, + "loss": 0.0042, + "step": 20900 + }, + { + "epoch": 9.576659038901601, + "grad_norm": 0.24845191836357117, + "learning_rate": 1.2765625e-06, + "loss": 0.0059, + "step": 20925 + }, + { + "epoch": 9.588100686498855, + "grad_norm": 0.3911905884742737, + "learning_rate": 1.2687500000000001e-06, + "loss": 0.0066, + "step": 20950 + }, + { + "epoch": 9.59954233409611, + "grad_norm": 4.166600704193115, + "learning_rate": 1.2609375000000002e-06, + "loss": 0.0102, + "step": 20975 + }, + { + "epoch": 9.610983981693364, + "grad_norm": 1.9600577354431152, + "learning_rate": 1.2531250000000002e-06, + "loss": 0.0055, + "step": 21000 + }, + { + "epoch": 9.622425629290618, + "grad_norm": 0.053466055542230606, + "learning_rate": 1.2453125000000002e-06, + "loss": 0.0081, + "step": 21025 + }, + { + "epoch": 9.633867276887871, + "grad_norm": 0.10502848774194717, + "learning_rate": 1.2375000000000002e-06, + "loss": 0.0046, + "step": 21050 + }, + { + "epoch": 9.645308924485127, + "grad_norm": 0.6094481348991394, + "learning_rate": 1.2296875e-06, + "loss": 0.0081, + "step": 21075 + }, + { + "epoch": 9.65675057208238, + "grad_norm": 0.2322821468114853, + "learning_rate": 1.221875e-06, + "loss": 0.0064, + "step": 21100 + }, + { + "epoch": 9.668192219679634, + "grad_norm": 0.5858094692230225, + "learning_rate": 1.2140625e-06, + "loss": 0.0099, + "step": 21125 + }, + { + "epoch": 9.679633867276888, + "grad_norm": 0.4338737428188324, + "learning_rate": 1.20625e-06, + "loss": 0.0053, + "step": 21150 + }, + { + "epoch": 9.691075514874141, + "grad_norm": 2.8088247776031494, + "learning_rate": 1.1984375000000001e-06, + "loss": 0.0096, + "step": 21175 + }, + { + "epoch": 9.702517162471397, + "grad_norm": 0.27802830934524536, + "learning_rate": 1.1906250000000001e-06, + "loss": 0.005, + "step": 21200 + }, + { + "epoch": 9.71395881006865, + "grad_norm": 0.3727511167526245, + "learning_rate": 1.1828125000000002e-06, + "loss": 0.012, + "step": 21225 + }, + { + "epoch": 9.725400457665904, + "grad_norm": 0.17081360518932343, + "learning_rate": 1.175e-06, + "loss": 0.0059, + "step": 21250 + }, + { + "epoch": 9.736842105263158, + "grad_norm": 2.140308380126953, + "learning_rate": 1.1671875e-06, + "loss": 0.0146, + "step": 21275 + }, + { + "epoch": 9.748283752860411, + "grad_norm": 0.04215441644191742, + "learning_rate": 1.159375e-06, + "loss": 0.0056, + "step": 21300 + }, + { + "epoch": 9.759725400457667, + "grad_norm": 1.3280128240585327, + "learning_rate": 1.1515625e-06, + "loss": 0.0049, + "step": 21325 + }, + { + "epoch": 9.77116704805492, + "grad_norm": 0.8858660459518433, + "learning_rate": 1.14375e-06, + "loss": 0.0061, + "step": 21350 + }, + { + "epoch": 9.782608695652174, + "grad_norm": 2.4523348808288574, + "learning_rate": 1.1359375e-06, + "loss": 0.0074, + "step": 21375 + }, + { + "epoch": 9.794050343249427, + "grad_norm": 0.11217644065618515, + "learning_rate": 1.128125e-06, + "loss": 0.0037, + "step": 21400 + }, + { + "epoch": 9.805491990846681, + "grad_norm": 0.23124535381793976, + "learning_rate": 1.120625e-06, + "loss": 0.008, + "step": 21425 + }, + { + "epoch": 9.816933638443937, + "grad_norm": 4.284655570983887, + "learning_rate": 1.1128125000000002e-06, + "loss": 0.0083, + "step": 21450 + }, + { + "epoch": 9.82837528604119, + "grad_norm": 1.4181989431381226, + "learning_rate": 1.1050000000000002e-06, + "loss": 0.0068, + "step": 21475 + }, + { + "epoch": 9.839816933638444, + "grad_norm": 0.043931469321250916, + "learning_rate": 1.0971875e-06, + "loss": 0.0056, + "step": 21500 + }, + { + "epoch": 9.851258581235697, + "grad_norm": 1.3745752573013306, + "learning_rate": 1.089375e-06, + "loss": 0.0053, + "step": 21525 + }, + { + "epoch": 9.862700228832953, + "grad_norm": 4.101833820343018, + "learning_rate": 1.0815625e-06, + "loss": 0.005, + "step": 21550 + }, + { + "epoch": 9.874141876430206, + "grad_norm": 0.13216155767440796, + "learning_rate": 1.07375e-06, + "loss": 0.0046, + "step": 21575 + }, + { + "epoch": 9.88558352402746, + "grad_norm": 0.1495962142944336, + "learning_rate": 1.0659375000000001e-06, + "loss": 0.006, + "step": 21600 + }, + { + "epoch": 9.897025171624714, + "grad_norm": 6.025510787963867, + "learning_rate": 1.0581250000000001e-06, + "loss": 0.0096, + "step": 21625 + }, + { + "epoch": 9.908466819221967, + "grad_norm": 1.0735716819763184, + "learning_rate": 1.0503125000000002e-06, + "loss": 0.006, + "step": 21650 + }, + { + "epoch": 9.919908466819223, + "grad_norm": 0.1290319263935089, + "learning_rate": 1.0425e-06, + "loss": 0.0062, + "step": 21675 + }, + { + "epoch": 9.931350114416476, + "grad_norm": 0.3358752429485321, + "learning_rate": 1.0346875e-06, + "loss": 0.0072, + "step": 21700 + }, + { + "epoch": 9.94279176201373, + "grad_norm": 0.3810190260410309, + "learning_rate": 1.026875e-06, + "loss": 0.0054, + "step": 21725 + }, + { + "epoch": 9.954233409610984, + "grad_norm": 0.14147138595581055, + "learning_rate": 1.0190625e-06, + "loss": 0.0041, + "step": 21750 + }, + { + "epoch": 9.965675057208237, + "grad_norm": 0.8347612619400024, + "learning_rate": 1.01125e-06, + "loss": 0.0087, + "step": 21775 + }, + { + "epoch": 9.977116704805493, + "grad_norm": 0.0881999209523201, + "learning_rate": 1.0034375e-06, + "loss": 0.0072, + "step": 21800 + }, + { + "epoch": 9.988558352402746, + "grad_norm": 0.05394062027335167, + "learning_rate": 9.95625e-07, + "loss": 0.0085, + "step": 21825 + }, + { + "epoch": 10.0, + "grad_norm": 5.682041645050049, + "learning_rate": 9.878125000000001e-07, + "loss": 0.008, + "step": 21850 + }, + { + "epoch": 10.011441647597254, + "grad_norm": 0.030163856223225594, + "learning_rate": 9.8e-07, + "loss": 0.0035, + "step": 21875 + }, + { + "epoch": 10.022883295194507, + "grad_norm": 1.5702613592147827, + "learning_rate": 9.721875e-07, + "loss": 0.0059, + "step": 21900 + }, + { + "epoch": 10.034324942791763, + "grad_norm": 1.831508994102478, + "learning_rate": 9.64375e-07, + "loss": 0.007, + "step": 21925 + }, + { + "epoch": 10.045766590389016, + "grad_norm": 3.454017162322998, + "learning_rate": 9.565625e-07, + "loss": 0.007, + "step": 21950 + }, + { + "epoch": 10.05720823798627, + "grad_norm": 0.42849200963974, + "learning_rate": 9.4875e-07, + "loss": 0.0029, + "step": 21975 + }, + { + "epoch": 10.068649885583524, + "grad_norm": 10.296757698059082, + "learning_rate": 9.409374999999999e-07, + "loss": 0.012, + "step": 22000 + }, + { + "epoch": 10.080091533180777, + "grad_norm": 2.301107168197632, + "learning_rate": 9.33125e-07, + "loss": 0.0064, + "step": 22025 + }, + { + "epoch": 10.091533180778033, + "grad_norm": 0.2854618728160858, + "learning_rate": 9.253125e-07, + "loss": 0.009, + "step": 22050 + }, + { + "epoch": 10.102974828375286, + "grad_norm": 3.725808620452881, + "learning_rate": 9.175000000000001e-07, + "loss": 0.0039, + "step": 22075 + }, + { + "epoch": 10.11441647597254, + "grad_norm": 5.867912769317627, + "learning_rate": 9.096875000000001e-07, + "loss": 0.0078, + "step": 22100 + }, + { + "epoch": 10.125858123569794, + "grad_norm": 3.184048652648926, + "learning_rate": 9.018750000000002e-07, + "loss": 0.0051, + "step": 22125 + }, + { + "epoch": 10.137299771167047, + "grad_norm": 4.497511386871338, + "learning_rate": 8.940625000000001e-07, + "loss": 0.0089, + "step": 22150 + }, + { + "epoch": 10.148741418764303, + "grad_norm": 0.0537056103348732, + "learning_rate": 8.862500000000001e-07, + "loss": 0.0027, + "step": 22175 + }, + { + "epoch": 10.160183066361556, + "grad_norm": 3.5934197902679443, + "learning_rate": 8.784375000000001e-07, + "loss": 0.009, + "step": 22200 + }, + { + "epoch": 10.17162471395881, + "grad_norm": 0.23545417189598083, + "learning_rate": 8.706250000000001e-07, + "loss": 0.0058, + "step": 22225 + }, + { + "epoch": 10.183066361556063, + "grad_norm": 5.6573004722595215, + "learning_rate": 8.628125e-07, + "loss": 0.0082, + "step": 22250 + }, + { + "epoch": 10.194508009153319, + "grad_norm": 1.168822169303894, + "learning_rate": 8.550000000000001e-07, + "loss": 0.0032, + "step": 22275 + }, + { + "epoch": 10.205949656750573, + "grad_norm": 4.02138090133667, + "learning_rate": 8.471875000000001e-07, + "loss": 0.011, + "step": 22300 + }, + { + "epoch": 10.217391304347826, + "grad_norm": 0.18906021118164062, + "learning_rate": 8.39375e-07, + "loss": 0.0042, + "step": 22325 + }, + { + "epoch": 10.22883295194508, + "grad_norm": 4.549426078796387, + "learning_rate": 8.315625e-07, + "loss": 0.0127, + "step": 22350 + }, + { + "epoch": 10.240274599542333, + "grad_norm": 0.021799422800540924, + "learning_rate": 8.237500000000001e-07, + "loss": 0.0056, + "step": 22375 + }, + { + "epoch": 10.251716247139589, + "grad_norm": 6.450737953186035, + "learning_rate": 8.159375000000001e-07, + "loss": 0.0037, + "step": 22400 + }, + { + "epoch": 10.263157894736842, + "grad_norm": 0.04190767928957939, + "learning_rate": 8.08125e-07, + "loss": 0.0087, + "step": 22425 + }, + { + "epoch": 10.274599542334096, + "grad_norm": 5.531692028045654, + "learning_rate": 8.003125e-07, + "loss": 0.0076, + "step": 22450 + }, + { + "epoch": 10.28604118993135, + "grad_norm": 0.1595929116010666, + "learning_rate": 7.925e-07, + "loss": 0.0043, + "step": 22475 + }, + { + "epoch": 10.297482837528603, + "grad_norm": 2.6428611278533936, + "learning_rate": 7.846875000000001e-07, + "loss": 0.0107, + "step": 22500 + }, + { + "epoch": 10.308924485125859, + "grad_norm": 0.043632976710796356, + "learning_rate": 7.76875e-07, + "loss": 0.0055, + "step": 22525 + }, + { + "epoch": 10.320366132723112, + "grad_norm": 4.209513187408447, + "learning_rate": 7.690625000000001e-07, + "loss": 0.014, + "step": 22550 + }, + { + "epoch": 10.331807780320366, + "grad_norm": 0.08739282190799713, + "learning_rate": 7.612500000000001e-07, + "loss": 0.0054, + "step": 22575 + }, + { + "epoch": 10.34324942791762, + "grad_norm": 5.106118202209473, + "learning_rate": 7.534375e-07, + "loss": 0.0048, + "step": 22600 + }, + { + "epoch": 10.354691075514873, + "grad_norm": 1.0995253324508667, + "learning_rate": 7.456250000000001e-07, + "loss": 0.0048, + "step": 22625 + }, + { + "epoch": 10.366132723112129, + "grad_norm": 4.488454341888428, + "learning_rate": 7.378125000000001e-07, + "loss": 0.0083, + "step": 22650 + }, + { + "epoch": 10.377574370709382, + "grad_norm": 0.04510258138179779, + "learning_rate": 7.3e-07, + "loss": 0.0054, + "step": 22675 + }, + { + "epoch": 10.389016018306636, + "grad_norm": 12.380441665649414, + "learning_rate": 7.221875e-07, + "loss": 0.0053, + "step": 22700 + }, + { + "epoch": 10.40045766590389, + "grad_norm": 0.06387301534414291, + "learning_rate": 7.14375e-07, + "loss": 0.0036, + "step": 22725 + }, + { + "epoch": 10.411899313501145, + "grad_norm": 5.258612155914307, + "learning_rate": 7.065625000000001e-07, + "loss": 0.0044, + "step": 22750 + }, + { + "epoch": 10.423340961098399, + "grad_norm": 1.0784777402877808, + "learning_rate": 6.9875e-07, + "loss": 0.0023, + "step": 22775 + }, + { + "epoch": 10.434782608695652, + "grad_norm": 9.012079238891602, + "learning_rate": 6.909375e-07, + "loss": 0.0085, + "step": 22800 + }, + { + "epoch": 10.446224256292906, + "grad_norm": 0.2832639515399933, + "learning_rate": 6.83125e-07, + "loss": 0.005, + "step": 22825 + }, + { + "epoch": 10.45766590389016, + "grad_norm": 1.8746007680892944, + "learning_rate": 6.753124999999999e-07, + "loss": 0.0067, + "step": 22850 + }, + { + "epoch": 10.469107551487415, + "grad_norm": 0.7013452649116516, + "learning_rate": 6.675000000000001e-07, + "loss": 0.0035, + "step": 22875 + }, + { + "epoch": 10.480549199084669, + "grad_norm": 4.465462684631348, + "learning_rate": 6.596875000000001e-07, + "loss": 0.0048, + "step": 22900 + }, + { + "epoch": 10.491990846681922, + "grad_norm": 3.2530734539031982, + "learning_rate": 6.51875e-07, + "loss": 0.0062, + "step": 22925 + }, + { + "epoch": 10.503432494279176, + "grad_norm": 11.731400489807129, + "learning_rate": 6.440625e-07, + "loss": 0.0157, + "step": 22950 + }, + { + "epoch": 10.51487414187643, + "grad_norm": 0.5493900775909424, + "learning_rate": 6.3625e-07, + "loss": 0.0039, + "step": 22975 + }, + { + "epoch": 10.526315789473685, + "grad_norm": 4.6817240715026855, + "learning_rate": 6.284375000000001e-07, + "loss": 0.0092, + "step": 23000 + }, + { + "epoch": 10.537757437070939, + "grad_norm": 1.2315994501113892, + "learning_rate": 6.20625e-07, + "loss": 0.0044, + "step": 23025 + }, + { + "epoch": 10.549199084668192, + "grad_norm": 0.23386216163635254, + "learning_rate": 6.128125e-07, + "loss": 0.0079, + "step": 23050 + }, + { + "epoch": 10.560640732265446, + "grad_norm": 0.387458860874176, + "learning_rate": 6.05e-07, + "loss": 0.0032, + "step": 23075 + }, + { + "epoch": 10.5720823798627, + "grad_norm": 3.660433769226074, + "learning_rate": 5.971875e-07, + "loss": 0.0036, + "step": 23100 + }, + { + "epoch": 10.583524027459955, + "grad_norm": 2.7630832195281982, + "learning_rate": 5.89375e-07, + "loss": 0.0068, + "step": 23125 + }, + { + "epoch": 10.594965675057209, + "grad_norm": 12.062283515930176, + "learning_rate": 5.815625e-07, + "loss": 0.0101, + "step": 23150 + }, + { + "epoch": 10.606407322654462, + "grad_norm": 0.07887246459722519, + "learning_rate": 5.737500000000001e-07, + "loss": 0.0036, + "step": 23175 + }, + { + "epoch": 10.617848970251716, + "grad_norm": 0.8604605793952942, + "learning_rate": 5.659375e-07, + "loss": 0.008, + "step": 23200 + }, + { + "epoch": 10.62929061784897, + "grad_norm": 0.21621568500995636, + "learning_rate": 5.581250000000001e-07, + "loss": 0.0036, + "step": 23225 + }, + { + "epoch": 10.640732265446225, + "grad_norm": 4.55475378036499, + "learning_rate": 5.503125000000001e-07, + "loss": 0.0049, + "step": 23250 + }, + { + "epoch": 10.652173913043478, + "grad_norm": 0.07743289321660995, + "learning_rate": 5.425e-07, + "loss": 0.0022, + "step": 23275 + }, + { + "epoch": 10.663615560640732, + "grad_norm": 2.8529226779937744, + "learning_rate": 5.346875e-07, + "loss": 0.0035, + "step": 23300 + }, + { + "epoch": 10.675057208237986, + "grad_norm": 1.4317353963851929, + "learning_rate": 5.26875e-07, + "loss": 0.0046, + "step": 23325 + }, + { + "epoch": 10.68649885583524, + "grad_norm": 4.198233127593994, + "learning_rate": 5.190625000000001e-07, + "loss": 0.0039, + "step": 23350 + }, + { + "epoch": 10.697940503432495, + "grad_norm": 1.182904601097107, + "learning_rate": 5.1125e-07, + "loss": 0.0053, + "step": 23375 + }, + { + "epoch": 10.709382151029748, + "grad_norm": 7.1977105140686035, + "learning_rate": 5.034375e-07, + "loss": 0.0103, + "step": 23400 + }, + { + "epoch": 10.720823798627002, + "grad_norm": 0.051739439368247986, + "learning_rate": 4.95625e-07, + "loss": 0.0044, + "step": 23425 + }, + { + "epoch": 10.732265446224256, + "grad_norm": 6.68768310546875, + "learning_rate": 4.878125e-07, + "loss": 0.0084, + "step": 23450 + }, + { + "epoch": 10.743707093821511, + "grad_norm": 0.04325169697403908, + "learning_rate": 4.8e-07, + "loss": 0.0054, + "step": 23475 + }, + { + "epoch": 10.755148741418765, + "grad_norm": 7.35172700881958, + "learning_rate": 4.721875000000001e-07, + "loss": 0.0173, + "step": 23500 + }, + { + "epoch": 10.766590389016018, + "grad_norm": 1.9433234930038452, + "learning_rate": 4.6437500000000005e-07, + "loss": 0.0032, + "step": 23525 + }, + { + "epoch": 10.778032036613272, + "grad_norm": 9.553720474243164, + "learning_rate": 4.565625e-07, + "loss": 0.0047, + "step": 23550 + }, + { + "epoch": 10.789473684210526, + "grad_norm": 0.2103455811738968, + "learning_rate": 4.4875000000000004e-07, + "loss": 0.0047, + "step": 23575 + }, + { + "epoch": 10.800915331807781, + "grad_norm": 0.601758599281311, + "learning_rate": 4.409375e-07, + "loss": 0.0091, + "step": 23600 + }, + { + "epoch": 10.812356979405035, + "grad_norm": 0.04723617061972618, + "learning_rate": 4.3312500000000004e-07, + "loss": 0.0043, + "step": 23625 + }, + { + "epoch": 10.823798627002288, + "grad_norm": 0.1756594479084015, + "learning_rate": 4.253125e-07, + "loss": 0.0068, + "step": 23650 + }, + { + "epoch": 10.835240274599542, + "grad_norm": 0.13647685945034027, + "learning_rate": 4.175e-07, + "loss": 0.0035, + "step": 23675 + }, + { + "epoch": 10.846681922196796, + "grad_norm": 6.657943248748779, + "learning_rate": 4.096875e-07, + "loss": 0.0086, + "step": 23700 + }, + { + "epoch": 10.858123569794051, + "grad_norm": 1.3252496719360352, + "learning_rate": 4.01875e-07, + "loss": 0.0046, + "step": 23725 + }, + { + "epoch": 10.869565217391305, + "grad_norm": 0.7701499462127686, + "learning_rate": 3.940625e-07, + "loss": 0.0059, + "step": 23750 + }, + { + "epoch": 10.881006864988558, + "grad_norm": 0.02328154630959034, + "learning_rate": 3.8625e-07, + "loss": 0.0063, + "step": 23775 + }, + { + "epoch": 10.892448512585812, + "grad_norm": 1.4766663312911987, + "learning_rate": 3.7843750000000003e-07, + "loss": 0.0058, + "step": 23800 + }, + { + "epoch": 10.903890160183066, + "grad_norm": 0.08133988082408905, + "learning_rate": 3.70625e-07, + "loss": 0.0032, + "step": 23825 + }, + { + "epoch": 10.915331807780321, + "grad_norm": 9.198525428771973, + "learning_rate": 3.628125e-07, + "loss": 0.0054, + "step": 23850 + }, + { + "epoch": 10.926773455377575, + "grad_norm": 0.09271799027919769, + "learning_rate": 3.5500000000000004e-07, + "loss": 0.0044, + "step": 23875 + }, + { + "epoch": 10.938215102974828, + "grad_norm": 1.473221778869629, + "learning_rate": 3.471875e-07, + "loss": 0.0049, + "step": 23900 + }, + { + "epoch": 10.949656750572082, + "grad_norm": 3.2264695167541504, + "learning_rate": 3.3937500000000003e-07, + "loss": 0.0049, + "step": 23925 + }, + { + "epoch": 10.961098398169337, + "grad_norm": 14.144686698913574, + "learning_rate": 3.315625e-07, + "loss": 0.0072, + "step": 23950 + }, + { + "epoch": 10.972540045766591, + "grad_norm": 2.536242723464966, + "learning_rate": 3.2375e-07, + "loss": 0.0047, + "step": 23975 + }, + { + "epoch": 10.983981693363845, + "grad_norm": 7.181889533996582, + "learning_rate": 3.159375e-07, + "loss": 0.0123, + "step": 24000 + }, + { + "epoch": 10.995423340961098, + "grad_norm": 0.28295642137527466, + "learning_rate": 3.08125e-07, + "loss": 0.005, + "step": 24025 + }, + { + "epoch": 11.006864988558352, + "grad_norm": 1.317877173423767, + "learning_rate": 3.0031250000000004e-07, + "loss": 0.005, + "step": 24050 + }, + { + "epoch": 11.018306636155607, + "grad_norm": 0.0772235319018364, + "learning_rate": 2.925e-07, + "loss": 0.0016, + "step": 24075 + }, + { + "epoch": 11.02974828375286, + "grad_norm": 2.499826192855835, + "learning_rate": 2.8468750000000003e-07, + "loss": 0.0084, + "step": 24100 + }, + { + "epoch": 11.041189931350115, + "grad_norm": 12.638914108276367, + "learning_rate": 2.76875e-07, + "loss": 0.004, + "step": 24125 + }, + { + "epoch": 11.052631578947368, + "grad_norm": 0.1981966644525528, + "learning_rate": 2.690625e-07, + "loss": 0.0083, + "step": 24150 + }, + { + "epoch": 11.064073226544622, + "grad_norm": 0.04735041409730911, + "learning_rate": 2.6125e-07, + "loss": 0.0054, + "step": 24175 + }, + { + "epoch": 11.075514874141877, + "grad_norm": 0.5447256565093994, + "learning_rate": 2.534375e-07, + "loss": 0.0038, + "step": 24200 + }, + { + "epoch": 11.08695652173913, + "grad_norm": 0.0347541943192482, + "learning_rate": 2.4562500000000003e-07, + "loss": 0.0029, + "step": 24225 + }, + { + "epoch": 11.098398169336384, + "grad_norm": 13.768309593200684, + "learning_rate": 2.3781250000000003e-07, + "loss": 0.0049, + "step": 24250 + }, + { + "epoch": 11.109839816933638, + "grad_norm": 1.426023244857788, + "learning_rate": 2.3e-07, + "loss": 0.0054, + "step": 24275 + }, + { + "epoch": 11.121281464530892, + "grad_norm": 0.021928640082478523, + "learning_rate": 2.221875e-07, + "loss": 0.0063, + "step": 24300 + }, + { + "epoch": 11.132723112128147, + "grad_norm": 2.05029296875, + "learning_rate": 2.1437499999999999e-07, + "loss": 0.0046, + "step": 24325 + }, + { + "epoch": 11.1441647597254, + "grad_norm": 0.16184775531291962, + "learning_rate": 2.0656250000000003e-07, + "loss": 0.0062, + "step": 24350 + }, + { + "epoch": 11.155606407322654, + "grad_norm": 1.0633113384246826, + "learning_rate": 1.9875000000000003e-07, + "loss": 0.0049, + "step": 24375 + }, + { + "epoch": 11.167048054919908, + "grad_norm": 0.7184084057807922, + "learning_rate": 1.9093750000000002e-07, + "loss": 0.0088, + "step": 24400 + }, + { + "epoch": 11.178489702517162, + "grad_norm": 0.5524642467498779, + "learning_rate": 1.83125e-07, + "loss": 0.0019, + "step": 24425 + }, + { + "epoch": 11.189931350114417, + "grad_norm": 0.1881658285856247, + "learning_rate": 1.753125e-07, + "loss": 0.0061, + "step": 24450 + }, + { + "epoch": 11.20137299771167, + "grad_norm": 3.7017602920532227, + "learning_rate": 1.675e-07, + "loss": 0.006, + "step": 24475 + }, + { + "epoch": 11.212814645308924, + "grad_norm": 1.5420652627944946, + "learning_rate": 1.596875e-07, + "loss": 0.0048, + "step": 24500 + }, + { + "epoch": 11.224256292906178, + "grad_norm": 5.843155384063721, + "learning_rate": 1.51875e-07, + "loss": 0.0064, + "step": 24525 + }, + { + "epoch": 11.235697940503432, + "grad_norm": 0.06349411606788635, + "learning_rate": 1.4406250000000002e-07, + "loss": 0.0092, + "step": 24550 + }, + { + "epoch": 11.247139588100687, + "grad_norm": 2.3772706985473633, + "learning_rate": 1.3625000000000002e-07, + "loss": 0.0029, + "step": 24575 + }, + { + "epoch": 11.25858123569794, + "grad_norm": 1.1983476877212524, + "learning_rate": 1.284375e-07, + "loss": 0.0071, + "step": 24600 + }, + { + "epoch": 11.270022883295194, + "grad_norm": 0.02243354730308056, + "learning_rate": 1.20625e-07, + "loss": 0.0035, + "step": 24625 + }, + { + "epoch": 11.281464530892448, + "grad_norm": 1.8328884840011597, + "learning_rate": 1.1312500000000002e-07, + "loss": 0.0097, + "step": 24650 + }, + { + "epoch": 11.292906178489703, + "grad_norm": 0.05551990494132042, + "learning_rate": 1.0531250000000001e-07, + "loss": 0.0034, + "step": 24675 + }, + { + "epoch": 11.304347826086957, + "grad_norm": 0.030004534870386124, + "learning_rate": 9.75e-08, + "loss": 0.0103, + "step": 24700 + }, + { + "epoch": 11.31578947368421, + "grad_norm": 0.03290963172912598, + "learning_rate": 8.96875e-08, + "loss": 0.0074, + "step": 24725 + }, + { + "epoch": 11.327231121281464, + "grad_norm": 0.29857203364372253, + "learning_rate": 8.187500000000001e-08, + "loss": 0.0057, + "step": 24750 + }, + { + "epoch": 11.338672768878718, + "grad_norm": 1.0099925994873047, + "learning_rate": 7.40625e-08, + "loss": 0.0066, + "step": 24775 + }, + { + "epoch": 11.350114416475973, + "grad_norm": 0.26381850242614746, + "learning_rate": 6.625e-08, + "loss": 0.0063, + "step": 24800 + }, + { + "epoch": 11.361556064073227, + "grad_norm": 0.07193097472190857, + "learning_rate": 5.843750000000001e-08, + "loss": 0.006, + "step": 24825 + }, + { + "epoch": 11.37299771167048, + "grad_norm": 0.05023621395230293, + "learning_rate": 5.0625e-08, + "loss": 0.0032, + "step": 24850 + }, + { + "epoch": 11.384439359267734, + "grad_norm": 0.03356996551156044, + "learning_rate": 4.28125e-08, + "loss": 0.0036, + "step": 24875 + }, + { + "epoch": 11.395881006864988, + "grad_norm": 1.0069869756698608, + "learning_rate": 3.5e-08, + "loss": 0.0078, + "step": 24900 + }, + { + "epoch": 11.407322654462243, + "grad_norm": 0.020418858155608177, + "learning_rate": 2.7187499999999998e-08, + "loss": 0.0024, + "step": 24925 + }, + { + "epoch": 11.418764302059497, + "grad_norm": 0.03717822954058647, + "learning_rate": 1.9375e-08, + "loss": 0.0051, + "step": 24950 + }, + { + "epoch": 11.43020594965675, + "grad_norm": 0.03731099143624306, + "learning_rate": 1.1562500000000002e-08, + "loss": 0.003, + "step": 24975 + }, + { + "epoch": 11.441647597254004, + "grad_norm": 1.1690711975097656, + "learning_rate": 3.75e-09, + "loss": 0.0043, + "step": 25000 + }, + { + "epoch": 11.441647597254004, + "eval_loss": 0.20993435382843018, + "eval_runtime": 5536.2723, + "eval_samples_per_second": 1.72, + "eval_steps_per_second": 0.215, + "eval_wer": 0.09534779548217828, + "step": 25000 + }, + { + "epoch": 11.441647597254004, + "step": 25000, + "total_flos": 4.081970563920691e+20, + "train_loss": 0.052786808650195596, + "train_runtime": 192279.8306, + "train_samples_per_second": 2.08, + "train_steps_per_second": 0.13 + } + ], + "logging_steps": 25, + "max_steps": 25000, + "num_input_tokens_seen": 0, + "num_train_epochs": 12, + "save_steps": 5000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 4.081970563920691e+20, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}