{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9999099150744839, "eval_steps": 500, "global_step": 7631, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 13.304811477661133, "learning_rate": 4.366812227074236e-08, "loss": 1.3583, "step": 1 }, { "epoch": 0.0, "grad_norm": 13.087164878845215, "learning_rate": 8.733624454148472e-08, "loss": 1.4085, "step": 2 }, { "epoch": 0.0, "grad_norm": 14.511214256286621, "learning_rate": 1.3100436681222707e-07, "loss": 1.3689, "step": 3 }, { "epoch": 0.0, "grad_norm": 14.423463821411133, "learning_rate": 1.7467248908296944e-07, "loss": 1.3941, "step": 4 }, { "epoch": 0.0, "grad_norm": 14.853147506713867, "learning_rate": 2.183406113537118e-07, "loss": 1.3823, "step": 5 }, { "epoch": 0.0, "grad_norm": 2.778461217880249, "learning_rate": 2.6200873362445414e-07, "loss": 1.2714, "step": 6 }, { "epoch": 0.0, "grad_norm": 16.974666595458984, "learning_rate": 3.056768558951965e-07, "loss": 1.3531, "step": 7 }, { "epoch": 0.0, "grad_norm": 13.01663589477539, "learning_rate": 3.4934497816593887e-07, "loss": 1.4299, "step": 8 }, { "epoch": 0.0, "grad_norm": 13.560871124267578, "learning_rate": 3.930131004366813e-07, "loss": 1.4018, "step": 9 }, { "epoch": 0.0, "grad_norm": 13.416844367980957, "learning_rate": 4.366812227074236e-07, "loss": 1.4584, "step": 10 }, { "epoch": 0.0, "grad_norm": 13.355825424194336, "learning_rate": 4.80349344978166e-07, "loss": 1.3404, "step": 11 }, { "epoch": 0.0, "grad_norm": 2.7600033283233643, "learning_rate": 5.240174672489083e-07, "loss": 1.2862, "step": 12 }, { "epoch": 0.0, "grad_norm": 2.891749858856201, "learning_rate": 5.676855895196507e-07, "loss": 1.2853, "step": 13 }, { "epoch": 0.0, "grad_norm": 12.469636917114258, "learning_rate": 6.11353711790393e-07, "loss": 1.3609, "step": 14 }, { "epoch": 0.0, "grad_norm": 11.10430908203125, "learning_rate": 6.550218340611354e-07, "loss": 1.399, "step": 15 }, { "epoch": 0.0, "grad_norm": 9.495098114013672, "learning_rate": 6.986899563318777e-07, "loss": 1.3412, "step": 16 }, { "epoch": 0.0, "grad_norm": 9.553077697753906, "learning_rate": 7.423580786026202e-07, "loss": 1.4116, "step": 17 }, { "epoch": 0.0, "grad_norm": 8.774883270263672, "learning_rate": 7.860262008733626e-07, "loss": 1.4215, "step": 18 }, { "epoch": 0.0, "grad_norm": 8.561419486999512, "learning_rate": 8.296943231441049e-07, "loss": 1.3365, "step": 19 }, { "epoch": 0.0, "grad_norm": 8.979242324829102, "learning_rate": 8.733624454148472e-07, "loss": 1.3371, "step": 20 }, { "epoch": 0.0, "grad_norm": 7.492669105529785, "learning_rate": 9.170305676855896e-07, "loss": 1.3081, "step": 21 }, { "epoch": 0.0, "grad_norm": 6.69644021987915, "learning_rate": 9.60698689956332e-07, "loss": 1.332, "step": 22 }, { "epoch": 0.0, "grad_norm": 6.4971394538879395, "learning_rate": 1.0043668122270742e-06, "loss": 1.326, "step": 23 }, { "epoch": 0.0, "grad_norm": 6.388171195983887, "learning_rate": 1.0480349344978166e-06, "loss": 1.3153, "step": 24 }, { "epoch": 0.0, "grad_norm": 2.559579372406006, "learning_rate": 1.091703056768559e-06, "loss": 1.2561, "step": 25 }, { "epoch": 0.0, "grad_norm": 6.097795009613037, "learning_rate": 1.1353711790393014e-06, "loss": 1.3206, "step": 26 }, { "epoch": 0.0, "grad_norm": 2.449272394180298, "learning_rate": 1.1790393013100437e-06, "loss": 1.2784, "step": 27 }, { "epoch": 0.0, "grad_norm": 5.802021503448486, "learning_rate": 1.222707423580786e-06, "loss": 1.3462, "step": 28 }, { "epoch": 0.0, "grad_norm": 5.36378288269043, "learning_rate": 1.2663755458515283e-06, "loss": 1.2855, "step": 29 }, { "epoch": 0.0, "grad_norm": 2.4155356884002686, "learning_rate": 1.3100436681222709e-06, "loss": 1.2978, "step": 30 }, { "epoch": 0.0, "grad_norm": 3.8745667934417725, "learning_rate": 1.3537117903930134e-06, "loss": 1.2834, "step": 31 }, { "epoch": 0.0, "grad_norm": 3.236354351043701, "learning_rate": 1.3973799126637555e-06, "loss": 1.2131, "step": 32 }, { "epoch": 0.0, "grad_norm": 3.5207133293151855, "learning_rate": 1.441048034934498e-06, "loss": 1.2493, "step": 33 }, { "epoch": 0.0, "grad_norm": 2.8835604190826416, "learning_rate": 1.4847161572052403e-06, "loss": 1.2107, "step": 34 }, { "epoch": 0.0, "grad_norm": 2.682253122329712, "learning_rate": 1.5283842794759826e-06, "loss": 1.2475, "step": 35 }, { "epoch": 0.0, "grad_norm": 3.009310245513916, "learning_rate": 1.5720524017467252e-06, "loss": 1.2478, "step": 36 }, { "epoch": 0.0, "grad_norm": 2.5312387943267822, "learning_rate": 1.6157205240174673e-06, "loss": 1.2272, "step": 37 }, { "epoch": 0.0, "grad_norm": 2.2449684143066406, "learning_rate": 1.6593886462882098e-06, "loss": 1.2504, "step": 38 }, { "epoch": 0.01, "grad_norm": 2.610213279724121, "learning_rate": 1.703056768558952e-06, "loss": 1.145, "step": 39 }, { "epoch": 0.01, "grad_norm": 2.628636598587036, "learning_rate": 1.7467248908296944e-06, "loss": 1.2302, "step": 40 }, { "epoch": 0.01, "grad_norm": 3.213498592376709, "learning_rate": 1.790393013100437e-06, "loss": 1.2693, "step": 41 }, { "epoch": 0.01, "grad_norm": 2.8643910884857178, "learning_rate": 1.8340611353711792e-06, "loss": 1.2073, "step": 42 }, { "epoch": 0.01, "grad_norm": 2.6910083293914795, "learning_rate": 1.8777292576419216e-06, "loss": 1.1566, "step": 43 }, { "epoch": 0.01, "grad_norm": 2.359659194946289, "learning_rate": 1.921397379912664e-06, "loss": 1.1937, "step": 44 }, { "epoch": 0.01, "grad_norm": 2.2783567905426025, "learning_rate": 1.965065502183406e-06, "loss": 1.1801, "step": 45 }, { "epoch": 0.01, "grad_norm": 2.343083381652832, "learning_rate": 2.0087336244541485e-06, "loss": 1.1887, "step": 46 }, { "epoch": 0.01, "grad_norm": 2.31093168258667, "learning_rate": 2.052401746724891e-06, "loss": 1.2036, "step": 47 }, { "epoch": 0.01, "grad_norm": 1.8689913749694824, "learning_rate": 2.096069868995633e-06, "loss": 1.1263, "step": 48 }, { "epoch": 0.01, "grad_norm": 2.5457167625427246, "learning_rate": 2.1397379912663756e-06, "loss": 1.2581, "step": 49 }, { "epoch": 0.01, "grad_norm": 2.393673896789551, "learning_rate": 2.183406113537118e-06, "loss": 1.2694, "step": 50 }, { "epoch": 0.01, "grad_norm": 1.98115074634552, "learning_rate": 2.2270742358078603e-06, "loss": 1.1327, "step": 51 }, { "epoch": 0.01, "grad_norm": 1.9964866638183594, "learning_rate": 2.270742358078603e-06, "loss": 1.1087, "step": 52 }, { "epoch": 0.01, "grad_norm": 2.034078598022461, "learning_rate": 2.3144104803493453e-06, "loss": 1.2636, "step": 53 }, { "epoch": 0.01, "grad_norm": 1.7923938035964966, "learning_rate": 2.3580786026200874e-06, "loss": 1.1473, "step": 54 }, { "epoch": 0.01, "grad_norm": 1.8316377401351929, "learning_rate": 2.40174672489083e-06, "loss": 1.1567, "step": 55 }, { "epoch": 0.01, "grad_norm": 1.9214555025100708, "learning_rate": 2.445414847161572e-06, "loss": 1.1324, "step": 56 }, { "epoch": 0.01, "grad_norm": 1.9386367797851562, "learning_rate": 2.4890829694323146e-06, "loss": 1.0922, "step": 57 }, { "epoch": 0.01, "grad_norm": 1.7853647470474243, "learning_rate": 2.5327510917030567e-06, "loss": 1.1615, "step": 58 }, { "epoch": 0.01, "grad_norm": 1.6762571334838867, "learning_rate": 2.576419213973799e-06, "loss": 1.1335, "step": 59 }, { "epoch": 0.01, "grad_norm": 1.9913884401321411, "learning_rate": 2.6200873362445417e-06, "loss": 1.0928, "step": 60 }, { "epoch": 0.01, "grad_norm": 1.8067705631256104, "learning_rate": 2.6637554585152842e-06, "loss": 1.1657, "step": 61 }, { "epoch": 0.01, "grad_norm": 1.5861470699310303, "learning_rate": 2.7074235807860268e-06, "loss": 1.1215, "step": 62 }, { "epoch": 0.01, "grad_norm": 1.502251148223877, "learning_rate": 2.7510917030567684e-06, "loss": 1.2568, "step": 63 }, { "epoch": 0.01, "grad_norm": 1.513132929801941, "learning_rate": 2.794759825327511e-06, "loss": 1.234, "step": 64 }, { "epoch": 0.01, "grad_norm": 1.6179192066192627, "learning_rate": 2.8384279475982535e-06, "loss": 1.1612, "step": 65 }, { "epoch": 0.01, "grad_norm": 1.8862794637680054, "learning_rate": 2.882096069868996e-06, "loss": 1.1059, "step": 66 }, { "epoch": 0.01, "grad_norm": 1.6959172487258911, "learning_rate": 2.9257641921397385e-06, "loss": 1.0608, "step": 67 }, { "epoch": 0.01, "grad_norm": 1.5873368978500366, "learning_rate": 2.9694323144104806e-06, "loss": 1.0896, "step": 68 }, { "epoch": 0.01, "grad_norm": 1.8314930200576782, "learning_rate": 3.0131004366812227e-06, "loss": 1.1112, "step": 69 }, { "epoch": 0.01, "grad_norm": 1.6843721866607666, "learning_rate": 3.0567685589519653e-06, "loss": 1.1016, "step": 70 }, { "epoch": 0.01, "grad_norm": 1.5256156921386719, "learning_rate": 3.100436681222708e-06, "loss": 1.0541, "step": 71 }, { "epoch": 0.01, "grad_norm": 1.6435916423797607, "learning_rate": 3.1441048034934503e-06, "loss": 1.1156, "step": 72 }, { "epoch": 0.01, "grad_norm": 1.6711663007736206, "learning_rate": 3.1877729257641924e-06, "loss": 1.0766, "step": 73 }, { "epoch": 0.01, "grad_norm": 3.3657314777374268, "learning_rate": 3.2314410480349345e-06, "loss": 1.102, "step": 74 }, { "epoch": 0.01, "grad_norm": 1.6692707538604736, "learning_rate": 3.275109170305677e-06, "loss": 1.0752, "step": 75 }, { "epoch": 0.01, "grad_norm": 1.5913656949996948, "learning_rate": 3.3187772925764196e-06, "loss": 1.1234, "step": 76 }, { "epoch": 0.01, "grad_norm": 1.0402421951293945, "learning_rate": 3.362445414847162e-06, "loss": 1.2329, "step": 77 }, { "epoch": 0.01, "grad_norm": 1.6331881284713745, "learning_rate": 3.406113537117904e-06, "loss": 1.0584, "step": 78 }, { "epoch": 0.01, "grad_norm": 0.9260515570640564, "learning_rate": 3.4497816593886467e-06, "loss": 1.2414, "step": 79 }, { "epoch": 0.01, "grad_norm": 0.8897286057472229, "learning_rate": 3.493449781659389e-06, "loss": 1.2293, "step": 80 }, { "epoch": 0.01, "grad_norm": 1.6475101709365845, "learning_rate": 3.5371179039301313e-06, "loss": 1.1182, "step": 81 }, { "epoch": 0.01, "grad_norm": 1.6854326725006104, "learning_rate": 3.580786026200874e-06, "loss": 1.0784, "step": 82 }, { "epoch": 0.01, "grad_norm": 1.5393686294555664, "learning_rate": 3.624454148471616e-06, "loss": 1.1113, "step": 83 }, { "epoch": 0.01, "grad_norm": 0.8464553952217102, "learning_rate": 3.6681222707423585e-06, "loss": 1.2281, "step": 84 }, { "epoch": 0.01, "grad_norm": 1.6644121408462524, "learning_rate": 3.7117903930131006e-06, "loss": 1.0415, "step": 85 }, { "epoch": 0.01, "grad_norm": 0.8101783990859985, "learning_rate": 3.755458515283843e-06, "loss": 1.213, "step": 86 }, { "epoch": 0.01, "grad_norm": 1.6632949113845825, "learning_rate": 3.7991266375545856e-06, "loss": 1.0876, "step": 87 }, { "epoch": 0.01, "grad_norm": 0.8269057869911194, "learning_rate": 3.842794759825328e-06, "loss": 1.2207, "step": 88 }, { "epoch": 0.01, "grad_norm": 0.7662131786346436, "learning_rate": 3.88646288209607e-06, "loss": 1.2051, "step": 89 }, { "epoch": 0.01, "grad_norm": 0.6866616606712341, "learning_rate": 3.930131004366812e-06, "loss": 1.2193, "step": 90 }, { "epoch": 0.01, "grad_norm": 0.6745291352272034, "learning_rate": 3.9737991266375545e-06, "loss": 1.2097, "step": 91 }, { "epoch": 0.01, "grad_norm": 0.6483508944511414, "learning_rate": 4.017467248908297e-06, "loss": 1.2232, "step": 92 }, { "epoch": 0.01, "grad_norm": 1.6614634990692139, "learning_rate": 4.0611353711790395e-06, "loss": 1.0544, "step": 93 }, { "epoch": 0.01, "grad_norm": 1.595900535583496, "learning_rate": 4.104803493449782e-06, "loss": 1.0578, "step": 94 }, { "epoch": 0.01, "grad_norm": 1.7429417371749878, "learning_rate": 4.1484716157205246e-06, "loss": 1.0448, "step": 95 }, { "epoch": 0.01, "grad_norm": 0.7087650299072266, "learning_rate": 4.192139737991266e-06, "loss": 1.2124, "step": 96 }, { "epoch": 0.01, "grad_norm": 1.6001737117767334, "learning_rate": 4.235807860262009e-06, "loss": 1.0349, "step": 97 }, { "epoch": 0.01, "grad_norm": 1.5752873420715332, "learning_rate": 4.279475982532751e-06, "loss": 1.0214, "step": 98 }, { "epoch": 0.01, "grad_norm": 0.7572908997535706, "learning_rate": 4.323144104803494e-06, "loss": 1.2117, "step": 99 }, { "epoch": 0.01, "grad_norm": 1.5496107339859009, "learning_rate": 4.366812227074236e-06, "loss": 1.0556, "step": 100 }, { "epoch": 0.01, "grad_norm": 1.4614073038101196, "learning_rate": 4.410480349344978e-06, "loss": 1.0029, "step": 101 }, { "epoch": 0.01, "grad_norm": 1.5224772691726685, "learning_rate": 4.4541484716157205e-06, "loss": 1.0215, "step": 102 }, { "epoch": 0.01, "grad_norm": 1.4642726182937622, "learning_rate": 4.497816593886463e-06, "loss": 1.0783, "step": 103 }, { "epoch": 0.01, "grad_norm": 1.6880898475646973, "learning_rate": 4.541484716157206e-06, "loss": 1.0709, "step": 104 }, { "epoch": 0.01, "grad_norm": 1.4050922393798828, "learning_rate": 4.585152838427948e-06, "loss": 1.0332, "step": 105 }, { "epoch": 0.01, "grad_norm": 0.810804009437561, "learning_rate": 4.628820960698691e-06, "loss": 1.1803, "step": 106 }, { "epoch": 0.01, "grad_norm": 2.9103569984436035, "learning_rate": 4.672489082969432e-06, "loss": 0.9966, "step": 107 }, { "epoch": 0.01, "grad_norm": 0.6876052021980286, "learning_rate": 4.716157205240175e-06, "loss": 1.2123, "step": 108 }, { "epoch": 0.01, "grad_norm": 1.461586594581604, "learning_rate": 4.759825327510917e-06, "loss": 1.0294, "step": 109 }, { "epoch": 0.01, "grad_norm": 1.562894344329834, "learning_rate": 4.80349344978166e-06, "loss": 1.0466, "step": 110 }, { "epoch": 0.01, "grad_norm": 1.6435930728912354, "learning_rate": 4.847161572052402e-06, "loss": 1.0512, "step": 111 }, { "epoch": 0.01, "grad_norm": 1.5058637857437134, "learning_rate": 4.890829694323144e-06, "loss": 1.0896, "step": 112 }, { "epoch": 0.01, "grad_norm": 1.5014902353286743, "learning_rate": 4.934497816593887e-06, "loss": 1.0426, "step": 113 }, { "epoch": 0.01, "grad_norm": 0.9436601996421814, "learning_rate": 4.978165938864629e-06, "loss": 1.2067, "step": 114 }, { "epoch": 0.02, "grad_norm": 0.7618218660354614, "learning_rate": 5.021834061135371e-06, "loss": 1.1894, "step": 115 }, { "epoch": 0.02, "grad_norm": 1.5726590156555176, "learning_rate": 5.065502183406113e-06, "loss": 0.9988, "step": 116 }, { "epoch": 0.02, "grad_norm": 1.7089123725891113, "learning_rate": 5.109170305676856e-06, "loss": 1.0178, "step": 117 }, { "epoch": 0.02, "grad_norm": 0.7760702967643738, "learning_rate": 5.152838427947598e-06, "loss": 1.2077, "step": 118 }, { "epoch": 0.02, "grad_norm": 1.4757777452468872, "learning_rate": 5.196506550218341e-06, "loss": 1.0626, "step": 119 }, { "epoch": 0.02, "grad_norm": 1.5556186437606812, "learning_rate": 5.2401746724890834e-06, "loss": 1.0365, "step": 120 }, { "epoch": 0.02, "grad_norm": 1.65530264377594, "learning_rate": 5.283842794759826e-06, "loss": 1.0402, "step": 121 }, { "epoch": 0.02, "grad_norm": 1.5536997318267822, "learning_rate": 5.3275109170305685e-06, "loss": 1.0107, "step": 122 }, { "epoch": 0.02, "grad_norm": 1.563592553138733, "learning_rate": 5.371179039301311e-06, "loss": 0.9905, "step": 123 }, { "epoch": 0.02, "grad_norm": 1.4085707664489746, "learning_rate": 5.4148471615720535e-06, "loss": 1.0525, "step": 124 }, { "epoch": 0.02, "grad_norm": 1.6952048540115356, "learning_rate": 5.458515283842796e-06, "loss": 1.0557, "step": 125 }, { "epoch": 0.02, "grad_norm": 0.9427299499511719, "learning_rate": 5.502183406113537e-06, "loss": 1.215, "step": 126 }, { "epoch": 0.02, "grad_norm": 1.5666266679763794, "learning_rate": 5.545851528384279e-06, "loss": 0.9897, "step": 127 }, { "epoch": 0.02, "grad_norm": 0.7353399991989136, "learning_rate": 5.589519650655022e-06, "loss": 1.1899, "step": 128 }, { "epoch": 0.02, "grad_norm": 1.560914397239685, "learning_rate": 5.6331877729257645e-06, "loss": 1.0941, "step": 129 }, { "epoch": 0.02, "grad_norm": 1.6984800100326538, "learning_rate": 5.676855895196507e-06, "loss": 1.0395, "step": 130 }, { "epoch": 0.02, "grad_norm": 0.9565629959106445, "learning_rate": 5.7205240174672495e-06, "loss": 1.1978, "step": 131 }, { "epoch": 0.02, "grad_norm": 0.8980270624160767, "learning_rate": 5.764192139737992e-06, "loss": 1.1994, "step": 132 }, { "epoch": 0.02, "grad_norm": 1.651293158531189, "learning_rate": 5.8078602620087346e-06, "loss": 1.0259, "step": 133 }, { "epoch": 0.02, "grad_norm": 0.6336477994918823, "learning_rate": 5.851528384279477e-06, "loss": 1.1976, "step": 134 }, { "epoch": 0.02, "grad_norm": 0.8076828122138977, "learning_rate": 5.89519650655022e-06, "loss": 1.2005, "step": 135 }, { "epoch": 0.02, "grad_norm": 1.5511376857757568, "learning_rate": 5.938864628820961e-06, "loss": 0.9869, "step": 136 }, { "epoch": 0.02, "grad_norm": 1.4748386144638062, "learning_rate": 5.982532751091703e-06, "loss": 1.0386, "step": 137 }, { "epoch": 0.02, "grad_norm": 0.8771867752075195, "learning_rate": 6.0262008733624455e-06, "loss": 1.2209, "step": 138 }, { "epoch": 0.02, "grad_norm": 1.6702499389648438, "learning_rate": 6.069868995633188e-06, "loss": 1.0339, "step": 139 }, { "epoch": 0.02, "grad_norm": 1.478677749633789, "learning_rate": 6.1135371179039305e-06, "loss": 1.029, "step": 140 }, { "epoch": 0.02, "grad_norm": 1.5998152494430542, "learning_rate": 6.157205240174673e-06, "loss": 1.0322, "step": 141 }, { "epoch": 0.02, "grad_norm": 1.3773895502090454, "learning_rate": 6.200873362445416e-06, "loss": 0.9776, "step": 142 }, { "epoch": 0.02, "grad_norm": 0.8842025995254517, "learning_rate": 6.244541484716158e-06, "loss": 1.2028, "step": 143 }, { "epoch": 0.02, "grad_norm": 0.7378107905387878, "learning_rate": 6.288209606986901e-06, "loss": 1.2018, "step": 144 }, { "epoch": 0.02, "grad_norm": 1.6874125003814697, "learning_rate": 6.331877729257643e-06, "loss": 0.9886, "step": 145 }, { "epoch": 0.02, "grad_norm": 1.620457649230957, "learning_rate": 6.375545851528385e-06, "loss": 0.9704, "step": 146 }, { "epoch": 0.02, "grad_norm": 1.5623376369476318, "learning_rate": 6.419213973799127e-06, "loss": 1.072, "step": 147 }, { "epoch": 0.02, "grad_norm": 1.4587401151657104, "learning_rate": 6.462882096069869e-06, "loss": 0.9875, "step": 148 }, { "epoch": 0.02, "grad_norm": 1.5273566246032715, "learning_rate": 6.5065502183406116e-06, "loss": 0.9749, "step": 149 }, { "epoch": 0.02, "grad_norm": 1.5816082954406738, "learning_rate": 6.550218340611354e-06, "loss": 0.9619, "step": 150 }, { "epoch": 0.02, "grad_norm": 1.505989909172058, "learning_rate": 6.593886462882097e-06, "loss": 1.0682, "step": 151 }, { "epoch": 0.02, "grad_norm": 1.4429112672805786, "learning_rate": 6.637554585152839e-06, "loss": 0.9825, "step": 152 }, { "epoch": 0.02, "grad_norm": 1.3323229551315308, "learning_rate": 6.681222707423582e-06, "loss": 1.2077, "step": 153 }, { "epoch": 0.02, "grad_norm": 0.9722976088523865, "learning_rate": 6.724890829694324e-06, "loss": 1.1747, "step": 154 }, { "epoch": 0.02, "grad_norm": 1.771681308746338, "learning_rate": 6.768558951965066e-06, "loss": 1.0381, "step": 155 }, { "epoch": 0.02, "grad_norm": 1.6915695667266846, "learning_rate": 6.812227074235808e-06, "loss": 1.0218, "step": 156 }, { "epoch": 0.02, "grad_norm": 1.6457403898239136, "learning_rate": 6.855895196506551e-06, "loss": 0.9891, "step": 157 }, { "epoch": 0.02, "grad_norm": 1.9012256860733032, "learning_rate": 6.8995633187772934e-06, "loss": 1.1657, "step": 158 }, { "epoch": 0.02, "grad_norm": 1.6828746795654297, "learning_rate": 6.943231441048035e-06, "loss": 1.0481, "step": 159 }, { "epoch": 0.02, "grad_norm": 1.5881805419921875, "learning_rate": 6.986899563318778e-06, "loss": 1.0033, "step": 160 }, { "epoch": 0.02, "grad_norm": 1.4730792045593262, "learning_rate": 7.03056768558952e-06, "loss": 1.0232, "step": 161 }, { "epoch": 0.02, "grad_norm": 1.3077380657196045, "learning_rate": 7.074235807860263e-06, "loss": 1.1809, "step": 162 }, { "epoch": 0.02, "grad_norm": 1.8247615098953247, "learning_rate": 7.117903930131005e-06, "loss": 0.9824, "step": 163 }, { "epoch": 0.02, "grad_norm": 2.3732621669769287, "learning_rate": 7.161572052401748e-06, "loss": 0.991, "step": 164 }, { "epoch": 0.02, "grad_norm": 1.5936084985733032, "learning_rate": 7.205240174672489e-06, "loss": 1.0119, "step": 165 }, { "epoch": 0.02, "grad_norm": 1.7444220781326294, "learning_rate": 7.248908296943232e-06, "loss": 1.0274, "step": 166 }, { "epoch": 0.02, "grad_norm": 1.5248379707336426, "learning_rate": 7.2925764192139745e-06, "loss": 0.963, "step": 167 }, { "epoch": 0.02, "grad_norm": 1.5232897996902466, "learning_rate": 7.336244541484717e-06, "loss": 0.98, "step": 168 }, { "epoch": 0.02, "grad_norm": 1.6525030136108398, "learning_rate": 7.3799126637554595e-06, "loss": 1.0068, "step": 169 }, { "epoch": 0.02, "grad_norm": 1.5497615337371826, "learning_rate": 7.423580786026201e-06, "loss": 1.0099, "step": 170 }, { "epoch": 0.02, "grad_norm": 1.6999753713607788, "learning_rate": 7.467248908296944e-06, "loss": 0.9776, "step": 171 }, { "epoch": 0.02, "grad_norm": 1.5410760641098022, "learning_rate": 7.510917030567686e-06, "loss": 1.003, "step": 172 }, { "epoch": 0.02, "grad_norm": 1.2797045707702637, "learning_rate": 7.554585152838429e-06, "loss": 1.1826, "step": 173 }, { "epoch": 0.02, "grad_norm": 1.0029257535934448, "learning_rate": 7.598253275109171e-06, "loss": 1.2077, "step": 174 }, { "epoch": 0.02, "grad_norm": 0.7270669341087341, "learning_rate": 7.641921397379913e-06, "loss": 1.2098, "step": 175 }, { "epoch": 0.02, "grad_norm": 1.7051515579223633, "learning_rate": 7.685589519650655e-06, "loss": 1.0284, "step": 176 }, { "epoch": 0.02, "grad_norm": 1.4038517475128174, "learning_rate": 7.729257641921398e-06, "loss": 1.1952, "step": 177 }, { "epoch": 0.02, "grad_norm": 1.6143279075622559, "learning_rate": 7.77292576419214e-06, "loss": 1.0552, "step": 178 }, { "epoch": 0.02, "grad_norm": 1.6958212852478027, "learning_rate": 7.816593886462883e-06, "loss": 1.0064, "step": 179 }, { "epoch": 0.02, "grad_norm": 1.3593459129333496, "learning_rate": 7.860262008733624e-06, "loss": 1.1588, "step": 180 }, { "epoch": 0.02, "grad_norm": 1.654365062713623, "learning_rate": 7.903930131004366e-06, "loss": 1.0109, "step": 181 }, { "epoch": 0.02, "grad_norm": 1.64669668674469, "learning_rate": 7.947598253275109e-06, "loss": 1.0333, "step": 182 }, { "epoch": 0.02, "grad_norm": 1.4429659843444824, "learning_rate": 7.991266375545851e-06, "loss": 0.9873, "step": 183 }, { "epoch": 0.02, "grad_norm": 1.5439213514328003, "learning_rate": 8.034934497816594e-06, "loss": 0.9583, "step": 184 }, { "epoch": 0.02, "grad_norm": 1.1317036151885986, "learning_rate": 8.078602620087337e-06, "loss": 1.1827, "step": 185 }, { "epoch": 0.02, "grad_norm": 0.8799869418144226, "learning_rate": 8.122270742358079e-06, "loss": 1.198, "step": 186 }, { "epoch": 0.02, "grad_norm": 0.723175048828125, "learning_rate": 8.165938864628822e-06, "loss": 1.1746, "step": 187 }, { "epoch": 0.02, "grad_norm": 0.8373928666114807, "learning_rate": 8.209606986899564e-06, "loss": 1.1646, "step": 188 }, { "epoch": 0.02, "grad_norm": 0.9782618880271912, "learning_rate": 8.253275109170307e-06, "loss": 1.1786, "step": 189 }, { "epoch": 0.02, "grad_norm": 2.0180017948150635, "learning_rate": 8.296943231441049e-06, "loss": 1.0337, "step": 190 }, { "epoch": 0.03, "grad_norm": 1.7367628812789917, "learning_rate": 8.34061135371179e-06, "loss": 1.0318, "step": 191 }, { "epoch": 0.03, "grad_norm": 1.5601131916046143, "learning_rate": 8.384279475982532e-06, "loss": 0.9767, "step": 192 }, { "epoch": 0.03, "grad_norm": 1.596856713294983, "learning_rate": 8.427947598253275e-06, "loss": 1.0048, "step": 193 }, { "epoch": 0.03, "grad_norm": 1.6807384490966797, "learning_rate": 8.471615720524018e-06, "loss": 0.9777, "step": 194 }, { "epoch": 0.03, "grad_norm": 1.6354954242706299, "learning_rate": 8.51528384279476e-06, "loss": 0.9927, "step": 195 }, { "epoch": 0.03, "grad_norm": 1.523680567741394, "learning_rate": 8.558951965065503e-06, "loss": 0.9108, "step": 196 }, { "epoch": 0.03, "grad_norm": 1.6078166961669922, "learning_rate": 8.602620087336245e-06, "loss": 0.9787, "step": 197 }, { "epoch": 0.03, "grad_norm": 1.5335301160812378, "learning_rate": 8.646288209606988e-06, "loss": 1.0147, "step": 198 }, { "epoch": 0.03, "grad_norm": 1.5344676971435547, "learning_rate": 8.68995633187773e-06, "loss": 0.9614, "step": 199 }, { "epoch": 0.03, "grad_norm": 1.6371797323226929, "learning_rate": 8.733624454148473e-06, "loss": 1.2076, "step": 200 }, { "epoch": 0.03, "grad_norm": 1.5231714248657227, "learning_rate": 8.777292576419215e-06, "loss": 0.9695, "step": 201 }, { "epoch": 0.03, "grad_norm": 1.5656837224960327, "learning_rate": 8.820960698689956e-06, "loss": 0.9084, "step": 202 }, { "epoch": 0.03, "grad_norm": 1.5318102836608887, "learning_rate": 8.864628820960699e-06, "loss": 0.9849, "step": 203 }, { "epoch": 0.03, "grad_norm": 0.860832691192627, "learning_rate": 8.908296943231441e-06, "loss": 1.1618, "step": 204 }, { "epoch": 0.03, "grad_norm": 1.9508543014526367, "learning_rate": 8.951965065502184e-06, "loss": 1.0034, "step": 205 }, { "epoch": 0.03, "grad_norm": 0.7524669766426086, "learning_rate": 8.995633187772926e-06, "loss": 1.175, "step": 206 }, { "epoch": 0.03, "grad_norm": 1.836098551750183, "learning_rate": 9.039301310043669e-06, "loss": 0.9854, "step": 207 }, { "epoch": 0.03, "grad_norm": 1.5345371961593628, "learning_rate": 9.082969432314411e-06, "loss": 0.9732, "step": 208 }, { "epoch": 0.03, "grad_norm": 1.4543159008026123, "learning_rate": 9.126637554585154e-06, "loss": 1.0012, "step": 209 }, { "epoch": 0.03, "grad_norm": 1.5667959451675415, "learning_rate": 9.170305676855896e-06, "loss": 0.9397, "step": 210 }, { "epoch": 0.03, "grad_norm": 1.5333147048950195, "learning_rate": 9.213973799126639e-06, "loss": 1.0074, "step": 211 }, { "epoch": 0.03, "grad_norm": 1.5981898307800293, "learning_rate": 9.257641921397381e-06, "loss": 1.0002, "step": 212 }, { "epoch": 0.03, "grad_norm": 1.9412317276000977, "learning_rate": 9.301310043668122e-06, "loss": 0.9788, "step": 213 }, { "epoch": 0.03, "grad_norm": 0.9766890406608582, "learning_rate": 9.344978165938865e-06, "loss": 1.1812, "step": 214 }, { "epoch": 0.03, "grad_norm": 1.7886097431182861, "learning_rate": 9.388646288209607e-06, "loss": 1.033, "step": 215 }, { "epoch": 0.03, "grad_norm": 0.7834906578063965, "learning_rate": 9.43231441048035e-06, "loss": 1.1693, "step": 216 }, { "epoch": 0.03, "grad_norm": 0.7616706490516663, "learning_rate": 9.475982532751092e-06, "loss": 1.1582, "step": 217 }, { "epoch": 0.03, "grad_norm": 1.5023159980773926, "learning_rate": 9.519650655021835e-06, "loss": 1.0164, "step": 218 }, { "epoch": 0.03, "grad_norm": 1.5782949924468994, "learning_rate": 9.563318777292577e-06, "loss": 0.9346, "step": 219 }, { "epoch": 0.03, "grad_norm": 1.5002120733261108, "learning_rate": 9.60698689956332e-06, "loss": 0.9796, "step": 220 }, { "epoch": 0.03, "grad_norm": 1.5034847259521484, "learning_rate": 9.650655021834062e-06, "loss": 0.9425, "step": 221 }, { "epoch": 0.03, "grad_norm": 1.5126230716705322, "learning_rate": 9.694323144104805e-06, "loss": 0.967, "step": 222 }, { "epoch": 0.03, "grad_norm": 1.492726445198059, "learning_rate": 9.737991266375547e-06, "loss": 0.9665, "step": 223 }, { "epoch": 0.03, "grad_norm": 1.1735299825668335, "learning_rate": 9.781659388646288e-06, "loss": 1.1734, "step": 224 }, { "epoch": 0.03, "grad_norm": 1.4979047775268555, "learning_rate": 9.82532751091703e-06, "loss": 0.9671, "step": 225 }, { "epoch": 0.03, "grad_norm": 1.5828872919082642, "learning_rate": 9.868995633187773e-06, "loss": 1.0051, "step": 226 }, { "epoch": 0.03, "grad_norm": 1.7658978700637817, "learning_rate": 9.912663755458516e-06, "loss": 1.0237, "step": 227 }, { "epoch": 0.03, "grad_norm": 0.9465129375457764, "learning_rate": 9.956331877729258e-06, "loss": 1.1678, "step": 228 }, { "epoch": 0.03, "grad_norm": 1.5850268602371216, "learning_rate": 1e-05, "loss": 0.8891, "step": 229 }, { "epoch": 0.03, "grad_norm": 1.5686501264572144, "learning_rate": 9.999999549658898e-06, "loss": 0.9337, "step": 230 }, { "epoch": 0.03, "grad_norm": 1.6235178709030151, "learning_rate": 9.999998198635674e-06, "loss": 0.9794, "step": 231 }, { "epoch": 0.03, "grad_norm": 0.9662335515022278, "learning_rate": 9.999995946930572e-06, "loss": 1.1738, "step": 232 }, { "epoch": 0.03, "grad_norm": 1.4978463649749756, "learning_rate": 9.999992794543995e-06, "loss": 0.992, "step": 233 }, { "epoch": 0.03, "grad_norm": 1.5580271482467651, "learning_rate": 9.999988741476512e-06, "loss": 0.8773, "step": 234 }, { "epoch": 0.03, "grad_norm": 1.581213116645813, "learning_rate": 9.999983787728855e-06, "loss": 1.0354, "step": 235 }, { "epoch": 0.03, "grad_norm": 1.485978603363037, "learning_rate": 9.999977933301912e-06, "loss": 0.9733, "step": 236 }, { "epoch": 0.03, "grad_norm": 1.0185579061508179, "learning_rate": 9.999971178196743e-06, "loss": 1.1816, "step": 237 }, { "epoch": 0.03, "grad_norm": 1.467269778251648, "learning_rate": 9.999963522414561e-06, "loss": 1.0128, "step": 238 }, { "epoch": 0.03, "grad_norm": 1.4987730979919434, "learning_rate": 9.999954965956747e-06, "loss": 0.9201, "step": 239 }, { "epoch": 0.03, "grad_norm": 1.4105852842330933, "learning_rate": 9.999945508824842e-06, "loss": 0.9594, "step": 240 }, { "epoch": 0.03, "grad_norm": 0.8526809215545654, "learning_rate": 9.999935151020548e-06, "loss": 1.1713, "step": 241 }, { "epoch": 0.03, "grad_norm": 1.6232208013534546, "learning_rate": 9.999923892545734e-06, "loss": 1.0042, "step": 242 }, { "epoch": 0.03, "grad_norm": 1.67000150680542, "learning_rate": 9.999911733402426e-06, "loss": 0.9825, "step": 243 }, { "epoch": 0.03, "grad_norm": 0.7694621086120605, "learning_rate": 9.999898673592812e-06, "loss": 1.1453, "step": 244 }, { "epoch": 0.03, "grad_norm": 1.594449758529663, "learning_rate": 9.999884713119248e-06, "loss": 0.9697, "step": 245 }, { "epoch": 0.03, "grad_norm": 1.566441297531128, "learning_rate": 9.99986985198425e-06, "loss": 0.9242, "step": 246 }, { "epoch": 0.03, "grad_norm": 1.4490538835525513, "learning_rate": 9.99985409019049e-06, "loss": 0.935, "step": 247 }, { "epoch": 0.03, "grad_norm": 0.9326693415641785, "learning_rate": 9.99983742774081e-06, "loss": 1.1595, "step": 248 }, { "epoch": 0.03, "grad_norm": 1.5367944240570068, "learning_rate": 9.999819864638214e-06, "loss": 1.0122, "step": 249 }, { "epoch": 0.03, "grad_norm": 1.550855278968811, "learning_rate": 9.999801400885862e-06, "loss": 0.9884, "step": 250 }, { "epoch": 0.03, "grad_norm": 1.526693344116211, "learning_rate": 9.999782036487082e-06, "loss": 1.0209, "step": 251 }, { "epoch": 0.03, "grad_norm": 1.1695066690444946, "learning_rate": 9.99976177144536e-06, "loss": 1.1589, "step": 252 }, { "epoch": 0.03, "grad_norm": 0.7873033285140991, "learning_rate": 9.99974060576435e-06, "loss": 1.1636, "step": 253 }, { "epoch": 0.03, "grad_norm": 0.8224154710769653, "learning_rate": 9.99971853944786e-06, "loss": 1.1742, "step": 254 }, { "epoch": 0.03, "grad_norm": 0.985554575920105, "learning_rate": 9.99969557249987e-06, "loss": 1.1709, "step": 255 }, { "epoch": 0.03, "grad_norm": 1.663857102394104, "learning_rate": 9.999671704924514e-06, "loss": 0.9167, "step": 256 }, { "epoch": 0.03, "grad_norm": 1.831000566482544, "learning_rate": 9.999646936726094e-06, "loss": 0.9101, "step": 257 }, { "epoch": 0.03, "grad_norm": 1.4246646165847778, "learning_rate": 9.999621267909068e-06, "loss": 0.9646, "step": 258 }, { "epoch": 0.03, "grad_norm": 1.5065090656280518, "learning_rate": 9.999594698478063e-06, "loss": 0.9995, "step": 259 }, { "epoch": 0.03, "grad_norm": 1.6437197923660278, "learning_rate": 9.999567228437864e-06, "loss": 0.955, "step": 260 }, { "epoch": 0.03, "grad_norm": 1.487500786781311, "learning_rate": 9.999538857793419e-06, "loss": 1.012, "step": 261 }, { "epoch": 0.03, "grad_norm": 1.858747124671936, "learning_rate": 9.999509586549837e-06, "loss": 0.9685, "step": 262 }, { "epoch": 0.03, "grad_norm": 1.6007310152053833, "learning_rate": 9.999479414712395e-06, "loss": 1.0056, "step": 263 }, { "epoch": 0.03, "grad_norm": 1.4403270483016968, "learning_rate": 9.999448342286526e-06, "loss": 0.9645, "step": 264 }, { "epoch": 0.03, "grad_norm": 1.6856467723846436, "learning_rate": 9.999416369277825e-06, "loss": 0.9902, "step": 265 }, { "epoch": 0.03, "grad_norm": 2.0750908851623535, "learning_rate": 9.999383495692056e-06, "loss": 0.9302, "step": 266 }, { "epoch": 0.03, "grad_norm": 1.5402804613113403, "learning_rate": 9.999349721535136e-06, "loss": 0.9524, "step": 267 }, { "epoch": 0.04, "grad_norm": 1.5318063497543335, "learning_rate": 9.999315046813153e-06, "loss": 1.0006, "step": 268 }, { "epoch": 0.04, "grad_norm": 1.5211050510406494, "learning_rate": 9.99927947153235e-06, "loss": 0.9085, "step": 269 }, { "epoch": 0.04, "grad_norm": 1.6457608938217163, "learning_rate": 9.99924299569914e-06, "loss": 0.9701, "step": 270 }, { "epoch": 0.04, "grad_norm": 1.5001360177993774, "learning_rate": 9.999205619320087e-06, "loss": 0.9196, "step": 271 }, { "epoch": 0.04, "grad_norm": 1.4735041856765747, "learning_rate": 9.99916734240193e-06, "loss": 1.1713, "step": 272 }, { "epoch": 0.04, "grad_norm": 1.0693626403808594, "learning_rate": 9.999128164951559e-06, "loss": 1.1818, "step": 273 }, { "epoch": 0.04, "grad_norm": 0.7612532377243042, "learning_rate": 9.999088086976034e-06, "loss": 1.1457, "step": 274 }, { "epoch": 0.04, "grad_norm": 1.1819087266921997, "learning_rate": 9.999047108482576e-06, "loss": 1.1722, "step": 275 }, { "epoch": 0.04, "grad_norm": 1.7210181951522827, "learning_rate": 9.999005229478561e-06, "loss": 0.9772, "step": 276 }, { "epoch": 0.04, "grad_norm": 1.6040396690368652, "learning_rate": 9.998962449971542e-06, "loss": 0.9654, "step": 277 }, { "epoch": 0.04, "grad_norm": 1.2411504983901978, "learning_rate": 9.998918769969216e-06, "loss": 1.1556, "step": 278 }, { "epoch": 0.04, "grad_norm": 1.4973070621490479, "learning_rate": 9.998874189479457e-06, "loss": 0.9852, "step": 279 }, { "epoch": 0.04, "grad_norm": 1.5554070472717285, "learning_rate": 9.998828708510295e-06, "loss": 0.9414, "step": 280 }, { "epoch": 0.04, "grad_norm": 1.043073296546936, "learning_rate": 9.99878232706992e-06, "loss": 1.1394, "step": 281 }, { "epoch": 0.04, "grad_norm": 1.7199949026107788, "learning_rate": 9.99873504516669e-06, "loss": 0.9496, "step": 282 }, { "epoch": 0.04, "grad_norm": 1.4713066816329956, "learning_rate": 9.998686862809119e-06, "loss": 0.9727, "step": 283 }, { "epoch": 0.04, "grad_norm": 0.8652296662330627, "learning_rate": 9.998637780005889e-06, "loss": 1.1811, "step": 284 }, { "epoch": 0.04, "grad_norm": 1.4764145612716675, "learning_rate": 9.998587796765842e-06, "loss": 0.9483, "step": 285 }, { "epoch": 0.04, "grad_norm": 1.8786189556121826, "learning_rate": 9.99853691309798e-06, "loss": 1.0122, "step": 286 }, { "epoch": 0.04, "grad_norm": 1.4919754266738892, "learning_rate": 9.998485129011468e-06, "loss": 0.9487, "step": 287 }, { "epoch": 0.04, "grad_norm": 1.482703447341919, "learning_rate": 9.998432444515638e-06, "loss": 1.0405, "step": 288 }, { "epoch": 0.04, "grad_norm": 1.3549461364746094, "learning_rate": 9.998378859619978e-06, "loss": 0.9446, "step": 289 }, { "epoch": 0.04, "grad_norm": 1.4813355207443237, "learning_rate": 9.99832437433414e-06, "loss": 0.9576, "step": 290 }, { "epoch": 0.04, "grad_norm": 1.5404571294784546, "learning_rate": 9.99826898866794e-06, "loss": 0.9753, "step": 291 }, { "epoch": 0.04, "grad_norm": 1.4476451873779297, "learning_rate": 9.998212702631355e-06, "loss": 0.9693, "step": 292 }, { "epoch": 0.04, "grad_norm": 1.3981077671051025, "learning_rate": 9.998155516234523e-06, "loss": 0.9312, "step": 293 }, { "epoch": 0.04, "grad_norm": 1.7836053371429443, "learning_rate": 9.998097429487748e-06, "loss": 0.9281, "step": 294 }, { "epoch": 0.04, "grad_norm": 1.3860063552856445, "learning_rate": 9.99803844240149e-06, "loss": 0.8845, "step": 295 }, { "epoch": 0.04, "grad_norm": 1.4524714946746826, "learning_rate": 9.997978554986376e-06, "loss": 0.9518, "step": 296 }, { "epoch": 0.04, "grad_norm": 1.4673844575881958, "learning_rate": 9.997917767253195e-06, "loss": 0.949, "step": 297 }, { "epoch": 0.04, "grad_norm": 1.2295887470245361, "learning_rate": 9.997856079212896e-06, "loss": 1.1514, "step": 298 }, { "epoch": 0.04, "grad_norm": 0.9245722889900208, "learning_rate": 9.997793490876593e-06, "loss": 1.1593, "step": 299 }, { "epoch": 0.04, "grad_norm": 1.5969908237457275, "learning_rate": 9.997730002255559e-06, "loss": 0.9996, "step": 300 }, { "epoch": 0.04, "grad_norm": 1.6836344003677368, "learning_rate": 9.99766561336123e-06, "loss": 1.0058, "step": 301 }, { "epoch": 0.04, "grad_norm": 1.602613091468811, "learning_rate": 9.997600324205204e-06, "loss": 0.9708, "step": 302 }, { "epoch": 0.04, "grad_norm": 1.5415093898773193, "learning_rate": 9.997534134799244e-06, "loss": 0.9575, "step": 303 }, { "epoch": 0.04, "grad_norm": 2.4839026927948, "learning_rate": 9.997467045155276e-06, "loss": 1.1673, "step": 304 }, { "epoch": 0.04, "grad_norm": 2.0586516857147217, "learning_rate": 9.997399055285379e-06, "loss": 1.1804, "step": 305 }, { "epoch": 0.04, "grad_norm": 1.673376202583313, "learning_rate": 9.997330165201805e-06, "loss": 0.9305, "step": 306 }, { "epoch": 0.04, "grad_norm": 1.727352499961853, "learning_rate": 9.99726037491696e-06, "loss": 0.9346, "step": 307 }, { "epoch": 0.04, "grad_norm": 0.8754662275314331, "learning_rate": 9.997189684443418e-06, "loss": 1.1804, "step": 308 }, { "epoch": 0.04, "grad_norm": 0.9189266562461853, "learning_rate": 9.997118093793912e-06, "loss": 1.1737, "step": 309 }, { "epoch": 0.04, "grad_norm": 0.9685351252555847, "learning_rate": 9.99704560298134e-06, "loss": 1.1709, "step": 310 }, { "epoch": 0.04, "grad_norm": 0.8709153532981873, "learning_rate": 9.996972212018758e-06, "loss": 1.186, "step": 311 }, { "epoch": 0.04, "grad_norm": 0.8711608648300171, "learning_rate": 9.996897920919388e-06, "loss": 1.1593, "step": 312 }, { "epoch": 0.04, "grad_norm": 1.723369836807251, "learning_rate": 9.996822729696613e-06, "loss": 0.8861, "step": 313 }, { "epoch": 0.04, "grad_norm": 1.5541276931762695, "learning_rate": 9.996746638363976e-06, "loss": 0.9378, "step": 314 }, { "epoch": 0.04, "grad_norm": 1.5653910636901855, "learning_rate": 9.996669646935185e-06, "loss": 0.8906, "step": 315 }, { "epoch": 0.04, "grad_norm": 4.508587837219238, "learning_rate": 9.996591755424105e-06, "loss": 1.0205, "step": 316 }, { "epoch": 0.04, "grad_norm": 1.6818132400512695, "learning_rate": 9.996512963844774e-06, "loss": 0.9658, "step": 317 }, { "epoch": 0.04, "grad_norm": 1.7247850894927979, "learning_rate": 9.996433272211381e-06, "loss": 0.9428, "step": 318 }, { "epoch": 0.04, "grad_norm": 1.9894436597824097, "learning_rate": 9.996352680538281e-06, "loss": 0.9124, "step": 319 }, { "epoch": 0.04, "grad_norm": 1.7595661878585815, "learning_rate": 9.996271188839994e-06, "loss": 0.9599, "step": 320 }, { "epoch": 0.04, "grad_norm": 1.8685007095336914, "learning_rate": 9.996188797131197e-06, "loss": 1.1779, "step": 321 }, { "epoch": 0.04, "grad_norm": 1.528751015663147, "learning_rate": 9.996105505426733e-06, "loss": 0.9492, "step": 322 }, { "epoch": 0.04, "grad_norm": 1.5063788890838623, "learning_rate": 9.996021313741606e-06, "loss": 0.979, "step": 323 }, { "epoch": 0.04, "grad_norm": 1.5781878232955933, "learning_rate": 9.995936222090982e-06, "loss": 0.914, "step": 324 }, { "epoch": 0.04, "grad_norm": 1.5012404918670654, "learning_rate": 9.995850230490188e-06, "loss": 0.9438, "step": 325 }, { "epoch": 0.04, "grad_norm": 1.5318392515182495, "learning_rate": 9.995763338954715e-06, "loss": 0.9231, "step": 326 }, { "epoch": 0.04, "grad_norm": 1.4163219928741455, "learning_rate": 9.995675547500216e-06, "loss": 0.9084, "step": 327 }, { "epoch": 0.04, "grad_norm": 0.950498640537262, "learning_rate": 9.995586856142507e-06, "loss": 1.1567, "step": 328 }, { "epoch": 0.04, "grad_norm": 1.5478583574295044, "learning_rate": 9.99549726489756e-06, "loss": 0.9254, "step": 329 }, { "epoch": 0.04, "grad_norm": 1.5884135961532593, "learning_rate": 9.995406773781518e-06, "loss": 0.9135, "step": 330 }, { "epoch": 0.04, "grad_norm": 1.365029215812683, "learning_rate": 9.995315382810677e-06, "loss": 0.9097, "step": 331 }, { "epoch": 0.04, "grad_norm": 1.0175745487213135, "learning_rate": 9.995223092001505e-06, "loss": 1.1583, "step": 332 }, { "epoch": 0.04, "grad_norm": 1.865436315536499, "learning_rate": 9.995129901370624e-06, "loss": 0.8634, "step": 333 }, { "epoch": 0.04, "grad_norm": 1.6106716394424438, "learning_rate": 9.995035810934821e-06, "loss": 0.9808, "step": 334 }, { "epoch": 0.04, "grad_norm": 0.7579584121704102, "learning_rate": 9.994940820711045e-06, "loss": 1.1363, "step": 335 }, { "epoch": 0.04, "grad_norm": 0.7074031233787537, "learning_rate": 9.99484493071641e-06, "loss": 1.152, "step": 336 }, { "epoch": 0.04, "grad_norm": 1.4812480211257935, "learning_rate": 9.994748140968187e-06, "loss": 0.9249, "step": 337 }, { "epoch": 0.04, "grad_norm": 1.520470380783081, "learning_rate": 9.99465045148381e-06, "loss": 0.9215, "step": 338 }, { "epoch": 0.04, "grad_norm": 1.5297267436981201, "learning_rate": 9.994551862280879e-06, "loss": 1.0324, "step": 339 }, { "epoch": 0.04, "grad_norm": 1.390197992324829, "learning_rate": 9.994452373377152e-06, "loss": 0.9262, "step": 340 }, { "epoch": 0.04, "grad_norm": 1.1602429151535034, "learning_rate": 9.994351984790552e-06, "loss": 1.1506, "step": 341 }, { "epoch": 0.04, "grad_norm": 0.8836959600448608, "learning_rate": 9.994250696539162e-06, "loss": 1.1655, "step": 342 }, { "epoch": 0.04, "grad_norm": 1.466088056564331, "learning_rate": 9.994148508641228e-06, "loss": 0.8691, "step": 343 }, { "epoch": 0.05, "grad_norm": 1.4872745275497437, "learning_rate": 9.994045421115155e-06, "loss": 0.9483, "step": 344 }, { "epoch": 0.05, "grad_norm": 1.7123887538909912, "learning_rate": 9.993941433979516e-06, "loss": 0.9594, "step": 345 }, { "epoch": 0.05, "grad_norm": 1.4231785535812378, "learning_rate": 9.993836547253041e-06, "loss": 1.1827, "step": 346 }, { "epoch": 0.05, "grad_norm": 1.2433327436447144, "learning_rate": 9.993730760954628e-06, "loss": 1.1477, "step": 347 }, { "epoch": 0.05, "grad_norm": 1.5125154256820679, "learning_rate": 9.993624075103328e-06, "loss": 1.0043, "step": 348 }, { "epoch": 0.05, "grad_norm": 1.771949052810669, "learning_rate": 9.99351648971836e-06, "loss": 0.9686, "step": 349 }, { "epoch": 0.05, "grad_norm": 1.3246914148330688, "learning_rate": 9.993408004819104e-06, "loss": 1.1367, "step": 350 }, { "epoch": 0.05, "grad_norm": 1.4710237979888916, "learning_rate": 9.993298620425106e-06, "loss": 0.9827, "step": 351 }, { "epoch": 0.05, "grad_norm": 1.540903925895691, "learning_rate": 9.993188336556065e-06, "loss": 0.8498, "step": 352 }, { "epoch": 0.05, "grad_norm": 1.129909873008728, "learning_rate": 9.99307715323185e-06, "loss": 1.1432, "step": 353 }, { "epoch": 0.05, "grad_norm": 1.7469611167907715, "learning_rate": 9.992965070472487e-06, "loss": 0.9341, "step": 354 }, { "epoch": 0.05, "grad_norm": 1.4255008697509766, "learning_rate": 9.992852088298171e-06, "loss": 0.8954, "step": 355 }, { "epoch": 0.05, "grad_norm": 1.5338915586471558, "learning_rate": 9.992738206729247e-06, "loss": 0.9102, "step": 356 }, { "epoch": 0.05, "grad_norm": 1.4899808168411255, "learning_rate": 9.992623425786235e-06, "loss": 0.942, "step": 357 }, { "epoch": 0.05, "grad_norm": 1.1174592971801758, "learning_rate": 9.992507745489808e-06, "loss": 1.1574, "step": 358 }, { "epoch": 0.05, "grad_norm": 1.4698232412338257, "learning_rate": 9.992391165860806e-06, "loss": 0.9327, "step": 359 }, { "epoch": 0.05, "grad_norm": 1.5327694416046143, "learning_rate": 9.992273686920229e-06, "loss": 0.9772, "step": 360 }, { "epoch": 0.05, "grad_norm": 1.564378023147583, "learning_rate": 9.992155308689239e-06, "loss": 0.8787, "step": 361 }, { "epoch": 0.05, "grad_norm": 1.4940303564071655, "learning_rate": 9.992036031189161e-06, "loss": 0.9205, "step": 362 }, { "epoch": 0.05, "grad_norm": 1.6388412714004517, "learning_rate": 9.991915854441478e-06, "loss": 0.9049, "step": 363 }, { "epoch": 0.05, "grad_norm": 1.4785562753677368, "learning_rate": 9.991794778467843e-06, "loss": 0.9362, "step": 364 }, { "epoch": 0.05, "grad_norm": 1.631843090057373, "learning_rate": 9.991672803290061e-06, "loss": 0.9161, "step": 365 }, { "epoch": 0.05, "grad_norm": 1.1704261302947998, "learning_rate": 9.991549928930111e-06, "loss": 1.1699, "step": 366 }, { "epoch": 0.05, "grad_norm": 1.8831162452697754, "learning_rate": 9.99142615541012e-06, "loss": 0.9026, "step": 367 }, { "epoch": 0.05, "grad_norm": 1.4302266836166382, "learning_rate": 9.99130148275239e-06, "loss": 0.9428, "step": 368 }, { "epoch": 0.05, "grad_norm": 1.4927130937576294, "learning_rate": 9.991175910979374e-06, "loss": 0.9486, "step": 369 }, { "epoch": 0.05, "grad_norm": 1.5271186828613281, "learning_rate": 9.991049440113693e-06, "loss": 0.9265, "step": 370 }, { "epoch": 0.05, "grad_norm": 1.5548104047775269, "learning_rate": 9.990922070178133e-06, "loss": 0.9347, "step": 371 }, { "epoch": 0.05, "grad_norm": 1.5833098888397217, "learning_rate": 9.990793801195636e-06, "loss": 0.9629, "step": 372 }, { "epoch": 0.05, "grad_norm": 1.2764489650726318, "learning_rate": 9.990664633189306e-06, "loss": 1.1384, "step": 373 }, { "epoch": 0.05, "grad_norm": 1.587994933128357, "learning_rate": 9.99053456618241e-06, "loss": 0.8956, "step": 374 }, { "epoch": 0.05, "grad_norm": 1.4944427013397217, "learning_rate": 9.990403600198383e-06, "loss": 0.9374, "step": 375 }, { "epoch": 0.05, "grad_norm": 1.8357075452804565, "learning_rate": 9.990271735260813e-06, "loss": 0.9042, "step": 376 }, { "epoch": 0.05, "grad_norm": 1.4558972120285034, "learning_rate": 9.990138971393456e-06, "loss": 0.8765, "step": 377 }, { "epoch": 0.05, "grad_norm": 0.996590256690979, "learning_rate": 9.990005308620223e-06, "loss": 1.1612, "step": 378 }, { "epoch": 0.05, "grad_norm": 1.5949198007583618, "learning_rate": 9.989870746965195e-06, "loss": 0.9797, "step": 379 }, { "epoch": 0.05, "grad_norm": 1.502863883972168, "learning_rate": 9.989735286452613e-06, "loss": 0.918, "step": 380 }, { "epoch": 0.05, "grad_norm": 1.491126298904419, "learning_rate": 9.989598927106874e-06, "loss": 0.9372, "step": 381 }, { "epoch": 0.05, "grad_norm": 1.3910324573516846, "learning_rate": 9.989461668952545e-06, "loss": 0.9229, "step": 382 }, { "epoch": 0.05, "grad_norm": 1.5401750802993774, "learning_rate": 9.989323512014351e-06, "loss": 0.9933, "step": 383 }, { "epoch": 0.05, "grad_norm": 1.6283583641052246, "learning_rate": 9.989184456317176e-06, "loss": 0.9373, "step": 384 }, { "epoch": 0.05, "grad_norm": 1.462511420249939, "learning_rate": 9.989044501886072e-06, "loss": 0.8593, "step": 385 }, { "epoch": 0.05, "grad_norm": 1.472151756286621, "learning_rate": 9.98890364874625e-06, "loss": 0.9667, "step": 386 }, { "epoch": 0.05, "grad_norm": 1.4696073532104492, "learning_rate": 9.988761896923078e-06, "loss": 0.9195, "step": 387 }, { "epoch": 0.05, "grad_norm": 1.5488590002059937, "learning_rate": 9.988619246442098e-06, "loss": 0.9655, "step": 388 }, { "epoch": 0.05, "grad_norm": 1.6673967838287354, "learning_rate": 9.988475697328999e-06, "loss": 0.928, "step": 389 }, { "epoch": 0.05, "grad_norm": 1.3978501558303833, "learning_rate": 9.988331249609646e-06, "loss": 0.9342, "step": 390 }, { "epoch": 0.05, "grad_norm": 1.3128303289413452, "learning_rate": 9.988185903310058e-06, "loss": 1.1496, "step": 391 }, { "epoch": 0.05, "grad_norm": 1.6885484457015991, "learning_rate": 9.988039658456411e-06, "loss": 0.9127, "step": 392 }, { "epoch": 0.05, "grad_norm": 1.5293642282485962, "learning_rate": 9.987892515075058e-06, "loss": 0.9202, "step": 393 }, { "epoch": 0.05, "grad_norm": 1.4735708236694336, "learning_rate": 9.9877444731925e-06, "loss": 0.9633, "step": 394 }, { "epoch": 0.05, "grad_norm": 1.4691985845565796, "learning_rate": 9.987595532835403e-06, "loss": 0.9012, "step": 395 }, { "epoch": 0.05, "grad_norm": 1.4186545610427856, "learning_rate": 9.987445694030601e-06, "loss": 0.9353, "step": 396 }, { "epoch": 0.05, "grad_norm": 1.6280784606933594, "learning_rate": 9.987294956805082e-06, "loss": 0.9758, "step": 397 }, { "epoch": 0.05, "grad_norm": 1.4731460809707642, "learning_rate": 9.987143321186002e-06, "loss": 1.1262, "step": 398 }, { "epoch": 0.05, "grad_norm": 1.4722166061401367, "learning_rate": 9.986990787200675e-06, "loss": 0.9472, "step": 399 }, { "epoch": 0.05, "grad_norm": 1.4632469415664673, "learning_rate": 9.986837354876578e-06, "loss": 0.9415, "step": 400 }, { "epoch": 0.05, "grad_norm": 1.6286556720733643, "learning_rate": 9.986683024241348e-06, "loss": 0.9395, "step": 401 }, { "epoch": 0.05, "grad_norm": 1.5162997245788574, "learning_rate": 9.986527795322787e-06, "loss": 0.9255, "step": 402 }, { "epoch": 0.05, "grad_norm": 1.4052927494049072, "learning_rate": 9.986371668148858e-06, "loss": 0.9403, "step": 403 }, { "epoch": 0.05, "grad_norm": 1.0754902362823486, "learning_rate": 9.986214642747685e-06, "loss": 1.1558, "step": 404 }, { "epoch": 0.05, "grad_norm": 1.4700806140899658, "learning_rate": 9.986056719147553e-06, "loss": 0.9742, "step": 405 }, { "epoch": 0.05, "grad_norm": 0.7152315378189087, "learning_rate": 9.98589789737691e-06, "loss": 1.1353, "step": 406 }, { "epoch": 0.05, "grad_norm": 1.5968958139419556, "learning_rate": 9.985738177464367e-06, "loss": 0.9202, "step": 407 }, { "epoch": 0.05, "grad_norm": 1.4840434789657593, "learning_rate": 9.985577559438694e-06, "loss": 0.8817, "step": 408 }, { "epoch": 0.05, "grad_norm": 1.0782058238983154, "learning_rate": 9.985416043328823e-06, "loss": 1.156, "step": 409 }, { "epoch": 0.05, "grad_norm": 1.4170235395431519, "learning_rate": 9.985253629163853e-06, "loss": 0.9715, "step": 410 }, { "epoch": 0.05, "grad_norm": 1.5115629434585571, "learning_rate": 9.985090316973035e-06, "loss": 0.922, "step": 411 }, { "epoch": 0.05, "grad_norm": 1.6028817892074585, "learning_rate": 9.984926106785795e-06, "loss": 0.9833, "step": 412 }, { "epoch": 0.05, "grad_norm": 1.5558334589004517, "learning_rate": 9.984760998631706e-06, "loss": 0.9736, "step": 413 }, { "epoch": 0.05, "grad_norm": 1.4095706939697266, "learning_rate": 9.984594992540513e-06, "loss": 0.8889, "step": 414 }, { "epoch": 0.05, "grad_norm": 1.4545128345489502, "learning_rate": 9.984428088542121e-06, "loss": 0.8975, "step": 415 }, { "epoch": 0.05, "grad_norm": 1.500214695930481, "learning_rate": 9.984260286666592e-06, "loss": 0.9097, "step": 416 }, { "epoch": 0.05, "grad_norm": 1.4830856323242188, "learning_rate": 9.984091586944158e-06, "loss": 0.8669, "step": 417 }, { "epoch": 0.05, "grad_norm": 1.7060056924819946, "learning_rate": 9.983921989405203e-06, "loss": 1.1517, "step": 418 }, { "epoch": 0.05, "grad_norm": 1.5446199178695679, "learning_rate": 9.98375149408028e-06, "loss": 0.8732, "step": 419 }, { "epoch": 0.06, "grad_norm": 1.526822805404663, "learning_rate": 9.983580101000103e-06, "loss": 0.9193, "step": 420 }, { "epoch": 0.06, "grad_norm": 0.7554365992546082, "learning_rate": 9.983407810195544e-06, "loss": 1.1289, "step": 421 }, { "epoch": 0.06, "grad_norm": 1.5259208679199219, "learning_rate": 9.983234621697639e-06, "loss": 0.9566, "step": 422 }, { "epoch": 0.06, "grad_norm": 1.3889814615249634, "learning_rate": 9.983060535537587e-06, "loss": 0.9081, "step": 423 }, { "epoch": 0.06, "grad_norm": 1.345273494720459, "learning_rate": 9.982885551746745e-06, "loss": 0.8948, "step": 424 }, { "epoch": 0.06, "grad_norm": 1.3663051128387451, "learning_rate": 9.982709670356636e-06, "loss": 0.9351, "step": 425 }, { "epoch": 0.06, "grad_norm": 1.5985093116760254, "learning_rate": 9.982532891398941e-06, "loss": 0.9187, "step": 426 }, { "epoch": 0.06, "grad_norm": 1.215554118156433, "learning_rate": 9.982355214905506e-06, "loss": 1.1368, "step": 427 }, { "epoch": 0.06, "grad_norm": 1.5529940128326416, "learning_rate": 9.982176640908337e-06, "loss": 0.9334, "step": 428 }, { "epoch": 0.06, "grad_norm": 0.8110408782958984, "learning_rate": 9.981997169439599e-06, "loss": 1.1472, "step": 429 }, { "epoch": 0.06, "grad_norm": 1.4708997011184692, "learning_rate": 9.981816800531623e-06, "loss": 0.9748, "step": 430 }, { "epoch": 0.06, "grad_norm": 1.62168550491333, "learning_rate": 9.981635534216902e-06, "loss": 0.8913, "step": 431 }, { "epoch": 0.06, "grad_norm": 1.4584624767303467, "learning_rate": 9.981453370528085e-06, "loss": 0.9969, "step": 432 }, { "epoch": 0.06, "grad_norm": 1.3661155700683594, "learning_rate": 9.98127030949799e-06, "loss": 0.8957, "step": 433 }, { "epoch": 0.06, "grad_norm": 1.5104775428771973, "learning_rate": 9.98108635115959e-06, "loss": 1.1582, "step": 434 }, { "epoch": 0.06, "grad_norm": 1.5411720275878906, "learning_rate": 9.980901495546025e-06, "loss": 0.957, "step": 435 }, { "epoch": 0.06, "grad_norm": 1.4284855127334595, "learning_rate": 9.980715742690592e-06, "loss": 0.902, "step": 436 }, { "epoch": 0.06, "grad_norm": 1.3668948411941528, "learning_rate": 9.980529092626754e-06, "loss": 0.9019, "step": 437 }, { "epoch": 0.06, "grad_norm": 1.7874689102172852, "learning_rate": 9.980341545388132e-06, "loss": 0.8988, "step": 438 }, { "epoch": 0.06, "grad_norm": 0.9790839552879333, "learning_rate": 9.980153101008509e-06, "loss": 1.1573, "step": 439 }, { "epoch": 0.06, "grad_norm": 1.5672273635864258, "learning_rate": 9.979963759521833e-06, "loss": 0.9774, "step": 440 }, { "epoch": 0.06, "grad_norm": 1.5062142610549927, "learning_rate": 9.979773520962212e-06, "loss": 0.9502, "step": 441 }, { "epoch": 0.06, "grad_norm": 3.0093607902526855, "learning_rate": 9.979582385363912e-06, "loss": 0.9505, "step": 442 }, { "epoch": 0.06, "grad_norm": 1.4821429252624512, "learning_rate": 9.979390352761364e-06, "loss": 0.9049, "step": 443 }, { "epoch": 0.06, "grad_norm": 1.493414044380188, "learning_rate": 9.979197423189163e-06, "loss": 0.8665, "step": 444 }, { "epoch": 0.06, "grad_norm": 1.5437967777252197, "learning_rate": 9.979003596682058e-06, "loss": 0.8631, "step": 445 }, { "epoch": 0.06, "grad_norm": 1.4784977436065674, "learning_rate": 9.978808873274968e-06, "loss": 0.881, "step": 446 }, { "epoch": 0.06, "grad_norm": 1.67030930519104, "learning_rate": 9.978613253002968e-06, "loss": 0.9372, "step": 447 }, { "epoch": 0.06, "grad_norm": 1.6668001413345337, "learning_rate": 9.9784167359013e-06, "loss": 0.9559, "step": 448 }, { "epoch": 0.06, "grad_norm": 1.4616689682006836, "learning_rate": 9.978219322005358e-06, "loss": 0.92, "step": 449 }, { "epoch": 0.06, "grad_norm": 1.1318022012710571, "learning_rate": 9.978021011350706e-06, "loss": 1.1651, "step": 450 }, { "epoch": 0.06, "grad_norm": 1.4155759811401367, "learning_rate": 9.977821803973067e-06, "loss": 0.9225, "step": 451 }, { "epoch": 0.06, "grad_norm": 0.7408362030982971, "learning_rate": 9.977621699908327e-06, "loss": 1.1585, "step": 452 }, { "epoch": 0.06, "grad_norm": 1.595224380493164, "learning_rate": 9.97742069919253e-06, "loss": 0.9268, "step": 453 }, { "epoch": 0.06, "grad_norm": 1.4310078620910645, "learning_rate": 9.977218801861884e-06, "loss": 0.9204, "step": 454 }, { "epoch": 0.06, "grad_norm": 1.4169831275939941, "learning_rate": 9.97701600795276e-06, "loss": 0.9334, "step": 455 }, { "epoch": 0.06, "grad_norm": 1.383875846862793, "learning_rate": 9.976812317501685e-06, "loss": 0.9153, "step": 456 }, { "epoch": 0.06, "grad_norm": 1.3240448236465454, "learning_rate": 9.976607730545354e-06, "loss": 1.1557, "step": 457 }, { "epoch": 0.06, "grad_norm": 1.522214412689209, "learning_rate": 9.97640224712062e-06, "loss": 0.9297, "step": 458 }, { "epoch": 0.06, "grad_norm": 1.58584463596344, "learning_rate": 9.976195867264496e-06, "loss": 0.9223, "step": 459 }, { "epoch": 0.06, "grad_norm": 1.4080723524093628, "learning_rate": 9.97598859101416e-06, "loss": 0.9162, "step": 460 }, { "epoch": 0.06, "grad_norm": 1.4770317077636719, "learning_rate": 9.975780418406953e-06, "loss": 0.9454, "step": 461 }, { "epoch": 0.06, "grad_norm": 1.3944971561431885, "learning_rate": 9.975571349480368e-06, "loss": 0.9163, "step": 462 }, { "epoch": 0.06, "grad_norm": 1.352354645729065, "learning_rate": 9.975361384272073e-06, "loss": 0.9465, "step": 463 }, { "epoch": 0.06, "grad_norm": 1.4349240064620972, "learning_rate": 9.975150522819886e-06, "loss": 0.9066, "step": 464 }, { "epoch": 0.06, "grad_norm": 1.4692420959472656, "learning_rate": 9.97493876516179e-06, "loss": 0.891, "step": 465 }, { "epoch": 0.06, "grad_norm": 1.450058102607727, "learning_rate": 9.974726111335932e-06, "loss": 0.9429, "step": 466 }, { "epoch": 0.06, "grad_norm": 1.1307307481765747, "learning_rate": 9.974512561380622e-06, "loss": 1.1353, "step": 467 }, { "epoch": 0.06, "grad_norm": 1.4313478469848633, "learning_rate": 9.974298115334322e-06, "loss": 0.9001, "step": 468 }, { "epoch": 0.06, "grad_norm": 1.4373480081558228, "learning_rate": 9.974082773235665e-06, "loss": 0.9186, "step": 469 }, { "epoch": 0.06, "grad_norm": 0.8010553121566772, "learning_rate": 9.973866535123442e-06, "loss": 1.1336, "step": 470 }, { "epoch": 0.06, "grad_norm": 0.7450042366981506, "learning_rate": 9.973649401036605e-06, "loss": 1.1436, "step": 471 }, { "epoch": 0.06, "grad_norm": 1.5333291292190552, "learning_rate": 9.973431371014266e-06, "loss": 0.9594, "step": 472 }, { "epoch": 0.06, "grad_norm": 0.6851996779441833, "learning_rate": 9.973212445095704e-06, "loss": 1.1147, "step": 473 }, { "epoch": 0.06, "grad_norm": 1.3614850044250488, "learning_rate": 9.97299262332035e-06, "loss": 0.864, "step": 474 }, { "epoch": 0.06, "grad_norm": 1.3757747411727905, "learning_rate": 9.972771905727807e-06, "loss": 0.8914, "step": 475 }, { "epoch": 0.06, "grad_norm": 0.8465238809585571, "learning_rate": 9.972550292357833e-06, "loss": 1.14, "step": 476 }, { "epoch": 0.06, "grad_norm": 0.792616605758667, "learning_rate": 9.972327783250348e-06, "loss": 1.1406, "step": 477 }, { "epoch": 0.06, "grad_norm": 0.6589449048042297, "learning_rate": 9.972104378445434e-06, "loss": 1.1422, "step": 478 }, { "epoch": 0.06, "grad_norm": 1.5642703771591187, "learning_rate": 9.971880077983334e-06, "loss": 0.9017, "step": 479 }, { "epoch": 0.06, "grad_norm": 1.4927376508712769, "learning_rate": 9.97165488190445e-06, "loss": 0.9239, "step": 480 }, { "epoch": 0.06, "grad_norm": 1.5184775590896606, "learning_rate": 9.971428790249355e-06, "loss": 0.9563, "step": 481 }, { "epoch": 0.06, "grad_norm": 1.1736501455307007, "learning_rate": 9.971201803058772e-06, "loss": 1.1429, "step": 482 }, { "epoch": 0.06, "grad_norm": 1.6412310600280762, "learning_rate": 9.97097392037359e-06, "loss": 1.0089, "step": 483 }, { "epoch": 0.06, "grad_norm": 1.406668782234192, "learning_rate": 9.970745142234859e-06, "loss": 0.9478, "step": 484 }, { "epoch": 0.06, "grad_norm": 1.5595468282699585, "learning_rate": 9.970515468683789e-06, "loss": 0.931, "step": 485 }, { "epoch": 0.06, "grad_norm": 1.5404413938522339, "learning_rate": 9.970284899761756e-06, "loss": 0.9377, "step": 486 }, { "epoch": 0.06, "grad_norm": 0.9446240067481995, "learning_rate": 9.97005343551029e-06, "loss": 1.1257, "step": 487 }, { "epoch": 0.06, "grad_norm": 0.8129511475563049, "learning_rate": 9.96982107597109e-06, "loss": 1.1451, "step": 488 }, { "epoch": 0.06, "grad_norm": 1.492637038230896, "learning_rate": 9.969587821186008e-06, "loss": 0.9456, "step": 489 }, { "epoch": 0.06, "grad_norm": 1.4750523567199707, "learning_rate": 9.969353671197064e-06, "loss": 0.8763, "step": 490 }, { "epoch": 0.06, "grad_norm": 1.4654428958892822, "learning_rate": 9.969118626046439e-06, "loss": 0.9109, "step": 491 }, { "epoch": 0.06, "grad_norm": 1.2809250354766846, "learning_rate": 9.96888268577647e-06, "loss": 1.1546, "step": 492 }, { "epoch": 0.06, "grad_norm": 1.3946822881698608, "learning_rate": 9.968645850429659e-06, "loss": 0.8957, "step": 493 }, { "epoch": 0.06, "grad_norm": 1.3612524271011353, "learning_rate": 9.968408120048671e-06, "loss": 0.8655, "step": 494 }, { "epoch": 0.06, "grad_norm": 0.8383128046989441, "learning_rate": 9.968169494676327e-06, "loss": 1.1374, "step": 495 }, { "epoch": 0.06, "grad_norm": 1.503581166267395, "learning_rate": 9.967929974355615e-06, "loss": 0.8953, "step": 496 }, { "epoch": 0.07, "grad_norm": 1.5719225406646729, "learning_rate": 9.967689559129677e-06, "loss": 1.0023, "step": 497 }, { "epoch": 0.07, "grad_norm": 1.562961459159851, "learning_rate": 9.967448249041825e-06, "loss": 0.9438, "step": 498 }, { "epoch": 0.07, "grad_norm": 1.2267963886260986, "learning_rate": 9.967206044135526e-06, "loss": 1.1251, "step": 499 }, { "epoch": 0.07, "grad_norm": 1.4219261407852173, "learning_rate": 9.966962944454408e-06, "loss": 0.8779, "step": 500 }, { "epoch": 0.07, "grad_norm": 1.4327058792114258, "learning_rate": 9.966718950042267e-06, "loss": 0.8656, "step": 501 }, { "epoch": 0.07, "grad_norm": 0.7444456815719604, "learning_rate": 9.96647406094305e-06, "loss": 1.1285, "step": 502 }, { "epoch": 0.07, "grad_norm": 0.7379494309425354, "learning_rate": 9.966228277200874e-06, "loss": 1.1286, "step": 503 }, { "epoch": 0.07, "grad_norm": 1.4499906301498413, "learning_rate": 9.965981598860012e-06, "loss": 0.9215, "step": 504 }, { "epoch": 0.07, "grad_norm": 1.4829331636428833, "learning_rate": 9.965734025964899e-06, "loss": 0.9651, "step": 505 }, { "epoch": 0.07, "grad_norm": 0.9941326379776001, "learning_rate": 9.965485558560134e-06, "loss": 1.1227, "step": 506 }, { "epoch": 0.07, "grad_norm": 0.8838790655136108, "learning_rate": 9.965236196690475e-06, "loss": 1.1614, "step": 507 }, { "epoch": 0.07, "grad_norm": 1.4430944919586182, "learning_rate": 9.96498594040084e-06, "loss": 0.9023, "step": 508 }, { "epoch": 0.07, "grad_norm": 0.9815982580184937, "learning_rate": 9.964734789736307e-06, "loss": 1.1444, "step": 509 }, { "epoch": 0.07, "grad_norm": 1.4260226488113403, "learning_rate": 9.964482744742123e-06, "loss": 0.8874, "step": 510 }, { "epoch": 0.07, "grad_norm": 0.9109838008880615, "learning_rate": 9.964229805463685e-06, "loss": 1.1404, "step": 511 }, { "epoch": 0.07, "grad_norm": 1.615838885307312, "learning_rate": 9.96397597194656e-06, "loss": 0.9743, "step": 512 }, { "epoch": 0.07, "grad_norm": 1.4941562414169312, "learning_rate": 9.963721244236473e-06, "loss": 0.9363, "step": 513 }, { "epoch": 0.07, "grad_norm": 1.0725109577178955, "learning_rate": 9.963465622379308e-06, "loss": 1.1213, "step": 514 }, { "epoch": 0.07, "grad_norm": 1.4225226640701294, "learning_rate": 9.963209106421109e-06, "loss": 0.8978, "step": 515 }, { "epoch": 0.07, "grad_norm": 1.3777199983596802, "learning_rate": 9.962951696408089e-06, "loss": 0.8646, "step": 516 }, { "epoch": 0.07, "grad_norm": 1.3811845779418945, "learning_rate": 9.962693392386615e-06, "loss": 0.9246, "step": 517 }, { "epoch": 0.07, "grad_norm": 1.4976787567138672, "learning_rate": 9.962434194403217e-06, "loss": 0.9238, "step": 518 }, { "epoch": 0.07, "grad_norm": 1.471773624420166, "learning_rate": 9.962174102504586e-06, "loss": 0.949, "step": 519 }, { "epoch": 0.07, "grad_norm": 1.6502408981323242, "learning_rate": 9.961913116737573e-06, "loss": 0.9587, "step": 520 }, { "epoch": 0.07, "grad_norm": 1.4394665956497192, "learning_rate": 9.961651237149194e-06, "loss": 0.9272, "step": 521 }, { "epoch": 0.07, "grad_norm": 1.6487658023834229, "learning_rate": 9.961388463786619e-06, "loss": 0.9549, "step": 522 }, { "epoch": 0.07, "grad_norm": 1.3483222723007202, "learning_rate": 9.961124796697185e-06, "loss": 0.9737, "step": 523 }, { "epoch": 0.07, "grad_norm": 1.5094635486602783, "learning_rate": 9.96086023592839e-06, "loss": 0.8476, "step": 524 }, { "epoch": 0.07, "grad_norm": 1.4152005910873413, "learning_rate": 9.960594781527887e-06, "loss": 0.8725, "step": 525 }, { "epoch": 0.07, "grad_norm": 1.5195025205612183, "learning_rate": 9.960328433543499e-06, "loss": 0.906, "step": 526 }, { "epoch": 0.07, "grad_norm": 1.3670496940612793, "learning_rate": 9.9600611920232e-06, "loss": 0.8784, "step": 527 }, { "epoch": 0.07, "grad_norm": 0.8869580030441284, "learning_rate": 9.959793057015133e-06, "loss": 1.1235, "step": 528 }, { "epoch": 0.07, "grad_norm": 1.4101290702819824, "learning_rate": 9.959524028567598e-06, "loss": 0.9213, "step": 529 }, { "epoch": 0.07, "grad_norm": 1.4678270816802979, "learning_rate": 9.959254106729057e-06, "loss": 0.9043, "step": 530 }, { "epoch": 0.07, "grad_norm": 0.715247392654419, "learning_rate": 9.958983291548131e-06, "loss": 1.1369, "step": 531 }, { "epoch": 0.07, "grad_norm": 1.4350565671920776, "learning_rate": 9.958711583073608e-06, "loss": 0.9125, "step": 532 }, { "epoch": 0.07, "grad_norm": 1.4310588836669922, "learning_rate": 9.958438981354427e-06, "loss": 0.8695, "step": 533 }, { "epoch": 0.07, "grad_norm": 1.4433536529541016, "learning_rate": 9.9581654864397e-06, "loss": 0.8983, "step": 534 }, { "epoch": 0.07, "grad_norm": 1.5280098915100098, "learning_rate": 9.957891098378688e-06, "loss": 0.9528, "step": 535 }, { "epoch": 0.07, "grad_norm": 1.426310420036316, "learning_rate": 9.957615817220818e-06, "loss": 0.9307, "step": 536 }, { "epoch": 0.07, "grad_norm": 1.0297839641571045, "learning_rate": 9.957339643015682e-06, "loss": 1.1136, "step": 537 }, { "epoch": 0.07, "grad_norm": 0.7066054344177246, "learning_rate": 9.957062575813027e-06, "loss": 1.1098, "step": 538 }, { "epoch": 0.07, "grad_norm": 1.4472213983535767, "learning_rate": 9.956784615662764e-06, "loss": 0.9919, "step": 539 }, { "epoch": 0.07, "grad_norm": 1.0490275621414185, "learning_rate": 9.956505762614963e-06, "loss": 1.1455, "step": 540 }, { "epoch": 0.07, "grad_norm": 1.4573273658752441, "learning_rate": 9.956226016719855e-06, "loss": 0.8611, "step": 541 }, { "epoch": 0.07, "grad_norm": 1.5085358619689941, "learning_rate": 9.955945378027832e-06, "loss": 0.9244, "step": 542 }, { "epoch": 0.07, "grad_norm": 1.3400354385375977, "learning_rate": 9.955663846589448e-06, "loss": 0.9273, "step": 543 }, { "epoch": 0.07, "grad_norm": 1.3608402013778687, "learning_rate": 9.955381422455418e-06, "loss": 0.88, "step": 544 }, { "epoch": 0.07, "grad_norm": 0.9880529046058655, "learning_rate": 9.955098105676617e-06, "loss": 1.1395, "step": 545 }, { "epoch": 0.07, "grad_norm": 1.3732532262802124, "learning_rate": 9.954813896304078e-06, "loss": 0.8685, "step": 546 }, { "epoch": 0.07, "grad_norm": 1.4943867921829224, "learning_rate": 9.954528794389002e-06, "loss": 0.9234, "step": 547 }, { "epoch": 0.07, "grad_norm": 0.6807183027267456, "learning_rate": 9.954242799982742e-06, "loss": 1.1478, "step": 548 }, { "epoch": 0.07, "grad_norm": 0.6960909366607666, "learning_rate": 9.953955913136817e-06, "loss": 1.1159, "step": 549 }, { "epoch": 0.07, "grad_norm": 1.6839532852172852, "learning_rate": 9.953668133902907e-06, "loss": 0.8628, "step": 550 }, { "epoch": 0.07, "grad_norm": 1.6231952905654907, "learning_rate": 9.95337946233285e-06, "loss": 0.9391, "step": 551 }, { "epoch": 0.07, "grad_norm": 1.5207356214523315, "learning_rate": 9.95308989847865e-06, "loss": 0.9191, "step": 552 }, { "epoch": 0.07, "grad_norm": 1.3624461889266968, "learning_rate": 9.952799442392462e-06, "loss": 0.8944, "step": 553 }, { "epoch": 0.07, "grad_norm": 1.4801981449127197, "learning_rate": 9.952508094126611e-06, "loss": 0.8802, "step": 554 }, { "epoch": 0.07, "grad_norm": 0.9643036127090454, "learning_rate": 9.952215853733581e-06, "loss": 1.147, "step": 555 }, { "epoch": 0.07, "grad_norm": 1.5695667266845703, "learning_rate": 9.951922721266013e-06, "loss": 0.8933, "step": 556 }, { "epoch": 0.07, "grad_norm": 1.4505219459533691, "learning_rate": 9.95162869677671e-06, "loss": 0.9574, "step": 557 }, { "epoch": 0.07, "grad_norm": 1.6177573204040527, "learning_rate": 9.95133378031864e-06, "loss": 0.9623, "step": 558 }, { "epoch": 0.07, "grad_norm": 1.316936731338501, "learning_rate": 9.951037971944923e-06, "loss": 0.8722, "step": 559 }, { "epoch": 0.07, "grad_norm": 1.5060988664627075, "learning_rate": 9.95074127170885e-06, "loss": 0.9274, "step": 560 }, { "epoch": 0.07, "grad_norm": 1.5979845523834229, "learning_rate": 9.950443679663867e-06, "loss": 0.9052, "step": 561 }, { "epoch": 0.07, "grad_norm": 2.349004030227661, "learning_rate": 9.950145195863577e-06, "loss": 0.9002, "step": 562 }, { "epoch": 0.07, "grad_norm": 1.539211392402649, "learning_rate": 9.94984582036175e-06, "loss": 0.925, "step": 563 }, { "epoch": 0.07, "grad_norm": 1.3603487014770508, "learning_rate": 9.949545553212317e-06, "loss": 0.8994, "step": 564 }, { "epoch": 0.07, "grad_norm": 1.3160780668258667, "learning_rate": 9.949244394469362e-06, "loss": 0.8749, "step": 565 }, { "epoch": 0.07, "grad_norm": 1.2745013236999512, "learning_rate": 9.948942344187141e-06, "loss": 1.1485, "step": 566 }, { "epoch": 0.07, "grad_norm": 1.4726169109344482, "learning_rate": 9.948639402420058e-06, "loss": 0.9015, "step": 567 }, { "epoch": 0.07, "grad_norm": 0.7058154940605164, "learning_rate": 9.948335569222687e-06, "loss": 1.1258, "step": 568 }, { "epoch": 0.07, "grad_norm": 0.7658041715621948, "learning_rate": 9.94803084464976e-06, "loss": 1.1254, "step": 569 }, { "epoch": 0.07, "grad_norm": 1.4049140214920044, "learning_rate": 9.947725228756168e-06, "loss": 0.915, "step": 570 }, { "epoch": 0.07, "grad_norm": 1.459114909172058, "learning_rate": 9.947418721596963e-06, "loss": 0.9072, "step": 571 }, { "epoch": 0.07, "grad_norm": 1.5731946229934692, "learning_rate": 9.94711132322736e-06, "loss": 0.9366, "step": 572 }, { "epoch": 0.08, "grad_norm": 1.4637694358825684, "learning_rate": 9.94680303370273e-06, "loss": 0.9217, "step": 573 }, { "epoch": 0.08, "grad_norm": 1.804619312286377, "learning_rate": 9.946493853078609e-06, "loss": 0.9373, "step": 574 }, { "epoch": 0.08, "grad_norm": 1.6396490335464478, "learning_rate": 9.946183781410692e-06, "loss": 0.9213, "step": 575 }, { "epoch": 0.08, "grad_norm": 2.054314136505127, "learning_rate": 9.945872818754831e-06, "loss": 0.9431, "step": 576 }, { "epoch": 0.08, "grad_norm": 1.4272584915161133, "learning_rate": 9.945560965167047e-06, "loss": 0.8665, "step": 577 }, { "epoch": 0.08, "grad_norm": 1.420850396156311, "learning_rate": 9.945248220703513e-06, "loss": 0.9326, "step": 578 }, { "epoch": 0.08, "grad_norm": 1.5234938859939575, "learning_rate": 9.944934585420565e-06, "loss": 0.8776, "step": 579 }, { "epoch": 0.08, "grad_norm": 1.3499335050582886, "learning_rate": 9.944620059374702e-06, "loss": 0.9127, "step": 580 }, { "epoch": 0.08, "grad_norm": 2.7517011165618896, "learning_rate": 9.94430464262258e-06, "loss": 0.9579, "step": 581 }, { "epoch": 0.08, "grad_norm": 1.511658787727356, "learning_rate": 9.94398833522102e-06, "loss": 0.8863, "step": 582 }, { "epoch": 0.08, "grad_norm": 1.575329065322876, "learning_rate": 9.943671137226996e-06, "loss": 0.9076, "step": 583 }, { "epoch": 0.08, "grad_norm": 1.893376350402832, "learning_rate": 9.943353048697649e-06, "loss": 1.148, "step": 584 }, { "epoch": 0.08, "grad_norm": 1.4626972675323486, "learning_rate": 9.94303406969028e-06, "loss": 0.9252, "step": 585 }, { "epoch": 0.08, "grad_norm": 1.372159719467163, "learning_rate": 9.942714200262347e-06, "loss": 0.9323, "step": 586 }, { "epoch": 0.08, "grad_norm": 1.4686838388442993, "learning_rate": 9.94239344047147e-06, "loss": 0.913, "step": 587 }, { "epoch": 0.08, "grad_norm": 1.3882702589035034, "learning_rate": 9.94207179037543e-06, "loss": 0.9352, "step": 588 }, { "epoch": 0.08, "grad_norm": 1.6097077131271362, "learning_rate": 9.941749250032168e-06, "loss": 0.9124, "step": 589 }, { "epoch": 0.08, "grad_norm": 1.4467363357543945, "learning_rate": 9.941425819499787e-06, "loss": 0.97, "step": 590 }, { "epoch": 0.08, "grad_norm": 1.3655301332473755, "learning_rate": 9.941101498836544e-06, "loss": 0.8578, "step": 591 }, { "epoch": 0.08, "grad_norm": 1.7663347721099854, "learning_rate": 9.940776288100865e-06, "loss": 0.9233, "step": 592 }, { "epoch": 0.08, "grad_norm": 1.5170968770980835, "learning_rate": 9.940450187351332e-06, "loss": 0.9219, "step": 593 }, { "epoch": 0.08, "grad_norm": 1.2887120246887207, "learning_rate": 9.940123196646686e-06, "loss": 1.1508, "step": 594 }, { "epoch": 0.08, "grad_norm": 1.5205132961273193, "learning_rate": 9.93979531604583e-06, "loss": 0.8711, "step": 595 }, { "epoch": 0.08, "grad_norm": 1.4376335144042969, "learning_rate": 9.939466545607827e-06, "loss": 0.9828, "step": 596 }, { "epoch": 0.08, "grad_norm": 1.4160596132278442, "learning_rate": 9.939136885391903e-06, "loss": 0.8854, "step": 597 }, { "epoch": 0.08, "grad_norm": 1.384126901626587, "learning_rate": 9.938806335457441e-06, "loss": 0.9042, "step": 598 }, { "epoch": 0.08, "grad_norm": 1.5627719163894653, "learning_rate": 9.938474895863984e-06, "loss": 0.9404, "step": 599 }, { "epoch": 0.08, "grad_norm": 1.4128080606460571, "learning_rate": 9.938142566671234e-06, "loss": 0.9119, "step": 600 }, { "epoch": 0.08, "grad_norm": 1.4634920358657837, "learning_rate": 9.93780934793906e-06, "loss": 0.8931, "step": 601 }, { "epoch": 0.08, "grad_norm": 1.447167158126831, "learning_rate": 9.937475239727486e-06, "loss": 0.9053, "step": 602 }, { "epoch": 0.08, "grad_norm": 1.3842171430587769, "learning_rate": 9.937140242096697e-06, "loss": 0.9447, "step": 603 }, { "epoch": 0.08, "grad_norm": 1.411766767501831, "learning_rate": 9.936804355107035e-06, "loss": 0.9384, "step": 604 }, { "epoch": 0.08, "grad_norm": 1.581775188446045, "learning_rate": 9.93646757881901e-06, "loss": 0.9516, "step": 605 }, { "epoch": 0.08, "grad_norm": 1.2439706325531006, "learning_rate": 9.936129913293284e-06, "loss": 0.9346, "step": 606 }, { "epoch": 0.08, "grad_norm": 1.5507838726043701, "learning_rate": 9.935791358590684e-06, "loss": 0.9335, "step": 607 }, { "epoch": 0.08, "grad_norm": 1.4203637838363647, "learning_rate": 9.935451914772199e-06, "loss": 0.9003, "step": 608 }, { "epoch": 0.08, "grad_norm": 1.3394794464111328, "learning_rate": 9.93511158189897e-06, "loss": 0.905, "step": 609 }, { "epoch": 0.08, "grad_norm": 1.4186174869537354, "learning_rate": 9.93477036003231e-06, "loss": 1.1164, "step": 610 }, { "epoch": 0.08, "grad_norm": 1.4027564525604248, "learning_rate": 9.934428249233676e-06, "loss": 0.9628, "step": 611 }, { "epoch": 0.08, "grad_norm": 1.5176805257797241, "learning_rate": 9.934085249564705e-06, "loss": 0.8742, "step": 612 }, { "epoch": 0.08, "grad_norm": 0.727710485458374, "learning_rate": 9.933741361087177e-06, "loss": 1.1273, "step": 613 }, { "epoch": 0.08, "grad_norm": 0.7581411004066467, "learning_rate": 9.933396583863039e-06, "loss": 1.1247, "step": 614 }, { "epoch": 0.08, "grad_norm": 1.4443426132202148, "learning_rate": 9.933050917954404e-06, "loss": 0.9506, "step": 615 }, { "epoch": 0.08, "grad_norm": 0.713009774684906, "learning_rate": 9.932704363423531e-06, "loss": 1.1187, "step": 616 }, { "epoch": 0.08, "grad_norm": 0.6805099844932556, "learning_rate": 9.932356920332852e-06, "loss": 1.1181, "step": 617 }, { "epoch": 0.08, "grad_norm": 1.409752607345581, "learning_rate": 9.932008588744955e-06, "loss": 0.8795, "step": 618 }, { "epoch": 0.08, "grad_norm": 0.7561904788017273, "learning_rate": 9.931659368722583e-06, "loss": 1.1091, "step": 619 }, { "epoch": 0.08, "grad_norm": 1.4316661357879639, "learning_rate": 9.931309260328645e-06, "loss": 0.9123, "step": 620 }, { "epoch": 0.08, "grad_norm": 1.4595361948013306, "learning_rate": 9.930958263626212e-06, "loss": 0.8683, "step": 621 }, { "epoch": 0.08, "grad_norm": 0.7839780449867249, "learning_rate": 9.930606378678504e-06, "loss": 1.1265, "step": 622 }, { "epoch": 0.08, "grad_norm": 1.4128811359405518, "learning_rate": 9.930253605548914e-06, "loss": 0.8762, "step": 623 }, { "epoch": 0.08, "grad_norm": 1.3933911323547363, "learning_rate": 9.929899944300988e-06, "loss": 0.9111, "step": 624 }, { "epoch": 0.08, "grad_norm": 1.360595464706421, "learning_rate": 9.929545394998433e-06, "loss": 0.9193, "step": 625 }, { "epoch": 0.08, "grad_norm": 2.1083121299743652, "learning_rate": 9.929189957705116e-06, "loss": 0.9444, "step": 626 }, { "epoch": 0.08, "grad_norm": 0.8485044836997986, "learning_rate": 9.928833632485064e-06, "loss": 1.1574, "step": 627 }, { "epoch": 0.08, "grad_norm": 0.7834640145301819, "learning_rate": 9.928476419402466e-06, "loss": 1.0924, "step": 628 }, { "epoch": 0.08, "grad_norm": 1.569897174835205, "learning_rate": 9.928118318521666e-06, "loss": 0.9204, "step": 629 }, { "epoch": 0.08, "grad_norm": 1.5619263648986816, "learning_rate": 9.927759329907174e-06, "loss": 0.9005, "step": 630 }, { "epoch": 0.08, "grad_norm": 1.3795418739318848, "learning_rate": 9.927399453623655e-06, "loss": 0.8982, "step": 631 }, { "epoch": 0.08, "grad_norm": 1.4037349224090576, "learning_rate": 9.927038689735934e-06, "loss": 0.9222, "step": 632 }, { "epoch": 0.08, "grad_norm": 1.1475216150283813, "learning_rate": 9.926677038309005e-06, "loss": 1.1245, "step": 633 }, { "epoch": 0.08, "grad_norm": 0.9332098960876465, "learning_rate": 9.926314499408007e-06, "loss": 1.1546, "step": 634 }, { "epoch": 0.08, "grad_norm": 0.7333803772926331, "learning_rate": 9.92595107309825e-06, "loss": 1.1032, "step": 635 }, { "epoch": 0.08, "grad_norm": 0.7873438000679016, "learning_rate": 9.925586759445199e-06, "loss": 1.099, "step": 636 }, { "epoch": 0.08, "grad_norm": 1.6364785432815552, "learning_rate": 9.92522155851448e-06, "loss": 0.9716, "step": 637 }, { "epoch": 0.08, "grad_norm": 1.0541969537734985, "learning_rate": 9.924855470371879e-06, "loss": 1.155, "step": 638 }, { "epoch": 0.08, "grad_norm": 1.4854607582092285, "learning_rate": 9.924488495083345e-06, "loss": 0.9438, "step": 639 }, { "epoch": 0.08, "grad_norm": 1.4189294576644897, "learning_rate": 9.92412063271498e-06, "loss": 0.895, "step": 640 }, { "epoch": 0.08, "grad_norm": 1.5169216394424438, "learning_rate": 9.92375188333305e-06, "loss": 0.9105, "step": 641 }, { "epoch": 0.08, "grad_norm": 1.3798799514770508, "learning_rate": 9.92338224700398e-06, "loss": 0.8767, "step": 642 }, { "epoch": 0.08, "grad_norm": 1.3165807723999023, "learning_rate": 9.923011723794358e-06, "loss": 1.1281, "step": 643 }, { "epoch": 0.08, "grad_norm": 1.4282736778259277, "learning_rate": 9.922640313770926e-06, "loss": 0.9101, "step": 644 }, { "epoch": 0.08, "grad_norm": 1.410033106803894, "learning_rate": 9.922268017000589e-06, "loss": 0.9414, "step": 645 }, { "epoch": 0.08, "grad_norm": 0.8061385750770569, "learning_rate": 9.92189483355041e-06, "loss": 1.1371, "step": 646 }, { "epoch": 0.08, "grad_norm": 1.393268346786499, "learning_rate": 9.921520763487614e-06, "loss": 0.8493, "step": 647 }, { "epoch": 0.08, "grad_norm": 1.4565198421478271, "learning_rate": 9.921145806879586e-06, "loss": 0.91, "step": 648 }, { "epoch": 0.09, "grad_norm": 1.6094073057174683, "learning_rate": 9.920769963793868e-06, "loss": 0.9335, "step": 649 }, { "epoch": 0.09, "grad_norm": 1.5463881492614746, "learning_rate": 9.920393234298163e-06, "loss": 0.842, "step": 650 }, { "epoch": 0.09, "grad_norm": 1.0696147680282593, "learning_rate": 9.920015618460333e-06, "loss": 1.1121, "step": 651 }, { "epoch": 0.09, "grad_norm": 0.8222506046295166, "learning_rate": 9.919637116348402e-06, "loss": 1.1223, "step": 652 }, { "epoch": 0.09, "grad_norm": 1.461240530014038, "learning_rate": 9.91925772803055e-06, "loss": 0.8956, "step": 653 }, { "epoch": 0.09, "grad_norm": 1.4640754461288452, "learning_rate": 9.918877453575121e-06, "loss": 0.8945, "step": 654 }, { "epoch": 0.09, "grad_norm": 1.6219547986984253, "learning_rate": 9.918496293050615e-06, "loss": 1.1303, "step": 655 }, { "epoch": 0.09, "grad_norm": 1.435442328453064, "learning_rate": 9.918114246525693e-06, "loss": 1.1205, "step": 656 }, { "epoch": 0.09, "grad_norm": 1.4334720373153687, "learning_rate": 9.917731314069176e-06, "loss": 0.8967, "step": 657 }, { "epoch": 0.09, "grad_norm": 1.3656113147735596, "learning_rate": 9.917347495750043e-06, "loss": 0.9593, "step": 658 }, { "epoch": 0.09, "grad_norm": 1.2096589803695679, "learning_rate": 9.916962791637434e-06, "loss": 1.1118, "step": 659 }, { "epoch": 0.09, "grad_norm": 1.5322846174240112, "learning_rate": 9.91657720180065e-06, "loss": 0.8532, "step": 660 }, { "epoch": 0.09, "grad_norm": 1.505578637123108, "learning_rate": 9.916190726309147e-06, "loss": 0.864, "step": 661 }, { "epoch": 0.09, "grad_norm": 1.4839987754821777, "learning_rate": 9.915803365232546e-06, "loss": 0.9669, "step": 662 }, { "epoch": 0.09, "grad_norm": 0.9134460687637329, "learning_rate": 9.915415118640623e-06, "loss": 1.1108, "step": 663 }, { "epoch": 0.09, "grad_norm": 1.4986017942428589, "learning_rate": 9.915025986603315e-06, "loss": 0.8547, "step": 664 }, { "epoch": 0.09, "grad_norm": 1.4104316234588623, "learning_rate": 9.91463596919072e-06, "loss": 0.8521, "step": 665 }, { "epoch": 0.09, "grad_norm": 0.8438710570335388, "learning_rate": 9.914245066473095e-06, "loss": 1.1205, "step": 666 }, { "epoch": 0.09, "grad_norm": 1.3536555767059326, "learning_rate": 9.913853278520855e-06, "loss": 0.8553, "step": 667 }, { "epoch": 0.09, "grad_norm": 1.445067048072815, "learning_rate": 9.913460605404574e-06, "loss": 0.9265, "step": 668 }, { "epoch": 0.09, "grad_norm": 1.403393030166626, "learning_rate": 9.913067047194988e-06, "loss": 0.8752, "step": 669 }, { "epoch": 0.09, "grad_norm": 1.3951669931411743, "learning_rate": 9.912672603962992e-06, "loss": 0.93, "step": 670 }, { "epoch": 0.09, "grad_norm": 1.4539226293563843, "learning_rate": 9.912277275779639e-06, "loss": 0.9184, "step": 671 }, { "epoch": 0.09, "grad_norm": 1.3986860513687134, "learning_rate": 9.91188106271614e-06, "loss": 0.9039, "step": 672 }, { "epoch": 0.09, "grad_norm": 1.4044756889343262, "learning_rate": 9.911483964843872e-06, "loss": 0.8417, "step": 673 }, { "epoch": 0.09, "grad_norm": 1.4084972143173218, "learning_rate": 9.91108598223436e-06, "loss": 0.9311, "step": 674 }, { "epoch": 0.09, "grad_norm": 1.3715561628341675, "learning_rate": 9.910687114959302e-06, "loss": 0.9041, "step": 675 }, { "epoch": 0.09, "grad_norm": 1.4749529361724854, "learning_rate": 9.910287363090544e-06, "loss": 0.8704, "step": 676 }, { "epoch": 0.09, "grad_norm": 1.4542763233184814, "learning_rate": 9.909886726700099e-06, "loss": 0.9172, "step": 677 }, { "epoch": 0.09, "grad_norm": 1.4592245817184448, "learning_rate": 9.909485205860134e-06, "loss": 0.9213, "step": 678 }, { "epoch": 0.09, "grad_norm": 1.3907123804092407, "learning_rate": 9.909082800642978e-06, "loss": 0.9056, "step": 679 }, { "epoch": 0.09, "grad_norm": 1.5260694026947021, "learning_rate": 9.90867951112112e-06, "loss": 0.9141, "step": 680 }, { "epoch": 0.09, "grad_norm": 1.4086151123046875, "learning_rate": 9.908275337367205e-06, "loss": 0.904, "step": 681 }, { "epoch": 0.09, "grad_norm": 1.3766196966171265, "learning_rate": 9.907870279454042e-06, "loss": 0.9364, "step": 682 }, { "epoch": 0.09, "grad_norm": 1.1790541410446167, "learning_rate": 9.907464337454595e-06, "loss": 1.114, "step": 683 }, { "epoch": 0.09, "grad_norm": 0.9067893624305725, "learning_rate": 9.907057511441988e-06, "loss": 1.1219, "step": 684 }, { "epoch": 0.09, "grad_norm": 1.7996842861175537, "learning_rate": 9.906649801489507e-06, "loss": 0.888, "step": 685 }, { "epoch": 0.09, "grad_norm": 0.9921464323997498, "learning_rate": 9.906241207670597e-06, "loss": 1.1319, "step": 686 }, { "epoch": 0.09, "grad_norm": 1.4900379180908203, "learning_rate": 9.905831730058857e-06, "loss": 0.9203, "step": 687 }, { "epoch": 0.09, "grad_norm": 1.3317217826843262, "learning_rate": 9.905421368728049e-06, "loss": 0.893, "step": 688 }, { "epoch": 0.09, "grad_norm": 1.6245999336242676, "learning_rate": 9.905010123752096e-06, "loss": 0.8438, "step": 689 }, { "epoch": 0.09, "grad_norm": 1.4188563823699951, "learning_rate": 9.90459799520508e-06, "loss": 0.8984, "step": 690 }, { "epoch": 0.09, "grad_norm": 1.8423871994018555, "learning_rate": 9.904184983161236e-06, "loss": 0.8637, "step": 691 }, { "epoch": 0.09, "grad_norm": 1.270402193069458, "learning_rate": 9.903771087694965e-06, "loss": 1.1153, "step": 692 }, { "epoch": 0.09, "grad_norm": 1.4932940006256104, "learning_rate": 9.903356308880824e-06, "loss": 0.8993, "step": 693 }, { "epoch": 0.09, "grad_norm": 1.4707093238830566, "learning_rate": 9.90294064679353e-06, "loss": 0.9377, "step": 694 }, { "epoch": 0.09, "grad_norm": 1.4020847082138062, "learning_rate": 9.902524101507959e-06, "loss": 0.9069, "step": 695 }, { "epoch": 0.09, "grad_norm": 0.7753880620002747, "learning_rate": 9.902106673099146e-06, "loss": 1.119, "step": 696 }, { "epoch": 0.09, "grad_norm": 1.425355315208435, "learning_rate": 9.901688361642283e-06, "loss": 0.8584, "step": 697 }, { "epoch": 0.09, "grad_norm": 1.4899628162384033, "learning_rate": 9.901269167212726e-06, "loss": 0.9454, "step": 698 }, { "epoch": 0.09, "grad_norm": 1.58889639377594, "learning_rate": 9.900849089885987e-06, "loss": 0.9295, "step": 699 }, { "epoch": 0.09, "grad_norm": 0.9516058564186096, "learning_rate": 9.900428129737737e-06, "loss": 1.078, "step": 700 }, { "epoch": 0.09, "grad_norm": 1.4897679090499878, "learning_rate": 9.900006286843804e-06, "loss": 0.9123, "step": 701 }, { "epoch": 0.09, "grad_norm": 1.4276320934295654, "learning_rate": 9.89958356128018e-06, "loss": 0.8932, "step": 702 }, { "epoch": 0.09, "grad_norm": 1.9791690111160278, "learning_rate": 9.899159953123012e-06, "loss": 0.8974, "step": 703 }, { "epoch": 0.09, "grad_norm": 1.8400890827178955, "learning_rate": 9.898735462448608e-06, "loss": 0.8825, "step": 704 }, { "epoch": 0.09, "grad_norm": 1.3817170858383179, "learning_rate": 9.898310089333433e-06, "loss": 0.8912, "step": 705 }, { "epoch": 0.09, "grad_norm": 1.509076714515686, "learning_rate": 9.897883833854113e-06, "loss": 0.9188, "step": 706 }, { "epoch": 0.09, "grad_norm": 0.9686797261238098, "learning_rate": 9.897456696087433e-06, "loss": 1.1068, "step": 707 }, { "epoch": 0.09, "grad_norm": 0.7312049865722656, "learning_rate": 9.897028676110337e-06, "loss": 1.1173, "step": 708 }, { "epoch": 0.09, "grad_norm": 1.5283945798873901, "learning_rate": 9.896599773999922e-06, "loss": 0.9091, "step": 709 }, { "epoch": 0.09, "grad_norm": 1.4477241039276123, "learning_rate": 9.896169989833453e-06, "loss": 0.9124, "step": 710 }, { "epoch": 0.09, "grad_norm": 1.4504585266113281, "learning_rate": 9.895739323688348e-06, "loss": 0.9131, "step": 711 }, { "epoch": 0.09, "grad_norm": 1.4241914749145508, "learning_rate": 9.895307775642187e-06, "loss": 0.852, "step": 712 }, { "epoch": 0.09, "grad_norm": 1.3404731750488281, "learning_rate": 9.894875345772707e-06, "loss": 0.8932, "step": 713 }, { "epoch": 0.09, "grad_norm": 1.7421597242355347, "learning_rate": 9.894442034157806e-06, "loss": 1.1253, "step": 714 }, { "epoch": 0.09, "grad_norm": 1.4540709257125854, "learning_rate": 9.894007840875537e-06, "loss": 0.8651, "step": 715 }, { "epoch": 0.09, "grad_norm": 1.0198307037353516, "learning_rate": 9.893572766004113e-06, "loss": 1.1138, "step": 716 }, { "epoch": 0.09, "grad_norm": 1.6160597801208496, "learning_rate": 9.893136809621911e-06, "loss": 0.9099, "step": 717 }, { "epoch": 0.09, "grad_norm": 2.072018623352051, "learning_rate": 9.89269997180746e-06, "loss": 0.9039, "step": 718 }, { "epoch": 0.09, "grad_norm": 1.4258978366851807, "learning_rate": 9.892262252639448e-06, "loss": 0.8806, "step": 719 }, { "epoch": 0.09, "grad_norm": 1.5746345520019531, "learning_rate": 9.891823652196728e-06, "loss": 0.9345, "step": 720 }, { "epoch": 0.09, "grad_norm": 1.4617137908935547, "learning_rate": 9.891384170558306e-06, "loss": 0.9202, "step": 721 }, { "epoch": 0.09, "grad_norm": 1.370076060295105, "learning_rate": 9.890943807803353e-06, "loss": 0.8345, "step": 722 }, { "epoch": 0.09, "grad_norm": 1.493896245956421, "learning_rate": 9.890502564011186e-06, "loss": 0.924, "step": 723 }, { "epoch": 0.09, "grad_norm": 1.4552048444747925, "learning_rate": 9.890060439261298e-06, "loss": 0.9303, "step": 724 }, { "epoch": 0.09, "grad_norm": 2.370812177658081, "learning_rate": 9.889617433633326e-06, "loss": 1.1093, "step": 725 }, { "epoch": 0.1, "grad_norm": 1.8691112995147705, "learning_rate": 9.889173547207072e-06, "loss": 1.1458, "step": 726 }, { "epoch": 0.1, "grad_norm": 1.5203043222427368, "learning_rate": 9.888728780062496e-06, "loss": 0.9404, "step": 727 }, { "epoch": 0.1, "grad_norm": 1.6341633796691895, "learning_rate": 9.888283132279719e-06, "loss": 0.9484, "step": 728 }, { "epoch": 0.1, "grad_norm": 2.035503387451172, "learning_rate": 9.887836603939019e-06, "loss": 0.8543, "step": 729 }, { "epoch": 0.1, "grad_norm": 1.5410994291305542, "learning_rate": 9.887389195120829e-06, "loss": 0.9361, "step": 730 }, { "epoch": 0.1, "grad_norm": 1.4091203212738037, "learning_rate": 9.886940905905745e-06, "loss": 0.9262, "step": 731 }, { "epoch": 0.1, "grad_norm": 1.4543572664260864, "learning_rate": 9.886491736374521e-06, "loss": 1.1253, "step": 732 }, { "epoch": 0.1, "grad_norm": 1.300648808479309, "learning_rate": 9.886041686608067e-06, "loss": 1.1113, "step": 733 }, { "epoch": 0.1, "grad_norm": 1.6162737607955933, "learning_rate": 9.885590756687456e-06, "loss": 0.8758, "step": 734 }, { "epoch": 0.1, "grad_norm": 1.734884262084961, "learning_rate": 9.885138946693912e-06, "loss": 0.8844, "step": 735 }, { "epoch": 0.1, "grad_norm": 1.579867959022522, "learning_rate": 9.884686256708829e-06, "loss": 0.8818, "step": 736 }, { "epoch": 0.1, "grad_norm": 1.5095404386520386, "learning_rate": 9.884232686813748e-06, "loss": 0.8942, "step": 737 }, { "epoch": 0.1, "grad_norm": 1.4160809516906738, "learning_rate": 9.883778237090375e-06, "loss": 0.9307, "step": 738 }, { "epoch": 0.1, "grad_norm": 1.3186970949172974, "learning_rate": 9.883322907620575e-06, "loss": 0.9792, "step": 739 }, { "epoch": 0.1, "grad_norm": 1.3688945770263672, "learning_rate": 9.882866698486365e-06, "loss": 0.9604, "step": 740 }, { "epoch": 0.1, "grad_norm": 1.1893435716629028, "learning_rate": 9.88240960976993e-06, "loss": 1.1373, "step": 741 }, { "epoch": 0.1, "grad_norm": 1.019599199295044, "learning_rate": 9.881951641553604e-06, "loss": 1.0977, "step": 742 }, { "epoch": 0.1, "grad_norm": 1.6349986791610718, "learning_rate": 9.881492793919885e-06, "loss": 0.9073, "step": 743 }, { "epoch": 0.1, "grad_norm": 1.5114550590515137, "learning_rate": 9.881033066951432e-06, "loss": 0.8568, "step": 744 }, { "epoch": 0.1, "grad_norm": 1.4682226181030273, "learning_rate": 9.88057246073105e-06, "loss": 0.8737, "step": 745 }, { "epoch": 0.1, "grad_norm": 1.437742829322815, "learning_rate": 9.880110975341721e-06, "loss": 0.9242, "step": 746 }, { "epoch": 0.1, "grad_norm": 1.4206645488739014, "learning_rate": 9.879648610866567e-06, "loss": 0.8694, "step": 747 }, { "epoch": 0.1, "grad_norm": 1.533833384513855, "learning_rate": 9.879185367388883e-06, "loss": 0.88, "step": 748 }, { "epoch": 0.1, "grad_norm": 1.4494678974151611, "learning_rate": 9.878721244992113e-06, "loss": 0.9179, "step": 749 }, { "epoch": 0.1, "grad_norm": 1.6451377868652344, "learning_rate": 9.878256243759863e-06, "loss": 0.8534, "step": 750 }, { "epoch": 0.1, "grad_norm": 1.6167254447937012, "learning_rate": 9.877790363775894e-06, "loss": 0.8926, "step": 751 }, { "epoch": 0.1, "grad_norm": 1.8674066066741943, "learning_rate": 9.877323605124131e-06, "loss": 1.1402, "step": 752 }, { "epoch": 0.1, "grad_norm": 1.4537386894226074, "learning_rate": 9.876855967888656e-06, "loss": 0.8918, "step": 753 }, { "epoch": 0.1, "grad_norm": 1.6709058284759521, "learning_rate": 9.876387452153702e-06, "loss": 0.8936, "step": 754 }, { "epoch": 0.1, "grad_norm": 1.7530885934829712, "learning_rate": 9.87591805800367e-06, "loss": 0.9095, "step": 755 }, { "epoch": 0.1, "grad_norm": 1.394126057624817, "learning_rate": 9.875447785523116e-06, "loss": 0.8426, "step": 756 }, { "epoch": 0.1, "grad_norm": 1.3686985969543457, "learning_rate": 9.874976634796747e-06, "loss": 0.9062, "step": 757 }, { "epoch": 0.1, "grad_norm": 1.4387723207473755, "learning_rate": 9.874504605909441e-06, "loss": 0.8298, "step": 758 }, { "epoch": 0.1, "grad_norm": 1.3426527976989746, "learning_rate": 9.874031698946225e-06, "loss": 0.8228, "step": 759 }, { "epoch": 0.1, "grad_norm": 1.0364350080490112, "learning_rate": 9.873557913992286e-06, "loss": 1.1105, "step": 760 }, { "epoch": 0.1, "grad_norm": 1.4421563148498535, "learning_rate": 9.87308325113297e-06, "loss": 0.8871, "step": 761 }, { "epoch": 0.1, "grad_norm": 1.4720102548599243, "learning_rate": 9.872607710453784e-06, "loss": 0.9113, "step": 762 }, { "epoch": 0.1, "grad_norm": 1.452410340309143, "learning_rate": 9.872131292040387e-06, "loss": 0.8886, "step": 763 }, { "epoch": 0.1, "grad_norm": 0.7727645635604858, "learning_rate": 9.871653995978599e-06, "loss": 1.0993, "step": 764 }, { "epoch": 0.1, "grad_norm": 1.3914252519607544, "learning_rate": 9.871175822354401e-06, "loss": 0.9064, "step": 765 }, { "epoch": 0.1, "grad_norm": 1.5801339149475098, "learning_rate": 9.87069677125393e-06, "loss": 0.9413, "step": 766 }, { "epoch": 0.1, "grad_norm": 0.7446367740631104, "learning_rate": 9.870216842763476e-06, "loss": 1.1251, "step": 767 }, { "epoch": 0.1, "grad_norm": 1.3629376888275146, "learning_rate": 9.869736036969495e-06, "loss": 0.8696, "step": 768 }, { "epoch": 0.1, "grad_norm": 0.6673344373703003, "learning_rate": 9.869254353958597e-06, "loss": 1.1369, "step": 769 }, { "epoch": 0.1, "grad_norm": 1.3835548162460327, "learning_rate": 9.86877179381755e-06, "loss": 0.9224, "step": 770 }, { "epoch": 0.1, "grad_norm": 1.3425287008285522, "learning_rate": 9.868288356633284e-06, "loss": 0.8938, "step": 771 }, { "epoch": 0.1, "grad_norm": 1.4843101501464844, "learning_rate": 9.867804042492878e-06, "loss": 0.914, "step": 772 }, { "epoch": 0.1, "grad_norm": 1.4511003494262695, "learning_rate": 9.86731885148358e-06, "loss": 0.8783, "step": 773 }, { "epoch": 0.1, "grad_norm": 1.3445926904678345, "learning_rate": 9.866832783692788e-06, "loss": 0.872, "step": 774 }, { "epoch": 0.1, "grad_norm": 1.3204299211502075, "learning_rate": 9.86634583920806e-06, "loss": 0.8409, "step": 775 }, { "epoch": 0.1, "grad_norm": 1.1074345111846924, "learning_rate": 9.865858018117113e-06, "loss": 1.1174, "step": 776 }, { "epoch": 0.1, "grad_norm": 1.5174676179885864, "learning_rate": 9.865369320507823e-06, "loss": 0.915, "step": 777 }, { "epoch": 0.1, "grad_norm": 1.3798490762710571, "learning_rate": 9.86487974646822e-06, "loss": 0.841, "step": 778 }, { "epoch": 0.1, "grad_norm": 0.7438609600067139, "learning_rate": 9.864389296086494e-06, "loss": 1.1339, "step": 779 }, { "epoch": 0.1, "grad_norm": 1.4742200374603271, "learning_rate": 9.863897969450994e-06, "loss": 0.897, "step": 780 }, { "epoch": 0.1, "grad_norm": 1.640984058380127, "learning_rate": 9.863405766650228e-06, "loss": 0.8767, "step": 781 }, { "epoch": 0.1, "grad_norm": 0.848603367805481, "learning_rate": 9.862912687772857e-06, "loss": 1.1075, "step": 782 }, { "epoch": 0.1, "grad_norm": 0.653846800327301, "learning_rate": 9.862418732907702e-06, "loss": 1.1293, "step": 783 }, { "epoch": 0.1, "grad_norm": 1.444520354270935, "learning_rate": 9.861923902143742e-06, "loss": 0.9026, "step": 784 }, { "epoch": 0.1, "grad_norm": 1.7201651334762573, "learning_rate": 9.861428195570116e-06, "loss": 0.9196, "step": 785 }, { "epoch": 0.1, "grad_norm": 1.4426990747451782, "learning_rate": 9.860931613276118e-06, "loss": 0.795, "step": 786 }, { "epoch": 0.1, "grad_norm": 1.0104615688323975, "learning_rate": 9.860434155351202e-06, "loss": 1.1379, "step": 787 }, { "epoch": 0.1, "grad_norm": 1.3463772535324097, "learning_rate": 9.859935821884975e-06, "loss": 0.8835, "step": 788 }, { "epoch": 0.1, "grad_norm": 1.4633132219314575, "learning_rate": 9.859436612967208e-06, "loss": 0.8508, "step": 789 }, { "epoch": 0.1, "grad_norm": 1.5253889560699463, "learning_rate": 9.858936528687826e-06, "loss": 0.8512, "step": 790 }, { "epoch": 0.1, "grad_norm": 1.4655793905258179, "learning_rate": 9.85843556913691e-06, "loss": 0.8624, "step": 791 }, { "epoch": 0.1, "grad_norm": 1.4208614826202393, "learning_rate": 9.857933734404705e-06, "loss": 0.8498, "step": 792 }, { "epoch": 0.1, "grad_norm": 1.3766721487045288, "learning_rate": 9.857431024581607e-06, "loss": 0.8427, "step": 793 }, { "epoch": 0.1, "grad_norm": 0.846320390701294, "learning_rate": 9.856927439758174e-06, "loss": 1.1235, "step": 794 }, { "epoch": 0.1, "grad_norm": 1.3872582912445068, "learning_rate": 9.856422980025119e-06, "loss": 0.8751, "step": 795 }, { "epoch": 0.1, "grad_norm": 1.5294125080108643, "learning_rate": 9.855917645473314e-06, "loss": 0.8376, "step": 796 }, { "epoch": 0.1, "grad_norm": 1.5088386535644531, "learning_rate": 9.855411436193788e-06, "loss": 0.8891, "step": 797 }, { "epoch": 0.1, "grad_norm": 1.569964051246643, "learning_rate": 9.854904352277725e-06, "loss": 0.894, "step": 798 }, { "epoch": 0.1, "grad_norm": 1.5247586965560913, "learning_rate": 9.854396393816474e-06, "loss": 0.9124, "step": 799 }, { "epoch": 0.1, "grad_norm": 1.393784999847412, "learning_rate": 9.853887560901537e-06, "loss": 0.9178, "step": 800 }, { "epoch": 0.1, "grad_norm": 1.3788208961486816, "learning_rate": 9.853377853624568e-06, "loss": 0.878, "step": 801 }, { "epoch": 0.11, "grad_norm": 1.3953132629394531, "learning_rate": 9.852867272077387e-06, "loss": 0.9155, "step": 802 }, { "epoch": 0.11, "grad_norm": 0.8565292954444885, "learning_rate": 9.85235581635197e-06, "loss": 1.1291, "step": 803 }, { "epoch": 0.11, "grad_norm": 1.408879041671753, "learning_rate": 9.851843486540447e-06, "loss": 0.902, "step": 804 }, { "epoch": 0.11, "grad_norm": 1.5604192018508911, "learning_rate": 9.851330282735106e-06, "loss": 0.8783, "step": 805 }, { "epoch": 0.11, "grad_norm": 1.4830436706542969, "learning_rate": 9.850816205028397e-06, "loss": 0.9254, "step": 806 }, { "epoch": 0.11, "grad_norm": 1.4956260919570923, "learning_rate": 9.85030125351292e-06, "loss": 0.9514, "step": 807 }, { "epoch": 0.11, "grad_norm": 1.3864753246307373, "learning_rate": 9.84978542828144e-06, "loss": 0.8138, "step": 808 }, { "epoch": 0.11, "grad_norm": 0.9673091173171997, "learning_rate": 9.849268729426874e-06, "loss": 1.1096, "step": 809 }, { "epoch": 0.11, "grad_norm": 1.4280766248703003, "learning_rate": 9.8487511570423e-06, "loss": 0.8483, "step": 810 }, { "epoch": 0.11, "grad_norm": 1.3508570194244385, "learning_rate": 9.84823271122095e-06, "loss": 0.9194, "step": 811 }, { "epoch": 0.11, "grad_norm": 0.7663477659225464, "learning_rate": 9.847713392056218e-06, "loss": 1.1029, "step": 812 }, { "epoch": 0.11, "grad_norm": 0.7287114858627319, "learning_rate": 9.847193199641648e-06, "loss": 1.1088, "step": 813 }, { "epoch": 0.11, "grad_norm": 1.6559542417526245, "learning_rate": 9.846672134070947e-06, "loss": 0.9129, "step": 814 }, { "epoch": 0.11, "grad_norm": 1.5535616874694824, "learning_rate": 9.846150195437978e-06, "loss": 0.8852, "step": 815 }, { "epoch": 0.11, "grad_norm": 1.543023943901062, "learning_rate": 9.845627383836763e-06, "loss": 0.8727, "step": 816 }, { "epoch": 0.11, "grad_norm": 1.3398256301879883, "learning_rate": 9.845103699361477e-06, "loss": 0.8994, "step": 817 }, { "epoch": 0.11, "grad_norm": 1.3506790399551392, "learning_rate": 9.844579142106456e-06, "loss": 0.856, "step": 818 }, { "epoch": 0.11, "grad_norm": 1.43305504322052, "learning_rate": 9.844053712166192e-06, "loss": 0.8964, "step": 819 }, { "epoch": 0.11, "grad_norm": 1.473858118057251, "learning_rate": 9.843527409635334e-06, "loss": 0.9196, "step": 820 }, { "epoch": 0.11, "grad_norm": 1.3658828735351562, "learning_rate": 9.843000234608686e-06, "loss": 0.9721, "step": 821 }, { "epoch": 0.11, "grad_norm": 1.367032527923584, "learning_rate": 9.842472187181214e-06, "loss": 0.8977, "step": 822 }, { "epoch": 0.11, "grad_norm": 1.48551607131958, "learning_rate": 9.84194326744804e-06, "loss": 0.9394, "step": 823 }, { "epoch": 0.11, "grad_norm": 1.366186261177063, "learning_rate": 9.841413475504438e-06, "loss": 1.1166, "step": 824 }, { "epoch": 0.11, "grad_norm": 1.7670689821243286, "learning_rate": 9.840882811445844e-06, "loss": 0.9073, "step": 825 }, { "epoch": 0.11, "grad_norm": 1.4508744478225708, "learning_rate": 9.84035127536785e-06, "loss": 0.9369, "step": 826 }, { "epoch": 0.11, "grad_norm": 1.5064738988876343, "learning_rate": 9.839818867366208e-06, "loss": 0.8705, "step": 827 }, { "epoch": 0.11, "grad_norm": 1.477091670036316, "learning_rate": 9.83928558753682e-06, "loss": 0.8514, "step": 828 }, { "epoch": 0.11, "grad_norm": 1.5831011533737183, "learning_rate": 9.838751435975751e-06, "loss": 0.9001, "step": 829 }, { "epoch": 0.11, "grad_norm": 1.4713047742843628, "learning_rate": 9.83821641277922e-06, "loss": 0.8966, "step": 830 }, { "epoch": 0.11, "grad_norm": 1.501590609550476, "learning_rate": 9.837680518043605e-06, "loss": 1.1236, "step": 831 }, { "epoch": 0.11, "grad_norm": 1.517800211906433, "learning_rate": 9.837143751865442e-06, "loss": 0.8892, "step": 832 }, { "epoch": 0.11, "grad_norm": 1.4091768264770508, "learning_rate": 9.836606114341418e-06, "loss": 0.92, "step": 833 }, { "epoch": 0.11, "grad_norm": 1.3287079334259033, "learning_rate": 9.836067605568384e-06, "loss": 0.8934, "step": 834 }, { "epoch": 0.11, "grad_norm": 0.7625091671943665, "learning_rate": 9.835528225643346e-06, "loss": 1.1259, "step": 835 }, { "epoch": 0.11, "grad_norm": 0.786653995513916, "learning_rate": 9.834987974663461e-06, "loss": 1.0868, "step": 836 }, { "epoch": 0.11, "grad_norm": 1.5339269638061523, "learning_rate": 9.834446852726053e-06, "loss": 0.8903, "step": 837 }, { "epoch": 0.11, "grad_norm": 1.3778269290924072, "learning_rate": 9.833904859928599e-06, "loss": 0.9228, "step": 838 }, { "epoch": 0.11, "grad_norm": 1.0042506456375122, "learning_rate": 9.833361996368725e-06, "loss": 1.1053, "step": 839 }, { "epoch": 0.11, "grad_norm": 2.47739577293396, "learning_rate": 9.832818262144226e-06, "loss": 0.9255, "step": 840 }, { "epoch": 0.11, "grad_norm": 0.6812829971313477, "learning_rate": 9.832273657353046e-06, "loss": 1.1068, "step": 841 }, { "epoch": 0.11, "grad_norm": 0.7001181244850159, "learning_rate": 9.831728182093289e-06, "loss": 1.1213, "step": 842 }, { "epoch": 0.11, "grad_norm": 1.466845989227295, "learning_rate": 9.831181836463216e-06, "loss": 0.9133, "step": 843 }, { "epoch": 0.11, "grad_norm": 1.5123558044433594, "learning_rate": 9.83063462056124e-06, "loss": 0.9037, "step": 844 }, { "epoch": 0.11, "grad_norm": 1.5116639137268066, "learning_rate": 9.83008653448594e-06, "loss": 0.8408, "step": 845 }, { "epoch": 0.11, "grad_norm": 1.3863539695739746, "learning_rate": 9.82953757833604e-06, "loss": 0.8617, "step": 846 }, { "epoch": 0.11, "grad_norm": 1.3345741033554077, "learning_rate": 9.828987752210432e-06, "loss": 0.8414, "step": 847 }, { "epoch": 0.11, "grad_norm": 1.4743430614471436, "learning_rate": 9.828437056208158e-06, "loss": 0.9196, "step": 848 }, { "epoch": 0.11, "grad_norm": 1.308967113494873, "learning_rate": 9.827885490428419e-06, "loss": 0.8683, "step": 849 }, { "epoch": 0.11, "grad_norm": 1.3381526470184326, "learning_rate": 9.827333054970572e-06, "loss": 0.8991, "step": 850 }, { "epoch": 0.11, "grad_norm": 1.35189950466156, "learning_rate": 9.826779749934128e-06, "loss": 0.8644, "step": 851 }, { "epoch": 0.11, "grad_norm": 1.3446916341781616, "learning_rate": 9.826225575418762e-06, "loss": 0.9632, "step": 852 }, { "epoch": 0.11, "grad_norm": 1.4919034242630005, "learning_rate": 9.825670531524295e-06, "loss": 0.8614, "step": 853 }, { "epoch": 0.11, "grad_norm": 1.5601634979248047, "learning_rate": 9.82511461835072e-06, "loss": 0.8568, "step": 854 }, { "epoch": 0.11, "grad_norm": 1.3073718547821045, "learning_rate": 9.824557835998168e-06, "loss": 0.8826, "step": 855 }, { "epoch": 0.11, "grad_norm": 1.421726942062378, "learning_rate": 9.82400018456694e-06, "loss": 0.8441, "step": 856 }, { "epoch": 0.11, "grad_norm": 1.2883778810501099, "learning_rate": 9.823441664157489e-06, "loss": 0.828, "step": 857 }, { "epoch": 0.11, "grad_norm": 1.251497507095337, "learning_rate": 9.822882274870424e-06, "loss": 1.1079, "step": 858 }, { "epoch": 0.11, "grad_norm": 1.6879445314407349, "learning_rate": 9.822322016806514e-06, "loss": 0.8602, "step": 859 }, { "epoch": 0.11, "grad_norm": 1.3709065914154053, "learning_rate": 9.821760890066677e-06, "loss": 0.8697, "step": 860 }, { "epoch": 0.11, "grad_norm": 1.3128007650375366, "learning_rate": 9.821198894751998e-06, "loss": 0.8223, "step": 861 }, { "epoch": 0.11, "grad_norm": 0.8389681577682495, "learning_rate": 9.820636030963708e-06, "loss": 1.1133, "step": 862 }, { "epoch": 0.11, "grad_norm": 1.4501254558563232, "learning_rate": 9.820072298803203e-06, "loss": 0.9014, "step": 863 }, { "epoch": 0.11, "grad_norm": 0.7097868323326111, "learning_rate": 9.819507698372028e-06, "loss": 1.0904, "step": 864 }, { "epoch": 0.11, "grad_norm": 1.330877661705017, "learning_rate": 9.818942229771892e-06, "loss": 0.8973, "step": 865 }, { "epoch": 0.11, "grad_norm": 1.4190847873687744, "learning_rate": 9.818375893104654e-06, "loss": 0.9032, "step": 866 }, { "epoch": 0.11, "grad_norm": 0.8320370316505432, "learning_rate": 9.817808688472331e-06, "loss": 1.1103, "step": 867 }, { "epoch": 0.11, "grad_norm": 1.3615670204162598, "learning_rate": 9.817240615977102e-06, "loss": 0.8836, "step": 868 }, { "epoch": 0.11, "grad_norm": 0.7136479616165161, "learning_rate": 9.816671675721292e-06, "loss": 1.1054, "step": 869 }, { "epoch": 0.11, "grad_norm": 1.4444630146026611, "learning_rate": 9.816101867807393e-06, "loss": 0.9042, "step": 870 }, { "epoch": 0.11, "grad_norm": 0.7079001069068909, "learning_rate": 9.815531192338042e-06, "loss": 1.081, "step": 871 }, { "epoch": 0.11, "grad_norm": 0.6935680508613586, "learning_rate": 9.814959649416044e-06, "loss": 1.0984, "step": 872 }, { "epoch": 0.11, "grad_norm": 1.4125357866287231, "learning_rate": 9.814387239144352e-06, "loss": 0.8965, "step": 873 }, { "epoch": 0.11, "grad_norm": 1.4230444431304932, "learning_rate": 9.813813961626078e-06, "loss": 0.8821, "step": 874 }, { "epoch": 0.11, "grad_norm": 0.6882942318916321, "learning_rate": 9.81323981696449e-06, "loss": 1.1098, "step": 875 }, { "epoch": 0.11, "grad_norm": 0.6731878519058228, "learning_rate": 9.812664805263015e-06, "loss": 1.0882, "step": 876 }, { "epoch": 0.11, "grad_norm": 0.6358586549758911, "learning_rate": 9.812088926625231e-06, "loss": 1.0946, "step": 877 }, { "epoch": 0.12, "grad_norm": 0.6346225738525391, "learning_rate": 9.811512181154876e-06, "loss": 1.1165, "step": 878 }, { "epoch": 0.12, "grad_norm": 1.3571605682373047, "learning_rate": 9.810934568955842e-06, "loss": 0.8788, "step": 879 }, { "epoch": 0.12, "grad_norm": 1.3564969301223755, "learning_rate": 9.810356090132176e-06, "loss": 0.9056, "step": 880 }, { "epoch": 0.12, "grad_norm": 1.6369811296463013, "learning_rate": 9.809776744788086e-06, "loss": 0.8588, "step": 881 }, { "epoch": 0.12, "grad_norm": 1.3714869022369385, "learning_rate": 9.809196533027934e-06, "loss": 0.8897, "step": 882 }, { "epoch": 0.12, "grad_norm": 0.840480625629425, "learning_rate": 9.808615454956235e-06, "loss": 1.1061, "step": 883 }, { "epoch": 0.12, "grad_norm": 1.4040305614471436, "learning_rate": 9.808033510677663e-06, "loss": 0.916, "step": 884 }, { "epoch": 0.12, "grad_norm": 1.3611023426055908, "learning_rate": 9.807450700297048e-06, "loss": 0.8898, "step": 885 }, { "epoch": 0.12, "grad_norm": 1.4571062326431274, "learning_rate": 9.806867023919376e-06, "loss": 0.9233, "step": 886 }, { "epoch": 0.12, "grad_norm": 0.7808288335800171, "learning_rate": 9.806282481649785e-06, "loss": 1.139, "step": 887 }, { "epoch": 0.12, "grad_norm": 1.4413596391677856, "learning_rate": 9.805697073593577e-06, "loss": 0.9221, "step": 888 }, { "epoch": 0.12, "grad_norm": 1.485229253768921, "learning_rate": 9.805110799856201e-06, "loss": 0.9826, "step": 889 }, { "epoch": 0.12, "grad_norm": 1.3383777141571045, "learning_rate": 9.804523660543269e-06, "loss": 0.8859, "step": 890 }, { "epoch": 0.12, "grad_norm": 1.3790655136108398, "learning_rate": 9.803935655760546e-06, "loss": 0.8721, "step": 891 }, { "epoch": 0.12, "grad_norm": 0.8521832823753357, "learning_rate": 9.803346785613952e-06, "loss": 1.0862, "step": 892 }, { "epoch": 0.12, "grad_norm": 0.7341249585151672, "learning_rate": 9.802757050209563e-06, "loss": 1.0937, "step": 893 }, { "epoch": 0.12, "grad_norm": 1.346872329711914, "learning_rate": 9.802166449653616e-06, "loss": 0.8922, "step": 894 }, { "epoch": 0.12, "grad_norm": 1.3798168897628784, "learning_rate": 9.801574984052495e-06, "loss": 0.97, "step": 895 }, { "epoch": 0.12, "grad_norm": 1.355987310409546, "learning_rate": 9.800982653512747e-06, "loss": 0.8928, "step": 896 }, { "epoch": 0.12, "grad_norm": 1.4280612468719482, "learning_rate": 9.800389458141073e-06, "loss": 0.8864, "step": 897 }, { "epoch": 0.12, "grad_norm": 1.41385817527771, "learning_rate": 9.799795398044328e-06, "loss": 0.8993, "step": 898 }, { "epoch": 0.12, "grad_norm": 1.3677647113800049, "learning_rate": 9.799200473329522e-06, "loss": 0.9071, "step": 899 }, { "epoch": 0.12, "grad_norm": 1.3201531171798706, "learning_rate": 9.798604684103824e-06, "loss": 0.9349, "step": 900 }, { "epoch": 0.12, "grad_norm": 1.3757593631744385, "learning_rate": 9.79800803047456e-06, "loss": 0.9221, "step": 901 }, { "epoch": 0.12, "grad_norm": 1.3938688039779663, "learning_rate": 9.797410512549207e-06, "loss": 0.8904, "step": 902 }, { "epoch": 0.12, "grad_norm": 1.3370075225830078, "learning_rate": 9.796812130435398e-06, "loss": 0.8838, "step": 903 }, { "epoch": 0.12, "grad_norm": 1.3492234945297241, "learning_rate": 9.796212884240925e-06, "loss": 0.9075, "step": 904 }, { "epoch": 0.12, "grad_norm": 1.527876377105713, "learning_rate": 9.795612774073737e-06, "loss": 0.9006, "step": 905 }, { "epoch": 0.12, "grad_norm": 1.3141062259674072, "learning_rate": 9.79501180004193e-06, "loss": 0.8934, "step": 906 }, { "epoch": 0.12, "grad_norm": 1.2802561521530151, "learning_rate": 9.794409962253764e-06, "loss": 0.8101, "step": 907 }, { "epoch": 0.12, "grad_norm": 1.2537505626678467, "learning_rate": 9.793807260817653e-06, "loss": 0.8547, "step": 908 }, { "epoch": 0.12, "grad_norm": 1.4451417922973633, "learning_rate": 9.793203695842163e-06, "loss": 0.9049, "step": 909 }, { "epoch": 0.12, "grad_norm": 1.3361825942993164, "learning_rate": 9.792599267436022e-06, "loss": 0.8722, "step": 910 }, { "epoch": 0.12, "grad_norm": 1.4640637636184692, "learning_rate": 9.791993975708107e-06, "loss": 0.8709, "step": 911 }, { "epoch": 0.12, "grad_norm": 1.3704324960708618, "learning_rate": 9.791387820767454e-06, "loss": 0.9204, "step": 912 }, { "epoch": 0.12, "grad_norm": 1.466179609298706, "learning_rate": 9.79078080272325e-06, "loss": 0.8574, "step": 913 }, { "epoch": 0.12, "grad_norm": 1.4459888935089111, "learning_rate": 9.790172921684844e-06, "loss": 0.9077, "step": 914 }, { "epoch": 0.12, "grad_norm": 1.3442072868347168, "learning_rate": 9.78956417776174e-06, "loss": 0.8587, "step": 915 }, { "epoch": 0.12, "grad_norm": 1.3378477096557617, "learning_rate": 9.78895457106359e-06, "loss": 0.8724, "step": 916 }, { "epoch": 0.12, "grad_norm": 1.3742400407791138, "learning_rate": 9.78834410170021e-06, "loss": 0.89, "step": 917 }, { "epoch": 0.12, "grad_norm": 1.699657678604126, "learning_rate": 9.787732769781566e-06, "loss": 1.0985, "step": 918 }, { "epoch": 0.12, "grad_norm": 1.366969347000122, "learning_rate": 9.787120575417783e-06, "loss": 0.8424, "step": 919 }, { "epoch": 0.12, "grad_norm": 1.3734949827194214, "learning_rate": 9.786507518719138e-06, "loss": 0.8769, "step": 920 }, { "epoch": 0.12, "grad_norm": 1.3316668272018433, "learning_rate": 9.785893599796064e-06, "loss": 0.8459, "step": 921 }, { "epoch": 0.12, "grad_norm": 1.3765772581100464, "learning_rate": 9.785278818759153e-06, "loss": 0.9453, "step": 922 }, { "epoch": 0.12, "grad_norm": 1.3415788412094116, "learning_rate": 9.784663175719145e-06, "loss": 0.8541, "step": 923 }, { "epoch": 0.12, "grad_norm": 1.39987313747406, "learning_rate": 9.784046670786944e-06, "loss": 0.862, "step": 924 }, { "epoch": 0.12, "grad_norm": 1.2061628103256226, "learning_rate": 9.783429304073604e-06, "loss": 1.1156, "step": 925 }, { "epoch": 0.12, "grad_norm": 0.8152620792388916, "learning_rate": 9.782811075690335e-06, "loss": 1.1141, "step": 926 }, { "epoch": 0.12, "grad_norm": 1.426748514175415, "learning_rate": 9.782191985748502e-06, "loss": 0.8909, "step": 927 }, { "epoch": 0.12, "grad_norm": 1.6210740804672241, "learning_rate": 9.781572034359624e-06, "loss": 0.9014, "step": 928 }, { "epoch": 0.12, "grad_norm": 1.3547043800354004, "learning_rate": 9.780951221635379e-06, "loss": 0.8828, "step": 929 }, { "epoch": 0.12, "grad_norm": 1.479280948638916, "learning_rate": 9.7803295476876e-06, "loss": 0.9137, "step": 930 }, { "epoch": 0.12, "grad_norm": 1.3838508129119873, "learning_rate": 9.77970701262827e-06, "loss": 0.9176, "step": 931 }, { "epoch": 0.12, "grad_norm": 2.4237453937530518, "learning_rate": 9.77908361656953e-06, "loss": 1.1174, "step": 932 }, { "epoch": 0.12, "grad_norm": 1.3491950035095215, "learning_rate": 9.778459359623678e-06, "loss": 0.9037, "step": 933 }, { "epoch": 0.12, "grad_norm": 1.4246842861175537, "learning_rate": 9.777834241903166e-06, "loss": 0.9236, "step": 934 }, { "epoch": 0.12, "grad_norm": 1.384809136390686, "learning_rate": 9.777208263520597e-06, "loss": 0.9077, "step": 935 }, { "epoch": 0.12, "grad_norm": 1.3866328001022339, "learning_rate": 9.776581424588736e-06, "loss": 0.8866, "step": 936 }, { "epoch": 0.12, "grad_norm": 1.347657322883606, "learning_rate": 9.7759537252205e-06, "loss": 0.8927, "step": 937 }, { "epoch": 0.12, "grad_norm": 1.4483144283294678, "learning_rate": 9.775325165528957e-06, "loss": 0.891, "step": 938 }, { "epoch": 0.12, "grad_norm": 1.36770761013031, "learning_rate": 9.774695745627335e-06, "loss": 0.8573, "step": 939 }, { "epoch": 0.12, "grad_norm": 1.4672572612762451, "learning_rate": 9.774065465629017e-06, "loss": 0.9302, "step": 940 }, { "epoch": 0.12, "grad_norm": 1.1163315773010254, "learning_rate": 9.773434325647537e-06, "loss": 1.1335, "step": 941 }, { "epoch": 0.12, "grad_norm": 1.369384765625, "learning_rate": 9.77280232579659e-06, "loss": 0.8478, "step": 942 }, { "epoch": 0.12, "grad_norm": 1.4033691883087158, "learning_rate": 9.772169466190017e-06, "loss": 0.8874, "step": 943 }, { "epoch": 0.12, "grad_norm": 1.2762619256973267, "learning_rate": 9.771535746941822e-06, "loss": 0.8339, "step": 944 }, { "epoch": 0.12, "grad_norm": 1.4283263683319092, "learning_rate": 9.770901168166163e-06, "loss": 0.8953, "step": 945 }, { "epoch": 0.12, "grad_norm": 1.288071870803833, "learning_rate": 9.770265729977347e-06, "loss": 0.8732, "step": 946 }, { "epoch": 0.12, "grad_norm": 1.265836477279663, "learning_rate": 9.769629432489842e-06, "loss": 1.1119, "step": 947 }, { "epoch": 0.12, "grad_norm": 1.4001754522323608, "learning_rate": 9.768992275818267e-06, "loss": 0.8848, "step": 948 }, { "epoch": 0.12, "grad_norm": 1.3709064722061157, "learning_rate": 9.768354260077398e-06, "loss": 0.8702, "step": 949 }, { "epoch": 0.12, "grad_norm": 0.7715767621994019, "learning_rate": 9.767715385382163e-06, "loss": 1.1043, "step": 950 }, { "epoch": 0.12, "grad_norm": 0.7078161239624023, "learning_rate": 9.76707565184765e-06, "loss": 1.1031, "step": 951 }, { "epoch": 0.12, "grad_norm": 1.3625563383102417, "learning_rate": 9.766435059589096e-06, "loss": 0.8809, "step": 952 }, { "epoch": 0.12, "grad_norm": 1.4119305610656738, "learning_rate": 9.765793608721894e-06, "loss": 0.8724, "step": 953 }, { "epoch": 0.13, "grad_norm": 1.4198050498962402, "learning_rate": 9.765151299361596e-06, "loss": 0.8778, "step": 954 }, { "epoch": 0.13, "grad_norm": 1.192926287651062, "learning_rate": 9.764508131623902e-06, "loss": 1.0911, "step": 955 }, { "epoch": 0.13, "grad_norm": 1.2825816869735718, "learning_rate": 9.763864105624673e-06, "loss": 0.8285, "step": 956 }, { "epoch": 0.13, "grad_norm": 1.3320741653442383, "learning_rate": 9.763219221479919e-06, "loss": 0.9096, "step": 957 }, { "epoch": 0.13, "grad_norm": 1.4244359731674194, "learning_rate": 9.762573479305807e-06, "loss": 0.9641, "step": 958 }, { "epoch": 0.13, "grad_norm": 0.8997495174407959, "learning_rate": 9.761926879218661e-06, "loss": 1.1166, "step": 959 }, { "epoch": 0.13, "grad_norm": 0.8105767369270325, "learning_rate": 9.761279421334957e-06, "loss": 1.1232, "step": 960 }, { "epoch": 0.13, "grad_norm": 1.391581654548645, "learning_rate": 9.760631105771322e-06, "loss": 0.8292, "step": 961 }, { "epoch": 0.13, "grad_norm": 1.4057297706604004, "learning_rate": 9.759981932644547e-06, "loss": 0.8617, "step": 962 }, { "epoch": 0.13, "grad_norm": 1.3650215864181519, "learning_rate": 9.759331902071569e-06, "loss": 0.9028, "step": 963 }, { "epoch": 0.13, "grad_norm": 1.126091718673706, "learning_rate": 9.75868101416948e-06, "loss": 1.1151, "step": 964 }, { "epoch": 0.13, "grad_norm": 1.046148419380188, "learning_rate": 9.758029269055531e-06, "loss": 1.0984, "step": 965 }, { "epoch": 0.13, "grad_norm": 1.6891731023788452, "learning_rate": 9.757376666847125e-06, "loss": 0.8463, "step": 966 }, { "epoch": 0.13, "grad_norm": 0.6902652382850647, "learning_rate": 9.756723207661818e-06, "loss": 1.0972, "step": 967 }, { "epoch": 0.13, "grad_norm": 0.8954395055770874, "learning_rate": 9.756068891617323e-06, "loss": 1.1201, "step": 968 }, { "epoch": 0.13, "grad_norm": 0.7314707040786743, "learning_rate": 9.755413718831507e-06, "loss": 1.0984, "step": 969 }, { "epoch": 0.13, "grad_norm": 1.4362481832504272, "learning_rate": 9.75475768942239e-06, "loss": 0.8431, "step": 970 }, { "epoch": 0.13, "grad_norm": 1.521817922592163, "learning_rate": 9.754100803508146e-06, "loss": 0.9149, "step": 971 }, { "epoch": 0.13, "grad_norm": 1.392622709274292, "learning_rate": 9.753443061207104e-06, "loss": 0.868, "step": 972 }, { "epoch": 0.13, "grad_norm": 1.388299822807312, "learning_rate": 9.752784462637748e-06, "loss": 0.9056, "step": 973 }, { "epoch": 0.13, "grad_norm": 1.3938041925430298, "learning_rate": 9.752125007918714e-06, "loss": 0.8889, "step": 974 }, { "epoch": 0.13, "grad_norm": 1.4239897727966309, "learning_rate": 9.751464697168798e-06, "loss": 0.9314, "step": 975 }, { "epoch": 0.13, "grad_norm": 1.347528338432312, "learning_rate": 9.75080353050694e-06, "loss": 1.1052, "step": 976 }, { "epoch": 0.13, "grad_norm": 1.440366506576538, "learning_rate": 9.750141508052246e-06, "loss": 0.8879, "step": 977 }, { "epoch": 0.13, "grad_norm": 1.5517972707748413, "learning_rate": 9.749478629923968e-06, "loss": 0.844, "step": 978 }, { "epoch": 0.13, "grad_norm": 1.4419828653335571, "learning_rate": 9.748814896241512e-06, "loss": 0.9096, "step": 979 }, { "epoch": 0.13, "grad_norm": 1.362219214439392, "learning_rate": 9.748150307124444e-06, "loss": 0.9271, "step": 980 }, { "epoch": 0.13, "grad_norm": 1.510644793510437, "learning_rate": 9.74748486269248e-06, "loss": 0.8581, "step": 981 }, { "epoch": 0.13, "grad_norm": 1.2699527740478516, "learning_rate": 9.746818563065491e-06, "loss": 0.8522, "step": 982 }, { "epoch": 0.13, "grad_norm": 0.8779464960098267, "learning_rate": 9.7461514083635e-06, "loss": 1.1189, "step": 983 }, { "epoch": 0.13, "grad_norm": 1.4538229703903198, "learning_rate": 9.74548339870669e-06, "loss": 0.8692, "step": 984 }, { "epoch": 0.13, "grad_norm": 1.5233659744262695, "learning_rate": 9.744814534215388e-06, "loss": 0.8547, "step": 985 }, { "epoch": 0.13, "grad_norm": 1.3874084949493408, "learning_rate": 9.744144815010085e-06, "loss": 0.8566, "step": 986 }, { "epoch": 0.13, "grad_norm": 1.3507319688796997, "learning_rate": 9.743474241211422e-06, "loss": 0.8824, "step": 987 }, { "epoch": 0.13, "grad_norm": 1.393507957458496, "learning_rate": 9.742802812940191e-06, "loss": 0.9305, "step": 988 }, { "epoch": 0.13, "grad_norm": 1.3943482637405396, "learning_rate": 9.742130530317342e-06, "loss": 0.8918, "step": 989 }, { "epoch": 0.13, "grad_norm": 1.3994193077087402, "learning_rate": 9.741457393463977e-06, "loss": 0.8655, "step": 990 }, { "epoch": 0.13, "grad_norm": 1.4928964376449585, "learning_rate": 9.740783402501356e-06, "loss": 0.92, "step": 991 }, { "epoch": 0.13, "grad_norm": 1.0207608938217163, "learning_rate": 9.740108557550885e-06, "loss": 1.1371, "step": 992 }, { "epoch": 0.13, "grad_norm": 1.3384218215942383, "learning_rate": 9.739432858734131e-06, "loss": 0.9263, "step": 993 }, { "epoch": 0.13, "grad_norm": 1.3722705841064453, "learning_rate": 9.73875630617281e-06, "loss": 0.8704, "step": 994 }, { "epoch": 0.13, "grad_norm": 0.8322850465774536, "learning_rate": 9.738078899988794e-06, "loss": 1.1126, "step": 995 }, { "epoch": 0.13, "grad_norm": 0.7224822640419006, "learning_rate": 9.737400640304108e-06, "loss": 1.0805, "step": 996 }, { "epoch": 0.13, "grad_norm": 1.433716893196106, "learning_rate": 9.736721527240935e-06, "loss": 0.8853, "step": 997 }, { "epoch": 0.13, "grad_norm": 1.4827104806900024, "learning_rate": 9.736041560921604e-06, "loss": 0.9112, "step": 998 }, { "epoch": 0.13, "grad_norm": 1.4176993370056152, "learning_rate": 9.735360741468604e-06, "loss": 0.904, "step": 999 }, { "epoch": 0.13, "grad_norm": 1.3996453285217285, "learning_rate": 9.734679069004575e-06, "loss": 0.8522, "step": 1000 }, { "epoch": 0.13, "grad_norm": 1.309475302696228, "learning_rate": 9.73399654365231e-06, "loss": 1.0985, "step": 1001 }, { "epoch": 0.13, "grad_norm": 1.6583755016326904, "learning_rate": 9.733313165534757e-06, "loss": 0.8669, "step": 1002 }, { "epoch": 0.13, "grad_norm": 1.3305031061172485, "learning_rate": 9.732628934775017e-06, "loss": 0.8832, "step": 1003 }, { "epoch": 0.13, "grad_norm": 0.762117326259613, "learning_rate": 9.731943851496347e-06, "loss": 1.1075, "step": 1004 }, { "epoch": 0.13, "grad_norm": 0.689617931842804, "learning_rate": 9.731257915822152e-06, "loss": 1.0895, "step": 1005 }, { "epoch": 0.13, "grad_norm": 1.5566413402557373, "learning_rate": 9.730571127875997e-06, "loss": 0.8499, "step": 1006 }, { "epoch": 0.13, "grad_norm": 1.4198592901229858, "learning_rate": 9.729883487781595e-06, "loss": 0.8886, "step": 1007 }, { "epoch": 0.13, "grad_norm": 1.4094185829162598, "learning_rate": 9.729194995662818e-06, "loss": 0.8987, "step": 1008 }, { "epoch": 0.13, "grad_norm": 1.372226357460022, "learning_rate": 9.728505651643687e-06, "loss": 0.9582, "step": 1009 }, { "epoch": 0.13, "grad_norm": 1.6364820003509521, "learning_rate": 9.727815455848376e-06, "loss": 0.8254, "step": 1010 }, { "epoch": 0.13, "grad_norm": 1.548100471496582, "learning_rate": 9.727124408401221e-06, "loss": 0.9512, "step": 1011 }, { "epoch": 0.13, "grad_norm": 1.5141242742538452, "learning_rate": 9.726432509426697e-06, "loss": 1.1245, "step": 1012 }, { "epoch": 0.13, "grad_norm": 1.6424754858016968, "learning_rate": 9.725739759049444e-06, "loss": 0.8859, "step": 1013 }, { "epoch": 0.13, "grad_norm": 2.5971102714538574, "learning_rate": 9.72504615739425e-06, "loss": 0.844, "step": 1014 }, { "epoch": 0.13, "grad_norm": 1.32912015914917, "learning_rate": 9.72435170458606e-06, "loss": 0.8766, "step": 1015 }, { "epoch": 0.13, "grad_norm": 1.478065013885498, "learning_rate": 9.72365640074997e-06, "loss": 0.8757, "step": 1016 }, { "epoch": 0.13, "grad_norm": 1.4246549606323242, "learning_rate": 9.722960246011226e-06, "loss": 0.9133, "step": 1017 }, { "epoch": 0.13, "grad_norm": 1.4088371992111206, "learning_rate": 9.722263240495236e-06, "loss": 0.8784, "step": 1018 }, { "epoch": 0.13, "grad_norm": 1.3428981304168701, "learning_rate": 9.721565384327554e-06, "loss": 0.8813, "step": 1019 }, { "epoch": 0.13, "grad_norm": 1.3067352771759033, "learning_rate": 9.720866677633889e-06, "loss": 1.1114, "step": 1020 }, { "epoch": 0.13, "grad_norm": 1.367276906967163, "learning_rate": 9.720167120540102e-06, "loss": 0.9134, "step": 1021 }, { "epoch": 0.13, "grad_norm": 1.4672311544418335, "learning_rate": 9.719466713172212e-06, "loss": 0.9271, "step": 1022 }, { "epoch": 0.13, "grad_norm": 1.5334101915359497, "learning_rate": 9.718765455656385e-06, "loss": 0.8476, "step": 1023 }, { "epoch": 0.13, "grad_norm": 1.3493796586990356, "learning_rate": 9.718063348118945e-06, "loss": 0.8188, "step": 1024 }, { "epoch": 0.13, "grad_norm": 1.5200425386428833, "learning_rate": 9.717360390686366e-06, "loss": 0.9318, "step": 1025 }, { "epoch": 0.13, "grad_norm": 1.1404281854629517, "learning_rate": 9.716656583485276e-06, "loss": 1.0914, "step": 1026 }, { "epoch": 0.13, "grad_norm": 0.8613666892051697, "learning_rate": 9.715951926642458e-06, "loss": 1.1092, "step": 1027 }, { "epoch": 0.13, "grad_norm": 1.4232321977615356, "learning_rate": 9.715246420284845e-06, "loss": 0.8002, "step": 1028 }, { "epoch": 0.13, "grad_norm": 1.5045092105865479, "learning_rate": 9.714540064539524e-06, "loss": 0.9302, "step": 1029 }, { "epoch": 0.13, "grad_norm": 1.409341812133789, "learning_rate": 9.713832859533737e-06, "loss": 0.8388, "step": 1030 }, { "epoch": 0.14, "grad_norm": 1.3205595016479492, "learning_rate": 9.713124805394877e-06, "loss": 0.7776, "step": 1031 }, { "epoch": 0.14, "grad_norm": 1.4238077402114868, "learning_rate": 9.71241590225049e-06, "loss": 0.8535, "step": 1032 }, { "epoch": 0.14, "grad_norm": 1.4523141384124756, "learning_rate": 9.711706150228273e-06, "loss": 0.8999, "step": 1033 }, { "epoch": 0.14, "grad_norm": 2.435572624206543, "learning_rate": 9.710995549456082e-06, "loss": 1.125, "step": 1034 }, { "epoch": 0.14, "grad_norm": 1.4363900423049927, "learning_rate": 9.71028410006192e-06, "loss": 0.8715, "step": 1035 }, { "epoch": 0.14, "grad_norm": 1.4128482341766357, "learning_rate": 9.709571802173947e-06, "loss": 0.8843, "step": 1036 }, { "epoch": 0.14, "grad_norm": 1.4995672702789307, "learning_rate": 9.70885865592047e-06, "loss": 0.89, "step": 1037 }, { "epoch": 0.14, "grad_norm": 1.1006680727005005, "learning_rate": 9.708144661429956e-06, "loss": 1.1078, "step": 1038 }, { "epoch": 0.14, "grad_norm": 0.6882006525993347, "learning_rate": 9.70742981883102e-06, "loss": 1.0801, "step": 1039 }, { "epoch": 0.14, "grad_norm": 1.5755130052566528, "learning_rate": 9.706714128252433e-06, "loss": 0.9075, "step": 1040 }, { "epoch": 0.14, "grad_norm": 1.509615421295166, "learning_rate": 9.705997589823112e-06, "loss": 0.9374, "step": 1041 }, { "epoch": 0.14, "grad_norm": 1.4925847053527832, "learning_rate": 9.705280203672139e-06, "loss": 0.8752, "step": 1042 }, { "epoch": 0.14, "grad_norm": 1.4956281185150146, "learning_rate": 9.704561969928736e-06, "loss": 0.8868, "step": 1043 }, { "epoch": 0.14, "grad_norm": 1.355008840560913, "learning_rate": 9.703842888722284e-06, "loss": 0.9339, "step": 1044 }, { "epoch": 0.14, "grad_norm": 1.4216502904891968, "learning_rate": 9.703122960182316e-06, "loss": 0.8839, "step": 1045 }, { "epoch": 0.14, "grad_norm": 1.4472242593765259, "learning_rate": 9.70240218443852e-06, "loss": 0.899, "step": 1046 }, { "epoch": 0.14, "grad_norm": 1.4498854875564575, "learning_rate": 9.70168056162073e-06, "loss": 0.9608, "step": 1047 }, { "epoch": 0.14, "grad_norm": 1.5080444812774658, "learning_rate": 9.700958091858937e-06, "loss": 0.8789, "step": 1048 }, { "epoch": 0.14, "grad_norm": 1.3233859539031982, "learning_rate": 9.700234775283287e-06, "loss": 0.8912, "step": 1049 }, { "epoch": 0.14, "grad_norm": 1.398271083831787, "learning_rate": 9.699510612024074e-06, "loss": 0.8594, "step": 1050 }, { "epoch": 0.14, "grad_norm": 1.4654033184051514, "learning_rate": 9.698785602211745e-06, "loss": 0.8752, "step": 1051 }, { "epoch": 0.14, "grad_norm": 2.9380886554718018, "learning_rate": 9.698059745976905e-06, "loss": 1.135, "step": 1052 }, { "epoch": 0.14, "grad_norm": 1.4990178346633911, "learning_rate": 9.697333043450301e-06, "loss": 0.9274, "step": 1053 }, { "epoch": 0.14, "grad_norm": 1.322774887084961, "learning_rate": 9.696605494762842e-06, "loss": 0.8201, "step": 1054 }, { "epoch": 0.14, "grad_norm": 1.4943960905075073, "learning_rate": 9.695877100045586e-06, "loss": 0.8466, "step": 1055 }, { "epoch": 0.14, "grad_norm": 1.348846435546875, "learning_rate": 9.695147859429744e-06, "loss": 0.8722, "step": 1056 }, { "epoch": 0.14, "grad_norm": 1.384539246559143, "learning_rate": 9.694417773046679e-06, "loss": 0.863, "step": 1057 }, { "epoch": 0.14, "grad_norm": 1.472916603088379, "learning_rate": 9.693686841027902e-06, "loss": 0.9454, "step": 1058 }, { "epoch": 0.14, "grad_norm": 1.3185468912124634, "learning_rate": 9.692955063505084e-06, "loss": 0.8544, "step": 1059 }, { "epoch": 0.14, "grad_norm": 1.339053750038147, "learning_rate": 9.692222440610046e-06, "loss": 0.8794, "step": 1060 }, { "epoch": 0.14, "grad_norm": 0.8552629947662354, "learning_rate": 9.691488972474758e-06, "loss": 1.0979, "step": 1061 }, { "epoch": 0.14, "grad_norm": 1.385477066040039, "learning_rate": 9.690754659231346e-06, "loss": 0.8757, "step": 1062 }, { "epoch": 0.14, "grad_norm": 1.4067034721374512, "learning_rate": 9.690019501012085e-06, "loss": 0.8962, "step": 1063 }, { "epoch": 0.14, "grad_norm": 1.9606120586395264, "learning_rate": 9.689283497949401e-06, "loss": 0.9111, "step": 1064 }, { "epoch": 0.14, "grad_norm": 1.4897401332855225, "learning_rate": 9.688546650175883e-06, "loss": 0.8773, "step": 1065 }, { "epoch": 0.14, "grad_norm": 1.4420095682144165, "learning_rate": 9.687808957824256e-06, "loss": 0.9054, "step": 1066 }, { "epoch": 0.14, "grad_norm": 1.7357510328292847, "learning_rate": 9.68707042102741e-06, "loss": 0.8947, "step": 1067 }, { "epoch": 0.14, "grad_norm": 1.0448522567749023, "learning_rate": 9.686331039918381e-06, "loss": 1.083, "step": 1068 }, { "epoch": 0.14, "grad_norm": 1.3617088794708252, "learning_rate": 9.685590814630356e-06, "loss": 0.9054, "step": 1069 }, { "epoch": 0.14, "grad_norm": 1.4369652271270752, "learning_rate": 9.684849745296682e-06, "loss": 0.9198, "step": 1070 }, { "epoch": 0.14, "grad_norm": 1.2767913341522217, "learning_rate": 9.684107832050849e-06, "loss": 0.8739, "step": 1071 }, { "epoch": 0.14, "grad_norm": 0.8045611381530762, "learning_rate": 9.683365075026503e-06, "loss": 1.115, "step": 1072 }, { "epoch": 0.14, "grad_norm": 1.4118460416793823, "learning_rate": 9.682621474357441e-06, "loss": 0.879, "step": 1073 }, { "epoch": 0.14, "grad_norm": 1.4772776365280151, "learning_rate": 9.681877030177614e-06, "loss": 0.9348, "step": 1074 }, { "epoch": 0.14, "grad_norm": 1.3972548246383667, "learning_rate": 9.681131742621123e-06, "loss": 0.839, "step": 1075 }, { "epoch": 0.14, "grad_norm": 1.6981782913208008, "learning_rate": 9.68038561182222e-06, "loss": 0.874, "step": 1076 }, { "epoch": 0.14, "grad_norm": 0.6889490485191345, "learning_rate": 9.679638637915314e-06, "loss": 1.0782, "step": 1077 }, { "epoch": 0.14, "grad_norm": 1.4622575044631958, "learning_rate": 9.678890821034959e-06, "loss": 0.8487, "step": 1078 }, { "epoch": 0.14, "grad_norm": 0.6205365657806396, "learning_rate": 9.678142161315864e-06, "loss": 1.0932, "step": 1079 }, { "epoch": 0.14, "grad_norm": 1.3641854524612427, "learning_rate": 9.677392658892892e-06, "loss": 0.8947, "step": 1080 }, { "epoch": 0.14, "grad_norm": 1.444913387298584, "learning_rate": 9.676642313901055e-06, "loss": 0.9133, "step": 1081 }, { "epoch": 0.14, "grad_norm": 1.2944564819335938, "learning_rate": 9.675891126475514e-06, "loss": 0.8414, "step": 1082 }, { "epoch": 0.14, "grad_norm": 1.3610739707946777, "learning_rate": 9.675139096751591e-06, "loss": 0.8208, "step": 1083 }, { "epoch": 0.14, "grad_norm": 1.3302583694458008, "learning_rate": 9.674386224864752e-06, "loss": 0.8533, "step": 1084 }, { "epoch": 0.14, "grad_norm": 0.8902223706245422, "learning_rate": 9.673632510950615e-06, "loss": 1.097, "step": 1085 }, { "epoch": 0.14, "grad_norm": 1.5245836973190308, "learning_rate": 9.672877955144952e-06, "loss": 0.8928, "step": 1086 }, { "epoch": 0.14, "grad_norm": 1.2985963821411133, "learning_rate": 9.672122557583685e-06, "loss": 0.8158, "step": 1087 }, { "epoch": 0.14, "grad_norm": 1.4060461521148682, "learning_rate": 9.671366318402892e-06, "loss": 0.9097, "step": 1088 }, { "epoch": 0.14, "grad_norm": 0.7629050612449646, "learning_rate": 9.670609237738795e-06, "loss": 1.1154, "step": 1089 }, { "epoch": 0.14, "grad_norm": 1.4348442554473877, "learning_rate": 9.669851315727777e-06, "loss": 0.9001, "step": 1090 }, { "epoch": 0.14, "grad_norm": 1.4148443937301636, "learning_rate": 9.669092552506361e-06, "loss": 0.8685, "step": 1091 }, { "epoch": 0.14, "grad_norm": 1.3683133125305176, "learning_rate": 9.668332948211234e-06, "loss": 0.8855, "step": 1092 }, { "epoch": 0.14, "grad_norm": 1.5008445978164673, "learning_rate": 9.667572502979223e-06, "loss": 0.8917, "step": 1093 }, { "epoch": 0.14, "grad_norm": 1.687303900718689, "learning_rate": 9.666811216947315e-06, "loss": 0.9219, "step": 1094 }, { "epoch": 0.14, "grad_norm": 0.7916784286499023, "learning_rate": 9.666049090252647e-06, "loss": 1.1009, "step": 1095 }, { "epoch": 0.14, "grad_norm": 0.7831093668937683, "learning_rate": 9.665286123032502e-06, "loss": 1.0951, "step": 1096 }, { "epoch": 0.14, "grad_norm": 1.3979437351226807, "learning_rate": 9.664522315424321e-06, "loss": 0.8953, "step": 1097 }, { "epoch": 0.14, "grad_norm": 1.5661232471466064, "learning_rate": 9.663757667565691e-06, "loss": 0.8453, "step": 1098 }, { "epoch": 0.14, "grad_norm": 0.776566207408905, "learning_rate": 9.662992179594355e-06, "loss": 1.1016, "step": 1099 }, { "epoch": 0.14, "grad_norm": 1.315075397491455, "learning_rate": 9.662225851648205e-06, "loss": 0.8378, "step": 1100 }, { "epoch": 0.14, "grad_norm": 1.3944978713989258, "learning_rate": 9.661458683865283e-06, "loss": 0.8705, "step": 1101 }, { "epoch": 0.14, "grad_norm": 0.7923712134361267, "learning_rate": 9.660690676383786e-06, "loss": 1.1093, "step": 1102 }, { "epoch": 0.14, "grad_norm": 0.6610912680625916, "learning_rate": 9.65992182934206e-06, "loss": 1.1003, "step": 1103 }, { "epoch": 0.14, "grad_norm": 0.640299916267395, "learning_rate": 9.659152142878601e-06, "loss": 1.0745, "step": 1104 }, { "epoch": 0.14, "grad_norm": 1.3734917640686035, "learning_rate": 9.658381617132058e-06, "loss": 0.8863, "step": 1105 }, { "epoch": 0.14, "grad_norm": 1.4019179344177246, "learning_rate": 9.657610252241232e-06, "loss": 0.8563, "step": 1106 }, { "epoch": 0.15, "grad_norm": 1.3423867225646973, "learning_rate": 9.656838048345072e-06, "loss": 0.8692, "step": 1107 }, { "epoch": 0.15, "grad_norm": 1.332181692123413, "learning_rate": 9.65606500558268e-06, "loss": 0.8085, "step": 1108 }, { "epoch": 0.15, "grad_norm": 1.0113447904586792, "learning_rate": 9.655291124093311e-06, "loss": 1.0972, "step": 1109 }, { "epoch": 0.15, "grad_norm": 1.40089750289917, "learning_rate": 9.65451640401637e-06, "loss": 0.9175, "step": 1110 }, { "epoch": 0.15, "grad_norm": 1.4321414232254028, "learning_rate": 9.653740845491408e-06, "loss": 0.8886, "step": 1111 }, { "epoch": 0.15, "grad_norm": 0.737697958946228, "learning_rate": 9.652964448658135e-06, "loss": 1.1021, "step": 1112 }, { "epoch": 0.15, "grad_norm": 1.3724920749664307, "learning_rate": 9.652187213656409e-06, "loss": 0.8703, "step": 1113 }, { "epoch": 0.15, "grad_norm": 0.7051709890365601, "learning_rate": 9.651409140626236e-06, "loss": 1.0755, "step": 1114 }, { "epoch": 0.15, "grad_norm": 0.6920014023780823, "learning_rate": 9.650630229707775e-06, "loss": 1.1025, "step": 1115 }, { "epoch": 0.15, "grad_norm": 1.4395391941070557, "learning_rate": 9.649850481041338e-06, "loss": 0.9236, "step": 1116 }, { "epoch": 0.15, "grad_norm": 1.3581922054290771, "learning_rate": 9.649069894767386e-06, "loss": 0.8432, "step": 1117 }, { "epoch": 0.15, "grad_norm": 1.3632484674453735, "learning_rate": 9.64828847102653e-06, "loss": 0.8608, "step": 1118 }, { "epoch": 0.15, "grad_norm": 1.3567981719970703, "learning_rate": 9.647506209959534e-06, "loss": 0.7983, "step": 1119 }, { "epoch": 0.15, "grad_norm": 1.8419607877731323, "learning_rate": 9.64672311170731e-06, "loss": 0.9316, "step": 1120 }, { "epoch": 0.15, "grad_norm": 1.3550816774368286, "learning_rate": 9.645939176410924e-06, "loss": 0.8938, "step": 1121 }, { "epoch": 0.15, "grad_norm": 1.4306718111038208, "learning_rate": 9.64515440421159e-06, "loss": 0.9796, "step": 1122 }, { "epoch": 0.15, "grad_norm": 1.3996896743774414, "learning_rate": 9.644368795250675e-06, "loss": 0.937, "step": 1123 }, { "epoch": 0.15, "grad_norm": 1.3689876794815063, "learning_rate": 9.643582349669698e-06, "loss": 0.8837, "step": 1124 }, { "epoch": 0.15, "grad_norm": 0.9149777889251709, "learning_rate": 9.642795067610324e-06, "loss": 1.0967, "step": 1125 }, { "epoch": 0.15, "grad_norm": 0.804594099521637, "learning_rate": 9.64200694921437e-06, "loss": 1.0908, "step": 1126 }, { "epoch": 0.15, "grad_norm": 1.4125851392745972, "learning_rate": 9.641217994623808e-06, "loss": 0.8828, "step": 1127 }, { "epoch": 0.15, "grad_norm": 1.3574293851852417, "learning_rate": 9.640428203980755e-06, "loss": 0.8892, "step": 1128 }, { "epoch": 0.15, "grad_norm": 1.4092024564743042, "learning_rate": 9.639637577427483e-06, "loss": 0.8836, "step": 1129 }, { "epoch": 0.15, "grad_norm": 1.2806117534637451, "learning_rate": 9.63884611510641e-06, "loss": 1.0998, "step": 1130 }, { "epoch": 0.15, "grad_norm": 1.6021461486816406, "learning_rate": 9.63805381716011e-06, "loss": 0.9063, "step": 1131 }, { "epoch": 0.15, "grad_norm": 1.5351502895355225, "learning_rate": 9.637260683731303e-06, "loss": 0.9051, "step": 1132 }, { "epoch": 0.15, "grad_norm": 1.709657073020935, "learning_rate": 9.63646671496286e-06, "loss": 0.8539, "step": 1133 }, { "epoch": 0.15, "grad_norm": 1.3224538564682007, "learning_rate": 9.635671910997808e-06, "loss": 0.8036, "step": 1134 }, { "epoch": 0.15, "grad_norm": 1.4146167039871216, "learning_rate": 9.634876271979317e-06, "loss": 0.8126, "step": 1135 }, { "epoch": 0.15, "grad_norm": 1.414717674255371, "learning_rate": 9.63407979805071e-06, "loss": 0.8857, "step": 1136 }, { "epoch": 0.15, "grad_norm": 1.533186912536621, "learning_rate": 9.633282489355463e-06, "loss": 0.8515, "step": 1137 }, { "epoch": 0.15, "grad_norm": 1.3627574443817139, "learning_rate": 9.6324843460372e-06, "loss": 0.8206, "step": 1138 }, { "epoch": 0.15, "grad_norm": 1.289175271987915, "learning_rate": 9.631685368239695e-06, "loss": 0.8942, "step": 1139 }, { "epoch": 0.15, "grad_norm": 1.424505352973938, "learning_rate": 9.630885556106873e-06, "loss": 0.8863, "step": 1140 }, { "epoch": 0.15, "grad_norm": 1.448671817779541, "learning_rate": 9.630084909782809e-06, "loss": 0.8983, "step": 1141 }, { "epoch": 0.15, "grad_norm": 0.9991753697395325, "learning_rate": 9.629283429411731e-06, "loss": 1.1039, "step": 1142 }, { "epoch": 0.15, "grad_norm": 1.3644537925720215, "learning_rate": 9.628481115138011e-06, "loss": 0.8403, "step": 1143 }, { "epoch": 0.15, "grad_norm": 0.6951291561126709, "learning_rate": 9.627677967106178e-06, "loss": 1.086, "step": 1144 }, { "epoch": 0.15, "grad_norm": 1.3709001541137695, "learning_rate": 9.626873985460908e-06, "loss": 0.9433, "step": 1145 }, { "epoch": 0.15, "grad_norm": 1.4347950220108032, "learning_rate": 9.626069170347022e-06, "loss": 0.842, "step": 1146 }, { "epoch": 0.15, "grad_norm": 1.3994264602661133, "learning_rate": 9.625263521909508e-06, "loss": 0.8856, "step": 1147 }, { "epoch": 0.15, "grad_norm": 1.3570325374603271, "learning_rate": 9.62445704029348e-06, "loss": 0.8481, "step": 1148 }, { "epoch": 0.15, "grad_norm": 1.3347605466842651, "learning_rate": 9.623649725644225e-06, "loss": 0.8783, "step": 1149 }, { "epoch": 0.15, "grad_norm": 1.4657697677612305, "learning_rate": 9.622841578107163e-06, "loss": 0.7826, "step": 1150 }, { "epoch": 0.15, "grad_norm": 1.3849912881851196, "learning_rate": 9.622032597827873e-06, "loss": 0.8836, "step": 1151 }, { "epoch": 0.15, "grad_norm": 1.3766971826553345, "learning_rate": 9.621222784952082e-06, "loss": 1.1094, "step": 1152 }, { "epoch": 0.15, "grad_norm": 1.4481498003005981, "learning_rate": 9.620412139625666e-06, "loss": 0.903, "step": 1153 }, { "epoch": 0.15, "grad_norm": 0.7020240426063538, "learning_rate": 9.619600661994655e-06, "loss": 1.0875, "step": 1154 }, { "epoch": 0.15, "grad_norm": 1.4497480392456055, "learning_rate": 9.618788352205221e-06, "loss": 0.893, "step": 1155 }, { "epoch": 0.15, "grad_norm": 1.5258890390396118, "learning_rate": 9.617975210403694e-06, "loss": 0.883, "step": 1156 }, { "epoch": 0.15, "grad_norm": 1.4306615591049194, "learning_rate": 9.617161236736549e-06, "loss": 0.8382, "step": 1157 }, { "epoch": 0.15, "grad_norm": 1.4290173053741455, "learning_rate": 9.616346431350413e-06, "loss": 0.8513, "step": 1158 }, { "epoch": 0.15, "grad_norm": 1.536266803741455, "learning_rate": 9.61553079439206e-06, "loss": 0.8916, "step": 1159 }, { "epoch": 0.15, "grad_norm": 1.3851081132888794, "learning_rate": 9.614714326008419e-06, "loss": 0.8955, "step": 1160 }, { "epoch": 0.15, "grad_norm": 1.3853864669799805, "learning_rate": 9.613897026346563e-06, "loss": 0.8679, "step": 1161 }, { "epoch": 0.15, "grad_norm": 1.3505232334136963, "learning_rate": 9.61307889555372e-06, "loss": 0.8194, "step": 1162 }, { "epoch": 0.15, "grad_norm": 1.3538168668746948, "learning_rate": 9.612259933777265e-06, "loss": 0.8606, "step": 1163 }, { "epoch": 0.15, "grad_norm": 1.2959727048873901, "learning_rate": 9.61144014116472e-06, "loss": 0.8837, "step": 1164 }, { "epoch": 0.15, "grad_norm": 1.5656203031539917, "learning_rate": 9.610619517863763e-06, "loss": 0.7861, "step": 1165 }, { "epoch": 0.15, "grad_norm": 1.9490981101989746, "learning_rate": 9.609798064022215e-06, "loss": 0.9549, "step": 1166 }, { "epoch": 0.15, "grad_norm": 1.3681564331054688, "learning_rate": 9.608975779788054e-06, "loss": 0.905, "step": 1167 }, { "epoch": 0.15, "grad_norm": 1.322548508644104, "learning_rate": 9.608152665309399e-06, "loss": 0.8321, "step": 1168 }, { "epoch": 0.15, "grad_norm": 1.8977123498916626, "learning_rate": 9.607328720734523e-06, "loss": 1.0957, "step": 1169 }, { "epoch": 0.15, "grad_norm": 1.3920536041259766, "learning_rate": 9.606503946211852e-06, "loss": 0.8515, "step": 1170 }, { "epoch": 0.15, "grad_norm": 1.3609968423843384, "learning_rate": 9.605678341889955e-06, "loss": 0.9056, "step": 1171 }, { "epoch": 0.15, "grad_norm": 1.4110909700393677, "learning_rate": 9.604851907917555e-06, "loss": 0.8713, "step": 1172 }, { "epoch": 0.15, "grad_norm": 0.7329552173614502, "learning_rate": 9.60402464444352e-06, "loss": 1.0998, "step": 1173 }, { "epoch": 0.15, "grad_norm": 0.7838072776794434, "learning_rate": 9.603196551616877e-06, "loss": 1.0799, "step": 1174 }, { "epoch": 0.15, "grad_norm": 0.7674273252487183, "learning_rate": 9.602367629586789e-06, "loss": 1.0915, "step": 1175 }, { "epoch": 0.15, "grad_norm": 1.3759715557098389, "learning_rate": 9.601537878502577e-06, "loss": 0.8487, "step": 1176 }, { "epoch": 0.15, "grad_norm": 1.3295668363571167, "learning_rate": 9.60070729851371e-06, "loss": 0.9231, "step": 1177 }, { "epoch": 0.15, "grad_norm": 1.4400553703308105, "learning_rate": 9.599875889769806e-06, "loss": 0.8946, "step": 1178 }, { "epoch": 0.15, "grad_norm": 0.9115793108940125, "learning_rate": 9.59904365242063e-06, "loss": 1.0905, "step": 1179 }, { "epoch": 0.15, "grad_norm": 4.654597282409668, "learning_rate": 9.598210586616101e-06, "loss": 0.8462, "step": 1180 }, { "epoch": 0.15, "grad_norm": 1.3919533491134644, "learning_rate": 9.597376692506282e-06, "loss": 0.8281, "step": 1181 }, { "epoch": 0.15, "grad_norm": 0.810049831867218, "learning_rate": 9.596541970241392e-06, "loss": 1.1169, "step": 1182 }, { "epoch": 0.16, "grad_norm": 1.3884869813919067, "learning_rate": 9.595706419971788e-06, "loss": 0.9006, "step": 1183 }, { "epoch": 0.16, "grad_norm": 1.4306480884552002, "learning_rate": 9.594870041847988e-06, "loss": 0.8555, "step": 1184 }, { "epoch": 0.16, "grad_norm": 0.737879753112793, "learning_rate": 9.594032836020654e-06, "loss": 1.1157, "step": 1185 }, { "epoch": 0.16, "grad_norm": 1.6053987741470337, "learning_rate": 9.593194802640596e-06, "loss": 0.8098, "step": 1186 }, { "epoch": 0.16, "grad_norm": 1.4402254819869995, "learning_rate": 9.592355941858775e-06, "loss": 0.9153, "step": 1187 }, { "epoch": 0.16, "grad_norm": 1.4629706144332886, "learning_rate": 9.591516253826301e-06, "loss": 0.8677, "step": 1188 }, { "epoch": 0.16, "grad_norm": 1.2206265926361084, "learning_rate": 9.590675738694431e-06, "loss": 0.8783, "step": 1189 }, { "epoch": 0.16, "grad_norm": 1.330177664756775, "learning_rate": 9.58983439661457e-06, "loss": 0.8553, "step": 1190 }, { "epoch": 0.16, "grad_norm": 0.9807674884796143, "learning_rate": 9.588992227738281e-06, "loss": 1.0787, "step": 1191 }, { "epoch": 0.16, "grad_norm": 1.352518081665039, "learning_rate": 9.588149232217264e-06, "loss": 0.8533, "step": 1192 }, { "epoch": 0.16, "grad_norm": 1.4558006525039673, "learning_rate": 9.587305410203376e-06, "loss": 0.9135, "step": 1193 }, { "epoch": 0.16, "grad_norm": 1.4733396768569946, "learning_rate": 9.586460761848617e-06, "loss": 0.8872, "step": 1194 }, { "epoch": 0.16, "grad_norm": 1.3392482995986938, "learning_rate": 9.585615287305144e-06, "loss": 0.877, "step": 1195 }, { "epoch": 0.16, "grad_norm": 1.4062153100967407, "learning_rate": 9.58476898672525e-06, "loss": 0.8854, "step": 1196 }, { "epoch": 0.16, "grad_norm": 1.1290230751037598, "learning_rate": 9.583921860261393e-06, "loss": 1.0788, "step": 1197 }, { "epoch": 0.16, "grad_norm": 1.4088383913040161, "learning_rate": 9.583073908066166e-06, "loss": 0.882, "step": 1198 }, { "epoch": 0.16, "grad_norm": 1.404361605644226, "learning_rate": 9.582225130292319e-06, "loss": 0.9261, "step": 1199 }, { "epoch": 0.16, "grad_norm": 1.5470279455184937, "learning_rate": 9.581375527092744e-06, "loss": 0.8708, "step": 1200 }, { "epoch": 0.16, "grad_norm": 1.469233751296997, "learning_rate": 9.58052509862049e-06, "loss": 0.7845, "step": 1201 }, { "epoch": 0.16, "grad_norm": 0.784518837928772, "learning_rate": 9.579673845028747e-06, "loss": 1.0632, "step": 1202 }, { "epoch": 0.16, "grad_norm": 1.377051591873169, "learning_rate": 9.578821766470859e-06, "loss": 0.8328, "step": 1203 }, { "epoch": 0.16, "grad_norm": 1.3076634407043457, "learning_rate": 9.577968863100315e-06, "loss": 0.8679, "step": 1204 }, { "epoch": 0.16, "grad_norm": 1.4298224449157715, "learning_rate": 9.577115135070753e-06, "loss": 0.8642, "step": 1205 }, { "epoch": 0.16, "grad_norm": 0.8042106032371521, "learning_rate": 9.576260582535963e-06, "loss": 1.0928, "step": 1206 }, { "epoch": 0.16, "grad_norm": 1.4885644912719727, "learning_rate": 9.575405205649881e-06, "loss": 0.8602, "step": 1207 }, { "epoch": 0.16, "grad_norm": 1.353499412536621, "learning_rate": 9.574549004566589e-06, "loss": 0.8239, "step": 1208 }, { "epoch": 0.16, "grad_norm": 0.7581660151481628, "learning_rate": 9.573691979440321e-06, "loss": 1.0795, "step": 1209 }, { "epoch": 0.16, "grad_norm": 0.7599761486053467, "learning_rate": 9.57283413042546e-06, "loss": 1.101, "step": 1210 }, { "epoch": 0.16, "grad_norm": 0.6496105194091797, "learning_rate": 9.571975457676534e-06, "loss": 1.0985, "step": 1211 }, { "epoch": 0.16, "grad_norm": 0.6685724258422852, "learning_rate": 9.571115961348222e-06, "loss": 1.09, "step": 1212 }, { "epoch": 0.16, "grad_norm": 1.3565123081207275, "learning_rate": 9.570255641595352e-06, "loss": 0.8589, "step": 1213 }, { "epoch": 0.16, "grad_norm": 1.3888427019119263, "learning_rate": 9.569394498572895e-06, "loss": 0.8999, "step": 1214 }, { "epoch": 0.16, "grad_norm": 1.4384242296218872, "learning_rate": 9.568532532435979e-06, "loss": 0.8601, "step": 1215 }, { "epoch": 0.16, "grad_norm": 1.2738661766052246, "learning_rate": 9.567669743339872e-06, "loss": 0.8048, "step": 1216 }, { "epoch": 0.16, "grad_norm": 1.550553321838379, "learning_rate": 9.566806131439995e-06, "loss": 0.8144, "step": 1217 }, { "epoch": 0.16, "grad_norm": 1.287669062614441, "learning_rate": 9.565941696891917e-06, "loss": 1.102, "step": 1218 }, { "epoch": 0.16, "grad_norm": 1.527320146560669, "learning_rate": 9.565076439851354e-06, "loss": 0.9279, "step": 1219 }, { "epoch": 0.16, "grad_norm": 0.908008337020874, "learning_rate": 9.564210360474166e-06, "loss": 1.0891, "step": 1220 }, { "epoch": 0.16, "grad_norm": 1.4923946857452393, "learning_rate": 9.563343458916372e-06, "loss": 0.8873, "step": 1221 }, { "epoch": 0.16, "grad_norm": 1.4323437213897705, "learning_rate": 9.562475735334128e-06, "loss": 0.8854, "step": 1222 }, { "epoch": 0.16, "grad_norm": 1.4052668809890747, "learning_rate": 9.561607189883746e-06, "loss": 0.8592, "step": 1223 }, { "epoch": 0.16, "grad_norm": 1.4379667043685913, "learning_rate": 9.56073782272168e-06, "loss": 0.9053, "step": 1224 }, { "epoch": 0.16, "grad_norm": 1.412378191947937, "learning_rate": 9.559867634004535e-06, "loss": 0.9031, "step": 1225 }, { "epoch": 0.16, "grad_norm": 1.4254385232925415, "learning_rate": 9.558996623889065e-06, "loss": 0.9022, "step": 1226 }, { "epoch": 0.16, "grad_norm": 1.377379059791565, "learning_rate": 9.558124792532169e-06, "loss": 0.8303, "step": 1227 }, { "epoch": 0.16, "grad_norm": 1.381487250328064, "learning_rate": 9.557252140090897e-06, "loss": 0.8702, "step": 1228 }, { "epoch": 0.16, "grad_norm": 1.3590375185012817, "learning_rate": 9.556378666722445e-06, "loss": 1.0962, "step": 1229 }, { "epoch": 0.16, "grad_norm": 0.9638091325759888, "learning_rate": 9.555504372584158e-06, "loss": 1.095, "step": 1230 }, { "epoch": 0.16, "grad_norm": 1.4935150146484375, "learning_rate": 9.554629257833528e-06, "loss": 0.8083, "step": 1231 }, { "epoch": 0.16, "grad_norm": 1.7808703184127808, "learning_rate": 9.553753322628192e-06, "loss": 0.8402, "step": 1232 }, { "epoch": 0.16, "grad_norm": 1.6804693937301636, "learning_rate": 9.552876567125943e-06, "loss": 0.9075, "step": 1233 }, { "epoch": 0.16, "grad_norm": 1.6109771728515625, "learning_rate": 9.551998991484715e-06, "loss": 0.8917, "step": 1234 }, { "epoch": 0.16, "grad_norm": 1.7475111484527588, "learning_rate": 9.551120595862588e-06, "loss": 1.0867, "step": 1235 }, { "epoch": 0.16, "grad_norm": 1.473875880241394, "learning_rate": 9.550241380417797e-06, "loss": 0.8429, "step": 1236 }, { "epoch": 0.16, "grad_norm": 1.5461609363555908, "learning_rate": 9.54936134530872e-06, "loss": 0.847, "step": 1237 }, { "epoch": 0.16, "grad_norm": 1.1336948871612549, "learning_rate": 9.54848049069388e-06, "loss": 1.0904, "step": 1238 }, { "epoch": 0.16, "grad_norm": 1.5614229440689087, "learning_rate": 9.547598816731956e-06, "loss": 0.8935, "step": 1239 }, { "epoch": 0.16, "grad_norm": 1.6721992492675781, "learning_rate": 9.546716323581766e-06, "loss": 0.8237, "step": 1240 }, { "epoch": 0.16, "grad_norm": 0.934751033782959, "learning_rate": 9.545833011402281e-06, "loss": 1.0947, "step": 1241 }, { "epoch": 0.16, "grad_norm": 1.5559734106063843, "learning_rate": 9.544948880352618e-06, "loss": 0.8356, "step": 1242 }, { "epoch": 0.16, "grad_norm": 1.5445928573608398, "learning_rate": 9.54406393059204e-06, "loss": 0.8843, "step": 1243 }, { "epoch": 0.16, "grad_norm": 2.0155062675476074, "learning_rate": 9.54317816227996e-06, "loss": 0.8984, "step": 1244 }, { "epoch": 0.16, "grad_norm": 0.8883624076843262, "learning_rate": 9.542291575575936e-06, "loss": 1.0829, "step": 1245 }, { "epoch": 0.16, "grad_norm": 1.418125033378601, "learning_rate": 9.541404170639672e-06, "loss": 0.8317, "step": 1246 }, { "epoch": 0.16, "grad_norm": 1.4356238842010498, "learning_rate": 9.540515947631028e-06, "loss": 0.9165, "step": 1247 }, { "epoch": 0.16, "grad_norm": 1.3577775955200195, "learning_rate": 9.539626906710002e-06, "loss": 0.874, "step": 1248 }, { "epoch": 0.16, "grad_norm": 1.4302442073822021, "learning_rate": 9.538737048036741e-06, "loss": 0.8785, "step": 1249 }, { "epoch": 0.16, "grad_norm": 1.4667354822158813, "learning_rate": 9.537846371771544e-06, "loss": 0.903, "step": 1250 }, { "epoch": 0.16, "grad_norm": 0.8807317614555359, "learning_rate": 9.536954878074853e-06, "loss": 1.0791, "step": 1251 }, { "epoch": 0.16, "grad_norm": 1.3856877088546753, "learning_rate": 9.536062567107258e-06, "loss": 0.8493, "step": 1252 }, { "epoch": 0.16, "grad_norm": 1.3566983938217163, "learning_rate": 9.5351694390295e-06, "loss": 0.84, "step": 1253 }, { "epoch": 0.16, "grad_norm": 1.3375083208084106, "learning_rate": 9.534275494002457e-06, "loss": 0.91, "step": 1254 }, { "epoch": 0.16, "grad_norm": 0.7618712782859802, "learning_rate": 9.533380732187168e-06, "loss": 1.0775, "step": 1255 }, { "epoch": 0.16, "grad_norm": 0.7333803772926331, "learning_rate": 9.532485153744808e-06, "loss": 1.0891, "step": 1256 }, { "epoch": 0.16, "grad_norm": 1.577810525894165, "learning_rate": 9.531588758836706e-06, "loss": 0.9277, "step": 1257 }, { "epoch": 0.16, "grad_norm": 1.4028518199920654, "learning_rate": 9.530691547624333e-06, "loss": 0.8058, "step": 1258 }, { "epoch": 0.16, "grad_norm": 1.383252501487732, "learning_rate": 9.529793520269312e-06, "loss": 0.8842, "step": 1259 }, { "epoch": 0.17, "grad_norm": 2.0394206047058105, "learning_rate": 9.528894676933408e-06, "loss": 0.837, "step": 1260 }, { "epoch": 0.17, "grad_norm": 0.9193066358566284, "learning_rate": 9.527995017778538e-06, "loss": 1.101, "step": 1261 }, { "epoch": 0.17, "grad_norm": 1.3813376426696777, "learning_rate": 9.52709454296676e-06, "loss": 0.8239, "step": 1262 }, { "epoch": 0.17, "grad_norm": 1.612668514251709, "learning_rate": 9.526193252660285e-06, "loss": 0.8816, "step": 1263 }, { "epoch": 0.17, "grad_norm": 1.4152127504348755, "learning_rate": 9.525291147021467e-06, "loss": 0.8901, "step": 1264 }, { "epoch": 0.17, "grad_norm": 0.8285928964614868, "learning_rate": 9.52438822621281e-06, "loss": 1.0797, "step": 1265 }, { "epoch": 0.17, "grad_norm": 1.4116026163101196, "learning_rate": 9.523484490396961e-06, "loss": 0.8717, "step": 1266 }, { "epoch": 0.17, "grad_norm": 1.4105387926101685, "learning_rate": 9.522579939736716e-06, "loss": 0.8868, "step": 1267 }, { "epoch": 0.17, "grad_norm": 1.3539414405822754, "learning_rate": 9.521674574395016e-06, "loss": 0.8893, "step": 1268 }, { "epoch": 0.17, "grad_norm": 1.4094806909561157, "learning_rate": 9.520768394534955e-06, "loss": 0.901, "step": 1269 }, { "epoch": 0.17, "grad_norm": 1.3346123695373535, "learning_rate": 9.519861400319765e-06, "loss": 0.9182, "step": 1270 }, { "epoch": 0.17, "grad_norm": 1.1204547882080078, "learning_rate": 9.51895359191283e-06, "loss": 1.1118, "step": 1271 }, { "epoch": 0.17, "grad_norm": 0.775888979434967, "learning_rate": 9.518044969477678e-06, "loss": 1.0986, "step": 1272 }, { "epoch": 0.17, "grad_norm": 0.6698940992355347, "learning_rate": 9.517135533177989e-06, "loss": 1.0844, "step": 1273 }, { "epoch": 0.17, "grad_norm": 1.4502182006835938, "learning_rate": 9.516225283177579e-06, "loss": 0.856, "step": 1274 }, { "epoch": 0.17, "grad_norm": 1.4956223964691162, "learning_rate": 9.515314219640424e-06, "loss": 0.884, "step": 1275 }, { "epoch": 0.17, "grad_norm": 1.5601767301559448, "learning_rate": 9.514402342730636e-06, "loss": 0.9294, "step": 1276 }, { "epoch": 0.17, "grad_norm": 1.4284145832061768, "learning_rate": 9.513489652612475e-06, "loss": 0.854, "step": 1277 }, { "epoch": 0.17, "grad_norm": 1.3464808464050293, "learning_rate": 9.512576149450356e-06, "loss": 0.8895, "step": 1278 }, { "epoch": 0.17, "grad_norm": 1.4209083318710327, "learning_rate": 9.511661833408831e-06, "loss": 0.8775, "step": 1279 }, { "epoch": 0.17, "grad_norm": 1.4815081357955933, "learning_rate": 9.5107467046526e-06, "loss": 1.1185, "step": 1280 }, { "epoch": 0.17, "grad_norm": 1.1996359825134277, "learning_rate": 9.509830763346514e-06, "loss": 1.1004, "step": 1281 }, { "epoch": 0.17, "grad_norm": 1.626787543296814, "learning_rate": 9.508914009655565e-06, "loss": 0.8962, "step": 1282 }, { "epoch": 0.17, "grad_norm": 1.7974363565444946, "learning_rate": 9.507996443744896e-06, "loss": 0.8818, "step": 1283 }, { "epoch": 0.17, "grad_norm": 1.5830916166305542, "learning_rate": 9.507078065779793e-06, "loss": 0.8711, "step": 1284 }, { "epoch": 0.17, "grad_norm": 1.3745226860046387, "learning_rate": 9.506158875925689e-06, "loss": 1.0614, "step": 1285 }, { "epoch": 0.17, "grad_norm": 1.4340379238128662, "learning_rate": 9.505238874348161e-06, "loss": 0.9099, "step": 1286 }, { "epoch": 0.17, "grad_norm": 1.3736696243286133, "learning_rate": 9.50431806121294e-06, "loss": 0.8944, "step": 1287 }, { "epoch": 0.17, "grad_norm": 1.4905014038085938, "learning_rate": 9.503396436685896e-06, "loss": 0.8278, "step": 1288 }, { "epoch": 0.17, "grad_norm": 1.4303690195083618, "learning_rate": 9.502474000933046e-06, "loss": 0.8464, "step": 1289 }, { "epoch": 0.17, "grad_norm": 1.5033752918243408, "learning_rate": 9.501550754120557e-06, "loss": 0.8897, "step": 1290 }, { "epoch": 0.17, "grad_norm": 1.4312320947647095, "learning_rate": 9.500626696414734e-06, "loss": 0.8611, "step": 1291 }, { "epoch": 0.17, "grad_norm": 1.3001493215560913, "learning_rate": 9.49970182798204e-06, "loss": 0.8965, "step": 1292 }, { "epoch": 0.17, "grad_norm": 0.7505031824111938, "learning_rate": 9.498776148989073e-06, "loss": 1.0837, "step": 1293 }, { "epoch": 0.17, "grad_norm": 1.404866099357605, "learning_rate": 9.497849659602582e-06, "loss": 0.8357, "step": 1294 }, { "epoch": 0.17, "grad_norm": 1.4876941442489624, "learning_rate": 9.496922359989464e-06, "loss": 0.8782, "step": 1295 }, { "epoch": 0.17, "grad_norm": 1.4467862844467163, "learning_rate": 9.495994250316756e-06, "loss": 0.8319, "step": 1296 }, { "epoch": 0.17, "grad_norm": 1.3827005624771118, "learning_rate": 9.495065330751649e-06, "loss": 0.8339, "step": 1297 }, { "epoch": 0.17, "grad_norm": 1.3773727416992188, "learning_rate": 9.494135601461469e-06, "loss": 0.8731, "step": 1298 }, { "epoch": 0.17, "grad_norm": 1.319554090499878, "learning_rate": 9.4932050626137e-06, "loss": 0.8506, "step": 1299 }, { "epoch": 0.17, "grad_norm": 0.7993850708007812, "learning_rate": 9.492273714375963e-06, "loss": 1.0967, "step": 1300 }, { "epoch": 0.17, "grad_norm": 0.70549476146698, "learning_rate": 9.49134155691603e-06, "loss": 1.092, "step": 1301 }, { "epoch": 0.17, "grad_norm": 0.6450845003128052, "learning_rate": 9.490408590401814e-06, "loss": 1.0942, "step": 1302 }, { "epoch": 0.17, "grad_norm": 1.425662875175476, "learning_rate": 9.489474815001377e-06, "loss": 0.8838, "step": 1303 }, { "epoch": 0.17, "grad_norm": 1.5318257808685303, "learning_rate": 9.488540230882925e-06, "loss": 0.934, "step": 1304 }, { "epoch": 0.17, "grad_norm": 1.3899426460266113, "learning_rate": 9.487604838214812e-06, "loss": 0.8284, "step": 1305 }, { "epoch": 0.17, "grad_norm": 1.4868003129959106, "learning_rate": 9.486668637165536e-06, "loss": 0.882, "step": 1306 }, { "epoch": 0.17, "grad_norm": 1.3504862785339355, "learning_rate": 9.485731627903743e-06, "loss": 0.8401, "step": 1307 }, { "epoch": 0.17, "grad_norm": 1.3063600063323975, "learning_rate": 9.484793810598218e-06, "loss": 0.8066, "step": 1308 }, { "epoch": 0.17, "grad_norm": 1.2924582958221436, "learning_rate": 9.4838551854179e-06, "loss": 0.798, "step": 1309 }, { "epoch": 0.17, "grad_norm": 1.6346269845962524, "learning_rate": 9.482915752531868e-06, "loss": 1.0892, "step": 1310 }, { "epoch": 0.17, "grad_norm": 1.4492403268814087, "learning_rate": 9.481975512109349e-06, "loss": 0.8295, "step": 1311 }, { "epoch": 0.17, "grad_norm": 1.378644585609436, "learning_rate": 9.481034464319713e-06, "loss": 0.8252, "step": 1312 }, { "epoch": 0.17, "grad_norm": 1.4652355909347534, "learning_rate": 9.480092609332477e-06, "loss": 0.8507, "step": 1313 }, { "epoch": 0.17, "grad_norm": 0.7880611419677734, "learning_rate": 9.479149947317306e-06, "loss": 1.0884, "step": 1314 }, { "epoch": 0.17, "grad_norm": 1.456468105316162, "learning_rate": 9.478206478444006e-06, "loss": 0.9272, "step": 1315 }, { "epoch": 0.17, "grad_norm": 0.8521767854690552, "learning_rate": 9.477262202882528e-06, "loss": 1.0957, "step": 1316 }, { "epoch": 0.17, "grad_norm": 0.7221609354019165, "learning_rate": 9.476317120802974e-06, "loss": 1.113, "step": 1317 }, { "epoch": 0.17, "grad_norm": 0.681101381778717, "learning_rate": 9.475371232375588e-06, "loss": 1.0788, "step": 1318 }, { "epoch": 0.17, "grad_norm": 1.410689353942871, "learning_rate": 9.474424537770755e-06, "loss": 0.8771, "step": 1319 }, { "epoch": 0.17, "grad_norm": 1.509526014328003, "learning_rate": 9.473477037159012e-06, "loss": 0.8564, "step": 1320 }, { "epoch": 0.17, "grad_norm": 1.5262292623519897, "learning_rate": 9.472528730711039e-06, "loss": 0.9028, "step": 1321 }, { "epoch": 0.17, "grad_norm": 1.350480318069458, "learning_rate": 9.471579618597657e-06, "loss": 0.8519, "step": 1322 }, { "epoch": 0.17, "grad_norm": 1.3522783517837524, "learning_rate": 9.47062970098984e-06, "loss": 0.8431, "step": 1323 }, { "epoch": 0.17, "grad_norm": 1.5727908611297607, "learning_rate": 9.4696789780587e-06, "loss": 1.0724, "step": 1324 }, { "epoch": 0.17, "grad_norm": 1.1654962301254272, "learning_rate": 9.4687274499755e-06, "loss": 1.0647, "step": 1325 }, { "epoch": 0.17, "grad_norm": 1.4370495080947876, "learning_rate": 9.46777511691164e-06, "loss": 0.846, "step": 1326 }, { "epoch": 0.17, "grad_norm": 1.5739305019378662, "learning_rate": 9.466821979038675e-06, "loss": 0.8768, "step": 1327 }, { "epoch": 0.17, "grad_norm": 1.6404635906219482, "learning_rate": 9.465868036528295e-06, "loss": 0.8739, "step": 1328 }, { "epoch": 0.17, "grad_norm": 1.4506582021713257, "learning_rate": 9.464913289552342e-06, "loss": 0.7952, "step": 1329 }, { "epoch": 0.17, "grad_norm": 1.9795327186584473, "learning_rate": 9.463957738282802e-06, "loss": 1.0879, "step": 1330 }, { "epoch": 0.17, "grad_norm": 1.3901766538619995, "learning_rate": 9.463001382891805e-06, "loss": 0.9208, "step": 1331 }, { "epoch": 0.17, "grad_norm": 1.484792709350586, "learning_rate": 9.462044223551621e-06, "loss": 0.8628, "step": 1332 }, { "epoch": 0.17, "grad_norm": 1.8345932960510254, "learning_rate": 9.461086260434674e-06, "loss": 0.8851, "step": 1333 }, { "epoch": 0.17, "grad_norm": 1.6713322401046753, "learning_rate": 9.460127493713526e-06, "loss": 0.8691, "step": 1334 }, { "epoch": 0.17, "grad_norm": 1.5497745275497437, "learning_rate": 9.459167923560886e-06, "loss": 0.8391, "step": 1335 }, { "epoch": 0.18, "grad_norm": 1.5734200477600098, "learning_rate": 9.458207550149608e-06, "loss": 0.9189, "step": 1336 }, { "epoch": 0.18, "grad_norm": 1.9211746454238892, "learning_rate": 9.45724637365269e-06, "loss": 0.8921, "step": 1337 }, { "epoch": 0.18, "grad_norm": 0.9419506788253784, "learning_rate": 9.456284394243273e-06, "loss": 1.0717, "step": 1338 }, { "epoch": 0.18, "grad_norm": 1.4980762004852295, "learning_rate": 9.455321612094649e-06, "loss": 0.9201, "step": 1339 }, { "epoch": 0.18, "grad_norm": 1.3906235694885254, "learning_rate": 9.454358027380246e-06, "loss": 0.869, "step": 1340 }, { "epoch": 0.18, "grad_norm": 1.6004528999328613, "learning_rate": 9.453393640273642e-06, "loss": 0.8617, "step": 1341 }, { "epoch": 0.18, "grad_norm": 1.5382592678070068, "learning_rate": 9.452428450948562e-06, "loss": 0.8771, "step": 1342 }, { "epoch": 0.18, "grad_norm": 1.3893702030181885, "learning_rate": 9.451462459578864e-06, "loss": 0.9222, "step": 1343 }, { "epoch": 0.18, "grad_norm": 1.1996320486068726, "learning_rate": 9.450495666338563e-06, "loss": 1.0728, "step": 1344 }, { "epoch": 0.18, "grad_norm": 1.5072020292282104, "learning_rate": 9.449528071401816e-06, "loss": 0.901, "step": 1345 }, { "epoch": 0.18, "grad_norm": 1.4380501508712769, "learning_rate": 9.448559674942916e-06, "loss": 0.8935, "step": 1346 }, { "epoch": 0.18, "grad_norm": 1.445411205291748, "learning_rate": 9.447590477136312e-06, "loss": 0.8918, "step": 1347 }, { "epoch": 0.18, "grad_norm": 0.7528914213180542, "learning_rate": 9.446620478156589e-06, "loss": 1.0793, "step": 1348 }, { "epoch": 0.18, "grad_norm": 1.3318743705749512, "learning_rate": 9.445649678178482e-06, "loss": 0.8635, "step": 1349 }, { "epoch": 0.18, "grad_norm": 0.7619823813438416, "learning_rate": 9.444678077376862e-06, "loss": 1.0921, "step": 1350 }, { "epoch": 0.18, "grad_norm": 0.7517566084861755, "learning_rate": 9.443705675926755e-06, "loss": 1.0472, "step": 1351 }, { "epoch": 0.18, "grad_norm": 1.3838615417480469, "learning_rate": 9.442732474003323e-06, "loss": 0.9098, "step": 1352 }, { "epoch": 0.18, "grad_norm": 1.3563282489776611, "learning_rate": 9.441758471781876e-06, "loss": 0.8517, "step": 1353 }, { "epoch": 0.18, "grad_norm": 1.3530272245407104, "learning_rate": 9.440783669437868e-06, "loss": 0.8477, "step": 1354 }, { "epoch": 0.18, "grad_norm": 0.9198051691055298, "learning_rate": 9.439808067146895e-06, "loss": 1.1183, "step": 1355 }, { "epoch": 0.18, "grad_norm": 0.8174216151237488, "learning_rate": 9.4388316650847e-06, "loss": 1.0744, "step": 1356 }, { "epoch": 0.18, "grad_norm": 1.2925758361816406, "learning_rate": 9.437854463427168e-06, "loss": 0.8573, "step": 1357 }, { "epoch": 0.18, "grad_norm": 3.7339630126953125, "learning_rate": 9.43687646235033e-06, "loss": 0.8523, "step": 1358 }, { "epoch": 0.18, "grad_norm": 0.785099446773529, "learning_rate": 9.435897662030354e-06, "loss": 1.0782, "step": 1359 }, { "epoch": 0.18, "grad_norm": 1.3142130374908447, "learning_rate": 9.434918062643564e-06, "loss": 0.8554, "step": 1360 }, { "epoch": 0.18, "grad_norm": 0.7320508360862732, "learning_rate": 9.433937664366419e-06, "loss": 1.0807, "step": 1361 }, { "epoch": 0.18, "grad_norm": 1.4443566799163818, "learning_rate": 9.432956467375525e-06, "loss": 0.923, "step": 1362 }, { "epoch": 0.18, "grad_norm": 1.2836289405822754, "learning_rate": 9.431974471847633e-06, "loss": 0.8565, "step": 1363 }, { "epoch": 0.18, "grad_norm": 1.3381601572036743, "learning_rate": 9.430991677959632e-06, "loss": 0.8958, "step": 1364 }, { "epoch": 0.18, "grad_norm": 0.8099682927131653, "learning_rate": 9.430008085888563e-06, "loss": 1.1081, "step": 1365 }, { "epoch": 0.18, "grad_norm": 0.7158101797103882, "learning_rate": 9.429023695811603e-06, "loss": 1.0856, "step": 1366 }, { "epoch": 0.18, "grad_norm": 1.3943763971328735, "learning_rate": 9.428038507906079e-06, "loss": 0.8356, "step": 1367 }, { "epoch": 0.18, "grad_norm": 0.7492722272872925, "learning_rate": 9.42705252234946e-06, "loss": 1.0608, "step": 1368 }, { "epoch": 0.18, "grad_norm": 1.3238840103149414, "learning_rate": 9.426065739319356e-06, "loss": 0.8574, "step": 1369 }, { "epoch": 0.18, "grad_norm": 1.4194629192352295, "learning_rate": 9.425078158993524e-06, "loss": 0.8839, "step": 1370 }, { "epoch": 0.18, "grad_norm": 1.3588755130767822, "learning_rate": 9.424089781549862e-06, "loss": 0.8866, "step": 1371 }, { "epoch": 0.18, "grad_norm": 1.3375402688980103, "learning_rate": 9.423100607166413e-06, "loss": 0.8785, "step": 1372 }, { "epoch": 0.18, "grad_norm": 1.3761132955551147, "learning_rate": 9.422110636021365e-06, "loss": 0.877, "step": 1373 }, { "epoch": 0.18, "grad_norm": 1.4334734678268433, "learning_rate": 9.421119868293045e-06, "loss": 0.8259, "step": 1374 }, { "epoch": 0.18, "grad_norm": 1.3586546182632446, "learning_rate": 9.420128304159927e-06, "loss": 0.8615, "step": 1375 }, { "epoch": 0.18, "grad_norm": 1.397455096244812, "learning_rate": 9.41913594380063e-06, "loss": 0.8946, "step": 1376 }, { "epoch": 0.18, "grad_norm": 1.458193063735962, "learning_rate": 9.418142787393913e-06, "loss": 0.8656, "step": 1377 }, { "epoch": 0.18, "grad_norm": 1.4000136852264404, "learning_rate": 9.417148835118681e-06, "loss": 0.8759, "step": 1378 }, { "epoch": 0.18, "grad_norm": 1.4124404191970825, "learning_rate": 9.416154087153978e-06, "loss": 0.9122, "step": 1379 }, { "epoch": 0.18, "grad_norm": 1.4329955577850342, "learning_rate": 9.415158543678998e-06, "loss": 0.8654, "step": 1380 }, { "epoch": 0.18, "grad_norm": 1.0442873239517212, "learning_rate": 9.41416220487307e-06, "loss": 1.1111, "step": 1381 }, { "epoch": 0.18, "grad_norm": 0.838023841381073, "learning_rate": 9.413165070915676e-06, "loss": 1.0952, "step": 1382 }, { "epoch": 0.18, "grad_norm": 1.481780767440796, "learning_rate": 9.412167141986433e-06, "loss": 0.8921, "step": 1383 }, { "epoch": 0.18, "grad_norm": 1.530687689781189, "learning_rate": 9.411168418265105e-06, "loss": 0.8948, "step": 1384 }, { "epoch": 0.18, "grad_norm": 1.56992769241333, "learning_rate": 9.410168899931599e-06, "loss": 0.8323, "step": 1385 }, { "epoch": 0.18, "grad_norm": 1.3853996992111206, "learning_rate": 9.409168587165964e-06, "loss": 0.813, "step": 1386 }, { "epoch": 0.18, "grad_norm": 1.4884214401245117, "learning_rate": 9.408167480148394e-06, "loss": 1.1021, "step": 1387 }, { "epoch": 0.18, "grad_norm": 1.4543163776397705, "learning_rate": 9.407165579059224e-06, "loss": 0.8753, "step": 1388 }, { "epoch": 0.18, "grad_norm": 1.6531388759613037, "learning_rate": 9.406162884078932e-06, "loss": 0.8603, "step": 1389 }, { "epoch": 0.18, "grad_norm": 1.0021418333053589, "learning_rate": 9.405159395388141e-06, "loss": 1.0973, "step": 1390 }, { "epoch": 0.18, "grad_norm": 1.4383387565612793, "learning_rate": 9.404155113167617e-06, "loss": 0.8312, "step": 1391 }, { "epoch": 0.18, "grad_norm": 1.3955416679382324, "learning_rate": 9.403150037598266e-06, "loss": 0.8922, "step": 1392 }, { "epoch": 0.18, "grad_norm": 1.4687221050262451, "learning_rate": 9.402144168861139e-06, "loss": 0.8603, "step": 1393 }, { "epoch": 0.18, "grad_norm": 1.4225596189498901, "learning_rate": 9.401137507137428e-06, "loss": 0.8521, "step": 1394 }, { "epoch": 0.18, "grad_norm": 1.490239143371582, "learning_rate": 9.400130052608473e-06, "loss": 0.9226, "step": 1395 }, { "epoch": 0.18, "grad_norm": 1.3788493871688843, "learning_rate": 9.399121805455753e-06, "loss": 0.8483, "step": 1396 }, { "epoch": 0.18, "grad_norm": 1.638862133026123, "learning_rate": 9.398112765860887e-06, "loss": 1.0993, "step": 1397 }, { "epoch": 0.18, "grad_norm": 1.5151784420013428, "learning_rate": 9.397102934005642e-06, "loss": 0.8914, "step": 1398 }, { "epoch": 0.18, "grad_norm": 1.3920468091964722, "learning_rate": 9.396092310071924e-06, "loss": 0.8842, "step": 1399 }, { "epoch": 0.18, "grad_norm": 0.9492843747138977, "learning_rate": 9.395080894241786e-06, "loss": 1.0952, "step": 1400 }, { "epoch": 0.18, "grad_norm": 0.767254114151001, "learning_rate": 9.394068686697418e-06, "loss": 1.1136, "step": 1401 }, { "epoch": 0.18, "grad_norm": 0.7739744782447815, "learning_rate": 9.393055687621155e-06, "loss": 1.096, "step": 1402 }, { "epoch": 0.18, "grad_norm": 0.8964375257492065, "learning_rate": 9.392041897195478e-06, "loss": 1.0746, "step": 1403 }, { "epoch": 0.18, "grad_norm": 1.5417048931121826, "learning_rate": 9.391027315603007e-06, "loss": 0.8499, "step": 1404 }, { "epoch": 0.18, "grad_norm": 1.6159498691558838, "learning_rate": 9.390011943026505e-06, "loss": 0.9166, "step": 1405 }, { "epoch": 0.18, "grad_norm": 1.5196301937103271, "learning_rate": 9.388995779648875e-06, "loss": 0.8667, "step": 1406 }, { "epoch": 0.18, "grad_norm": 1.1233733892440796, "learning_rate": 9.387978825653169e-06, "loss": 1.0745, "step": 1407 }, { "epoch": 0.18, "grad_norm": 1.3710861206054688, "learning_rate": 9.386961081222572e-06, "loss": 0.9303, "step": 1408 }, { "epoch": 0.18, "grad_norm": 0.8156165480613708, "learning_rate": 9.38594254654042e-06, "loss": 1.0508, "step": 1409 }, { "epoch": 0.18, "grad_norm": 0.8172760605812073, "learning_rate": 9.384923221790192e-06, "loss": 1.0816, "step": 1410 }, { "epoch": 0.18, "grad_norm": 0.8552246689796448, "learning_rate": 9.3839031071555e-06, "loss": 1.0622, "step": 1411 }, { "epoch": 0.19, "grad_norm": 1.5248708724975586, "learning_rate": 9.382882202820107e-06, "loss": 0.8541, "step": 1412 }, { "epoch": 0.19, "grad_norm": 0.7196890711784363, "learning_rate": 9.381860508967913e-06, "loss": 1.1102, "step": 1413 }, { "epoch": 0.19, "grad_norm": 0.7516447305679321, "learning_rate": 9.380838025782964e-06, "loss": 1.0886, "step": 1414 }, { "epoch": 0.19, "grad_norm": 1.4082685708999634, "learning_rate": 9.379814753449445e-06, "loss": 0.8493, "step": 1415 }, { "epoch": 0.19, "grad_norm": 1.3968108892440796, "learning_rate": 9.378790692151685e-06, "loss": 0.8768, "step": 1416 }, { "epoch": 0.19, "grad_norm": 1.426833152770996, "learning_rate": 9.377765842074155e-06, "loss": 0.8613, "step": 1417 }, { "epoch": 0.19, "grad_norm": 1.369166612625122, "learning_rate": 9.376740203401468e-06, "loss": 0.8839, "step": 1418 }, { "epoch": 0.19, "grad_norm": 1.3933422565460205, "learning_rate": 9.375713776318381e-06, "loss": 0.8685, "step": 1419 }, { "epoch": 0.19, "grad_norm": 1.2996107339859009, "learning_rate": 9.374686561009786e-06, "loss": 0.8646, "step": 1420 }, { "epoch": 0.19, "grad_norm": 1.4129893779754639, "learning_rate": 9.373658557660726e-06, "loss": 0.8552, "step": 1421 }, { "epoch": 0.19, "grad_norm": 1.621646761894226, "learning_rate": 9.37262976645638e-06, "loss": 0.8769, "step": 1422 }, { "epoch": 0.19, "grad_norm": 1.393646240234375, "learning_rate": 9.37160018758207e-06, "loss": 1.0797, "step": 1423 }, { "epoch": 0.19, "grad_norm": 1.3885618448257446, "learning_rate": 9.370569821223264e-06, "loss": 0.8682, "step": 1424 }, { "epoch": 0.19, "grad_norm": 0.825641930103302, "learning_rate": 9.369538667565566e-06, "loss": 1.0948, "step": 1425 }, { "epoch": 0.19, "grad_norm": 1.4548072814941406, "learning_rate": 9.368506726794725e-06, "loss": 0.8893, "step": 1426 }, { "epoch": 0.19, "grad_norm": 1.5937960147857666, "learning_rate": 9.36747399909663e-06, "loss": 0.8454, "step": 1427 }, { "epoch": 0.19, "grad_norm": 1.3376721143722534, "learning_rate": 9.366440484657316e-06, "loss": 0.8325, "step": 1428 }, { "epoch": 0.19, "grad_norm": 1.324638843536377, "learning_rate": 9.365406183662952e-06, "loss": 0.8361, "step": 1429 }, { "epoch": 0.19, "grad_norm": 1.3351768255233765, "learning_rate": 9.364371096299855e-06, "loss": 0.8382, "step": 1430 }, { "epoch": 0.19, "grad_norm": 1.400344729423523, "learning_rate": 9.363335222754486e-06, "loss": 0.892, "step": 1431 }, { "epoch": 0.19, "grad_norm": 1.5255377292633057, "learning_rate": 9.36229856321344e-06, "loss": 0.8539, "step": 1432 }, { "epoch": 0.19, "grad_norm": 1.355898380279541, "learning_rate": 9.361261117863457e-06, "loss": 0.9081, "step": 1433 }, { "epoch": 0.19, "grad_norm": 1.5066516399383545, "learning_rate": 9.360222886891417e-06, "loss": 0.8861, "step": 1434 }, { "epoch": 0.19, "grad_norm": 1.340270757675171, "learning_rate": 9.359183870484347e-06, "loss": 0.855, "step": 1435 }, { "epoch": 0.19, "grad_norm": 1.6599119901657104, "learning_rate": 9.35814406882941e-06, "loss": 1.0673, "step": 1436 }, { "epoch": 0.19, "grad_norm": 1.3317086696624756, "learning_rate": 9.357103482113912e-06, "loss": 0.9044, "step": 1437 }, { "epoch": 0.19, "grad_norm": 1.5031999349594116, "learning_rate": 9.356062110525302e-06, "loss": 0.8475, "step": 1438 }, { "epoch": 0.19, "grad_norm": 1.3913251161575317, "learning_rate": 9.355019954251165e-06, "loss": 0.8577, "step": 1439 }, { "epoch": 0.19, "grad_norm": 0.8534239530563354, "learning_rate": 9.353977013479238e-06, "loss": 1.0789, "step": 1440 }, { "epoch": 0.19, "grad_norm": 1.4796650409698486, "learning_rate": 9.352933288397387e-06, "loss": 0.9584, "step": 1441 }, { "epoch": 0.19, "grad_norm": 1.4318147897720337, "learning_rate": 9.351888779193626e-06, "loss": 0.8938, "step": 1442 }, { "epoch": 0.19, "grad_norm": 1.3935739994049072, "learning_rate": 9.35084348605611e-06, "loss": 0.914, "step": 1443 }, { "epoch": 0.19, "grad_norm": 1.2370229959487915, "learning_rate": 9.349797409173136e-06, "loss": 0.8509, "step": 1444 }, { "epoch": 0.19, "grad_norm": 1.366385817527771, "learning_rate": 9.348750548733138e-06, "loss": 0.8362, "step": 1445 }, { "epoch": 0.19, "grad_norm": 1.3907369375228882, "learning_rate": 9.347702904924694e-06, "loss": 0.8463, "step": 1446 }, { "epoch": 0.19, "grad_norm": 1.352317452430725, "learning_rate": 9.346654477936523e-06, "loss": 1.0722, "step": 1447 }, { "epoch": 0.19, "grad_norm": 1.3289299011230469, "learning_rate": 9.345605267957487e-06, "loss": 0.8573, "step": 1448 }, { "epoch": 0.19, "grad_norm": 1.6326111555099487, "learning_rate": 9.344555275176585e-06, "loss": 0.8521, "step": 1449 }, { "epoch": 0.19, "grad_norm": 0.7701502442359924, "learning_rate": 9.343504499782958e-06, "loss": 1.0785, "step": 1450 }, { "epoch": 0.19, "grad_norm": 1.3656245470046997, "learning_rate": 9.342452941965892e-06, "loss": 0.8376, "step": 1451 }, { "epoch": 0.19, "grad_norm": 0.6701757907867432, "learning_rate": 9.341400601914807e-06, "loss": 1.075, "step": 1452 }, { "epoch": 0.19, "grad_norm": 1.4024765491485596, "learning_rate": 9.340347479819272e-06, "loss": 0.8499, "step": 1453 }, { "epoch": 0.19, "grad_norm": 0.7564851641654968, "learning_rate": 9.339293575868988e-06, "loss": 1.0637, "step": 1454 }, { "epoch": 0.19, "grad_norm": 1.6117775440216064, "learning_rate": 9.338238890253805e-06, "loss": 0.8915, "step": 1455 }, { "epoch": 0.19, "grad_norm": 1.6304636001586914, "learning_rate": 9.33718342316371e-06, "loss": 0.8336, "step": 1456 }, { "epoch": 0.19, "grad_norm": 0.7551613450050354, "learning_rate": 9.33612717478883e-06, "loss": 1.0925, "step": 1457 }, { "epoch": 0.19, "grad_norm": 0.7011208534240723, "learning_rate": 9.335070145319433e-06, "loss": 1.077, "step": 1458 }, { "epoch": 0.19, "grad_norm": 0.6193296909332275, "learning_rate": 9.334012334945932e-06, "loss": 1.0741, "step": 1459 }, { "epoch": 0.19, "grad_norm": 0.7227243185043335, "learning_rate": 9.332953743858874e-06, "loss": 1.0817, "step": 1460 }, { "epoch": 0.19, "grad_norm": 1.4116978645324707, "learning_rate": 9.33189437224895e-06, "loss": 0.8783, "step": 1461 }, { "epoch": 0.19, "grad_norm": 1.4247901439666748, "learning_rate": 9.330834220306995e-06, "loss": 0.889, "step": 1462 }, { "epoch": 0.19, "grad_norm": 1.4029723405838013, "learning_rate": 9.329773288223976e-06, "loss": 0.8823, "step": 1463 }, { "epoch": 0.19, "grad_norm": 1.420514464378357, "learning_rate": 9.328711576191007e-06, "loss": 0.8472, "step": 1464 }, { "epoch": 0.19, "grad_norm": 1.4213595390319824, "learning_rate": 9.327649084399341e-06, "loss": 0.796, "step": 1465 }, { "epoch": 0.19, "grad_norm": 1.3259559869766235, "learning_rate": 9.326585813040377e-06, "loss": 0.874, "step": 1466 }, { "epoch": 0.19, "grad_norm": 1.419091820716858, "learning_rate": 9.325521762305639e-06, "loss": 0.8505, "step": 1467 }, { "epoch": 0.19, "grad_norm": 1.4818387031555176, "learning_rate": 9.324456932386809e-06, "loss": 0.8164, "step": 1468 }, { "epoch": 0.19, "grad_norm": 1.4225730895996094, "learning_rate": 9.323391323475699e-06, "loss": 0.9278, "step": 1469 }, { "epoch": 0.19, "grad_norm": 1.380539059638977, "learning_rate": 9.322324935764264e-06, "loss": 0.89, "step": 1470 }, { "epoch": 0.19, "grad_norm": 1.4754348993301392, "learning_rate": 9.321257769444598e-06, "loss": 0.873, "step": 1471 }, { "epoch": 0.19, "grad_norm": 1.3761647939682007, "learning_rate": 9.32018982470894e-06, "loss": 0.8005, "step": 1472 }, { "epoch": 0.19, "grad_norm": 1.4271438121795654, "learning_rate": 9.319121101749664e-06, "loss": 0.9029, "step": 1473 }, { "epoch": 0.19, "grad_norm": 1.4456696510314941, "learning_rate": 9.318051600759284e-06, "loss": 0.8791, "step": 1474 }, { "epoch": 0.19, "grad_norm": 1.6439743041992188, "learning_rate": 9.316981321930459e-06, "loss": 0.8825, "step": 1475 }, { "epoch": 0.19, "grad_norm": 1.251871943473816, "learning_rate": 9.315910265455983e-06, "loss": 0.856, "step": 1476 }, { "epoch": 0.19, "grad_norm": 1.4335570335388184, "learning_rate": 9.314838431528794e-06, "loss": 0.8596, "step": 1477 }, { "epoch": 0.19, "grad_norm": 1.3701590299606323, "learning_rate": 9.313765820341968e-06, "loss": 0.8299, "step": 1478 }, { "epoch": 0.19, "grad_norm": 1.0720679759979248, "learning_rate": 9.31269243208872e-06, "loss": 1.0656, "step": 1479 }, { "epoch": 0.19, "grad_norm": 1.3450844287872314, "learning_rate": 9.311618266962408e-06, "loss": 0.8223, "step": 1480 }, { "epoch": 0.19, "grad_norm": 1.6016170978546143, "learning_rate": 9.310543325156529e-06, "loss": 0.8913, "step": 1481 }, { "epoch": 0.19, "grad_norm": 1.5905847549438477, "learning_rate": 9.309467606864715e-06, "loss": 0.8873, "step": 1482 }, { "epoch": 0.19, "grad_norm": 1.3575339317321777, "learning_rate": 9.308391112280747e-06, "loss": 0.8787, "step": 1483 }, { "epoch": 0.19, "grad_norm": 0.8812870979309082, "learning_rate": 9.307313841598538e-06, "loss": 1.0835, "step": 1484 }, { "epoch": 0.19, "grad_norm": 0.8017166256904602, "learning_rate": 9.306235795012144e-06, "loss": 1.1044, "step": 1485 }, { "epoch": 0.19, "grad_norm": 0.6602761149406433, "learning_rate": 9.305156972715762e-06, "loss": 1.0679, "step": 1486 }, { "epoch": 0.19, "grad_norm": 1.4622806310653687, "learning_rate": 9.304077374903726e-06, "loss": 0.9013, "step": 1487 }, { "epoch": 0.19, "grad_norm": 0.9760235548019409, "learning_rate": 9.302997001770511e-06, "loss": 1.0648, "step": 1488 }, { "epoch": 0.2, "grad_norm": 1.3890328407287598, "learning_rate": 9.301915853510733e-06, "loss": 0.8507, "step": 1489 }, { "epoch": 0.2, "grad_norm": 1.3539079427719116, "learning_rate": 9.300833930319142e-06, "loss": 0.9274, "step": 1490 }, { "epoch": 0.2, "grad_norm": 1.3317608833312988, "learning_rate": 9.299751232390638e-06, "loss": 0.8894, "step": 1491 }, { "epoch": 0.2, "grad_norm": 0.8934832811355591, "learning_rate": 9.298667759920248e-06, "loss": 1.1021, "step": 1492 }, { "epoch": 0.2, "grad_norm": 1.4396631717681885, "learning_rate": 9.29758351310315e-06, "loss": 0.9069, "step": 1493 }, { "epoch": 0.2, "grad_norm": 1.4599579572677612, "learning_rate": 9.296498492134653e-06, "loss": 0.8399, "step": 1494 }, { "epoch": 0.2, "grad_norm": 1.5318454504013062, "learning_rate": 9.29541269721021e-06, "loss": 0.8688, "step": 1495 }, { "epoch": 0.2, "grad_norm": 1.3299007415771484, "learning_rate": 9.294326128525413e-06, "loss": 0.8665, "step": 1496 }, { "epoch": 0.2, "grad_norm": 1.3672480583190918, "learning_rate": 9.29323878627599e-06, "loss": 0.8889, "step": 1497 }, { "epoch": 0.2, "grad_norm": 1.4886951446533203, "learning_rate": 9.292150670657817e-06, "loss": 0.8131, "step": 1498 }, { "epoch": 0.2, "grad_norm": 1.288751482963562, "learning_rate": 9.291061781866896e-06, "loss": 0.8748, "step": 1499 }, { "epoch": 0.2, "grad_norm": 1.4925278425216675, "learning_rate": 9.28997212009938e-06, "loss": 0.8654, "step": 1500 }, { "epoch": 0.2, "grad_norm": 1.3489280939102173, "learning_rate": 9.288881685551554e-06, "loss": 0.8665, "step": 1501 }, { "epoch": 0.2, "grad_norm": 1.371575117111206, "learning_rate": 9.287790478419849e-06, "loss": 0.8396, "step": 1502 }, { "epoch": 0.2, "grad_norm": 0.8957815170288086, "learning_rate": 9.286698498900825e-06, "loss": 1.0767, "step": 1503 }, { "epoch": 0.2, "grad_norm": 1.4231122732162476, "learning_rate": 9.285605747191193e-06, "loss": 0.8884, "step": 1504 }, { "epoch": 0.2, "grad_norm": 1.353060007095337, "learning_rate": 9.284512223487797e-06, "loss": 0.8168, "step": 1505 }, { "epoch": 0.2, "grad_norm": 1.380187749862671, "learning_rate": 9.283417927987616e-06, "loss": 0.8932, "step": 1506 }, { "epoch": 0.2, "grad_norm": 1.3582398891448975, "learning_rate": 9.282322860887778e-06, "loss": 0.906, "step": 1507 }, { "epoch": 0.2, "grad_norm": 0.7671614289283752, "learning_rate": 9.28122702238554e-06, "loss": 1.0929, "step": 1508 }, { "epoch": 0.2, "grad_norm": 1.6624308824539185, "learning_rate": 9.280130412678306e-06, "loss": 0.8821, "step": 1509 }, { "epoch": 0.2, "grad_norm": 0.7030964493751526, "learning_rate": 9.27903303196361e-06, "loss": 1.1111, "step": 1510 }, { "epoch": 0.2, "grad_norm": 1.3878520727157593, "learning_rate": 9.277934880439138e-06, "loss": 0.896, "step": 1511 }, { "epoch": 0.2, "grad_norm": 1.3275320529937744, "learning_rate": 9.276835958302702e-06, "loss": 0.846, "step": 1512 }, { "epoch": 0.2, "grad_norm": 1.2899580001831055, "learning_rate": 9.275736265752257e-06, "loss": 0.8229, "step": 1513 }, { "epoch": 0.2, "grad_norm": 1.3326547145843506, "learning_rate": 9.2746358029859e-06, "loss": 0.9178, "step": 1514 }, { "epoch": 0.2, "grad_norm": 1.3702659606933594, "learning_rate": 9.273534570201867e-06, "loss": 0.9047, "step": 1515 }, { "epoch": 0.2, "grad_norm": 1.2587273120880127, "learning_rate": 9.272432567598523e-06, "loss": 0.8651, "step": 1516 }, { "epoch": 0.2, "grad_norm": 1.370261311531067, "learning_rate": 9.271329795374386e-06, "loss": 0.854, "step": 1517 }, { "epoch": 0.2, "grad_norm": 1.3478281497955322, "learning_rate": 9.270226253728104e-06, "loss": 0.8989, "step": 1518 }, { "epoch": 0.2, "grad_norm": 1.4169342517852783, "learning_rate": 9.26912194285846e-06, "loss": 0.8302, "step": 1519 }, { "epoch": 0.2, "grad_norm": 1.2408888339996338, "learning_rate": 9.268016862964386e-06, "loss": 0.7731, "step": 1520 }, { "epoch": 0.2, "grad_norm": 1.3736968040466309, "learning_rate": 9.266911014244946e-06, "loss": 0.7581, "step": 1521 }, { "epoch": 0.2, "grad_norm": 1.3736655712127686, "learning_rate": 9.26580439689934e-06, "loss": 0.8043, "step": 1522 }, { "epoch": 0.2, "grad_norm": 1.205118179321289, "learning_rate": 9.264697011126916e-06, "loss": 1.0861, "step": 1523 }, { "epoch": 0.2, "grad_norm": 0.8760619759559631, "learning_rate": 9.263588857127152e-06, "loss": 1.0981, "step": 1524 }, { "epoch": 0.2, "grad_norm": 1.375063419342041, "learning_rate": 9.262479935099665e-06, "loss": 0.8276, "step": 1525 }, { "epoch": 0.2, "grad_norm": 1.463663101196289, "learning_rate": 9.261370245244214e-06, "loss": 0.8214, "step": 1526 }, { "epoch": 0.2, "grad_norm": 1.4292796850204468, "learning_rate": 9.260259787760695e-06, "loss": 0.8525, "step": 1527 }, { "epoch": 0.2, "grad_norm": 1.3911385536193848, "learning_rate": 9.25914856284914e-06, "loss": 0.932, "step": 1528 }, { "epoch": 0.2, "grad_norm": 1.3465534448623657, "learning_rate": 9.258036570709724e-06, "loss": 0.8711, "step": 1529 }, { "epoch": 0.2, "grad_norm": 1.6794798374176025, "learning_rate": 9.256923811542755e-06, "loss": 1.1138, "step": 1530 }, { "epoch": 0.2, "grad_norm": 1.3002338409423828, "learning_rate": 9.255810285548684e-06, "loss": 0.7927, "step": 1531 }, { "epoch": 0.2, "grad_norm": 1.6096510887145996, "learning_rate": 9.254695992928094e-06, "loss": 0.8172, "step": 1532 }, { "epoch": 0.2, "grad_norm": 1.4025838375091553, "learning_rate": 9.253580933881711e-06, "loss": 0.8428, "step": 1533 }, { "epoch": 0.2, "grad_norm": 1.468554139137268, "learning_rate": 9.252465108610397e-06, "loss": 0.8709, "step": 1534 }, { "epoch": 0.2, "grad_norm": 1.4448219537734985, "learning_rate": 9.251348517315158e-06, "loss": 0.8774, "step": 1535 }, { "epoch": 0.2, "grad_norm": 1.336504578590393, "learning_rate": 9.250231160197127e-06, "loss": 0.8212, "step": 1536 }, { "epoch": 0.2, "grad_norm": 1.303905725479126, "learning_rate": 9.249113037457582e-06, "loss": 0.8925, "step": 1537 }, { "epoch": 0.2, "grad_norm": 1.3818609714508057, "learning_rate": 9.247994149297938e-06, "loss": 0.859, "step": 1538 }, { "epoch": 0.2, "grad_norm": 0.9422876834869385, "learning_rate": 9.246874495919749e-06, "loss": 1.1018, "step": 1539 }, { "epoch": 0.2, "grad_norm": 1.6498897075653076, "learning_rate": 9.245754077524702e-06, "loss": 0.906, "step": 1540 }, { "epoch": 0.2, "grad_norm": 1.380821943283081, "learning_rate": 9.244632894314628e-06, "loss": 0.8677, "step": 1541 }, { "epoch": 0.2, "grad_norm": 0.758248507976532, "learning_rate": 9.243510946491494e-06, "loss": 1.1007, "step": 1542 }, { "epoch": 0.2, "grad_norm": 1.4063104391098022, "learning_rate": 9.242388234257402e-06, "loss": 0.8699, "step": 1543 }, { "epoch": 0.2, "grad_norm": 1.2846015691757202, "learning_rate": 9.241264757814591e-06, "loss": 0.8412, "step": 1544 }, { "epoch": 0.2, "grad_norm": 1.365959644317627, "learning_rate": 9.240140517365445e-06, "loss": 0.8648, "step": 1545 }, { "epoch": 0.2, "grad_norm": 0.7958374619483948, "learning_rate": 9.239015513112476e-06, "loss": 1.0709, "step": 1546 }, { "epoch": 0.2, "grad_norm": 0.6515523195266724, "learning_rate": 9.237889745258342e-06, "loss": 1.0596, "step": 1547 }, { "epoch": 0.2, "grad_norm": 0.64351886510849, "learning_rate": 9.236763214005832e-06, "loss": 1.1066, "step": 1548 }, { "epoch": 0.2, "grad_norm": 1.4398231506347656, "learning_rate": 9.235635919557876e-06, "loss": 0.8387, "step": 1549 }, { "epoch": 0.2, "grad_norm": 1.3798593282699585, "learning_rate": 9.234507862117542e-06, "loss": 0.838, "step": 1550 }, { "epoch": 0.2, "grad_norm": 1.2602958679199219, "learning_rate": 9.233379041888034e-06, "loss": 0.8516, "step": 1551 }, { "epoch": 0.2, "grad_norm": 1.338721752166748, "learning_rate": 9.232249459072693e-06, "loss": 0.8981, "step": 1552 }, { "epoch": 0.2, "grad_norm": 1.062838077545166, "learning_rate": 9.231119113874998e-06, "loss": 1.0758, "step": 1553 }, { "epoch": 0.2, "grad_norm": 0.8939120769500732, "learning_rate": 9.229988006498564e-06, "loss": 1.0901, "step": 1554 }, { "epoch": 0.2, "grad_norm": 0.6826174855232239, "learning_rate": 9.228856137147149e-06, "loss": 1.0557, "step": 1555 }, { "epoch": 0.2, "grad_norm": 1.4734843969345093, "learning_rate": 9.22772350602464e-06, "loss": 0.8514, "step": 1556 }, { "epoch": 0.2, "grad_norm": 1.4075063467025757, "learning_rate": 9.226590113335065e-06, "loss": 0.8684, "step": 1557 }, { "epoch": 0.2, "grad_norm": 1.449388027191162, "learning_rate": 9.225455959282592e-06, "loss": 0.8837, "step": 1558 }, { "epoch": 0.2, "grad_norm": 1.4233202934265137, "learning_rate": 9.224321044071522e-06, "loss": 0.8284, "step": 1559 }, { "epoch": 0.2, "grad_norm": 2.6876840591430664, "learning_rate": 9.223185367906293e-06, "loss": 0.7997, "step": 1560 }, { "epoch": 0.2, "grad_norm": 1.9703329801559448, "learning_rate": 9.222048930991485e-06, "loss": 1.0995, "step": 1561 }, { "epoch": 0.2, "grad_norm": 1.39277982711792, "learning_rate": 9.22091173353181e-06, "loss": 0.8395, "step": 1562 }, { "epoch": 0.2, "grad_norm": 1.3539443016052246, "learning_rate": 9.21977377573212e-06, "loss": 0.8032, "step": 1563 }, { "epoch": 0.2, "grad_norm": 1.3999414443969727, "learning_rate": 9.218635057797399e-06, "loss": 0.8788, "step": 1564 }, { "epoch": 0.21, "grad_norm": 1.3268327713012695, "learning_rate": 9.217495579932775e-06, "loss": 0.8177, "step": 1565 }, { "epoch": 0.21, "grad_norm": 1.3150286674499512, "learning_rate": 9.216355342343508e-06, "loss": 0.8757, "step": 1566 }, { "epoch": 0.21, "grad_norm": 1.4113048315048218, "learning_rate": 9.215214345234998e-06, "loss": 0.7913, "step": 1567 }, { "epoch": 0.21, "grad_norm": 0.956439197063446, "learning_rate": 9.214072588812779e-06, "loss": 1.0906, "step": 1568 }, { "epoch": 0.21, "grad_norm": 1.4223291873931885, "learning_rate": 9.21293007328252e-06, "loss": 0.8662, "step": 1569 }, { "epoch": 0.21, "grad_norm": 1.3996999263763428, "learning_rate": 9.211786798850036e-06, "loss": 0.9015, "step": 1570 }, { "epoch": 0.21, "grad_norm": 1.4007549285888672, "learning_rate": 9.210642765721267e-06, "loss": 0.8354, "step": 1571 }, { "epoch": 0.21, "grad_norm": 0.7561197280883789, "learning_rate": 9.209497974102298e-06, "loss": 1.0598, "step": 1572 }, { "epoch": 0.21, "grad_norm": 1.338181495666504, "learning_rate": 9.208352424199347e-06, "loss": 0.9148, "step": 1573 }, { "epoch": 0.21, "grad_norm": 1.2763174772262573, "learning_rate": 9.207206116218769e-06, "loss": 0.8752, "step": 1574 }, { "epoch": 0.21, "grad_norm": 0.7563900947570801, "learning_rate": 9.206059050367055e-06, "loss": 1.0632, "step": 1575 }, { "epoch": 0.21, "grad_norm": 0.7244670987129211, "learning_rate": 9.204911226850835e-06, "loss": 1.0592, "step": 1576 }, { "epoch": 0.21, "grad_norm": 1.3825713396072388, "learning_rate": 9.203762645876872e-06, "loss": 0.8706, "step": 1577 }, { "epoch": 0.21, "grad_norm": 1.41713547706604, "learning_rate": 9.202613307652068e-06, "loss": 0.8364, "step": 1578 }, { "epoch": 0.21, "grad_norm": 1.5038472414016724, "learning_rate": 9.201463212383463e-06, "loss": 0.8772, "step": 1579 }, { "epoch": 0.21, "grad_norm": 1.3805770874023438, "learning_rate": 9.200312360278227e-06, "loss": 0.8598, "step": 1580 }, { "epoch": 0.21, "grad_norm": 1.4555543661117554, "learning_rate": 9.199160751543674e-06, "loss": 0.8456, "step": 1581 }, { "epoch": 0.21, "grad_norm": 1.384750485420227, "learning_rate": 9.19800838638725e-06, "loss": 0.8868, "step": 1582 }, { "epoch": 0.21, "grad_norm": 1.0839464664459229, "learning_rate": 9.196855265016535e-06, "loss": 1.0826, "step": 1583 }, { "epoch": 0.21, "grad_norm": 0.9104017615318298, "learning_rate": 9.195701387639253e-06, "loss": 1.0699, "step": 1584 }, { "epoch": 0.21, "grad_norm": 0.6760470867156982, "learning_rate": 9.194546754463255e-06, "loss": 1.0956, "step": 1585 }, { "epoch": 0.21, "grad_norm": 1.4556983709335327, "learning_rate": 9.193391365696534e-06, "loss": 0.8413, "step": 1586 }, { "epoch": 0.21, "grad_norm": 1.4947521686553955, "learning_rate": 9.192235221547218e-06, "loss": 0.8923, "step": 1587 }, { "epoch": 0.21, "grad_norm": 1.229750633239746, "learning_rate": 9.191078322223572e-06, "loss": 1.0611, "step": 1588 }, { "epoch": 0.21, "grad_norm": 1.4097541570663452, "learning_rate": 9.189920667933991e-06, "loss": 0.8843, "step": 1589 }, { "epoch": 0.21, "grad_norm": 1.3242521286010742, "learning_rate": 9.188762258887016e-06, "loss": 0.8369, "step": 1590 }, { "epoch": 0.21, "grad_norm": 1.0717897415161133, "learning_rate": 9.187603095291318e-06, "loss": 1.0829, "step": 1591 }, { "epoch": 0.21, "grad_norm": 1.4355746507644653, "learning_rate": 9.186443177355702e-06, "loss": 0.8571, "step": 1592 }, { "epoch": 0.21, "grad_norm": 1.3125669956207275, "learning_rate": 9.185282505289113e-06, "loss": 0.9163, "step": 1593 }, { "epoch": 0.21, "grad_norm": 0.8319694399833679, "learning_rate": 9.184121079300629e-06, "loss": 1.0837, "step": 1594 }, { "epoch": 0.21, "grad_norm": 1.4622987508773804, "learning_rate": 9.182958899599466e-06, "loss": 0.8172, "step": 1595 }, { "epoch": 0.21, "grad_norm": 1.2991074323654175, "learning_rate": 9.181795966394977e-06, "loss": 0.8651, "step": 1596 }, { "epoch": 0.21, "grad_norm": 1.318095326423645, "learning_rate": 9.180632279896646e-06, "loss": 0.8511, "step": 1597 }, { "epoch": 0.21, "grad_norm": 1.3805311918258667, "learning_rate": 9.179467840314096e-06, "loss": 0.9097, "step": 1598 }, { "epoch": 0.21, "grad_norm": 1.2825580835342407, "learning_rate": 9.178302647857085e-06, "loss": 0.8416, "step": 1599 }, { "epoch": 0.21, "grad_norm": 1.4156479835510254, "learning_rate": 9.177136702735507e-06, "loss": 0.8233, "step": 1600 }, { "epoch": 0.21, "grad_norm": 1.379021167755127, "learning_rate": 9.17597000515939e-06, "loss": 0.8095, "step": 1601 }, { "epoch": 0.21, "grad_norm": 1.4249799251556396, "learning_rate": 9.174802555338901e-06, "loss": 0.8924, "step": 1602 }, { "epoch": 0.21, "grad_norm": 1.300891399383545, "learning_rate": 9.17363435348434e-06, "loss": 1.0809, "step": 1603 }, { "epoch": 0.21, "grad_norm": 1.555829405784607, "learning_rate": 9.172465399806141e-06, "loss": 0.9352, "step": 1604 }, { "epoch": 0.21, "grad_norm": 1.4161720275878906, "learning_rate": 9.171295694514875e-06, "loss": 0.856, "step": 1605 }, { "epoch": 0.21, "grad_norm": 1.4698694944381714, "learning_rate": 9.17012523782125e-06, "loss": 0.8359, "step": 1606 }, { "epoch": 0.21, "grad_norm": 0.7653726935386658, "learning_rate": 9.168954029936105e-06, "loss": 1.1044, "step": 1607 }, { "epoch": 0.21, "grad_norm": 1.375754952430725, "learning_rate": 9.167782071070423e-06, "loss": 0.8044, "step": 1608 }, { "epoch": 0.21, "grad_norm": 1.3788819313049316, "learning_rate": 9.16660936143531e-06, "loss": 0.8721, "step": 1609 }, { "epoch": 0.21, "grad_norm": 1.4104304313659668, "learning_rate": 9.165435901242019e-06, "loss": 0.8444, "step": 1610 }, { "epoch": 0.21, "grad_norm": 1.4523276090621948, "learning_rate": 9.16426169070193e-06, "loss": 0.8852, "step": 1611 }, { "epoch": 0.21, "grad_norm": 1.0971168279647827, "learning_rate": 9.163086730026562e-06, "loss": 1.082, "step": 1612 }, { "epoch": 0.21, "grad_norm": 1.3154042959213257, "learning_rate": 9.16191101942757e-06, "loss": 0.8287, "step": 1613 }, { "epoch": 0.21, "grad_norm": 1.3367840051651, "learning_rate": 9.160734559116738e-06, "loss": 0.8294, "step": 1614 }, { "epoch": 0.21, "grad_norm": 1.3805824518203735, "learning_rate": 9.159557349305991e-06, "loss": 0.8918, "step": 1615 }, { "epoch": 0.21, "grad_norm": 1.5855810642242432, "learning_rate": 9.158379390207389e-06, "loss": 0.8562, "step": 1616 }, { "epoch": 0.21, "grad_norm": 1.3834675550460815, "learning_rate": 9.157200682033126e-06, "loss": 0.9395, "step": 1617 }, { "epoch": 0.21, "grad_norm": 1.3400195837020874, "learning_rate": 9.156021224995526e-06, "loss": 0.8317, "step": 1618 }, { "epoch": 0.21, "grad_norm": 1.361720085144043, "learning_rate": 9.154841019307057e-06, "loss": 0.8174, "step": 1619 }, { "epoch": 0.21, "grad_norm": 1.2807002067565918, "learning_rate": 9.153660065180314e-06, "loss": 0.8143, "step": 1620 }, { "epoch": 0.21, "grad_norm": 1.3166614770889282, "learning_rate": 9.15247836282803e-06, "loss": 0.8373, "step": 1621 }, { "epoch": 0.21, "grad_norm": 1.3475087881088257, "learning_rate": 9.151295912463076e-06, "loss": 0.8946, "step": 1622 }, { "epoch": 0.21, "grad_norm": 1.110578179359436, "learning_rate": 9.150112714298451e-06, "loss": 1.0814, "step": 1623 }, { "epoch": 0.21, "grad_norm": 1.3773102760314941, "learning_rate": 9.148928768547292e-06, "loss": 0.8603, "step": 1624 }, { "epoch": 0.21, "grad_norm": 1.3450895547866821, "learning_rate": 9.147744075422874e-06, "loss": 0.8531, "step": 1625 }, { "epoch": 0.21, "grad_norm": 1.353878140449524, "learning_rate": 9.1465586351386e-06, "loss": 0.8988, "step": 1626 }, { "epoch": 0.21, "grad_norm": 1.4774506092071533, "learning_rate": 9.145372447908011e-06, "loss": 0.8063, "step": 1627 }, { "epoch": 0.21, "grad_norm": 0.9731611609458923, "learning_rate": 9.144185513944785e-06, "loss": 1.0703, "step": 1628 }, { "epoch": 0.21, "grad_norm": 1.609370231628418, "learning_rate": 9.142997833462733e-06, "loss": 0.8086, "step": 1629 }, { "epoch": 0.21, "grad_norm": 0.745705783367157, "learning_rate": 9.141809406675794e-06, "loss": 1.097, "step": 1630 }, { "epoch": 0.21, "grad_norm": 1.3642488718032837, "learning_rate": 9.140620233798053e-06, "loss": 0.8094, "step": 1631 }, { "epoch": 0.21, "grad_norm": 0.8063977956771851, "learning_rate": 9.139430315043718e-06, "loss": 1.057, "step": 1632 }, { "epoch": 0.21, "grad_norm": 1.3725179433822632, "learning_rate": 9.13823965062714e-06, "loss": 0.9008, "step": 1633 }, { "epoch": 0.21, "grad_norm": 1.3574985265731812, "learning_rate": 9.137048240762802e-06, "loss": 0.8405, "step": 1634 }, { "epoch": 0.21, "grad_norm": 0.9390093684196472, "learning_rate": 9.135856085665317e-06, "loss": 1.0742, "step": 1635 }, { "epoch": 0.21, "grad_norm": 1.6234182119369507, "learning_rate": 9.13466318554944e-06, "loss": 0.9046, "step": 1636 }, { "epoch": 0.21, "grad_norm": 1.4994686841964722, "learning_rate": 9.133469540630051e-06, "loss": 0.8152, "step": 1637 }, { "epoch": 0.21, "grad_norm": 0.8142198324203491, "learning_rate": 9.132275151122173e-06, "loss": 1.0748, "step": 1638 }, { "epoch": 0.21, "grad_norm": 0.7481842041015625, "learning_rate": 9.131080017240956e-06, "loss": 1.074, "step": 1639 }, { "epoch": 0.21, "grad_norm": 0.6287876963615417, "learning_rate": 9.129884139201688e-06, "loss": 1.0671, "step": 1640 }, { "epoch": 0.22, "grad_norm": 1.2884410619735718, "learning_rate": 9.12868751721979e-06, "loss": 0.859, "step": 1641 }, { "epoch": 0.22, "grad_norm": 1.372111201286316, "learning_rate": 9.127490151510819e-06, "loss": 0.897, "step": 1642 }, { "epoch": 0.22, "grad_norm": 1.2954853773117065, "learning_rate": 9.126292042290464e-06, "loss": 0.8757, "step": 1643 }, { "epoch": 0.22, "grad_norm": 1.326271653175354, "learning_rate": 9.125093189774545e-06, "loss": 0.8936, "step": 1644 }, { "epoch": 0.22, "grad_norm": 1.2454047203063965, "learning_rate": 9.123893594179022e-06, "loss": 0.7999, "step": 1645 }, { "epoch": 0.22, "grad_norm": 1.5273449420928955, "learning_rate": 9.122693255719986e-06, "loss": 0.8839, "step": 1646 }, { "epoch": 0.22, "grad_norm": 1.4241482019424438, "learning_rate": 9.121492174613659e-06, "loss": 0.8747, "step": 1647 }, { "epoch": 0.22, "grad_norm": 1.3649232387542725, "learning_rate": 9.120290351076402e-06, "loss": 0.8809, "step": 1648 }, { "epoch": 0.22, "grad_norm": 1.4139227867126465, "learning_rate": 9.119087785324708e-06, "loss": 0.8914, "step": 1649 }, { "epoch": 0.22, "grad_norm": 1.1393394470214844, "learning_rate": 9.117884477575198e-06, "loss": 1.0456, "step": 1650 }, { "epoch": 0.22, "grad_norm": 1.424619197845459, "learning_rate": 9.116680428044638e-06, "loss": 0.8904, "step": 1651 }, { "epoch": 0.22, "grad_norm": 0.6952387094497681, "learning_rate": 9.115475636949916e-06, "loss": 1.0777, "step": 1652 }, { "epoch": 0.22, "grad_norm": 1.451910138130188, "learning_rate": 9.114270104508063e-06, "loss": 0.9287, "step": 1653 }, { "epoch": 0.22, "grad_norm": 1.488861083984375, "learning_rate": 9.113063830936234e-06, "loss": 0.8052, "step": 1654 }, { "epoch": 0.22, "grad_norm": 4.55080509185791, "learning_rate": 9.11185681645173e-06, "loss": 0.8396, "step": 1655 }, { "epoch": 0.22, "grad_norm": 1.351517915725708, "learning_rate": 9.110649061271971e-06, "loss": 0.7947, "step": 1656 }, { "epoch": 0.22, "grad_norm": 1.4005862474441528, "learning_rate": 9.109440565614522e-06, "loss": 0.854, "step": 1657 }, { "epoch": 0.22, "grad_norm": 1.1725647449493408, "learning_rate": 9.108231329697077e-06, "loss": 1.0958, "step": 1658 }, { "epoch": 0.22, "grad_norm": 1.3552590608596802, "learning_rate": 9.107021353737463e-06, "loss": 0.8713, "step": 1659 }, { "epoch": 0.22, "grad_norm": 1.4483329057693481, "learning_rate": 9.10581063795364e-06, "loss": 0.8302, "step": 1660 }, { "epoch": 0.22, "grad_norm": 0.6898401379585266, "learning_rate": 9.104599182563701e-06, "loss": 1.0911, "step": 1661 }, { "epoch": 0.22, "grad_norm": 1.3341988325119019, "learning_rate": 9.103386987785876e-06, "loss": 0.889, "step": 1662 }, { "epoch": 0.22, "grad_norm": 1.342226266860962, "learning_rate": 9.102174053838523e-06, "loss": 0.8433, "step": 1663 }, { "epoch": 0.22, "grad_norm": 0.9272589087486267, "learning_rate": 9.100960380940138e-06, "loss": 1.0896, "step": 1664 }, { "epoch": 0.22, "grad_norm": 1.5275981426239014, "learning_rate": 9.099745969309346e-06, "loss": 0.9123, "step": 1665 }, { "epoch": 0.22, "grad_norm": 1.3263906240463257, "learning_rate": 9.098530819164907e-06, "loss": 0.8762, "step": 1666 }, { "epoch": 0.22, "grad_norm": 1.3727529048919678, "learning_rate": 9.097314930725717e-06, "loss": 0.8413, "step": 1667 }, { "epoch": 0.22, "grad_norm": 1.3477314710617065, "learning_rate": 9.096098304210794e-06, "loss": 0.9418, "step": 1668 }, { "epoch": 0.22, "grad_norm": 1.411854863166809, "learning_rate": 9.094880939839303e-06, "loss": 0.8578, "step": 1669 }, { "epoch": 0.22, "grad_norm": 1.3559421300888062, "learning_rate": 9.093662837830537e-06, "loss": 0.9229, "step": 1670 }, { "epoch": 0.22, "grad_norm": 1.3525563478469849, "learning_rate": 9.092443998403915e-06, "loss": 0.7979, "step": 1671 }, { "epoch": 0.22, "grad_norm": 0.9244986176490784, "learning_rate": 9.091224421779e-06, "loss": 1.0653, "step": 1672 }, { "epoch": 0.22, "grad_norm": 1.4077785015106201, "learning_rate": 9.090004108175476e-06, "loss": 0.8805, "step": 1673 }, { "epoch": 0.22, "grad_norm": 1.2509574890136719, "learning_rate": 9.088783057813172e-06, "loss": 0.843, "step": 1674 }, { "epoch": 0.22, "grad_norm": 1.6673318147659302, "learning_rate": 9.08756127091204e-06, "loss": 0.855, "step": 1675 }, { "epoch": 0.22, "grad_norm": 1.373148798942566, "learning_rate": 9.08633874769217e-06, "loss": 0.7947, "step": 1676 }, { "epoch": 0.22, "grad_norm": 1.3497064113616943, "learning_rate": 9.085115488373783e-06, "loss": 0.7952, "step": 1677 }, { "epoch": 0.22, "grad_norm": 0.9620547294616699, "learning_rate": 9.08389149317723e-06, "loss": 1.0744, "step": 1678 }, { "epoch": 0.22, "grad_norm": 1.339611291885376, "learning_rate": 9.082666762323e-06, "loss": 0.8371, "step": 1679 }, { "epoch": 0.22, "grad_norm": 1.2670882940292358, "learning_rate": 9.081441296031713e-06, "loss": 0.8297, "step": 1680 }, { "epoch": 0.22, "grad_norm": 1.4200752973556519, "learning_rate": 9.080215094524116e-06, "loss": 0.8422, "step": 1681 }, { "epoch": 0.22, "grad_norm": 1.2393498420715332, "learning_rate": 9.078988158021094e-06, "loss": 0.8141, "step": 1682 }, { "epoch": 0.22, "grad_norm": 1.4653388261795044, "learning_rate": 9.077760486743663e-06, "loss": 0.8775, "step": 1683 }, { "epoch": 0.22, "grad_norm": 1.0229191780090332, "learning_rate": 9.076532080912974e-06, "loss": 1.088, "step": 1684 }, { "epoch": 0.22, "grad_norm": 1.4045383930206299, "learning_rate": 9.075302940750304e-06, "loss": 0.8644, "step": 1685 }, { "epoch": 0.22, "grad_norm": 1.2444570064544678, "learning_rate": 9.074073066477068e-06, "loss": 0.8138, "step": 1686 }, { "epoch": 0.22, "grad_norm": 1.3546322584152222, "learning_rate": 9.07284245831481e-06, "loss": 0.8514, "step": 1687 }, { "epoch": 0.22, "grad_norm": 1.486791968345642, "learning_rate": 9.071611116485209e-06, "loss": 0.8338, "step": 1688 }, { "epoch": 0.22, "grad_norm": 1.4057968854904175, "learning_rate": 9.070379041210072e-06, "loss": 0.8397, "step": 1689 }, { "epoch": 0.22, "grad_norm": 1.3046547174453735, "learning_rate": 9.069146232711343e-06, "loss": 0.8155, "step": 1690 }, { "epoch": 0.22, "grad_norm": 1.663824200630188, "learning_rate": 9.067912691211095e-06, "loss": 0.8444, "step": 1691 }, { "epoch": 0.22, "grad_norm": 1.4769788980484009, "learning_rate": 9.066678416931533e-06, "loss": 0.892, "step": 1692 }, { "epoch": 0.22, "grad_norm": 1.3306864500045776, "learning_rate": 9.065443410094997e-06, "loss": 0.8448, "step": 1693 }, { "epoch": 0.22, "grad_norm": 1.257284164428711, "learning_rate": 9.064207670923953e-06, "loss": 0.8025, "step": 1694 }, { "epoch": 0.22, "grad_norm": 1.394649624824524, "learning_rate": 9.062971199641007e-06, "loss": 0.8295, "step": 1695 }, { "epoch": 0.22, "grad_norm": 1.386700987815857, "learning_rate": 9.061733996468889e-06, "loss": 0.8599, "step": 1696 }, { "epoch": 0.22, "grad_norm": 1.3818525075912476, "learning_rate": 9.060496061630466e-06, "loss": 0.8516, "step": 1697 }, { "epoch": 0.22, "grad_norm": 1.3758352994918823, "learning_rate": 9.059257395348736e-06, "loss": 1.0717, "step": 1698 }, { "epoch": 0.22, "grad_norm": 1.435915231704712, "learning_rate": 9.058017997846825e-06, "loss": 0.8338, "step": 1699 }, { "epoch": 0.22, "grad_norm": 1.568253993988037, "learning_rate": 9.056777869347995e-06, "loss": 0.863, "step": 1700 }, { "epoch": 0.22, "grad_norm": 1.3014200925827026, "learning_rate": 9.05553701007564e-06, "loss": 0.8089, "step": 1701 }, { "epoch": 0.22, "grad_norm": 1.3845064640045166, "learning_rate": 9.054295420253283e-06, "loss": 0.8385, "step": 1702 }, { "epoch": 0.22, "grad_norm": 2.0574421882629395, "learning_rate": 9.05305310010458e-06, "loss": 0.9114, "step": 1703 }, { "epoch": 0.22, "grad_norm": 0.9495604634284973, "learning_rate": 9.051810049853316e-06, "loss": 1.068, "step": 1704 }, { "epoch": 0.22, "grad_norm": 0.7932641506195068, "learning_rate": 9.050566269723412e-06, "loss": 1.0799, "step": 1705 }, { "epoch": 0.22, "grad_norm": 1.411979079246521, "learning_rate": 9.049321759938918e-06, "loss": 0.8855, "step": 1706 }, { "epoch": 0.22, "grad_norm": 1.3891080617904663, "learning_rate": 9.048076520724014e-06, "loss": 0.8385, "step": 1707 }, { "epoch": 0.22, "grad_norm": 1.4067484140396118, "learning_rate": 9.046830552303012e-06, "loss": 0.8377, "step": 1708 }, { "epoch": 0.22, "grad_norm": 1.4414639472961426, "learning_rate": 9.04558385490036e-06, "loss": 0.8566, "step": 1709 }, { "epoch": 0.22, "grad_norm": 1.4933208227157593, "learning_rate": 9.044336428740632e-06, "loss": 1.0652, "step": 1710 }, { "epoch": 0.22, "grad_norm": 2.5693514347076416, "learning_rate": 9.043088274048535e-06, "loss": 0.9059, "step": 1711 }, { "epoch": 0.22, "grad_norm": 1.4897791147232056, "learning_rate": 9.041839391048905e-06, "loss": 0.8717, "step": 1712 }, { "epoch": 0.22, "grad_norm": 1.4964066743850708, "learning_rate": 9.040589779966716e-06, "loss": 0.8726, "step": 1713 }, { "epoch": 0.22, "grad_norm": 1.4312390089035034, "learning_rate": 9.039339441027064e-06, "loss": 0.8674, "step": 1714 }, { "epoch": 0.22, "grad_norm": 1.3464858531951904, "learning_rate": 9.038088374455181e-06, "loss": 0.8742, "step": 1715 }, { "epoch": 0.22, "grad_norm": 1.3045086860656738, "learning_rate": 9.036836580476432e-06, "loss": 0.8177, "step": 1716 }, { "epoch": 0.22, "grad_norm": 1.3890578746795654, "learning_rate": 9.03558405931631e-06, "loss": 0.8565, "step": 1717 }, { "epoch": 0.23, "grad_norm": 1.4172301292419434, "learning_rate": 9.034330811200438e-06, "loss": 0.886, "step": 1718 }, { "epoch": 0.23, "grad_norm": 1.4381407499313354, "learning_rate": 9.033076836354575e-06, "loss": 0.8815, "step": 1719 }, { "epoch": 0.23, "grad_norm": 1.3186039924621582, "learning_rate": 9.031822135004604e-06, "loss": 0.8623, "step": 1720 }, { "epoch": 0.23, "grad_norm": 1.3996469974517822, "learning_rate": 9.030566707376545e-06, "loss": 0.8804, "step": 1721 }, { "epoch": 0.23, "grad_norm": 0.8340597152709961, "learning_rate": 9.029310553696545e-06, "loss": 1.0759, "step": 1722 }, { "epoch": 0.23, "grad_norm": 1.3442422151565552, "learning_rate": 9.028053674190883e-06, "loss": 0.8665, "step": 1723 }, { "epoch": 0.23, "grad_norm": 1.3600280284881592, "learning_rate": 9.02679606908597e-06, "loss": 0.8089, "step": 1724 }, { "epoch": 0.23, "grad_norm": 1.3509466648101807, "learning_rate": 9.025537738608344e-06, "loss": 0.8534, "step": 1725 }, { "epoch": 0.23, "grad_norm": 1.8929088115692139, "learning_rate": 9.024278682984679e-06, "loss": 0.8503, "step": 1726 }, { "epoch": 0.23, "grad_norm": 1.433030605316162, "learning_rate": 9.023018902441774e-06, "loss": 0.8632, "step": 1727 }, { "epoch": 0.23, "grad_norm": 1.3138136863708496, "learning_rate": 9.021758397206565e-06, "loss": 0.902, "step": 1728 }, { "epoch": 0.23, "grad_norm": 1.4001853466033936, "learning_rate": 9.020497167506113e-06, "loss": 0.8741, "step": 1729 }, { "epoch": 0.23, "grad_norm": 1.288833737373352, "learning_rate": 9.01923521356761e-06, "loss": 0.8522, "step": 1730 }, { "epoch": 0.23, "grad_norm": 1.3289155960083008, "learning_rate": 9.017972535618381e-06, "loss": 0.8769, "step": 1731 }, { "epoch": 0.23, "grad_norm": 1.337571620941162, "learning_rate": 9.01670913388588e-06, "loss": 0.8442, "step": 1732 }, { "epoch": 0.23, "grad_norm": 1.9636164903640747, "learning_rate": 9.015445008597694e-06, "loss": 0.8539, "step": 1733 }, { "epoch": 0.23, "grad_norm": 1.3017714023590088, "learning_rate": 9.014180159981536e-06, "loss": 0.853, "step": 1734 }, { "epoch": 0.23, "grad_norm": 1.4562413692474365, "learning_rate": 9.01291458826525e-06, "loss": 0.9201, "step": 1735 }, { "epoch": 0.23, "grad_norm": 1.3196742534637451, "learning_rate": 9.011648293676816e-06, "loss": 0.814, "step": 1736 }, { "epoch": 0.23, "grad_norm": 1.3768370151519775, "learning_rate": 9.010381276444336e-06, "loss": 0.8805, "step": 1737 }, { "epoch": 0.23, "grad_norm": 1.384937047958374, "learning_rate": 9.009113536796046e-06, "loss": 0.8884, "step": 1738 }, { "epoch": 0.23, "grad_norm": 1.3528721332550049, "learning_rate": 9.007845074960314e-06, "loss": 0.8495, "step": 1739 }, { "epoch": 0.23, "grad_norm": 1.133043885231018, "learning_rate": 9.006575891165633e-06, "loss": 1.0797, "step": 1740 }, { "epoch": 0.23, "grad_norm": 0.8193289041519165, "learning_rate": 9.005305985640633e-06, "loss": 1.0798, "step": 1741 }, { "epoch": 0.23, "grad_norm": 0.6784392595291138, "learning_rate": 9.00403535861407e-06, "loss": 1.0592, "step": 1742 }, { "epoch": 0.23, "grad_norm": 1.555119514465332, "learning_rate": 9.002764010314828e-06, "loss": 0.8844, "step": 1743 }, { "epoch": 0.23, "grad_norm": 1.157651662826538, "learning_rate": 9.001491940971922e-06, "loss": 1.0718, "step": 1744 }, { "epoch": 0.23, "grad_norm": 2.106734037399292, "learning_rate": 9.000219150814504e-06, "loss": 0.8308, "step": 1745 }, { "epoch": 0.23, "grad_norm": 2.262172222137451, "learning_rate": 8.998945640071844e-06, "loss": 0.88, "step": 1746 }, { "epoch": 0.23, "grad_norm": 1.3913600444793701, "learning_rate": 8.99767140897335e-06, "loss": 0.8343, "step": 1747 }, { "epoch": 0.23, "grad_norm": 0.9041345715522766, "learning_rate": 8.996396457748556e-06, "loss": 1.0865, "step": 1748 }, { "epoch": 0.23, "grad_norm": 1.4514657258987427, "learning_rate": 8.995120786627129e-06, "loss": 0.8758, "step": 1749 }, { "epoch": 0.23, "grad_norm": 0.799793541431427, "learning_rate": 8.993844395838865e-06, "loss": 1.0522, "step": 1750 }, { "epoch": 0.23, "grad_norm": 1.3990520238876343, "learning_rate": 8.992567285613686e-06, "loss": 0.8225, "step": 1751 }, { "epoch": 0.23, "grad_norm": 1.4110051393508911, "learning_rate": 8.991289456181647e-06, "loss": 0.7704, "step": 1752 }, { "epoch": 0.23, "grad_norm": 0.7728286981582642, "learning_rate": 8.99001090777293e-06, "loss": 1.0535, "step": 1753 }, { "epoch": 0.23, "grad_norm": 1.471675992012024, "learning_rate": 8.98873164061785e-06, "loss": 0.8302, "step": 1754 }, { "epoch": 0.23, "grad_norm": 1.3585525751113892, "learning_rate": 8.98745165494685e-06, "loss": 0.8469, "step": 1755 }, { "epoch": 0.23, "grad_norm": 1.413468837738037, "learning_rate": 8.986170950990502e-06, "loss": 0.8303, "step": 1756 }, { "epoch": 0.23, "grad_norm": 1.3952977657318115, "learning_rate": 8.984889528979505e-06, "loss": 0.8568, "step": 1757 }, { "epoch": 0.23, "grad_norm": 1.416661024093628, "learning_rate": 8.983607389144692e-06, "loss": 0.8016, "step": 1758 }, { "epoch": 0.23, "grad_norm": 1.3889414072036743, "learning_rate": 8.982324531717024e-06, "loss": 0.9198, "step": 1759 }, { "epoch": 0.23, "grad_norm": 1.3263397216796875, "learning_rate": 8.981040956927588e-06, "loss": 0.8104, "step": 1760 }, { "epoch": 0.23, "grad_norm": 1.383980631828308, "learning_rate": 8.979756665007604e-06, "loss": 0.8765, "step": 1761 }, { "epoch": 0.23, "grad_norm": 1.4375208616256714, "learning_rate": 8.97847165618842e-06, "loss": 0.8541, "step": 1762 }, { "epoch": 0.23, "grad_norm": 1.37940514087677, "learning_rate": 8.977185930701513e-06, "loss": 0.8576, "step": 1763 }, { "epoch": 0.23, "grad_norm": 1.2787059545516968, "learning_rate": 8.975899488778487e-06, "loss": 0.8505, "step": 1764 }, { "epoch": 0.23, "grad_norm": 1.4484447240829468, "learning_rate": 8.974612330651078e-06, "loss": 0.8526, "step": 1765 }, { "epoch": 0.23, "grad_norm": 1.285022258758545, "learning_rate": 8.973324456551152e-06, "loss": 0.8911, "step": 1766 }, { "epoch": 0.23, "grad_norm": 1.401607632637024, "learning_rate": 8.972035866710699e-06, "loss": 0.8689, "step": 1767 }, { "epoch": 0.23, "grad_norm": 1.32899010181427, "learning_rate": 8.970746561361842e-06, "loss": 0.9119, "step": 1768 }, { "epoch": 0.23, "grad_norm": 1.3023254871368408, "learning_rate": 8.969456540736835e-06, "loss": 0.82, "step": 1769 }, { "epoch": 0.23, "grad_norm": 1.3774874210357666, "learning_rate": 8.968165805068054e-06, "loss": 0.9117, "step": 1770 }, { "epoch": 0.23, "grad_norm": 1.0129311084747314, "learning_rate": 8.966874354588008e-06, "loss": 1.076, "step": 1771 }, { "epoch": 0.23, "grad_norm": 1.3180665969848633, "learning_rate": 8.965582189529336e-06, "loss": 0.8716, "step": 1772 }, { "epoch": 0.23, "grad_norm": 1.3135488033294678, "learning_rate": 8.964289310124803e-06, "loss": 0.8764, "step": 1773 }, { "epoch": 0.23, "grad_norm": 1.4084582328796387, "learning_rate": 8.962995716607304e-06, "loss": 0.8793, "step": 1774 }, { "epoch": 0.23, "grad_norm": 1.5004761219024658, "learning_rate": 8.96170140920986e-06, "loss": 0.9001, "step": 1775 }, { "epoch": 0.23, "grad_norm": 1.2926511764526367, "learning_rate": 8.960406388165626e-06, "loss": 0.882, "step": 1776 }, { "epoch": 0.23, "grad_norm": 1.2417036294937134, "learning_rate": 8.959110653707883e-06, "loss": 0.8451, "step": 1777 }, { "epoch": 0.23, "grad_norm": 1.4351919889450073, "learning_rate": 8.957814206070037e-06, "loss": 0.8628, "step": 1778 }, { "epoch": 0.23, "grad_norm": 1.1053410768508911, "learning_rate": 8.956517045485626e-06, "loss": 1.0764, "step": 1779 }, { "epoch": 0.23, "grad_norm": 1.3560283184051514, "learning_rate": 8.955219172188319e-06, "loss": 0.9284, "step": 1780 }, { "epoch": 0.23, "grad_norm": 1.4007644653320312, "learning_rate": 8.953920586411906e-06, "loss": 0.8354, "step": 1781 }, { "epoch": 0.23, "grad_norm": 0.7367246150970459, "learning_rate": 8.952621288390313e-06, "loss": 1.068, "step": 1782 }, { "epoch": 0.23, "grad_norm": 0.7594488859176636, "learning_rate": 8.951321278357589e-06, "loss": 1.0653, "step": 1783 }, { "epoch": 0.23, "grad_norm": 0.6915542483329773, "learning_rate": 8.950020556547913e-06, "loss": 1.0511, "step": 1784 }, { "epoch": 0.23, "grad_norm": 2.0256969928741455, "learning_rate": 8.948719123195596e-06, "loss": 0.798, "step": 1785 }, { "epoch": 0.23, "grad_norm": 1.3629860877990723, "learning_rate": 8.947416978535069e-06, "loss": 0.8307, "step": 1786 }, { "epoch": 0.23, "grad_norm": 1.4111140966415405, "learning_rate": 8.946114122800896e-06, "loss": 0.8607, "step": 1787 }, { "epoch": 0.23, "grad_norm": 0.9956296682357788, "learning_rate": 8.944810556227771e-06, "loss": 1.062, "step": 1788 }, { "epoch": 0.23, "grad_norm": 1.310139536857605, "learning_rate": 8.943506279050515e-06, "loss": 0.8391, "step": 1789 }, { "epoch": 0.23, "grad_norm": 1.3316633701324463, "learning_rate": 8.942201291504071e-06, "loss": 0.8965, "step": 1790 }, { "epoch": 0.23, "grad_norm": 0.7765719294548035, "learning_rate": 8.940895593823519e-06, "loss": 1.0662, "step": 1791 }, { "epoch": 0.23, "grad_norm": 1.3152003288269043, "learning_rate": 8.939589186244063e-06, "loss": 0.8088, "step": 1792 }, { "epoch": 0.23, "grad_norm": 0.7157859206199646, "learning_rate": 8.938282069001031e-06, "loss": 1.0882, "step": 1793 }, { "epoch": 0.24, "grad_norm": 0.6919587254524231, "learning_rate": 8.936974242329887e-06, "loss": 1.0633, "step": 1794 }, { "epoch": 0.24, "grad_norm": 1.3148657083511353, "learning_rate": 8.935665706466214e-06, "loss": 0.807, "step": 1795 }, { "epoch": 0.24, "grad_norm": 0.6805819272994995, "learning_rate": 8.93435646164573e-06, "loss": 1.0552, "step": 1796 }, { "epoch": 0.24, "grad_norm": 1.4271432161331177, "learning_rate": 8.933046508104275e-06, "loss": 0.9025, "step": 1797 }, { "epoch": 0.24, "grad_norm": 1.2702491283416748, "learning_rate": 8.93173584607782e-06, "loss": 0.8529, "step": 1798 }, { "epoch": 0.24, "grad_norm": 1.3106601238250732, "learning_rate": 8.930424475802466e-06, "loss": 0.8509, "step": 1799 }, { "epoch": 0.24, "grad_norm": 1.2336549758911133, "learning_rate": 8.929112397514436e-06, "loss": 0.8774, "step": 1800 }, { "epoch": 0.24, "grad_norm": 1.2906410694122314, "learning_rate": 8.927799611450083e-06, "loss": 0.8141, "step": 1801 }, { "epoch": 0.24, "grad_norm": 0.8819374442100525, "learning_rate": 8.926486117845886e-06, "loss": 1.0803, "step": 1802 }, { "epoch": 0.24, "grad_norm": 1.4250092506408691, "learning_rate": 8.925171916938459e-06, "loss": 0.8576, "step": 1803 }, { "epoch": 0.24, "grad_norm": 1.4200950860977173, "learning_rate": 8.923857008964534e-06, "loss": 0.7781, "step": 1804 }, { "epoch": 0.24, "grad_norm": 1.337685465812683, "learning_rate": 8.92254139416097e-06, "loss": 0.8306, "step": 1805 }, { "epoch": 0.24, "grad_norm": 1.3017419576644897, "learning_rate": 8.921225072764764e-06, "loss": 0.8552, "step": 1806 }, { "epoch": 0.24, "grad_norm": 1.3570330142974854, "learning_rate": 8.919908045013028e-06, "loss": 0.8121, "step": 1807 }, { "epoch": 0.24, "grad_norm": 1.26701021194458, "learning_rate": 8.91859031114301e-06, "loss": 0.8422, "step": 1808 }, { "epoch": 0.24, "grad_norm": 1.8744664192199707, "learning_rate": 8.917271871392079e-06, "loss": 0.8649, "step": 1809 }, { "epoch": 0.24, "grad_norm": 0.8543540835380554, "learning_rate": 8.915952725997736e-06, "loss": 1.0757, "step": 1810 }, { "epoch": 0.24, "grad_norm": 1.4309229850769043, "learning_rate": 8.914632875197608e-06, "loss": 0.8326, "step": 1811 }, { "epoch": 0.24, "grad_norm": 0.6401118636131287, "learning_rate": 8.913312319229448e-06, "loss": 1.0765, "step": 1812 }, { "epoch": 0.24, "grad_norm": 0.6902464032173157, "learning_rate": 8.911991058331133e-06, "loss": 1.0787, "step": 1813 }, { "epoch": 0.24, "grad_norm": 1.3781843185424805, "learning_rate": 8.910669092740675e-06, "loss": 0.8362, "step": 1814 }, { "epoch": 0.24, "grad_norm": 1.3079488277435303, "learning_rate": 8.909346422696204e-06, "loss": 0.8366, "step": 1815 }, { "epoch": 0.24, "grad_norm": 1.362828016281128, "learning_rate": 8.908023048435983e-06, "loss": 0.9178, "step": 1816 }, { "epoch": 0.24, "grad_norm": 1.335433840751648, "learning_rate": 8.906698970198402e-06, "loss": 0.8453, "step": 1817 }, { "epoch": 0.24, "grad_norm": 0.9516278505325317, "learning_rate": 8.905374188221971e-06, "loss": 1.0778, "step": 1818 }, { "epoch": 0.24, "grad_norm": 0.7621197700500488, "learning_rate": 8.904048702745336e-06, "loss": 1.0681, "step": 1819 }, { "epoch": 0.24, "grad_norm": 0.6694728136062622, "learning_rate": 8.902722514007262e-06, "loss": 1.0665, "step": 1820 }, { "epoch": 0.24, "grad_norm": 1.3069301843643188, "learning_rate": 8.901395622246645e-06, "loss": 0.8653, "step": 1821 }, { "epoch": 0.24, "grad_norm": 0.8950925469398499, "learning_rate": 8.900068027702508e-06, "loss": 1.0574, "step": 1822 }, { "epoch": 0.24, "grad_norm": 1.2797092199325562, "learning_rate": 8.898739730613996e-06, "loss": 0.8949, "step": 1823 }, { "epoch": 0.24, "grad_norm": 1.3859115839004517, "learning_rate": 8.897410731220388e-06, "loss": 0.8434, "step": 1824 }, { "epoch": 0.24, "grad_norm": 1.3121079206466675, "learning_rate": 8.896081029761081e-06, "loss": 0.8136, "step": 1825 }, { "epoch": 0.24, "grad_norm": 1.293921947479248, "learning_rate": 8.894750626475605e-06, "loss": 0.8364, "step": 1826 }, { "epoch": 0.24, "grad_norm": 1.2566838264465332, "learning_rate": 8.893419521603615e-06, "loss": 0.8166, "step": 1827 }, { "epoch": 0.24, "grad_norm": 1.5384953022003174, "learning_rate": 8.892087715384888e-06, "loss": 0.8262, "step": 1828 }, { "epoch": 0.24, "grad_norm": 1.3254806995391846, "learning_rate": 8.890755208059334e-06, "loss": 0.8883, "step": 1829 }, { "epoch": 0.24, "grad_norm": 1.3893588781356812, "learning_rate": 8.889421999866985e-06, "loss": 0.8523, "step": 1830 }, { "epoch": 0.24, "grad_norm": 1.4557651281356812, "learning_rate": 8.888088091048001e-06, "loss": 0.9116, "step": 1831 }, { "epoch": 0.24, "grad_norm": 1.3399736881256104, "learning_rate": 8.886753481842667e-06, "loss": 0.8505, "step": 1832 }, { "epoch": 0.24, "grad_norm": 1.0357614755630493, "learning_rate": 8.885418172491396e-06, "loss": 1.091, "step": 1833 }, { "epoch": 0.24, "grad_norm": 1.4086112976074219, "learning_rate": 8.884082163234722e-06, "loss": 0.8291, "step": 1834 }, { "epoch": 0.24, "grad_norm": 1.4774903059005737, "learning_rate": 8.882745454313313e-06, "loss": 0.8678, "step": 1835 }, { "epoch": 0.24, "grad_norm": 0.6981735825538635, "learning_rate": 8.88140804596796e-06, "loss": 1.0623, "step": 1836 }, { "epoch": 0.24, "grad_norm": 1.5059090852737427, "learning_rate": 8.880069938439573e-06, "loss": 0.8628, "step": 1837 }, { "epoch": 0.24, "grad_norm": 1.3500643968582153, "learning_rate": 8.878731131969199e-06, "loss": 0.8589, "step": 1838 }, { "epoch": 0.24, "grad_norm": 1.3967907428741455, "learning_rate": 8.877391626798004e-06, "loss": 0.8879, "step": 1839 }, { "epoch": 0.24, "grad_norm": 1.4178146123886108, "learning_rate": 8.876051423167281e-06, "loss": 0.8941, "step": 1840 }, { "epoch": 0.24, "grad_norm": 0.8623592853546143, "learning_rate": 8.874710521318452e-06, "loss": 1.0867, "step": 1841 }, { "epoch": 0.24, "grad_norm": 1.459934115409851, "learning_rate": 8.87336892149306e-06, "loss": 0.83, "step": 1842 }, { "epoch": 0.24, "grad_norm": 1.282278299331665, "learning_rate": 8.872026623932779e-06, "loss": 0.8076, "step": 1843 }, { "epoch": 0.24, "grad_norm": 1.3266260623931885, "learning_rate": 8.870683628879401e-06, "loss": 0.7971, "step": 1844 }, { "epoch": 0.24, "grad_norm": 0.7775670886039734, "learning_rate": 8.869339936574852e-06, "loss": 1.0768, "step": 1845 }, { "epoch": 0.24, "grad_norm": 1.3063470125198364, "learning_rate": 8.867995547261179e-06, "loss": 0.8725, "step": 1846 }, { "epoch": 0.24, "grad_norm": 1.3244198560714722, "learning_rate": 8.866650461180557e-06, "loss": 0.8853, "step": 1847 }, { "epoch": 0.24, "grad_norm": 0.7214176654815674, "learning_rate": 8.865304678575281e-06, "loss": 1.0479, "step": 1848 }, { "epoch": 0.24, "grad_norm": 1.326221227645874, "learning_rate": 8.863958199687779e-06, "loss": 0.8933, "step": 1849 }, { "epoch": 0.24, "grad_norm": 1.484511375427246, "learning_rate": 8.862611024760598e-06, "loss": 0.8604, "step": 1850 }, { "epoch": 0.24, "grad_norm": 1.442515254020691, "learning_rate": 8.861263154036417e-06, "loss": 0.8946, "step": 1851 }, { "epoch": 0.24, "grad_norm": 1.3853068351745605, "learning_rate": 8.859914587758035e-06, "loss": 0.8459, "step": 1852 }, { "epoch": 0.24, "grad_norm": 0.7454867362976074, "learning_rate": 8.858565326168377e-06, "loss": 1.0836, "step": 1853 }, { "epoch": 0.24, "grad_norm": 0.6899653077125549, "learning_rate": 8.857215369510496e-06, "loss": 1.0712, "step": 1854 }, { "epoch": 0.24, "grad_norm": 1.4565489292144775, "learning_rate": 8.855864718027565e-06, "loss": 0.944, "step": 1855 }, { "epoch": 0.24, "grad_norm": 1.34996497631073, "learning_rate": 8.85451337196289e-06, "loss": 0.8933, "step": 1856 }, { "epoch": 0.24, "grad_norm": 1.2652060985565186, "learning_rate": 8.853161331559894e-06, "loss": 0.8628, "step": 1857 }, { "epoch": 0.24, "grad_norm": 1.4151182174682617, "learning_rate": 8.851808597062131e-06, "loss": 0.8824, "step": 1858 }, { "epoch": 0.24, "grad_norm": 1.2311006784439087, "learning_rate": 8.850455168713277e-06, "loss": 0.8392, "step": 1859 }, { "epoch": 0.24, "grad_norm": 1.41880202293396, "learning_rate": 8.849101046757134e-06, "loss": 0.8522, "step": 1860 }, { "epoch": 0.24, "grad_norm": 1.3640975952148438, "learning_rate": 8.847746231437629e-06, "loss": 0.9077, "step": 1861 }, { "epoch": 0.24, "grad_norm": 1.4848065376281738, "learning_rate": 8.84639072299881e-06, "loss": 0.8946, "step": 1862 }, { "epoch": 0.24, "grad_norm": 1.1968719959259033, "learning_rate": 8.845034521684859e-06, "loss": 1.0745, "step": 1863 }, { "epoch": 0.24, "grad_norm": 1.4318548440933228, "learning_rate": 8.843677627740075e-06, "loss": 0.8863, "step": 1864 }, { "epoch": 0.24, "grad_norm": 1.3733289241790771, "learning_rate": 8.842320041408882e-06, "loss": 0.8161, "step": 1865 }, { "epoch": 0.24, "grad_norm": 0.6759650111198425, "learning_rate": 8.840961762935836e-06, "loss": 1.0712, "step": 1866 }, { "epoch": 0.24, "grad_norm": 1.5036410093307495, "learning_rate": 8.839602792565606e-06, "loss": 0.8034, "step": 1867 }, { "epoch": 0.24, "grad_norm": 1.458937168121338, "learning_rate": 8.838243130542995e-06, "loss": 0.8515, "step": 1868 }, { "epoch": 0.24, "grad_norm": 1.383972406387329, "learning_rate": 8.83688277711293e-06, "loss": 0.8066, "step": 1869 }, { "epoch": 0.25, "grad_norm": 0.9902024269104004, "learning_rate": 8.835521732520459e-06, "loss": 1.0944, "step": 1870 }, { "epoch": 0.25, "grad_norm": 0.7496927976608276, "learning_rate": 8.834159997010751e-06, "loss": 1.0389, "step": 1871 }, { "epoch": 0.25, "grad_norm": 1.4867339134216309, "learning_rate": 8.83279757082911e-06, "loss": 0.8585, "step": 1872 }, { "epoch": 0.25, "grad_norm": 1.3416062593460083, "learning_rate": 8.831434454220957e-06, "loss": 0.8681, "step": 1873 }, { "epoch": 0.25, "grad_norm": 1.7162106037139893, "learning_rate": 8.830070647431839e-06, "loss": 0.8558, "step": 1874 }, { "epoch": 0.25, "grad_norm": 1.303488850593567, "learning_rate": 8.828706150707424e-06, "loss": 0.7956, "step": 1875 }, { "epoch": 0.25, "grad_norm": 1.4261656999588013, "learning_rate": 8.827340964293514e-06, "loss": 1.0696, "step": 1876 }, { "epoch": 0.25, "grad_norm": 1.3487672805786133, "learning_rate": 8.825975088436023e-06, "loss": 0.8546, "step": 1877 }, { "epoch": 0.25, "grad_norm": 1.4807608127593994, "learning_rate": 8.824608523380998e-06, "loss": 0.9348, "step": 1878 }, { "epoch": 0.25, "grad_norm": 1.3934472799301147, "learning_rate": 8.823241269374606e-06, "loss": 0.8092, "step": 1879 }, { "epoch": 0.25, "grad_norm": 1.4481914043426514, "learning_rate": 8.82187332666314e-06, "loss": 0.8667, "step": 1880 }, { "epoch": 0.25, "grad_norm": 0.7389448881149292, "learning_rate": 8.820504695493018e-06, "loss": 1.0675, "step": 1881 }, { "epoch": 0.25, "grad_norm": 1.4814488887786865, "learning_rate": 8.819135376110776e-06, "loss": 0.852, "step": 1882 }, { "epoch": 0.25, "grad_norm": 1.3754545450210571, "learning_rate": 8.817765368763081e-06, "loss": 0.8871, "step": 1883 }, { "epoch": 0.25, "grad_norm": 0.6755757927894592, "learning_rate": 8.816394673696722e-06, "loss": 1.0821, "step": 1884 }, { "epoch": 0.25, "grad_norm": 1.329973578453064, "learning_rate": 8.815023291158608e-06, "loss": 0.8666, "step": 1885 }, { "epoch": 0.25, "grad_norm": 1.4273056983947754, "learning_rate": 8.813651221395779e-06, "loss": 0.8199, "step": 1886 }, { "epoch": 0.25, "grad_norm": 0.7661797404289246, "learning_rate": 8.812278464655394e-06, "loss": 1.0715, "step": 1887 }, { "epoch": 0.25, "grad_norm": 1.2908546924591064, "learning_rate": 8.810905021184732e-06, "loss": 0.8314, "step": 1888 }, { "epoch": 0.25, "grad_norm": 1.3341591358184814, "learning_rate": 8.809530891231206e-06, "loss": 0.9049, "step": 1889 }, { "epoch": 0.25, "grad_norm": 1.2940512895584106, "learning_rate": 8.808156075042343e-06, "loss": 0.881, "step": 1890 }, { "epoch": 0.25, "grad_norm": 1.399583339691162, "learning_rate": 8.806780572865801e-06, "loss": 0.847, "step": 1891 }, { "epoch": 0.25, "grad_norm": 1.3356722593307495, "learning_rate": 8.805404384949354e-06, "loss": 0.805, "step": 1892 }, { "epoch": 0.25, "grad_norm": 0.9751436710357666, "learning_rate": 8.804027511540906e-06, "loss": 1.0528, "step": 1893 }, { "epoch": 0.25, "grad_norm": 1.3217172622680664, "learning_rate": 8.802649952888483e-06, "loss": 0.8373, "step": 1894 }, { "epoch": 0.25, "grad_norm": 1.3506028652191162, "learning_rate": 8.80127170924023e-06, "loss": 0.8894, "step": 1895 }, { "epoch": 0.25, "grad_norm": 0.7691227197647095, "learning_rate": 8.799892780844424e-06, "loss": 1.0667, "step": 1896 }, { "epoch": 0.25, "grad_norm": 0.7706899046897888, "learning_rate": 8.798513167949453e-06, "loss": 1.0622, "step": 1897 }, { "epoch": 0.25, "grad_norm": 1.400685429573059, "learning_rate": 8.797132870803844e-06, "loss": 0.8605, "step": 1898 }, { "epoch": 0.25, "grad_norm": 1.2662901878356934, "learning_rate": 8.795751889656232e-06, "loss": 0.8108, "step": 1899 }, { "epoch": 0.25, "grad_norm": 1.4304981231689453, "learning_rate": 8.794370224755386e-06, "loss": 0.8587, "step": 1900 }, { "epoch": 0.25, "grad_norm": 1.2928080558776855, "learning_rate": 8.792987876350194e-06, "loss": 0.8641, "step": 1901 }, { "epoch": 0.25, "grad_norm": 1.3297828435897827, "learning_rate": 8.791604844689664e-06, "loss": 0.8138, "step": 1902 }, { "epoch": 0.25, "grad_norm": 1.3151434659957886, "learning_rate": 8.790221130022934e-06, "loss": 0.8358, "step": 1903 }, { "epoch": 0.25, "grad_norm": 1.343634009361267, "learning_rate": 8.78883673259926e-06, "loss": 0.8595, "step": 1904 }, { "epoch": 0.25, "grad_norm": 1.817884922027588, "learning_rate": 8.787451652668023e-06, "loss": 0.8766, "step": 1905 }, { "epoch": 0.25, "grad_norm": 1.434847354888916, "learning_rate": 8.786065890478725e-06, "loss": 1.0935, "step": 1906 }, { "epoch": 0.25, "grad_norm": 1.3718390464782715, "learning_rate": 8.784679446280994e-06, "loss": 0.8339, "step": 1907 }, { "epoch": 0.25, "grad_norm": 1.3952771425247192, "learning_rate": 8.783292320324578e-06, "loss": 0.8689, "step": 1908 }, { "epoch": 0.25, "grad_norm": 1.362883448600769, "learning_rate": 8.781904512859348e-06, "loss": 0.7863, "step": 1909 }, { "epoch": 0.25, "grad_norm": 0.7913500666618347, "learning_rate": 8.780516024135301e-06, "loss": 1.0704, "step": 1910 }, { "epoch": 0.25, "grad_norm": 1.28773832321167, "learning_rate": 8.779126854402554e-06, "loss": 0.7741, "step": 1911 }, { "epoch": 0.25, "grad_norm": 1.324953317642212, "learning_rate": 8.777737003911346e-06, "loss": 0.8408, "step": 1912 }, { "epoch": 0.25, "grad_norm": 0.7817246913909912, "learning_rate": 8.776346472912039e-06, "loss": 1.063, "step": 1913 }, { "epoch": 0.25, "grad_norm": 1.3268537521362305, "learning_rate": 8.77495526165512e-06, "loss": 0.8122, "step": 1914 }, { "epoch": 0.25, "grad_norm": 0.6371838450431824, "learning_rate": 8.773563370391198e-06, "loss": 1.0906, "step": 1915 }, { "epoch": 0.25, "grad_norm": 0.7121062874794006, "learning_rate": 8.772170799371e-06, "loss": 1.0881, "step": 1916 }, { "epoch": 0.25, "grad_norm": 1.6411666870117188, "learning_rate": 8.77077754884538e-06, "loss": 0.8435, "step": 1917 }, { "epoch": 0.25, "grad_norm": 1.4103935956954956, "learning_rate": 8.769383619065315e-06, "loss": 0.8375, "step": 1918 }, { "epoch": 0.25, "grad_norm": 1.3667819499969482, "learning_rate": 8.767989010281899e-06, "loss": 0.8062, "step": 1919 }, { "epoch": 0.25, "grad_norm": 1.271324872970581, "learning_rate": 8.766593722746355e-06, "loss": 0.8142, "step": 1920 }, { "epoch": 0.25, "grad_norm": 2.0019378662109375, "learning_rate": 8.765197756710025e-06, "loss": 0.8004, "step": 1921 }, { "epoch": 0.25, "grad_norm": 1.0294692516326904, "learning_rate": 8.763801112424372e-06, "loss": 1.0621, "step": 1922 }, { "epoch": 0.25, "grad_norm": 1.7858046293258667, "learning_rate": 8.762403790140982e-06, "loss": 0.806, "step": 1923 }, { "epoch": 0.25, "grad_norm": 1.2916327714920044, "learning_rate": 8.761005790111567e-06, "loss": 0.8519, "step": 1924 }, { "epoch": 0.25, "grad_norm": 1.3280445337295532, "learning_rate": 8.759607112587956e-06, "loss": 0.8502, "step": 1925 }, { "epoch": 0.25, "grad_norm": 1.3798775672912598, "learning_rate": 8.7582077578221e-06, "loss": 0.7402, "step": 1926 }, { "epoch": 0.25, "grad_norm": 0.6891633868217468, "learning_rate": 8.756807726066073e-06, "loss": 1.0531, "step": 1927 }, { "epoch": 0.25, "grad_norm": 0.7236558198928833, "learning_rate": 8.755407017572075e-06, "loss": 1.0641, "step": 1928 }, { "epoch": 0.25, "grad_norm": 0.6407458186149597, "learning_rate": 8.754005632592425e-06, "loss": 1.0629, "step": 1929 }, { "epoch": 0.25, "grad_norm": 1.4049153327941895, "learning_rate": 8.752603571379559e-06, "loss": 0.81, "step": 1930 }, { "epoch": 0.25, "grad_norm": 1.3429954051971436, "learning_rate": 8.751200834186044e-06, "loss": 0.8129, "step": 1931 }, { "epoch": 0.25, "grad_norm": 1.3749715089797974, "learning_rate": 8.749797421264563e-06, "loss": 0.868, "step": 1932 }, { "epoch": 0.25, "grad_norm": 1.3110910654067993, "learning_rate": 8.748393332867919e-06, "loss": 0.8541, "step": 1933 }, { "epoch": 0.25, "grad_norm": 0.9394099712371826, "learning_rate": 8.746988569249044e-06, "loss": 1.0442, "step": 1934 }, { "epoch": 0.25, "grad_norm": 1.3344838619232178, "learning_rate": 8.745583130660983e-06, "loss": 0.8326, "step": 1935 }, { "epoch": 0.25, "grad_norm": 1.3943880796432495, "learning_rate": 8.74417701735691e-06, "loss": 0.8519, "step": 1936 }, { "epoch": 0.25, "grad_norm": 1.3427174091339111, "learning_rate": 8.742770229590113e-06, "loss": 0.8364, "step": 1937 }, { "epoch": 0.25, "grad_norm": 1.3663225173950195, "learning_rate": 8.741362767614012e-06, "loss": 0.8498, "step": 1938 }, { "epoch": 0.25, "grad_norm": 0.7354006767272949, "learning_rate": 8.739954631682136e-06, "loss": 1.074, "step": 1939 }, { "epoch": 0.25, "grad_norm": 1.704529047012329, "learning_rate": 8.738545822048144e-06, "loss": 0.8649, "step": 1940 }, { "epoch": 0.25, "grad_norm": 0.658925473690033, "learning_rate": 8.737136338965816e-06, "loss": 1.0778, "step": 1941 }, { "epoch": 0.25, "grad_norm": 1.4096746444702148, "learning_rate": 8.735726182689048e-06, "loss": 0.7937, "step": 1942 }, { "epoch": 0.25, "grad_norm": 0.6370680332183838, "learning_rate": 8.734315353471863e-06, "loss": 1.0504, "step": 1943 }, { "epoch": 0.25, "grad_norm": 1.355211615562439, "learning_rate": 8.732903851568401e-06, "loss": 0.8603, "step": 1944 }, { "epoch": 0.25, "grad_norm": 1.3467758893966675, "learning_rate": 8.731491677232926e-06, "loss": 0.8196, "step": 1945 }, { "epoch": 0.25, "grad_norm": 0.7179825901985168, "learning_rate": 8.73007883071982e-06, "loss": 1.0578, "step": 1946 }, { "epoch": 0.26, "grad_norm": 1.463333249092102, "learning_rate": 8.728665312283593e-06, "loss": 0.8866, "step": 1947 }, { "epoch": 0.26, "grad_norm": 0.6143915057182312, "learning_rate": 8.727251122178866e-06, "loss": 1.0647, "step": 1948 }, { "epoch": 0.26, "grad_norm": 1.3445799350738525, "learning_rate": 8.725836260660387e-06, "loss": 0.9242, "step": 1949 }, { "epoch": 0.26, "grad_norm": 1.3200148344039917, "learning_rate": 8.724420727983028e-06, "loss": 0.9012, "step": 1950 }, { "epoch": 0.26, "grad_norm": 1.4495422840118408, "learning_rate": 8.723004524401774e-06, "loss": 0.867, "step": 1951 }, { "epoch": 0.26, "grad_norm": 1.2997546195983887, "learning_rate": 8.721587650171737e-06, "loss": 0.8421, "step": 1952 }, { "epoch": 0.26, "grad_norm": 1.4150729179382324, "learning_rate": 8.720170105548145e-06, "loss": 0.8746, "step": 1953 }, { "epoch": 0.26, "grad_norm": 0.8063339591026306, "learning_rate": 8.718751890786354e-06, "loss": 1.0459, "step": 1954 }, { "epoch": 0.26, "grad_norm": 0.6677082180976868, "learning_rate": 8.717333006141831e-06, "loss": 1.0427, "step": 1955 }, { "epoch": 0.26, "grad_norm": 0.631256103515625, "learning_rate": 8.715913451870173e-06, "loss": 1.069, "step": 1956 }, { "epoch": 0.26, "grad_norm": 1.4785112142562866, "learning_rate": 8.71449322822709e-06, "loss": 0.8984, "step": 1957 }, { "epoch": 0.26, "grad_norm": 0.7585464715957642, "learning_rate": 8.71307233546842e-06, "loss": 1.0575, "step": 1958 }, { "epoch": 0.26, "grad_norm": 0.7330500483512878, "learning_rate": 8.711650773850114e-06, "loss": 1.0673, "step": 1959 }, { "epoch": 0.26, "grad_norm": 1.2173906564712524, "learning_rate": 8.710228543628248e-06, "loss": 0.773, "step": 1960 }, { "epoch": 0.26, "grad_norm": 1.3719147443771362, "learning_rate": 8.708805645059019e-06, "loss": 0.8891, "step": 1961 }, { "epoch": 0.26, "grad_norm": 0.733723521232605, "learning_rate": 8.70738207839874e-06, "loss": 1.0589, "step": 1962 }, { "epoch": 0.26, "grad_norm": 1.2917729616165161, "learning_rate": 8.705957843903852e-06, "loss": 0.8845, "step": 1963 }, { "epoch": 0.26, "grad_norm": 0.7072991728782654, "learning_rate": 8.704532941830905e-06, "loss": 1.0571, "step": 1964 }, { "epoch": 0.26, "grad_norm": 1.2768408060073853, "learning_rate": 8.703107372436583e-06, "loss": 0.8411, "step": 1965 }, { "epoch": 0.26, "grad_norm": 1.3167437314987183, "learning_rate": 8.701681135977678e-06, "loss": 0.8394, "step": 1966 }, { "epoch": 0.26, "grad_norm": 0.7386677265167236, "learning_rate": 8.700254232711107e-06, "loss": 1.0798, "step": 1967 }, { "epoch": 0.26, "grad_norm": 1.2386975288391113, "learning_rate": 8.69882666289391e-06, "loss": 0.8598, "step": 1968 }, { "epoch": 0.26, "grad_norm": 0.6555726528167725, "learning_rate": 8.697398426783243e-06, "loss": 1.0368, "step": 1969 }, { "epoch": 0.26, "grad_norm": 0.616493284702301, "learning_rate": 8.695969524636383e-06, "loss": 1.0832, "step": 1970 }, { "epoch": 0.26, "grad_norm": 1.2765209674835205, "learning_rate": 8.694539956710728e-06, "loss": 0.7883, "step": 1971 }, { "epoch": 0.26, "grad_norm": 1.3102673292160034, "learning_rate": 8.693109723263794e-06, "loss": 0.8092, "step": 1972 }, { "epoch": 0.26, "grad_norm": 1.4754810333251953, "learning_rate": 8.691678824553223e-06, "loss": 0.872, "step": 1973 }, { "epoch": 0.26, "grad_norm": 0.7046512961387634, "learning_rate": 8.690247260836765e-06, "loss": 1.0574, "step": 1974 }, { "epoch": 0.26, "grad_norm": 1.319474697113037, "learning_rate": 8.688815032372302e-06, "loss": 0.8929, "step": 1975 }, { "epoch": 0.26, "grad_norm": 0.7119693756103516, "learning_rate": 8.687382139417827e-06, "loss": 1.099, "step": 1976 }, { "epoch": 0.26, "grad_norm": 1.5243884325027466, "learning_rate": 8.685948582231458e-06, "loss": 0.8326, "step": 1977 }, { "epoch": 0.26, "grad_norm": 1.3662028312683105, "learning_rate": 8.68451436107143e-06, "loss": 0.9059, "step": 1978 }, { "epoch": 0.26, "grad_norm": 0.7417484521865845, "learning_rate": 8.683079476196101e-06, "loss": 1.0699, "step": 1979 }, { "epoch": 0.26, "grad_norm": 1.38494873046875, "learning_rate": 8.681643927863944e-06, "loss": 0.8288, "step": 1980 }, { "epoch": 0.26, "grad_norm": 1.3764697313308716, "learning_rate": 8.680207716333552e-06, "loss": 0.8314, "step": 1981 }, { "epoch": 0.26, "grad_norm": 0.7261018753051758, "learning_rate": 8.678770841863644e-06, "loss": 1.0766, "step": 1982 }, { "epoch": 0.26, "grad_norm": 1.4471360445022583, "learning_rate": 8.677333304713048e-06, "loss": 0.8358, "step": 1983 }, { "epoch": 0.26, "grad_norm": 0.6594396829605103, "learning_rate": 8.67589510514072e-06, "loss": 1.0522, "step": 1984 }, { "epoch": 0.26, "grad_norm": 1.366003394126892, "learning_rate": 8.674456243405732e-06, "loss": 0.8184, "step": 1985 }, { "epoch": 0.26, "grad_norm": 0.6405418515205383, "learning_rate": 8.673016719767274e-06, "loss": 1.0745, "step": 1986 }, { "epoch": 0.26, "grad_norm": 1.214858889579773, "learning_rate": 8.671576534484655e-06, "loss": 0.8168, "step": 1987 }, { "epoch": 0.26, "grad_norm": 1.2560150623321533, "learning_rate": 8.670135687817312e-06, "loss": 0.7952, "step": 1988 }, { "epoch": 0.26, "grad_norm": 1.3469221591949463, "learning_rate": 8.668694180024785e-06, "loss": 0.8687, "step": 1989 }, { "epoch": 0.26, "grad_norm": 0.7842597365379333, "learning_rate": 8.66725201136675e-06, "loss": 1.0495, "step": 1990 }, { "epoch": 0.26, "grad_norm": 1.3332122564315796, "learning_rate": 8.665809182102987e-06, "loss": 0.8826, "step": 1991 }, { "epoch": 0.26, "grad_norm": 1.2870562076568604, "learning_rate": 8.664365692493408e-06, "loss": 0.8392, "step": 1992 }, { "epoch": 0.26, "grad_norm": 1.5315020084381104, "learning_rate": 8.662921542798032e-06, "loss": 0.8199, "step": 1993 }, { "epoch": 0.26, "grad_norm": 1.3886914253234863, "learning_rate": 8.661476733277011e-06, "loss": 0.8023, "step": 1994 }, { "epoch": 0.26, "grad_norm": 1.3647223711013794, "learning_rate": 8.660031264190603e-06, "loss": 0.8718, "step": 1995 }, { "epoch": 0.26, "grad_norm": 1.3864790201187134, "learning_rate": 8.658585135799188e-06, "loss": 0.8618, "step": 1996 }, { "epoch": 0.26, "grad_norm": 1.35519278049469, "learning_rate": 8.65713834836327e-06, "loss": 0.8436, "step": 1997 }, { "epoch": 0.26, "grad_norm": 1.368944764137268, "learning_rate": 8.655690902143466e-06, "loss": 0.8567, "step": 1998 }, { "epoch": 0.26, "grad_norm": 1.5342183113098145, "learning_rate": 8.654242797400516e-06, "loss": 0.8692, "step": 1999 }, { "epoch": 0.26, "grad_norm": 1.5314011573791504, "learning_rate": 8.652794034395274e-06, "loss": 0.8556, "step": 2000 }, { "epoch": 0.26, "grad_norm": 1.3479560613632202, "learning_rate": 8.651344613388714e-06, "loss": 0.8385, "step": 2001 }, { "epoch": 0.26, "grad_norm": 1.3432526588439941, "learning_rate": 8.649894534641933e-06, "loss": 0.7854, "step": 2002 }, { "epoch": 0.26, "grad_norm": 0.8297284245491028, "learning_rate": 8.648443798416142e-06, "loss": 1.0605, "step": 2003 }, { "epoch": 0.26, "grad_norm": 0.7156286239624023, "learning_rate": 8.646992404972671e-06, "loss": 1.0827, "step": 2004 }, { "epoch": 0.26, "grad_norm": 1.3795721530914307, "learning_rate": 8.645540354572969e-06, "loss": 0.8707, "step": 2005 }, { "epoch": 0.26, "grad_norm": 1.423746109008789, "learning_rate": 8.644087647478602e-06, "loss": 0.9001, "step": 2006 }, { "epoch": 0.26, "grad_norm": 0.9395588040351868, "learning_rate": 8.642634283951257e-06, "loss": 1.0583, "step": 2007 }, { "epoch": 0.26, "grad_norm": 1.33549165725708, "learning_rate": 8.641180264252738e-06, "loss": 0.8795, "step": 2008 }, { "epoch": 0.26, "grad_norm": 1.406492829322815, "learning_rate": 8.639725588644965e-06, "loss": 0.8697, "step": 2009 }, { "epoch": 0.26, "grad_norm": 2.134129524230957, "learning_rate": 8.63827025738998e-06, "loss": 0.8241, "step": 2010 }, { "epoch": 0.26, "grad_norm": 1.3849151134490967, "learning_rate": 8.63681427074994e-06, "loss": 0.8962, "step": 2011 }, { "epoch": 0.26, "grad_norm": 1.409122109413147, "learning_rate": 8.635357628987122e-06, "loss": 0.8404, "step": 2012 }, { "epoch": 0.26, "grad_norm": 1.4498237371444702, "learning_rate": 8.63390033236392e-06, "loss": 0.8407, "step": 2013 }, { "epoch": 0.26, "grad_norm": 1.2208402156829834, "learning_rate": 8.632442381142846e-06, "loss": 0.768, "step": 2014 }, { "epoch": 0.26, "grad_norm": 1.4424431324005127, "learning_rate": 8.630983775586532e-06, "loss": 0.8233, "step": 2015 }, { "epoch": 0.26, "grad_norm": 0.8585935235023499, "learning_rate": 8.629524515957724e-06, "loss": 1.0275, "step": 2016 }, { "epoch": 0.26, "grad_norm": 1.3133069276809692, "learning_rate": 8.628064602519287e-06, "loss": 0.8254, "step": 2017 }, { "epoch": 0.26, "grad_norm": 1.4858465194702148, "learning_rate": 8.626604035534205e-06, "loss": 0.8552, "step": 2018 }, { "epoch": 0.26, "grad_norm": 1.2947896718978882, "learning_rate": 8.625142815265581e-06, "loss": 0.7976, "step": 2019 }, { "epoch": 0.26, "grad_norm": 1.468495488166809, "learning_rate": 8.623680941976634e-06, "loss": 0.8541, "step": 2020 }, { "epoch": 0.26, "grad_norm": 1.3705620765686035, "learning_rate": 8.622218415930698e-06, "loss": 0.8703, "step": 2021 }, { "epoch": 0.26, "grad_norm": 1.4735009670257568, "learning_rate": 8.62075523739123e-06, "loss": 0.8964, "step": 2022 }, { "epoch": 0.27, "grad_norm": 0.8667446970939636, "learning_rate": 8.619291406621801e-06, "loss": 1.0561, "step": 2023 }, { "epoch": 0.27, "grad_norm": 1.4080414772033691, "learning_rate": 8.617826923886102e-06, "loss": 0.8564, "step": 2024 }, { "epoch": 0.27, "grad_norm": 1.4493528604507446, "learning_rate": 8.616361789447935e-06, "loss": 0.9149, "step": 2025 }, { "epoch": 0.27, "grad_norm": 1.3023327589035034, "learning_rate": 8.614896003571227e-06, "loss": 0.8117, "step": 2026 }, { "epoch": 0.27, "grad_norm": 1.3060028553009033, "learning_rate": 8.61342956652002e-06, "loss": 0.8693, "step": 2027 }, { "epoch": 0.27, "grad_norm": 1.3496217727661133, "learning_rate": 8.611962478558473e-06, "loss": 0.9087, "step": 2028 }, { "epoch": 0.27, "grad_norm": 0.9623264074325562, "learning_rate": 8.61049473995086e-06, "loss": 1.0741, "step": 2029 }, { "epoch": 0.27, "grad_norm": 0.7353186011314392, "learning_rate": 8.609026350961576e-06, "loss": 1.0621, "step": 2030 }, { "epoch": 0.27, "grad_norm": 1.4427518844604492, "learning_rate": 8.60755731185513e-06, "loss": 0.8177, "step": 2031 }, { "epoch": 0.27, "grad_norm": 1.5418291091918945, "learning_rate": 8.60608762289615e-06, "loss": 0.8553, "step": 2032 }, { "epoch": 0.27, "grad_norm": 1.858298420906067, "learning_rate": 8.604617284349381e-06, "loss": 0.8564, "step": 2033 }, { "epoch": 0.27, "grad_norm": 1.4488192796707153, "learning_rate": 8.603146296479683e-06, "loss": 0.8573, "step": 2034 }, { "epoch": 0.27, "grad_norm": 1.4380223751068115, "learning_rate": 8.601674659552038e-06, "loss": 0.8429, "step": 2035 }, { "epoch": 0.27, "grad_norm": 1.3580344915390015, "learning_rate": 8.600202373831538e-06, "loss": 0.8352, "step": 2036 }, { "epoch": 0.27, "grad_norm": 1.632461428642273, "learning_rate": 8.598729439583398e-06, "loss": 0.8599, "step": 2037 }, { "epoch": 0.27, "grad_norm": 1.4178675413131714, "learning_rate": 8.597255857072944e-06, "loss": 0.8417, "step": 2038 }, { "epoch": 0.27, "grad_norm": 1.3782179355621338, "learning_rate": 8.595781626565624e-06, "loss": 0.822, "step": 2039 }, { "epoch": 0.27, "grad_norm": 1.4437721967697144, "learning_rate": 8.594306748327e-06, "loss": 0.8342, "step": 2040 }, { "epoch": 0.27, "grad_norm": 1.3284863233566284, "learning_rate": 8.592831222622752e-06, "loss": 0.8755, "step": 2041 }, { "epoch": 0.27, "grad_norm": 2.0764710903167725, "learning_rate": 8.591355049718675e-06, "loss": 1.0652, "step": 2042 }, { "epoch": 0.27, "grad_norm": 1.7139768600463867, "learning_rate": 8.589878229880682e-06, "loss": 1.0673, "step": 2043 }, { "epoch": 0.27, "grad_norm": 1.4210734367370605, "learning_rate": 8.588400763374802e-06, "loss": 0.8231, "step": 2044 }, { "epoch": 0.27, "grad_norm": 0.7233688831329346, "learning_rate": 8.586922650467182e-06, "loss": 1.0307, "step": 2045 }, { "epoch": 0.27, "grad_norm": 0.8553022742271423, "learning_rate": 8.58544389142408e-06, "loss": 1.0667, "step": 2046 }, { "epoch": 0.27, "grad_norm": 1.6232562065124512, "learning_rate": 8.58396448651188e-06, "loss": 0.8552, "step": 2047 }, { "epoch": 0.27, "grad_norm": 1.453912615776062, "learning_rate": 8.582484435997071e-06, "loss": 0.85, "step": 2048 }, { "epoch": 0.27, "grad_norm": 1.4262022972106934, "learning_rate": 8.581003740146269e-06, "loss": 0.8448, "step": 2049 }, { "epoch": 0.27, "grad_norm": 1.344477653503418, "learning_rate": 8.579522399226196e-06, "loss": 0.8897, "step": 2050 }, { "epoch": 0.27, "grad_norm": 1.4276745319366455, "learning_rate": 8.578040413503702e-06, "loss": 0.852, "step": 2051 }, { "epoch": 0.27, "grad_norm": 1.438550353050232, "learning_rate": 8.576557783245743e-06, "loss": 0.8476, "step": 2052 }, { "epoch": 0.27, "grad_norm": 2.4150166511535645, "learning_rate": 8.575074508719392e-06, "loss": 0.8538, "step": 2053 }, { "epoch": 0.27, "grad_norm": 1.434360384941101, "learning_rate": 8.573590590191844e-06, "loss": 0.8338, "step": 2054 }, { "epoch": 0.27, "grad_norm": 1.3972821235656738, "learning_rate": 8.572106027930409e-06, "loss": 0.8619, "step": 2055 }, { "epoch": 0.27, "grad_norm": 1.8920388221740723, "learning_rate": 8.570620822202507e-06, "loss": 1.0825, "step": 2056 }, { "epoch": 0.27, "grad_norm": 1.5255622863769531, "learning_rate": 8.569134973275678e-06, "loss": 0.8629, "step": 2057 }, { "epoch": 0.27, "grad_norm": 1.5614593029022217, "learning_rate": 8.567648481417581e-06, "loss": 0.8719, "step": 2058 }, { "epoch": 0.27, "grad_norm": 0.966897189617157, "learning_rate": 8.566161346895983e-06, "loss": 1.0546, "step": 2059 }, { "epoch": 0.27, "grad_norm": 1.4450279474258423, "learning_rate": 8.564673569978774e-06, "loss": 0.8792, "step": 2060 }, { "epoch": 0.27, "grad_norm": 1.4523415565490723, "learning_rate": 8.563185150933955e-06, "loss": 0.8006, "step": 2061 }, { "epoch": 0.27, "grad_norm": 1.083490252494812, "learning_rate": 8.561696090029645e-06, "loss": 1.059, "step": 2062 }, { "epoch": 0.27, "grad_norm": 1.0497690439224243, "learning_rate": 8.560206387534078e-06, "loss": 1.0395, "step": 2063 }, { "epoch": 0.27, "grad_norm": 0.8158340454101562, "learning_rate": 8.558716043715606e-06, "loss": 1.0619, "step": 2064 }, { "epoch": 0.27, "grad_norm": 1.4991624355316162, "learning_rate": 8.557225058842691e-06, "loss": 0.865, "step": 2065 }, { "epoch": 0.27, "grad_norm": 1.4536428451538086, "learning_rate": 8.555733433183915e-06, "loss": 0.817, "step": 2066 }, { "epoch": 0.27, "grad_norm": 1.640952706336975, "learning_rate": 8.554241167007976e-06, "loss": 0.8841, "step": 2067 }, { "epoch": 0.27, "grad_norm": 1.3122966289520264, "learning_rate": 8.55274826058368e-06, "loss": 1.0508, "step": 2068 }, { "epoch": 0.27, "grad_norm": 1.3404159545898438, "learning_rate": 8.551254714179961e-06, "loss": 0.8344, "step": 2069 }, { "epoch": 0.27, "grad_norm": 0.8526004552841187, "learning_rate": 8.549760528065855e-06, "loss": 1.0362, "step": 2070 }, { "epoch": 0.27, "grad_norm": 0.6790236830711365, "learning_rate": 8.548265702510524e-06, "loss": 1.06, "step": 2071 }, { "epoch": 0.27, "grad_norm": 0.8032899498939514, "learning_rate": 8.54677023778324e-06, "loss": 1.0795, "step": 2072 }, { "epoch": 0.27, "grad_norm": 1.6170380115509033, "learning_rate": 8.545274134153386e-06, "loss": 0.855, "step": 2073 }, { "epoch": 0.27, "grad_norm": 0.9709977507591248, "learning_rate": 8.54377739189047e-06, "loss": 1.0639, "step": 2074 }, { "epoch": 0.27, "grad_norm": 0.7723125219345093, "learning_rate": 8.542280011264108e-06, "loss": 1.0578, "step": 2075 }, { "epoch": 0.27, "grad_norm": 0.6967212557792664, "learning_rate": 8.540781992544032e-06, "loss": 1.0705, "step": 2076 }, { "epoch": 0.27, "grad_norm": 1.4744096994400024, "learning_rate": 8.539283336000091e-06, "loss": 0.8553, "step": 2077 }, { "epoch": 0.27, "grad_norm": 0.9403405785560608, "learning_rate": 8.537784041902247e-06, "loss": 1.0499, "step": 2078 }, { "epoch": 0.27, "grad_norm": 1.3928614854812622, "learning_rate": 8.536284110520577e-06, "loss": 0.8978, "step": 2079 }, { "epoch": 0.27, "grad_norm": 1.34027099609375, "learning_rate": 8.534783542125276e-06, "loss": 0.8278, "step": 2080 }, { "epoch": 0.27, "grad_norm": 1.3252215385437012, "learning_rate": 8.533282336986646e-06, "loss": 0.841, "step": 2081 }, { "epoch": 0.27, "grad_norm": 1.4211759567260742, "learning_rate": 8.531780495375112e-06, "loss": 0.806, "step": 2082 }, { "epoch": 0.27, "grad_norm": 1.411467432975769, "learning_rate": 8.530278017561214e-06, "loss": 0.8097, "step": 2083 }, { "epoch": 0.27, "grad_norm": 1.3622697591781616, "learning_rate": 8.528774903815596e-06, "loss": 0.7978, "step": 2084 }, { "epoch": 0.27, "grad_norm": 1.2796224355697632, "learning_rate": 8.527271154409027e-06, "loss": 0.8453, "step": 2085 }, { "epoch": 0.27, "grad_norm": 1.0858200788497925, "learning_rate": 8.525766769612387e-06, "loss": 1.0588, "step": 2086 }, { "epoch": 0.27, "grad_norm": 1.9377305507659912, "learning_rate": 8.52426174969667e-06, "loss": 0.875, "step": 2087 }, { "epoch": 0.27, "grad_norm": 1.4002759456634521, "learning_rate": 8.522756094932987e-06, "loss": 0.7964, "step": 2088 }, { "epoch": 0.27, "grad_norm": 0.7780790328979492, "learning_rate": 8.521249805592558e-06, "loss": 1.0797, "step": 2089 }, { "epoch": 0.27, "grad_norm": 1.477901816368103, "learning_rate": 8.519742881946723e-06, "loss": 0.8201, "step": 2090 }, { "epoch": 0.27, "grad_norm": 1.318721055984497, "learning_rate": 8.518235324266931e-06, "loss": 0.7735, "step": 2091 }, { "epoch": 0.27, "grad_norm": 1.361155390739441, "learning_rate": 8.51672713282475e-06, "loss": 0.8098, "step": 2092 }, { "epoch": 0.27, "grad_norm": 1.332892656326294, "learning_rate": 8.515218307891864e-06, "loss": 0.8338, "step": 2093 }, { "epoch": 0.27, "grad_norm": 1.4213495254516602, "learning_rate": 8.513708849740061e-06, "loss": 0.8658, "step": 2094 }, { "epoch": 0.27, "grad_norm": 1.386252760887146, "learning_rate": 8.512198758641252e-06, "loss": 0.7828, "step": 2095 }, { "epoch": 0.27, "grad_norm": 1.2942347526550293, "learning_rate": 8.510688034867461e-06, "loss": 0.815, "step": 2096 }, { "epoch": 0.27, "grad_norm": 0.9874261617660522, "learning_rate": 8.509176678690823e-06, "loss": 1.0541, "step": 2097 }, { "epoch": 0.27, "grad_norm": 1.3494826555252075, "learning_rate": 8.507664690383585e-06, "loss": 0.846, "step": 2098 }, { "epoch": 0.28, "grad_norm": 0.6890168190002441, "learning_rate": 8.506152070218119e-06, "loss": 1.0684, "step": 2099 }, { "epoch": 0.28, "grad_norm": 1.4464980363845825, "learning_rate": 8.504638818466895e-06, "loss": 0.7999, "step": 2100 }, { "epoch": 0.28, "grad_norm": 1.4026312828063965, "learning_rate": 8.50312493540251e-06, "loss": 0.8214, "step": 2101 }, { "epoch": 0.28, "grad_norm": 1.4190572500228882, "learning_rate": 8.501610421297666e-06, "loss": 0.7878, "step": 2102 }, { "epoch": 0.28, "grad_norm": 1.3254218101501465, "learning_rate": 8.500095276425187e-06, "loss": 0.814, "step": 2103 }, { "epoch": 0.28, "grad_norm": 1.2522624731063843, "learning_rate": 8.498579501058e-06, "loss": 0.7922, "step": 2104 }, { "epoch": 0.28, "grad_norm": 1.280304193496704, "learning_rate": 8.497063095469155e-06, "loss": 0.8559, "step": 2105 }, { "epoch": 0.28, "grad_norm": 1.4736685752868652, "learning_rate": 8.49554605993181e-06, "loss": 0.8438, "step": 2106 }, { "epoch": 0.28, "grad_norm": 1.4051668643951416, "learning_rate": 8.494028394719241e-06, "loss": 0.8915, "step": 2107 }, { "epoch": 0.28, "grad_norm": 1.3262768983840942, "learning_rate": 8.492510100104833e-06, "loss": 0.8059, "step": 2108 }, { "epoch": 0.28, "grad_norm": 1.2059708833694458, "learning_rate": 8.490991176362086e-06, "loss": 1.0548, "step": 2109 }, { "epoch": 0.28, "grad_norm": 1.3432248830795288, "learning_rate": 8.489471623764613e-06, "loss": 0.8436, "step": 2110 }, { "epoch": 0.28, "grad_norm": 1.3582614660263062, "learning_rate": 8.487951442586141e-06, "loss": 0.8622, "step": 2111 }, { "epoch": 0.28, "grad_norm": 1.480358362197876, "learning_rate": 8.486430633100513e-06, "loss": 0.8485, "step": 2112 }, { "epoch": 0.28, "grad_norm": 0.7447559833526611, "learning_rate": 8.484909195581682e-06, "loss": 1.0415, "step": 2113 }, { "epoch": 0.28, "grad_norm": 1.3987295627593994, "learning_rate": 8.483387130303708e-06, "loss": 0.7831, "step": 2114 }, { "epoch": 0.28, "grad_norm": 1.3728179931640625, "learning_rate": 8.481864437540777e-06, "loss": 0.9246, "step": 2115 }, { "epoch": 0.28, "grad_norm": 0.7584182620048523, "learning_rate": 8.480341117567177e-06, "loss": 1.0686, "step": 2116 }, { "epoch": 0.28, "grad_norm": 1.3672726154327393, "learning_rate": 8.478817170657319e-06, "loss": 0.8864, "step": 2117 }, { "epoch": 0.28, "grad_norm": 1.44571852684021, "learning_rate": 8.477292597085714e-06, "loss": 0.8362, "step": 2118 }, { "epoch": 0.28, "grad_norm": 1.383514404296875, "learning_rate": 8.475767397126999e-06, "loss": 0.8744, "step": 2119 }, { "epoch": 0.28, "grad_norm": 1.2967406511306763, "learning_rate": 8.474241571055916e-06, "loss": 0.8858, "step": 2120 }, { "epoch": 0.28, "grad_norm": 1.3434549570083618, "learning_rate": 8.472715119147324e-06, "loss": 0.8512, "step": 2121 }, { "epoch": 0.28, "grad_norm": 1.309775710105896, "learning_rate": 8.47118804167619e-06, "loss": 0.8809, "step": 2122 }, { "epoch": 0.28, "grad_norm": 0.9290735125541687, "learning_rate": 8.469660338917596e-06, "loss": 1.0678, "step": 2123 }, { "epoch": 0.28, "grad_norm": 1.4108799695968628, "learning_rate": 8.468132011146739e-06, "loss": 0.9003, "step": 2124 }, { "epoch": 0.28, "grad_norm": 1.2199444770812988, "learning_rate": 8.466603058638925e-06, "loss": 0.9034, "step": 2125 }, { "epoch": 0.28, "grad_norm": 1.4046916961669922, "learning_rate": 8.465073481669574e-06, "loss": 0.8768, "step": 2126 }, { "epoch": 0.28, "grad_norm": 1.38951575756073, "learning_rate": 8.46354328051422e-06, "loss": 0.8646, "step": 2127 }, { "epoch": 0.28, "grad_norm": 1.2389930486679077, "learning_rate": 8.462012455448507e-06, "loss": 0.89, "step": 2128 }, { "epoch": 0.28, "grad_norm": 1.2976946830749512, "learning_rate": 8.460481006748194e-06, "loss": 0.8677, "step": 2129 }, { "epoch": 0.28, "grad_norm": 1.3627876043319702, "learning_rate": 8.458948934689145e-06, "loss": 0.8304, "step": 2130 }, { "epoch": 0.28, "grad_norm": 1.3075031042099, "learning_rate": 8.457416239547352e-06, "loss": 0.8406, "step": 2131 }, { "epoch": 0.28, "grad_norm": 1.2747400999069214, "learning_rate": 8.4558829215989e-06, "loss": 0.8786, "step": 2132 }, { "epoch": 0.28, "grad_norm": 0.9742590188980103, "learning_rate": 8.45434898112e-06, "loss": 1.0755, "step": 2133 }, { "epoch": 0.28, "grad_norm": 1.4171844720840454, "learning_rate": 8.45281441838697e-06, "loss": 0.8848, "step": 2134 }, { "epoch": 0.28, "grad_norm": 1.543287754058838, "learning_rate": 8.45127923367624e-06, "loss": 0.8327, "step": 2135 }, { "epoch": 0.28, "grad_norm": 1.3402224779129028, "learning_rate": 8.449743427264351e-06, "loss": 0.8387, "step": 2136 }, { "epoch": 0.28, "grad_norm": 1.2989925146102905, "learning_rate": 8.448206999427962e-06, "loss": 0.8032, "step": 2137 }, { "epoch": 0.28, "grad_norm": 1.3505730628967285, "learning_rate": 8.446669950443836e-06, "loss": 0.7962, "step": 2138 }, { "epoch": 0.28, "grad_norm": 1.2642104625701904, "learning_rate": 8.445132280588853e-06, "loss": 0.8078, "step": 2139 }, { "epoch": 0.28, "grad_norm": 1.1471048593521118, "learning_rate": 8.443593990140001e-06, "loss": 1.0576, "step": 2140 }, { "epoch": 0.28, "grad_norm": 1.3224451541900635, "learning_rate": 8.442055079374389e-06, "loss": 0.8407, "step": 2141 }, { "epoch": 0.28, "grad_norm": 1.3817311525344849, "learning_rate": 8.440515548569224e-06, "loss": 0.766, "step": 2142 }, { "epoch": 0.28, "grad_norm": 1.422425389289856, "learning_rate": 8.438975398001834e-06, "loss": 0.8356, "step": 2143 }, { "epoch": 0.28, "grad_norm": 1.3900418281555176, "learning_rate": 8.437434627949655e-06, "loss": 0.8323, "step": 2144 }, { "epoch": 0.28, "grad_norm": 0.8761746287345886, "learning_rate": 8.435893238690238e-06, "loss": 1.0754, "step": 2145 }, { "epoch": 0.28, "grad_norm": 1.5290266275405884, "learning_rate": 8.434351230501242e-06, "loss": 0.87, "step": 2146 }, { "epoch": 0.28, "grad_norm": 0.6761625409126282, "learning_rate": 8.432808603660439e-06, "loss": 1.0726, "step": 2147 }, { "epoch": 0.28, "grad_norm": 1.3217034339904785, "learning_rate": 8.431265358445713e-06, "loss": 0.8046, "step": 2148 }, { "epoch": 0.28, "grad_norm": 1.4207561016082764, "learning_rate": 8.429721495135057e-06, "loss": 0.8568, "step": 2149 }, { "epoch": 0.28, "grad_norm": 1.2803086042404175, "learning_rate": 8.42817701400658e-06, "loss": 0.844, "step": 2150 }, { "epoch": 0.28, "grad_norm": 0.8299134373664856, "learning_rate": 8.426631915338497e-06, "loss": 1.0839, "step": 2151 }, { "epoch": 0.28, "grad_norm": 0.7272237539291382, "learning_rate": 8.425086199409135e-06, "loss": 1.0755, "step": 2152 }, { "epoch": 0.28, "grad_norm": 1.5773663520812988, "learning_rate": 8.423539866496938e-06, "loss": 0.8349, "step": 2153 }, { "epoch": 0.28, "grad_norm": 1.5386892557144165, "learning_rate": 8.421992916880454e-06, "loss": 0.8414, "step": 2154 }, { "epoch": 0.28, "grad_norm": 1.2634762525558472, "learning_rate": 8.420445350838346e-06, "loss": 0.7969, "step": 2155 }, { "epoch": 0.28, "grad_norm": 1.3293406963348389, "learning_rate": 8.418897168649388e-06, "loss": 0.8414, "step": 2156 }, { "epoch": 0.28, "grad_norm": 1.4372118711471558, "learning_rate": 8.417348370592462e-06, "loss": 0.8037, "step": 2157 }, { "epoch": 0.28, "grad_norm": 1.3640446662902832, "learning_rate": 8.415798956946564e-06, "loss": 0.8315, "step": 2158 }, { "epoch": 0.28, "grad_norm": 1.386733889579773, "learning_rate": 8.414248927990797e-06, "loss": 0.8363, "step": 2159 }, { "epoch": 0.28, "grad_norm": 1.2860084772109985, "learning_rate": 8.412698284004385e-06, "loss": 1.062, "step": 2160 }, { "epoch": 0.28, "grad_norm": 1.3468931913375854, "learning_rate": 8.411147025266649e-06, "loss": 0.845, "step": 2161 }, { "epoch": 0.28, "grad_norm": 1.3426918983459473, "learning_rate": 8.409595152057028e-06, "loss": 0.835, "step": 2162 }, { "epoch": 0.28, "grad_norm": 0.7761290073394775, "learning_rate": 8.408042664655071e-06, "loss": 1.0594, "step": 2163 }, { "epoch": 0.28, "grad_norm": 1.3625916242599487, "learning_rate": 8.406489563340442e-06, "loss": 0.8275, "step": 2164 }, { "epoch": 0.28, "grad_norm": 1.4580817222595215, "learning_rate": 8.404935848392905e-06, "loss": 0.8367, "step": 2165 }, { "epoch": 0.28, "grad_norm": 0.938602864742279, "learning_rate": 8.403381520092346e-06, "loss": 1.0561, "step": 2166 }, { "epoch": 0.28, "grad_norm": 0.8267118334770203, "learning_rate": 8.401826578718752e-06, "loss": 1.0636, "step": 2167 }, { "epoch": 0.28, "grad_norm": 1.391079068183899, "learning_rate": 8.400271024552225e-06, "loss": 0.8663, "step": 2168 }, { "epoch": 0.28, "grad_norm": 1.3954970836639404, "learning_rate": 8.39871485787298e-06, "loss": 0.8473, "step": 2169 }, { "epoch": 0.28, "grad_norm": 1.397945761680603, "learning_rate": 8.397158078961336e-06, "loss": 0.8463, "step": 2170 }, { "epoch": 0.28, "grad_norm": 1.3637231588363647, "learning_rate": 8.39560068809773e-06, "loss": 0.8398, "step": 2171 }, { "epoch": 0.28, "grad_norm": 1.5613937377929688, "learning_rate": 8.394042685562699e-06, "loss": 1.0661, "step": 2172 }, { "epoch": 0.28, "grad_norm": 1.3499689102172852, "learning_rate": 8.3924840716369e-06, "loss": 0.8626, "step": 2173 }, { "epoch": 0.28, "grad_norm": 1.363057255744934, "learning_rate": 8.390924846601095e-06, "loss": 0.8342, "step": 2174 }, { "epoch": 0.28, "grad_norm": 1.321439504623413, "learning_rate": 8.389365010736157e-06, "loss": 0.8367, "step": 2175 }, { "epoch": 0.29, "grad_norm": 0.809124231338501, "learning_rate": 8.38780456432307e-06, "loss": 1.0521, "step": 2176 }, { "epoch": 0.29, "grad_norm": 0.7260259985923767, "learning_rate": 8.386243507642926e-06, "loss": 1.0368, "step": 2177 }, { "epoch": 0.29, "grad_norm": 1.3361968994140625, "learning_rate": 8.384681840976932e-06, "loss": 0.8291, "step": 2178 }, { "epoch": 0.29, "grad_norm": 1.3307143449783325, "learning_rate": 8.383119564606394e-06, "loss": 0.8202, "step": 2179 }, { "epoch": 0.29, "grad_norm": 1.100263237953186, "learning_rate": 8.381556678812742e-06, "loss": 1.0635, "step": 2180 }, { "epoch": 0.29, "grad_norm": 1.2523822784423828, "learning_rate": 8.379993183877506e-06, "loss": 0.8038, "step": 2181 }, { "epoch": 0.29, "grad_norm": 1.4461878538131714, "learning_rate": 8.378429080082327e-06, "loss": 0.8823, "step": 2182 }, { "epoch": 0.29, "grad_norm": 1.5214444398880005, "learning_rate": 8.376864367708958e-06, "loss": 0.86, "step": 2183 }, { "epoch": 0.29, "grad_norm": 1.6556941270828247, "learning_rate": 8.375299047039263e-06, "loss": 0.8115, "step": 2184 }, { "epoch": 0.29, "grad_norm": 1.34066903591156, "learning_rate": 8.37373311835521e-06, "loss": 0.8424, "step": 2185 }, { "epoch": 0.29, "grad_norm": 1.372091293334961, "learning_rate": 8.372166581938882e-06, "loss": 0.9333, "step": 2186 }, { "epoch": 0.29, "grad_norm": 0.9069502353668213, "learning_rate": 8.370599438072465e-06, "loss": 1.0464, "step": 2187 }, { "epoch": 0.29, "grad_norm": 1.3166158199310303, "learning_rate": 8.369031687038267e-06, "loss": 0.8283, "step": 2188 }, { "epoch": 0.29, "grad_norm": 1.4170655012130737, "learning_rate": 8.367463329118688e-06, "loss": 0.8988, "step": 2189 }, { "epoch": 0.29, "grad_norm": 1.5105592012405396, "learning_rate": 8.365894364596253e-06, "loss": 0.8631, "step": 2190 }, { "epoch": 0.29, "grad_norm": 0.7398719787597656, "learning_rate": 8.364324793753587e-06, "loss": 1.0809, "step": 2191 }, { "epoch": 0.29, "grad_norm": 2.6576220989227295, "learning_rate": 8.362754616873428e-06, "loss": 0.8323, "step": 2192 }, { "epoch": 0.29, "grad_norm": 1.4136877059936523, "learning_rate": 8.36118383423862e-06, "loss": 0.8488, "step": 2193 }, { "epoch": 0.29, "grad_norm": 1.344748854637146, "learning_rate": 8.35961244613212e-06, "loss": 0.8482, "step": 2194 }, { "epoch": 0.29, "grad_norm": 1.3448710441589355, "learning_rate": 8.358040452836991e-06, "loss": 0.8537, "step": 2195 }, { "epoch": 0.29, "grad_norm": 0.6713296175003052, "learning_rate": 8.356467854636408e-06, "loss": 1.0549, "step": 2196 }, { "epoch": 0.29, "grad_norm": 1.291822910308838, "learning_rate": 8.354894651813653e-06, "loss": 0.8686, "step": 2197 }, { "epoch": 0.29, "grad_norm": 0.6700169444084167, "learning_rate": 8.353320844652114e-06, "loss": 1.0316, "step": 2198 }, { "epoch": 0.29, "grad_norm": 1.3525851964950562, "learning_rate": 8.351746433435295e-06, "loss": 0.8665, "step": 2199 }, { "epoch": 0.29, "grad_norm": 1.3144645690917969, "learning_rate": 8.350171418446803e-06, "loss": 0.9072, "step": 2200 }, { "epoch": 0.29, "grad_norm": 1.3424752950668335, "learning_rate": 8.348595799970356e-06, "loss": 0.8638, "step": 2201 }, { "epoch": 0.29, "grad_norm": 0.7323353886604309, "learning_rate": 8.347019578289781e-06, "loss": 1.0593, "step": 2202 }, { "epoch": 0.29, "grad_norm": 1.3689687252044678, "learning_rate": 8.34544275368901e-06, "loss": 0.8584, "step": 2203 }, { "epoch": 0.29, "grad_norm": 0.6408782601356506, "learning_rate": 8.34386532645209e-06, "loss": 1.0659, "step": 2204 }, { "epoch": 0.29, "grad_norm": 0.6436393857002258, "learning_rate": 8.342287296863173e-06, "loss": 1.0703, "step": 2205 }, { "epoch": 0.29, "grad_norm": 1.2768141031265259, "learning_rate": 8.340708665206517e-06, "loss": 0.8187, "step": 2206 }, { "epoch": 0.29, "grad_norm": 1.4711109399795532, "learning_rate": 8.339129431766492e-06, "loss": 0.8507, "step": 2207 }, { "epoch": 0.29, "grad_norm": 0.8155707716941833, "learning_rate": 8.337549596827578e-06, "loss": 1.0512, "step": 2208 }, { "epoch": 0.29, "grad_norm": 1.4509330987930298, "learning_rate": 8.335969160674357e-06, "loss": 0.8476, "step": 2209 }, { "epoch": 0.29, "grad_norm": 1.3331685066223145, "learning_rate": 8.334388123591528e-06, "loss": 0.8795, "step": 2210 }, { "epoch": 0.29, "grad_norm": 1.3731067180633545, "learning_rate": 8.332806485863888e-06, "loss": 0.8531, "step": 2211 }, { "epoch": 0.29, "grad_norm": 1.396159052848816, "learning_rate": 8.331224247776351e-06, "loss": 0.8384, "step": 2212 }, { "epoch": 0.29, "grad_norm": 1.2687067985534668, "learning_rate": 8.329641409613933e-06, "loss": 0.8219, "step": 2213 }, { "epoch": 0.29, "grad_norm": 0.7254360914230347, "learning_rate": 8.328057971661765e-06, "loss": 1.0458, "step": 2214 }, { "epoch": 0.29, "grad_norm": 1.4443700313568115, "learning_rate": 8.326473934205077e-06, "loss": 0.8446, "step": 2215 }, { "epoch": 0.29, "grad_norm": 1.3932666778564453, "learning_rate": 8.324889297529217e-06, "loss": 0.862, "step": 2216 }, { "epoch": 0.29, "grad_norm": 1.2706048488616943, "learning_rate": 8.323304061919633e-06, "loss": 0.8338, "step": 2217 }, { "epoch": 0.29, "grad_norm": 0.6874626278877258, "learning_rate": 8.321718227661881e-06, "loss": 1.0847, "step": 2218 }, { "epoch": 0.29, "grad_norm": 1.2987312078475952, "learning_rate": 8.320131795041633e-06, "loss": 0.8207, "step": 2219 }, { "epoch": 0.29, "grad_norm": 1.363914132118225, "learning_rate": 8.31854476434466e-06, "loss": 0.8012, "step": 2220 }, { "epoch": 0.29, "grad_norm": 1.3315943479537964, "learning_rate": 8.316957135856845e-06, "loss": 0.7372, "step": 2221 }, { "epoch": 0.29, "grad_norm": 1.2819491624832153, "learning_rate": 8.315368909864177e-06, "loss": 0.8547, "step": 2222 }, { "epoch": 0.29, "grad_norm": 1.3146238327026367, "learning_rate": 8.313780086652753e-06, "loss": 0.8178, "step": 2223 }, { "epoch": 0.29, "grad_norm": 1.2729042768478394, "learning_rate": 8.31219066650878e-06, "loss": 0.8195, "step": 2224 }, { "epoch": 0.29, "grad_norm": 0.7487366795539856, "learning_rate": 8.31060064971857e-06, "loss": 1.0597, "step": 2225 }, { "epoch": 0.29, "grad_norm": 1.3102303743362427, "learning_rate": 8.30901003656854e-06, "loss": 0.7915, "step": 2226 }, { "epoch": 0.29, "grad_norm": 1.3731590509414673, "learning_rate": 8.307418827345224e-06, "loss": 0.8462, "step": 2227 }, { "epoch": 0.29, "grad_norm": 0.6539173126220703, "learning_rate": 8.305827022335249e-06, "loss": 1.0449, "step": 2228 }, { "epoch": 0.29, "grad_norm": 1.4858068227767944, "learning_rate": 8.30423462182536e-06, "loss": 0.8489, "step": 2229 }, { "epoch": 0.29, "grad_norm": 1.4141820669174194, "learning_rate": 8.30264162610241e-06, "loss": 0.8664, "step": 2230 }, { "epoch": 0.29, "grad_norm": 1.2307460308074951, "learning_rate": 8.30104803545335e-06, "loss": 0.8167, "step": 2231 }, { "epoch": 0.29, "grad_norm": 0.7746099829673767, "learning_rate": 8.299453850165247e-06, "loss": 1.0694, "step": 2232 }, { "epoch": 0.29, "grad_norm": 1.4137424230575562, "learning_rate": 8.297859070525271e-06, "loss": 0.82, "step": 2233 }, { "epoch": 0.29, "grad_norm": 1.3785054683685303, "learning_rate": 8.296263696820701e-06, "loss": 0.8701, "step": 2234 }, { "epoch": 0.29, "grad_norm": 1.4390004873275757, "learning_rate": 8.29466772933892e-06, "loss": 0.8488, "step": 2235 }, { "epoch": 0.29, "grad_norm": 1.3524168729782104, "learning_rate": 8.293071168367422e-06, "loss": 0.8709, "step": 2236 }, { "epoch": 0.29, "grad_norm": 1.4094816446304321, "learning_rate": 8.291474014193803e-06, "loss": 0.8784, "step": 2237 }, { "epoch": 0.29, "grad_norm": 1.2895920276641846, "learning_rate": 8.289876267105771e-06, "loss": 0.851, "step": 2238 }, { "epoch": 0.29, "grad_norm": 1.5393768548965454, "learning_rate": 8.288277927391138e-06, "loss": 0.8686, "step": 2239 }, { "epoch": 0.29, "grad_norm": 1.4619711637496948, "learning_rate": 8.286678995337823e-06, "loss": 0.8793, "step": 2240 }, { "epoch": 0.29, "grad_norm": 1.378340244293213, "learning_rate": 8.285079471233852e-06, "loss": 0.8538, "step": 2241 }, { "epoch": 0.29, "grad_norm": 1.3138666152954102, "learning_rate": 8.283479355367357e-06, "loss": 0.8303, "step": 2242 }, { "epoch": 0.29, "grad_norm": 1.4475826025009155, "learning_rate": 8.281878648026577e-06, "loss": 0.841, "step": 2243 }, { "epoch": 0.29, "grad_norm": 1.3164142370224, "learning_rate": 8.280277349499863e-06, "loss": 0.8309, "step": 2244 }, { "epoch": 0.29, "grad_norm": 1.3644750118255615, "learning_rate": 8.278675460075658e-06, "loss": 0.8751, "step": 2245 }, { "epoch": 0.29, "grad_norm": 1.3504383563995361, "learning_rate": 8.277072980042528e-06, "loss": 0.8125, "step": 2246 }, { "epoch": 0.29, "grad_norm": 1.020343542098999, "learning_rate": 8.275469909689134e-06, "loss": 1.0501, "step": 2247 }, { "epoch": 0.29, "grad_norm": 1.2835469245910645, "learning_rate": 8.27386624930425e-06, "loss": 0.8291, "step": 2248 }, { "epoch": 0.29, "grad_norm": 0.6611838340759277, "learning_rate": 8.27226199917675e-06, "loss": 1.0602, "step": 2249 }, { "epoch": 0.29, "grad_norm": 0.7130618691444397, "learning_rate": 8.270657159595623e-06, "loss": 1.0336, "step": 2250 }, { "epoch": 0.29, "grad_norm": 1.2863787412643433, "learning_rate": 8.269051730849955e-06, "loss": 0.7761, "step": 2251 }, { "epoch": 0.3, "grad_norm": 0.7202798128128052, "learning_rate": 8.267445713228943e-06, "loss": 1.0481, "step": 2252 }, { "epoch": 0.3, "grad_norm": 1.3048022985458374, "learning_rate": 8.265839107021893e-06, "loss": 0.8616, "step": 2253 }, { "epoch": 0.3, "grad_norm": 1.9376673698425293, "learning_rate": 8.264231912518207e-06, "loss": 0.8758, "step": 2254 }, { "epoch": 0.3, "grad_norm": 0.728493332862854, "learning_rate": 8.262624130007404e-06, "loss": 1.0705, "step": 2255 }, { "epoch": 0.3, "grad_norm": 0.7590440511703491, "learning_rate": 8.261015759779103e-06, "loss": 1.0529, "step": 2256 }, { "epoch": 0.3, "grad_norm": 1.444746971130371, "learning_rate": 8.259406802123029e-06, "loss": 0.8228, "step": 2257 }, { "epoch": 0.3, "grad_norm": 1.300676941871643, "learning_rate": 8.257797257329017e-06, "loss": 0.8318, "step": 2258 }, { "epoch": 0.3, "grad_norm": 0.6737804412841797, "learning_rate": 8.256187125687001e-06, "loss": 1.054, "step": 2259 }, { "epoch": 0.3, "grad_norm": 1.2791285514831543, "learning_rate": 8.254576407487026e-06, "loss": 0.8121, "step": 2260 }, { "epoch": 0.3, "grad_norm": 1.426448941230774, "learning_rate": 8.25296510301924e-06, "loss": 0.8171, "step": 2261 }, { "epoch": 0.3, "grad_norm": 1.3177324533462524, "learning_rate": 8.251353212573899e-06, "loss": 0.7614, "step": 2262 }, { "epoch": 0.3, "grad_norm": 1.2996282577514648, "learning_rate": 8.249740736441362e-06, "loss": 0.7991, "step": 2263 }, { "epoch": 0.3, "grad_norm": 0.839264452457428, "learning_rate": 8.248127674912097e-06, "loss": 1.0826, "step": 2264 }, { "epoch": 0.3, "grad_norm": 1.3135242462158203, "learning_rate": 8.246514028276675e-06, "loss": 0.8975, "step": 2265 }, { "epoch": 0.3, "grad_norm": 1.2558568716049194, "learning_rate": 8.244899796825767e-06, "loss": 0.8404, "step": 2266 }, { "epoch": 0.3, "grad_norm": 1.3396544456481934, "learning_rate": 8.243284980850164e-06, "loss": 0.8094, "step": 2267 }, { "epoch": 0.3, "grad_norm": 1.3931080102920532, "learning_rate": 8.241669580640746e-06, "loss": 0.8375, "step": 2268 }, { "epoch": 0.3, "grad_norm": 1.2619189023971558, "learning_rate": 8.240053596488507e-06, "loss": 0.8428, "step": 2269 }, { "epoch": 0.3, "grad_norm": 1.4188592433929443, "learning_rate": 8.238437028684546e-06, "loss": 0.8703, "step": 2270 }, { "epoch": 0.3, "grad_norm": 1.2911392450332642, "learning_rate": 8.236819877520064e-06, "loss": 0.8124, "step": 2271 }, { "epoch": 0.3, "grad_norm": 1.3610613346099854, "learning_rate": 8.235202143286371e-06, "loss": 0.8664, "step": 2272 }, { "epoch": 0.3, "grad_norm": 1.7924094200134277, "learning_rate": 8.23358382627488e-06, "loss": 0.894, "step": 2273 }, { "epoch": 0.3, "grad_norm": 0.8347455859184265, "learning_rate": 8.231964926777106e-06, "loss": 1.0319, "step": 2274 }, { "epoch": 0.3, "grad_norm": 1.7406535148620605, "learning_rate": 8.230345445084675e-06, "loss": 0.829, "step": 2275 }, { "epoch": 0.3, "grad_norm": 1.4192242622375488, "learning_rate": 8.228725381489312e-06, "loss": 0.8504, "step": 2276 }, { "epoch": 0.3, "grad_norm": 0.6551278829574585, "learning_rate": 8.227104736282851e-06, "loss": 1.0684, "step": 2277 }, { "epoch": 0.3, "grad_norm": 1.453993797302246, "learning_rate": 8.225483509757231e-06, "loss": 0.814, "step": 2278 }, { "epoch": 0.3, "grad_norm": 1.369248390197754, "learning_rate": 8.22386170220449e-06, "loss": 0.8514, "step": 2279 }, { "epoch": 0.3, "grad_norm": 1.3353925943374634, "learning_rate": 8.222239313916776e-06, "loss": 0.8494, "step": 2280 }, { "epoch": 0.3, "grad_norm": 1.3484148979187012, "learning_rate": 8.220616345186342e-06, "loss": 0.8334, "step": 2281 }, { "epoch": 0.3, "grad_norm": 1.4887903928756714, "learning_rate": 8.218992796305544e-06, "loss": 0.8747, "step": 2282 }, { "epoch": 0.3, "grad_norm": 1.0294548273086548, "learning_rate": 8.217368667566838e-06, "loss": 1.0539, "step": 2283 }, { "epoch": 0.3, "grad_norm": 1.3509122133255005, "learning_rate": 8.215743959262793e-06, "loss": 0.8691, "step": 2284 }, { "epoch": 0.3, "grad_norm": 1.3153961896896362, "learning_rate": 8.214118671686078e-06, "loss": 0.8736, "step": 2285 }, { "epoch": 0.3, "grad_norm": 0.7255154252052307, "learning_rate": 8.212492805129465e-06, "loss": 1.0625, "step": 2286 }, { "epoch": 0.3, "grad_norm": 1.3935507535934448, "learning_rate": 8.21086635988583e-06, "loss": 0.8474, "step": 2287 }, { "epoch": 0.3, "grad_norm": 0.7571918368339539, "learning_rate": 8.209239336248159e-06, "loss": 1.0459, "step": 2288 }, { "epoch": 0.3, "grad_norm": 0.6699935793876648, "learning_rate": 8.207611734509537e-06, "loss": 1.0548, "step": 2289 }, { "epoch": 0.3, "grad_norm": 0.6535113453865051, "learning_rate": 8.205983554963152e-06, "loss": 1.0547, "step": 2290 }, { "epoch": 0.3, "grad_norm": 1.395221471786499, "learning_rate": 8.204354797902304e-06, "loss": 0.8484, "step": 2291 }, { "epoch": 0.3, "grad_norm": 1.3916034698486328, "learning_rate": 8.202725463620384e-06, "loss": 0.8588, "step": 2292 }, { "epoch": 0.3, "grad_norm": 0.9634713530540466, "learning_rate": 8.2010955524109e-06, "loss": 1.0535, "step": 2293 }, { "epoch": 0.3, "grad_norm": 1.2412477731704712, "learning_rate": 8.199465064567457e-06, "loss": 0.8243, "step": 2294 }, { "epoch": 0.3, "grad_norm": 1.4024105072021484, "learning_rate": 8.197834000383766e-06, "loss": 0.7941, "step": 2295 }, { "epoch": 0.3, "grad_norm": 1.428576946258545, "learning_rate": 8.19620236015364e-06, "loss": 0.9155, "step": 2296 }, { "epoch": 0.3, "grad_norm": 1.3855758905410767, "learning_rate": 8.194570144170997e-06, "loss": 0.9016, "step": 2297 }, { "epoch": 0.3, "grad_norm": 1.3113726377487183, "learning_rate": 8.192937352729858e-06, "loss": 0.8695, "step": 2298 }, { "epoch": 0.3, "grad_norm": 0.9188327789306641, "learning_rate": 8.191303986124347e-06, "loss": 1.032, "step": 2299 }, { "epoch": 0.3, "grad_norm": 1.35025954246521, "learning_rate": 8.189670044648696e-06, "loss": 0.8236, "step": 2300 }, { "epoch": 0.3, "grad_norm": 0.7007113695144653, "learning_rate": 8.188035528597237e-06, "loss": 1.06, "step": 2301 }, { "epoch": 0.3, "grad_norm": 1.3743822574615479, "learning_rate": 8.186400438264405e-06, "loss": 0.8564, "step": 2302 }, { "epoch": 0.3, "grad_norm": 1.4081745147705078, "learning_rate": 8.184764773944737e-06, "loss": 0.7815, "step": 2303 }, { "epoch": 0.3, "grad_norm": 1.4448531866073608, "learning_rate": 8.183128535932882e-06, "loss": 0.8475, "step": 2304 }, { "epoch": 0.3, "grad_norm": 1.4194762706756592, "learning_rate": 8.18149172452358e-06, "loss": 0.8482, "step": 2305 }, { "epoch": 0.3, "grad_norm": 1.3547059297561646, "learning_rate": 8.179854340011683e-06, "loss": 0.9015, "step": 2306 }, { "epoch": 0.3, "grad_norm": 1.3098593950271606, "learning_rate": 8.178216382692143e-06, "loss": 0.8481, "step": 2307 }, { "epoch": 0.3, "grad_norm": 1.662400722503662, "learning_rate": 8.176577852860015e-06, "loss": 0.8282, "step": 2308 }, { "epoch": 0.3, "grad_norm": 1.2198748588562012, "learning_rate": 8.174938750810461e-06, "loss": 1.0566, "step": 2309 }, { "epoch": 0.3, "grad_norm": 0.9344524145126343, "learning_rate": 8.173299076838739e-06, "loss": 1.0446, "step": 2310 }, { "epoch": 0.3, "grad_norm": 0.7774750590324402, "learning_rate": 8.171658831240216e-06, "loss": 1.0451, "step": 2311 }, { "epoch": 0.3, "grad_norm": 0.7832086086273193, "learning_rate": 8.170018014310361e-06, "loss": 1.0395, "step": 2312 }, { "epoch": 0.3, "grad_norm": 1.5134721994400024, "learning_rate": 8.168376626344743e-06, "loss": 0.8961, "step": 2313 }, { "epoch": 0.3, "grad_norm": 1.4737492799758911, "learning_rate": 8.166734667639039e-06, "loss": 0.8688, "step": 2314 }, { "epoch": 0.3, "grad_norm": 1.285037636756897, "learning_rate": 8.16509213848902e-06, "loss": 0.7841, "step": 2315 }, { "epoch": 0.3, "grad_norm": 1.4818613529205322, "learning_rate": 8.163449039190569e-06, "loss": 1.0715, "step": 2316 }, { "epoch": 0.3, "grad_norm": 1.3240011930465698, "learning_rate": 8.161805370039671e-06, "loss": 0.7898, "step": 2317 }, { "epoch": 0.3, "grad_norm": 1.6011556386947632, "learning_rate": 8.160161131332404e-06, "loss": 0.7884, "step": 2318 }, { "epoch": 0.3, "grad_norm": 1.487117886543274, "learning_rate": 8.15851632336496e-06, "loss": 0.8278, "step": 2319 }, { "epoch": 0.3, "grad_norm": 1.3158347606658936, "learning_rate": 8.156870946433628e-06, "loss": 0.8352, "step": 2320 }, { "epoch": 0.3, "grad_norm": 1.3481274843215942, "learning_rate": 8.1552250008348e-06, "loss": 0.8362, "step": 2321 }, { "epoch": 0.3, "grad_norm": 1.3890187740325928, "learning_rate": 8.153578486864967e-06, "loss": 0.8474, "step": 2322 }, { "epoch": 0.3, "grad_norm": 1.3059630393981934, "learning_rate": 8.151931404820734e-06, "loss": 0.8564, "step": 2323 }, { "epoch": 0.3, "grad_norm": 1.1939616203308105, "learning_rate": 8.150283754998795e-06, "loss": 1.0558, "step": 2324 }, { "epoch": 0.3, "grad_norm": 0.8964215517044067, "learning_rate": 8.148635537695955e-06, "loss": 1.053, "step": 2325 }, { "epoch": 0.3, "grad_norm": 0.7117035388946533, "learning_rate": 8.146986753209114e-06, "loss": 1.0406, "step": 2326 }, { "epoch": 0.3, "grad_norm": 1.9298481941223145, "learning_rate": 8.14533740183528e-06, "loss": 0.865, "step": 2327 }, { "epoch": 0.31, "grad_norm": 1.4362845420837402, "learning_rate": 8.143687483871563e-06, "loss": 0.8213, "step": 2328 }, { "epoch": 0.31, "grad_norm": 1.464491367340088, "learning_rate": 8.142036999615171e-06, "loss": 0.8215, "step": 2329 }, { "epoch": 0.31, "grad_norm": 1.4191770553588867, "learning_rate": 8.14038594936342e-06, "loss": 0.804, "step": 2330 }, { "epoch": 0.31, "grad_norm": 1.3844040632247925, "learning_rate": 8.138734333413715e-06, "loss": 0.8424, "step": 2331 }, { "epoch": 0.31, "grad_norm": 1.307418704032898, "learning_rate": 8.137082152063584e-06, "loss": 0.8646, "step": 2332 }, { "epoch": 0.31, "grad_norm": 1.3769235610961914, "learning_rate": 8.13542940561064e-06, "loss": 0.8778, "step": 2333 }, { "epoch": 0.31, "grad_norm": 2.48341965675354, "learning_rate": 8.133776094352602e-06, "loss": 0.8567, "step": 2334 }, { "epoch": 0.31, "grad_norm": 1.4418153762817383, "learning_rate": 8.13212221858729e-06, "loss": 0.8705, "step": 2335 }, { "epoch": 0.31, "grad_norm": 1.4209741353988647, "learning_rate": 8.130467778612634e-06, "loss": 0.8565, "step": 2336 }, { "epoch": 0.31, "grad_norm": 1.4381884336471558, "learning_rate": 8.128812774726651e-06, "loss": 0.8599, "step": 2337 }, { "epoch": 0.31, "grad_norm": 1.4004125595092773, "learning_rate": 8.127157207227473e-06, "loss": 0.8445, "step": 2338 }, { "epoch": 0.31, "grad_norm": 1.2675215005874634, "learning_rate": 8.125501076413325e-06, "loss": 0.8414, "step": 2339 }, { "epoch": 0.31, "grad_norm": 2.347928285598755, "learning_rate": 8.123844382582538e-06, "loss": 1.0555, "step": 2340 }, { "epoch": 0.31, "grad_norm": 2.0362441539764404, "learning_rate": 8.122187126033542e-06, "loss": 1.0869, "step": 2341 }, { "epoch": 0.31, "grad_norm": 1.402421236038208, "learning_rate": 8.12052930706487e-06, "loss": 0.8342, "step": 2342 }, { "epoch": 0.31, "grad_norm": 1.3436527252197266, "learning_rate": 8.118870925975158e-06, "loss": 0.8351, "step": 2343 }, { "epoch": 0.31, "grad_norm": 1.3607608079910278, "learning_rate": 8.117211983063134e-06, "loss": 0.883, "step": 2344 }, { "epoch": 0.31, "grad_norm": 1.384363055229187, "learning_rate": 8.115552478627642e-06, "loss": 0.8862, "step": 2345 }, { "epoch": 0.31, "grad_norm": 1.3540140390396118, "learning_rate": 8.113892412967614e-06, "loss": 0.7536, "step": 2346 }, { "epoch": 0.31, "grad_norm": 1.355247974395752, "learning_rate": 8.11223178638209e-06, "loss": 0.8347, "step": 2347 }, { "epoch": 0.31, "grad_norm": 1.3579199314117432, "learning_rate": 8.11057059917021e-06, "loss": 0.87, "step": 2348 }, { "epoch": 0.31, "grad_norm": 1.4453706741333008, "learning_rate": 8.108908851631213e-06, "loss": 0.8231, "step": 2349 }, { "epoch": 0.31, "grad_norm": 1.3087586164474487, "learning_rate": 8.10724654406444e-06, "loss": 0.8187, "step": 2350 }, { "epoch": 0.31, "grad_norm": 1.6567524671554565, "learning_rate": 8.105583676769336e-06, "loss": 0.91, "step": 2351 }, { "epoch": 0.31, "grad_norm": 1.3838709592819214, "learning_rate": 8.103920250045442e-06, "loss": 0.8646, "step": 2352 }, { "epoch": 0.31, "grad_norm": 2.138636589050293, "learning_rate": 8.1022562641924e-06, "loss": 1.0719, "step": 2353 }, { "epoch": 0.31, "grad_norm": 1.3818163871765137, "learning_rate": 8.100591719509959e-06, "loss": 0.8622, "step": 2354 }, { "epoch": 0.31, "grad_norm": 1.3608489036560059, "learning_rate": 8.09892661629796e-06, "loss": 0.8774, "step": 2355 }, { "epoch": 0.31, "grad_norm": 1.2475318908691406, "learning_rate": 8.09726095485635e-06, "loss": 1.0494, "step": 2356 }, { "epoch": 0.31, "grad_norm": 0.7996786832809448, "learning_rate": 8.095594735485177e-06, "loss": 1.0624, "step": 2357 }, { "epoch": 0.31, "grad_norm": 0.8337989449501038, "learning_rate": 8.093927958484585e-06, "loss": 1.0436, "step": 2358 }, { "epoch": 0.31, "grad_norm": 1.5361480712890625, "learning_rate": 8.092260624154822e-06, "loss": 0.8823, "step": 2359 }, { "epoch": 0.31, "grad_norm": 1.4595774412155151, "learning_rate": 8.09059273279624e-06, "loss": 0.8956, "step": 2360 }, { "epoch": 0.31, "grad_norm": 1.5412917137145996, "learning_rate": 8.08892428470928e-06, "loss": 0.8942, "step": 2361 }, { "epoch": 0.31, "grad_norm": 1.8968170881271362, "learning_rate": 8.087255280194494e-06, "loss": 1.0487, "step": 2362 }, { "epoch": 0.31, "grad_norm": 1.4414489269256592, "learning_rate": 8.08558571955253e-06, "loss": 0.8217, "step": 2363 }, { "epoch": 0.31, "grad_norm": 1.6500309705734253, "learning_rate": 8.083915603084138e-06, "loss": 0.9323, "step": 2364 }, { "epoch": 0.31, "grad_norm": 1.6758666038513184, "learning_rate": 8.082244931090167e-06, "loss": 0.8143, "step": 2365 }, { "epoch": 0.31, "grad_norm": 1.4272942543029785, "learning_rate": 8.080573703871563e-06, "loss": 0.837, "step": 2366 }, { "epoch": 0.31, "grad_norm": 1.415355920791626, "learning_rate": 8.078901921729377e-06, "loss": 0.8733, "step": 2367 }, { "epoch": 0.31, "grad_norm": 1.2753634452819824, "learning_rate": 8.077229584964757e-06, "loss": 1.0563, "step": 2368 }, { "epoch": 0.31, "grad_norm": 0.9524385929107666, "learning_rate": 8.075556693878954e-06, "loss": 1.0491, "step": 2369 }, { "epoch": 0.31, "grad_norm": 0.7275992035865784, "learning_rate": 8.073883248773313e-06, "loss": 1.0571, "step": 2370 }, { "epoch": 0.31, "grad_norm": 1.446068286895752, "learning_rate": 8.072209249949285e-06, "loss": 0.8688, "step": 2371 }, { "epoch": 0.31, "grad_norm": 1.4021942615509033, "learning_rate": 8.07053469770842e-06, "loss": 0.8194, "step": 2372 }, { "epoch": 0.31, "grad_norm": 1.3680686950683594, "learning_rate": 8.068859592352359e-06, "loss": 1.0657, "step": 2373 }, { "epoch": 0.31, "grad_norm": 1.4358683824539185, "learning_rate": 8.067183934182856e-06, "loss": 0.8602, "step": 2374 }, { "epoch": 0.31, "grad_norm": 1.3173518180847168, "learning_rate": 8.065507723501757e-06, "loss": 0.842, "step": 2375 }, { "epoch": 0.31, "grad_norm": 1.3087056875228882, "learning_rate": 8.063830960611005e-06, "loss": 0.87, "step": 2376 }, { "epoch": 0.31, "grad_norm": 1.3683677911758423, "learning_rate": 8.062153645812651e-06, "loss": 0.8259, "step": 2377 }, { "epoch": 0.31, "grad_norm": 1.4444836378097534, "learning_rate": 8.060475779408839e-06, "loss": 0.8458, "step": 2378 }, { "epoch": 0.31, "grad_norm": 1.3532943725585938, "learning_rate": 8.058797361701812e-06, "loss": 0.8098, "step": 2379 }, { "epoch": 0.31, "grad_norm": 1.3386350870132446, "learning_rate": 8.057118392993916e-06, "loss": 0.8074, "step": 2380 }, { "epoch": 0.31, "grad_norm": 1.3303593397140503, "learning_rate": 8.055438873587593e-06, "loss": 0.7943, "step": 2381 }, { "epoch": 0.31, "grad_norm": 1.2905372381210327, "learning_rate": 8.053758803785387e-06, "loss": 0.7859, "step": 2382 }, { "epoch": 0.31, "grad_norm": 1.2911192178726196, "learning_rate": 8.052078183889938e-06, "loss": 1.0658, "step": 2383 }, { "epoch": 0.31, "grad_norm": 1.6722050905227661, "learning_rate": 8.05039701420399e-06, "loss": 0.7672, "step": 2384 }, { "epoch": 0.31, "grad_norm": 0.9247140288352966, "learning_rate": 8.04871529503038e-06, "loss": 1.0387, "step": 2385 }, { "epoch": 0.31, "grad_norm": 0.7474027276039124, "learning_rate": 8.047033026672049e-06, "loss": 1.0289, "step": 2386 }, { "epoch": 0.31, "grad_norm": 1.3855465650558472, "learning_rate": 8.045350209432033e-06, "loss": 0.8622, "step": 2387 }, { "epoch": 0.31, "grad_norm": 1.5333259105682373, "learning_rate": 8.04366684361347e-06, "loss": 0.8106, "step": 2388 }, { "epoch": 0.31, "grad_norm": 1.0067236423492432, "learning_rate": 8.041982929519595e-06, "loss": 1.0886, "step": 2389 }, { "epoch": 0.31, "grad_norm": 1.3393597602844238, "learning_rate": 8.040298467453742e-06, "loss": 0.7814, "step": 2390 }, { "epoch": 0.31, "grad_norm": 0.9999254941940308, "learning_rate": 8.038613457719344e-06, "loss": 1.0686, "step": 2391 }, { "epoch": 0.31, "grad_norm": 1.313879370689392, "learning_rate": 8.036927900619932e-06, "loss": 0.8232, "step": 2392 }, { "epoch": 0.31, "grad_norm": 1.3661153316497803, "learning_rate": 8.035241796459142e-06, "loss": 0.8591, "step": 2393 }, { "epoch": 0.31, "grad_norm": 1.316706895828247, "learning_rate": 8.033555145540693e-06, "loss": 0.8746, "step": 2394 }, { "epoch": 0.31, "grad_norm": 1.5023155212402344, "learning_rate": 8.031867948168418e-06, "loss": 0.8651, "step": 2395 }, { "epoch": 0.31, "grad_norm": 1.344814658164978, "learning_rate": 8.030180204646245e-06, "loss": 0.8244, "step": 2396 }, { "epoch": 0.31, "grad_norm": 1.3563851118087769, "learning_rate": 8.028491915278193e-06, "loss": 0.8248, "step": 2397 }, { "epoch": 0.31, "grad_norm": 1.3161944150924683, "learning_rate": 8.026803080368387e-06, "loss": 0.8248, "step": 2398 }, { "epoch": 0.31, "grad_norm": 1.0391982793807983, "learning_rate": 8.025113700221047e-06, "loss": 1.0613, "step": 2399 }, { "epoch": 0.31, "grad_norm": 1.3312629461288452, "learning_rate": 8.023423775140494e-06, "loss": 0.8422, "step": 2400 }, { "epoch": 0.31, "grad_norm": 1.3649829626083374, "learning_rate": 8.021733305431141e-06, "loss": 0.8466, "step": 2401 }, { "epoch": 0.31, "grad_norm": 1.4432228803634644, "learning_rate": 8.020042291397507e-06, "loss": 0.8977, "step": 2402 }, { "epoch": 0.31, "grad_norm": 1.4087326526641846, "learning_rate": 8.018350733344203e-06, "loss": 0.8761, "step": 2403 }, { "epoch": 0.32, "grad_norm": 0.6844815611839294, "learning_rate": 8.016658631575942e-06, "loss": 1.0248, "step": 2404 }, { "epoch": 0.32, "grad_norm": 0.6906821131706238, "learning_rate": 8.01496598639753e-06, "loss": 1.053, "step": 2405 }, { "epoch": 0.32, "grad_norm": 1.296325445175171, "learning_rate": 8.01327279811388e-06, "loss": 0.8058, "step": 2406 }, { "epoch": 0.32, "grad_norm": 1.377892017364502, "learning_rate": 8.01157906702999e-06, "loss": 0.8175, "step": 2407 }, { "epoch": 0.32, "grad_norm": 1.372336506843567, "learning_rate": 8.009884793450968e-06, "loss": 0.8394, "step": 2408 }, { "epoch": 0.32, "grad_norm": 0.703333854675293, "learning_rate": 8.008189977682013e-06, "loss": 1.0565, "step": 2409 }, { "epoch": 0.32, "grad_norm": 1.5397179126739502, "learning_rate": 8.00649462002842e-06, "loss": 0.7871, "step": 2410 }, { "epoch": 0.32, "grad_norm": 1.2783769369125366, "learning_rate": 8.004798720795588e-06, "loss": 0.8768, "step": 2411 }, { "epoch": 0.32, "grad_norm": 1.442527174949646, "learning_rate": 8.00310228028901e-06, "loss": 0.8684, "step": 2412 }, { "epoch": 0.32, "grad_norm": 0.7591844797134399, "learning_rate": 8.001405298814275e-06, "loss": 1.056, "step": 2413 }, { "epoch": 0.32, "grad_norm": 1.4070191383361816, "learning_rate": 7.999707776677072e-06, "loss": 0.8657, "step": 2414 }, { "epoch": 0.32, "grad_norm": 1.2698726654052734, "learning_rate": 7.998009714183188e-06, "loss": 0.8244, "step": 2415 }, { "epoch": 0.32, "grad_norm": 1.331717848777771, "learning_rate": 7.996311111638505e-06, "loss": 0.8203, "step": 2416 }, { "epoch": 0.32, "grad_norm": 1.3903571367263794, "learning_rate": 7.994611969349004e-06, "loss": 0.925, "step": 2417 }, { "epoch": 0.32, "grad_norm": 0.6970469951629639, "learning_rate": 7.99291228762076e-06, "loss": 1.0371, "step": 2418 }, { "epoch": 0.32, "grad_norm": 1.3843894004821777, "learning_rate": 7.991212066759949e-06, "loss": 0.8825, "step": 2419 }, { "epoch": 0.32, "grad_norm": 1.2827684879302979, "learning_rate": 7.989511307072842e-06, "loss": 0.8343, "step": 2420 }, { "epoch": 0.32, "grad_norm": 1.3491034507751465, "learning_rate": 7.98781000886581e-06, "loss": 0.8822, "step": 2421 }, { "epoch": 0.32, "grad_norm": 2.695162534713745, "learning_rate": 7.986108172445318e-06, "loss": 0.8457, "step": 2422 }, { "epoch": 0.32, "grad_norm": 0.7667326331138611, "learning_rate": 7.984405798117928e-06, "loss": 1.0407, "step": 2423 }, { "epoch": 0.32, "grad_norm": 1.3128838539123535, "learning_rate": 7.982702886190299e-06, "loss": 0.821, "step": 2424 }, { "epoch": 0.32, "grad_norm": 0.6551548838615417, "learning_rate": 7.980999436969187e-06, "loss": 1.0504, "step": 2425 }, { "epoch": 0.32, "grad_norm": 1.5275671482086182, "learning_rate": 7.97929545076145e-06, "loss": 0.8647, "step": 2426 }, { "epoch": 0.32, "grad_norm": 1.3617916107177734, "learning_rate": 7.977590927874033e-06, "loss": 0.8208, "step": 2427 }, { "epoch": 0.32, "grad_norm": 0.6816791892051697, "learning_rate": 7.975885868613985e-06, "loss": 1.079, "step": 2428 }, { "epoch": 0.32, "grad_norm": 1.2921150922775269, "learning_rate": 7.974180273288447e-06, "loss": 0.8592, "step": 2429 }, { "epoch": 0.32, "grad_norm": 1.342817783355713, "learning_rate": 7.972474142204663e-06, "loss": 0.8211, "step": 2430 }, { "epoch": 0.32, "grad_norm": 1.2404446601867676, "learning_rate": 7.970767475669966e-06, "loss": 0.7965, "step": 2431 }, { "epoch": 0.32, "grad_norm": 1.2241322994232178, "learning_rate": 7.969060273991787e-06, "loss": 0.7693, "step": 2432 }, { "epoch": 0.32, "grad_norm": 1.2389676570892334, "learning_rate": 7.96735253747766e-06, "loss": 0.8068, "step": 2433 }, { "epoch": 0.32, "grad_norm": 0.7357007265090942, "learning_rate": 7.965644266435209e-06, "loss": 1.0393, "step": 2434 }, { "epoch": 0.32, "grad_norm": 0.719470202922821, "learning_rate": 7.963935461172156e-06, "loss": 1.0554, "step": 2435 }, { "epoch": 0.32, "grad_norm": 1.3436802625656128, "learning_rate": 7.962226121996317e-06, "loss": 0.8288, "step": 2436 }, { "epoch": 0.32, "grad_norm": 1.2764242887496948, "learning_rate": 7.960516249215609e-06, "loss": 0.8855, "step": 2437 }, { "epoch": 0.32, "grad_norm": 1.3447227478027344, "learning_rate": 7.95880584313804e-06, "loss": 0.8736, "step": 2438 }, { "epoch": 0.32, "grad_norm": 1.3928941488265991, "learning_rate": 7.957094904071718e-06, "loss": 0.8505, "step": 2439 }, { "epoch": 0.32, "grad_norm": 1.3534959554672241, "learning_rate": 7.955383432324846e-06, "loss": 0.8627, "step": 2440 }, { "epoch": 0.32, "grad_norm": 1.4158580303192139, "learning_rate": 7.95367142820572e-06, "loss": 0.8409, "step": 2441 }, { "epoch": 0.32, "grad_norm": 0.9437967538833618, "learning_rate": 7.951958892022738e-06, "loss": 1.0592, "step": 2442 }, { "epoch": 0.32, "grad_norm": 0.8400505185127258, "learning_rate": 7.950245824084386e-06, "loss": 1.0542, "step": 2443 }, { "epoch": 0.32, "grad_norm": 1.4142165184020996, "learning_rate": 7.948532224699253e-06, "loss": 0.8546, "step": 2444 }, { "epoch": 0.32, "grad_norm": 1.3097450733184814, "learning_rate": 7.946818094176019e-06, "loss": 0.8421, "step": 2445 }, { "epoch": 0.32, "grad_norm": 1.3363726139068604, "learning_rate": 7.945103432823463e-06, "loss": 0.7711, "step": 2446 }, { "epoch": 0.32, "grad_norm": 1.4097660779953003, "learning_rate": 7.943388240950456e-06, "loss": 0.8785, "step": 2447 }, { "epoch": 0.32, "grad_norm": 1.4939074516296387, "learning_rate": 7.941672518865969e-06, "loss": 1.0651, "step": 2448 }, { "epoch": 0.32, "grad_norm": 1.195152759552002, "learning_rate": 7.939956266879063e-06, "loss": 1.0737, "step": 2449 }, { "epoch": 0.32, "grad_norm": 1.3830504417419434, "learning_rate": 7.9382394852989e-06, "loss": 0.7956, "step": 2450 }, { "epoch": 0.32, "grad_norm": 1.4765712022781372, "learning_rate": 7.936522174434736e-06, "loss": 0.9033, "step": 2451 }, { "epoch": 0.32, "grad_norm": 1.0728126764297485, "learning_rate": 7.934804334595915e-06, "loss": 1.0394, "step": 2452 }, { "epoch": 0.32, "grad_norm": 1.43229079246521, "learning_rate": 7.93308596609189e-06, "loss": 0.8753, "step": 2453 }, { "epoch": 0.32, "grad_norm": 1.316386342048645, "learning_rate": 7.9313670692322e-06, "loss": 0.8369, "step": 2454 }, { "epoch": 0.32, "grad_norm": 1.3197741508483887, "learning_rate": 7.929647644326477e-06, "loss": 0.8798, "step": 2455 }, { "epoch": 0.32, "grad_norm": 1.3374468088150024, "learning_rate": 7.927927691684457e-06, "loss": 0.8156, "step": 2456 }, { "epoch": 0.32, "grad_norm": 1.9249439239501953, "learning_rate": 7.926207211615962e-06, "loss": 0.8691, "step": 2457 }, { "epoch": 0.32, "grad_norm": 1.4411239624023438, "learning_rate": 7.924486204430917e-06, "loss": 0.8355, "step": 2458 }, { "epoch": 0.32, "grad_norm": 1.3282010555267334, "learning_rate": 7.922764670439335e-06, "loss": 0.8291, "step": 2459 }, { "epoch": 0.32, "grad_norm": 1.3894848823547363, "learning_rate": 7.921042609951328e-06, "loss": 0.8564, "step": 2460 }, { "epoch": 0.32, "grad_norm": 1.384006142616272, "learning_rate": 7.919320023277101e-06, "loss": 0.8237, "step": 2461 }, { "epoch": 0.32, "grad_norm": 1.3143770694732666, "learning_rate": 7.917596910726958e-06, "loss": 0.8338, "step": 2462 }, { "epoch": 0.32, "grad_norm": 1.4485615491867065, "learning_rate": 7.91587327261129e-06, "loss": 0.8379, "step": 2463 }, { "epoch": 0.32, "grad_norm": 1.340031385421753, "learning_rate": 7.914149109240592e-06, "loss": 0.8805, "step": 2464 }, { "epoch": 0.32, "grad_norm": 1.329480528831482, "learning_rate": 7.912424420925444e-06, "loss": 0.8132, "step": 2465 }, { "epoch": 0.32, "grad_norm": 1.133155107498169, "learning_rate": 7.910699207976527e-06, "loss": 1.0515, "step": 2466 }, { "epoch": 0.32, "grad_norm": 1.3315707445144653, "learning_rate": 7.908973470704614e-06, "loss": 0.8261, "step": 2467 }, { "epoch": 0.32, "grad_norm": 1.3212783336639404, "learning_rate": 7.907247209420575e-06, "loss": 0.7947, "step": 2468 }, { "epoch": 0.32, "grad_norm": 0.7098699808120728, "learning_rate": 7.905520424435371e-06, "loss": 1.063, "step": 2469 }, { "epoch": 0.32, "grad_norm": 1.649369716644287, "learning_rate": 7.90379311606006e-06, "loss": 0.7448, "step": 2470 }, { "epoch": 0.32, "grad_norm": 0.6937199831008911, "learning_rate": 7.90206528460579e-06, "loss": 1.0611, "step": 2471 }, { "epoch": 0.32, "grad_norm": 0.6834391355514526, "learning_rate": 7.90033693038381e-06, "loss": 1.0462, "step": 2472 }, { "epoch": 0.32, "grad_norm": 1.379929542541504, "learning_rate": 7.89860805370546e-06, "loss": 0.8668, "step": 2473 }, { "epoch": 0.32, "grad_norm": 1.2659275531768799, "learning_rate": 7.89687865488217e-06, "loss": 0.7769, "step": 2474 }, { "epoch": 0.32, "grad_norm": 1.4018336534500122, "learning_rate": 7.895148734225472e-06, "loss": 0.8245, "step": 2475 }, { "epoch": 0.32, "grad_norm": 1.2842915058135986, "learning_rate": 7.893418292046984e-06, "loss": 0.8158, "step": 2476 }, { "epoch": 0.32, "grad_norm": 1.3243769407272339, "learning_rate": 7.891687328658423e-06, "loss": 0.8516, "step": 2477 }, { "epoch": 0.32, "grad_norm": 1.3487465381622314, "learning_rate": 7.889955844371601e-06, "loss": 0.8568, "step": 2478 }, { "epoch": 0.32, "grad_norm": 1.47812020778656, "learning_rate": 7.88822383949842e-06, "loss": 0.862, "step": 2479 }, { "epoch": 0.32, "grad_norm": 1.0568212270736694, "learning_rate": 7.886491314350874e-06, "loss": 1.0496, "step": 2480 }, { "epoch": 0.33, "grad_norm": 1.389512300491333, "learning_rate": 7.884758269241057e-06, "loss": 0.8178, "step": 2481 }, { "epoch": 0.33, "grad_norm": 0.7327361106872559, "learning_rate": 7.883024704481154e-06, "loss": 1.0313, "step": 2482 }, { "epoch": 0.33, "grad_norm": 1.4553368091583252, "learning_rate": 7.881290620383442e-06, "loss": 0.8538, "step": 2483 }, { "epoch": 0.33, "grad_norm": 1.3594433069229126, "learning_rate": 7.879556017260293e-06, "loss": 0.8453, "step": 2484 }, { "epoch": 0.33, "grad_norm": 0.8345894813537598, "learning_rate": 7.87782089542417e-06, "loss": 1.0387, "step": 2485 }, { "epoch": 0.33, "grad_norm": 1.3334484100341797, "learning_rate": 7.876085255187638e-06, "loss": 0.8517, "step": 2486 }, { "epoch": 0.33, "grad_norm": 1.7254889011383057, "learning_rate": 7.874349096863341e-06, "loss": 0.8797, "step": 2487 }, { "epoch": 0.33, "grad_norm": 0.7313980460166931, "learning_rate": 7.872612420764031e-06, "loss": 1.0519, "step": 2488 }, { "epoch": 0.33, "grad_norm": 0.7298480272293091, "learning_rate": 7.870875227202544e-06, "loss": 1.0544, "step": 2489 }, { "epoch": 0.33, "grad_norm": 0.6634610295295715, "learning_rate": 7.86913751649181e-06, "loss": 1.0308, "step": 2490 }, { "epoch": 0.33, "grad_norm": 1.3785877227783203, "learning_rate": 7.867399288944855e-06, "loss": 0.8256, "step": 2491 }, { "epoch": 0.33, "grad_norm": 1.2847341299057007, "learning_rate": 7.8656605448748e-06, "loss": 0.7465, "step": 2492 }, { "epoch": 0.33, "grad_norm": 1.3627618551254272, "learning_rate": 7.863921284594856e-06, "loss": 0.8318, "step": 2493 }, { "epoch": 0.33, "grad_norm": 1.3745925426483154, "learning_rate": 7.86218150841832e-06, "loss": 0.88, "step": 2494 }, { "epoch": 0.33, "grad_norm": 1.805799961090088, "learning_rate": 7.8604412166586e-06, "loss": 0.7786, "step": 2495 }, { "epoch": 0.33, "grad_norm": 1.3787671327590942, "learning_rate": 7.858700409629177e-06, "loss": 0.8332, "step": 2496 }, { "epoch": 0.33, "grad_norm": 1.1518840789794922, "learning_rate": 7.856959087643638e-06, "loss": 1.0274, "step": 2497 }, { "epoch": 0.33, "grad_norm": 1.3593764305114746, "learning_rate": 7.855217251015658e-06, "loss": 0.8603, "step": 2498 }, { "epoch": 0.33, "grad_norm": 1.2621194124221802, "learning_rate": 7.853474900059005e-06, "loss": 0.7645, "step": 2499 }, { "epoch": 0.33, "grad_norm": 1.274614691734314, "learning_rate": 7.851732035087537e-06, "loss": 0.8047, "step": 2500 }, { "epoch": 0.33, "grad_norm": 1.460108995437622, "learning_rate": 7.849988656415214e-06, "loss": 0.8244, "step": 2501 }, { "epoch": 0.33, "grad_norm": 0.7456743121147156, "learning_rate": 7.848244764356076e-06, "loss": 1.0567, "step": 2502 }, { "epoch": 0.33, "grad_norm": 1.3053079843521118, "learning_rate": 7.846500359224266e-06, "loss": 0.7913, "step": 2503 }, { "epoch": 0.33, "grad_norm": 1.4072761535644531, "learning_rate": 7.844755441334012e-06, "loss": 0.869, "step": 2504 }, { "epoch": 0.33, "grad_norm": 1.60353422164917, "learning_rate": 7.843010010999636e-06, "loss": 0.7936, "step": 2505 }, { "epoch": 0.33, "grad_norm": 1.2585444450378418, "learning_rate": 7.841264068535559e-06, "loss": 0.8268, "step": 2506 }, { "epoch": 0.33, "grad_norm": 1.3344731330871582, "learning_rate": 7.839517614256282e-06, "loss": 0.8616, "step": 2507 }, { "epoch": 0.33, "grad_norm": 1.2837364673614502, "learning_rate": 7.837770648476411e-06, "loss": 0.8331, "step": 2508 }, { "epoch": 0.33, "grad_norm": 1.3993449211120605, "learning_rate": 7.836023171510637e-06, "loss": 0.8578, "step": 2509 }, { "epoch": 0.33, "grad_norm": 1.8877943754196167, "learning_rate": 7.83427518367374e-06, "loss": 0.8353, "step": 2510 }, { "epoch": 0.33, "grad_norm": 1.3677664995193481, "learning_rate": 7.832526685280601e-06, "loss": 0.8121, "step": 2511 }, { "epoch": 0.33, "grad_norm": 1.3919084072113037, "learning_rate": 7.830777676646187e-06, "loss": 0.8521, "step": 2512 }, { "epoch": 0.33, "grad_norm": 1.431662917137146, "learning_rate": 7.829028158085556e-06, "loss": 0.8197, "step": 2513 }, { "epoch": 0.33, "grad_norm": 1.4091217517852783, "learning_rate": 7.827278129913865e-06, "loss": 0.7799, "step": 2514 }, { "epoch": 0.33, "grad_norm": 1.3398250341415405, "learning_rate": 7.825527592446353e-06, "loss": 0.7925, "step": 2515 }, { "epoch": 0.33, "grad_norm": 1.2664474248886108, "learning_rate": 7.823776545998357e-06, "loss": 0.8838, "step": 2516 }, { "epoch": 0.33, "grad_norm": 1.3594770431518555, "learning_rate": 7.822024990885304e-06, "loss": 0.8098, "step": 2517 }, { "epoch": 0.33, "grad_norm": 0.9194474220275879, "learning_rate": 7.820272927422717e-06, "loss": 1.088, "step": 2518 }, { "epoch": 0.33, "grad_norm": 0.7120622992515564, "learning_rate": 7.818520355926198e-06, "loss": 1.0585, "step": 2519 }, { "epoch": 0.33, "grad_norm": 1.4133563041687012, "learning_rate": 7.816767276711458e-06, "loss": 0.8409, "step": 2520 }, { "epoch": 0.33, "grad_norm": 1.364540457725525, "learning_rate": 7.815013690094284e-06, "loss": 0.7958, "step": 2521 }, { "epoch": 0.33, "grad_norm": 1.0152313709259033, "learning_rate": 7.813259596390565e-06, "loss": 1.0696, "step": 2522 }, { "epoch": 0.33, "grad_norm": 0.8452303409576416, "learning_rate": 7.811504995916275e-06, "loss": 1.0368, "step": 2523 }, { "epoch": 0.33, "grad_norm": 1.4601562023162842, "learning_rate": 7.809749888987482e-06, "loss": 0.8724, "step": 2524 }, { "epoch": 0.33, "grad_norm": 6.988308906555176, "learning_rate": 7.807994275920345e-06, "loss": 1.0445, "step": 2525 }, { "epoch": 0.33, "grad_norm": 1.7820959091186523, "learning_rate": 7.806238157031112e-06, "loss": 0.8488, "step": 2526 }, { "epoch": 0.33, "grad_norm": 1.3502261638641357, "learning_rate": 7.804481532636128e-06, "loss": 0.7596, "step": 2527 }, { "epoch": 0.33, "grad_norm": 1.862627625465393, "learning_rate": 7.80272440305182e-06, "loss": 0.7759, "step": 2528 }, { "epoch": 0.33, "grad_norm": 1.3807685375213623, "learning_rate": 7.800966768594715e-06, "loss": 0.8054, "step": 2529 }, { "epoch": 0.33, "grad_norm": 1.3990918397903442, "learning_rate": 7.799208629581426e-06, "loss": 0.863, "step": 2530 }, { "epoch": 0.33, "grad_norm": 1.3672034740447998, "learning_rate": 7.797449986328657e-06, "loss": 0.8576, "step": 2531 }, { "epoch": 0.33, "grad_norm": 1.509293556213379, "learning_rate": 7.795690839153205e-06, "loss": 1.0633, "step": 2532 }, { "epoch": 0.33, "grad_norm": 1.4536961317062378, "learning_rate": 7.793931188371955e-06, "loss": 0.8246, "step": 2533 }, { "epoch": 0.33, "grad_norm": 0.8627890348434448, "learning_rate": 7.792171034301885e-06, "loss": 1.0767, "step": 2534 }, { "epoch": 0.33, "grad_norm": 1.4928886890411377, "learning_rate": 7.790410377260065e-06, "loss": 0.9063, "step": 2535 }, { "epoch": 0.33, "grad_norm": 1.5045751333236694, "learning_rate": 7.78864921756365e-06, "loss": 0.8625, "step": 2536 }, { "epoch": 0.33, "grad_norm": 1.3289844989776611, "learning_rate": 7.78688755552989e-06, "loss": 0.8852, "step": 2537 }, { "epoch": 0.33, "grad_norm": 1.3063031435012817, "learning_rate": 7.785125391476125e-06, "loss": 0.7718, "step": 2538 }, { "epoch": 0.33, "grad_norm": 2.3264644145965576, "learning_rate": 7.783362725719788e-06, "loss": 0.8725, "step": 2539 }, { "epoch": 0.33, "grad_norm": 1.533951997756958, "learning_rate": 7.781599558578394e-06, "loss": 1.0622, "step": 2540 }, { "epoch": 0.33, "grad_norm": 1.6534215211868286, "learning_rate": 7.779835890369555e-06, "loss": 0.8814, "step": 2541 }, { "epoch": 0.33, "grad_norm": 1.6434799432754517, "learning_rate": 7.778071721410973e-06, "loss": 0.804, "step": 2542 }, { "epoch": 0.33, "grad_norm": 1.473069667816162, "learning_rate": 7.776307052020442e-06, "loss": 0.8295, "step": 2543 }, { "epoch": 0.33, "grad_norm": 1.42866849899292, "learning_rate": 7.774541882515839e-06, "loss": 0.8447, "step": 2544 }, { "epoch": 0.33, "grad_norm": 1.495346188545227, "learning_rate": 7.772776213215136e-06, "loss": 0.841, "step": 2545 }, { "epoch": 0.33, "grad_norm": 1.425960898399353, "learning_rate": 7.771010044436393e-06, "loss": 0.8669, "step": 2546 }, { "epoch": 0.33, "grad_norm": 1.4401205778121948, "learning_rate": 7.769243376497766e-06, "loss": 0.839, "step": 2547 }, { "epoch": 0.33, "grad_norm": 1.442757487297058, "learning_rate": 7.767476209717495e-06, "loss": 0.8516, "step": 2548 }, { "epoch": 0.33, "grad_norm": 0.7462306022644043, "learning_rate": 7.765708544413907e-06, "loss": 1.0613, "step": 2549 }, { "epoch": 0.33, "grad_norm": 1.3948954343795776, "learning_rate": 7.763940380905427e-06, "loss": 0.8138, "step": 2550 }, { "epoch": 0.33, "grad_norm": 1.4263001680374146, "learning_rate": 7.762171719510564e-06, "loss": 0.8154, "step": 2551 }, { "epoch": 0.33, "grad_norm": 1.2735543251037598, "learning_rate": 7.760402560547919e-06, "loss": 0.7852, "step": 2552 }, { "epoch": 0.33, "grad_norm": 1.446519136428833, "learning_rate": 7.758632904336183e-06, "loss": 0.8156, "step": 2553 }, { "epoch": 0.33, "grad_norm": 1.4787596464157104, "learning_rate": 7.756862751194132e-06, "loss": 0.8131, "step": 2554 }, { "epoch": 0.33, "grad_norm": 0.8239370584487915, "learning_rate": 7.755092101440638e-06, "loss": 1.039, "step": 2555 }, { "epoch": 0.33, "grad_norm": 1.38565194606781, "learning_rate": 7.75332095539466e-06, "loss": 0.8411, "step": 2556 }, { "epoch": 0.34, "grad_norm": 1.3863023519515991, "learning_rate": 7.751549313375246e-06, "loss": 0.8219, "step": 2557 }, { "epoch": 0.34, "grad_norm": 1.512394666671753, "learning_rate": 7.74977717570153e-06, "loss": 0.8866, "step": 2558 }, { "epoch": 0.34, "grad_norm": 0.7309461832046509, "learning_rate": 7.748004542692742e-06, "loss": 1.0354, "step": 2559 }, { "epoch": 0.34, "grad_norm": 1.412311315536499, "learning_rate": 7.746231414668195e-06, "loss": 0.8592, "step": 2560 }, { "epoch": 0.34, "grad_norm": 1.308187484741211, "learning_rate": 7.744457791947297e-06, "loss": 0.8833, "step": 2561 }, { "epoch": 0.34, "grad_norm": 0.691969633102417, "learning_rate": 7.742683674849541e-06, "loss": 1.043, "step": 2562 }, { "epoch": 0.34, "grad_norm": 1.335222840309143, "learning_rate": 7.74090906369451e-06, "loss": 0.8352, "step": 2563 }, { "epoch": 0.34, "grad_norm": 0.7244397401809692, "learning_rate": 7.739133958801874e-06, "loss": 1.0441, "step": 2564 }, { "epoch": 0.34, "grad_norm": 1.383521318435669, "learning_rate": 7.737358360491396e-06, "loss": 0.8567, "step": 2565 }, { "epoch": 0.34, "grad_norm": 1.3872034549713135, "learning_rate": 7.735582269082928e-06, "loss": 0.8331, "step": 2566 }, { "epoch": 0.34, "grad_norm": 1.3368312120437622, "learning_rate": 7.733805684896407e-06, "loss": 0.7886, "step": 2567 }, { "epoch": 0.34, "grad_norm": 0.7287898659706116, "learning_rate": 7.732028608251859e-06, "loss": 1.0356, "step": 2568 }, { "epoch": 0.34, "grad_norm": 1.3885807991027832, "learning_rate": 7.730251039469401e-06, "loss": 0.7727, "step": 2569 }, { "epoch": 0.34, "grad_norm": 1.3486676216125488, "learning_rate": 7.728472978869239e-06, "loss": 0.8634, "step": 2570 }, { "epoch": 0.34, "grad_norm": 1.353318214416504, "learning_rate": 7.726694426771667e-06, "loss": 0.8444, "step": 2571 }, { "epoch": 0.34, "grad_norm": 1.6152186393737793, "learning_rate": 7.724915383497064e-06, "loss": 0.8262, "step": 2572 }, { "epoch": 0.34, "grad_norm": 1.4606307744979858, "learning_rate": 7.723135849365906e-06, "loss": 0.8217, "step": 2573 }, { "epoch": 0.34, "grad_norm": 1.33113431930542, "learning_rate": 7.721355824698746e-06, "loss": 0.7972, "step": 2574 }, { "epoch": 0.34, "grad_norm": 1.570235013961792, "learning_rate": 7.719575309816232e-06, "loss": 0.8724, "step": 2575 }, { "epoch": 0.34, "grad_norm": 1.4664751291275024, "learning_rate": 7.717794305039105e-06, "loss": 0.9053, "step": 2576 }, { "epoch": 0.34, "grad_norm": 1.3814932107925415, "learning_rate": 7.716012810688183e-06, "loss": 0.8279, "step": 2577 }, { "epoch": 0.34, "grad_norm": 1.427255392074585, "learning_rate": 7.714230827084383e-06, "loss": 0.873, "step": 2578 }, { "epoch": 0.34, "grad_norm": 1.438101887702942, "learning_rate": 7.7124483545487e-06, "loss": 0.8752, "step": 2579 }, { "epoch": 0.34, "grad_norm": 1.4593545198440552, "learning_rate": 7.710665393402226e-06, "loss": 0.8601, "step": 2580 }, { "epoch": 0.34, "grad_norm": 0.9271795153617859, "learning_rate": 7.708881943966135e-06, "loss": 1.0645, "step": 2581 }, { "epoch": 0.34, "grad_norm": 1.4257786273956299, "learning_rate": 7.707098006561692e-06, "loss": 0.8425, "step": 2582 }, { "epoch": 0.34, "grad_norm": 1.416089415550232, "learning_rate": 7.70531358151025e-06, "loss": 0.7917, "step": 2583 }, { "epoch": 0.34, "grad_norm": 1.3240020275115967, "learning_rate": 7.703528669133248e-06, "loss": 0.8115, "step": 2584 }, { "epoch": 0.34, "grad_norm": 1.3279213905334473, "learning_rate": 7.701743269752217e-06, "loss": 0.8161, "step": 2585 }, { "epoch": 0.34, "grad_norm": 1.319289207458496, "learning_rate": 7.699957383688767e-06, "loss": 0.8093, "step": 2586 }, { "epoch": 0.34, "grad_norm": 1.3753303289413452, "learning_rate": 7.698171011264604e-06, "loss": 0.7639, "step": 2587 }, { "epoch": 0.34, "grad_norm": 1.4769556522369385, "learning_rate": 7.696384152801519e-06, "loss": 0.8598, "step": 2588 }, { "epoch": 0.34, "grad_norm": 1.4266202449798584, "learning_rate": 7.694596808621389e-06, "loss": 0.8679, "step": 2589 }, { "epoch": 0.34, "grad_norm": 0.9061298370361328, "learning_rate": 7.692808979046182e-06, "loss": 1.043, "step": 2590 }, { "epoch": 0.34, "grad_norm": 1.472448468208313, "learning_rate": 7.69102066439795e-06, "loss": 0.8143, "step": 2591 }, { "epoch": 0.34, "grad_norm": 1.525087833404541, "learning_rate": 7.689231864998832e-06, "loss": 0.8561, "step": 2592 }, { "epoch": 0.34, "grad_norm": 0.6360167860984802, "learning_rate": 7.687442581171059e-06, "loss": 1.0355, "step": 2593 }, { "epoch": 0.34, "grad_norm": 0.6319229602813721, "learning_rate": 7.685652813236946e-06, "loss": 1.0345, "step": 2594 }, { "epoch": 0.34, "grad_norm": 1.3747011423110962, "learning_rate": 7.683862561518892e-06, "loss": 0.8662, "step": 2595 }, { "epoch": 0.34, "grad_norm": 0.700996994972229, "learning_rate": 7.68207182633939e-06, "loss": 1.0697, "step": 2596 }, { "epoch": 0.34, "grad_norm": 0.6639552116394043, "learning_rate": 7.680280608021014e-06, "loss": 1.0396, "step": 2597 }, { "epoch": 0.34, "grad_norm": 1.3489950895309448, "learning_rate": 7.67848890688643e-06, "loss": 0.8241, "step": 2598 }, { "epoch": 0.34, "grad_norm": 1.6559405326843262, "learning_rate": 7.676696723258387e-06, "loss": 0.7956, "step": 2599 }, { "epoch": 0.34, "grad_norm": 1.5203449726104736, "learning_rate": 7.674904057459726e-06, "loss": 0.8967, "step": 2600 }, { "epoch": 0.34, "grad_norm": 1.391599178314209, "learning_rate": 7.673110909813366e-06, "loss": 0.8176, "step": 2601 }, { "epoch": 0.34, "grad_norm": 0.9329496026039124, "learning_rate": 7.671317280642323e-06, "loss": 1.0411, "step": 2602 }, { "epoch": 0.34, "grad_norm": 1.2814639806747437, "learning_rate": 7.66952317026969e-06, "loss": 0.8453, "step": 2603 }, { "epoch": 0.34, "grad_norm": 1.2747368812561035, "learning_rate": 7.66772857901866e-06, "loss": 0.7737, "step": 2604 }, { "epoch": 0.34, "grad_norm": 1.3470447063446045, "learning_rate": 7.665933507212495e-06, "loss": 0.8019, "step": 2605 }, { "epoch": 0.34, "grad_norm": 1.3187211751937866, "learning_rate": 7.664137955174558e-06, "loss": 0.8562, "step": 2606 }, { "epoch": 0.34, "grad_norm": 1.4177896976470947, "learning_rate": 7.66234192322829e-06, "loss": 0.8294, "step": 2607 }, { "epoch": 0.34, "grad_norm": 1.3627793788909912, "learning_rate": 7.660545411697227e-06, "loss": 0.8162, "step": 2608 }, { "epoch": 0.34, "grad_norm": 1.4108549356460571, "learning_rate": 7.658748420904983e-06, "loss": 0.9074, "step": 2609 }, { "epoch": 0.34, "grad_norm": 1.2915631532669067, "learning_rate": 7.65695095117526e-06, "loss": 0.8163, "step": 2610 }, { "epoch": 0.34, "grad_norm": 1.266932487487793, "learning_rate": 7.655153002831849e-06, "loss": 0.8552, "step": 2611 }, { "epoch": 0.34, "grad_norm": 1.2981318235397339, "learning_rate": 7.653354576198629e-06, "loss": 0.8469, "step": 2612 }, { "epoch": 0.34, "grad_norm": 1.3572070598602295, "learning_rate": 7.651555671599557e-06, "loss": 0.84, "step": 2613 }, { "epoch": 0.34, "grad_norm": 1.3740835189819336, "learning_rate": 7.649756289358686e-06, "loss": 0.8436, "step": 2614 }, { "epoch": 0.34, "grad_norm": 2.7811224460601807, "learning_rate": 7.647956429800147e-06, "loss": 0.8779, "step": 2615 }, { "epoch": 0.34, "grad_norm": 1.349076271057129, "learning_rate": 7.646156093248162e-06, "loss": 0.8275, "step": 2616 }, { "epoch": 0.34, "grad_norm": 1.3638482093811035, "learning_rate": 7.644355280027037e-06, "loss": 0.8252, "step": 2617 }, { "epoch": 0.34, "grad_norm": 1.3318886756896973, "learning_rate": 7.642553990461165e-06, "loss": 0.7998, "step": 2618 }, { "epoch": 0.34, "grad_norm": 1.3568922281265259, "learning_rate": 7.640752224875022e-06, "loss": 0.9094, "step": 2619 }, { "epoch": 0.34, "grad_norm": 1.3179395198822021, "learning_rate": 7.63894998359317e-06, "loss": 0.8955, "step": 2620 }, { "epoch": 0.34, "grad_norm": 1.0422979593276978, "learning_rate": 7.637147266940264e-06, "loss": 1.0459, "step": 2621 }, { "epoch": 0.34, "grad_norm": 1.543636441230774, "learning_rate": 7.635344075241034e-06, "loss": 0.8588, "step": 2622 }, { "epoch": 0.34, "grad_norm": 1.2906429767608643, "learning_rate": 7.633540408820301e-06, "loss": 0.7853, "step": 2623 }, { "epoch": 0.34, "grad_norm": 1.2571903467178345, "learning_rate": 7.631736268002975e-06, "loss": 0.8612, "step": 2624 }, { "epoch": 0.34, "grad_norm": 0.7457429766654968, "learning_rate": 7.629931653114044e-06, "loss": 1.0633, "step": 2625 }, { "epoch": 0.34, "grad_norm": 1.287019968032837, "learning_rate": 7.628126564478584e-06, "loss": 0.8051, "step": 2626 }, { "epoch": 0.34, "grad_norm": 1.379317045211792, "learning_rate": 7.626321002421761e-06, "loss": 0.8601, "step": 2627 }, { "epoch": 0.34, "grad_norm": 1.328911542892456, "learning_rate": 7.624514967268819e-06, "loss": 0.8517, "step": 2628 }, { "epoch": 0.34, "grad_norm": 1.2855830192565918, "learning_rate": 7.622708459345093e-06, "loss": 0.7739, "step": 2629 }, { "epoch": 0.34, "grad_norm": 1.286238431930542, "learning_rate": 7.6209014789759995e-06, "loss": 0.7949, "step": 2630 }, { "epoch": 0.34, "grad_norm": 1.3222965002059937, "learning_rate": 7.619094026487042e-06, "loss": 0.7957, "step": 2631 }, { "epoch": 0.34, "grad_norm": 1.4939385652542114, "learning_rate": 7.617286102203809e-06, "loss": 0.8095, "step": 2632 }, { "epoch": 0.35, "grad_norm": 1.3079543113708496, "learning_rate": 7.615477706451975e-06, "loss": 0.8504, "step": 2633 }, { "epoch": 0.35, "grad_norm": 0.7333444952964783, "learning_rate": 7.613668839557294e-06, "loss": 1.0245, "step": 2634 }, { "epoch": 0.35, "grad_norm": 1.3699227571487427, "learning_rate": 7.611859501845611e-06, "loss": 0.8699, "step": 2635 }, { "epoch": 0.35, "grad_norm": 1.6639150381088257, "learning_rate": 7.6100496936428545e-06, "loss": 0.8596, "step": 2636 }, { "epoch": 0.35, "grad_norm": 1.3180615901947021, "learning_rate": 7.6082394152750365e-06, "loss": 0.9008, "step": 2637 }, { "epoch": 0.35, "grad_norm": 1.303541660308838, "learning_rate": 7.606428667068253e-06, "loss": 0.8423, "step": 2638 }, { "epoch": 0.35, "grad_norm": 1.4360311031341553, "learning_rate": 7.6046174493486854e-06, "loss": 0.8395, "step": 2639 }, { "epoch": 0.35, "grad_norm": 1.5038973093032837, "learning_rate": 7.6028057624426025e-06, "loss": 0.7763, "step": 2640 }, { "epoch": 0.35, "grad_norm": 1.3912527561187744, "learning_rate": 7.600993606676355e-06, "loss": 0.8268, "step": 2641 }, { "epoch": 0.35, "grad_norm": 1.5755800008773804, "learning_rate": 7.599180982376373e-06, "loss": 0.7886, "step": 2642 }, { "epoch": 0.35, "grad_norm": 0.8078329563140869, "learning_rate": 7.597367889869183e-06, "loss": 1.0436, "step": 2643 }, { "epoch": 0.35, "grad_norm": 1.309241771697998, "learning_rate": 7.595554329481385e-06, "loss": 0.7618, "step": 2644 }, { "epoch": 0.35, "grad_norm": 1.2907609939575195, "learning_rate": 7.593740301539667e-06, "loss": 0.8661, "step": 2645 }, { "epoch": 0.35, "grad_norm": 1.3790518045425415, "learning_rate": 7.5919258063708055e-06, "loss": 0.801, "step": 2646 }, { "epoch": 0.35, "grad_norm": 1.3473786115646362, "learning_rate": 7.590110844301655e-06, "loss": 0.8892, "step": 2647 }, { "epoch": 0.35, "grad_norm": 0.7707178592681885, "learning_rate": 7.588295415659154e-06, "loss": 1.0324, "step": 2648 }, { "epoch": 0.35, "grad_norm": 1.3513009548187256, "learning_rate": 7.586479520770329e-06, "loss": 0.813, "step": 2649 }, { "epoch": 0.35, "grad_norm": 1.481490969657898, "learning_rate": 7.58466315996229e-06, "loss": 0.8548, "step": 2650 }, { "epoch": 0.35, "grad_norm": 1.5003479719161987, "learning_rate": 7.582846333562231e-06, "loss": 0.8609, "step": 2651 }, { "epoch": 0.35, "grad_norm": 1.7983958721160889, "learning_rate": 7.581029041897423e-06, "loss": 0.8038, "step": 2652 }, { "epoch": 0.35, "grad_norm": 1.3376060724258423, "learning_rate": 7.579211285295231e-06, "loss": 0.7971, "step": 2653 }, { "epoch": 0.35, "grad_norm": 0.8038067817687988, "learning_rate": 7.577393064083101e-06, "loss": 1.0221, "step": 2654 }, { "epoch": 0.35, "grad_norm": 0.6759063005447388, "learning_rate": 7.575574378588555e-06, "loss": 1.0354, "step": 2655 }, { "epoch": 0.35, "grad_norm": 1.3555853366851807, "learning_rate": 7.573755229139208e-06, "loss": 0.8215, "step": 2656 }, { "epoch": 0.35, "grad_norm": 1.4031643867492676, "learning_rate": 7.571935616062756e-06, "loss": 0.8923, "step": 2657 }, { "epoch": 0.35, "grad_norm": 0.7329822778701782, "learning_rate": 7.570115539686975e-06, "loss": 1.0713, "step": 2658 }, { "epoch": 0.35, "grad_norm": 1.360599160194397, "learning_rate": 7.568295000339729e-06, "loss": 0.8506, "step": 2659 }, { "epoch": 0.35, "grad_norm": 1.3883612155914307, "learning_rate": 7.566473998348963e-06, "loss": 0.8337, "step": 2660 }, { "epoch": 0.35, "grad_norm": 0.6834741234779358, "learning_rate": 7.5646525340427054e-06, "loss": 1.0582, "step": 2661 }, { "epoch": 0.35, "grad_norm": 1.303406000137329, "learning_rate": 7.56283060774907e-06, "loss": 0.8285, "step": 2662 }, { "epoch": 0.35, "grad_norm": 0.6577319502830505, "learning_rate": 7.561008219796249e-06, "loss": 1.0448, "step": 2663 }, { "epoch": 0.35, "grad_norm": 1.5726022720336914, "learning_rate": 7.559185370512523e-06, "loss": 0.8344, "step": 2664 }, { "epoch": 0.35, "grad_norm": 1.4480431079864502, "learning_rate": 7.557362060226254e-06, "loss": 0.8725, "step": 2665 }, { "epoch": 0.35, "grad_norm": 1.4089685678482056, "learning_rate": 7.555538289265885e-06, "loss": 0.8498, "step": 2666 }, { "epoch": 0.35, "grad_norm": 0.6999285221099854, "learning_rate": 7.553714057959944e-06, "loss": 1.057, "step": 2667 }, { "epoch": 0.35, "grad_norm": 1.3200336694717407, "learning_rate": 7.551889366637043e-06, "loss": 0.8275, "step": 2668 }, { "epoch": 0.35, "grad_norm": 1.275614857673645, "learning_rate": 7.550064215625874e-06, "loss": 0.8782, "step": 2669 }, { "epoch": 0.35, "grad_norm": 1.2901808023452759, "learning_rate": 7.548238605255215e-06, "loss": 0.8463, "step": 2670 }, { "epoch": 0.35, "grad_norm": 1.3514444828033447, "learning_rate": 7.5464125358539216e-06, "loss": 0.8385, "step": 2671 }, { "epoch": 0.35, "grad_norm": 1.3909103870391846, "learning_rate": 7.544586007750938e-06, "loss": 0.8273, "step": 2672 }, { "epoch": 0.35, "grad_norm": 1.3866055011749268, "learning_rate": 7.542759021275289e-06, "loss": 0.8385, "step": 2673 }, { "epoch": 0.35, "grad_norm": 1.3378056287765503, "learning_rate": 7.540931576756079e-06, "loss": 0.8536, "step": 2674 }, { "epoch": 0.35, "grad_norm": 1.7419037818908691, "learning_rate": 7.539103674522499e-06, "loss": 0.8683, "step": 2675 }, { "epoch": 0.35, "grad_norm": 0.7728086113929749, "learning_rate": 7.53727531490382e-06, "loss": 1.0355, "step": 2676 }, { "epoch": 0.35, "grad_norm": 0.702224850654602, "learning_rate": 7.5354464982293975e-06, "loss": 1.0552, "step": 2677 }, { "epoch": 0.35, "grad_norm": 1.3621621131896973, "learning_rate": 7.533617224828667e-06, "loss": 0.8484, "step": 2678 }, { "epoch": 0.35, "grad_norm": 1.3960362672805786, "learning_rate": 7.531787495031148e-06, "loss": 0.8317, "step": 2679 }, { "epoch": 0.35, "grad_norm": 1.3486826419830322, "learning_rate": 7.52995730916644e-06, "loss": 0.8625, "step": 2680 }, { "epoch": 0.35, "grad_norm": 1.3445136547088623, "learning_rate": 7.528126667564226e-06, "loss": 0.837, "step": 2681 }, { "epoch": 0.35, "grad_norm": 1.4311063289642334, "learning_rate": 7.526295570554274e-06, "loss": 0.7936, "step": 2682 }, { "epoch": 0.35, "grad_norm": 0.9335439205169678, "learning_rate": 7.524464018466428e-06, "loss": 1.0687, "step": 2683 }, { "epoch": 0.35, "grad_norm": 1.395727515220642, "learning_rate": 7.522632011630622e-06, "loss": 0.8143, "step": 2684 }, { "epoch": 0.35, "grad_norm": 0.704051673412323, "learning_rate": 7.520799550376863e-06, "loss": 1.0209, "step": 2685 }, { "epoch": 0.35, "grad_norm": 1.2879384756088257, "learning_rate": 7.518966635035246e-06, "loss": 0.8041, "step": 2686 }, { "epoch": 0.35, "grad_norm": 1.4573758840560913, "learning_rate": 7.517133265935943e-06, "loss": 0.8255, "step": 2687 }, { "epoch": 0.35, "grad_norm": 1.3225946426391602, "learning_rate": 7.515299443409214e-06, "loss": 0.7815, "step": 2688 }, { "epoch": 0.35, "grad_norm": 1.3844631910324097, "learning_rate": 7.513465167785394e-06, "loss": 0.8741, "step": 2689 }, { "epoch": 0.35, "grad_norm": 1.0207690000534058, "learning_rate": 7.511630439394908e-06, "loss": 1.0507, "step": 2690 }, { "epoch": 0.35, "grad_norm": 1.364751935005188, "learning_rate": 7.509795258568251e-06, "loss": 0.8603, "step": 2691 }, { "epoch": 0.35, "grad_norm": 0.7445399165153503, "learning_rate": 7.507959625636012e-06, "loss": 1.0524, "step": 2692 }, { "epoch": 0.35, "grad_norm": 2.1156532764434814, "learning_rate": 7.5061235409288515e-06, "loss": 0.8531, "step": 2693 }, { "epoch": 0.35, "grad_norm": 1.4447886943817139, "learning_rate": 7.504287004777517e-06, "loss": 0.8718, "step": 2694 }, { "epoch": 0.35, "grad_norm": 0.8812224864959717, "learning_rate": 7.5024500175128336e-06, "loss": 1.0613, "step": 2695 }, { "epoch": 0.35, "grad_norm": 1.3136683702468872, "learning_rate": 7.500612579465711e-06, "loss": 0.7965, "step": 2696 }, { "epoch": 0.35, "grad_norm": 1.3073822259902954, "learning_rate": 7.498774690967139e-06, "loss": 0.868, "step": 2697 }, { "epoch": 0.35, "grad_norm": 1.3610388040542603, "learning_rate": 7.49693635234819e-06, "loss": 0.7931, "step": 2698 }, { "epoch": 0.35, "grad_norm": 1.3401392698287964, "learning_rate": 7.49509756394001e-06, "loss": 0.8677, "step": 2699 }, { "epoch": 0.35, "grad_norm": 0.7749993205070496, "learning_rate": 7.493258326073838e-06, "loss": 1.0399, "step": 2700 }, { "epoch": 0.35, "grad_norm": 0.6728084683418274, "learning_rate": 7.491418639080985e-06, "loss": 1.0634, "step": 2701 }, { "epoch": 0.35, "grad_norm": 1.5497844219207764, "learning_rate": 7.489578503292847e-06, "loss": 0.8551, "step": 2702 }, { "epoch": 0.35, "grad_norm": 1.4256402254104614, "learning_rate": 7.487737919040897e-06, "loss": 0.8917, "step": 2703 }, { "epoch": 0.35, "grad_norm": 1.5041227340698242, "learning_rate": 7.4858968866566925e-06, "loss": 0.8418, "step": 2704 }, { "epoch": 0.35, "grad_norm": 1.501463770866394, "learning_rate": 7.484055406471872e-06, "loss": 0.88, "step": 2705 }, { "epoch": 0.35, "grad_norm": 1.365180253982544, "learning_rate": 7.482213478818151e-06, "loss": 0.8361, "step": 2706 }, { "epoch": 0.35, "grad_norm": 1.2665188312530518, "learning_rate": 7.480371104027331e-06, "loss": 0.8043, "step": 2707 }, { "epoch": 0.35, "grad_norm": 0.9927685856819153, "learning_rate": 7.478528282431287e-06, "loss": 1.0531, "step": 2708 }, { "epoch": 0.35, "grad_norm": 1.3721137046813965, "learning_rate": 7.4766850143619795e-06, "loss": 0.8377, "step": 2709 }, { "epoch": 0.36, "grad_norm": 1.3489160537719727, "learning_rate": 7.47484130015145e-06, "loss": 0.812, "step": 2710 }, { "epoch": 0.36, "grad_norm": 1.3632639646530151, "learning_rate": 7.472997140131816e-06, "loss": 0.8049, "step": 2711 }, { "epoch": 0.36, "grad_norm": 1.4046945571899414, "learning_rate": 7.471152534635281e-06, "loss": 0.7822, "step": 2712 }, { "epoch": 0.36, "grad_norm": 1.5674452781677246, "learning_rate": 7.469307483994122e-06, "loss": 0.8785, "step": 2713 }, { "epoch": 0.36, "grad_norm": 1.5071920156478882, "learning_rate": 7.467461988540704e-06, "loss": 0.845, "step": 2714 }, { "epoch": 0.36, "grad_norm": 0.795006513595581, "learning_rate": 7.465616048607464e-06, "loss": 1.0419, "step": 2715 }, { "epoch": 0.36, "grad_norm": 1.392560362815857, "learning_rate": 7.463769664526926e-06, "loss": 0.8714, "step": 2716 }, { "epoch": 0.36, "grad_norm": 0.648163914680481, "learning_rate": 7.461922836631688e-06, "loss": 1.0589, "step": 2717 }, { "epoch": 0.36, "grad_norm": 1.7239807844161987, "learning_rate": 7.4600755652544344e-06, "loss": 0.8256, "step": 2718 }, { "epoch": 0.36, "grad_norm": 1.4367743730545044, "learning_rate": 7.458227850727925e-06, "loss": 0.8688, "step": 2719 }, { "epoch": 0.36, "grad_norm": 1.4681731462478638, "learning_rate": 7.456379693384999e-06, "loss": 0.8432, "step": 2720 }, { "epoch": 0.36, "grad_norm": 0.7628823518753052, "learning_rate": 7.454531093558579e-06, "loss": 1.0589, "step": 2721 }, { "epoch": 0.36, "grad_norm": 1.4833554029464722, "learning_rate": 7.4526820515816635e-06, "loss": 0.8165, "step": 2722 }, { "epoch": 0.36, "grad_norm": 1.3233978748321533, "learning_rate": 7.450832567787333e-06, "loss": 0.7901, "step": 2723 }, { "epoch": 0.36, "grad_norm": 1.4906229972839355, "learning_rate": 7.448982642508746e-06, "loss": 0.8309, "step": 2724 }, { "epoch": 0.36, "grad_norm": 0.6369411945343018, "learning_rate": 7.447132276079143e-06, "loss": 1.0315, "step": 2725 }, { "epoch": 0.36, "grad_norm": 1.408367395401001, "learning_rate": 7.445281468831843e-06, "loss": 0.8932, "step": 2726 }, { "epoch": 0.36, "grad_norm": 1.630934715270996, "learning_rate": 7.443430221100242e-06, "loss": 0.7788, "step": 2727 }, { "epoch": 0.36, "grad_norm": 0.6265681982040405, "learning_rate": 7.441578533217817e-06, "loss": 1.0323, "step": 2728 }, { "epoch": 0.36, "grad_norm": 1.303134560585022, "learning_rate": 7.439726405518126e-06, "loss": 0.8377, "step": 2729 }, { "epoch": 0.36, "grad_norm": 0.6204424500465393, "learning_rate": 7.437873838334805e-06, "loss": 1.0496, "step": 2730 }, { "epoch": 0.36, "grad_norm": 1.2800440788269043, "learning_rate": 7.436020832001567e-06, "loss": 0.8042, "step": 2731 }, { "epoch": 0.36, "grad_norm": 1.3000479936599731, "learning_rate": 7.434167386852206e-06, "loss": 0.8679, "step": 2732 }, { "epoch": 0.36, "grad_norm": 1.39506995677948, "learning_rate": 7.432313503220597e-06, "loss": 0.8199, "step": 2733 }, { "epoch": 0.36, "grad_norm": 0.677017331123352, "learning_rate": 7.43045918144069e-06, "loss": 1.0158, "step": 2734 }, { "epoch": 0.36, "grad_norm": 1.243342399597168, "learning_rate": 7.428604421846517e-06, "loss": 0.8575, "step": 2735 }, { "epoch": 0.36, "grad_norm": 0.6713952422142029, "learning_rate": 7.426749224772188e-06, "loss": 1.0359, "step": 2736 }, { "epoch": 0.36, "grad_norm": 1.2251205444335938, "learning_rate": 7.424893590551891e-06, "loss": 0.7864, "step": 2737 }, { "epoch": 0.36, "grad_norm": 1.3584080934524536, "learning_rate": 7.423037519519892e-06, "loss": 0.8145, "step": 2738 }, { "epoch": 0.36, "grad_norm": 1.4708906412124634, "learning_rate": 7.42118101201054e-06, "loss": 0.861, "step": 2739 }, { "epoch": 0.36, "grad_norm": 1.269092082977295, "learning_rate": 7.4193240683582576e-06, "loss": 0.7534, "step": 2740 }, { "epoch": 0.36, "grad_norm": 0.692176342010498, "learning_rate": 7.417466688897549e-06, "loss": 1.0335, "step": 2741 }, { "epoch": 0.36, "grad_norm": 0.638009786605835, "learning_rate": 7.415608873962995e-06, "loss": 1.0242, "step": 2742 }, { "epoch": 0.36, "grad_norm": 0.6258887648582458, "learning_rate": 7.413750623889257e-06, "loss": 1.055, "step": 2743 }, { "epoch": 0.36, "grad_norm": 1.6199005842208862, "learning_rate": 7.411891939011073e-06, "loss": 0.8683, "step": 2744 }, { "epoch": 0.36, "grad_norm": 1.3712468147277832, "learning_rate": 7.410032819663259e-06, "loss": 0.8214, "step": 2745 }, { "epoch": 0.36, "grad_norm": 1.3842699527740479, "learning_rate": 7.40817326618071e-06, "loss": 0.8476, "step": 2746 }, { "epoch": 0.36, "grad_norm": 0.7986129522323608, "learning_rate": 7.406313278898402e-06, "loss": 1.0366, "step": 2747 }, { "epoch": 0.36, "grad_norm": 1.3226032257080078, "learning_rate": 7.404452858151383e-06, "loss": 0.7997, "step": 2748 }, { "epoch": 0.36, "grad_norm": 1.401343822479248, "learning_rate": 7.402592004274785e-06, "loss": 0.8398, "step": 2749 }, { "epoch": 0.36, "grad_norm": 1.2829153537750244, "learning_rate": 7.4007307176038156e-06, "loss": 0.8685, "step": 2750 }, { "epoch": 0.36, "grad_norm": 0.6376968622207642, "learning_rate": 7.398868998473758e-06, "loss": 1.0287, "step": 2751 }, { "epoch": 0.36, "grad_norm": 0.6610245704650879, "learning_rate": 7.397006847219978e-06, "loss": 1.065, "step": 2752 }, { "epoch": 0.36, "grad_norm": 0.6444390416145325, "learning_rate": 7.395144264177918e-06, "loss": 1.0546, "step": 2753 }, { "epoch": 0.36, "grad_norm": 1.6282514333724976, "learning_rate": 7.393281249683093e-06, "loss": 0.872, "step": 2754 }, { "epoch": 0.36, "grad_norm": 1.3328979015350342, "learning_rate": 7.391417804071104e-06, "loss": 0.8819, "step": 2755 }, { "epoch": 0.36, "grad_norm": 1.3334386348724365, "learning_rate": 7.389553927677622e-06, "loss": 0.8215, "step": 2756 }, { "epoch": 0.36, "grad_norm": 1.363668441772461, "learning_rate": 7.3876896208384e-06, "loss": 0.8067, "step": 2757 }, { "epoch": 0.36, "grad_norm": 1.3014483451843262, "learning_rate": 7.38582488388927e-06, "loss": 0.8023, "step": 2758 }, { "epoch": 0.36, "grad_norm": 1.3989754915237427, "learning_rate": 7.383959717166136e-06, "loss": 0.8345, "step": 2759 }, { "epoch": 0.36, "grad_norm": 1.3744498491287231, "learning_rate": 7.382094121004983e-06, "loss": 0.8654, "step": 2760 }, { "epoch": 0.36, "grad_norm": 1.3863519430160522, "learning_rate": 7.380228095741875e-06, "loss": 0.8223, "step": 2761 }, { "epoch": 0.36, "grad_norm": 1.2905850410461426, "learning_rate": 7.378361641712948e-06, "loss": 0.8328, "step": 2762 }, { "epoch": 0.36, "grad_norm": 1.3353687524795532, "learning_rate": 7.376494759254422e-06, "loss": 0.7968, "step": 2763 }, { "epoch": 0.36, "grad_norm": 1.3495123386383057, "learning_rate": 7.3746274487025885e-06, "loss": 0.8306, "step": 2764 }, { "epoch": 0.36, "grad_norm": 1.3722673654556274, "learning_rate": 7.3727597103938175e-06, "loss": 0.8418, "step": 2765 }, { "epoch": 0.36, "grad_norm": 1.3599368333816528, "learning_rate": 7.370891544664558e-06, "loss": 0.815, "step": 2766 }, { "epoch": 0.36, "grad_norm": 1.3626376390457153, "learning_rate": 7.369022951851333e-06, "loss": 0.8181, "step": 2767 }, { "epoch": 0.36, "grad_norm": 1.2919408082962036, "learning_rate": 7.367153932290748e-06, "loss": 0.7847, "step": 2768 }, { "epoch": 0.36, "grad_norm": 1.4622271060943604, "learning_rate": 7.365284486319477e-06, "loss": 0.9503, "step": 2769 }, { "epoch": 0.36, "grad_norm": 0.8147485256195068, "learning_rate": 7.363414614274278e-06, "loss": 1.0301, "step": 2770 }, { "epoch": 0.36, "grad_norm": 1.5441794395446777, "learning_rate": 7.361544316491982e-06, "loss": 0.8128, "step": 2771 }, { "epoch": 0.36, "grad_norm": 0.6732627153396606, "learning_rate": 7.3596735933095e-06, "loss": 1.0498, "step": 2772 }, { "epoch": 0.36, "grad_norm": 1.35495126247406, "learning_rate": 7.357802445063814e-06, "loss": 0.7888, "step": 2773 }, { "epoch": 0.36, "grad_norm": 1.4058078527450562, "learning_rate": 7.35593087209199e-06, "loss": 0.7823, "step": 2774 }, { "epoch": 0.36, "grad_norm": 1.3937978744506836, "learning_rate": 7.354058874731161e-06, "loss": 0.8041, "step": 2775 }, { "epoch": 0.36, "grad_norm": 1.3081984519958496, "learning_rate": 7.352186453318546e-06, "loss": 0.8123, "step": 2776 }, { "epoch": 0.36, "grad_norm": 0.8005050420761108, "learning_rate": 7.350313608191436e-06, "loss": 1.0536, "step": 2777 }, { "epoch": 0.36, "grad_norm": 1.3855814933776855, "learning_rate": 7.3484403396871975e-06, "loss": 0.8402, "step": 2778 }, { "epoch": 0.36, "grad_norm": 1.4566665887832642, "learning_rate": 7.346566648143275e-06, "loss": 0.8447, "step": 2779 }, { "epoch": 0.36, "grad_norm": 1.3285702466964722, "learning_rate": 7.344692533897188e-06, "loss": 0.8097, "step": 2780 }, { "epoch": 0.36, "grad_norm": 1.3451201915740967, "learning_rate": 7.3428179972865345e-06, "loss": 0.8238, "step": 2781 }, { "epoch": 0.36, "grad_norm": 1.276929497718811, "learning_rate": 7.340943038648986e-06, "loss": 0.8273, "step": 2782 }, { "epoch": 0.36, "grad_norm": 1.3052208423614502, "learning_rate": 7.33906765832229e-06, "loss": 0.8341, "step": 2783 }, { "epoch": 0.36, "grad_norm": 1.3123648166656494, "learning_rate": 7.33719185664427e-06, "loss": 0.847, "step": 2784 }, { "epoch": 0.36, "grad_norm": 0.8734331130981445, "learning_rate": 7.33531563395283e-06, "loss": 1.0533, "step": 2785 }, { "epoch": 0.37, "grad_norm": 1.3579270839691162, "learning_rate": 7.333438990585942e-06, "loss": 0.8425, "step": 2786 }, { "epoch": 0.37, "grad_norm": 1.3903274536132812, "learning_rate": 7.331561926881661e-06, "loss": 0.8151, "step": 2787 }, { "epoch": 0.37, "grad_norm": 1.3690022230148315, "learning_rate": 7.329684443178112e-06, "loss": 0.8479, "step": 2788 }, { "epoch": 0.37, "grad_norm": 1.4144541025161743, "learning_rate": 7.3278065398135e-06, "loss": 0.8688, "step": 2789 }, { "epoch": 0.37, "grad_norm": 1.2832231521606445, "learning_rate": 7.3259282171261035e-06, "loss": 0.7404, "step": 2790 }, { "epoch": 0.37, "grad_norm": 0.7724542617797852, "learning_rate": 7.324049475454277e-06, "loss": 1.0456, "step": 2791 }, { "epoch": 0.37, "grad_norm": 1.307246446609497, "learning_rate": 7.3221703151364485e-06, "loss": 0.8326, "step": 2792 }, { "epoch": 0.37, "grad_norm": 1.3806105852127075, "learning_rate": 7.320290736511125e-06, "loss": 0.797, "step": 2793 }, { "epoch": 0.37, "grad_norm": 1.3191146850585938, "learning_rate": 7.318410739916888e-06, "loss": 0.789, "step": 2794 }, { "epoch": 0.37, "grad_norm": 1.326403021812439, "learning_rate": 7.316530325692392e-06, "loss": 0.8803, "step": 2795 }, { "epoch": 0.37, "grad_norm": 1.3882569074630737, "learning_rate": 7.314649494176368e-06, "loss": 0.8434, "step": 2796 }, { "epoch": 0.37, "grad_norm": 0.7490984797477722, "learning_rate": 7.312768245707622e-06, "loss": 1.0381, "step": 2797 }, { "epoch": 0.37, "grad_norm": 1.3228522539138794, "learning_rate": 7.310886580625036e-06, "loss": 0.8165, "step": 2798 }, { "epoch": 0.37, "grad_norm": 1.3550381660461426, "learning_rate": 7.309004499267567e-06, "loss": 0.8317, "step": 2799 }, { "epoch": 0.37, "grad_norm": 1.393107533454895, "learning_rate": 7.307122001974245e-06, "loss": 0.7497, "step": 2800 }, { "epoch": 0.37, "grad_norm": 1.3795772790908813, "learning_rate": 7.305239089084179e-06, "loss": 0.8295, "step": 2801 }, { "epoch": 0.37, "grad_norm": 0.6593143343925476, "learning_rate": 7.303355760936547e-06, "loss": 1.0686, "step": 2802 }, { "epoch": 0.37, "grad_norm": 0.6243030428886414, "learning_rate": 7.301472017870607e-06, "loss": 1.0404, "step": 2803 }, { "epoch": 0.37, "grad_norm": 1.336880087852478, "learning_rate": 7.2995878602256885e-06, "loss": 0.7696, "step": 2804 }, { "epoch": 0.37, "grad_norm": 1.76347815990448, "learning_rate": 7.297703288341199e-06, "loss": 0.7581, "step": 2805 }, { "epoch": 0.37, "grad_norm": 0.6215465068817139, "learning_rate": 7.295818302556617e-06, "loss": 1.0612, "step": 2806 }, { "epoch": 0.37, "grad_norm": 1.3000422716140747, "learning_rate": 7.293932903211497e-06, "loss": 0.7935, "step": 2807 }, { "epoch": 0.37, "grad_norm": 1.430633544921875, "learning_rate": 7.292047090645469e-06, "loss": 0.8254, "step": 2808 }, { "epoch": 0.37, "grad_norm": 1.375624656677246, "learning_rate": 7.290160865198234e-06, "loss": 0.7788, "step": 2809 }, { "epoch": 0.37, "grad_norm": 0.6632499694824219, "learning_rate": 7.288274227209575e-06, "loss": 1.0368, "step": 2810 }, { "epoch": 0.37, "grad_norm": 0.6383571028709412, "learning_rate": 7.286387177019341e-06, "loss": 1.0577, "step": 2811 }, { "epoch": 0.37, "grad_norm": 1.2919286489486694, "learning_rate": 7.284499714967458e-06, "loss": 0.8634, "step": 2812 }, { "epoch": 0.37, "grad_norm": 1.4843512773513794, "learning_rate": 7.2826118413939275e-06, "loss": 0.8316, "step": 2813 }, { "epoch": 0.37, "grad_norm": 1.435305118560791, "learning_rate": 7.2807235566388245e-06, "loss": 0.8703, "step": 2814 }, { "epoch": 0.37, "grad_norm": 0.695947527885437, "learning_rate": 7.278834861042299e-06, "loss": 1.0473, "step": 2815 }, { "epoch": 0.37, "grad_norm": 1.239074468612671, "learning_rate": 7.276945754944571e-06, "loss": 0.8316, "step": 2816 }, { "epoch": 0.37, "grad_norm": 1.6823779344558716, "learning_rate": 7.275056238685939e-06, "loss": 0.8713, "step": 2817 }, { "epoch": 0.37, "grad_norm": 1.3182501792907715, "learning_rate": 7.2731663126067745e-06, "loss": 0.8227, "step": 2818 }, { "epoch": 0.37, "grad_norm": 1.3072055578231812, "learning_rate": 7.271275977047521e-06, "loss": 0.8207, "step": 2819 }, { "epoch": 0.37, "grad_norm": 1.3331211805343628, "learning_rate": 7.269385232348697e-06, "loss": 0.8536, "step": 2820 }, { "epoch": 0.37, "grad_norm": 1.3581324815750122, "learning_rate": 7.267494078850894e-06, "loss": 0.8781, "step": 2821 }, { "epoch": 0.37, "grad_norm": 1.2513459920883179, "learning_rate": 7.265602516894778e-06, "loss": 0.8376, "step": 2822 }, { "epoch": 0.37, "grad_norm": 1.347963809967041, "learning_rate": 7.26371054682109e-06, "loss": 0.8077, "step": 2823 }, { "epoch": 0.37, "grad_norm": 0.6693441271781921, "learning_rate": 7.261818168970642e-06, "loss": 1.0374, "step": 2824 }, { "epoch": 0.37, "grad_norm": 0.6352295279502869, "learning_rate": 7.259925383684318e-06, "loss": 1.0285, "step": 2825 }, { "epoch": 0.37, "grad_norm": 1.3348710536956787, "learning_rate": 7.25803219130308e-06, "loss": 0.8144, "step": 2826 }, { "epoch": 0.37, "grad_norm": 0.6278809309005737, "learning_rate": 7.256138592167958e-06, "loss": 1.0303, "step": 2827 }, { "epoch": 0.37, "grad_norm": 1.3238322734832764, "learning_rate": 7.254244586620061e-06, "loss": 0.7702, "step": 2828 }, { "epoch": 0.37, "grad_norm": 1.3484621047973633, "learning_rate": 7.252350175000569e-06, "loss": 0.8369, "step": 2829 }, { "epoch": 0.37, "grad_norm": 1.3334599733352661, "learning_rate": 7.250455357650732e-06, "loss": 0.8369, "step": 2830 }, { "epoch": 0.37, "grad_norm": 1.3309293985366821, "learning_rate": 7.2485601349118775e-06, "loss": 0.8087, "step": 2831 }, { "epoch": 0.37, "grad_norm": 1.3438478708267212, "learning_rate": 7.246664507125403e-06, "loss": 0.8405, "step": 2832 }, { "epoch": 0.37, "grad_norm": 1.932469129562378, "learning_rate": 7.244768474632782e-06, "loss": 0.8191, "step": 2833 }, { "epoch": 0.37, "grad_norm": 1.3696272373199463, "learning_rate": 7.242872037775557e-06, "loss": 0.8133, "step": 2834 }, { "epoch": 0.37, "grad_norm": 1.314755916595459, "learning_rate": 7.240975196895346e-06, "loss": 0.7961, "step": 2835 }, { "epoch": 0.37, "grad_norm": 1.3923640251159668, "learning_rate": 7.23907795233384e-06, "loss": 0.8647, "step": 2836 }, { "epoch": 0.37, "grad_norm": 1.32186758518219, "learning_rate": 7.2371803044328e-06, "loss": 0.9192, "step": 2837 }, { "epoch": 0.37, "grad_norm": 1.3223069906234741, "learning_rate": 7.2352822535340635e-06, "loss": 0.8095, "step": 2838 }, { "epoch": 0.37, "grad_norm": 0.8231055736541748, "learning_rate": 7.233383799979537e-06, "loss": 1.0357, "step": 2839 }, { "epoch": 0.37, "grad_norm": 0.7594298720359802, "learning_rate": 7.2314849441112024e-06, "loss": 1.0328, "step": 2840 }, { "epoch": 0.37, "grad_norm": 1.2191270589828491, "learning_rate": 7.2295856862711125e-06, "loss": 0.8, "step": 2841 }, { "epoch": 0.37, "grad_norm": 1.313835620880127, "learning_rate": 7.2276860268013925e-06, "loss": 0.7924, "step": 2842 }, { "epoch": 0.37, "grad_norm": 0.7656760811805725, "learning_rate": 7.225785966044242e-06, "loss": 1.0297, "step": 2843 }, { "epoch": 0.37, "grad_norm": 1.4086486101150513, "learning_rate": 7.223885504341927e-06, "loss": 0.7567, "step": 2844 }, { "epoch": 0.37, "grad_norm": 0.7645559906959534, "learning_rate": 7.2219846420367956e-06, "loss": 1.0339, "step": 2845 }, { "epoch": 0.37, "grad_norm": 0.6779087781906128, "learning_rate": 7.2200833794712574e-06, "loss": 1.0365, "step": 2846 }, { "epoch": 0.37, "grad_norm": 0.6342766880989075, "learning_rate": 7.218181716987803e-06, "loss": 1.0603, "step": 2847 }, { "epoch": 0.37, "grad_norm": 1.3729419708251953, "learning_rate": 7.216279654928988e-06, "loss": 0.8702, "step": 2848 }, { "epoch": 0.37, "grad_norm": 0.736022412776947, "learning_rate": 7.2143771936374445e-06, "loss": 1.046, "step": 2849 }, { "epoch": 0.37, "grad_norm": 1.414380431175232, "learning_rate": 7.212474333455875e-06, "loss": 0.8795, "step": 2850 }, { "epoch": 0.37, "grad_norm": 1.3410040140151978, "learning_rate": 7.210571074727053e-06, "loss": 0.8631, "step": 2851 }, { "epoch": 0.37, "grad_norm": 1.3118903636932373, "learning_rate": 7.208667417793827e-06, "loss": 0.8051, "step": 2852 }, { "epoch": 0.37, "grad_norm": 1.4190006256103516, "learning_rate": 7.206763362999114e-06, "loss": 0.7845, "step": 2853 }, { "epoch": 0.37, "grad_norm": 0.824127197265625, "learning_rate": 7.204858910685902e-06, "loss": 1.0286, "step": 2854 }, { "epoch": 0.37, "grad_norm": 1.4371787309646606, "learning_rate": 7.202954061197254e-06, "loss": 0.8546, "step": 2855 }, { "epoch": 0.37, "grad_norm": 0.7142112851142883, "learning_rate": 7.201048814876303e-06, "loss": 1.0221, "step": 2856 }, { "epoch": 0.37, "grad_norm": 0.6517556309700012, "learning_rate": 7.199143172066252e-06, "loss": 1.0614, "step": 2857 }, { "epoch": 0.37, "grad_norm": 1.29354989528656, "learning_rate": 7.197237133110377e-06, "loss": 0.7661, "step": 2858 }, { "epoch": 0.37, "grad_norm": 2.194183826446533, "learning_rate": 7.195330698352026e-06, "loss": 0.8622, "step": 2859 }, { "epoch": 0.37, "grad_norm": 1.3205757141113281, "learning_rate": 7.193423868134616e-06, "loss": 0.7943, "step": 2860 }, { "epoch": 0.37, "grad_norm": 0.8551307916641235, "learning_rate": 7.191516642801639e-06, "loss": 1.0613, "step": 2861 }, { "epoch": 0.38, "grad_norm": 1.3765863180160522, "learning_rate": 7.189609022696653e-06, "loss": 0.7926, "step": 2862 }, { "epoch": 0.38, "grad_norm": 1.3860585689544678, "learning_rate": 7.18770100816329e-06, "loss": 0.8177, "step": 2863 }, { "epoch": 0.38, "grad_norm": 1.4710423946380615, "learning_rate": 7.185792599545256e-06, "loss": 0.8828, "step": 2864 }, { "epoch": 0.38, "grad_norm": 0.7440875172615051, "learning_rate": 7.183883797186322e-06, "loss": 1.0566, "step": 2865 }, { "epoch": 0.38, "grad_norm": 1.3633331060409546, "learning_rate": 7.181974601430335e-06, "loss": 0.8081, "step": 2866 }, { "epoch": 0.38, "grad_norm": 1.3743667602539062, "learning_rate": 7.180065012621208e-06, "loss": 0.8215, "step": 2867 }, { "epoch": 0.38, "grad_norm": 0.6479548811912537, "learning_rate": 7.178155031102929e-06, "loss": 1.0415, "step": 2868 }, { "epoch": 0.38, "grad_norm": 1.3852967023849487, "learning_rate": 7.176244657219556e-06, "loss": 0.8671, "step": 2869 }, { "epoch": 0.38, "grad_norm": 1.6163464784622192, "learning_rate": 7.174333891315216e-06, "loss": 0.8291, "step": 2870 }, { "epoch": 0.38, "grad_norm": 0.6778782606124878, "learning_rate": 7.172422733734109e-06, "loss": 1.027, "step": 2871 }, { "epoch": 0.38, "grad_norm": 1.501255989074707, "learning_rate": 7.170511184820502e-06, "loss": 0.8264, "step": 2872 }, { "epoch": 0.38, "grad_norm": 0.6319093704223633, "learning_rate": 7.168599244918736e-06, "loss": 1.0415, "step": 2873 }, { "epoch": 0.38, "grad_norm": 1.84703528881073, "learning_rate": 7.166686914373221e-06, "loss": 0.8347, "step": 2874 }, { "epoch": 0.38, "grad_norm": 1.3253200054168701, "learning_rate": 7.164774193528436e-06, "loss": 0.8896, "step": 2875 }, { "epoch": 0.38, "grad_norm": 0.6666940450668335, "learning_rate": 7.162861082728934e-06, "loss": 1.0402, "step": 2876 }, { "epoch": 0.38, "grad_norm": 1.4653780460357666, "learning_rate": 7.160947582319335e-06, "loss": 0.8697, "step": 2877 }, { "epoch": 0.38, "grad_norm": 1.6360176801681519, "learning_rate": 7.15903369264433e-06, "loss": 0.7628, "step": 2878 }, { "epoch": 0.38, "grad_norm": 1.4585880041122437, "learning_rate": 7.157119414048679e-06, "loss": 0.8419, "step": 2879 }, { "epoch": 0.38, "grad_norm": 0.6804687976837158, "learning_rate": 7.155204746877214e-06, "loss": 1.04, "step": 2880 }, { "epoch": 0.38, "grad_norm": 0.6691820025444031, "learning_rate": 7.153289691474839e-06, "loss": 1.0294, "step": 2881 }, { "epoch": 0.38, "grad_norm": 1.324181318283081, "learning_rate": 7.151374248186521e-06, "loss": 0.8232, "step": 2882 }, { "epoch": 0.38, "grad_norm": 0.6433263421058655, "learning_rate": 7.149458417357305e-06, "loss": 1.0442, "step": 2883 }, { "epoch": 0.38, "grad_norm": 1.5022608041763306, "learning_rate": 7.147542199332298e-06, "loss": 0.9109, "step": 2884 }, { "epoch": 0.38, "grad_norm": 1.2793737649917603, "learning_rate": 7.1456255944566856e-06, "loss": 0.8822, "step": 2885 }, { "epoch": 0.38, "grad_norm": 1.41126549243927, "learning_rate": 7.1437086030757144e-06, "loss": 0.8749, "step": 2886 }, { "epoch": 0.38, "grad_norm": 0.7367566823959351, "learning_rate": 7.1417912255347054e-06, "loss": 1.0322, "step": 2887 }, { "epoch": 0.38, "grad_norm": 1.6441401243209839, "learning_rate": 7.139873462179048e-06, "loss": 0.86, "step": 2888 }, { "epoch": 0.38, "grad_norm": 1.4183189868927002, "learning_rate": 7.137955313354202e-06, "loss": 0.8706, "step": 2889 }, { "epoch": 0.38, "grad_norm": 1.2902112007141113, "learning_rate": 7.136036779405695e-06, "loss": 0.8012, "step": 2890 }, { "epoch": 0.38, "grad_norm": 1.309104084968567, "learning_rate": 7.1341178606791265e-06, "loss": 0.8171, "step": 2891 }, { "epoch": 0.38, "grad_norm": 1.3125743865966797, "learning_rate": 7.13219855752016e-06, "loss": 0.7528, "step": 2892 }, { "epoch": 0.38, "grad_norm": 1.288561463356018, "learning_rate": 7.130278870274538e-06, "loss": 0.8165, "step": 2893 }, { "epoch": 0.38, "grad_norm": 0.7084964513778687, "learning_rate": 7.12835879928806e-06, "loss": 1.0371, "step": 2894 }, { "epoch": 0.38, "grad_norm": 0.6878177523612976, "learning_rate": 7.126438344906603e-06, "loss": 1.053, "step": 2895 }, { "epoch": 0.38, "grad_norm": 1.2558704614639282, "learning_rate": 7.124517507476113e-06, "loss": 0.7731, "step": 2896 }, { "epoch": 0.38, "grad_norm": 0.6352792382240295, "learning_rate": 7.1225962873426e-06, "loss": 1.0308, "step": 2897 }, { "epoch": 0.38, "grad_norm": 1.4317227602005005, "learning_rate": 7.120674684852147e-06, "loss": 0.888, "step": 2898 }, { "epoch": 0.38, "grad_norm": 0.685522198677063, "learning_rate": 7.118752700350904e-06, "loss": 1.0307, "step": 2899 }, { "epoch": 0.38, "grad_norm": 0.6724451780319214, "learning_rate": 7.116830334185091e-06, "loss": 1.0332, "step": 2900 }, { "epoch": 0.38, "grad_norm": 1.3257560729980469, "learning_rate": 7.1149075867009965e-06, "loss": 0.8184, "step": 2901 }, { "epoch": 0.38, "grad_norm": 1.2317135334014893, "learning_rate": 7.112984458244976e-06, "loss": 0.8137, "step": 2902 }, { "epoch": 0.38, "grad_norm": 1.2714718580245972, "learning_rate": 7.111060949163456e-06, "loss": 0.8073, "step": 2903 }, { "epoch": 0.38, "grad_norm": 0.8051239252090454, "learning_rate": 7.109137059802931e-06, "loss": 1.0316, "step": 2904 }, { "epoch": 0.38, "grad_norm": 1.5279055833816528, "learning_rate": 7.107212790509962e-06, "loss": 0.7874, "step": 2905 }, { "epoch": 0.38, "grad_norm": 1.258653998374939, "learning_rate": 7.1052881416311815e-06, "loss": 0.8325, "step": 2906 }, { "epoch": 0.38, "grad_norm": 1.318384051322937, "learning_rate": 7.103363113513288e-06, "loss": 0.7561, "step": 2907 }, { "epoch": 0.38, "grad_norm": 0.6922521591186523, "learning_rate": 7.101437706503052e-06, "loss": 1.0479, "step": 2908 }, { "epoch": 0.38, "grad_norm": 1.2557677030563354, "learning_rate": 7.099511920947305e-06, "loss": 0.8009, "step": 2909 }, { "epoch": 0.38, "grad_norm": 0.670120358467102, "learning_rate": 7.0975857571929545e-06, "loss": 1.0402, "step": 2910 }, { "epoch": 0.38, "grad_norm": 1.2845499515533447, "learning_rate": 7.09565921558697e-06, "loss": 0.8411, "step": 2911 }, { "epoch": 0.38, "grad_norm": 1.4157682657241821, "learning_rate": 7.093732296476394e-06, "loss": 0.8042, "step": 2912 }, { "epoch": 0.38, "grad_norm": 0.694938063621521, "learning_rate": 7.091805000208336e-06, "loss": 1.0484, "step": 2913 }, { "epoch": 0.38, "grad_norm": 1.2337640523910522, "learning_rate": 7.089877327129969e-06, "loss": 0.7981, "step": 2914 }, { "epoch": 0.38, "grad_norm": 1.3737177848815918, "learning_rate": 7.087949277588539e-06, "loss": 0.8011, "step": 2915 }, { "epoch": 0.38, "grad_norm": 1.2780396938323975, "learning_rate": 7.086020851931359e-06, "loss": 0.7648, "step": 2916 }, { "epoch": 0.38, "grad_norm": 1.390142798423767, "learning_rate": 7.084092050505807e-06, "loss": 0.8208, "step": 2917 }, { "epoch": 0.38, "grad_norm": 1.5270711183547974, "learning_rate": 7.08216287365933e-06, "loss": 0.846, "step": 2918 }, { "epoch": 0.38, "grad_norm": 1.267449140548706, "learning_rate": 7.080233321739443e-06, "loss": 0.8015, "step": 2919 }, { "epoch": 0.38, "grad_norm": 0.7827869653701782, "learning_rate": 7.078303395093731e-06, "loss": 1.0606, "step": 2920 }, { "epoch": 0.38, "grad_norm": 1.2946395874023438, "learning_rate": 7.076373094069842e-06, "loss": 0.8181, "step": 2921 }, { "epoch": 0.38, "grad_norm": 0.6634315848350525, "learning_rate": 7.074442419015496e-06, "loss": 1.0201, "step": 2922 }, { "epoch": 0.38, "grad_norm": 1.334916353225708, "learning_rate": 7.072511370278474e-06, "loss": 0.8423, "step": 2923 }, { "epoch": 0.38, "grad_norm": 1.322635293006897, "learning_rate": 7.0705799482066305e-06, "loss": 0.8227, "step": 2924 }, { "epoch": 0.38, "grad_norm": 1.4151252508163452, "learning_rate": 7.068648153147886e-06, "loss": 0.8533, "step": 2925 }, { "epoch": 0.38, "grad_norm": 1.39988112449646, "learning_rate": 7.0667159854502254e-06, "loss": 0.8206, "step": 2926 }, { "epoch": 0.38, "grad_norm": 0.9026666283607483, "learning_rate": 7.064783445461705e-06, "loss": 1.032, "step": 2927 }, { "epoch": 0.38, "grad_norm": 1.5926473140716553, "learning_rate": 7.062850533530441e-06, "loss": 0.8315, "step": 2928 }, { "epoch": 0.38, "grad_norm": 1.2862131595611572, "learning_rate": 7.060917250004627e-06, "loss": 0.7922, "step": 2929 }, { "epoch": 0.38, "grad_norm": 1.4851632118225098, "learning_rate": 7.058983595232512e-06, "loss": 0.8919, "step": 2930 }, { "epoch": 0.38, "grad_norm": 0.7548871636390686, "learning_rate": 7.057049569562422e-06, "loss": 1.0594, "step": 2931 }, { "epoch": 0.38, "grad_norm": 1.3939210176467896, "learning_rate": 7.0551151733427446e-06, "loss": 0.7577, "step": 2932 }, { "epoch": 0.38, "grad_norm": 1.4236606359481812, "learning_rate": 7.053180406921934e-06, "loss": 0.8511, "step": 2933 }, { "epoch": 0.38, "grad_norm": 1.2682400941848755, "learning_rate": 7.051245270648514e-06, "loss": 0.7698, "step": 2934 }, { "epoch": 0.38, "grad_norm": 0.6439751982688904, "learning_rate": 7.049309764871071e-06, "loss": 1.0285, "step": 2935 }, { "epoch": 0.38, "grad_norm": 1.2702243328094482, "learning_rate": 7.04737388993826e-06, "loss": 0.8161, "step": 2936 }, { "epoch": 0.38, "grad_norm": 1.391436219215393, "learning_rate": 7.045437646198806e-06, "loss": 0.8951, "step": 2937 }, { "epoch": 0.38, "grad_norm": 1.3841065168380737, "learning_rate": 7.043501034001493e-06, "loss": 0.834, "step": 2938 }, { "epoch": 0.39, "grad_norm": 0.6498528122901917, "learning_rate": 7.041564053695178e-06, "loss": 1.0375, "step": 2939 }, { "epoch": 0.39, "grad_norm": 0.6492411494255066, "learning_rate": 7.039626705628779e-06, "loss": 1.0455, "step": 2940 }, { "epoch": 0.39, "grad_norm": 1.3455462455749512, "learning_rate": 7.037688990151288e-06, "loss": 0.8274, "step": 2941 }, { "epoch": 0.39, "grad_norm": 1.2558889389038086, "learning_rate": 7.035750907611752e-06, "loss": 0.834, "step": 2942 }, { "epoch": 0.39, "grad_norm": 0.6303099989891052, "learning_rate": 7.033812458359295e-06, "loss": 1.0505, "step": 2943 }, { "epoch": 0.39, "grad_norm": 1.3916655778884888, "learning_rate": 7.0318736427431e-06, "loss": 0.8228, "step": 2944 }, { "epoch": 0.39, "grad_norm": 0.671491801738739, "learning_rate": 7.029934461112419e-06, "loss": 1.0147, "step": 2945 }, { "epoch": 0.39, "grad_norm": 1.3327358961105347, "learning_rate": 7.02799491381657e-06, "loss": 0.8888, "step": 2946 }, { "epoch": 0.39, "grad_norm": 1.28776216506958, "learning_rate": 7.026055001204935e-06, "loss": 0.8559, "step": 2947 }, { "epoch": 0.39, "grad_norm": 1.299415111541748, "learning_rate": 7.024114723626962e-06, "loss": 0.7974, "step": 2948 }, { "epoch": 0.39, "grad_norm": 1.3773621320724487, "learning_rate": 7.0221740814321695e-06, "loss": 0.7796, "step": 2949 }, { "epoch": 0.39, "grad_norm": 0.6488187909126282, "learning_rate": 7.020233074970134e-06, "loss": 1.0306, "step": 2950 }, { "epoch": 0.39, "grad_norm": 0.6863738894462585, "learning_rate": 7.018291704590503e-06, "loss": 1.0383, "step": 2951 }, { "epoch": 0.39, "grad_norm": 1.29067063331604, "learning_rate": 7.0163499706429885e-06, "loss": 0.7528, "step": 2952 }, { "epoch": 0.39, "grad_norm": 1.3637830018997192, "learning_rate": 7.014407873477367e-06, "loss": 0.8331, "step": 2953 }, { "epoch": 0.39, "grad_norm": 1.3028217554092407, "learning_rate": 7.012465413443481e-06, "loss": 0.817, "step": 2954 }, { "epoch": 0.39, "grad_norm": 0.6502650380134583, "learning_rate": 7.010522590891238e-06, "loss": 1.0263, "step": 2955 }, { "epoch": 0.39, "grad_norm": 0.6738719344139099, "learning_rate": 7.008579406170612e-06, "loss": 1.0348, "step": 2956 }, { "epoch": 0.39, "grad_norm": 0.6274803876876831, "learning_rate": 7.006635859631641e-06, "loss": 1.0416, "step": 2957 }, { "epoch": 0.39, "grad_norm": 1.584547996520996, "learning_rate": 7.004691951624429e-06, "loss": 0.8135, "step": 2958 }, { "epoch": 0.39, "grad_norm": 0.614427387714386, "learning_rate": 7.002747682499141e-06, "loss": 1.0361, "step": 2959 }, { "epoch": 0.39, "grad_norm": 1.5101336240768433, "learning_rate": 7.000803052606018e-06, "loss": 0.7711, "step": 2960 }, { "epoch": 0.39, "grad_norm": 1.3826760053634644, "learning_rate": 6.998858062295352e-06, "loss": 0.7776, "step": 2961 }, { "epoch": 0.39, "grad_norm": 0.6948124766349792, "learning_rate": 6.996912711917509e-06, "loss": 1.04, "step": 2962 }, { "epoch": 0.39, "grad_norm": 1.3169699907302856, "learning_rate": 6.994967001822918e-06, "loss": 0.8776, "step": 2963 }, { "epoch": 0.39, "grad_norm": 1.3551775217056274, "learning_rate": 6.9930209323620735e-06, "loss": 0.8685, "step": 2964 }, { "epoch": 0.39, "grad_norm": 0.6264030337333679, "learning_rate": 6.99107450388553e-06, "loss": 1.042, "step": 2965 }, { "epoch": 0.39, "grad_norm": 1.2663745880126953, "learning_rate": 6.989127716743913e-06, "loss": 0.8385, "step": 2966 }, { "epoch": 0.39, "grad_norm": 1.3437057733535767, "learning_rate": 6.987180571287908e-06, "loss": 0.794, "step": 2967 }, { "epoch": 0.39, "grad_norm": 1.2964979410171509, "learning_rate": 6.985233067868269e-06, "loss": 0.9018, "step": 2968 }, { "epoch": 0.39, "grad_norm": 0.645410418510437, "learning_rate": 6.98328520683581e-06, "loss": 1.0338, "step": 2969 }, { "epoch": 0.39, "grad_norm": 1.4339087009429932, "learning_rate": 6.981336988541414e-06, "loss": 0.8247, "step": 2970 }, { "epoch": 0.39, "grad_norm": 0.6335654854774475, "learning_rate": 6.979388413336025e-06, "loss": 1.0337, "step": 2971 }, { "epoch": 0.39, "grad_norm": 1.4241324663162231, "learning_rate": 6.977439481570651e-06, "loss": 0.7993, "step": 2972 }, { "epoch": 0.39, "grad_norm": 1.3399946689605713, "learning_rate": 6.9754901935963695e-06, "loss": 0.8315, "step": 2973 }, { "epoch": 0.39, "grad_norm": 0.6523324847221375, "learning_rate": 6.973540549764315e-06, "loss": 1.0256, "step": 2974 }, { "epoch": 0.39, "grad_norm": 0.6535820960998535, "learning_rate": 6.971590550425688e-06, "loss": 1.0407, "step": 2975 }, { "epoch": 0.39, "grad_norm": 1.3284869194030762, "learning_rate": 6.96964019593176e-06, "loss": 0.7954, "step": 2976 }, { "epoch": 0.39, "grad_norm": 1.332450270652771, "learning_rate": 6.967689486633856e-06, "loss": 0.857, "step": 2977 }, { "epoch": 0.39, "grad_norm": 0.6758267879486084, "learning_rate": 6.965738422883372e-06, "loss": 1.0329, "step": 2978 }, { "epoch": 0.39, "grad_norm": 1.270972728729248, "learning_rate": 6.963787005031765e-06, "loss": 0.8127, "step": 2979 }, { "epoch": 0.39, "grad_norm": 1.3625787496566772, "learning_rate": 6.961835233430556e-06, "loss": 0.7766, "step": 2980 }, { "epoch": 0.39, "grad_norm": 0.6337068676948547, "learning_rate": 6.959883108431331e-06, "loss": 1.0188, "step": 2981 }, { "epoch": 0.39, "grad_norm": 1.4439054727554321, "learning_rate": 6.957930630385738e-06, "loss": 0.8733, "step": 2982 }, { "epoch": 0.39, "grad_norm": 1.3237640857696533, "learning_rate": 6.955977799645491e-06, "loss": 0.8471, "step": 2983 }, { "epoch": 0.39, "grad_norm": 1.3639600276947021, "learning_rate": 6.954024616562365e-06, "loss": 0.8144, "step": 2984 }, { "epoch": 0.39, "grad_norm": 0.7655502557754517, "learning_rate": 6.952071081488199e-06, "loss": 1.0438, "step": 2985 }, { "epoch": 0.39, "grad_norm": 0.6442192792892456, "learning_rate": 6.9501171947748955e-06, "loss": 1.0352, "step": 2986 }, { "epoch": 0.39, "grad_norm": 1.3024489879608154, "learning_rate": 6.9481629567744224e-06, "loss": 0.798, "step": 2987 }, { "epoch": 0.39, "grad_norm": 1.317404866218567, "learning_rate": 6.94620836783881e-06, "loss": 0.8115, "step": 2988 }, { "epoch": 0.39, "grad_norm": 1.3425871133804321, "learning_rate": 6.944253428320148e-06, "loss": 0.7838, "step": 2989 }, { "epoch": 0.39, "grad_norm": 0.740271270275116, "learning_rate": 6.942298138570594e-06, "loss": 1.0261, "step": 2990 }, { "epoch": 0.39, "grad_norm": 0.646454393863678, "learning_rate": 6.940342498942365e-06, "loss": 1.0357, "step": 2991 }, { "epoch": 0.39, "grad_norm": 1.314817190170288, "learning_rate": 6.938386509787745e-06, "loss": 0.8455, "step": 2992 }, { "epoch": 0.39, "grad_norm": 1.4021885395050049, "learning_rate": 6.936430171459079e-06, "loss": 0.8316, "step": 2993 }, { "epoch": 0.39, "grad_norm": 0.6918042898178101, "learning_rate": 6.934473484308775e-06, "loss": 1.02, "step": 2994 }, { "epoch": 0.39, "grad_norm": 1.426990270614624, "learning_rate": 6.9325164486893015e-06, "loss": 0.8408, "step": 2995 }, { "epoch": 0.39, "grad_norm": 0.6689214110374451, "learning_rate": 6.930559064953194e-06, "loss": 1.0199, "step": 2996 }, { "epoch": 0.39, "grad_norm": 1.3108935356140137, "learning_rate": 6.928601333453049e-06, "loss": 0.8689, "step": 2997 }, { "epoch": 0.39, "grad_norm": 1.2841562032699585, "learning_rate": 6.926643254541522e-06, "loss": 0.8237, "step": 2998 }, { "epoch": 0.39, "grad_norm": 1.2996211051940918, "learning_rate": 6.924684828571338e-06, "loss": 0.8277, "step": 2999 }, { "epoch": 0.39, "grad_norm": 1.2803982496261597, "learning_rate": 6.922726055895278e-06, "loss": 0.8913, "step": 3000 }, { "epoch": 0.39, "grad_norm": 1.4913057088851929, "learning_rate": 6.92076693686619e-06, "loss": 0.8356, "step": 3001 }, { "epoch": 0.39, "grad_norm": 1.2952152490615845, "learning_rate": 6.918807471836985e-06, "loss": 0.856, "step": 3002 }, { "epoch": 0.39, "grad_norm": 1.3793503046035767, "learning_rate": 6.916847661160628e-06, "loss": 0.8622, "step": 3003 }, { "epoch": 0.39, "grad_norm": 1.308638572692871, "learning_rate": 6.9148875051901584e-06, "loss": 0.7887, "step": 3004 }, { "epoch": 0.39, "grad_norm": 1.3594554662704468, "learning_rate": 6.912927004278669e-06, "loss": 0.84, "step": 3005 }, { "epoch": 0.39, "grad_norm": 1.4221129417419434, "learning_rate": 6.9109661587793176e-06, "loss": 0.8871, "step": 3006 }, { "epoch": 0.39, "grad_norm": 1.3580089807510376, "learning_rate": 6.909004969045323e-06, "loss": 0.8138, "step": 3007 }, { "epoch": 0.39, "grad_norm": 1.3221536874771118, "learning_rate": 6.9070434354299685e-06, "loss": 0.8394, "step": 3008 }, { "epoch": 0.39, "grad_norm": 1.2587980031967163, "learning_rate": 6.905081558286598e-06, "loss": 0.8306, "step": 3009 }, { "epoch": 0.39, "grad_norm": 1.497908592224121, "learning_rate": 6.903119337968615e-06, "loss": 0.8152, "step": 3010 }, { "epoch": 0.39, "grad_norm": 1.5451631546020508, "learning_rate": 6.90115677482949e-06, "loss": 0.8379, "step": 3011 }, { "epoch": 0.39, "grad_norm": 1.5830111503601074, "learning_rate": 6.899193869222749e-06, "loss": 0.8115, "step": 3012 }, { "epoch": 0.39, "grad_norm": 0.8083904385566711, "learning_rate": 6.8972306215019846e-06, "loss": 1.0359, "step": 3013 }, { "epoch": 0.39, "grad_norm": 1.2679452896118164, "learning_rate": 6.895267032020849e-06, "loss": 0.8333, "step": 3014 }, { "epoch": 0.4, "grad_norm": 1.2542626857757568, "learning_rate": 6.893303101133055e-06, "loss": 0.8249, "step": 3015 }, { "epoch": 0.4, "grad_norm": 1.3150299787521362, "learning_rate": 6.89133882919238e-06, "loss": 0.8361, "step": 3016 }, { "epoch": 0.4, "grad_norm": 0.685942530632019, "learning_rate": 6.889374216552661e-06, "loss": 1.0569, "step": 3017 }, { "epoch": 0.4, "grad_norm": 1.3236654996871948, "learning_rate": 6.887409263567795e-06, "loss": 0.8593, "step": 3018 }, { "epoch": 0.4, "grad_norm": 0.6138343811035156, "learning_rate": 6.885443970591741e-06, "loss": 1.0241, "step": 3019 }, { "epoch": 0.4, "grad_norm": 1.405121922492981, "learning_rate": 6.883478337978522e-06, "loss": 0.8577, "step": 3020 }, { "epoch": 0.4, "grad_norm": 1.3468194007873535, "learning_rate": 6.881512366082219e-06, "loss": 0.8299, "step": 3021 }, { "epoch": 0.4, "grad_norm": 1.2219507694244385, "learning_rate": 6.8795460552569735e-06, "loss": 0.7952, "step": 3022 }, { "epoch": 0.4, "grad_norm": 0.6933592557907104, "learning_rate": 6.877579405856993e-06, "loss": 1.0203, "step": 3023 }, { "epoch": 0.4, "grad_norm": 1.285374641418457, "learning_rate": 6.8756124182365405e-06, "loss": 0.7604, "step": 3024 }, { "epoch": 0.4, "grad_norm": 1.9715851545333862, "learning_rate": 6.8736450927499446e-06, "loss": 0.8214, "step": 3025 }, { "epoch": 0.4, "grad_norm": 1.3972442150115967, "learning_rate": 6.871677429751588e-06, "loss": 0.8425, "step": 3026 }, { "epoch": 0.4, "grad_norm": 1.5148839950561523, "learning_rate": 6.869709429595922e-06, "loss": 0.8126, "step": 3027 }, { "epoch": 0.4, "grad_norm": 1.3834803104400635, "learning_rate": 6.867741092637454e-06, "loss": 0.8399, "step": 3028 }, { "epoch": 0.4, "grad_norm": 1.3274935483932495, "learning_rate": 6.865772419230754e-06, "loss": 0.8421, "step": 3029 }, { "epoch": 0.4, "grad_norm": 1.2929531335830688, "learning_rate": 6.863803409730452e-06, "loss": 0.8221, "step": 3030 }, { "epoch": 0.4, "grad_norm": 1.3582453727722168, "learning_rate": 6.861834064491235e-06, "loss": 0.8448, "step": 3031 }, { "epoch": 0.4, "grad_norm": 0.723221480846405, "learning_rate": 6.859864383867858e-06, "loss": 1.0385, "step": 3032 }, { "epoch": 0.4, "grad_norm": 1.373902440071106, "learning_rate": 6.85789436821513e-06, "loss": 0.8034, "step": 3033 }, { "epoch": 0.4, "grad_norm": 1.2340296506881714, "learning_rate": 6.855924017887925e-06, "loss": 0.8192, "step": 3034 }, { "epoch": 0.4, "grad_norm": 1.301221251487732, "learning_rate": 6.853953333241172e-06, "loss": 0.8953, "step": 3035 }, { "epoch": 0.4, "grad_norm": 0.6503003835678101, "learning_rate": 6.8519823146298645e-06, "loss": 1.035, "step": 3036 }, { "epoch": 0.4, "grad_norm": 1.2782890796661377, "learning_rate": 6.850010962409055e-06, "loss": 0.7953, "step": 3037 }, { "epoch": 0.4, "grad_norm": 1.2805858850479126, "learning_rate": 6.848039276933855e-06, "loss": 0.8161, "step": 3038 }, { "epoch": 0.4, "grad_norm": 1.3105943202972412, "learning_rate": 6.8460672585594365e-06, "loss": 0.825, "step": 3039 }, { "epoch": 0.4, "grad_norm": 1.2774021625518799, "learning_rate": 6.844094907641033e-06, "loss": 0.8053, "step": 3040 }, { "epoch": 0.4, "grad_norm": 0.6553773283958435, "learning_rate": 6.842122224533937e-06, "loss": 1.0383, "step": 3041 }, { "epoch": 0.4, "grad_norm": 1.3210464715957642, "learning_rate": 6.8401492095935e-06, "loss": 0.8858, "step": 3042 }, { "epoch": 0.4, "grad_norm": 1.3631806373596191, "learning_rate": 6.838175863175134e-06, "loss": 0.8961, "step": 3043 }, { "epoch": 0.4, "grad_norm": 1.3425346612930298, "learning_rate": 6.8362021856343104e-06, "loss": 0.877, "step": 3044 }, { "epoch": 0.4, "grad_norm": 1.325446605682373, "learning_rate": 6.834228177326561e-06, "loss": 0.8581, "step": 3045 }, { "epoch": 0.4, "grad_norm": 1.4050432443618774, "learning_rate": 6.832253838607475e-06, "loss": 0.8414, "step": 3046 }, { "epoch": 0.4, "grad_norm": 1.2982819080352783, "learning_rate": 6.830279169832705e-06, "loss": 0.8471, "step": 3047 }, { "epoch": 0.4, "grad_norm": 1.316251277923584, "learning_rate": 6.82830417135796e-06, "loss": 0.863, "step": 3048 }, { "epoch": 0.4, "grad_norm": 1.3731440305709839, "learning_rate": 6.826328843539009e-06, "loss": 0.7892, "step": 3049 }, { "epoch": 0.4, "grad_norm": 1.3967102766036987, "learning_rate": 6.824353186731681e-06, "loss": 0.8015, "step": 3050 }, { "epoch": 0.4, "grad_norm": 0.7116036415100098, "learning_rate": 6.822377201291863e-06, "loss": 1.0295, "step": 3051 }, { "epoch": 0.4, "grad_norm": 1.2957472801208496, "learning_rate": 6.8204008875755024e-06, "loss": 0.831, "step": 3052 }, { "epoch": 0.4, "grad_norm": 1.2553579807281494, "learning_rate": 6.818424245938605e-06, "loss": 0.8196, "step": 3053 }, { "epoch": 0.4, "grad_norm": 1.3728446960449219, "learning_rate": 6.816447276737236e-06, "loss": 0.8727, "step": 3054 }, { "epoch": 0.4, "grad_norm": 1.2821924686431885, "learning_rate": 6.814469980327519e-06, "loss": 0.8259, "step": 3055 }, { "epoch": 0.4, "grad_norm": 1.3650710582733154, "learning_rate": 6.812492357065641e-06, "loss": 0.8006, "step": 3056 }, { "epoch": 0.4, "grad_norm": 1.2386242151260376, "learning_rate": 6.810514407307838e-06, "loss": 0.8231, "step": 3057 }, { "epoch": 0.4, "grad_norm": 1.273422122001648, "learning_rate": 6.808536131410416e-06, "loss": 0.7815, "step": 3058 }, { "epoch": 0.4, "grad_norm": 1.532668113708496, "learning_rate": 6.8065575297297305e-06, "loss": 0.8246, "step": 3059 }, { "epoch": 0.4, "grad_norm": 1.4557819366455078, "learning_rate": 6.804578602622202e-06, "loss": 0.8166, "step": 3060 }, { "epoch": 0.4, "grad_norm": 1.3585940599441528, "learning_rate": 6.802599350444306e-06, "loss": 0.8814, "step": 3061 }, { "epoch": 0.4, "grad_norm": 0.7225295305252075, "learning_rate": 6.800619773552582e-06, "loss": 1.0156, "step": 3062 }, { "epoch": 0.4, "grad_norm": 1.4638649225234985, "learning_rate": 6.798639872303618e-06, "loss": 0.8126, "step": 3063 }, { "epoch": 0.4, "grad_norm": 1.3213427066802979, "learning_rate": 6.7966596470540715e-06, "loss": 0.8524, "step": 3064 }, { "epoch": 0.4, "grad_norm": 1.261141061782837, "learning_rate": 6.794679098160649e-06, "loss": 0.8375, "step": 3065 }, { "epoch": 0.4, "grad_norm": 0.6731187105178833, "learning_rate": 6.792698225980122e-06, "loss": 1.0602, "step": 3066 }, { "epoch": 0.4, "grad_norm": 1.219870924949646, "learning_rate": 6.790717030869319e-06, "loss": 0.7983, "step": 3067 }, { "epoch": 0.4, "grad_norm": 0.6151318550109863, "learning_rate": 6.788735513185122e-06, "loss": 1.0347, "step": 3068 }, { "epoch": 0.4, "grad_norm": 0.6280215978622437, "learning_rate": 6.786753673284476e-06, "loss": 1.0348, "step": 3069 }, { "epoch": 0.4, "grad_norm": 0.6401231288909912, "learning_rate": 6.784771511524384e-06, "loss": 1.0191, "step": 3070 }, { "epoch": 0.4, "grad_norm": 1.346914529800415, "learning_rate": 6.7827890282619035e-06, "loss": 0.8224, "step": 3071 }, { "epoch": 0.4, "grad_norm": 1.2988818883895874, "learning_rate": 6.780806223854154e-06, "loss": 0.8378, "step": 3072 }, { "epoch": 0.4, "grad_norm": 1.2990446090698242, "learning_rate": 6.778823098658309e-06, "loss": 0.8025, "step": 3073 }, { "epoch": 0.4, "grad_norm": 0.6831691861152649, "learning_rate": 6.776839653031602e-06, "loss": 1.0391, "step": 3074 }, { "epoch": 0.4, "grad_norm": 1.295408844947815, "learning_rate": 6.774855887331325e-06, "loss": 0.7852, "step": 3075 }, { "epoch": 0.4, "grad_norm": 2.9541473388671875, "learning_rate": 6.772871801914826e-06, "loss": 0.8325, "step": 3076 }, { "epoch": 0.4, "grad_norm": 1.256531834602356, "learning_rate": 6.7708873971395095e-06, "loss": 0.7887, "step": 3077 }, { "epoch": 0.4, "grad_norm": 1.4372678995132446, "learning_rate": 6.768902673362841e-06, "loss": 0.8871, "step": 3078 }, { "epoch": 0.4, "grad_norm": 1.4540200233459473, "learning_rate": 6.766917630942341e-06, "loss": 0.8334, "step": 3079 }, { "epoch": 0.4, "grad_norm": 1.294981598854065, "learning_rate": 6.764932270235587e-06, "loss": 0.7986, "step": 3080 }, { "epoch": 0.4, "grad_norm": 1.3686540126800537, "learning_rate": 6.762946591600218e-06, "loss": 0.8035, "step": 3081 }, { "epoch": 0.4, "grad_norm": 1.288496732711792, "learning_rate": 6.7609605953939225e-06, "loss": 0.8067, "step": 3082 }, { "epoch": 0.4, "grad_norm": 1.3613455295562744, "learning_rate": 6.758974281974454e-06, "loss": 0.8613, "step": 3083 }, { "epoch": 0.4, "grad_norm": 1.2899293899536133, "learning_rate": 6.756987651699618e-06, "loss": 0.8125, "step": 3084 }, { "epoch": 0.4, "grad_norm": 1.3755841255187988, "learning_rate": 6.755000704927281e-06, "loss": 0.8336, "step": 3085 }, { "epoch": 0.4, "grad_norm": 1.3433868885040283, "learning_rate": 6.753013442015365e-06, "loss": 0.8058, "step": 3086 }, { "epoch": 0.4, "grad_norm": 1.342294454574585, "learning_rate": 6.751025863321845e-06, "loss": 0.8411, "step": 3087 }, { "epoch": 0.4, "grad_norm": 1.428605079650879, "learning_rate": 6.749037969204759e-06, "loss": 0.8362, "step": 3088 }, { "epoch": 0.4, "grad_norm": 1.3121016025543213, "learning_rate": 6.747049760022197e-06, "loss": 0.8014, "step": 3089 }, { "epoch": 0.4, "grad_norm": 1.2571885585784912, "learning_rate": 6.745061236132311e-06, "loss": 0.7803, "step": 3090 }, { "epoch": 0.41, "grad_norm": 0.83385169506073, "learning_rate": 6.743072397893307e-06, "loss": 1.0385, "step": 3091 }, { "epoch": 0.41, "grad_norm": 0.7392719388008118, "learning_rate": 6.741083245663443e-06, "loss": 1.0496, "step": 3092 }, { "epoch": 0.41, "grad_norm": 1.3274468183517456, "learning_rate": 6.739093779801039e-06, "loss": 0.8091, "step": 3093 }, { "epoch": 0.41, "grad_norm": 1.344325065612793, "learning_rate": 6.737104000664474e-06, "loss": 0.8042, "step": 3094 }, { "epoch": 0.41, "grad_norm": 0.7413623929023743, "learning_rate": 6.735113908612176e-06, "loss": 1.0382, "step": 3095 }, { "epoch": 0.41, "grad_norm": 1.3872320652008057, "learning_rate": 6.733123504002635e-06, "loss": 0.8505, "step": 3096 }, { "epoch": 0.41, "grad_norm": 1.32227623462677, "learning_rate": 6.731132787194394e-06, "loss": 0.7922, "step": 3097 }, { "epoch": 0.41, "grad_norm": 0.7188237905502319, "learning_rate": 6.729141758546054e-06, "loss": 1.0356, "step": 3098 }, { "epoch": 0.41, "grad_norm": 1.2883341312408447, "learning_rate": 6.727150418416273e-06, "loss": 0.7758, "step": 3099 }, { "epoch": 0.41, "grad_norm": 1.3959237337112427, "learning_rate": 6.7251587671637616e-06, "loss": 0.8088, "step": 3100 }, { "epoch": 0.41, "grad_norm": 1.285133719444275, "learning_rate": 6.723166805147292e-06, "loss": 0.8032, "step": 3101 }, { "epoch": 0.41, "grad_norm": 1.25136399269104, "learning_rate": 6.721174532725686e-06, "loss": 0.8617, "step": 3102 }, { "epoch": 0.41, "grad_norm": 1.313409686088562, "learning_rate": 6.719181950257826e-06, "loss": 0.8345, "step": 3103 }, { "epoch": 0.41, "grad_norm": 0.7530161142349243, "learning_rate": 6.717189058102649e-06, "loss": 1.0371, "step": 3104 }, { "epoch": 0.41, "grad_norm": 0.7062869668006897, "learning_rate": 6.715195856619146e-06, "loss": 1.0342, "step": 3105 }, { "epoch": 0.41, "grad_norm": 0.6131823062896729, "learning_rate": 6.713202346166368e-06, "loss": 1.0125, "step": 3106 }, { "epoch": 0.41, "grad_norm": 1.2950925827026367, "learning_rate": 6.7112085271034154e-06, "loss": 0.8761, "step": 3107 }, { "epoch": 0.41, "grad_norm": 0.7147043347358704, "learning_rate": 6.70921439978945e-06, "loss": 1.0291, "step": 3108 }, { "epoch": 0.41, "grad_norm": 1.3354145288467407, "learning_rate": 6.707219964583688e-06, "loss": 0.8638, "step": 3109 }, { "epoch": 0.41, "grad_norm": 0.7261008024215698, "learning_rate": 6.705225221845395e-06, "loss": 1.0297, "step": 3110 }, { "epoch": 0.41, "grad_norm": 1.3466765880584717, "learning_rate": 6.703230171933902e-06, "loss": 0.7925, "step": 3111 }, { "epoch": 0.41, "grad_norm": 1.3634177446365356, "learning_rate": 6.701234815208587e-06, "loss": 0.779, "step": 3112 }, { "epoch": 0.41, "grad_norm": 1.4172844886779785, "learning_rate": 6.699239152028888e-06, "loss": 0.8614, "step": 3113 }, { "epoch": 0.41, "grad_norm": 1.331575632095337, "learning_rate": 6.697243182754297e-06, "loss": 0.8406, "step": 3114 }, { "epoch": 0.41, "grad_norm": 1.2521390914916992, "learning_rate": 6.69524690774436e-06, "loss": 0.7528, "step": 3115 }, { "epoch": 0.41, "grad_norm": 1.3504738807678223, "learning_rate": 6.693250327358678e-06, "loss": 0.8583, "step": 3116 }, { "epoch": 0.41, "grad_norm": 1.3666514158248901, "learning_rate": 6.691253441956909e-06, "loss": 0.8008, "step": 3117 }, { "epoch": 0.41, "grad_norm": 1.3719969987869263, "learning_rate": 6.689256251898763e-06, "loss": 0.8476, "step": 3118 }, { "epoch": 0.41, "grad_norm": 1.5221904516220093, "learning_rate": 6.687258757544012e-06, "loss": 0.8493, "step": 3119 }, { "epoch": 0.41, "grad_norm": 1.3568106889724731, "learning_rate": 6.6852609592524706e-06, "loss": 0.8255, "step": 3120 }, { "epoch": 0.41, "grad_norm": 1.4190478324890137, "learning_rate": 6.68326285738402e-06, "loss": 0.8777, "step": 3121 }, { "epoch": 0.41, "grad_norm": 1.5395170450210571, "learning_rate": 6.6812644522985885e-06, "loss": 0.8249, "step": 3122 }, { "epoch": 0.41, "grad_norm": 1.42179536819458, "learning_rate": 6.679265744356164e-06, "loss": 0.7169, "step": 3123 }, { "epoch": 0.41, "grad_norm": 1.4188179969787598, "learning_rate": 6.677266733916784e-06, "loss": 0.8126, "step": 3124 }, { "epoch": 0.41, "grad_norm": 1.3629815578460693, "learning_rate": 6.675267421340544e-06, "loss": 0.8643, "step": 3125 }, { "epoch": 0.41, "grad_norm": 1.6050795316696167, "learning_rate": 6.673267806987594e-06, "loss": 0.838, "step": 3126 }, { "epoch": 0.41, "grad_norm": 1.291519284248352, "learning_rate": 6.671267891218137e-06, "loss": 0.7741, "step": 3127 }, { "epoch": 0.41, "grad_norm": 1.212667465209961, "learning_rate": 6.669267674392432e-06, "loss": 0.8232, "step": 3128 }, { "epoch": 0.41, "grad_norm": 1.3191518783569336, "learning_rate": 6.667267156870788e-06, "loss": 0.8443, "step": 3129 }, { "epoch": 0.41, "grad_norm": 1.3946170806884766, "learning_rate": 6.665266339013573e-06, "loss": 0.8294, "step": 3130 }, { "epoch": 0.41, "grad_norm": 0.9373865723609924, "learning_rate": 6.663265221181205e-06, "loss": 1.0464, "step": 3131 }, { "epoch": 0.41, "grad_norm": 1.3816618919372559, "learning_rate": 6.661263803734163e-06, "loss": 0.7924, "step": 3132 }, { "epoch": 0.41, "grad_norm": 1.3197535276412964, "learning_rate": 6.659262087032972e-06, "loss": 0.8816, "step": 3133 }, { "epoch": 0.41, "grad_norm": 1.349848747253418, "learning_rate": 6.657260071438213e-06, "loss": 0.8181, "step": 3134 }, { "epoch": 0.41, "grad_norm": 1.339754581451416, "learning_rate": 6.6552577573105236e-06, "loss": 0.876, "step": 3135 }, { "epoch": 0.41, "grad_norm": 1.258653998374939, "learning_rate": 6.653255145010594e-06, "loss": 0.8254, "step": 3136 }, { "epoch": 0.41, "grad_norm": 1.3142905235290527, "learning_rate": 6.651252234899167e-06, "loss": 0.7727, "step": 3137 }, { "epoch": 0.41, "grad_norm": 0.7056183815002441, "learning_rate": 6.64924902733704e-06, "loss": 1.0547, "step": 3138 }, { "epoch": 0.41, "grad_norm": 1.6333707571029663, "learning_rate": 6.647245522685062e-06, "loss": 0.8344, "step": 3139 }, { "epoch": 0.41, "grad_norm": 1.3609552383422852, "learning_rate": 6.6452417213041394e-06, "loss": 0.7987, "step": 3140 }, { "epoch": 0.41, "grad_norm": 0.657126784324646, "learning_rate": 6.643237623555228e-06, "loss": 1.0562, "step": 3141 }, { "epoch": 0.41, "grad_norm": 1.459162950515747, "learning_rate": 6.641233229799342e-06, "loss": 0.7749, "step": 3142 }, { "epoch": 0.41, "grad_norm": 0.6612762808799744, "learning_rate": 6.639228540397542e-06, "loss": 1.0267, "step": 3143 }, { "epoch": 0.41, "grad_norm": 1.3329381942749023, "learning_rate": 6.6372235557109475e-06, "loss": 0.8587, "step": 3144 }, { "epoch": 0.41, "grad_norm": 0.701859712600708, "learning_rate": 6.6352182761007286e-06, "loss": 1.0413, "step": 3145 }, { "epoch": 0.41, "grad_norm": 1.3839365243911743, "learning_rate": 6.63321270192811e-06, "loss": 0.8206, "step": 3146 }, { "epoch": 0.41, "grad_norm": 1.311676263809204, "learning_rate": 6.6312068335543685e-06, "loss": 0.8189, "step": 3147 }, { "epoch": 0.41, "grad_norm": 1.3722854852676392, "learning_rate": 6.629200671340834e-06, "loss": 0.8763, "step": 3148 }, { "epoch": 0.41, "grad_norm": 1.6748658418655396, "learning_rate": 6.627194215648889e-06, "loss": 0.7303, "step": 3149 }, { "epoch": 0.41, "grad_norm": 1.3677434921264648, "learning_rate": 6.625187466839969e-06, "loss": 0.861, "step": 3150 }, { "epoch": 0.41, "grad_norm": 1.321062445640564, "learning_rate": 6.623180425275565e-06, "loss": 0.8834, "step": 3151 }, { "epoch": 0.41, "grad_norm": 1.297183632850647, "learning_rate": 6.621173091317216e-06, "loss": 0.7764, "step": 3152 }, { "epoch": 0.41, "grad_norm": 1.4805543422698975, "learning_rate": 6.619165465326516e-06, "loss": 0.8287, "step": 3153 }, { "epoch": 0.41, "grad_norm": 0.9308131337165833, "learning_rate": 6.617157547665112e-06, "loss": 1.02, "step": 3154 }, { "epoch": 0.41, "grad_norm": 1.3743587732315063, "learning_rate": 6.615149338694704e-06, "loss": 0.793, "step": 3155 }, { "epoch": 0.41, "grad_norm": 1.3237159252166748, "learning_rate": 6.613140838777041e-06, "loss": 0.8348, "step": 3156 }, { "epoch": 0.41, "grad_norm": 1.386305332183838, "learning_rate": 6.611132048273932e-06, "loss": 0.8093, "step": 3157 }, { "epoch": 0.41, "grad_norm": 1.3263633251190186, "learning_rate": 6.609122967547228e-06, "loss": 0.8289, "step": 3158 }, { "epoch": 0.41, "grad_norm": 1.2590391635894775, "learning_rate": 6.607113596958841e-06, "loss": 0.8136, "step": 3159 }, { "epoch": 0.41, "grad_norm": 1.4011849164962769, "learning_rate": 6.605103936870731e-06, "loss": 0.8726, "step": 3160 }, { "epoch": 0.41, "grad_norm": 1.444329857826233, "learning_rate": 6.60309398764491e-06, "loss": 0.8678, "step": 3161 }, { "epoch": 0.41, "grad_norm": 0.7018819451332092, "learning_rate": 6.601083749643444e-06, "loss": 1.0245, "step": 3162 }, { "epoch": 0.41, "grad_norm": 1.4407756328582764, "learning_rate": 6.599073223228449e-06, "loss": 0.8768, "step": 3163 }, { "epoch": 0.41, "grad_norm": 0.655733585357666, "learning_rate": 6.5970624087620975e-06, "loss": 1.0693, "step": 3164 }, { "epoch": 0.41, "grad_norm": 1.279342532157898, "learning_rate": 6.595051306606606e-06, "loss": 0.8393, "step": 3165 }, { "epoch": 0.41, "grad_norm": 1.38214910030365, "learning_rate": 6.593039917124251e-06, "loss": 0.8177, "step": 3166 }, { "epoch": 0.41, "grad_norm": 1.3109978437423706, "learning_rate": 6.591028240677355e-06, "loss": 0.7785, "step": 3167 }, { "epoch": 0.42, "grad_norm": 1.3190829753875732, "learning_rate": 6.589016277628294e-06, "loss": 0.7624, "step": 3168 }, { "epoch": 0.42, "grad_norm": 0.6804970502853394, "learning_rate": 6.587004028339496e-06, "loss": 1.0482, "step": 3169 }, { "epoch": 0.42, "grad_norm": 1.4246000051498413, "learning_rate": 6.584991493173443e-06, "loss": 0.8247, "step": 3170 }, { "epoch": 0.42, "grad_norm": 1.3366767168045044, "learning_rate": 6.582978672492662e-06, "loss": 0.8344, "step": 3171 }, { "epoch": 0.42, "grad_norm": 1.4099924564361572, "learning_rate": 6.580965566659739e-06, "loss": 0.8964, "step": 3172 }, { "epoch": 0.42, "grad_norm": 0.6554224491119385, "learning_rate": 6.5789521760373035e-06, "loss": 1.009, "step": 3173 }, { "epoch": 0.42, "grad_norm": 1.3145605325698853, "learning_rate": 6.576938500988045e-06, "loss": 0.8741, "step": 3174 }, { "epoch": 0.42, "grad_norm": 0.6517271995544434, "learning_rate": 6.5749245418746965e-06, "loss": 1.0543, "step": 3175 }, { "epoch": 0.42, "grad_norm": 1.3602795600891113, "learning_rate": 6.5729102990600466e-06, "loss": 0.8238, "step": 3176 }, { "epoch": 0.42, "grad_norm": 0.649824857711792, "learning_rate": 6.5708957729069355e-06, "loss": 1.0545, "step": 3177 }, { "epoch": 0.42, "grad_norm": 0.6336112022399902, "learning_rate": 6.568880963778249e-06, "loss": 1.0211, "step": 3178 }, { "epoch": 0.42, "grad_norm": 1.4168168306350708, "learning_rate": 6.56686587203693e-06, "loss": 0.7994, "step": 3179 }, { "epoch": 0.42, "grad_norm": 1.5932613611221313, "learning_rate": 6.564850498045971e-06, "loss": 0.8445, "step": 3180 }, { "epoch": 0.42, "grad_norm": 1.457050085067749, "learning_rate": 6.562834842168413e-06, "loss": 0.8771, "step": 3181 }, { "epoch": 0.42, "grad_norm": 1.2822223901748657, "learning_rate": 6.560818904767349e-06, "loss": 0.8061, "step": 3182 }, { "epoch": 0.42, "grad_norm": 1.33260977268219, "learning_rate": 6.558802686205922e-06, "loss": 0.8534, "step": 3183 }, { "epoch": 0.42, "grad_norm": 0.7134935259819031, "learning_rate": 6.556786186847328e-06, "loss": 1.0401, "step": 3184 }, { "epoch": 0.42, "grad_norm": 1.4673945903778076, "learning_rate": 6.554769407054811e-06, "loss": 0.8362, "step": 3185 }, { "epoch": 0.42, "grad_norm": 1.3204113245010376, "learning_rate": 6.552752347191667e-06, "loss": 0.8558, "step": 3186 }, { "epoch": 0.42, "grad_norm": 1.2809948921203613, "learning_rate": 6.5507350076212405e-06, "loss": 0.7992, "step": 3187 }, { "epoch": 0.42, "grad_norm": 1.5493910312652588, "learning_rate": 6.548717388706931e-06, "loss": 0.8367, "step": 3188 }, { "epoch": 0.42, "grad_norm": 1.5571216344833374, "learning_rate": 6.546699490812184e-06, "loss": 0.8609, "step": 3189 }, { "epoch": 0.42, "grad_norm": 0.7519665360450745, "learning_rate": 6.544681314300494e-06, "loss": 1.0432, "step": 3190 }, { "epoch": 0.42, "grad_norm": 1.8125487565994263, "learning_rate": 6.542662859535411e-06, "loss": 0.8283, "step": 3191 }, { "epoch": 0.42, "grad_norm": 1.3663281202316284, "learning_rate": 6.540644126880531e-06, "loss": 0.839, "step": 3192 }, { "epoch": 0.42, "grad_norm": 1.32797110080719, "learning_rate": 6.5386251166995006e-06, "loss": 0.7594, "step": 3193 }, { "epoch": 0.42, "grad_norm": 1.3719232082366943, "learning_rate": 6.53660582935602e-06, "loss": 0.7889, "step": 3194 }, { "epoch": 0.42, "grad_norm": 1.4320629835128784, "learning_rate": 6.534586265213834e-06, "loss": 0.8554, "step": 3195 }, { "epoch": 0.42, "grad_norm": 0.6705951690673828, "learning_rate": 6.53256642463674e-06, "loss": 1.043, "step": 3196 }, { "epoch": 0.42, "grad_norm": 1.3335826396942139, "learning_rate": 6.530546307988586e-06, "loss": 0.8347, "step": 3197 }, { "epoch": 0.42, "grad_norm": 1.3267672061920166, "learning_rate": 6.528525915633267e-06, "loss": 0.8095, "step": 3198 }, { "epoch": 0.42, "grad_norm": 1.2365691661834717, "learning_rate": 6.526505247934729e-06, "loss": 0.8066, "step": 3199 }, { "epoch": 0.42, "grad_norm": 1.3788385391235352, "learning_rate": 6.5244843052569696e-06, "loss": 0.833, "step": 3200 }, { "epoch": 0.42, "grad_norm": 1.2803863286972046, "learning_rate": 6.522463087964033e-06, "loss": 0.7785, "step": 3201 }, { "epoch": 0.42, "grad_norm": 0.6923340559005737, "learning_rate": 6.520441596420015e-06, "loss": 1.0326, "step": 3202 }, { "epoch": 0.42, "grad_norm": 1.5473921298980713, "learning_rate": 6.51841983098906e-06, "loss": 0.8007, "step": 3203 }, { "epoch": 0.42, "grad_norm": 1.3011642694473267, "learning_rate": 6.516397792035361e-06, "loss": 0.8031, "step": 3204 }, { "epoch": 0.42, "grad_norm": 1.3969727754592896, "learning_rate": 6.51437547992316e-06, "loss": 0.847, "step": 3205 }, { "epoch": 0.42, "grad_norm": 1.3877761363983154, "learning_rate": 6.512352895016751e-06, "loss": 0.7632, "step": 3206 }, { "epoch": 0.42, "grad_norm": 0.6583780646324158, "learning_rate": 6.510330037680475e-06, "loss": 1.0588, "step": 3207 }, { "epoch": 0.42, "grad_norm": 1.4175677299499512, "learning_rate": 6.5083069082787205e-06, "loss": 0.8541, "step": 3208 }, { "epoch": 0.42, "grad_norm": 1.350285291671753, "learning_rate": 6.506283507175929e-06, "loss": 0.7918, "step": 3209 }, { "epoch": 0.42, "grad_norm": 1.3725324869155884, "learning_rate": 6.504259834736587e-06, "loss": 0.8671, "step": 3210 }, { "epoch": 0.42, "grad_norm": 0.6294450759887695, "learning_rate": 6.502235891325233e-06, "loss": 1.0254, "step": 3211 }, { "epoch": 0.42, "grad_norm": 0.6291966438293457, "learning_rate": 6.500211677306451e-06, "loss": 1.0206, "step": 3212 }, { "epoch": 0.42, "grad_norm": 1.3748674392700195, "learning_rate": 6.4981871930448795e-06, "loss": 0.8334, "step": 3213 }, { "epoch": 0.42, "grad_norm": 0.6170034408569336, "learning_rate": 6.496162438905198e-06, "loss": 1.0329, "step": 3214 }, { "epoch": 0.42, "grad_norm": 1.324426293373108, "learning_rate": 6.49413741525214e-06, "loss": 0.8345, "step": 3215 }, { "epoch": 0.42, "grad_norm": 1.341865062713623, "learning_rate": 6.492112122450487e-06, "loss": 0.8126, "step": 3216 }, { "epoch": 0.42, "grad_norm": 1.290331244468689, "learning_rate": 6.4900865608650666e-06, "loss": 0.8619, "step": 3217 }, { "epoch": 0.42, "grad_norm": 1.4014853239059448, "learning_rate": 6.488060730860757e-06, "loss": 0.7776, "step": 3218 }, { "epoch": 0.42, "grad_norm": 0.7874432802200317, "learning_rate": 6.486034632802484e-06, "loss": 1.0287, "step": 3219 }, { "epoch": 0.42, "grad_norm": 0.7762148380279541, "learning_rate": 6.48400826705522e-06, "loss": 1.0311, "step": 3220 }, { "epoch": 0.42, "grad_norm": 1.2876163721084595, "learning_rate": 6.481981633983989e-06, "loss": 0.8099, "step": 3221 }, { "epoch": 0.42, "grad_norm": 1.4704384803771973, "learning_rate": 6.479954733953864e-06, "loss": 0.86, "step": 3222 }, { "epoch": 0.42, "grad_norm": 1.3579204082489014, "learning_rate": 6.477927567329958e-06, "loss": 0.834, "step": 3223 }, { "epoch": 0.42, "grad_norm": 1.37025785446167, "learning_rate": 6.475900134477441e-06, "loss": 0.7776, "step": 3224 }, { "epoch": 0.42, "grad_norm": 0.8639868497848511, "learning_rate": 6.473872435761527e-06, "loss": 1.0185, "step": 3225 }, { "epoch": 0.42, "grad_norm": 1.3492709398269653, "learning_rate": 6.471844471547479e-06, "loss": 0.8169, "step": 3226 }, { "epoch": 0.42, "grad_norm": 1.335512399673462, "learning_rate": 6.469816242200606e-06, "loss": 0.8536, "step": 3227 }, { "epoch": 0.42, "grad_norm": 1.3927565813064575, "learning_rate": 6.467787748086266e-06, "loss": 0.8039, "step": 3228 }, { "epoch": 0.42, "grad_norm": 1.381166934967041, "learning_rate": 6.465758989569866e-06, "loss": 0.8999, "step": 3229 }, { "epoch": 0.42, "grad_norm": 0.7031058669090271, "learning_rate": 6.463729967016858e-06, "loss": 1.0562, "step": 3230 }, { "epoch": 0.42, "grad_norm": 0.6391482353210449, "learning_rate": 6.461700680792745e-06, "loss": 1.0571, "step": 3231 }, { "epoch": 0.42, "grad_norm": 1.3266435861587524, "learning_rate": 6.459671131263073e-06, "loss": 0.8153, "step": 3232 }, { "epoch": 0.42, "grad_norm": 1.4592418670654297, "learning_rate": 6.4576413187934394e-06, "loss": 0.8825, "step": 3233 }, { "epoch": 0.42, "grad_norm": 1.999729871749878, "learning_rate": 6.455611243749486e-06, "loss": 0.8203, "step": 3234 }, { "epoch": 0.42, "grad_norm": 1.604971170425415, "learning_rate": 6.453580906496905e-06, "loss": 0.8029, "step": 3235 }, { "epoch": 0.42, "grad_norm": 1.3187990188598633, "learning_rate": 6.451550307401432e-06, "loss": 0.8524, "step": 3236 }, { "epoch": 0.42, "grad_norm": 1.3451588153839111, "learning_rate": 6.449519446828855e-06, "loss": 0.8225, "step": 3237 }, { "epoch": 0.42, "grad_norm": 0.8727684617042542, "learning_rate": 6.4474883251450015e-06, "loss": 1.0357, "step": 3238 }, { "epoch": 0.42, "grad_norm": 1.2638819217681885, "learning_rate": 6.445456942715756e-06, "loss": 0.8112, "step": 3239 }, { "epoch": 0.42, "grad_norm": 1.649640679359436, "learning_rate": 6.4434252999070406e-06, "loss": 0.7558, "step": 3240 }, { "epoch": 0.42, "grad_norm": 1.3298883438110352, "learning_rate": 6.441393397084827e-06, "loss": 0.7938, "step": 3241 }, { "epoch": 0.42, "grad_norm": 0.7288616895675659, "learning_rate": 6.439361234615139e-06, "loss": 1.0207, "step": 3242 }, { "epoch": 0.42, "grad_norm": 1.4768553972244263, "learning_rate": 6.437328812864041e-06, "loss": 0.8439, "step": 3243 }, { "epoch": 0.43, "grad_norm": 1.2992362976074219, "learning_rate": 6.435296132197647e-06, "loss": 0.863, "step": 3244 }, { "epoch": 0.43, "grad_norm": 1.2810314893722534, "learning_rate": 6.433263192982116e-06, "loss": 0.8129, "step": 3245 }, { "epoch": 0.43, "grad_norm": 0.710665762424469, "learning_rate": 6.431229995583656e-06, "loss": 1.0431, "step": 3246 }, { "epoch": 0.43, "grad_norm": 1.3376959562301636, "learning_rate": 6.429196540368516e-06, "loss": 0.8081, "step": 3247 }, { "epoch": 0.43, "grad_norm": 1.4261306524276733, "learning_rate": 6.427162827702999e-06, "loss": 0.8169, "step": 3248 }, { "epoch": 0.43, "grad_norm": 1.352046251296997, "learning_rate": 6.425128857953451e-06, "loss": 0.8282, "step": 3249 }, { "epoch": 0.43, "grad_norm": 1.3785591125488281, "learning_rate": 6.423094631486264e-06, "loss": 0.8421, "step": 3250 }, { "epoch": 0.43, "grad_norm": 0.7034136652946472, "learning_rate": 6.421060148667874e-06, "loss": 1.0344, "step": 3251 }, { "epoch": 0.43, "grad_norm": 1.3613054752349854, "learning_rate": 6.419025409864766e-06, "loss": 0.8716, "step": 3252 }, { "epoch": 0.43, "grad_norm": 1.3814903497695923, "learning_rate": 6.416990415443471e-06, "loss": 0.8778, "step": 3253 }, { "epoch": 0.43, "grad_norm": 1.3692011833190918, "learning_rate": 6.414955165770569e-06, "loss": 0.797, "step": 3254 }, { "epoch": 0.43, "grad_norm": 1.3446727991104126, "learning_rate": 6.4129196612126765e-06, "loss": 0.7739, "step": 3255 }, { "epoch": 0.43, "grad_norm": 0.6221069097518921, "learning_rate": 6.4108839021364665e-06, "loss": 1.0126, "step": 3256 }, { "epoch": 0.43, "grad_norm": 1.44303560256958, "learning_rate": 6.4088478889086525e-06, "loss": 0.7969, "step": 3257 }, { "epoch": 0.43, "grad_norm": 1.534533143043518, "learning_rate": 6.406811621895994e-06, "loss": 0.778, "step": 3258 }, { "epoch": 0.43, "grad_norm": 1.4505834579467773, "learning_rate": 6.404775101465298e-06, "loss": 0.8167, "step": 3259 }, { "epoch": 0.43, "grad_norm": 1.310742735862732, "learning_rate": 6.402738327983413e-06, "loss": 0.7867, "step": 3260 }, { "epoch": 0.43, "grad_norm": 0.6623648405075073, "learning_rate": 6.40070130181724e-06, "loss": 1.0203, "step": 3261 }, { "epoch": 0.43, "grad_norm": 1.25936758518219, "learning_rate": 6.398664023333719e-06, "loss": 0.7562, "step": 3262 }, { "epoch": 0.43, "grad_norm": 1.3637988567352295, "learning_rate": 6.39662649289984e-06, "loss": 0.8976, "step": 3263 }, { "epoch": 0.43, "grad_norm": 1.3535276651382446, "learning_rate": 6.394588710882635e-06, "loss": 0.8124, "step": 3264 }, { "epoch": 0.43, "grad_norm": 0.6456485390663147, "learning_rate": 6.392550677649183e-06, "loss": 1.0474, "step": 3265 }, { "epoch": 0.43, "grad_norm": 0.6338167786598206, "learning_rate": 6.390512393566609e-06, "loss": 1.0146, "step": 3266 }, { "epoch": 0.43, "grad_norm": 0.5927530527114868, "learning_rate": 6.388473859002082e-06, "loss": 1.0242, "step": 3267 }, { "epoch": 0.43, "grad_norm": 1.3792439699172974, "learning_rate": 6.3864350743228165e-06, "loss": 0.8292, "step": 3268 }, { "epoch": 0.43, "grad_norm": 0.6262499690055847, "learning_rate": 6.38439603989607e-06, "loss": 1.0506, "step": 3269 }, { "epoch": 0.43, "grad_norm": 1.338195562362671, "learning_rate": 6.382356756089149e-06, "loss": 0.7955, "step": 3270 }, { "epoch": 0.43, "grad_norm": 0.6228491067886353, "learning_rate": 6.380317223269401e-06, "loss": 1.048, "step": 3271 }, { "epoch": 0.43, "grad_norm": 1.4320586919784546, "learning_rate": 6.3782774418042216e-06, "loss": 0.8789, "step": 3272 }, { "epoch": 0.43, "grad_norm": 1.238569736480713, "learning_rate": 6.376237412061051e-06, "loss": 0.8041, "step": 3273 }, { "epoch": 0.43, "grad_norm": 1.3326939344406128, "learning_rate": 6.37419713440737e-06, "loss": 0.7923, "step": 3274 }, { "epoch": 0.43, "grad_norm": 1.4015470743179321, "learning_rate": 6.372156609210709e-06, "loss": 0.814, "step": 3275 }, { "epoch": 0.43, "grad_norm": 1.457828402519226, "learning_rate": 6.370115836838642e-06, "loss": 0.8083, "step": 3276 }, { "epoch": 0.43, "grad_norm": 0.8969069719314575, "learning_rate": 6.368074817658783e-06, "loss": 1.0268, "step": 3277 }, { "epoch": 0.43, "grad_norm": 1.6287429332733154, "learning_rate": 6.366033552038796e-06, "loss": 0.8929, "step": 3278 }, { "epoch": 0.43, "grad_norm": 1.3739510774612427, "learning_rate": 6.363992040346388e-06, "loss": 0.8874, "step": 3279 }, { "epoch": 0.43, "grad_norm": 1.3447799682617188, "learning_rate": 6.361950282949306e-06, "loss": 0.8173, "step": 3280 }, { "epoch": 0.43, "grad_norm": 1.3969242572784424, "learning_rate": 6.359908280215349e-06, "loss": 0.8433, "step": 3281 }, { "epoch": 0.43, "grad_norm": 0.7040032148361206, "learning_rate": 6.357866032512355e-06, "loss": 1.0308, "step": 3282 }, { "epoch": 0.43, "grad_norm": 1.2270593643188477, "learning_rate": 6.355823540208207e-06, "loss": 0.8275, "step": 3283 }, { "epoch": 0.43, "grad_norm": 1.318561315536499, "learning_rate": 6.3537808036708314e-06, "loss": 0.8531, "step": 3284 }, { "epoch": 0.43, "grad_norm": 1.4604517221450806, "learning_rate": 6.3517378232682005e-06, "loss": 0.8284, "step": 3285 }, { "epoch": 0.43, "grad_norm": 1.4601794481277466, "learning_rate": 6.34969459936833e-06, "loss": 0.8169, "step": 3286 }, { "epoch": 0.43, "grad_norm": 1.3990041017532349, "learning_rate": 6.347651132339279e-06, "loss": 0.7858, "step": 3287 }, { "epoch": 0.43, "grad_norm": 0.683517336845398, "learning_rate": 6.345607422549148e-06, "loss": 1.0282, "step": 3288 }, { "epoch": 0.43, "grad_norm": 1.2905068397521973, "learning_rate": 6.3435634703660855e-06, "loss": 0.7944, "step": 3289 }, { "epoch": 0.43, "grad_norm": 1.3019174337387085, "learning_rate": 6.341519276158282e-06, "loss": 0.8258, "step": 3290 }, { "epoch": 0.43, "grad_norm": 1.4439465999603271, "learning_rate": 6.339474840293971e-06, "loss": 0.8087, "step": 3291 }, { "epoch": 0.43, "grad_norm": 1.4316948652267456, "learning_rate": 6.337430163141431e-06, "loss": 0.8611, "step": 3292 }, { "epoch": 0.43, "grad_norm": 1.3415764570236206, "learning_rate": 6.335385245068979e-06, "loss": 0.7927, "step": 3293 }, { "epoch": 0.43, "grad_norm": 1.4662672281265259, "learning_rate": 6.333340086444984e-06, "loss": 0.8599, "step": 3294 }, { "epoch": 0.43, "grad_norm": 0.7206083536148071, "learning_rate": 6.33129468763785e-06, "loss": 1.0561, "step": 3295 }, { "epoch": 0.43, "grad_norm": 1.4158295392990112, "learning_rate": 6.329249049016032e-06, "loss": 0.8925, "step": 3296 }, { "epoch": 0.43, "grad_norm": 0.6528066396713257, "learning_rate": 6.32720317094802e-06, "loss": 1.0539, "step": 3297 }, { "epoch": 0.43, "grad_norm": 0.6362607479095459, "learning_rate": 6.325157053802354e-06, "loss": 1.0059, "step": 3298 }, { "epoch": 0.43, "grad_norm": 1.5277903079986572, "learning_rate": 6.323110697947611e-06, "loss": 0.7718, "step": 3299 }, { "epoch": 0.43, "grad_norm": 1.294355034828186, "learning_rate": 6.321064103752415e-06, "loss": 0.7701, "step": 3300 }, { "epoch": 0.43, "grad_norm": 2.4681155681610107, "learning_rate": 6.319017271585437e-06, "loss": 0.8791, "step": 3301 }, { "epoch": 0.43, "grad_norm": 1.330368161201477, "learning_rate": 6.31697020181538e-06, "loss": 0.8109, "step": 3302 }, { "epoch": 0.43, "grad_norm": 1.4085710048675537, "learning_rate": 6.314922894810999e-06, "loss": 0.8494, "step": 3303 }, { "epoch": 0.43, "grad_norm": 0.9782917499542236, "learning_rate": 6.312875350941088e-06, "loss": 1.0263, "step": 3304 }, { "epoch": 0.43, "grad_norm": 0.7896094918251038, "learning_rate": 6.310827570574484e-06, "loss": 1.0291, "step": 3305 }, { "epoch": 0.43, "grad_norm": 0.631192684173584, "learning_rate": 6.3087795540800664e-06, "loss": 1.055, "step": 3306 }, { "epoch": 0.43, "grad_norm": 1.688755989074707, "learning_rate": 6.306731301826759e-06, "loss": 0.8112, "step": 3307 }, { "epoch": 0.43, "grad_norm": 0.8476356267929077, "learning_rate": 6.304682814183526e-06, "loss": 1.0318, "step": 3308 }, { "epoch": 0.43, "grad_norm": 0.8669450283050537, "learning_rate": 6.302634091519374e-06, "loss": 1.0456, "step": 3309 }, { "epoch": 0.43, "grad_norm": 1.3388358354568481, "learning_rate": 6.300585134203354e-06, "loss": 0.7646, "step": 3310 }, { "epoch": 0.43, "grad_norm": 1.2789109945297241, "learning_rate": 6.298535942604556e-06, "loss": 0.7897, "step": 3311 }, { "epoch": 0.43, "grad_norm": 1.2523869276046753, "learning_rate": 6.296486517092116e-06, "loss": 0.8065, "step": 3312 }, { "epoch": 0.43, "grad_norm": 1.3447215557098389, "learning_rate": 6.2944368580352085e-06, "loss": 0.7751, "step": 3313 }, { "epoch": 0.43, "grad_norm": 0.7826550006866455, "learning_rate": 6.292386965803054e-06, "loss": 1.0462, "step": 3314 }, { "epoch": 0.43, "grad_norm": 0.7210400104522705, "learning_rate": 6.290336840764913e-06, "loss": 1.0152, "step": 3315 }, { "epoch": 0.43, "grad_norm": 1.2935162782669067, "learning_rate": 6.288286483290082e-06, "loss": 0.7883, "step": 3316 }, { "epoch": 0.43, "grad_norm": 1.2939590215682983, "learning_rate": 6.286235893747913e-06, "loss": 0.8139, "step": 3317 }, { "epoch": 0.43, "grad_norm": 1.3251445293426514, "learning_rate": 6.284185072507786e-06, "loss": 0.8383, "step": 3318 }, { "epoch": 0.43, "grad_norm": 1.388974905014038, "learning_rate": 6.28213401993913e-06, "loss": 0.8354, "step": 3319 }, { "epoch": 0.44, "grad_norm": 1.4946433305740356, "learning_rate": 6.280082736411418e-06, "loss": 0.819, "step": 3320 }, { "epoch": 0.44, "grad_norm": 1.3067846298217773, "learning_rate": 6.278031222294157e-06, "loss": 0.797, "step": 3321 }, { "epoch": 0.44, "grad_norm": 1.345949649810791, "learning_rate": 6.275979477956899e-06, "loss": 0.8127, "step": 3322 }, { "epoch": 0.44, "grad_norm": 1.4739019870758057, "learning_rate": 6.2739275037692394e-06, "loss": 0.8375, "step": 3323 }, { "epoch": 0.44, "grad_norm": 1.6420999765396118, "learning_rate": 6.2718753001008145e-06, "loss": 0.8217, "step": 3324 }, { "epoch": 0.44, "grad_norm": 1.3843618631362915, "learning_rate": 6.2698228673213005e-06, "loss": 0.7799, "step": 3325 }, { "epoch": 0.44, "grad_norm": 1.8590850830078125, "learning_rate": 6.2677702058004145e-06, "loss": 0.8714, "step": 3326 }, { "epoch": 0.44, "grad_norm": 1.2797625064849854, "learning_rate": 6.265717315907915e-06, "loss": 0.7834, "step": 3327 }, { "epoch": 0.44, "grad_norm": 0.8991274833679199, "learning_rate": 6.263664198013604e-06, "loss": 1.0454, "step": 3328 }, { "epoch": 0.44, "grad_norm": 1.4311339855194092, "learning_rate": 6.261610852487323e-06, "loss": 0.7998, "step": 3329 }, { "epoch": 0.44, "grad_norm": 1.2735165357589722, "learning_rate": 6.259557279698952e-06, "loss": 0.8401, "step": 3330 }, { "epoch": 0.44, "grad_norm": 1.3959288597106934, "learning_rate": 6.257503480018417e-06, "loss": 0.7952, "step": 3331 }, { "epoch": 0.44, "grad_norm": 1.4218652248382568, "learning_rate": 6.255449453815679e-06, "loss": 0.7981, "step": 3332 }, { "epoch": 0.44, "grad_norm": 1.3976161479949951, "learning_rate": 6.253395201460746e-06, "loss": 0.753, "step": 3333 }, { "epoch": 0.44, "grad_norm": 1.5426791906356812, "learning_rate": 6.251340723323663e-06, "loss": 0.9075, "step": 3334 }, { "epoch": 0.44, "grad_norm": 0.8787147402763367, "learning_rate": 6.249286019774517e-06, "loss": 1.014, "step": 3335 }, { "epoch": 0.44, "grad_norm": 1.271324872970581, "learning_rate": 6.247231091183432e-06, "loss": 0.8521, "step": 3336 }, { "epoch": 0.44, "grad_norm": 1.3454504013061523, "learning_rate": 6.245175937920579e-06, "loss": 0.8583, "step": 3337 }, { "epoch": 0.44, "grad_norm": 1.3622400760650635, "learning_rate": 6.2431205603561626e-06, "loss": 0.8282, "step": 3338 }, { "epoch": 0.44, "grad_norm": 1.3514363765716553, "learning_rate": 6.241064958860435e-06, "loss": 0.8861, "step": 3339 }, { "epoch": 0.44, "grad_norm": 1.379241704940796, "learning_rate": 6.239009133803682e-06, "loss": 0.8006, "step": 3340 }, { "epoch": 0.44, "grad_norm": 1.291581153869629, "learning_rate": 6.2369530855562335e-06, "loss": 0.7671, "step": 3341 }, { "epoch": 0.44, "grad_norm": 0.8399065732955933, "learning_rate": 6.234896814488459e-06, "loss": 1.042, "step": 3342 }, { "epoch": 0.44, "grad_norm": 1.3275765180587769, "learning_rate": 6.232840320970769e-06, "loss": 0.8579, "step": 3343 }, { "epoch": 0.44, "grad_norm": 0.6625158190727234, "learning_rate": 6.2307836053736096e-06, "loss": 1.0223, "step": 3344 }, { "epoch": 0.44, "grad_norm": 1.6785231828689575, "learning_rate": 6.228726668067473e-06, "loss": 0.8063, "step": 3345 }, { "epoch": 0.44, "grad_norm": 1.3462579250335693, "learning_rate": 6.2266695094228866e-06, "loss": 0.8096, "step": 3346 }, { "epoch": 0.44, "grad_norm": 1.3516314029693604, "learning_rate": 6.224612129810422e-06, "loss": 0.825, "step": 3347 }, { "epoch": 0.44, "grad_norm": 0.8151007890701294, "learning_rate": 6.222554529600689e-06, "loss": 1.0338, "step": 3348 }, { "epoch": 0.44, "grad_norm": 1.3664895296096802, "learning_rate": 6.220496709164332e-06, "loss": 0.8472, "step": 3349 }, { "epoch": 0.44, "grad_norm": 1.342450737953186, "learning_rate": 6.218438668872042e-06, "loss": 0.8494, "step": 3350 }, { "epoch": 0.44, "grad_norm": 1.6772407293319702, "learning_rate": 6.2163804090945465e-06, "loss": 0.8314, "step": 3351 }, { "epoch": 0.44, "grad_norm": 1.324462652206421, "learning_rate": 6.214321930202615e-06, "loss": 0.8548, "step": 3352 }, { "epoch": 0.44, "grad_norm": 1.383480429649353, "learning_rate": 6.212263232567053e-06, "loss": 0.8374, "step": 3353 }, { "epoch": 0.44, "grad_norm": 1.350412130355835, "learning_rate": 6.210204316558707e-06, "loss": 0.8555, "step": 3354 }, { "epoch": 0.44, "grad_norm": 1.4329869747161865, "learning_rate": 6.208145182548462e-06, "loss": 0.833, "step": 3355 }, { "epoch": 0.44, "grad_norm": 1.6140118837356567, "learning_rate": 6.2060858309072445e-06, "loss": 0.8009, "step": 3356 }, { "epoch": 0.44, "grad_norm": 1.2510898113250732, "learning_rate": 6.204026262006017e-06, "loss": 0.8219, "step": 3357 }, { "epoch": 0.44, "grad_norm": 0.7338991165161133, "learning_rate": 6.201966476215788e-06, "loss": 1.0516, "step": 3358 }, { "epoch": 0.44, "grad_norm": 0.6930378675460815, "learning_rate": 6.199906473907593e-06, "loss": 1.0436, "step": 3359 }, { "epoch": 0.44, "grad_norm": 0.618396520614624, "learning_rate": 6.197846255452516e-06, "loss": 1.0287, "step": 3360 }, { "epoch": 0.44, "grad_norm": 1.5938575267791748, "learning_rate": 6.195785821221679e-06, "loss": 0.8371, "step": 3361 }, { "epoch": 0.44, "grad_norm": 0.6495134234428406, "learning_rate": 6.193725171586241e-06, "loss": 1.0478, "step": 3362 }, { "epoch": 0.44, "grad_norm": 1.42080819606781, "learning_rate": 6.191664306917399e-06, "loss": 0.8765, "step": 3363 }, { "epoch": 0.44, "grad_norm": 1.5270957946777344, "learning_rate": 6.189603227586389e-06, "loss": 0.859, "step": 3364 }, { "epoch": 0.44, "grad_norm": 1.3291418552398682, "learning_rate": 6.187541933964489e-06, "loss": 0.8246, "step": 3365 }, { "epoch": 0.44, "grad_norm": 1.3282158374786377, "learning_rate": 6.18548042642301e-06, "loss": 0.7766, "step": 3366 }, { "epoch": 0.44, "grad_norm": 1.3800276517868042, "learning_rate": 6.1834187053333095e-06, "loss": 0.854, "step": 3367 }, { "epoch": 0.44, "grad_norm": 1.3733588457107544, "learning_rate": 6.181356771066773e-06, "loss": 0.7874, "step": 3368 }, { "epoch": 0.44, "grad_norm": 1.3036857843399048, "learning_rate": 6.179294623994833e-06, "loss": 0.8211, "step": 3369 }, { "epoch": 0.44, "grad_norm": 1.3688148260116577, "learning_rate": 6.177232264488958e-06, "loss": 0.7864, "step": 3370 }, { "epoch": 0.44, "grad_norm": 1.7607518434524536, "learning_rate": 6.175169692920653e-06, "loss": 0.8911, "step": 3371 }, { "epoch": 0.44, "grad_norm": 1.4063549041748047, "learning_rate": 6.173106909661461e-06, "loss": 0.8796, "step": 3372 }, { "epoch": 0.44, "grad_norm": 1.2794535160064697, "learning_rate": 6.171043915082967e-06, "loss": 0.8045, "step": 3373 }, { "epoch": 0.44, "grad_norm": 1.3572874069213867, "learning_rate": 6.168980709556789e-06, "loss": 0.8162, "step": 3374 }, { "epoch": 0.44, "grad_norm": 1.3308416604995728, "learning_rate": 6.1669172934545875e-06, "loss": 0.7761, "step": 3375 }, { "epoch": 0.44, "grad_norm": 1.388364315032959, "learning_rate": 6.164853667148059e-06, "loss": 0.8665, "step": 3376 }, { "epoch": 0.44, "grad_norm": 1.3668382167816162, "learning_rate": 6.162789831008935e-06, "loss": 0.8347, "step": 3377 }, { "epoch": 0.44, "grad_norm": 1.3325508832931519, "learning_rate": 6.160725785408991e-06, "loss": 0.8065, "step": 3378 }, { "epoch": 0.44, "grad_norm": 1.29705011844635, "learning_rate": 6.158661530720035e-06, "loss": 0.8499, "step": 3379 }, { "epoch": 0.44, "grad_norm": 0.9400076866149902, "learning_rate": 6.156597067313914e-06, "loss": 1.0362, "step": 3380 }, { "epoch": 0.44, "grad_norm": 1.5152268409729004, "learning_rate": 6.154532395562515e-06, "loss": 0.8099, "step": 3381 }, { "epoch": 0.44, "grad_norm": 0.6963512301445007, "learning_rate": 6.152467515837759e-06, "loss": 1.0394, "step": 3382 }, { "epoch": 0.44, "grad_norm": 1.3132047653198242, "learning_rate": 6.150402428511607e-06, "loss": 0.8027, "step": 3383 }, { "epoch": 0.44, "grad_norm": 1.7174354791641235, "learning_rate": 6.1483371339560545e-06, "loss": 0.7876, "step": 3384 }, { "epoch": 0.44, "grad_norm": 0.7634620666503906, "learning_rate": 6.146271632543142e-06, "loss": 1.015, "step": 3385 }, { "epoch": 0.44, "grad_norm": 0.7493101954460144, "learning_rate": 6.144205924644934e-06, "loss": 1.0226, "step": 3386 }, { "epoch": 0.44, "grad_norm": 1.325758934020996, "learning_rate": 6.1421400106335435e-06, "loss": 0.7642, "step": 3387 }, { "epoch": 0.44, "grad_norm": 0.6514648199081421, "learning_rate": 6.140073890881117e-06, "loss": 1.0148, "step": 3388 }, { "epoch": 0.44, "grad_norm": 1.345051884651184, "learning_rate": 6.138007565759838e-06, "loss": 0.8206, "step": 3389 }, { "epoch": 0.44, "grad_norm": 1.517386555671692, "learning_rate": 6.135941035641927e-06, "loss": 0.7771, "step": 3390 }, { "epoch": 0.44, "grad_norm": 0.9065435528755188, "learning_rate": 6.133874300899641e-06, "loss": 1.0309, "step": 3391 }, { "epoch": 0.44, "grad_norm": 1.4120314121246338, "learning_rate": 6.1318073619052735e-06, "loss": 0.8063, "step": 3392 }, { "epoch": 0.44, "grad_norm": 1.4328341484069824, "learning_rate": 6.129740219031156e-06, "loss": 0.8282, "step": 3393 }, { "epoch": 0.44, "grad_norm": 1.3102046251296997, "learning_rate": 6.127672872649657e-06, "loss": 0.8077, "step": 3394 }, { "epoch": 0.44, "grad_norm": 0.7246078252792358, "learning_rate": 6.12560532313318e-06, "loss": 1.0448, "step": 3395 }, { "epoch": 0.44, "grad_norm": 0.7243000268936157, "learning_rate": 6.123537570854168e-06, "loss": 1.0389, "step": 3396 }, { "epoch": 0.45, "grad_norm": 1.2391215562820435, "learning_rate": 6.121469616185096e-06, "loss": 0.8361, "step": 3397 }, { "epoch": 0.45, "grad_norm": 1.3430075645446777, "learning_rate": 6.119401459498479e-06, "loss": 0.8427, "step": 3398 }, { "epoch": 0.45, "grad_norm": 2.0286598205566406, "learning_rate": 6.117333101166867e-06, "loss": 0.8045, "step": 3399 }, { "epoch": 0.45, "grad_norm": 0.7114830613136292, "learning_rate": 6.115264541562847e-06, "loss": 1.0378, "step": 3400 }, { "epoch": 0.45, "grad_norm": 1.3307654857635498, "learning_rate": 6.1131957810590425e-06, "loss": 0.8426, "step": 3401 }, { "epoch": 0.45, "grad_norm": 1.5030142068862915, "learning_rate": 6.1111268200281115e-06, "loss": 0.8154, "step": 3402 }, { "epoch": 0.45, "grad_norm": 1.313955307006836, "learning_rate": 6.1090576588427495e-06, "loss": 0.7848, "step": 3403 }, { "epoch": 0.45, "grad_norm": 1.3088035583496094, "learning_rate": 6.106988297875691e-06, "loss": 0.8026, "step": 3404 }, { "epoch": 0.45, "grad_norm": 1.315166711807251, "learning_rate": 6.104918737499699e-06, "loss": 0.7781, "step": 3405 }, { "epoch": 0.45, "grad_norm": 1.3600138425827026, "learning_rate": 6.102848978087576e-06, "loss": 0.7881, "step": 3406 }, { "epoch": 0.45, "grad_norm": 1.2426856756210327, "learning_rate": 6.100779020012167e-06, "loss": 0.8182, "step": 3407 }, { "epoch": 0.45, "grad_norm": 0.7555664777755737, "learning_rate": 6.0987088636463406e-06, "loss": 1.0252, "step": 3408 }, { "epoch": 0.45, "grad_norm": 1.5038059949874878, "learning_rate": 6.096638509363013e-06, "loss": 0.8288, "step": 3409 }, { "epoch": 0.45, "grad_norm": 1.293471097946167, "learning_rate": 6.0945679575351255e-06, "loss": 0.8344, "step": 3410 }, { "epoch": 0.45, "grad_norm": 1.3606244325637817, "learning_rate": 6.092497208535662e-06, "loss": 0.8156, "step": 3411 }, { "epoch": 0.45, "grad_norm": 1.5609970092773438, "learning_rate": 6.090426262737639e-06, "loss": 0.9063, "step": 3412 }, { "epoch": 0.45, "grad_norm": 1.4405966997146606, "learning_rate": 6.088355120514113e-06, "loss": 0.8148, "step": 3413 }, { "epoch": 0.45, "grad_norm": 1.4339927434921265, "learning_rate": 6.086283782238168e-06, "loss": 0.8677, "step": 3414 }, { "epoch": 0.45, "grad_norm": 1.4183459281921387, "learning_rate": 6.084212248282928e-06, "loss": 0.7823, "step": 3415 }, { "epoch": 0.45, "grad_norm": 1.3340709209442139, "learning_rate": 6.0821405190215535e-06, "loss": 0.8125, "step": 3416 }, { "epoch": 0.45, "grad_norm": 1.6586419343948364, "learning_rate": 6.080068594827238e-06, "loss": 0.8442, "step": 3417 }, { "epoch": 0.45, "grad_norm": 1.4311574697494507, "learning_rate": 6.0779964760732095e-06, "loss": 0.8039, "step": 3418 }, { "epoch": 0.45, "grad_norm": 1.9793933629989624, "learning_rate": 6.075924163132734e-06, "loss": 0.7701, "step": 3419 }, { "epoch": 0.45, "grad_norm": 0.774562656879425, "learning_rate": 6.073851656379107e-06, "loss": 1.0338, "step": 3420 }, { "epoch": 0.45, "grad_norm": 1.2647894620895386, "learning_rate": 6.071778956185666e-06, "loss": 0.8038, "step": 3421 }, { "epoch": 0.45, "grad_norm": 1.3197096586227417, "learning_rate": 6.069706062925777e-06, "loss": 0.8437, "step": 3422 }, { "epoch": 0.45, "grad_norm": 1.3986598253250122, "learning_rate": 6.067632976972848e-06, "loss": 0.8147, "step": 3423 }, { "epoch": 0.45, "grad_norm": 1.5399433374404907, "learning_rate": 6.065559698700312e-06, "loss": 0.7955, "step": 3424 }, { "epoch": 0.45, "grad_norm": 0.6291626691818237, "learning_rate": 6.063486228481645e-06, "loss": 1.0165, "step": 3425 }, { "epoch": 0.45, "grad_norm": 0.6336756348609924, "learning_rate": 6.061412566690355e-06, "loss": 1.0303, "step": 3426 }, { "epoch": 0.45, "grad_norm": 1.7807645797729492, "learning_rate": 6.059338713699983e-06, "loss": 0.7903, "step": 3427 }, { "epoch": 0.45, "grad_norm": 2.176802635192871, "learning_rate": 6.057264669884105e-06, "loss": 0.8491, "step": 3428 }, { "epoch": 0.45, "grad_norm": 0.6384469270706177, "learning_rate": 6.055190435616333e-06, "loss": 1.0608, "step": 3429 }, { "epoch": 0.45, "grad_norm": 1.2861963510513306, "learning_rate": 6.053116011270311e-06, "loss": 0.8532, "step": 3430 }, { "epoch": 0.45, "grad_norm": 0.6439231038093567, "learning_rate": 6.051041397219719e-06, "loss": 1.0411, "step": 3431 }, { "epoch": 0.45, "grad_norm": 1.5067089796066284, "learning_rate": 6.048966593838272e-06, "loss": 0.8147, "step": 3432 }, { "epoch": 0.45, "grad_norm": 1.3961549997329712, "learning_rate": 6.0468916014997156e-06, "loss": 0.7886, "step": 3433 }, { "epoch": 0.45, "grad_norm": 1.654340147972107, "learning_rate": 6.0448164205778315e-06, "loss": 0.859, "step": 3434 }, { "epoch": 0.45, "grad_norm": 1.3194180727005005, "learning_rate": 6.042741051446437e-06, "loss": 0.8672, "step": 3435 }, { "epoch": 0.45, "grad_norm": 1.3328090906143188, "learning_rate": 6.040665494479381e-06, "loss": 0.9181, "step": 3436 }, { "epoch": 0.45, "grad_norm": 1.2703129053115845, "learning_rate": 6.038589750050546e-06, "loss": 0.8009, "step": 3437 }, { "epoch": 0.45, "grad_norm": 0.7295884490013123, "learning_rate": 6.036513818533852e-06, "loss": 1.0119, "step": 3438 }, { "epoch": 0.45, "grad_norm": 1.3430297374725342, "learning_rate": 6.034437700303246e-06, "loss": 0.7503, "step": 3439 }, { "epoch": 0.45, "grad_norm": 0.6762939095497131, "learning_rate": 6.0323613957327156e-06, "loss": 1.0292, "step": 3440 }, { "epoch": 0.45, "grad_norm": 1.3392648696899414, "learning_rate": 6.0302849051962765e-06, "loss": 0.799, "step": 3441 }, { "epoch": 0.45, "grad_norm": 1.4510736465454102, "learning_rate": 6.028208229067983e-06, "loss": 0.852, "step": 3442 }, { "epoch": 0.45, "grad_norm": 1.5267901420593262, "learning_rate": 6.026131367721919e-06, "loss": 0.7913, "step": 3443 }, { "epoch": 0.45, "grad_norm": 1.3665494918823242, "learning_rate": 6.024054321532201e-06, "loss": 0.8577, "step": 3444 }, { "epoch": 0.45, "grad_norm": 1.310630202293396, "learning_rate": 6.021977090872983e-06, "loss": 0.8022, "step": 3445 }, { "epoch": 0.45, "grad_norm": 1.3336774110794067, "learning_rate": 6.0198996761184515e-06, "loss": 0.7843, "step": 3446 }, { "epoch": 0.45, "grad_norm": 1.27853524684906, "learning_rate": 6.017822077642821e-06, "loss": 0.8788, "step": 3447 }, { "epoch": 0.45, "grad_norm": 2.2226247787475586, "learning_rate": 6.015744295820343e-06, "loss": 0.8217, "step": 3448 }, { "epoch": 0.45, "grad_norm": 1.285783052444458, "learning_rate": 6.013666331025303e-06, "loss": 0.7668, "step": 3449 }, { "epoch": 0.45, "grad_norm": 1.343577265739441, "learning_rate": 6.011588183632019e-06, "loss": 0.8119, "step": 3450 }, { "epoch": 0.45, "grad_norm": 1.3371126651763916, "learning_rate": 6.009509854014841e-06, "loss": 0.8518, "step": 3451 }, { "epoch": 0.45, "grad_norm": 1.35457181930542, "learning_rate": 6.0074313425481504e-06, "loss": 0.8405, "step": 3452 }, { "epoch": 0.45, "grad_norm": 1.3255585432052612, "learning_rate": 6.005352649606362e-06, "loss": 0.8106, "step": 3453 }, { "epoch": 0.45, "grad_norm": 1.3026137351989746, "learning_rate": 6.003273775563927e-06, "loss": 0.8132, "step": 3454 }, { "epoch": 0.45, "grad_norm": 1.3130903244018555, "learning_rate": 6.001194720795324e-06, "loss": 0.9074, "step": 3455 }, { "epoch": 0.45, "grad_norm": 1.6757898330688477, "learning_rate": 5.999115485675069e-06, "loss": 0.8037, "step": 3456 }, { "epoch": 0.45, "grad_norm": 2.455749988555908, "learning_rate": 5.997036070577705e-06, "loss": 0.8, "step": 3457 }, { "epoch": 0.45, "grad_norm": 3.9813475608825684, "learning_rate": 5.994956475877814e-06, "loss": 0.826, "step": 3458 }, { "epoch": 0.45, "grad_norm": 1.3809030055999756, "learning_rate": 5.992876701950003e-06, "loss": 0.8202, "step": 3459 }, { "epoch": 0.45, "grad_norm": 1.32774817943573, "learning_rate": 5.990796749168919e-06, "loss": 0.8028, "step": 3460 }, { "epoch": 0.45, "grad_norm": 1.3394927978515625, "learning_rate": 5.988716617909233e-06, "loss": 0.836, "step": 3461 }, { "epoch": 0.45, "grad_norm": 1.3459781408309937, "learning_rate": 5.986636308545655e-06, "loss": 0.8933, "step": 3462 }, { "epoch": 0.45, "grad_norm": 1.351047396659851, "learning_rate": 5.984555821452923e-06, "loss": 0.8228, "step": 3463 }, { "epoch": 0.45, "grad_norm": 0.9066408276557922, "learning_rate": 5.982475157005812e-06, "loss": 1.0364, "step": 3464 }, { "epoch": 0.45, "grad_norm": 0.753197431564331, "learning_rate": 5.9803943155791225e-06, "loss": 1.0358, "step": 3465 }, { "epoch": 0.45, "grad_norm": 0.6539208292961121, "learning_rate": 5.978313297547688e-06, "loss": 1.0103, "step": 3466 }, { "epoch": 0.45, "grad_norm": 1.46469247341156, "learning_rate": 5.9762321032863816e-06, "loss": 0.8433, "step": 3467 }, { "epoch": 0.45, "grad_norm": 1.4199100732803345, "learning_rate": 5.974150733170096e-06, "loss": 0.8248, "step": 3468 }, { "epoch": 0.45, "grad_norm": 1.3588638305664062, "learning_rate": 5.972069187573766e-06, "loss": 0.7801, "step": 3469 }, { "epoch": 0.45, "grad_norm": 1.348817229270935, "learning_rate": 5.969987466872354e-06, "loss": 0.8328, "step": 3470 }, { "epoch": 0.45, "grad_norm": 1.274083137512207, "learning_rate": 5.967905571440849e-06, "loss": 0.8433, "step": 3471 }, { "epoch": 0.45, "grad_norm": 1.3519933223724365, "learning_rate": 5.9658235016542806e-06, "loss": 0.8145, "step": 3472 }, { "epoch": 0.46, "grad_norm": 1.3305457830429077, "learning_rate": 5.963741257887704e-06, "loss": 0.7827, "step": 3473 }, { "epoch": 0.46, "grad_norm": 1.4223048686981201, "learning_rate": 5.961658840516209e-06, "loss": 0.8717, "step": 3474 }, { "epoch": 0.46, "grad_norm": 1.3691673278808594, "learning_rate": 5.959576249914913e-06, "loss": 0.7436, "step": 3475 }, { "epoch": 0.46, "grad_norm": 1.7387336492538452, "learning_rate": 5.957493486458965e-06, "loss": 1.0509, "step": 3476 }, { "epoch": 0.46, "grad_norm": 1.3514131307601929, "learning_rate": 5.95541055052355e-06, "loss": 0.8582, "step": 3477 }, { "epoch": 0.46, "grad_norm": 1.4235661029815674, "learning_rate": 5.953327442483878e-06, "loss": 0.8029, "step": 3478 }, { "epoch": 0.46, "grad_norm": 1.448529601097107, "learning_rate": 5.951244162715194e-06, "loss": 0.808, "step": 3479 }, { "epoch": 0.46, "grad_norm": 0.8217684030532837, "learning_rate": 5.949160711592773e-06, "loss": 1.0275, "step": 3480 }, { "epoch": 0.46, "grad_norm": 1.3901269435882568, "learning_rate": 5.947077089491919e-06, "loss": 0.7837, "step": 3481 }, { "epoch": 0.46, "grad_norm": 2.6162307262420654, "learning_rate": 5.9449932967879686e-06, "loss": 0.8592, "step": 3482 }, { "epoch": 0.46, "grad_norm": 0.7771351337432861, "learning_rate": 5.9429093338562895e-06, "loss": 1.0414, "step": 3483 }, { "epoch": 0.46, "grad_norm": 1.3429449796676636, "learning_rate": 5.94082520107228e-06, "loss": 0.7665, "step": 3484 }, { "epoch": 0.46, "grad_norm": 0.691087007522583, "learning_rate": 5.9387408988113656e-06, "loss": 1.0396, "step": 3485 }, { "epoch": 0.46, "grad_norm": 0.6359138488769531, "learning_rate": 5.936656427449009e-06, "loss": 1.0124, "step": 3486 }, { "epoch": 0.46, "grad_norm": 1.745116949081421, "learning_rate": 5.934571787360696e-06, "loss": 0.8175, "step": 3487 }, { "epoch": 0.46, "grad_norm": 4.5665059089660645, "learning_rate": 5.932486978921948e-06, "loss": 0.8252, "step": 3488 }, { "epoch": 0.46, "grad_norm": 1.3799591064453125, "learning_rate": 5.930402002508315e-06, "loss": 0.7914, "step": 3489 }, { "epoch": 0.46, "grad_norm": 1.3539732694625854, "learning_rate": 5.928316858495377e-06, "loss": 0.8455, "step": 3490 }, { "epoch": 0.46, "grad_norm": 1.441145658493042, "learning_rate": 5.926231547258743e-06, "loss": 0.7469, "step": 3491 }, { "epoch": 0.46, "grad_norm": 1.0655736923217773, "learning_rate": 5.924146069174055e-06, "loss": 1.0356, "step": 3492 }, { "epoch": 0.46, "grad_norm": 1.3451131582260132, "learning_rate": 5.922060424616985e-06, "loss": 0.7935, "step": 3493 }, { "epoch": 0.46, "grad_norm": 0.7669339179992676, "learning_rate": 5.91997461396323e-06, "loss": 1.0232, "step": 3494 }, { "epoch": 0.46, "grad_norm": 1.341977834701538, "learning_rate": 5.9178886375885235e-06, "loss": 0.8516, "step": 3495 }, { "epoch": 0.46, "grad_norm": 0.690029501914978, "learning_rate": 5.915802495868624e-06, "loss": 1.0401, "step": 3496 }, { "epoch": 0.46, "grad_norm": 1.2835488319396973, "learning_rate": 5.913716189179324e-06, "loss": 0.8237, "step": 3497 }, { "epoch": 0.46, "grad_norm": 0.8485487103462219, "learning_rate": 5.911629717896441e-06, "loss": 1.045, "step": 3498 }, { "epoch": 0.46, "grad_norm": 1.2942215204238892, "learning_rate": 5.909543082395826e-06, "loss": 0.8041, "step": 3499 }, { "epoch": 0.46, "grad_norm": 1.3065595626831055, "learning_rate": 5.907456283053357e-06, "loss": 0.8303, "step": 3500 }, { "epoch": 0.46, "grad_norm": 1.372756004333496, "learning_rate": 5.905369320244942e-06, "loss": 0.8448, "step": 3501 }, { "epoch": 0.46, "grad_norm": 1.3255903720855713, "learning_rate": 5.903282194346522e-06, "loss": 0.8189, "step": 3502 }, { "epoch": 0.46, "grad_norm": 1.3833034038543701, "learning_rate": 5.9011949057340624e-06, "loss": 0.808, "step": 3503 }, { "epoch": 0.46, "grad_norm": 1.2899292707443237, "learning_rate": 5.899107454783558e-06, "loss": 0.802, "step": 3504 }, { "epoch": 0.46, "grad_norm": 0.869970977306366, "learning_rate": 5.897019841871039e-06, "loss": 1.022, "step": 3505 }, { "epoch": 0.46, "grad_norm": 1.4362244606018066, "learning_rate": 5.894932067372556e-06, "loss": 0.8036, "step": 3506 }, { "epoch": 0.46, "grad_norm": 0.7393714189529419, "learning_rate": 5.892844131664199e-06, "loss": 1.0179, "step": 3507 }, { "epoch": 0.46, "grad_norm": 1.3386245965957642, "learning_rate": 5.890756035122076e-06, "loss": 0.884, "step": 3508 }, { "epoch": 0.46, "grad_norm": 0.6740692257881165, "learning_rate": 5.888667778122331e-06, "loss": 1.0566, "step": 3509 }, { "epoch": 0.46, "grad_norm": 0.6697698831558228, "learning_rate": 5.886579361041136e-06, "loss": 1.0231, "step": 3510 }, { "epoch": 0.46, "grad_norm": 1.3699710369110107, "learning_rate": 5.884490784254691e-06, "loss": 0.8177, "step": 3511 }, { "epoch": 0.46, "grad_norm": 0.6860848069190979, "learning_rate": 5.882402048139224e-06, "loss": 1.0317, "step": 3512 }, { "epoch": 0.46, "grad_norm": 1.3789589405059814, "learning_rate": 5.880313153070992e-06, "loss": 0.8641, "step": 3513 }, { "epoch": 0.46, "grad_norm": 1.4246124029159546, "learning_rate": 5.878224099426283e-06, "loss": 0.784, "step": 3514 }, { "epoch": 0.46, "grad_norm": 1.345728874206543, "learning_rate": 5.87613488758141e-06, "loss": 0.7575, "step": 3515 }, { "epoch": 0.46, "grad_norm": 1.3256776332855225, "learning_rate": 5.874045517912717e-06, "loss": 0.7734, "step": 3516 }, { "epoch": 0.46, "grad_norm": 1.5106977224349976, "learning_rate": 5.871955990796575e-06, "loss": 0.8504, "step": 3517 }, { "epoch": 0.46, "grad_norm": 1.3109921216964722, "learning_rate": 5.869866306609385e-06, "loss": 0.782, "step": 3518 }, { "epoch": 0.46, "grad_norm": 1.4421850442886353, "learning_rate": 5.867776465727574e-06, "loss": 0.8196, "step": 3519 }, { "epoch": 0.46, "grad_norm": 1.4866379499435425, "learning_rate": 5.865686468527599e-06, "loss": 0.8186, "step": 3520 }, { "epoch": 0.46, "grad_norm": 1.301222801208496, "learning_rate": 5.863596315385947e-06, "loss": 0.8525, "step": 3521 }, { "epoch": 0.46, "grad_norm": 0.8317051529884338, "learning_rate": 5.861506006679125e-06, "loss": 1.0324, "step": 3522 }, { "epoch": 0.46, "grad_norm": 0.7250036001205444, "learning_rate": 5.859415542783678e-06, "loss": 1.0198, "step": 3523 }, { "epoch": 0.46, "grad_norm": 1.4641340970993042, "learning_rate": 5.857324924076175e-06, "loss": 0.7873, "step": 3524 }, { "epoch": 0.46, "grad_norm": 1.365883469581604, "learning_rate": 5.8552341509332125e-06, "loss": 0.824, "step": 3525 }, { "epoch": 0.46, "grad_norm": 1.814273476600647, "learning_rate": 5.8531432237314135e-06, "loss": 0.7817, "step": 3526 }, { "epoch": 0.46, "grad_norm": 1.3965750932693481, "learning_rate": 5.85105214284743e-06, "loss": 0.8041, "step": 3527 }, { "epoch": 0.46, "grad_norm": 0.8868900537490845, "learning_rate": 5.8489609086579425e-06, "loss": 1.033, "step": 3528 }, { "epoch": 0.46, "grad_norm": 1.323270559310913, "learning_rate": 5.846869521539659e-06, "loss": 0.835, "step": 3529 }, { "epoch": 0.46, "grad_norm": 0.6773298978805542, "learning_rate": 5.844777981869315e-06, "loss": 1.0437, "step": 3530 }, { "epoch": 0.46, "grad_norm": 1.270607829093933, "learning_rate": 5.842686290023671e-06, "loss": 0.7975, "step": 3531 }, { "epoch": 0.46, "grad_norm": 1.5098727941513062, "learning_rate": 5.840594446379518e-06, "loss": 0.8714, "step": 3532 }, { "epoch": 0.46, "grad_norm": 1.2867287397384644, "learning_rate": 5.838502451313675e-06, "loss": 0.8252, "step": 3533 }, { "epoch": 0.46, "grad_norm": 1.4469044208526611, "learning_rate": 5.836410305202983e-06, "loss": 0.8443, "step": 3534 }, { "epoch": 0.46, "grad_norm": 1.318569302558899, "learning_rate": 5.834318008424316e-06, "loss": 0.7926, "step": 3535 }, { "epoch": 0.46, "grad_norm": 1.3592772483825684, "learning_rate": 5.832225561354573e-06, "loss": 0.7992, "step": 3536 }, { "epoch": 0.46, "grad_norm": 1.3645431995391846, "learning_rate": 5.8301329643706794e-06, "loss": 0.8116, "step": 3537 }, { "epoch": 0.46, "grad_norm": 1.3967509269714355, "learning_rate": 5.828040217849589e-06, "loss": 0.8624, "step": 3538 }, { "epoch": 0.46, "grad_norm": 0.9165219664573669, "learning_rate": 5.825947322168281e-06, "loss": 1.0229, "step": 3539 }, { "epoch": 0.46, "grad_norm": 1.457206130027771, "learning_rate": 5.823854277703761e-06, "loss": 0.8037, "step": 3540 }, { "epoch": 0.46, "grad_norm": 1.4392482042312622, "learning_rate": 5.821761084833064e-06, "loss": 0.7674, "step": 3541 }, { "epoch": 0.46, "grad_norm": 0.6759335994720459, "learning_rate": 5.819667743933251e-06, "loss": 1.0342, "step": 3542 }, { "epoch": 0.46, "grad_norm": 1.3982634544372559, "learning_rate": 5.8175742553814065e-06, "loss": 0.7702, "step": 3543 }, { "epoch": 0.46, "grad_norm": 1.500092625617981, "learning_rate": 5.815480619554646e-06, "loss": 0.819, "step": 3544 }, { "epoch": 0.46, "grad_norm": 1.3659428358078003, "learning_rate": 5.81338683683011e-06, "loss": 0.8007, "step": 3545 }, { "epoch": 0.46, "grad_norm": 1.35895836353302, "learning_rate": 5.811292907584963e-06, "loss": 0.8468, "step": 3546 }, { "epoch": 0.46, "grad_norm": 1.465488076210022, "learning_rate": 5.809198832196401e-06, "loss": 0.8202, "step": 3547 }, { "epoch": 0.46, "grad_norm": 1.3628437519073486, "learning_rate": 5.80710461104164e-06, "loss": 0.87, "step": 3548 }, { "epoch": 0.47, "grad_norm": 1.2993667125701904, "learning_rate": 5.805010244497928e-06, "loss": 0.8323, "step": 3549 }, { "epoch": 0.47, "grad_norm": 1.501900553703308, "learning_rate": 5.802915732942536e-06, "loss": 0.8189, "step": 3550 }, { "epoch": 0.47, "grad_norm": 1.388518214225769, "learning_rate": 5.800821076752761e-06, "loss": 0.7884, "step": 3551 }, { "epoch": 0.47, "grad_norm": 0.9831971526145935, "learning_rate": 5.798726276305927e-06, "loss": 1.0388, "step": 3552 }, { "epoch": 0.47, "grad_norm": 1.4261077642440796, "learning_rate": 5.796631331979385e-06, "loss": 0.8614, "step": 3553 }, { "epoch": 0.47, "grad_norm": 1.6250520944595337, "learning_rate": 5.794536244150512e-06, "loss": 0.7962, "step": 3554 }, { "epoch": 0.47, "grad_norm": 1.3109210729599, "learning_rate": 5.792441013196706e-06, "loss": 0.8498, "step": 3555 }, { "epoch": 0.47, "grad_norm": 1.414172887802124, "learning_rate": 5.790345639495396e-06, "loss": 0.8356, "step": 3556 }, { "epoch": 0.47, "grad_norm": 0.7623609900474548, "learning_rate": 5.788250123424037e-06, "loss": 1.0462, "step": 3557 }, { "epoch": 0.47, "grad_norm": 0.7828123569488525, "learning_rate": 5.786154465360107e-06, "loss": 1.0279, "step": 3558 }, { "epoch": 0.47, "grad_norm": 1.3761621713638306, "learning_rate": 5.784058665681108e-06, "loss": 0.8074, "step": 3559 }, { "epoch": 0.47, "grad_norm": 1.3298510313034058, "learning_rate": 5.781962724764572e-06, "loss": 0.7776, "step": 3560 }, { "epoch": 0.47, "grad_norm": 1.4422979354858398, "learning_rate": 5.779866642988055e-06, "loss": 0.7824, "step": 3561 }, { "epoch": 0.47, "grad_norm": 1.522512674331665, "learning_rate": 5.777770420729135e-06, "loss": 0.8169, "step": 3562 }, { "epoch": 0.47, "grad_norm": 0.7178623080253601, "learning_rate": 5.775674058365423e-06, "loss": 1.0164, "step": 3563 }, { "epoch": 0.47, "grad_norm": 1.4041744470596313, "learning_rate": 5.773577556274543e-06, "loss": 0.8518, "step": 3564 }, { "epoch": 0.47, "grad_norm": 0.7120608687400818, "learning_rate": 5.771480914834158e-06, "loss": 1.0479, "step": 3565 }, { "epoch": 0.47, "grad_norm": 1.4387189149856567, "learning_rate": 5.769384134421945e-06, "loss": 0.8254, "step": 3566 }, { "epoch": 0.47, "grad_norm": 1.3745111227035522, "learning_rate": 5.767287215415613e-06, "loss": 0.7932, "step": 3567 }, { "epoch": 0.47, "grad_norm": 0.6779829263687134, "learning_rate": 5.765190158192895e-06, "loss": 1.0361, "step": 3568 }, { "epoch": 0.47, "grad_norm": 0.6395149827003479, "learning_rate": 5.7630929631315424e-06, "loss": 1.0175, "step": 3569 }, { "epoch": 0.47, "grad_norm": 0.6363388895988464, "learning_rate": 5.76099563060934e-06, "loss": 1.0137, "step": 3570 }, { "epoch": 0.47, "grad_norm": 1.3308758735656738, "learning_rate": 5.758898161004095e-06, "loss": 0.7859, "step": 3571 }, { "epoch": 0.47, "grad_norm": 1.5323657989501953, "learning_rate": 5.756800554693636e-06, "loss": 0.8193, "step": 3572 }, { "epoch": 0.47, "grad_norm": 0.6871373653411865, "learning_rate": 5.754702812055819e-06, "loss": 1.0186, "step": 3573 }, { "epoch": 0.47, "grad_norm": 1.3896219730377197, "learning_rate": 5.752604933468523e-06, "loss": 0.7905, "step": 3574 }, { "epoch": 0.47, "grad_norm": 0.7092670202255249, "learning_rate": 5.750506919309652e-06, "loss": 1.0229, "step": 3575 }, { "epoch": 0.47, "grad_norm": 1.5270966291427612, "learning_rate": 5.748408769957137e-06, "loss": 0.7914, "step": 3576 }, { "epoch": 0.47, "grad_norm": 1.356990933418274, "learning_rate": 5.746310485788931e-06, "loss": 0.8252, "step": 3577 }, { "epoch": 0.47, "grad_norm": 1.2914187908172607, "learning_rate": 5.744212067183008e-06, "loss": 0.8553, "step": 3578 }, { "epoch": 0.47, "grad_norm": 1.3070354461669922, "learning_rate": 5.742113514517375e-06, "loss": 0.8408, "step": 3579 }, { "epoch": 0.47, "grad_norm": 0.6662927865982056, "learning_rate": 5.740014828170053e-06, "loss": 1.0413, "step": 3580 }, { "epoch": 0.47, "grad_norm": 1.272708535194397, "learning_rate": 5.737916008519093e-06, "loss": 0.7374, "step": 3581 }, { "epoch": 0.47, "grad_norm": 1.286575198173523, "learning_rate": 5.735817055942572e-06, "loss": 0.8318, "step": 3582 }, { "epoch": 0.47, "grad_norm": 1.5324275493621826, "learning_rate": 5.733717970818584e-06, "loss": 0.888, "step": 3583 }, { "epoch": 0.47, "grad_norm": 1.2797898054122925, "learning_rate": 5.731618753525252e-06, "loss": 0.8227, "step": 3584 }, { "epoch": 0.47, "grad_norm": 1.3481653928756714, "learning_rate": 5.729519404440722e-06, "loss": 0.8108, "step": 3585 }, { "epoch": 0.47, "grad_norm": 0.6181612610816956, "learning_rate": 5.727419923943163e-06, "loss": 1.0394, "step": 3586 }, { "epoch": 0.47, "grad_norm": 1.290080189704895, "learning_rate": 5.7253203124107695e-06, "loss": 0.8566, "step": 3587 }, { "epoch": 0.47, "grad_norm": 0.610013484954834, "learning_rate": 5.7232205702217545e-06, "loss": 1.0244, "step": 3588 }, { "epoch": 0.47, "grad_norm": 0.6220468878746033, "learning_rate": 5.721120697754361e-06, "loss": 1.038, "step": 3589 }, { "epoch": 0.47, "grad_norm": 1.3096345663070679, "learning_rate": 5.719020695386851e-06, "loss": 0.7789, "step": 3590 }, { "epoch": 0.47, "grad_norm": 1.336406946182251, "learning_rate": 5.716920563497514e-06, "loss": 0.8244, "step": 3591 }, { "epoch": 0.47, "grad_norm": 1.3485491275787354, "learning_rate": 5.714820302464656e-06, "loss": 0.8244, "step": 3592 }, { "epoch": 0.47, "grad_norm": 1.4130549430847168, "learning_rate": 5.712719912666615e-06, "loss": 0.9009, "step": 3593 }, { "epoch": 0.47, "grad_norm": 1.1555261611938477, "learning_rate": 5.710619394481745e-06, "loss": 0.7794, "step": 3594 }, { "epoch": 0.47, "grad_norm": 1.3198789358139038, "learning_rate": 5.708518748288425e-06, "loss": 0.8119, "step": 3595 }, { "epoch": 0.47, "grad_norm": 1.346423625946045, "learning_rate": 5.706417974465061e-06, "loss": 0.8013, "step": 3596 }, { "epoch": 0.47, "grad_norm": 0.7999628186225891, "learning_rate": 5.7043170733900775e-06, "loss": 1.0413, "step": 3597 }, { "epoch": 0.47, "grad_norm": 0.7058148384094238, "learning_rate": 5.702216045441924e-06, "loss": 1.0537, "step": 3598 }, { "epoch": 0.47, "grad_norm": 1.3323801755905151, "learning_rate": 5.700114890999071e-06, "loss": 0.8038, "step": 3599 }, { "epoch": 0.47, "grad_norm": 0.6278709173202515, "learning_rate": 5.698013610440012e-06, "loss": 1.032, "step": 3600 }, { "epoch": 0.47, "grad_norm": 1.3129332065582275, "learning_rate": 5.695912204143268e-06, "loss": 0.8368, "step": 3601 }, { "epoch": 0.47, "grad_norm": 1.422566533088684, "learning_rate": 5.6938106724873745e-06, "loss": 0.7765, "step": 3602 }, { "epoch": 0.47, "grad_norm": 1.4299649000167847, "learning_rate": 5.691709015850897e-06, "loss": 0.7781, "step": 3603 }, { "epoch": 0.47, "grad_norm": 1.2825186252593994, "learning_rate": 5.689607234612418e-06, "loss": 0.8268, "step": 3604 }, { "epoch": 0.47, "grad_norm": 0.9749084711074829, "learning_rate": 5.687505329150548e-06, "loss": 1.0208, "step": 3605 }, { "epoch": 0.47, "grad_norm": 1.4346954822540283, "learning_rate": 5.685403299843914e-06, "loss": 0.8296, "step": 3606 }, { "epoch": 0.47, "grad_norm": 1.338948130607605, "learning_rate": 5.68330114707117e-06, "loss": 0.8404, "step": 3607 }, { "epoch": 0.47, "grad_norm": 1.4959017038345337, "learning_rate": 5.681198871210988e-06, "loss": 0.8497, "step": 3608 }, { "epoch": 0.47, "grad_norm": 0.6707041263580322, "learning_rate": 5.679096472642066e-06, "loss": 1.0425, "step": 3609 }, { "epoch": 0.47, "grad_norm": 1.3372937440872192, "learning_rate": 5.676993951743124e-06, "loss": 0.7767, "step": 3610 }, { "epoch": 0.47, "grad_norm": 1.29006826877594, "learning_rate": 5.6748913088929005e-06, "loss": 0.7808, "step": 3611 }, { "epoch": 0.47, "grad_norm": 1.2429653406143188, "learning_rate": 5.6727885444701584e-06, "loss": 0.7538, "step": 3612 }, { "epoch": 0.47, "grad_norm": 1.3649749755859375, "learning_rate": 5.670685658853682e-06, "loss": 0.8521, "step": 3613 }, { "epoch": 0.47, "grad_norm": 1.277970314025879, "learning_rate": 5.6685826524222785e-06, "loss": 0.7936, "step": 3614 }, { "epoch": 0.47, "grad_norm": 1.7802608013153076, "learning_rate": 5.666479525554777e-06, "loss": 0.8268, "step": 3615 }, { "epoch": 0.47, "grad_norm": 1.3418983221054077, "learning_rate": 5.664376278630025e-06, "loss": 0.8136, "step": 3616 }, { "epoch": 0.47, "grad_norm": 1.3078199625015259, "learning_rate": 5.662272912026894e-06, "loss": 0.8102, "step": 3617 }, { "epoch": 0.47, "grad_norm": 1.365081548690796, "learning_rate": 5.66016942612428e-06, "loss": 0.7991, "step": 3618 }, { "epoch": 0.47, "grad_norm": 0.7964968681335449, "learning_rate": 5.658065821301093e-06, "loss": 1.042, "step": 3619 }, { "epoch": 0.47, "grad_norm": 1.2892001867294312, "learning_rate": 5.6559620979362736e-06, "loss": 0.7807, "step": 3620 }, { "epoch": 0.47, "grad_norm": 1.2798486948013306, "learning_rate": 5.653858256408775e-06, "loss": 0.7855, "step": 3621 }, { "epoch": 0.47, "grad_norm": 1.4527311325073242, "learning_rate": 5.651754297097577e-06, "loss": 0.8885, "step": 3622 }, { "epoch": 0.47, "grad_norm": 1.3718225955963135, "learning_rate": 5.64965022038168e-06, "loss": 0.777, "step": 3623 }, { "epoch": 0.47, "grad_norm": 1.8455184698104858, "learning_rate": 5.647546026640106e-06, "loss": 0.7552, "step": 3624 }, { "epoch": 0.47, "grad_norm": 0.6334443092346191, "learning_rate": 5.6454417162518935e-06, "loss": 1.0092, "step": 3625 }, { "epoch": 0.48, "grad_norm": 0.6134970784187317, "learning_rate": 5.6433372895961095e-06, "loss": 1.0489, "step": 3626 }, { "epoch": 0.48, "grad_norm": 0.6007273197174072, "learning_rate": 5.641232747051834e-06, "loss": 1.0179, "step": 3627 }, { "epoch": 0.48, "grad_norm": 1.3300715684890747, "learning_rate": 5.639128088998175e-06, "loss": 0.7858, "step": 3628 }, { "epoch": 0.48, "grad_norm": 0.6061898469924927, "learning_rate": 5.637023315814257e-06, "loss": 0.9975, "step": 3629 }, { "epoch": 0.48, "grad_norm": 1.3190665245056152, "learning_rate": 5.634918427879225e-06, "loss": 0.8671, "step": 3630 }, { "epoch": 0.48, "grad_norm": 0.6535469889640808, "learning_rate": 5.632813425572247e-06, "loss": 1.0228, "step": 3631 }, { "epoch": 0.48, "grad_norm": 1.425696611404419, "learning_rate": 5.630708309272513e-06, "loss": 0.8641, "step": 3632 }, { "epoch": 0.48, "grad_norm": 1.9688973426818848, "learning_rate": 5.62860307935923e-06, "loss": 0.7907, "step": 3633 }, { "epoch": 0.48, "grad_norm": 1.4289450645446777, "learning_rate": 5.626497736211622e-06, "loss": 0.8293, "step": 3634 }, { "epoch": 0.48, "grad_norm": 1.3546644449234009, "learning_rate": 5.624392280208943e-06, "loss": 0.7966, "step": 3635 }, { "epoch": 0.48, "grad_norm": 1.3653162717819214, "learning_rate": 5.622286711730461e-06, "loss": 0.8298, "step": 3636 }, { "epoch": 0.48, "grad_norm": 1.2690832614898682, "learning_rate": 5.620181031155467e-06, "loss": 0.8131, "step": 3637 }, { "epoch": 0.48, "grad_norm": 0.6617457270622253, "learning_rate": 5.618075238863269e-06, "loss": 1.0414, "step": 3638 }, { "epoch": 0.48, "grad_norm": 1.363630771636963, "learning_rate": 5.615969335233197e-06, "loss": 0.8297, "step": 3639 }, { "epoch": 0.48, "grad_norm": 1.336689829826355, "learning_rate": 5.613863320644601e-06, "loss": 0.8298, "step": 3640 }, { "epoch": 0.48, "grad_norm": 1.5576332807540894, "learning_rate": 5.611757195476852e-06, "loss": 0.7918, "step": 3641 }, { "epoch": 0.48, "grad_norm": 1.3682236671447754, "learning_rate": 5.609650960109338e-06, "loss": 0.8726, "step": 3642 }, { "epoch": 0.48, "grad_norm": 1.2861316204071045, "learning_rate": 5.607544614921472e-06, "loss": 0.8661, "step": 3643 }, { "epoch": 0.48, "grad_norm": 1.3983091115951538, "learning_rate": 5.60543816029268e-06, "loss": 0.8285, "step": 3644 }, { "epoch": 0.48, "grad_norm": 0.6541066765785217, "learning_rate": 5.603331596602412e-06, "loss": 1.0249, "step": 3645 }, { "epoch": 0.48, "grad_norm": 1.3123716115951538, "learning_rate": 5.601224924230139e-06, "loss": 0.82, "step": 3646 }, { "epoch": 0.48, "grad_norm": 1.2442485094070435, "learning_rate": 5.5991181435553476e-06, "loss": 0.8387, "step": 3647 }, { "epoch": 0.48, "grad_norm": 1.2651419639587402, "learning_rate": 5.5970112549575465e-06, "loss": 0.8696, "step": 3648 }, { "epoch": 0.48, "grad_norm": 1.20163893699646, "learning_rate": 5.594904258816262e-06, "loss": 0.7947, "step": 3649 }, { "epoch": 0.48, "grad_norm": 1.370469570159912, "learning_rate": 5.592797155511043e-06, "loss": 0.8207, "step": 3650 }, { "epoch": 0.48, "grad_norm": 1.3484078645706177, "learning_rate": 5.590689945421452e-06, "loss": 0.8305, "step": 3651 }, { "epoch": 0.48, "grad_norm": 1.2805687189102173, "learning_rate": 5.58858262892708e-06, "loss": 0.8159, "step": 3652 }, { "epoch": 0.48, "grad_norm": 1.255721092224121, "learning_rate": 5.586475206407527e-06, "loss": 0.8267, "step": 3653 }, { "epoch": 0.48, "grad_norm": 0.7003995776176453, "learning_rate": 5.5843676782424176e-06, "loss": 1.0519, "step": 3654 }, { "epoch": 0.48, "grad_norm": 1.329129695892334, "learning_rate": 5.5822600448113926e-06, "loss": 0.8549, "step": 3655 }, { "epoch": 0.48, "grad_norm": 1.3641172647476196, "learning_rate": 5.5801523064941175e-06, "loss": 0.7964, "step": 3656 }, { "epoch": 0.48, "grad_norm": 0.692423403263092, "learning_rate": 5.578044463670271e-06, "loss": 1.0298, "step": 3657 }, { "epoch": 0.48, "grad_norm": 0.6522066593170166, "learning_rate": 5.575936516719552e-06, "loss": 1.0494, "step": 3658 }, { "epoch": 0.48, "grad_norm": 1.3027796745300293, "learning_rate": 5.573828466021679e-06, "loss": 0.8788, "step": 3659 }, { "epoch": 0.48, "grad_norm": 1.4300432205200195, "learning_rate": 5.5717203119563875e-06, "loss": 0.8235, "step": 3660 }, { "epoch": 0.48, "grad_norm": 0.6690620183944702, "learning_rate": 5.569612054903436e-06, "loss": 1.0145, "step": 3661 }, { "epoch": 0.48, "grad_norm": 1.9080073833465576, "learning_rate": 5.567503695242595e-06, "loss": 0.8132, "step": 3662 }, { "epoch": 0.48, "grad_norm": 0.7089280486106873, "learning_rate": 5.5653952333536575e-06, "loss": 1.0445, "step": 3663 }, { "epoch": 0.48, "grad_norm": 0.6400163173675537, "learning_rate": 5.563286669616435e-06, "loss": 1.0132, "step": 3664 }, { "epoch": 0.48, "grad_norm": 0.6006166934967041, "learning_rate": 5.561178004410758e-06, "loss": 1.0247, "step": 3665 }, { "epoch": 0.48, "grad_norm": 0.6297223567962646, "learning_rate": 5.559069238116472e-06, "loss": 1.0352, "step": 3666 }, { "epoch": 0.48, "grad_norm": 1.4678150415420532, "learning_rate": 5.556960371113445e-06, "loss": 0.7141, "step": 3667 }, { "epoch": 0.48, "grad_norm": 1.3954142332077026, "learning_rate": 5.5548514037815556e-06, "loss": 0.8271, "step": 3668 }, { "epoch": 0.48, "grad_norm": 1.2824068069458008, "learning_rate": 5.552742336500711e-06, "loss": 0.8469, "step": 3669 }, { "epoch": 0.48, "grad_norm": 1.4069658517837524, "learning_rate": 5.55063316965083e-06, "loss": 0.8022, "step": 3670 }, { "epoch": 0.48, "grad_norm": 1.3895798921585083, "learning_rate": 5.5485239036118495e-06, "loss": 0.7744, "step": 3671 }, { "epoch": 0.48, "grad_norm": 1.3696274757385254, "learning_rate": 5.546414538763725e-06, "loss": 0.8374, "step": 3672 }, { "epoch": 0.48, "grad_norm": 1.3042855262756348, "learning_rate": 5.54430507548643e-06, "loss": 0.8084, "step": 3673 }, { "epoch": 0.48, "grad_norm": 1.2608016729354858, "learning_rate": 5.542195514159956e-06, "loss": 0.7913, "step": 3674 }, { "epoch": 0.48, "grad_norm": 1.2304441928863525, "learning_rate": 5.540085855164312e-06, "loss": 0.8063, "step": 3675 }, { "epoch": 0.48, "grad_norm": 0.949249267578125, "learning_rate": 5.5379760988795265e-06, "loss": 1.0081, "step": 3676 }, { "epoch": 0.48, "grad_norm": 2.318824052810669, "learning_rate": 5.5358662456856385e-06, "loss": 0.8171, "step": 3677 }, { "epoch": 0.48, "grad_norm": 1.414650797843933, "learning_rate": 5.533756295962715e-06, "loss": 0.7984, "step": 3678 }, { "epoch": 0.48, "grad_norm": 1.4001164436340332, "learning_rate": 5.531646250090829e-06, "loss": 0.8351, "step": 3679 }, { "epoch": 0.48, "grad_norm": 1.5238442420959473, "learning_rate": 5.5295361084500806e-06, "loss": 0.7971, "step": 3680 }, { "epoch": 0.48, "grad_norm": 2.1327297687530518, "learning_rate": 5.527425871420584e-06, "loss": 0.8087, "step": 3681 }, { "epoch": 0.48, "grad_norm": 0.7743279933929443, "learning_rate": 5.5253155393824676e-06, "loss": 1.034, "step": 3682 }, { "epoch": 0.48, "grad_norm": 1.7819241285324097, "learning_rate": 5.523205112715878e-06, "loss": 0.8878, "step": 3683 }, { "epoch": 0.48, "grad_norm": 1.4481420516967773, "learning_rate": 5.521094591800983e-06, "loss": 0.7844, "step": 3684 }, { "epoch": 0.48, "grad_norm": 1.2390854358673096, "learning_rate": 5.5189839770179645e-06, "loss": 0.7978, "step": 3685 }, { "epoch": 0.48, "grad_norm": 1.3516643047332764, "learning_rate": 5.5168732687470175e-06, "loss": 0.7921, "step": 3686 }, { "epoch": 0.48, "grad_norm": 1.3149749040603638, "learning_rate": 5.51476246736836e-06, "loss": 0.819, "step": 3687 }, { "epoch": 0.48, "grad_norm": 1.3355677127838135, "learning_rate": 5.512651573262223e-06, "loss": 0.8327, "step": 3688 }, { "epoch": 0.48, "grad_norm": 1.7913932800292969, "learning_rate": 5.510540586808857e-06, "loss": 0.7929, "step": 3689 }, { "epoch": 0.48, "grad_norm": 1.3763339519500732, "learning_rate": 5.508429508388528e-06, "loss": 0.8459, "step": 3690 }, { "epoch": 0.48, "grad_norm": 1.2982076406478882, "learning_rate": 5.506318338381516e-06, "loss": 0.7935, "step": 3691 }, { "epoch": 0.48, "grad_norm": 1.3070337772369385, "learning_rate": 5.50420707716812e-06, "loss": 0.858, "step": 3692 }, { "epoch": 0.48, "grad_norm": 0.7103028297424316, "learning_rate": 5.502095725128656e-06, "loss": 1.0319, "step": 3693 }, { "epoch": 0.48, "grad_norm": 1.3374412059783936, "learning_rate": 5.499984282643456e-06, "loss": 0.7801, "step": 3694 }, { "epoch": 0.48, "grad_norm": 1.3099545240402222, "learning_rate": 5.497872750092866e-06, "loss": 0.8087, "step": 3695 }, { "epoch": 0.48, "grad_norm": 1.3419878482818604, "learning_rate": 5.495761127857251e-06, "loss": 0.7998, "step": 3696 }, { "epoch": 0.48, "grad_norm": 1.3948562145233154, "learning_rate": 5.493649416316991e-06, "loss": 0.8315, "step": 3697 }, { "epoch": 0.48, "grad_norm": 0.6876908540725708, "learning_rate": 5.491537615852483e-06, "loss": 1.0192, "step": 3698 }, { "epoch": 0.48, "grad_norm": 1.4748759269714355, "learning_rate": 5.489425726844138e-06, "loss": 0.8251, "step": 3699 }, { "epoch": 0.48, "grad_norm": 1.3514493703842163, "learning_rate": 5.487313749672385e-06, "loss": 0.7975, "step": 3700 }, { "epoch": 0.48, "grad_norm": 0.6227331757545471, "learning_rate": 5.485201684717667e-06, "loss": 1.0319, "step": 3701 }, { "epoch": 0.49, "grad_norm": 2.1879711151123047, "learning_rate": 5.483089532360444e-06, "loss": 0.8744, "step": 3702 }, { "epoch": 0.49, "grad_norm": 1.3860113620758057, "learning_rate": 5.480977292981192e-06, "loss": 0.7749, "step": 3703 }, { "epoch": 0.49, "grad_norm": 1.3635897636413574, "learning_rate": 5.4788649669604044e-06, "loss": 0.7838, "step": 3704 }, { "epoch": 0.49, "grad_norm": 1.3794395923614502, "learning_rate": 5.476752554678584e-06, "loss": 0.8395, "step": 3705 }, { "epoch": 0.49, "grad_norm": 1.3911168575286865, "learning_rate": 5.474640056516256e-06, "loss": 0.7791, "step": 3706 }, { "epoch": 0.49, "grad_norm": 1.326120138168335, "learning_rate": 5.472527472853957e-06, "loss": 0.8594, "step": 3707 }, { "epoch": 0.49, "grad_norm": 1.2985206842422485, "learning_rate": 5.470414804072243e-06, "loss": 0.7951, "step": 3708 }, { "epoch": 0.49, "grad_norm": 1.6564805507659912, "learning_rate": 5.468302050551678e-06, "loss": 0.819, "step": 3709 }, { "epoch": 0.49, "grad_norm": 1.309251308441162, "learning_rate": 5.466189212672849e-06, "loss": 0.8236, "step": 3710 }, { "epoch": 0.49, "grad_norm": 0.7483739256858826, "learning_rate": 5.464076290816354e-06, "loss": 1.039, "step": 3711 }, { "epoch": 0.49, "grad_norm": 1.3930197954177856, "learning_rate": 5.461963285362809e-06, "loss": 0.8306, "step": 3712 }, { "epoch": 0.49, "grad_norm": 0.6327620148658752, "learning_rate": 5.459850196692843e-06, "loss": 1.0297, "step": 3713 }, { "epoch": 0.49, "grad_norm": 0.6209039688110352, "learning_rate": 5.457737025187098e-06, "loss": 1.0285, "step": 3714 }, { "epoch": 0.49, "grad_norm": 0.6391229629516602, "learning_rate": 5.455623771226234e-06, "loss": 1.0091, "step": 3715 }, { "epoch": 0.49, "grad_norm": 1.3672538995742798, "learning_rate": 5.453510435190925e-06, "loss": 0.8016, "step": 3716 }, { "epoch": 0.49, "grad_norm": 1.539162516593933, "learning_rate": 5.451397017461862e-06, "loss": 0.8077, "step": 3717 }, { "epoch": 0.49, "grad_norm": 1.3150593042373657, "learning_rate": 5.4492835184197465e-06, "loss": 0.8596, "step": 3718 }, { "epoch": 0.49, "grad_norm": 1.376542568206787, "learning_rate": 5.447169938445297e-06, "loss": 0.7731, "step": 3719 }, { "epoch": 0.49, "grad_norm": 1.3063125610351562, "learning_rate": 5.445056277919247e-06, "loss": 0.84, "step": 3720 }, { "epoch": 0.49, "grad_norm": 1.353556752204895, "learning_rate": 5.442942537222342e-06, "loss": 0.8546, "step": 3721 }, { "epoch": 0.49, "grad_norm": 0.8055543303489685, "learning_rate": 5.440828716735347e-06, "loss": 1.0446, "step": 3722 }, { "epoch": 0.49, "grad_norm": 0.8102369904518127, "learning_rate": 5.438714816839035e-06, "loss": 0.9818, "step": 3723 }, { "epoch": 0.49, "grad_norm": 1.2907265424728394, "learning_rate": 5.436600837914195e-06, "loss": 0.8726, "step": 3724 }, { "epoch": 0.49, "grad_norm": 1.2958788871765137, "learning_rate": 5.4344867803416365e-06, "loss": 0.7871, "step": 3725 }, { "epoch": 0.49, "grad_norm": 0.6688776016235352, "learning_rate": 5.4323726445021755e-06, "loss": 1.0265, "step": 3726 }, { "epoch": 0.49, "grad_norm": 1.3665721416473389, "learning_rate": 5.430258430776645e-06, "loss": 0.7692, "step": 3727 }, { "epoch": 0.49, "grad_norm": 1.3658479452133179, "learning_rate": 5.428144139545892e-06, "loss": 0.8294, "step": 3728 }, { "epoch": 0.49, "grad_norm": 1.4380038976669312, "learning_rate": 5.426029771190777e-06, "loss": 0.8284, "step": 3729 }, { "epoch": 0.49, "grad_norm": 1.4918469190597534, "learning_rate": 5.4239153260921764e-06, "loss": 0.8602, "step": 3730 }, { "epoch": 0.49, "grad_norm": 1.337720513343811, "learning_rate": 5.421800804630978e-06, "loss": 0.8547, "step": 3731 }, { "epoch": 0.49, "grad_norm": 1.5596288442611694, "learning_rate": 5.419686207188084e-06, "loss": 0.8155, "step": 3732 }, { "epoch": 0.49, "grad_norm": 1.3019658327102661, "learning_rate": 5.417571534144409e-06, "loss": 0.8062, "step": 3733 }, { "epoch": 0.49, "grad_norm": 1.4571846723556519, "learning_rate": 5.415456785880885e-06, "loss": 0.7732, "step": 3734 }, { "epoch": 0.49, "grad_norm": 1.338686227798462, "learning_rate": 5.4133419627784545e-06, "loss": 0.8263, "step": 3735 }, { "epoch": 0.49, "grad_norm": 0.8230371475219727, "learning_rate": 5.411227065218074e-06, "loss": 1.0056, "step": 3736 }, { "epoch": 0.49, "grad_norm": 1.4119374752044678, "learning_rate": 5.409112093580714e-06, "loss": 0.8024, "step": 3737 }, { "epoch": 0.49, "grad_norm": 1.2993584871292114, "learning_rate": 5.406997048247357e-06, "loss": 0.7414, "step": 3738 }, { "epoch": 0.49, "grad_norm": 1.3022115230560303, "learning_rate": 5.404881929599001e-06, "loss": 0.7852, "step": 3739 }, { "epoch": 0.49, "grad_norm": 0.6760143637657166, "learning_rate": 5.402766738016654e-06, "loss": 1.0379, "step": 3740 }, { "epoch": 0.49, "grad_norm": 1.3409786224365234, "learning_rate": 5.400651473881342e-06, "loss": 0.7933, "step": 3741 }, { "epoch": 0.49, "grad_norm": 1.324107050895691, "learning_rate": 5.398536137574099e-06, "loss": 0.7968, "step": 3742 }, { "epoch": 0.49, "grad_norm": 1.7264823913574219, "learning_rate": 5.396420729475974e-06, "loss": 0.8164, "step": 3743 }, { "epoch": 0.49, "grad_norm": 0.6479153633117676, "learning_rate": 5.394305249968029e-06, "loss": 1.0193, "step": 3744 }, { "epoch": 0.49, "grad_norm": 0.6570194959640503, "learning_rate": 5.39218969943134e-06, "loss": 1.0306, "step": 3745 }, { "epoch": 0.49, "grad_norm": 1.491721749305725, "learning_rate": 5.390074078246997e-06, "loss": 0.8034, "step": 3746 }, { "epoch": 0.49, "grad_norm": 1.452089548110962, "learning_rate": 5.387958386796096e-06, "loss": 0.8154, "step": 3747 }, { "epoch": 0.49, "grad_norm": 1.3155728578567505, "learning_rate": 5.385842625459752e-06, "loss": 0.794, "step": 3748 }, { "epoch": 0.49, "grad_norm": 1.3248852491378784, "learning_rate": 5.383726794619091e-06, "loss": 0.8565, "step": 3749 }, { "epoch": 0.49, "grad_norm": 1.330849528312683, "learning_rate": 5.381610894655251e-06, "loss": 0.8107, "step": 3750 }, { "epoch": 0.49, "grad_norm": 1.6562135219573975, "learning_rate": 5.379494925949383e-06, "loss": 0.8526, "step": 3751 }, { "epoch": 0.49, "grad_norm": 1.385236382484436, "learning_rate": 5.37737888888265e-06, "loss": 0.8132, "step": 3752 }, { "epoch": 0.49, "grad_norm": 0.8663389086723328, "learning_rate": 5.375262783836226e-06, "loss": 1.0385, "step": 3753 }, { "epoch": 0.49, "grad_norm": 1.3394358158111572, "learning_rate": 5.3731466111913e-06, "loss": 0.8364, "step": 3754 }, { "epoch": 0.49, "grad_norm": 1.5712083578109741, "learning_rate": 5.371030371329073e-06, "loss": 0.8204, "step": 3755 }, { "epoch": 0.49, "grad_norm": 5.218817710876465, "learning_rate": 5.368914064630755e-06, "loss": 0.8458, "step": 3756 }, { "epoch": 0.49, "grad_norm": 1.411035418510437, "learning_rate": 5.366797691477568e-06, "loss": 0.8209, "step": 3757 }, { "epoch": 0.49, "grad_norm": 1.40706467628479, "learning_rate": 5.364681252250754e-06, "loss": 0.7989, "step": 3758 }, { "epoch": 0.49, "grad_norm": 1.4340620040893555, "learning_rate": 5.362564747331556e-06, "loss": 0.8214, "step": 3759 }, { "epoch": 0.49, "grad_norm": 1.414812684059143, "learning_rate": 5.360448177101234e-06, "loss": 0.856, "step": 3760 }, { "epoch": 0.49, "grad_norm": 0.7438288331031799, "learning_rate": 5.358331541941062e-06, "loss": 1.0256, "step": 3761 }, { "epoch": 0.49, "grad_norm": 1.2729005813598633, "learning_rate": 5.356214842232319e-06, "loss": 0.7979, "step": 3762 }, { "epoch": 0.49, "grad_norm": 1.3868350982666016, "learning_rate": 5.354098078356304e-06, "loss": 0.7949, "step": 3763 }, { "epoch": 0.49, "grad_norm": 1.3303277492523193, "learning_rate": 5.3519812506943215e-06, "loss": 0.8828, "step": 3764 }, { "epoch": 0.49, "grad_norm": 0.629899799823761, "learning_rate": 5.349864359627689e-06, "loss": 1.0558, "step": 3765 }, { "epoch": 0.49, "grad_norm": 1.432153344154358, "learning_rate": 5.347747405537735e-06, "loss": 0.7834, "step": 3766 }, { "epoch": 0.49, "grad_norm": 0.617239773273468, "learning_rate": 5.345630388805802e-06, "loss": 1.0239, "step": 3767 }, { "epoch": 0.49, "grad_norm": 1.4641958475112915, "learning_rate": 5.343513309813239e-06, "loss": 0.7871, "step": 3768 }, { "epoch": 0.49, "grad_norm": 1.2467657327651978, "learning_rate": 5.341396168941415e-06, "loss": 0.7849, "step": 3769 }, { "epoch": 0.49, "grad_norm": 1.3174598217010498, "learning_rate": 5.339278966571697e-06, "loss": 0.8448, "step": 3770 }, { "epoch": 0.49, "grad_norm": 1.395891785621643, "learning_rate": 5.337161703085475e-06, "loss": 0.8162, "step": 3771 }, { "epoch": 0.49, "grad_norm": 0.6543887853622437, "learning_rate": 5.3350443788641434e-06, "loss": 0.9928, "step": 3772 }, { "epoch": 0.49, "grad_norm": 1.6372538805007935, "learning_rate": 5.332926994289109e-06, "loss": 0.8443, "step": 3773 }, { "epoch": 0.49, "grad_norm": 0.6849886178970337, "learning_rate": 5.330809549741793e-06, "loss": 1.0219, "step": 3774 }, { "epoch": 0.49, "grad_norm": 1.8390272855758667, "learning_rate": 5.32869204560362e-06, "loss": 0.8424, "step": 3775 }, { "epoch": 0.49, "grad_norm": 0.5964027643203735, "learning_rate": 5.326574482256034e-06, "loss": 1.0253, "step": 3776 }, { "epoch": 0.49, "grad_norm": 1.5048627853393555, "learning_rate": 5.324456860080481e-06, "loss": 0.8806, "step": 3777 }, { "epoch": 0.5, "grad_norm": 0.6968114972114563, "learning_rate": 5.322339179458427e-06, "loss": 1.0089, "step": 3778 }, { "epoch": 0.5, "grad_norm": 0.6927106976509094, "learning_rate": 5.320221440771337e-06, "loss": 1.0306, "step": 3779 }, { "epoch": 0.5, "grad_norm": 1.380940318107605, "learning_rate": 5.318103644400699e-06, "loss": 0.7791, "step": 3780 }, { "epoch": 0.5, "grad_norm": 1.351540446281433, "learning_rate": 5.315985790728002e-06, "loss": 0.8065, "step": 3781 }, { "epoch": 0.5, "grad_norm": 0.626133918762207, "learning_rate": 5.313867880134749e-06, "loss": 1.0131, "step": 3782 }, { "epoch": 0.5, "grad_norm": 1.485500693321228, "learning_rate": 5.3117499130024534e-06, "loss": 0.8003, "step": 3783 }, { "epoch": 0.5, "grad_norm": 0.6417425274848938, "learning_rate": 5.3096318897126375e-06, "loss": 1.0256, "step": 3784 }, { "epoch": 0.5, "grad_norm": 1.312713623046875, "learning_rate": 5.307513810646836e-06, "loss": 0.7834, "step": 3785 }, { "epoch": 0.5, "grad_norm": 1.4661059379577637, "learning_rate": 5.3053956761865895e-06, "loss": 0.8432, "step": 3786 }, { "epoch": 0.5, "grad_norm": 1.3086326122283936, "learning_rate": 5.303277486713454e-06, "loss": 0.8175, "step": 3787 }, { "epoch": 0.5, "grad_norm": 1.2735064029693604, "learning_rate": 5.301159242608992e-06, "loss": 0.8186, "step": 3788 }, { "epoch": 0.5, "grad_norm": 1.3450218439102173, "learning_rate": 5.2990409442547755e-06, "loss": 0.8294, "step": 3789 }, { "epoch": 0.5, "grad_norm": 0.7078295946121216, "learning_rate": 5.296922592032388e-06, "loss": 1.0334, "step": 3790 }, { "epoch": 0.5, "grad_norm": 0.7010477781295776, "learning_rate": 5.29480418632342e-06, "loss": 1.0176, "step": 3791 }, { "epoch": 0.5, "grad_norm": 1.3176854848861694, "learning_rate": 5.292685727509478e-06, "loss": 0.7934, "step": 3792 }, { "epoch": 0.5, "grad_norm": 0.6009455919265747, "learning_rate": 5.2905672159721685e-06, "loss": 1.0495, "step": 3793 }, { "epoch": 0.5, "grad_norm": 1.4652202129364014, "learning_rate": 5.288448652093115e-06, "loss": 0.8034, "step": 3794 }, { "epoch": 0.5, "grad_norm": 1.4406064748764038, "learning_rate": 5.28633003625395e-06, "loss": 0.7474, "step": 3795 }, { "epoch": 0.5, "grad_norm": 1.3274712562561035, "learning_rate": 5.2842113688363086e-06, "loss": 0.7878, "step": 3796 }, { "epoch": 0.5, "grad_norm": 1.7104809284210205, "learning_rate": 5.282092650221845e-06, "loss": 0.7645, "step": 3797 }, { "epoch": 0.5, "grad_norm": 1.4253349304199219, "learning_rate": 5.279973880792214e-06, "loss": 0.8465, "step": 3798 }, { "epoch": 0.5, "grad_norm": 1.5690999031066895, "learning_rate": 5.277855060929086e-06, "loss": 0.7887, "step": 3799 }, { "epoch": 0.5, "grad_norm": 1.3876367807388306, "learning_rate": 5.275736191014136e-06, "loss": 0.8436, "step": 3800 }, { "epoch": 0.5, "grad_norm": 1.4117573499679565, "learning_rate": 5.273617271429049e-06, "loss": 0.7598, "step": 3801 }, { "epoch": 0.5, "grad_norm": 1.6289429664611816, "learning_rate": 5.2714983025555226e-06, "loss": 0.8231, "step": 3802 }, { "epoch": 0.5, "grad_norm": 0.8010244369506836, "learning_rate": 5.269379284775256e-06, "loss": 1.0366, "step": 3803 }, { "epoch": 0.5, "grad_norm": 1.5395234823226929, "learning_rate": 5.267260218469964e-06, "loss": 0.8444, "step": 3804 }, { "epoch": 0.5, "grad_norm": 0.6775914430618286, "learning_rate": 5.265141104021369e-06, "loss": 1.0346, "step": 3805 }, { "epoch": 0.5, "grad_norm": 0.6697185039520264, "learning_rate": 5.263021941811197e-06, "loss": 0.9995, "step": 3806 }, { "epoch": 0.5, "grad_norm": 1.4855434894561768, "learning_rate": 5.260902732221192e-06, "loss": 0.8437, "step": 3807 }, { "epoch": 0.5, "grad_norm": 1.4130275249481201, "learning_rate": 5.258783475633094e-06, "loss": 0.8359, "step": 3808 }, { "epoch": 0.5, "grad_norm": 1.3499064445495605, "learning_rate": 5.256664172428664e-06, "loss": 0.847, "step": 3809 }, { "epoch": 0.5, "grad_norm": 1.3478420972824097, "learning_rate": 5.254544822989662e-06, "loss": 0.8091, "step": 3810 }, { "epoch": 0.5, "grad_norm": 1.3778825998306274, "learning_rate": 5.252425427697863e-06, "loss": 0.7608, "step": 3811 }, { "epoch": 0.5, "grad_norm": 0.8692454099655151, "learning_rate": 5.250305986935045e-06, "loss": 1.0259, "step": 3812 }, { "epoch": 0.5, "grad_norm": 0.7812676429748535, "learning_rate": 5.2481865010829966e-06, "loss": 1.0167, "step": 3813 }, { "epoch": 0.5, "grad_norm": 0.6948541402816772, "learning_rate": 5.246066970523517e-06, "loss": 1.0265, "step": 3814 }, { "epoch": 0.5, "grad_norm": 1.3995743989944458, "learning_rate": 5.243947395638408e-06, "loss": 0.7943, "step": 3815 }, { "epoch": 0.5, "grad_norm": 1.3683090209960938, "learning_rate": 5.241827776809485e-06, "loss": 0.7959, "step": 3816 }, { "epoch": 0.5, "grad_norm": 0.8398767113685608, "learning_rate": 5.239708114418566e-06, "loss": 1.036, "step": 3817 }, { "epoch": 0.5, "grad_norm": 1.4058207273483276, "learning_rate": 5.237588408847479e-06, "loss": 0.8862, "step": 3818 }, { "epoch": 0.5, "grad_norm": 0.7538043856620789, "learning_rate": 5.235468660478065e-06, "loss": 1.0256, "step": 3819 }, { "epoch": 0.5, "grad_norm": 1.333658218383789, "learning_rate": 5.233348869692162e-06, "loss": 0.7816, "step": 3820 }, { "epoch": 0.5, "grad_norm": 1.312824010848999, "learning_rate": 5.2312290368716256e-06, "loss": 0.7805, "step": 3821 }, { "epoch": 0.5, "grad_norm": 0.7630980610847473, "learning_rate": 5.229109162398313e-06, "loss": 1.0207, "step": 3822 }, { "epoch": 0.5, "grad_norm": 1.3048995733261108, "learning_rate": 5.226989246654091e-06, "loss": 0.835, "step": 3823 }, { "epoch": 0.5, "grad_norm": 1.3708609342575073, "learning_rate": 5.224869290020833e-06, "loss": 0.8247, "step": 3824 }, { "epoch": 0.5, "grad_norm": 3.0430755615234375, "learning_rate": 5.2227492928804235e-06, "loss": 0.7872, "step": 3825 }, { "epoch": 0.5, "grad_norm": 1.423831582069397, "learning_rate": 5.220629255614747e-06, "loss": 0.8011, "step": 3826 }, { "epoch": 0.5, "grad_norm": 1.4111593961715698, "learning_rate": 5.218509178605704e-06, "loss": 0.7736, "step": 3827 }, { "epoch": 0.5, "grad_norm": 1.3835350275039673, "learning_rate": 5.216389062235193e-06, "loss": 0.7795, "step": 3828 }, { "epoch": 0.5, "grad_norm": 1.429492473602295, "learning_rate": 5.214268906885128e-06, "loss": 0.8067, "step": 3829 }, { "epoch": 0.5, "grad_norm": 1.3577390909194946, "learning_rate": 5.212148712937425e-06, "loss": 0.8176, "step": 3830 }, { "epoch": 0.5, "grad_norm": 0.8659578561782837, "learning_rate": 5.210028480774008e-06, "loss": 1.0313, "step": 3831 }, { "epoch": 0.5, "grad_norm": 1.3941007852554321, "learning_rate": 5.207908210776807e-06, "loss": 0.8247, "step": 3832 }, { "epoch": 0.5, "grad_norm": 1.2579097747802734, "learning_rate": 5.205787903327761e-06, "loss": 0.794, "step": 3833 }, { "epoch": 0.5, "grad_norm": 1.3234256505966187, "learning_rate": 5.203667558808815e-06, "loss": 0.7794, "step": 3834 }, { "epoch": 0.5, "grad_norm": 1.4277094602584839, "learning_rate": 5.20154717760192e-06, "loss": 0.8359, "step": 3835 }, { "epoch": 0.5, "grad_norm": 0.6450001001358032, "learning_rate": 5.1994267600890336e-06, "loss": 1.0223, "step": 3836 }, { "epoch": 0.5, "grad_norm": 1.3776756525039673, "learning_rate": 5.1973063066521205e-06, "loss": 0.8139, "step": 3837 }, { "epoch": 0.5, "grad_norm": 1.3054097890853882, "learning_rate": 5.195185817673151e-06, "loss": 0.8205, "step": 3838 }, { "epoch": 0.5, "grad_norm": 1.4993736743927002, "learning_rate": 5.193065293534104e-06, "loss": 0.8025, "step": 3839 }, { "epoch": 0.5, "grad_norm": 0.6765356063842773, "learning_rate": 5.1909447346169614e-06, "loss": 1.0388, "step": 3840 }, { "epoch": 0.5, "grad_norm": 0.6361967921257019, "learning_rate": 5.188824141303715e-06, "loss": 1.0094, "step": 3841 }, { "epoch": 0.5, "grad_norm": 1.4014209508895874, "learning_rate": 5.18670351397636e-06, "loss": 0.818, "step": 3842 }, { "epoch": 0.5, "grad_norm": 0.5936711430549622, "learning_rate": 5.184582853016897e-06, "loss": 1.0177, "step": 3843 }, { "epoch": 0.5, "grad_norm": 0.6208168268203735, "learning_rate": 5.182462158807338e-06, "loss": 1.022, "step": 3844 }, { "epoch": 0.5, "grad_norm": 1.344887137413025, "learning_rate": 5.180341431729693e-06, "loss": 0.8206, "step": 3845 }, { "epoch": 0.5, "grad_norm": 1.3045891523361206, "learning_rate": 5.178220672165985e-06, "loss": 0.7701, "step": 3846 }, { "epoch": 0.5, "grad_norm": 1.3605225086212158, "learning_rate": 5.176099880498239e-06, "loss": 0.8199, "step": 3847 }, { "epoch": 0.5, "grad_norm": 1.3798514604568481, "learning_rate": 5.173979057108487e-06, "loss": 0.8039, "step": 3848 }, { "epoch": 0.5, "grad_norm": 1.364398717880249, "learning_rate": 5.171858202378769e-06, "loss": 0.8575, "step": 3849 }, { "epoch": 0.5, "grad_norm": 0.7877320647239685, "learning_rate": 5.169737316691125e-06, "loss": 1.0186, "step": 3850 }, { "epoch": 0.5, "grad_norm": 1.3397114276885986, "learning_rate": 5.167616400427604e-06, "loss": 0.9063, "step": 3851 }, { "epoch": 0.5, "grad_norm": 1.3272969722747803, "learning_rate": 5.165495453970262e-06, "loss": 0.8494, "step": 3852 }, { "epoch": 0.5, "grad_norm": 1.3611959218978882, "learning_rate": 5.163374477701159e-06, "loss": 0.8378, "step": 3853 }, { "epoch": 0.5, "grad_norm": 1.4452593326568604, "learning_rate": 5.161253472002358e-06, "loss": 0.8807, "step": 3854 }, { "epoch": 0.51, "grad_norm": 1.3197680711746216, "learning_rate": 5.1591324372559294e-06, "loss": 0.7981, "step": 3855 }, { "epoch": 0.51, "grad_norm": 1.410827398300171, "learning_rate": 5.157011373843951e-06, "loss": 0.7911, "step": 3856 }, { "epoch": 0.51, "grad_norm": 1.3858360052108765, "learning_rate": 5.154890282148502e-06, "loss": 0.7733, "step": 3857 }, { "epoch": 0.51, "grad_norm": 1.2739986181259155, "learning_rate": 5.152769162551671e-06, "loss": 0.7947, "step": 3858 }, { "epoch": 0.51, "grad_norm": 1.5389373302459717, "learning_rate": 5.150648015435544e-06, "loss": 0.813, "step": 3859 }, { "epoch": 0.51, "grad_norm": 1.2494876384735107, "learning_rate": 5.148526841182222e-06, "loss": 0.7887, "step": 3860 }, { "epoch": 0.51, "grad_norm": 1.3201303482055664, "learning_rate": 5.146405640173802e-06, "loss": 0.7423, "step": 3861 }, { "epoch": 0.51, "grad_norm": 1.297305703163147, "learning_rate": 5.1442844127923915e-06, "loss": 0.8274, "step": 3862 }, { "epoch": 0.51, "grad_norm": 1.3748242855072021, "learning_rate": 5.142163159420102e-06, "loss": 0.8365, "step": 3863 }, { "epoch": 0.51, "grad_norm": 1.3967629671096802, "learning_rate": 5.140041880439045e-06, "loss": 0.8383, "step": 3864 }, { "epoch": 0.51, "grad_norm": 1.3670928478240967, "learning_rate": 5.137920576231342e-06, "loss": 0.8453, "step": 3865 }, { "epoch": 0.51, "grad_norm": 1.339058518409729, "learning_rate": 5.135799247179118e-06, "loss": 0.8347, "step": 3866 }, { "epoch": 0.51, "grad_norm": 1.3622868061065674, "learning_rate": 5.133677893664501e-06, "loss": 0.8302, "step": 3867 }, { "epoch": 0.51, "grad_norm": 0.6960027813911438, "learning_rate": 5.131556516069625e-06, "loss": 0.9882, "step": 3868 }, { "epoch": 0.51, "grad_norm": 1.292918086051941, "learning_rate": 5.129435114776624e-06, "loss": 0.7988, "step": 3869 }, { "epoch": 0.51, "grad_norm": 1.3152223825454712, "learning_rate": 5.127313690167643e-06, "loss": 0.8241, "step": 3870 }, { "epoch": 0.51, "grad_norm": 1.3719329833984375, "learning_rate": 5.125192242624828e-06, "loss": 0.8612, "step": 3871 }, { "epoch": 0.51, "grad_norm": 1.380407452583313, "learning_rate": 5.123070772530328e-06, "loss": 0.8256, "step": 3872 }, { "epoch": 0.51, "grad_norm": 1.3068009614944458, "learning_rate": 5.1209492802662966e-06, "loss": 0.8254, "step": 3873 }, { "epoch": 0.51, "grad_norm": 1.3886265754699707, "learning_rate": 5.118827766214891e-06, "loss": 0.8157, "step": 3874 }, { "epoch": 0.51, "grad_norm": 1.3921568393707275, "learning_rate": 5.116706230758276e-06, "loss": 0.8015, "step": 3875 }, { "epoch": 0.51, "grad_norm": 1.337790608406067, "learning_rate": 5.114584674278616e-06, "loss": 0.8728, "step": 3876 }, { "epoch": 0.51, "grad_norm": 1.3348222970962524, "learning_rate": 5.11246309715808e-06, "loss": 0.8443, "step": 3877 }, { "epoch": 0.51, "grad_norm": 1.3084266185760498, "learning_rate": 5.1103414997788424e-06, "loss": 0.8278, "step": 3878 }, { "epoch": 0.51, "grad_norm": 0.7426634430885315, "learning_rate": 5.10821988252308e-06, "loss": 1.0355, "step": 3879 }, { "epoch": 0.51, "grad_norm": 1.3258490562438965, "learning_rate": 5.106098245772974e-06, "loss": 0.8375, "step": 3880 }, { "epoch": 0.51, "grad_norm": 1.3952897787094116, "learning_rate": 5.1039765899107055e-06, "loss": 0.8172, "step": 3881 }, { "epoch": 0.51, "grad_norm": 0.5981616973876953, "learning_rate": 5.101854915318466e-06, "loss": 1.0148, "step": 3882 }, { "epoch": 0.51, "grad_norm": 1.3151518106460571, "learning_rate": 5.099733222378443e-06, "loss": 0.805, "step": 3883 }, { "epoch": 0.51, "grad_norm": 1.3759174346923828, "learning_rate": 5.097611511472833e-06, "loss": 0.8468, "step": 3884 }, { "epoch": 0.51, "grad_norm": 0.6641613841056824, "learning_rate": 5.095489782983833e-06, "loss": 1.0414, "step": 3885 }, { "epoch": 0.51, "grad_norm": 1.2646328210830688, "learning_rate": 5.0933680372936445e-06, "loss": 0.8153, "step": 3886 }, { "epoch": 0.51, "grad_norm": 1.3356201648712158, "learning_rate": 5.091246274784468e-06, "loss": 0.8304, "step": 3887 }, { "epoch": 0.51, "grad_norm": 1.289957046508789, "learning_rate": 5.089124495838513e-06, "loss": 0.8339, "step": 3888 }, { "epoch": 0.51, "grad_norm": 1.3259456157684326, "learning_rate": 5.087002700837988e-06, "loss": 0.8133, "step": 3889 }, { "epoch": 0.51, "grad_norm": 1.2786617279052734, "learning_rate": 5.084880890165107e-06, "loss": 0.8292, "step": 3890 }, { "epoch": 0.51, "grad_norm": 1.6324208974838257, "learning_rate": 5.0827590642020856e-06, "loss": 0.7954, "step": 3891 }, { "epoch": 0.51, "grad_norm": 0.8396244645118713, "learning_rate": 5.080637223331138e-06, "loss": 1.0187, "step": 3892 }, { "epoch": 0.51, "grad_norm": 1.4023345708847046, "learning_rate": 5.07851536793449e-06, "loss": 0.7944, "step": 3893 }, { "epoch": 0.51, "grad_norm": 1.2990020513534546, "learning_rate": 5.0763934983943634e-06, "loss": 0.8473, "step": 3894 }, { "epoch": 0.51, "grad_norm": 1.5014960765838623, "learning_rate": 5.074271615092984e-06, "loss": 0.8063, "step": 3895 }, { "epoch": 0.51, "grad_norm": 1.2322580814361572, "learning_rate": 5.072149718412579e-06, "loss": 0.816, "step": 3896 }, { "epoch": 0.51, "grad_norm": 1.260581612586975, "learning_rate": 5.07002780873538e-06, "loss": 0.7922, "step": 3897 }, { "epoch": 0.51, "grad_norm": 1.4136369228363037, "learning_rate": 5.067905886443621e-06, "loss": 0.7985, "step": 3898 }, { "epoch": 0.51, "grad_norm": 1.3836851119995117, "learning_rate": 5.0657839519195396e-06, "loss": 0.853, "step": 3899 }, { "epoch": 0.51, "grad_norm": 1.3496990203857422, "learning_rate": 5.06366200554537e-06, "loss": 0.8323, "step": 3900 }, { "epoch": 0.51, "grad_norm": 1.4420685768127441, "learning_rate": 5.061540047703354e-06, "loss": 0.8072, "step": 3901 }, { "epoch": 0.51, "grad_norm": 1.2343335151672363, "learning_rate": 5.0594180787757305e-06, "loss": 0.8017, "step": 3902 }, { "epoch": 0.51, "grad_norm": 1.248738408088684, "learning_rate": 5.057296099144747e-06, "loss": 0.814, "step": 3903 }, { "epoch": 0.51, "grad_norm": 1.3049319982528687, "learning_rate": 5.055174109192649e-06, "loss": 0.7998, "step": 3904 }, { "epoch": 0.51, "grad_norm": 0.726016104221344, "learning_rate": 5.053052109301682e-06, "loss": 1.0136, "step": 3905 }, { "epoch": 0.51, "grad_norm": 0.6646234393119812, "learning_rate": 5.0509300998540974e-06, "loss": 1.0085, "step": 3906 }, { "epoch": 0.51, "grad_norm": 1.3058379888534546, "learning_rate": 5.048808081232146e-06, "loss": 0.8172, "step": 3907 }, { "epoch": 0.51, "grad_norm": 1.3852169513702393, "learning_rate": 5.046686053818081e-06, "loss": 0.8203, "step": 3908 }, { "epoch": 0.51, "grad_norm": 1.3887712955474854, "learning_rate": 5.044564017994157e-06, "loss": 0.8012, "step": 3909 }, { "epoch": 0.51, "grad_norm": 1.4770618677139282, "learning_rate": 5.042441974142628e-06, "loss": 0.7709, "step": 3910 }, { "epoch": 0.51, "grad_norm": 1.2580626010894775, "learning_rate": 5.040319922645752e-06, "loss": 0.7882, "step": 3911 }, { "epoch": 0.51, "grad_norm": 1.4822030067443848, "learning_rate": 5.03819786388579e-06, "loss": 0.8766, "step": 3912 }, { "epoch": 0.51, "grad_norm": 0.7721188068389893, "learning_rate": 5.0360757982450005e-06, "loss": 1.0242, "step": 3913 }, { "epoch": 0.51, "grad_norm": 1.3465229272842407, "learning_rate": 5.033953726105646e-06, "loss": 0.7983, "step": 3914 }, { "epoch": 0.51, "grad_norm": 1.4272263050079346, "learning_rate": 5.031831647849987e-06, "loss": 0.8137, "step": 3915 }, { "epoch": 0.51, "grad_norm": 0.6769410371780396, "learning_rate": 5.0297095638602885e-06, "loss": 1.0148, "step": 3916 }, { "epoch": 0.51, "grad_norm": 0.6069938540458679, "learning_rate": 5.0275874745188145e-06, "loss": 1.0275, "step": 3917 }, { "epoch": 0.51, "grad_norm": 1.2727214097976685, "learning_rate": 5.025465380207832e-06, "loss": 0.7785, "step": 3918 }, { "epoch": 0.51, "grad_norm": 1.788883090019226, "learning_rate": 5.023343281309607e-06, "loss": 0.837, "step": 3919 }, { "epoch": 0.51, "grad_norm": 1.2518349885940552, "learning_rate": 5.021221178206405e-06, "loss": 0.7985, "step": 3920 }, { "epoch": 0.51, "grad_norm": 1.2398122549057007, "learning_rate": 5.019099071280496e-06, "loss": 0.8376, "step": 3921 }, { "epoch": 0.51, "grad_norm": 1.3686500787734985, "learning_rate": 5.016976960914149e-06, "loss": 0.7886, "step": 3922 }, { "epoch": 0.51, "grad_norm": 1.3219542503356934, "learning_rate": 5.014854847489633e-06, "loss": 0.8917, "step": 3923 }, { "epoch": 0.51, "grad_norm": 1.6338194608688354, "learning_rate": 5.0127327313892164e-06, "loss": 0.8263, "step": 3924 }, { "epoch": 0.51, "grad_norm": 1.245627999305725, "learning_rate": 5.0106106129951704e-06, "loss": 0.7905, "step": 3925 }, { "epoch": 0.51, "grad_norm": 2.3723316192626953, "learning_rate": 5.008488492689767e-06, "loss": 0.7653, "step": 3926 }, { "epoch": 0.51, "grad_norm": 1.372410774230957, "learning_rate": 5.006366370855277e-06, "loss": 0.7951, "step": 3927 }, { "epoch": 0.51, "grad_norm": 1.3207619190216064, "learning_rate": 5.0042442478739715e-06, "loss": 0.7869, "step": 3928 }, { "epoch": 0.51, "grad_norm": 1.4101073741912842, "learning_rate": 5.002122124128122e-06, "loss": 0.8209, "step": 3929 }, { "epoch": 0.51, "grad_norm": 0.9055166840553284, "learning_rate": 5e-06, "loss": 1.0194, "step": 3930 }, { "epoch": 0.52, "grad_norm": 1.4007892608642578, "learning_rate": 4.99787787587188e-06, "loss": 0.8463, "step": 3931 }, { "epoch": 0.52, "grad_norm": 1.6445558071136475, "learning_rate": 4.995755752126029e-06, "loss": 0.8781, "step": 3932 }, { "epoch": 0.52, "grad_norm": 0.6664135456085205, "learning_rate": 4.993633629144726e-06, "loss": 1.0109, "step": 3933 }, { "epoch": 0.52, "grad_norm": 1.2962361574172974, "learning_rate": 4.991511507310233e-06, "loss": 0.7971, "step": 3934 }, { "epoch": 0.52, "grad_norm": 1.468740463256836, "learning_rate": 4.989389387004831e-06, "loss": 0.8175, "step": 3935 }, { "epoch": 0.52, "grad_norm": 1.2916414737701416, "learning_rate": 4.987267268610786e-06, "loss": 0.7884, "step": 3936 }, { "epoch": 0.52, "grad_norm": 0.7539187669754028, "learning_rate": 4.985145152510369e-06, "loss": 1.0429, "step": 3937 }, { "epoch": 0.52, "grad_norm": 1.3510656356811523, "learning_rate": 4.9830230390858525e-06, "loss": 0.7773, "step": 3938 }, { "epoch": 0.52, "grad_norm": 1.3877100944519043, "learning_rate": 4.9809009287195045e-06, "loss": 0.7937, "step": 3939 }, { "epoch": 0.52, "grad_norm": 0.6881576776504517, "learning_rate": 4.978778821793596e-06, "loss": 1.0161, "step": 3940 }, { "epoch": 0.52, "grad_norm": 1.298241376876831, "learning_rate": 4.976656718690395e-06, "loss": 0.8189, "step": 3941 }, { "epoch": 0.52, "grad_norm": 0.6622557640075684, "learning_rate": 4.974534619792169e-06, "loss": 1.0356, "step": 3942 }, { "epoch": 0.52, "grad_norm": 1.2883514165878296, "learning_rate": 4.972412525481186e-06, "loss": 0.8173, "step": 3943 }, { "epoch": 0.52, "grad_norm": 0.629728376865387, "learning_rate": 4.9702904361397114e-06, "loss": 1.0252, "step": 3944 }, { "epoch": 0.52, "grad_norm": 1.4238393306732178, "learning_rate": 4.968168352150014e-06, "loss": 0.8612, "step": 3945 }, { "epoch": 0.52, "grad_norm": 1.3077847957611084, "learning_rate": 4.966046273894356e-06, "loss": 0.8117, "step": 3946 }, { "epoch": 0.52, "grad_norm": 1.332853078842163, "learning_rate": 4.9639242017549995e-06, "loss": 0.8367, "step": 3947 }, { "epoch": 0.52, "grad_norm": 1.520127534866333, "learning_rate": 4.961802136114211e-06, "loss": 0.8039, "step": 3948 }, { "epoch": 0.52, "grad_norm": 1.2393156290054321, "learning_rate": 4.95968007735425e-06, "loss": 0.7774, "step": 3949 }, { "epoch": 0.52, "grad_norm": 1.4811437129974365, "learning_rate": 4.957558025857374e-06, "loss": 0.8034, "step": 3950 }, { "epoch": 0.52, "grad_norm": 0.7378031611442566, "learning_rate": 4.9554359820058465e-06, "loss": 1.0363, "step": 3951 }, { "epoch": 0.52, "grad_norm": 1.4766961336135864, "learning_rate": 4.95331394618192e-06, "loss": 0.7723, "step": 3952 }, { "epoch": 0.52, "grad_norm": 1.3697009086608887, "learning_rate": 4.951191918767855e-06, "loss": 0.8226, "step": 3953 }, { "epoch": 0.52, "grad_norm": 1.4041955471038818, "learning_rate": 4.949069900145904e-06, "loss": 0.8316, "step": 3954 }, { "epoch": 0.52, "grad_norm": 1.4117976427078247, "learning_rate": 4.946947890698318e-06, "loss": 0.8067, "step": 3955 }, { "epoch": 0.52, "grad_norm": 1.3917016983032227, "learning_rate": 4.944825890807353e-06, "loss": 0.8184, "step": 3956 }, { "epoch": 0.52, "grad_norm": 1.3696125745773315, "learning_rate": 4.942703900855254e-06, "loss": 0.87, "step": 3957 }, { "epoch": 0.52, "grad_norm": 1.375594973564148, "learning_rate": 4.94058192122427e-06, "loss": 0.8359, "step": 3958 }, { "epoch": 0.52, "grad_norm": 1.7024248838424683, "learning_rate": 4.938459952296649e-06, "loss": 0.7965, "step": 3959 }, { "epoch": 0.52, "grad_norm": 1.4214141368865967, "learning_rate": 4.936337994454632e-06, "loss": 0.7747, "step": 3960 }, { "epoch": 0.52, "grad_norm": 1.382473111152649, "learning_rate": 4.934216048080462e-06, "loss": 0.8031, "step": 3961 }, { "epoch": 0.52, "grad_norm": 1.3366432189941406, "learning_rate": 4.9320941135563786e-06, "loss": 0.7473, "step": 3962 }, { "epoch": 0.52, "grad_norm": 1.4242229461669922, "learning_rate": 4.929972191264621e-06, "loss": 0.8391, "step": 3963 }, { "epoch": 0.52, "grad_norm": 1.3640762567520142, "learning_rate": 4.927850281587424e-06, "loss": 0.8325, "step": 3964 }, { "epoch": 0.52, "grad_norm": 1.5454593896865845, "learning_rate": 4.925728384907018e-06, "loss": 0.8165, "step": 3965 }, { "epoch": 0.52, "grad_norm": 0.6855883598327637, "learning_rate": 4.923606501605639e-06, "loss": 1.0245, "step": 3966 }, { "epoch": 0.52, "grad_norm": 1.6510921716690063, "learning_rate": 4.92148463206551e-06, "loss": 0.8274, "step": 3967 }, { "epoch": 0.52, "grad_norm": 1.293726921081543, "learning_rate": 4.9193627766688625e-06, "loss": 0.828, "step": 3968 }, { "epoch": 0.52, "grad_norm": 1.3239390850067139, "learning_rate": 4.917240935797918e-06, "loss": 0.7448, "step": 3969 }, { "epoch": 0.52, "grad_norm": 1.3733911514282227, "learning_rate": 4.915119109834894e-06, "loss": 0.7665, "step": 3970 }, { "epoch": 0.52, "grad_norm": 1.352418065071106, "learning_rate": 4.9129972991620135e-06, "loss": 0.7469, "step": 3971 }, { "epoch": 0.52, "grad_norm": 0.6438746452331543, "learning_rate": 4.910875504161488e-06, "loss": 1.0261, "step": 3972 }, { "epoch": 0.52, "grad_norm": 1.3514320850372314, "learning_rate": 4.9087537252155335e-06, "loss": 0.8093, "step": 3973 }, { "epoch": 0.52, "grad_norm": 0.6229220628738403, "learning_rate": 4.906631962706359e-06, "loss": 1.0307, "step": 3974 }, { "epoch": 0.52, "grad_norm": 1.4448904991149902, "learning_rate": 4.904510217016168e-06, "loss": 0.8069, "step": 3975 }, { "epoch": 0.52, "grad_norm": 0.6525881886482239, "learning_rate": 4.9023884885271675e-06, "loss": 1.0376, "step": 3976 }, { "epoch": 0.52, "grad_norm": 1.3884700536727905, "learning_rate": 4.900266777621559e-06, "loss": 0.7795, "step": 3977 }, { "epoch": 0.52, "grad_norm": 1.401213526725769, "learning_rate": 4.898145084681536e-06, "loss": 0.7764, "step": 3978 }, { "epoch": 0.52, "grad_norm": 1.3201115131378174, "learning_rate": 4.896023410089295e-06, "loss": 0.8737, "step": 3979 }, { "epoch": 0.52, "grad_norm": 0.6429000496864319, "learning_rate": 4.893901754227028e-06, "loss": 1.0297, "step": 3980 }, { "epoch": 0.52, "grad_norm": 0.6586340069770813, "learning_rate": 4.891780117476921e-06, "loss": 1.0018, "step": 3981 }, { "epoch": 0.52, "grad_norm": 0.6299601197242737, "learning_rate": 4.889658500221159e-06, "loss": 1.0226, "step": 3982 }, { "epoch": 0.52, "grad_norm": 0.6060304045677185, "learning_rate": 4.887536902841921e-06, "loss": 1.0089, "step": 3983 }, { "epoch": 0.52, "grad_norm": 1.337146282196045, "learning_rate": 4.885415325721386e-06, "loss": 0.8121, "step": 3984 }, { "epoch": 0.52, "grad_norm": 1.45977783203125, "learning_rate": 4.883293769241725e-06, "loss": 0.7937, "step": 3985 }, { "epoch": 0.52, "grad_norm": 1.2933069467544556, "learning_rate": 4.88117223378511e-06, "loss": 0.7981, "step": 3986 }, { "epoch": 0.52, "grad_norm": 1.3541896343231201, "learning_rate": 4.879050719733707e-06, "loss": 0.7908, "step": 3987 }, { "epoch": 0.52, "grad_norm": 1.332101821899414, "learning_rate": 4.876929227469673e-06, "loss": 0.7912, "step": 3988 }, { "epoch": 0.52, "grad_norm": 1.3751411437988281, "learning_rate": 4.874807757375174e-06, "loss": 0.7535, "step": 3989 }, { "epoch": 0.52, "grad_norm": 0.8159030079841614, "learning_rate": 4.8726863098323566e-06, "loss": 1.002, "step": 3990 }, { "epoch": 0.52, "grad_norm": 1.3634120225906372, "learning_rate": 4.870564885223377e-06, "loss": 0.8023, "step": 3991 }, { "epoch": 0.52, "grad_norm": 0.6395212411880493, "learning_rate": 4.868443483930379e-06, "loss": 1.0171, "step": 3992 }, { "epoch": 0.52, "grad_norm": 1.7028542757034302, "learning_rate": 4.8663221063355e-06, "loss": 0.8246, "step": 3993 }, { "epoch": 0.52, "grad_norm": 0.66887366771698, "learning_rate": 4.8642007528208835e-06, "loss": 1.0402, "step": 3994 }, { "epoch": 0.52, "grad_norm": 1.4012254476547241, "learning_rate": 4.862079423768658e-06, "loss": 0.7727, "step": 3995 }, { "epoch": 0.52, "grad_norm": 1.408952236175537, "learning_rate": 4.859958119560956e-06, "loss": 0.7572, "step": 3996 }, { "epoch": 0.52, "grad_norm": 1.2494210004806519, "learning_rate": 4.8578368405799e-06, "loss": 0.8512, "step": 3997 }, { "epoch": 0.52, "grad_norm": 0.6830434203147888, "learning_rate": 4.8557155872076085e-06, "loss": 1.0315, "step": 3998 }, { "epoch": 0.52, "grad_norm": 0.6256372928619385, "learning_rate": 4.8535943598261984e-06, "loss": 1.0363, "step": 3999 }, { "epoch": 0.52, "grad_norm": 1.3226521015167236, "learning_rate": 4.8514731588177784e-06, "loss": 0.7686, "step": 4000 }, { "epoch": 0.52, "grad_norm": 1.4286394119262695, "learning_rate": 4.849351984564457e-06, "loss": 0.8372, "step": 4001 }, { "epoch": 0.52, "grad_norm": 1.312138557434082, "learning_rate": 4.847230837448331e-06, "loss": 0.7993, "step": 4002 }, { "epoch": 0.52, "grad_norm": 0.6605648398399353, "learning_rate": 4.845109717851498e-06, "loss": 1.0348, "step": 4003 }, { "epoch": 0.52, "grad_norm": 1.4150105714797974, "learning_rate": 4.842988626156051e-06, "loss": 0.79, "step": 4004 }, { "epoch": 0.52, "grad_norm": 1.2941406965255737, "learning_rate": 4.840867562744073e-06, "loss": 0.8026, "step": 4005 }, { "epoch": 0.52, "grad_norm": 1.3729265928268433, "learning_rate": 4.8387465279976446e-06, "loss": 0.8135, "step": 4006 }, { "epoch": 0.53, "grad_norm": 0.6516928672790527, "learning_rate": 4.836625522298844e-06, "loss": 1.0581, "step": 4007 }, { "epoch": 0.53, "grad_norm": 1.3973379135131836, "learning_rate": 4.834504546029739e-06, "loss": 0.7988, "step": 4008 }, { "epoch": 0.53, "grad_norm": 1.2840696573257446, "learning_rate": 4.832383599572397e-06, "loss": 0.8074, "step": 4009 }, { "epoch": 0.53, "grad_norm": 1.3911471366882324, "learning_rate": 4.830262683308878e-06, "loss": 0.8142, "step": 4010 }, { "epoch": 0.53, "grad_norm": 0.6323899030685425, "learning_rate": 4.828141797621232e-06, "loss": 1.0201, "step": 4011 }, { "epoch": 0.53, "grad_norm": 1.3085185289382935, "learning_rate": 4.826020942891514e-06, "loss": 0.8534, "step": 4012 }, { "epoch": 0.53, "grad_norm": 1.3114265203475952, "learning_rate": 4.823900119501762e-06, "loss": 0.7583, "step": 4013 }, { "epoch": 0.53, "grad_norm": 0.6076804995536804, "learning_rate": 4.821779327834017e-06, "loss": 1.0413, "step": 4014 }, { "epoch": 0.53, "grad_norm": 1.3365609645843506, "learning_rate": 4.81965856827031e-06, "loss": 0.823, "step": 4015 }, { "epoch": 0.53, "grad_norm": 1.292633295059204, "learning_rate": 4.817537841192664e-06, "loss": 0.8147, "step": 4016 }, { "epoch": 0.53, "grad_norm": 1.5098648071289062, "learning_rate": 4.815417146983104e-06, "loss": 0.8746, "step": 4017 }, { "epoch": 0.53, "grad_norm": 0.6139063835144043, "learning_rate": 4.813296486023642e-06, "loss": 1.0149, "step": 4018 }, { "epoch": 0.53, "grad_norm": 1.2808787822723389, "learning_rate": 4.811175858696286e-06, "loss": 0.7523, "step": 4019 }, { "epoch": 0.53, "grad_norm": 0.5975820422172546, "learning_rate": 4.809055265383039e-06, "loss": 1.0554, "step": 4020 }, { "epoch": 0.53, "grad_norm": 1.2132126092910767, "learning_rate": 4.806934706465897e-06, "loss": 0.7961, "step": 4021 }, { "epoch": 0.53, "grad_norm": 1.3768726587295532, "learning_rate": 4.80481418232685e-06, "loss": 0.8354, "step": 4022 }, { "epoch": 0.53, "grad_norm": 0.5966764688491821, "learning_rate": 4.80269369334788e-06, "loss": 1.0131, "step": 4023 }, { "epoch": 0.53, "grad_norm": 1.25246000289917, "learning_rate": 4.800573239910968e-06, "loss": 0.8174, "step": 4024 }, { "epoch": 0.53, "grad_norm": 0.6300486922264099, "learning_rate": 4.798452822398083e-06, "loss": 1.0372, "step": 4025 }, { "epoch": 0.53, "grad_norm": 1.2447855472564697, "learning_rate": 4.796332441191185e-06, "loss": 0.7502, "step": 4026 }, { "epoch": 0.53, "grad_norm": 1.3656519651412964, "learning_rate": 4.7942120966722405e-06, "loss": 0.8537, "step": 4027 }, { "epoch": 0.53, "grad_norm": 1.283416986465454, "learning_rate": 4.792091789223193e-06, "loss": 0.8463, "step": 4028 }, { "epoch": 0.53, "grad_norm": 1.349194049835205, "learning_rate": 4.789971519225994e-06, "loss": 0.7407, "step": 4029 }, { "epoch": 0.53, "grad_norm": 0.6687452793121338, "learning_rate": 4.7878512870625775e-06, "loss": 1.0364, "step": 4030 }, { "epoch": 0.53, "grad_norm": 0.7476613521575928, "learning_rate": 4.785731093114873e-06, "loss": 1.0415, "step": 4031 }, { "epoch": 0.53, "grad_norm": 0.6361120343208313, "learning_rate": 4.783610937764808e-06, "loss": 1.0332, "step": 4032 }, { "epoch": 0.53, "grad_norm": 1.4003660678863525, "learning_rate": 4.781490821394299e-06, "loss": 0.8063, "step": 4033 }, { "epoch": 0.53, "grad_norm": 1.3489421606063843, "learning_rate": 4.7793707443852536e-06, "loss": 0.8123, "step": 4034 }, { "epoch": 0.53, "grad_norm": 1.4418102502822876, "learning_rate": 4.77725070711958e-06, "loss": 0.7446, "step": 4035 }, { "epoch": 0.53, "grad_norm": 1.537082552909851, "learning_rate": 4.775130709979167e-06, "loss": 0.8075, "step": 4036 }, { "epoch": 0.53, "grad_norm": 1.341296672821045, "learning_rate": 4.7730107533459115e-06, "loss": 0.8171, "step": 4037 }, { "epoch": 0.53, "grad_norm": 2.0174169540405273, "learning_rate": 4.770890837601691e-06, "loss": 0.8273, "step": 4038 }, { "epoch": 0.53, "grad_norm": 1.7598177194595337, "learning_rate": 4.768770963128376e-06, "loss": 0.8185, "step": 4039 }, { "epoch": 0.53, "grad_norm": 1.3835866451263428, "learning_rate": 4.76665113030784e-06, "loss": 0.8706, "step": 4040 }, { "epoch": 0.53, "grad_norm": 1.3246086835861206, "learning_rate": 4.764531339521938e-06, "loss": 0.8122, "step": 4041 }, { "epoch": 0.53, "grad_norm": 1.8975584506988525, "learning_rate": 4.762411591152522e-06, "loss": 0.8186, "step": 4042 }, { "epoch": 0.53, "grad_norm": 1.3409621715545654, "learning_rate": 4.7602918855814365e-06, "loss": 0.8026, "step": 4043 }, { "epoch": 0.53, "grad_norm": 1.0145652294158936, "learning_rate": 4.7581722231905165e-06, "loss": 1.0254, "step": 4044 }, { "epoch": 0.53, "grad_norm": 1.3215423822402954, "learning_rate": 4.756052604361594e-06, "loss": 0.8573, "step": 4045 }, { "epoch": 0.53, "grad_norm": 0.6969495415687561, "learning_rate": 4.753933029476484e-06, "loss": 1.0118, "step": 4046 }, { "epoch": 0.53, "grad_norm": 1.4220337867736816, "learning_rate": 4.751813498917004e-06, "loss": 0.8085, "step": 4047 }, { "epoch": 0.53, "grad_norm": 1.3670133352279663, "learning_rate": 4.749694013064958e-06, "loss": 0.8301, "step": 4048 }, { "epoch": 0.53, "grad_norm": 1.2906596660614014, "learning_rate": 4.747574572302138e-06, "loss": 0.8828, "step": 4049 }, { "epoch": 0.53, "grad_norm": 0.9372462034225464, "learning_rate": 4.74545517701034e-06, "loss": 0.9947, "step": 4050 }, { "epoch": 0.53, "grad_norm": 0.8811337351799011, "learning_rate": 4.743335827571337e-06, "loss": 1.0102, "step": 4051 }, { "epoch": 0.53, "grad_norm": 1.3011633157730103, "learning_rate": 4.7412165243669065e-06, "loss": 0.8501, "step": 4052 }, { "epoch": 0.53, "grad_norm": 0.6099061369895935, "learning_rate": 4.739097267778811e-06, "loss": 1.0376, "step": 4053 }, { "epoch": 0.53, "grad_norm": 0.7517300844192505, "learning_rate": 4.736978058188803e-06, "loss": 1.0219, "step": 4054 }, { "epoch": 0.53, "grad_norm": 1.286749243736267, "learning_rate": 4.734858895978633e-06, "loss": 0.8168, "step": 4055 }, { "epoch": 0.53, "grad_norm": 0.7266089916229248, "learning_rate": 4.732739781530036e-06, "loss": 1.0446, "step": 4056 }, { "epoch": 0.53, "grad_norm": 1.794356346130371, "learning_rate": 4.730620715224745e-06, "loss": 0.8289, "step": 4057 }, { "epoch": 0.53, "grad_norm": 1.339590311050415, "learning_rate": 4.72850169744448e-06, "loss": 0.8024, "step": 4058 }, { "epoch": 0.53, "grad_norm": 1.3399733304977417, "learning_rate": 4.726382728570952e-06, "loss": 0.8409, "step": 4059 }, { "epoch": 0.53, "grad_norm": 1.3268476724624634, "learning_rate": 4.724263808985866e-06, "loss": 0.8298, "step": 4060 }, { "epoch": 0.53, "grad_norm": 1.238791584968567, "learning_rate": 4.722144939070915e-06, "loss": 0.7877, "step": 4061 }, { "epoch": 0.53, "grad_norm": 1.4104033708572388, "learning_rate": 4.720026119207787e-06, "loss": 0.7872, "step": 4062 }, { "epoch": 0.53, "grad_norm": 1.3051927089691162, "learning_rate": 4.717907349778156e-06, "loss": 0.7698, "step": 4063 }, { "epoch": 0.53, "grad_norm": 1.449804663658142, "learning_rate": 4.715788631163692e-06, "loss": 0.7654, "step": 4064 }, { "epoch": 0.53, "grad_norm": 1.349682331085205, "learning_rate": 4.713669963746053e-06, "loss": 0.7973, "step": 4065 }, { "epoch": 0.53, "grad_norm": 0.9173910021781921, "learning_rate": 4.7115513479068866e-06, "loss": 1.05, "step": 4066 }, { "epoch": 0.53, "grad_norm": 0.7607377767562866, "learning_rate": 4.709432784027832e-06, "loss": 1.0277, "step": 4067 }, { "epoch": 0.53, "grad_norm": 0.6257115602493286, "learning_rate": 4.707314272490524e-06, "loss": 1.0186, "step": 4068 }, { "epoch": 0.53, "grad_norm": 1.3205797672271729, "learning_rate": 4.70519581367658e-06, "loss": 0.8547, "step": 4069 }, { "epoch": 0.53, "grad_norm": 1.3512089252471924, "learning_rate": 4.703077407967614e-06, "loss": 0.7933, "step": 4070 }, { "epoch": 0.53, "grad_norm": 1.5714268684387207, "learning_rate": 4.700959055745227e-06, "loss": 0.8073, "step": 4071 }, { "epoch": 0.53, "grad_norm": 1.4268062114715576, "learning_rate": 4.698840757391009e-06, "loss": 0.7877, "step": 4072 }, { "epoch": 0.53, "grad_norm": 1.5585349798202515, "learning_rate": 4.696722513286547e-06, "loss": 0.8698, "step": 4073 }, { "epoch": 0.53, "grad_norm": 1.104549765586853, "learning_rate": 4.6946043238134104e-06, "loss": 1.0156, "step": 4074 }, { "epoch": 0.53, "grad_norm": 1.5751581192016602, "learning_rate": 4.692486189353166e-06, "loss": 0.8325, "step": 4075 }, { "epoch": 0.53, "grad_norm": 1.3374918699264526, "learning_rate": 4.690368110287365e-06, "loss": 0.816, "step": 4076 }, { "epoch": 0.53, "grad_norm": 0.6603697538375854, "learning_rate": 4.688250086997547e-06, "loss": 1.0035, "step": 4077 }, { "epoch": 0.53, "grad_norm": 0.6249760985374451, "learning_rate": 4.6861321198652535e-06, "loss": 0.9866, "step": 4078 }, { "epoch": 0.53, "grad_norm": 0.6389419436454773, "learning_rate": 4.684014209272e-06, "loss": 1.0184, "step": 4079 }, { "epoch": 0.53, "grad_norm": 1.6370724439620972, "learning_rate": 4.681896355599302e-06, "loss": 0.8069, "step": 4080 }, { "epoch": 0.53, "grad_norm": 1.4028815031051636, "learning_rate": 4.679778559228664e-06, "loss": 0.7742, "step": 4081 }, { "epoch": 0.53, "grad_norm": 1.412598967552185, "learning_rate": 4.677660820541575e-06, "loss": 0.8147, "step": 4082 }, { "epoch": 0.54, "grad_norm": 1.4817713499069214, "learning_rate": 4.67554313991952e-06, "loss": 0.8579, "step": 4083 }, { "epoch": 0.54, "grad_norm": 1.3354454040527344, "learning_rate": 4.673425517743967e-06, "loss": 0.8413, "step": 4084 }, { "epoch": 0.54, "grad_norm": 1.29461669921875, "learning_rate": 4.67130795439638e-06, "loss": 0.7863, "step": 4085 }, { "epoch": 0.54, "grad_norm": 1.4417413473129272, "learning_rate": 4.6691904502582095e-06, "loss": 0.7773, "step": 4086 }, { "epoch": 0.54, "grad_norm": 1.4146419763565063, "learning_rate": 4.66707300571089e-06, "loss": 0.8596, "step": 4087 }, { "epoch": 0.54, "grad_norm": 0.9204492568969727, "learning_rate": 4.664955621135858e-06, "loss": 1.0452, "step": 4088 }, { "epoch": 0.54, "grad_norm": 0.7967346906661987, "learning_rate": 4.662838296914525e-06, "loss": 1.0173, "step": 4089 }, { "epoch": 0.54, "grad_norm": 0.6445587277412415, "learning_rate": 4.6607210334283036e-06, "loss": 1.036, "step": 4090 }, { "epoch": 0.54, "grad_norm": 1.447426438331604, "learning_rate": 4.658603831058588e-06, "loss": 0.8219, "step": 4091 }, { "epoch": 0.54, "grad_norm": 1.5061777830123901, "learning_rate": 4.65648669018676e-06, "loss": 0.8281, "step": 4092 }, { "epoch": 0.54, "grad_norm": 1.3950726985931396, "learning_rate": 4.6543696111942e-06, "loss": 0.7991, "step": 4093 }, { "epoch": 0.54, "grad_norm": 1.710413932800293, "learning_rate": 4.6522525944622675e-06, "loss": 0.8495, "step": 4094 }, { "epoch": 0.54, "grad_norm": 2.090404510498047, "learning_rate": 4.650135640372313e-06, "loss": 0.843, "step": 4095 }, { "epoch": 0.54, "grad_norm": 1.3186376094818115, "learning_rate": 4.648018749305681e-06, "loss": 0.8577, "step": 4096 }, { "epoch": 0.54, "grad_norm": 1.4714927673339844, "learning_rate": 4.645901921643697e-06, "loss": 0.7804, "step": 4097 }, { "epoch": 0.54, "grad_norm": 1.370436429977417, "learning_rate": 4.643785157767682e-06, "loss": 0.8133, "step": 4098 }, { "epoch": 0.54, "grad_norm": 1.4491318464279175, "learning_rate": 4.641668458058941e-06, "loss": 0.7755, "step": 4099 }, { "epoch": 0.54, "grad_norm": 1.4076610803604126, "learning_rate": 4.6395518228987676e-06, "loss": 0.8042, "step": 4100 }, { "epoch": 0.54, "grad_norm": 1.8670079708099365, "learning_rate": 4.637435252668446e-06, "loss": 0.8288, "step": 4101 }, { "epoch": 0.54, "grad_norm": 1.211161494255066, "learning_rate": 4.635318747749247e-06, "loss": 1.0368, "step": 4102 }, { "epoch": 0.54, "grad_norm": 1.3310173749923706, "learning_rate": 4.633202308522432e-06, "loss": 0.7725, "step": 4103 }, { "epoch": 0.54, "grad_norm": 1.4166717529296875, "learning_rate": 4.631085935369247e-06, "loss": 0.8008, "step": 4104 }, { "epoch": 0.54, "grad_norm": 1.3017066717147827, "learning_rate": 4.6289696286709284e-06, "loss": 0.8243, "step": 4105 }, { "epoch": 0.54, "grad_norm": 0.696027398109436, "learning_rate": 4.626853388808702e-06, "loss": 0.9986, "step": 4106 }, { "epoch": 0.54, "grad_norm": 1.4157204627990723, "learning_rate": 4.624737216163774e-06, "loss": 0.8306, "step": 4107 }, { "epoch": 0.54, "grad_norm": 1.4030438661575317, "learning_rate": 4.622621111117352e-06, "loss": 0.7954, "step": 4108 }, { "epoch": 0.54, "grad_norm": 0.6538476347923279, "learning_rate": 4.620505074050619e-06, "loss": 1.0054, "step": 4109 }, { "epoch": 0.54, "grad_norm": 0.6472955942153931, "learning_rate": 4.6183891053447495e-06, "loss": 1.0119, "step": 4110 }, { "epoch": 0.54, "grad_norm": 1.382773995399475, "learning_rate": 4.616273205380911e-06, "loss": 0.8023, "step": 4111 }, { "epoch": 0.54, "grad_norm": 0.6314696669578552, "learning_rate": 4.614157374540248e-06, "loss": 1.0262, "step": 4112 }, { "epoch": 0.54, "grad_norm": 1.5243539810180664, "learning_rate": 4.6120416132039055e-06, "loss": 0.8085, "step": 4113 }, { "epoch": 0.54, "grad_norm": 1.258512258529663, "learning_rate": 4.6099259217530055e-06, "loss": 0.7259, "step": 4114 }, { "epoch": 0.54, "grad_norm": 1.4330527782440186, "learning_rate": 4.607810300568659e-06, "loss": 0.85, "step": 4115 }, { "epoch": 0.54, "grad_norm": 1.641115665435791, "learning_rate": 4.605694750031972e-06, "loss": 0.805, "step": 4116 }, { "epoch": 0.54, "grad_norm": 1.4867925643920898, "learning_rate": 4.603579270524028e-06, "loss": 0.8115, "step": 4117 }, { "epoch": 0.54, "grad_norm": 6.9820733070373535, "learning_rate": 4.601463862425903e-06, "loss": 0.7721, "step": 4118 }, { "epoch": 0.54, "grad_norm": 1.3584741353988647, "learning_rate": 4.59934852611866e-06, "loss": 0.8526, "step": 4119 }, { "epoch": 0.54, "grad_norm": 0.8050498366355896, "learning_rate": 4.597233261983347e-06, "loss": 1.0057, "step": 4120 }, { "epoch": 0.54, "grad_norm": 1.3844667673110962, "learning_rate": 4.595118070401001e-06, "loss": 0.8195, "step": 4121 }, { "epoch": 0.54, "grad_norm": 1.2972798347473145, "learning_rate": 4.593002951752645e-06, "loss": 0.8077, "step": 4122 }, { "epoch": 0.54, "grad_norm": 1.3097772598266602, "learning_rate": 4.590887906419288e-06, "loss": 0.8407, "step": 4123 }, { "epoch": 0.54, "grad_norm": 1.463160514831543, "learning_rate": 4.588772934781929e-06, "loss": 0.849, "step": 4124 }, { "epoch": 0.54, "grad_norm": 1.2991145849227905, "learning_rate": 4.586658037221546e-06, "loss": 0.852, "step": 4125 }, { "epoch": 0.54, "grad_norm": 1.3877451419830322, "learning_rate": 4.584543214119117e-06, "loss": 0.8043, "step": 4126 }, { "epoch": 0.54, "grad_norm": 0.7310925722122192, "learning_rate": 4.582428465855594e-06, "loss": 1.0601, "step": 4127 }, { "epoch": 0.54, "grad_norm": 1.4778413772583008, "learning_rate": 4.5803137928119176e-06, "loss": 0.8052, "step": 4128 }, { "epoch": 0.54, "grad_norm": 1.5148558616638184, "learning_rate": 4.5781991953690245e-06, "loss": 0.8496, "step": 4129 }, { "epoch": 0.54, "grad_norm": 1.2794114351272583, "learning_rate": 4.576084673907824e-06, "loss": 0.8522, "step": 4130 }, { "epoch": 0.54, "grad_norm": 1.3258659839630127, "learning_rate": 4.573970228809224e-06, "loss": 0.8225, "step": 4131 }, { "epoch": 0.54, "grad_norm": 0.6248547434806824, "learning_rate": 4.571855860454111e-06, "loss": 1.0409, "step": 4132 }, { "epoch": 0.54, "grad_norm": 1.4847437143325806, "learning_rate": 4.569741569223356e-06, "loss": 0.824, "step": 4133 }, { "epoch": 0.54, "grad_norm": 1.3946453332901, "learning_rate": 4.567627355497827e-06, "loss": 0.8598, "step": 4134 }, { "epoch": 0.54, "grad_norm": 1.3614919185638428, "learning_rate": 4.565513219658364e-06, "loss": 0.8304, "step": 4135 }, { "epoch": 0.54, "grad_norm": 0.6920589804649353, "learning_rate": 4.563399162085806e-06, "loss": 1.0363, "step": 4136 }, { "epoch": 0.54, "grad_norm": 0.6203770041465759, "learning_rate": 4.5612851831609684e-06, "loss": 1.0271, "step": 4137 }, { "epoch": 0.54, "grad_norm": 1.2838424444198608, "learning_rate": 4.559171283264654e-06, "loss": 0.8443, "step": 4138 }, { "epoch": 0.54, "grad_norm": 1.3101203441619873, "learning_rate": 4.5570574627776584e-06, "loss": 0.7598, "step": 4139 }, { "epoch": 0.54, "grad_norm": 1.3143922090530396, "learning_rate": 4.554943722080754e-06, "loss": 0.7958, "step": 4140 }, { "epoch": 0.54, "grad_norm": 1.4300651550292969, "learning_rate": 4.5528300615547045e-06, "loss": 0.8277, "step": 4141 }, { "epoch": 0.54, "grad_norm": 0.834246814250946, "learning_rate": 4.550716481580254e-06, "loss": 1.0171, "step": 4142 }, { "epoch": 0.54, "grad_norm": 1.4855374097824097, "learning_rate": 4.548602982538139e-06, "loss": 0.6938, "step": 4143 }, { "epoch": 0.54, "grad_norm": 1.5284886360168457, "learning_rate": 4.546489564809076e-06, "loss": 0.8253, "step": 4144 }, { "epoch": 0.54, "grad_norm": 0.6430706977844238, "learning_rate": 4.544376228773766e-06, "loss": 1.0297, "step": 4145 }, { "epoch": 0.54, "grad_norm": 0.6010593771934509, "learning_rate": 4.542262974812904e-06, "loss": 1.0066, "step": 4146 }, { "epoch": 0.54, "grad_norm": 0.6066120266914368, "learning_rate": 4.54014980330716e-06, "loss": 1.0096, "step": 4147 }, { "epoch": 0.54, "grad_norm": 0.6302679777145386, "learning_rate": 4.538036714637191e-06, "loss": 1.024, "step": 4148 }, { "epoch": 0.54, "grad_norm": 1.4875017404556274, "learning_rate": 4.5359237091836465e-06, "loss": 0.7655, "step": 4149 }, { "epoch": 0.54, "grad_norm": 1.4187966585159302, "learning_rate": 4.533810787327153e-06, "loss": 0.8492, "step": 4150 }, { "epoch": 0.54, "grad_norm": 1.658725380897522, "learning_rate": 4.531697949448323e-06, "loss": 0.7932, "step": 4151 }, { "epoch": 0.54, "grad_norm": 0.6327226161956787, "learning_rate": 4.529585195927761e-06, "loss": 0.9996, "step": 4152 }, { "epoch": 0.54, "grad_norm": 1.331268310546875, "learning_rate": 4.527472527146043e-06, "loss": 0.8252, "step": 4153 }, { "epoch": 0.54, "grad_norm": 0.6149207949638367, "learning_rate": 4.5253599434837456e-06, "loss": 1.008, "step": 4154 }, { "epoch": 0.54, "grad_norm": 1.3062243461608887, "learning_rate": 4.523247445321418e-06, "loss": 0.7689, "step": 4155 }, { "epoch": 0.54, "grad_norm": 2.2429521083831787, "learning_rate": 4.521135033039597e-06, "loss": 0.8094, "step": 4156 }, { "epoch": 0.54, "grad_norm": 1.2728514671325684, "learning_rate": 4.519022707018809e-06, "loss": 0.7842, "step": 4157 }, { "epoch": 0.54, "grad_norm": 1.4241442680358887, "learning_rate": 4.516910467639556e-06, "loss": 0.7951, "step": 4158 }, { "epoch": 0.54, "grad_norm": 1.352217435836792, "learning_rate": 4.514798315282335e-06, "loss": 0.7697, "step": 4159 }, { "epoch": 0.55, "grad_norm": 1.3422425985336304, "learning_rate": 4.512686250327618e-06, "loss": 0.795, "step": 4160 }, { "epoch": 0.55, "grad_norm": 1.3782076835632324, "learning_rate": 4.510574273155864e-06, "loss": 0.7746, "step": 4161 }, { "epoch": 0.55, "grad_norm": 1.5768612623214722, "learning_rate": 4.5084623841475184e-06, "loss": 0.7791, "step": 4162 }, { "epoch": 0.55, "grad_norm": 1.268567681312561, "learning_rate": 4.50635058368301e-06, "loss": 0.817, "step": 4163 }, { "epoch": 0.55, "grad_norm": 0.7384819388389587, "learning_rate": 4.504238872142751e-06, "loss": 1.0249, "step": 4164 }, { "epoch": 0.55, "grad_norm": 1.3748087882995605, "learning_rate": 4.502127249907136e-06, "loss": 0.7838, "step": 4165 }, { "epoch": 0.55, "grad_norm": 0.6538206338882446, "learning_rate": 4.500015717356545e-06, "loss": 1.0368, "step": 4166 }, { "epoch": 0.55, "grad_norm": 0.6262401938438416, "learning_rate": 4.497904274871346e-06, "loss": 1.045, "step": 4167 }, { "epoch": 0.55, "grad_norm": 1.60027015209198, "learning_rate": 4.495792922831881e-06, "loss": 0.8139, "step": 4168 }, { "epoch": 0.55, "grad_norm": 1.3536641597747803, "learning_rate": 4.493681661618487e-06, "loss": 0.883, "step": 4169 }, { "epoch": 0.55, "grad_norm": 2.736670970916748, "learning_rate": 4.491570491611475e-06, "loss": 0.8271, "step": 4170 }, { "epoch": 0.55, "grad_norm": 0.668231189250946, "learning_rate": 4.489459413191143e-06, "loss": 1.0132, "step": 4171 }, { "epoch": 0.55, "grad_norm": 1.3291345834732056, "learning_rate": 4.487348426737779e-06, "loss": 0.8371, "step": 4172 }, { "epoch": 0.55, "grad_norm": 1.2332088947296143, "learning_rate": 4.485237532631641e-06, "loss": 0.811, "step": 4173 }, { "epoch": 0.55, "grad_norm": 1.5204814672470093, "learning_rate": 4.483126731252984e-06, "loss": 0.8528, "step": 4174 }, { "epoch": 0.55, "grad_norm": 1.3204675912857056, "learning_rate": 4.481016022982039e-06, "loss": 0.7857, "step": 4175 }, { "epoch": 0.55, "grad_norm": 1.2813506126403809, "learning_rate": 4.478905408199017e-06, "loss": 0.7676, "step": 4176 }, { "epoch": 0.55, "grad_norm": 0.7250071167945862, "learning_rate": 4.476794887284123e-06, "loss": 1.0497, "step": 4177 }, { "epoch": 0.55, "grad_norm": 0.6510939598083496, "learning_rate": 4.474684460617536e-06, "loss": 1.0086, "step": 4178 }, { "epoch": 0.55, "grad_norm": 1.335830807685852, "learning_rate": 4.4725741285794175e-06, "loss": 0.7754, "step": 4179 }, { "epoch": 0.55, "grad_norm": 1.4499088525772095, "learning_rate": 4.47046389154992e-06, "loss": 0.754, "step": 4180 }, { "epoch": 0.55, "grad_norm": 1.426069974899292, "learning_rate": 4.468353749909172e-06, "loss": 0.8067, "step": 4181 }, { "epoch": 0.55, "grad_norm": 1.3605799674987793, "learning_rate": 4.466243704037287e-06, "loss": 0.7643, "step": 4182 }, { "epoch": 0.55, "grad_norm": 1.3240011930465698, "learning_rate": 4.464133754314363e-06, "loss": 0.7998, "step": 4183 }, { "epoch": 0.55, "grad_norm": 1.533372402191162, "learning_rate": 4.462023901120476e-06, "loss": 0.7479, "step": 4184 }, { "epoch": 0.55, "grad_norm": 1.3850969076156616, "learning_rate": 4.459914144835689e-06, "loss": 0.7857, "step": 4185 }, { "epoch": 0.55, "grad_norm": 0.8041591644287109, "learning_rate": 4.457804485840044e-06, "loss": 1.0267, "step": 4186 }, { "epoch": 0.55, "grad_norm": 1.3840903043746948, "learning_rate": 4.4556949245135715e-06, "loss": 0.8254, "step": 4187 }, { "epoch": 0.55, "grad_norm": 0.680663526058197, "learning_rate": 4.453585461236277e-06, "loss": 0.9994, "step": 4188 }, { "epoch": 0.55, "grad_norm": 1.4690946340560913, "learning_rate": 4.451476096388151e-06, "loss": 0.7904, "step": 4189 }, { "epoch": 0.55, "grad_norm": 1.285264015197754, "learning_rate": 4.449366830349172e-06, "loss": 0.7733, "step": 4190 }, { "epoch": 0.55, "grad_norm": 1.3845562934875488, "learning_rate": 4.447257663499289e-06, "loss": 0.8406, "step": 4191 }, { "epoch": 0.55, "grad_norm": 1.4481498003005981, "learning_rate": 4.445148596218445e-06, "loss": 0.8436, "step": 4192 }, { "epoch": 0.55, "grad_norm": 1.3347175121307373, "learning_rate": 4.443039628886559e-06, "loss": 0.793, "step": 4193 }, { "epoch": 0.55, "grad_norm": 1.4866362810134888, "learning_rate": 4.4409307618835284e-06, "loss": 0.8376, "step": 4194 }, { "epoch": 0.55, "grad_norm": 0.9204904437065125, "learning_rate": 4.4388219955892436e-06, "loss": 1.0231, "step": 4195 }, { "epoch": 0.55, "grad_norm": 1.3493797779083252, "learning_rate": 4.4367133303835646e-06, "loss": 0.7672, "step": 4196 }, { "epoch": 0.55, "grad_norm": 1.4499503374099731, "learning_rate": 4.434604766646344e-06, "loss": 0.8059, "step": 4197 }, { "epoch": 0.55, "grad_norm": 1.333446741104126, "learning_rate": 4.432496304757408e-06, "loss": 0.835, "step": 4198 }, { "epoch": 0.55, "grad_norm": 1.417589545249939, "learning_rate": 4.430387945096566e-06, "loss": 0.8293, "step": 4199 }, { "epoch": 0.55, "grad_norm": 0.6923410296440125, "learning_rate": 4.428279688043613e-06, "loss": 1.0611, "step": 4200 }, { "epoch": 0.55, "grad_norm": 1.5787951946258545, "learning_rate": 4.426171533978321e-06, "loss": 0.8488, "step": 4201 }, { "epoch": 0.55, "grad_norm": 1.2822757959365845, "learning_rate": 4.4240634832804496e-06, "loss": 0.8448, "step": 4202 }, { "epoch": 0.55, "grad_norm": 0.622704803943634, "learning_rate": 4.42195553632973e-06, "loss": 1.0361, "step": 4203 }, { "epoch": 0.55, "grad_norm": 1.3691539764404297, "learning_rate": 4.419847693505883e-06, "loss": 0.7776, "step": 4204 }, { "epoch": 0.55, "grad_norm": 1.4938726425170898, "learning_rate": 4.417739955188608e-06, "loss": 0.7789, "step": 4205 }, { "epoch": 0.55, "grad_norm": 1.313355803489685, "learning_rate": 4.415632321757583e-06, "loss": 0.8279, "step": 4206 }, { "epoch": 0.55, "grad_norm": 0.628459632396698, "learning_rate": 4.413524793592475e-06, "loss": 1.0216, "step": 4207 }, { "epoch": 0.55, "grad_norm": 1.304477334022522, "learning_rate": 4.411417371072922e-06, "loss": 0.8181, "step": 4208 }, { "epoch": 0.55, "grad_norm": 0.6046699285507202, "learning_rate": 4.409310054578547e-06, "loss": 1.0267, "step": 4209 }, { "epoch": 0.55, "grad_norm": 1.6021637916564941, "learning_rate": 4.407202844488959e-06, "loss": 0.7497, "step": 4210 }, { "epoch": 0.55, "grad_norm": 1.3039582967758179, "learning_rate": 4.40509574118374e-06, "loss": 0.7548, "step": 4211 }, { "epoch": 0.55, "grad_norm": 1.3424180746078491, "learning_rate": 4.402988745042454e-06, "loss": 0.8033, "step": 4212 }, { "epoch": 0.55, "grad_norm": 1.341144323348999, "learning_rate": 4.400881856444654e-06, "loss": 0.8096, "step": 4213 }, { "epoch": 0.55, "grad_norm": 1.5476642847061157, "learning_rate": 4.398775075769862e-06, "loss": 0.8093, "step": 4214 }, { "epoch": 0.55, "grad_norm": 1.3334033489227295, "learning_rate": 4.396668403397589e-06, "loss": 0.8525, "step": 4215 }, { "epoch": 0.55, "grad_norm": 0.6208717823028564, "learning_rate": 4.394561839707323e-06, "loss": 1.0245, "step": 4216 }, { "epoch": 0.55, "grad_norm": 1.7750823497772217, "learning_rate": 4.392455385078529e-06, "loss": 0.8104, "step": 4217 }, { "epoch": 0.55, "grad_norm": 1.374553918838501, "learning_rate": 4.3903490398906634e-06, "loss": 0.7736, "step": 4218 }, { "epoch": 0.55, "grad_norm": 1.3968383073806763, "learning_rate": 4.388242804523149e-06, "loss": 0.7787, "step": 4219 }, { "epoch": 0.55, "grad_norm": 0.6151182651519775, "learning_rate": 4.3861366793554e-06, "loss": 1.0154, "step": 4220 }, { "epoch": 0.55, "grad_norm": 1.4162403345108032, "learning_rate": 4.3840306647668055e-06, "loss": 0.8217, "step": 4221 }, { "epoch": 0.55, "grad_norm": 1.3100520372390747, "learning_rate": 4.381924761136733e-06, "loss": 0.7861, "step": 4222 }, { "epoch": 0.55, "grad_norm": 0.6150359511375427, "learning_rate": 4.379818968844534e-06, "loss": 1.0174, "step": 4223 }, { "epoch": 0.55, "grad_norm": 1.3832969665527344, "learning_rate": 4.377713288269539e-06, "loss": 0.8681, "step": 4224 }, { "epoch": 0.55, "grad_norm": 1.3602310419082642, "learning_rate": 4.375607719791058e-06, "loss": 0.8327, "step": 4225 }, { "epoch": 0.55, "grad_norm": 1.3407196998596191, "learning_rate": 4.37350226378838e-06, "loss": 0.8396, "step": 4226 }, { "epoch": 0.55, "grad_norm": 1.409629464149475, "learning_rate": 4.371396920640773e-06, "loss": 0.8171, "step": 4227 }, { "epoch": 0.55, "grad_norm": 1.2517467737197876, "learning_rate": 4.369291690727489e-06, "loss": 0.7722, "step": 4228 }, { "epoch": 0.55, "grad_norm": 1.3646395206451416, "learning_rate": 4.367186574427752e-06, "loss": 0.8383, "step": 4229 }, { "epoch": 0.55, "grad_norm": 1.304740071296692, "learning_rate": 4.365081572120776e-06, "loss": 0.8096, "step": 4230 }, { "epoch": 0.55, "grad_norm": 1.3472042083740234, "learning_rate": 4.362976684185746e-06, "loss": 0.7478, "step": 4231 }, { "epoch": 0.55, "grad_norm": 1.443901777267456, "learning_rate": 4.360871911001826e-06, "loss": 0.8079, "step": 4232 }, { "epoch": 0.55, "grad_norm": 1.3880940675735474, "learning_rate": 4.358767252948167e-06, "loss": 0.773, "step": 4233 }, { "epoch": 0.55, "grad_norm": 0.6520529389381409, "learning_rate": 4.356662710403891e-06, "loss": 1.0022, "step": 4234 }, { "epoch": 0.55, "grad_norm": 1.429814100265503, "learning_rate": 4.354558283748107e-06, "loss": 0.8293, "step": 4235 }, { "epoch": 0.56, "grad_norm": 1.3764238357543945, "learning_rate": 4.352453973359897e-06, "loss": 0.8457, "step": 4236 }, { "epoch": 0.56, "grad_norm": 1.391581654548645, "learning_rate": 4.3503497796183204e-06, "loss": 0.8357, "step": 4237 }, { "epoch": 0.56, "grad_norm": 1.3260918855667114, "learning_rate": 4.348245702902425e-06, "loss": 0.7902, "step": 4238 }, { "epoch": 0.56, "grad_norm": 1.311713457107544, "learning_rate": 4.346141743591228e-06, "loss": 0.8214, "step": 4239 }, { "epoch": 0.56, "grad_norm": 0.6426883339881897, "learning_rate": 4.344037902063728e-06, "loss": 1.0457, "step": 4240 }, { "epoch": 0.56, "grad_norm": 1.2935212850570679, "learning_rate": 4.341934178698908e-06, "loss": 0.7976, "step": 4241 }, { "epoch": 0.56, "grad_norm": 1.6754037141799927, "learning_rate": 4.339830573875722e-06, "loss": 0.8254, "step": 4242 }, { "epoch": 0.56, "grad_norm": 1.3708540201187134, "learning_rate": 4.337727087973107e-06, "loss": 0.8369, "step": 4243 }, { "epoch": 0.56, "grad_norm": 0.6375563144683838, "learning_rate": 4.335623721369977e-06, "loss": 1.0352, "step": 4244 }, { "epoch": 0.56, "grad_norm": 1.3196309804916382, "learning_rate": 4.333520474445225e-06, "loss": 0.8628, "step": 4245 }, { "epoch": 0.56, "grad_norm": 0.5943576097488403, "learning_rate": 4.331417347577723e-06, "loss": 1.0263, "step": 4246 }, { "epoch": 0.56, "grad_norm": 1.6916296482086182, "learning_rate": 4.329314341146319e-06, "loss": 0.7726, "step": 4247 }, { "epoch": 0.56, "grad_norm": 0.612793505191803, "learning_rate": 4.327211455529844e-06, "loss": 1.0092, "step": 4248 }, { "epoch": 0.56, "grad_norm": 0.5743474960327148, "learning_rate": 4.325108691107103e-06, "loss": 1.0051, "step": 4249 }, { "epoch": 0.56, "grad_norm": 1.337005853652954, "learning_rate": 4.323006048256878e-06, "loss": 0.8623, "step": 4250 }, { "epoch": 0.56, "grad_norm": 0.6004089117050171, "learning_rate": 4.320903527357936e-06, "loss": 1.0487, "step": 4251 }, { "epoch": 0.56, "grad_norm": 1.435333251953125, "learning_rate": 4.318801128789013e-06, "loss": 0.8254, "step": 4252 }, { "epoch": 0.56, "grad_norm": 0.5930620431900024, "learning_rate": 4.316698852928832e-06, "loss": 1.0302, "step": 4253 }, { "epoch": 0.56, "grad_norm": 1.3702768087387085, "learning_rate": 4.314596700156088e-06, "loss": 0.7828, "step": 4254 }, { "epoch": 0.56, "grad_norm": 1.3212385177612305, "learning_rate": 4.312494670849453e-06, "loss": 0.7994, "step": 4255 }, { "epoch": 0.56, "grad_norm": 1.3044434785842896, "learning_rate": 4.310392765387583e-06, "loss": 0.7611, "step": 4256 }, { "epoch": 0.56, "grad_norm": 1.4402908086776733, "learning_rate": 4.3082909841491034e-06, "loss": 0.8073, "step": 4257 }, { "epoch": 0.56, "grad_norm": 1.3803527355194092, "learning_rate": 4.306189327512626e-06, "loss": 0.8291, "step": 4258 }, { "epoch": 0.56, "grad_norm": 1.3394255638122559, "learning_rate": 4.304087795856735e-06, "loss": 0.8199, "step": 4259 }, { "epoch": 0.56, "grad_norm": 0.6692004203796387, "learning_rate": 4.301986389559989e-06, "loss": 1.0119, "step": 4260 }, { "epoch": 0.56, "grad_norm": 0.6321772933006287, "learning_rate": 4.299885109000931e-06, "loss": 1.0357, "step": 4261 }, { "epoch": 0.56, "grad_norm": 0.638935923576355, "learning_rate": 4.297783954558077e-06, "loss": 0.9822, "step": 4262 }, { "epoch": 0.56, "grad_norm": 1.4182791709899902, "learning_rate": 4.295682926609923e-06, "loss": 0.8454, "step": 4263 }, { "epoch": 0.56, "grad_norm": 1.4357877969741821, "learning_rate": 4.29358202553494e-06, "loss": 0.7895, "step": 4264 }, { "epoch": 0.56, "grad_norm": 0.6374244093894958, "learning_rate": 4.291481251711575e-06, "loss": 1.0024, "step": 4265 }, { "epoch": 0.56, "grad_norm": 1.361905574798584, "learning_rate": 4.289380605518258e-06, "loss": 0.8089, "step": 4266 }, { "epoch": 0.56, "grad_norm": 0.6439118981361389, "learning_rate": 4.287280087333388e-06, "loss": 1.0109, "step": 4267 }, { "epoch": 0.56, "grad_norm": 0.6283375024795532, "learning_rate": 4.285179697535345e-06, "loss": 1.017, "step": 4268 }, { "epoch": 0.56, "grad_norm": 1.305338978767395, "learning_rate": 4.283079436502489e-06, "loss": 0.8293, "step": 4269 }, { "epoch": 0.56, "grad_norm": 1.3896353244781494, "learning_rate": 4.280979304613149e-06, "loss": 0.8339, "step": 4270 }, { "epoch": 0.56, "grad_norm": 1.4250093698501587, "learning_rate": 4.278879302245641e-06, "loss": 0.827, "step": 4271 }, { "epoch": 0.56, "grad_norm": 1.3145192861557007, "learning_rate": 4.276779429778248e-06, "loss": 0.791, "step": 4272 }, { "epoch": 0.56, "grad_norm": 1.6219812631607056, "learning_rate": 4.274679687589232e-06, "loss": 0.8436, "step": 4273 }, { "epoch": 0.56, "grad_norm": 0.6495571732521057, "learning_rate": 4.272580076056838e-06, "loss": 1.0135, "step": 4274 }, { "epoch": 0.56, "grad_norm": 1.3222122192382812, "learning_rate": 4.270480595559278e-06, "loss": 0.8416, "step": 4275 }, { "epoch": 0.56, "grad_norm": 1.2665660381317139, "learning_rate": 4.268381246474749e-06, "loss": 0.8693, "step": 4276 }, { "epoch": 0.56, "grad_norm": 0.6226847171783447, "learning_rate": 4.266282029181419e-06, "loss": 1.0091, "step": 4277 }, { "epoch": 0.56, "grad_norm": 1.3569785356521606, "learning_rate": 4.264182944057429e-06, "loss": 0.8036, "step": 4278 }, { "epoch": 0.56, "grad_norm": 1.2652463912963867, "learning_rate": 4.262083991480908e-06, "loss": 0.7437, "step": 4279 }, { "epoch": 0.56, "grad_norm": 0.6183117628097534, "learning_rate": 4.259985171829949e-06, "loss": 1.0061, "step": 4280 }, { "epoch": 0.56, "grad_norm": 0.5995633006095886, "learning_rate": 4.257886485482627e-06, "loss": 0.9937, "step": 4281 }, { "epoch": 0.56, "grad_norm": 1.392499327659607, "learning_rate": 4.255787932816993e-06, "loss": 0.8332, "step": 4282 }, { "epoch": 0.56, "grad_norm": 0.608741819858551, "learning_rate": 4.253689514211071e-06, "loss": 1.0074, "step": 4283 }, { "epoch": 0.56, "grad_norm": 1.3082225322723389, "learning_rate": 4.251591230042865e-06, "loss": 0.817, "step": 4284 }, { "epoch": 0.56, "grad_norm": 1.3420381546020508, "learning_rate": 4.249493080690348e-06, "loss": 0.8117, "step": 4285 }, { "epoch": 0.56, "grad_norm": 1.4039044380187988, "learning_rate": 4.247395066531479e-06, "loss": 0.8213, "step": 4286 }, { "epoch": 0.56, "grad_norm": 1.4192852973937988, "learning_rate": 4.245297187944184e-06, "loss": 0.7654, "step": 4287 }, { "epoch": 0.56, "grad_norm": 1.4084302186965942, "learning_rate": 4.243199445306365e-06, "loss": 0.7964, "step": 4288 }, { "epoch": 0.56, "grad_norm": 0.6754562258720398, "learning_rate": 4.241101838995907e-06, "loss": 1.0099, "step": 4289 }, { "epoch": 0.56, "grad_norm": 1.6170326471328735, "learning_rate": 4.239004369390659e-06, "loss": 0.8472, "step": 4290 }, { "epoch": 0.56, "grad_norm": 1.3385486602783203, "learning_rate": 4.236907036868458e-06, "loss": 0.8148, "step": 4291 }, { "epoch": 0.56, "grad_norm": 1.2364990711212158, "learning_rate": 4.234809841807108e-06, "loss": 0.7869, "step": 4292 }, { "epoch": 0.56, "grad_norm": 0.6258172392845154, "learning_rate": 4.2327127845843874e-06, "loss": 1.0285, "step": 4293 }, { "epoch": 0.56, "grad_norm": 1.3202751874923706, "learning_rate": 4.230615865578057e-06, "loss": 0.8292, "step": 4294 }, { "epoch": 0.56, "grad_norm": 1.4176750183105469, "learning_rate": 4.228519085165846e-06, "loss": 0.8483, "step": 4295 }, { "epoch": 0.56, "grad_norm": 1.3495759963989258, "learning_rate": 4.226422443725458e-06, "loss": 0.8124, "step": 4296 }, { "epoch": 0.56, "grad_norm": 1.4443000555038452, "learning_rate": 4.2243259416345805e-06, "loss": 0.8218, "step": 4297 }, { "epoch": 0.56, "grad_norm": 0.6992434859275818, "learning_rate": 4.222229579270865e-06, "loss": 1.0017, "step": 4298 }, { "epoch": 0.56, "grad_norm": 1.3618414402008057, "learning_rate": 4.220133357011946e-06, "loss": 0.8072, "step": 4299 }, { "epoch": 0.56, "grad_norm": 0.6252961754798889, "learning_rate": 4.218037275235429e-06, "loss": 1.0059, "step": 4300 }, { "epoch": 0.56, "grad_norm": 1.4308202266693115, "learning_rate": 4.2159413343188935e-06, "loss": 0.8044, "step": 4301 }, { "epoch": 0.56, "grad_norm": 1.3149831295013428, "learning_rate": 4.213845534639895e-06, "loss": 0.7908, "step": 4302 }, { "epoch": 0.56, "grad_norm": 0.7074105143547058, "learning_rate": 4.211749876575964e-06, "loss": 1.0247, "step": 4303 }, { "epoch": 0.56, "grad_norm": 0.7240758538246155, "learning_rate": 4.209654360504605e-06, "loss": 1.0342, "step": 4304 }, { "epoch": 0.56, "grad_norm": 0.6258951425552368, "learning_rate": 4.207558986803295e-06, "loss": 1.0375, "step": 4305 }, { "epoch": 0.56, "grad_norm": 1.252048373222351, "learning_rate": 4.20546375584949e-06, "loss": 0.7929, "step": 4306 }, { "epoch": 0.56, "grad_norm": 0.6212829947471619, "learning_rate": 4.203368668020616e-06, "loss": 1.007, "step": 4307 }, { "epoch": 0.56, "grad_norm": 1.3119285106658936, "learning_rate": 4.201273723694073e-06, "loss": 0.8742, "step": 4308 }, { "epoch": 0.56, "grad_norm": 0.7352123260498047, "learning_rate": 4.199178923247241e-06, "loss": 1.0196, "step": 4309 }, { "epoch": 0.56, "grad_norm": 0.702736496925354, "learning_rate": 4.197084267057466e-06, "loss": 1.0248, "step": 4310 }, { "epoch": 0.56, "grad_norm": 0.6146852970123291, "learning_rate": 4.194989755502073e-06, "loss": 1.0085, "step": 4311 }, { "epoch": 0.57, "grad_norm": 0.5995909571647644, "learning_rate": 4.192895388958361e-06, "loss": 1.0197, "step": 4312 }, { "epoch": 0.57, "grad_norm": 1.43799889087677, "learning_rate": 4.1908011678036e-06, "loss": 0.8435, "step": 4313 }, { "epoch": 0.57, "grad_norm": 1.6933066844940186, "learning_rate": 4.188707092415038e-06, "loss": 0.8541, "step": 4314 }, { "epoch": 0.57, "grad_norm": 1.2657175064086914, "learning_rate": 4.186613163169892e-06, "loss": 0.7828, "step": 4315 }, { "epoch": 0.57, "grad_norm": 1.3512996435165405, "learning_rate": 4.184519380445355e-06, "loss": 0.8155, "step": 4316 }, { "epoch": 0.57, "grad_norm": 0.706534743309021, "learning_rate": 4.182425744618595e-06, "loss": 1.0117, "step": 4317 }, { "epoch": 0.57, "grad_norm": 0.6801413893699646, "learning_rate": 4.18033225606675e-06, "loss": 1.0218, "step": 4318 }, { "epoch": 0.57, "grad_norm": 1.2989065647125244, "learning_rate": 4.178238915166937e-06, "loss": 0.7935, "step": 4319 }, { "epoch": 0.57, "grad_norm": 1.3260531425476074, "learning_rate": 4.176145722296241e-06, "loss": 0.8403, "step": 4320 }, { "epoch": 0.57, "grad_norm": 0.6071808934211731, "learning_rate": 4.174052677831722e-06, "loss": 1.0286, "step": 4321 }, { "epoch": 0.57, "grad_norm": 1.5386906862258911, "learning_rate": 4.171959782150412e-06, "loss": 0.839, "step": 4322 }, { "epoch": 0.57, "grad_norm": 1.2889105081558228, "learning_rate": 4.1698670356293205e-06, "loss": 0.7946, "step": 4323 }, { "epoch": 0.57, "grad_norm": 3.81603741645813, "learning_rate": 4.167774438645428e-06, "loss": 0.769, "step": 4324 }, { "epoch": 0.57, "grad_norm": 1.3448208570480347, "learning_rate": 4.165681991575686e-06, "loss": 0.8056, "step": 4325 }, { "epoch": 0.57, "grad_norm": 1.2788931131362915, "learning_rate": 4.163589694797018e-06, "loss": 0.8213, "step": 4326 }, { "epoch": 0.57, "grad_norm": 1.3174769878387451, "learning_rate": 4.161497548686327e-06, "loss": 0.8217, "step": 4327 }, { "epoch": 0.57, "grad_norm": 1.3659909963607788, "learning_rate": 4.1594055536204835e-06, "loss": 0.8285, "step": 4328 }, { "epoch": 0.57, "grad_norm": 0.682087242603302, "learning_rate": 4.15731370997633e-06, "loss": 1.0309, "step": 4329 }, { "epoch": 0.57, "grad_norm": 1.347281575202942, "learning_rate": 4.1552220181306875e-06, "loss": 0.771, "step": 4330 }, { "epoch": 0.57, "grad_norm": 1.4668480157852173, "learning_rate": 4.153130478460341e-06, "loss": 0.7847, "step": 4331 }, { "epoch": 0.57, "grad_norm": 1.4707832336425781, "learning_rate": 4.151039091342058e-06, "loss": 0.84, "step": 4332 }, { "epoch": 0.57, "grad_norm": 1.2560495138168335, "learning_rate": 4.148947857152572e-06, "loss": 0.8009, "step": 4333 }, { "epoch": 0.57, "grad_norm": 1.5764814615249634, "learning_rate": 4.146856776268588e-06, "loss": 0.8437, "step": 4334 }, { "epoch": 0.57, "grad_norm": 1.3249306678771973, "learning_rate": 4.14476584906679e-06, "loss": 0.7877, "step": 4335 }, { "epoch": 0.57, "grad_norm": 1.4472252130508423, "learning_rate": 4.142675075923825e-06, "loss": 0.7975, "step": 4336 }, { "epoch": 0.57, "grad_norm": 0.6482588648796082, "learning_rate": 4.1405844572163225e-06, "loss": 1.036, "step": 4337 }, { "epoch": 0.57, "grad_norm": 1.2961256504058838, "learning_rate": 4.138493993320877e-06, "loss": 0.7877, "step": 4338 }, { "epoch": 0.57, "grad_norm": 1.343406319618225, "learning_rate": 4.136403684614055e-06, "loss": 0.7953, "step": 4339 }, { "epoch": 0.57, "grad_norm": 1.6249215602874756, "learning_rate": 4.1343135314724026e-06, "loss": 0.7932, "step": 4340 }, { "epoch": 0.57, "grad_norm": 1.7494763135910034, "learning_rate": 4.132223534272427e-06, "loss": 0.8446, "step": 4341 }, { "epoch": 0.57, "grad_norm": 0.5875198841094971, "learning_rate": 4.130133693390616e-06, "loss": 1.0061, "step": 4342 }, { "epoch": 0.57, "grad_norm": 1.2670328617095947, "learning_rate": 4.128044009203427e-06, "loss": 0.8812, "step": 4343 }, { "epoch": 0.57, "grad_norm": 1.34520423412323, "learning_rate": 4.125954482087284e-06, "loss": 0.8215, "step": 4344 }, { "epoch": 0.57, "grad_norm": 1.360505223274231, "learning_rate": 4.123865112418593e-06, "loss": 0.8705, "step": 4345 }, { "epoch": 0.57, "grad_norm": 1.271131157875061, "learning_rate": 4.121775900573718e-06, "loss": 0.7756, "step": 4346 }, { "epoch": 0.57, "grad_norm": 1.3460065126419067, "learning_rate": 4.119686846929009e-06, "loss": 0.8073, "step": 4347 }, { "epoch": 0.57, "grad_norm": 0.6122519969940186, "learning_rate": 4.117597951860779e-06, "loss": 1.0329, "step": 4348 }, { "epoch": 0.57, "grad_norm": 1.346786618232727, "learning_rate": 4.11550921574531e-06, "loss": 0.9143, "step": 4349 }, { "epoch": 0.57, "grad_norm": 0.6104592084884644, "learning_rate": 4.113420638958866e-06, "loss": 1.0138, "step": 4350 }, { "epoch": 0.57, "grad_norm": 1.4033366441726685, "learning_rate": 4.111332221877669e-06, "loss": 0.8613, "step": 4351 }, { "epoch": 0.57, "grad_norm": 1.5774457454681396, "learning_rate": 4.109243964877926e-06, "loss": 0.8566, "step": 4352 }, { "epoch": 0.57, "grad_norm": 1.3513001203536987, "learning_rate": 4.1071558683358035e-06, "loss": 0.8651, "step": 4353 }, { "epoch": 0.57, "grad_norm": 0.6082331538200378, "learning_rate": 4.105067932627443e-06, "loss": 1.033, "step": 4354 }, { "epoch": 0.57, "grad_norm": 1.325066328048706, "learning_rate": 4.102980158128963e-06, "loss": 0.798, "step": 4355 }, { "epoch": 0.57, "grad_norm": 1.3354668617248535, "learning_rate": 4.1008925452164435e-06, "loss": 0.7518, "step": 4356 }, { "epoch": 0.57, "grad_norm": 1.271214485168457, "learning_rate": 4.098805094265939e-06, "loss": 0.8204, "step": 4357 }, { "epoch": 0.57, "grad_norm": 1.4414782524108887, "learning_rate": 4.09671780565348e-06, "loss": 0.8707, "step": 4358 }, { "epoch": 0.57, "grad_norm": 1.4223568439483643, "learning_rate": 4.094630679755058e-06, "loss": 0.7347, "step": 4359 }, { "epoch": 0.57, "grad_norm": 1.3076305389404297, "learning_rate": 4.092543716946644e-06, "loss": 0.8008, "step": 4360 }, { "epoch": 0.57, "grad_norm": 1.4482965469360352, "learning_rate": 4.090456917604176e-06, "loss": 0.8052, "step": 4361 }, { "epoch": 0.57, "grad_norm": 1.4592161178588867, "learning_rate": 4.08837028210356e-06, "loss": 0.8262, "step": 4362 }, { "epoch": 0.57, "grad_norm": 1.2225009202957153, "learning_rate": 4.086283810820677e-06, "loss": 0.8145, "step": 4363 }, { "epoch": 0.57, "grad_norm": 0.641649603843689, "learning_rate": 4.0841975041313766e-06, "loss": 1.0173, "step": 4364 }, { "epoch": 0.57, "grad_norm": 0.6311721801757812, "learning_rate": 4.082111362411478e-06, "loss": 1.0171, "step": 4365 }, { "epoch": 0.57, "grad_norm": 1.332265019416809, "learning_rate": 4.0800253860367725e-06, "loss": 0.8044, "step": 4366 }, { "epoch": 0.57, "grad_norm": 1.4203451871871948, "learning_rate": 4.077939575383017e-06, "loss": 0.8651, "step": 4367 }, { "epoch": 0.57, "grad_norm": 1.2936303615570068, "learning_rate": 4.075853930825947e-06, "loss": 0.8058, "step": 4368 }, { "epoch": 0.57, "grad_norm": 1.3616831302642822, "learning_rate": 4.073768452741258e-06, "loss": 0.8106, "step": 4369 }, { "epoch": 0.57, "grad_norm": 1.5567481517791748, "learning_rate": 4.071683141504625e-06, "loss": 0.8024, "step": 4370 }, { "epoch": 0.57, "grad_norm": 1.349467158317566, "learning_rate": 4.069597997491688e-06, "loss": 0.8044, "step": 4371 }, { "epoch": 0.57, "grad_norm": 1.3184027671813965, "learning_rate": 4.0675130210780525e-06, "loss": 0.8386, "step": 4372 }, { "epoch": 0.57, "grad_norm": 1.3433098793029785, "learning_rate": 4.065428212639306e-06, "loss": 0.8188, "step": 4373 }, { "epoch": 0.57, "grad_norm": 1.397481083869934, "learning_rate": 4.063343572550992e-06, "loss": 0.7941, "step": 4374 }, { "epoch": 0.57, "grad_norm": 1.3679115772247314, "learning_rate": 4.061259101188635e-06, "loss": 0.8519, "step": 4375 }, { "epoch": 0.57, "grad_norm": 3.977595329284668, "learning_rate": 4.0591747989277225e-06, "loss": 0.7876, "step": 4376 }, { "epoch": 0.57, "grad_norm": 1.6541920900344849, "learning_rate": 4.057090666143711e-06, "loss": 0.787, "step": 4377 }, { "epoch": 0.57, "grad_norm": 0.6595613360404968, "learning_rate": 4.055006703212033e-06, "loss": 1.0096, "step": 4378 }, { "epoch": 0.57, "grad_norm": 1.4696171283721924, "learning_rate": 4.052922910508083e-06, "loss": 0.7447, "step": 4379 }, { "epoch": 0.57, "grad_norm": 1.3726862668991089, "learning_rate": 4.0508392884072285e-06, "loss": 0.8573, "step": 4380 }, { "epoch": 0.57, "grad_norm": 1.5516504049301147, "learning_rate": 4.048755837284807e-06, "loss": 0.8026, "step": 4381 }, { "epoch": 0.57, "grad_norm": 1.3300420045852661, "learning_rate": 4.0466725575161235e-06, "loss": 0.8323, "step": 4382 }, { "epoch": 0.57, "grad_norm": 0.6063814759254456, "learning_rate": 4.044589449476451e-06, "loss": 1.0258, "step": 4383 }, { "epoch": 0.57, "grad_norm": 1.3703210353851318, "learning_rate": 4.0425065135410364e-06, "loss": 0.8429, "step": 4384 }, { "epoch": 0.57, "grad_norm": 1.4473999738693237, "learning_rate": 4.040423750085089e-06, "loss": 0.8268, "step": 4385 }, { "epoch": 0.57, "grad_norm": 0.6152052879333496, "learning_rate": 4.038341159483793e-06, "loss": 1.0244, "step": 4386 }, { "epoch": 0.57, "grad_norm": 1.3105634450912476, "learning_rate": 4.036258742112296e-06, "loss": 0.7893, "step": 4387 }, { "epoch": 0.57, "grad_norm": 0.6074528098106384, "learning_rate": 4.03417649834572e-06, "loss": 1.0326, "step": 4388 }, { "epoch": 0.58, "grad_norm": 0.5896666049957275, "learning_rate": 4.032094428559153e-06, "loss": 1.0155, "step": 4389 }, { "epoch": 0.58, "grad_norm": 1.3433622121810913, "learning_rate": 4.030012533127649e-06, "loss": 0.7962, "step": 4390 }, { "epoch": 0.58, "grad_norm": 1.5763168334960938, "learning_rate": 4.027930812426236e-06, "loss": 0.8578, "step": 4391 }, { "epoch": 0.58, "grad_norm": 1.3486087322235107, "learning_rate": 4.0258492668299045e-06, "loss": 0.8556, "step": 4392 }, { "epoch": 0.58, "grad_norm": 0.5917470455169678, "learning_rate": 4.023767896713621e-06, "loss": 1.013, "step": 4393 }, { "epoch": 0.58, "grad_norm": 1.328068494796753, "learning_rate": 4.021686702452313e-06, "loss": 0.85, "step": 4394 }, { "epoch": 0.58, "grad_norm": 1.3487101793289185, "learning_rate": 4.01960568442088e-06, "loss": 0.8561, "step": 4395 }, { "epoch": 0.58, "grad_norm": 1.3159582614898682, "learning_rate": 4.017524842994191e-06, "loss": 0.812, "step": 4396 }, { "epoch": 0.58, "grad_norm": 0.6232182383537292, "learning_rate": 4.015444178547076e-06, "loss": 1.0072, "step": 4397 }, { "epoch": 0.58, "grad_norm": 1.4825984239578247, "learning_rate": 4.013363691454347e-06, "loss": 0.7908, "step": 4398 }, { "epoch": 0.58, "grad_norm": 1.4330769777297974, "learning_rate": 4.01128338209077e-06, "loss": 0.7659, "step": 4399 }, { "epoch": 0.58, "grad_norm": 1.317531943321228, "learning_rate": 4.009203250831083e-06, "loss": 0.823, "step": 4400 }, { "epoch": 0.58, "grad_norm": 1.5624619722366333, "learning_rate": 4.007123298049998e-06, "loss": 0.7636, "step": 4401 }, { "epoch": 0.58, "grad_norm": 1.356276512145996, "learning_rate": 4.0050435241221876e-06, "loss": 0.8339, "step": 4402 }, { "epoch": 0.58, "grad_norm": 0.6121082901954651, "learning_rate": 4.002963929422296e-06, "loss": 0.9935, "step": 4403 }, { "epoch": 0.58, "grad_norm": 1.3596795797348022, "learning_rate": 4.0008845143249325e-06, "loss": 0.8368, "step": 4404 }, { "epoch": 0.58, "grad_norm": 1.3271245956420898, "learning_rate": 3.998805279204676e-06, "loss": 0.8149, "step": 4405 }, { "epoch": 0.58, "grad_norm": 0.5875277519226074, "learning_rate": 3.996726224436075e-06, "loss": 1.0133, "step": 4406 }, { "epoch": 0.58, "grad_norm": 1.4769247770309448, "learning_rate": 3.994647350393638e-06, "loss": 0.7971, "step": 4407 }, { "epoch": 0.58, "grad_norm": 1.388622760772705, "learning_rate": 3.992568657451852e-06, "loss": 0.8232, "step": 4408 }, { "epoch": 0.58, "grad_norm": 1.288396954536438, "learning_rate": 3.990490145985162e-06, "loss": 0.7539, "step": 4409 }, { "epoch": 0.58, "grad_norm": 2.0285587310791016, "learning_rate": 3.988411816367981e-06, "loss": 0.797, "step": 4410 }, { "epoch": 0.58, "grad_norm": 0.6061114072799683, "learning_rate": 3.986333668974698e-06, "loss": 1.0359, "step": 4411 }, { "epoch": 0.58, "grad_norm": 1.3517541885375977, "learning_rate": 3.9842557041796595e-06, "loss": 0.8061, "step": 4412 }, { "epoch": 0.58, "grad_norm": 1.2550302743911743, "learning_rate": 3.982177922357181e-06, "loss": 0.7915, "step": 4413 }, { "epoch": 0.58, "grad_norm": 0.6039450168609619, "learning_rate": 3.980100323881551e-06, "loss": 1.0229, "step": 4414 }, { "epoch": 0.58, "grad_norm": 1.3580433130264282, "learning_rate": 3.978022909127017e-06, "loss": 0.7916, "step": 4415 }, { "epoch": 0.58, "grad_norm": 1.3228150606155396, "learning_rate": 3.9759456784678e-06, "loss": 0.8212, "step": 4416 }, { "epoch": 0.58, "grad_norm": 1.3218952417373657, "learning_rate": 3.973868632278084e-06, "loss": 0.8026, "step": 4417 }, { "epoch": 0.58, "grad_norm": 0.5795385837554932, "learning_rate": 3.971791770932018e-06, "loss": 1.0268, "step": 4418 }, { "epoch": 0.58, "grad_norm": 1.315985918045044, "learning_rate": 3.969715094803724e-06, "loss": 0.8039, "step": 4419 }, { "epoch": 0.58, "grad_norm": 0.5887387990951538, "learning_rate": 3.967638604267286e-06, "loss": 1.0082, "step": 4420 }, { "epoch": 0.58, "grad_norm": 0.5944545269012451, "learning_rate": 3.965562299696754e-06, "loss": 1.0224, "step": 4421 }, { "epoch": 0.58, "grad_norm": 1.580729365348816, "learning_rate": 3.963486181466151e-06, "loss": 0.7742, "step": 4422 }, { "epoch": 0.58, "grad_norm": 1.3557617664337158, "learning_rate": 3.961410249949455e-06, "loss": 0.7985, "step": 4423 }, { "epoch": 0.58, "grad_norm": 1.3003971576690674, "learning_rate": 3.95933450552062e-06, "loss": 0.7507, "step": 4424 }, { "epoch": 0.58, "grad_norm": 1.3263554573059082, "learning_rate": 3.957258948553563e-06, "loss": 0.8525, "step": 4425 }, { "epoch": 0.58, "grad_norm": 0.5893410444259644, "learning_rate": 3.95518357942217e-06, "loss": 1.0333, "step": 4426 }, { "epoch": 0.58, "grad_norm": 1.2497491836547852, "learning_rate": 3.953108398500287e-06, "loss": 0.7618, "step": 4427 }, { "epoch": 0.58, "grad_norm": 1.3615446090698242, "learning_rate": 3.951033406161729e-06, "loss": 0.8368, "step": 4428 }, { "epoch": 0.58, "grad_norm": 1.3425074815750122, "learning_rate": 3.948958602780283e-06, "loss": 0.8128, "step": 4429 }, { "epoch": 0.58, "grad_norm": 1.3057061433792114, "learning_rate": 3.9468839887296895e-06, "loss": 0.7767, "step": 4430 }, { "epoch": 0.58, "grad_norm": 0.6339139342308044, "learning_rate": 3.944809564383669e-06, "loss": 1.0278, "step": 4431 }, { "epoch": 0.58, "grad_norm": 1.3796815872192383, "learning_rate": 3.942735330115897e-06, "loss": 0.9016, "step": 4432 }, { "epoch": 0.58, "grad_norm": 1.3486393690109253, "learning_rate": 3.940661286300018e-06, "loss": 0.7758, "step": 4433 }, { "epoch": 0.58, "grad_norm": 1.3545244932174683, "learning_rate": 3.938587433309646e-06, "loss": 0.8211, "step": 4434 }, { "epoch": 0.58, "grad_norm": 1.2505152225494385, "learning_rate": 3.936513771518354e-06, "loss": 0.7796, "step": 4435 }, { "epoch": 0.58, "grad_norm": 1.3937406539916992, "learning_rate": 3.934440301299689e-06, "loss": 0.8268, "step": 4436 }, { "epoch": 0.58, "grad_norm": 1.2906252145767212, "learning_rate": 3.932367023027155e-06, "loss": 0.8048, "step": 4437 }, { "epoch": 0.58, "grad_norm": 1.4086486101150513, "learning_rate": 3.930293937074224e-06, "loss": 0.8245, "step": 4438 }, { "epoch": 0.58, "grad_norm": 1.3579462766647339, "learning_rate": 3.928221043814336e-06, "loss": 0.7898, "step": 4439 }, { "epoch": 0.58, "grad_norm": 0.6125370264053345, "learning_rate": 3.926148343620896e-06, "loss": 1.0226, "step": 4440 }, { "epoch": 0.58, "grad_norm": 1.3530592918395996, "learning_rate": 3.924075836867268e-06, "loss": 0.8037, "step": 4441 }, { "epoch": 0.58, "grad_norm": 0.6086733341217041, "learning_rate": 3.922003523926791e-06, "loss": 1.008, "step": 4442 }, { "epoch": 0.58, "grad_norm": 1.3099123239517212, "learning_rate": 3.919931405172763e-06, "loss": 0.7971, "step": 4443 }, { "epoch": 0.58, "grad_norm": 1.4142457246780396, "learning_rate": 3.917859480978447e-06, "loss": 0.7855, "step": 4444 }, { "epoch": 0.58, "grad_norm": 1.3510361909866333, "learning_rate": 3.915787751717073e-06, "loss": 0.7787, "step": 4445 }, { "epoch": 0.58, "grad_norm": 1.3288941383361816, "learning_rate": 3.9137162177618335e-06, "loss": 0.8367, "step": 4446 }, { "epoch": 0.58, "grad_norm": 1.5817816257476807, "learning_rate": 3.911644879485889e-06, "loss": 0.83, "step": 4447 }, { "epoch": 0.58, "grad_norm": 1.3708271980285645, "learning_rate": 3.90957373726236e-06, "loss": 0.7671, "step": 4448 }, { "epoch": 0.58, "grad_norm": 1.3236862421035767, "learning_rate": 3.907502791464339e-06, "loss": 0.8115, "step": 4449 }, { "epoch": 0.58, "grad_norm": 1.454714298248291, "learning_rate": 3.905432042464877e-06, "loss": 0.8283, "step": 4450 }, { "epoch": 0.58, "grad_norm": 1.3885167837142944, "learning_rate": 3.903361490636989e-06, "loss": 0.7597, "step": 4451 }, { "epoch": 0.58, "grad_norm": 0.6798139214515686, "learning_rate": 3.90129113635366e-06, "loss": 1.0117, "step": 4452 }, { "epoch": 0.58, "grad_norm": 1.3447096347808838, "learning_rate": 3.899220979987834e-06, "loss": 0.7895, "step": 4453 }, { "epoch": 0.58, "grad_norm": 1.3172804117202759, "learning_rate": 3.8971510219124246e-06, "loss": 0.8248, "step": 4454 }, { "epoch": 0.58, "grad_norm": 0.629353940486908, "learning_rate": 3.8950812625003044e-06, "loss": 1.0148, "step": 4455 }, { "epoch": 0.58, "grad_norm": 1.5140397548675537, "learning_rate": 3.893011702124311e-06, "loss": 0.7451, "step": 4456 }, { "epoch": 0.58, "grad_norm": 1.3281476497650146, "learning_rate": 3.890942341157251e-06, "loss": 0.8299, "step": 4457 }, { "epoch": 0.58, "grad_norm": 0.6086899042129517, "learning_rate": 3.8888731799718885e-06, "loss": 1.0099, "step": 4458 }, { "epoch": 0.58, "grad_norm": 0.6223453283309937, "learning_rate": 3.886804218940959e-06, "loss": 1.0299, "step": 4459 }, { "epoch": 0.58, "grad_norm": 0.6219561696052551, "learning_rate": 3.884735458437155e-06, "loss": 1.0018, "step": 4460 }, { "epoch": 0.58, "grad_norm": 0.609808087348938, "learning_rate": 3.882666898833135e-06, "loss": 1.0302, "step": 4461 }, { "epoch": 0.58, "grad_norm": 1.4326395988464355, "learning_rate": 3.880598540501523e-06, "loss": 0.8435, "step": 4462 }, { "epoch": 0.58, "grad_norm": 0.6007850766181946, "learning_rate": 3.878530383814905e-06, "loss": 1.0292, "step": 4463 }, { "epoch": 0.58, "grad_norm": 1.252296805381775, "learning_rate": 3.8764624291458346e-06, "loss": 0.7841, "step": 4464 }, { "epoch": 0.59, "grad_norm": 0.6009907722473145, "learning_rate": 3.874394676866821e-06, "loss": 1.0278, "step": 4465 }, { "epoch": 0.59, "grad_norm": 1.292341947555542, "learning_rate": 3.872327127350344e-06, "loss": 0.7759, "step": 4466 }, { "epoch": 0.59, "grad_norm": 1.4297378063201904, "learning_rate": 3.870259780968846e-06, "loss": 0.7885, "step": 4467 }, { "epoch": 0.59, "grad_norm": 0.6073304414749146, "learning_rate": 3.868192638094727e-06, "loss": 1.0347, "step": 4468 }, { "epoch": 0.59, "grad_norm": 1.558402180671692, "learning_rate": 3.866125699100361e-06, "loss": 0.8633, "step": 4469 }, { "epoch": 0.59, "grad_norm": 0.6552836894989014, "learning_rate": 3.864058964358075e-06, "loss": 1.0319, "step": 4470 }, { "epoch": 0.59, "grad_norm": 1.3014247417449951, "learning_rate": 3.8619924342401625e-06, "loss": 0.8033, "step": 4471 }, { "epoch": 0.59, "grad_norm": 1.5332834720611572, "learning_rate": 3.859926109118884e-06, "loss": 0.815, "step": 4472 }, { "epoch": 0.59, "grad_norm": 0.626205325126648, "learning_rate": 3.857859989366459e-06, "loss": 1.0079, "step": 4473 }, { "epoch": 0.59, "grad_norm": 1.3308205604553223, "learning_rate": 3.855794075355068e-06, "loss": 0.8506, "step": 4474 }, { "epoch": 0.59, "grad_norm": 1.3062560558319092, "learning_rate": 3.853728367456862e-06, "loss": 0.7675, "step": 4475 }, { "epoch": 0.59, "grad_norm": 1.4192287921905518, "learning_rate": 3.851662866043945e-06, "loss": 0.8308, "step": 4476 }, { "epoch": 0.59, "grad_norm": 1.362353801727295, "learning_rate": 3.849597571488395e-06, "loss": 0.8405, "step": 4477 }, { "epoch": 0.59, "grad_norm": 1.4324305057525635, "learning_rate": 3.847532484162244e-06, "loss": 0.7826, "step": 4478 }, { "epoch": 0.59, "grad_norm": 1.2560057640075684, "learning_rate": 3.845467604437486e-06, "loss": 0.7781, "step": 4479 }, { "epoch": 0.59, "grad_norm": 0.6737470030784607, "learning_rate": 3.843402932686088e-06, "loss": 1.0127, "step": 4480 }, { "epoch": 0.59, "grad_norm": 0.6571162343025208, "learning_rate": 3.841338469279967e-06, "loss": 1.0146, "step": 4481 }, { "epoch": 0.59, "grad_norm": 1.3001970052719116, "learning_rate": 3.839274214591011e-06, "loss": 0.836, "step": 4482 }, { "epoch": 0.59, "grad_norm": 0.5892510414123535, "learning_rate": 3.8372101689910666e-06, "loss": 1.0257, "step": 4483 }, { "epoch": 0.59, "grad_norm": 1.5124861001968384, "learning_rate": 3.835146332851943e-06, "loss": 0.7639, "step": 4484 }, { "epoch": 0.59, "grad_norm": 1.3854639530181885, "learning_rate": 3.833082706545414e-06, "loss": 0.8245, "step": 4485 }, { "epoch": 0.59, "grad_norm": 1.4329053163528442, "learning_rate": 3.831019290443211e-06, "loss": 0.8551, "step": 4486 }, { "epoch": 0.59, "grad_norm": 0.6532583236694336, "learning_rate": 3.828956084917035e-06, "loss": 1.0058, "step": 4487 }, { "epoch": 0.59, "grad_norm": 1.3857988119125366, "learning_rate": 3.826893090338541e-06, "loss": 0.8668, "step": 4488 }, { "epoch": 0.59, "grad_norm": 1.354088306427002, "learning_rate": 3.824830307079348e-06, "loss": 0.7557, "step": 4489 }, { "epoch": 0.59, "grad_norm": 1.4077534675598145, "learning_rate": 3.822767735511043e-06, "loss": 0.8094, "step": 4490 }, { "epoch": 0.59, "grad_norm": 1.297497272491455, "learning_rate": 3.820705376005166e-06, "loss": 0.7291, "step": 4491 }, { "epoch": 0.59, "grad_norm": 0.6404273509979248, "learning_rate": 3.818643228933228e-06, "loss": 1.0367, "step": 4492 }, { "epoch": 0.59, "grad_norm": 1.4107903242111206, "learning_rate": 3.816581294666693e-06, "loss": 0.7832, "step": 4493 }, { "epoch": 0.59, "grad_norm": 0.5989241600036621, "learning_rate": 3.814519573576989e-06, "loss": 1.0303, "step": 4494 }, { "epoch": 0.59, "grad_norm": 0.59532231092453, "learning_rate": 3.8124580660355128e-06, "loss": 1.0392, "step": 4495 }, { "epoch": 0.59, "grad_norm": 1.3175424337387085, "learning_rate": 3.810396772413611e-06, "loss": 0.8213, "step": 4496 }, { "epoch": 0.59, "grad_norm": 1.8350951671600342, "learning_rate": 3.8083356930826025e-06, "loss": 0.7491, "step": 4497 }, { "epoch": 0.59, "grad_norm": 1.6274460554122925, "learning_rate": 3.8062748284137612e-06, "loss": 0.8136, "step": 4498 }, { "epoch": 0.59, "grad_norm": 1.3977975845336914, "learning_rate": 3.8042141787783215e-06, "loss": 0.8938, "step": 4499 }, { "epoch": 0.59, "grad_norm": 1.4024738073349, "learning_rate": 3.8021537445474854e-06, "loss": 0.8218, "step": 4500 }, { "epoch": 0.59, "grad_norm": 0.6522252559661865, "learning_rate": 3.80009352609241e-06, "loss": 1.0252, "step": 4501 }, { "epoch": 0.59, "grad_norm": 1.5786385536193848, "learning_rate": 3.798033523784214e-06, "loss": 0.8347, "step": 4502 }, { "epoch": 0.59, "grad_norm": 1.3243705034255981, "learning_rate": 3.795973737993983e-06, "loss": 0.7952, "step": 4503 }, { "epoch": 0.59, "grad_norm": 1.326534628868103, "learning_rate": 3.7939141690927567e-06, "loss": 0.7269, "step": 4504 }, { "epoch": 0.59, "grad_norm": 0.6229495406150818, "learning_rate": 3.79185481745154e-06, "loss": 1.0303, "step": 4505 }, { "epoch": 0.59, "grad_norm": 0.6168175935745239, "learning_rate": 3.789795683441295e-06, "loss": 1.0224, "step": 4506 }, { "epoch": 0.59, "grad_norm": 1.397003412246704, "learning_rate": 3.7877367674329484e-06, "loss": 0.8505, "step": 4507 }, { "epoch": 0.59, "grad_norm": 1.3589049577713013, "learning_rate": 3.785678069797387e-06, "loss": 0.7787, "step": 4508 }, { "epoch": 0.59, "grad_norm": 0.594740629196167, "learning_rate": 3.7836195909054534e-06, "loss": 1.0122, "step": 4509 }, { "epoch": 0.59, "grad_norm": 1.374300479888916, "learning_rate": 3.78156133112796e-06, "loss": 0.815, "step": 4510 }, { "epoch": 0.59, "grad_norm": 1.2428579330444336, "learning_rate": 3.779503290835671e-06, "loss": 0.7457, "step": 4511 }, { "epoch": 0.59, "grad_norm": 0.6289659738540649, "learning_rate": 3.777445470399313e-06, "loss": 1.0179, "step": 4512 }, { "epoch": 0.59, "grad_norm": 1.520715355873108, "learning_rate": 3.7753878701895792e-06, "loss": 0.8841, "step": 4513 }, { "epoch": 0.59, "grad_norm": 1.3257266283035278, "learning_rate": 3.773330490577113e-06, "loss": 0.8653, "step": 4514 }, { "epoch": 0.59, "grad_norm": 1.6515445709228516, "learning_rate": 3.7712733319325288e-06, "loss": 0.8067, "step": 4515 }, { "epoch": 0.59, "grad_norm": 0.6086208820343018, "learning_rate": 3.7692163946263925e-06, "loss": 1.0099, "step": 4516 }, { "epoch": 0.59, "grad_norm": 0.6056521534919739, "learning_rate": 3.767159679029233e-06, "loss": 1.0268, "step": 4517 }, { "epoch": 0.59, "grad_norm": 1.3972415924072266, "learning_rate": 3.765103185511543e-06, "loss": 0.7841, "step": 4518 }, { "epoch": 0.59, "grad_norm": 0.5893301367759705, "learning_rate": 3.7630469144437664e-06, "loss": 0.9915, "step": 4519 }, { "epoch": 0.59, "grad_norm": 1.2943602800369263, "learning_rate": 3.760990866196319e-06, "loss": 0.8303, "step": 4520 }, { "epoch": 0.59, "grad_norm": 0.6172223091125488, "learning_rate": 3.7589350411395674e-06, "loss": 1.0161, "step": 4521 }, { "epoch": 0.59, "grad_norm": 1.311293363571167, "learning_rate": 3.7568794396438387e-06, "loss": 0.7846, "step": 4522 }, { "epoch": 0.59, "grad_norm": 1.3673932552337646, "learning_rate": 3.7548240620794228e-06, "loss": 0.8, "step": 4523 }, { "epoch": 0.59, "grad_norm": 1.409480333328247, "learning_rate": 3.7527689088165687e-06, "loss": 0.8027, "step": 4524 }, { "epoch": 0.59, "grad_norm": 1.397377371788025, "learning_rate": 3.7507139802254855e-06, "loss": 0.7805, "step": 4525 }, { "epoch": 0.59, "grad_norm": 1.3574386835098267, "learning_rate": 3.7486592766763375e-06, "loss": 0.7692, "step": 4526 }, { "epoch": 0.59, "grad_norm": 1.4153128862380981, "learning_rate": 3.746604798539254e-06, "loss": 0.824, "step": 4527 }, { "epoch": 0.59, "grad_norm": 1.4014190435409546, "learning_rate": 3.7445505461843225e-06, "loss": 0.8394, "step": 4528 }, { "epoch": 0.59, "grad_norm": 1.284553050994873, "learning_rate": 3.7424965199815866e-06, "loss": 0.7555, "step": 4529 }, { "epoch": 0.59, "grad_norm": 2.0265564918518066, "learning_rate": 3.7404427203010497e-06, "loss": 0.7638, "step": 4530 }, { "epoch": 0.59, "grad_norm": 1.3111498355865479, "learning_rate": 3.73838914751268e-06, "loss": 0.8244, "step": 4531 }, { "epoch": 0.59, "grad_norm": 1.3353689908981323, "learning_rate": 3.736335801986396e-06, "loss": 0.7758, "step": 4532 }, { "epoch": 0.59, "grad_norm": 1.3047102689743042, "learning_rate": 3.7342826840920863e-06, "loss": 0.7869, "step": 4533 }, { "epoch": 0.59, "grad_norm": 0.7612444162368774, "learning_rate": 3.7322297941995884e-06, "loss": 1.0225, "step": 4534 }, { "epoch": 0.59, "grad_norm": 0.678650975227356, "learning_rate": 3.7301771326787008e-06, "loss": 1.0169, "step": 4535 }, { "epoch": 0.59, "grad_norm": 1.6244608163833618, "learning_rate": 3.728124699899187e-06, "loss": 0.87, "step": 4536 }, { "epoch": 0.59, "grad_norm": 1.4374046325683594, "learning_rate": 3.7260724962307605e-06, "loss": 0.8062, "step": 4537 }, { "epoch": 0.59, "grad_norm": 1.4085144996643066, "learning_rate": 3.7240205220431025e-06, "loss": 0.7473, "step": 4538 }, { "epoch": 0.59, "grad_norm": 1.3737232685089111, "learning_rate": 3.7219687777058468e-06, "loss": 0.7695, "step": 4539 }, { "epoch": 0.59, "grad_norm": 0.6626834273338318, "learning_rate": 3.719917263588584e-06, "loss": 1.0292, "step": 4540 }, { "epoch": 0.6, "grad_norm": 1.357556939125061, "learning_rate": 3.7178659800608706e-06, "loss": 0.823, "step": 4541 }, { "epoch": 0.6, "grad_norm": 1.3650926351547241, "learning_rate": 3.715814927492216e-06, "loss": 0.7636, "step": 4542 }, { "epoch": 0.6, "grad_norm": 1.3564598560333252, "learning_rate": 3.7137641062520892e-06, "loss": 0.8884, "step": 4543 }, { "epoch": 0.6, "grad_norm": 0.6233653426170349, "learning_rate": 3.7117135167099194e-06, "loss": 1.0202, "step": 4544 }, { "epoch": 0.6, "grad_norm": 0.6568004488945007, "learning_rate": 3.7096631592350895e-06, "loss": 1.0036, "step": 4545 }, { "epoch": 0.6, "grad_norm": 1.3103337287902832, "learning_rate": 3.7076130341969474e-06, "loss": 0.786, "step": 4546 }, { "epoch": 0.6, "grad_norm": 1.3764628171920776, "learning_rate": 3.705563141964791e-06, "loss": 0.763, "step": 4547 }, { "epoch": 0.6, "grad_norm": 1.4520107507705688, "learning_rate": 3.7035134829078844e-06, "loss": 0.8489, "step": 4548 }, { "epoch": 0.6, "grad_norm": 1.3889461755752563, "learning_rate": 3.7014640573954454e-06, "loss": 0.8719, "step": 4549 }, { "epoch": 0.6, "grad_norm": 1.3720885515213013, "learning_rate": 3.6994148657966465e-06, "loss": 0.8083, "step": 4550 }, { "epoch": 0.6, "grad_norm": 1.5041898488998413, "learning_rate": 3.697365908480627e-06, "loss": 0.8247, "step": 4551 }, { "epoch": 0.6, "grad_norm": 1.3176435232162476, "learning_rate": 3.6953171858164745e-06, "loss": 0.7465, "step": 4552 }, { "epoch": 0.6, "grad_norm": 0.8105617165565491, "learning_rate": 3.693268698173242e-06, "loss": 1.0032, "step": 4553 }, { "epoch": 0.6, "grad_norm": 1.4273395538330078, "learning_rate": 3.6912204459199352e-06, "loss": 0.8302, "step": 4554 }, { "epoch": 0.6, "grad_norm": 1.3310753107070923, "learning_rate": 3.689172429425517e-06, "loss": 0.7928, "step": 4555 }, { "epoch": 0.6, "grad_norm": 1.267284631729126, "learning_rate": 3.6871246490589143e-06, "loss": 0.7659, "step": 4556 }, { "epoch": 0.6, "grad_norm": 1.239939570426941, "learning_rate": 3.6850771051890035e-06, "loss": 0.7951, "step": 4557 }, { "epoch": 0.6, "grad_norm": 1.5074236392974854, "learning_rate": 3.6830297981846206e-06, "loss": 0.809, "step": 4558 }, { "epoch": 0.6, "grad_norm": 1.2956362962722778, "learning_rate": 3.6809827284145655e-06, "loss": 0.7868, "step": 4559 }, { "epoch": 0.6, "grad_norm": 0.6375194787979126, "learning_rate": 3.678935896247584e-06, "loss": 1.0122, "step": 4560 }, { "epoch": 0.6, "grad_norm": 1.3413816690444946, "learning_rate": 3.676889302052391e-06, "loss": 0.8256, "step": 4561 }, { "epoch": 0.6, "grad_norm": 1.3507691621780396, "learning_rate": 3.6748429461976493e-06, "loss": 0.8421, "step": 4562 }, { "epoch": 0.6, "grad_norm": 1.301371455192566, "learning_rate": 3.6727968290519816e-06, "loss": 0.8225, "step": 4563 }, { "epoch": 0.6, "grad_norm": 1.8170088529586792, "learning_rate": 3.6707509509839696e-06, "loss": 0.8269, "step": 4564 }, { "epoch": 0.6, "grad_norm": 1.315710186958313, "learning_rate": 3.6687053123621497e-06, "loss": 0.726, "step": 4565 }, { "epoch": 0.6, "grad_norm": 1.4389961957931519, "learning_rate": 3.6666599135550172e-06, "loss": 0.8381, "step": 4566 }, { "epoch": 0.6, "grad_norm": 1.3829551935195923, "learning_rate": 3.6646147549310217e-06, "loss": 0.8362, "step": 4567 }, { "epoch": 0.6, "grad_norm": 1.359209656715393, "learning_rate": 3.662569836858571e-06, "loss": 0.8419, "step": 4568 }, { "epoch": 0.6, "grad_norm": 1.3417870998382568, "learning_rate": 3.660525159706031e-06, "loss": 0.721, "step": 4569 }, { "epoch": 0.6, "grad_norm": 0.6396344304084778, "learning_rate": 3.6584807238417185e-06, "loss": 1.0131, "step": 4570 }, { "epoch": 0.6, "grad_norm": 1.4543061256408691, "learning_rate": 3.6564365296339153e-06, "loss": 0.8412, "step": 4571 }, { "epoch": 0.6, "grad_norm": 1.195123553276062, "learning_rate": 3.6543925774508546e-06, "loss": 0.7301, "step": 4572 }, { "epoch": 0.6, "grad_norm": 0.629848301410675, "learning_rate": 3.6523488676607235e-06, "loss": 1.0066, "step": 4573 }, { "epoch": 0.6, "grad_norm": 1.3733114004135132, "learning_rate": 3.650305400631672e-06, "loss": 0.8065, "step": 4574 }, { "epoch": 0.6, "grad_norm": 1.2561017274856567, "learning_rate": 3.6482621767317995e-06, "loss": 0.7804, "step": 4575 }, { "epoch": 0.6, "grad_norm": 0.6008248329162598, "learning_rate": 3.64621919632917e-06, "loss": 1.0268, "step": 4576 }, { "epoch": 0.6, "grad_norm": 0.5942851305007935, "learning_rate": 3.6441764597917954e-06, "loss": 1.0015, "step": 4577 }, { "epoch": 0.6, "grad_norm": 0.6149254441261292, "learning_rate": 3.642133967487646e-06, "loss": 1.0208, "step": 4578 }, { "epoch": 0.6, "grad_norm": 1.2823816537857056, "learning_rate": 3.640091719784653e-06, "loss": 0.7997, "step": 4579 }, { "epoch": 0.6, "grad_norm": 1.3422155380249023, "learning_rate": 3.6380497170506957e-06, "loss": 0.7713, "step": 4580 }, { "epoch": 0.6, "grad_norm": 1.3409382104873657, "learning_rate": 3.636007959653615e-06, "loss": 0.7942, "step": 4581 }, { "epoch": 0.6, "grad_norm": 1.3831769227981567, "learning_rate": 3.6339664479612064e-06, "loss": 0.7854, "step": 4582 }, { "epoch": 0.6, "grad_norm": 1.2925865650177002, "learning_rate": 3.6319251823412187e-06, "loss": 0.8016, "step": 4583 }, { "epoch": 0.6, "grad_norm": 1.260257601737976, "learning_rate": 3.6298841631613598e-06, "loss": 0.7355, "step": 4584 }, { "epoch": 0.6, "grad_norm": 1.2646899223327637, "learning_rate": 3.6278433907892906e-06, "loss": 0.7825, "step": 4585 }, { "epoch": 0.6, "grad_norm": 1.3230252265930176, "learning_rate": 3.62580286559263e-06, "loss": 0.8335, "step": 4586 }, { "epoch": 0.6, "grad_norm": 1.447801947593689, "learning_rate": 3.623762587938951e-06, "loss": 0.8998, "step": 4587 }, { "epoch": 0.6, "grad_norm": 1.5001429319381714, "learning_rate": 3.6217225581957784e-06, "loss": 0.8231, "step": 4588 }, { "epoch": 0.6, "grad_norm": 1.4759608507156372, "learning_rate": 3.6196827767306003e-06, "loss": 0.8443, "step": 4589 }, { "epoch": 0.6, "grad_norm": 1.333182692527771, "learning_rate": 3.6176432439108545e-06, "loss": 0.8002, "step": 4590 }, { "epoch": 0.6, "grad_norm": 1.3089336156845093, "learning_rate": 3.6156039601039315e-06, "loss": 0.7967, "step": 4591 }, { "epoch": 0.6, "grad_norm": 1.3289176225662231, "learning_rate": 3.6135649256771865e-06, "loss": 0.7869, "step": 4592 }, { "epoch": 0.6, "grad_norm": 1.3531394004821777, "learning_rate": 3.6115261409979183e-06, "loss": 0.8179, "step": 4593 }, { "epoch": 0.6, "grad_norm": 0.6518003940582275, "learning_rate": 3.6094876064333917e-06, "loss": 1.0085, "step": 4594 }, { "epoch": 0.6, "grad_norm": 1.2977043390274048, "learning_rate": 3.607449322350819e-06, "loss": 0.8431, "step": 4595 }, { "epoch": 0.6, "grad_norm": 1.3618454933166504, "learning_rate": 3.6054112891173657e-06, "loss": 0.8431, "step": 4596 }, { "epoch": 0.6, "grad_norm": 1.3105827569961548, "learning_rate": 3.603373507100162e-06, "loss": 0.8247, "step": 4597 }, { "epoch": 0.6, "grad_norm": 1.2237982749938965, "learning_rate": 3.6013359766662814e-06, "loss": 0.8591, "step": 4598 }, { "epoch": 0.6, "grad_norm": 1.3203120231628418, "learning_rate": 3.5992986981827615e-06, "loss": 0.8001, "step": 4599 }, { "epoch": 0.6, "grad_norm": 1.394968032836914, "learning_rate": 3.597261672016589e-06, "loss": 0.8266, "step": 4600 }, { "epoch": 0.6, "grad_norm": 1.2748194932937622, "learning_rate": 3.5952248985347037e-06, "loss": 0.8036, "step": 4601 }, { "epoch": 0.6, "grad_norm": 1.318142056465149, "learning_rate": 3.5931883781040078e-06, "loss": 0.7465, "step": 4602 }, { "epoch": 0.6, "grad_norm": 1.365724802017212, "learning_rate": 3.5911521110913483e-06, "loss": 0.84, "step": 4603 }, { "epoch": 0.6, "grad_norm": 1.5042192935943604, "learning_rate": 3.5891160978635348e-06, "loss": 0.7798, "step": 4604 }, { "epoch": 0.6, "grad_norm": 1.2962918281555176, "learning_rate": 3.5870803387873243e-06, "loss": 0.7988, "step": 4605 }, { "epoch": 0.6, "grad_norm": 0.6267488598823547, "learning_rate": 3.585044834229433e-06, "loss": 0.9864, "step": 4606 }, { "epoch": 0.6, "grad_norm": 0.6179139018058777, "learning_rate": 3.5830095845565297e-06, "loss": 1.0301, "step": 4607 }, { "epoch": 0.6, "grad_norm": 1.2571016550064087, "learning_rate": 3.5809745901352344e-06, "loss": 0.7904, "step": 4608 }, { "epoch": 0.6, "grad_norm": 1.3422294855117798, "learning_rate": 3.5789398513321284e-06, "loss": 0.7871, "step": 4609 }, { "epoch": 0.6, "grad_norm": 1.2920584678649902, "learning_rate": 3.576905368513739e-06, "loss": 0.7908, "step": 4610 }, { "epoch": 0.6, "grad_norm": 1.5711665153503418, "learning_rate": 3.5748711420465487e-06, "loss": 0.8333, "step": 4611 }, { "epoch": 0.6, "grad_norm": 0.640822172164917, "learning_rate": 3.572837172297001e-06, "loss": 1.0171, "step": 4612 }, { "epoch": 0.6, "grad_norm": 0.6269561648368835, "learning_rate": 3.570803459631484e-06, "loss": 1.0069, "step": 4613 }, { "epoch": 0.6, "grad_norm": 1.33266282081604, "learning_rate": 3.5687700044163464e-06, "loss": 0.8762, "step": 4614 }, { "epoch": 0.6, "grad_norm": 1.277841567993164, "learning_rate": 3.566736807017886e-06, "loss": 0.7662, "step": 4615 }, { "epoch": 0.6, "grad_norm": 1.4009482860565186, "learning_rate": 3.5647038678023537e-06, "loss": 0.8735, "step": 4616 }, { "epoch": 0.6, "grad_norm": 1.3957898616790771, "learning_rate": 3.5626711871359598e-06, "loss": 0.8502, "step": 4617 }, { "epoch": 0.61, "grad_norm": 1.27216637134552, "learning_rate": 3.5606387653848627e-06, "loss": 0.789, "step": 4618 }, { "epoch": 0.61, "grad_norm": 1.3270184993743896, "learning_rate": 3.5586066029151735e-06, "loss": 0.8058, "step": 4619 }, { "epoch": 0.61, "grad_norm": 1.5154587030410767, "learning_rate": 3.556574700092963e-06, "loss": 0.8201, "step": 4620 }, { "epoch": 0.61, "grad_norm": 1.2992651462554932, "learning_rate": 3.5545430572842467e-06, "loss": 0.8069, "step": 4621 }, { "epoch": 0.61, "grad_norm": 0.6577392220497131, "learning_rate": 3.5525116748549994e-06, "loss": 1.0175, "step": 4622 }, { "epoch": 0.61, "grad_norm": 1.4081279039382935, "learning_rate": 3.550480553171148e-06, "loss": 0.8179, "step": 4623 }, { "epoch": 0.61, "grad_norm": 1.3154408931732178, "learning_rate": 3.548449692598569e-06, "loss": 0.7936, "step": 4624 }, { "epoch": 0.61, "grad_norm": 1.358473539352417, "learning_rate": 3.5464190935030967e-06, "loss": 0.8074, "step": 4625 }, { "epoch": 0.61, "grad_norm": 1.4284448623657227, "learning_rate": 3.5443887562505153e-06, "loss": 0.7933, "step": 4626 }, { "epoch": 0.61, "grad_norm": 1.476716160774231, "learning_rate": 3.5423586812065626e-06, "loss": 0.8121, "step": 4627 }, { "epoch": 0.61, "grad_norm": 1.440077781677246, "learning_rate": 3.5403288687369296e-06, "loss": 0.8437, "step": 4628 }, { "epoch": 0.61, "grad_norm": 1.2854118347167969, "learning_rate": 3.538299319207256e-06, "loss": 0.7906, "step": 4629 }, { "epoch": 0.61, "grad_norm": 1.3911751508712769, "learning_rate": 3.5362700329831436e-06, "loss": 0.7741, "step": 4630 }, { "epoch": 0.61, "grad_norm": 1.8321467638015747, "learning_rate": 3.5342410104301346e-06, "loss": 0.8168, "step": 4631 }, { "epoch": 0.61, "grad_norm": 1.2833434343338013, "learning_rate": 3.5322122519137354e-06, "loss": 0.8016, "step": 4632 }, { "epoch": 0.61, "grad_norm": 1.3474739789962769, "learning_rate": 3.530183757799397e-06, "loss": 0.8338, "step": 4633 }, { "epoch": 0.61, "grad_norm": 1.2977416515350342, "learning_rate": 3.5281555284525227e-06, "loss": 0.8281, "step": 4634 }, { "epoch": 0.61, "grad_norm": 1.480543613433838, "learning_rate": 3.5261275642384745e-06, "loss": 0.8164, "step": 4635 }, { "epoch": 0.61, "grad_norm": 1.3531054258346558, "learning_rate": 3.5240998655225593e-06, "loss": 0.764, "step": 4636 }, { "epoch": 0.61, "grad_norm": 0.6516178846359253, "learning_rate": 3.5220724326700434e-06, "loss": 1.0298, "step": 4637 }, { "epoch": 0.61, "grad_norm": 1.418807029724121, "learning_rate": 3.5200452660461394e-06, "loss": 0.8442, "step": 4638 }, { "epoch": 0.61, "grad_norm": 0.6180102825164795, "learning_rate": 3.5180183660160106e-06, "loss": 1.0342, "step": 4639 }, { "epoch": 0.61, "grad_norm": 1.530503273010254, "learning_rate": 3.5159917329447814e-06, "loss": 0.8365, "step": 4640 }, { "epoch": 0.61, "grad_norm": 1.3073903322219849, "learning_rate": 3.5139653671975175e-06, "loss": 0.7502, "step": 4641 }, { "epoch": 0.61, "grad_norm": 0.6099628806114197, "learning_rate": 3.511939269139244e-06, "loss": 1.0518, "step": 4642 }, { "epoch": 0.61, "grad_norm": 1.527377963066101, "learning_rate": 3.5099134391349347e-06, "loss": 0.8312, "step": 4643 }, { "epoch": 0.61, "grad_norm": 1.277100682258606, "learning_rate": 3.507887877549514e-06, "loss": 0.7877, "step": 4644 }, { "epoch": 0.61, "grad_norm": 1.2902940511703491, "learning_rate": 3.505862584747861e-06, "loss": 0.829, "step": 4645 }, { "epoch": 0.61, "grad_norm": 1.3200956583023071, "learning_rate": 3.5038375610948037e-06, "loss": 0.8508, "step": 4646 }, { "epoch": 0.61, "grad_norm": 1.309497594833374, "learning_rate": 3.501812806955122e-06, "loss": 0.829, "step": 4647 }, { "epoch": 0.61, "grad_norm": 0.6567745208740234, "learning_rate": 3.4997883226935503e-06, "loss": 1.0244, "step": 4648 }, { "epoch": 0.61, "grad_norm": 0.6489103436470032, "learning_rate": 3.497764108674768e-06, "loss": 1.0003, "step": 4649 }, { "epoch": 0.61, "grad_norm": 1.3293988704681396, "learning_rate": 3.4957401652634147e-06, "loss": 0.7869, "step": 4650 }, { "epoch": 0.61, "grad_norm": 1.8150571584701538, "learning_rate": 3.493716492824074e-06, "loss": 0.8158, "step": 4651 }, { "epoch": 0.61, "grad_norm": 0.6065269708633423, "learning_rate": 3.4916930917212803e-06, "loss": 1.0082, "step": 4652 }, { "epoch": 0.61, "grad_norm": 1.2387123107910156, "learning_rate": 3.4896699623195274e-06, "loss": 0.7362, "step": 4653 }, { "epoch": 0.61, "grad_norm": 0.6211508512496948, "learning_rate": 3.4876471049832495e-06, "loss": 0.9922, "step": 4654 }, { "epoch": 0.61, "grad_norm": 1.2619400024414062, "learning_rate": 3.4856245200768413e-06, "loss": 0.7647, "step": 4655 }, { "epoch": 0.61, "grad_norm": 1.4031834602355957, "learning_rate": 3.4836022079646414e-06, "loss": 0.8945, "step": 4656 }, { "epoch": 0.61, "grad_norm": 1.320438027381897, "learning_rate": 3.481580169010941e-06, "loss": 0.7573, "step": 4657 }, { "epoch": 0.61, "grad_norm": 1.3323241472244263, "learning_rate": 3.479558403579987e-06, "loss": 0.8221, "step": 4658 }, { "epoch": 0.61, "grad_norm": 1.279031753540039, "learning_rate": 3.477536912035967e-06, "loss": 0.8249, "step": 4659 }, { "epoch": 0.61, "grad_norm": 1.2545925378799438, "learning_rate": 3.4755156947430325e-06, "loss": 0.7806, "step": 4660 }, { "epoch": 0.61, "grad_norm": 1.3384888172149658, "learning_rate": 3.4734947520652735e-06, "loss": 0.7933, "step": 4661 }, { "epoch": 0.61, "grad_norm": 0.6251102685928345, "learning_rate": 3.471474084366735e-06, "loss": 1.0027, "step": 4662 }, { "epoch": 0.61, "grad_norm": 0.598251223564148, "learning_rate": 3.4694536920114164e-06, "loss": 1.0079, "step": 4663 }, { "epoch": 0.61, "grad_norm": 1.231247901916504, "learning_rate": 3.467433575363261e-06, "loss": 0.7469, "step": 4664 }, { "epoch": 0.61, "grad_norm": 1.3278571367263794, "learning_rate": 3.4654137347861676e-06, "loss": 0.8019, "step": 4665 }, { "epoch": 0.61, "grad_norm": 1.4119466543197632, "learning_rate": 3.463394170643981e-06, "loss": 0.8084, "step": 4666 }, { "epoch": 0.61, "grad_norm": 1.3315013647079468, "learning_rate": 3.461374883300499e-06, "loss": 0.8006, "step": 4667 }, { "epoch": 0.61, "grad_norm": 1.2724580764770508, "learning_rate": 3.459355873119471e-06, "loss": 0.8118, "step": 4668 }, { "epoch": 0.61, "grad_norm": 1.3402506113052368, "learning_rate": 3.4573371404645895e-06, "loss": 0.7689, "step": 4669 }, { "epoch": 0.61, "grad_norm": 1.3983403444290161, "learning_rate": 3.455318685699507e-06, "loss": 0.8736, "step": 4670 }, { "epoch": 0.61, "grad_norm": 1.3470370769500732, "learning_rate": 3.453300509187819e-06, "loss": 0.7809, "step": 4671 }, { "epoch": 0.61, "grad_norm": 1.349684238433838, "learning_rate": 3.4512826112930696e-06, "loss": 0.8261, "step": 4672 }, { "epoch": 0.61, "grad_norm": 0.6445335745811462, "learning_rate": 3.4492649923787603e-06, "loss": 1.0059, "step": 4673 }, { "epoch": 0.61, "grad_norm": 1.4070192575454712, "learning_rate": 3.447247652808336e-06, "loss": 0.8796, "step": 4674 }, { "epoch": 0.61, "grad_norm": 0.6035231351852417, "learning_rate": 3.44523059294519e-06, "loss": 1.0261, "step": 4675 }, { "epoch": 0.61, "grad_norm": 1.2733941078186035, "learning_rate": 3.4432138131526742e-06, "loss": 0.7633, "step": 4676 }, { "epoch": 0.61, "grad_norm": 1.3933554887771606, "learning_rate": 3.441197313794079e-06, "loss": 0.7967, "step": 4677 }, { "epoch": 0.61, "grad_norm": 1.3508424758911133, "learning_rate": 3.439181095232653e-06, "loss": 0.7534, "step": 4678 }, { "epoch": 0.61, "grad_norm": 1.4120612144470215, "learning_rate": 3.437165157831589e-06, "loss": 0.8088, "step": 4679 }, { "epoch": 0.61, "grad_norm": 1.440738558769226, "learning_rate": 3.435149501954029e-06, "loss": 0.8223, "step": 4680 }, { "epoch": 0.61, "grad_norm": 1.2871413230895996, "learning_rate": 3.43313412796307e-06, "loss": 0.8114, "step": 4681 }, { "epoch": 0.61, "grad_norm": 1.379665732383728, "learning_rate": 3.431119036221752e-06, "loss": 0.8114, "step": 4682 }, { "epoch": 0.61, "grad_norm": 1.7311244010925293, "learning_rate": 3.429104227093066e-06, "loss": 0.8269, "step": 4683 }, { "epoch": 0.61, "grad_norm": 0.72144615650177, "learning_rate": 3.427089700939955e-06, "loss": 1.0202, "step": 4684 }, { "epoch": 0.61, "grad_norm": 1.3238856792449951, "learning_rate": 3.4250754581253056e-06, "loss": 0.75, "step": 4685 }, { "epoch": 0.61, "grad_norm": 0.6502658724784851, "learning_rate": 3.4230614990119564e-06, "loss": 1.0363, "step": 4686 }, { "epoch": 0.61, "grad_norm": 1.2457716464996338, "learning_rate": 3.421047823962697e-06, "loss": 0.7601, "step": 4687 }, { "epoch": 0.61, "grad_norm": 1.3728748559951782, "learning_rate": 3.419034433340264e-06, "loss": 0.7845, "step": 4688 }, { "epoch": 0.61, "grad_norm": 1.3717539310455322, "learning_rate": 3.4170213275073406e-06, "loss": 0.7983, "step": 4689 }, { "epoch": 0.61, "grad_norm": 1.368064284324646, "learning_rate": 3.415008506826558e-06, "loss": 0.8258, "step": 4690 }, { "epoch": 0.61, "grad_norm": 1.4376325607299805, "learning_rate": 3.4129959716605054e-06, "loss": 0.7771, "step": 4691 }, { "epoch": 0.61, "grad_norm": 1.287112832069397, "learning_rate": 3.410983722371707e-06, "loss": 0.8591, "step": 4692 }, { "epoch": 0.61, "grad_norm": 3.230195999145508, "learning_rate": 3.4089717593226467e-06, "loss": 0.7876, "step": 4693 }, { "epoch": 0.62, "grad_norm": 1.386665940284729, "learning_rate": 3.406960082875751e-06, "loss": 0.7846, "step": 4694 }, { "epoch": 0.62, "grad_norm": 1.2640024423599243, "learning_rate": 3.404948693393394e-06, "loss": 0.8432, "step": 4695 }, { "epoch": 0.62, "grad_norm": 1.3313403129577637, "learning_rate": 3.4029375912379046e-06, "loss": 0.7923, "step": 4696 }, { "epoch": 0.62, "grad_norm": 1.2773669958114624, "learning_rate": 3.4009267767715505e-06, "loss": 0.8023, "step": 4697 }, { "epoch": 0.62, "grad_norm": 1.3088948726654053, "learning_rate": 3.3989162503565577e-06, "loss": 0.8868, "step": 4698 }, { "epoch": 0.62, "grad_norm": 0.7894348502159119, "learning_rate": 3.396906012355092e-06, "loss": 1.0396, "step": 4699 }, { "epoch": 0.62, "grad_norm": 1.284944772720337, "learning_rate": 3.3948960631292705e-06, "loss": 0.7871, "step": 4700 }, { "epoch": 0.62, "grad_norm": 1.3380625247955322, "learning_rate": 3.392886403041161e-06, "loss": 0.7601, "step": 4701 }, { "epoch": 0.62, "grad_norm": 0.6695156097412109, "learning_rate": 3.3908770324527727e-06, "loss": 1.0191, "step": 4702 }, { "epoch": 0.62, "grad_norm": 1.3756645917892456, "learning_rate": 3.388867951726069e-06, "loss": 0.8561, "step": 4703 }, { "epoch": 0.62, "grad_norm": 1.411241054534912, "learning_rate": 3.3868591612229595e-06, "loss": 0.8433, "step": 4704 }, { "epoch": 0.62, "grad_norm": 1.6223194599151611, "learning_rate": 3.384850661305298e-06, "loss": 0.8144, "step": 4705 }, { "epoch": 0.62, "grad_norm": 1.261899471282959, "learning_rate": 3.38284245233489e-06, "loss": 0.7604, "step": 4706 }, { "epoch": 0.62, "grad_norm": 1.3999546766281128, "learning_rate": 3.380834534673486e-06, "loss": 0.806, "step": 4707 }, { "epoch": 0.62, "grad_norm": 1.354354739189148, "learning_rate": 3.3788269086827856e-06, "loss": 0.8215, "step": 4708 }, { "epoch": 0.62, "grad_norm": 1.365188479423523, "learning_rate": 3.376819574724437e-06, "loss": 0.8165, "step": 4709 }, { "epoch": 0.62, "grad_norm": 2.250487804412842, "learning_rate": 3.374812533160031e-06, "loss": 0.8081, "step": 4710 }, { "epoch": 0.62, "grad_norm": 1.298168659210205, "learning_rate": 3.3728057843511125e-06, "loss": 0.7759, "step": 4711 }, { "epoch": 0.62, "grad_norm": 1.476869821548462, "learning_rate": 3.3707993286591683e-06, "loss": 0.8362, "step": 4712 }, { "epoch": 0.62, "grad_norm": 1.4185608625411987, "learning_rate": 3.3687931664456323e-06, "loss": 0.7775, "step": 4713 }, { "epoch": 0.62, "grad_norm": 1.465653419494629, "learning_rate": 3.366787298071892e-06, "loss": 0.8065, "step": 4714 }, { "epoch": 0.62, "grad_norm": 0.8826681971549988, "learning_rate": 3.364781723899272e-06, "loss": 1.0257, "step": 4715 }, { "epoch": 0.62, "grad_norm": 1.3113493919372559, "learning_rate": 3.362776444289054e-06, "loss": 0.7909, "step": 4716 }, { "epoch": 0.62, "grad_norm": 1.4060701131820679, "learning_rate": 3.3607714596024602e-06, "loss": 0.8218, "step": 4717 }, { "epoch": 0.62, "grad_norm": 1.5624862909317017, "learning_rate": 3.3587667702006594e-06, "loss": 0.7704, "step": 4718 }, { "epoch": 0.62, "grad_norm": 1.3239266872406006, "learning_rate": 3.356762376444773e-06, "loss": 0.7445, "step": 4719 }, { "epoch": 0.62, "grad_norm": 0.7321733236312866, "learning_rate": 3.354758278695861e-06, "loss": 1.0186, "step": 4720 }, { "epoch": 0.62, "grad_norm": 1.4109143018722534, "learning_rate": 3.352754477314939e-06, "loss": 0.7863, "step": 4721 }, { "epoch": 0.62, "grad_norm": 0.6944002509117126, "learning_rate": 3.3507509726629627e-06, "loss": 1.0104, "step": 4722 }, { "epoch": 0.62, "grad_norm": 1.44301438331604, "learning_rate": 3.3487477651008348e-06, "loss": 0.8895, "step": 4723 }, { "epoch": 0.62, "grad_norm": 0.660940408706665, "learning_rate": 3.3467448549894064e-06, "loss": 1.0216, "step": 4724 }, { "epoch": 0.62, "grad_norm": 1.3643509149551392, "learning_rate": 3.3447422426894773e-06, "loss": 0.8039, "step": 4725 }, { "epoch": 0.62, "grad_norm": 1.3159124851226807, "learning_rate": 3.3427399285617884e-06, "loss": 0.8462, "step": 4726 }, { "epoch": 0.62, "grad_norm": 0.7673836350440979, "learning_rate": 3.34073791296703e-06, "loss": 1.0363, "step": 4727 }, { "epoch": 0.62, "grad_norm": 0.705376148223877, "learning_rate": 3.338736196265838e-06, "loss": 0.9899, "step": 4728 }, { "epoch": 0.62, "grad_norm": 1.34028160572052, "learning_rate": 3.3367347788187953e-06, "loss": 0.8572, "step": 4729 }, { "epoch": 0.62, "grad_norm": 1.4276206493377686, "learning_rate": 3.3347336609864276e-06, "loss": 0.7723, "step": 4730 }, { "epoch": 0.62, "grad_norm": 1.33878755569458, "learning_rate": 3.3327328431292133e-06, "loss": 0.7685, "step": 4731 }, { "epoch": 0.62, "grad_norm": 1.2680259943008423, "learning_rate": 3.3307323256075708e-06, "loss": 0.821, "step": 4732 }, { "epoch": 0.62, "grad_norm": 1.2951006889343262, "learning_rate": 3.328732108781863e-06, "loss": 0.8246, "step": 4733 }, { "epoch": 0.62, "grad_norm": 1.335298776626587, "learning_rate": 3.326732193012407e-06, "loss": 0.7853, "step": 4734 }, { "epoch": 0.62, "grad_norm": 0.7192938327789307, "learning_rate": 3.3247325786594583e-06, "loss": 1.0346, "step": 4735 }, { "epoch": 0.62, "grad_norm": 1.3241987228393555, "learning_rate": 3.3227332660832173e-06, "loss": 0.8681, "step": 4736 }, { "epoch": 0.62, "grad_norm": 1.3255177736282349, "learning_rate": 3.3207342556438387e-06, "loss": 0.8533, "step": 4737 }, { "epoch": 0.62, "grad_norm": 0.6360120177268982, "learning_rate": 3.3187355477014114e-06, "loss": 1.0133, "step": 4738 }, { "epoch": 0.62, "grad_norm": 1.500754475593567, "learning_rate": 3.3167371426159816e-06, "loss": 0.7758, "step": 4739 }, { "epoch": 0.62, "grad_norm": 1.5417128801345825, "learning_rate": 3.314739040747531e-06, "loss": 0.7483, "step": 4740 }, { "epoch": 0.62, "grad_norm": 0.6351470947265625, "learning_rate": 3.31274124245599e-06, "loss": 0.998, "step": 4741 }, { "epoch": 0.62, "grad_norm": 1.3919765949249268, "learning_rate": 3.310743748101237e-06, "loss": 0.7832, "step": 4742 }, { "epoch": 0.62, "grad_norm": 1.350653886795044, "learning_rate": 3.3087465580430927e-06, "loss": 0.7666, "step": 4743 }, { "epoch": 0.62, "grad_norm": 0.6515694260597229, "learning_rate": 3.3067496726413236e-06, "loss": 1.0052, "step": 4744 }, { "epoch": 0.62, "grad_norm": 1.2353107929229736, "learning_rate": 3.3047530922556424e-06, "loss": 0.8181, "step": 4745 }, { "epoch": 0.62, "grad_norm": 1.2786250114440918, "learning_rate": 3.3027568172457047e-06, "loss": 0.7991, "step": 4746 }, { "epoch": 0.62, "grad_norm": 0.6103754639625549, "learning_rate": 3.300760847971114e-06, "loss": 1.0157, "step": 4747 }, { "epoch": 0.62, "grad_norm": 1.6201928853988647, "learning_rate": 3.298765184791414e-06, "loss": 0.8075, "step": 4748 }, { "epoch": 0.62, "grad_norm": 0.6107372045516968, "learning_rate": 3.2967698280660997e-06, "loss": 1.008, "step": 4749 }, { "epoch": 0.62, "grad_norm": 1.2873523235321045, "learning_rate": 3.2947747781546068e-06, "loss": 0.7714, "step": 4750 }, { "epoch": 0.62, "grad_norm": 1.3738963603973389, "learning_rate": 3.2927800354163143e-06, "loss": 0.7511, "step": 4751 }, { "epoch": 0.62, "grad_norm": 1.6384501457214355, "learning_rate": 3.290785600210551e-06, "loss": 0.7877, "step": 4752 }, { "epoch": 0.62, "grad_norm": 1.3047568798065186, "learning_rate": 3.2887914728965845e-06, "loss": 0.7761, "step": 4753 }, { "epoch": 0.62, "grad_norm": 0.6087550520896912, "learning_rate": 3.286797653833633e-06, "loss": 1.0118, "step": 4754 }, { "epoch": 0.62, "grad_norm": 1.4818646907806396, "learning_rate": 3.2848041433808554e-06, "loss": 0.7829, "step": 4755 }, { "epoch": 0.62, "grad_norm": 1.4601621627807617, "learning_rate": 3.282810941897352e-06, "loss": 0.7927, "step": 4756 }, { "epoch": 0.62, "grad_norm": 1.471396803855896, "learning_rate": 3.2808180497421755e-06, "loss": 0.7772, "step": 4757 }, { "epoch": 0.62, "grad_norm": 0.6353058218955994, "learning_rate": 3.2788254672743146e-06, "loss": 1.0117, "step": 4758 }, { "epoch": 0.62, "grad_norm": 2.3324882984161377, "learning_rate": 3.27683319485271e-06, "loss": 0.7817, "step": 4759 }, { "epoch": 0.62, "grad_norm": 1.2898002862930298, "learning_rate": 3.27484123283624e-06, "loss": 0.8014, "step": 4760 }, { "epoch": 0.62, "grad_norm": 1.3974236249923706, "learning_rate": 3.272849581583728e-06, "loss": 0.7618, "step": 4761 }, { "epoch": 0.62, "grad_norm": 0.6058230400085449, "learning_rate": 3.2708582414539473e-06, "loss": 0.9789, "step": 4762 }, { "epoch": 0.62, "grad_norm": 1.4278067350387573, "learning_rate": 3.2688672128056086e-06, "loss": 0.8559, "step": 4763 }, { "epoch": 0.62, "grad_norm": 1.2816461324691772, "learning_rate": 3.2668764959973677e-06, "loss": 0.811, "step": 4764 }, { "epoch": 0.62, "grad_norm": 0.6029966473579407, "learning_rate": 3.2648860913878254e-06, "loss": 1.0205, "step": 4765 }, { "epoch": 0.62, "grad_norm": 0.6258345246315002, "learning_rate": 3.262895999335527e-06, "loss": 1.0204, "step": 4766 }, { "epoch": 0.62, "grad_norm": 0.5922671556472778, "learning_rate": 3.2609062201989618e-06, "loss": 1.0317, "step": 4767 }, { "epoch": 0.62, "grad_norm": 1.2748783826828003, "learning_rate": 3.258916754336559e-06, "loss": 0.7648, "step": 4768 }, { "epoch": 0.62, "grad_norm": 0.5942237973213196, "learning_rate": 3.2569276021066953e-06, "loss": 1.0217, "step": 4769 }, { "epoch": 0.63, "grad_norm": 1.3744215965270996, "learning_rate": 3.25493876386769e-06, "loss": 0.7824, "step": 4770 }, { "epoch": 0.63, "grad_norm": 1.4685139656066895, "learning_rate": 3.252950239977802e-06, "loss": 0.7946, "step": 4771 }, { "epoch": 0.63, "grad_norm": 1.3883311748504639, "learning_rate": 3.250962030795243e-06, "loss": 0.8436, "step": 4772 }, { "epoch": 0.63, "grad_norm": 1.294008493423462, "learning_rate": 3.248974136678158e-06, "loss": 0.7759, "step": 4773 }, { "epoch": 0.63, "grad_norm": 0.6043446063995361, "learning_rate": 3.246986557984637e-06, "loss": 0.9919, "step": 4774 }, { "epoch": 0.63, "grad_norm": 1.3140052556991577, "learning_rate": 3.2449992950727207e-06, "loss": 0.7795, "step": 4775 }, { "epoch": 0.63, "grad_norm": 1.333595871925354, "learning_rate": 3.2430123483003816e-06, "loss": 0.8132, "step": 4776 }, { "epoch": 0.63, "grad_norm": 1.2564085721969604, "learning_rate": 3.2410257180255477e-06, "loss": 0.7711, "step": 4777 }, { "epoch": 0.63, "grad_norm": 1.563035011291504, "learning_rate": 3.2390394046060796e-06, "loss": 0.8255, "step": 4778 }, { "epoch": 0.63, "grad_norm": 1.2409294843673706, "learning_rate": 3.237053408399784e-06, "loss": 0.8487, "step": 4779 }, { "epoch": 0.63, "grad_norm": 1.3276145458221436, "learning_rate": 3.235067729764414e-06, "loss": 0.8049, "step": 4780 }, { "epoch": 0.63, "grad_norm": 1.335748314857483, "learning_rate": 3.2330823690576596e-06, "loss": 0.8053, "step": 4781 }, { "epoch": 0.63, "grad_norm": 1.987518548965454, "learning_rate": 3.2310973266371603e-06, "loss": 0.7931, "step": 4782 }, { "epoch": 0.63, "grad_norm": 1.291741967201233, "learning_rate": 3.2291126028604926e-06, "loss": 0.7874, "step": 4783 }, { "epoch": 0.63, "grad_norm": 0.621894121170044, "learning_rate": 3.227128198085176e-06, "loss": 1.0023, "step": 4784 }, { "epoch": 0.63, "grad_norm": 0.6353855133056641, "learning_rate": 3.225144112668676e-06, "loss": 0.9911, "step": 4785 }, { "epoch": 0.63, "grad_norm": 1.32016122341156, "learning_rate": 3.2231603469683982e-06, "loss": 0.7867, "step": 4786 }, { "epoch": 0.63, "grad_norm": 1.408892035484314, "learning_rate": 3.221176901341693e-06, "loss": 0.7707, "step": 4787 }, { "epoch": 0.63, "grad_norm": 1.694120168685913, "learning_rate": 3.219193776145849e-06, "loss": 0.7731, "step": 4788 }, { "epoch": 0.63, "grad_norm": 1.3819000720977783, "learning_rate": 3.2172109717380973e-06, "loss": 0.8254, "step": 4789 }, { "epoch": 0.63, "grad_norm": 1.2875280380249023, "learning_rate": 3.215228488475618e-06, "loss": 0.7516, "step": 4790 }, { "epoch": 0.63, "grad_norm": 1.5067076683044434, "learning_rate": 3.213246326715526e-06, "loss": 0.7859, "step": 4791 }, { "epoch": 0.63, "grad_norm": 1.25938880443573, "learning_rate": 3.2112644868148794e-06, "loss": 0.8211, "step": 4792 }, { "epoch": 0.63, "grad_norm": 1.485992193222046, "learning_rate": 3.209282969130684e-06, "loss": 0.8122, "step": 4793 }, { "epoch": 0.63, "grad_norm": 1.3307185173034668, "learning_rate": 3.2073017740198787e-06, "loss": 0.7923, "step": 4794 }, { "epoch": 0.63, "grad_norm": 1.3493257761001587, "learning_rate": 3.205320901839353e-06, "loss": 0.8085, "step": 4795 }, { "epoch": 0.63, "grad_norm": 1.3054062128067017, "learning_rate": 3.203340352945932e-06, "loss": 0.8009, "step": 4796 }, { "epoch": 0.63, "grad_norm": 1.272583246231079, "learning_rate": 3.2013601276963825e-06, "loss": 0.7628, "step": 4797 }, { "epoch": 0.63, "grad_norm": 1.3730665445327759, "learning_rate": 3.1993802264474205e-06, "loss": 0.8097, "step": 4798 }, { "epoch": 0.63, "grad_norm": 1.3167681694030762, "learning_rate": 3.1974006495556936e-06, "loss": 0.8344, "step": 4799 }, { "epoch": 0.63, "grad_norm": 1.293603777885437, "learning_rate": 3.195421397377799e-06, "loss": 0.7597, "step": 4800 }, { "epoch": 0.63, "grad_norm": 1.2800182104110718, "learning_rate": 3.193442470270272e-06, "loss": 0.8273, "step": 4801 }, { "epoch": 0.63, "grad_norm": 0.6805842518806458, "learning_rate": 3.191463868589586e-06, "loss": 1.0073, "step": 4802 }, { "epoch": 0.63, "grad_norm": 0.6568797826766968, "learning_rate": 3.189485592692163e-06, "loss": 1.0308, "step": 4803 }, { "epoch": 0.63, "grad_norm": 0.5991194248199463, "learning_rate": 3.187507642934361e-06, "loss": 1.0154, "step": 4804 }, { "epoch": 0.63, "grad_norm": 0.6168190240859985, "learning_rate": 3.1855300196724815e-06, "loss": 1.0264, "step": 4805 }, { "epoch": 0.63, "grad_norm": 1.3327418565750122, "learning_rate": 3.1835527232627657e-06, "loss": 0.7944, "step": 4806 }, { "epoch": 0.63, "grad_norm": 0.65186607837677, "learning_rate": 3.181575754061397e-06, "loss": 1.0029, "step": 4807 }, { "epoch": 0.63, "grad_norm": 0.6215276718139648, "learning_rate": 3.1795991124244996e-06, "loss": 1.0166, "step": 4808 }, { "epoch": 0.63, "grad_norm": 1.3907694816589355, "learning_rate": 3.1776227987081375e-06, "loss": 0.7921, "step": 4809 }, { "epoch": 0.63, "grad_norm": 1.3705371618270874, "learning_rate": 3.1756468132683206e-06, "loss": 0.7766, "step": 4810 }, { "epoch": 0.63, "grad_norm": 1.6826335191726685, "learning_rate": 3.1736711564609924e-06, "loss": 0.8352, "step": 4811 }, { "epoch": 0.63, "grad_norm": 0.6445711255073547, "learning_rate": 3.1716958286420397e-06, "loss": 1.0212, "step": 4812 }, { "epoch": 0.63, "grad_norm": 1.384311556816101, "learning_rate": 3.1697208301672957e-06, "loss": 0.7142, "step": 4813 }, { "epoch": 0.63, "grad_norm": 1.270889163017273, "learning_rate": 3.167746161392524e-06, "loss": 0.7822, "step": 4814 }, { "epoch": 0.63, "grad_norm": 1.289448857307434, "learning_rate": 3.16577182267344e-06, "loss": 0.7845, "step": 4815 }, { "epoch": 0.63, "grad_norm": 1.4031966924667358, "learning_rate": 3.1637978143656912e-06, "loss": 0.7559, "step": 4816 }, { "epoch": 0.63, "grad_norm": 1.3284924030303955, "learning_rate": 3.161824136824867e-06, "loss": 0.801, "step": 4817 }, { "epoch": 0.63, "grad_norm": 1.248366355895996, "learning_rate": 3.1598507904065013e-06, "loss": 0.8151, "step": 4818 }, { "epoch": 0.63, "grad_norm": 1.5125787258148193, "learning_rate": 3.157877775466065e-06, "loss": 0.844, "step": 4819 }, { "epoch": 0.63, "grad_norm": 1.7758939266204834, "learning_rate": 3.155905092358968e-06, "loss": 0.875, "step": 4820 }, { "epoch": 0.63, "grad_norm": 0.6531118750572205, "learning_rate": 3.153932741440565e-06, "loss": 1.0064, "step": 4821 }, { "epoch": 0.63, "grad_norm": 0.6488597989082336, "learning_rate": 3.1519607230661465e-06, "loss": 1.0211, "step": 4822 }, { "epoch": 0.63, "grad_norm": 1.3101258277893066, "learning_rate": 3.1499890375909468e-06, "loss": 0.7939, "step": 4823 }, { "epoch": 0.63, "grad_norm": 1.3071246147155762, "learning_rate": 3.148017685370137e-06, "loss": 0.8079, "step": 4824 }, { "epoch": 0.63, "grad_norm": 1.3202226161956787, "learning_rate": 3.14604666675883e-06, "loss": 0.7442, "step": 4825 }, { "epoch": 0.63, "grad_norm": 1.418071985244751, "learning_rate": 3.1440759821120764e-06, "loss": 0.8135, "step": 4826 }, { "epoch": 0.63, "grad_norm": 0.5953696370124817, "learning_rate": 3.1421056317848698e-06, "loss": 1.0428, "step": 4827 }, { "epoch": 0.63, "grad_norm": 0.6040500402450562, "learning_rate": 3.1401356161321433e-06, "loss": 1.0084, "step": 4828 }, { "epoch": 0.63, "grad_norm": 0.609856367111206, "learning_rate": 3.1381659355087666e-06, "loss": 1.019, "step": 4829 }, { "epoch": 0.63, "grad_norm": 0.6260685324668884, "learning_rate": 3.1361965902695503e-06, "loss": 1.0219, "step": 4830 }, { "epoch": 0.63, "grad_norm": 1.2248731851577759, "learning_rate": 3.134227580769248e-06, "loss": 0.752, "step": 4831 }, { "epoch": 0.63, "grad_norm": 1.331159234046936, "learning_rate": 3.1322589073625464e-06, "loss": 0.7553, "step": 4832 }, { "epoch": 0.63, "grad_norm": 1.2557357549667358, "learning_rate": 3.1302905704040794e-06, "loss": 0.7781, "step": 4833 }, { "epoch": 0.63, "grad_norm": 1.3200229406356812, "learning_rate": 3.1283225702484145e-06, "loss": 0.8172, "step": 4834 }, { "epoch": 0.63, "grad_norm": 1.3612710237503052, "learning_rate": 3.1263549072500575e-06, "loss": 0.7436, "step": 4835 }, { "epoch": 0.63, "grad_norm": 1.493969440460205, "learning_rate": 3.1243875817634607e-06, "loss": 0.7591, "step": 4836 }, { "epoch": 0.63, "grad_norm": 1.5922291278839111, "learning_rate": 3.1224205941430074e-06, "loss": 0.7619, "step": 4837 }, { "epoch": 0.63, "grad_norm": 1.2976548671722412, "learning_rate": 3.1204539447430273e-06, "loss": 0.813, "step": 4838 }, { "epoch": 0.63, "grad_norm": 1.3002780675888062, "learning_rate": 3.1184876339177837e-06, "loss": 0.7652, "step": 4839 }, { "epoch": 0.63, "grad_norm": 1.3099873065948486, "learning_rate": 3.116521662021479e-06, "loss": 0.7778, "step": 4840 }, { "epoch": 0.63, "grad_norm": 0.7354375123977661, "learning_rate": 3.1145560294082604e-06, "loss": 1.0207, "step": 4841 }, { "epoch": 0.63, "grad_norm": 1.4452317953109741, "learning_rate": 3.1125907364322074e-06, "loss": 0.7467, "step": 4842 }, { "epoch": 0.63, "grad_norm": 0.6811464428901672, "learning_rate": 3.1106257834473397e-06, "loss": 0.9968, "step": 4843 }, { "epoch": 0.63, "grad_norm": 1.3393726348876953, "learning_rate": 3.108661170807621e-06, "loss": 0.7434, "step": 4844 }, { "epoch": 0.63, "grad_norm": 1.2958579063415527, "learning_rate": 3.1066968988669456e-06, "loss": 0.8281, "step": 4845 }, { "epoch": 0.63, "grad_norm": 1.4858875274658203, "learning_rate": 3.1047329679791522e-06, "loss": 0.7614, "step": 4846 }, { "epoch": 0.64, "grad_norm": 1.2574501037597656, "learning_rate": 3.1027693784980163e-06, "loss": 0.8307, "step": 4847 }, { "epoch": 0.64, "grad_norm": 1.4741876125335693, "learning_rate": 3.100806130777252e-06, "loss": 0.8092, "step": 4848 }, { "epoch": 0.64, "grad_norm": 1.3685505390167236, "learning_rate": 3.0988432251705127e-06, "loss": 0.7598, "step": 4849 }, { "epoch": 0.64, "grad_norm": 1.3352304697036743, "learning_rate": 3.096880662031385e-06, "loss": 0.7529, "step": 4850 }, { "epoch": 0.64, "grad_norm": 1.2717170715332031, "learning_rate": 3.094918441713404e-06, "loss": 0.756, "step": 4851 }, { "epoch": 0.64, "grad_norm": 0.6965552568435669, "learning_rate": 3.092956564570033e-06, "loss": 1.0275, "step": 4852 }, { "epoch": 0.64, "grad_norm": 1.2982747554779053, "learning_rate": 3.090995030954678e-06, "loss": 0.8158, "step": 4853 }, { "epoch": 0.64, "grad_norm": 1.4165852069854736, "learning_rate": 3.0890338412206845e-06, "loss": 0.8, "step": 4854 }, { "epoch": 0.64, "grad_norm": 1.3546533584594727, "learning_rate": 3.087072995721332e-06, "loss": 0.7849, "step": 4855 }, { "epoch": 0.64, "grad_norm": 1.3791559934616089, "learning_rate": 3.0851124948098432e-06, "loss": 0.7783, "step": 4856 }, { "epoch": 0.64, "grad_norm": 1.5116771459579468, "learning_rate": 3.0831523388393737e-06, "loss": 0.8054, "step": 4857 }, { "epoch": 0.64, "grad_norm": 0.6169607043266296, "learning_rate": 3.0811925281630172e-06, "loss": 1.0134, "step": 4858 }, { "epoch": 0.64, "grad_norm": 0.6091222763061523, "learning_rate": 3.0792330631338107e-06, "loss": 1.0028, "step": 4859 }, { "epoch": 0.64, "grad_norm": 1.3080697059631348, "learning_rate": 3.077273944104723e-06, "loss": 0.8683, "step": 4860 }, { "epoch": 0.64, "grad_norm": 1.368828535079956, "learning_rate": 3.0753151714286644e-06, "loss": 0.7914, "step": 4861 }, { "epoch": 0.64, "grad_norm": 1.3928991556167603, "learning_rate": 3.0733567454584805e-06, "loss": 0.8311, "step": 4862 }, { "epoch": 0.64, "grad_norm": 2.0681750774383545, "learning_rate": 3.071398666546953e-06, "loss": 0.7742, "step": 4863 }, { "epoch": 0.64, "grad_norm": 1.3063843250274658, "learning_rate": 3.069440935046808e-06, "loss": 0.8367, "step": 4864 }, { "epoch": 0.64, "grad_norm": 1.341728687286377, "learning_rate": 3.0674835513106993e-06, "loss": 0.7159, "step": 4865 }, { "epoch": 0.64, "grad_norm": 0.6341086030006409, "learning_rate": 3.0655265156912274e-06, "loss": 1.0056, "step": 4866 }, { "epoch": 0.64, "grad_norm": 0.60065096616745, "learning_rate": 3.063569828540922e-06, "loss": 1.0071, "step": 4867 }, { "epoch": 0.64, "grad_norm": 0.5843819379806519, "learning_rate": 3.061613490212256e-06, "loss": 0.9954, "step": 4868 }, { "epoch": 0.64, "grad_norm": 1.256561279296875, "learning_rate": 3.0596575010576368e-06, "loss": 0.8759, "step": 4869 }, { "epoch": 0.64, "grad_norm": 2.622250556945801, "learning_rate": 3.0577018614294074e-06, "loss": 0.7676, "step": 4870 }, { "epoch": 0.64, "grad_norm": 1.5253881216049194, "learning_rate": 3.055746571679854e-06, "loss": 0.8438, "step": 4871 }, { "epoch": 0.64, "grad_norm": 0.6097161769866943, "learning_rate": 3.0537916321611925e-06, "loss": 1.0245, "step": 4872 }, { "epoch": 0.64, "grad_norm": 1.2511454820632935, "learning_rate": 3.051837043225577e-06, "loss": 0.7433, "step": 4873 }, { "epoch": 0.64, "grad_norm": 1.291351556777954, "learning_rate": 3.049882805225105e-06, "loss": 0.8349, "step": 4874 }, { "epoch": 0.64, "grad_norm": 1.3743137121200562, "learning_rate": 3.0479289185118016e-06, "loss": 0.7906, "step": 4875 }, { "epoch": 0.64, "grad_norm": 1.443542718887329, "learning_rate": 3.0459753834376364e-06, "loss": 0.7698, "step": 4876 }, { "epoch": 0.64, "grad_norm": 1.2895833253860474, "learning_rate": 3.0440222003545113e-06, "loss": 0.7831, "step": 4877 }, { "epoch": 0.64, "grad_norm": 1.529422640800476, "learning_rate": 3.042069369614262e-06, "loss": 0.7791, "step": 4878 }, { "epoch": 0.64, "grad_norm": 1.3306092023849487, "learning_rate": 3.0401168915686707e-06, "loss": 0.8085, "step": 4879 }, { "epoch": 0.64, "grad_norm": 1.6746021509170532, "learning_rate": 3.038164766569447e-06, "loss": 0.7956, "step": 4880 }, { "epoch": 0.64, "grad_norm": 0.6195342540740967, "learning_rate": 3.0362129949682368e-06, "loss": 1.0521, "step": 4881 }, { "epoch": 0.64, "grad_norm": 1.620025396347046, "learning_rate": 3.0342615771166303e-06, "loss": 0.8021, "step": 4882 }, { "epoch": 0.64, "grad_norm": 1.3760724067687988, "learning_rate": 3.0323105133661457e-06, "loss": 0.8155, "step": 4883 }, { "epoch": 0.64, "grad_norm": 0.6269130706787109, "learning_rate": 3.0303598040682413e-06, "loss": 0.9943, "step": 4884 }, { "epoch": 0.64, "grad_norm": 0.620469868183136, "learning_rate": 3.028409449574312e-06, "loss": 1.0297, "step": 4885 }, { "epoch": 0.64, "grad_norm": 0.61760014295578, "learning_rate": 3.026459450235687e-06, "loss": 1.0302, "step": 4886 }, { "epoch": 0.64, "grad_norm": 1.2424863576889038, "learning_rate": 3.0245098064036317e-06, "loss": 0.7448, "step": 4887 }, { "epoch": 0.64, "grad_norm": 1.469536304473877, "learning_rate": 3.022560518429348e-06, "loss": 0.8614, "step": 4888 }, { "epoch": 0.64, "grad_norm": 1.32646906375885, "learning_rate": 3.0206115866639762e-06, "loss": 0.7753, "step": 4889 }, { "epoch": 0.64, "grad_norm": 0.6137300133705139, "learning_rate": 3.018663011458588e-06, "loss": 1.0051, "step": 4890 }, { "epoch": 0.64, "grad_norm": 1.5984352827072144, "learning_rate": 3.0167147931641906e-06, "loss": 0.7823, "step": 4891 }, { "epoch": 0.64, "grad_norm": 1.3747384548187256, "learning_rate": 3.014766932131733e-06, "loss": 0.7738, "step": 4892 }, { "epoch": 0.64, "grad_norm": 0.6198129057884216, "learning_rate": 3.0128194287120925e-06, "loss": 1.0094, "step": 4893 }, { "epoch": 0.64, "grad_norm": 0.595280110836029, "learning_rate": 3.0108722832560887e-06, "loss": 1.0039, "step": 4894 }, { "epoch": 0.64, "grad_norm": 1.3610888719558716, "learning_rate": 3.0089254961144725e-06, "loss": 0.7936, "step": 4895 }, { "epoch": 0.64, "grad_norm": 0.5991052985191345, "learning_rate": 3.0069790676379286e-06, "loss": 0.9999, "step": 4896 }, { "epoch": 0.64, "grad_norm": 1.3463698625564575, "learning_rate": 3.0050329981770833e-06, "loss": 0.8262, "step": 4897 }, { "epoch": 0.64, "grad_norm": 1.353022813796997, "learning_rate": 3.0030872880824914e-06, "loss": 0.8152, "step": 4898 }, { "epoch": 0.64, "grad_norm": 0.61745685338974, "learning_rate": 3.0011419377046496e-06, "loss": 1.0016, "step": 4899 }, { "epoch": 0.64, "grad_norm": 1.4912675619125366, "learning_rate": 2.999196947393985e-06, "loss": 0.7665, "step": 4900 }, { "epoch": 0.64, "grad_norm": 1.330254316329956, "learning_rate": 2.9972523175008584e-06, "loss": 0.8067, "step": 4901 }, { "epoch": 0.64, "grad_norm": 0.6070032119750977, "learning_rate": 2.9953080483755735e-06, "loss": 0.9817, "step": 4902 }, { "epoch": 0.64, "grad_norm": 1.5680288076400757, "learning_rate": 2.99336414036836e-06, "loss": 0.7974, "step": 4903 }, { "epoch": 0.64, "grad_norm": 0.6238321661949158, "learning_rate": 2.9914205938293883e-06, "loss": 1.0076, "step": 4904 }, { "epoch": 0.64, "grad_norm": 1.4201401472091675, "learning_rate": 2.989477409108763e-06, "loss": 0.7449, "step": 4905 }, { "epoch": 0.64, "grad_norm": 1.3074421882629395, "learning_rate": 2.9875345865565198e-06, "loss": 0.8707, "step": 4906 }, { "epoch": 0.64, "grad_norm": 1.3748250007629395, "learning_rate": 2.9855921265226353e-06, "loss": 0.8289, "step": 4907 }, { "epoch": 0.64, "grad_norm": 0.5891549587249756, "learning_rate": 2.9836500293570123e-06, "loss": 0.9788, "step": 4908 }, { "epoch": 0.64, "grad_norm": 0.6122168302536011, "learning_rate": 2.981708295409498e-06, "loss": 0.9979, "step": 4909 }, { "epoch": 0.64, "grad_norm": 0.6139310002326965, "learning_rate": 2.9797669250298677e-06, "loss": 0.9957, "step": 4910 }, { "epoch": 0.64, "grad_norm": 1.5439718961715698, "learning_rate": 2.9778259185678317e-06, "loss": 0.773, "step": 4911 }, { "epoch": 0.64, "grad_norm": 0.605506420135498, "learning_rate": 2.9758852763730385e-06, "loss": 0.996, "step": 4912 }, { "epoch": 0.64, "grad_norm": 1.3073219060897827, "learning_rate": 2.9739449987950675e-06, "loss": 0.8013, "step": 4913 }, { "epoch": 0.64, "grad_norm": 1.3016263246536255, "learning_rate": 2.9720050861834315e-06, "loss": 0.8355, "step": 4914 }, { "epoch": 0.64, "grad_norm": 0.6225219964981079, "learning_rate": 2.970065538887583e-06, "loss": 1.0041, "step": 4915 }, { "epoch": 0.64, "grad_norm": 1.2868826389312744, "learning_rate": 2.9681263572569007e-06, "loss": 0.8522, "step": 4916 }, { "epoch": 0.64, "grad_norm": 0.6262460350990295, "learning_rate": 2.9661875416407064e-06, "loss": 1.0039, "step": 4917 }, { "epoch": 0.64, "grad_norm": 0.6386151313781738, "learning_rate": 2.9642490923882495e-06, "loss": 1.0363, "step": 4918 }, { "epoch": 0.64, "grad_norm": 1.3040752410888672, "learning_rate": 2.962311009848714e-06, "loss": 0.7984, "step": 4919 }, { "epoch": 0.64, "grad_norm": 0.5798511505126953, "learning_rate": 2.9603732943712215e-06, "loss": 1.0056, "step": 4920 }, { "epoch": 0.64, "grad_norm": 1.3378441333770752, "learning_rate": 2.958435946304823e-06, "loss": 0.8228, "step": 4921 }, { "epoch": 0.64, "grad_norm": 1.2615402936935425, "learning_rate": 2.9564989659985078e-06, "loss": 0.8095, "step": 4922 }, { "epoch": 0.65, "grad_norm": 1.3026248216629028, "learning_rate": 2.954562353801196e-06, "loss": 0.8175, "step": 4923 }, { "epoch": 0.65, "grad_norm": 0.6232566833496094, "learning_rate": 2.95262611006174e-06, "loss": 1.0059, "step": 4924 }, { "epoch": 0.65, "grad_norm": 1.2802703380584717, "learning_rate": 2.95069023512893e-06, "loss": 0.7849, "step": 4925 }, { "epoch": 0.65, "grad_norm": 1.4670367240905762, "learning_rate": 2.9487547293514874e-06, "loss": 0.8263, "step": 4926 }, { "epoch": 0.65, "grad_norm": 1.5759799480438232, "learning_rate": 2.946819593078067e-06, "loss": 0.8088, "step": 4927 }, { "epoch": 0.65, "grad_norm": 2.128509759902954, "learning_rate": 2.9448848266572562e-06, "loss": 0.7972, "step": 4928 }, { "epoch": 0.65, "grad_norm": 0.6054830551147461, "learning_rate": 2.942950430437579e-06, "loss": 1.0056, "step": 4929 }, { "epoch": 0.65, "grad_norm": 1.3780899047851562, "learning_rate": 2.9410164047674894e-06, "loss": 0.8297, "step": 4930 }, { "epoch": 0.65, "grad_norm": 1.4664714336395264, "learning_rate": 2.9390827499953746e-06, "loss": 0.8396, "step": 4931 }, { "epoch": 0.65, "grad_norm": 1.370278000831604, "learning_rate": 2.93714946646956e-06, "loss": 0.7805, "step": 4932 }, { "epoch": 0.65, "grad_norm": 1.484582781791687, "learning_rate": 2.9352165545382983e-06, "loss": 0.7442, "step": 4933 }, { "epoch": 0.65, "grad_norm": 1.3646318912506104, "learning_rate": 2.933284014549775e-06, "loss": 0.7727, "step": 4934 }, { "epoch": 0.65, "grad_norm": 1.4170479774475098, "learning_rate": 2.931351846852115e-06, "loss": 0.6794, "step": 4935 }, { "epoch": 0.65, "grad_norm": 1.2991387844085693, "learning_rate": 2.929420051793371e-06, "loss": 0.8147, "step": 4936 }, { "epoch": 0.65, "grad_norm": 1.4067739248275757, "learning_rate": 2.9274886297215273e-06, "loss": 0.7867, "step": 4937 }, { "epoch": 0.65, "grad_norm": 1.3732141256332397, "learning_rate": 2.9255575809845067e-06, "loss": 0.8347, "step": 4938 }, { "epoch": 0.65, "grad_norm": 1.5170944929122925, "learning_rate": 2.923626905930158e-06, "loss": 0.798, "step": 4939 }, { "epoch": 0.65, "grad_norm": 1.2989834547042847, "learning_rate": 2.92169660490627e-06, "loss": 0.7944, "step": 4940 }, { "epoch": 0.65, "grad_norm": 1.3680731058120728, "learning_rate": 2.9197666782605588e-06, "loss": 0.8605, "step": 4941 }, { "epoch": 0.65, "grad_norm": 0.6208709478378296, "learning_rate": 2.917837126340671e-06, "loss": 1.0472, "step": 4942 }, { "epoch": 0.65, "grad_norm": 0.6115047931671143, "learning_rate": 2.9159079494941957e-06, "loss": 1.0285, "step": 4943 }, { "epoch": 0.65, "grad_norm": 0.6304863095283508, "learning_rate": 2.913979148068642e-06, "loss": 1.007, "step": 4944 }, { "epoch": 0.65, "grad_norm": 0.5892921686172485, "learning_rate": 2.9120507224114613e-06, "loss": 1.0331, "step": 4945 }, { "epoch": 0.65, "grad_norm": 1.5760958194732666, "learning_rate": 2.910122672870032e-06, "loss": 0.7983, "step": 4946 }, { "epoch": 0.65, "grad_norm": 1.3164342641830444, "learning_rate": 2.9081949997916642e-06, "loss": 0.8247, "step": 4947 }, { "epoch": 0.65, "grad_norm": 0.6207318305969238, "learning_rate": 2.906267703523606e-06, "loss": 0.9997, "step": 4948 }, { "epoch": 0.65, "grad_norm": 1.3666470050811768, "learning_rate": 2.904340784413029e-06, "loss": 0.7294, "step": 4949 }, { "epoch": 0.65, "grad_norm": 1.2882081270217896, "learning_rate": 2.9024142428070468e-06, "loss": 0.8049, "step": 4950 }, { "epoch": 0.65, "grad_norm": 1.4513921737670898, "learning_rate": 2.9004880790526957e-06, "loss": 0.817, "step": 4951 }, { "epoch": 0.65, "grad_norm": 1.420433521270752, "learning_rate": 2.8985622934969483e-06, "loss": 0.7848, "step": 4952 }, { "epoch": 0.65, "grad_norm": 1.4926797151565552, "learning_rate": 2.8966368864867116e-06, "loss": 0.8736, "step": 4953 }, { "epoch": 0.65, "grad_norm": 0.6281828284263611, "learning_rate": 2.8947118583688193e-06, "loss": 1.0256, "step": 4954 }, { "epoch": 0.65, "grad_norm": 1.2840039730072021, "learning_rate": 2.8927872094900384e-06, "loss": 0.7466, "step": 4955 }, { "epoch": 0.65, "grad_norm": 0.6521816253662109, "learning_rate": 2.8908629401970706e-06, "loss": 1.0058, "step": 4956 }, { "epoch": 0.65, "grad_norm": 1.4130706787109375, "learning_rate": 2.8889390508365462e-06, "loss": 0.7951, "step": 4957 }, { "epoch": 0.65, "grad_norm": 1.3003350496292114, "learning_rate": 2.8870155417550274e-06, "loss": 0.8168, "step": 4958 }, { "epoch": 0.65, "grad_norm": 1.378037691116333, "learning_rate": 2.8850924132990056e-06, "loss": 0.8673, "step": 4959 }, { "epoch": 0.65, "grad_norm": 0.602182924747467, "learning_rate": 2.883169665814911e-06, "loss": 0.9898, "step": 4960 }, { "epoch": 0.65, "grad_norm": 1.281585693359375, "learning_rate": 2.8812472996490993e-06, "loss": 0.8009, "step": 4961 }, { "epoch": 0.65, "grad_norm": 1.33250093460083, "learning_rate": 2.8793253151478553e-06, "loss": 0.7861, "step": 4962 }, { "epoch": 0.65, "grad_norm": 1.264006495475769, "learning_rate": 2.8774037126574023e-06, "loss": 0.7681, "step": 4963 }, { "epoch": 0.65, "grad_norm": 0.6241231560707092, "learning_rate": 2.8754824925238876e-06, "loss": 1.0244, "step": 4964 }, { "epoch": 0.65, "grad_norm": 0.6029791235923767, "learning_rate": 2.8735616550933975e-06, "loss": 1.0256, "step": 4965 }, { "epoch": 0.65, "grad_norm": 1.2958064079284668, "learning_rate": 2.8716412007119427e-06, "loss": 0.7941, "step": 4966 }, { "epoch": 0.65, "grad_norm": 1.2849152088165283, "learning_rate": 2.869721129725464e-06, "loss": 0.8379, "step": 4967 }, { "epoch": 0.65, "grad_norm": 0.60988450050354, "learning_rate": 2.8678014424798406e-06, "loss": 1.0206, "step": 4968 }, { "epoch": 0.65, "grad_norm": 1.3150018453598022, "learning_rate": 2.865882139320877e-06, "loss": 0.8004, "step": 4969 }, { "epoch": 0.65, "grad_norm": 1.3564602136611938, "learning_rate": 2.863963220594306e-06, "loss": 0.8154, "step": 4970 }, { "epoch": 0.65, "grad_norm": 1.3817458152770996, "learning_rate": 2.8620446866457995e-06, "loss": 0.789, "step": 4971 }, { "epoch": 0.65, "grad_norm": 1.4485585689544678, "learning_rate": 2.8601265378209523e-06, "loss": 0.807, "step": 4972 }, { "epoch": 0.65, "grad_norm": 1.4209673404693604, "learning_rate": 2.8582087744652962e-06, "loss": 0.8093, "step": 4973 }, { "epoch": 0.65, "grad_norm": 1.4893566370010376, "learning_rate": 2.8562913969242876e-06, "loss": 0.8323, "step": 4974 }, { "epoch": 0.65, "grad_norm": 1.3925226926803589, "learning_rate": 2.8543744055433153e-06, "loss": 0.8214, "step": 4975 }, { "epoch": 0.65, "grad_norm": 0.6331270933151245, "learning_rate": 2.852457800667703e-06, "loss": 1.0203, "step": 4976 }, { "epoch": 0.65, "grad_norm": 0.6357619762420654, "learning_rate": 2.8505415826426963e-06, "loss": 1.0117, "step": 4977 }, { "epoch": 0.65, "grad_norm": 0.6027647852897644, "learning_rate": 2.8486257518134795e-06, "loss": 1.0135, "step": 4978 }, { "epoch": 0.65, "grad_norm": 1.2604771852493286, "learning_rate": 2.8467103085251633e-06, "loss": 0.7671, "step": 4979 }, { "epoch": 0.65, "grad_norm": 1.3856886625289917, "learning_rate": 2.844795253122786e-06, "loss": 0.7837, "step": 4980 }, { "epoch": 0.65, "grad_norm": 1.3393521308898926, "learning_rate": 2.842880585951323e-06, "loss": 0.7534, "step": 4981 }, { "epoch": 0.65, "grad_norm": 1.4118633270263672, "learning_rate": 2.8409663073556715e-06, "loss": 0.8095, "step": 4982 }, { "epoch": 0.65, "grad_norm": 1.324476957321167, "learning_rate": 2.8390524176806665e-06, "loss": 0.8018, "step": 4983 }, { "epoch": 0.65, "grad_norm": 1.354972004890442, "learning_rate": 2.8371389172710674e-06, "loss": 0.7769, "step": 4984 }, { "epoch": 0.65, "grad_norm": 1.3531267642974854, "learning_rate": 2.8352258064715638e-06, "loss": 0.7318, "step": 4985 }, { "epoch": 0.65, "grad_norm": 1.487191915512085, "learning_rate": 2.83331308562678e-06, "loss": 0.8171, "step": 4986 }, { "epoch": 0.65, "grad_norm": 0.6858446598052979, "learning_rate": 2.8314007550812635e-06, "loss": 1.0056, "step": 4987 }, { "epoch": 0.65, "grad_norm": 1.4477643966674805, "learning_rate": 2.829488815179498e-06, "loss": 0.7687, "step": 4988 }, { "epoch": 0.65, "grad_norm": 1.3073265552520752, "learning_rate": 2.827577266265892e-06, "loss": 0.8024, "step": 4989 }, { "epoch": 0.65, "grad_norm": 1.3382123708724976, "learning_rate": 2.825666108684783e-06, "loss": 0.8264, "step": 4990 }, { "epoch": 0.65, "grad_norm": 1.3304368257522583, "learning_rate": 2.8237553427804443e-06, "loss": 0.8008, "step": 4991 }, { "epoch": 0.65, "grad_norm": 1.2945234775543213, "learning_rate": 2.821844968897072e-06, "loss": 0.7998, "step": 4992 }, { "epoch": 0.65, "grad_norm": 1.4183226823806763, "learning_rate": 2.819934987378793e-06, "loss": 0.909, "step": 4993 }, { "epoch": 0.65, "grad_norm": 1.4764856100082397, "learning_rate": 2.818025398569667e-06, "loss": 0.8194, "step": 4994 }, { "epoch": 0.65, "grad_norm": 0.6287428736686707, "learning_rate": 2.8161162028136803e-06, "loss": 1.025, "step": 4995 }, { "epoch": 0.65, "grad_norm": 1.316732406616211, "learning_rate": 2.8142074004547447e-06, "loss": 0.8172, "step": 4996 }, { "epoch": 0.65, "grad_norm": 0.5955269932746887, "learning_rate": 2.8122989918367106e-06, "loss": 1.0209, "step": 4997 }, { "epoch": 0.65, "grad_norm": 1.3232847452163696, "learning_rate": 2.81039097730335e-06, "loss": 0.8347, "step": 4998 }, { "epoch": 0.66, "grad_norm": 1.3405439853668213, "learning_rate": 2.808483357198365e-06, "loss": 0.7987, "step": 4999 }, { "epoch": 0.66, "grad_norm": 1.340009331703186, "learning_rate": 2.806576131865385e-06, "loss": 0.7447, "step": 5000 }, { "epoch": 0.66, "grad_norm": 1.4184764623641968, "learning_rate": 2.8046693016479763e-06, "loss": 0.7967, "step": 5001 }, { "epoch": 0.66, "grad_norm": 1.4145455360412598, "learning_rate": 2.8027628668896263e-06, "loss": 0.8619, "step": 5002 }, { "epoch": 0.66, "grad_norm": 1.2791633605957031, "learning_rate": 2.80085682793375e-06, "loss": 0.8154, "step": 5003 }, { "epoch": 0.66, "grad_norm": 1.7820409536361694, "learning_rate": 2.7989511851237e-06, "loss": 0.8128, "step": 5004 }, { "epoch": 0.66, "grad_norm": 0.6181132793426514, "learning_rate": 2.797045938802747e-06, "loss": 1.0318, "step": 5005 }, { "epoch": 0.66, "grad_norm": 0.6694237589836121, "learning_rate": 2.7951410893140997e-06, "loss": 1.0177, "step": 5006 }, { "epoch": 0.66, "grad_norm": 1.2920870780944824, "learning_rate": 2.793236637000889e-06, "loss": 0.8102, "step": 5007 }, { "epoch": 0.66, "grad_norm": 1.3186949491500854, "learning_rate": 2.791332582206174e-06, "loss": 0.7825, "step": 5008 }, { "epoch": 0.66, "grad_norm": 1.2816376686096191, "learning_rate": 2.789428925272948e-06, "loss": 0.8481, "step": 5009 }, { "epoch": 0.66, "grad_norm": 1.3354904651641846, "learning_rate": 2.7875256665441256e-06, "loss": 0.8236, "step": 5010 }, { "epoch": 0.66, "grad_norm": 0.6050139665603638, "learning_rate": 2.7856228063625568e-06, "loss": 1.0118, "step": 5011 }, { "epoch": 0.66, "grad_norm": 1.4474411010742188, "learning_rate": 2.783720345071014e-06, "loss": 0.7903, "step": 5012 }, { "epoch": 0.66, "grad_norm": 1.2726237773895264, "learning_rate": 2.7818182830121983e-06, "loss": 0.7909, "step": 5013 }, { "epoch": 0.66, "grad_norm": 0.6387267708778381, "learning_rate": 2.779916620528744e-06, "loss": 1.0252, "step": 5014 }, { "epoch": 0.66, "grad_norm": 0.6089372038841248, "learning_rate": 2.7780153579632052e-06, "loss": 1.0275, "step": 5015 }, { "epoch": 0.66, "grad_norm": 1.9493200778961182, "learning_rate": 2.776114495658073e-06, "loss": 0.789, "step": 5016 }, { "epoch": 0.66, "grad_norm": 1.3324040174484253, "learning_rate": 2.7742140339557607e-06, "loss": 0.7819, "step": 5017 }, { "epoch": 0.66, "grad_norm": 1.3221895694732666, "learning_rate": 2.7723139731986084e-06, "loss": 0.7745, "step": 5018 }, { "epoch": 0.66, "grad_norm": 0.5892423987388611, "learning_rate": 2.7704143137288887e-06, "loss": 1.0291, "step": 5019 }, { "epoch": 0.66, "grad_norm": 0.598829448223114, "learning_rate": 2.768515055888798e-06, "loss": 1.0036, "step": 5020 }, { "epoch": 0.66, "grad_norm": 1.4370005130767822, "learning_rate": 2.7666162000204645e-06, "loss": 0.7825, "step": 5021 }, { "epoch": 0.66, "grad_norm": 1.4201762676239014, "learning_rate": 2.764717746465939e-06, "loss": 0.7857, "step": 5022 }, { "epoch": 0.66, "grad_norm": 1.3351374864578247, "learning_rate": 2.7628196955672004e-06, "loss": 0.8361, "step": 5023 }, { "epoch": 0.66, "grad_norm": 0.5927944779396057, "learning_rate": 2.760922047666162e-06, "loss": 1.0038, "step": 5024 }, { "epoch": 0.66, "grad_norm": 1.3824808597564697, "learning_rate": 2.759024803104656e-06, "loss": 0.7933, "step": 5025 }, { "epoch": 0.66, "grad_norm": 1.390897274017334, "learning_rate": 2.757127962224444e-06, "loss": 0.8113, "step": 5026 }, { "epoch": 0.66, "grad_norm": 1.2765854597091675, "learning_rate": 2.755231525367219e-06, "loss": 0.8327, "step": 5027 }, { "epoch": 0.66, "grad_norm": 1.4853274822235107, "learning_rate": 2.7533354928745958e-06, "loss": 0.7372, "step": 5028 }, { "epoch": 0.66, "grad_norm": 1.2632309198379517, "learning_rate": 2.751439865088123e-06, "loss": 0.7802, "step": 5029 }, { "epoch": 0.66, "grad_norm": 1.3806732892990112, "learning_rate": 2.749544642349269e-06, "loss": 0.818, "step": 5030 }, { "epoch": 0.66, "grad_norm": 1.3192805051803589, "learning_rate": 2.747649824999431e-06, "loss": 0.8034, "step": 5031 }, { "epoch": 0.66, "grad_norm": 0.6170989871025085, "learning_rate": 2.745755413379939e-06, "loss": 1.0046, "step": 5032 }, { "epoch": 0.66, "grad_norm": 3.984863758087158, "learning_rate": 2.7438614078320413e-06, "loss": 0.8044, "step": 5033 }, { "epoch": 0.66, "grad_norm": 1.8458807468414307, "learning_rate": 2.7419678086969216e-06, "loss": 0.7766, "step": 5034 }, { "epoch": 0.66, "grad_norm": 1.3279837369918823, "learning_rate": 2.7400746163156837e-06, "loss": 0.8001, "step": 5035 }, { "epoch": 0.66, "grad_norm": 1.341094970703125, "learning_rate": 2.7381818310293604e-06, "loss": 0.7729, "step": 5036 }, { "epoch": 0.66, "grad_norm": 0.6369876265525818, "learning_rate": 2.7362894531789104e-06, "loss": 0.9975, "step": 5037 }, { "epoch": 0.66, "grad_norm": 0.6174123287200928, "learning_rate": 2.734397483105222e-06, "loss": 1.021, "step": 5038 }, { "epoch": 0.66, "grad_norm": 0.6034179925918579, "learning_rate": 2.7325059211491077e-06, "loss": 1.011, "step": 5039 }, { "epoch": 0.66, "grad_norm": 1.3911556005477905, "learning_rate": 2.730614767651306e-06, "loss": 0.7887, "step": 5040 }, { "epoch": 0.66, "grad_norm": 2.4281973838806152, "learning_rate": 2.7287240229524813e-06, "loss": 0.8174, "step": 5041 }, { "epoch": 0.66, "grad_norm": 0.665155291557312, "learning_rate": 2.7268336873932285e-06, "loss": 1.0063, "step": 5042 }, { "epoch": 0.66, "grad_norm": 1.3111966848373413, "learning_rate": 2.7249437613140622e-06, "loss": 0.7988, "step": 5043 }, { "epoch": 0.66, "grad_norm": 1.3872679471969604, "learning_rate": 2.7230542450554313e-06, "loss": 0.8331, "step": 5044 }, { "epoch": 0.66, "grad_norm": 0.637099027633667, "learning_rate": 2.7211651389577044e-06, "loss": 1.0208, "step": 5045 }, { "epoch": 0.66, "grad_norm": 1.6755454540252686, "learning_rate": 2.7192764433611763e-06, "loss": 0.85, "step": 5046 }, { "epoch": 0.66, "grad_norm": 0.6169679760932922, "learning_rate": 2.7173881586060746e-06, "loss": 1.0127, "step": 5047 }, { "epoch": 0.66, "grad_norm": 1.337418794631958, "learning_rate": 2.7155002850325428e-06, "loss": 0.8208, "step": 5048 }, { "epoch": 0.66, "grad_norm": 1.2504940032958984, "learning_rate": 2.7136128229806607e-06, "loss": 0.8016, "step": 5049 }, { "epoch": 0.66, "grad_norm": 1.5378974676132202, "learning_rate": 2.7117257727904267e-06, "loss": 0.8052, "step": 5050 }, { "epoch": 0.66, "grad_norm": 0.6180280447006226, "learning_rate": 2.7098391348017656e-06, "loss": 1.0132, "step": 5051 }, { "epoch": 0.66, "grad_norm": 1.1997349262237549, "learning_rate": 2.707952909354533e-06, "loss": 0.8405, "step": 5052 }, { "epoch": 0.66, "grad_norm": 1.3436461687088013, "learning_rate": 2.7060670967885062e-06, "loss": 0.7663, "step": 5053 }, { "epoch": 0.66, "grad_norm": 1.5760018825531006, "learning_rate": 2.7041816974433843e-06, "loss": 0.828, "step": 5054 }, { "epoch": 0.66, "grad_norm": 1.8787574768066406, "learning_rate": 2.7022967116588036e-06, "loss": 0.8371, "step": 5055 }, { "epoch": 0.66, "grad_norm": 1.6680785417556763, "learning_rate": 2.7004121397743123e-06, "loss": 0.8089, "step": 5056 }, { "epoch": 0.66, "grad_norm": 2.06536602973938, "learning_rate": 2.6985279821293954e-06, "loss": 0.8089, "step": 5057 }, { "epoch": 0.66, "grad_norm": 1.6429613828659058, "learning_rate": 2.6966442390634557e-06, "loss": 0.792, "step": 5058 }, { "epoch": 0.66, "grad_norm": 1.3268071413040161, "learning_rate": 2.694760910915823e-06, "loss": 0.8433, "step": 5059 }, { "epoch": 0.66, "grad_norm": 1.4843719005584717, "learning_rate": 2.6928779980257567e-06, "loss": 0.7905, "step": 5060 }, { "epoch": 0.66, "grad_norm": 1.346014380455017, "learning_rate": 2.690995500732434e-06, "loss": 0.7909, "step": 5061 }, { "epoch": 0.66, "grad_norm": 0.6423817276954651, "learning_rate": 2.6891134193749657e-06, "loss": 0.9952, "step": 5062 }, { "epoch": 0.66, "grad_norm": 1.5004247426986694, "learning_rate": 2.687231754292381e-06, "loss": 0.7824, "step": 5063 }, { "epoch": 0.66, "grad_norm": 1.344205617904663, "learning_rate": 2.685350505823634e-06, "loss": 0.8104, "step": 5064 }, { "epoch": 0.66, "grad_norm": 1.277514934539795, "learning_rate": 2.68346967430761e-06, "loss": 0.7831, "step": 5065 }, { "epoch": 0.66, "grad_norm": 1.349753975868225, "learning_rate": 2.6815892600831124e-06, "loss": 0.7587, "step": 5066 }, { "epoch": 0.66, "grad_norm": 1.2792794704437256, "learning_rate": 2.6797092634888754e-06, "loss": 0.8303, "step": 5067 }, { "epoch": 0.66, "grad_norm": 1.3730604648590088, "learning_rate": 2.677829684863553e-06, "loss": 0.8036, "step": 5068 }, { "epoch": 0.66, "grad_norm": 1.3036714792251587, "learning_rate": 2.675950524545724e-06, "loss": 0.8195, "step": 5069 }, { "epoch": 0.66, "grad_norm": 1.242684006690979, "learning_rate": 2.6740717828738973e-06, "loss": 0.7758, "step": 5070 }, { "epoch": 0.66, "grad_norm": 0.6111409664154053, "learning_rate": 2.6721934601864997e-06, "loss": 1.0162, "step": 5071 }, { "epoch": 0.66, "grad_norm": 1.2909494638442993, "learning_rate": 2.670315556821888e-06, "loss": 0.8283, "step": 5072 }, { "epoch": 0.66, "grad_norm": 1.305766224861145, "learning_rate": 2.6684380731183403e-06, "loss": 0.7883, "step": 5073 }, { "epoch": 0.66, "grad_norm": 0.6082508563995361, "learning_rate": 2.666561009414057e-06, "loss": 1.011, "step": 5074 }, { "epoch": 0.66, "grad_norm": 0.6001378893852234, "learning_rate": 2.664684366047171e-06, "loss": 0.9994, "step": 5075 }, { "epoch": 0.67, "grad_norm": 1.350361943244934, "learning_rate": 2.6628081433557307e-06, "loss": 0.7523, "step": 5076 }, { "epoch": 0.67, "grad_norm": 1.3611680269241333, "learning_rate": 2.660932341677713e-06, "loss": 0.7856, "step": 5077 }, { "epoch": 0.67, "grad_norm": 1.4612137079238892, "learning_rate": 2.6590569613510152e-06, "loss": 0.7919, "step": 5078 }, { "epoch": 0.67, "grad_norm": 0.6456922888755798, "learning_rate": 2.6571820027134663e-06, "loss": 0.9879, "step": 5079 }, { "epoch": 0.67, "grad_norm": 1.3779678344726562, "learning_rate": 2.6553074661028133e-06, "loss": 0.8534, "step": 5080 }, { "epoch": 0.67, "grad_norm": 1.5595383644104004, "learning_rate": 2.6534333518567262e-06, "loss": 0.7455, "step": 5081 }, { "epoch": 0.67, "grad_norm": 1.3363876342773438, "learning_rate": 2.6515596603128045e-06, "loss": 0.8531, "step": 5082 }, { "epoch": 0.67, "grad_norm": 1.40597665309906, "learning_rate": 2.6496863918085667e-06, "loss": 0.8741, "step": 5083 }, { "epoch": 0.67, "grad_norm": 1.325095772743225, "learning_rate": 2.6478135466814555e-06, "loss": 0.815, "step": 5084 }, { "epoch": 0.67, "grad_norm": 0.6087805032730103, "learning_rate": 2.6459411252688414e-06, "loss": 1.0196, "step": 5085 }, { "epoch": 0.67, "grad_norm": 1.2849822044372559, "learning_rate": 2.6440691279080144e-06, "loss": 0.8051, "step": 5086 }, { "epoch": 0.67, "grad_norm": 1.2285566329956055, "learning_rate": 2.6421975549361867e-06, "loss": 0.8038, "step": 5087 }, { "epoch": 0.67, "grad_norm": 1.2998322248458862, "learning_rate": 2.640326406690502e-06, "loss": 0.7692, "step": 5088 }, { "epoch": 0.67, "grad_norm": 1.2701292037963867, "learning_rate": 2.638455683508018e-06, "loss": 0.7785, "step": 5089 }, { "epoch": 0.67, "grad_norm": 1.4188865423202515, "learning_rate": 2.6365853857257227e-06, "loss": 0.8045, "step": 5090 }, { "epoch": 0.67, "grad_norm": 1.577505111694336, "learning_rate": 2.6347155136805247e-06, "loss": 0.827, "step": 5091 }, { "epoch": 0.67, "grad_norm": 1.55299973487854, "learning_rate": 2.6328460677092537e-06, "loss": 0.8427, "step": 5092 }, { "epoch": 0.67, "grad_norm": 1.3942341804504395, "learning_rate": 2.630977048148668e-06, "loss": 0.798, "step": 5093 }, { "epoch": 0.67, "grad_norm": 1.297163724899292, "learning_rate": 2.629108455335443e-06, "loss": 0.7636, "step": 5094 }, { "epoch": 0.67, "grad_norm": 1.278268575668335, "learning_rate": 2.6272402896061846e-06, "loss": 0.7957, "step": 5095 }, { "epoch": 0.67, "grad_norm": 0.6427402496337891, "learning_rate": 2.6253725512974144e-06, "loss": 0.9891, "step": 5096 }, { "epoch": 0.67, "grad_norm": 1.3928200006484985, "learning_rate": 2.623505240745579e-06, "loss": 0.7881, "step": 5097 }, { "epoch": 0.67, "grad_norm": 0.6280611157417297, "learning_rate": 2.621638358287053e-06, "loss": 1.0102, "step": 5098 }, { "epoch": 0.67, "grad_norm": 0.6236205697059631, "learning_rate": 2.619771904258126e-06, "loss": 1.0413, "step": 5099 }, { "epoch": 0.67, "grad_norm": 0.5952662229537964, "learning_rate": 2.6179058789950184e-06, "loss": 1.0232, "step": 5100 }, { "epoch": 0.67, "grad_norm": 0.5989642143249512, "learning_rate": 2.6160402828338673e-06, "loss": 1.0109, "step": 5101 }, { "epoch": 0.67, "grad_norm": 1.3444979190826416, "learning_rate": 2.614175116110732e-06, "loss": 0.8016, "step": 5102 }, { "epoch": 0.67, "grad_norm": 1.3254592418670654, "learning_rate": 2.612310379161601e-06, "loss": 0.8095, "step": 5103 }, { "epoch": 0.67, "grad_norm": 1.2505232095718384, "learning_rate": 2.610446072322379e-06, "loss": 0.7611, "step": 5104 }, { "epoch": 0.67, "grad_norm": 0.6457368731498718, "learning_rate": 2.608582195928898e-06, "loss": 1.0055, "step": 5105 }, { "epoch": 0.67, "grad_norm": 1.4894897937774658, "learning_rate": 2.6067187503169085e-06, "loss": 0.8298, "step": 5106 }, { "epoch": 0.67, "grad_norm": 1.4467014074325562, "learning_rate": 2.6048557358220826e-06, "loss": 0.8309, "step": 5107 }, { "epoch": 0.67, "grad_norm": 1.3005067110061646, "learning_rate": 2.6029931527800217e-06, "loss": 0.7467, "step": 5108 }, { "epoch": 0.67, "grad_norm": 1.2912085056304932, "learning_rate": 2.601131001526241e-06, "loss": 0.8051, "step": 5109 }, { "epoch": 0.67, "grad_norm": 1.3406978845596313, "learning_rate": 2.5992692823961853e-06, "loss": 0.8135, "step": 5110 }, { "epoch": 0.67, "grad_norm": 1.2914279699325562, "learning_rate": 2.597407995725216e-06, "loss": 0.7995, "step": 5111 }, { "epoch": 0.67, "grad_norm": 1.3006747961044312, "learning_rate": 2.5955471418486167e-06, "loss": 0.8553, "step": 5112 }, { "epoch": 0.67, "grad_norm": 0.6278954744338989, "learning_rate": 2.5936867211015993e-06, "loss": 0.996, "step": 5113 }, { "epoch": 0.67, "grad_norm": 1.2347558736801147, "learning_rate": 2.591826733819291e-06, "loss": 0.7958, "step": 5114 }, { "epoch": 0.67, "grad_norm": 0.6190347075462341, "learning_rate": 2.589967180336742e-06, "loss": 1.0012, "step": 5115 }, { "epoch": 0.67, "grad_norm": 0.5955314040184021, "learning_rate": 2.5881080609889286e-06, "loss": 0.9968, "step": 5116 }, { "epoch": 0.67, "grad_norm": 1.3648959398269653, "learning_rate": 2.5862493761107448e-06, "loss": 0.7353, "step": 5117 }, { "epoch": 0.67, "grad_norm": 0.5825329422950745, "learning_rate": 2.584391126037007e-06, "loss": 1.003, "step": 5118 }, { "epoch": 0.67, "grad_norm": 1.3256193399429321, "learning_rate": 2.582533311102452e-06, "loss": 0.7547, "step": 5119 }, { "epoch": 0.67, "grad_norm": 1.3386263847351074, "learning_rate": 2.580675931641744e-06, "loss": 0.8372, "step": 5120 }, { "epoch": 0.67, "grad_norm": 1.4185227155685425, "learning_rate": 2.5788189879894625e-06, "loss": 0.8035, "step": 5121 }, { "epoch": 0.67, "grad_norm": 1.380842924118042, "learning_rate": 2.576962480480109e-06, "loss": 0.7667, "step": 5122 }, { "epoch": 0.67, "grad_norm": 1.318367838859558, "learning_rate": 2.575106409448112e-06, "loss": 0.849, "step": 5123 }, { "epoch": 0.67, "grad_norm": 1.2759650945663452, "learning_rate": 2.5732507752278157e-06, "loss": 0.7678, "step": 5124 }, { "epoch": 0.67, "grad_norm": 0.6679052710533142, "learning_rate": 2.571395578153485e-06, "loss": 1.0166, "step": 5125 }, { "epoch": 0.67, "grad_norm": 1.39897882938385, "learning_rate": 2.5695408185593126e-06, "loss": 0.7888, "step": 5126 }, { "epoch": 0.67, "grad_norm": 1.3241512775421143, "learning_rate": 2.5676864967794047e-06, "loss": 0.7977, "step": 5127 }, { "epoch": 0.67, "grad_norm": 1.342632532119751, "learning_rate": 2.5658326131477955e-06, "loss": 0.787, "step": 5128 }, { "epoch": 0.67, "grad_norm": 1.326037049293518, "learning_rate": 2.563979167998436e-06, "loss": 0.8291, "step": 5129 }, { "epoch": 0.67, "grad_norm": 0.6008538603782654, "learning_rate": 2.5621261616651965e-06, "loss": 0.9905, "step": 5130 }, { "epoch": 0.67, "grad_norm": 1.2735378742218018, "learning_rate": 2.5602735944818754e-06, "loss": 0.761, "step": 5131 }, { "epoch": 0.67, "grad_norm": 1.5467537641525269, "learning_rate": 2.558421466782183e-06, "loss": 0.7669, "step": 5132 }, { "epoch": 0.67, "grad_norm": 0.5756959319114685, "learning_rate": 2.5565697788997595e-06, "loss": 1.0197, "step": 5133 }, { "epoch": 0.67, "grad_norm": 1.368331789970398, "learning_rate": 2.5547185311681588e-06, "loss": 0.7827, "step": 5134 }, { "epoch": 0.67, "grad_norm": 1.4264875650405884, "learning_rate": 2.552867723920857e-06, "loss": 0.8147, "step": 5135 }, { "epoch": 0.67, "grad_norm": 1.345211148262024, "learning_rate": 2.5510173574912554e-06, "loss": 0.8772, "step": 5136 }, { "epoch": 0.67, "grad_norm": 1.6850656270980835, "learning_rate": 2.549167432212668e-06, "loss": 0.7774, "step": 5137 }, { "epoch": 0.67, "grad_norm": 0.6085002422332764, "learning_rate": 2.5473179484183386e-06, "loss": 1.0127, "step": 5138 }, { "epoch": 0.67, "grad_norm": 1.4387880563735962, "learning_rate": 2.5454689064414233e-06, "loss": 0.7877, "step": 5139 }, { "epoch": 0.67, "grad_norm": 1.3404450416564941, "learning_rate": 2.5436203066150017e-06, "loss": 0.8589, "step": 5140 }, { "epoch": 0.67, "grad_norm": 1.3103632926940918, "learning_rate": 2.541772149272077e-06, "loss": 0.7923, "step": 5141 }, { "epoch": 0.67, "grad_norm": 1.3270940780639648, "learning_rate": 2.5399244347455677e-06, "loss": 0.7985, "step": 5142 }, { "epoch": 0.67, "grad_norm": 0.6257933378219604, "learning_rate": 2.538077163368312e-06, "loss": 1.0145, "step": 5143 }, { "epoch": 0.67, "grad_norm": 0.6149579882621765, "learning_rate": 2.5362303354730768e-06, "loss": 1.0188, "step": 5144 }, { "epoch": 0.67, "grad_norm": 1.3289908170700073, "learning_rate": 2.534383951392536e-06, "loss": 0.7543, "step": 5145 }, { "epoch": 0.67, "grad_norm": 1.3842376470565796, "learning_rate": 2.532538011459298e-06, "loss": 0.8143, "step": 5146 }, { "epoch": 0.67, "grad_norm": 1.3801053762435913, "learning_rate": 2.5306925160058794e-06, "loss": 0.8021, "step": 5147 }, { "epoch": 0.67, "grad_norm": 0.5970144271850586, "learning_rate": 2.52884746536472e-06, "loss": 1.0181, "step": 5148 }, { "epoch": 0.67, "grad_norm": 1.4017754793167114, "learning_rate": 2.5270028598681846e-06, "loss": 0.7948, "step": 5149 }, { "epoch": 0.67, "grad_norm": 1.5316712856292725, "learning_rate": 2.52515869984855e-06, "loss": 0.8353, "step": 5150 }, { "epoch": 0.67, "grad_norm": 1.4606083631515503, "learning_rate": 2.523314985638021e-06, "loss": 0.7316, "step": 5151 }, { "epoch": 0.68, "grad_norm": 1.4387553930282593, "learning_rate": 2.521471717568715e-06, "loss": 0.8819, "step": 5152 }, { "epoch": 0.68, "grad_norm": 1.2716232538223267, "learning_rate": 2.51962889597267e-06, "loss": 0.7273, "step": 5153 }, { "epoch": 0.68, "grad_norm": 0.6474739909172058, "learning_rate": 2.5177865211818486e-06, "loss": 1.0006, "step": 5154 }, { "epoch": 0.68, "grad_norm": 1.3452452421188354, "learning_rate": 2.515944593528129e-06, "loss": 0.7754, "step": 5155 }, { "epoch": 0.68, "grad_norm": 1.3160622119903564, "learning_rate": 2.514103113343307e-06, "loss": 0.7338, "step": 5156 }, { "epoch": 0.68, "grad_norm": 1.357749581336975, "learning_rate": 2.512262080959105e-06, "loss": 0.8057, "step": 5157 }, { "epoch": 0.68, "grad_norm": 0.5993131399154663, "learning_rate": 2.5104214967071554e-06, "loss": 1.0096, "step": 5158 }, { "epoch": 0.68, "grad_norm": 1.3958460092544556, "learning_rate": 2.5085813609190173e-06, "loss": 0.7793, "step": 5159 }, { "epoch": 0.68, "grad_norm": 1.305528163909912, "learning_rate": 2.506741673926163e-06, "loss": 0.8069, "step": 5160 }, { "epoch": 0.68, "grad_norm": 1.6262602806091309, "learning_rate": 2.5049024360599914e-06, "loss": 0.8121, "step": 5161 }, { "epoch": 0.68, "grad_norm": 1.2856719493865967, "learning_rate": 2.5030636476518145e-06, "loss": 0.7996, "step": 5162 }, { "epoch": 0.68, "grad_norm": 1.2815899848937988, "learning_rate": 2.501225309032862e-06, "loss": 0.782, "step": 5163 }, { "epoch": 0.68, "grad_norm": 2.221837043762207, "learning_rate": 2.499387420534291e-06, "loss": 0.823, "step": 5164 }, { "epoch": 0.68, "grad_norm": 1.330805778503418, "learning_rate": 2.497549982487168e-06, "loss": 0.7595, "step": 5165 }, { "epoch": 0.68, "grad_norm": 1.4635227918624878, "learning_rate": 2.495712995222486e-06, "loss": 0.8044, "step": 5166 }, { "epoch": 0.68, "grad_norm": 0.6268345713615417, "learning_rate": 2.493876459071151e-06, "loss": 0.993, "step": 5167 }, { "epoch": 0.68, "grad_norm": 0.6317769289016724, "learning_rate": 2.4920403743639893e-06, "loss": 1.0192, "step": 5168 }, { "epoch": 0.68, "grad_norm": 1.34707510471344, "learning_rate": 2.4902047414317497e-06, "loss": 0.7935, "step": 5169 }, { "epoch": 0.68, "grad_norm": 0.591114342212677, "learning_rate": 2.4883695606050952e-06, "loss": 1.0214, "step": 5170 }, { "epoch": 0.68, "grad_norm": 1.30140221118927, "learning_rate": 2.4865348322146064e-06, "loss": 0.7626, "step": 5171 }, { "epoch": 0.68, "grad_norm": 1.306078314781189, "learning_rate": 2.4847005565907886e-06, "loss": 0.7912, "step": 5172 }, { "epoch": 0.68, "grad_norm": 1.3689621686935425, "learning_rate": 2.482866734064058e-06, "loss": 0.7558, "step": 5173 }, { "epoch": 0.68, "grad_norm": 0.6081058979034424, "learning_rate": 2.4810333649647565e-06, "loss": 0.9948, "step": 5174 }, { "epoch": 0.68, "grad_norm": 1.5576086044311523, "learning_rate": 2.479200449623139e-06, "loss": 0.8332, "step": 5175 }, { "epoch": 0.68, "grad_norm": 1.4227908849716187, "learning_rate": 2.4773679883693785e-06, "loss": 0.8585, "step": 5176 }, { "epoch": 0.68, "grad_norm": 1.3220828771591187, "learning_rate": 2.475535981533572e-06, "loss": 0.8443, "step": 5177 }, { "epoch": 0.68, "grad_norm": 1.5411372184753418, "learning_rate": 2.473704429445726e-06, "loss": 0.7088, "step": 5178 }, { "epoch": 0.68, "grad_norm": 0.6230736374855042, "learning_rate": 2.4718733324357746e-06, "loss": 0.9881, "step": 5179 }, { "epoch": 0.68, "grad_norm": 1.2917759418487549, "learning_rate": 2.4700426908335627e-06, "loss": 0.7852, "step": 5180 }, { "epoch": 0.68, "grad_norm": 0.6401428580284119, "learning_rate": 2.4682125049688536e-06, "loss": 1.0209, "step": 5181 }, { "epoch": 0.68, "grad_norm": 1.2861263751983643, "learning_rate": 2.466382775171335e-06, "loss": 0.8073, "step": 5182 }, { "epoch": 0.68, "grad_norm": 1.2392990589141846, "learning_rate": 2.464553501770603e-06, "loss": 0.7552, "step": 5183 }, { "epoch": 0.68, "grad_norm": 1.4558981657028198, "learning_rate": 2.4627246850961804e-06, "loss": 0.8156, "step": 5184 }, { "epoch": 0.68, "grad_norm": 1.3135181665420532, "learning_rate": 2.4608963254775027e-06, "loss": 0.799, "step": 5185 }, { "epoch": 0.68, "grad_norm": 1.386446237564087, "learning_rate": 2.4590684232439216e-06, "loss": 0.8012, "step": 5186 }, { "epoch": 0.68, "grad_norm": 1.2946645021438599, "learning_rate": 2.4572409787247127e-06, "loss": 0.8252, "step": 5187 }, { "epoch": 0.68, "grad_norm": 0.6938742399215698, "learning_rate": 2.4554139922490617e-06, "loss": 1.0423, "step": 5188 }, { "epoch": 0.68, "grad_norm": 1.5285277366638184, "learning_rate": 2.453587464146079e-06, "loss": 0.7772, "step": 5189 }, { "epoch": 0.68, "grad_norm": 0.6216615438461304, "learning_rate": 2.4517613947447867e-06, "loss": 1.013, "step": 5190 }, { "epoch": 0.68, "grad_norm": 1.3805325031280518, "learning_rate": 2.449935784374125e-06, "loss": 0.8296, "step": 5191 }, { "epoch": 0.68, "grad_norm": 1.8922278881072998, "learning_rate": 2.4481106333629573e-06, "loss": 0.7572, "step": 5192 }, { "epoch": 0.68, "grad_norm": 0.6000376343727112, "learning_rate": 2.4462859420400547e-06, "loss": 1.0108, "step": 5193 }, { "epoch": 0.68, "grad_norm": 1.3836199045181274, "learning_rate": 2.4444617107341157e-06, "loss": 0.7844, "step": 5194 }, { "epoch": 0.68, "grad_norm": 1.431630253791809, "learning_rate": 2.4426379397737474e-06, "loss": 0.7847, "step": 5195 }, { "epoch": 0.68, "grad_norm": 0.6036073565483093, "learning_rate": 2.4408146294874785e-06, "loss": 1.0076, "step": 5196 }, { "epoch": 0.68, "grad_norm": 1.6561604738235474, "learning_rate": 2.4389917802037518e-06, "loss": 0.7782, "step": 5197 }, { "epoch": 0.68, "grad_norm": 0.5939946174621582, "learning_rate": 2.437169392250932e-06, "loss": 1.0227, "step": 5198 }, { "epoch": 0.68, "grad_norm": 1.3805607557296753, "learning_rate": 2.4353474659572962e-06, "loss": 0.7716, "step": 5199 }, { "epoch": 0.68, "grad_norm": 0.5924229621887207, "learning_rate": 2.43352600165104e-06, "loss": 1.0171, "step": 5200 }, { "epoch": 0.68, "grad_norm": 1.320926547050476, "learning_rate": 2.4317049996602727e-06, "loss": 0.8431, "step": 5201 }, { "epoch": 0.68, "grad_norm": 1.5276570320129395, "learning_rate": 2.429884460313027e-06, "loss": 0.8359, "step": 5202 }, { "epoch": 0.68, "grad_norm": 1.4486172199249268, "learning_rate": 2.4280643839372476e-06, "loss": 0.7905, "step": 5203 }, { "epoch": 0.68, "grad_norm": 0.6228839755058289, "learning_rate": 2.426244770860794e-06, "loss": 0.9897, "step": 5204 }, { "epoch": 0.68, "grad_norm": 0.6286916136741638, "learning_rate": 2.424425621411448e-06, "loss": 1.0183, "step": 5205 }, { "epoch": 0.68, "grad_norm": 1.3226220607757568, "learning_rate": 2.4226069359169015e-06, "loss": 0.8055, "step": 5206 }, { "epoch": 0.68, "grad_norm": 1.260408878326416, "learning_rate": 2.4207887147047696e-06, "loss": 0.849, "step": 5207 }, { "epoch": 0.68, "grad_norm": 1.3619710206985474, "learning_rate": 2.4189709581025794e-06, "loss": 0.7921, "step": 5208 }, { "epoch": 0.68, "grad_norm": 1.320326328277588, "learning_rate": 2.4171536664377714e-06, "loss": 0.7551, "step": 5209 }, { "epoch": 0.68, "grad_norm": 1.3535339832305908, "learning_rate": 2.4153368400377114e-06, "loss": 0.8088, "step": 5210 }, { "epoch": 0.68, "grad_norm": 1.3298029899597168, "learning_rate": 2.4135204792296714e-06, "loss": 0.7929, "step": 5211 }, { "epoch": 0.68, "grad_norm": 1.3339184522628784, "learning_rate": 2.4117045843408476e-06, "loss": 0.7918, "step": 5212 }, { "epoch": 0.68, "grad_norm": 1.2595858573913574, "learning_rate": 2.4098891556983483e-06, "loss": 0.751, "step": 5213 }, { "epoch": 0.68, "grad_norm": 0.614092230796814, "learning_rate": 2.4080741936291953e-06, "loss": 1.0013, "step": 5214 }, { "epoch": 0.68, "grad_norm": 1.3777928352355957, "learning_rate": 2.406259698460333e-06, "loss": 0.7885, "step": 5215 }, { "epoch": 0.68, "grad_norm": 1.5409637689590454, "learning_rate": 2.404445670518616e-06, "loss": 0.832, "step": 5216 }, { "epoch": 0.68, "grad_norm": 1.3659110069274902, "learning_rate": 2.4026321101308186e-06, "loss": 0.7729, "step": 5217 }, { "epoch": 0.68, "grad_norm": 1.2731610536575317, "learning_rate": 2.4008190176236287e-06, "loss": 0.8008, "step": 5218 }, { "epoch": 0.68, "grad_norm": 1.5122157335281372, "learning_rate": 2.3990063933236475e-06, "loss": 0.7622, "step": 5219 }, { "epoch": 0.68, "grad_norm": 1.7965084314346313, "learning_rate": 2.397194237557399e-06, "loss": 0.8311, "step": 5220 }, { "epoch": 0.68, "grad_norm": 1.6391831636428833, "learning_rate": 2.3953825506513145e-06, "loss": 0.7985, "step": 5221 }, { "epoch": 0.68, "grad_norm": 1.4224395751953125, "learning_rate": 2.3935713329317486e-06, "loss": 0.8322, "step": 5222 }, { "epoch": 0.68, "grad_norm": 1.346585750579834, "learning_rate": 2.391760584724966e-06, "loss": 0.821, "step": 5223 }, { "epoch": 0.68, "grad_norm": 1.3159562349319458, "learning_rate": 2.3899503063571463e-06, "loss": 0.7836, "step": 5224 }, { "epoch": 0.68, "grad_norm": 0.6176594495773315, "learning_rate": 2.3881404981543898e-06, "loss": 1.0052, "step": 5225 }, { "epoch": 0.68, "grad_norm": 1.2816754579544067, "learning_rate": 2.3863311604427063e-06, "loss": 0.752, "step": 5226 }, { "epoch": 0.68, "grad_norm": 1.2831398248672485, "learning_rate": 2.3845222935480266e-06, "loss": 0.8288, "step": 5227 }, { "epoch": 0.69, "grad_norm": 1.3212766647338867, "learning_rate": 2.3827138977961915e-06, "loss": 0.7704, "step": 5228 }, { "epoch": 0.69, "grad_norm": 1.5767954587936401, "learning_rate": 2.3809059735129575e-06, "loss": 0.8192, "step": 5229 }, { "epoch": 0.69, "grad_norm": 1.3088175058364868, "learning_rate": 2.379098521024001e-06, "loss": 0.8203, "step": 5230 }, { "epoch": 0.69, "grad_norm": 1.301761269569397, "learning_rate": 2.3772915406549084e-06, "loss": 0.8056, "step": 5231 }, { "epoch": 0.69, "grad_norm": 1.262900948524475, "learning_rate": 2.375485032731181e-06, "loss": 0.7718, "step": 5232 }, { "epoch": 0.69, "grad_norm": 1.3061878681182861, "learning_rate": 2.3736789975782404e-06, "loss": 0.8052, "step": 5233 }, { "epoch": 0.69, "grad_norm": 1.399498462677002, "learning_rate": 2.3718734355214157e-06, "loss": 0.8063, "step": 5234 }, { "epoch": 0.69, "grad_norm": 0.6175563931465149, "learning_rate": 2.3700683468859576e-06, "loss": 1.0162, "step": 5235 }, { "epoch": 0.69, "grad_norm": 1.2824084758758545, "learning_rate": 2.3682637319970265e-06, "loss": 0.8298, "step": 5236 }, { "epoch": 0.69, "grad_norm": 1.2573856115341187, "learning_rate": 2.3664595911797e-06, "loss": 0.8061, "step": 5237 }, { "epoch": 0.69, "grad_norm": 1.3804668188095093, "learning_rate": 2.364655924758967e-06, "loss": 0.8267, "step": 5238 }, { "epoch": 0.69, "grad_norm": 1.311626672744751, "learning_rate": 2.3628527330597385e-06, "loss": 0.8444, "step": 5239 }, { "epoch": 0.69, "grad_norm": 1.7728450298309326, "learning_rate": 2.3610500164068317e-06, "loss": 0.8459, "step": 5240 }, { "epoch": 0.69, "grad_norm": 0.6242385506629944, "learning_rate": 2.3592477751249827e-06, "loss": 1.0222, "step": 5241 }, { "epoch": 0.69, "grad_norm": 1.4905623197555542, "learning_rate": 2.3574460095388375e-06, "loss": 0.7829, "step": 5242 }, { "epoch": 0.69, "grad_norm": 1.3010307550430298, "learning_rate": 2.3556447199729648e-06, "loss": 0.7996, "step": 5243 }, { "epoch": 0.69, "grad_norm": 1.3912711143493652, "learning_rate": 2.3538439067518388e-06, "loss": 0.7259, "step": 5244 }, { "epoch": 0.69, "grad_norm": 1.8654717206954956, "learning_rate": 2.3520435701998544e-06, "loss": 0.7506, "step": 5245 }, { "epoch": 0.69, "grad_norm": 1.3504595756530762, "learning_rate": 2.350243710641317e-06, "loss": 0.8284, "step": 5246 }, { "epoch": 0.69, "grad_norm": 1.3476035594940186, "learning_rate": 2.348444328400444e-06, "loss": 0.8109, "step": 5247 }, { "epoch": 0.69, "grad_norm": 1.386000394821167, "learning_rate": 2.3466454238013743e-06, "loss": 0.8155, "step": 5248 }, { "epoch": 0.69, "grad_norm": 1.311568260192871, "learning_rate": 2.3448469971681514e-06, "loss": 0.8065, "step": 5249 }, { "epoch": 0.69, "grad_norm": 0.6098197102546692, "learning_rate": 2.3430490488247427e-06, "loss": 1.006, "step": 5250 }, { "epoch": 0.69, "grad_norm": 1.392672061920166, "learning_rate": 2.3412515790950207e-06, "loss": 0.7622, "step": 5251 }, { "epoch": 0.69, "grad_norm": 1.3092347383499146, "learning_rate": 2.3394545883027747e-06, "loss": 0.8785, "step": 5252 }, { "epoch": 0.69, "grad_norm": 1.4896091222763062, "learning_rate": 2.337658076771711e-06, "loss": 0.7449, "step": 5253 }, { "epoch": 0.69, "grad_norm": 1.318793773651123, "learning_rate": 2.3358620448254437e-06, "loss": 0.807, "step": 5254 }, { "epoch": 0.69, "grad_norm": 0.5921041965484619, "learning_rate": 2.3340664927875067e-06, "loss": 1.0165, "step": 5255 }, { "epoch": 0.69, "grad_norm": 1.311013102531433, "learning_rate": 2.3322714209813433e-06, "loss": 0.8299, "step": 5256 }, { "epoch": 0.69, "grad_norm": 1.2561407089233398, "learning_rate": 2.3304768297303094e-06, "loss": 0.7806, "step": 5257 }, { "epoch": 0.69, "grad_norm": 1.294939398765564, "learning_rate": 2.328682719357679e-06, "loss": 0.7424, "step": 5258 }, { "epoch": 0.69, "grad_norm": 1.3084477186203003, "learning_rate": 2.326889090186636e-06, "loss": 0.8234, "step": 5259 }, { "epoch": 0.69, "grad_norm": 1.333473563194275, "learning_rate": 2.325095942540276e-06, "loss": 0.792, "step": 5260 }, { "epoch": 0.69, "grad_norm": 1.2884433269500732, "learning_rate": 2.3233032767416136e-06, "loss": 0.8285, "step": 5261 }, { "epoch": 0.69, "grad_norm": 1.3531895875930786, "learning_rate": 2.3215110931135702e-06, "loss": 0.778, "step": 5262 }, { "epoch": 0.69, "grad_norm": 1.2594470977783203, "learning_rate": 2.3197193919789873e-06, "loss": 0.7675, "step": 5263 }, { "epoch": 0.69, "grad_norm": 1.2842466831207275, "learning_rate": 2.317928173660613e-06, "loss": 0.7586, "step": 5264 }, { "epoch": 0.69, "grad_norm": 0.6462261080741882, "learning_rate": 2.316137438481109e-06, "loss": 1.0004, "step": 5265 }, { "epoch": 0.69, "grad_norm": 0.6159448027610779, "learning_rate": 2.3143471867630563e-06, "loss": 0.9919, "step": 5266 }, { "epoch": 0.69, "grad_norm": 1.5030324459075928, "learning_rate": 2.3125574188289408e-06, "loss": 0.794, "step": 5267 }, { "epoch": 0.69, "grad_norm": 1.4074751138687134, "learning_rate": 2.3107681350011677e-06, "loss": 0.8577, "step": 5268 }, { "epoch": 0.69, "grad_norm": 1.3051588535308838, "learning_rate": 2.308979335602052e-06, "loss": 0.772, "step": 5269 }, { "epoch": 0.69, "grad_norm": 0.5878707766532898, "learning_rate": 2.3071910209538177e-06, "loss": 0.9935, "step": 5270 }, { "epoch": 0.69, "grad_norm": 1.5980933904647827, "learning_rate": 2.3054031913786113e-06, "loss": 0.7782, "step": 5271 }, { "epoch": 0.69, "grad_norm": 0.6104573607444763, "learning_rate": 2.303615847198481e-06, "loss": 1.0134, "step": 5272 }, { "epoch": 0.69, "grad_norm": 1.346009373664856, "learning_rate": 2.3018289887353964e-06, "loss": 0.7458, "step": 5273 }, { "epoch": 0.69, "grad_norm": 1.3749687671661377, "learning_rate": 2.3000426163112342e-06, "loss": 0.7912, "step": 5274 }, { "epoch": 0.69, "grad_norm": 1.4456398487091064, "learning_rate": 2.298256730247784e-06, "loss": 0.8795, "step": 5275 }, { "epoch": 0.69, "grad_norm": 1.3451329469680786, "learning_rate": 2.2964713308667514e-06, "loss": 0.7605, "step": 5276 }, { "epoch": 0.69, "grad_norm": 1.4475089311599731, "learning_rate": 2.2946864184897505e-06, "loss": 0.8277, "step": 5277 }, { "epoch": 0.69, "grad_norm": 1.3910155296325684, "learning_rate": 2.2929019934383095e-06, "loss": 0.7744, "step": 5278 }, { "epoch": 0.69, "grad_norm": 1.3517754077911377, "learning_rate": 2.291118056033866e-06, "loss": 0.7374, "step": 5279 }, { "epoch": 0.69, "grad_norm": 1.3783420324325562, "learning_rate": 2.2893346065977764e-06, "loss": 0.8211, "step": 5280 }, { "epoch": 0.69, "grad_norm": 2.552692413330078, "learning_rate": 2.2875516454513023e-06, "loss": 0.8468, "step": 5281 }, { "epoch": 0.69, "grad_norm": 0.6142101287841797, "learning_rate": 2.285769172915619e-06, "loss": 1.011, "step": 5282 }, { "epoch": 0.69, "grad_norm": 1.30226731300354, "learning_rate": 2.2839871893118175e-06, "loss": 0.7997, "step": 5283 }, { "epoch": 0.69, "grad_norm": 0.60770183801651, "learning_rate": 2.282205694960897e-06, "loss": 0.9936, "step": 5284 }, { "epoch": 0.69, "grad_norm": 1.389857292175293, "learning_rate": 2.280424690183768e-06, "loss": 0.7862, "step": 5285 }, { "epoch": 0.69, "grad_norm": 1.3250150680541992, "learning_rate": 2.2786441753012565e-06, "loss": 0.7749, "step": 5286 }, { "epoch": 0.69, "grad_norm": 1.2741087675094604, "learning_rate": 2.2768641506340977e-06, "loss": 0.8401, "step": 5287 }, { "epoch": 0.69, "grad_norm": 1.3912265300750732, "learning_rate": 2.275084616502936e-06, "loss": 0.7569, "step": 5288 }, { "epoch": 0.69, "grad_norm": 1.3175275325775146, "learning_rate": 2.2733055732283356e-06, "loss": 0.7897, "step": 5289 }, { "epoch": 0.69, "grad_norm": 0.6002609133720398, "learning_rate": 2.271527021130762e-06, "loss": 1.0012, "step": 5290 }, { "epoch": 0.69, "grad_norm": 1.5885951519012451, "learning_rate": 2.2697489605306007e-06, "loss": 0.8618, "step": 5291 }, { "epoch": 0.69, "grad_norm": 1.3668084144592285, "learning_rate": 2.2679713917481438e-06, "loss": 0.7858, "step": 5292 }, { "epoch": 0.69, "grad_norm": 1.3833026885986328, "learning_rate": 2.266194315103595e-06, "loss": 0.7879, "step": 5293 }, { "epoch": 0.69, "grad_norm": 1.2063846588134766, "learning_rate": 2.2644177309170733e-06, "loss": 0.7438, "step": 5294 }, { "epoch": 0.69, "grad_norm": 1.2645909786224365, "learning_rate": 2.262641639508603e-06, "loss": 0.7313, "step": 5295 }, { "epoch": 0.69, "grad_norm": 0.5945859551429749, "learning_rate": 2.260866041198127e-06, "loss": 1.0173, "step": 5296 }, { "epoch": 0.69, "grad_norm": 1.3262280225753784, "learning_rate": 2.259090936305493e-06, "loss": 0.8441, "step": 5297 }, { "epoch": 0.69, "grad_norm": 1.3387279510498047, "learning_rate": 2.25731632515046e-06, "loss": 0.8673, "step": 5298 }, { "epoch": 0.69, "grad_norm": 0.596829891204834, "learning_rate": 2.255542208052704e-06, "loss": 1.0111, "step": 5299 }, { "epoch": 0.69, "grad_norm": 1.3735532760620117, "learning_rate": 2.2537685853318046e-06, "loss": 0.8177, "step": 5300 }, { "epoch": 0.69, "grad_norm": 1.3213001489639282, "learning_rate": 2.2519954573072596e-06, "loss": 0.8083, "step": 5301 }, { "epoch": 0.69, "grad_norm": 1.3846781253814697, "learning_rate": 2.250222824298472e-06, "loss": 0.816, "step": 5302 }, { "epoch": 0.69, "grad_norm": 1.2739760875701904, "learning_rate": 2.2484506866247556e-06, "loss": 0.7996, "step": 5303 }, { "epoch": 0.69, "grad_norm": 1.5751457214355469, "learning_rate": 2.2466790446053412e-06, "loss": 0.8037, "step": 5304 }, { "epoch": 0.7, "grad_norm": 1.319508671760559, "learning_rate": 2.2449078985593614e-06, "loss": 0.7723, "step": 5305 }, { "epoch": 0.7, "grad_norm": 1.5292177200317383, "learning_rate": 2.243137248805868e-06, "loss": 0.8597, "step": 5306 }, { "epoch": 0.7, "grad_norm": 1.312096118927002, "learning_rate": 2.2413670956638193e-06, "loss": 0.8628, "step": 5307 }, { "epoch": 0.7, "grad_norm": 0.6172523498535156, "learning_rate": 2.239597439452081e-06, "loss": 1.0154, "step": 5308 }, { "epoch": 0.7, "grad_norm": 1.4044897556304932, "learning_rate": 2.237828280489437e-06, "loss": 0.8298, "step": 5309 }, { "epoch": 0.7, "grad_norm": 1.5189130306243896, "learning_rate": 2.2360596190945727e-06, "loss": 0.8015, "step": 5310 }, { "epoch": 0.7, "grad_norm": 1.4030284881591797, "learning_rate": 2.2342914555860934e-06, "loss": 0.7917, "step": 5311 }, { "epoch": 0.7, "grad_norm": 0.6002864241600037, "learning_rate": 2.232523790282507e-06, "loss": 1.0081, "step": 5312 }, { "epoch": 0.7, "grad_norm": 0.5811753273010254, "learning_rate": 2.230756623502233e-06, "loss": 1.0099, "step": 5313 }, { "epoch": 0.7, "grad_norm": 1.3407487869262695, "learning_rate": 2.2289899555636062e-06, "loss": 0.8021, "step": 5314 }, { "epoch": 0.7, "grad_norm": 1.3531041145324707, "learning_rate": 2.2272237867848664e-06, "loss": 0.8059, "step": 5315 }, { "epoch": 0.7, "grad_norm": 1.358747959136963, "learning_rate": 2.225458117484162e-06, "loss": 0.7803, "step": 5316 }, { "epoch": 0.7, "grad_norm": 1.3327401876449585, "learning_rate": 2.2236929479795592e-06, "loss": 0.7602, "step": 5317 }, { "epoch": 0.7, "grad_norm": 0.581963837146759, "learning_rate": 2.2219282785890273e-06, "loss": 0.9937, "step": 5318 }, { "epoch": 0.7, "grad_norm": 1.3129160404205322, "learning_rate": 2.220164109630447e-06, "loss": 0.7681, "step": 5319 }, { "epoch": 0.7, "grad_norm": 1.3589664697647095, "learning_rate": 2.218400441421608e-06, "loss": 0.8118, "step": 5320 }, { "epoch": 0.7, "grad_norm": 1.3163988590240479, "learning_rate": 2.216637274280215e-06, "loss": 0.7842, "step": 5321 }, { "epoch": 0.7, "grad_norm": 0.5881624817848206, "learning_rate": 2.2148746085238763e-06, "loss": 1.0243, "step": 5322 }, { "epoch": 0.7, "grad_norm": 1.302986979484558, "learning_rate": 2.213112444470111e-06, "loss": 0.7162, "step": 5323 }, { "epoch": 0.7, "grad_norm": 1.3729223012924194, "learning_rate": 2.211350782436353e-06, "loss": 0.8182, "step": 5324 }, { "epoch": 0.7, "grad_norm": 1.6548506021499634, "learning_rate": 2.209589622739938e-06, "loss": 0.7905, "step": 5325 }, { "epoch": 0.7, "grad_norm": 1.7875896692276, "learning_rate": 2.2078289656981163e-06, "loss": 0.8209, "step": 5326 }, { "epoch": 0.7, "grad_norm": 1.5876820087432861, "learning_rate": 2.206068811628047e-06, "loss": 0.8152, "step": 5327 }, { "epoch": 0.7, "grad_norm": 1.3052946329116821, "learning_rate": 2.204309160846797e-06, "loss": 0.803, "step": 5328 }, { "epoch": 0.7, "grad_norm": 1.5579384565353394, "learning_rate": 2.2025500136713453e-06, "loss": 0.742, "step": 5329 }, { "epoch": 0.7, "grad_norm": 0.5905042290687561, "learning_rate": 2.2007913704185774e-06, "loss": 0.9786, "step": 5330 }, { "epoch": 0.7, "grad_norm": 1.3294352293014526, "learning_rate": 2.1990332314052866e-06, "loss": 0.7358, "step": 5331 }, { "epoch": 0.7, "grad_norm": 1.4134302139282227, "learning_rate": 2.1972755969481823e-06, "loss": 0.7977, "step": 5332 }, { "epoch": 0.7, "grad_norm": 1.3460352420806885, "learning_rate": 2.1955184673638745e-06, "loss": 0.8169, "step": 5333 }, { "epoch": 0.7, "grad_norm": 1.309977650642395, "learning_rate": 2.1937618429688896e-06, "loss": 0.8192, "step": 5334 }, { "epoch": 0.7, "grad_norm": 1.2493188381195068, "learning_rate": 2.1920057240796583e-06, "loss": 0.7207, "step": 5335 }, { "epoch": 0.7, "grad_norm": 1.3144419193267822, "learning_rate": 2.1902501110125196e-06, "loss": 0.7175, "step": 5336 }, { "epoch": 0.7, "grad_norm": 1.3549994230270386, "learning_rate": 2.188495004083727e-06, "loss": 0.7905, "step": 5337 }, { "epoch": 0.7, "grad_norm": 1.3118253946304321, "learning_rate": 2.186740403609436e-06, "loss": 0.7764, "step": 5338 }, { "epoch": 0.7, "grad_norm": 1.349724531173706, "learning_rate": 2.1849863099057167e-06, "loss": 0.8138, "step": 5339 }, { "epoch": 0.7, "grad_norm": 1.2790251970291138, "learning_rate": 2.1832327232885442e-06, "loss": 0.8584, "step": 5340 }, { "epoch": 0.7, "grad_norm": 1.22278892993927, "learning_rate": 2.1814796440738016e-06, "loss": 0.7785, "step": 5341 }, { "epoch": 0.7, "grad_norm": 1.3943015336990356, "learning_rate": 2.1797270725772856e-06, "loss": 0.7838, "step": 5342 }, { "epoch": 0.7, "grad_norm": 1.42366361618042, "learning_rate": 2.177975009114695e-06, "loss": 0.8473, "step": 5343 }, { "epoch": 0.7, "grad_norm": 1.2672208547592163, "learning_rate": 2.1762234540016434e-06, "loss": 0.8244, "step": 5344 }, { "epoch": 0.7, "grad_norm": 1.2596771717071533, "learning_rate": 2.1744724075536488e-06, "loss": 0.8139, "step": 5345 }, { "epoch": 0.7, "grad_norm": 0.6045711040496826, "learning_rate": 2.172721870086136e-06, "loss": 1.0047, "step": 5346 }, { "epoch": 0.7, "grad_norm": 1.5074121952056885, "learning_rate": 2.170971841914444e-06, "loss": 0.8424, "step": 5347 }, { "epoch": 0.7, "grad_norm": 1.2746787071228027, "learning_rate": 2.169222323353815e-06, "loss": 0.8185, "step": 5348 }, { "epoch": 0.7, "grad_norm": 1.3454893827438354, "learning_rate": 2.1674733147193995e-06, "loss": 0.7311, "step": 5349 }, { "epoch": 0.7, "grad_norm": 0.5809167623519897, "learning_rate": 2.1657248163262607e-06, "loss": 1.0291, "step": 5350 }, { "epoch": 0.7, "grad_norm": 1.7259368896484375, "learning_rate": 2.163976828489364e-06, "loss": 0.8333, "step": 5351 }, { "epoch": 0.7, "grad_norm": 1.3238763809204102, "learning_rate": 2.1622293515235893e-06, "loss": 0.8106, "step": 5352 }, { "epoch": 0.7, "grad_norm": 1.334159255027771, "learning_rate": 2.1604823857437185e-06, "loss": 0.7628, "step": 5353 }, { "epoch": 0.7, "grad_norm": 0.5787563323974609, "learning_rate": 2.1587359314644423e-06, "loss": 1.0109, "step": 5354 }, { "epoch": 0.7, "grad_norm": 1.3978673219680786, "learning_rate": 2.156989989000364e-06, "loss": 0.8238, "step": 5355 }, { "epoch": 0.7, "grad_norm": 1.3915470838546753, "learning_rate": 2.1552445586659904e-06, "loss": 0.7736, "step": 5356 }, { "epoch": 0.7, "grad_norm": 1.3105334043502808, "learning_rate": 2.1534996407757354e-06, "loss": 0.7734, "step": 5357 }, { "epoch": 0.7, "grad_norm": 1.3526208400726318, "learning_rate": 2.151755235643925e-06, "loss": 0.8123, "step": 5358 }, { "epoch": 0.7, "grad_norm": 1.3288912773132324, "learning_rate": 2.150011343584788e-06, "loss": 0.7972, "step": 5359 }, { "epoch": 0.7, "grad_norm": 1.3366230726242065, "learning_rate": 2.148267964912464e-06, "loss": 0.8078, "step": 5360 }, { "epoch": 0.7, "grad_norm": 1.2055765390396118, "learning_rate": 2.1465250999409975e-06, "loss": 0.7667, "step": 5361 }, { "epoch": 0.7, "grad_norm": 1.263532042503357, "learning_rate": 2.144782748984345e-06, "loss": 0.7716, "step": 5362 }, { "epoch": 0.7, "grad_norm": 1.810561180114746, "learning_rate": 2.1430409123563652e-06, "loss": 0.7578, "step": 5363 }, { "epoch": 0.7, "grad_norm": 1.3129507303237915, "learning_rate": 2.141299590370825e-06, "loss": 0.7683, "step": 5364 }, { "epoch": 0.7, "grad_norm": 1.5654371976852417, "learning_rate": 2.1395587833414033e-06, "loss": 0.7817, "step": 5365 }, { "epoch": 0.7, "grad_norm": 1.3651907444000244, "learning_rate": 2.1378184915816795e-06, "loss": 0.7576, "step": 5366 }, { "epoch": 0.7, "grad_norm": 1.4385913610458374, "learning_rate": 2.136078715405147e-06, "loss": 0.8504, "step": 5367 }, { "epoch": 0.7, "grad_norm": 0.6318222284317017, "learning_rate": 2.1343394551252018e-06, "loss": 1.0294, "step": 5368 }, { "epoch": 0.7, "grad_norm": 2.5997908115386963, "learning_rate": 2.132600711055145e-06, "loss": 0.8236, "step": 5369 }, { "epoch": 0.7, "grad_norm": 1.6201547384262085, "learning_rate": 2.1308624835081924e-06, "loss": 0.7576, "step": 5370 }, { "epoch": 0.7, "grad_norm": 0.6136257648468018, "learning_rate": 2.129124772797458e-06, "loss": 1.0135, "step": 5371 }, { "epoch": 0.7, "grad_norm": 1.2696644067764282, "learning_rate": 2.1273875792359704e-06, "loss": 0.8222, "step": 5372 }, { "epoch": 0.7, "grad_norm": 1.3474565744400024, "learning_rate": 2.12565090313666e-06, "loss": 0.7521, "step": 5373 }, { "epoch": 0.7, "grad_norm": 1.4685442447662354, "learning_rate": 2.1239147448123637e-06, "loss": 0.7707, "step": 5374 }, { "epoch": 0.7, "grad_norm": 1.2622227668762207, "learning_rate": 2.1221791045758306e-06, "loss": 0.8165, "step": 5375 }, { "epoch": 0.7, "grad_norm": 1.6976429224014282, "learning_rate": 2.12044398273971e-06, "loss": 0.8267, "step": 5376 }, { "epoch": 0.7, "grad_norm": 1.3482861518859863, "learning_rate": 2.118709379616559e-06, "loss": 0.7529, "step": 5377 }, { "epoch": 0.7, "grad_norm": 1.465763807296753, "learning_rate": 2.1169752955188476e-06, "loss": 0.7864, "step": 5378 }, { "epoch": 0.7, "grad_norm": 1.3030273914337158, "learning_rate": 2.115241730758943e-06, "loss": 0.8128, "step": 5379 }, { "epoch": 0.7, "grad_norm": 1.2471928596496582, "learning_rate": 2.1135086856491273e-06, "loss": 0.8526, "step": 5380 }, { "epoch": 0.71, "grad_norm": 1.3727768659591675, "learning_rate": 2.111776160501583e-06, "loss": 0.7968, "step": 5381 }, { "epoch": 0.71, "grad_norm": 1.36199152469635, "learning_rate": 2.1100441556283992e-06, "loss": 0.8105, "step": 5382 }, { "epoch": 0.71, "grad_norm": 1.3875483274459839, "learning_rate": 2.1083126713415767e-06, "loss": 0.8114, "step": 5383 }, { "epoch": 0.71, "grad_norm": 1.2746984958648682, "learning_rate": 2.106581707953016e-06, "loss": 0.7601, "step": 5384 }, { "epoch": 0.71, "grad_norm": 1.2513339519500732, "learning_rate": 2.104851265774529e-06, "loss": 0.7856, "step": 5385 }, { "epoch": 0.71, "grad_norm": 4.816081523895264, "learning_rate": 2.1031213451178312e-06, "loss": 0.7457, "step": 5386 }, { "epoch": 0.71, "grad_norm": 0.6136434078216553, "learning_rate": 2.101391946294541e-06, "loss": 0.9983, "step": 5387 }, { "epoch": 0.71, "grad_norm": 0.6169759631156921, "learning_rate": 2.099663069616191e-06, "loss": 1.0154, "step": 5388 }, { "epoch": 0.71, "grad_norm": 1.4387648105621338, "learning_rate": 2.0979347153942105e-06, "loss": 0.8474, "step": 5389 }, { "epoch": 0.71, "grad_norm": 1.3481119871139526, "learning_rate": 2.0962068839399423e-06, "loss": 0.8119, "step": 5390 }, { "epoch": 0.71, "grad_norm": 0.58546382188797, "learning_rate": 2.0944795755646315e-06, "loss": 1.0206, "step": 5391 }, { "epoch": 0.71, "grad_norm": 2.357866048812866, "learning_rate": 2.0927527905794258e-06, "loss": 0.7769, "step": 5392 }, { "epoch": 0.71, "grad_norm": 0.5810564160346985, "learning_rate": 2.091026529295387e-06, "loss": 0.9932, "step": 5393 }, { "epoch": 0.71, "grad_norm": 1.3419708013534546, "learning_rate": 2.089300792023473e-06, "loss": 0.7793, "step": 5394 }, { "epoch": 0.71, "grad_norm": 1.2688156366348267, "learning_rate": 2.0875755790745573e-06, "loss": 0.777, "step": 5395 }, { "epoch": 0.71, "grad_norm": 0.5734976530075073, "learning_rate": 2.0858508907594092e-06, "loss": 1.0179, "step": 5396 }, { "epoch": 0.71, "grad_norm": 1.414757490158081, "learning_rate": 2.0841267273887108e-06, "loss": 0.7719, "step": 5397 }, { "epoch": 0.71, "grad_norm": 1.4980695247650146, "learning_rate": 2.082403089273042e-06, "loss": 0.7684, "step": 5398 }, { "epoch": 0.71, "grad_norm": 0.5795096158981323, "learning_rate": 2.080679976722899e-06, "loss": 1.0077, "step": 5399 }, { "epoch": 0.71, "grad_norm": 0.6086354851722717, "learning_rate": 2.078957390048674e-06, "loss": 0.9998, "step": 5400 }, { "epoch": 0.71, "grad_norm": 0.5932402610778809, "learning_rate": 2.0772353295606684e-06, "loss": 1.0112, "step": 5401 }, { "epoch": 0.71, "grad_norm": 1.2642900943756104, "learning_rate": 2.075513795569085e-06, "loss": 0.8149, "step": 5402 }, { "epoch": 0.71, "grad_norm": 1.439705491065979, "learning_rate": 2.0737927883840397e-06, "loss": 0.7849, "step": 5403 }, { "epoch": 0.71, "grad_norm": 1.3058404922485352, "learning_rate": 2.0720723083155464e-06, "loss": 0.7975, "step": 5404 }, { "epoch": 0.71, "grad_norm": 1.44630765914917, "learning_rate": 2.070352355673524e-06, "loss": 0.8005, "step": 5405 }, { "epoch": 0.71, "grad_norm": 1.5993390083312988, "learning_rate": 2.0686329307678025e-06, "loss": 0.8083, "step": 5406 }, { "epoch": 0.71, "grad_norm": 0.6196556091308594, "learning_rate": 2.06691403390811e-06, "loss": 0.9986, "step": 5407 }, { "epoch": 0.71, "grad_norm": 1.37267005443573, "learning_rate": 2.0651956654040855e-06, "loss": 0.8358, "step": 5408 }, { "epoch": 0.71, "grad_norm": 1.2816277742385864, "learning_rate": 2.0634778255652676e-06, "loss": 0.7678, "step": 5409 }, { "epoch": 0.71, "grad_norm": 1.4121479988098145, "learning_rate": 2.061760514701101e-06, "loss": 0.7616, "step": 5410 }, { "epoch": 0.71, "grad_norm": 1.4282406568527222, "learning_rate": 2.060043733120939e-06, "loss": 0.8268, "step": 5411 }, { "epoch": 0.71, "grad_norm": 0.6113808155059814, "learning_rate": 2.0583274811340327e-06, "loss": 1.0079, "step": 5412 }, { "epoch": 0.71, "grad_norm": 0.5832378268241882, "learning_rate": 2.0566117590495454e-06, "loss": 0.9902, "step": 5413 }, { "epoch": 0.71, "grad_norm": 1.3599674701690674, "learning_rate": 2.05489656717654e-06, "loss": 0.7951, "step": 5414 }, { "epoch": 0.71, "grad_norm": 1.2877154350280762, "learning_rate": 2.053181905823982e-06, "loss": 0.7644, "step": 5415 }, { "epoch": 0.71, "grad_norm": 1.483573317527771, "learning_rate": 2.051467775300749e-06, "loss": 0.8269, "step": 5416 }, { "epoch": 0.71, "grad_norm": 1.2778923511505127, "learning_rate": 2.049754175915615e-06, "loss": 0.7802, "step": 5417 }, { "epoch": 0.71, "grad_norm": 1.4663991928100586, "learning_rate": 2.0480411079772645e-06, "loss": 0.8116, "step": 5418 }, { "epoch": 0.71, "grad_norm": 2.4945828914642334, "learning_rate": 2.0463285717942816e-06, "loss": 0.8119, "step": 5419 }, { "epoch": 0.71, "grad_norm": 1.8748928308486938, "learning_rate": 2.044616567675155e-06, "loss": 0.8122, "step": 5420 }, { "epoch": 0.71, "grad_norm": 0.622241735458374, "learning_rate": 2.0429050959282832e-06, "loss": 1.0002, "step": 5421 }, { "epoch": 0.71, "grad_norm": 1.3508284091949463, "learning_rate": 2.0411941568619602e-06, "loss": 0.7616, "step": 5422 }, { "epoch": 0.71, "grad_norm": 1.3316978216171265, "learning_rate": 2.039483750784392e-06, "loss": 0.8328, "step": 5423 }, { "epoch": 0.71, "grad_norm": 1.2829504013061523, "learning_rate": 2.0377738780036843e-06, "loss": 0.7645, "step": 5424 }, { "epoch": 0.71, "grad_norm": 1.3329179286956787, "learning_rate": 2.036064538827845e-06, "loss": 0.7123, "step": 5425 }, { "epoch": 0.71, "grad_norm": 1.3582043647766113, "learning_rate": 2.0343557335647916e-06, "loss": 0.7896, "step": 5426 }, { "epoch": 0.71, "grad_norm": 1.461238145828247, "learning_rate": 2.032647462522339e-06, "loss": 0.8255, "step": 5427 }, { "epoch": 0.71, "grad_norm": 1.4373283386230469, "learning_rate": 2.030939726008213e-06, "loss": 0.8576, "step": 5428 }, { "epoch": 0.71, "grad_norm": 1.3217517137527466, "learning_rate": 2.029232524330037e-06, "loss": 0.8148, "step": 5429 }, { "epoch": 0.71, "grad_norm": 1.2813374996185303, "learning_rate": 2.0275258577953383e-06, "loss": 0.782, "step": 5430 }, { "epoch": 0.71, "grad_norm": 1.3704835176467896, "learning_rate": 2.0258197267115538e-06, "loss": 0.8095, "step": 5431 }, { "epoch": 0.71, "grad_norm": 1.3185316324234009, "learning_rate": 2.024114131386017e-06, "loss": 0.8149, "step": 5432 }, { "epoch": 0.71, "grad_norm": 1.3932374715805054, "learning_rate": 2.0224090721259675e-06, "loss": 0.7949, "step": 5433 }, { "epoch": 0.71, "grad_norm": 1.4147619009017944, "learning_rate": 2.0207045492385514e-06, "loss": 0.8002, "step": 5434 }, { "epoch": 0.71, "grad_norm": 0.6382246613502502, "learning_rate": 2.0190005630308113e-06, "loss": 1.0043, "step": 5435 }, { "epoch": 0.71, "grad_norm": 1.2652771472930908, "learning_rate": 2.017297113809702e-06, "loss": 0.7988, "step": 5436 }, { "epoch": 0.71, "grad_norm": 1.3199372291564941, "learning_rate": 2.015594201882074e-06, "loss": 0.7856, "step": 5437 }, { "epoch": 0.71, "grad_norm": 1.4433226585388184, "learning_rate": 2.013891827554684e-06, "loss": 0.8402, "step": 5438 }, { "epoch": 0.71, "grad_norm": 3.1753740310668945, "learning_rate": 2.01218999113419e-06, "loss": 0.8567, "step": 5439 }, { "epoch": 0.71, "grad_norm": 1.370019555091858, "learning_rate": 2.010488692927159e-06, "loss": 0.8025, "step": 5440 }, { "epoch": 0.71, "grad_norm": 0.5906922221183777, "learning_rate": 2.0087879332400533e-06, "loss": 1.0036, "step": 5441 }, { "epoch": 0.71, "grad_norm": 0.5953747630119324, "learning_rate": 2.007087712379242e-06, "loss": 1.0308, "step": 5442 }, { "epoch": 0.71, "grad_norm": 0.6074784398078918, "learning_rate": 2.005388030650999e-06, "loss": 1.018, "step": 5443 }, { "epoch": 0.71, "grad_norm": 1.3673667907714844, "learning_rate": 2.003688888361497e-06, "loss": 0.7822, "step": 5444 }, { "epoch": 0.71, "grad_norm": 0.597752571105957, "learning_rate": 2.001990285816813e-06, "loss": 1.0382, "step": 5445 }, { "epoch": 0.71, "grad_norm": 1.2741578817367554, "learning_rate": 2.0002922233229288e-06, "loss": 0.8422, "step": 5446 }, { "epoch": 0.71, "grad_norm": 1.367706298828125, "learning_rate": 1.9985947011857276e-06, "loss": 0.8074, "step": 5447 }, { "epoch": 0.71, "grad_norm": 1.3298181295394897, "learning_rate": 1.9968977197109916e-06, "loss": 0.7687, "step": 5448 }, { "epoch": 0.71, "grad_norm": 0.6212354302406311, "learning_rate": 1.995201279204414e-06, "loss": 0.9752, "step": 5449 }, { "epoch": 0.71, "grad_norm": 0.6065380573272705, "learning_rate": 1.993505379971581e-06, "loss": 1.0084, "step": 5450 }, { "epoch": 0.71, "grad_norm": 1.741956114768982, "learning_rate": 1.99181002231799e-06, "loss": 0.8289, "step": 5451 }, { "epoch": 0.71, "grad_norm": 1.301721215248108, "learning_rate": 1.9901152065490337e-06, "loss": 0.7213, "step": 5452 }, { "epoch": 0.71, "grad_norm": 1.3139214515686035, "learning_rate": 1.98842093297001e-06, "loss": 0.7708, "step": 5453 }, { "epoch": 0.71, "grad_norm": 1.2576422691345215, "learning_rate": 1.986727201886122e-06, "loss": 0.8015, "step": 5454 }, { "epoch": 0.71, "grad_norm": 0.5957777500152588, "learning_rate": 1.9850340136024688e-06, "loss": 1.0165, "step": 5455 }, { "epoch": 0.71, "grad_norm": 0.5886528491973877, "learning_rate": 1.9833413684240593e-06, "loss": 1.006, "step": 5456 }, { "epoch": 0.72, "grad_norm": 0.5935488939285278, "learning_rate": 1.9816492666557983e-06, "loss": 1.0048, "step": 5457 }, { "epoch": 0.72, "grad_norm": 0.5952271223068237, "learning_rate": 1.979957708602494e-06, "loss": 1.0178, "step": 5458 }, { "epoch": 0.72, "grad_norm": 1.2714720964431763, "learning_rate": 1.97826669456886e-06, "loss": 0.8067, "step": 5459 }, { "epoch": 0.72, "grad_norm": 1.5459668636322021, "learning_rate": 1.976576224859509e-06, "loss": 0.8169, "step": 5460 }, { "epoch": 0.72, "grad_norm": 0.569124162197113, "learning_rate": 1.974886299778954e-06, "loss": 0.9978, "step": 5461 }, { "epoch": 0.72, "grad_norm": 1.375249981880188, "learning_rate": 1.9731969196316146e-06, "loss": 0.8192, "step": 5462 }, { "epoch": 0.72, "grad_norm": 1.2460439205169678, "learning_rate": 1.971508084721808e-06, "loss": 0.7404, "step": 5463 }, { "epoch": 0.72, "grad_norm": 1.316490888595581, "learning_rate": 1.9698197953537566e-06, "loss": 0.8505, "step": 5464 }, { "epoch": 0.72, "grad_norm": 1.3677533864974976, "learning_rate": 1.968132051831583e-06, "loss": 0.7793, "step": 5465 }, { "epoch": 0.72, "grad_norm": 1.6151691675186157, "learning_rate": 1.9664448544593072e-06, "loss": 0.7921, "step": 5466 }, { "epoch": 0.72, "grad_norm": 1.5128692388534546, "learning_rate": 1.964758203540861e-06, "loss": 0.7723, "step": 5467 }, { "epoch": 0.72, "grad_norm": 3.4765877723693848, "learning_rate": 1.9630720993800667e-06, "loss": 0.8362, "step": 5468 }, { "epoch": 0.72, "grad_norm": 1.4413764476776123, "learning_rate": 1.961386542280657e-06, "loss": 0.7399, "step": 5469 }, { "epoch": 0.72, "grad_norm": 0.6186902523040771, "learning_rate": 1.9597015325462597e-06, "loss": 1.0127, "step": 5470 }, { "epoch": 0.72, "grad_norm": 0.6585413813591003, "learning_rate": 1.9580170704804058e-06, "loss": 1.0313, "step": 5471 }, { "epoch": 0.72, "grad_norm": 1.3860814571380615, "learning_rate": 1.956333156386531e-06, "loss": 0.814, "step": 5472 }, { "epoch": 0.72, "grad_norm": 1.3070236444473267, "learning_rate": 1.954649790567967e-06, "loss": 0.7404, "step": 5473 }, { "epoch": 0.72, "grad_norm": 0.6066024303436279, "learning_rate": 1.952966973327952e-06, "loss": 1.0163, "step": 5474 }, { "epoch": 0.72, "grad_norm": 1.3673336505889893, "learning_rate": 1.951284704969621e-06, "loss": 0.8019, "step": 5475 }, { "epoch": 0.72, "grad_norm": 0.5863862037658691, "learning_rate": 1.9496029857960103e-06, "loss": 0.9976, "step": 5476 }, { "epoch": 0.72, "grad_norm": 1.292531132698059, "learning_rate": 1.947921816110062e-06, "loss": 0.7772, "step": 5477 }, { "epoch": 0.72, "grad_norm": 1.2856606245040894, "learning_rate": 1.946241196214615e-06, "loss": 0.7918, "step": 5478 }, { "epoch": 0.72, "grad_norm": 0.6085711121559143, "learning_rate": 1.944561126412409e-06, "loss": 1.0125, "step": 5479 }, { "epoch": 0.72, "grad_norm": 0.5853901505470276, "learning_rate": 1.942881607006086e-06, "loss": 0.9829, "step": 5480 }, { "epoch": 0.72, "grad_norm": 0.5852105617523193, "learning_rate": 1.9412026382981897e-06, "loss": 1.0059, "step": 5481 }, { "epoch": 0.72, "grad_norm": 0.5862083435058594, "learning_rate": 1.9395242205911634e-06, "loss": 1.0054, "step": 5482 }, { "epoch": 0.72, "grad_norm": 1.33817458152771, "learning_rate": 1.9378463541873494e-06, "loss": 0.8021, "step": 5483 }, { "epoch": 0.72, "grad_norm": 1.4130566120147705, "learning_rate": 1.936169039388996e-06, "loss": 0.792, "step": 5484 }, { "epoch": 0.72, "grad_norm": 0.5940247178077698, "learning_rate": 1.934492276498246e-06, "loss": 1.0137, "step": 5485 }, { "epoch": 0.72, "grad_norm": 1.3831619024276733, "learning_rate": 1.932816065817145e-06, "loss": 0.7888, "step": 5486 }, { "epoch": 0.72, "grad_norm": 1.3079365491867065, "learning_rate": 1.9311404076476433e-06, "loss": 0.8052, "step": 5487 }, { "epoch": 0.72, "grad_norm": 0.619125485420227, "learning_rate": 1.929465302291583e-06, "loss": 1.0285, "step": 5488 }, { "epoch": 0.72, "grad_norm": 1.2766584157943726, "learning_rate": 1.9277907500507166e-06, "loss": 0.7757, "step": 5489 }, { "epoch": 0.72, "grad_norm": 1.3515067100524902, "learning_rate": 1.9261167512266893e-06, "loss": 0.7687, "step": 5490 }, { "epoch": 0.72, "grad_norm": 1.9091755151748657, "learning_rate": 1.9244433061210477e-06, "loss": 0.7615, "step": 5491 }, { "epoch": 0.72, "grad_norm": 1.3967093229293823, "learning_rate": 1.9227704150352437e-06, "loss": 0.8019, "step": 5492 }, { "epoch": 0.72, "grad_norm": 0.5718570351600647, "learning_rate": 1.921098078270625e-06, "loss": 1.0241, "step": 5493 }, { "epoch": 0.72, "grad_norm": 1.3479472398757935, "learning_rate": 1.919426296128438e-06, "loss": 0.8076, "step": 5494 }, { "epoch": 0.72, "grad_norm": 1.481327772140503, "learning_rate": 1.917755068909835e-06, "loss": 0.8493, "step": 5495 }, { "epoch": 0.72, "grad_norm": 0.5868615508079529, "learning_rate": 1.9160843969158615e-06, "loss": 1.0329, "step": 5496 }, { "epoch": 0.72, "grad_norm": 1.3049540519714355, "learning_rate": 1.91441428044747e-06, "loss": 0.8245, "step": 5497 }, { "epoch": 0.72, "grad_norm": 1.381830096244812, "learning_rate": 1.9127447198055076e-06, "loss": 0.7624, "step": 5498 }, { "epoch": 0.72, "grad_norm": 0.5925220251083374, "learning_rate": 1.9110757152907215e-06, "loss": 1.0068, "step": 5499 }, { "epoch": 0.72, "grad_norm": 1.7464269399642944, "learning_rate": 1.909407267203763e-06, "loss": 0.8037, "step": 5500 }, { "epoch": 0.72, "grad_norm": 1.262747049331665, "learning_rate": 1.9077393758451777e-06, "loss": 0.8329, "step": 5501 }, { "epoch": 0.72, "grad_norm": 1.2337957620620728, "learning_rate": 1.9060720415154165e-06, "loss": 0.8227, "step": 5502 }, { "epoch": 0.72, "grad_norm": 1.3761627674102783, "learning_rate": 1.904405264514826e-06, "loss": 0.8277, "step": 5503 }, { "epoch": 0.72, "grad_norm": 1.2690550088882446, "learning_rate": 1.9027390451436505e-06, "loss": 0.7799, "step": 5504 }, { "epoch": 0.72, "grad_norm": 1.3013635873794556, "learning_rate": 1.9010733837020417e-06, "loss": 0.7917, "step": 5505 }, { "epoch": 0.72, "grad_norm": 1.274188756942749, "learning_rate": 1.8994082804900422e-06, "loss": 0.7748, "step": 5506 }, { "epoch": 0.72, "grad_norm": 0.587648332118988, "learning_rate": 1.8977437358076002e-06, "loss": 1.0333, "step": 5507 }, { "epoch": 0.72, "grad_norm": 1.552088975906372, "learning_rate": 1.8960797499545602e-06, "loss": 0.802, "step": 5508 }, { "epoch": 0.72, "grad_norm": 1.2887197732925415, "learning_rate": 1.8944163232306645e-06, "loss": 0.8014, "step": 5509 }, { "epoch": 0.72, "grad_norm": 1.2629165649414062, "learning_rate": 1.8927534559355603e-06, "loss": 0.828, "step": 5510 }, { "epoch": 0.72, "grad_norm": 1.35712468624115, "learning_rate": 1.8910911483687872e-06, "loss": 0.8089, "step": 5511 }, { "epoch": 0.72, "grad_norm": 0.6041145324707031, "learning_rate": 1.889429400829791e-06, "loss": 1.0134, "step": 5512 }, { "epoch": 0.72, "grad_norm": 0.5926774740219116, "learning_rate": 1.8877682136179115e-06, "loss": 1.0095, "step": 5513 }, { "epoch": 0.72, "grad_norm": 1.352151870727539, "learning_rate": 1.8861075870323864e-06, "loss": 0.7631, "step": 5514 }, { "epoch": 0.72, "grad_norm": 1.4510918855667114, "learning_rate": 1.8844475213723595e-06, "loss": 0.82, "step": 5515 }, { "epoch": 0.72, "grad_norm": 1.2858302593231201, "learning_rate": 1.8827880169368667e-06, "loss": 0.8037, "step": 5516 }, { "epoch": 0.72, "grad_norm": 1.2782431840896606, "learning_rate": 1.881129074024844e-06, "loss": 0.75, "step": 5517 }, { "epoch": 0.72, "grad_norm": 1.296211838722229, "learning_rate": 1.8794706929351303e-06, "loss": 0.7873, "step": 5518 }, { "epoch": 0.72, "grad_norm": 1.357788324356079, "learning_rate": 1.8778128739664591e-06, "loss": 0.786, "step": 5519 }, { "epoch": 0.72, "grad_norm": 1.2873504161834717, "learning_rate": 1.8761556174174644e-06, "loss": 0.7847, "step": 5520 }, { "epoch": 0.72, "grad_norm": 1.439873218536377, "learning_rate": 1.8744989235866766e-06, "loss": 0.8066, "step": 5521 }, { "epoch": 0.72, "grad_norm": 1.2689391374588013, "learning_rate": 1.8728427927725295e-06, "loss": 0.8478, "step": 5522 }, { "epoch": 0.72, "grad_norm": 1.3148393630981445, "learning_rate": 1.8711872252733514e-06, "loss": 0.7791, "step": 5523 }, { "epoch": 0.72, "grad_norm": 1.3280106782913208, "learning_rate": 1.8695322213873684e-06, "loss": 0.8119, "step": 5524 }, { "epoch": 0.72, "grad_norm": 0.6198797225952148, "learning_rate": 1.8678777814127108e-06, "loss": 1.0122, "step": 5525 }, { "epoch": 0.72, "grad_norm": 1.345835566520691, "learning_rate": 1.8662239056474013e-06, "loss": 0.7795, "step": 5526 }, { "epoch": 0.72, "grad_norm": 0.6112748384475708, "learning_rate": 1.8645705943893615e-06, "loss": 1.0007, "step": 5527 }, { "epoch": 0.72, "grad_norm": 0.5965059995651245, "learning_rate": 1.8629178479364174e-06, "loss": 0.9883, "step": 5528 }, { "epoch": 0.72, "grad_norm": 1.3463702201843262, "learning_rate": 1.8612656665862843e-06, "loss": 0.8094, "step": 5529 }, { "epoch": 0.72, "grad_norm": 1.4365036487579346, "learning_rate": 1.8596140506365839e-06, "loss": 0.8151, "step": 5530 }, { "epoch": 0.72, "grad_norm": 0.5684584975242615, "learning_rate": 1.8579630003848309e-06, "loss": 0.9916, "step": 5531 }, { "epoch": 0.72, "grad_norm": 0.5888611078262329, "learning_rate": 1.8563125161284384e-06, "loss": 1.0021, "step": 5532 }, { "epoch": 0.73, "grad_norm": 1.3277522325515747, "learning_rate": 1.8546625981647215e-06, "loss": 0.7563, "step": 5533 }, { "epoch": 0.73, "grad_norm": 1.339870572090149, "learning_rate": 1.8530132467908868e-06, "loss": 0.809, "step": 5534 }, { "epoch": 0.73, "grad_norm": 0.603844940662384, "learning_rate": 1.8513644623040472e-06, "loss": 1.0127, "step": 5535 }, { "epoch": 0.73, "grad_norm": 1.4183224439620972, "learning_rate": 1.8497162450012062e-06, "loss": 0.8187, "step": 5536 }, { "epoch": 0.73, "grad_norm": 0.5919128656387329, "learning_rate": 1.8480685951792659e-06, "loss": 0.9958, "step": 5537 }, { "epoch": 0.73, "grad_norm": 1.3322303295135498, "learning_rate": 1.8464215131350327e-06, "loss": 0.8343, "step": 5538 }, { "epoch": 0.73, "grad_norm": 1.6103956699371338, "learning_rate": 1.8447749991652014e-06, "loss": 0.7418, "step": 5539 }, { "epoch": 0.73, "grad_norm": 1.4522744417190552, "learning_rate": 1.8431290535663737e-06, "loss": 0.8162, "step": 5540 }, { "epoch": 0.73, "grad_norm": 1.4307352304458618, "learning_rate": 1.841483676635042e-06, "loss": 0.7936, "step": 5541 }, { "epoch": 0.73, "grad_norm": 1.4448667764663696, "learning_rate": 1.8398388686675965e-06, "loss": 0.8545, "step": 5542 }, { "epoch": 0.73, "grad_norm": 1.3088436126708984, "learning_rate": 1.8381946299603316e-06, "loss": 0.7715, "step": 5543 }, { "epoch": 0.73, "grad_norm": 1.3030961751937866, "learning_rate": 1.83655096080943e-06, "loss": 0.7477, "step": 5544 }, { "epoch": 0.73, "grad_norm": 0.6133267283439636, "learning_rate": 1.8349078615109805e-06, "loss": 1.003, "step": 5545 }, { "epoch": 0.73, "grad_norm": 1.3796601295471191, "learning_rate": 1.8332653323609638e-06, "loss": 0.8223, "step": 5546 }, { "epoch": 0.73, "grad_norm": 1.3822211027145386, "learning_rate": 1.8316233736552568e-06, "loss": 0.7666, "step": 5547 }, { "epoch": 0.73, "grad_norm": 1.427581787109375, "learning_rate": 1.82998198568964e-06, "loss": 0.8322, "step": 5548 }, { "epoch": 0.73, "grad_norm": 1.5166966915130615, "learning_rate": 1.8283411687597851e-06, "loss": 0.7908, "step": 5549 }, { "epoch": 0.73, "grad_norm": 1.2846858501434326, "learning_rate": 1.826700923161262e-06, "loss": 0.8451, "step": 5550 }, { "epoch": 0.73, "grad_norm": 1.3180629014968872, "learning_rate": 1.8250612491895415e-06, "loss": 0.8304, "step": 5551 }, { "epoch": 0.73, "grad_norm": 1.261322021484375, "learning_rate": 1.8234221471399849e-06, "loss": 0.7598, "step": 5552 }, { "epoch": 0.73, "grad_norm": 0.5980848670005798, "learning_rate": 1.8217836173078584e-06, "loss": 1.001, "step": 5553 }, { "epoch": 0.73, "grad_norm": 1.328591227531433, "learning_rate": 1.8201456599883188e-06, "loss": 0.7789, "step": 5554 }, { "epoch": 0.73, "grad_norm": 1.393600583076477, "learning_rate": 1.8185082754764205e-06, "loss": 0.7656, "step": 5555 }, { "epoch": 0.73, "grad_norm": 1.3991601467132568, "learning_rate": 1.8168714640671193e-06, "loss": 0.8514, "step": 5556 }, { "epoch": 0.73, "grad_norm": 1.409458875656128, "learning_rate": 1.815235226055263e-06, "loss": 0.803, "step": 5557 }, { "epoch": 0.73, "grad_norm": 1.2977787256240845, "learning_rate": 1.8135995617355957e-06, "loss": 0.8228, "step": 5558 }, { "epoch": 0.73, "grad_norm": 1.3277878761291504, "learning_rate": 1.811964471402764e-06, "loss": 0.8337, "step": 5559 }, { "epoch": 0.73, "grad_norm": 1.3850107192993164, "learning_rate": 1.8103299553513048e-06, "loss": 0.8186, "step": 5560 }, { "epoch": 0.73, "grad_norm": 1.3808022737503052, "learning_rate": 1.8086960138756531e-06, "loss": 0.7722, "step": 5561 }, { "epoch": 0.73, "grad_norm": 0.602184534072876, "learning_rate": 1.8070626472701446e-06, "loss": 1.0371, "step": 5562 }, { "epoch": 0.73, "grad_norm": 0.6170496344566345, "learning_rate": 1.8054298558290063e-06, "loss": 0.994, "step": 5563 }, { "epoch": 0.73, "grad_norm": 1.3103477954864502, "learning_rate": 1.803797639846363e-06, "loss": 0.7774, "step": 5564 }, { "epoch": 0.73, "grad_norm": 1.2059061527252197, "learning_rate": 1.802165999616235e-06, "loss": 0.7693, "step": 5565 }, { "epoch": 0.73, "grad_norm": 0.5988016128540039, "learning_rate": 1.8005349354325435e-06, "loss": 0.9886, "step": 5566 }, { "epoch": 0.73, "grad_norm": 1.420951008796692, "learning_rate": 1.7989044475891e-06, "loss": 0.7237, "step": 5567 }, { "epoch": 0.73, "grad_norm": 1.4422167539596558, "learning_rate": 1.7972745363796168e-06, "loss": 0.7876, "step": 5568 }, { "epoch": 0.73, "grad_norm": 3.2454519271850586, "learning_rate": 1.795645202097699e-06, "loss": 0.8106, "step": 5569 }, { "epoch": 0.73, "grad_norm": 1.2918957471847534, "learning_rate": 1.794016445036848e-06, "loss": 0.7823, "step": 5570 }, { "epoch": 0.73, "grad_norm": 1.4203861951828003, "learning_rate": 1.792388265490465e-06, "loss": 0.7925, "step": 5571 }, { "epoch": 0.73, "grad_norm": 1.2868614196777344, "learning_rate": 1.7907606637518415e-06, "loss": 0.8482, "step": 5572 }, { "epoch": 0.73, "grad_norm": 1.3954732418060303, "learning_rate": 1.7891336401141712e-06, "loss": 0.7495, "step": 5573 }, { "epoch": 0.73, "grad_norm": 1.3393940925598145, "learning_rate": 1.7875071948705386e-06, "loss": 0.7992, "step": 5574 }, { "epoch": 0.73, "grad_norm": 1.2755155563354492, "learning_rate": 1.7858813283139237e-06, "loss": 0.7831, "step": 5575 }, { "epoch": 0.73, "grad_norm": 1.2724751234054565, "learning_rate": 1.7842560407372083e-06, "loss": 0.7692, "step": 5576 }, { "epoch": 0.73, "grad_norm": 1.3474339246749878, "learning_rate": 1.7826313324331646e-06, "loss": 0.8322, "step": 5577 }, { "epoch": 0.73, "grad_norm": 0.6065584421157837, "learning_rate": 1.7810072036944588e-06, "loss": 1.0081, "step": 5578 }, { "epoch": 0.73, "grad_norm": 1.5325623750686646, "learning_rate": 1.7793836548136595e-06, "loss": 0.7775, "step": 5579 }, { "epoch": 0.73, "grad_norm": 1.336632490158081, "learning_rate": 1.7777606860832242e-06, "loss": 0.8149, "step": 5580 }, { "epoch": 0.73, "grad_norm": 0.5792363286018372, "learning_rate": 1.7761382977955117e-06, "loss": 1.0048, "step": 5581 }, { "epoch": 0.73, "grad_norm": 1.407347321510315, "learning_rate": 1.774516490242772e-06, "loss": 0.7709, "step": 5582 }, { "epoch": 0.73, "grad_norm": 1.3684868812561035, "learning_rate": 1.7728952637171488e-06, "loss": 0.8398, "step": 5583 }, { "epoch": 0.73, "grad_norm": 0.5998273491859436, "learning_rate": 1.7712746185106888e-06, "loss": 0.9953, "step": 5584 }, { "epoch": 0.73, "grad_norm": 1.351933240890503, "learning_rate": 1.7696545549153255e-06, "loss": 0.7901, "step": 5585 }, { "epoch": 0.73, "grad_norm": 1.4227925539016724, "learning_rate": 1.7680350732228945e-06, "loss": 0.8105, "step": 5586 }, { "epoch": 0.73, "grad_norm": 1.3428065776824951, "learning_rate": 1.7664161737251217e-06, "loss": 0.7927, "step": 5587 }, { "epoch": 0.73, "grad_norm": 1.389904260635376, "learning_rate": 1.7647978567136292e-06, "loss": 0.7774, "step": 5588 }, { "epoch": 0.73, "grad_norm": 1.4323999881744385, "learning_rate": 1.7631801224799367e-06, "loss": 0.8077, "step": 5589 }, { "epoch": 0.73, "grad_norm": 1.3540470600128174, "learning_rate": 1.7615629713154548e-06, "loss": 0.7979, "step": 5590 }, { "epoch": 0.73, "grad_norm": 0.586200475692749, "learning_rate": 1.7599464035114944e-06, "loss": 1.0062, "step": 5591 }, { "epoch": 0.73, "grad_norm": 1.407182216644287, "learning_rate": 1.7583304193592566e-06, "loss": 0.7743, "step": 5592 }, { "epoch": 0.73, "grad_norm": 1.2437548637390137, "learning_rate": 1.7567150191498372e-06, "loss": 0.7441, "step": 5593 }, { "epoch": 0.73, "grad_norm": 0.5851925015449524, "learning_rate": 1.7551002031742325e-06, "loss": 1.0122, "step": 5594 }, { "epoch": 0.73, "grad_norm": 1.3066326379776, "learning_rate": 1.753485971723326e-06, "loss": 0.7709, "step": 5595 }, { "epoch": 0.73, "grad_norm": 1.3402764797210693, "learning_rate": 1.7518723250879027e-06, "loss": 0.7728, "step": 5596 }, { "epoch": 0.73, "grad_norm": 1.606950283050537, "learning_rate": 1.750259263558638e-06, "loss": 0.8205, "step": 5597 }, { "epoch": 0.73, "grad_norm": 0.5866311192512512, "learning_rate": 1.7486467874261025e-06, "loss": 1.0024, "step": 5598 }, { "epoch": 0.73, "grad_norm": 1.4426753520965576, "learning_rate": 1.747034896980761e-06, "loss": 0.7573, "step": 5599 }, { "epoch": 0.73, "grad_norm": 1.390463948249817, "learning_rate": 1.7454235925129764e-06, "loss": 0.7867, "step": 5600 }, { "epoch": 0.73, "grad_norm": 1.4292323589324951, "learning_rate": 1.7438128743130018e-06, "loss": 0.8345, "step": 5601 }, { "epoch": 0.73, "grad_norm": 1.3462961912155151, "learning_rate": 1.7422027426709848e-06, "loss": 0.7552, "step": 5602 }, { "epoch": 0.73, "grad_norm": 1.5547751188278198, "learning_rate": 1.7405931978769719e-06, "loss": 0.7955, "step": 5603 }, { "epoch": 0.73, "grad_norm": 0.6060029864311218, "learning_rate": 1.738984240220899e-06, "loss": 1.0293, "step": 5604 }, { "epoch": 0.73, "grad_norm": 1.432005763053894, "learning_rate": 1.737375869992597e-06, "loss": 0.7814, "step": 5605 }, { "epoch": 0.73, "grad_norm": 1.250579833984375, "learning_rate": 1.7357680874817944e-06, "loss": 0.7861, "step": 5606 }, { "epoch": 0.73, "grad_norm": 1.2750803232192993, "learning_rate": 1.7341608929781107e-06, "loss": 0.8367, "step": 5607 }, { "epoch": 0.73, "grad_norm": 1.2841323614120483, "learning_rate": 1.7325542867710576e-06, "loss": 0.7891, "step": 5608 }, { "epoch": 0.73, "grad_norm": 1.3862614631652832, "learning_rate": 1.7309482691500473e-06, "loss": 0.7752, "step": 5609 }, { "epoch": 0.74, "grad_norm": 1.300353765487671, "learning_rate": 1.72934284040438e-06, "loss": 0.7721, "step": 5610 }, { "epoch": 0.74, "grad_norm": 1.2643624544143677, "learning_rate": 1.7277380008232508e-06, "loss": 0.7734, "step": 5611 }, { "epoch": 0.74, "grad_norm": 4.881514549255371, "learning_rate": 1.7261337506957532e-06, "loss": 0.7589, "step": 5612 }, { "epoch": 0.74, "grad_norm": 0.6066707968711853, "learning_rate": 1.7245300903108669e-06, "loss": 0.9988, "step": 5613 }, { "epoch": 0.74, "grad_norm": 1.2389832735061646, "learning_rate": 1.7229270199574743e-06, "loss": 0.8084, "step": 5614 }, { "epoch": 0.74, "grad_norm": 0.6007294058799744, "learning_rate": 1.7213245399243439e-06, "loss": 1.0078, "step": 5615 }, { "epoch": 0.74, "grad_norm": 1.2838423252105713, "learning_rate": 1.719722650500139e-06, "loss": 0.8124, "step": 5616 }, { "epoch": 0.74, "grad_norm": 1.4337854385375977, "learning_rate": 1.7181213519734225e-06, "loss": 0.7612, "step": 5617 }, { "epoch": 0.74, "grad_norm": 0.5699495077133179, "learning_rate": 1.7165206446326433e-06, "loss": 1.0162, "step": 5618 }, { "epoch": 0.74, "grad_norm": 1.2559514045715332, "learning_rate": 1.7149205287661492e-06, "loss": 0.7692, "step": 5619 }, { "epoch": 0.74, "grad_norm": 1.8447778224945068, "learning_rate": 1.713321004662179e-06, "loss": 0.8442, "step": 5620 }, { "epoch": 0.74, "grad_norm": 0.5917696356773376, "learning_rate": 1.7117220726088628e-06, "loss": 0.9806, "step": 5621 }, { "epoch": 0.74, "grad_norm": 0.5878071784973145, "learning_rate": 1.7101237328942305e-06, "loss": 0.9806, "step": 5622 }, { "epoch": 0.74, "grad_norm": 1.397134780883789, "learning_rate": 1.7085259858061977e-06, "loss": 0.8278, "step": 5623 }, { "epoch": 0.74, "grad_norm": 1.4041309356689453, "learning_rate": 1.7069288316325799e-06, "loss": 0.852, "step": 5624 }, { "epoch": 0.74, "grad_norm": 1.4326568841934204, "learning_rate": 1.705332270661082e-06, "loss": 0.805, "step": 5625 }, { "epoch": 0.74, "grad_norm": 1.3206511735916138, "learning_rate": 1.7037363031793003e-06, "loss": 0.7921, "step": 5626 }, { "epoch": 0.74, "grad_norm": 0.5906815528869629, "learning_rate": 1.7021409294747299e-06, "loss": 1.0289, "step": 5627 }, { "epoch": 0.74, "grad_norm": 1.3678984642028809, "learning_rate": 1.7005461498347532e-06, "loss": 0.8379, "step": 5628 }, { "epoch": 0.74, "grad_norm": 0.5741339921951294, "learning_rate": 1.6989519645466507e-06, "loss": 0.9934, "step": 5629 }, { "epoch": 0.74, "grad_norm": 1.325785756111145, "learning_rate": 1.6973583738975925e-06, "loss": 0.8354, "step": 5630 }, { "epoch": 0.74, "grad_norm": 1.387882947921753, "learning_rate": 1.6957653781746397e-06, "loss": 0.7948, "step": 5631 }, { "epoch": 0.74, "grad_norm": 1.324587345123291, "learning_rate": 1.6941729776647524e-06, "loss": 0.8666, "step": 5632 }, { "epoch": 0.74, "grad_norm": 1.360060214996338, "learning_rate": 1.6925811726547774e-06, "loss": 0.8356, "step": 5633 }, { "epoch": 0.74, "grad_norm": 1.309194803237915, "learning_rate": 1.6909899634314592e-06, "loss": 0.8745, "step": 5634 }, { "epoch": 0.74, "grad_norm": 1.3470100164413452, "learning_rate": 1.6893993502814316e-06, "loss": 0.7751, "step": 5635 }, { "epoch": 0.74, "grad_norm": 1.3145571947097778, "learning_rate": 1.6878093334912199e-06, "loss": 0.7742, "step": 5636 }, { "epoch": 0.74, "grad_norm": 1.3748095035552979, "learning_rate": 1.6862199133472473e-06, "loss": 0.8073, "step": 5637 }, { "epoch": 0.74, "grad_norm": 1.5230497121810913, "learning_rate": 1.6846310901358248e-06, "loss": 0.7843, "step": 5638 }, { "epoch": 0.74, "grad_norm": 1.3721216917037964, "learning_rate": 1.683042864143158e-06, "loss": 0.7588, "step": 5639 }, { "epoch": 0.74, "grad_norm": 0.5940048098564148, "learning_rate": 1.6814552356553416e-06, "loss": 1.0291, "step": 5640 }, { "epoch": 0.74, "grad_norm": 5.112137317657471, "learning_rate": 1.6798682049583687e-06, "loss": 0.8145, "step": 5641 }, { "epoch": 0.74, "grad_norm": 1.3157790899276733, "learning_rate": 1.6782817723381206e-06, "loss": 0.8243, "step": 5642 }, { "epoch": 0.74, "grad_norm": 1.3536481857299805, "learning_rate": 1.6766959380803693e-06, "loss": 0.7865, "step": 5643 }, { "epoch": 0.74, "grad_norm": 0.6024211049079895, "learning_rate": 1.6751107024707846e-06, "loss": 1.0072, "step": 5644 }, { "epoch": 0.74, "grad_norm": 1.3661246299743652, "learning_rate": 1.6735260657949243e-06, "loss": 0.7939, "step": 5645 }, { "epoch": 0.74, "grad_norm": 0.5781591534614563, "learning_rate": 1.6719420283382366e-06, "loss": 1.0298, "step": 5646 }, { "epoch": 0.74, "grad_norm": 1.31074857711792, "learning_rate": 1.6703585903860681e-06, "loss": 0.8093, "step": 5647 }, { "epoch": 0.74, "grad_norm": 1.3388675451278687, "learning_rate": 1.668775752223652e-06, "loss": 0.8057, "step": 5648 }, { "epoch": 0.74, "grad_norm": 1.2917900085449219, "learning_rate": 1.6671935141361134e-06, "loss": 0.7785, "step": 5649 }, { "epoch": 0.74, "grad_norm": 1.283867359161377, "learning_rate": 1.6656118764084745e-06, "loss": 0.8307, "step": 5650 }, { "epoch": 0.74, "grad_norm": 1.864347219467163, "learning_rate": 1.6640308393256427e-06, "loss": 0.7792, "step": 5651 }, { "epoch": 0.74, "grad_norm": 1.2689999341964722, "learning_rate": 1.662450403172423e-06, "loss": 0.8349, "step": 5652 }, { "epoch": 0.74, "grad_norm": 1.3212203979492188, "learning_rate": 1.6608705682335092e-06, "loss": 0.7755, "step": 5653 }, { "epoch": 0.74, "grad_norm": 1.29672110080719, "learning_rate": 1.6592913347934836e-06, "loss": 0.8034, "step": 5654 }, { "epoch": 0.74, "grad_norm": 1.3948447704315186, "learning_rate": 1.657712703136829e-06, "loss": 0.8593, "step": 5655 }, { "epoch": 0.74, "grad_norm": 1.2610478401184082, "learning_rate": 1.65613467354791e-06, "loss": 0.7683, "step": 5656 }, { "epoch": 0.74, "grad_norm": 1.343430757522583, "learning_rate": 1.6545572463109904e-06, "loss": 0.7762, "step": 5657 }, { "epoch": 0.74, "grad_norm": 1.3130717277526855, "learning_rate": 1.6529804217102214e-06, "loss": 0.8011, "step": 5658 }, { "epoch": 0.74, "grad_norm": 1.248701810836792, "learning_rate": 1.6514042000296448e-06, "loss": 0.735, "step": 5659 }, { "epoch": 0.74, "grad_norm": 1.2637665271759033, "learning_rate": 1.6498285815531984e-06, "loss": 0.7453, "step": 5660 }, { "epoch": 0.74, "grad_norm": 1.4572263956069946, "learning_rate": 1.6482535665647053e-06, "loss": 0.7543, "step": 5661 }, { "epoch": 0.74, "grad_norm": 1.346194863319397, "learning_rate": 1.6466791553478872e-06, "loss": 0.804, "step": 5662 }, { "epoch": 0.74, "grad_norm": 1.3273423910140991, "learning_rate": 1.6451053481863498e-06, "loss": 0.7759, "step": 5663 }, { "epoch": 0.74, "grad_norm": 1.3906010389328003, "learning_rate": 1.6435321453635934e-06, "loss": 0.7149, "step": 5664 }, { "epoch": 0.74, "grad_norm": 1.320391058921814, "learning_rate": 1.6419595471630107e-06, "loss": 0.8645, "step": 5665 }, { "epoch": 0.74, "grad_norm": 1.2793775796890259, "learning_rate": 1.6403875538678827e-06, "loss": 0.8498, "step": 5666 }, { "epoch": 0.74, "grad_norm": 1.3723427057266235, "learning_rate": 1.6388161657613816e-06, "loss": 0.812, "step": 5667 }, { "epoch": 0.74, "grad_norm": 1.396850347518921, "learning_rate": 1.6372453831265744e-06, "loss": 0.9095, "step": 5668 }, { "epoch": 0.74, "grad_norm": 1.467105507850647, "learning_rate": 1.6356752062464137e-06, "loss": 0.7951, "step": 5669 }, { "epoch": 0.74, "grad_norm": 0.6094633936882019, "learning_rate": 1.6341056354037482e-06, "loss": 1.0187, "step": 5670 }, { "epoch": 0.74, "grad_norm": 1.3800514936447144, "learning_rate": 1.6325366708813135e-06, "loss": 0.8148, "step": 5671 }, { "epoch": 0.74, "grad_norm": 0.6000351309776306, "learning_rate": 1.630968312961735e-06, "loss": 1.0433, "step": 5672 }, { "epoch": 0.74, "grad_norm": 1.3182547092437744, "learning_rate": 1.6294005619275354e-06, "loss": 0.8029, "step": 5673 }, { "epoch": 0.74, "grad_norm": 1.2944884300231934, "learning_rate": 1.62783341806112e-06, "loss": 0.8063, "step": 5674 }, { "epoch": 0.74, "grad_norm": 1.3299055099487305, "learning_rate": 1.6262668816447918e-06, "loss": 0.8222, "step": 5675 }, { "epoch": 0.74, "grad_norm": 1.4148657321929932, "learning_rate": 1.6247009529607394e-06, "loss": 0.8543, "step": 5676 }, { "epoch": 0.74, "grad_norm": 0.5929334163665771, "learning_rate": 1.6231356322910418e-06, "loss": 0.9849, "step": 5677 }, { "epoch": 0.74, "grad_norm": 1.2701022624969482, "learning_rate": 1.6215709199176744e-06, "loss": 0.7506, "step": 5678 }, { "epoch": 0.74, "grad_norm": 1.4561363458633423, "learning_rate": 1.6200068161224963e-06, "loss": 0.8084, "step": 5679 }, { "epoch": 0.74, "grad_norm": 0.5994031429290771, "learning_rate": 1.6184433211872597e-06, "loss": 1.0024, "step": 5680 }, { "epoch": 0.74, "grad_norm": 1.8892019987106323, "learning_rate": 1.6168804353936063e-06, "loss": 0.8812, "step": 5681 }, { "epoch": 0.74, "grad_norm": 1.319070816040039, "learning_rate": 1.615318159023071e-06, "loss": 0.8378, "step": 5682 }, { "epoch": 0.74, "grad_norm": 1.2919851541519165, "learning_rate": 1.613756492357076e-06, "loss": 0.7628, "step": 5683 }, { "epoch": 0.74, "grad_norm": 1.457441806793213, "learning_rate": 1.6121954356769314e-06, "loss": 0.7553, "step": 5684 }, { "epoch": 0.74, "grad_norm": 0.5886672735214233, "learning_rate": 1.6106349892638451e-06, "loss": 1.0126, "step": 5685 }, { "epoch": 0.75, "grad_norm": 1.3450721502304077, "learning_rate": 1.6090751533989084e-06, "loss": 0.8056, "step": 5686 }, { "epoch": 0.75, "grad_norm": 1.4432209730148315, "learning_rate": 1.607515928363102e-06, "loss": 0.8293, "step": 5687 }, { "epoch": 0.75, "grad_norm": 0.5696353316307068, "learning_rate": 1.6059573144373035e-06, "loss": 1.002, "step": 5688 }, { "epoch": 0.75, "grad_norm": 1.2769510746002197, "learning_rate": 1.6043993119022722e-06, "loss": 0.7647, "step": 5689 }, { "epoch": 0.75, "grad_norm": 0.5891345143318176, "learning_rate": 1.6028419210386648e-06, "loss": 0.9941, "step": 5690 }, { "epoch": 0.75, "grad_norm": 1.441389560699463, "learning_rate": 1.601285142127022e-06, "loss": 0.8718, "step": 5691 }, { "epoch": 0.75, "grad_norm": 0.5820150375366211, "learning_rate": 1.5997289754477758e-06, "loss": 0.9922, "step": 5692 }, { "epoch": 0.75, "grad_norm": 2.067706346511841, "learning_rate": 1.5981734212812506e-06, "loss": 0.7766, "step": 5693 }, { "epoch": 0.75, "grad_norm": 0.5747236013412476, "learning_rate": 1.5966184799076567e-06, "loss": 1.0048, "step": 5694 }, { "epoch": 0.75, "grad_norm": 0.5649420619010925, "learning_rate": 1.5950641516070953e-06, "loss": 1.0279, "step": 5695 }, { "epoch": 0.75, "grad_norm": 1.3932156562805176, "learning_rate": 1.59351043665956e-06, "loss": 0.8503, "step": 5696 }, { "epoch": 0.75, "grad_norm": 1.3567396402359009, "learning_rate": 1.591957335344928e-06, "loss": 0.8134, "step": 5697 }, { "epoch": 0.75, "grad_norm": 1.8029769659042358, "learning_rate": 1.5904048479429734e-06, "loss": 0.771, "step": 5698 }, { "epoch": 0.75, "grad_norm": 1.2815278768539429, "learning_rate": 1.5888529747333537e-06, "loss": 0.8144, "step": 5699 }, { "epoch": 0.75, "grad_norm": 1.3930268287658691, "learning_rate": 1.5873017159956167e-06, "loss": 0.703, "step": 5700 }, { "epoch": 0.75, "grad_norm": 1.3645144701004028, "learning_rate": 1.585751072009203e-06, "loss": 0.7616, "step": 5701 }, { "epoch": 0.75, "grad_norm": 1.3949958086013794, "learning_rate": 1.584201043053437e-06, "loss": 0.7303, "step": 5702 }, { "epoch": 0.75, "grad_norm": 1.328474760055542, "learning_rate": 1.5826516294075394e-06, "loss": 0.7736, "step": 5703 }, { "epoch": 0.75, "grad_norm": 1.5312252044677734, "learning_rate": 1.5811028313506138e-06, "loss": 0.7991, "step": 5704 }, { "epoch": 0.75, "grad_norm": 0.5920330286026001, "learning_rate": 1.5795546491616537e-06, "loss": 0.9995, "step": 5705 }, { "epoch": 0.75, "grad_norm": 1.3339109420776367, "learning_rate": 1.5780070831195466e-06, "loss": 0.8176, "step": 5706 }, { "epoch": 0.75, "grad_norm": 1.401720404624939, "learning_rate": 1.5764601335030615e-06, "loss": 0.8499, "step": 5707 }, { "epoch": 0.75, "grad_norm": 1.3738751411437988, "learning_rate": 1.5749138005908648e-06, "loss": 0.7709, "step": 5708 }, { "epoch": 0.75, "grad_norm": 0.5915643572807312, "learning_rate": 1.573368084661505e-06, "loss": 1.0004, "step": 5709 }, { "epoch": 0.75, "grad_norm": 1.355438232421875, "learning_rate": 1.571822985993421e-06, "loss": 0.8219, "step": 5710 }, { "epoch": 0.75, "grad_norm": 1.5761815309524536, "learning_rate": 1.5702785048649432e-06, "loss": 0.8118, "step": 5711 }, { "epoch": 0.75, "grad_norm": 0.5785830020904541, "learning_rate": 1.568734641554287e-06, "loss": 1.0089, "step": 5712 }, { "epoch": 0.75, "grad_norm": 0.5895341634750366, "learning_rate": 1.5671913963395612e-06, "loss": 0.9972, "step": 5713 }, { "epoch": 0.75, "grad_norm": 1.3230422735214233, "learning_rate": 1.5656487694987593e-06, "loss": 0.8822, "step": 5714 }, { "epoch": 0.75, "grad_norm": 0.5743219256401062, "learning_rate": 1.5641067613097621e-06, "loss": 1.0258, "step": 5715 }, { "epoch": 0.75, "grad_norm": 1.866860032081604, "learning_rate": 1.5625653720503458e-06, "loss": 0.7744, "step": 5716 }, { "epoch": 0.75, "grad_norm": 1.3569918870925903, "learning_rate": 1.561024601998168e-06, "loss": 0.8373, "step": 5717 }, { "epoch": 0.75, "grad_norm": 1.4399479627609253, "learning_rate": 1.5594844514307767e-06, "loss": 0.8373, "step": 5718 }, { "epoch": 0.75, "grad_norm": 1.4347025156021118, "learning_rate": 1.557944920625612e-06, "loss": 0.8064, "step": 5719 }, { "epoch": 0.75, "grad_norm": 1.4409246444702148, "learning_rate": 1.5564060098599986e-06, "loss": 0.7892, "step": 5720 }, { "epoch": 0.75, "grad_norm": 1.326335072517395, "learning_rate": 1.5548677194111477e-06, "loss": 0.8702, "step": 5721 }, { "epoch": 0.75, "grad_norm": 1.3781486749649048, "learning_rate": 1.5533300495561653e-06, "loss": 0.7853, "step": 5722 }, { "epoch": 0.75, "grad_norm": 1.3263145685195923, "learning_rate": 1.5517930005720404e-06, "loss": 0.8024, "step": 5723 }, { "epoch": 0.75, "grad_norm": 1.298985242843628, "learning_rate": 1.550256572735651e-06, "loss": 0.8554, "step": 5724 }, { "epoch": 0.75, "grad_norm": 0.6624327898025513, "learning_rate": 1.5487207663237625e-06, "loss": 0.9984, "step": 5725 }, { "epoch": 0.75, "grad_norm": 2.2485671043395996, "learning_rate": 1.5471855816130328e-06, "loss": 0.8122, "step": 5726 }, { "epoch": 0.75, "grad_norm": 0.6306642889976501, "learning_rate": 1.5456510188800023e-06, "loss": 0.997, "step": 5727 }, { "epoch": 0.75, "grad_norm": 1.3298956155776978, "learning_rate": 1.544117078401101e-06, "loss": 0.7669, "step": 5728 }, { "epoch": 0.75, "grad_norm": 1.2502905130386353, "learning_rate": 1.5425837604526506e-06, "loss": 0.7954, "step": 5729 }, { "epoch": 0.75, "grad_norm": 1.3008549213409424, "learning_rate": 1.541051065310854e-06, "loss": 0.7562, "step": 5730 }, { "epoch": 0.75, "grad_norm": 0.5997135043144226, "learning_rate": 1.5395189932518085e-06, "loss": 0.9942, "step": 5731 }, { "epoch": 0.75, "grad_norm": 1.2991734743118286, "learning_rate": 1.5379875445514947e-06, "loss": 0.7858, "step": 5732 }, { "epoch": 0.75, "grad_norm": 1.3368172645568848, "learning_rate": 1.536456719485781e-06, "loss": 0.7969, "step": 5733 }, { "epoch": 0.75, "grad_norm": 0.5769408941268921, "learning_rate": 1.5349265183304273e-06, "loss": 1.0067, "step": 5734 }, { "epoch": 0.75, "grad_norm": 1.2739698886871338, "learning_rate": 1.5333969413610766e-06, "loss": 0.7903, "step": 5735 }, { "epoch": 0.75, "grad_norm": 1.3133176565170288, "learning_rate": 1.5318679888532633e-06, "loss": 0.7488, "step": 5736 }, { "epoch": 0.75, "grad_norm": 1.3065721988677979, "learning_rate": 1.5303396610824061e-06, "loss": 0.8233, "step": 5737 }, { "epoch": 0.75, "grad_norm": 1.2927088737487793, "learning_rate": 1.528811958323812e-06, "loss": 0.7914, "step": 5738 }, { "epoch": 0.75, "grad_norm": 1.7547574043273926, "learning_rate": 1.527284880852678e-06, "loss": 0.8076, "step": 5739 }, { "epoch": 0.75, "grad_norm": 1.3058931827545166, "learning_rate": 1.5257584289440835e-06, "loss": 0.8263, "step": 5740 }, { "epoch": 0.75, "grad_norm": 0.6215991973876953, "learning_rate": 1.5242326028730015e-06, "loss": 0.9963, "step": 5741 }, { "epoch": 0.75, "grad_norm": 0.6036487817764282, "learning_rate": 1.5227074029142874e-06, "loss": 1.011, "step": 5742 }, { "epoch": 0.75, "grad_norm": 0.5965812802314758, "learning_rate": 1.5211828293426833e-06, "loss": 1.0091, "step": 5743 }, { "epoch": 0.75, "grad_norm": 1.333753228187561, "learning_rate": 1.519658882432824e-06, "loss": 0.8187, "step": 5744 }, { "epoch": 0.75, "grad_norm": 1.318195104598999, "learning_rate": 1.5181355624592242e-06, "loss": 0.8041, "step": 5745 }, { "epoch": 0.75, "grad_norm": 0.5771324038505554, "learning_rate": 1.5166128696962929e-06, "loss": 1.0158, "step": 5746 }, { "epoch": 0.75, "grad_norm": 1.3503345251083374, "learning_rate": 1.515090804418321e-06, "loss": 0.7387, "step": 5747 }, { "epoch": 0.75, "grad_norm": 1.30474054813385, "learning_rate": 1.5135693668994866e-06, "loss": 0.827, "step": 5748 }, { "epoch": 0.75, "grad_norm": 0.6058622598648071, "learning_rate": 1.5120485574138583e-06, "loss": 1.0053, "step": 5749 }, { "epoch": 0.75, "grad_norm": 1.3747936487197876, "learning_rate": 1.5105283762353866e-06, "loss": 0.7946, "step": 5750 }, { "epoch": 0.75, "grad_norm": 1.193913459777832, "learning_rate": 1.5090088236379151e-06, "loss": 0.7523, "step": 5751 }, { "epoch": 0.75, "grad_norm": 1.3963285684585571, "learning_rate": 1.5074898998951686e-06, "loss": 0.8018, "step": 5752 }, { "epoch": 0.75, "grad_norm": 1.4195276498794556, "learning_rate": 1.5059716052807593e-06, "loss": 0.8054, "step": 5753 }, { "epoch": 0.75, "grad_norm": 1.3360220193862915, "learning_rate": 1.5044539400681901e-06, "loss": 0.8053, "step": 5754 }, { "epoch": 0.75, "grad_norm": 1.4829891920089722, "learning_rate": 1.5029369045308466e-06, "loss": 0.7945, "step": 5755 }, { "epoch": 0.75, "grad_norm": 1.6203913688659668, "learning_rate": 1.5014204989420005e-06, "loss": 0.7921, "step": 5756 }, { "epoch": 0.75, "grad_norm": 0.6024445295333862, "learning_rate": 1.4999047235748149e-06, "loss": 0.9842, "step": 5757 }, { "epoch": 0.75, "grad_norm": 1.2503794431686401, "learning_rate": 1.4983895787023345e-06, "loss": 0.7232, "step": 5758 }, { "epoch": 0.75, "grad_norm": 1.2966883182525635, "learning_rate": 1.496875064597491e-06, "loss": 0.7704, "step": 5759 }, { "epoch": 0.75, "grad_norm": 0.5840237140655518, "learning_rate": 1.495361181533106e-06, "loss": 1.0073, "step": 5760 }, { "epoch": 0.75, "grad_norm": 0.5861770510673523, "learning_rate": 1.4938479297818837e-06, "loss": 1.0191, "step": 5761 }, { "epoch": 0.76, "grad_norm": 1.4378658533096313, "learning_rate": 1.4923353096164144e-06, "loss": 0.8255, "step": 5762 }, { "epoch": 0.76, "grad_norm": 1.3363529443740845, "learning_rate": 1.4908233213091793e-06, "loss": 0.7952, "step": 5763 }, { "epoch": 0.76, "grad_norm": 1.3527425527572632, "learning_rate": 1.4893119651325405e-06, "loss": 0.8127, "step": 5764 }, { "epoch": 0.76, "grad_norm": 0.6010879874229431, "learning_rate": 1.4878012413587495e-06, "loss": 0.998, "step": 5765 }, { "epoch": 0.76, "grad_norm": 1.2600860595703125, "learning_rate": 1.4862911502599404e-06, "loss": 0.7959, "step": 5766 }, { "epoch": 0.76, "grad_norm": 1.4337431192398071, "learning_rate": 1.4847816921081382e-06, "loss": 0.825, "step": 5767 }, { "epoch": 0.76, "grad_norm": 1.5739651918411255, "learning_rate": 1.4832728671752488e-06, "loss": 0.8124, "step": 5768 }, { "epoch": 0.76, "grad_norm": 0.6011191606521606, "learning_rate": 1.48176467573307e-06, "loss": 0.9965, "step": 5769 }, { "epoch": 0.76, "grad_norm": 0.58770352602005, "learning_rate": 1.4802571180532798e-06, "loss": 1.0141, "step": 5770 }, { "epoch": 0.76, "grad_norm": 1.309753179550171, "learning_rate": 1.4787501944074435e-06, "loss": 0.7592, "step": 5771 }, { "epoch": 0.76, "grad_norm": 1.2828128337860107, "learning_rate": 1.4772439050670157e-06, "loss": 0.7895, "step": 5772 }, { "epoch": 0.76, "grad_norm": 0.5770124793052673, "learning_rate": 1.47573825030333e-06, "loss": 1.0265, "step": 5773 }, { "epoch": 0.76, "grad_norm": 0.5667924284934998, "learning_rate": 1.4742332303876144e-06, "loss": 1.0303, "step": 5774 }, { "epoch": 0.76, "grad_norm": 1.2779595851898193, "learning_rate": 1.472728845590975e-06, "loss": 0.7361, "step": 5775 }, { "epoch": 0.76, "grad_norm": 0.5909060835838318, "learning_rate": 1.4712250961844054e-06, "loss": 0.9938, "step": 5776 }, { "epoch": 0.76, "grad_norm": 0.5945279002189636, "learning_rate": 1.4697219824387887e-06, "loss": 1.0114, "step": 5777 }, { "epoch": 0.76, "grad_norm": 1.375620722770691, "learning_rate": 1.4682195046248871e-06, "loss": 0.8465, "step": 5778 }, { "epoch": 0.76, "grad_norm": 1.4058698415756226, "learning_rate": 1.466717663013355e-06, "loss": 0.8865, "step": 5779 }, { "epoch": 0.76, "grad_norm": 1.329715609550476, "learning_rate": 1.4652164578747274e-06, "loss": 0.7697, "step": 5780 }, { "epoch": 0.76, "grad_norm": 1.3355653285980225, "learning_rate": 1.463715889479424e-06, "loss": 0.7431, "step": 5781 }, { "epoch": 0.76, "grad_norm": 1.6756675243377686, "learning_rate": 1.4622159580977546e-06, "loss": 0.7859, "step": 5782 }, { "epoch": 0.76, "grad_norm": 1.2889306545257568, "learning_rate": 1.4607166639999114e-06, "loss": 0.8102, "step": 5783 }, { "epoch": 0.76, "grad_norm": 1.277280569076538, "learning_rate": 1.4592180074559692e-06, "loss": 0.7289, "step": 5784 }, { "epoch": 0.76, "grad_norm": 1.3062422275543213, "learning_rate": 1.4577199887358939e-06, "loss": 0.7482, "step": 5785 }, { "epoch": 0.76, "grad_norm": 0.5742699503898621, "learning_rate": 1.4562226081095304e-06, "loss": 1.0192, "step": 5786 }, { "epoch": 0.76, "grad_norm": 1.322646975517273, "learning_rate": 1.4547258658466147e-06, "loss": 0.7623, "step": 5787 }, { "epoch": 0.76, "grad_norm": 1.619138240814209, "learning_rate": 1.453229762216763e-06, "loss": 0.8247, "step": 5788 }, { "epoch": 0.76, "grad_norm": 1.5821067094802856, "learning_rate": 1.451734297489476e-06, "loss": 0.7753, "step": 5789 }, { "epoch": 0.76, "grad_norm": 1.297759771347046, "learning_rate": 1.450239471934145e-06, "loss": 0.8367, "step": 5790 }, { "epoch": 0.76, "grad_norm": 1.5656774044036865, "learning_rate": 1.4487452858200395e-06, "loss": 0.8444, "step": 5791 }, { "epoch": 0.76, "grad_norm": 1.303881287574768, "learning_rate": 1.44725173941632e-06, "loss": 0.8446, "step": 5792 }, { "epoch": 0.76, "grad_norm": 1.3475663661956787, "learning_rate": 1.445758832992027e-06, "loss": 0.7678, "step": 5793 }, { "epoch": 0.76, "grad_norm": 0.5880793333053589, "learning_rate": 1.4442665668160855e-06, "loss": 1.0014, "step": 5794 }, { "epoch": 0.76, "grad_norm": 1.3143463134765625, "learning_rate": 1.44277494115731e-06, "loss": 0.8509, "step": 5795 }, { "epoch": 0.76, "grad_norm": 0.5737321376800537, "learning_rate": 1.4412839562843944e-06, "loss": 1.0129, "step": 5796 }, { "epoch": 0.76, "grad_norm": 0.6095173358917236, "learning_rate": 1.4397936124659218e-06, "loss": 0.9933, "step": 5797 }, { "epoch": 0.76, "grad_norm": 1.2888370752334595, "learning_rate": 1.438303909970356e-06, "loss": 0.808, "step": 5798 }, { "epoch": 0.76, "grad_norm": 0.579931914806366, "learning_rate": 1.436814849066047e-06, "loss": 1.0148, "step": 5799 }, { "epoch": 0.76, "grad_norm": 1.2524925470352173, "learning_rate": 1.4353264300212273e-06, "loss": 0.8402, "step": 5800 }, { "epoch": 0.76, "grad_norm": 1.2922186851501465, "learning_rate": 1.4338386531040176e-06, "loss": 0.7808, "step": 5801 }, { "epoch": 0.76, "grad_norm": 1.3427610397338867, "learning_rate": 1.4323515185824204e-06, "loss": 0.6969, "step": 5802 }, { "epoch": 0.76, "grad_norm": 1.276766061782837, "learning_rate": 1.4308650267243213e-06, "loss": 0.7852, "step": 5803 }, { "epoch": 0.76, "grad_norm": 1.6158946752548218, "learning_rate": 1.4293791777974936e-06, "loss": 0.8492, "step": 5804 }, { "epoch": 0.76, "grad_norm": 1.342363715171814, "learning_rate": 1.4278939720695923e-06, "loss": 0.8044, "step": 5805 }, { "epoch": 0.76, "grad_norm": 1.3377900123596191, "learning_rate": 1.4264094098081554e-06, "loss": 0.7602, "step": 5806 }, { "epoch": 0.76, "grad_norm": 1.2690436840057373, "learning_rate": 1.4249254912806092e-06, "loss": 0.7654, "step": 5807 }, { "epoch": 0.76, "grad_norm": 1.3291325569152832, "learning_rate": 1.4234422167542604e-06, "loss": 0.8036, "step": 5808 }, { "epoch": 0.76, "grad_norm": 1.3342899084091187, "learning_rate": 1.4219595864962993e-06, "loss": 0.7546, "step": 5809 }, { "epoch": 0.76, "grad_norm": 1.426802635192871, "learning_rate": 1.420477600773804e-06, "loss": 0.8591, "step": 5810 }, { "epoch": 0.76, "grad_norm": 1.4962778091430664, "learning_rate": 1.4189962598537338e-06, "loss": 0.8089, "step": 5811 }, { "epoch": 0.76, "grad_norm": 1.30381178855896, "learning_rate": 1.4175155640029298e-06, "loss": 0.7712, "step": 5812 }, { "epoch": 0.76, "grad_norm": 1.2953115701675415, "learning_rate": 1.4160355134881227e-06, "loss": 0.7925, "step": 5813 }, { "epoch": 0.76, "grad_norm": 1.2727209329605103, "learning_rate": 1.4145561085759201e-06, "loss": 0.7754, "step": 5814 }, { "epoch": 0.76, "grad_norm": 1.380553126335144, "learning_rate": 1.41307734953282e-06, "loss": 0.7888, "step": 5815 }, { "epoch": 0.76, "grad_norm": 1.2530661821365356, "learning_rate": 1.4115992366251996e-06, "loss": 0.7636, "step": 5816 }, { "epoch": 0.76, "grad_norm": 1.2861392498016357, "learning_rate": 1.4101217701193188e-06, "loss": 0.8427, "step": 5817 }, { "epoch": 0.76, "grad_norm": 1.2771718502044678, "learning_rate": 1.4086449502813265e-06, "loss": 0.7856, "step": 5818 }, { "epoch": 0.76, "grad_norm": 1.3790676593780518, "learning_rate": 1.4071687773772485e-06, "loss": 0.7507, "step": 5819 }, { "epoch": 0.76, "grad_norm": 0.6318563222885132, "learning_rate": 1.405693251673001e-06, "loss": 1.013, "step": 5820 }, { "epoch": 0.76, "grad_norm": 0.6068992018699646, "learning_rate": 1.4042183734343779e-06, "loss": 0.9967, "step": 5821 }, { "epoch": 0.76, "grad_norm": 1.4906160831451416, "learning_rate": 1.4027441429270566e-06, "loss": 0.8399, "step": 5822 }, { "epoch": 0.76, "grad_norm": 1.2709884643554688, "learning_rate": 1.4012705604166038e-06, "loss": 0.7939, "step": 5823 }, { "epoch": 0.76, "grad_norm": 1.3939069509506226, "learning_rate": 1.3997976261684615e-06, "loss": 0.8093, "step": 5824 }, { "epoch": 0.76, "grad_norm": 1.4670621156692505, "learning_rate": 1.3983253404479625e-06, "loss": 0.7736, "step": 5825 }, { "epoch": 0.76, "grad_norm": 1.45097017288208, "learning_rate": 1.3968537035203173e-06, "loss": 0.8319, "step": 5826 }, { "epoch": 0.76, "grad_norm": 1.2880184650421143, "learning_rate": 1.3953827156506194e-06, "loss": 0.8014, "step": 5827 }, { "epoch": 0.76, "grad_norm": 0.587405264377594, "learning_rate": 1.3939123771038515e-06, "loss": 1.0287, "step": 5828 }, { "epoch": 0.76, "grad_norm": 1.2253308296203613, "learning_rate": 1.392442688144871e-06, "loss": 0.7778, "step": 5829 }, { "epoch": 0.76, "grad_norm": 1.345903992652893, "learning_rate": 1.3909736490384258e-06, "loss": 0.7773, "step": 5830 }, { "epoch": 0.76, "grad_norm": 1.2805379629135132, "learning_rate": 1.3895052600491421e-06, "loss": 0.8438, "step": 5831 }, { "epoch": 0.76, "grad_norm": 1.4632360935211182, "learning_rate": 1.3880375214415281e-06, "loss": 0.8171, "step": 5832 }, { "epoch": 0.76, "grad_norm": 1.450731635093689, "learning_rate": 1.3865704334799807e-06, "loss": 0.7938, "step": 5833 }, { "epoch": 0.76, "grad_norm": 0.5917348265647888, "learning_rate": 1.3851039964287733e-06, "loss": 0.9912, "step": 5834 }, { "epoch": 0.76, "grad_norm": 1.566248893737793, "learning_rate": 1.3836382105520663e-06, "loss": 0.8164, "step": 5835 }, { "epoch": 0.76, "grad_norm": 1.3307698965072632, "learning_rate": 1.3821730761139008e-06, "loss": 0.742, "step": 5836 }, { "epoch": 0.76, "grad_norm": 0.6008026599884033, "learning_rate": 1.3807085933781988e-06, "loss": 1.0036, "step": 5837 }, { "epoch": 0.76, "grad_norm": 1.3283597230911255, "learning_rate": 1.3792447626087702e-06, "loss": 0.8464, "step": 5838 }, { "epoch": 0.77, "grad_norm": 0.5942666530609131, "learning_rate": 1.3777815840693026e-06, "loss": 1.0059, "step": 5839 }, { "epoch": 0.77, "grad_norm": 1.3265557289123535, "learning_rate": 1.3763190580233687e-06, "loss": 0.7618, "step": 5840 }, { "epoch": 0.77, "grad_norm": 1.2970008850097656, "learning_rate": 1.37485718473442e-06, "loss": 0.8529, "step": 5841 }, { "epoch": 0.77, "grad_norm": 1.2413841485977173, "learning_rate": 1.3733959644657964e-06, "loss": 0.7386, "step": 5842 }, { "epoch": 0.77, "grad_norm": 1.3074026107788086, "learning_rate": 1.3719353974807159e-06, "loss": 0.7753, "step": 5843 }, { "epoch": 0.77, "grad_norm": 1.3485804796218872, "learning_rate": 1.370475484042278e-06, "loss": 0.8007, "step": 5844 }, { "epoch": 0.77, "grad_norm": 0.5854542255401611, "learning_rate": 1.3690162244134697e-06, "loss": 1.0245, "step": 5845 }, { "epoch": 0.77, "grad_norm": 1.3493131399154663, "learning_rate": 1.3675576188571548e-06, "loss": 0.7996, "step": 5846 }, { "epoch": 0.77, "grad_norm": 2.19130277633667, "learning_rate": 1.36609966763608e-06, "loss": 0.7729, "step": 5847 }, { "epoch": 0.77, "grad_norm": 1.4471193552017212, "learning_rate": 1.3646423710128786e-06, "loss": 0.7614, "step": 5848 }, { "epoch": 0.77, "grad_norm": 0.5762377977371216, "learning_rate": 1.363185729250061e-06, "loss": 1.0182, "step": 5849 }, { "epoch": 0.77, "grad_norm": 1.3593144416809082, "learning_rate": 1.3617297426100208e-06, "loss": 0.8013, "step": 5850 }, { "epoch": 0.77, "grad_norm": 0.5963528156280518, "learning_rate": 1.360274411355036e-06, "loss": 1.0046, "step": 5851 }, { "epoch": 0.77, "grad_norm": 1.246561050415039, "learning_rate": 1.3588197357472628e-06, "loss": 0.7753, "step": 5852 }, { "epoch": 0.77, "grad_norm": 1.3125818967819214, "learning_rate": 1.3573657160487436e-06, "loss": 0.7462, "step": 5853 }, { "epoch": 0.77, "grad_norm": 1.361146092414856, "learning_rate": 1.3559123525213997e-06, "loss": 0.8055, "step": 5854 }, { "epoch": 0.77, "grad_norm": 1.4558393955230713, "learning_rate": 1.354459645427032e-06, "loss": 0.7717, "step": 5855 }, { "epoch": 0.77, "grad_norm": 1.3343006372451782, "learning_rate": 1.3530075950273303e-06, "loss": 0.8484, "step": 5856 }, { "epoch": 0.77, "grad_norm": 0.5933970212936401, "learning_rate": 1.3515562015838585e-06, "loss": 1.0234, "step": 5857 }, { "epoch": 0.77, "grad_norm": 1.306185245513916, "learning_rate": 1.3501054653580675e-06, "loss": 0.8072, "step": 5858 }, { "epoch": 0.77, "grad_norm": 1.334793210029602, "learning_rate": 1.3486553866112872e-06, "loss": 0.8228, "step": 5859 }, { "epoch": 0.77, "grad_norm": 1.3544604778289795, "learning_rate": 1.347205965604728e-06, "loss": 0.7788, "step": 5860 }, { "epoch": 0.77, "grad_norm": 1.5510765314102173, "learning_rate": 1.3457572025994864e-06, "loss": 0.8545, "step": 5861 }, { "epoch": 0.77, "grad_norm": 1.3287917375564575, "learning_rate": 1.3443090978565344e-06, "loss": 0.8064, "step": 5862 }, { "epoch": 0.77, "grad_norm": 1.3769797086715698, "learning_rate": 1.3428616516367316e-06, "loss": 0.793, "step": 5863 }, { "epoch": 0.77, "grad_norm": 1.387160062789917, "learning_rate": 1.3414148642008134e-06, "loss": 0.7741, "step": 5864 }, { "epoch": 0.77, "grad_norm": 1.2930328845977783, "learning_rate": 1.3399687358093987e-06, "loss": 0.7752, "step": 5865 }, { "epoch": 0.77, "grad_norm": 1.4460489749908447, "learning_rate": 1.3385232667229904e-06, "loss": 0.7651, "step": 5866 }, { "epoch": 0.77, "grad_norm": 1.3168561458587646, "learning_rate": 1.3370784572019663e-06, "loss": 0.8193, "step": 5867 }, { "epoch": 0.77, "grad_norm": 0.6046688556671143, "learning_rate": 1.3356343075065937e-06, "loss": 1.0212, "step": 5868 }, { "epoch": 0.77, "grad_norm": 1.3532068729400635, "learning_rate": 1.3341908178970141e-06, "loss": 0.7986, "step": 5869 }, { "epoch": 0.77, "grad_norm": 0.6073417067527771, "learning_rate": 1.3327479886332522e-06, "loss": 1.0168, "step": 5870 }, { "epoch": 0.77, "grad_norm": 1.4243968725204468, "learning_rate": 1.3313058199752154e-06, "loss": 0.7535, "step": 5871 }, { "epoch": 0.77, "grad_norm": 1.4426363706588745, "learning_rate": 1.329864312182691e-06, "loss": 0.8238, "step": 5872 }, { "epoch": 0.77, "grad_norm": 1.3493560552597046, "learning_rate": 1.3284234655153443e-06, "loss": 0.7607, "step": 5873 }, { "epoch": 0.77, "grad_norm": 1.3521627187728882, "learning_rate": 1.326983280232728e-06, "loss": 0.7733, "step": 5874 }, { "epoch": 0.77, "grad_norm": 1.3281702995300293, "learning_rate": 1.3255437565942692e-06, "loss": 0.7851, "step": 5875 }, { "epoch": 0.77, "grad_norm": 1.4623819589614868, "learning_rate": 1.3241048948592805e-06, "loss": 0.7713, "step": 5876 }, { "epoch": 0.77, "grad_norm": 1.2828166484832764, "learning_rate": 1.3226666952869532e-06, "loss": 0.7633, "step": 5877 }, { "epoch": 0.77, "grad_norm": 0.585110068321228, "learning_rate": 1.3212291581363567e-06, "loss": 1.0222, "step": 5878 }, { "epoch": 0.77, "grad_norm": 1.3305892944335938, "learning_rate": 1.3197922836664473e-06, "loss": 0.8179, "step": 5879 }, { "epoch": 0.77, "grad_norm": 1.3580206632614136, "learning_rate": 1.3183560721360577e-06, "loss": 0.8441, "step": 5880 }, { "epoch": 0.77, "grad_norm": 1.4255685806274414, "learning_rate": 1.316920523803899e-06, "loss": 0.8364, "step": 5881 }, { "epoch": 0.77, "grad_norm": 1.4987274408340454, "learning_rate": 1.3154856389285702e-06, "loss": 0.851, "step": 5882 }, { "epoch": 0.77, "grad_norm": 1.3492183685302734, "learning_rate": 1.3140514177685438e-06, "loss": 0.7596, "step": 5883 }, { "epoch": 0.77, "grad_norm": 1.2381666898727417, "learning_rate": 1.3126178605821754e-06, "loss": 0.7783, "step": 5884 }, { "epoch": 0.77, "grad_norm": 1.4262123107910156, "learning_rate": 1.3111849676277e-06, "loss": 0.7541, "step": 5885 }, { "epoch": 0.77, "grad_norm": 1.3398419618606567, "learning_rate": 1.3097527391632365e-06, "loss": 0.8088, "step": 5886 }, { "epoch": 0.77, "grad_norm": 1.3383190631866455, "learning_rate": 1.3083211754467801e-06, "loss": 0.7375, "step": 5887 }, { "epoch": 0.77, "grad_norm": 1.297507882118225, "learning_rate": 1.3068902767362057e-06, "loss": 0.7759, "step": 5888 }, { "epoch": 0.77, "grad_norm": 0.5822311639785767, "learning_rate": 1.3054600432892734e-06, "loss": 0.9992, "step": 5889 }, { "epoch": 0.77, "grad_norm": 1.5857821702957153, "learning_rate": 1.304030475363618e-06, "loss": 0.7796, "step": 5890 }, { "epoch": 0.77, "grad_norm": 1.4157729148864746, "learning_rate": 1.3026015732167586e-06, "loss": 0.7742, "step": 5891 }, { "epoch": 0.77, "grad_norm": 1.3562119007110596, "learning_rate": 1.301173337106092e-06, "loss": 0.752, "step": 5892 }, { "epoch": 0.77, "grad_norm": 1.3223198652267456, "learning_rate": 1.299745767288894e-06, "loss": 0.8263, "step": 5893 }, { "epoch": 0.77, "grad_norm": 1.751035213470459, "learning_rate": 1.2983188640223248e-06, "loss": 0.7622, "step": 5894 }, { "epoch": 0.77, "grad_norm": 1.4121713638305664, "learning_rate": 1.296892627563418e-06, "loss": 0.7716, "step": 5895 }, { "epoch": 0.77, "grad_norm": 1.3097952604293823, "learning_rate": 1.2954670581690947e-06, "loss": 0.7828, "step": 5896 }, { "epoch": 0.77, "grad_norm": 1.332562804222107, "learning_rate": 1.2940421560961501e-06, "loss": 0.7521, "step": 5897 }, { "epoch": 0.77, "grad_norm": 1.3170368671417236, "learning_rate": 1.2926179216012596e-06, "loss": 0.8206, "step": 5898 }, { "epoch": 0.77, "grad_norm": 1.4210290908813477, "learning_rate": 1.2911943549409827e-06, "loss": 0.8027, "step": 5899 }, { "epoch": 0.77, "grad_norm": 1.2368905544281006, "learning_rate": 1.289771456371754e-06, "loss": 0.8062, "step": 5900 }, { "epoch": 0.77, "grad_norm": 1.4254677295684814, "learning_rate": 1.2883492261498876e-06, "loss": 0.7614, "step": 5901 }, { "epoch": 0.77, "grad_norm": 1.3791615962982178, "learning_rate": 1.286927664531582e-06, "loss": 0.7702, "step": 5902 }, { "epoch": 0.77, "grad_norm": 1.303736686706543, "learning_rate": 1.2855067717729103e-06, "loss": 0.7963, "step": 5903 }, { "epoch": 0.77, "grad_norm": 0.6079103350639343, "learning_rate": 1.2840865481298288e-06, "loss": 1.0206, "step": 5904 }, { "epoch": 0.77, "grad_norm": 1.3317543268203735, "learning_rate": 1.2826669938581704e-06, "loss": 0.8544, "step": 5905 }, { "epoch": 0.77, "grad_norm": 1.3727269172668457, "learning_rate": 1.2812481092136476e-06, "loss": 0.7882, "step": 5906 }, { "epoch": 0.77, "grad_norm": 0.603192150592804, "learning_rate": 1.2798298944518555e-06, "loss": 0.9911, "step": 5907 }, { "epoch": 0.77, "grad_norm": 1.3365495204925537, "learning_rate": 1.278412349828264e-06, "loss": 0.8328, "step": 5908 }, { "epoch": 0.77, "grad_norm": 1.8223514556884766, "learning_rate": 1.2769954755982268e-06, "loss": 0.7816, "step": 5909 }, { "epoch": 0.77, "grad_norm": 1.3478187322616577, "learning_rate": 1.2755792720169736e-06, "loss": 0.7949, "step": 5910 }, { "epoch": 0.77, "grad_norm": 1.5371490716934204, "learning_rate": 1.2741637393396122e-06, "loss": 0.8141, "step": 5911 }, { "epoch": 0.77, "grad_norm": 1.3957459926605225, "learning_rate": 1.2727488778211356e-06, "loss": 0.7807, "step": 5912 }, { "epoch": 0.77, "grad_norm": 1.400453805923462, "learning_rate": 1.271334687716408e-06, "loss": 0.8208, "step": 5913 }, { "epoch": 0.77, "grad_norm": 1.4223103523254395, "learning_rate": 1.2699211692801794e-06, "loss": 0.792, "step": 5914 }, { "epoch": 0.78, "grad_norm": 0.5835632085800171, "learning_rate": 1.2685083227670757e-06, "loss": 0.9978, "step": 5915 }, { "epoch": 0.78, "grad_norm": 1.2977303266525269, "learning_rate": 1.2670961484315996e-06, "loss": 0.756, "step": 5916 }, { "epoch": 0.78, "grad_norm": 1.276464581489563, "learning_rate": 1.265684646528138e-06, "loss": 0.8399, "step": 5917 }, { "epoch": 0.78, "grad_norm": 1.4174407720565796, "learning_rate": 1.2642738173109531e-06, "loss": 0.8076, "step": 5918 }, { "epoch": 0.78, "grad_norm": 0.5692289471626282, "learning_rate": 1.262863661034185e-06, "loss": 1.0158, "step": 5919 }, { "epoch": 0.78, "grad_norm": 0.5658013820648193, "learning_rate": 1.2614541779518563e-06, "loss": 0.9929, "step": 5920 }, { "epoch": 0.78, "grad_norm": 1.34083890914917, "learning_rate": 1.2600453683178659e-06, "loss": 0.8126, "step": 5921 }, { "epoch": 0.78, "grad_norm": 1.6255234479904175, "learning_rate": 1.2586372323859896e-06, "loss": 0.7811, "step": 5922 }, { "epoch": 0.78, "grad_norm": 1.3969268798828125, "learning_rate": 1.2572297704098873e-06, "loss": 0.807, "step": 5923 }, { "epoch": 0.78, "grad_norm": 1.3206826448440552, "learning_rate": 1.2558229826430922e-06, "loss": 0.7288, "step": 5924 }, { "epoch": 0.78, "grad_norm": 1.4748902320861816, "learning_rate": 1.2544168693390191e-06, "loss": 0.7856, "step": 5925 }, { "epoch": 0.78, "grad_norm": 1.8583861589431763, "learning_rate": 1.2530114307509577e-06, "loss": 0.7446, "step": 5926 }, { "epoch": 0.78, "grad_norm": 1.5012261867523193, "learning_rate": 1.251606667132082e-06, "loss": 0.8553, "step": 5927 }, { "epoch": 0.78, "grad_norm": 1.3297982215881348, "learning_rate": 1.25020257873544e-06, "loss": 0.7974, "step": 5928 }, { "epoch": 0.78, "grad_norm": 0.5955520868301392, "learning_rate": 1.248799165813957e-06, "loss": 1.0167, "step": 5929 }, { "epoch": 0.78, "grad_norm": 1.355428695678711, "learning_rate": 1.2473964286204426e-06, "loss": 0.7946, "step": 5930 }, { "epoch": 0.78, "grad_norm": 0.6081669926643372, "learning_rate": 1.245994367407577e-06, "loss": 1.015, "step": 5931 }, { "epoch": 0.78, "grad_norm": 1.3027476072311401, "learning_rate": 1.2445929824279262e-06, "loss": 0.7794, "step": 5932 }, { "epoch": 0.78, "grad_norm": 1.2524229288101196, "learning_rate": 1.2431922739339293e-06, "loss": 0.7487, "step": 5933 }, { "epoch": 0.78, "grad_norm": 0.5923529863357544, "learning_rate": 1.2417922421779027e-06, "loss": 1.0169, "step": 5934 }, { "epoch": 0.78, "grad_norm": 1.329870581626892, "learning_rate": 1.240392887412047e-06, "loss": 0.781, "step": 5935 }, { "epoch": 0.78, "grad_norm": 1.3054615259170532, "learning_rate": 1.2389942098884334e-06, "loss": 0.8039, "step": 5936 }, { "epoch": 0.78, "grad_norm": 0.5929877161979675, "learning_rate": 1.2375962098590178e-06, "loss": 1.0595, "step": 5937 }, { "epoch": 0.78, "grad_norm": 1.3064830303192139, "learning_rate": 1.2361988875756292e-06, "loss": 0.7654, "step": 5938 }, { "epoch": 0.78, "grad_norm": 1.3748586177825928, "learning_rate": 1.2348022432899754e-06, "loss": 0.7603, "step": 5939 }, { "epoch": 0.78, "grad_norm": 1.3037976026535034, "learning_rate": 1.2334062772536454e-06, "loss": 0.8476, "step": 5940 }, { "epoch": 0.78, "grad_norm": 1.4769887924194336, "learning_rate": 1.232010989718101e-06, "loss": 0.8378, "step": 5941 }, { "epoch": 0.78, "grad_norm": 1.3270574808120728, "learning_rate": 1.2306163809346867e-06, "loss": 0.8215, "step": 5942 }, { "epoch": 0.78, "grad_norm": 0.5941171050071716, "learning_rate": 1.2292224511546219e-06, "loss": 1.0221, "step": 5943 }, { "epoch": 0.78, "grad_norm": 1.502066731452942, "learning_rate": 1.2278292006290015e-06, "loss": 0.8173, "step": 5944 }, { "epoch": 0.78, "grad_norm": 1.3911794424057007, "learning_rate": 1.2264366296088043e-06, "loss": 0.7476, "step": 5945 }, { "epoch": 0.78, "grad_norm": 1.3631789684295654, "learning_rate": 1.2250447383448798e-06, "loss": 0.8785, "step": 5946 }, { "epoch": 0.78, "grad_norm": 1.3021173477172852, "learning_rate": 1.2236535270879617e-06, "loss": 0.8377, "step": 5947 }, { "epoch": 0.78, "grad_norm": 0.5843448042869568, "learning_rate": 1.2222629960886561e-06, "loss": 1.0123, "step": 5948 }, { "epoch": 0.78, "grad_norm": 1.4048362970352173, "learning_rate": 1.2208731455974466e-06, "loss": 0.7619, "step": 5949 }, { "epoch": 0.78, "grad_norm": 1.2811239957809448, "learning_rate": 1.2194839758646999e-06, "loss": 0.7598, "step": 5950 }, { "epoch": 0.78, "grad_norm": 1.2338786125183105, "learning_rate": 1.2180954871406514e-06, "loss": 0.7833, "step": 5951 }, { "epoch": 0.78, "grad_norm": 1.3096908330917358, "learning_rate": 1.216707679675423e-06, "loss": 0.7457, "step": 5952 }, { "epoch": 0.78, "grad_norm": 1.2282480001449585, "learning_rate": 1.2153205537190077e-06, "loss": 0.8194, "step": 5953 }, { "epoch": 0.78, "grad_norm": 1.501909852027893, "learning_rate": 1.2139341095212754e-06, "loss": 0.7597, "step": 5954 }, { "epoch": 0.78, "grad_norm": 1.3138833045959473, "learning_rate": 1.2125483473319782e-06, "loss": 0.8046, "step": 5955 }, { "epoch": 0.78, "grad_norm": 1.3845723867416382, "learning_rate": 1.2111632674007412e-06, "loss": 0.7439, "step": 5956 }, { "epoch": 0.78, "grad_norm": 1.4905935525894165, "learning_rate": 1.2097788699770658e-06, "loss": 0.7998, "step": 5957 }, { "epoch": 0.78, "grad_norm": 1.2697036266326904, "learning_rate": 1.2083951553103362e-06, "loss": 0.7504, "step": 5958 }, { "epoch": 0.78, "grad_norm": 1.3456772565841675, "learning_rate": 1.2070121236498078e-06, "loss": 0.8506, "step": 5959 }, { "epoch": 0.78, "grad_norm": 1.3317480087280273, "learning_rate": 1.2056297752446139e-06, "loss": 0.8015, "step": 5960 }, { "epoch": 0.78, "grad_norm": 1.5904892683029175, "learning_rate": 1.2042481103437686e-06, "loss": 0.7677, "step": 5961 }, { "epoch": 0.78, "grad_norm": 1.3641557693481445, "learning_rate": 1.202867129196158e-06, "loss": 0.8214, "step": 5962 }, { "epoch": 0.78, "grad_norm": 1.3063782453536987, "learning_rate": 1.2014868320505468e-06, "loss": 0.7943, "step": 5963 }, { "epoch": 0.78, "grad_norm": 1.4153708219528198, "learning_rate": 1.200107219155579e-06, "loss": 0.7966, "step": 5964 }, { "epoch": 0.78, "grad_norm": 1.364876627922058, "learning_rate": 1.198728290759772e-06, "loss": 0.8054, "step": 5965 }, { "epoch": 0.78, "grad_norm": 0.5890673995018005, "learning_rate": 1.19735004711152e-06, "loss": 1.0184, "step": 5966 }, { "epoch": 0.78, "grad_norm": 1.3107397556304932, "learning_rate": 1.1959724884590957e-06, "loss": 0.7372, "step": 5967 }, { "epoch": 0.78, "grad_norm": 1.2604182958602905, "learning_rate": 1.1945956150506482e-06, "loss": 0.7525, "step": 5968 }, { "epoch": 0.78, "grad_norm": 1.3694415092468262, "learning_rate": 1.1932194271342006e-06, "loss": 0.7922, "step": 5969 }, { "epoch": 0.78, "grad_norm": 1.3329870700836182, "learning_rate": 1.191843924957658e-06, "loss": 0.8723, "step": 5970 }, { "epoch": 0.78, "grad_norm": 1.2631592750549316, "learning_rate": 1.1904691087687964e-06, "loss": 0.8078, "step": 5971 }, { "epoch": 0.78, "grad_norm": 1.1816706657409668, "learning_rate": 1.189094978815269e-06, "loss": 0.7807, "step": 5972 }, { "epoch": 0.78, "grad_norm": 1.2607836723327637, "learning_rate": 1.1877215353446092e-06, "loss": 0.7945, "step": 5973 }, { "epoch": 0.78, "grad_norm": 1.5002049207687378, "learning_rate": 1.1863487786042215e-06, "loss": 0.7712, "step": 5974 }, { "epoch": 0.78, "grad_norm": 0.6107935309410095, "learning_rate": 1.1849767088413926e-06, "loss": 0.9754, "step": 5975 }, { "epoch": 0.78, "grad_norm": 1.2993624210357666, "learning_rate": 1.1836053263032804e-06, "loss": 0.8033, "step": 5976 }, { "epoch": 0.78, "grad_norm": 1.226801872253418, "learning_rate": 1.1822346312369198e-06, "loss": 0.7698, "step": 5977 }, { "epoch": 0.78, "grad_norm": 0.5889541506767273, "learning_rate": 1.1808646238892257e-06, "loss": 1.0221, "step": 5978 }, { "epoch": 0.78, "grad_norm": 1.2878388166427612, "learning_rate": 1.1794953045069835e-06, "loss": 0.7696, "step": 5979 }, { "epoch": 0.78, "grad_norm": 0.5721207857131958, "learning_rate": 1.1781266733368602e-06, "loss": 1.0207, "step": 5980 }, { "epoch": 0.78, "grad_norm": 1.3401927947998047, "learning_rate": 1.1767587306253952e-06, "loss": 0.7662, "step": 5981 }, { "epoch": 0.78, "grad_norm": 1.2749004364013672, "learning_rate": 1.1753914766190028e-06, "loss": 0.8133, "step": 5982 }, { "epoch": 0.78, "grad_norm": 0.5889723896980286, "learning_rate": 1.174024911563978e-06, "loss": 1.0206, "step": 5983 }, { "epoch": 0.78, "grad_norm": 1.3059889078140259, "learning_rate": 1.172659035706487e-06, "loss": 0.8004, "step": 5984 }, { "epoch": 0.78, "grad_norm": 0.5895581841468811, "learning_rate": 1.1712938492925757e-06, "loss": 0.9856, "step": 5985 }, { "epoch": 0.78, "grad_norm": 1.2614268064498901, "learning_rate": 1.1699293525681632e-06, "loss": 0.7762, "step": 5986 }, { "epoch": 0.78, "grad_norm": 0.5860753655433655, "learning_rate": 1.1685655457790435e-06, "loss": 1.0114, "step": 5987 }, { "epoch": 0.78, "grad_norm": 1.2715110778808594, "learning_rate": 1.1672024291708906e-06, "loss": 0.7637, "step": 5988 }, { "epoch": 0.78, "grad_norm": 1.2769508361816406, "learning_rate": 1.1658400029892502e-06, "loss": 0.7859, "step": 5989 }, { "epoch": 0.78, "grad_norm": 1.3273723125457764, "learning_rate": 1.1644782674795435e-06, "loss": 0.7484, "step": 5990 }, { "epoch": 0.79, "grad_norm": 1.2614575624465942, "learning_rate": 1.1631172228870706e-06, "loss": 0.8274, "step": 5991 }, { "epoch": 0.79, "grad_norm": 0.5705573558807373, "learning_rate": 1.161756869457004e-06, "loss": 1.002, "step": 5992 }, { "epoch": 0.79, "grad_norm": 0.5698104500770569, "learning_rate": 1.160397207434395e-06, "loss": 1.0248, "step": 5993 }, { "epoch": 0.79, "grad_norm": 1.2665637731552124, "learning_rate": 1.1590382370641663e-06, "loss": 0.7662, "step": 5994 }, { "epoch": 0.79, "grad_norm": 1.61711847782135, "learning_rate": 1.1576799585911175e-06, "loss": 0.8681, "step": 5995 }, { "epoch": 0.79, "grad_norm": 1.413494348526001, "learning_rate": 1.1563223722599266e-06, "loss": 0.7787, "step": 5996 }, { "epoch": 0.79, "grad_norm": 1.630902886390686, "learning_rate": 1.154965478315141e-06, "loss": 0.7871, "step": 5997 }, { "epoch": 0.79, "grad_norm": 1.3063185214996338, "learning_rate": 1.1536092770011898e-06, "loss": 0.809, "step": 5998 }, { "epoch": 0.79, "grad_norm": 1.318007469177246, "learning_rate": 1.1522537685623735e-06, "loss": 0.7663, "step": 5999 }, { "epoch": 0.79, "grad_norm": 1.2894973754882812, "learning_rate": 1.1508989532428683e-06, "loss": 0.7976, "step": 6000 }, { "epoch": 0.79, "grad_norm": 1.367490291595459, "learning_rate": 1.1495448312867235e-06, "loss": 0.7643, "step": 6001 }, { "epoch": 0.79, "grad_norm": 1.3551409244537354, "learning_rate": 1.1481914029378698e-06, "loss": 0.816, "step": 6002 }, { "epoch": 0.79, "grad_norm": 2.5250799655914307, "learning_rate": 1.1468386684401074e-06, "loss": 0.793, "step": 6003 }, { "epoch": 0.79, "grad_norm": 1.7191073894500732, "learning_rate": 1.1454866280371107e-06, "loss": 0.7943, "step": 6004 }, { "epoch": 0.79, "grad_norm": 1.2516992092132568, "learning_rate": 1.1441352819724355e-06, "loss": 0.766, "step": 6005 }, { "epoch": 0.79, "grad_norm": 1.272121787071228, "learning_rate": 1.1427846304895063e-06, "loss": 0.8085, "step": 6006 }, { "epoch": 0.79, "grad_norm": 1.261834740638733, "learning_rate": 1.1414346738316234e-06, "loss": 0.8055, "step": 6007 }, { "epoch": 0.79, "grad_norm": 1.1964716911315918, "learning_rate": 1.1400854122419659e-06, "loss": 0.7449, "step": 6008 }, { "epoch": 0.79, "grad_norm": 1.3677688837051392, "learning_rate": 1.138736845963584e-06, "loss": 0.81, "step": 6009 }, { "epoch": 0.79, "grad_norm": 2.0938217639923096, "learning_rate": 1.1373889752394018e-06, "loss": 0.7751, "step": 6010 }, { "epoch": 0.79, "grad_norm": 0.6066261529922485, "learning_rate": 1.1360418003122231e-06, "loss": 1.0281, "step": 6011 }, { "epoch": 0.79, "grad_norm": 1.3538470268249512, "learning_rate": 1.1346953214247197e-06, "loss": 0.8128, "step": 6012 }, { "epoch": 0.79, "grad_norm": 1.3024636507034302, "learning_rate": 1.1333495388194454e-06, "loss": 0.7479, "step": 6013 }, { "epoch": 0.79, "grad_norm": 0.5837989449501038, "learning_rate": 1.1320044527388223e-06, "loss": 1.0004, "step": 6014 }, { "epoch": 0.79, "grad_norm": 1.3887118101119995, "learning_rate": 1.1306600634251486e-06, "loss": 0.8216, "step": 6015 }, { "epoch": 0.79, "grad_norm": 0.5638948082923889, "learning_rate": 1.1293163711206e-06, "loss": 0.9738, "step": 6016 }, { "epoch": 0.79, "grad_norm": 0.5848968625068665, "learning_rate": 1.1279733760672235e-06, "loss": 1.0106, "step": 6017 }, { "epoch": 0.79, "grad_norm": 1.3822104930877686, "learning_rate": 1.1266310785069406e-06, "loss": 0.7414, "step": 6018 }, { "epoch": 0.79, "grad_norm": 1.425295352935791, "learning_rate": 1.1252894786815494e-06, "loss": 0.794, "step": 6019 }, { "epoch": 0.79, "grad_norm": 1.325864553451538, "learning_rate": 1.1239485768327195e-06, "loss": 0.7545, "step": 6020 }, { "epoch": 0.79, "grad_norm": 0.5730368494987488, "learning_rate": 1.1226083732019981e-06, "loss": 1.0269, "step": 6021 }, { "epoch": 0.79, "grad_norm": 1.3795198202133179, "learning_rate": 1.1212688680308042e-06, "loss": 0.8191, "step": 6022 }, { "epoch": 0.79, "grad_norm": 0.5726265907287598, "learning_rate": 1.1199300615604291e-06, "loss": 1.0024, "step": 6023 }, { "epoch": 0.79, "grad_norm": 1.3362727165222168, "learning_rate": 1.118591954032044e-06, "loss": 0.8283, "step": 6024 }, { "epoch": 0.79, "grad_norm": 1.3465731143951416, "learning_rate": 1.1172545456866873e-06, "loss": 0.8423, "step": 6025 }, { "epoch": 0.79, "grad_norm": 1.2604238986968994, "learning_rate": 1.115917836765279e-06, "loss": 0.8035, "step": 6026 }, { "epoch": 0.79, "grad_norm": 1.2898756265640259, "learning_rate": 1.1145818275086068e-06, "loss": 0.7893, "step": 6027 }, { "epoch": 0.79, "grad_norm": 0.5749006271362305, "learning_rate": 1.1132465181573332e-06, "loss": 1.0285, "step": 6028 }, { "epoch": 0.79, "grad_norm": 1.4596349000930786, "learning_rate": 1.1119119089519997e-06, "loss": 0.8338, "step": 6029 }, { "epoch": 0.79, "grad_norm": 1.4275457859039307, "learning_rate": 1.1105780001330147e-06, "loss": 0.7876, "step": 6030 }, { "epoch": 0.79, "grad_norm": 1.3254634141921997, "learning_rate": 1.1092447919406662e-06, "loss": 0.8211, "step": 6031 }, { "epoch": 0.79, "grad_norm": 1.3608224391937256, "learning_rate": 1.107912284615113e-06, "loss": 0.8205, "step": 6032 }, { "epoch": 0.79, "grad_norm": 0.5590670704841614, "learning_rate": 1.1065804783963856e-06, "loss": 0.9844, "step": 6033 }, { "epoch": 0.79, "grad_norm": 1.472211480140686, "learning_rate": 1.1052493735243952e-06, "loss": 0.793, "step": 6034 }, { "epoch": 0.79, "grad_norm": 1.324992299079895, "learning_rate": 1.1039189702389187e-06, "loss": 0.7646, "step": 6035 }, { "epoch": 0.79, "grad_norm": 0.5662062764167786, "learning_rate": 1.102589268779613e-06, "loss": 1.005, "step": 6036 }, { "epoch": 0.79, "grad_norm": 1.2731062173843384, "learning_rate": 1.1012602693860047e-06, "loss": 0.7944, "step": 6037 }, { "epoch": 0.79, "grad_norm": 1.2262043952941895, "learning_rate": 1.0999319722974928e-06, "loss": 0.8048, "step": 6038 }, { "epoch": 0.79, "grad_norm": 1.325471043586731, "learning_rate": 1.0986043777533556e-06, "loss": 0.7669, "step": 6039 }, { "epoch": 0.79, "grad_norm": 1.326026439666748, "learning_rate": 1.0972774859927398e-06, "loss": 0.8516, "step": 6040 }, { "epoch": 0.79, "grad_norm": 0.5901515483856201, "learning_rate": 1.095951297254665e-06, "loss": 0.9914, "step": 6041 }, { "epoch": 0.79, "grad_norm": 0.582938551902771, "learning_rate": 1.0946258117780294e-06, "loss": 1.0353, "step": 6042 }, { "epoch": 0.79, "grad_norm": 1.3223601579666138, "learning_rate": 1.0933010298016e-06, "loss": 0.8005, "step": 6043 }, { "epoch": 0.79, "grad_norm": 1.3686940670013428, "learning_rate": 1.0919769515640183e-06, "loss": 0.7987, "step": 6044 }, { "epoch": 0.79, "grad_norm": 0.5808348655700684, "learning_rate": 1.0906535773037969e-06, "loss": 1.0125, "step": 6045 }, { "epoch": 0.79, "grad_norm": 0.5851852893829346, "learning_rate": 1.0893309072593273e-06, "loss": 1.0114, "step": 6046 }, { "epoch": 0.79, "grad_norm": 0.5810065269470215, "learning_rate": 1.0880089416688682e-06, "loss": 1.0199, "step": 6047 }, { "epoch": 0.79, "grad_norm": 1.3313502073287964, "learning_rate": 1.086687680770554e-06, "loss": 0.7794, "step": 6048 }, { "epoch": 0.79, "grad_norm": 1.2004178762435913, "learning_rate": 1.0853671248023934e-06, "loss": 0.7863, "step": 6049 }, { "epoch": 0.79, "grad_norm": 1.323646903038025, "learning_rate": 1.0840472740022657e-06, "loss": 0.8466, "step": 6050 }, { "epoch": 0.79, "grad_norm": 1.3183708190917969, "learning_rate": 1.082728128607922e-06, "loss": 0.73, "step": 6051 }, { "epoch": 0.79, "grad_norm": 0.574501097202301, "learning_rate": 1.0814096888569931e-06, "loss": 0.9889, "step": 6052 }, { "epoch": 0.79, "grad_norm": 0.5617794394493103, "learning_rate": 1.0800919549869737e-06, "loss": 0.994, "step": 6053 }, { "epoch": 0.79, "grad_norm": 1.3638224601745605, "learning_rate": 1.0787749272352381e-06, "loss": 0.7536, "step": 6054 }, { "epoch": 0.79, "grad_norm": 0.5693981051445007, "learning_rate": 1.0774586058390314e-06, "loss": 1.0084, "step": 6055 }, { "epoch": 0.79, "grad_norm": 0.5692901611328125, "learning_rate": 1.0761429910354682e-06, "loss": 0.996, "step": 6056 }, { "epoch": 0.79, "grad_norm": 1.3116204738616943, "learning_rate": 1.074828083061542e-06, "loss": 0.7867, "step": 6057 }, { "epoch": 0.79, "grad_norm": 0.5833879113197327, "learning_rate": 1.0735138821541125e-06, "loss": 1.0071, "step": 6058 }, { "epoch": 0.79, "grad_norm": 1.2420251369476318, "learning_rate": 1.0722003885499184e-06, "loss": 0.7512, "step": 6059 }, { "epoch": 0.79, "grad_norm": 1.2680039405822754, "learning_rate": 1.0708876024855664e-06, "loss": 0.7861, "step": 6060 }, { "epoch": 0.79, "grad_norm": 0.5926173329353333, "learning_rate": 1.0695755241975347e-06, "loss": 1.0382, "step": 6061 }, { "epoch": 0.79, "grad_norm": 1.368786096572876, "learning_rate": 1.0682641539221805e-06, "loss": 0.7895, "step": 6062 }, { "epoch": 0.79, "grad_norm": 0.593513011932373, "learning_rate": 1.0669534918957258e-06, "loss": 0.9823, "step": 6063 }, { "epoch": 0.79, "grad_norm": 1.2760016918182373, "learning_rate": 1.0656435383542724e-06, "loss": 0.773, "step": 6064 }, { "epoch": 0.79, "grad_norm": 0.5822018384933472, "learning_rate": 1.0643342935337876e-06, "loss": 1.0071, "step": 6065 }, { "epoch": 0.79, "grad_norm": 1.25194251537323, "learning_rate": 1.0630257576701142e-06, "loss": 0.8142, "step": 6066 }, { "epoch": 0.79, "grad_norm": 1.284792423248291, "learning_rate": 1.0617179309989688e-06, "loss": 0.8609, "step": 6067 }, { "epoch": 0.8, "grad_norm": 1.4404727220535278, "learning_rate": 1.060410813755937e-06, "loss": 0.8302, "step": 6068 }, { "epoch": 0.8, "grad_norm": 1.319764494895935, "learning_rate": 1.0591044061764804e-06, "loss": 0.7777, "step": 6069 }, { "epoch": 0.8, "grad_norm": 0.571600079536438, "learning_rate": 1.0577987084959296e-06, "loss": 1.0383, "step": 6070 }, { "epoch": 0.8, "grad_norm": 0.5686830878257751, "learning_rate": 1.0564937209494863e-06, "loss": 0.9753, "step": 6071 }, { "epoch": 0.8, "grad_norm": 1.2360931634902954, "learning_rate": 1.0551894437722294e-06, "loss": 0.7996, "step": 6072 }, { "epoch": 0.8, "grad_norm": 0.5755792260169983, "learning_rate": 1.0538858771991056e-06, "loss": 0.9831, "step": 6073 }, { "epoch": 0.8, "grad_norm": 1.3405144214630127, "learning_rate": 1.052583021464933e-06, "loss": 0.7529, "step": 6074 }, { "epoch": 0.8, "grad_norm": 1.3758955001831055, "learning_rate": 1.0512808768044063e-06, "loss": 0.791, "step": 6075 }, { "epoch": 0.8, "grad_norm": 0.5777812004089355, "learning_rate": 1.0499794434520867e-06, "loss": 1.0183, "step": 6076 }, { "epoch": 0.8, "grad_norm": 1.2608489990234375, "learning_rate": 1.048678721642412e-06, "loss": 0.7685, "step": 6077 }, { "epoch": 0.8, "grad_norm": 1.282855749130249, "learning_rate": 1.0473787116096885e-06, "loss": 0.7712, "step": 6078 }, { "epoch": 0.8, "grad_norm": 0.5680213570594788, "learning_rate": 1.0460794135880941e-06, "loss": 0.9845, "step": 6079 }, { "epoch": 0.8, "grad_norm": 1.8012757301330566, "learning_rate": 1.0447808278116828e-06, "loss": 0.7821, "step": 6080 }, { "epoch": 0.8, "grad_norm": 0.5738063454627991, "learning_rate": 1.0434829545143748e-06, "loss": 0.9973, "step": 6081 }, { "epoch": 0.8, "grad_norm": 1.3021752834320068, "learning_rate": 1.042185793929964e-06, "loss": 0.7667, "step": 6082 }, { "epoch": 0.8, "grad_norm": 0.5925012826919556, "learning_rate": 1.0408893462921187e-06, "loss": 1.0225, "step": 6083 }, { "epoch": 0.8, "grad_norm": 1.321461796760559, "learning_rate": 1.0395936118343747e-06, "loss": 0.8161, "step": 6084 }, { "epoch": 0.8, "grad_norm": 1.2935646772384644, "learning_rate": 1.0382985907901416e-06, "loss": 0.8335, "step": 6085 }, { "epoch": 0.8, "grad_norm": 1.3711656332015991, "learning_rate": 1.037004283392698e-06, "loss": 0.8245, "step": 6086 }, { "epoch": 0.8, "grad_norm": 1.401155710220337, "learning_rate": 1.035710689875199e-06, "loss": 0.8156, "step": 6087 }, { "epoch": 0.8, "grad_norm": 1.3378831148147583, "learning_rate": 1.034417810470666e-06, "loss": 0.7991, "step": 6088 }, { "epoch": 0.8, "grad_norm": 1.3685660362243652, "learning_rate": 1.0331256454119927e-06, "loss": 0.7604, "step": 6089 }, { "epoch": 0.8, "grad_norm": 1.4341095685958862, "learning_rate": 1.0318341949319478e-06, "loss": 0.8409, "step": 6090 }, { "epoch": 0.8, "grad_norm": 1.2994322776794434, "learning_rate": 1.0305434592631658e-06, "loss": 0.786, "step": 6091 }, { "epoch": 0.8, "grad_norm": 1.3568379878997803, "learning_rate": 1.0292534386381581e-06, "loss": 0.8024, "step": 6092 }, { "epoch": 0.8, "grad_norm": 1.2298648357391357, "learning_rate": 1.0279641332893032e-06, "loss": 0.8023, "step": 6093 }, { "epoch": 0.8, "grad_norm": 0.5871374607086182, "learning_rate": 1.0266755434488501e-06, "loss": 0.9989, "step": 6094 }, { "epoch": 0.8, "grad_norm": 0.564456582069397, "learning_rate": 1.0253876693489234e-06, "loss": 1.0333, "step": 6095 }, { "epoch": 0.8, "grad_norm": 1.3213578462600708, "learning_rate": 1.024100511221514e-06, "loss": 0.7713, "step": 6096 }, { "epoch": 0.8, "grad_norm": 0.5811501741409302, "learning_rate": 1.022814069298489e-06, "loss": 1.0056, "step": 6097 }, { "epoch": 0.8, "grad_norm": 1.2443293333053589, "learning_rate": 1.0215283438115813e-06, "loss": 0.7812, "step": 6098 }, { "epoch": 0.8, "grad_norm": 0.5701594948768616, "learning_rate": 1.0202433349923957e-06, "loss": 1.0177, "step": 6099 }, { "epoch": 0.8, "grad_norm": 1.3879570960998535, "learning_rate": 1.0189590430724123e-06, "loss": 0.8078, "step": 6100 }, { "epoch": 0.8, "grad_norm": 1.3713561296463013, "learning_rate": 1.0176754682829759e-06, "loss": 0.826, "step": 6101 }, { "epoch": 0.8, "grad_norm": 1.3600250482559204, "learning_rate": 1.016392610855308e-06, "loss": 0.8283, "step": 6102 }, { "epoch": 0.8, "grad_norm": 1.3204537630081177, "learning_rate": 1.0151104710204957e-06, "loss": 0.8369, "step": 6103 }, { "epoch": 0.8, "grad_norm": 1.4391838312149048, "learning_rate": 1.0138290490094992e-06, "loss": 0.8142, "step": 6104 }, { "epoch": 0.8, "grad_norm": 1.3730846643447876, "learning_rate": 1.0125483450531509e-06, "loss": 0.7787, "step": 6105 }, { "epoch": 0.8, "grad_norm": 1.2786661386489868, "learning_rate": 1.011268359382151e-06, "loss": 0.8304, "step": 6106 }, { "epoch": 0.8, "grad_norm": 1.2834824323654175, "learning_rate": 1.009989092227071e-06, "loss": 0.7708, "step": 6107 }, { "epoch": 0.8, "grad_norm": 1.295733094215393, "learning_rate": 1.0087105438183554e-06, "loss": 0.797, "step": 6108 }, { "epoch": 0.8, "grad_norm": 1.229718804359436, "learning_rate": 1.007432714386315e-06, "loss": 0.79, "step": 6109 }, { "epoch": 0.8, "grad_norm": 1.2414944171905518, "learning_rate": 1.0061556041611358e-06, "loss": 0.726, "step": 6110 }, { "epoch": 0.8, "grad_norm": 1.3375904560089111, "learning_rate": 1.0048792133728713e-06, "loss": 0.7115, "step": 6111 }, { "epoch": 0.8, "grad_norm": 1.331658124923706, "learning_rate": 1.003603542251444e-06, "loss": 0.789, "step": 6112 }, { "epoch": 0.8, "grad_norm": 1.3193087577819824, "learning_rate": 1.0023285910266517e-06, "loss": 0.7942, "step": 6113 }, { "epoch": 0.8, "grad_norm": 1.323050856590271, "learning_rate": 1.0010543599281569e-06, "loss": 0.7839, "step": 6114 }, { "epoch": 0.8, "grad_norm": 1.269890308380127, "learning_rate": 9.997808491854977e-07, "loss": 0.7368, "step": 6115 }, { "epoch": 0.8, "grad_norm": 1.3524274826049805, "learning_rate": 9.985080590280783e-07, "loss": 0.77, "step": 6116 }, { "epoch": 0.8, "grad_norm": 0.617642343044281, "learning_rate": 9.972359896851736e-07, "loss": 1.0169, "step": 6117 }, { "epoch": 0.8, "grad_norm": 1.3202481269836426, "learning_rate": 9.95964641385932e-07, "loss": 0.7772, "step": 6118 }, { "epoch": 0.8, "grad_norm": 1.3102591037750244, "learning_rate": 9.94694014359368e-07, "loss": 0.8464, "step": 6119 }, { "epoch": 0.8, "grad_norm": 1.2959182262420654, "learning_rate": 9.934241088343671e-07, "loss": 0.799, "step": 6120 }, { "epoch": 0.8, "grad_norm": 1.308376431465149, "learning_rate": 9.921549250396884e-07, "loss": 0.783, "step": 6121 }, { "epoch": 0.8, "grad_norm": 0.6061881184577942, "learning_rate": 9.90886463203956e-07, "loss": 0.9934, "step": 6122 }, { "epoch": 0.8, "grad_norm": 0.5993947982788086, "learning_rate": 9.896187235556653e-07, "loss": 1.0183, "step": 6123 }, { "epoch": 0.8, "grad_norm": 1.2901314496994019, "learning_rate": 9.88351706323185e-07, "loss": 0.758, "step": 6124 }, { "epoch": 0.8, "grad_norm": 0.5810291767120361, "learning_rate": 9.8708541173475e-07, "loss": 1.0099, "step": 6125 }, { "epoch": 0.8, "grad_norm": 0.5789811015129089, "learning_rate": 9.85819840018466e-07, "loss": 0.9953, "step": 6126 }, { "epoch": 0.8, "grad_norm": 0.5648767352104187, "learning_rate": 9.845549914023067e-07, "loss": 1.002, "step": 6127 }, { "epoch": 0.8, "grad_norm": 1.2618038654327393, "learning_rate": 9.832908661141204e-07, "loss": 0.8049, "step": 6128 }, { "epoch": 0.8, "grad_norm": 0.5905703902244568, "learning_rate": 9.820274643816197e-07, "loss": 1.0135, "step": 6129 }, { "epoch": 0.8, "grad_norm": 1.3657526969909668, "learning_rate": 9.80764786432392e-07, "loss": 0.7688, "step": 6130 }, { "epoch": 0.8, "grad_norm": 1.302578091621399, "learning_rate": 9.795028324938899e-07, "loss": 0.7258, "step": 6131 }, { "epoch": 0.8, "grad_norm": 1.3464787006378174, "learning_rate": 9.78241602793436e-07, "loss": 0.8169, "step": 6132 }, { "epoch": 0.8, "grad_norm": 1.5420200824737549, "learning_rate": 9.769810975582268e-07, "loss": 0.8593, "step": 6133 }, { "epoch": 0.8, "grad_norm": 1.251189112663269, "learning_rate": 9.757213170153234e-07, "loss": 0.7347, "step": 6134 }, { "epoch": 0.8, "grad_norm": 1.3317124843597412, "learning_rate": 9.744622613916572e-07, "loss": 0.7984, "step": 6135 }, { "epoch": 0.8, "grad_norm": 1.4653650522232056, "learning_rate": 9.73203930914033e-07, "loss": 0.8023, "step": 6136 }, { "epoch": 0.8, "grad_norm": 1.3808934688568115, "learning_rate": 9.719463258091182e-07, "loss": 0.7503, "step": 6137 }, { "epoch": 0.8, "grad_norm": 0.5772836208343506, "learning_rate": 9.706894463034567e-07, "loss": 1.0046, "step": 6138 }, { "epoch": 0.8, "grad_norm": 1.7885345220565796, "learning_rate": 9.694332926234567e-07, "loss": 0.8598, "step": 6139 }, { "epoch": 0.8, "grad_norm": 1.361107349395752, "learning_rate": 9.681778649953966e-07, "loss": 0.8225, "step": 6140 }, { "epoch": 0.8, "grad_norm": 1.406907081604004, "learning_rate": 9.66923163645427e-07, "loss": 0.7616, "step": 6141 }, { "epoch": 0.8, "grad_norm": 1.2952024936676025, "learning_rate": 9.656691887995618e-07, "loss": 0.8144, "step": 6142 }, { "epoch": 0.8, "grad_norm": 1.3776792287826538, "learning_rate": 9.644159406836912e-07, "loss": 0.816, "step": 6143 }, { "epoch": 0.81, "grad_norm": 0.5802117586135864, "learning_rate": 9.631634195235696e-07, "loss": 1.0091, "step": 6144 }, { "epoch": 0.81, "grad_norm": 1.4785270690917969, "learning_rate": 9.619116255448196e-07, "loss": 0.8394, "step": 6145 }, { "epoch": 0.81, "grad_norm": 1.5055124759674072, "learning_rate": 9.606605589729385e-07, "loss": 0.7756, "step": 6146 }, { "epoch": 0.81, "grad_norm": 1.328513503074646, "learning_rate": 9.594102200332855e-07, "loss": 0.8168, "step": 6147 }, { "epoch": 0.81, "grad_norm": 1.2745280265808105, "learning_rate": 9.581606089510952e-07, "loss": 0.751, "step": 6148 }, { "epoch": 0.81, "grad_norm": 1.2610840797424316, "learning_rate": 9.569117259514666e-07, "loss": 0.7769, "step": 6149 }, { "epoch": 0.81, "grad_norm": 0.5721755623817444, "learning_rate": 9.556635712593682e-07, "loss": 1.006, "step": 6150 }, { "epoch": 0.81, "grad_norm": 1.3550083637237549, "learning_rate": 9.544161450996402e-07, "loss": 0.7645, "step": 6151 }, { "epoch": 0.81, "grad_norm": 1.7028005123138428, "learning_rate": 9.531694476969872e-07, "loss": 0.7754, "step": 6152 }, { "epoch": 0.81, "grad_norm": 1.3483515977859497, "learning_rate": 9.51923479275988e-07, "loss": 0.8366, "step": 6153 }, { "epoch": 0.81, "grad_norm": 1.3020504713058472, "learning_rate": 9.506782400610842e-07, "loss": 0.8463, "step": 6154 }, { "epoch": 0.81, "grad_norm": 1.3345447778701782, "learning_rate": 9.494337302765883e-07, "loss": 0.7891, "step": 6155 }, { "epoch": 0.81, "grad_norm": 0.5792674422264099, "learning_rate": 9.481899501466846e-07, "loss": 1.0061, "step": 6156 }, { "epoch": 0.81, "grad_norm": 1.3843839168548584, "learning_rate": 9.469468998954206e-07, "loss": 0.8106, "step": 6157 }, { "epoch": 0.81, "grad_norm": 0.5688767433166504, "learning_rate": 9.457045797467173e-07, "loss": 1.0025, "step": 6158 }, { "epoch": 0.81, "grad_norm": 1.6310837268829346, "learning_rate": 9.444629899243607e-07, "loss": 0.7544, "step": 6159 }, { "epoch": 0.81, "grad_norm": 1.3754609823226929, "learning_rate": 9.43222130652005e-07, "loss": 0.7803, "step": 6160 }, { "epoch": 0.81, "grad_norm": 1.3126211166381836, "learning_rate": 9.419820021531767e-07, "loss": 0.8039, "step": 6161 }, { "epoch": 0.81, "grad_norm": 1.3078187704086304, "learning_rate": 9.407426046512669e-07, "loss": 0.7915, "step": 6162 }, { "epoch": 0.81, "grad_norm": 1.272973656654358, "learning_rate": 9.395039383695359e-07, "loss": 0.769, "step": 6163 }, { "epoch": 0.81, "grad_norm": 1.3472896814346313, "learning_rate": 9.382660035311125e-07, "loss": 0.8005, "step": 6164 }, { "epoch": 0.81, "grad_norm": 1.2807612419128418, "learning_rate": 9.37028800358995e-07, "loss": 0.7264, "step": 6165 }, { "epoch": 0.81, "grad_norm": 0.5913375020027161, "learning_rate": 9.357923290760484e-07, "loss": 1.0059, "step": 6166 }, { "epoch": 0.81, "grad_norm": 0.6129940152168274, "learning_rate": 9.345565899050058e-07, "loss": 1.0175, "step": 6167 }, { "epoch": 0.81, "grad_norm": 1.3488343954086304, "learning_rate": 9.333215830684678e-07, "loss": 0.8108, "step": 6168 }, { "epoch": 0.81, "grad_norm": 1.283746600151062, "learning_rate": 9.320873087889071e-07, "loss": 0.8039, "step": 6169 }, { "epoch": 0.81, "grad_norm": 1.3007540702819824, "learning_rate": 9.308537672886581e-07, "loss": 0.8193, "step": 6170 }, { "epoch": 0.81, "grad_norm": 1.3006606101989746, "learning_rate": 9.296209587899291e-07, "loss": 0.7904, "step": 6171 }, { "epoch": 0.81, "grad_norm": 1.4396908283233643, "learning_rate": 9.283888835147936e-07, "loss": 0.7358, "step": 6172 }, { "epoch": 0.81, "grad_norm": 1.3969091176986694, "learning_rate": 9.271575416851908e-07, "loss": 0.7635, "step": 6173 }, { "epoch": 0.81, "grad_norm": 0.5608701705932617, "learning_rate": 9.259269335229338e-07, "loss": 0.9882, "step": 6174 }, { "epoch": 0.81, "grad_norm": 1.4980642795562744, "learning_rate": 9.246970592496968e-07, "loss": 0.8255, "step": 6175 }, { "epoch": 0.81, "grad_norm": 1.3033393621444702, "learning_rate": 9.234679190870277e-07, "loss": 0.8357, "step": 6176 }, { "epoch": 0.81, "grad_norm": 0.5914415717124939, "learning_rate": 9.222395132563378e-07, "loss": 1.015, "step": 6177 }, { "epoch": 0.81, "grad_norm": 1.3393656015396118, "learning_rate": 9.210118419789071e-07, "loss": 0.8038, "step": 6178 }, { "epoch": 0.81, "grad_norm": 1.4128754138946533, "learning_rate": 9.197849054758861e-07, "loss": 0.8492, "step": 6179 }, { "epoch": 0.81, "grad_norm": 1.251688003540039, "learning_rate": 9.185587039682886e-07, "loss": 0.7829, "step": 6180 }, { "epoch": 0.81, "grad_norm": 1.340187907218933, "learning_rate": 9.173332376770006e-07, "loss": 0.8113, "step": 6181 }, { "epoch": 0.81, "grad_norm": 1.3561654090881348, "learning_rate": 9.16108506822771e-07, "loss": 0.842, "step": 6182 }, { "epoch": 0.81, "grad_norm": 1.385120153427124, "learning_rate": 9.148845116262184e-07, "loss": 0.7875, "step": 6183 }, { "epoch": 0.81, "grad_norm": 0.5800309777259827, "learning_rate": 9.136612523078314e-07, "loss": 0.998, "step": 6184 }, { "epoch": 0.81, "grad_norm": 1.3408849239349365, "learning_rate": 9.124387290879605e-07, "loss": 0.8252, "step": 6185 }, { "epoch": 0.81, "grad_norm": 1.2979806661605835, "learning_rate": 9.112169421868289e-07, "loss": 0.7818, "step": 6186 }, { "epoch": 0.81, "grad_norm": 1.4668824672698975, "learning_rate": 9.099958918245249e-07, "loss": 0.8465, "step": 6187 }, { "epoch": 0.81, "grad_norm": 1.296739935874939, "learning_rate": 9.087755782210017e-07, "loss": 0.8542, "step": 6188 }, { "epoch": 0.81, "grad_norm": 1.344461441040039, "learning_rate": 9.075560015960855e-07, "loss": 0.7787, "step": 6189 }, { "epoch": 0.81, "grad_norm": 1.351488709449768, "learning_rate": 9.06337162169465e-07, "loss": 0.741, "step": 6190 }, { "epoch": 0.81, "grad_norm": 1.3347054719924927, "learning_rate": 9.051190601606968e-07, "loss": 0.7356, "step": 6191 }, { "epoch": 0.81, "grad_norm": 3.038213014602661, "learning_rate": 9.039016957892071e-07, "loss": 0.8105, "step": 6192 }, { "epoch": 0.81, "grad_norm": 1.453277587890625, "learning_rate": 9.026850692742856e-07, "loss": 0.7732, "step": 6193 }, { "epoch": 0.81, "grad_norm": 1.355228066444397, "learning_rate": 9.014691808350934e-07, "loss": 0.7906, "step": 6194 }, { "epoch": 0.81, "grad_norm": 1.403239130973816, "learning_rate": 9.002540306906549e-07, "loss": 0.835, "step": 6195 }, { "epoch": 0.81, "grad_norm": 1.321146011352539, "learning_rate": 8.990396190598622e-07, "loss": 0.7302, "step": 6196 }, { "epoch": 0.81, "grad_norm": 1.3554145097732544, "learning_rate": 8.978259461614769e-07, "loss": 0.8156, "step": 6197 }, { "epoch": 0.81, "grad_norm": 0.5976943969726562, "learning_rate": 8.966130122141242e-07, "loss": 1.0083, "step": 6198 }, { "epoch": 0.81, "grad_norm": 1.639625072479248, "learning_rate": 8.954008174362994e-07, "loss": 0.8253, "step": 6199 }, { "epoch": 0.81, "grad_norm": 1.3156384229660034, "learning_rate": 8.94189362046362e-07, "loss": 0.8485, "step": 6200 }, { "epoch": 0.81, "grad_norm": 0.5886688232421875, "learning_rate": 8.929786462625379e-07, "loss": 1.0183, "step": 6201 }, { "epoch": 0.81, "grad_norm": 1.4030237197875977, "learning_rate": 8.917686703029233e-07, "loss": 0.7735, "step": 6202 }, { "epoch": 0.81, "grad_norm": 1.3407561779022217, "learning_rate": 8.905594343854785e-07, "loss": 0.7869, "step": 6203 }, { "epoch": 0.81, "grad_norm": 1.2460527420043945, "learning_rate": 8.893509387280302e-07, "loss": 0.8454, "step": 6204 }, { "epoch": 0.81, "grad_norm": 1.2591795921325684, "learning_rate": 8.881431835482718e-07, "loss": 0.7027, "step": 6205 }, { "epoch": 0.81, "grad_norm": 1.4957209825515747, "learning_rate": 8.869361690637668e-07, "loss": 0.8413, "step": 6206 }, { "epoch": 0.81, "grad_norm": 1.311622142791748, "learning_rate": 8.857298954919402e-07, "loss": 0.8282, "step": 6207 }, { "epoch": 0.81, "grad_norm": 1.3186688423156738, "learning_rate": 8.84524363050085e-07, "loss": 0.7431, "step": 6208 }, { "epoch": 0.81, "grad_norm": 1.301349401473999, "learning_rate": 8.833195719553645e-07, "loss": 0.781, "step": 6209 }, { "epoch": 0.81, "grad_norm": 1.3446627855300903, "learning_rate": 8.821155224248035e-07, "loss": 0.7768, "step": 6210 }, { "epoch": 0.81, "grad_norm": 1.374963641166687, "learning_rate": 8.809122146752946e-07, "loss": 0.7928, "step": 6211 }, { "epoch": 0.81, "grad_norm": 1.449046015739441, "learning_rate": 8.797096489235995e-07, "loss": 0.7864, "step": 6212 }, { "epoch": 0.81, "grad_norm": 1.5630691051483154, "learning_rate": 8.785078253863422e-07, "loss": 0.8478, "step": 6213 }, { "epoch": 0.81, "grad_norm": 0.5695932507514954, "learning_rate": 8.773067442800164e-07, "loss": 1.0159, "step": 6214 }, { "epoch": 0.81, "grad_norm": 1.3195761442184448, "learning_rate": 8.7610640582098e-07, "loss": 0.8129, "step": 6215 }, { "epoch": 0.81, "grad_norm": 0.5904609560966492, "learning_rate": 8.749068102254559e-07, "loss": 0.9734, "step": 6216 }, { "epoch": 0.81, "grad_norm": 0.5654212832450867, "learning_rate": 8.737079577095386e-07, "loss": 1.0199, "step": 6217 }, { "epoch": 0.81, "grad_norm": 0.6047806143760681, "learning_rate": 8.725098484891825e-07, "loss": 0.9868, "step": 6218 }, { "epoch": 0.81, "grad_norm": 0.5765145421028137, "learning_rate": 8.713124827802105e-07, "loss": 1.0026, "step": 6219 }, { "epoch": 0.82, "grad_norm": 1.2927794456481934, "learning_rate": 8.70115860798314e-07, "loss": 0.8226, "step": 6220 }, { "epoch": 0.82, "grad_norm": 1.6756672859191895, "learning_rate": 8.689199827590455e-07, "loss": 0.8152, "step": 6221 }, { "epoch": 0.82, "grad_norm": 0.583538293838501, "learning_rate": 8.67724848877829e-07, "loss": 0.9837, "step": 6222 }, { "epoch": 0.82, "grad_norm": 1.3229966163635254, "learning_rate": 8.665304593699503e-07, "loss": 0.7964, "step": 6223 }, { "epoch": 0.82, "grad_norm": 2.1543266773223877, "learning_rate": 8.653368144505614e-07, "loss": 0.7947, "step": 6224 }, { "epoch": 0.82, "grad_norm": 1.3968933820724487, "learning_rate": 8.64143914334683e-07, "loss": 0.8934, "step": 6225 }, { "epoch": 0.82, "grad_norm": 3.0069429874420166, "learning_rate": 8.629517592371983e-07, "loss": 0.8473, "step": 6226 }, { "epoch": 0.82, "grad_norm": 0.5759766101837158, "learning_rate": 8.617603493728604e-07, "loss": 1.0187, "step": 6227 }, { "epoch": 0.82, "grad_norm": 1.386659860610962, "learning_rate": 8.605696849562834e-07, "loss": 0.7341, "step": 6228 }, { "epoch": 0.82, "grad_norm": 1.1665022373199463, "learning_rate": 8.593797662019493e-07, "loss": 0.7276, "step": 6229 }, { "epoch": 0.82, "grad_norm": 1.3093265295028687, "learning_rate": 8.581905933242068e-07, "loss": 0.7952, "step": 6230 }, { "epoch": 0.82, "grad_norm": 0.5891158580780029, "learning_rate": 8.570021665372686e-07, "loss": 1.0083, "step": 6231 }, { "epoch": 0.82, "grad_norm": 1.4379734992980957, "learning_rate": 8.558144860552147e-07, "loss": 0.7954, "step": 6232 }, { "epoch": 0.82, "grad_norm": 1.3612134456634521, "learning_rate": 8.546275520919894e-07, "loss": 0.7576, "step": 6233 }, { "epoch": 0.82, "grad_norm": 1.289618730545044, "learning_rate": 8.534413648614009e-07, "loss": 0.7895, "step": 6234 }, { "epoch": 0.82, "grad_norm": 1.3137823343276978, "learning_rate": 8.522559245771272e-07, "loss": 0.8087, "step": 6235 }, { "epoch": 0.82, "grad_norm": 1.5856035947799683, "learning_rate": 8.51071231452707e-07, "loss": 0.8677, "step": 6236 }, { "epoch": 0.82, "grad_norm": 1.3104639053344727, "learning_rate": 8.498872857015495e-07, "loss": 0.772, "step": 6237 }, { "epoch": 0.82, "grad_norm": 1.2736313343048096, "learning_rate": 8.487040875369251e-07, "loss": 0.806, "step": 6238 }, { "epoch": 0.82, "grad_norm": 0.5783594846725464, "learning_rate": 8.475216371719691e-07, "loss": 1.0205, "step": 6239 }, { "epoch": 0.82, "grad_norm": 0.5776702761650085, "learning_rate": 8.463399348196871e-07, "loss": 0.9765, "step": 6240 }, { "epoch": 0.82, "grad_norm": 1.5921727418899536, "learning_rate": 8.45158980692945e-07, "loss": 0.7404, "step": 6241 }, { "epoch": 0.82, "grad_norm": 1.3077398538589478, "learning_rate": 8.439787750044742e-07, "loss": 0.8138, "step": 6242 }, { "epoch": 0.82, "grad_norm": 2.7983717918395996, "learning_rate": 8.427993179668764e-07, "loss": 0.7631, "step": 6243 }, { "epoch": 0.82, "grad_norm": 1.3176352977752686, "learning_rate": 8.416206097926122e-07, "loss": 0.7902, "step": 6244 }, { "epoch": 0.82, "grad_norm": 1.2209609746932983, "learning_rate": 8.404426506940111e-07, "loss": 0.7794, "step": 6245 }, { "epoch": 0.82, "grad_norm": 1.5702427625656128, "learning_rate": 8.39265440883264e-07, "loss": 0.7898, "step": 6246 }, { "epoch": 0.82, "grad_norm": 1.3263423442840576, "learning_rate": 8.380889805724329e-07, "loss": 0.8105, "step": 6247 }, { "epoch": 0.82, "grad_norm": 1.5121827125549316, "learning_rate": 8.369132699734394e-07, "loss": 0.7731, "step": 6248 }, { "epoch": 0.82, "grad_norm": 1.334979772567749, "learning_rate": 8.357383092980703e-07, "loss": 0.7788, "step": 6249 }, { "epoch": 0.82, "grad_norm": 1.3099690675735474, "learning_rate": 8.345640987579817e-07, "loss": 0.7824, "step": 6250 }, { "epoch": 0.82, "grad_norm": 1.6373556852340698, "learning_rate": 8.333906385646906e-07, "loss": 0.702, "step": 6251 }, { "epoch": 0.82, "grad_norm": 1.3489748239517212, "learning_rate": 8.322179289295784e-07, "loss": 0.7951, "step": 6252 }, { "epoch": 0.82, "grad_norm": 1.2544219493865967, "learning_rate": 8.310459700638957e-07, "loss": 0.8141, "step": 6253 }, { "epoch": 0.82, "grad_norm": 1.3747020959854126, "learning_rate": 8.298747621787517e-07, "loss": 0.8287, "step": 6254 }, { "epoch": 0.82, "grad_norm": 1.3444294929504395, "learning_rate": 8.287043054851268e-07, "loss": 0.7992, "step": 6255 }, { "epoch": 0.82, "grad_norm": 1.4902831315994263, "learning_rate": 8.275346001938617e-07, "loss": 0.8277, "step": 6256 }, { "epoch": 0.82, "grad_norm": 0.5801898241043091, "learning_rate": 8.263656465156611e-07, "loss": 0.9937, "step": 6257 }, { "epoch": 0.82, "grad_norm": 1.3474385738372803, "learning_rate": 8.251974446610994e-07, "loss": 0.7682, "step": 6258 }, { "epoch": 0.82, "grad_norm": 1.6636170148849487, "learning_rate": 8.240299948406094e-07, "loss": 0.8158, "step": 6259 }, { "epoch": 0.82, "grad_norm": 1.514907717704773, "learning_rate": 8.22863297264494e-07, "loss": 0.7588, "step": 6260 }, { "epoch": 0.82, "grad_norm": 1.5914080142974854, "learning_rate": 8.216973521429161e-07, "loss": 0.8071, "step": 6261 }, { "epoch": 0.82, "grad_norm": 1.3211050033569336, "learning_rate": 8.205321596859045e-07, "loss": 0.7776, "step": 6262 }, { "epoch": 0.82, "grad_norm": 1.4829673767089844, "learning_rate": 8.193677201033551e-07, "loss": 0.8369, "step": 6263 }, { "epoch": 0.82, "grad_norm": 0.5804111957550049, "learning_rate": 8.182040336050234e-07, "loss": 0.9912, "step": 6264 }, { "epoch": 0.82, "grad_norm": 1.2800045013427734, "learning_rate": 8.170411004005341e-07, "loss": 0.8231, "step": 6265 }, { "epoch": 0.82, "grad_norm": 1.2298510074615479, "learning_rate": 8.158789206993728e-07, "loss": 0.7625, "step": 6266 }, { "epoch": 0.82, "grad_norm": 0.59841388463974, "learning_rate": 8.147174947108888e-07, "loss": 0.9859, "step": 6267 }, { "epoch": 0.82, "grad_norm": 0.5738855004310608, "learning_rate": 8.135568226442997e-07, "loss": 0.9841, "step": 6268 }, { "epoch": 0.82, "grad_norm": 0.5729653239250183, "learning_rate": 8.123969047086828e-07, "loss": 1.0178, "step": 6269 }, { "epoch": 0.82, "grad_norm": 1.3173435926437378, "learning_rate": 8.11237741112984e-07, "loss": 0.8117, "step": 6270 }, { "epoch": 0.82, "grad_norm": 0.5763965249061584, "learning_rate": 8.100793320660095e-07, "loss": 1.0265, "step": 6271 }, { "epoch": 0.82, "grad_norm": 1.3302538394927979, "learning_rate": 8.089216777764297e-07, "loss": 0.8077, "step": 6272 }, { "epoch": 0.82, "grad_norm": 1.3661686182022095, "learning_rate": 8.077647784527831e-07, "loss": 0.779, "step": 6273 }, { "epoch": 0.82, "grad_norm": 1.2655417919158936, "learning_rate": 8.066086343034663e-07, "loss": 0.8262, "step": 6274 }, { "epoch": 0.82, "grad_norm": 0.5725481510162354, "learning_rate": 8.054532455367459e-07, "loss": 1.001, "step": 6275 }, { "epoch": 0.82, "grad_norm": 1.259192943572998, "learning_rate": 8.042986123607488e-07, "loss": 0.7952, "step": 6276 }, { "epoch": 0.82, "grad_norm": 1.2937753200531006, "learning_rate": 8.031447349834643e-07, "loss": 0.7852, "step": 6277 }, { "epoch": 0.82, "grad_norm": 1.340461015701294, "learning_rate": 8.019916136127509e-07, "loss": 0.7786, "step": 6278 }, { "epoch": 0.82, "grad_norm": 1.4514542818069458, "learning_rate": 8.00839248456326e-07, "loss": 0.8106, "step": 6279 }, { "epoch": 0.82, "grad_norm": 0.5615679621696472, "learning_rate": 7.996876397217723e-07, "loss": 0.9966, "step": 6280 }, { "epoch": 0.82, "grad_norm": 1.3553141355514526, "learning_rate": 7.985367876165378e-07, "loss": 0.8203, "step": 6281 }, { "epoch": 0.82, "grad_norm": 0.5799345970153809, "learning_rate": 7.973866923479323e-07, "loss": 0.998, "step": 6282 }, { "epoch": 0.82, "grad_norm": 0.5731193423271179, "learning_rate": 7.962373541231283e-07, "loss": 1.0124, "step": 6283 }, { "epoch": 0.82, "grad_norm": 1.6452656984329224, "learning_rate": 7.950887731491669e-07, "loss": 0.8283, "step": 6284 }, { "epoch": 0.82, "grad_norm": 1.595248818397522, "learning_rate": 7.939409496329465e-07, "loss": 0.768, "step": 6285 }, { "epoch": 0.82, "grad_norm": 1.3816814422607422, "learning_rate": 7.927938837812338e-07, "loss": 0.8482, "step": 6286 }, { "epoch": 0.82, "grad_norm": 0.5770306587219238, "learning_rate": 7.916475758006543e-07, "loss": 0.9972, "step": 6287 }, { "epoch": 0.82, "grad_norm": 0.5862593650817871, "learning_rate": 7.905020258977031e-07, "loss": 1.0095, "step": 6288 }, { "epoch": 0.82, "grad_norm": 1.719469666481018, "learning_rate": 7.893572342787342e-07, "loss": 0.7666, "step": 6289 }, { "epoch": 0.82, "grad_norm": 1.2543666362762451, "learning_rate": 7.882132011499654e-07, "loss": 0.7544, "step": 6290 }, { "epoch": 0.82, "grad_norm": 0.5869361758232117, "learning_rate": 7.870699267174809e-07, "loss": 1.0133, "step": 6291 }, { "epoch": 0.82, "grad_norm": 0.5811319351196289, "learning_rate": 7.859274111872228e-07, "loss": 1.0167, "step": 6292 }, { "epoch": 0.82, "grad_norm": 0.5642651319503784, "learning_rate": 7.84785654765004e-07, "loss": 0.9817, "step": 6293 }, { "epoch": 0.82, "grad_norm": 1.324582815170288, "learning_rate": 7.836446576564932e-07, "loss": 0.844, "step": 6294 }, { "epoch": 0.82, "grad_norm": 0.5825919508934021, "learning_rate": 7.82504420067226e-07, "loss": 1.0073, "step": 6295 }, { "epoch": 0.82, "grad_norm": 1.328710913658142, "learning_rate": 7.813649422026026e-07, "loss": 0.75, "step": 6296 }, { "epoch": 0.83, "grad_norm": 1.4422303438186646, "learning_rate": 7.802262242678815e-07, "loss": 0.753, "step": 6297 }, { "epoch": 0.83, "grad_norm": 0.5722624659538269, "learning_rate": 7.790882664681909e-07, "loss": 1.003, "step": 6298 }, { "epoch": 0.83, "grad_norm": 1.4272574186325073, "learning_rate": 7.779510690085157e-07, "loss": 0.8017, "step": 6299 }, { "epoch": 0.83, "grad_norm": 1.249852180480957, "learning_rate": 7.76814632093707e-07, "loss": 0.757, "step": 6300 }, { "epoch": 0.83, "grad_norm": 0.557258129119873, "learning_rate": 7.7567895592848e-07, "loss": 1.0104, "step": 6301 }, { "epoch": 0.83, "grad_norm": 1.238234281539917, "learning_rate": 7.745440407174088e-07, "loss": 0.7295, "step": 6302 }, { "epoch": 0.83, "grad_norm": 0.5644288659095764, "learning_rate": 7.734098866649358e-07, "loss": 1.0007, "step": 6303 }, { "epoch": 0.83, "grad_norm": 0.5622349977493286, "learning_rate": 7.722764939753624e-07, "loss": 1.0043, "step": 6304 }, { "epoch": 0.83, "grad_norm": 0.6069150567054749, "learning_rate": 7.711438628528523e-07, "loss": 1.0167, "step": 6305 }, { "epoch": 0.83, "grad_norm": 1.282017469406128, "learning_rate": 7.700119935014367e-07, "loss": 0.8354, "step": 6306 }, { "epoch": 0.83, "grad_norm": 1.4332244396209717, "learning_rate": 7.688808861250047e-07, "loss": 0.7775, "step": 6307 }, { "epoch": 0.83, "grad_norm": 1.5242946147918701, "learning_rate": 7.677505409273084e-07, "loss": 0.7637, "step": 6308 }, { "epoch": 0.83, "grad_norm": 0.5633425116539001, "learning_rate": 7.666209581119677e-07, "loss": 0.9884, "step": 6309 }, { "epoch": 0.83, "grad_norm": 1.4692240953445435, "learning_rate": 7.654921378824582e-07, "loss": 0.8253, "step": 6310 }, { "epoch": 0.83, "grad_norm": 0.57038813829422, "learning_rate": 7.643640804421248e-07, "loss": 0.9928, "step": 6311 }, { "epoch": 0.83, "grad_norm": 0.5793243050575256, "learning_rate": 7.632367859941698e-07, "loss": 0.9938, "step": 6312 }, { "epoch": 0.83, "grad_norm": 1.4633991718292236, "learning_rate": 7.621102547416592e-07, "loss": 0.7712, "step": 6313 }, { "epoch": 0.83, "grad_norm": 1.43671715259552, "learning_rate": 7.609844868875249e-07, "loss": 0.8422, "step": 6314 }, { "epoch": 0.83, "grad_norm": 0.5593546032905579, "learning_rate": 7.598594826345557e-07, "loss": 1.0164, "step": 6315 }, { "epoch": 0.83, "grad_norm": 1.2034969329833984, "learning_rate": 7.587352421854088e-07, "loss": 0.8118, "step": 6316 }, { "epoch": 0.83, "grad_norm": 1.3076305389404297, "learning_rate": 7.576117657425997e-07, "loss": 0.7588, "step": 6317 }, { "epoch": 0.83, "grad_norm": 1.4100196361541748, "learning_rate": 7.564890535085056e-07, "loss": 0.797, "step": 6318 }, { "epoch": 0.83, "grad_norm": 1.3585578203201294, "learning_rate": 7.553671056853712e-07, "loss": 0.8074, "step": 6319 }, { "epoch": 0.83, "grad_norm": 1.4542145729064941, "learning_rate": 7.542459224752973e-07, "loss": 0.7821, "step": 6320 }, { "epoch": 0.83, "grad_norm": 0.563062310218811, "learning_rate": 7.53125504080252e-07, "loss": 0.99, "step": 6321 }, { "epoch": 0.83, "grad_norm": 0.5760937333106995, "learning_rate": 7.520058507020627e-07, "loss": 1.0078, "step": 6322 }, { "epoch": 0.83, "grad_norm": 1.3733282089233398, "learning_rate": 7.508869625424198e-07, "loss": 0.8185, "step": 6323 }, { "epoch": 0.83, "grad_norm": 1.3510328531265259, "learning_rate": 7.497688398028746e-07, "loss": 0.7879, "step": 6324 }, { "epoch": 0.83, "grad_norm": 3.40812349319458, "learning_rate": 7.48651482684844e-07, "loss": 0.8362, "step": 6325 }, { "epoch": 0.83, "grad_norm": 0.5671582221984863, "learning_rate": 7.47534891389603e-07, "loss": 0.9747, "step": 6326 }, { "epoch": 0.83, "grad_norm": 1.3245258331298828, "learning_rate": 7.464190661182918e-07, "loss": 0.7917, "step": 6327 }, { "epoch": 0.83, "grad_norm": 1.4348945617675781, "learning_rate": 7.453040070719087e-07, "loss": 0.7627, "step": 6328 }, { "epoch": 0.83, "grad_norm": 1.3153984546661377, "learning_rate": 7.441897144513188e-07, "loss": 0.7491, "step": 6329 }, { "epoch": 0.83, "grad_norm": 1.4534099102020264, "learning_rate": 7.430761884572457e-07, "loss": 0.8057, "step": 6330 }, { "epoch": 0.83, "grad_norm": 1.3221378326416016, "learning_rate": 7.419634292902772e-07, "loss": 0.8168, "step": 6331 }, { "epoch": 0.83, "grad_norm": 1.3669015169143677, "learning_rate": 7.408514371508613e-07, "loss": 0.7989, "step": 6332 }, { "epoch": 0.83, "grad_norm": 1.2853295803070068, "learning_rate": 7.397402122393066e-07, "loss": 0.7853, "step": 6333 }, { "epoch": 0.83, "grad_norm": 1.3213846683502197, "learning_rate": 7.386297547557874e-07, "loss": 0.7394, "step": 6334 }, { "epoch": 0.83, "grad_norm": 0.5899839401245117, "learning_rate": 7.375200649003372e-07, "loss": 1.0097, "step": 6335 }, { "epoch": 0.83, "grad_norm": 0.5715935230255127, "learning_rate": 7.364111428728504e-07, "loss": 1.0097, "step": 6336 }, { "epoch": 0.83, "grad_norm": 1.3874372243881226, "learning_rate": 7.353029888730856e-07, "loss": 0.7775, "step": 6337 }, { "epoch": 0.83, "grad_norm": 0.5850542187690735, "learning_rate": 7.341956031006597e-07, "loss": 1.0102, "step": 6338 }, { "epoch": 0.83, "grad_norm": 1.3041342496871948, "learning_rate": 7.330889857550566e-07, "loss": 0.7594, "step": 6339 }, { "epoch": 0.83, "grad_norm": 1.3840669393539429, "learning_rate": 7.319831370356162e-07, "loss": 0.8174, "step": 6340 }, { "epoch": 0.83, "grad_norm": 1.2807865142822266, "learning_rate": 7.308780571415408e-07, "loss": 0.7783, "step": 6341 }, { "epoch": 0.83, "grad_norm": 1.3074554204940796, "learning_rate": 7.297737462718985e-07, "loss": 0.7313, "step": 6342 }, { "epoch": 0.83, "grad_norm": 1.271699070930481, "learning_rate": 7.286702046256139e-07, "loss": 0.8273, "step": 6343 }, { "epoch": 0.83, "grad_norm": 1.348691463470459, "learning_rate": 7.275674324014764e-07, "loss": 0.8011, "step": 6344 }, { "epoch": 0.83, "grad_norm": 1.2617135047912598, "learning_rate": 7.264654297981355e-07, "loss": 0.7305, "step": 6345 }, { "epoch": 0.83, "grad_norm": 1.32358980178833, "learning_rate": 7.253641970140995e-07, "loss": 0.7721, "step": 6346 }, { "epoch": 0.83, "grad_norm": 1.288977026939392, "learning_rate": 7.242637342477437e-07, "loss": 0.7831, "step": 6347 }, { "epoch": 0.83, "grad_norm": 1.376826286315918, "learning_rate": 7.231640416972996e-07, "loss": 0.7565, "step": 6348 }, { "epoch": 0.83, "grad_norm": 1.3904669284820557, "learning_rate": 7.220651195608635e-07, "loss": 0.7344, "step": 6349 }, { "epoch": 0.83, "grad_norm": 1.2461448907852173, "learning_rate": 7.209669680363901e-07, "loss": 0.755, "step": 6350 }, { "epoch": 0.83, "grad_norm": 1.3598053455352783, "learning_rate": 7.19869587321696e-07, "loss": 0.81, "step": 6351 }, { "epoch": 0.83, "grad_norm": 1.4280829429626465, "learning_rate": 7.187729776144608e-07, "loss": 0.7581, "step": 6352 }, { "epoch": 0.83, "grad_norm": 1.3311545848846436, "learning_rate": 7.176771391122228e-07, "loss": 0.8243, "step": 6353 }, { "epoch": 0.83, "grad_norm": 0.5625548958778381, "learning_rate": 7.16582072012384e-07, "loss": 1.0036, "step": 6354 }, { "epoch": 0.83, "grad_norm": 1.3739908933639526, "learning_rate": 7.154877765122048e-07, "loss": 0.8451, "step": 6355 }, { "epoch": 0.83, "grad_norm": 1.304229497909546, "learning_rate": 7.143942528088066e-07, "loss": 0.7363, "step": 6356 }, { "epoch": 0.83, "grad_norm": 1.4100868701934814, "learning_rate": 7.13301501099175e-07, "loss": 0.7879, "step": 6357 }, { "epoch": 0.83, "grad_norm": 1.3083559274673462, "learning_rate": 7.122095215801522e-07, "loss": 0.788, "step": 6358 }, { "epoch": 0.83, "grad_norm": 0.6017577052116394, "learning_rate": 7.111183144484462e-07, "loss": 1.009, "step": 6359 }, { "epoch": 0.83, "grad_norm": 0.5715031623840332, "learning_rate": 7.100278799006216e-07, "loss": 1.0012, "step": 6360 }, { "epoch": 0.83, "grad_norm": 1.3711113929748535, "learning_rate": 7.089382181331045e-07, "loss": 0.7525, "step": 6361 }, { "epoch": 0.83, "grad_norm": 1.5639822483062744, "learning_rate": 7.078493293421845e-07, "loss": 0.8159, "step": 6362 }, { "epoch": 0.83, "grad_norm": 1.364988088607788, "learning_rate": 7.067612137240098e-07, "loss": 0.8149, "step": 6363 }, { "epoch": 0.83, "grad_norm": 1.3165377378463745, "learning_rate": 7.056738714745887e-07, "loss": 0.7302, "step": 6364 }, { "epoch": 0.83, "grad_norm": 1.2800250053405762, "learning_rate": 7.045873027897909e-07, "loss": 0.8244, "step": 6365 }, { "epoch": 0.83, "grad_norm": 1.3176155090332031, "learning_rate": 7.03501507865349e-07, "loss": 0.8027, "step": 6366 }, { "epoch": 0.83, "grad_norm": 1.2888495922088623, "learning_rate": 7.024164868968525e-07, "loss": 0.8118, "step": 6367 }, { "epoch": 0.83, "grad_norm": 0.5747355818748474, "learning_rate": 7.01332240079754e-07, "loss": 1.0044, "step": 6368 }, { "epoch": 0.83, "grad_norm": 1.701808214187622, "learning_rate": 7.002487676093645e-07, "loss": 0.8319, "step": 6369 }, { "epoch": 0.83, "grad_norm": 1.2839224338531494, "learning_rate": 6.991660696808594e-07, "loss": 0.818, "step": 6370 }, { "epoch": 0.83, "grad_norm": 0.5872213244438171, "learning_rate": 6.98084146489269e-07, "loss": 0.9922, "step": 6371 }, { "epoch": 0.83, "grad_norm": 1.3755979537963867, "learning_rate": 6.970029982294902e-07, "loss": 0.8294, "step": 6372 }, { "epoch": 0.84, "grad_norm": 0.5710161328315735, "learning_rate": 6.959226250962759e-07, "loss": 1.0039, "step": 6373 }, { "epoch": 0.84, "grad_norm": 0.5716859102249146, "learning_rate": 6.948430272842388e-07, "loss": 1.0262, "step": 6374 }, { "epoch": 0.84, "grad_norm": 1.3763967752456665, "learning_rate": 6.937642049878568e-07, "loss": 0.8395, "step": 6375 }, { "epoch": 0.84, "grad_norm": 1.2404714822769165, "learning_rate": 6.926861584014627e-07, "loss": 0.7654, "step": 6376 }, { "epoch": 0.84, "grad_norm": 0.5812270641326904, "learning_rate": 6.916088877192545e-07, "loss": 1.0247, "step": 6377 }, { "epoch": 0.84, "grad_norm": 1.4309134483337402, "learning_rate": 6.905323931352864e-07, "loss": 0.833, "step": 6378 }, { "epoch": 0.84, "grad_norm": 1.4771361351013184, "learning_rate": 6.894566748434734e-07, "loss": 0.8352, "step": 6379 }, { "epoch": 0.84, "grad_norm": 1.488682746887207, "learning_rate": 6.883817330375931e-07, "loss": 0.8026, "step": 6380 }, { "epoch": 0.84, "grad_norm": 1.255196213722229, "learning_rate": 6.873075679112806e-07, "loss": 0.745, "step": 6381 }, { "epoch": 0.84, "grad_norm": 1.3859655857086182, "learning_rate": 6.862341796580335e-07, "loss": 0.7916, "step": 6382 }, { "epoch": 0.84, "grad_norm": 1.3966842889785767, "learning_rate": 6.851615684712077e-07, "loss": 0.7769, "step": 6383 }, { "epoch": 0.84, "grad_norm": 0.5654913783073425, "learning_rate": 6.840897345440173e-07, "loss": 0.9922, "step": 6384 }, { "epoch": 0.84, "grad_norm": 1.2412333488464355, "learning_rate": 6.830186780695425e-07, "loss": 0.8095, "step": 6385 }, { "epoch": 0.84, "grad_norm": 1.294084072113037, "learning_rate": 6.819483992407166e-07, "loss": 0.7934, "step": 6386 }, { "epoch": 0.84, "grad_norm": 1.3549853563308716, "learning_rate": 6.808788982503379e-07, "loss": 0.7846, "step": 6387 }, { "epoch": 0.84, "grad_norm": 1.299815058708191, "learning_rate": 6.798101752910613e-07, "loss": 0.7173, "step": 6388 }, { "epoch": 0.84, "grad_norm": 1.37784743309021, "learning_rate": 6.787422305554015e-07, "loss": 0.7771, "step": 6389 }, { "epoch": 0.84, "grad_norm": 1.3809354305267334, "learning_rate": 6.776750642357372e-07, "loss": 0.8179, "step": 6390 }, { "epoch": 0.84, "grad_norm": 1.3081543445587158, "learning_rate": 6.766086765243013e-07, "loss": 0.8092, "step": 6391 }, { "epoch": 0.84, "grad_norm": 1.3595818281173706, "learning_rate": 6.755430676131919e-07, "loss": 0.7768, "step": 6392 }, { "epoch": 0.84, "grad_norm": 1.3227970600128174, "learning_rate": 6.744782376943615e-07, "loss": 0.7738, "step": 6393 }, { "epoch": 0.84, "grad_norm": 1.483218789100647, "learning_rate": 6.73414186959625e-07, "loss": 0.7773, "step": 6394 }, { "epoch": 0.84, "grad_norm": 1.4302576780319214, "learning_rate": 6.723509156006585e-07, "loss": 0.7723, "step": 6395 }, { "epoch": 0.84, "grad_norm": 1.2545514106750488, "learning_rate": 6.712884238089945e-07, "loss": 0.7633, "step": 6396 }, { "epoch": 0.84, "grad_norm": 0.5650122761726379, "learning_rate": 6.702267117760253e-07, "loss": 1.0012, "step": 6397 }, { "epoch": 0.84, "grad_norm": 0.582394003868103, "learning_rate": 6.691657796930074e-07, "loss": 0.9896, "step": 6398 }, { "epoch": 0.84, "grad_norm": 1.3776496648788452, "learning_rate": 6.681056277510495e-07, "loss": 0.7711, "step": 6399 }, { "epoch": 0.84, "grad_norm": 1.553163766860962, "learning_rate": 6.670462561411262e-07, "loss": 0.789, "step": 6400 }, { "epoch": 0.84, "grad_norm": 1.3533604145050049, "learning_rate": 6.659876650540687e-07, "loss": 0.8063, "step": 6401 }, { "epoch": 0.84, "grad_norm": 0.5814450979232788, "learning_rate": 6.64929854680566e-07, "loss": 1.017, "step": 6402 }, { "epoch": 0.84, "grad_norm": 1.4021135568618774, "learning_rate": 6.638728252111704e-07, "loss": 0.7935, "step": 6403 }, { "epoch": 0.84, "grad_norm": 1.3180439472198486, "learning_rate": 6.62816576836291e-07, "loss": 0.7658, "step": 6404 }, { "epoch": 0.84, "grad_norm": 0.5756613612174988, "learning_rate": 6.617611097461962e-07, "loss": 1.0078, "step": 6405 }, { "epoch": 0.84, "grad_norm": 1.4768955707550049, "learning_rate": 6.607064241310124e-07, "loss": 0.7569, "step": 6406 }, { "epoch": 0.84, "grad_norm": 0.5715254545211792, "learning_rate": 6.596525201807302e-07, "loss": 1.0064, "step": 6407 }, { "epoch": 0.84, "grad_norm": 1.380645513534546, "learning_rate": 6.585993980851945e-07, "loss": 0.8873, "step": 6408 }, { "epoch": 0.84, "grad_norm": 1.3097741603851318, "learning_rate": 6.575470580341098e-07, "loss": 0.747, "step": 6409 }, { "epoch": 0.84, "grad_norm": 1.2875298261642456, "learning_rate": 6.564955002170425e-07, "loss": 0.7758, "step": 6410 }, { "epoch": 0.84, "grad_norm": 1.2516248226165771, "learning_rate": 6.55444724823417e-07, "loss": 0.8589, "step": 6411 }, { "epoch": 0.84, "grad_norm": 1.3766353130340576, "learning_rate": 6.543947320425137e-07, "loss": 0.8074, "step": 6412 }, { "epoch": 0.84, "grad_norm": 0.5803237557411194, "learning_rate": 6.533455220634771e-07, "loss": 1.0063, "step": 6413 }, { "epoch": 0.84, "grad_norm": 1.3249845504760742, "learning_rate": 6.522970950753066e-07, "loss": 0.8784, "step": 6414 }, { "epoch": 0.84, "grad_norm": 1.4248160123825073, "learning_rate": 6.512494512668633e-07, "loss": 0.7632, "step": 6415 }, { "epoch": 0.84, "grad_norm": 0.5724253058433533, "learning_rate": 6.502025908268655e-07, "loss": 0.9932, "step": 6416 }, { "epoch": 0.84, "grad_norm": 1.3396645784378052, "learning_rate": 6.491565139438899e-07, "loss": 0.7829, "step": 6417 }, { "epoch": 0.84, "grad_norm": 1.3662902116775513, "learning_rate": 6.48111220806375e-07, "loss": 0.825, "step": 6418 }, { "epoch": 0.84, "grad_norm": 0.5835832953453064, "learning_rate": 6.470667116026141e-07, "loss": 1.028, "step": 6419 }, { "epoch": 0.84, "grad_norm": 1.394944429397583, "learning_rate": 6.460229865207635e-07, "loss": 0.8179, "step": 6420 }, { "epoch": 0.84, "grad_norm": 1.3166890144348145, "learning_rate": 6.449800457488354e-07, "loss": 0.7785, "step": 6421 }, { "epoch": 0.84, "grad_norm": 0.5932551622390747, "learning_rate": 6.439378894746995e-07, "loss": 1.0211, "step": 6422 }, { "epoch": 0.84, "grad_norm": 1.3026576042175293, "learning_rate": 6.428965178860891e-07, "loss": 0.7453, "step": 6423 }, { "epoch": 0.84, "grad_norm": 0.5521374940872192, "learning_rate": 6.418559311705919e-07, "loss": 1.0102, "step": 6424 }, { "epoch": 0.84, "grad_norm": 1.327549934387207, "learning_rate": 6.408161295156545e-07, "loss": 0.7484, "step": 6425 }, { "epoch": 0.84, "grad_norm": 1.254152536392212, "learning_rate": 6.397771131085845e-07, "loss": 0.8038, "step": 6426 }, { "epoch": 0.84, "grad_norm": 1.2409318685531616, "learning_rate": 6.387388821365453e-07, "loss": 0.7586, "step": 6427 }, { "epoch": 0.84, "grad_norm": 1.28146231174469, "learning_rate": 6.37701436786562e-07, "loss": 0.7835, "step": 6428 }, { "epoch": 0.84, "grad_norm": 1.334086298942566, "learning_rate": 6.366647772455154e-07, "loss": 0.8356, "step": 6429 }, { "epoch": 0.84, "grad_norm": 0.5601619482040405, "learning_rate": 6.35628903700144e-07, "loss": 0.9962, "step": 6430 }, { "epoch": 0.84, "grad_norm": 1.33486008644104, "learning_rate": 6.345938163370491e-07, "loss": 0.7577, "step": 6431 }, { "epoch": 0.84, "grad_norm": 1.360546350479126, "learning_rate": 6.335595153426854e-07, "loss": 0.7904, "step": 6432 }, { "epoch": 0.84, "grad_norm": 1.382261872291565, "learning_rate": 6.325260009033707e-07, "loss": 0.8061, "step": 6433 }, { "epoch": 0.84, "grad_norm": 1.4302315711975098, "learning_rate": 6.314932732052764e-07, "loss": 0.7646, "step": 6434 }, { "epoch": 0.84, "grad_norm": 1.3459560871124268, "learning_rate": 6.304613324344344e-07, "loss": 0.7808, "step": 6435 }, { "epoch": 0.84, "grad_norm": 1.4286580085754395, "learning_rate": 6.294301787767365e-07, "loss": 0.7791, "step": 6436 }, { "epoch": 0.84, "grad_norm": 1.2865303754806519, "learning_rate": 6.283998124179292e-07, "loss": 0.8344, "step": 6437 }, { "epoch": 0.84, "grad_norm": 1.3279920816421509, "learning_rate": 6.27370233543621e-07, "loss": 0.7576, "step": 6438 }, { "epoch": 0.84, "grad_norm": 1.3554351329803467, "learning_rate": 6.263414423392755e-07, "loss": 0.8501, "step": 6439 }, { "epoch": 0.84, "grad_norm": 0.5759028196334839, "learning_rate": 6.253134389902143e-07, "loss": 1.0105, "step": 6440 }, { "epoch": 0.84, "grad_norm": 1.331900954246521, "learning_rate": 6.242862236816205e-07, "loss": 0.7976, "step": 6441 }, { "epoch": 0.84, "grad_norm": 1.252202033996582, "learning_rate": 6.232597965985321e-07, "loss": 0.7848, "step": 6442 }, { "epoch": 0.84, "grad_norm": 0.5868815779685974, "learning_rate": 6.222341579258451e-07, "loss": 1.0396, "step": 6443 }, { "epoch": 0.84, "grad_norm": 0.5735925436019897, "learning_rate": 6.212093078483155e-07, "loss": 1.0198, "step": 6444 }, { "epoch": 0.84, "grad_norm": 0.5797534584999084, "learning_rate": 6.201852465505564e-07, "loss": 1.0021, "step": 6445 }, { "epoch": 0.84, "grad_norm": 1.4020627737045288, "learning_rate": 6.19161974217038e-07, "loss": 0.8107, "step": 6446 }, { "epoch": 0.84, "grad_norm": 0.5686027407646179, "learning_rate": 6.181394910320876e-07, "loss": 0.9944, "step": 6447 }, { "epoch": 0.84, "grad_norm": 0.5765783786773682, "learning_rate": 6.171177971798942e-07, "loss": 1.016, "step": 6448 }, { "epoch": 0.85, "grad_norm": 1.3181792497634888, "learning_rate": 6.16096892844501e-07, "loss": 0.8342, "step": 6449 }, { "epoch": 0.85, "grad_norm": 1.4761582612991333, "learning_rate": 6.150767782098088e-07, "loss": 0.8163, "step": 6450 }, { "epoch": 0.85, "grad_norm": 1.229125738143921, "learning_rate": 6.1405745345958e-07, "loss": 0.7295, "step": 6451 }, { "epoch": 0.85, "grad_norm": 1.3470278978347778, "learning_rate": 6.130389187774305e-07, "loss": 0.7906, "step": 6452 }, { "epoch": 0.85, "grad_norm": 1.3162237405776978, "learning_rate": 6.12021174346834e-07, "loss": 0.7513, "step": 6453 }, { "epoch": 0.85, "grad_norm": 1.395850419998169, "learning_rate": 6.110042203511268e-07, "loss": 0.8351, "step": 6454 }, { "epoch": 0.85, "grad_norm": 1.4031178951263428, "learning_rate": 6.099880569734967e-07, "loss": 0.7805, "step": 6455 }, { "epoch": 0.85, "grad_norm": 1.3008137941360474, "learning_rate": 6.08972684396994e-07, "loss": 0.7434, "step": 6456 }, { "epoch": 0.85, "grad_norm": 0.577159583568573, "learning_rate": 6.079581028045223e-07, "loss": 0.9819, "step": 6457 }, { "epoch": 0.85, "grad_norm": 1.4225080013275146, "learning_rate": 6.069443123788449e-07, "loss": 0.8237, "step": 6458 }, { "epoch": 0.85, "grad_norm": 1.3080788850784302, "learning_rate": 6.059313133025841e-07, "loss": 0.7797, "step": 6459 }, { "epoch": 0.85, "grad_norm": 0.5844112634658813, "learning_rate": 6.049191057582155e-07, "loss": 1.0151, "step": 6460 }, { "epoch": 0.85, "grad_norm": 1.6405692100524902, "learning_rate": 6.039076899280766e-07, "loss": 0.8028, "step": 6461 }, { "epoch": 0.85, "grad_norm": 1.3352142572402954, "learning_rate": 6.028970659943601e-07, "loss": 0.8221, "step": 6462 }, { "epoch": 0.85, "grad_norm": 1.4994747638702393, "learning_rate": 6.018872341391141e-07, "loss": 0.7541, "step": 6463 }, { "epoch": 0.85, "grad_norm": 1.2705589532852173, "learning_rate": 6.008781945442488e-07, "loss": 0.7502, "step": 6464 }, { "epoch": 0.85, "grad_norm": 1.3824870586395264, "learning_rate": 5.99869947391527e-07, "loss": 0.8257, "step": 6465 }, { "epoch": 0.85, "grad_norm": 1.3332685232162476, "learning_rate": 5.988624928625725e-07, "loss": 0.7744, "step": 6466 }, { "epoch": 0.85, "grad_norm": 1.342923879623413, "learning_rate": 5.978558311388639e-07, "loss": 0.8206, "step": 6467 }, { "epoch": 0.85, "grad_norm": 0.5639646053314209, "learning_rate": 5.96849962401736e-07, "loss": 1.0014, "step": 6468 }, { "epoch": 0.85, "grad_norm": 0.564798891544342, "learning_rate": 5.958448868323846e-07, "loss": 1.0097, "step": 6469 }, { "epoch": 0.85, "grad_norm": 0.5853749513626099, "learning_rate": 5.948406046118594e-07, "loss": 1.0123, "step": 6470 }, { "epoch": 0.85, "grad_norm": 1.4151952266693115, "learning_rate": 5.938371159210687e-07, "loss": 0.8117, "step": 6471 }, { "epoch": 0.85, "grad_norm": 2.107938766479492, "learning_rate": 5.928344209407777e-07, "loss": 0.7556, "step": 6472 }, { "epoch": 0.85, "grad_norm": 1.3193310499191284, "learning_rate": 5.918325198516067e-07, "loss": 0.8225, "step": 6473 }, { "epoch": 0.85, "grad_norm": 1.29067862033844, "learning_rate": 5.908314128340364e-07, "loss": 0.8054, "step": 6474 }, { "epoch": 0.85, "grad_norm": 1.4361451864242554, "learning_rate": 5.89831100068401e-07, "loss": 0.776, "step": 6475 }, { "epoch": 0.85, "grad_norm": 1.3686466217041016, "learning_rate": 5.888315817348955e-07, "loss": 0.8147, "step": 6476 }, { "epoch": 0.85, "grad_norm": 1.4988876581192017, "learning_rate": 5.878328580135679e-07, "loss": 0.8349, "step": 6477 }, { "epoch": 0.85, "grad_norm": 1.4796217679977417, "learning_rate": 5.868349290843239e-07, "loss": 0.7836, "step": 6478 }, { "epoch": 0.85, "grad_norm": 0.580723226070404, "learning_rate": 5.858377951269295e-07, "loss": 1.0132, "step": 6479 }, { "epoch": 0.85, "grad_norm": 1.312092900276184, "learning_rate": 5.848414563210036e-07, "loss": 0.8391, "step": 6480 }, { "epoch": 0.85, "grad_norm": 1.2824649810791016, "learning_rate": 5.838459128460216e-07, "loss": 0.815, "step": 6481 }, { "epoch": 0.85, "grad_norm": 1.3451807498931885, "learning_rate": 5.828511648813201e-07, "loss": 0.7963, "step": 6482 }, { "epoch": 0.85, "grad_norm": 0.5848342180252075, "learning_rate": 5.818572126060873e-07, "loss": 1.0415, "step": 6483 }, { "epoch": 0.85, "grad_norm": 1.3679380416870117, "learning_rate": 5.8086405619937e-07, "loss": 0.8006, "step": 6484 }, { "epoch": 0.85, "grad_norm": 1.4079482555389404, "learning_rate": 5.798716958400735e-07, "loss": 0.7803, "step": 6485 }, { "epoch": 0.85, "grad_norm": 1.4818421602249146, "learning_rate": 5.788801317069575e-07, "loss": 0.8016, "step": 6486 }, { "epoch": 0.85, "grad_norm": 1.3661935329437256, "learning_rate": 5.778893639786381e-07, "loss": 0.8094, "step": 6487 }, { "epoch": 0.85, "grad_norm": 1.3196179866790771, "learning_rate": 5.768993928335887e-07, "loss": 0.7564, "step": 6488 }, { "epoch": 0.85, "grad_norm": 1.6194216012954712, "learning_rate": 5.759102184501403e-07, "loss": 0.7272, "step": 6489 }, { "epoch": 0.85, "grad_norm": 1.3494818210601807, "learning_rate": 5.749218410064789e-07, "loss": 0.7996, "step": 6490 }, { "epoch": 0.85, "grad_norm": 0.5714158415794373, "learning_rate": 5.739342606806453e-07, "loss": 1.0083, "step": 6491 }, { "epoch": 0.85, "grad_norm": 1.3529719114303589, "learning_rate": 5.729474776505423e-07, "loss": 0.8281, "step": 6492 }, { "epoch": 0.85, "grad_norm": 1.9778244495391846, "learning_rate": 5.719614920939215e-07, "loss": 0.8162, "step": 6493 }, { "epoch": 0.85, "grad_norm": 1.3712315559387207, "learning_rate": 5.709763041883987e-07, "loss": 0.7253, "step": 6494 }, { "epoch": 0.85, "grad_norm": 1.4173943996429443, "learning_rate": 5.699919141114401e-07, "loss": 0.8333, "step": 6495 }, { "epoch": 0.85, "grad_norm": 0.563456118106842, "learning_rate": 5.690083220403692e-07, "loss": 0.9989, "step": 6496 }, { "epoch": 0.85, "grad_norm": 1.4365543127059937, "learning_rate": 5.680255281523694e-07, "loss": 0.7602, "step": 6497 }, { "epoch": 0.85, "grad_norm": 1.3687658309936523, "learning_rate": 5.67043532624475e-07, "loss": 0.8026, "step": 6498 }, { "epoch": 0.85, "grad_norm": 1.4961981773376465, "learning_rate": 5.660623356335815e-07, "loss": 0.8266, "step": 6499 }, { "epoch": 0.85, "grad_norm": 1.3002750873565674, "learning_rate": 5.650819373564376e-07, "loss": 0.8045, "step": 6500 }, { "epoch": 0.85, "grad_norm": 1.4003525972366333, "learning_rate": 5.641023379696469e-07, "loss": 0.7881, "step": 6501 }, { "epoch": 0.85, "grad_norm": 0.5660541653633118, "learning_rate": 5.631235376496735e-07, "loss": 0.9938, "step": 6502 }, { "epoch": 0.85, "grad_norm": 1.3780570030212402, "learning_rate": 5.621455365728329e-07, "loss": 0.7957, "step": 6503 }, { "epoch": 0.85, "grad_norm": 1.297361969947815, "learning_rate": 5.611683349153013e-07, "loss": 0.7811, "step": 6504 }, { "epoch": 0.85, "grad_norm": 1.3168710470199585, "learning_rate": 5.601919328531064e-07, "loss": 0.8031, "step": 6505 }, { "epoch": 0.85, "grad_norm": 0.5699018239974976, "learning_rate": 5.592163305621329e-07, "loss": 0.9929, "step": 6506 }, { "epoch": 0.85, "grad_norm": 1.3412408828735352, "learning_rate": 5.582415282181253e-07, "loss": 0.7494, "step": 6507 }, { "epoch": 0.85, "grad_norm": 1.4648374319076538, "learning_rate": 5.572675259966775e-07, "loss": 0.7847, "step": 6508 }, { "epoch": 0.85, "grad_norm": 0.5816910862922668, "learning_rate": 5.562943240732466e-07, "loss": 1.0118, "step": 6509 }, { "epoch": 0.85, "grad_norm": 1.546495795249939, "learning_rate": 5.553219226231393e-07, "loss": 0.7521, "step": 6510 }, { "epoch": 0.85, "grad_norm": 1.3995321989059448, "learning_rate": 5.543503218215196e-07, "loss": 0.7791, "step": 6511 }, { "epoch": 0.85, "grad_norm": 0.5720011591911316, "learning_rate": 5.533795218434113e-07, "loss": 1.0178, "step": 6512 }, { "epoch": 0.85, "grad_norm": 1.6577500104904175, "learning_rate": 5.524095228636889e-07, "loss": 0.863, "step": 6513 }, { "epoch": 0.85, "grad_norm": 1.4057364463806152, "learning_rate": 5.514403250570832e-07, "loss": 0.8184, "step": 6514 }, { "epoch": 0.85, "grad_norm": 0.5697141885757446, "learning_rate": 5.504719285981858e-07, "loss": 1.0077, "step": 6515 }, { "epoch": 0.85, "grad_norm": 2.0980405807495117, "learning_rate": 5.495043336614364e-07, "loss": 0.8422, "step": 6516 }, { "epoch": 0.85, "grad_norm": 1.3857768774032593, "learning_rate": 5.485375404211369e-07, "loss": 0.7671, "step": 6517 }, { "epoch": 0.85, "grad_norm": 1.4544843435287476, "learning_rate": 5.475715490514405e-07, "loss": 0.8146, "step": 6518 }, { "epoch": 0.85, "grad_norm": 0.5543302297592163, "learning_rate": 5.466063597263571e-07, "loss": 1.0096, "step": 6519 }, { "epoch": 0.85, "grad_norm": 1.3081713914871216, "learning_rate": 5.456419726197542e-07, "loss": 0.7462, "step": 6520 }, { "epoch": 0.85, "grad_norm": 1.2988812923431396, "learning_rate": 5.446783879053513e-07, "loss": 0.796, "step": 6521 }, { "epoch": 0.85, "grad_norm": 1.3315284252166748, "learning_rate": 5.437156057567267e-07, "loss": 0.8172, "step": 6522 }, { "epoch": 0.85, "grad_norm": 0.5641986727714539, "learning_rate": 5.427536263473115e-07, "loss": 0.9782, "step": 6523 }, { "epoch": 0.85, "grad_norm": 1.2802464962005615, "learning_rate": 5.417924498503935e-07, "loss": 0.8425, "step": 6524 }, { "epoch": 0.85, "grad_norm": 1.2742087841033936, "learning_rate": 5.408320764391146e-07, "loss": 0.8092, "step": 6525 }, { "epoch": 0.86, "grad_norm": 0.5668755769729614, "learning_rate": 5.398725062864751e-07, "loss": 1.0163, "step": 6526 }, { "epoch": 0.86, "grad_norm": 0.5560587048530579, "learning_rate": 5.389137395653271e-07, "loss": 1.0113, "step": 6527 }, { "epoch": 0.86, "grad_norm": 1.3488357067108154, "learning_rate": 5.3795577644838e-07, "loss": 0.8285, "step": 6528 }, { "epoch": 0.86, "grad_norm": 0.5797984600067139, "learning_rate": 5.36998617108197e-07, "loss": 1.0048, "step": 6529 }, { "epoch": 0.86, "grad_norm": 0.5657799243927002, "learning_rate": 5.36042261717199e-07, "loss": 0.9985, "step": 6530 }, { "epoch": 0.86, "grad_norm": 1.3438498973846436, "learning_rate": 5.350867104476581e-07, "loss": 0.8055, "step": 6531 }, { "epoch": 0.86, "grad_norm": 1.2991963624954224, "learning_rate": 5.341319634717063e-07, "loss": 0.7994, "step": 6532 }, { "epoch": 0.86, "grad_norm": 1.5298489332199097, "learning_rate": 5.331780209613275e-07, "loss": 0.8198, "step": 6533 }, { "epoch": 0.86, "grad_norm": 1.6837257146835327, "learning_rate": 5.322248830883603e-07, "loss": 0.846, "step": 6534 }, { "epoch": 0.86, "grad_norm": 1.684136986732483, "learning_rate": 5.31272550024502e-07, "loss": 0.8103, "step": 6535 }, { "epoch": 0.86, "grad_norm": 1.376766324043274, "learning_rate": 5.303210219412996e-07, "loss": 0.7402, "step": 6536 }, { "epoch": 0.86, "grad_norm": 1.2856957912445068, "learning_rate": 5.293702990101607e-07, "loss": 0.8036, "step": 6537 }, { "epoch": 0.86, "grad_norm": 1.4173306226730347, "learning_rate": 5.284203814023436e-07, "loss": 0.8183, "step": 6538 }, { "epoch": 0.86, "grad_norm": 1.3025643825531006, "learning_rate": 5.274712692889627e-07, "loss": 0.7786, "step": 6539 }, { "epoch": 0.86, "grad_norm": 0.5653534531593323, "learning_rate": 5.265229628409891e-07, "loss": 0.9867, "step": 6540 }, { "epoch": 0.86, "grad_norm": 1.424048662185669, "learning_rate": 5.255754622292464e-07, "loss": 0.8226, "step": 6541 }, { "epoch": 0.86, "grad_norm": 1.3828989267349243, "learning_rate": 5.246287676244138e-07, "loss": 0.729, "step": 6542 }, { "epoch": 0.86, "grad_norm": 1.422162652015686, "learning_rate": 5.236828791970266e-07, "loss": 0.7893, "step": 6543 }, { "epoch": 0.86, "grad_norm": 1.3770406246185303, "learning_rate": 5.227377971174718e-07, "loss": 0.7925, "step": 6544 }, { "epoch": 0.86, "grad_norm": 0.5745840668678284, "learning_rate": 5.217935215559961e-07, "loss": 0.992, "step": 6545 }, { "epoch": 0.86, "grad_norm": 1.1952831745147705, "learning_rate": 5.208500526826954e-07, "loss": 0.7167, "step": 6546 }, { "epoch": 0.86, "grad_norm": 1.327162742614746, "learning_rate": 5.199073906675234e-07, "loss": 0.7863, "step": 6547 }, { "epoch": 0.86, "grad_norm": 1.3072590827941895, "learning_rate": 5.189655356802886e-07, "loss": 0.8403, "step": 6548 }, { "epoch": 0.86, "grad_norm": 1.398939609527588, "learning_rate": 5.18024487890652e-07, "loss": 0.8292, "step": 6549 }, { "epoch": 0.86, "grad_norm": 1.2792524099349976, "learning_rate": 5.170842474681326e-07, "loss": 0.7901, "step": 6550 }, { "epoch": 0.86, "grad_norm": 1.5122140645980835, "learning_rate": 5.161448145821013e-07, "loss": 0.8311, "step": 6551 }, { "epoch": 0.86, "grad_norm": 0.5755533576011658, "learning_rate": 5.152061894017823e-07, "loss": 0.9981, "step": 6552 }, { "epoch": 0.86, "grad_norm": 0.570173442363739, "learning_rate": 5.142683720962588e-07, "loss": 1.0252, "step": 6553 }, { "epoch": 0.86, "grad_norm": 0.5899632573127747, "learning_rate": 5.133313628344638e-07, "loss": 1.0004, "step": 6554 }, { "epoch": 0.86, "grad_norm": 1.3191064596176147, "learning_rate": 5.123951617851891e-07, "loss": 0.8001, "step": 6555 }, { "epoch": 0.86, "grad_norm": 1.215479850769043, "learning_rate": 5.11459769117077e-07, "loss": 0.714, "step": 6556 }, { "epoch": 0.86, "grad_norm": 0.5791468620300293, "learning_rate": 5.105251849986249e-07, "loss": 1.01, "step": 6557 }, { "epoch": 0.86, "grad_norm": 1.3848836421966553, "learning_rate": 5.095914095981879e-07, "loss": 0.8611, "step": 6558 }, { "epoch": 0.86, "grad_norm": 1.4522383213043213, "learning_rate": 5.086584430839708e-07, "loss": 0.8429, "step": 6559 }, { "epoch": 0.86, "grad_norm": 1.5544782876968384, "learning_rate": 5.077262856240367e-07, "loss": 0.7385, "step": 6560 }, { "epoch": 0.86, "grad_norm": 1.3622593879699707, "learning_rate": 5.067949373863007e-07, "loss": 0.8229, "step": 6561 }, { "epoch": 0.86, "grad_norm": 0.5598632097244263, "learning_rate": 5.058643985385303e-07, "loss": 1.0058, "step": 6562 }, { "epoch": 0.86, "grad_norm": 1.2468072175979614, "learning_rate": 5.049346692483525e-07, "loss": 0.8129, "step": 6563 }, { "epoch": 0.86, "grad_norm": 0.5792249441146851, "learning_rate": 5.040057496832446e-07, "loss": 1.0147, "step": 6564 }, { "epoch": 0.86, "grad_norm": 0.5782164931297302, "learning_rate": 5.03077640010538e-07, "loss": 1.04, "step": 6565 }, { "epoch": 0.86, "grad_norm": 0.5741206407546997, "learning_rate": 5.021503403974187e-07, "loss": 0.9986, "step": 6566 }, { "epoch": 0.86, "grad_norm": 1.3742902278900146, "learning_rate": 5.01223851010929e-07, "loss": 0.8089, "step": 6567 }, { "epoch": 0.86, "grad_norm": 0.5573077201843262, "learning_rate": 5.002981720179622e-07, "loss": 0.9965, "step": 6568 }, { "epoch": 0.86, "grad_norm": 0.5734115839004517, "learning_rate": 4.993733035852666e-07, "loss": 1.0076, "step": 6569 }, { "epoch": 0.86, "grad_norm": 1.2795827388763428, "learning_rate": 4.984492458794454e-07, "loss": 0.8036, "step": 6570 }, { "epoch": 0.86, "grad_norm": 1.3185673952102661, "learning_rate": 4.975259990669551e-07, "loss": 0.7863, "step": 6571 }, { "epoch": 0.86, "grad_norm": 1.28365159034729, "learning_rate": 4.966035633141047e-07, "loss": 0.7922, "step": 6572 }, { "epoch": 0.86, "grad_norm": 1.5870633125305176, "learning_rate": 4.956819387870604e-07, "loss": 0.7768, "step": 6573 }, { "epoch": 0.86, "grad_norm": 1.350748896598816, "learning_rate": 4.947611256518392e-07, "loss": 0.7823, "step": 6574 }, { "epoch": 0.86, "grad_norm": 1.3584731817245483, "learning_rate": 4.93841124074313e-07, "loss": 0.8448, "step": 6575 }, { "epoch": 0.86, "grad_norm": 1.4706071615219116, "learning_rate": 4.929219342202091e-07, "loss": 0.7921, "step": 6576 }, { "epoch": 0.86, "grad_norm": 1.3930565118789673, "learning_rate": 4.920035562551045e-07, "loss": 0.8361, "step": 6577 }, { "epoch": 0.86, "grad_norm": 1.3271297216415405, "learning_rate": 4.910859903444354e-07, "loss": 0.8134, "step": 6578 }, { "epoch": 0.86, "grad_norm": 0.5964890122413635, "learning_rate": 4.901692366534872e-07, "loss": 1.0076, "step": 6579 }, { "epoch": 0.86, "grad_norm": 1.5623162984848022, "learning_rate": 4.892532953474e-07, "loss": 0.7945, "step": 6580 }, { "epoch": 0.86, "grad_norm": 1.4055778980255127, "learning_rate": 4.883381665911708e-07, "loss": 0.8142, "step": 6581 }, { "epoch": 0.86, "grad_norm": 1.2850110530853271, "learning_rate": 4.874238505496442e-07, "loss": 0.8105, "step": 6582 }, { "epoch": 0.86, "grad_norm": 0.5892237424850464, "learning_rate": 4.865103473875248e-07, "loss": 1.0124, "step": 6583 }, { "epoch": 0.86, "grad_norm": 0.5650753378868103, "learning_rate": 4.855976572693666e-07, "loss": 1.0095, "step": 6584 }, { "epoch": 0.86, "grad_norm": 0.5857819318771362, "learning_rate": 4.846857803595778e-07, "loss": 1.0052, "step": 6585 }, { "epoch": 0.86, "grad_norm": 1.2634825706481934, "learning_rate": 4.837747168224222e-07, "loss": 0.8284, "step": 6586 }, { "epoch": 0.86, "grad_norm": 1.3489644527435303, "learning_rate": 4.828644668220134e-07, "loss": 0.748, "step": 6587 }, { "epoch": 0.86, "grad_norm": 1.3591305017471313, "learning_rate": 4.819550305223225e-07, "loss": 0.7499, "step": 6588 }, { "epoch": 0.86, "grad_norm": 0.5898870229721069, "learning_rate": 4.810464080871719e-07, "loss": 1.018, "step": 6589 }, { "epoch": 0.86, "grad_norm": 0.5902773141860962, "learning_rate": 4.80138599680236e-07, "loss": 1.0078, "step": 6590 }, { "epoch": 0.86, "grad_norm": 0.575167179107666, "learning_rate": 4.792316054650458e-07, "loss": 1.0018, "step": 6591 }, { "epoch": 0.86, "grad_norm": 1.4042102098464966, "learning_rate": 4.78325425604983e-07, "loss": 0.784, "step": 6592 }, { "epoch": 0.86, "grad_norm": 1.3522498607635498, "learning_rate": 4.774200602632851e-07, "loss": 0.8107, "step": 6593 }, { "epoch": 0.86, "grad_norm": 1.380479335784912, "learning_rate": 4.765155096030405e-07, "loss": 0.6865, "step": 6594 }, { "epoch": 0.86, "grad_norm": 1.3817880153656006, "learning_rate": 4.7561177378719016e-07, "loss": 0.7839, "step": 6595 }, { "epoch": 0.86, "grad_norm": 1.3591402769088745, "learning_rate": 4.7470885297853265e-07, "loss": 0.7835, "step": 6596 }, { "epoch": 0.86, "grad_norm": 0.568789005279541, "learning_rate": 4.738067473397157e-07, "loss": 1.0171, "step": 6597 }, { "epoch": 0.86, "grad_norm": 1.2813853025436401, "learning_rate": 4.7290545703324033e-07, "loss": 0.8086, "step": 6598 }, { "epoch": 0.86, "grad_norm": 0.572780191898346, "learning_rate": 4.720049822214634e-07, "loss": 1.0028, "step": 6599 }, { "epoch": 0.86, "grad_norm": 1.641053318977356, "learning_rate": 4.7110532306659184e-07, "loss": 0.796, "step": 6600 }, { "epoch": 0.86, "grad_norm": 0.5772932767868042, "learning_rate": 4.702064797306888e-07, "loss": 1.0118, "step": 6601 }, { "epoch": 0.87, "grad_norm": 1.368701457977295, "learning_rate": 4.693084523756675e-07, "loss": 0.8115, "step": 6602 }, { "epoch": 0.87, "grad_norm": 1.317931890487671, "learning_rate": 4.684112411632946e-07, "loss": 0.7477, "step": 6603 }, { "epoch": 0.87, "grad_norm": 1.4836101531982422, "learning_rate": 4.675148462551926e-07, "loss": 0.7932, "step": 6604 }, { "epoch": 0.87, "grad_norm": 1.3320754766464233, "learning_rate": 4.666192678128334e-07, "loss": 0.8163, "step": 6605 }, { "epoch": 0.87, "grad_norm": 0.5833955407142639, "learning_rate": 4.6572450599754403e-07, "loss": 0.9776, "step": 6606 }, { "epoch": 0.87, "grad_norm": 1.411440372467041, "learning_rate": 4.6483056097050286e-07, "loss": 0.7183, "step": 6607 }, { "epoch": 0.87, "grad_norm": 1.357493281364441, "learning_rate": 4.639374328927426e-07, "loss": 0.797, "step": 6608 }, { "epoch": 0.87, "grad_norm": 1.6616384983062744, "learning_rate": 4.630451219251486e-07, "loss": 0.7755, "step": 6609 }, { "epoch": 0.87, "grad_norm": 1.4453405141830444, "learning_rate": 4.621536282284572e-07, "loss": 0.7402, "step": 6610 }, { "epoch": 0.87, "grad_norm": 1.3349932432174683, "learning_rate": 4.6126295196326e-07, "loss": 0.8792, "step": 6611 }, { "epoch": 0.87, "grad_norm": 1.3131369352340698, "learning_rate": 4.6037309329000077e-07, "loss": 0.7594, "step": 6612 }, { "epoch": 0.87, "grad_norm": 1.5776432752609253, "learning_rate": 4.5948405236897366e-07, "loss": 0.8031, "step": 6613 }, { "epoch": 0.87, "grad_norm": 0.5686835050582886, "learning_rate": 4.5859582936032877e-07, "loss": 0.9926, "step": 6614 }, { "epoch": 0.87, "grad_norm": 1.2948354482650757, "learning_rate": 4.577084244240665e-07, "loss": 0.777, "step": 6615 }, { "epoch": 0.87, "grad_norm": 0.575505793094635, "learning_rate": 4.568218377200423e-07, "loss": 1.0068, "step": 6616 }, { "epoch": 0.87, "grad_norm": 0.5674161314964294, "learning_rate": 4.5593606940796175e-07, "loss": 0.988, "step": 6617 }, { "epoch": 0.87, "grad_norm": 1.4318643808364868, "learning_rate": 4.550511196473828e-07, "loss": 0.7884, "step": 6618 }, { "epoch": 0.87, "grad_norm": 1.3999290466308594, "learning_rate": 4.541669885977196e-07, "loss": 0.7775, "step": 6619 }, { "epoch": 0.87, "grad_norm": 1.319238305091858, "learning_rate": 4.532836764182347e-07, "loss": 0.8217, "step": 6620 }, { "epoch": 0.87, "grad_norm": 1.2560969591140747, "learning_rate": 4.524011832680458e-07, "loss": 0.8415, "step": 6621 }, { "epoch": 0.87, "grad_norm": 1.2954057455062866, "learning_rate": 4.5151950930612144e-07, "loss": 0.7826, "step": 6622 }, { "epoch": 0.87, "grad_norm": 1.390169382095337, "learning_rate": 4.5063865469128276e-07, "loss": 0.8393, "step": 6623 }, { "epoch": 0.87, "grad_norm": 1.4491060972213745, "learning_rate": 4.497586195822046e-07, "loss": 0.7468, "step": 6624 }, { "epoch": 0.87, "grad_norm": 1.3088239431381226, "learning_rate": 4.488794041374128e-07, "loss": 0.7767, "step": 6625 }, { "epoch": 0.87, "grad_norm": 5.061985492706299, "learning_rate": 4.480010085152875e-07, "loss": 0.828, "step": 6626 }, { "epoch": 0.87, "grad_norm": 1.4603346586227417, "learning_rate": 4.471234328740581e-07, "loss": 0.7601, "step": 6627 }, { "epoch": 0.87, "grad_norm": 1.4501750469207764, "learning_rate": 4.4624667737180827e-07, "loss": 0.8343, "step": 6628 }, { "epoch": 0.87, "grad_norm": 1.31398606300354, "learning_rate": 4.4537074216647434e-07, "loss": 0.7782, "step": 6629 }, { "epoch": 0.87, "grad_norm": 0.5646894574165344, "learning_rate": 4.4449562741584353e-07, "loss": 1.0121, "step": 6630 }, { "epoch": 0.87, "grad_norm": 0.5678205490112305, "learning_rate": 4.436213332775557e-07, "loss": 1.0101, "step": 6631 }, { "epoch": 0.87, "grad_norm": 1.3356508016586304, "learning_rate": 4.42747859909104e-07, "loss": 0.7329, "step": 6632 }, { "epoch": 0.87, "grad_norm": 1.3668127059936523, "learning_rate": 4.4187520746783177e-07, "loss": 0.796, "step": 6633 }, { "epoch": 0.87, "grad_norm": 0.5614054799079895, "learning_rate": 4.410033761109361e-07, "loss": 0.9993, "step": 6634 }, { "epoch": 0.87, "grad_norm": 0.5791052579879761, "learning_rate": 4.401323659954665e-07, "loss": 1.0268, "step": 6635 }, { "epoch": 0.87, "grad_norm": 1.2929999828338623, "learning_rate": 4.392621772783212e-07, "loss": 0.8609, "step": 6636 }, { "epoch": 0.87, "grad_norm": 1.4761860370635986, "learning_rate": 4.383928101162549e-07, "loss": 0.7809, "step": 6637 }, { "epoch": 0.87, "grad_norm": 1.3446091413497925, "learning_rate": 4.3752426466587174e-07, "loss": 0.7622, "step": 6638 }, { "epoch": 0.87, "grad_norm": 1.2898160219192505, "learning_rate": 4.3665654108362877e-07, "loss": 0.7599, "step": 6639 }, { "epoch": 0.87, "grad_norm": 1.3351162672042847, "learning_rate": 4.357896395258343e-07, "loss": 0.7866, "step": 6640 }, { "epoch": 0.87, "grad_norm": 0.5554015636444092, "learning_rate": 4.349235601486479e-07, "loss": 0.9984, "step": 6641 }, { "epoch": 0.87, "grad_norm": 1.3205262422561646, "learning_rate": 4.340583031080836e-07, "loss": 0.793, "step": 6642 }, { "epoch": 0.87, "grad_norm": 1.4378845691680908, "learning_rate": 4.3319386856000565e-07, "loss": 0.8209, "step": 6643 }, { "epoch": 0.87, "grad_norm": 1.3013942241668701, "learning_rate": 4.323302566601284e-07, "loss": 0.7776, "step": 6644 }, { "epoch": 0.87, "grad_norm": 1.2750236988067627, "learning_rate": 4.3146746756402236e-07, "loss": 0.7416, "step": 6645 }, { "epoch": 0.87, "grad_norm": 1.3396613597869873, "learning_rate": 4.30605501427106e-07, "loss": 0.7342, "step": 6646 }, { "epoch": 0.87, "grad_norm": 1.515760898590088, "learning_rate": 4.297443584046507e-07, "loss": 0.7919, "step": 6647 }, { "epoch": 0.87, "grad_norm": 1.294283151626587, "learning_rate": 4.28884038651779e-07, "loss": 0.834, "step": 6648 }, { "epoch": 0.87, "grad_norm": 1.2521146535873413, "learning_rate": 4.280245423234675e-07, "loss": 0.7814, "step": 6649 }, { "epoch": 0.87, "grad_norm": 1.29165780544281, "learning_rate": 4.2716586957454174e-07, "loss": 0.7161, "step": 6650 }, { "epoch": 0.87, "grad_norm": 1.37162446975708, "learning_rate": 4.263080205596798e-07, "loss": 0.7947, "step": 6651 }, { "epoch": 0.87, "grad_norm": 1.6997190713882446, "learning_rate": 4.254509954334124e-07, "loss": 0.815, "step": 6652 }, { "epoch": 0.87, "grad_norm": 0.5686938762664795, "learning_rate": 4.2459479435011995e-07, "loss": 0.9956, "step": 6653 }, { "epoch": 0.87, "grad_norm": 0.5740544199943542, "learning_rate": 4.237394174640369e-07, "loss": 0.979, "step": 6654 }, { "epoch": 0.87, "grad_norm": 1.4443252086639404, "learning_rate": 4.228848649292472e-07, "loss": 0.8215, "step": 6655 }, { "epoch": 0.87, "grad_norm": 1.6556274890899658, "learning_rate": 4.22031136899686e-07, "loss": 0.769, "step": 6656 }, { "epoch": 0.87, "grad_norm": 0.5721803903579712, "learning_rate": 4.21178233529142e-07, "loss": 0.9919, "step": 6657 }, { "epoch": 0.87, "grad_norm": 1.5846774578094482, "learning_rate": 4.2032615497125396e-07, "loss": 0.792, "step": 6658 }, { "epoch": 0.87, "grad_norm": 0.5985622406005859, "learning_rate": 4.1947490137951076e-07, "loss": 1.0275, "step": 6659 }, { "epoch": 0.87, "grad_norm": 1.6981091499328613, "learning_rate": 4.186244729072564e-07, "loss": 0.8274, "step": 6660 }, { "epoch": 0.87, "grad_norm": 0.5859979391098022, "learning_rate": 4.177748697076822e-07, "loss": 0.9977, "step": 6661 }, { "epoch": 0.87, "grad_norm": 0.5639109015464783, "learning_rate": 4.169260919338347e-07, "loss": 1.024, "step": 6662 }, { "epoch": 0.87, "grad_norm": 1.2998236417770386, "learning_rate": 4.160781397386082e-07, "loss": 0.7956, "step": 6663 }, { "epoch": 0.87, "grad_norm": 0.5638720393180847, "learning_rate": 4.1523101327474945e-07, "loss": 1.007, "step": 6664 }, { "epoch": 0.87, "grad_norm": 1.3610804080963135, "learning_rate": 4.143847126948586e-07, "loss": 0.7704, "step": 6665 }, { "epoch": 0.87, "grad_norm": 1.3529527187347412, "learning_rate": 4.135392381513831e-07, "loss": 0.7417, "step": 6666 }, { "epoch": 0.87, "grad_norm": 1.38226318359375, "learning_rate": 4.126945897966256e-07, "loss": 0.7968, "step": 6667 }, { "epoch": 0.87, "grad_norm": 1.3436946868896484, "learning_rate": 4.118507677827371e-07, "loss": 0.7511, "step": 6668 }, { "epoch": 0.87, "grad_norm": 0.5782281160354614, "learning_rate": 4.110077722617195e-07, "loss": 0.9929, "step": 6669 }, { "epoch": 0.87, "grad_norm": 1.3106111288070679, "learning_rate": 4.1016560338542954e-07, "loss": 0.777, "step": 6670 }, { "epoch": 0.87, "grad_norm": 0.5732954144477844, "learning_rate": 4.093242613055709e-07, "loss": 0.9919, "step": 6671 }, { "epoch": 0.87, "grad_norm": 1.6637353897094727, "learning_rate": 4.084837461737007e-07, "loss": 0.8449, "step": 6672 }, { "epoch": 0.87, "grad_norm": 1.3016884326934814, "learning_rate": 4.0764405814122556e-07, "loss": 0.8195, "step": 6673 }, { "epoch": 0.87, "grad_norm": 1.2911585569381714, "learning_rate": 4.0680519735940394e-07, "loss": 0.7914, "step": 6674 }, { "epoch": 0.87, "grad_norm": 0.5637033581733704, "learning_rate": 4.0596716397934665e-07, "loss": 0.9773, "step": 6675 }, { "epoch": 0.87, "grad_norm": 1.2998234033584595, "learning_rate": 4.0512995815201185e-07, "loss": 0.7474, "step": 6676 }, { "epoch": 0.87, "grad_norm": 0.588404655456543, "learning_rate": 4.0429358002821275e-07, "loss": 1.0361, "step": 6677 }, { "epoch": 0.88, "grad_norm": 1.2757920026779175, "learning_rate": 4.034580297586105e-07, "loss": 0.7991, "step": 6678 }, { "epoch": 0.88, "grad_norm": 1.2843424081802368, "learning_rate": 4.0262330749371805e-07, "loss": 0.7564, "step": 6679 }, { "epoch": 0.88, "grad_norm": 0.5620914697647095, "learning_rate": 4.017894133839001e-07, "loss": 1.0103, "step": 6680 }, { "epoch": 0.88, "grad_norm": 1.4094116687774658, "learning_rate": 4.009563475793704e-07, "loss": 0.8297, "step": 6681 }, { "epoch": 0.88, "grad_norm": 0.5854859352111816, "learning_rate": 4.001241102301956e-07, "loss": 0.9856, "step": 6682 }, { "epoch": 0.88, "grad_norm": 1.282716155052185, "learning_rate": 3.992927014862913e-07, "loss": 0.8547, "step": 6683 }, { "epoch": 0.88, "grad_norm": 1.2791111469268799, "learning_rate": 3.9846212149742433e-07, "loss": 0.7891, "step": 6684 }, { "epoch": 0.88, "grad_norm": 0.552767276763916, "learning_rate": 3.9763237041321233e-07, "loss": 0.983, "step": 6685 }, { "epoch": 0.88, "grad_norm": 1.3387540578842163, "learning_rate": 3.9680344838312456e-07, "loss": 0.7495, "step": 6686 }, { "epoch": 0.88, "grad_norm": 1.3646479845046997, "learning_rate": 3.959753555564794e-07, "loss": 0.8399, "step": 6687 }, { "epoch": 0.88, "grad_norm": 1.240920901298523, "learning_rate": 3.9514809208244685e-07, "loss": 0.7384, "step": 6688 }, { "epoch": 0.88, "grad_norm": 1.3721970319747925, "learning_rate": 3.943216581100462e-07, "loss": 0.8014, "step": 6689 }, { "epoch": 0.88, "grad_norm": 0.566590428352356, "learning_rate": 3.934960537881499e-07, "loss": 0.9925, "step": 6690 }, { "epoch": 0.88, "grad_norm": 1.5587753057479858, "learning_rate": 3.9267127926547856e-07, "loss": 0.8146, "step": 6691 }, { "epoch": 0.88, "grad_norm": 1.2629493474960327, "learning_rate": 3.918473346906032e-07, "loss": 0.7543, "step": 6692 }, { "epoch": 0.88, "grad_norm": 0.5764155387878418, "learning_rate": 3.9102422021194853e-07, "loss": 1.0151, "step": 6693 }, { "epoch": 0.88, "grad_norm": 1.293750524520874, "learning_rate": 3.902019359777848e-07, "loss": 0.8133, "step": 6694 }, { "epoch": 0.88, "grad_norm": 1.653839111328125, "learning_rate": 3.893804821362379e-07, "loss": 0.7941, "step": 6695 }, { "epoch": 0.88, "grad_norm": 1.3280179500579834, "learning_rate": 3.885598588352807e-07, "loss": 0.8657, "step": 6696 }, { "epoch": 0.88, "grad_norm": 1.238293170928955, "learning_rate": 3.8774006622273597e-07, "loss": 0.8189, "step": 6697 }, { "epoch": 0.88, "grad_norm": 1.3003686666488647, "learning_rate": 3.869211044462806e-07, "loss": 0.7768, "step": 6698 }, { "epoch": 0.88, "grad_norm": 1.8978168964385986, "learning_rate": 3.861029736534366e-07, "loss": 0.7797, "step": 6699 }, { "epoch": 0.88, "grad_norm": 1.6005581617355347, "learning_rate": 3.8528567399158225e-07, "loss": 0.8054, "step": 6700 }, { "epoch": 0.88, "grad_norm": 1.29892098903656, "learning_rate": 3.844692056079413e-07, "loss": 0.7748, "step": 6701 }, { "epoch": 0.88, "grad_norm": 1.316267490386963, "learning_rate": 3.8365356864958903e-07, "loss": 0.8221, "step": 6702 }, { "epoch": 0.88, "grad_norm": 1.5415458679199219, "learning_rate": 3.8283876326345283e-07, "loss": 0.7841, "step": 6703 }, { "epoch": 0.88, "grad_norm": 1.316549301147461, "learning_rate": 3.8202478959630694e-07, "loss": 0.8795, "step": 6704 }, { "epoch": 0.88, "grad_norm": 1.2714810371398926, "learning_rate": 3.812116477947803e-07, "loss": 0.7772, "step": 6705 }, { "epoch": 0.88, "grad_norm": 1.3124080896377563, "learning_rate": 3.8039933800534724e-07, "loss": 0.7907, "step": 6706 }, { "epoch": 0.88, "grad_norm": 1.4778109788894653, "learning_rate": 3.7958786037433413e-07, "loss": 0.8011, "step": 6707 }, { "epoch": 0.88, "grad_norm": 0.5880893468856812, "learning_rate": 3.7877721504791963e-07, "loss": 1.025, "step": 6708 }, { "epoch": 0.88, "grad_norm": 1.898126244544983, "learning_rate": 3.7796740217212804e-07, "loss": 0.773, "step": 6709 }, { "epoch": 0.88, "grad_norm": 0.5791788697242737, "learning_rate": 3.771584218928387e-07, "loss": 0.992, "step": 6710 }, { "epoch": 0.88, "grad_norm": 1.8124028444290161, "learning_rate": 3.7635027435577734e-07, "loss": 0.7638, "step": 6711 }, { "epoch": 0.88, "grad_norm": 1.4038804769515991, "learning_rate": 3.755429597065197e-07, "loss": 0.7935, "step": 6712 }, { "epoch": 0.88, "grad_norm": 1.3439767360687256, "learning_rate": 3.747364780904944e-07, "loss": 0.8251, "step": 6713 }, { "epoch": 0.88, "grad_norm": 1.3194085359573364, "learning_rate": 3.7393082965297754e-07, "loss": 0.7756, "step": 6714 }, { "epoch": 0.88, "grad_norm": 1.3655339479446411, "learning_rate": 3.731260145390941e-07, "loss": 0.7624, "step": 6715 }, { "epoch": 0.88, "grad_norm": 1.3892273902893066, "learning_rate": 3.7232203289382374e-07, "loss": 0.8054, "step": 6716 }, { "epoch": 0.88, "grad_norm": 0.5673899054527283, "learning_rate": 3.715188848619894e-07, "loss": 0.9996, "step": 6717 }, { "epoch": 0.88, "grad_norm": 1.3154135942459106, "learning_rate": 3.707165705882704e-07, "loss": 0.7519, "step": 6718 }, { "epoch": 0.88, "grad_norm": 1.4543832540512085, "learning_rate": 3.6991509021719163e-07, "loss": 0.7623, "step": 6719 }, { "epoch": 0.88, "grad_norm": 1.2605581283569336, "learning_rate": 3.6911444389312766e-07, "loss": 0.795, "step": 6720 }, { "epoch": 0.88, "grad_norm": 1.2255284786224365, "learning_rate": 3.683146317603059e-07, "loss": 0.7755, "step": 6721 }, { "epoch": 0.88, "grad_norm": 1.3671444654464722, "learning_rate": 3.6751565396280055e-07, "loss": 0.7715, "step": 6722 }, { "epoch": 0.88, "grad_norm": 1.3588005304336548, "learning_rate": 3.6671751064453756e-07, "loss": 0.8292, "step": 6723 }, { "epoch": 0.88, "grad_norm": 0.5853177905082703, "learning_rate": 3.659202019492908e-07, "loss": 1.0192, "step": 6724 }, { "epoch": 0.88, "grad_norm": 1.405012607574463, "learning_rate": 3.6512372802068486e-07, "loss": 0.8706, "step": 6725 }, { "epoch": 0.88, "grad_norm": 1.361867070198059, "learning_rate": 3.643280890021933e-07, "loss": 0.7466, "step": 6726 }, { "epoch": 0.88, "grad_norm": 0.5570843815803528, "learning_rate": 3.6353328503714036e-07, "loss": 1.0215, "step": 6727 }, { "epoch": 0.88, "grad_norm": 1.9243186712265015, "learning_rate": 3.6273931626869927e-07, "loss": 0.8041, "step": 6728 }, { "epoch": 0.88, "grad_norm": 0.5834921002388, "learning_rate": 3.6194618283989183e-07, "loss": 0.9697, "step": 6729 }, { "epoch": 0.88, "grad_norm": 1.4061843156814575, "learning_rate": 3.6115388489359137e-07, "loss": 0.7989, "step": 6730 }, { "epoch": 0.88, "grad_norm": 1.4894931316375732, "learning_rate": 3.603624225725194e-07, "loss": 0.783, "step": 6731 }, { "epoch": 0.88, "grad_norm": 0.5739057064056396, "learning_rate": 3.5957179601924574e-07, "loss": 1.0238, "step": 6732 }, { "epoch": 0.88, "grad_norm": 1.4782277345657349, "learning_rate": 3.587820053761931e-07, "loss": 0.8259, "step": 6733 }, { "epoch": 0.88, "grad_norm": 1.2851173877716064, "learning_rate": 3.5799305078563104e-07, "loss": 0.797, "step": 6734 }, { "epoch": 0.88, "grad_norm": 1.436591386795044, "learning_rate": 3.57204932389677e-07, "loss": 0.7615, "step": 6735 }, { "epoch": 0.88, "grad_norm": 1.4266740083694458, "learning_rate": 3.5641765033030294e-07, "loss": 0.7754, "step": 6736 }, { "epoch": 0.88, "grad_norm": 2.304534673690796, "learning_rate": 3.5563120474932435e-07, "loss": 0.7991, "step": 6737 }, { "epoch": 0.88, "grad_norm": 0.5669252872467041, "learning_rate": 3.548455957884106e-07, "loss": 1.0095, "step": 6738 }, { "epoch": 0.88, "grad_norm": 1.508277416229248, "learning_rate": 3.54060823589078e-07, "loss": 0.7815, "step": 6739 }, { "epoch": 0.88, "grad_norm": 1.3577282428741455, "learning_rate": 3.532768882926918e-07, "loss": 0.7011, "step": 6740 }, { "epoch": 0.88, "grad_norm": 0.5753604173660278, "learning_rate": 3.5249379004046847e-07, "loss": 1.0338, "step": 6741 }, { "epoch": 0.88, "grad_norm": 2.231173515319824, "learning_rate": 3.5171152897347125e-07, "loss": 0.7613, "step": 6742 }, { "epoch": 0.88, "grad_norm": 0.5643740296363831, "learning_rate": 3.5093010523261583e-07, "loss": 1.0027, "step": 6743 }, { "epoch": 0.88, "grad_norm": 0.5657911896705627, "learning_rate": 3.5014951895866347e-07, "loss": 1.0104, "step": 6744 }, { "epoch": 0.88, "grad_norm": 1.416417121887207, "learning_rate": 3.4936977029222607e-07, "loss": 0.7767, "step": 6745 }, { "epoch": 0.88, "grad_norm": 0.5620242953300476, "learning_rate": 3.485908593737658e-07, "loss": 0.9973, "step": 6746 }, { "epoch": 0.88, "grad_norm": 0.5848862528800964, "learning_rate": 3.478127863435926e-07, "loss": 1.0098, "step": 6747 }, { "epoch": 0.88, "grad_norm": 1.4371715784072876, "learning_rate": 3.4703555134186496e-07, "loss": 0.745, "step": 6748 }, { "epoch": 0.88, "grad_norm": 0.5697247982025146, "learning_rate": 3.462591545085925e-07, "loss": 0.9813, "step": 6749 }, { "epoch": 0.88, "grad_norm": 1.3408024311065674, "learning_rate": 3.454835959836317e-07, "loss": 0.8356, "step": 6750 }, { "epoch": 0.88, "grad_norm": 1.273849368095398, "learning_rate": 3.447088759066891e-07, "loss": 0.7246, "step": 6751 }, { "epoch": 0.88, "grad_norm": 1.471662163734436, "learning_rate": 3.439349944173209e-07, "loss": 0.7944, "step": 6752 }, { "epoch": 0.88, "grad_norm": 1.3738250732421875, "learning_rate": 3.4316195165492947e-07, "loss": 0.839, "step": 6753 }, { "epoch": 0.88, "grad_norm": 1.6916700601577759, "learning_rate": 3.4238974775876954e-07, "loss": 0.8135, "step": 6754 }, { "epoch": 0.89, "grad_norm": 0.5604421496391296, "learning_rate": 3.4161838286794256e-07, "loss": 0.9763, "step": 6755 }, { "epoch": 0.89, "grad_norm": 1.3554527759552002, "learning_rate": 3.4084785712139967e-07, "loss": 0.789, "step": 6756 }, { "epoch": 0.89, "grad_norm": 1.4432140588760376, "learning_rate": 3.4007817065794146e-07, "loss": 0.8418, "step": 6757 }, { "epoch": 0.89, "grad_norm": 1.2687126398086548, "learning_rate": 3.3930932361621426e-07, "loss": 0.7832, "step": 6758 }, { "epoch": 0.89, "grad_norm": 0.5707587599754333, "learning_rate": 3.385413161347173e-07, "loss": 1.0049, "step": 6759 }, { "epoch": 0.89, "grad_norm": 0.5778396129608154, "learning_rate": 3.37774148351796e-07, "loss": 1.0057, "step": 6760 }, { "epoch": 0.89, "grad_norm": 1.5036507844924927, "learning_rate": 3.3700782040564594e-07, "loss": 0.8193, "step": 6761 }, { "epoch": 0.89, "grad_norm": 0.5607236623764038, "learning_rate": 3.3624233243431004e-07, "loss": 0.9978, "step": 6762 }, { "epoch": 0.89, "grad_norm": 1.3219143152236938, "learning_rate": 3.3547768457568017e-07, "loss": 0.7513, "step": 6763 }, { "epoch": 0.89, "grad_norm": 1.3014214038848877, "learning_rate": 3.3471387696749844e-07, "loss": 0.7388, "step": 6764 }, { "epoch": 0.89, "grad_norm": 1.3322495222091675, "learning_rate": 3.3395090974735423e-07, "loss": 0.7553, "step": 6765 }, { "epoch": 0.89, "grad_norm": 0.5700018405914307, "learning_rate": 3.331887830526853e-07, "loss": 0.9888, "step": 6766 }, { "epoch": 0.89, "grad_norm": 1.5241154432296753, "learning_rate": 3.324274970207775e-07, "loss": 0.7525, "step": 6767 }, { "epoch": 0.89, "grad_norm": 1.316815733909607, "learning_rate": 3.3166705178876825e-07, "loss": 0.7977, "step": 6768 }, { "epoch": 0.89, "grad_norm": 1.3016785383224487, "learning_rate": 3.309074474936402e-07, "loss": 0.7723, "step": 6769 }, { "epoch": 0.89, "grad_norm": 0.583554208278656, "learning_rate": 3.3014868427222513e-07, "loss": 0.9836, "step": 6770 }, { "epoch": 0.89, "grad_norm": 0.5667036175727844, "learning_rate": 3.2939076226120524e-07, "loss": 1.0031, "step": 6771 }, { "epoch": 0.89, "grad_norm": 1.3745378255844116, "learning_rate": 3.2863368159710975e-07, "loss": 0.7983, "step": 6772 }, { "epoch": 0.89, "grad_norm": 1.4240857362747192, "learning_rate": 3.2787744241631556e-07, "loss": 0.8088, "step": 6773 }, { "epoch": 0.89, "grad_norm": 1.3648990392684937, "learning_rate": 3.2712204485504994e-07, "loss": 0.8108, "step": 6774 }, { "epoch": 0.89, "grad_norm": 0.5749523639678955, "learning_rate": 3.263674890493873e-07, "loss": 0.9885, "step": 6775 }, { "epoch": 0.89, "grad_norm": 1.6064362525939941, "learning_rate": 3.256137751352495e-07, "loss": 0.7949, "step": 6776 }, { "epoch": 0.89, "grad_norm": 0.5624083876609802, "learning_rate": 3.2486090324840957e-07, "loss": 1.0218, "step": 6777 }, { "epoch": 0.89, "grad_norm": 1.2871674299240112, "learning_rate": 3.241088735244857e-07, "loss": 0.7536, "step": 6778 }, { "epoch": 0.89, "grad_norm": 1.3030165433883667, "learning_rate": 3.2335768609894735e-07, "loss": 0.7474, "step": 6779 }, { "epoch": 0.89, "grad_norm": 1.3670275211334229, "learning_rate": 3.226073411071096e-07, "loss": 0.8092, "step": 6780 }, { "epoch": 0.89, "grad_norm": 1.5379582643508911, "learning_rate": 3.21857838684137e-07, "loss": 0.8443, "step": 6781 }, { "epoch": 0.89, "grad_norm": 1.6201484203338623, "learning_rate": 3.2110917896504277e-07, "loss": 0.7977, "step": 6782 }, { "epoch": 0.89, "grad_norm": 1.3806884288787842, "learning_rate": 3.2036136208468684e-07, "loss": 0.7539, "step": 6783 }, { "epoch": 0.89, "grad_norm": 1.3500224351882935, "learning_rate": 3.196143881777802e-07, "loss": 0.8198, "step": 6784 }, { "epoch": 0.89, "grad_norm": 1.373577356338501, "learning_rate": 3.188682573788787e-07, "loss": 0.8153, "step": 6785 }, { "epoch": 0.89, "grad_norm": 1.629842758178711, "learning_rate": 3.18122969822387e-07, "loss": 0.7318, "step": 6786 }, { "epoch": 0.89, "grad_norm": 1.7040927410125732, "learning_rate": 3.173785256425599e-07, "loss": 0.7335, "step": 6787 }, { "epoch": 0.89, "grad_norm": 1.420386552810669, "learning_rate": 3.166349249734979e-07, "loss": 0.775, "step": 6788 }, { "epoch": 0.89, "grad_norm": 0.5741977095603943, "learning_rate": 3.158921679491522e-07, "loss": 1.0046, "step": 6789 }, { "epoch": 0.89, "grad_norm": 1.3860310316085815, "learning_rate": 3.151502547033186e-07, "loss": 0.8316, "step": 6790 }, { "epoch": 0.89, "grad_norm": 1.3896547555923462, "learning_rate": 3.144091853696435e-07, "loss": 0.7756, "step": 6791 }, { "epoch": 0.89, "grad_norm": 1.2843544483184814, "learning_rate": 3.1366896008162064e-07, "loss": 0.7915, "step": 6792 }, { "epoch": 0.89, "grad_norm": 1.2796025276184082, "learning_rate": 3.129295789725906e-07, "loss": 0.7686, "step": 6793 }, { "epoch": 0.89, "grad_norm": 0.5555355548858643, "learning_rate": 3.121910421757446e-07, "loss": 0.999, "step": 6794 }, { "epoch": 0.89, "grad_norm": 0.5629092454910278, "learning_rate": 3.1145334982411845e-07, "loss": 0.9842, "step": 6795 }, { "epoch": 0.89, "grad_norm": 0.5876675248146057, "learning_rate": 3.1071650205059755e-07, "loss": 0.978, "step": 6796 }, { "epoch": 0.89, "grad_norm": 0.5467488765716553, "learning_rate": 3.0998049898791685e-07, "loss": 0.9874, "step": 6797 }, { "epoch": 0.89, "grad_norm": 1.3142004013061523, "learning_rate": 3.0924534076865464e-07, "loss": 0.8063, "step": 6798 }, { "epoch": 0.89, "grad_norm": 1.4331101179122925, "learning_rate": 3.0851102752524176e-07, "loss": 0.7981, "step": 6799 }, { "epoch": 0.89, "grad_norm": 1.365189790725708, "learning_rate": 3.07777559389954e-07, "loss": 0.7683, "step": 6800 }, { "epoch": 0.89, "grad_norm": 0.577072024345398, "learning_rate": 3.070449364949152e-07, "loss": 1.0027, "step": 6801 }, { "epoch": 0.89, "grad_norm": 1.2610220909118652, "learning_rate": 3.0631315897209854e-07, "loss": 0.8281, "step": 6802 }, { "epoch": 0.89, "grad_norm": 1.3460348844528198, "learning_rate": 3.055822269533237e-07, "loss": 0.8064, "step": 6803 }, { "epoch": 0.89, "grad_norm": 1.667944073677063, "learning_rate": 3.048521405702565e-07, "loss": 0.7684, "step": 6804 }, { "epoch": 0.89, "grad_norm": 0.5501877665519714, "learning_rate": 3.0412289995441437e-07, "loss": 1.0115, "step": 6805 }, { "epoch": 0.89, "grad_norm": 1.4973629713058472, "learning_rate": 3.0339450523715896e-07, "loss": 0.7816, "step": 6806 }, { "epoch": 0.89, "grad_norm": 0.5758787989616394, "learning_rate": 3.026669565497009e-07, "loss": 0.9998, "step": 6807 }, { "epoch": 0.89, "grad_norm": 1.287689447402954, "learning_rate": 3.0194025402309757e-07, "loss": 0.8133, "step": 6808 }, { "epoch": 0.89, "grad_norm": 0.576535165309906, "learning_rate": 3.012143977882553e-07, "loss": 1.0313, "step": 6809 }, { "epoch": 0.89, "grad_norm": 1.3721294403076172, "learning_rate": 3.00489387975928e-07, "loss": 0.8022, "step": 6810 }, { "epoch": 0.89, "grad_norm": 1.3095033168792725, "learning_rate": 2.997652247167138e-07, "loss": 0.7928, "step": 6811 }, { "epoch": 0.89, "grad_norm": 1.3815100193023682, "learning_rate": 2.990419081410639e-07, "loss": 0.8242, "step": 6812 }, { "epoch": 0.89, "grad_norm": 1.8801233768463135, "learning_rate": 2.9831943837927257e-07, "loss": 0.7948, "step": 6813 }, { "epoch": 0.89, "grad_norm": 1.3889516592025757, "learning_rate": 2.9759781556148224e-07, "loss": 0.7708, "step": 6814 }, { "epoch": 0.89, "grad_norm": 1.3270223140716553, "learning_rate": 2.968770398176851e-07, "loss": 0.8166, "step": 6815 }, { "epoch": 0.89, "grad_norm": 1.3059725761413574, "learning_rate": 2.961571112777173e-07, "loss": 0.7613, "step": 6816 }, { "epoch": 0.89, "grad_norm": 0.563771665096283, "learning_rate": 2.9543803007126613e-07, "loss": 1.006, "step": 6817 }, { "epoch": 0.89, "grad_norm": 1.3571711778640747, "learning_rate": 2.947197963278631e-07, "loss": 0.73, "step": 6818 }, { "epoch": 0.89, "grad_norm": 1.3557252883911133, "learning_rate": 2.940024101768879e-07, "loss": 0.7828, "step": 6819 }, { "epoch": 0.89, "grad_norm": 1.326245665550232, "learning_rate": 2.932858717475695e-07, "loss": 0.8554, "step": 6820 }, { "epoch": 0.89, "grad_norm": 0.5688807368278503, "learning_rate": 2.9257018116898074e-07, "loss": 1.01, "step": 6821 }, { "epoch": 0.89, "grad_norm": 0.5659614205360413, "learning_rate": 2.918553385700451e-07, "loss": 1.0245, "step": 6822 }, { "epoch": 0.89, "grad_norm": 1.3064976930618286, "learning_rate": 2.911413440795313e-07, "loss": 0.7846, "step": 6823 }, { "epoch": 0.89, "grad_norm": 1.3398923873901367, "learning_rate": 2.904281978260548e-07, "loss": 0.8008, "step": 6824 }, { "epoch": 0.89, "grad_norm": 1.484010934829712, "learning_rate": 2.897158999380806e-07, "loss": 0.7871, "step": 6825 }, { "epoch": 0.89, "grad_norm": 0.5589016675949097, "learning_rate": 2.8900445054391877e-07, "loss": 1.0207, "step": 6826 }, { "epoch": 0.89, "grad_norm": 1.3283754587173462, "learning_rate": 2.8829384977172747e-07, "loss": 0.787, "step": 6827 }, { "epoch": 0.89, "grad_norm": 1.2828294038772583, "learning_rate": 2.875840977495126e-07, "loss": 0.7893, "step": 6828 }, { "epoch": 0.89, "grad_norm": 0.567528247833252, "learning_rate": 2.86875194605124e-07, "loss": 0.9723, "step": 6829 }, { "epoch": 0.89, "grad_norm": 1.3162801265716553, "learning_rate": 2.861671404662636e-07, "loss": 0.801, "step": 6830 }, { "epoch": 0.9, "grad_norm": 1.336905598640442, "learning_rate": 2.85459935460477e-07, "loss": 0.8264, "step": 6831 }, { "epoch": 0.9, "grad_norm": 1.3452762365341187, "learning_rate": 2.847535797151563e-07, "loss": 0.8179, "step": 6832 }, { "epoch": 0.9, "grad_norm": 1.303302526473999, "learning_rate": 2.8404807335754346e-07, "loss": 0.7677, "step": 6833 }, { "epoch": 0.9, "grad_norm": 1.322798252105713, "learning_rate": 2.8334341651472477e-07, "loss": 0.796, "step": 6834 }, { "epoch": 0.9, "grad_norm": 1.4207946062088013, "learning_rate": 2.8263960931363524e-07, "loss": 0.8819, "step": 6835 }, { "epoch": 0.9, "grad_norm": 1.3335299491882324, "learning_rate": 2.819366518810568e-07, "loss": 0.7175, "step": 6836 }, { "epoch": 0.9, "grad_norm": 0.5629952549934387, "learning_rate": 2.8123454434361597e-07, "loss": 1.0009, "step": 6837 }, { "epoch": 0.9, "grad_norm": 1.4595781564712524, "learning_rate": 2.8053328682778936e-07, "loss": 0.7849, "step": 6838 }, { "epoch": 0.9, "grad_norm": 1.2745976448059082, "learning_rate": 2.7983287945989813e-07, "loss": 0.8144, "step": 6839 }, { "epoch": 0.9, "grad_norm": 1.349273920059204, "learning_rate": 2.7913332236611243e-07, "loss": 0.8163, "step": 6840 }, { "epoch": 0.9, "grad_norm": 0.567363977432251, "learning_rate": 2.784346156724471e-07, "loss": 1.0011, "step": 6841 }, { "epoch": 0.9, "grad_norm": 1.3003852367401123, "learning_rate": 2.7773675950476366e-07, "loss": 0.8078, "step": 6842 }, { "epoch": 0.9, "grad_norm": 1.3309009075164795, "learning_rate": 2.770397539887737e-07, "loss": 0.737, "step": 6843 }, { "epoch": 0.9, "grad_norm": 1.3419904708862305, "learning_rate": 2.763435992500318e-07, "loss": 0.7722, "step": 6844 }, { "epoch": 0.9, "grad_norm": 1.3046460151672363, "learning_rate": 2.7564829541394057e-07, "loss": 0.7212, "step": 6845 }, { "epoch": 0.9, "grad_norm": 0.5749820470809937, "learning_rate": 2.7495384260575077e-07, "loss": 1.0008, "step": 6846 }, { "epoch": 0.9, "grad_norm": 1.381966471672058, "learning_rate": 2.7426024095055794e-07, "loss": 0.8318, "step": 6847 }, { "epoch": 0.9, "grad_norm": 0.5764818787574768, "learning_rate": 2.7356749057330433e-07, "loss": 1.0024, "step": 6848 }, { "epoch": 0.9, "grad_norm": 1.4759047031402588, "learning_rate": 2.7287559159878076e-07, "loss": 0.7876, "step": 6849 }, { "epoch": 0.9, "grad_norm": 1.2985358238220215, "learning_rate": 2.72184544151623e-07, "loss": 0.7889, "step": 6850 }, { "epoch": 0.9, "grad_norm": 1.3375309705734253, "learning_rate": 2.7149434835631426e-07, "loss": 0.8296, "step": 6851 }, { "epoch": 0.9, "grad_norm": 1.3177578449249268, "learning_rate": 2.708050043371824e-07, "loss": 0.8076, "step": 6852 }, { "epoch": 0.9, "grad_norm": 1.7482917308807373, "learning_rate": 2.7011651221840574e-07, "loss": 0.7495, "step": 6853 }, { "epoch": 0.9, "grad_norm": 1.3811593055725098, "learning_rate": 2.694288721240046e-07, "loss": 0.7736, "step": 6854 }, { "epoch": 0.9, "grad_norm": 1.401901364326477, "learning_rate": 2.687420841778493e-07, "loss": 0.8332, "step": 6855 }, { "epoch": 0.9, "grad_norm": 1.4779149293899536, "learning_rate": 2.6805614850365536e-07, "loss": 0.7849, "step": 6856 }, { "epoch": 0.9, "grad_norm": 1.3524835109710693, "learning_rate": 2.673710652249839e-07, "loss": 0.8109, "step": 6857 }, { "epoch": 0.9, "grad_norm": 1.2827621698379517, "learning_rate": 2.6668683446524456e-07, "loss": 0.8016, "step": 6858 }, { "epoch": 0.9, "grad_norm": 1.4023268222808838, "learning_rate": 2.660034563476921e-07, "loss": 0.8275, "step": 6859 }, { "epoch": 0.9, "grad_norm": 1.314373254776001, "learning_rate": 2.653209309954263e-07, "loss": 0.7859, "step": 6860 }, { "epoch": 0.9, "grad_norm": 1.2683727741241455, "learning_rate": 2.6463925853139673e-07, "loss": 0.7408, "step": 6861 }, { "epoch": 0.9, "grad_norm": 0.5767613649368286, "learning_rate": 2.639584390783956e-07, "loss": 1.0272, "step": 6862 }, { "epoch": 0.9, "grad_norm": 1.495545744895935, "learning_rate": 2.632784727590654e-07, "loss": 0.7973, "step": 6863 }, { "epoch": 0.9, "grad_norm": 1.2562079429626465, "learning_rate": 2.6259935969589203e-07, "loss": 0.7432, "step": 6864 }, { "epoch": 0.9, "grad_norm": 1.485282301902771, "learning_rate": 2.6192110001120707e-07, "loss": 0.7743, "step": 6865 }, { "epoch": 0.9, "grad_norm": 1.4292296171188354, "learning_rate": 2.612436938271917e-07, "loss": 0.803, "step": 6866 }, { "epoch": 0.9, "grad_norm": 1.4235001802444458, "learning_rate": 2.605671412658706e-07, "loss": 0.8362, "step": 6867 }, { "epoch": 0.9, "grad_norm": 1.354128360748291, "learning_rate": 2.598914424491161e-07, "loss": 0.8182, "step": 6868 }, { "epoch": 0.9, "grad_norm": 1.2769708633422852, "learning_rate": 2.592165974986455e-07, "loss": 0.7803, "step": 6869 }, { "epoch": 0.9, "grad_norm": 1.3120710849761963, "learning_rate": 2.5854260653602306e-07, "loss": 0.7792, "step": 6870 }, { "epoch": 0.9, "grad_norm": 0.5662367939949036, "learning_rate": 2.578694696826595e-07, "loss": 1.0096, "step": 6871 }, { "epoch": 0.9, "grad_norm": 1.3107452392578125, "learning_rate": 2.5719718705981055e-07, "loss": 0.7645, "step": 6872 }, { "epoch": 0.9, "grad_norm": 1.2356359958648682, "learning_rate": 2.5652575878858e-07, "loss": 0.809, "step": 6873 }, { "epoch": 0.9, "grad_norm": 1.3869001865386963, "learning_rate": 2.5585518498991604e-07, "loss": 0.8231, "step": 6874 }, { "epoch": 0.9, "grad_norm": 1.3922672271728516, "learning_rate": 2.5518546578461256e-07, "loss": 0.7285, "step": 6875 }, { "epoch": 0.9, "grad_norm": 1.3500428199768066, "learning_rate": 2.5451660129331204e-07, "loss": 0.7692, "step": 6876 }, { "epoch": 0.9, "grad_norm": 1.5296992063522339, "learning_rate": 2.5384859163649966e-07, "loss": 0.8099, "step": 6877 }, { "epoch": 0.9, "grad_norm": 2.0319578647613525, "learning_rate": 2.531814369345098e-07, "loss": 0.7685, "step": 6878 }, { "epoch": 0.9, "grad_norm": 1.339532494544983, "learning_rate": 2.525151373075202e-07, "loss": 0.8106, "step": 6879 }, { "epoch": 0.9, "grad_norm": 1.309309720993042, "learning_rate": 2.5184969287555595e-07, "loss": 0.8085, "step": 6880 }, { "epoch": 0.9, "grad_norm": 1.256106972694397, "learning_rate": 2.5118510375848834e-07, "loss": 0.7621, "step": 6881 }, { "epoch": 0.9, "grad_norm": 0.5673219561576843, "learning_rate": 2.5052137007603337e-07, "loss": 1.0084, "step": 6882 }, { "epoch": 0.9, "grad_norm": 1.4262760877609253, "learning_rate": 2.498584919477548e-07, "loss": 0.8562, "step": 6883 }, { "epoch": 0.9, "grad_norm": 1.5426260232925415, "learning_rate": 2.491964694930604e-07, "loss": 0.826, "step": 6884 }, { "epoch": 0.9, "grad_norm": 1.2721911668777466, "learning_rate": 2.485353028312043e-07, "loss": 0.73, "step": 6885 }, { "epoch": 0.9, "grad_norm": 0.5608689785003662, "learning_rate": 2.4787499208128616e-07, "loss": 0.9894, "step": 6886 }, { "epoch": 0.9, "grad_norm": 1.3566066026687622, "learning_rate": 2.472155373622537e-07, "loss": 0.7679, "step": 6887 }, { "epoch": 0.9, "grad_norm": 1.3184854984283447, "learning_rate": 2.4655693879289787e-07, "loss": 0.775, "step": 6888 }, { "epoch": 0.9, "grad_norm": 0.5940840244293213, "learning_rate": 2.4589919649185556e-07, "loss": 1.0264, "step": 6889 }, { "epoch": 0.9, "grad_norm": 0.5598962306976318, "learning_rate": 2.4524231057761136e-07, "loss": 0.957, "step": 6890 }, { "epoch": 0.9, "grad_norm": 1.370180606842041, "learning_rate": 2.44586281168494e-07, "loss": 0.8588, "step": 6891 }, { "epoch": 0.9, "grad_norm": 1.5709056854248047, "learning_rate": 2.439311083826779e-07, "loss": 0.8159, "step": 6892 }, { "epoch": 0.9, "grad_norm": 0.5715129375457764, "learning_rate": 2.4327679233818294e-07, "loss": 1.0123, "step": 6893 }, { "epoch": 0.9, "grad_norm": 1.4843922853469849, "learning_rate": 2.426233331528771e-07, "loss": 0.791, "step": 6894 }, { "epoch": 0.9, "grad_norm": 1.739524006843567, "learning_rate": 2.419707309444702e-07, "loss": 0.8119, "step": 6895 }, { "epoch": 0.9, "grad_norm": 1.399310827255249, "learning_rate": 2.413189858305215e-07, "loss": 0.7594, "step": 6896 }, { "epoch": 0.9, "grad_norm": 1.3377704620361328, "learning_rate": 2.406680979284337e-07, "loss": 0.7755, "step": 6897 }, { "epoch": 0.9, "grad_norm": 0.5597560405731201, "learning_rate": 2.4001806735545365e-07, "loss": 1.001, "step": 6898 }, { "epoch": 0.9, "grad_norm": 0.599470853805542, "learning_rate": 2.393688942286776e-07, "loss": 0.9974, "step": 6899 }, { "epoch": 0.9, "grad_norm": 1.2682673931121826, "learning_rate": 2.387205786650437e-07, "loss": 0.7885, "step": 6900 }, { "epoch": 0.9, "grad_norm": 1.2938576936721802, "learning_rate": 2.3807312078133914e-07, "loss": 0.8621, "step": 6901 }, { "epoch": 0.9, "grad_norm": 0.5675923228263855, "learning_rate": 2.3742652069419337e-07, "loss": 1.0077, "step": 6902 }, { "epoch": 0.9, "grad_norm": 1.7042375802993774, "learning_rate": 2.3678077852008206e-07, "loss": 0.7823, "step": 6903 }, { "epoch": 0.9, "grad_norm": 0.5835651159286499, "learning_rate": 2.3613589437532835e-07, "loss": 1.02, "step": 6904 }, { "epoch": 0.9, "grad_norm": 1.3589048385620117, "learning_rate": 2.3549186837609817e-07, "loss": 0.7957, "step": 6905 }, { "epoch": 0.9, "grad_norm": 1.3505563735961914, "learning_rate": 2.3484870063840483e-07, "loss": 0.8025, "step": 6906 }, { "epoch": 0.91, "grad_norm": 0.560308575630188, "learning_rate": 2.3420639127810618e-07, "loss": 1.0285, "step": 6907 }, { "epoch": 0.91, "grad_norm": 1.3215073347091675, "learning_rate": 2.3356494041090527e-07, "loss": 0.7963, "step": 6908 }, { "epoch": 0.91, "grad_norm": 1.3960121870040894, "learning_rate": 2.329243481523513e-07, "loss": 0.7949, "step": 6909 }, { "epoch": 0.91, "grad_norm": 0.5661881566047668, "learning_rate": 2.3228461461783702e-07, "loss": 1.009, "step": 6910 }, { "epoch": 0.91, "grad_norm": 0.566831111907959, "learning_rate": 2.3164573992260352e-07, "loss": 1.0097, "step": 6911 }, { "epoch": 0.91, "grad_norm": 1.3291717767715454, "learning_rate": 2.3100772418173434e-07, "loss": 0.8041, "step": 6912 }, { "epoch": 0.91, "grad_norm": 0.5793535709381104, "learning_rate": 2.3037056751015864e-07, "loss": 1.0042, "step": 6913 }, { "epoch": 0.91, "grad_norm": 1.2946608066558838, "learning_rate": 2.2973427002265348e-07, "loss": 0.7504, "step": 6914 }, { "epoch": 0.91, "grad_norm": 1.5393754243850708, "learning_rate": 2.2909883183383774e-07, "loss": 0.8093, "step": 6915 }, { "epoch": 0.91, "grad_norm": 2.3984925746917725, "learning_rate": 2.2846425305817766e-07, "loss": 0.7717, "step": 6916 }, { "epoch": 0.91, "grad_norm": 1.3823590278625488, "learning_rate": 2.27830533809984e-07, "loss": 0.7951, "step": 6917 }, { "epoch": 0.91, "grad_norm": 0.5640837550163269, "learning_rate": 2.2719767420341155e-07, "loss": 1.0089, "step": 6918 }, { "epoch": 0.91, "grad_norm": 1.726491093635559, "learning_rate": 2.2656567435246357e-07, "loss": 0.7543, "step": 6919 }, { "epoch": 0.91, "grad_norm": 1.3295010328292847, "learning_rate": 2.2593453437098457e-07, "loss": 0.7916, "step": 6920 }, { "epoch": 0.91, "grad_norm": 1.4020764827728271, "learning_rate": 2.253042543726658e-07, "loss": 0.8187, "step": 6921 }, { "epoch": 0.91, "grad_norm": 1.5001956224441528, "learning_rate": 2.2467483447104476e-07, "loss": 0.7796, "step": 6922 }, { "epoch": 0.91, "grad_norm": 1.3383545875549316, "learning_rate": 2.2404627477950135e-07, "loss": 0.8447, "step": 6923 }, { "epoch": 0.91, "grad_norm": 1.2923396825790405, "learning_rate": 2.2341857541126445e-07, "loss": 0.7173, "step": 6924 }, { "epoch": 0.91, "grad_norm": 0.5807055830955505, "learning_rate": 2.2279173647940354e-07, "loss": 1.0153, "step": 6925 }, { "epoch": 0.91, "grad_norm": 1.3592529296875, "learning_rate": 2.2216575809683617e-07, "loss": 0.8058, "step": 6926 }, { "epoch": 0.91, "grad_norm": 0.5667950510978699, "learning_rate": 2.215406403763226e-07, "loss": 1.0072, "step": 6927 }, { "epoch": 0.91, "grad_norm": 1.3969162702560425, "learning_rate": 2.2091638343047062e-07, "loss": 0.7614, "step": 6928 }, { "epoch": 0.91, "grad_norm": 1.2563512325286865, "learning_rate": 2.2029298737173188e-07, "loss": 0.7982, "step": 6929 }, { "epoch": 0.91, "grad_norm": 0.5645936131477356, "learning_rate": 2.1967045231240103e-07, "loss": 1.006, "step": 6930 }, { "epoch": 0.91, "grad_norm": 1.3179190158843994, "learning_rate": 2.1904877836462059e-07, "loss": 0.8118, "step": 6931 }, { "epoch": 0.91, "grad_norm": 1.2701841592788696, "learning_rate": 2.1842796564037706e-07, "loss": 0.7666, "step": 6932 }, { "epoch": 0.91, "grad_norm": 0.5692049860954285, "learning_rate": 2.178080142514999e-07, "loss": 1.0124, "step": 6933 }, { "epoch": 0.91, "grad_norm": 1.5007908344268799, "learning_rate": 2.1718892430966698e-07, "loss": 0.7687, "step": 6934 }, { "epoch": 0.91, "grad_norm": 0.5733991265296936, "learning_rate": 2.1657069592639747e-07, "loss": 0.9996, "step": 6935 }, { "epoch": 0.91, "grad_norm": 1.27334725856781, "learning_rate": 2.1595332921305666e-07, "loss": 0.7725, "step": 6936 }, { "epoch": 0.91, "grad_norm": 0.5668216943740845, "learning_rate": 2.1533682428085568e-07, "loss": 1.0174, "step": 6937 }, { "epoch": 0.91, "grad_norm": 0.5877717733383179, "learning_rate": 2.1472118124084895e-07, "loss": 0.9969, "step": 6938 }, { "epoch": 0.91, "grad_norm": 1.3764543533325195, "learning_rate": 2.1410640020393726e-07, "loss": 0.8129, "step": 6939 }, { "epoch": 0.91, "grad_norm": 1.4152346849441528, "learning_rate": 2.134924812808642e-07, "loss": 0.7775, "step": 6940 }, { "epoch": 0.91, "grad_norm": 1.437232255935669, "learning_rate": 2.12879424582218e-07, "loss": 0.7635, "step": 6941 }, { "epoch": 0.91, "grad_norm": 1.5180660486221313, "learning_rate": 2.1226723021843475e-07, "loss": 0.7764, "step": 6942 }, { "epoch": 0.91, "grad_norm": 1.3863449096679688, "learning_rate": 2.1165589829979072e-07, "loss": 0.8285, "step": 6943 }, { "epoch": 0.91, "grad_norm": 0.5735527276992798, "learning_rate": 2.1104542893641055e-07, "loss": 0.991, "step": 6944 }, { "epoch": 0.91, "grad_norm": 1.4818694591522217, "learning_rate": 2.1043582223826185e-07, "loss": 0.7368, "step": 6945 }, { "epoch": 0.91, "grad_norm": 0.5757002830505371, "learning_rate": 2.0982707831515625e-07, "loss": 0.9986, "step": 6946 }, { "epoch": 0.91, "grad_norm": 0.5688075423240662, "learning_rate": 2.0921919727675154e-07, "loss": 0.9968, "step": 6947 }, { "epoch": 0.91, "grad_norm": 1.4700547456741333, "learning_rate": 2.0861217923254906e-07, "loss": 0.7984, "step": 6948 }, { "epoch": 0.91, "grad_norm": 1.4899206161499023, "learning_rate": 2.080060242918941e-07, "loss": 0.7745, "step": 6949 }, { "epoch": 0.91, "grad_norm": 1.3791823387145996, "learning_rate": 2.0740073256397876e-07, "loss": 0.7699, "step": 6950 }, { "epoch": 0.91, "grad_norm": 1.2796334028244019, "learning_rate": 2.0679630415783637e-07, "loss": 0.8344, "step": 6951 }, { "epoch": 0.91, "grad_norm": 1.3911619186401367, "learning_rate": 2.0619273918234762e-07, "loss": 0.8229, "step": 6952 }, { "epoch": 0.91, "grad_norm": 1.3377277851104736, "learning_rate": 2.0559003774623664e-07, "loss": 0.7475, "step": 6953 }, { "epoch": 0.91, "grad_norm": 1.349591851234436, "learning_rate": 2.0498819995807107e-07, "loss": 0.7996, "step": 6954 }, { "epoch": 0.91, "grad_norm": 1.3470200300216675, "learning_rate": 2.0438722592626524e-07, "loss": 0.777, "step": 6955 }, { "epoch": 0.91, "grad_norm": 1.2928985357284546, "learning_rate": 2.0378711575907428e-07, "loss": 0.8253, "step": 6956 }, { "epoch": 0.91, "grad_norm": 1.3591691255569458, "learning_rate": 2.0318786956460223e-07, "loss": 0.8139, "step": 6957 }, { "epoch": 0.91, "grad_norm": 0.5515346527099609, "learning_rate": 2.0258948745079444e-07, "loss": 1.0145, "step": 6958 }, { "epoch": 0.91, "grad_norm": 1.367506742477417, "learning_rate": 2.019919695254402e-07, "loss": 0.8258, "step": 6959 }, { "epoch": 0.91, "grad_norm": 0.5811647772789001, "learning_rate": 2.0139531589617567e-07, "loss": 1.0048, "step": 6960 }, { "epoch": 0.91, "grad_norm": 1.3627498149871826, "learning_rate": 2.0079952667047875e-07, "loss": 0.8298, "step": 6961 }, { "epoch": 0.91, "grad_norm": 1.542041540145874, "learning_rate": 2.002046019556736e-07, "loss": 0.8551, "step": 6962 }, { "epoch": 0.91, "grad_norm": 1.2165879011154175, "learning_rate": 1.9961054185892836e-07, "loss": 0.7955, "step": 6963 }, { "epoch": 0.91, "grad_norm": 0.5621658563613892, "learning_rate": 1.9901734648725303e-07, "loss": 0.9551, "step": 6964 }, { "epoch": 0.91, "grad_norm": 1.2761167287826538, "learning_rate": 1.984250159475054e-07, "loss": 0.8166, "step": 6965 }, { "epoch": 0.91, "grad_norm": 1.5082591772079468, "learning_rate": 1.9783355034638574e-07, "loss": 0.818, "step": 6966 }, { "epoch": 0.91, "grad_norm": 1.2949223518371582, "learning_rate": 1.972429497904377e-07, "loss": 0.7658, "step": 6967 }, { "epoch": 0.91, "grad_norm": 1.2542201280593872, "learning_rate": 1.9665321438605e-07, "loss": 0.7587, "step": 6968 }, { "epoch": 0.91, "grad_norm": 0.5652559995651245, "learning_rate": 1.960643442394561e-07, "loss": 0.9813, "step": 6969 }, { "epoch": 0.91, "grad_norm": 1.2362326383590698, "learning_rate": 1.9547633945673272e-07, "loss": 0.735, "step": 6970 }, { "epoch": 0.91, "grad_norm": 1.3238157033920288, "learning_rate": 1.948892001438002e-07, "loss": 0.7945, "step": 6971 }, { "epoch": 0.91, "grad_norm": 1.2796614170074463, "learning_rate": 1.94302926406425e-07, "loss": 0.7551, "step": 6972 }, { "epoch": 0.91, "grad_norm": 1.349491834640503, "learning_rate": 1.9371751835021603e-07, "loss": 0.8, "step": 6973 }, { "epoch": 0.91, "grad_norm": 0.5587652921676636, "learning_rate": 1.9313297608062608e-07, "loss": 1.0029, "step": 6974 }, { "epoch": 0.91, "grad_norm": 1.4068673849105835, "learning_rate": 1.925492997029532e-07, "loss": 0.7959, "step": 6975 }, { "epoch": 0.91, "grad_norm": 1.7608360052108765, "learning_rate": 1.919664893223383e-07, "loss": 0.8326, "step": 6976 }, { "epoch": 0.91, "grad_norm": 1.37606680393219, "learning_rate": 1.9138454504376625e-07, "loss": 0.7139, "step": 6977 }, { "epoch": 0.91, "grad_norm": 1.3895493745803833, "learning_rate": 1.9080346697206765e-07, "loss": 0.829, "step": 6978 }, { "epoch": 0.91, "grad_norm": 0.5773962736129761, "learning_rate": 1.9022325521191432e-07, "loss": 0.9859, "step": 6979 }, { "epoch": 0.91, "grad_norm": 0.57344651222229, "learning_rate": 1.8964390986782545e-07, "loss": 0.989, "step": 6980 }, { "epoch": 0.91, "grad_norm": 1.2467246055603027, "learning_rate": 1.8906543104416086e-07, "loss": 0.8119, "step": 6981 }, { "epoch": 0.91, "grad_norm": 0.5670983791351318, "learning_rate": 1.8848781884512613e-07, "loss": 1.0152, "step": 6982 }, { "epoch": 0.92, "grad_norm": 0.5729318261146545, "learning_rate": 1.8791107337477022e-07, "loss": 1.0004, "step": 6983 }, { "epoch": 0.92, "grad_norm": 1.3393099308013916, "learning_rate": 1.8733519473698557e-07, "loss": 0.8259, "step": 6984 }, { "epoch": 0.92, "grad_norm": 1.26008939743042, "learning_rate": 1.867601830355098e-07, "loss": 0.7952, "step": 6985 }, { "epoch": 0.92, "grad_norm": 0.5710600018501282, "learning_rate": 1.8618603837392336e-07, "loss": 0.9911, "step": 6986 }, { "epoch": 0.92, "grad_norm": 1.3464112281799316, "learning_rate": 1.8561276085564906e-07, "loss": 0.7891, "step": 6987 }, { "epoch": 0.92, "grad_norm": 1.3312067985534668, "learning_rate": 1.8504035058395708e-07, "loss": 0.8066, "step": 6988 }, { "epoch": 0.92, "grad_norm": 1.3762949705123901, "learning_rate": 1.8446880766195828e-07, "loss": 0.8561, "step": 6989 }, { "epoch": 0.92, "grad_norm": 0.5765984654426575, "learning_rate": 1.8389813219260865e-07, "loss": 1.0014, "step": 6990 }, { "epoch": 0.92, "grad_norm": 1.2493102550506592, "learning_rate": 1.8332832427870817e-07, "loss": 0.8328, "step": 6991 }, { "epoch": 0.92, "grad_norm": 0.5761697292327881, "learning_rate": 1.827593840228986e-07, "loss": 1.0279, "step": 6992 }, { "epoch": 0.92, "grad_norm": 1.3871737718582153, "learning_rate": 1.8219131152766855e-07, "loss": 0.8165, "step": 6993 }, { "epoch": 0.92, "grad_norm": 0.566443920135498, "learning_rate": 1.816241068953467e-07, "loss": 0.9986, "step": 6994 }, { "epoch": 0.92, "grad_norm": 1.3047508001327515, "learning_rate": 1.8105777022810911e-07, "loss": 0.843, "step": 6995 }, { "epoch": 0.92, "grad_norm": 0.5770310163497925, "learning_rate": 1.8049230162797305e-07, "loss": 0.9991, "step": 6996 }, { "epoch": 0.92, "grad_norm": 2.299257755279541, "learning_rate": 1.799277011967987e-07, "loss": 0.8735, "step": 6997 }, { "epoch": 0.92, "grad_norm": 1.2992324829101562, "learning_rate": 1.7936396903629304e-07, "loss": 0.738, "step": 6998 }, { "epoch": 0.92, "grad_norm": 1.404500961303711, "learning_rate": 1.788011052480032e-07, "loss": 0.8356, "step": 6999 }, { "epoch": 0.92, "grad_norm": 1.3185287714004517, "learning_rate": 1.7823910993332306e-07, "loss": 0.789, "step": 7000 }, { "epoch": 0.92, "grad_norm": 0.5724669694900513, "learning_rate": 1.776779831934877e-07, "loss": 1.0131, "step": 7001 }, { "epoch": 0.92, "grad_norm": 0.5770564675331116, "learning_rate": 1.771177251295758e-07, "loss": 1.0096, "step": 7002 }, { "epoch": 0.92, "grad_norm": 0.5833758115768433, "learning_rate": 1.7655833584251157e-07, "loss": 1.0072, "step": 7003 }, { "epoch": 0.92, "grad_norm": 1.2901856899261475, "learning_rate": 1.759998154330611e-07, "loss": 0.7952, "step": 7004 }, { "epoch": 0.92, "grad_norm": 1.2011462450027466, "learning_rate": 1.7544216400183277e-07, "loss": 0.7493, "step": 7005 }, { "epoch": 0.92, "grad_norm": 1.3822747468948364, "learning_rate": 1.748853816492818e-07, "loss": 0.815, "step": 7006 }, { "epoch": 0.92, "grad_norm": 1.4086987972259521, "learning_rate": 1.7432946847570408e-07, "loss": 0.8082, "step": 7007 }, { "epoch": 0.92, "grad_norm": 1.34428870677948, "learning_rate": 1.737744245812395e-07, "loss": 0.8589, "step": 7008 }, { "epoch": 0.92, "grad_norm": 1.2859050035476685, "learning_rate": 1.7322025006587306e-07, "loss": 0.7605, "step": 7009 }, { "epoch": 0.92, "grad_norm": 1.2956557273864746, "learning_rate": 1.7266694502943048e-07, "loss": 0.788, "step": 7010 }, { "epoch": 0.92, "grad_norm": 0.5696055293083191, "learning_rate": 1.721145095715826e-07, "loss": 0.9983, "step": 7011 }, { "epoch": 0.92, "grad_norm": 1.7650820016860962, "learning_rate": 1.7156294379184257e-07, "loss": 0.8052, "step": 7012 }, { "epoch": 0.92, "grad_norm": 0.5651748776435852, "learning_rate": 1.7101224778956927e-07, "loss": 0.9983, "step": 7013 }, { "epoch": 0.92, "grad_norm": 0.5788602232933044, "learning_rate": 1.704624216639611e-07, "loss": 1.0159, "step": 7014 }, { "epoch": 0.92, "grad_norm": 1.3168662786483765, "learning_rate": 1.6991346551406275e-07, "loss": 0.8352, "step": 7015 }, { "epoch": 0.92, "grad_norm": 1.7619441747665405, "learning_rate": 1.693653794387612e-07, "loss": 0.773, "step": 7016 }, { "epoch": 0.92, "grad_norm": 1.344864010810852, "learning_rate": 1.6881816353678583e-07, "loss": 0.7478, "step": 7017 }, { "epoch": 0.92, "grad_norm": 0.5800348520278931, "learning_rate": 1.682718179067122e-07, "loss": 1.0178, "step": 7018 }, { "epoch": 0.92, "grad_norm": 1.3721952438354492, "learning_rate": 1.6772634264695553e-07, "loss": 0.7753, "step": 7019 }, { "epoch": 0.92, "grad_norm": 0.5815547704696655, "learning_rate": 1.6718173785577496e-07, "loss": 1.0024, "step": 7020 }, { "epoch": 0.92, "grad_norm": 0.5640113353729248, "learning_rate": 1.6663800363127592e-07, "loss": 1.0192, "step": 7021 }, { "epoch": 0.92, "grad_norm": 1.4676835536956787, "learning_rate": 1.6609514007140282e-07, "loss": 0.7722, "step": 7022 }, { "epoch": 0.92, "grad_norm": 1.3151487112045288, "learning_rate": 1.655531472739469e-07, "loss": 0.807, "step": 7023 }, { "epoch": 0.92, "grad_norm": 2.372936487197876, "learning_rate": 1.6501202533653948e-07, "loss": 0.8081, "step": 7024 }, { "epoch": 0.92, "grad_norm": 1.3353826999664307, "learning_rate": 1.644717743566565e-07, "loss": 0.7466, "step": 7025 }, { "epoch": 0.92, "grad_norm": 1.3115134239196777, "learning_rate": 1.6393239443161735e-07, "loss": 0.7448, "step": 7026 }, { "epoch": 0.92, "grad_norm": 0.5784138441085815, "learning_rate": 1.6339388565858317e-07, "loss": 0.9733, "step": 7027 }, { "epoch": 0.92, "grad_norm": 0.5718473792076111, "learning_rate": 1.6285624813456025e-07, "loss": 1.0038, "step": 7028 }, { "epoch": 0.92, "grad_norm": 0.5676963329315186, "learning_rate": 1.6231948195639613e-07, "loss": 0.9836, "step": 7029 }, { "epoch": 0.92, "grad_norm": 1.3462791442871094, "learning_rate": 1.6178358722078125e-07, "loss": 0.8014, "step": 7030 }, { "epoch": 0.92, "grad_norm": 1.3571604490280151, "learning_rate": 1.612485640242506e-07, "loss": 0.7718, "step": 7031 }, { "epoch": 0.92, "grad_norm": 1.3812756538391113, "learning_rate": 1.60714412463181e-07, "loss": 0.7767, "step": 7032 }, { "epoch": 0.92, "grad_norm": 0.5698915719985962, "learning_rate": 1.601811326337932e-07, "loss": 0.9872, "step": 7033 }, { "epoch": 0.92, "grad_norm": 0.5715164542198181, "learning_rate": 1.5964872463214985e-07, "loss": 1.0011, "step": 7034 }, { "epoch": 0.92, "grad_norm": 0.5595239996910095, "learning_rate": 1.5911718855415648e-07, "loss": 0.9948, "step": 7035 }, { "epoch": 0.92, "grad_norm": 0.5656161308288574, "learning_rate": 1.5858652449556368e-07, "loss": 0.9973, "step": 7036 }, { "epoch": 0.92, "grad_norm": 1.382992148399353, "learning_rate": 1.5805673255196164e-07, "loss": 0.7482, "step": 7037 }, { "epoch": 0.92, "grad_norm": 1.8660573959350586, "learning_rate": 1.5752781281878626e-07, "loss": 0.7643, "step": 7038 }, { "epoch": 0.92, "grad_norm": 1.3193473815917969, "learning_rate": 1.5699976539131467e-07, "loss": 0.7819, "step": 7039 }, { "epoch": 0.92, "grad_norm": 1.3780491352081299, "learning_rate": 1.56472590364668e-07, "loss": 0.8079, "step": 7040 }, { "epoch": 0.92, "grad_norm": 1.3805923461914062, "learning_rate": 1.5594628783380917e-07, "loss": 0.7905, "step": 7041 }, { "epoch": 0.92, "grad_norm": 1.3360012769699097, "learning_rate": 1.554208578935451e-07, "loss": 0.7717, "step": 7042 }, { "epoch": 0.92, "grad_norm": 1.4138227701187134, "learning_rate": 1.5489630063852402e-07, "loss": 0.7751, "step": 7043 }, { "epoch": 0.92, "grad_norm": 1.2636280059814453, "learning_rate": 1.5437261616323861e-07, "loss": 0.7647, "step": 7044 }, { "epoch": 0.92, "grad_norm": 0.5827427506446838, "learning_rate": 1.5384980456202348e-07, "loss": 0.9956, "step": 7045 }, { "epoch": 0.92, "grad_norm": 1.2867599725723267, "learning_rate": 1.533278659290549e-07, "loss": 0.8317, "step": 7046 }, { "epoch": 0.92, "grad_norm": 1.3661538362503052, "learning_rate": 1.5280680035835438e-07, "loss": 0.733, "step": 7047 }, { "epoch": 0.92, "grad_norm": 1.3943713903427124, "learning_rate": 1.522866079437846e-07, "loss": 0.8252, "step": 7048 }, { "epoch": 0.92, "grad_norm": 1.6478981971740723, "learning_rate": 1.5176728877905e-07, "loss": 0.8633, "step": 7049 }, { "epoch": 0.92, "grad_norm": 1.3314963579177856, "learning_rate": 1.512488429577008e-07, "loss": 0.7757, "step": 7050 }, { "epoch": 0.92, "grad_norm": 1.408504843711853, "learning_rate": 1.507312705731262e-07, "loss": 0.799, "step": 7051 }, { "epoch": 0.92, "grad_norm": 1.364835500717163, "learning_rate": 1.5021457171856103e-07, "loss": 0.8945, "step": 7052 }, { "epoch": 0.92, "grad_norm": 1.3591300249099731, "learning_rate": 1.4969874648708082e-07, "loss": 0.76, "step": 7053 }, { "epoch": 0.92, "grad_norm": 1.33800208568573, "learning_rate": 1.4918379497160517e-07, "loss": 0.8267, "step": 7054 }, { "epoch": 0.92, "grad_norm": 1.382048487663269, "learning_rate": 1.4866971726489486e-07, "loss": 0.769, "step": 7055 }, { "epoch": 0.92, "grad_norm": 1.3182936906814575, "learning_rate": 1.4815651345955474e-07, "loss": 0.7805, "step": 7056 }, { "epoch": 0.92, "grad_norm": 1.2957514524459839, "learning_rate": 1.4764418364803134e-07, "loss": 0.8212, "step": 7057 }, { "epoch": 0.92, "grad_norm": 0.5698044896125793, "learning_rate": 1.4713272792261368e-07, "loss": 0.9948, "step": 7058 }, { "epoch": 0.92, "grad_norm": 1.4076645374298096, "learning_rate": 1.466221463754336e-07, "loss": 0.816, "step": 7059 }, { "epoch": 0.93, "grad_norm": 1.3484408855438232, "learning_rate": 1.4611243909846529e-07, "loss": 0.7694, "step": 7060 }, { "epoch": 0.93, "grad_norm": 1.2943674325942993, "learning_rate": 1.4560360618352643e-07, "loss": 0.7606, "step": 7061 }, { "epoch": 0.93, "grad_norm": 0.5674381256103516, "learning_rate": 1.4509564772227535e-07, "loss": 0.9853, "step": 7062 }, { "epoch": 0.93, "grad_norm": 1.3615009784698486, "learning_rate": 1.445885638062139e-07, "loss": 0.8169, "step": 7063 }, { "epoch": 0.93, "grad_norm": 1.4704917669296265, "learning_rate": 1.4408235452668783e-07, "loss": 0.7916, "step": 7064 }, { "epoch": 0.93, "grad_norm": 1.274056077003479, "learning_rate": 1.435770199748826e-07, "loss": 0.8049, "step": 7065 }, { "epoch": 0.93, "grad_norm": 0.5737782120704651, "learning_rate": 1.4307256024182703e-07, "loss": 0.9994, "step": 7066 }, { "epoch": 0.93, "grad_norm": 0.5778781771659851, "learning_rate": 1.4256897541839343e-07, "loss": 0.9842, "step": 7067 }, { "epoch": 0.93, "grad_norm": 1.4367882013320923, "learning_rate": 1.420662655952959e-07, "loss": 0.7873, "step": 7068 }, { "epoch": 0.93, "grad_norm": 1.3676972389221191, "learning_rate": 1.4156443086309034e-07, "loss": 0.8086, "step": 7069 }, { "epoch": 0.93, "grad_norm": 1.3037816286087036, "learning_rate": 1.4106347131217612e-07, "loss": 0.7993, "step": 7070 }, { "epoch": 0.93, "grad_norm": 1.5994442701339722, "learning_rate": 1.4056338703279326e-07, "loss": 0.7814, "step": 7071 }, { "epoch": 0.93, "grad_norm": 0.5608426928520203, "learning_rate": 1.4006417811502582e-07, "loss": 1.0196, "step": 7072 }, { "epoch": 0.93, "grad_norm": 1.241241216659546, "learning_rate": 1.3956584464879964e-07, "loss": 0.8082, "step": 7073 }, { "epoch": 0.93, "grad_norm": 1.4704604148864746, "learning_rate": 1.3906838672388233e-07, "loss": 0.8192, "step": 7074 }, { "epoch": 0.93, "grad_norm": 1.2919594049453735, "learning_rate": 1.3857180442988506e-07, "loss": 0.8145, "step": 7075 }, { "epoch": 0.93, "grad_norm": 0.5715610980987549, "learning_rate": 1.3807609785625897e-07, "loss": 0.9955, "step": 7076 }, { "epoch": 0.93, "grad_norm": 1.3566553592681885, "learning_rate": 1.3758126709229992e-07, "loss": 0.7865, "step": 7077 }, { "epoch": 0.93, "grad_norm": 0.5510352253913879, "learning_rate": 1.370873122271449e-07, "loss": 1.0062, "step": 7078 }, { "epoch": 0.93, "grad_norm": 0.5572065114974976, "learning_rate": 1.3659423334977274e-07, "loss": 0.9954, "step": 7079 }, { "epoch": 0.93, "grad_norm": 1.7002222537994385, "learning_rate": 1.3610203054900571e-07, "loss": 0.7823, "step": 7080 }, { "epoch": 0.93, "grad_norm": 1.3156801462173462, "learning_rate": 1.3561070391350573e-07, "loss": 0.7873, "step": 7081 }, { "epoch": 0.93, "grad_norm": 1.3442686796188354, "learning_rate": 1.351202535317808e-07, "loss": 0.7752, "step": 7082 }, { "epoch": 0.93, "grad_norm": 1.4261780977249146, "learning_rate": 1.3463067949217756e-07, "loss": 0.8424, "step": 7083 }, { "epoch": 0.93, "grad_norm": 1.3849754333496094, "learning_rate": 1.3414198188288708e-07, "loss": 0.8601, "step": 7084 }, { "epoch": 0.93, "grad_norm": 1.294586420059204, "learning_rate": 1.336541607919406e-07, "loss": 0.7759, "step": 7085 }, { "epoch": 0.93, "grad_norm": 1.2957004308700562, "learning_rate": 1.3316721630721285e-07, "loss": 0.8654, "step": 7086 }, { "epoch": 0.93, "grad_norm": 1.3746182918548584, "learning_rate": 1.3268114851642032e-07, "loss": 0.8305, "step": 7087 }, { "epoch": 0.93, "grad_norm": 1.3170340061187744, "learning_rate": 1.3219595750712133e-07, "loss": 0.7572, "step": 7088 }, { "epoch": 0.93, "grad_norm": 2.8450067043304443, "learning_rate": 1.3171164336671705e-07, "loss": 0.8349, "step": 7089 }, { "epoch": 0.93, "grad_norm": 0.572658121585846, "learning_rate": 1.3122820618244936e-07, "loss": 0.9899, "step": 7090 }, { "epoch": 0.93, "grad_norm": 1.338787317276001, "learning_rate": 1.3074564604140362e-07, "loss": 0.7714, "step": 7091 }, { "epoch": 0.93, "grad_norm": 1.3111226558685303, "learning_rate": 1.3026396303050637e-07, "loss": 0.8303, "step": 7092 }, { "epoch": 0.93, "grad_norm": 1.222118854522705, "learning_rate": 1.2978315723652601e-07, "loss": 0.8025, "step": 7093 }, { "epoch": 0.93, "grad_norm": 1.3423233032226562, "learning_rate": 1.2930322874607272e-07, "loss": 0.7913, "step": 7094 }, { "epoch": 0.93, "grad_norm": 1.3717546463012695, "learning_rate": 1.288241776455995e-07, "loss": 0.7689, "step": 7095 }, { "epoch": 0.93, "grad_norm": 1.2571496963500977, "learning_rate": 1.2834600402140128e-07, "loss": 0.7555, "step": 7096 }, { "epoch": 0.93, "grad_norm": 1.277923583984375, "learning_rate": 1.278687079596147e-07, "loss": 0.793, "step": 7097 }, { "epoch": 0.93, "grad_norm": 1.326122760772705, "learning_rate": 1.2739228954621764e-07, "loss": 0.8072, "step": 7098 }, { "epoch": 0.93, "grad_norm": 0.5603311657905579, "learning_rate": 1.2691674886702976e-07, "loss": 0.9893, "step": 7099 }, { "epoch": 0.93, "grad_norm": 1.3401535749435425, "learning_rate": 1.2644208600771534e-07, "loss": 0.7705, "step": 7100 }, { "epoch": 0.93, "grad_norm": 1.343932867050171, "learning_rate": 1.25968301053776e-07, "loss": 0.8651, "step": 7101 }, { "epoch": 0.93, "grad_norm": 1.5183576345443726, "learning_rate": 1.2549539409055955e-07, "loss": 0.8252, "step": 7102 }, { "epoch": 0.93, "grad_norm": 1.395361304283142, "learning_rate": 1.250233652032534e-07, "loss": 0.7738, "step": 7103 }, { "epoch": 0.93, "grad_norm": 1.3911696672439575, "learning_rate": 1.2455221447688571e-07, "loss": 0.7794, "step": 7104 }, { "epoch": 0.93, "grad_norm": 0.5495386719703674, "learning_rate": 1.2408194199633016e-07, "loss": 1.0141, "step": 7105 }, { "epoch": 0.93, "grad_norm": 0.5733239054679871, "learning_rate": 1.2361254784629794e-07, "loss": 0.9813, "step": 7106 }, { "epoch": 0.93, "grad_norm": 1.395258903503418, "learning_rate": 1.2314403211134528e-07, "loss": 0.7379, "step": 7107 }, { "epoch": 0.93, "grad_norm": 0.5747294425964355, "learning_rate": 1.2267639487586858e-07, "loss": 1.0049, "step": 7108 }, { "epoch": 0.93, "grad_norm": 1.7336031198501587, "learning_rate": 1.2220963622410598e-07, "loss": 0.8352, "step": 7109 }, { "epoch": 0.93, "grad_norm": 1.3369791507720947, "learning_rate": 1.2174375624013858e-07, "loss": 0.7748, "step": 7110 }, { "epoch": 0.93, "grad_norm": 1.2412315607070923, "learning_rate": 1.2127875500788754e-07, "loss": 0.7307, "step": 7111 }, { "epoch": 0.93, "grad_norm": 1.3827316761016846, "learning_rate": 1.2081463261111704e-07, "loss": 0.8462, "step": 7112 }, { "epoch": 0.93, "grad_norm": 1.3331955671310425, "learning_rate": 1.2035138913343236e-07, "loss": 0.8286, "step": 7113 }, { "epoch": 0.93, "grad_norm": 1.2332327365875244, "learning_rate": 1.1988902465828013e-07, "loss": 0.7436, "step": 7114 }, { "epoch": 0.93, "grad_norm": 1.3611539602279663, "learning_rate": 1.194275392689498e-07, "loss": 0.8002, "step": 7115 }, { "epoch": 0.93, "grad_norm": 3.0514583587646484, "learning_rate": 1.1896693304857044e-07, "loss": 0.8296, "step": 7116 }, { "epoch": 0.93, "grad_norm": 0.5551863312721252, "learning_rate": 1.1850720608011513e-07, "loss": 0.9957, "step": 7117 }, { "epoch": 0.93, "grad_norm": 1.3644860982894897, "learning_rate": 1.1804835844639761e-07, "loss": 0.8063, "step": 7118 }, { "epoch": 0.93, "grad_norm": 1.420326590538025, "learning_rate": 1.175903902300718e-07, "loss": 0.7983, "step": 7119 }, { "epoch": 0.93, "grad_norm": 1.3886346817016602, "learning_rate": 1.1713330151363555e-07, "loss": 0.7267, "step": 7120 }, { "epoch": 0.93, "grad_norm": 1.269600510597229, "learning_rate": 1.1667709237942637e-07, "loss": 0.7806, "step": 7121 }, { "epoch": 0.93, "grad_norm": 1.4490602016448975, "learning_rate": 1.1622176290962517e-07, "loss": 0.8336, "step": 7122 }, { "epoch": 0.93, "grad_norm": 1.4503173828125, "learning_rate": 1.1576731318625301e-07, "loss": 0.8231, "step": 7123 }, { "epoch": 0.93, "grad_norm": 1.387004017829895, "learning_rate": 1.1531374329117217e-07, "loss": 0.7993, "step": 7124 }, { "epoch": 0.93, "grad_norm": 1.3047518730163574, "learning_rate": 1.1486105330608844e-07, "loss": 0.7922, "step": 7125 }, { "epoch": 0.93, "grad_norm": 1.453872799873352, "learning_rate": 1.1440924331254655e-07, "loss": 0.808, "step": 7126 }, { "epoch": 0.93, "grad_norm": 1.3516085147857666, "learning_rate": 1.1395831339193419e-07, "loss": 0.8655, "step": 7127 }, { "epoch": 0.93, "grad_norm": 0.559583306312561, "learning_rate": 1.1350826362548028e-07, "loss": 1.0206, "step": 7128 }, { "epoch": 0.93, "grad_norm": 1.362630844116211, "learning_rate": 1.1305909409425608e-07, "loss": 0.8561, "step": 7129 }, { "epoch": 0.93, "grad_norm": 1.3404490947723389, "learning_rate": 1.1261080487917241e-07, "loss": 0.8025, "step": 7130 }, { "epoch": 0.93, "grad_norm": 1.316226840019226, "learning_rate": 1.1216339606098247e-07, "loss": 0.8455, "step": 7131 }, { "epoch": 0.93, "grad_norm": 1.4255635738372803, "learning_rate": 1.1171686772028123e-07, "loss": 0.7951, "step": 7132 }, { "epoch": 0.93, "grad_norm": 0.5696969032287598, "learning_rate": 1.112712199375049e-07, "loss": 1.0026, "step": 7133 }, { "epoch": 0.93, "grad_norm": 1.3952422142028809, "learning_rate": 1.1082645279293036e-07, "loss": 0.8553, "step": 7134 }, { "epoch": 0.93, "grad_norm": 1.2690383195877075, "learning_rate": 1.1038256636667687e-07, "loss": 0.8028, "step": 7135 }, { "epoch": 0.94, "grad_norm": 1.3067325353622437, "learning_rate": 1.0993956073870438e-07, "loss": 0.8232, "step": 7136 }, { "epoch": 0.94, "grad_norm": 0.559071958065033, "learning_rate": 1.0949743598881401e-07, "loss": 0.9959, "step": 7137 }, { "epoch": 0.94, "grad_norm": 1.4937598705291748, "learning_rate": 1.0905619219664932e-07, "loss": 0.799, "step": 7138 }, { "epoch": 0.94, "grad_norm": 0.57149738073349, "learning_rate": 1.0861582944169335e-07, "loss": 1.0069, "step": 7139 }, { "epoch": 0.94, "grad_norm": 1.280074119567871, "learning_rate": 1.0817634780327213e-07, "loss": 0.7596, "step": 7140 }, { "epoch": 0.94, "grad_norm": 1.5910214185714722, "learning_rate": 1.0773774736055287e-07, "loss": 0.7928, "step": 7141 }, { "epoch": 0.94, "grad_norm": 2.085857391357422, "learning_rate": 1.0730002819254236e-07, "loss": 0.8195, "step": 7142 }, { "epoch": 0.94, "grad_norm": 1.5303434133529663, "learning_rate": 1.0686319037809034e-07, "loss": 0.7807, "step": 7143 }, { "epoch": 0.94, "grad_norm": 1.295912265777588, "learning_rate": 1.0642723399588662e-07, "loss": 0.7577, "step": 7144 }, { "epoch": 0.94, "grad_norm": 1.4153860807418823, "learning_rate": 1.0599215912446447e-07, "loss": 0.7853, "step": 7145 }, { "epoch": 0.94, "grad_norm": 1.4465440511703491, "learning_rate": 1.0555796584219514e-07, "loss": 0.7768, "step": 7146 }, { "epoch": 0.94, "grad_norm": 0.5677692294120789, "learning_rate": 1.0512465422729267e-07, "loss": 1.0103, "step": 7147 }, { "epoch": 0.94, "grad_norm": 1.4147875308990479, "learning_rate": 1.0469222435781357e-07, "loss": 0.7745, "step": 7148 }, { "epoch": 0.94, "grad_norm": 0.5542568564414978, "learning_rate": 1.0426067631165271e-07, "loss": 0.9972, "step": 7149 }, { "epoch": 0.94, "grad_norm": 5.567292213439941, "learning_rate": 1.0383001016654848e-07, "loss": 0.7878, "step": 7150 }, { "epoch": 0.94, "grad_norm": 1.29555344581604, "learning_rate": 1.0340022600007992e-07, "loss": 0.8209, "step": 7151 }, { "epoch": 0.94, "grad_norm": 0.57980877161026, "learning_rate": 1.0297132388966513e-07, "loss": 0.9887, "step": 7152 }, { "epoch": 0.94, "grad_norm": 1.3072631359100342, "learning_rate": 1.0254330391256729e-07, "loss": 0.8324, "step": 7153 }, { "epoch": 0.94, "grad_norm": 1.33707857131958, "learning_rate": 1.0211616614588693e-07, "loss": 0.8002, "step": 7154 }, { "epoch": 0.94, "grad_norm": 1.354453206062317, "learning_rate": 1.0168991066656698e-07, "loss": 0.8002, "step": 7155 }, { "epoch": 0.94, "grad_norm": 1.3956657648086548, "learning_rate": 1.0126453755139266e-07, "loss": 0.7949, "step": 7156 }, { "epoch": 0.94, "grad_norm": 1.5704582929611206, "learning_rate": 1.0084004687698823e-07, "loss": 0.8394, "step": 7157 }, { "epoch": 0.94, "grad_norm": 1.4237987995147705, "learning_rate": 1.0041643871982032e-07, "loss": 0.8572, "step": 7158 }, { "epoch": 0.94, "grad_norm": 1.2544734477996826, "learning_rate": 9.999371315619622e-08, "loss": 0.7182, "step": 7159 }, { "epoch": 0.94, "grad_norm": 1.321500301361084, "learning_rate": 9.957187026226389e-08, "loss": 0.8003, "step": 7160 }, { "epoch": 0.94, "grad_norm": 1.3648598194122314, "learning_rate": 9.915091011401313e-08, "loss": 0.8349, "step": 7161 }, { "epoch": 0.94, "grad_norm": 0.5702435970306396, "learning_rate": 9.87308327872738e-08, "loss": 1.0003, "step": 7162 }, { "epoch": 0.94, "grad_norm": 0.558138370513916, "learning_rate": 9.831163835771706e-08, "loss": 1.0154, "step": 7163 }, { "epoch": 0.94, "grad_norm": 0.5578525066375732, "learning_rate": 9.789332690085585e-08, "loss": 1.0151, "step": 7164 }, { "epoch": 0.94, "grad_norm": 0.5483888387680054, "learning_rate": 9.74758984920421e-08, "loss": 1.0186, "step": 7165 }, { "epoch": 0.94, "grad_norm": 3.143605947494507, "learning_rate": 9.70593532064712e-08, "loss": 0.8019, "step": 7166 }, { "epoch": 0.94, "grad_norm": 1.3693132400512695, "learning_rate": 9.664369111917704e-08, "loss": 0.7696, "step": 7167 }, { "epoch": 0.94, "grad_norm": 1.2800546884536743, "learning_rate": 9.622891230503584e-08, "loss": 0.7925, "step": 7168 }, { "epoch": 0.94, "grad_norm": 1.4018089771270752, "learning_rate": 9.581501683876504e-08, "loss": 0.7824, "step": 7169 }, { "epoch": 0.94, "grad_norm": 1.340295433998108, "learning_rate": 9.540200479492112e-08, "loss": 0.8404, "step": 7170 }, { "epoch": 0.94, "grad_norm": 1.3375816345214844, "learning_rate": 9.498987624790401e-08, "loss": 0.7804, "step": 7171 }, { "epoch": 0.94, "grad_norm": 1.3300743103027344, "learning_rate": 9.457863127195099e-08, "loss": 0.8249, "step": 7172 }, { "epoch": 0.94, "grad_norm": 1.4531288146972656, "learning_rate": 9.416826994114447e-08, "loss": 0.7786, "step": 7173 }, { "epoch": 0.94, "grad_norm": 0.5523495674133301, "learning_rate": 9.375879232940477e-08, "loss": 0.9881, "step": 7174 }, { "epoch": 0.94, "grad_norm": 1.250806450843811, "learning_rate": 9.335019851049288e-08, "loss": 0.7904, "step": 7175 }, { "epoch": 0.94, "grad_norm": 1.5574228763580322, "learning_rate": 9.294248855801269e-08, "loss": 0.7717, "step": 7176 }, { "epoch": 0.94, "grad_norm": 1.405004620552063, "learning_rate": 9.253566254540603e-08, "loss": 0.7444, "step": 7177 }, { "epoch": 0.94, "grad_norm": 1.3706046342849731, "learning_rate": 9.21297205459587e-08, "loss": 0.8133, "step": 7178 }, { "epoch": 0.94, "grad_norm": 0.569243311882019, "learning_rate": 9.172466263279555e-08, "loss": 1.0122, "step": 7179 }, { "epoch": 0.94, "grad_norm": 0.5752335786819458, "learning_rate": 9.132048887888101e-08, "loss": 1.008, "step": 7180 }, { "epoch": 0.94, "grad_norm": 1.3636524677276611, "learning_rate": 9.091719935702236e-08, "loss": 0.7907, "step": 7181 }, { "epoch": 0.94, "grad_norm": 1.3961395025253296, "learning_rate": 9.051479413986708e-08, "loss": 0.7727, "step": 7182 }, { "epoch": 0.94, "grad_norm": 1.419762134552002, "learning_rate": 9.01132732999016e-08, "loss": 0.7847, "step": 7183 }, { "epoch": 0.94, "grad_norm": 0.5689228773117065, "learning_rate": 8.971263690945587e-08, "loss": 1.0257, "step": 7184 }, { "epoch": 0.94, "grad_norm": 1.319989562034607, "learning_rate": 8.931288504069879e-08, "loss": 0.7813, "step": 7185 }, { "epoch": 0.94, "grad_norm": 1.285649061203003, "learning_rate": 8.891401776564001e-08, "loss": 0.7853, "step": 7186 }, { "epoch": 0.94, "grad_norm": 1.5328887701034546, "learning_rate": 8.851603515613038e-08, "loss": 0.7691, "step": 7187 }, { "epoch": 0.94, "grad_norm": 0.5729638338088989, "learning_rate": 8.811893728386033e-08, "loss": 0.9768, "step": 7188 }, { "epoch": 0.94, "grad_norm": 2.0220093727111816, "learning_rate": 8.772272422036265e-08, "loss": 0.7909, "step": 7189 }, { "epoch": 0.94, "grad_norm": 1.3351677656173706, "learning_rate": 8.732739603700857e-08, "loss": 0.7969, "step": 7190 }, { "epoch": 0.94, "grad_norm": 1.3526506423950195, "learning_rate": 8.693295280501224e-08, "loss": 0.8236, "step": 7191 }, { "epoch": 0.94, "grad_norm": 1.3325660228729248, "learning_rate": 8.653939459542682e-08, "loss": 0.7813, "step": 7192 }, { "epoch": 0.94, "grad_norm": 0.5733469724655151, "learning_rate": 8.614672147914616e-08, "loss": 0.9768, "step": 7193 }, { "epoch": 0.94, "grad_norm": 1.2913010120391846, "learning_rate": 8.57549335269059e-08, "loss": 0.7359, "step": 7194 }, { "epoch": 0.94, "grad_norm": 1.363594889640808, "learning_rate": 8.536403080928013e-08, "loss": 0.8145, "step": 7195 }, { "epoch": 0.94, "grad_norm": 1.3088243007659912, "learning_rate": 8.497401339668531e-08, "loss": 0.8028, "step": 7196 }, { "epoch": 0.94, "grad_norm": 1.598638892173767, "learning_rate": 8.458488135937859e-08, "loss": 0.7759, "step": 7197 }, { "epoch": 0.94, "grad_norm": 1.2907365560531616, "learning_rate": 8.419663476745499e-08, "loss": 0.7872, "step": 7198 }, { "epoch": 0.94, "grad_norm": 0.571816086769104, "learning_rate": 8.3809273690853e-08, "loss": 1.0067, "step": 7199 }, { "epoch": 0.94, "grad_norm": 1.2822012901306152, "learning_rate": 8.342279819935073e-08, "loss": 0.7827, "step": 7200 }, { "epoch": 0.94, "grad_norm": 0.5739362239837646, "learning_rate": 8.303720836256579e-08, "loss": 0.9972, "step": 7201 }, { "epoch": 0.94, "grad_norm": 1.3293633460998535, "learning_rate": 8.265250424995763e-08, "loss": 0.856, "step": 7202 }, { "epoch": 0.94, "grad_norm": 1.4024474620819092, "learning_rate": 8.22686859308247e-08, "loss": 0.8291, "step": 7203 }, { "epoch": 0.94, "grad_norm": 1.6134666204452515, "learning_rate": 8.188575347430783e-08, "loss": 0.7772, "step": 7204 }, { "epoch": 0.94, "grad_norm": 1.2956373691558838, "learning_rate": 8.150370694938626e-08, "loss": 0.8246, "step": 7205 }, { "epoch": 0.94, "grad_norm": 0.5637441277503967, "learning_rate": 8.112254642487993e-08, "loss": 0.9995, "step": 7206 }, { "epoch": 0.94, "grad_norm": 1.4777454137802124, "learning_rate": 8.074227196945061e-08, "loss": 0.7813, "step": 7207 }, { "epoch": 0.94, "grad_norm": 1.3416774272918701, "learning_rate": 8.03628836515996e-08, "loss": 0.7993, "step": 7208 }, { "epoch": 0.94, "grad_norm": 0.5644564628601074, "learning_rate": 7.998438153966837e-08, "loss": 1.0094, "step": 7209 }, { "epoch": 0.94, "grad_norm": 0.5684171915054321, "learning_rate": 7.960676570183901e-08, "loss": 1.0037, "step": 7210 }, { "epoch": 0.94, "grad_norm": 0.5602476596832275, "learning_rate": 7.923003620613379e-08, "loss": 1.0021, "step": 7211 }, { "epoch": 0.95, "grad_norm": 3.5400710105895996, "learning_rate": 7.885419312041509e-08, "loss": 0.793, "step": 7212 }, { "epoch": 0.95, "grad_norm": 1.3223555088043213, "learning_rate": 7.847923651238654e-08, "loss": 0.7663, "step": 7213 }, { "epoch": 0.95, "grad_norm": 1.2735298871994019, "learning_rate": 7.810516644959131e-08, "loss": 0.7759, "step": 7214 }, { "epoch": 0.95, "grad_norm": 0.5516714453697205, "learning_rate": 7.773198299941276e-08, "loss": 1.0296, "step": 7215 }, { "epoch": 0.95, "grad_norm": 0.5638626217842102, "learning_rate": 7.735968622907486e-08, "loss": 0.9927, "step": 7216 }, { "epoch": 0.95, "grad_norm": 1.2519932985305786, "learning_rate": 7.698827620564287e-08, "loss": 0.8134, "step": 7217 }, { "epoch": 0.95, "grad_norm": 0.5497806668281555, "learning_rate": 7.661775299601936e-08, "loss": 1.0115, "step": 7218 }, { "epoch": 0.95, "grad_norm": 1.275247573852539, "learning_rate": 7.624811666695098e-08, "loss": 0.7489, "step": 7219 }, { "epoch": 0.95, "grad_norm": 1.5456774234771729, "learning_rate": 7.587936728502166e-08, "loss": 0.7521, "step": 7220 }, { "epoch": 0.95, "grad_norm": 0.5667796730995178, "learning_rate": 7.55115049166566e-08, "loss": 0.9964, "step": 7221 }, { "epoch": 0.95, "grad_norm": 0.56172114610672, "learning_rate": 7.514452962812113e-08, "loss": 1.0147, "step": 7222 }, { "epoch": 0.95, "grad_norm": 1.3530352115631104, "learning_rate": 7.477844148552127e-08, "loss": 0.7766, "step": 7223 }, { "epoch": 0.95, "grad_norm": 0.5615549683570862, "learning_rate": 7.441324055480314e-08, "loss": 1.0292, "step": 7224 }, { "epoch": 0.95, "grad_norm": 1.4339919090270996, "learning_rate": 7.404892690175247e-08, "loss": 0.81, "step": 7225 }, { "epoch": 0.95, "grad_norm": 1.2388334274291992, "learning_rate": 7.368550059199453e-08, "loss": 0.7967, "step": 7226 }, { "epoch": 0.95, "grad_norm": 1.400059461593628, "learning_rate": 7.332296169099695e-08, "loss": 0.7952, "step": 7227 }, { "epoch": 0.95, "grad_norm": 1.3962647914886475, "learning_rate": 7.296131026406472e-08, "loss": 0.7747, "step": 7228 }, { "epoch": 0.95, "grad_norm": 1.316717505455017, "learning_rate": 7.260054637634628e-08, "loss": 0.8038, "step": 7229 }, { "epoch": 0.95, "grad_norm": 0.5621477365493774, "learning_rate": 7.224067009282742e-08, "loss": 1.0121, "step": 7230 }, { "epoch": 0.95, "grad_norm": 1.4132295846939087, "learning_rate": 7.188168147833464e-08, "loss": 0.7447, "step": 7231 }, { "epoch": 0.95, "grad_norm": 1.4075685739517212, "learning_rate": 7.152358059753506e-08, "loss": 0.7729, "step": 7232 }, { "epoch": 0.95, "grad_norm": 1.3671691417694092, "learning_rate": 7.1166367514936e-08, "loss": 0.7703, "step": 7233 }, { "epoch": 0.95, "grad_norm": 0.5672999620437622, "learning_rate": 7.081004229488431e-08, "loss": 1.0033, "step": 7234 }, { "epoch": 0.95, "grad_norm": 1.3290058374404907, "learning_rate": 7.045460500156753e-08, "loss": 0.821, "step": 7235 }, { "epoch": 0.95, "grad_norm": 1.3649070262908936, "learning_rate": 7.010005569901223e-08, "loss": 0.7913, "step": 7236 }, { "epoch": 0.95, "grad_norm": 2.1469168663024902, "learning_rate": 6.974639445108621e-08, "loss": 0.8044, "step": 7237 }, { "epoch": 0.95, "grad_norm": 1.3658604621887207, "learning_rate": 6.93936213214963e-08, "loss": 0.7743, "step": 7238 }, { "epoch": 0.95, "grad_norm": 1.2466174364089966, "learning_rate": 6.904173637378997e-08, "loss": 0.7829, "step": 7239 }, { "epoch": 0.95, "grad_norm": 1.8473166227340698, "learning_rate": 6.869073967135487e-08, "loss": 0.7982, "step": 7240 }, { "epoch": 0.95, "grad_norm": 0.5745725035667419, "learning_rate": 6.834063127741764e-08, "loss": 1.0152, "step": 7241 }, { "epoch": 0.95, "grad_norm": 1.2015951871871948, "learning_rate": 6.799141125504671e-08, "loss": 0.8077, "step": 7242 }, { "epoch": 0.95, "grad_norm": 1.419611930847168, "learning_rate": 6.76430796671479e-08, "loss": 0.7972, "step": 7243 }, { "epoch": 0.95, "grad_norm": 1.385027527809143, "learning_rate": 6.729563657646931e-08, "loss": 0.8323, "step": 7244 }, { "epoch": 0.95, "grad_norm": 1.443211317062378, "learning_rate": 6.694908204559758e-08, "loss": 0.7627, "step": 7245 }, { "epoch": 0.95, "grad_norm": 1.2804120779037476, "learning_rate": 6.660341613696053e-08, "loss": 0.7822, "step": 7246 }, { "epoch": 0.95, "grad_norm": 1.7137216329574585, "learning_rate": 6.625863891282447e-08, "loss": 0.7877, "step": 7247 }, { "epoch": 0.95, "grad_norm": 1.5343129634857178, "learning_rate": 6.591475043529638e-08, "loss": 0.826, "step": 7248 }, { "epoch": 0.95, "grad_norm": 0.5763525366783142, "learning_rate": 6.557175076632394e-08, "loss": 1.0016, "step": 7249 }, { "epoch": 0.95, "grad_norm": 0.5624454021453857, "learning_rate": 6.522963996769272e-08, "loss": 1.0122, "step": 7250 }, { "epoch": 0.95, "grad_norm": 1.2646101713180542, "learning_rate": 6.488841810103008e-08, "loss": 0.8021, "step": 7251 }, { "epoch": 0.95, "grad_norm": 1.5667364597320557, "learning_rate": 6.454808522780298e-08, "loss": 0.7826, "step": 7252 }, { "epoch": 0.95, "grad_norm": 1.3728066682815552, "learning_rate": 6.420864140931682e-08, "loss": 0.7671, "step": 7253 }, { "epoch": 0.95, "grad_norm": 1.3285659551620483, "learning_rate": 6.387008670671769e-08, "loss": 0.8062, "step": 7254 }, { "epoch": 0.95, "grad_norm": 1.348832607269287, "learning_rate": 6.353242118099234e-08, "loss": 0.788, "step": 7255 }, { "epoch": 0.95, "grad_norm": 1.7933588027954102, "learning_rate": 6.319564489296603e-08, "loss": 0.8049, "step": 7256 }, { "epoch": 0.95, "grad_norm": 1.3157188892364502, "learning_rate": 6.285975790330522e-08, "loss": 0.8234, "step": 7257 }, { "epoch": 0.95, "grad_norm": 1.3327863216400146, "learning_rate": 6.252476027251487e-08, "loss": 0.7497, "step": 7258 }, { "epoch": 0.95, "grad_norm": 1.2938780784606934, "learning_rate": 6.219065206094e-08, "loss": 0.7907, "step": 7259 }, { "epoch": 0.95, "grad_norm": 1.4357622861862183, "learning_rate": 6.185743332876637e-08, "loss": 0.8446, "step": 7260 }, { "epoch": 0.95, "grad_norm": 1.2818305492401123, "learning_rate": 6.15251041360182e-08, "loss": 0.7843, "step": 7261 }, { "epoch": 0.95, "grad_norm": 1.385692834854126, "learning_rate": 6.119366454256037e-08, "loss": 0.7861, "step": 7262 }, { "epoch": 0.95, "grad_norm": 1.2438286542892456, "learning_rate": 6.086311460809791e-08, "loss": 0.768, "step": 7263 }, { "epoch": 0.95, "grad_norm": 1.3396053314208984, "learning_rate": 6.053345439217318e-08, "loss": 0.8501, "step": 7264 }, { "epoch": 0.95, "grad_norm": 1.312435269355774, "learning_rate": 6.020468395417145e-08, "loss": 0.8219, "step": 7265 }, { "epoch": 0.95, "grad_norm": 1.3122931718826294, "learning_rate": 5.987680335331592e-08, "loss": 0.7862, "step": 7266 }, { "epoch": 0.95, "grad_norm": 1.652187705039978, "learning_rate": 5.954981264866988e-08, "loss": 0.8194, "step": 7267 }, { "epoch": 0.95, "grad_norm": 1.3075060844421387, "learning_rate": 5.922371189913567e-08, "loss": 0.8008, "step": 7268 }, { "epoch": 0.95, "grad_norm": 1.2688645124435425, "learning_rate": 5.889850116345686e-08, "loss": 0.7755, "step": 7269 }, { "epoch": 0.95, "grad_norm": 1.3217352628707886, "learning_rate": 5.857418050021491e-08, "loss": 0.7213, "step": 7270 }, { "epoch": 0.95, "grad_norm": 1.2682294845581055, "learning_rate": 5.825074996783253e-08, "loss": 0.8225, "step": 7271 }, { "epoch": 0.95, "grad_norm": 0.5702010989189148, "learning_rate": 5.7928209624570356e-08, "loss": 0.9808, "step": 7272 }, { "epoch": 0.95, "grad_norm": 1.5074584484100342, "learning_rate": 5.760655952853078e-08, "loss": 0.7937, "step": 7273 }, { "epoch": 0.95, "grad_norm": 1.5002999305725098, "learning_rate": 5.728579973765413e-08, "loss": 0.7926, "step": 7274 }, { "epoch": 0.95, "grad_norm": 0.5625356435775757, "learning_rate": 5.6965930309720866e-08, "loss": 1.0164, "step": 7275 }, { "epoch": 0.95, "grad_norm": 1.3525060415267944, "learning_rate": 5.664695130235154e-08, "loss": 0.8109, "step": 7276 }, { "epoch": 0.95, "grad_norm": 0.5639030337333679, "learning_rate": 5.63288627730052e-08, "loss": 1.0243, "step": 7277 }, { "epoch": 0.95, "grad_norm": 1.875187635421753, "learning_rate": 5.601166477898212e-08, "loss": 0.8458, "step": 7278 }, { "epoch": 0.95, "grad_norm": 1.3054765462875366, "learning_rate": 5.569535737742049e-08, "loss": 0.7295, "step": 7279 }, { "epoch": 0.95, "grad_norm": 1.269845724105835, "learning_rate": 5.537994062529917e-08, "loss": 0.8625, "step": 7280 }, { "epoch": 0.95, "grad_norm": 1.3612498044967651, "learning_rate": 5.506541457943548e-08, "loss": 0.8198, "step": 7281 }, { "epoch": 0.95, "grad_norm": 0.5639675855636597, "learning_rate": 5.4751779296488006e-08, "loss": 0.9934, "step": 7282 }, { "epoch": 0.95, "grad_norm": 1.2947946786880493, "learning_rate": 5.4439034832953764e-08, "loss": 0.7479, "step": 7283 }, { "epoch": 0.95, "grad_norm": 0.5616413354873657, "learning_rate": 5.412718124516825e-08, "loss": 0.9903, "step": 7284 }, { "epoch": 0.95, "grad_norm": 1.2516119480133057, "learning_rate": 5.3816218589309297e-08, "loss": 0.7531, "step": 7285 }, { "epoch": 0.95, "grad_norm": 0.5705171227455139, "learning_rate": 5.350614692139211e-08, "loss": 0.9807, "step": 7286 }, { "epoch": 0.95, "grad_norm": 1.3138827085494995, "learning_rate": 5.319696629727089e-08, "loss": 0.785, "step": 7287 }, { "epoch": 0.95, "grad_norm": 0.560437798500061, "learning_rate": 5.2888676772641644e-08, "loss": 1.0135, "step": 7288 }, { "epoch": 0.96, "grad_norm": 1.4809362888336182, "learning_rate": 5.258127840303773e-08, "loss": 0.8079, "step": 7289 }, { "epoch": 0.96, "grad_norm": 1.3466014862060547, "learning_rate": 5.227477124383318e-08, "loss": 0.7743, "step": 7290 }, { "epoch": 0.96, "grad_norm": 1.3315690755844116, "learning_rate": 5.196915535024105e-08, "loss": 0.8287, "step": 7291 }, { "epoch": 0.96, "grad_norm": 1.3826357126235962, "learning_rate": 5.1664430777313424e-08, "loss": 0.8242, "step": 7292 }, { "epoch": 0.96, "grad_norm": 1.3007802963256836, "learning_rate": 5.1360597579943606e-08, "loss": 0.8395, "step": 7293 }, { "epoch": 0.96, "grad_norm": 0.5752896070480347, "learning_rate": 5.105765581286115e-08, "loss": 1.0086, "step": 7294 }, { "epoch": 0.96, "grad_norm": 0.5730921030044556, "learning_rate": 5.075560553063852e-08, "loss": 1.0094, "step": 7295 }, { "epoch": 0.96, "grad_norm": 1.3325971364974976, "learning_rate": 5.045444678768496e-08, "loss": 0.7766, "step": 7296 }, { "epoch": 0.96, "grad_norm": 0.5725293159484863, "learning_rate": 5.0154179638250424e-08, "loss": 0.9981, "step": 7297 }, { "epoch": 0.96, "grad_norm": 1.3436943292617798, "learning_rate": 4.985480413642496e-08, "loss": 0.758, "step": 7298 }, { "epoch": 0.96, "grad_norm": 1.1979182958602905, "learning_rate": 4.955632033613544e-08, "loss": 0.7431, "step": 7299 }, { "epoch": 0.96, "grad_norm": 0.5599373579025269, "learning_rate": 4.925872829114997e-08, "loss": 1.0138, "step": 7300 }, { "epoch": 0.96, "grad_norm": 1.7423348426818848, "learning_rate": 4.896202805507677e-08, "loss": 0.7868, "step": 7301 }, { "epoch": 0.96, "grad_norm": 1.226104736328125, "learning_rate": 4.866621968136142e-08, "loss": 0.7502, "step": 7302 }, { "epoch": 0.96, "grad_norm": 1.523617148399353, "learning_rate": 4.83713032232902e-08, "loss": 0.8249, "step": 7303 }, { "epoch": 0.96, "grad_norm": 1.2697505950927734, "learning_rate": 4.807727873398838e-08, "loss": 0.6969, "step": 7304 }, { "epoch": 0.96, "grad_norm": 0.5770564079284668, "learning_rate": 4.7784146266419695e-08, "loss": 1.001, "step": 7305 }, { "epoch": 0.96, "grad_norm": 0.5683111548423767, "learning_rate": 4.7491905873389144e-08, "loss": 0.981, "step": 7306 }, { "epoch": 0.96, "grad_norm": 0.5765595436096191, "learning_rate": 4.72005576075385e-08, "loss": 0.9957, "step": 7307 }, { "epoch": 0.96, "grad_norm": 1.361348032951355, "learning_rate": 4.691010152135189e-08, "loss": 0.8088, "step": 7308 }, { "epoch": 0.96, "grad_norm": 1.39741051197052, "learning_rate": 4.662053766714969e-08, "loss": 0.7571, "step": 7309 }, { "epoch": 0.96, "grad_norm": 1.3718299865722656, "learning_rate": 4.6331866097093505e-08, "loss": 0.7729, "step": 7310 }, { "epoch": 0.96, "grad_norm": 0.5672656297683716, "learning_rate": 4.6044086863183403e-08, "loss": 1.0019, "step": 7311 }, { "epoch": 0.96, "grad_norm": 1.2675246000289917, "learning_rate": 4.575720001725903e-08, "loss": 0.8296, "step": 7312 }, { "epoch": 0.96, "grad_norm": 1.3607016801834106, "learning_rate": 4.547120561099905e-08, "loss": 0.7684, "step": 7313 }, { "epoch": 0.96, "grad_norm": 0.5592149496078491, "learning_rate": 4.518610369592169e-08, "loss": 1.0275, "step": 7314 }, { "epoch": 0.96, "grad_norm": 0.5581881403923035, "learning_rate": 4.490189432338365e-08, "loss": 0.9829, "step": 7315 }, { "epoch": 0.96, "grad_norm": 1.458592176437378, "learning_rate": 4.461857754458232e-08, "loss": 0.8278, "step": 7316 }, { "epoch": 0.96, "grad_norm": 0.5796078443527222, "learning_rate": 4.4336153410552416e-08, "loss": 1.0059, "step": 7317 }, { "epoch": 0.96, "grad_norm": 1.5281449556350708, "learning_rate": 4.405462197216937e-08, "loss": 0.7421, "step": 7318 }, { "epoch": 0.96, "grad_norm": 1.2698532342910767, "learning_rate": 4.377398328014704e-08, "loss": 0.816, "step": 7319 }, { "epoch": 0.96, "grad_norm": 1.2871546745300293, "learning_rate": 4.349423738503888e-08, "loss": 0.7865, "step": 7320 }, { "epoch": 0.96, "grad_norm": 0.5722664594650269, "learning_rate": 4.321538433723682e-08, "loss": 1.0022, "step": 7321 }, { "epoch": 0.96, "grad_norm": 0.5616286993026733, "learning_rate": 4.293742418697344e-08, "loss": 1.0055, "step": 7322 }, { "epoch": 0.96, "grad_norm": 1.3231550455093384, "learning_rate": 4.26603569843187e-08, "loss": 0.8393, "step": 7323 }, { "epoch": 0.96, "grad_norm": 1.3179271221160889, "learning_rate": 4.238418277918266e-08, "loss": 0.7546, "step": 7324 }, { "epoch": 0.96, "grad_norm": 1.3786520957946777, "learning_rate": 4.210890162131387e-08, "loss": 0.8034, "step": 7325 }, { "epoch": 0.96, "grad_norm": 1.2640987634658813, "learning_rate": 4.1834513560301545e-08, "loss": 0.7443, "step": 7326 }, { "epoch": 0.96, "grad_norm": 1.451143741607666, "learning_rate": 4.156101864557227e-08, "loss": 0.8295, "step": 7327 }, { "epoch": 0.96, "grad_norm": 0.5723615288734436, "learning_rate": 4.128841692639274e-08, "loss": 1.0099, "step": 7328 }, { "epoch": 0.96, "grad_norm": 1.2711105346679688, "learning_rate": 4.101670845186812e-08, "loss": 0.8408, "step": 7329 }, { "epoch": 0.96, "grad_norm": 1.6838395595550537, "learning_rate": 4.074589327094369e-08, "loss": 0.8224, "step": 7330 }, { "epoch": 0.96, "grad_norm": 1.2898602485656738, "learning_rate": 4.0475971432402675e-08, "loss": 0.8478, "step": 7331 }, { "epoch": 0.96, "grad_norm": 1.676215648651123, "learning_rate": 4.020694298486727e-08, "loss": 0.7584, "step": 7332 }, { "epoch": 0.96, "grad_norm": 1.310735821723938, "learning_rate": 3.993880797679983e-08, "loss": 0.8091, "step": 7333 }, { "epoch": 0.96, "grad_norm": 1.300054907798767, "learning_rate": 3.967156645650172e-08, "loss": 0.7943, "step": 7334 }, { "epoch": 0.96, "grad_norm": 0.5573461055755615, "learning_rate": 3.940521847211221e-08, "loss": 0.9989, "step": 7335 }, { "epoch": 0.96, "grad_norm": 1.2776504755020142, "learning_rate": 3.913976407161069e-08, "loss": 0.8081, "step": 7336 }, { "epoch": 0.96, "grad_norm": 1.4116599559783936, "learning_rate": 3.8875203302815024e-08, "loss": 0.8032, "step": 7337 }, { "epoch": 0.96, "grad_norm": 1.3764346837997437, "learning_rate": 3.861153621338154e-08, "loss": 0.7941, "step": 7338 }, { "epoch": 0.96, "grad_norm": 1.3619072437286377, "learning_rate": 3.8348762850807794e-08, "loss": 0.8272, "step": 7339 }, { "epoch": 0.96, "grad_norm": 0.5683006644248962, "learning_rate": 3.808688326242704e-08, "loss": 1.0128, "step": 7340 }, { "epoch": 0.96, "grad_norm": 1.4054405689239502, "learning_rate": 3.7825897495414874e-08, "loss": 0.7147, "step": 7341 }, { "epoch": 0.96, "grad_norm": 1.295374870300293, "learning_rate": 3.756580559678369e-08, "loss": 0.7995, "step": 7342 }, { "epoch": 0.96, "grad_norm": 1.429324746131897, "learning_rate": 3.730660761338545e-08, "loss": 0.7926, "step": 7343 }, { "epoch": 0.96, "grad_norm": 1.2779573202133179, "learning_rate": 3.7048303591911695e-08, "loss": 0.7932, "step": 7344 }, { "epoch": 0.96, "grad_norm": 0.5559040904045105, "learning_rate": 3.6790893578891874e-08, "loss": 0.9993, "step": 7345 }, { "epoch": 0.96, "grad_norm": 0.5601115226745605, "learning_rate": 3.6534377620695e-08, "loss": 1.0354, "step": 7346 }, { "epoch": 0.96, "grad_norm": 1.3040690422058105, "learning_rate": 3.627875576352857e-08, "loss": 0.8336, "step": 7347 }, { "epoch": 0.96, "grad_norm": 1.367562174797058, "learning_rate": 3.602402805344018e-08, "loss": 0.7574, "step": 7348 }, { "epoch": 0.96, "grad_norm": 1.554534673690796, "learning_rate": 3.5770194536314803e-08, "loss": 0.7822, "step": 7349 }, { "epoch": 0.96, "grad_norm": 1.633453130722046, "learning_rate": 3.551725525787808e-08, "loss": 0.7913, "step": 7350 }, { "epoch": 0.96, "grad_norm": 1.2975246906280518, "learning_rate": 3.526521026369245e-08, "loss": 0.7805, "step": 7351 }, { "epoch": 0.96, "grad_norm": 1.3334450721740723, "learning_rate": 3.50140595991616e-08, "loss": 0.7815, "step": 7352 }, { "epoch": 0.96, "grad_norm": 1.2630163431167603, "learning_rate": 3.4763803309525446e-08, "loss": 0.7334, "step": 7353 }, { "epoch": 0.96, "grad_norm": 1.2686779499053955, "learning_rate": 3.451444143986571e-08, "loss": 0.7865, "step": 7354 }, { "epoch": 0.96, "grad_norm": 1.340585470199585, "learning_rate": 3.42659740351009e-08, "loss": 0.7475, "step": 7355 }, { "epoch": 0.96, "grad_norm": 1.2894644737243652, "learning_rate": 3.401840113998855e-08, "loss": 0.7403, "step": 7356 }, { "epoch": 0.96, "grad_norm": 1.3759181499481201, "learning_rate": 3.3771722799126885e-08, "loss": 0.7353, "step": 7357 }, { "epoch": 0.96, "grad_norm": 1.3089823722839355, "learning_rate": 3.3525939056949786e-08, "loss": 0.7551, "step": 7358 }, { "epoch": 0.96, "grad_norm": 1.4788516759872437, "learning_rate": 3.328104995773407e-08, "loss": 0.7635, "step": 7359 }, { "epoch": 0.96, "grad_norm": 0.5732138752937317, "learning_rate": 3.303705554559167e-08, "loss": 0.9984, "step": 7360 }, { "epoch": 0.96, "grad_norm": 0.5666022300720215, "learning_rate": 3.27939558644752e-08, "loss": 1.0279, "step": 7361 }, { "epoch": 0.96, "grad_norm": 1.2540889978408813, "learning_rate": 3.255175095817575e-08, "loss": 0.8198, "step": 7362 }, { "epoch": 0.96, "grad_norm": 0.5627378225326538, "learning_rate": 3.2310440870323426e-08, "loss": 1.0026, "step": 7363 }, { "epoch": 0.96, "grad_norm": 0.5642465353012085, "learning_rate": 3.207002564438677e-08, "loss": 1.0256, "step": 7364 }, { "epoch": 0.97, "grad_norm": 1.3432929515838623, "learning_rate": 3.183050532367338e-08, "loss": 0.7662, "step": 7365 }, { "epoch": 0.97, "grad_norm": 1.269675612449646, "learning_rate": 3.159187995132984e-08, "loss": 0.7277, "step": 7366 }, { "epoch": 0.97, "grad_norm": 1.3370822668075562, "learning_rate": 3.135414957034122e-08, "loss": 0.8406, "step": 7367 }, { "epoch": 0.97, "grad_norm": 1.3802158832550049, "learning_rate": 3.1117314223531036e-08, "loss": 0.7241, "step": 7368 }, { "epoch": 0.97, "grad_norm": 1.2921582460403442, "learning_rate": 3.088137395356239e-08, "loss": 0.8071, "step": 7369 }, { "epoch": 0.97, "grad_norm": 1.340376615524292, "learning_rate": 3.0646328802936274e-08, "loss": 0.7909, "step": 7370 }, { "epoch": 0.97, "grad_norm": 1.3399347066879272, "learning_rate": 3.0412178813993274e-08, "loss": 0.8233, "step": 7371 }, { "epoch": 0.97, "grad_norm": 1.4239940643310547, "learning_rate": 3.017892402891243e-08, "loss": 0.7928, "step": 7372 }, { "epoch": 0.97, "grad_norm": 1.3291717767715454, "learning_rate": 2.994656448971123e-08, "loss": 0.7811, "step": 7373 }, { "epoch": 0.97, "grad_norm": 1.2867785692214966, "learning_rate": 2.9715100238245642e-08, "loss": 0.7646, "step": 7374 }, { "epoch": 0.97, "grad_norm": 1.2551723718643188, "learning_rate": 2.948453131621176e-08, "loss": 0.7707, "step": 7375 }, { "epoch": 0.97, "grad_norm": 1.27448570728302, "learning_rate": 2.9254857765142474e-08, "loss": 0.8146, "step": 7376 }, { "epoch": 0.97, "grad_norm": 1.2845827341079712, "learning_rate": 2.902607962641135e-08, "loss": 0.7995, "step": 7377 }, { "epoch": 0.97, "grad_norm": 1.3360058069229126, "learning_rate": 2.8798196941228762e-08, "loss": 0.8501, "step": 7378 }, { "epoch": 0.97, "grad_norm": 1.3051987886428833, "learning_rate": 2.8571209750645202e-08, "loss": 0.8312, "step": 7379 }, { "epoch": 0.97, "grad_norm": 1.2859607934951782, "learning_rate": 2.8345118095549073e-08, "loss": 0.8104, "step": 7380 }, { "epoch": 0.97, "grad_norm": 1.3219596147537231, "learning_rate": 2.81199220166678e-08, "loss": 0.8383, "step": 7381 }, { "epoch": 0.97, "grad_norm": 0.5569418668746948, "learning_rate": 2.789562155456782e-08, "loss": 0.9811, "step": 7382 }, { "epoch": 0.97, "grad_norm": 1.4647917747497559, "learning_rate": 2.7672216749652925e-08, "loss": 0.7575, "step": 7383 }, { "epoch": 0.97, "grad_norm": 1.3805371522903442, "learning_rate": 2.744970764216759e-08, "loss": 0.7561, "step": 7384 }, { "epoch": 0.97, "grad_norm": 1.3347785472869873, "learning_rate": 2.7228094272193086e-08, "loss": 0.8112, "step": 7385 }, { "epoch": 0.97, "grad_norm": 0.5620371699333191, "learning_rate": 2.70073766796497e-08, "loss": 1.0147, "step": 7386 }, { "epoch": 0.97, "grad_norm": 1.4590967893600464, "learning_rate": 2.6787554904297852e-08, "loss": 0.8314, "step": 7387 }, { "epoch": 0.97, "grad_norm": 0.5638197064399719, "learning_rate": 2.65686289857342e-08, "loss": 0.9976, "step": 7388 }, { "epoch": 0.97, "grad_norm": 0.5779762864112854, "learning_rate": 2.635059896339609e-08, "loss": 1.0185, "step": 7389 }, { "epoch": 0.97, "grad_norm": 0.582581102848053, "learning_rate": 2.6133464876558768e-08, "loss": 1.0024, "step": 7390 }, { "epoch": 0.97, "grad_norm": 1.3467961549758911, "learning_rate": 2.591722676433539e-08, "loss": 0.8062, "step": 7391 }, { "epoch": 0.97, "grad_norm": 1.3913062810897827, "learning_rate": 2.570188466567869e-08, "loss": 0.7785, "step": 7392 }, { "epoch": 0.97, "grad_norm": 0.5575605034828186, "learning_rate": 2.5487438619379303e-08, "loss": 0.9789, "step": 7393 }, { "epoch": 0.97, "grad_norm": 1.3527270555496216, "learning_rate": 2.5273888664067437e-08, "loss": 0.8167, "step": 7394 }, { "epoch": 0.97, "grad_norm": 1.2857071161270142, "learning_rate": 2.5061234838210656e-08, "loss": 0.8053, "step": 7395 }, { "epoch": 0.97, "grad_norm": 1.3077257871627808, "learning_rate": 2.4849477180115544e-08, "loss": 0.7748, "step": 7396 }, { "epoch": 0.97, "grad_norm": 0.5741165280342102, "learning_rate": 2.463861572792825e-08, "loss": 1.0263, "step": 7397 }, { "epoch": 0.97, "grad_norm": 3.6821157932281494, "learning_rate": 2.4428650519631725e-08, "loss": 0.8325, "step": 7398 }, { "epoch": 0.97, "grad_norm": 1.359393835067749, "learning_rate": 2.421958159304849e-08, "loss": 0.8155, "step": 7399 }, { "epoch": 0.97, "grad_norm": 1.3023029565811157, "learning_rate": 2.4011408985839535e-08, "loss": 0.7462, "step": 7400 }, { "epoch": 0.97, "grad_norm": 1.9251415729522705, "learning_rate": 2.380413273550486e-08, "loss": 0.783, "step": 7401 }, { "epoch": 0.97, "grad_norm": 1.3454855680465698, "learning_rate": 2.3597752879381818e-08, "loss": 0.7607, "step": 7402 }, { "epoch": 0.97, "grad_norm": 1.3317538499832153, "learning_rate": 2.339226945464734e-08, "loss": 0.8091, "step": 7403 }, { "epoch": 0.97, "grad_norm": 1.3881996870040894, "learning_rate": 2.318768249831571e-08, "loss": 0.7894, "step": 7404 }, { "epoch": 0.97, "grad_norm": 0.5779578685760498, "learning_rate": 2.2983992047241334e-08, "loss": 0.9811, "step": 7405 }, { "epoch": 0.97, "grad_norm": 0.5741367936134338, "learning_rate": 2.278119813811652e-08, "loss": 1.0324, "step": 7406 }, { "epoch": 0.97, "grad_norm": 1.6149898767471313, "learning_rate": 2.257930080747095e-08, "loss": 0.8023, "step": 7407 }, { "epoch": 0.97, "grad_norm": 1.3171443939208984, "learning_rate": 2.2378300091673853e-08, "loss": 0.7678, "step": 7408 }, { "epoch": 0.97, "grad_norm": 0.5674026012420654, "learning_rate": 2.2178196026933495e-08, "loss": 0.9946, "step": 7409 }, { "epoch": 0.97, "grad_norm": 1.3053455352783203, "learning_rate": 2.197898864929493e-08, "loss": 0.7815, "step": 7410 }, { "epoch": 0.97, "grad_norm": 1.2832109928131104, "learning_rate": 2.1780677994643895e-08, "loss": 0.7896, "step": 7411 }, { "epoch": 0.97, "grad_norm": 1.3852967023849487, "learning_rate": 2.1583264098702374e-08, "loss": 0.7953, "step": 7412 }, { "epoch": 0.97, "grad_norm": 1.3239359855651855, "learning_rate": 2.1386746997031914e-08, "loss": 0.7884, "step": 7413 }, { "epoch": 0.97, "grad_norm": 0.5758322477340698, "learning_rate": 2.1191126725031984e-08, "loss": 1.0126, "step": 7414 }, { "epoch": 0.97, "grad_norm": 1.2928119897842407, "learning_rate": 2.0996403317942172e-08, "loss": 0.7701, "step": 7415 }, { "epoch": 0.97, "grad_norm": 1.2635372877120972, "learning_rate": 2.080257681083886e-08, "loss": 0.8079, "step": 7416 }, { "epoch": 0.97, "grad_norm": 1.2961390018463135, "learning_rate": 2.0609647238636343e-08, "loss": 0.8199, "step": 7417 }, { "epoch": 0.97, "grad_norm": 1.4856441020965576, "learning_rate": 2.04176146360896e-08, "loss": 0.7822, "step": 7418 }, { "epoch": 0.97, "grad_norm": 1.2591180801391602, "learning_rate": 2.0226479037789292e-08, "loss": 0.7362, "step": 7419 }, { "epoch": 0.97, "grad_norm": 1.3347519636154175, "learning_rate": 2.0036240478167323e-08, "loss": 0.8382, "step": 7420 }, { "epoch": 0.97, "grad_norm": 1.4020726680755615, "learning_rate": 1.9846898991491283e-08, "loss": 0.8016, "step": 7421 }, { "epoch": 0.97, "grad_norm": 1.3231815099716187, "learning_rate": 1.965845461186944e-08, "loss": 0.7566, "step": 7422 }, { "epoch": 0.97, "grad_norm": 1.6475144624710083, "learning_rate": 1.947090737324686e-08, "loss": 0.7813, "step": 7423 }, { "epoch": 0.97, "grad_norm": 0.5670250654220581, "learning_rate": 1.9284257309408194e-08, "loss": 1.017, "step": 7424 }, { "epoch": 0.97, "grad_norm": 0.5654461979866028, "learning_rate": 1.909850445397543e-08, "loss": 1.0039, "step": 7425 }, { "epoch": 0.97, "grad_norm": 0.5602194666862488, "learning_rate": 1.8913648840410136e-08, "loss": 1.0097, "step": 7426 }, { "epoch": 0.97, "grad_norm": 1.340325117111206, "learning_rate": 1.872969050201068e-08, "loss": 0.8032, "step": 7427 }, { "epoch": 0.97, "grad_norm": 1.2958513498306274, "learning_rate": 1.854662947191499e-08, "loss": 0.8394, "step": 7428 }, { "epoch": 0.97, "grad_norm": 1.4217232465744019, "learning_rate": 1.8364465783098918e-08, "loss": 0.7889, "step": 7429 }, { "epoch": 0.97, "grad_norm": 1.37424635887146, "learning_rate": 1.8183199468377322e-08, "loss": 0.7877, "step": 7430 }, { "epoch": 0.97, "grad_norm": 1.3729945421218872, "learning_rate": 1.800283056040242e-08, "loss": 0.7774, "step": 7431 }, { "epoch": 0.97, "grad_norm": 1.4345204830169678, "learning_rate": 1.7823359091664882e-08, "loss": 0.7424, "step": 7432 }, { "epoch": 0.97, "grad_norm": 1.396033525466919, "learning_rate": 1.7644785094494964e-08, "loss": 0.7979, "step": 7433 }, { "epoch": 0.97, "grad_norm": 1.2599576711654663, "learning_rate": 1.7467108601059713e-08, "loss": 0.7603, "step": 7434 }, { "epoch": 0.97, "grad_norm": 1.5000718832015991, "learning_rate": 1.729032964336519e-08, "loss": 0.7149, "step": 7435 }, { "epoch": 0.97, "grad_norm": 1.4378604888916016, "learning_rate": 1.7114448253255924e-08, "loss": 0.7848, "step": 7436 }, { "epoch": 0.97, "grad_norm": 0.5621634721755981, "learning_rate": 1.6939464462414347e-08, "loss": 0.9849, "step": 7437 }, { "epoch": 0.97, "grad_norm": 1.3458172082901, "learning_rate": 1.6765378302361357e-08, "loss": 0.7958, "step": 7438 }, { "epoch": 0.97, "grad_norm": 1.3142374753952026, "learning_rate": 1.659218980445687e-08, "loss": 0.8035, "step": 7439 }, { "epoch": 0.97, "grad_norm": 1.278994083404541, "learning_rate": 1.64198989998976e-08, "loss": 0.7728, "step": 7440 }, { "epoch": 0.98, "grad_norm": 1.337620735168457, "learning_rate": 1.6248505919719825e-08, "loss": 0.7825, "step": 7441 }, { "epoch": 0.98, "grad_norm": 1.368904709815979, "learning_rate": 1.6078010594797742e-08, "loss": 0.7376, "step": 7442 }, { "epoch": 0.98, "grad_norm": 1.4009228944778442, "learning_rate": 1.590841305584345e-08, "loss": 0.8518, "step": 7443 }, { "epoch": 0.98, "grad_norm": 0.5687233805656433, "learning_rate": 1.5739713333408068e-08, "loss": 1.0028, "step": 7444 }, { "epoch": 0.98, "grad_norm": 1.4437930583953857, "learning_rate": 1.557191145788006e-08, "loss": 0.8201, "step": 7445 }, { "epoch": 0.98, "grad_norm": 1.3508877754211426, "learning_rate": 1.5405007459487475e-08, "loss": 0.7646, "step": 7446 }, { "epoch": 0.98, "grad_norm": 1.3431992530822754, "learning_rate": 1.5239001368294593e-08, "loss": 0.7171, "step": 7447 }, { "epoch": 0.98, "grad_norm": 1.3955057859420776, "learning_rate": 1.5073893214206383e-08, "loss": 0.8054, "step": 7448 }, { "epoch": 0.98, "grad_norm": 1.559317708015442, "learning_rate": 1.490968302696405e-08, "loss": 0.8202, "step": 7449 }, { "epoch": 0.98, "grad_norm": 0.5802846550941467, "learning_rate": 1.4746370836147827e-08, "loss": 0.989, "step": 7450 }, { "epoch": 0.98, "grad_norm": 1.2621636390686035, "learning_rate": 1.4583956671176957e-08, "loss": 0.7647, "step": 7451 }, { "epoch": 0.98, "grad_norm": 1.267615556716919, "learning_rate": 1.4422440561307482e-08, "loss": 0.7439, "step": 7452 }, { "epoch": 0.98, "grad_norm": 1.357882022857666, "learning_rate": 1.4261822535633907e-08, "loss": 0.7901, "step": 7453 }, { "epoch": 0.98, "grad_norm": 0.5702875852584839, "learning_rate": 1.4102102623090863e-08, "loss": 1.0116, "step": 7454 }, { "epoch": 0.98, "grad_norm": 1.2555882930755615, "learning_rate": 1.3943280852448116e-08, "loss": 0.8283, "step": 7455 }, { "epoch": 0.98, "grad_norm": 1.3888946771621704, "learning_rate": 1.3785357252316666e-08, "loss": 0.8348, "step": 7456 }, { "epoch": 0.98, "grad_norm": 1.3774981498718262, "learning_rate": 1.3628331851142651e-08, "loss": 0.7886, "step": 7457 }, { "epoch": 0.98, "grad_norm": 1.4366710186004639, "learning_rate": 1.3472204677213995e-08, "loss": 0.8386, "step": 7458 }, { "epoch": 0.98, "grad_norm": 1.375590205192566, "learning_rate": 1.3316975758653761e-08, "loss": 0.8009, "step": 7459 }, { "epoch": 0.98, "grad_norm": 1.2910913228988647, "learning_rate": 1.3162645123424022e-08, "loss": 0.7488, "step": 7460 }, { "epoch": 0.98, "grad_norm": 1.4226053953170776, "learning_rate": 1.3009212799325877e-08, "loss": 0.8042, "step": 7461 }, { "epoch": 0.98, "grad_norm": 1.4259495735168457, "learning_rate": 1.2856678813998325e-08, "loss": 0.8469, "step": 7462 }, { "epoch": 0.98, "grad_norm": 1.296939492225647, "learning_rate": 1.2705043194918276e-08, "loss": 0.8469, "step": 7463 }, { "epoch": 0.98, "grad_norm": 1.548366665840149, "learning_rate": 1.2554305969399993e-08, "loss": 0.8168, "step": 7464 }, { "epoch": 0.98, "grad_norm": 0.5766439437866211, "learning_rate": 1.2404467164597311e-08, "loss": 1.0366, "step": 7465 }, { "epoch": 0.98, "grad_norm": 1.299241304397583, "learning_rate": 1.2255526807501971e-08, "loss": 0.7406, "step": 7466 }, { "epoch": 0.98, "grad_norm": 1.3744921684265137, "learning_rate": 1.210748492494307e-08, "loss": 0.7938, "step": 7467 }, { "epoch": 0.98, "grad_norm": 1.2404675483703613, "learning_rate": 1.1960341543588716e-08, "loss": 0.8158, "step": 7468 }, { "epoch": 0.98, "grad_norm": 1.3273786306381226, "learning_rate": 1.1814096689944932e-08, "loss": 0.7823, "step": 7469 }, { "epoch": 0.98, "grad_norm": 1.3411787748336792, "learning_rate": 1.1668750390354533e-08, "loss": 0.7633, "step": 7470 }, { "epoch": 0.98, "grad_norm": 1.2939366102218628, "learning_rate": 1.152430267100102e-08, "loss": 0.7536, "step": 7471 }, { "epoch": 0.98, "grad_norm": 1.4231412410736084, "learning_rate": 1.138075355790469e-08, "loss": 0.7912, "step": 7472 }, { "epoch": 0.98, "grad_norm": 1.327249526977539, "learning_rate": 1.1238103076922635e-08, "loss": 0.8543, "step": 7473 }, { "epoch": 0.98, "grad_norm": 1.3493508100509644, "learning_rate": 1.1096351253752635e-08, "loss": 0.7775, "step": 7474 }, { "epoch": 0.98, "grad_norm": 1.3303964138031006, "learning_rate": 1.0955498113929264e-08, "loss": 0.7899, "step": 7475 }, { "epoch": 0.98, "grad_norm": 1.6778603792190552, "learning_rate": 1.0815543682825003e-08, "loss": 0.8086, "step": 7476 }, { "epoch": 0.98, "grad_norm": 0.580108106136322, "learning_rate": 1.0676487985650242e-08, "loss": 0.9929, "step": 7477 }, { "epoch": 0.98, "grad_norm": 1.3141077756881714, "learning_rate": 1.0538331047454941e-08, "loss": 0.7799, "step": 7478 }, { "epoch": 0.98, "grad_norm": 1.3226172924041748, "learning_rate": 1.0401072893125863e-08, "loss": 0.7663, "step": 7479 }, { "epoch": 0.98, "grad_norm": 1.283625602722168, "learning_rate": 1.0264713547388227e-08, "loss": 0.7934, "step": 7480 }, { "epoch": 0.98, "grad_norm": 1.2599881887435913, "learning_rate": 1.0129253034805165e-08, "loss": 0.8067, "step": 7481 }, { "epoch": 0.98, "grad_norm": 1.3861299753189087, "learning_rate": 9.99469137977771e-09, "loss": 0.8454, "step": 7482 }, { "epoch": 0.98, "grad_norm": 1.312813639640808, "learning_rate": 9.86102860654592e-09, "loss": 0.8213, "step": 7483 }, { "epoch": 0.98, "grad_norm": 1.2739429473876953, "learning_rate": 9.728264739187198e-09, "loss": 0.7971, "step": 7484 }, { "epoch": 0.98, "grad_norm": 1.790583848953247, "learning_rate": 9.596399801616862e-09, "loss": 0.825, "step": 7485 }, { "epoch": 0.98, "grad_norm": 1.2985254526138306, "learning_rate": 9.465433817589243e-09, "loss": 0.8221, "step": 7486 }, { "epoch": 0.98, "grad_norm": 0.5570783019065857, "learning_rate": 9.335366810695467e-09, "loss": 1.0107, "step": 7487 }, { "epoch": 0.98, "grad_norm": 1.3118103742599487, "learning_rate": 9.206198804365684e-09, "loss": 0.7713, "step": 7488 }, { "epoch": 0.98, "grad_norm": 1.4294533729553223, "learning_rate": 9.077929821867392e-09, "loss": 0.8307, "step": 7489 }, { "epoch": 0.98, "grad_norm": 1.23483145236969, "learning_rate": 8.950559886306553e-09, "loss": 0.786, "step": 7490 }, { "epoch": 0.98, "grad_norm": 1.428750991821289, "learning_rate": 8.82408902062759e-09, "loss": 0.8329, "step": 7491 }, { "epoch": 0.98, "grad_norm": 1.3393510580062866, "learning_rate": 8.698517247612282e-09, "loss": 0.7229, "step": 7492 }, { "epoch": 0.98, "grad_norm": 1.358005166053772, "learning_rate": 8.573844589880308e-09, "loss": 0.7828, "step": 7493 }, { "epoch": 0.98, "grad_norm": 1.3042155504226685, "learning_rate": 8.450071069890375e-09, "loss": 0.7677, "step": 7494 }, { "epoch": 0.98, "grad_norm": 1.2462702989578247, "learning_rate": 8.327196709937978e-09, "loss": 0.8234, "step": 7495 }, { "epoch": 0.98, "grad_norm": 1.7257542610168457, "learning_rate": 8.205221532158192e-09, "loss": 0.773, "step": 7496 }, { "epoch": 0.98, "grad_norm": 0.5731106996536255, "learning_rate": 8.084145558522327e-09, "loss": 0.9988, "step": 7497 }, { "epoch": 0.98, "grad_norm": 1.294878602027893, "learning_rate": 7.963968810840162e-09, "loss": 0.7633, "step": 7498 }, { "epoch": 0.98, "grad_norm": 1.3346526622772217, "learning_rate": 7.8446913107616e-09, "loss": 0.7565, "step": 7499 }, { "epoch": 0.98, "grad_norm": 1.3438961505889893, "learning_rate": 7.726313079771119e-09, "loss": 0.8299, "step": 7500 }, { "epoch": 0.98, "grad_norm": 0.5609399080276489, "learning_rate": 7.608834139194443e-09, "loss": 1.0033, "step": 7501 }, { "epoch": 0.98, "grad_norm": 1.5286401510238647, "learning_rate": 7.49225451019242e-09, "loss": 0.7893, "step": 7502 }, { "epoch": 0.98, "grad_norm": 1.2675989866256714, "learning_rate": 7.37657421376603e-09, "loss": 0.8047, "step": 7503 }, { "epoch": 0.98, "grad_norm": 1.4020730257034302, "learning_rate": 7.261793270753603e-09, "loss": 0.784, "step": 7504 }, { "epoch": 0.98, "grad_norm": 1.3191328048706055, "learning_rate": 7.147911701830823e-09, "loss": 0.7827, "step": 7505 }, { "epoch": 0.98, "grad_norm": 1.3452471494674683, "learning_rate": 7.034929527512946e-09, "loss": 0.8238, "step": 7506 }, { "epoch": 0.98, "grad_norm": 1.8417794704437256, "learning_rate": 6.9228467681509145e-09, "loss": 0.7825, "step": 7507 }, { "epoch": 0.98, "grad_norm": 0.5776745080947876, "learning_rate": 6.811663443935801e-09, "loss": 1.0022, "step": 7508 }, { "epoch": 0.98, "grad_norm": 1.3354034423828125, "learning_rate": 6.701379574894917e-09, "loss": 0.8312, "step": 7509 }, { "epoch": 0.98, "grad_norm": 0.5744615793228149, "learning_rate": 6.591995180895705e-09, "loss": 1.0243, "step": 7510 }, { "epoch": 0.98, "grad_norm": 1.378778338432312, "learning_rate": 6.483510281641292e-09, "loss": 0.8294, "step": 7511 }, { "epoch": 0.98, "grad_norm": 0.5605784058570862, "learning_rate": 6.375924896673824e-09, "loss": 0.9988, "step": 7512 }, { "epoch": 0.98, "grad_norm": 1.428117036819458, "learning_rate": 6.2692390453733545e-09, "loss": 0.8135, "step": 7513 }, { "epoch": 0.98, "grad_norm": 1.346508502960205, "learning_rate": 6.163452746958398e-09, "loss": 0.7313, "step": 7514 }, { "epoch": 0.98, "grad_norm": 0.5696964859962463, "learning_rate": 6.058566020484824e-09, "loss": 1.0234, "step": 7515 }, { "epoch": 0.98, "grad_norm": 0.567163348197937, "learning_rate": 5.954578884845852e-09, "loss": 1.0043, "step": 7516 }, { "epoch": 0.98, "grad_norm": 1.5041577816009521, "learning_rate": 5.8514913587737195e-09, "loss": 0.7929, "step": 7517 }, { "epoch": 0.99, "grad_norm": 0.5619074106216431, "learning_rate": 5.749303460838573e-09, "loss": 1.0103, "step": 7518 }, { "epoch": 0.99, "grad_norm": 1.3918572664260864, "learning_rate": 5.648015209448465e-09, "loss": 0.821, "step": 7519 }, { "epoch": 0.99, "grad_norm": 0.5567570924758911, "learning_rate": 5.547626622848246e-09, "loss": 1.0074, "step": 7520 }, { "epoch": 0.99, "grad_norm": 0.5613381266593933, "learning_rate": 5.448137719121782e-09, "loss": 1.0117, "step": 7521 }, { "epoch": 0.99, "grad_norm": 0.5659520626068115, "learning_rate": 5.34954851619085e-09, "loss": 1.0112, "step": 7522 }, { "epoch": 0.99, "grad_norm": 1.300559401512146, "learning_rate": 5.251859031814577e-09, "loss": 0.7214, "step": 7523 }, { "epoch": 0.99, "grad_norm": 0.5681010484695435, "learning_rate": 5.155069283591108e-09, "loss": 1.005, "step": 7524 }, { "epoch": 0.99, "grad_norm": 0.5715436339378357, "learning_rate": 5.0591792889553845e-09, "loss": 1.0135, "step": 7525 }, { "epoch": 0.99, "grad_norm": 0.5813396573066711, "learning_rate": 4.964189065180258e-09, "loss": 1.0202, "step": 7526 }, { "epoch": 0.99, "grad_norm": 0.5832331776618958, "learning_rate": 4.870098629377596e-09, "loss": 1.0201, "step": 7527 }, { "epoch": 0.99, "grad_norm": 1.5577771663665771, "learning_rate": 4.776907998496616e-09, "loss": 0.8049, "step": 7528 }, { "epoch": 0.99, "grad_norm": 0.5759887099266052, "learning_rate": 4.684617189323892e-09, "loss": 1.0178, "step": 7529 }, { "epoch": 0.99, "grad_norm": 1.4789025783538818, "learning_rate": 4.593226218483904e-09, "loss": 0.7924, "step": 7530 }, { "epoch": 0.99, "grad_norm": 1.2636927366256714, "learning_rate": 4.502735102440703e-09, "loss": 0.7783, "step": 7531 }, { "epoch": 0.99, "grad_norm": 0.5546414852142334, "learning_rate": 4.413143857494584e-09, "loss": 1.0171, "step": 7532 }, { "epoch": 0.99, "grad_norm": 1.2189403772354126, "learning_rate": 4.324452499783749e-09, "loss": 0.785, "step": 7533 }, { "epoch": 0.99, "grad_norm": 0.5886033177375793, "learning_rate": 4.2366610452848625e-09, "loss": 1.0139, "step": 7534 }, { "epoch": 0.99, "grad_norm": 1.349228024482727, "learning_rate": 4.1497695098124954e-09, "loss": 0.7982, "step": 7535 }, { "epoch": 0.99, "grad_norm": 0.5772801637649536, "learning_rate": 4.063777909019129e-09, "loss": 0.9825, "step": 7536 }, { "epoch": 0.99, "grad_norm": 1.4024220705032349, "learning_rate": 3.978686258395148e-09, "loss": 0.804, "step": 7537 }, { "epoch": 0.99, "grad_norm": 1.3287653923034668, "learning_rate": 3.8944945732682925e-09, "loss": 0.7607, "step": 7538 }, { "epoch": 0.99, "grad_norm": 1.3519747257232666, "learning_rate": 3.811202868804209e-09, "loss": 0.7715, "step": 7539 }, { "epoch": 0.99, "grad_norm": 1.604196548461914, "learning_rate": 3.728811160007562e-09, "loss": 0.7907, "step": 7540 }, { "epoch": 0.99, "grad_norm": 1.3407996892929077, "learning_rate": 3.6473194617198117e-09, "loss": 0.8144, "step": 7541 }, { "epoch": 0.99, "grad_norm": 0.573903501033783, "learning_rate": 3.5667277886203277e-09, "loss": 1.0114, "step": 7542 }, { "epoch": 0.99, "grad_norm": 1.2898436784744263, "learning_rate": 3.4870361552269415e-09, "loss": 0.7447, "step": 7543 }, { "epoch": 0.99, "grad_norm": 1.3324494361877441, "learning_rate": 3.408244575894837e-09, "loss": 0.7524, "step": 7544 }, { "epoch": 0.99, "grad_norm": 1.4304150342941284, "learning_rate": 3.3303530648171047e-09, "loss": 0.7861, "step": 7545 }, { "epoch": 0.99, "grad_norm": 1.2686588764190674, "learning_rate": 3.253361636024743e-09, "loss": 0.7816, "step": 7546 }, { "epoch": 0.99, "grad_norm": 1.7557690143585205, "learning_rate": 3.1772703033877693e-09, "loss": 0.8289, "step": 7547 }, { "epoch": 0.99, "grad_norm": 0.5758945345878601, "learning_rate": 3.1020790806118863e-09, "loss": 0.9877, "step": 7548 }, { "epoch": 0.99, "grad_norm": 0.5616769194602966, "learning_rate": 3.0277879812418145e-09, "loss": 1.0017, "step": 7549 }, { "epoch": 0.99, "grad_norm": 1.2851473093032837, "learning_rate": 2.9543970186607375e-09, "loss": 0.8514, "step": 7550 }, { "epoch": 0.99, "grad_norm": 1.3969848155975342, "learning_rate": 2.881906206088081e-09, "loss": 0.7475, "step": 7551 }, { "epoch": 0.99, "grad_norm": 1.3114389181137085, "learning_rate": 2.8103155565833982e-09, "loss": 0.8062, "step": 7552 }, { "epoch": 0.99, "grad_norm": 1.5429713726043701, "learning_rate": 2.7396250830419302e-09, "loss": 0.7476, "step": 7553 }, { "epoch": 0.99, "grad_norm": 1.315517544746399, "learning_rate": 2.6698347981973794e-09, "loss": 0.7918, "step": 7554 }, { "epoch": 0.99, "grad_norm": 1.4056555032730103, "learning_rate": 2.6009447146224664e-09, "loss": 0.7846, "step": 7555 }, { "epoch": 0.99, "grad_norm": 1.3257495164871216, "learning_rate": 2.532954844725599e-09, "loss": 0.788, "step": 7556 }, { "epoch": 0.99, "grad_norm": 1.2826660871505737, "learning_rate": 2.465865200755313e-09, "loss": 0.7076, "step": 7557 }, { "epoch": 0.99, "grad_norm": 0.577693521976471, "learning_rate": 2.3996757947963855e-09, "loss": 1.0087, "step": 7558 }, { "epoch": 0.99, "grad_norm": 0.5769581198692322, "learning_rate": 2.334386638771502e-09, "loss": 1.0032, "step": 7559 }, { "epoch": 0.99, "grad_norm": 1.4333399534225464, "learning_rate": 2.26999774444292e-09, "loss": 0.7638, "step": 7560 }, { "epoch": 0.99, "grad_norm": 0.5683351159095764, "learning_rate": 2.2065091234080295e-09, "loss": 0.9937, "step": 7561 }, { "epoch": 0.99, "grad_norm": 0.5784765481948853, "learning_rate": 2.1439207871043477e-09, "loss": 1.0238, "step": 7562 }, { "epoch": 0.99, "grad_norm": 1.2391443252563477, "learning_rate": 2.0822327468056347e-09, "loss": 0.7648, "step": 7563 }, { "epoch": 0.99, "grad_norm": 1.3604588508605957, "learning_rate": 2.0214450136246676e-09, "loss": 0.7622, "step": 7564 }, { "epoch": 0.99, "grad_norm": 1.3658896684646606, "learning_rate": 1.9615575985115766e-09, "loss": 0.8205, "step": 7565 }, { "epoch": 0.99, "grad_norm": 1.3353943824768066, "learning_rate": 1.902570512253843e-09, "loss": 0.8639, "step": 7566 }, { "epoch": 0.99, "grad_norm": 1.3285435438156128, "learning_rate": 1.8444837654774118e-09, "loss": 0.7872, "step": 7567 }, { "epoch": 0.99, "grad_norm": 0.5822870135307312, "learning_rate": 1.7872973686461348e-09, "loss": 1.0046, "step": 7568 }, { "epoch": 0.99, "grad_norm": 0.5532281398773193, "learning_rate": 1.7310113320606613e-09, "loss": 1.0036, "step": 7569 }, { "epoch": 0.99, "grad_norm": 0.5676046013832092, "learning_rate": 1.6756256658612135e-09, "loss": 1.0161, "step": 7570 }, { "epoch": 0.99, "grad_norm": 0.5847841501235962, "learning_rate": 1.6211403800237003e-09, "loss": 1.0207, "step": 7571 }, { "epoch": 0.99, "grad_norm": 1.2838746309280396, "learning_rate": 1.5675554843630481e-09, "loss": 0.7421, "step": 7572 }, { "epoch": 0.99, "grad_norm": 1.375592827796936, "learning_rate": 1.5148709885326463e-09, "loss": 0.7174, "step": 7573 }, { "epoch": 0.99, "grad_norm": 1.499038577079773, "learning_rate": 1.463086902021571e-09, "loss": 0.8365, "step": 7574 }, { "epoch": 0.99, "grad_norm": 1.2521615028381348, "learning_rate": 1.4122032341595816e-09, "loss": 0.8131, "step": 7575 }, { "epoch": 0.99, "grad_norm": 0.5546854734420776, "learning_rate": 1.3622199941115689e-09, "loss": 0.9995, "step": 7576 }, { "epoch": 0.99, "grad_norm": 0.561747670173645, "learning_rate": 1.3131371908819968e-09, "loss": 1.0068, "step": 7577 }, { "epoch": 0.99, "grad_norm": 0.5543844699859619, "learning_rate": 1.2649548333115713e-09, "loss": 1.0175, "step": 7578 }, { "epoch": 0.99, "grad_norm": 1.3222146034240723, "learning_rate": 1.2176729300811264e-09, "loss": 0.7622, "step": 7579 }, { "epoch": 0.99, "grad_norm": 1.2586721181869507, "learning_rate": 1.1712914897060723e-09, "loss": 0.7981, "step": 7580 }, { "epoch": 0.99, "grad_norm": 0.5752388834953308, "learning_rate": 1.125810520543058e-09, "loss": 1.0105, "step": 7581 }, { "epoch": 0.99, "grad_norm": 1.3019036054611206, "learning_rate": 1.0812300307838642e-09, "loss": 0.7488, "step": 7582 }, { "epoch": 0.99, "grad_norm": 3.47212553024292, "learning_rate": 1.0375500284592888e-09, "loss": 0.8155, "step": 7583 }, { "epoch": 0.99, "grad_norm": 1.3388773202896118, "learning_rate": 9.947705214380377e-10, "loss": 0.8144, "step": 7584 }, { "epoch": 0.99, "grad_norm": 1.2941739559173584, "learning_rate": 9.528915174256137e-10, "loss": 0.7918, "step": 7585 }, { "epoch": 0.99, "grad_norm": 1.2494287490844727, "learning_rate": 9.119130239665374e-10, "loss": 0.7879, "step": 7586 }, { "epoch": 0.99, "grad_norm": 2.6123921871185303, "learning_rate": 8.718350484421267e-10, "loss": 0.8043, "step": 7587 }, { "epoch": 0.99, "grad_norm": 1.3923317193984985, "learning_rate": 8.326575980716067e-10, "loss": 0.8049, "step": 7588 }, { "epoch": 0.99, "grad_norm": 0.5803012251853943, "learning_rate": 7.943806799132203e-10, "loss": 1.0166, "step": 7589 }, { "epoch": 0.99, "grad_norm": 1.3406081199645996, "learning_rate": 7.570043008614525e-10, "loss": 0.7698, "step": 7590 }, { "epoch": 0.99, "grad_norm": 1.3016037940979004, "learning_rate": 7.205284676492508e-10, "loss": 0.8038, "step": 7591 }, { "epoch": 0.99, "grad_norm": 1.3599004745483398, "learning_rate": 6.849531868474702e-10, "loss": 0.8227, "step": 7592 }, { "epoch": 0.99, "grad_norm": 1.2353869676589966, "learning_rate": 6.502784648637628e-10, "loss": 0.7402, "step": 7593 }, { "epoch": 1.0, "grad_norm": 1.3789799213409424, "learning_rate": 6.165043079447986e-10, "loss": 0.7799, "step": 7594 }, { "epoch": 1.0, "grad_norm": 1.4002841711044312, "learning_rate": 5.836307221751547e-10, "loss": 0.8422, "step": 7595 }, { "epoch": 1.0, "grad_norm": 0.5974023342132568, "learning_rate": 5.516577134750955e-10, "loss": 0.9919, "step": 7596 }, { "epoch": 1.0, "grad_norm": 0.5520154237747192, "learning_rate": 5.205852876055684e-10, "loss": 0.9889, "step": 7597 }, { "epoch": 1.0, "grad_norm": 1.4545469284057617, "learning_rate": 4.904134501632074e-10, "loss": 0.8213, "step": 7598 }, { "epoch": 1.0, "grad_norm": 1.4077341556549072, "learning_rate": 4.6114220658310946e-10, "loss": 0.7258, "step": 7599 }, { "epoch": 1.0, "grad_norm": 1.4131639003753662, "learning_rate": 4.3277156213772377e-10, "loss": 0.711, "step": 7600 }, { "epoch": 1.0, "grad_norm": 1.3725367784500122, "learning_rate": 4.053015219379619e-10, "loss": 0.776, "step": 7601 }, { "epoch": 1.0, "grad_norm": 1.3032304048538208, "learning_rate": 3.787320909326431e-10, "loss": 0.7621, "step": 7602 }, { "epoch": 1.0, "grad_norm": 1.9573750495910645, "learning_rate": 3.530632739073836e-10, "loss": 0.8006, "step": 7603 }, { "epoch": 1.0, "grad_norm": 0.555254340171814, "learning_rate": 3.282950754862624e-10, "loss": 1.0021, "step": 7604 }, { "epoch": 1.0, "grad_norm": 0.5682860612869263, "learning_rate": 3.0442750013071065e-10, "loss": 0.9951, "step": 7605 }, { "epoch": 1.0, "grad_norm": 1.7569857835769653, "learning_rate": 2.8146055214006705e-10, "loss": 0.8537, "step": 7606 }, { "epoch": 1.0, "grad_norm": 1.5172277688980103, "learning_rate": 2.5939423565157773e-10, "loss": 0.8166, "step": 7607 }, { "epoch": 1.0, "grad_norm": 1.4937715530395508, "learning_rate": 2.382285546409513e-10, "loss": 0.7437, "step": 7608 }, { "epoch": 1.0, "grad_norm": 1.3650805950164795, "learning_rate": 2.179635129195834e-10, "loss": 0.7883, "step": 7609 }, { "epoch": 1.0, "grad_norm": 1.755894422531128, "learning_rate": 1.9859911413899758e-10, "loss": 0.8123, "step": 7610 }, { "epoch": 1.0, "grad_norm": 1.354868769645691, "learning_rate": 1.8013536178695946e-10, "loss": 0.8184, "step": 7611 }, { "epoch": 1.0, "grad_norm": 1.7309300899505615, "learning_rate": 1.625722591896972e-10, "loss": 0.7793, "step": 7612 }, { "epoch": 1.0, "grad_norm": 1.5323681831359863, "learning_rate": 1.4590980951079137e-10, "loss": 0.7408, "step": 7613 }, { "epoch": 1.0, "grad_norm": 1.354314923286438, "learning_rate": 1.3014801575228497e-10, "loss": 0.8007, "step": 7614 }, { "epoch": 1.0, "grad_norm": 1.476895809173584, "learning_rate": 1.1528688075246319e-10, "loss": 0.7795, "step": 7615 }, { "epoch": 1.0, "grad_norm": 1.41563880443573, "learning_rate": 1.0132640718862885e-10, "loss": 0.7865, "step": 7616 }, { "epoch": 1.0, "grad_norm": 1.3769913911819458, "learning_rate": 8.82665975759922e-11, "loss": 0.7871, "step": 7617 }, { "epoch": 1.0, "grad_norm": 0.562637984752655, "learning_rate": 7.610745426711585e-11, "loss": 1.0029, "step": 7618 }, { "epoch": 1.0, "grad_norm": 0.5681291818618774, "learning_rate": 6.484897945191471e-11, "loss": 0.9779, "step": 7619 }, { "epoch": 1.0, "grad_norm": 1.3247727155685425, "learning_rate": 5.449117515876623e-11, "loss": 0.787, "step": 7620 }, { "epoch": 1.0, "grad_norm": 1.723408579826355, "learning_rate": 4.5034043253400214e-11, "loss": 0.8124, "step": 7621 }, { "epoch": 1.0, "grad_norm": 1.345506191253662, "learning_rate": 3.64775854394539e-11, "loss": 0.8084, "step": 7622 }, { "epoch": 1.0, "grad_norm": 0.56166672706604, "learning_rate": 2.8821803257916835e-11, "loss": 0.9957, "step": 7623 }, { "epoch": 1.0, "grad_norm": 1.2432947158813477, "learning_rate": 2.2066698087686022e-11, "loss": 0.7994, "step": 7624 }, { "epoch": 1.0, "grad_norm": 1.2729653120040894, "learning_rate": 1.6212271146676116e-11, "loss": 0.7809, "step": 7625 }, { "epoch": 1.0, "grad_norm": 0.5768424272537231, "learning_rate": 1.1258523488488771e-11, "loss": 1.006, "step": 7626 }, { "epoch": 1.0, "grad_norm": 1.2291783094406128, "learning_rate": 7.205456006298406e-12, "loss": 0.7834, "step": 7627 }, { "epoch": 1.0, "grad_norm": 0.5812944769859314, "learning_rate": 4.05306942952155e-12, "loss": 0.9799, "step": 7628 }, { "epoch": 1.0, "grad_norm": 1.539451003074646, "learning_rate": 1.8013643260372804e-12, "loss": 0.8442, "step": 7629 }, { "epoch": 1.0, "grad_norm": 0.5540574193000793, "learning_rate": 4.5034110163211243e-13, "loss": 0.9891, "step": 7630 }, { "epoch": 1.0, "grad_norm": 1.383908748626709, "learning_rate": 0.0, "loss": 0.6961, "step": 7631 }, { "epoch": 1.0, "step": 7631, "total_flos": 8.5089712268095e+18, "train_loss": 0.22531458207061572, "train_runtime": 12969.348, "train_samples_per_second": 301.284, "train_steps_per_second": 0.588 } ], "logging_steps": 1.0, "max_steps": 7631, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "total_flos": 8.5089712268095e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }