{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.7536168347216132, "eval_steps": 500, "global_step": 3000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001169077889814409, "grad_norm": 6.78125, "learning_rate": 3.898635477582846e-07, "loss": 5.0851, "step": 2 }, { "epoch": 0.002338155779628818, "grad_norm": 7.0625, "learning_rate": 7.797270955165692e-07, "loss": 5.1921, "step": 4 }, { "epoch": 0.0035072336694432268, "grad_norm": 7.34375, "learning_rate": 1.1695906432748538e-06, "loss": 5.0756, "step": 6 }, { "epoch": 0.004676311559257636, "grad_norm": 7.03125, "learning_rate": 1.5594541910331385e-06, "loss": 5.1097, "step": 8 }, { "epoch": 0.005845389449072045, "grad_norm": 6.40625, "learning_rate": 1.949317738791423e-06, "loss": 5.0591, "step": 10 }, { "epoch": 0.0070144673388864535, "grad_norm": 5.8125, "learning_rate": 2.3391812865497075e-06, "loss": 5.096, "step": 12 }, { "epoch": 0.008183545228700862, "grad_norm": 5.1875, "learning_rate": 2.729044834307992e-06, "loss": 5.0234, "step": 14 }, { "epoch": 0.009352623118515271, "grad_norm": 5.03125, "learning_rate": 3.118908382066277e-06, "loss": 5.0266, "step": 16 }, { "epoch": 0.01052170100832968, "grad_norm": 4.90625, "learning_rate": 3.5087719298245615e-06, "loss": 5.031, "step": 18 }, { "epoch": 0.01169077889814409, "grad_norm": 4.90625, "learning_rate": 3.898635477582846e-06, "loss": 4.9845, "step": 20 }, { "epoch": 0.012859856787958498, "grad_norm": 4.6875, "learning_rate": 4.2884990253411305e-06, "loss": 4.9933, "step": 22 }, { "epoch": 0.014028934677772907, "grad_norm": 4.625, "learning_rate": 4.678362573099415e-06, "loss": 5.0822, "step": 24 }, { "epoch": 0.015198012567587316, "grad_norm": 4.71875, "learning_rate": 5.0682261208576995e-06, "loss": 4.9392, "step": 26 }, { "epoch": 0.016367090457401725, "grad_norm": 4.65625, "learning_rate": 5.458089668615984e-06, "loss": 5.0441, "step": 28 }, { "epoch": 0.017536168347216132, "grad_norm": 4.6875, "learning_rate": 5.8479532163742686e-06, "loss": 4.8865, "step": 30 }, { "epoch": 0.018705246237030543, "grad_norm": 4.59375, "learning_rate": 6.237816764132554e-06, "loss": 4.9227, "step": 32 }, { "epoch": 0.01987432412684495, "grad_norm": 4.65625, "learning_rate": 6.6276803118908384e-06, "loss": 4.9035, "step": 34 }, { "epoch": 0.02104340201665936, "grad_norm": 4.59375, "learning_rate": 7.017543859649123e-06, "loss": 4.863, "step": 36 }, { "epoch": 0.022212479906473768, "grad_norm": 4.71875, "learning_rate": 7.4074074074074075e-06, "loss": 4.7735, "step": 38 }, { "epoch": 0.02338155779628818, "grad_norm": 4.84375, "learning_rate": 7.797270955165692e-06, "loss": 4.7932, "step": 40 }, { "epoch": 0.024550635686102586, "grad_norm": 5.3125, "learning_rate": 8.187134502923977e-06, "loss": 4.6933, "step": 42 }, { "epoch": 0.025719713575916996, "grad_norm": 7.15625, "learning_rate": 8.576998050682261e-06, "loss": 4.5906, "step": 44 }, { "epoch": 0.026888791465731404, "grad_norm": 8.8125, "learning_rate": 8.966861598440546e-06, "loss": 4.4291, "step": 46 }, { "epoch": 0.028057869355545814, "grad_norm": 9.6875, "learning_rate": 9.35672514619883e-06, "loss": 4.221, "step": 48 }, { "epoch": 0.02922694724536022, "grad_norm": 8.8125, "learning_rate": 9.746588693957115e-06, "loss": 4.032, "step": 50 }, { "epoch": 0.030396025135174632, "grad_norm": 9.0, "learning_rate": 1.0136452241715399e-05, "loss": 3.8603, "step": 52 }, { "epoch": 0.03156510302498904, "grad_norm": 7.46875, "learning_rate": 1.0526315789473684e-05, "loss": 3.6522, "step": 54 }, { "epoch": 0.03273418091480345, "grad_norm": 8.375, "learning_rate": 1.0916179337231968e-05, "loss": 3.4394, "step": 56 }, { "epoch": 0.03390325880461786, "grad_norm": 6.4375, "learning_rate": 1.1306042884990253e-05, "loss": 3.2265, "step": 58 }, { "epoch": 0.035072336694432264, "grad_norm": 5.75, "learning_rate": 1.1695906432748537e-05, "loss": 3.091, "step": 60 }, { "epoch": 0.03624141458424668, "grad_norm": 4.8125, "learning_rate": 1.2085769980506823e-05, "loss": 2.8856, "step": 62 }, { "epoch": 0.037410492474061086, "grad_norm": 3.8125, "learning_rate": 1.2475633528265108e-05, "loss": 2.6299, "step": 64 }, { "epoch": 0.03857957036387549, "grad_norm": 3.234375, "learning_rate": 1.2865497076023392e-05, "loss": 2.5917, "step": 66 }, { "epoch": 0.0397486482536899, "grad_norm": 4.0625, "learning_rate": 1.3255360623781677e-05, "loss": 2.3757, "step": 68 }, { "epoch": 0.040917726143504314, "grad_norm": 2.453125, "learning_rate": 1.364522417153996e-05, "loss": 2.308, "step": 70 }, { "epoch": 0.04208680403331872, "grad_norm": 2.390625, "learning_rate": 1.4035087719298246e-05, "loss": 2.3042, "step": 72 }, { "epoch": 0.04325588192313313, "grad_norm": 2.109375, "learning_rate": 1.442495126705653e-05, "loss": 2.2149, "step": 74 }, { "epoch": 0.044424959812947536, "grad_norm": 2.71875, "learning_rate": 1.4814814814814815e-05, "loss": 2.0738, "step": 76 }, { "epoch": 0.04559403770276195, "grad_norm": 2.46875, "learning_rate": 1.5204678362573099e-05, "loss": 1.9663, "step": 78 }, { "epoch": 0.04676311559257636, "grad_norm": 2.640625, "learning_rate": 1.5594541910331384e-05, "loss": 1.9384, "step": 80 }, { "epoch": 0.047932193482390764, "grad_norm": 2.0, "learning_rate": 1.5984405458089668e-05, "loss": 1.894, "step": 82 }, { "epoch": 0.04910127137220517, "grad_norm": 2.078125, "learning_rate": 1.6374269005847955e-05, "loss": 1.8204, "step": 84 }, { "epoch": 0.050270349262019585, "grad_norm": 2.953125, "learning_rate": 1.676413255360624e-05, "loss": 1.8051, "step": 86 }, { "epoch": 0.05143942715183399, "grad_norm": 2.046875, "learning_rate": 1.7153996101364522e-05, "loss": 1.5832, "step": 88 }, { "epoch": 0.0526085050416484, "grad_norm": 1.75, "learning_rate": 1.7543859649122806e-05, "loss": 1.5298, "step": 90 }, { "epoch": 0.05377758293146281, "grad_norm": 1.71875, "learning_rate": 1.7933723196881093e-05, "loss": 1.4964, "step": 92 }, { "epoch": 0.054946660821277214, "grad_norm": 1.3515625, "learning_rate": 1.8323586744639376e-05, "loss": 1.3987, "step": 94 }, { "epoch": 0.05611573871109163, "grad_norm": 1.9140625, "learning_rate": 1.871345029239766e-05, "loss": 1.3959, "step": 96 }, { "epoch": 0.057284816600906036, "grad_norm": 1.34375, "learning_rate": 1.9103313840155944e-05, "loss": 1.3673, "step": 98 }, { "epoch": 0.05845389449072044, "grad_norm": 1.09375, "learning_rate": 1.949317738791423e-05, "loss": 1.2864, "step": 100 }, { "epoch": 0.05962297238053485, "grad_norm": 1.1328125, "learning_rate": 1.9883040935672515e-05, "loss": 1.2535, "step": 102 }, { "epoch": 0.060792050270349264, "grad_norm": 1.2109375, "learning_rate": 2.0272904483430798e-05, "loss": 1.2242, "step": 104 }, { "epoch": 0.06196112816016367, "grad_norm": 2.59375, "learning_rate": 2.0662768031189085e-05, "loss": 1.1936, "step": 106 }, { "epoch": 0.06313020604997809, "grad_norm": 1.953125, "learning_rate": 2.105263157894737e-05, "loss": 1.1858, "step": 108 }, { "epoch": 0.06429928393979249, "grad_norm": 2.140625, "learning_rate": 2.1442495126705653e-05, "loss": 1.1438, "step": 110 }, { "epoch": 0.0654683618296069, "grad_norm": 1.09375, "learning_rate": 2.1832358674463936e-05, "loss": 1.1217, "step": 112 }, { "epoch": 0.06663743971942131, "grad_norm": 1.8671875, "learning_rate": 2.2222222222222223e-05, "loss": 1.1508, "step": 114 }, { "epoch": 0.06780651760923571, "grad_norm": 1.046875, "learning_rate": 2.2612085769980507e-05, "loss": 1.0976, "step": 116 }, { "epoch": 0.06897559549905012, "grad_norm": 1.234375, "learning_rate": 2.300194931773879e-05, "loss": 1.0716, "step": 118 }, { "epoch": 0.07014467338886453, "grad_norm": 1.6171875, "learning_rate": 2.3391812865497074e-05, "loss": 1.0152, "step": 120 }, { "epoch": 0.07131375127867894, "grad_norm": 3.125, "learning_rate": 2.378167641325536e-05, "loss": 1.0265, "step": 122 }, { "epoch": 0.07248282916849336, "grad_norm": 3.546875, "learning_rate": 2.4171539961013645e-05, "loss": 1.2355, "step": 124 }, { "epoch": 0.07365190705830776, "grad_norm": 1.1796875, "learning_rate": 2.456140350877193e-05, "loss": 1.0111, "step": 126 }, { "epoch": 0.07482098494812217, "grad_norm": 2.453125, "learning_rate": 2.4951267056530216e-05, "loss": 1.2317, "step": 128 }, { "epoch": 0.07599006283793658, "grad_norm": 1.09375, "learning_rate": 2.53411306042885e-05, "loss": 0.9988, "step": 130 }, { "epoch": 0.07715914072775099, "grad_norm": 1.09375, "learning_rate": 2.5730994152046783e-05, "loss": 1.0076, "step": 132 }, { "epoch": 0.07832821861756539, "grad_norm": 1.4375, "learning_rate": 2.6120857699805067e-05, "loss": 1.0155, "step": 134 }, { "epoch": 0.0794972965073798, "grad_norm": 1.796875, "learning_rate": 2.6510721247563354e-05, "loss": 1.1216, "step": 136 }, { "epoch": 0.08066637439719421, "grad_norm": 0.9921875, "learning_rate": 2.6900584795321637e-05, "loss": 0.9788, "step": 138 }, { "epoch": 0.08183545228700863, "grad_norm": 1.328125, "learning_rate": 2.729044834307992e-05, "loss": 0.9372, "step": 140 }, { "epoch": 0.08300453017682304, "grad_norm": 1.296875, "learning_rate": 2.7680311890838205e-05, "loss": 0.897, "step": 142 }, { "epoch": 0.08417360806663744, "grad_norm": 1.0703125, "learning_rate": 2.8070175438596492e-05, "loss": 0.9766, "step": 144 }, { "epoch": 0.08534268595645185, "grad_norm": 0.84765625, "learning_rate": 2.8460038986354776e-05, "loss": 0.9315, "step": 146 }, { "epoch": 0.08651176384626626, "grad_norm": 0.9140625, "learning_rate": 2.884990253411306e-05, "loss": 0.9425, "step": 148 }, { "epoch": 0.08768084173608066, "grad_norm": 0.99609375, "learning_rate": 2.9239766081871346e-05, "loss": 0.9242, "step": 150 }, { "epoch": 0.08884991962589507, "grad_norm": 1.8046875, "learning_rate": 2.962962962962963e-05, "loss": 0.9396, "step": 152 }, { "epoch": 0.09001899751570948, "grad_norm": 0.92578125, "learning_rate": 3.0019493177387914e-05, "loss": 0.8932, "step": 154 }, { "epoch": 0.0911880754055239, "grad_norm": 72.5, "learning_rate": 3.0409356725146197e-05, "loss": 0.9062, "step": 156 }, { "epoch": 0.0923571532953383, "grad_norm": 1.703125, "learning_rate": 3.0799220272904484e-05, "loss": 0.9534, "step": 158 }, { "epoch": 0.09352623118515271, "grad_norm": 1.453125, "learning_rate": 3.118908382066277e-05, "loss": 0.8715, "step": 160 }, { "epoch": 0.09469530907496712, "grad_norm": 1.015625, "learning_rate": 3.157894736842105e-05, "loss": 0.9029, "step": 162 }, { "epoch": 0.09586438696478153, "grad_norm": 1.15625, "learning_rate": 3.1968810916179335e-05, "loss": 0.8968, "step": 164 }, { "epoch": 0.09703346485459594, "grad_norm": 1.4296875, "learning_rate": 3.235867446393762e-05, "loss": 0.9784, "step": 166 }, { "epoch": 0.09820254274441034, "grad_norm": 1.0078125, "learning_rate": 3.274853801169591e-05, "loss": 0.9012, "step": 168 }, { "epoch": 0.09937162063422475, "grad_norm": 0.8359375, "learning_rate": 3.313840155945419e-05, "loss": 0.8895, "step": 170 }, { "epoch": 0.10054069852403917, "grad_norm": 0.8046875, "learning_rate": 3.352826510721248e-05, "loss": 0.8858, "step": 172 }, { "epoch": 0.10170977641385358, "grad_norm": 0.83203125, "learning_rate": 3.391812865497076e-05, "loss": 0.8408, "step": 174 }, { "epoch": 0.10287885430366799, "grad_norm": 0.8671875, "learning_rate": 3.4307992202729044e-05, "loss": 0.8896, "step": 176 }, { "epoch": 0.10404793219348239, "grad_norm": 1.78125, "learning_rate": 3.469785575048733e-05, "loss": 0.8941, "step": 178 }, { "epoch": 0.1052170100832968, "grad_norm": 0.8984375, "learning_rate": 3.508771929824561e-05, "loss": 0.9013, "step": 180 }, { "epoch": 0.1063860879731112, "grad_norm": 1.2734375, "learning_rate": 3.5477582846003895e-05, "loss": 0.8585, "step": 182 }, { "epoch": 0.10755516586292561, "grad_norm": 2.015625, "learning_rate": 3.5867446393762186e-05, "loss": 1.0231, "step": 184 }, { "epoch": 0.10872424375274002, "grad_norm": 1.7578125, "learning_rate": 3.625730994152047e-05, "loss": 0.8835, "step": 186 }, { "epoch": 0.10989332164255443, "grad_norm": 1.078125, "learning_rate": 3.664717348927875e-05, "loss": 0.8465, "step": 188 }, { "epoch": 0.11106239953236885, "grad_norm": 2.75, "learning_rate": 3.7037037037037037e-05, "loss": 1.1014, "step": 190 }, { "epoch": 0.11223147742218326, "grad_norm": 2.09375, "learning_rate": 3.742690058479532e-05, "loss": 0.8691, "step": 192 }, { "epoch": 0.11340055531199766, "grad_norm": 0.96875, "learning_rate": 3.7816764132553604e-05, "loss": 0.8547, "step": 194 }, { "epoch": 0.11456963320181207, "grad_norm": 1.5859375, "learning_rate": 3.820662768031189e-05, "loss": 0.8725, "step": 196 }, { "epoch": 0.11573871109162648, "grad_norm": 0.88671875, "learning_rate": 3.859649122807018e-05, "loss": 0.8393, "step": 198 }, { "epoch": 0.11690778898144089, "grad_norm": 1.015625, "learning_rate": 3.898635477582846e-05, "loss": 0.8345, "step": 200 }, { "epoch": 0.11807686687125529, "grad_norm": 0.82421875, "learning_rate": 3.9376218323586745e-05, "loss": 0.8966, "step": 202 }, { "epoch": 0.1192459447610697, "grad_norm": 0.79296875, "learning_rate": 3.976608187134503e-05, "loss": 0.8291, "step": 204 }, { "epoch": 0.12041502265088412, "grad_norm": 0.83203125, "learning_rate": 4.015594541910331e-05, "loss": 0.8175, "step": 206 }, { "epoch": 0.12158410054069853, "grad_norm": 0.74609375, "learning_rate": 4.0545808966861596e-05, "loss": 0.8353, "step": 208 }, { "epoch": 0.12275317843051294, "grad_norm": 1.796875, "learning_rate": 4.093567251461988e-05, "loss": 1.0084, "step": 210 }, { "epoch": 0.12392225632032734, "grad_norm": 59.75, "learning_rate": 4.132553606237817e-05, "loss": 0.8478, "step": 212 }, { "epoch": 0.12509133421014176, "grad_norm": 1.265625, "learning_rate": 4.1715399610136454e-05, "loss": 0.8267, "step": 214 }, { "epoch": 0.12626041209995617, "grad_norm": 1.0625, "learning_rate": 4.210526315789474e-05, "loss": 0.836, "step": 216 }, { "epoch": 0.12742948998977058, "grad_norm": 1.0078125, "learning_rate": 4.249512670565302e-05, "loss": 0.857, "step": 218 }, { "epoch": 0.12859856787958499, "grad_norm": 0.92578125, "learning_rate": 4.2884990253411305e-05, "loss": 0.8478, "step": 220 }, { "epoch": 0.1297676457693994, "grad_norm": 0.953125, "learning_rate": 4.327485380116959e-05, "loss": 0.844, "step": 222 }, { "epoch": 0.1309367236592138, "grad_norm": 0.796875, "learning_rate": 4.366471734892787e-05, "loss": 0.8128, "step": 224 }, { "epoch": 0.1321058015490282, "grad_norm": 1.28125, "learning_rate": 4.4054580896686156e-05, "loss": 0.8276, "step": 226 }, { "epoch": 0.13327487943884261, "grad_norm": 0.94921875, "learning_rate": 4.4444444444444447e-05, "loss": 0.8478, "step": 228 }, { "epoch": 0.13444395732865702, "grad_norm": 1.0703125, "learning_rate": 4.483430799220273e-05, "loss": 0.814, "step": 230 }, { "epoch": 0.13561303521847143, "grad_norm": 1.1328125, "learning_rate": 4.5224171539961014e-05, "loss": 0.765, "step": 232 }, { "epoch": 0.13678211310828584, "grad_norm": 0.87109375, "learning_rate": 4.56140350877193e-05, "loss": 0.8277, "step": 234 }, { "epoch": 0.13795119099810024, "grad_norm": 0.921875, "learning_rate": 4.600389863547758e-05, "loss": 0.8103, "step": 236 }, { "epoch": 0.13912026888791465, "grad_norm": 1.4921875, "learning_rate": 4.6393762183235865e-05, "loss": 0.7897, "step": 238 }, { "epoch": 0.14028934677772906, "grad_norm": 0.7734375, "learning_rate": 4.678362573099415e-05, "loss": 0.7798, "step": 240 }, { "epoch": 0.14145842466754346, "grad_norm": 0.98828125, "learning_rate": 4.717348927875244e-05, "loss": 0.8442, "step": 242 }, { "epoch": 0.14262750255735787, "grad_norm": 0.91796875, "learning_rate": 4.756335282651072e-05, "loss": 0.7936, "step": 244 }, { "epoch": 0.1437965804471723, "grad_norm": 0.859375, "learning_rate": 4.7953216374269006e-05, "loss": 0.7726, "step": 246 }, { "epoch": 0.1449656583369867, "grad_norm": 1.3671875, "learning_rate": 4.834307992202729e-05, "loss": 0.9626, "step": 248 }, { "epoch": 0.14613473622680112, "grad_norm": 0.921875, "learning_rate": 4.8732943469785574e-05, "loss": 0.7929, "step": 250 }, { "epoch": 0.14730381411661553, "grad_norm": 0.92578125, "learning_rate": 4.912280701754386e-05, "loss": 0.8589, "step": 252 }, { "epoch": 0.14847289200642994, "grad_norm": 0.83203125, "learning_rate": 4.951267056530214e-05, "loss": 0.7383, "step": 254 }, { "epoch": 0.14964196989624434, "grad_norm": 1.046875, "learning_rate": 4.990253411306043e-05, "loss": 0.8011, "step": 256 }, { "epoch": 0.15081104778605875, "grad_norm": 1.3359375, "learning_rate": 5.0292397660818715e-05, "loss": 0.8404, "step": 258 }, { "epoch": 0.15198012567587316, "grad_norm": 1.21875, "learning_rate": 5.0682261208577e-05, "loss": 0.8301, "step": 260 }, { "epoch": 0.15314920356568756, "grad_norm": 0.93359375, "learning_rate": 5.107212475633528e-05, "loss": 0.8157, "step": 262 }, { "epoch": 0.15431828145550197, "grad_norm": 1.3203125, "learning_rate": 5.1461988304093566e-05, "loss": 0.7724, "step": 264 }, { "epoch": 0.15548735934531638, "grad_norm": 1.09375, "learning_rate": 5.185185185185185e-05, "loss": 0.7586, "step": 266 }, { "epoch": 0.15665643723513079, "grad_norm": 1.1484375, "learning_rate": 5.2241715399610133e-05, "loss": 0.7812, "step": 268 }, { "epoch": 0.1578255151249452, "grad_norm": 0.74609375, "learning_rate": 5.2631578947368424e-05, "loss": 0.7486, "step": 270 }, { "epoch": 0.1589945930147596, "grad_norm": 0.81640625, "learning_rate": 5.302144249512671e-05, "loss": 0.7339, "step": 272 }, { "epoch": 0.160163670904574, "grad_norm": 1.046875, "learning_rate": 5.341130604288499e-05, "loss": 0.8381, "step": 274 }, { "epoch": 0.16133274879438841, "grad_norm": 0.88671875, "learning_rate": 5.3801169590643275e-05, "loss": 0.774, "step": 276 }, { "epoch": 0.16250182668420282, "grad_norm": 0.796875, "learning_rate": 5.419103313840156e-05, "loss": 0.8114, "step": 278 }, { "epoch": 0.16367090457401726, "grad_norm": 0.9140625, "learning_rate": 5.458089668615984e-05, "loss": 0.7549, "step": 280 }, { "epoch": 0.16483998246383166, "grad_norm": 1.0625, "learning_rate": 5.4970760233918126e-05, "loss": 0.7393, "step": 282 }, { "epoch": 0.16600906035364607, "grad_norm": 0.88671875, "learning_rate": 5.536062378167641e-05, "loss": 0.7607, "step": 284 }, { "epoch": 0.16717813824346048, "grad_norm": 1.046875, "learning_rate": 5.57504873294347e-05, "loss": 0.8158, "step": 286 }, { "epoch": 0.16834721613327489, "grad_norm": 1.734375, "learning_rate": 5.6140350877192984e-05, "loss": 0.7403, "step": 288 }, { "epoch": 0.1695162940230893, "grad_norm": 1.1015625, "learning_rate": 5.653021442495127e-05, "loss": 0.7278, "step": 290 }, { "epoch": 0.1706853719129037, "grad_norm": 1.7578125, "learning_rate": 5.692007797270955e-05, "loss": 0.821, "step": 292 }, { "epoch": 0.1718544498027181, "grad_norm": 3.1875, "learning_rate": 5.7309941520467835e-05, "loss": 0.8276, "step": 294 }, { "epoch": 0.1730235276925325, "grad_norm": 1.1484375, "learning_rate": 5.769980506822612e-05, "loss": 0.8505, "step": 296 }, { "epoch": 0.17419260558234692, "grad_norm": 1.2265625, "learning_rate": 5.80896686159844e-05, "loss": 0.7941, "step": 298 }, { "epoch": 0.17536168347216133, "grad_norm": 1.4765625, "learning_rate": 5.847953216374269e-05, "loss": 0.8621, "step": 300 }, { "epoch": 0.17653076136197574, "grad_norm": 2.25, "learning_rate": 5.8869395711500976e-05, "loss": 0.8071, "step": 302 }, { "epoch": 0.17769983925179014, "grad_norm": 0.92578125, "learning_rate": 5.925925925925926e-05, "loss": 0.803, "step": 304 }, { "epoch": 0.17886891714160455, "grad_norm": 0.9921875, "learning_rate": 5.9649122807017544e-05, "loss": 0.7569, "step": 306 }, { "epoch": 0.18003799503141896, "grad_norm": 1.09375, "learning_rate": 6.003898635477583e-05, "loss": 0.7722, "step": 308 }, { "epoch": 0.18120707292123336, "grad_norm": 0.984375, "learning_rate": 6.042884990253411e-05, "loss": 0.7746, "step": 310 }, { "epoch": 0.1823761508110478, "grad_norm": 2.296875, "learning_rate": 6.0818713450292395e-05, "loss": 0.7292, "step": 312 }, { "epoch": 0.1835452287008622, "grad_norm": 1.140625, "learning_rate": 6.120857699805068e-05, "loss": 0.7751, "step": 314 }, { "epoch": 0.1847143065906766, "grad_norm": 0.8203125, "learning_rate": 6.159844054580897e-05, "loss": 0.7473, "step": 316 }, { "epoch": 0.18588338448049102, "grad_norm": 1.0625, "learning_rate": 6.198830409356725e-05, "loss": 0.7718, "step": 318 }, { "epoch": 0.18705246237030543, "grad_norm": 1.0390625, "learning_rate": 6.237816764132554e-05, "loss": 0.7812, "step": 320 }, { "epoch": 0.18822154026011984, "grad_norm": 0.859375, "learning_rate": 6.276803118908382e-05, "loss": 0.7552, "step": 322 }, { "epoch": 0.18939061814993424, "grad_norm": 0.765625, "learning_rate": 6.31578947368421e-05, "loss": 0.7244, "step": 324 }, { "epoch": 0.19055969603974865, "grad_norm": 1.2734375, "learning_rate": 6.354775828460039e-05, "loss": 0.8182, "step": 326 }, { "epoch": 0.19172877392956306, "grad_norm": 0.93359375, "learning_rate": 6.393762183235867e-05, "loss": 0.7578, "step": 328 }, { "epoch": 0.19289785181937746, "grad_norm": 0.875, "learning_rate": 6.432748538011695e-05, "loss": 0.8755, "step": 330 }, { "epoch": 0.19406692970919187, "grad_norm": 1.4296875, "learning_rate": 6.471734892787524e-05, "loss": 0.7436, "step": 332 }, { "epoch": 0.19523600759900628, "grad_norm": 1.859375, "learning_rate": 6.510721247563352e-05, "loss": 0.7415, "step": 334 }, { "epoch": 0.19640508548882069, "grad_norm": 1.265625, "learning_rate": 6.549707602339182e-05, "loss": 0.8082, "step": 336 }, { "epoch": 0.1975741633786351, "grad_norm": 1.4453125, "learning_rate": 6.58869395711501e-05, "loss": 0.7226, "step": 338 }, { "epoch": 0.1987432412684495, "grad_norm": 0.8203125, "learning_rate": 6.627680311890839e-05, "loss": 0.7408, "step": 340 }, { "epoch": 0.1999123191582639, "grad_norm": 1.1875, "learning_rate": 6.666666666666667e-05, "loss": 0.7752, "step": 342 }, { "epoch": 0.20108139704807834, "grad_norm": 1.0234375, "learning_rate": 6.705653021442495e-05, "loss": 0.7313, "step": 344 }, { "epoch": 0.20225047493789275, "grad_norm": 1.1015625, "learning_rate": 6.744639376218324e-05, "loss": 0.7164, "step": 346 }, { "epoch": 0.20341955282770716, "grad_norm": 0.9296875, "learning_rate": 6.783625730994152e-05, "loss": 0.7836, "step": 348 }, { "epoch": 0.20458863071752156, "grad_norm": 1.9140625, "learning_rate": 6.82261208576998e-05, "loss": 0.8936, "step": 350 }, { "epoch": 0.20575770860733597, "grad_norm": 0.8984375, "learning_rate": 6.861598440545809e-05, "loss": 0.7386, "step": 352 }, { "epoch": 0.20692678649715038, "grad_norm": 0.9609375, "learning_rate": 6.900584795321637e-05, "loss": 0.8893, "step": 354 }, { "epoch": 0.20809586438696479, "grad_norm": 0.9140625, "learning_rate": 6.939571150097466e-05, "loss": 0.715, "step": 356 }, { "epoch": 0.2092649422767792, "grad_norm": 0.94921875, "learning_rate": 6.978557504873294e-05, "loss": 0.7087, "step": 358 }, { "epoch": 0.2104340201665936, "grad_norm": 0.96484375, "learning_rate": 7.017543859649122e-05, "loss": 0.7067, "step": 360 }, { "epoch": 0.211603098056408, "grad_norm": 0.9375, "learning_rate": 7.05653021442495e-05, "loss": 0.6979, "step": 362 }, { "epoch": 0.2127721759462224, "grad_norm": 1.8046875, "learning_rate": 7.095516569200779e-05, "loss": 0.8533, "step": 364 }, { "epoch": 0.21394125383603682, "grad_norm": 0.95703125, "learning_rate": 7.134502923976609e-05, "loss": 0.6971, "step": 366 }, { "epoch": 0.21511033172585123, "grad_norm": 1.1640625, "learning_rate": 7.173489278752437e-05, "loss": 0.7152, "step": 368 }, { "epoch": 0.21627940961566564, "grad_norm": 1.3984375, "learning_rate": 7.212475633528265e-05, "loss": 0.7176, "step": 370 }, { "epoch": 0.21744848750548004, "grad_norm": 0.84375, "learning_rate": 7.251461988304094e-05, "loss": 0.7054, "step": 372 }, { "epoch": 0.21861756539529445, "grad_norm": 0.859375, "learning_rate": 7.290448343079922e-05, "loss": 0.7157, "step": 374 }, { "epoch": 0.21978664328510886, "grad_norm": 0.99609375, "learning_rate": 7.32943469785575e-05, "loss": 0.742, "step": 376 }, { "epoch": 0.2209557211749233, "grad_norm": 1.1171875, "learning_rate": 7.368421052631579e-05, "loss": 0.7016, "step": 378 }, { "epoch": 0.2221247990647377, "grad_norm": 1.3359375, "learning_rate": 7.407407407407407e-05, "loss": 0.7263, "step": 380 }, { "epoch": 0.2232938769545521, "grad_norm": 0.8515625, "learning_rate": 7.446393762183236e-05, "loss": 0.7386, "step": 382 }, { "epoch": 0.2244629548443665, "grad_norm": 1.3359375, "learning_rate": 7.485380116959064e-05, "loss": 0.7488, "step": 384 }, { "epoch": 0.22563203273418092, "grad_norm": 0.98828125, "learning_rate": 7.524366471734892e-05, "loss": 0.7304, "step": 386 }, { "epoch": 0.22680111062399533, "grad_norm": 0.90625, "learning_rate": 7.563352826510721e-05, "loss": 0.6926, "step": 388 }, { "epoch": 0.22797018851380974, "grad_norm": 1.578125, "learning_rate": 7.602339181286549e-05, "loss": 0.7258, "step": 390 }, { "epoch": 0.22913926640362414, "grad_norm": 0.82421875, "learning_rate": 7.641325536062378e-05, "loss": 0.7169, "step": 392 }, { "epoch": 0.23030834429343855, "grad_norm": 0.78125, "learning_rate": 7.680311890838207e-05, "loss": 0.6962, "step": 394 }, { "epoch": 0.23147742218325296, "grad_norm": 1.0703125, "learning_rate": 7.719298245614036e-05, "loss": 0.718, "step": 396 }, { "epoch": 0.23264650007306736, "grad_norm": 0.84375, "learning_rate": 7.758284600389864e-05, "loss": 0.7106, "step": 398 }, { "epoch": 0.23381557796288177, "grad_norm": 1.109375, "learning_rate": 7.797270955165692e-05, "loss": 0.739, "step": 400 }, { "epoch": 0.23498465585269618, "grad_norm": 0.9140625, "learning_rate": 7.836257309941521e-05, "loss": 0.7037, "step": 402 }, { "epoch": 0.23615373374251059, "grad_norm": 0.76953125, "learning_rate": 7.875243664717349e-05, "loss": 0.717, "step": 404 }, { "epoch": 0.237322811632325, "grad_norm": 1.0546875, "learning_rate": 7.914230019493177e-05, "loss": 0.7041, "step": 406 }, { "epoch": 0.2384918895221394, "grad_norm": 0.921875, "learning_rate": 7.953216374269006e-05, "loss": 0.7016, "step": 408 }, { "epoch": 0.23966096741195383, "grad_norm": 0.8515625, "learning_rate": 7.992202729044834e-05, "loss": 0.7423, "step": 410 }, { "epoch": 0.24083004530176824, "grad_norm": 0.88671875, "learning_rate": 8.031189083820663e-05, "loss": 0.7128, "step": 412 }, { "epoch": 0.24199912319158265, "grad_norm": 0.9453125, "learning_rate": 8.070175438596491e-05, "loss": 0.7144, "step": 414 }, { "epoch": 0.24316820108139706, "grad_norm": 1.1328125, "learning_rate": 8.109161793372319e-05, "loss": 0.6987, "step": 416 }, { "epoch": 0.24433727897121146, "grad_norm": 0.9296875, "learning_rate": 8.148148148148148e-05, "loss": 0.6946, "step": 418 }, { "epoch": 0.24550635686102587, "grad_norm": 0.9296875, "learning_rate": 8.187134502923976e-05, "loss": 0.6954, "step": 420 }, { "epoch": 0.24667543475084028, "grad_norm": 0.84765625, "learning_rate": 8.226120857699804e-05, "loss": 0.7507, "step": 422 }, { "epoch": 0.24784451264065469, "grad_norm": 2.171875, "learning_rate": 8.265107212475634e-05, "loss": 0.9267, "step": 424 }, { "epoch": 0.2490135905304691, "grad_norm": 0.81640625, "learning_rate": 8.304093567251462e-05, "loss": 0.7222, "step": 426 }, { "epoch": 0.2501826684202835, "grad_norm": 1.7421875, "learning_rate": 8.343079922027291e-05, "loss": 0.8919, "step": 428 }, { "epoch": 0.25135174631009793, "grad_norm": 1.140625, "learning_rate": 8.382066276803119e-05, "loss": 0.7451, "step": 430 }, { "epoch": 0.25252082419991234, "grad_norm": 0.81640625, "learning_rate": 8.421052631578948e-05, "loss": 0.7027, "step": 432 }, { "epoch": 0.25368990208972675, "grad_norm": 0.79296875, "learning_rate": 8.460038986354776e-05, "loss": 0.712, "step": 434 }, { "epoch": 0.25485897997954116, "grad_norm": 0.8984375, "learning_rate": 8.499025341130604e-05, "loss": 0.7122, "step": 436 }, { "epoch": 0.25602805786935556, "grad_norm": 0.72265625, "learning_rate": 8.538011695906433e-05, "loss": 0.6871, "step": 438 }, { "epoch": 0.25719713575916997, "grad_norm": 0.77734375, "learning_rate": 8.576998050682261e-05, "loss": 0.7278, "step": 440 }, { "epoch": 0.2583662136489844, "grad_norm": 1.1171875, "learning_rate": 8.61598440545809e-05, "loss": 0.7023, "step": 442 }, { "epoch": 0.2595352915387988, "grad_norm": 0.8359375, "learning_rate": 8.654970760233918e-05, "loss": 0.7167, "step": 444 }, { "epoch": 0.2607043694286132, "grad_norm": 0.8203125, "learning_rate": 8.693957115009746e-05, "loss": 0.7346, "step": 446 }, { "epoch": 0.2618734473184276, "grad_norm": 0.9765625, "learning_rate": 8.732943469785574e-05, "loss": 0.7223, "step": 448 }, { "epoch": 0.263042525208242, "grad_norm": 0.96875, "learning_rate": 8.771929824561403e-05, "loss": 0.7032, "step": 450 }, { "epoch": 0.2642116030980564, "grad_norm": 0.9609375, "learning_rate": 8.810916179337231e-05, "loss": 0.6804, "step": 452 }, { "epoch": 0.2653806809878708, "grad_norm": 0.875, "learning_rate": 8.849902534113061e-05, "loss": 0.7224, "step": 454 }, { "epoch": 0.26654975887768523, "grad_norm": 0.8359375, "learning_rate": 8.888888888888889e-05, "loss": 0.7138, "step": 456 }, { "epoch": 0.26771883676749963, "grad_norm": 0.859375, "learning_rate": 8.927875243664718e-05, "loss": 0.7065, "step": 458 }, { "epoch": 0.26888791465731404, "grad_norm": 0.96484375, "learning_rate": 8.966861598440546e-05, "loss": 0.7289, "step": 460 }, { "epoch": 0.27005699254712845, "grad_norm": 1.0703125, "learning_rate": 9.005847953216374e-05, "loss": 0.6917, "step": 462 }, { "epoch": 0.27122607043694286, "grad_norm": 1.0078125, "learning_rate": 9.044834307992203e-05, "loss": 0.7493, "step": 464 }, { "epoch": 0.27239514832675726, "grad_norm": 1.0546875, "learning_rate": 9.083820662768031e-05, "loss": 0.7075, "step": 466 }, { "epoch": 0.27356422621657167, "grad_norm": 0.88671875, "learning_rate": 9.12280701754386e-05, "loss": 0.6926, "step": 468 }, { "epoch": 0.2747333041063861, "grad_norm": 0.734375, "learning_rate": 9.161793372319688e-05, "loss": 0.6714, "step": 470 }, { "epoch": 0.2759023819962005, "grad_norm": 0.73046875, "learning_rate": 9.200779727095516e-05, "loss": 0.6973, "step": 472 }, { "epoch": 0.2770714598860149, "grad_norm": 0.7578125, "learning_rate": 9.239766081871345e-05, "loss": 0.7289, "step": 474 }, { "epoch": 0.2782405377758293, "grad_norm": 0.84765625, "learning_rate": 9.278752436647173e-05, "loss": 0.7296, "step": 476 }, { "epoch": 0.2794096156656437, "grad_norm": 0.99609375, "learning_rate": 9.317738791423001e-05, "loss": 0.7229, "step": 478 }, { "epoch": 0.2805786935554581, "grad_norm": 1.0859375, "learning_rate": 9.35672514619883e-05, "loss": 0.7158, "step": 480 }, { "epoch": 0.2817477714452725, "grad_norm": 0.8671875, "learning_rate": 9.39571150097466e-05, "loss": 0.6862, "step": 482 }, { "epoch": 0.28291684933508693, "grad_norm": 0.77734375, "learning_rate": 9.434697855750488e-05, "loss": 0.6535, "step": 484 }, { "epoch": 0.28408592722490134, "grad_norm": 0.90234375, "learning_rate": 9.473684210526316e-05, "loss": 0.703, "step": 486 }, { "epoch": 0.28525500511471574, "grad_norm": 0.80078125, "learning_rate": 9.512670565302145e-05, "loss": 0.7296, "step": 488 }, { "epoch": 0.28642408300453015, "grad_norm": 1.3203125, "learning_rate": 9.551656920077973e-05, "loss": 0.7061, "step": 490 }, { "epoch": 0.2875931608943446, "grad_norm": 0.76953125, "learning_rate": 9.590643274853801e-05, "loss": 0.6731, "step": 492 }, { "epoch": 0.288762238784159, "grad_norm": 1.0234375, "learning_rate": 9.62962962962963e-05, "loss": 0.7141, "step": 494 }, { "epoch": 0.2899313166739734, "grad_norm": 1.8359375, "learning_rate": 9.668615984405458e-05, "loss": 0.7874, "step": 496 }, { "epoch": 0.29110039456378783, "grad_norm": 0.97265625, "learning_rate": 9.707602339181286e-05, "loss": 0.7179, "step": 498 }, { "epoch": 0.29226947245360224, "grad_norm": 0.8984375, "learning_rate": 9.746588693957115e-05, "loss": 0.6819, "step": 500 }, { "epoch": 0.29343855034341665, "grad_norm": 0.8359375, "learning_rate": 9.785575048732943e-05, "loss": 0.688, "step": 502 }, { "epoch": 0.29460762823323106, "grad_norm": 2.21875, "learning_rate": 9.824561403508771e-05, "loss": 0.8356, "step": 504 }, { "epoch": 0.29577670612304546, "grad_norm": 2.171875, "learning_rate": 9.8635477582846e-05, "loss": 0.7733, "step": 506 }, { "epoch": 0.29694578401285987, "grad_norm": 0.8203125, "learning_rate": 9.902534113060428e-05, "loss": 0.7087, "step": 508 }, { "epoch": 0.2981148619026743, "grad_norm": 1.2109375, "learning_rate": 9.941520467836257e-05, "loss": 0.7023, "step": 510 }, { "epoch": 0.2992839397924887, "grad_norm": 1.28125, "learning_rate": 9.980506822612086e-05, "loss": 0.7087, "step": 512 }, { "epoch": 0.3004530176823031, "grad_norm": 1.328125, "learning_rate": 0.00010019493177387915, "loss": 0.7002, "step": 514 }, { "epoch": 0.3016220955721175, "grad_norm": 0.8359375, "learning_rate": 0.00010058479532163743, "loss": 0.7355, "step": 516 }, { "epoch": 0.3027911734619319, "grad_norm": 1.21875, "learning_rate": 0.00010097465886939573, "loss": 0.6983, "step": 518 }, { "epoch": 0.3039602513517463, "grad_norm": 1.65625, "learning_rate": 0.000101364522417154, "loss": 0.7016, "step": 520 }, { "epoch": 0.3051293292415607, "grad_norm": 0.765625, "learning_rate": 0.0001017543859649123, "loss": 0.7096, "step": 522 }, { "epoch": 0.3062984071313751, "grad_norm": 1.46875, "learning_rate": 0.00010214424951267056, "loss": 0.7099, "step": 524 }, { "epoch": 0.30746748502118953, "grad_norm": 1.4375, "learning_rate": 0.00010253411306042886, "loss": 0.6931, "step": 526 }, { "epoch": 0.30863656291100394, "grad_norm": 0.96875, "learning_rate": 0.00010292397660818713, "loss": 0.7427, "step": 528 }, { "epoch": 0.30980564080081835, "grad_norm": 1.234375, "learning_rate": 0.00010331384015594543, "loss": 0.6609, "step": 530 }, { "epoch": 0.31097471869063276, "grad_norm": 1.0703125, "learning_rate": 0.0001037037037037037, "loss": 0.7195, "step": 532 }, { "epoch": 0.31214379658044716, "grad_norm": 0.7109375, "learning_rate": 0.000104093567251462, "loss": 0.6735, "step": 534 }, { "epoch": 0.31331287447026157, "grad_norm": 0.82421875, "learning_rate": 0.00010448343079922027, "loss": 0.7455, "step": 536 }, { "epoch": 0.314481952360076, "grad_norm": 0.91015625, "learning_rate": 0.00010487329434697856, "loss": 0.6683, "step": 538 }, { "epoch": 0.3156510302498904, "grad_norm": 0.7265625, "learning_rate": 0.00010526315789473685, "loss": 0.6981, "step": 540 }, { "epoch": 0.3168201081397048, "grad_norm": 0.69921875, "learning_rate": 0.00010565302144249513, "loss": 0.65, "step": 542 }, { "epoch": 0.3179891860295192, "grad_norm": 2.203125, "learning_rate": 0.00010604288499025342, "loss": 0.7997, "step": 544 }, { "epoch": 0.3191582639193336, "grad_norm": 0.890625, "learning_rate": 0.00010643274853801171, "loss": 0.7143, "step": 546 }, { "epoch": 0.320327341809148, "grad_norm": 0.98046875, "learning_rate": 0.00010682261208576998, "loss": 0.6619, "step": 548 }, { "epoch": 0.3214964196989624, "grad_norm": 0.97265625, "learning_rate": 0.00010721247563352828, "loss": 0.6632, "step": 550 }, { "epoch": 0.32266549758877683, "grad_norm": 0.765625, "learning_rate": 0.00010760233918128655, "loss": 0.6844, "step": 552 }, { "epoch": 0.32383457547859124, "grad_norm": 0.828125, "learning_rate": 0.00010799220272904485, "loss": 0.698, "step": 554 }, { "epoch": 0.32500365336840564, "grad_norm": 0.94140625, "learning_rate": 0.00010838206627680312, "loss": 0.6132, "step": 556 }, { "epoch": 0.3261727312582201, "grad_norm": 0.765625, "learning_rate": 0.00010877192982456141, "loss": 0.6507, "step": 558 }, { "epoch": 0.3273418091480345, "grad_norm": 1.03125, "learning_rate": 0.00010916179337231968, "loss": 0.696, "step": 560 }, { "epoch": 0.3285108870378489, "grad_norm": 3.0, "learning_rate": 0.00010955165692007798, "loss": 0.926, "step": 562 }, { "epoch": 0.3296799649276633, "grad_norm": 1.453125, "learning_rate": 0.00010994152046783625, "loss": 0.7797, "step": 564 }, { "epoch": 0.33084904281747773, "grad_norm": 1.71875, "learning_rate": 0.00011033138401559455, "loss": 0.7379, "step": 566 }, { "epoch": 0.33201812070729214, "grad_norm": 0.9921875, "learning_rate": 0.00011072124756335282, "loss": 0.679, "step": 568 }, { "epoch": 0.33318719859710655, "grad_norm": 0.77734375, "learning_rate": 0.00011111111111111112, "loss": 0.6668, "step": 570 }, { "epoch": 0.33435627648692096, "grad_norm": 0.796875, "learning_rate": 0.0001115009746588694, "loss": 0.7378, "step": 572 }, { "epoch": 0.33552535437673536, "grad_norm": 0.8046875, "learning_rate": 0.0001118908382066277, "loss": 0.7341, "step": 574 }, { "epoch": 0.33669443226654977, "grad_norm": 0.71875, "learning_rate": 0.00011228070175438597, "loss": 0.6793, "step": 576 }, { "epoch": 0.3378635101563642, "grad_norm": 0.82421875, "learning_rate": 0.00011267056530214426, "loss": 0.6926, "step": 578 }, { "epoch": 0.3390325880461786, "grad_norm": 0.89453125, "learning_rate": 0.00011306042884990253, "loss": 0.6393, "step": 580 }, { "epoch": 0.340201665935993, "grad_norm": 0.9375, "learning_rate": 0.00011345029239766083, "loss": 0.7207, "step": 582 }, { "epoch": 0.3413707438258074, "grad_norm": 0.75390625, "learning_rate": 0.0001138401559454191, "loss": 0.6168, "step": 584 }, { "epoch": 0.3425398217156218, "grad_norm": 0.7265625, "learning_rate": 0.0001142300194931774, "loss": 0.712, "step": 586 }, { "epoch": 0.3437088996054362, "grad_norm": 0.66796875, "learning_rate": 0.00011461988304093567, "loss": 0.6532, "step": 588 }, { "epoch": 0.3448779774952506, "grad_norm": 0.734375, "learning_rate": 0.00011500974658869397, "loss": 0.6766, "step": 590 }, { "epoch": 0.346047055385065, "grad_norm": 0.72265625, "learning_rate": 0.00011539961013645224, "loss": 0.6986, "step": 592 }, { "epoch": 0.34721613327487943, "grad_norm": 0.7109375, "learning_rate": 0.00011578947368421053, "loss": 0.6867, "step": 594 }, { "epoch": 0.34838521116469384, "grad_norm": 1.484375, "learning_rate": 0.0001161793372319688, "loss": 0.8175, "step": 596 }, { "epoch": 0.34955428905450825, "grad_norm": 0.8828125, "learning_rate": 0.0001165692007797271, "loss": 0.8122, "step": 598 }, { "epoch": 0.35072336694432266, "grad_norm": 0.96875, "learning_rate": 0.00011695906432748539, "loss": 0.7642, "step": 600 }, { "epoch": 0.35189244483413706, "grad_norm": 0.91796875, "learning_rate": 0.00011734892787524367, "loss": 0.703, "step": 602 }, { "epoch": 0.35306152272395147, "grad_norm": 0.7890625, "learning_rate": 0.00011773879142300195, "loss": 0.7031, "step": 604 }, { "epoch": 0.3542306006137659, "grad_norm": 0.98828125, "learning_rate": 0.00011812865497076025, "loss": 0.686, "step": 606 }, { "epoch": 0.3553996785035803, "grad_norm": 0.84375, "learning_rate": 0.00011851851851851852, "loss": 0.6708, "step": 608 }, { "epoch": 0.3565687563933947, "grad_norm": 0.8515625, "learning_rate": 0.00011890838206627682, "loss": 0.8084, "step": 610 }, { "epoch": 0.3577378342832091, "grad_norm": 0.75, "learning_rate": 0.00011929824561403509, "loss": 0.6705, "step": 612 }, { "epoch": 0.3589069121730235, "grad_norm": 0.6875, "learning_rate": 0.00011968810916179338, "loss": 0.6409, "step": 614 }, { "epoch": 0.3600759900628379, "grad_norm": 0.9375, "learning_rate": 0.00012007797270955165, "loss": 0.6805, "step": 616 }, { "epoch": 0.3612450679526523, "grad_norm": 1.59375, "learning_rate": 0.00012046783625730995, "loss": 0.7129, "step": 618 }, { "epoch": 0.36241414584246673, "grad_norm": 0.9609375, "learning_rate": 0.00012085769980506822, "loss": 0.6819, "step": 620 }, { "epoch": 0.36358322373228114, "grad_norm": 0.828125, "learning_rate": 0.00012124756335282652, "loss": 0.6424, "step": 622 }, { "epoch": 0.3647523016220956, "grad_norm": 0.7890625, "learning_rate": 0.00012163742690058479, "loss": 0.704, "step": 624 }, { "epoch": 0.36592137951191, "grad_norm": 1.3359375, "learning_rate": 0.00012202729044834309, "loss": 0.8172, "step": 626 }, { "epoch": 0.3670904574017244, "grad_norm": 0.92578125, "learning_rate": 0.00012241715399610137, "loss": 0.6486, "step": 628 }, { "epoch": 0.3682595352915388, "grad_norm": 1.3046875, "learning_rate": 0.00012280701754385965, "loss": 0.6772, "step": 630 }, { "epoch": 0.3694286131813532, "grad_norm": 1.1171875, "learning_rate": 0.00012319688109161794, "loss": 0.6997, "step": 632 }, { "epoch": 0.37059769107116763, "grad_norm": 0.94921875, "learning_rate": 0.00012358674463937622, "loss": 0.6984, "step": 634 }, { "epoch": 0.37176676896098204, "grad_norm": 0.72265625, "learning_rate": 0.0001239766081871345, "loss": 0.6796, "step": 636 }, { "epoch": 0.37293584685079645, "grad_norm": 0.92578125, "learning_rate": 0.0001243664717348928, "loss": 0.8194, "step": 638 }, { "epoch": 0.37410492474061086, "grad_norm": 0.84375, "learning_rate": 0.00012475633528265107, "loss": 0.7189, "step": 640 }, { "epoch": 0.37527400263042526, "grad_norm": 0.93359375, "learning_rate": 0.00012514619883040936, "loss": 0.6934, "step": 642 }, { "epoch": 0.37644308052023967, "grad_norm": 0.93359375, "learning_rate": 0.00012553606237816764, "loss": 0.6885, "step": 644 }, { "epoch": 0.3776121584100541, "grad_norm": 0.96875, "learning_rate": 0.00012592592592592592, "loss": 0.671, "step": 646 }, { "epoch": 0.3787812362998685, "grad_norm": 0.74609375, "learning_rate": 0.0001263157894736842, "loss": 0.6515, "step": 648 }, { "epoch": 0.3799503141896829, "grad_norm": 0.765625, "learning_rate": 0.00012670565302144252, "loss": 0.6833, "step": 650 }, { "epoch": 0.3811193920794973, "grad_norm": 0.76953125, "learning_rate": 0.00012709551656920077, "loss": 0.6363, "step": 652 }, { "epoch": 0.3822884699693117, "grad_norm": 0.70703125, "learning_rate": 0.00012748538011695908, "loss": 0.6848, "step": 654 }, { "epoch": 0.3834575478591261, "grad_norm": 0.93359375, "learning_rate": 0.00012787524366471734, "loss": 0.7383, "step": 656 }, { "epoch": 0.3846266257489405, "grad_norm": 0.92578125, "learning_rate": 0.00012826510721247565, "loss": 0.7057, "step": 658 }, { "epoch": 0.3857957036387549, "grad_norm": 0.734375, "learning_rate": 0.0001286549707602339, "loss": 0.7039, "step": 660 }, { "epoch": 0.38696478152856933, "grad_norm": 0.86328125, "learning_rate": 0.00012904483430799222, "loss": 0.727, "step": 662 }, { "epoch": 0.38813385941838374, "grad_norm": 1.2265625, "learning_rate": 0.00012943469785575048, "loss": 0.7128, "step": 664 }, { "epoch": 0.38930293730819815, "grad_norm": 0.671875, "learning_rate": 0.0001298245614035088, "loss": 0.6446, "step": 666 }, { "epoch": 0.39047201519801256, "grad_norm": 0.65234375, "learning_rate": 0.00013021442495126704, "loss": 0.6434, "step": 668 }, { "epoch": 0.39164109308782696, "grad_norm": 0.671875, "learning_rate": 0.00013060428849902535, "loss": 0.642, "step": 670 }, { "epoch": 0.39281017097764137, "grad_norm": 0.9609375, "learning_rate": 0.00013099415204678364, "loss": 0.6805, "step": 672 }, { "epoch": 0.3939792488674558, "grad_norm": 0.83984375, "learning_rate": 0.00013138401559454192, "loss": 0.6923, "step": 674 }, { "epoch": 0.3951483267572702, "grad_norm": 0.921875, "learning_rate": 0.0001317738791423002, "loss": 0.7944, "step": 676 }, { "epoch": 0.3963174046470846, "grad_norm": 0.94921875, "learning_rate": 0.0001321637426900585, "loss": 0.6876, "step": 678 }, { "epoch": 0.397486482536899, "grad_norm": 0.73046875, "learning_rate": 0.00013255360623781677, "loss": 0.6857, "step": 680 }, { "epoch": 0.3986555604267134, "grad_norm": 0.83984375, "learning_rate": 0.00013294346978557506, "loss": 0.6796, "step": 682 }, { "epoch": 0.3998246383165278, "grad_norm": 0.65234375, "learning_rate": 0.00013333333333333334, "loss": 0.6604, "step": 684 }, { "epoch": 0.4009937162063422, "grad_norm": 0.703125, "learning_rate": 0.00013372319688109162, "loss": 0.6999, "step": 686 }, { "epoch": 0.4021627940961567, "grad_norm": 1.0234375, "learning_rate": 0.0001341130604288499, "loss": 0.663, "step": 688 }, { "epoch": 0.4033318719859711, "grad_norm": 0.91796875, "learning_rate": 0.0001345029239766082, "loss": 0.6553, "step": 690 }, { "epoch": 0.4045009498757855, "grad_norm": 0.77734375, "learning_rate": 0.00013489278752436647, "loss": 0.6892, "step": 692 }, { "epoch": 0.4056700277655999, "grad_norm": 0.609375, "learning_rate": 0.00013528265107212476, "loss": 0.6529, "step": 694 }, { "epoch": 0.4068391056554143, "grad_norm": 0.7109375, "learning_rate": 0.00013567251461988304, "loss": 0.6728, "step": 696 }, { "epoch": 0.4080081835452287, "grad_norm": 0.671875, "learning_rate": 0.00013606237816764133, "loss": 0.6847, "step": 698 }, { "epoch": 0.4091772614350431, "grad_norm": 0.63671875, "learning_rate": 0.0001364522417153996, "loss": 0.7475, "step": 700 }, { "epoch": 0.41034633932485753, "grad_norm": 0.76171875, "learning_rate": 0.0001368421052631579, "loss": 0.6805, "step": 702 }, { "epoch": 0.41151541721467194, "grad_norm": 0.875, "learning_rate": 0.00013723196881091618, "loss": 0.8327, "step": 704 }, { "epoch": 0.41268449510448635, "grad_norm": 0.94921875, "learning_rate": 0.0001376218323586745, "loss": 0.663, "step": 706 }, { "epoch": 0.41385357299430076, "grad_norm": 0.6640625, "learning_rate": 0.00013801169590643274, "loss": 0.6675, "step": 708 }, { "epoch": 0.41502265088411516, "grad_norm": 0.95703125, "learning_rate": 0.00013840155945419105, "loss": 0.7458, "step": 710 }, { "epoch": 0.41619172877392957, "grad_norm": 0.66015625, "learning_rate": 0.0001387914230019493, "loss": 0.6385, "step": 712 }, { "epoch": 0.417360806663744, "grad_norm": 0.99609375, "learning_rate": 0.00013918128654970762, "loss": 0.7979, "step": 714 }, { "epoch": 0.4185298845535584, "grad_norm": 0.92578125, "learning_rate": 0.00013957115009746588, "loss": 0.6949, "step": 716 }, { "epoch": 0.4196989624433728, "grad_norm": 0.6953125, "learning_rate": 0.0001399610136452242, "loss": 0.6713, "step": 718 }, { "epoch": 0.4208680403331872, "grad_norm": 0.64453125, "learning_rate": 0.00014035087719298245, "loss": 0.6503, "step": 720 }, { "epoch": 0.4220371182230016, "grad_norm": 0.6640625, "learning_rate": 0.00014074074074074076, "loss": 0.6641, "step": 722 }, { "epoch": 0.423206196112816, "grad_norm": 0.6640625, "learning_rate": 0.000141130604288499, "loss": 0.7159, "step": 724 }, { "epoch": 0.4243752740026304, "grad_norm": 0.69140625, "learning_rate": 0.00014152046783625732, "loss": 0.786, "step": 726 }, { "epoch": 0.4255443518924448, "grad_norm": 0.92578125, "learning_rate": 0.00014191033138401558, "loss": 0.6511, "step": 728 }, { "epoch": 0.42671342978225923, "grad_norm": 0.890625, "learning_rate": 0.0001423001949317739, "loss": 0.8375, "step": 730 }, { "epoch": 0.42788250767207364, "grad_norm": 1.609375, "learning_rate": 0.00014269005847953217, "loss": 0.7475, "step": 732 }, { "epoch": 0.42905158556188805, "grad_norm": 1.3046875, "learning_rate": 0.00014307992202729046, "loss": 0.6989, "step": 734 }, { "epoch": 0.43022066345170246, "grad_norm": 0.9453125, "learning_rate": 0.00014346978557504874, "loss": 0.6792, "step": 736 }, { "epoch": 0.43138974134151686, "grad_norm": 0.734375, "learning_rate": 0.00014385964912280703, "loss": 0.7084, "step": 738 }, { "epoch": 0.43255881923133127, "grad_norm": 0.83203125, "learning_rate": 0.0001442495126705653, "loss": 0.7263, "step": 740 }, { "epoch": 0.4337278971211457, "grad_norm": 0.78515625, "learning_rate": 0.0001446393762183236, "loss": 0.706, "step": 742 }, { "epoch": 0.4348969750109601, "grad_norm": 0.671875, "learning_rate": 0.00014502923976608188, "loss": 0.6989, "step": 744 }, { "epoch": 0.4360660529007745, "grad_norm": 0.77734375, "learning_rate": 0.00014541910331384016, "loss": 0.7097, "step": 746 }, { "epoch": 0.4372351307905889, "grad_norm": 0.58984375, "learning_rate": 0.00014580896686159844, "loss": 0.6767, "step": 748 }, { "epoch": 0.4384042086804033, "grad_norm": 0.6796875, "learning_rate": 0.00014619883040935673, "loss": 0.7045, "step": 750 }, { "epoch": 0.4395732865702177, "grad_norm": 0.828125, "learning_rate": 0.000146588693957115, "loss": 0.6492, "step": 752 }, { "epoch": 0.4407423644600322, "grad_norm": 0.609375, "learning_rate": 0.0001469785575048733, "loss": 0.6776, "step": 754 }, { "epoch": 0.4419114423498466, "grad_norm": 0.77734375, "learning_rate": 0.00014736842105263158, "loss": 0.8008, "step": 756 }, { "epoch": 0.443080520239661, "grad_norm": 0.76953125, "learning_rate": 0.00014775828460038986, "loss": 0.6677, "step": 758 }, { "epoch": 0.4442495981294754, "grad_norm": 0.6796875, "learning_rate": 0.00014814814814814815, "loss": 0.6799, "step": 760 }, { "epoch": 0.4454186760192898, "grad_norm": 0.796875, "learning_rate": 0.00014853801169590643, "loss": 0.6626, "step": 762 }, { "epoch": 0.4465877539091042, "grad_norm": 0.8828125, "learning_rate": 0.0001489278752436647, "loss": 0.6814, "step": 764 }, { "epoch": 0.4477568317989186, "grad_norm": 0.671875, "learning_rate": 0.00014931773879142302, "loss": 0.7181, "step": 766 }, { "epoch": 0.448925909688733, "grad_norm": 0.8046875, "learning_rate": 0.00014970760233918128, "loss": 0.681, "step": 768 }, { "epoch": 0.45009498757854743, "grad_norm": 0.734375, "learning_rate": 0.0001500974658869396, "loss": 0.6625, "step": 770 }, { "epoch": 0.45126406546836184, "grad_norm": 0.69140625, "learning_rate": 0.00015048732943469785, "loss": 0.6558, "step": 772 }, { "epoch": 0.45243314335817625, "grad_norm": 0.66796875, "learning_rate": 0.00015087719298245616, "loss": 0.6968, "step": 774 }, { "epoch": 0.45360222124799066, "grad_norm": 1.171875, "learning_rate": 0.00015126705653021442, "loss": 0.7073, "step": 776 }, { "epoch": 0.45477129913780506, "grad_norm": 0.6875, "learning_rate": 0.00015165692007797273, "loss": 0.7151, "step": 778 }, { "epoch": 0.45594037702761947, "grad_norm": 0.66796875, "learning_rate": 0.00015204678362573098, "loss": 0.7192, "step": 780 }, { "epoch": 0.4571094549174339, "grad_norm": 0.70703125, "learning_rate": 0.0001524366471734893, "loss": 0.6937, "step": 782 }, { "epoch": 0.4582785328072483, "grad_norm": 0.76171875, "learning_rate": 0.00015282651072124755, "loss": 0.6859, "step": 784 }, { "epoch": 0.4594476106970627, "grad_norm": 0.7265625, "learning_rate": 0.00015321637426900586, "loss": 0.6491, "step": 786 }, { "epoch": 0.4606166885868771, "grad_norm": 0.6953125, "learning_rate": 0.00015360623781676414, "loss": 0.719, "step": 788 }, { "epoch": 0.4617857664766915, "grad_norm": 0.65625, "learning_rate": 0.00015399610136452243, "loss": 0.6906, "step": 790 }, { "epoch": 0.4629548443665059, "grad_norm": 0.72265625, "learning_rate": 0.0001543859649122807, "loss": 0.6957, "step": 792 }, { "epoch": 0.4641239222563203, "grad_norm": 0.59375, "learning_rate": 0.000154775828460039, "loss": 0.6863, "step": 794 }, { "epoch": 0.4652930001461347, "grad_norm": 0.69921875, "learning_rate": 0.00015516569200779728, "loss": 0.6845, "step": 796 }, { "epoch": 0.46646207803594913, "grad_norm": 0.625, "learning_rate": 0.00015555555555555556, "loss": 0.6393, "step": 798 }, { "epoch": 0.46763115592576354, "grad_norm": 0.67578125, "learning_rate": 0.00015594541910331385, "loss": 0.6739, "step": 800 }, { "epoch": 0.46880023381557795, "grad_norm": 0.62890625, "learning_rate": 0.00015633528265107213, "loss": 0.6919, "step": 802 }, { "epoch": 0.46996931170539236, "grad_norm": 0.6171875, "learning_rate": 0.00015672514619883041, "loss": 0.6885, "step": 804 }, { "epoch": 0.47113838959520676, "grad_norm": 0.64453125, "learning_rate": 0.0001571150097465887, "loss": 0.7075, "step": 806 }, { "epoch": 0.47230746748502117, "grad_norm": 0.6484375, "learning_rate": 0.00015750487329434698, "loss": 0.6716, "step": 808 }, { "epoch": 0.4734765453748356, "grad_norm": 0.5859375, "learning_rate": 0.00015789473684210527, "loss": 0.6595, "step": 810 }, { "epoch": 0.47464562326465, "grad_norm": 0.5390625, "learning_rate": 0.00015828460038986355, "loss": 0.6272, "step": 812 }, { "epoch": 0.4758147011544644, "grad_norm": 0.65234375, "learning_rate": 0.00015867446393762183, "loss": 0.6849, "step": 814 }, { "epoch": 0.4769837790442788, "grad_norm": 0.6953125, "learning_rate": 0.00015906432748538012, "loss": 0.6523, "step": 816 }, { "epoch": 0.4781528569340932, "grad_norm": 1.40625, "learning_rate": 0.0001594541910331384, "loss": 0.7112, "step": 818 }, { "epoch": 0.47932193482390767, "grad_norm": 0.625, "learning_rate": 0.00015984405458089668, "loss": 0.7073, "step": 820 }, { "epoch": 0.4804910127137221, "grad_norm": 0.5703125, "learning_rate": 0.000160233918128655, "loss": 0.6436, "step": 822 }, { "epoch": 0.4816600906035365, "grad_norm": 0.58203125, "learning_rate": 0.00016062378167641325, "loss": 0.669, "step": 824 }, { "epoch": 0.4828291684933509, "grad_norm": 0.94921875, "learning_rate": 0.00016101364522417156, "loss": 0.8224, "step": 826 }, { "epoch": 0.4839982463831653, "grad_norm": 0.75390625, "learning_rate": 0.00016140350877192982, "loss": 0.7526, "step": 828 }, { "epoch": 0.4851673242729797, "grad_norm": 0.64453125, "learning_rate": 0.00016179337231968813, "loss": 0.6986, "step": 830 }, { "epoch": 0.4863364021627941, "grad_norm": 0.625, "learning_rate": 0.00016218323586744639, "loss": 0.6496, "step": 832 }, { "epoch": 0.4875054800526085, "grad_norm": 0.6953125, "learning_rate": 0.0001625730994152047, "loss": 0.7165, "step": 834 }, { "epoch": 0.4886745579424229, "grad_norm": 0.58203125, "learning_rate": 0.00016296296296296295, "loss": 0.6453, "step": 836 }, { "epoch": 0.48984363583223733, "grad_norm": 0.61328125, "learning_rate": 0.00016335282651072126, "loss": 0.617, "step": 838 }, { "epoch": 0.49101271372205174, "grad_norm": 0.60546875, "learning_rate": 0.00016374269005847952, "loss": 0.6841, "step": 840 }, { "epoch": 0.49218179161186615, "grad_norm": 0.671875, "learning_rate": 0.00016413255360623783, "loss": 0.6668, "step": 842 }, { "epoch": 0.49335086950168056, "grad_norm": 0.71875, "learning_rate": 0.0001645224171539961, "loss": 0.7092, "step": 844 }, { "epoch": 0.49451994739149496, "grad_norm": 0.71484375, "learning_rate": 0.0001649122807017544, "loss": 0.6479, "step": 846 }, { "epoch": 0.49568902528130937, "grad_norm": 0.81640625, "learning_rate": 0.00016530214424951268, "loss": 0.6687, "step": 848 }, { "epoch": 0.4968581031711238, "grad_norm": 0.70703125, "learning_rate": 0.00016569200779727097, "loss": 0.6704, "step": 850 }, { "epoch": 0.4980271810609382, "grad_norm": 0.65234375, "learning_rate": 0.00016608187134502925, "loss": 0.6678, "step": 852 }, { "epoch": 0.4991962589507526, "grad_norm": 0.69140625, "learning_rate": 0.00016647173489278753, "loss": 0.6744, "step": 854 }, { "epoch": 0.500365336840567, "grad_norm": 0.79296875, "learning_rate": 0.00016686159844054582, "loss": 0.6565, "step": 856 }, { "epoch": 0.5015344147303814, "grad_norm": 0.62109375, "learning_rate": 0.0001672514619883041, "loss": 0.6499, "step": 858 }, { "epoch": 0.5027034926201959, "grad_norm": 0.84375, "learning_rate": 0.00016764132553606238, "loss": 0.6619, "step": 860 }, { "epoch": 0.5038725705100102, "grad_norm": 1.703125, "learning_rate": 0.00016803118908382067, "loss": 0.8945, "step": 862 }, { "epoch": 0.5050416483998247, "grad_norm": 0.9296875, "learning_rate": 0.00016842105263157895, "loss": 0.7597, "step": 864 }, { "epoch": 0.506210726289639, "grad_norm": 0.80078125, "learning_rate": 0.00016881091617933723, "loss": 0.6698, "step": 866 }, { "epoch": 0.5073798041794535, "grad_norm": 1.1953125, "learning_rate": 0.00016920077972709552, "loss": 0.7135, "step": 868 }, { "epoch": 0.5085488820692678, "grad_norm": 1.0859375, "learning_rate": 0.0001695906432748538, "loss": 0.8298, "step": 870 }, { "epoch": 0.5097179599590823, "grad_norm": 0.75390625, "learning_rate": 0.00016998050682261209, "loss": 0.6749, "step": 872 }, { "epoch": 0.5108870378488967, "grad_norm": 1.09375, "learning_rate": 0.00017037037037037037, "loss": 0.8318, "step": 874 }, { "epoch": 0.5120561157387111, "grad_norm": 0.7109375, "learning_rate": 0.00017076023391812865, "loss": 0.7329, "step": 876 }, { "epoch": 0.5132251936285255, "grad_norm": 0.87109375, "learning_rate": 0.00017115009746588694, "loss": 0.7127, "step": 878 }, { "epoch": 0.5143942715183399, "grad_norm": 9.375, "learning_rate": 0.00017153996101364522, "loss": 0.7969, "step": 880 }, { "epoch": 0.5155633494081543, "grad_norm": 2.015625, "learning_rate": 0.00017192982456140353, "loss": 0.8313, "step": 882 }, { "epoch": 0.5167324272979688, "grad_norm": 0.71875, "learning_rate": 0.0001723196881091618, "loss": 0.7693, "step": 884 }, { "epoch": 0.5179015051877831, "grad_norm": 0.625, "learning_rate": 0.0001727095516569201, "loss": 0.7127, "step": 886 }, { "epoch": 0.5190705830775976, "grad_norm": 0.72265625, "learning_rate": 0.00017309941520467836, "loss": 0.7283, "step": 888 }, { "epoch": 0.5202396609674119, "grad_norm": 0.70703125, "learning_rate": 0.00017348927875243667, "loss": 0.6591, "step": 890 }, { "epoch": 0.5214087388572264, "grad_norm": 0.66015625, "learning_rate": 0.00017387914230019492, "loss": 0.6794, "step": 892 }, { "epoch": 0.5225778167470407, "grad_norm": 0.57421875, "learning_rate": 0.00017426900584795323, "loss": 0.6299, "step": 894 }, { "epoch": 0.5237468946368552, "grad_norm": 0.92578125, "learning_rate": 0.0001746588693957115, "loss": 0.73, "step": 896 }, { "epoch": 0.5249159725266695, "grad_norm": 0.65625, "learning_rate": 0.0001750487329434698, "loss": 0.6538, "step": 898 }, { "epoch": 0.526085050416484, "grad_norm": 0.60546875, "learning_rate": 0.00017543859649122806, "loss": 0.7013, "step": 900 }, { "epoch": 0.5272541283062984, "grad_norm": 0.66015625, "learning_rate": 0.00017582846003898637, "loss": 0.6633, "step": 902 }, { "epoch": 0.5284232061961128, "grad_norm": 0.5234375, "learning_rate": 0.00017621832358674462, "loss": 0.6638, "step": 904 }, { "epoch": 0.5295922840859272, "grad_norm": 0.65234375, "learning_rate": 0.00017660818713450294, "loss": 0.6616, "step": 906 }, { "epoch": 0.5307613619757416, "grad_norm": 0.765625, "learning_rate": 0.00017699805068226122, "loss": 0.6935, "step": 908 }, { "epoch": 0.531930439865556, "grad_norm": 0.65625, "learning_rate": 0.0001773879142300195, "loss": 0.6795, "step": 910 }, { "epoch": 0.5330995177553705, "grad_norm": 0.87890625, "learning_rate": 0.00017777777777777779, "loss": 0.6797, "step": 912 }, { "epoch": 0.5342685956451848, "grad_norm": 0.75, "learning_rate": 0.00017816764132553607, "loss": 0.6583, "step": 914 }, { "epoch": 0.5354376735349993, "grad_norm": 0.640625, "learning_rate": 0.00017855750487329435, "loss": 0.6592, "step": 916 }, { "epoch": 0.5366067514248137, "grad_norm": 0.6328125, "learning_rate": 0.00017894736842105264, "loss": 0.7534, "step": 918 }, { "epoch": 0.5377758293146281, "grad_norm": 0.66796875, "learning_rate": 0.00017933723196881092, "loss": 0.699, "step": 920 }, { "epoch": 0.5389449072044425, "grad_norm": 0.7109375, "learning_rate": 0.0001797270955165692, "loss": 0.6759, "step": 922 }, { "epoch": 0.5401139850942569, "grad_norm": 0.58984375, "learning_rate": 0.0001801169590643275, "loss": 0.6676, "step": 924 }, { "epoch": 0.5412830629840714, "grad_norm": 0.6328125, "learning_rate": 0.00018050682261208577, "loss": 0.6734, "step": 926 }, { "epoch": 0.5424521408738857, "grad_norm": 0.72265625, "learning_rate": 0.00018089668615984406, "loss": 0.6545, "step": 928 }, { "epoch": 0.5436212187637002, "grad_norm": 0.65234375, "learning_rate": 0.00018128654970760234, "loss": 0.6863, "step": 930 }, { "epoch": 0.5447902966535145, "grad_norm": 0.54296875, "learning_rate": 0.00018167641325536062, "loss": 0.6536, "step": 932 }, { "epoch": 0.545959374543329, "grad_norm": 0.77734375, "learning_rate": 0.0001820662768031189, "loss": 0.6825, "step": 934 }, { "epoch": 0.5471284524331433, "grad_norm": 0.69921875, "learning_rate": 0.0001824561403508772, "loss": 0.7075, "step": 936 }, { "epoch": 0.5482975303229578, "grad_norm": 1.265625, "learning_rate": 0.00018284600389863547, "loss": 0.8103, "step": 938 }, { "epoch": 0.5494666082127722, "grad_norm": 1.4765625, "learning_rate": 0.00018323586744639376, "loss": 0.7081, "step": 940 }, { "epoch": 0.5506356861025866, "grad_norm": 0.734375, "learning_rate": 0.00018362573099415207, "loss": 0.6647, "step": 942 }, { "epoch": 0.551804763992401, "grad_norm": 0.98828125, "learning_rate": 0.00018401559454191032, "loss": 0.7202, "step": 944 }, { "epoch": 0.5529738418822154, "grad_norm": 1.09375, "learning_rate": 0.00018440545808966864, "loss": 0.7734, "step": 946 }, { "epoch": 0.5541429197720298, "grad_norm": 0.64453125, "learning_rate": 0.0001847953216374269, "loss": 0.6381, "step": 948 }, { "epoch": 0.5553119976618442, "grad_norm": 0.6796875, "learning_rate": 0.0001851851851851852, "loss": 0.7235, "step": 950 }, { "epoch": 0.5564810755516586, "grad_norm": 0.66796875, "learning_rate": 0.00018557504873294346, "loss": 0.8296, "step": 952 }, { "epoch": 0.5576501534414731, "grad_norm": 0.65234375, "learning_rate": 0.00018596491228070177, "loss": 0.6901, "step": 954 }, { "epoch": 0.5588192313312874, "grad_norm": 0.6796875, "learning_rate": 0.00018635477582846003, "loss": 0.6628, "step": 956 }, { "epoch": 0.5599883092211019, "grad_norm": 0.671875, "learning_rate": 0.00018674463937621834, "loss": 0.6689, "step": 958 }, { "epoch": 0.5611573871109162, "grad_norm": 0.6875, "learning_rate": 0.0001871345029239766, "loss": 0.7377, "step": 960 }, { "epoch": 0.5623264650007307, "grad_norm": 0.62890625, "learning_rate": 0.0001875243664717349, "loss": 0.7217, "step": 962 }, { "epoch": 0.563495542890545, "grad_norm": 0.65625, "learning_rate": 0.0001879142300194932, "loss": 0.6776, "step": 964 }, { "epoch": 0.5646646207803595, "grad_norm": 0.74609375, "learning_rate": 0.00018830409356725147, "loss": 0.694, "step": 966 }, { "epoch": 0.5658336986701739, "grad_norm": 0.5859375, "learning_rate": 0.00018869395711500976, "loss": 0.7305, "step": 968 }, { "epoch": 0.5670027765599883, "grad_norm": 0.671875, "learning_rate": 0.00018908382066276804, "loss": 0.6659, "step": 970 }, { "epoch": 0.5681718544498027, "grad_norm": 0.7578125, "learning_rate": 0.00018947368421052632, "loss": 0.705, "step": 972 }, { "epoch": 0.5693409323396171, "grad_norm": 0.64453125, "learning_rate": 0.0001898635477582846, "loss": 0.8348, "step": 974 }, { "epoch": 0.5705100102294315, "grad_norm": 0.65625, "learning_rate": 0.0001902534113060429, "loss": 0.7383, "step": 976 }, { "epoch": 0.571679088119246, "grad_norm": 0.609375, "learning_rate": 0.00019064327485380117, "loss": 0.7039, "step": 978 }, { "epoch": 0.5728481660090603, "grad_norm": 0.69921875, "learning_rate": 0.00019103313840155946, "loss": 0.7163, "step": 980 }, { "epoch": 0.5740172438988748, "grad_norm": 0.69140625, "learning_rate": 0.00019142300194931774, "loss": 0.6648, "step": 982 }, { "epoch": 0.5751863217886892, "grad_norm": 0.54296875, "learning_rate": 0.00019181286549707603, "loss": 0.6896, "step": 984 }, { "epoch": 0.5763553996785036, "grad_norm": 0.70703125, "learning_rate": 0.0001922027290448343, "loss": 0.7127, "step": 986 }, { "epoch": 0.577524477568318, "grad_norm": 0.8828125, "learning_rate": 0.0001925925925925926, "loss": 0.6549, "step": 988 }, { "epoch": 0.5786935554581324, "grad_norm": 0.77734375, "learning_rate": 0.00019298245614035088, "loss": 0.6581, "step": 990 }, { "epoch": 0.5798626333479469, "grad_norm": 0.578125, "learning_rate": 0.00019337231968810916, "loss": 0.6802, "step": 992 }, { "epoch": 0.5810317112377612, "grad_norm": 1.03125, "learning_rate": 0.00019376218323586744, "loss": 0.6906, "step": 994 }, { "epoch": 0.5822007891275757, "grad_norm": 0.7265625, "learning_rate": 0.00019415204678362573, "loss": 0.7383, "step": 996 }, { "epoch": 0.58336986701739, "grad_norm": 0.65234375, "learning_rate": 0.00019454191033138404, "loss": 0.7141, "step": 998 }, { "epoch": 0.5845389449072045, "grad_norm": 0.58984375, "learning_rate": 0.0001949317738791423, "loss": 0.6729, "step": 1000 }, { "epoch": 0.5857080227970188, "grad_norm": 0.59375, "learning_rate": 0.0001953216374269006, "loss": 0.6561, "step": 1002 }, { "epoch": 0.5868771006868333, "grad_norm": 1.109375, "learning_rate": 0.00019571150097465886, "loss": 0.9202, "step": 1004 }, { "epoch": 0.5880461785766476, "grad_norm": 0.921875, "learning_rate": 0.00019610136452241717, "loss": 0.7142, "step": 1006 }, { "epoch": 0.5892152564664621, "grad_norm": 105.0, "learning_rate": 0.00019649122807017543, "loss": 1.1271, "step": 1008 }, { "epoch": 0.5903843343562765, "grad_norm": 2.125, "learning_rate": 0.00019688109161793374, "loss": 1.1422, "step": 1010 }, { "epoch": 0.5915534122460909, "grad_norm": 1.0390625, "learning_rate": 0.000197270955165692, "loss": 0.7422, "step": 1012 }, { "epoch": 0.5927224901359053, "grad_norm": 0.84375, "learning_rate": 0.0001976608187134503, "loss": 0.7443, "step": 1014 }, { "epoch": 0.5938915680257197, "grad_norm": 0.6875, "learning_rate": 0.00019805068226120856, "loss": 0.7227, "step": 1016 }, { "epoch": 0.5950606459155341, "grad_norm": 1.28125, "learning_rate": 0.00019844054580896687, "loss": 0.7005, "step": 1018 }, { "epoch": 0.5962297238053486, "grad_norm": 0.7578125, "learning_rate": 0.00019883040935672513, "loss": 0.6514, "step": 1020 }, { "epoch": 0.5973988016951629, "grad_norm": 0.96875, "learning_rate": 0.00019922027290448344, "loss": 0.7211, "step": 1022 }, { "epoch": 0.5985678795849774, "grad_norm": 0.87890625, "learning_rate": 0.00019961013645224173, "loss": 0.6627, "step": 1024 }, { "epoch": 0.5997369574747917, "grad_norm": 0.6953125, "learning_rate": 0.0002, "loss": 0.6844, "step": 1026 }, { "epoch": 0.6009060353646062, "grad_norm": 1.0234375, "learning_rate": 0.00019990253411306043, "loss": 0.7113, "step": 1028 }, { "epoch": 0.6020751132544205, "grad_norm": 12.5, "learning_rate": 0.00019980506822612085, "loss": 0.751, "step": 1030 }, { "epoch": 0.603244191144235, "grad_norm": 1.9296875, "learning_rate": 0.00019970760233918128, "loss": 0.7035, "step": 1032 }, { "epoch": 0.6044132690340493, "grad_norm": 3.96875, "learning_rate": 0.00019961013645224173, "loss": 0.8057, "step": 1034 }, { "epoch": 0.6055823469238638, "grad_norm": 2.125, "learning_rate": 0.00019951267056530218, "loss": 1.0457, "step": 1036 }, { "epoch": 0.6067514248136782, "grad_norm": 0.640625, "learning_rate": 0.0001994152046783626, "loss": 0.7618, "step": 1038 }, { "epoch": 0.6079205027034926, "grad_norm": 1.3515625, "learning_rate": 0.00019931773879142302, "loss": 0.8979, "step": 1040 }, { "epoch": 0.609089580593307, "grad_norm": 1.1484375, "learning_rate": 0.00019922027290448344, "loss": 0.7494, "step": 1042 }, { "epoch": 0.6102586584831214, "grad_norm": 1.7109375, "learning_rate": 0.00019912280701754386, "loss": 0.7505, "step": 1044 }, { "epoch": 0.6114277363729358, "grad_norm": 1.2578125, "learning_rate": 0.0001990253411306043, "loss": 0.7356, "step": 1046 }, { "epoch": 0.6125968142627503, "grad_norm": 0.74609375, "learning_rate": 0.0001989278752436647, "loss": 0.7106, "step": 1048 }, { "epoch": 0.6137658921525647, "grad_norm": 0.71484375, "learning_rate": 0.00019883040935672513, "loss": 0.7151, "step": 1050 }, { "epoch": 0.6149349700423791, "grad_norm": 2.609375, "learning_rate": 0.00019873294346978558, "loss": 0.7294, "step": 1052 }, { "epoch": 0.6161040479321935, "grad_norm": 0.91796875, "learning_rate": 0.00019863547758284603, "loss": 0.7628, "step": 1054 }, { "epoch": 0.6172731258220079, "grad_norm": 0.6796875, "learning_rate": 0.00019853801169590645, "loss": 0.6701, "step": 1056 }, { "epoch": 0.6184422037118223, "grad_norm": 11.5, "learning_rate": 0.00019844054580896687, "loss": 0.8591, "step": 1058 }, { "epoch": 0.6196112816016367, "grad_norm": 6.125, "learning_rate": 0.0001983430799220273, "loss": 0.7596, "step": 1060 }, { "epoch": 0.6207803594914512, "grad_norm": 10.625, "learning_rate": 0.00019824561403508772, "loss": 0.7751, "step": 1062 }, { "epoch": 0.6219494373812655, "grad_norm": 1.59375, "learning_rate": 0.00019814814814814814, "loss": 0.8455, "step": 1064 }, { "epoch": 0.62311851527108, "grad_norm": 0.88671875, "learning_rate": 0.00019805068226120856, "loss": 0.7299, "step": 1066 }, { "epoch": 0.6242875931608943, "grad_norm": 0.69140625, "learning_rate": 0.000197953216374269, "loss": 0.7077, "step": 1068 }, { "epoch": 0.6254566710507088, "grad_norm": 0.76171875, "learning_rate": 0.00019785575048732946, "loss": 0.6911, "step": 1070 }, { "epoch": 0.6266257489405231, "grad_norm": 10.375, "learning_rate": 0.00019775828460038989, "loss": 0.6806, "step": 1072 }, { "epoch": 0.6277948268303376, "grad_norm": 0.8984375, "learning_rate": 0.0001976608187134503, "loss": 0.7528, "step": 1074 }, { "epoch": 0.628963904720152, "grad_norm": 1.6875, "learning_rate": 0.00019756335282651073, "loss": 0.6915, "step": 1076 }, { "epoch": 0.6301329826099664, "grad_norm": 4.125, "learning_rate": 0.00019746588693957115, "loss": 0.7393, "step": 1078 }, { "epoch": 0.6313020604997808, "grad_norm": 1.8515625, "learning_rate": 0.00019736842105263157, "loss": 0.698, "step": 1080 }, { "epoch": 0.6324711383895952, "grad_norm": 1.0546875, "learning_rate": 0.000197270955165692, "loss": 0.701, "step": 1082 }, { "epoch": 0.6336402162794096, "grad_norm": 2.15625, "learning_rate": 0.00019717348927875245, "loss": 0.6641, "step": 1084 }, { "epoch": 0.634809294169224, "grad_norm": 2.4375, "learning_rate": 0.00019707602339181287, "loss": 0.7128, "step": 1086 }, { "epoch": 0.6359783720590384, "grad_norm": 1.265625, "learning_rate": 0.00019697855750487332, "loss": 0.8843, "step": 1088 }, { "epoch": 0.6371474499488529, "grad_norm": 5.5625, "learning_rate": 0.00019688109161793374, "loss": 0.7482, "step": 1090 }, { "epoch": 0.6383165278386672, "grad_norm": 1.4765625, "learning_rate": 0.00019678362573099416, "loss": 0.8855, "step": 1092 }, { "epoch": 0.6394856057284817, "grad_norm": 0.98046875, "learning_rate": 0.00019668615984405458, "loss": 0.7744, "step": 1094 }, { "epoch": 0.640654683618296, "grad_norm": 0.76953125, "learning_rate": 0.000196588693957115, "loss": 0.72, "step": 1096 }, { "epoch": 0.6418237615081105, "grad_norm": 0.65625, "learning_rate": 0.00019649122807017543, "loss": 0.6577, "step": 1098 }, { "epoch": 0.6429928393979248, "grad_norm": 0.9765625, "learning_rate": 0.00019639376218323588, "loss": 0.736, "step": 1100 }, { "epoch": 0.6441619172877393, "grad_norm": 0.62109375, "learning_rate": 0.0001962962962962963, "loss": 0.7025, "step": 1102 }, { "epoch": 0.6453309951775537, "grad_norm": 0.80078125, "learning_rate": 0.00019619883040935675, "loss": 0.7076, "step": 1104 }, { "epoch": 0.6465000730673681, "grad_norm": 1.609375, "learning_rate": 0.00019610136452241717, "loss": 0.7121, "step": 1106 }, { "epoch": 0.6476691509571825, "grad_norm": 1.1015625, "learning_rate": 0.0001960038986354776, "loss": 0.705, "step": 1108 }, { "epoch": 0.6488382288469969, "grad_norm": 1.734375, "learning_rate": 0.00019590643274853802, "loss": 0.8037, "step": 1110 }, { "epoch": 0.6500073067368113, "grad_norm": 0.671875, "learning_rate": 0.00019580896686159844, "loss": 0.7236, "step": 1112 }, { "epoch": 0.6511763846266257, "grad_norm": 0.7265625, "learning_rate": 0.00019571150097465886, "loss": 0.764, "step": 1114 }, { "epoch": 0.6523454625164402, "grad_norm": 0.58984375, "learning_rate": 0.0001956140350877193, "loss": 0.6921, "step": 1116 }, { "epoch": 0.6535145404062546, "grad_norm": 0.61328125, "learning_rate": 0.00019551656920077973, "loss": 0.7026, "step": 1118 }, { "epoch": 0.654683618296069, "grad_norm": 0.921875, "learning_rate": 0.00019541910331384016, "loss": 0.6641, "step": 1120 }, { "epoch": 0.6558526961858834, "grad_norm": 0.63671875, "learning_rate": 0.0001953216374269006, "loss": 0.7031, "step": 1122 }, { "epoch": 0.6570217740756978, "grad_norm": 0.64453125, "learning_rate": 0.00019522417153996103, "loss": 0.6727, "step": 1124 }, { "epoch": 0.6581908519655122, "grad_norm": 0.62109375, "learning_rate": 0.00019512670565302145, "loss": 0.6963, "step": 1126 }, { "epoch": 0.6593599298553267, "grad_norm": 0.7265625, "learning_rate": 0.00019502923976608187, "loss": 0.665, "step": 1128 }, { "epoch": 0.660529007745141, "grad_norm": 0.68359375, "learning_rate": 0.0001949317738791423, "loss": 0.6752, "step": 1130 }, { "epoch": 0.6616980856349555, "grad_norm": 0.75390625, "learning_rate": 0.00019483430799220274, "loss": 0.6616, "step": 1132 }, { "epoch": 0.6628671635247698, "grad_norm": 0.68359375, "learning_rate": 0.00019473684210526317, "loss": 0.6426, "step": 1134 }, { "epoch": 0.6640362414145843, "grad_norm": 0.57421875, "learning_rate": 0.0001946393762183236, "loss": 0.675, "step": 1136 }, { "epoch": 0.6652053193043986, "grad_norm": 0.92578125, "learning_rate": 0.00019454191033138404, "loss": 0.732, "step": 1138 }, { "epoch": 0.6663743971942131, "grad_norm": 0.703125, "learning_rate": 0.00019444444444444446, "loss": 0.6701, "step": 1140 }, { "epoch": 0.6675434750840274, "grad_norm": 0.640625, "learning_rate": 0.00019434697855750488, "loss": 0.6637, "step": 1142 }, { "epoch": 0.6687125529738419, "grad_norm": 0.80078125, "learning_rate": 0.0001942495126705653, "loss": 0.8085, "step": 1144 }, { "epoch": 0.6698816308636563, "grad_norm": 0.625, "learning_rate": 0.00019415204678362573, "loss": 0.6828, "step": 1146 }, { "epoch": 0.6710507087534707, "grad_norm": 0.92578125, "learning_rate": 0.00019405458089668618, "loss": 0.7319, "step": 1148 }, { "epoch": 0.6722197866432851, "grad_norm": 0.72265625, "learning_rate": 0.0001939571150097466, "loss": 0.6919, "step": 1150 }, { "epoch": 0.6733888645330995, "grad_norm": 0.63671875, "learning_rate": 0.00019385964912280702, "loss": 0.7407, "step": 1152 }, { "epoch": 0.6745579424229139, "grad_norm": 1.171875, "learning_rate": 0.00019376218323586744, "loss": 0.7096, "step": 1154 }, { "epoch": 0.6757270203127284, "grad_norm": 0.62109375, "learning_rate": 0.0001936647173489279, "loss": 0.6948, "step": 1156 }, { "epoch": 0.6768960982025427, "grad_norm": 0.6171875, "learning_rate": 0.00019356725146198832, "loss": 0.6902, "step": 1158 }, { "epoch": 0.6780651760923572, "grad_norm": 0.51171875, "learning_rate": 0.00019346978557504874, "loss": 0.6991, "step": 1160 }, { "epoch": 0.6792342539821715, "grad_norm": 0.59375, "learning_rate": 0.00019337231968810916, "loss": 0.7059, "step": 1162 }, { "epoch": 0.680403331871986, "grad_norm": 0.953125, "learning_rate": 0.0001932748538011696, "loss": 0.8268, "step": 1164 }, { "epoch": 0.6815724097618003, "grad_norm": 0.69140625, "learning_rate": 0.00019317738791423003, "loss": 0.6697, "step": 1166 }, { "epoch": 0.6827414876516148, "grad_norm": 0.6328125, "learning_rate": 0.00019307992202729045, "loss": 0.6828, "step": 1168 }, { "epoch": 0.6839105655414291, "grad_norm": 0.76953125, "learning_rate": 0.00019298245614035088, "loss": 0.8377, "step": 1170 }, { "epoch": 0.6850796434312436, "grad_norm": 0.6953125, "learning_rate": 0.00019288499025341133, "loss": 0.7405, "step": 1172 }, { "epoch": 0.686248721321058, "grad_norm": 0.82421875, "learning_rate": 0.00019278752436647175, "loss": 0.7039, "step": 1174 }, { "epoch": 0.6874177992108724, "grad_norm": 0.71484375, "learning_rate": 0.00019269005847953217, "loss": 0.7217, "step": 1176 }, { "epoch": 0.6885868771006868, "grad_norm": 0.984375, "learning_rate": 0.0001925925925925926, "loss": 0.6667, "step": 1178 }, { "epoch": 0.6897559549905012, "grad_norm": 0.73046875, "learning_rate": 0.00019249512670565304, "loss": 0.7186, "step": 1180 }, { "epoch": 0.6909250328803157, "grad_norm": 0.74609375, "learning_rate": 0.00019239766081871346, "loss": 0.7037, "step": 1182 }, { "epoch": 0.69209411077013, "grad_norm": 0.57421875, "learning_rate": 0.0001923001949317739, "loss": 0.6953, "step": 1184 }, { "epoch": 0.6932631886599445, "grad_norm": 0.59765625, "learning_rate": 0.0001922027290448343, "loss": 0.6952, "step": 1186 }, { "epoch": 0.6944322665497589, "grad_norm": 0.5703125, "learning_rate": 0.00019210526315789473, "loss": 0.7018, "step": 1188 }, { "epoch": 0.6956013444395733, "grad_norm": 0.63671875, "learning_rate": 0.00019200779727095518, "loss": 0.675, "step": 1190 }, { "epoch": 0.6967704223293877, "grad_norm": 0.5703125, "learning_rate": 0.0001919103313840156, "loss": 0.6897, "step": 1192 }, { "epoch": 0.6979395002192021, "grad_norm": 0.5703125, "learning_rate": 0.00019181286549707603, "loss": 0.6644, "step": 1194 }, { "epoch": 0.6991085781090165, "grad_norm": 0.80078125, "learning_rate": 0.00019171539961013647, "loss": 0.7869, "step": 1196 }, { "epoch": 0.700277655998831, "grad_norm": 0.6640625, "learning_rate": 0.0001916179337231969, "loss": 0.6748, "step": 1198 }, { "epoch": 0.7014467338886453, "grad_norm": 0.5703125, "learning_rate": 0.00019152046783625732, "loss": 0.6426, "step": 1200 }, { "epoch": 0.7026158117784598, "grad_norm": 0.546875, "learning_rate": 0.00019142300194931774, "loss": 0.6642, "step": 1202 }, { "epoch": 0.7037848896682741, "grad_norm": 0.57421875, "learning_rate": 0.00019132553606237816, "loss": 0.7248, "step": 1204 }, { "epoch": 0.7049539675580886, "grad_norm": 0.52734375, "learning_rate": 0.0001912280701754386, "loss": 0.6599, "step": 1206 }, { "epoch": 0.7061230454479029, "grad_norm": 0.60546875, "learning_rate": 0.00019113060428849904, "loss": 0.6483, "step": 1208 }, { "epoch": 0.7072921233377174, "grad_norm": 0.53515625, "learning_rate": 0.00019103313840155946, "loss": 0.7163, "step": 1210 }, { "epoch": 0.7084612012275318, "grad_norm": 0.48828125, "learning_rate": 0.0001909356725146199, "loss": 0.6607, "step": 1212 }, { "epoch": 0.7096302791173462, "grad_norm": 0.61328125, "learning_rate": 0.00019083820662768033, "loss": 0.6647, "step": 1214 }, { "epoch": 0.7107993570071606, "grad_norm": 0.5390625, "learning_rate": 0.00019074074074074075, "loss": 0.7187, "step": 1216 }, { "epoch": 0.711968434896975, "grad_norm": 0.51953125, "learning_rate": 0.00019064327485380117, "loss": 0.6364, "step": 1218 }, { "epoch": 0.7131375127867894, "grad_norm": 0.61328125, "learning_rate": 0.0001905458089668616, "loss": 0.6514, "step": 1220 }, { "epoch": 0.7143065906766038, "grad_norm": 0.51953125, "learning_rate": 0.00019044834307992202, "loss": 0.6375, "step": 1222 }, { "epoch": 0.7154756685664182, "grad_norm": 0.52734375, "learning_rate": 0.00019035087719298247, "loss": 0.653, "step": 1224 }, { "epoch": 0.7166447464562327, "grad_norm": 0.58203125, "learning_rate": 0.0001902534113060429, "loss": 0.711, "step": 1226 }, { "epoch": 0.717813824346047, "grad_norm": 0.5078125, "learning_rate": 0.00019015594541910334, "loss": 0.6696, "step": 1228 }, { "epoch": 0.7189829022358615, "grad_norm": 0.51171875, "learning_rate": 0.00019005847953216376, "loss": 0.6747, "step": 1230 }, { "epoch": 0.7201519801256758, "grad_norm": 0.578125, "learning_rate": 0.00018996101364522418, "loss": 0.6863, "step": 1232 }, { "epoch": 0.7213210580154903, "grad_norm": 2.953125, "learning_rate": 0.0001898635477582846, "loss": 0.7781, "step": 1234 }, { "epoch": 0.7224901359053046, "grad_norm": 0.609375, "learning_rate": 0.00018976608187134503, "loss": 0.7072, "step": 1236 }, { "epoch": 0.7236592137951191, "grad_norm": 2.359375, "learning_rate": 0.00018966861598440545, "loss": 0.8231, "step": 1238 }, { "epoch": 0.7248282916849335, "grad_norm": 0.53125, "learning_rate": 0.0001895711500974659, "loss": 0.6973, "step": 1240 }, { "epoch": 0.7259973695747479, "grad_norm": 0.63671875, "learning_rate": 0.00018947368421052632, "loss": 0.6941, "step": 1242 }, { "epoch": 0.7271664474645623, "grad_norm": 0.546875, "learning_rate": 0.00018937621832358677, "loss": 0.6481, "step": 1244 }, { "epoch": 0.7283355253543767, "grad_norm": 0.59765625, "learning_rate": 0.0001892787524366472, "loss": 0.6984, "step": 1246 }, { "epoch": 0.7295046032441912, "grad_norm": 0.6484375, "learning_rate": 0.00018918128654970762, "loss": 0.7148, "step": 1248 }, { "epoch": 0.7306736811340055, "grad_norm": 0.52734375, "learning_rate": 0.00018908382066276804, "loss": 0.6359, "step": 1250 }, { "epoch": 0.73184275902382, "grad_norm": 0.51953125, "learning_rate": 0.00018898635477582846, "loss": 0.6469, "step": 1252 }, { "epoch": 0.7330118369136344, "grad_norm": 0.52734375, "learning_rate": 0.00018888888888888888, "loss": 0.6634, "step": 1254 }, { "epoch": 0.7341809148034488, "grad_norm": 0.5234375, "learning_rate": 0.0001887914230019493, "loss": 0.788, "step": 1256 }, { "epoch": 0.7353499926932632, "grad_norm": 0.490234375, "learning_rate": 0.00018869395711500976, "loss": 0.6582, "step": 1258 }, { "epoch": 0.7365190705830776, "grad_norm": 0.80078125, "learning_rate": 0.0001885964912280702, "loss": 0.6793, "step": 1260 }, { "epoch": 0.737688148472892, "grad_norm": 0.74609375, "learning_rate": 0.00018849902534113063, "loss": 0.6758, "step": 1262 }, { "epoch": 0.7388572263627065, "grad_norm": 0.58203125, "learning_rate": 0.00018840155945419105, "loss": 0.6362, "step": 1264 }, { "epoch": 0.7400263042525208, "grad_norm": 0.55078125, "learning_rate": 0.00018830409356725147, "loss": 0.692, "step": 1266 }, { "epoch": 0.7411953821423353, "grad_norm": 0.5546875, "learning_rate": 0.0001882066276803119, "loss": 0.6731, "step": 1268 }, { "epoch": 0.7423644600321496, "grad_norm": 0.75390625, "learning_rate": 0.00018810916179337232, "loss": 0.7731, "step": 1270 }, { "epoch": 0.7435335379219641, "grad_norm": 0.5625, "learning_rate": 0.00018801169590643274, "loss": 0.6723, "step": 1272 }, { "epoch": 0.7447026158117784, "grad_norm": 0.5703125, "learning_rate": 0.0001879142300194932, "loss": 0.6414, "step": 1274 }, { "epoch": 0.7458716937015929, "grad_norm": 0.50390625, "learning_rate": 0.0001878167641325536, "loss": 0.6392, "step": 1276 }, { "epoch": 0.7470407715914072, "grad_norm": 0.55859375, "learning_rate": 0.00018771929824561406, "loss": 0.6615, "step": 1278 }, { "epoch": 0.7482098494812217, "grad_norm": 0.546875, "learning_rate": 0.00018762183235867448, "loss": 0.6489, "step": 1280 }, { "epoch": 0.7493789273710361, "grad_norm": 0.46484375, "learning_rate": 0.0001875243664717349, "loss": 0.6432, "step": 1282 }, { "epoch": 0.7505480052608505, "grad_norm": 0.484375, "learning_rate": 0.00018742690058479533, "loss": 0.7061, "step": 1284 }, { "epoch": 0.7517170831506649, "grad_norm": 0.546875, "learning_rate": 0.00018732943469785575, "loss": 0.692, "step": 1286 }, { "epoch": 0.7528861610404793, "grad_norm": 0.60546875, "learning_rate": 0.00018723196881091617, "loss": 0.641, "step": 1288 }, { "epoch": 0.7540552389302937, "grad_norm": 0.4921875, "learning_rate": 0.0001871345029239766, "loss": 0.6842, "step": 1290 }, { "epoch": 0.7552243168201082, "grad_norm": 0.5078125, "learning_rate": 0.00018703703703703704, "loss": 0.6471, "step": 1292 }, { "epoch": 0.7563933947099225, "grad_norm": 0.54296875, "learning_rate": 0.0001869395711500975, "loss": 0.6824, "step": 1294 }, { "epoch": 0.757562472599737, "grad_norm": 0.515625, "learning_rate": 0.00018684210526315792, "loss": 0.6166, "step": 1296 }, { "epoch": 0.7587315504895513, "grad_norm": 0.82421875, "learning_rate": 0.00018674463937621834, "loss": 0.749, "step": 1298 }, { "epoch": 0.7599006283793658, "grad_norm": 1.21875, "learning_rate": 0.00018664717348927876, "loss": 0.7687, "step": 1300 }, { "epoch": 0.7610697062691801, "grad_norm": 1.046875, "learning_rate": 0.00018654970760233918, "loss": 0.7456, "step": 1302 }, { "epoch": 0.7622387841589946, "grad_norm": 0.8828125, "learning_rate": 0.0001864522417153996, "loss": 0.706, "step": 1304 }, { "epoch": 0.763407862048809, "grad_norm": 0.734375, "learning_rate": 0.00018635477582846003, "loss": 0.9545, "step": 1306 }, { "epoch": 0.7645769399386234, "grad_norm": 0.671875, "learning_rate": 0.00018625730994152048, "loss": 0.6858, "step": 1308 }, { "epoch": 0.7657460178284379, "grad_norm": 0.5625, "learning_rate": 0.0001861598440545809, "loss": 0.6233, "step": 1310 }, { "epoch": 0.7669150957182522, "grad_norm": 0.5703125, "learning_rate": 0.00018606237816764135, "loss": 0.6781, "step": 1312 }, { "epoch": 0.7680841736080667, "grad_norm": 0.58984375, "learning_rate": 0.00018596491228070177, "loss": 0.6687, "step": 1314 }, { "epoch": 0.769253251497881, "grad_norm": 0.58203125, "learning_rate": 0.0001858674463937622, "loss": 0.7021, "step": 1316 }, { "epoch": 0.7704223293876955, "grad_norm": 0.4921875, "learning_rate": 0.00018576998050682262, "loss": 0.6391, "step": 1318 }, { "epoch": 0.7715914072775099, "grad_norm": 0.49609375, "learning_rate": 0.00018567251461988304, "loss": 0.6663, "step": 1320 }, { "epoch": 0.7727604851673243, "grad_norm": 0.98046875, "learning_rate": 0.00018557504873294346, "loss": 0.6668, "step": 1322 }, { "epoch": 0.7739295630571387, "grad_norm": 0.46875, "learning_rate": 0.0001854775828460039, "loss": 0.6051, "step": 1324 }, { "epoch": 0.7750986409469531, "grad_norm": 0.48046875, "learning_rate": 0.00018538011695906433, "loss": 0.6486, "step": 1326 }, { "epoch": 0.7762677188367675, "grad_norm": 0.515625, "learning_rate": 0.00018528265107212478, "loss": 0.6645, "step": 1328 }, { "epoch": 0.777436796726582, "grad_norm": 0.5234375, "learning_rate": 0.0001851851851851852, "loss": 0.6235, "step": 1330 }, { "epoch": 0.7786058746163963, "grad_norm": 1.46875, "learning_rate": 0.00018508771929824563, "loss": 0.7003, "step": 1332 }, { "epoch": 0.7797749525062108, "grad_norm": 0.890625, "learning_rate": 0.00018499025341130605, "loss": 0.7938, "step": 1334 }, { "epoch": 0.7809440303960251, "grad_norm": 0.5859375, "learning_rate": 0.00018489278752436647, "loss": 0.6972, "step": 1336 }, { "epoch": 0.7821131082858396, "grad_norm": 0.51953125, "learning_rate": 0.0001847953216374269, "loss": 0.6443, "step": 1338 }, { "epoch": 0.7832821861756539, "grad_norm": 0.60546875, "learning_rate": 0.00018469785575048734, "loss": 0.5978, "step": 1340 }, { "epoch": 0.7844512640654684, "grad_norm": 1.2109375, "learning_rate": 0.00018460038986354776, "loss": 0.7214, "step": 1342 }, { "epoch": 0.7856203419552827, "grad_norm": 0.6484375, "learning_rate": 0.00018450292397660819, "loss": 0.6563, "step": 1344 }, { "epoch": 0.7867894198450972, "grad_norm": 0.5625, "learning_rate": 0.00018440545808966864, "loss": 0.6051, "step": 1346 }, { "epoch": 0.7879584977349116, "grad_norm": 0.470703125, "learning_rate": 0.00018430799220272906, "loss": 0.6615, "step": 1348 }, { "epoch": 0.789127575624726, "grad_norm": 0.6171875, "learning_rate": 0.00018421052631578948, "loss": 0.727, "step": 1350 }, { "epoch": 0.7902966535145404, "grad_norm": 0.7734375, "learning_rate": 0.0001841130604288499, "loss": 0.6712, "step": 1352 }, { "epoch": 0.7914657314043548, "grad_norm": 1.59375, "learning_rate": 0.00018401559454191032, "loss": 0.7094, "step": 1354 }, { "epoch": 0.7926348092941692, "grad_norm": 0.6640625, "learning_rate": 0.00018391812865497077, "loss": 0.6863, "step": 1356 }, { "epoch": 0.7938038871839836, "grad_norm": 0.62890625, "learning_rate": 0.0001838206627680312, "loss": 0.6378, "step": 1358 }, { "epoch": 0.794972965073798, "grad_norm": 0.640625, "learning_rate": 0.00018372319688109162, "loss": 0.8658, "step": 1360 }, { "epoch": 0.7961420429636125, "grad_norm": 0.54296875, "learning_rate": 0.00018362573099415207, "loss": 0.6507, "step": 1362 }, { "epoch": 0.7973111208534268, "grad_norm": 0.6328125, "learning_rate": 0.0001835282651072125, "loss": 0.7661, "step": 1364 }, { "epoch": 0.7984801987432413, "grad_norm": 0.625, "learning_rate": 0.0001834307992202729, "loss": 0.6932, "step": 1366 }, { "epoch": 0.7996492766330556, "grad_norm": 0.62890625, "learning_rate": 0.00018333333333333334, "loss": 0.681, "step": 1368 }, { "epoch": 0.8008183545228701, "grad_norm": 0.57421875, "learning_rate": 0.00018323586744639376, "loss": 0.6374, "step": 1370 }, { "epoch": 0.8019874324126844, "grad_norm": 0.6328125, "learning_rate": 0.0001831384015594542, "loss": 0.657, "step": 1372 }, { "epoch": 0.8031565103024989, "grad_norm": 0.83203125, "learning_rate": 0.00018304093567251463, "loss": 0.6511, "step": 1374 }, { "epoch": 0.8043255881923134, "grad_norm": 0.55078125, "learning_rate": 0.00018294346978557505, "loss": 0.6286, "step": 1376 }, { "epoch": 0.8054946660821277, "grad_norm": 0.6640625, "learning_rate": 0.00018284600389863547, "loss": 0.6964, "step": 1378 }, { "epoch": 0.8066637439719422, "grad_norm": 0.65625, "learning_rate": 0.00018274853801169592, "loss": 0.6383, "step": 1380 }, { "epoch": 0.8078328218617565, "grad_norm": 0.546875, "learning_rate": 0.00018265107212475635, "loss": 0.6459, "step": 1382 }, { "epoch": 0.809001899751571, "grad_norm": 0.49609375, "learning_rate": 0.00018255360623781677, "loss": 0.7644, "step": 1384 }, { "epoch": 0.8101709776413853, "grad_norm": 0.60546875, "learning_rate": 0.0001824561403508772, "loss": 0.6872, "step": 1386 }, { "epoch": 0.8113400555311998, "grad_norm": 0.5625, "learning_rate": 0.00018235867446393764, "loss": 0.7238, "step": 1388 }, { "epoch": 0.8125091334210142, "grad_norm": 0.671875, "learning_rate": 0.00018226120857699806, "loss": 0.6579, "step": 1390 }, { "epoch": 0.8136782113108286, "grad_norm": 0.546875, "learning_rate": 0.00018216374269005848, "loss": 0.6741, "step": 1392 }, { "epoch": 0.814847289200643, "grad_norm": 0.51953125, "learning_rate": 0.0001820662768031189, "loss": 0.6218, "step": 1394 }, { "epoch": 0.8160163670904574, "grad_norm": 0.4921875, "learning_rate": 0.00018196881091617936, "loss": 0.6077, "step": 1396 }, { "epoch": 0.8171854449802718, "grad_norm": 0.62109375, "learning_rate": 0.00018187134502923978, "loss": 0.6797, "step": 1398 }, { "epoch": 0.8183545228700863, "grad_norm": 0.55078125, "learning_rate": 0.0001817738791423002, "loss": 0.6416, "step": 1400 }, { "epoch": 0.8195236007599006, "grad_norm": 0.546875, "learning_rate": 0.00018167641325536062, "loss": 0.6664, "step": 1402 }, { "epoch": 0.8206926786497151, "grad_norm": 0.67578125, "learning_rate": 0.00018157894736842107, "loss": 0.6795, "step": 1404 }, { "epoch": 0.8218617565395294, "grad_norm": 0.5859375, "learning_rate": 0.0001814814814814815, "loss": 0.7478, "step": 1406 }, { "epoch": 0.8230308344293439, "grad_norm": 0.58203125, "learning_rate": 0.00018138401559454192, "loss": 0.6926, "step": 1408 }, { "epoch": 0.8241999123191582, "grad_norm": 0.671875, "learning_rate": 0.00018128654970760234, "loss": 0.6235, "step": 1410 }, { "epoch": 0.8253689902089727, "grad_norm": 1.203125, "learning_rate": 0.00018118908382066276, "loss": 0.7571, "step": 1412 }, { "epoch": 0.826538068098787, "grad_norm": 1.046875, "learning_rate": 0.0001810916179337232, "loss": 0.6671, "step": 1414 }, { "epoch": 0.8277071459886015, "grad_norm": 0.5390625, "learning_rate": 0.00018099415204678363, "loss": 0.715, "step": 1416 }, { "epoch": 0.8288762238784159, "grad_norm": 0.5859375, "learning_rate": 0.00018089668615984406, "loss": 0.643, "step": 1418 }, { "epoch": 0.8300453017682303, "grad_norm": 0.53125, "learning_rate": 0.0001807992202729045, "loss": 0.6701, "step": 1420 }, { "epoch": 0.8312143796580447, "grad_norm": 0.53125, "learning_rate": 0.00018070175438596493, "loss": 0.6674, "step": 1422 }, { "epoch": 0.8323834575478591, "grad_norm": 0.498046875, "learning_rate": 0.00018060428849902535, "loss": 0.6442, "step": 1424 }, { "epoch": 0.8335525354376735, "grad_norm": 0.58203125, "learning_rate": 0.00018050682261208577, "loss": 0.6345, "step": 1426 }, { "epoch": 0.834721613327488, "grad_norm": 0.640625, "learning_rate": 0.0001804093567251462, "loss": 0.7375, "step": 1428 }, { "epoch": 0.8358906912173023, "grad_norm": 0.5703125, "learning_rate": 0.00018031189083820664, "loss": 0.6404, "step": 1430 }, { "epoch": 0.8370597691071168, "grad_norm": 0.57421875, "learning_rate": 0.00018021442495126707, "loss": 0.6356, "step": 1432 }, { "epoch": 0.8382288469969311, "grad_norm": 0.51953125, "learning_rate": 0.0001801169590643275, "loss": 0.6207, "step": 1434 }, { "epoch": 0.8393979248867456, "grad_norm": 0.6171875, "learning_rate": 0.00018001949317738794, "loss": 0.6375, "step": 1436 }, { "epoch": 0.8405670027765599, "grad_norm": 1.453125, "learning_rate": 0.00017992202729044836, "loss": 0.8081, "step": 1438 }, { "epoch": 0.8417360806663744, "grad_norm": 0.86328125, "learning_rate": 0.00017982456140350878, "loss": 0.7778, "step": 1440 }, { "epoch": 0.8429051585561889, "grad_norm": 0.63671875, "learning_rate": 0.0001797270955165692, "loss": 0.7142, "step": 1442 }, { "epoch": 0.8440742364460032, "grad_norm": 0.55078125, "learning_rate": 0.00017962962962962963, "loss": 0.6502, "step": 1444 }, { "epoch": 0.8452433143358177, "grad_norm": 0.55078125, "learning_rate": 0.00017953216374269005, "loss": 0.6668, "step": 1446 }, { "epoch": 0.846412392225632, "grad_norm": 0.57421875, "learning_rate": 0.0001794346978557505, "loss": 0.6663, "step": 1448 }, { "epoch": 0.8475814701154465, "grad_norm": 3.65625, "learning_rate": 0.00017933723196881092, "loss": 0.7231, "step": 1450 }, { "epoch": 0.8487505480052608, "grad_norm": 0.5390625, "learning_rate": 0.00017923976608187137, "loss": 0.6152, "step": 1452 }, { "epoch": 0.8499196258950753, "grad_norm": 0.5390625, "learning_rate": 0.0001791423001949318, "loss": 0.7494, "step": 1454 }, { "epoch": 0.8510887037848897, "grad_norm": 0.5234375, "learning_rate": 0.00017904483430799221, "loss": 0.6777, "step": 1456 }, { "epoch": 0.8522577816747041, "grad_norm": 0.515625, "learning_rate": 0.00017894736842105264, "loss": 0.6667, "step": 1458 }, { "epoch": 0.8534268595645185, "grad_norm": 0.51171875, "learning_rate": 0.00017884990253411306, "loss": 0.6389, "step": 1460 }, { "epoch": 0.8545959374543329, "grad_norm": 0.5234375, "learning_rate": 0.00017875243664717348, "loss": 0.6435, "step": 1462 }, { "epoch": 0.8557650153441473, "grad_norm": 0.55859375, "learning_rate": 0.00017865497076023393, "loss": 0.654, "step": 1464 }, { "epoch": 0.8569340932339617, "grad_norm": 0.58984375, "learning_rate": 0.00017855750487329435, "loss": 0.6692, "step": 1466 }, { "epoch": 0.8581031711237761, "grad_norm": 0.58203125, "learning_rate": 0.0001784600389863548, "loss": 0.6807, "step": 1468 }, { "epoch": 0.8592722490135906, "grad_norm": 0.765625, "learning_rate": 0.00017836257309941523, "loss": 0.6314, "step": 1470 }, { "epoch": 0.8604413269034049, "grad_norm": 0.625, "learning_rate": 0.00017826510721247565, "loss": 0.6068, "step": 1472 }, { "epoch": 0.8616104047932194, "grad_norm": 0.59765625, "learning_rate": 0.00017816764132553607, "loss": 0.6545, "step": 1474 }, { "epoch": 0.8627794826830337, "grad_norm": 0.55078125, "learning_rate": 0.0001780701754385965, "loss": 0.6898, "step": 1476 }, { "epoch": 0.8639485605728482, "grad_norm": 0.58203125, "learning_rate": 0.00017797270955165691, "loss": 0.6213, "step": 1478 }, { "epoch": 0.8651176384626625, "grad_norm": 0.53125, "learning_rate": 0.00017787524366471734, "loss": 0.6487, "step": 1480 }, { "epoch": 0.866286716352477, "grad_norm": 0.546875, "learning_rate": 0.00017777777777777779, "loss": 0.6895, "step": 1482 }, { "epoch": 0.8674557942422914, "grad_norm": 0.515625, "learning_rate": 0.00017768031189083824, "loss": 0.6522, "step": 1484 }, { "epoch": 0.8686248721321058, "grad_norm": 0.56640625, "learning_rate": 0.00017758284600389866, "loss": 0.6484, "step": 1486 }, { "epoch": 0.8697939500219202, "grad_norm": 0.55078125, "learning_rate": 0.00017748538011695908, "loss": 0.6517, "step": 1488 }, { "epoch": 0.8709630279117346, "grad_norm": 0.5, "learning_rate": 0.0001773879142300195, "loss": 0.6644, "step": 1490 }, { "epoch": 0.872132105801549, "grad_norm": 0.50390625, "learning_rate": 0.00017729044834307992, "loss": 0.6383, "step": 1492 }, { "epoch": 0.8733011836913634, "grad_norm": 0.5703125, "learning_rate": 0.00017719298245614035, "loss": 0.7941, "step": 1494 }, { "epoch": 0.8744702615811778, "grad_norm": 0.6015625, "learning_rate": 0.00017709551656920077, "loss": 0.6882, "step": 1496 }, { "epoch": 0.8756393394709923, "grad_norm": 0.51171875, "learning_rate": 0.00017699805068226122, "loss": 0.627, "step": 1498 }, { "epoch": 0.8768084173608066, "grad_norm": 0.5390625, "learning_rate": 0.00017690058479532167, "loss": 0.6141, "step": 1500 }, { "epoch": 0.8779774952506211, "grad_norm": 0.482421875, "learning_rate": 0.0001768031189083821, "loss": 0.6226, "step": 1502 }, { "epoch": 0.8791465731404354, "grad_norm": 0.5078125, "learning_rate": 0.0001767056530214425, "loss": 0.6834, "step": 1504 }, { "epoch": 0.8803156510302499, "grad_norm": 0.5703125, "learning_rate": 0.00017660818713450294, "loss": 0.6269, "step": 1506 }, { "epoch": 0.8814847289200644, "grad_norm": 0.498046875, "learning_rate": 0.00017651072124756336, "loss": 0.6494, "step": 1508 }, { "epoch": 0.8826538068098787, "grad_norm": 0.49609375, "learning_rate": 0.00017641325536062378, "loss": 0.6028, "step": 1510 }, { "epoch": 0.8838228846996932, "grad_norm": 0.458984375, "learning_rate": 0.0001763157894736842, "loss": 0.6428, "step": 1512 }, { "epoch": 0.8849919625895075, "grad_norm": 0.58984375, "learning_rate": 0.00017621832358674462, "loss": 0.6303, "step": 1514 }, { "epoch": 0.886161040479322, "grad_norm": 0.546875, "learning_rate": 0.00017612085769980507, "loss": 0.6522, "step": 1516 }, { "epoch": 0.8873301183691363, "grad_norm": 0.494140625, "learning_rate": 0.00017602339181286552, "loss": 0.6279, "step": 1518 }, { "epoch": 0.8884991962589508, "grad_norm": 0.458984375, "learning_rate": 0.00017592592592592595, "loss": 0.6507, "step": 1520 }, { "epoch": 0.8896682741487651, "grad_norm": 0.47265625, "learning_rate": 0.00017582846003898637, "loss": 0.6137, "step": 1522 }, { "epoch": 0.8908373520385796, "grad_norm": 0.578125, "learning_rate": 0.0001757309941520468, "loss": 0.649, "step": 1524 }, { "epoch": 0.892006429928394, "grad_norm": 0.470703125, "learning_rate": 0.0001756335282651072, "loss": 0.6498, "step": 1526 }, { "epoch": 0.8931755078182084, "grad_norm": 1.59375, "learning_rate": 0.00017553606237816763, "loss": 0.6804, "step": 1528 }, { "epoch": 0.8943445857080228, "grad_norm": 0.47265625, "learning_rate": 0.00017543859649122806, "loss": 0.6186, "step": 1530 }, { "epoch": 0.8955136635978372, "grad_norm": 0.47265625, "learning_rate": 0.0001753411306042885, "loss": 0.6518, "step": 1532 }, { "epoch": 0.8966827414876516, "grad_norm": 0.44921875, "learning_rate": 0.00017524366471734896, "loss": 0.6458, "step": 1534 }, { "epoch": 0.897851819377466, "grad_norm": 0.482421875, "learning_rate": 0.00017514619883040938, "loss": 0.6642, "step": 1536 }, { "epoch": 0.8990208972672804, "grad_norm": 0.470703125, "learning_rate": 0.0001750487329434698, "loss": 0.6254, "step": 1538 }, { "epoch": 0.9001899751570949, "grad_norm": 0.50390625, "learning_rate": 0.00017495126705653022, "loss": 0.6158, "step": 1540 }, { "epoch": 0.9013590530469092, "grad_norm": 0.65234375, "learning_rate": 0.00017485380116959065, "loss": 0.6635, "step": 1542 }, { "epoch": 0.9025281309367237, "grad_norm": 0.4765625, "learning_rate": 0.00017475633528265107, "loss": 0.6242, "step": 1544 }, { "epoch": 0.903697208826538, "grad_norm": 0.61328125, "learning_rate": 0.0001746588693957115, "loss": 0.6503, "step": 1546 }, { "epoch": 0.9048662867163525, "grad_norm": 0.65625, "learning_rate": 0.00017456140350877194, "loss": 0.5875, "step": 1548 }, { "epoch": 0.9060353646061668, "grad_norm": 0.48046875, "learning_rate": 0.00017446393762183236, "loss": 0.6683, "step": 1550 }, { "epoch": 0.9072044424959813, "grad_norm": 0.466796875, "learning_rate": 0.0001743664717348928, "loss": 0.6176, "step": 1552 }, { "epoch": 0.9083735203857957, "grad_norm": 0.455078125, "learning_rate": 0.00017426900584795323, "loss": 0.6351, "step": 1554 }, { "epoch": 0.9095425982756101, "grad_norm": 0.5703125, "learning_rate": 0.00017417153996101366, "loss": 0.668, "step": 1556 }, { "epoch": 0.9107116761654245, "grad_norm": 0.53515625, "learning_rate": 0.00017407407407407408, "loss": 0.6425, "step": 1558 }, { "epoch": 0.9118807540552389, "grad_norm": 0.482421875, "learning_rate": 0.0001739766081871345, "loss": 0.6127, "step": 1560 }, { "epoch": 0.9130498319450533, "grad_norm": 0.61328125, "learning_rate": 0.00017387914230019492, "loss": 0.6055, "step": 1562 }, { "epoch": 0.9142189098348678, "grad_norm": 0.48828125, "learning_rate": 0.00017378167641325537, "loss": 0.6276, "step": 1564 }, { "epoch": 0.9153879877246821, "grad_norm": 0.55859375, "learning_rate": 0.0001736842105263158, "loss": 0.6818, "step": 1566 }, { "epoch": 0.9165570656144966, "grad_norm": 0.66015625, "learning_rate": 0.00017358674463937624, "loss": 0.76, "step": 1568 }, { "epoch": 0.9177261435043109, "grad_norm": 4.78125, "learning_rate": 0.00017348927875243667, "loss": 0.6739, "step": 1570 }, { "epoch": 0.9188952213941254, "grad_norm": 2.078125, "learning_rate": 0.0001733918128654971, "loss": 0.6578, "step": 1572 }, { "epoch": 0.9200642992839398, "grad_norm": 0.8515625, "learning_rate": 0.0001732943469785575, "loss": 0.6705, "step": 1574 }, { "epoch": 0.9212333771737542, "grad_norm": 0.80078125, "learning_rate": 0.00017319688109161793, "loss": 0.6605, "step": 1576 }, { "epoch": 0.9224024550635687, "grad_norm": 0.55859375, "learning_rate": 0.00017309941520467836, "loss": 0.6346, "step": 1578 }, { "epoch": 0.923571532953383, "grad_norm": 0.77734375, "learning_rate": 0.0001730019493177388, "loss": 0.6291, "step": 1580 }, { "epoch": 0.9247406108431975, "grad_norm": 1.0625, "learning_rate": 0.00017290448343079923, "loss": 0.6774, "step": 1582 }, { "epoch": 0.9259096887330118, "grad_norm": 0.58203125, "learning_rate": 0.00017280701754385965, "loss": 0.6603, "step": 1584 }, { "epoch": 0.9270787666228263, "grad_norm": 0.96484375, "learning_rate": 0.0001727095516569201, "loss": 0.6435, "step": 1586 }, { "epoch": 0.9282478445126406, "grad_norm": 0.70703125, "learning_rate": 0.00017261208576998052, "loss": 0.7842, "step": 1588 }, { "epoch": 0.9294169224024551, "grad_norm": 0.494140625, "learning_rate": 0.00017251461988304094, "loss": 0.6378, "step": 1590 }, { "epoch": 0.9305860002922695, "grad_norm": 0.515625, "learning_rate": 0.00017241715399610137, "loss": 0.6809, "step": 1592 }, { "epoch": 0.9317550781820839, "grad_norm": 0.48046875, "learning_rate": 0.0001723196881091618, "loss": 0.6226, "step": 1594 }, { "epoch": 0.9329241560718983, "grad_norm": 0.48828125, "learning_rate": 0.00017222222222222224, "loss": 0.6763, "step": 1596 }, { "epoch": 0.9340932339617127, "grad_norm": 0.49609375, "learning_rate": 0.00017212475633528266, "loss": 0.7279, "step": 1598 }, { "epoch": 0.9352623118515271, "grad_norm": 0.671875, "learning_rate": 0.00017202729044834308, "loss": 0.6655, "step": 1600 }, { "epoch": 0.9364313897413415, "grad_norm": 0.494140625, "learning_rate": 0.00017192982456140353, "loss": 0.6658, "step": 1602 }, { "epoch": 0.9376004676311559, "grad_norm": 0.53515625, "learning_rate": 0.00017183235867446395, "loss": 0.7698, "step": 1604 }, { "epoch": 0.9387695455209704, "grad_norm": 0.62890625, "learning_rate": 0.00017173489278752438, "loss": 0.6555, "step": 1606 }, { "epoch": 0.9399386234107847, "grad_norm": 0.470703125, "learning_rate": 0.0001716374269005848, "loss": 0.6465, "step": 1608 }, { "epoch": 0.9411077013005992, "grad_norm": 0.56640625, "learning_rate": 0.00017153996101364522, "loss": 0.6642, "step": 1610 }, { "epoch": 0.9422767791904135, "grad_norm": 0.47265625, "learning_rate": 0.00017144249512670567, "loss": 0.6579, "step": 1612 }, { "epoch": 0.943445857080228, "grad_norm": 0.53515625, "learning_rate": 0.0001713450292397661, "loss": 0.6427, "step": 1614 }, { "epoch": 0.9446149349700423, "grad_norm": 0.48828125, "learning_rate": 0.00017124756335282651, "loss": 0.6114, "step": 1616 }, { "epoch": 0.9457840128598568, "grad_norm": 0.486328125, "learning_rate": 0.00017115009746588694, "loss": 0.5887, "step": 1618 }, { "epoch": 0.9469530907496712, "grad_norm": 0.4765625, "learning_rate": 0.00017105263157894739, "loss": 0.6266, "step": 1620 }, { "epoch": 0.9481221686394856, "grad_norm": 0.55078125, "learning_rate": 0.0001709551656920078, "loss": 0.642, "step": 1622 }, { "epoch": 0.9492912465293, "grad_norm": 0.482421875, "learning_rate": 0.00017085769980506823, "loss": 0.605, "step": 1624 }, { "epoch": 0.9504603244191144, "grad_norm": 0.5078125, "learning_rate": 0.00017076023391812865, "loss": 0.5981, "step": 1626 }, { "epoch": 0.9516294023089288, "grad_norm": 0.515625, "learning_rate": 0.0001706627680311891, "loss": 0.7126, "step": 1628 }, { "epoch": 0.9527984801987432, "grad_norm": 0.78125, "learning_rate": 0.00017056530214424952, "loss": 0.7645, "step": 1630 }, { "epoch": 0.9539675580885576, "grad_norm": 0.74609375, "learning_rate": 0.00017046783625730995, "loss": 0.7024, "step": 1632 }, { "epoch": 0.9551366359783721, "grad_norm": 0.99609375, "learning_rate": 0.00017037037037037037, "loss": 0.7639, "step": 1634 }, { "epoch": 0.9563057138681864, "grad_norm": 0.53125, "learning_rate": 0.00017027290448343082, "loss": 0.624, "step": 1636 }, { "epoch": 0.9574747917580009, "grad_norm": 0.5234375, "learning_rate": 0.00017017543859649124, "loss": 0.6589, "step": 1638 }, { "epoch": 0.9586438696478153, "grad_norm": 0.53515625, "learning_rate": 0.00017007797270955166, "loss": 0.6068, "step": 1640 }, { "epoch": 0.9598129475376297, "grad_norm": 0.46484375, "learning_rate": 0.00016998050682261209, "loss": 0.6223, "step": 1642 }, { "epoch": 0.9609820254274442, "grad_norm": 0.47265625, "learning_rate": 0.00016988304093567254, "loss": 0.6217, "step": 1644 }, { "epoch": 0.9621511033172585, "grad_norm": 0.48046875, "learning_rate": 0.00016978557504873296, "loss": 0.6249, "step": 1646 }, { "epoch": 0.963320181207073, "grad_norm": 0.51171875, "learning_rate": 0.00016968810916179338, "loss": 0.6676, "step": 1648 }, { "epoch": 0.9644892590968873, "grad_norm": 0.4375, "learning_rate": 0.0001695906432748538, "loss": 0.6327, "step": 1650 }, { "epoch": 0.9656583369867018, "grad_norm": 0.482421875, "learning_rate": 0.00016949317738791422, "loss": 0.6233, "step": 1652 }, { "epoch": 0.9668274148765161, "grad_norm": 0.5703125, "learning_rate": 0.00016939571150097467, "loss": 0.6466, "step": 1654 }, { "epoch": 0.9679964927663306, "grad_norm": 0.47265625, "learning_rate": 0.0001692982456140351, "loss": 0.6173, "step": 1656 }, { "epoch": 0.969165570656145, "grad_norm": 0.5078125, "learning_rate": 0.00016920077972709552, "loss": 0.6279, "step": 1658 }, { "epoch": 0.9703346485459594, "grad_norm": 4.78125, "learning_rate": 0.00016910331384015597, "loss": 0.8359, "step": 1660 }, { "epoch": 0.9715037264357738, "grad_norm": 0.74609375, "learning_rate": 0.0001690058479532164, "loss": 0.6702, "step": 1662 }, { "epoch": 0.9726728043255882, "grad_norm": 0.73046875, "learning_rate": 0.0001689083820662768, "loss": 0.6553, "step": 1664 }, { "epoch": 0.9738418822154026, "grad_norm": 1.9453125, "learning_rate": 0.00016881091617933723, "loss": 0.7719, "step": 1666 }, { "epoch": 0.975010960105217, "grad_norm": 0.55859375, "learning_rate": 0.00016871345029239766, "loss": 0.6581, "step": 1668 }, { "epoch": 0.9761800379950314, "grad_norm": 0.5078125, "learning_rate": 0.0001686159844054581, "loss": 0.6169, "step": 1670 }, { "epoch": 0.9773491158848459, "grad_norm": 0.4921875, "learning_rate": 0.00016851851851851853, "loss": 0.6449, "step": 1672 }, { "epoch": 0.9785181937746602, "grad_norm": 0.54296875, "learning_rate": 0.00016842105263157895, "loss": 0.6011, "step": 1674 }, { "epoch": 0.9796872716644747, "grad_norm": 0.53515625, "learning_rate": 0.0001683235867446394, "loss": 0.6331, "step": 1676 }, { "epoch": 0.980856349554289, "grad_norm": 0.55078125, "learning_rate": 0.00016822612085769982, "loss": 0.635, "step": 1678 }, { "epoch": 0.9820254274441035, "grad_norm": 0.490234375, "learning_rate": 0.00016812865497076025, "loss": 0.6567, "step": 1680 }, { "epoch": 0.9831945053339178, "grad_norm": 0.46484375, "learning_rate": 0.00016803118908382067, "loss": 0.592, "step": 1682 }, { "epoch": 0.9843635832237323, "grad_norm": 0.4921875, "learning_rate": 0.0001679337231968811, "loss": 0.6291, "step": 1684 }, { "epoch": 0.9855326611135466, "grad_norm": 0.5234375, "learning_rate": 0.0001678362573099415, "loss": 0.6125, "step": 1686 }, { "epoch": 0.9867017390033611, "grad_norm": 0.451171875, "learning_rate": 0.00016773879142300196, "loss": 0.5989, "step": 1688 }, { "epoch": 0.9878708168931755, "grad_norm": 0.494140625, "learning_rate": 0.00016764132553606238, "loss": 0.6214, "step": 1690 }, { "epoch": 0.9890398947829899, "grad_norm": 0.4375, "learning_rate": 0.00016754385964912283, "loss": 0.5794, "step": 1692 }, { "epoch": 0.9902089726728043, "grad_norm": 0.42578125, "learning_rate": 0.00016744639376218326, "loss": 0.6138, "step": 1694 }, { "epoch": 0.9913780505626187, "grad_norm": 0.765625, "learning_rate": 0.00016734892787524368, "loss": 0.5763, "step": 1696 }, { "epoch": 0.9925471284524331, "grad_norm": 0.44921875, "learning_rate": 0.0001672514619883041, "loss": 0.5501, "step": 1698 }, { "epoch": 0.9937162063422476, "grad_norm": 0.462890625, "learning_rate": 0.00016715399610136452, "loss": 0.605, "step": 1700 }, { "epoch": 0.9948852842320619, "grad_norm": 0.470703125, "learning_rate": 0.00016705653021442494, "loss": 0.5973, "step": 1702 }, { "epoch": 0.9960543621218764, "grad_norm": 0.4453125, "learning_rate": 0.0001669590643274854, "loss": 0.5792, "step": 1704 }, { "epoch": 0.9972234400116908, "grad_norm": 0.435546875, "learning_rate": 0.00016686159844054582, "loss": 0.5833, "step": 1706 }, { "epoch": 0.9983925179015052, "grad_norm": 0.6640625, "learning_rate": 0.00016676413255360627, "loss": 0.6092, "step": 1708 }, { "epoch": 0.9995615957913196, "grad_norm": 0.470703125, "learning_rate": 0.0001666666666666667, "loss": 0.6591, "step": 1710 }, { "epoch": 1.000730673681134, "grad_norm": 0.46875, "learning_rate": 0.0001665692007797271, "loss": 0.5507, "step": 1712 }, { "epoch": 1.0018997515709485, "grad_norm": 0.453125, "learning_rate": 0.00016647173489278753, "loss": 0.5538, "step": 1714 }, { "epoch": 1.0030688294607628, "grad_norm": 0.4609375, "learning_rate": 0.00016637426900584796, "loss": 0.511, "step": 1716 }, { "epoch": 1.0042379073505772, "grad_norm": 0.45703125, "learning_rate": 0.00016627680311890838, "loss": 0.5208, "step": 1718 }, { "epoch": 1.0054069852403917, "grad_norm": 0.47265625, "learning_rate": 0.0001661793372319688, "loss": 0.5428, "step": 1720 }, { "epoch": 1.006576063130206, "grad_norm": 2.71875, "learning_rate": 0.00016608187134502925, "loss": 0.5979, "step": 1722 }, { "epoch": 1.0077451410200204, "grad_norm": 0.46484375, "learning_rate": 0.0001659844054580897, "loss": 0.5444, "step": 1724 }, { "epoch": 1.0089142189098348, "grad_norm": 0.51953125, "learning_rate": 0.00016588693957115012, "loss": 0.574, "step": 1726 }, { "epoch": 1.0100832967996494, "grad_norm": 0.58203125, "learning_rate": 0.00016578947368421054, "loss": 0.4927, "step": 1728 }, { "epoch": 1.0112523746894637, "grad_norm": 0.48828125, "learning_rate": 0.00016569200779727097, "loss": 0.5525, "step": 1730 }, { "epoch": 1.012421452579278, "grad_norm": 0.49609375, "learning_rate": 0.0001655945419103314, "loss": 0.5884, "step": 1732 }, { "epoch": 1.0135905304690924, "grad_norm": 0.455078125, "learning_rate": 0.0001654970760233918, "loss": 0.4956, "step": 1734 }, { "epoch": 1.014759608358907, "grad_norm": 0.546875, "learning_rate": 0.00016539961013645223, "loss": 0.5446, "step": 1736 }, { "epoch": 1.0159286862487213, "grad_norm": 0.486328125, "learning_rate": 0.00016530214424951268, "loss": 0.5602, "step": 1738 }, { "epoch": 1.0170977641385357, "grad_norm": 0.6328125, "learning_rate": 0.0001652046783625731, "loss": 0.5462, "step": 1740 }, { "epoch": 1.01826684202835, "grad_norm": 0.54296875, "learning_rate": 0.00016510721247563355, "loss": 0.53, "step": 1742 }, { "epoch": 1.0194359199181646, "grad_norm": 0.61328125, "learning_rate": 0.00016500974658869398, "loss": 0.5374, "step": 1744 }, { "epoch": 1.020604997807979, "grad_norm": 0.498046875, "learning_rate": 0.0001649122807017544, "loss": 0.5258, "step": 1746 }, { "epoch": 1.0217740756977933, "grad_norm": 0.51953125, "learning_rate": 0.00016481481481481482, "loss": 0.5118, "step": 1748 }, { "epoch": 1.0229431535876077, "grad_norm": 0.462890625, "learning_rate": 0.00016471734892787524, "loss": 0.4822, "step": 1750 }, { "epoch": 1.0241122314774223, "grad_norm": 0.56640625, "learning_rate": 0.00016461988304093567, "loss": 0.5435, "step": 1752 }, { "epoch": 1.0252813093672366, "grad_norm": 0.515625, "learning_rate": 0.0001645224171539961, "loss": 0.5703, "step": 1754 }, { "epoch": 1.026450387257051, "grad_norm": 0.4921875, "learning_rate": 0.00016442495126705654, "loss": 0.587, "step": 1756 }, { "epoch": 1.0276194651468653, "grad_norm": 0.453125, "learning_rate": 0.00016432748538011699, "loss": 0.5146, "step": 1758 }, { "epoch": 1.0287885430366799, "grad_norm": 0.462890625, "learning_rate": 0.0001642300194931774, "loss": 0.5102, "step": 1760 }, { "epoch": 1.0299576209264942, "grad_norm": 0.52734375, "learning_rate": 0.00016413255360623783, "loss": 0.5178, "step": 1762 }, { "epoch": 1.0311266988163086, "grad_norm": 0.4765625, "learning_rate": 0.00016403508771929825, "loss": 0.4984, "step": 1764 }, { "epoch": 1.032295776706123, "grad_norm": 0.609375, "learning_rate": 0.00016393762183235868, "loss": 0.4915, "step": 1766 }, { "epoch": 1.0334648545959375, "grad_norm": 0.49609375, "learning_rate": 0.0001638401559454191, "loss": 0.5493, "step": 1768 }, { "epoch": 1.0346339324857519, "grad_norm": 0.52734375, "learning_rate": 0.00016374269005847952, "loss": 0.4906, "step": 1770 }, { "epoch": 1.0358030103755662, "grad_norm": 0.546875, "learning_rate": 0.00016364522417153997, "loss": 0.5478, "step": 1772 }, { "epoch": 1.0369720882653808, "grad_norm": 0.5625, "learning_rate": 0.0001635477582846004, "loss": 0.5187, "step": 1774 }, { "epoch": 1.0381411661551951, "grad_norm": 0.50390625, "learning_rate": 0.00016345029239766084, "loss": 0.5421, "step": 1776 }, { "epoch": 1.0393102440450095, "grad_norm": 0.61328125, "learning_rate": 0.00016335282651072126, "loss": 0.5583, "step": 1778 }, { "epoch": 1.0404793219348238, "grad_norm": 0.470703125, "learning_rate": 0.00016325536062378169, "loss": 0.5283, "step": 1780 }, { "epoch": 1.0416483998246384, "grad_norm": 0.7578125, "learning_rate": 0.0001631578947368421, "loss": 0.5294, "step": 1782 }, { "epoch": 1.0428174777144528, "grad_norm": 0.54296875, "learning_rate": 0.00016306042884990253, "loss": 0.5682, "step": 1784 }, { "epoch": 1.0439865556042671, "grad_norm": 0.490234375, "learning_rate": 0.00016296296296296295, "loss": 0.5088, "step": 1786 }, { "epoch": 1.0451556334940815, "grad_norm": 0.490234375, "learning_rate": 0.0001628654970760234, "loss": 0.4887, "step": 1788 }, { "epoch": 1.046324711383896, "grad_norm": 0.5703125, "learning_rate": 0.00016276803118908382, "loss": 0.5478, "step": 1790 }, { "epoch": 1.0474937892737104, "grad_norm": 0.466796875, "learning_rate": 0.00016267056530214427, "loss": 0.5156, "step": 1792 }, { "epoch": 1.0486628671635247, "grad_norm": 0.51171875, "learning_rate": 0.0001625730994152047, "loss": 0.528, "step": 1794 }, { "epoch": 1.049831945053339, "grad_norm": 0.45703125, "learning_rate": 0.00016247563352826512, "loss": 0.5, "step": 1796 }, { "epoch": 1.0510010229431537, "grad_norm": 0.984375, "learning_rate": 0.00016237816764132554, "loss": 0.6486, "step": 1798 }, { "epoch": 1.052170100832968, "grad_norm": 0.55078125, "learning_rate": 0.00016228070175438596, "loss": 0.5215, "step": 1800 }, { "epoch": 1.0533391787227824, "grad_norm": 0.5078125, "learning_rate": 0.00016218323586744639, "loss": 0.5179, "step": 1802 }, { "epoch": 1.0545082566125967, "grad_norm": 0.62109375, "learning_rate": 0.00016208576998050683, "loss": 0.5772, "step": 1804 }, { "epoch": 1.0556773345024113, "grad_norm": 0.45703125, "learning_rate": 0.00016198830409356726, "loss": 0.504, "step": 1806 }, { "epoch": 1.0568464123922257, "grad_norm": 0.59765625, "learning_rate": 0.0001618908382066277, "loss": 0.5463, "step": 1808 }, { "epoch": 1.05801549028204, "grad_norm": 0.5703125, "learning_rate": 0.00016179337231968813, "loss": 0.5146, "step": 1810 }, { "epoch": 1.0591845681718544, "grad_norm": 0.5390625, "learning_rate": 0.00016169590643274855, "loss": 0.5626, "step": 1812 }, { "epoch": 1.060353646061669, "grad_norm": 0.5078125, "learning_rate": 0.00016159844054580897, "loss": 0.5224, "step": 1814 }, { "epoch": 1.0615227239514833, "grad_norm": 0.44140625, "learning_rate": 0.0001615009746588694, "loss": 0.5456, "step": 1816 }, { "epoch": 1.0626918018412976, "grad_norm": 0.8125, "learning_rate": 0.00016140350877192982, "loss": 0.5831, "step": 1818 }, { "epoch": 1.063860879731112, "grad_norm": 0.515625, "learning_rate": 0.00016130604288499027, "loss": 0.5131, "step": 1820 }, { "epoch": 1.0650299576209266, "grad_norm": 0.55859375, "learning_rate": 0.0001612085769980507, "loss": 0.6097, "step": 1822 }, { "epoch": 1.066199035510741, "grad_norm": 0.62890625, "learning_rate": 0.0001611111111111111, "loss": 0.5895, "step": 1824 }, { "epoch": 1.0673681134005553, "grad_norm": 0.5546875, "learning_rate": 0.00016101364522417156, "loss": 0.5048, "step": 1826 }, { "epoch": 1.0685371912903698, "grad_norm": 0.5078125, "learning_rate": 0.00016091617933723198, "loss": 0.5041, "step": 1828 }, { "epoch": 1.0697062691801842, "grad_norm": 0.51953125, "learning_rate": 0.0001608187134502924, "loss": 0.5293, "step": 1830 }, { "epoch": 1.0708753470699985, "grad_norm": 0.50390625, "learning_rate": 0.00016072124756335283, "loss": 0.5054, "step": 1832 }, { "epoch": 1.072044424959813, "grad_norm": 0.5, "learning_rate": 0.00016062378167641325, "loss": 0.5345, "step": 1834 }, { "epoch": 1.0732135028496272, "grad_norm": 0.48828125, "learning_rate": 0.0001605263157894737, "loss": 0.5057, "step": 1836 }, { "epoch": 1.0743825807394418, "grad_norm": 0.55859375, "learning_rate": 0.00016042884990253412, "loss": 0.5478, "step": 1838 }, { "epoch": 1.0755516586292562, "grad_norm": 0.6640625, "learning_rate": 0.00016033138401559454, "loss": 0.6639, "step": 1840 }, { "epoch": 1.0767207365190705, "grad_norm": 0.53125, "learning_rate": 0.000160233918128655, "loss": 0.529, "step": 1842 }, { "epoch": 1.077889814408885, "grad_norm": 0.6875, "learning_rate": 0.00016013645224171542, "loss": 0.5433, "step": 1844 }, { "epoch": 1.0790588922986994, "grad_norm": 0.5859375, "learning_rate": 0.00016003898635477584, "loss": 0.6416, "step": 1846 }, { "epoch": 1.0802279701885138, "grad_norm": 0.447265625, "learning_rate": 0.00015994152046783626, "loss": 0.5125, "step": 1848 }, { "epoch": 1.0813970480783281, "grad_norm": 0.5546875, "learning_rate": 0.00015984405458089668, "loss": 0.5199, "step": 1850 }, { "epoch": 1.0825661259681427, "grad_norm": 0.46484375, "learning_rate": 0.00015974658869395713, "loss": 0.5225, "step": 1852 }, { "epoch": 1.083735203857957, "grad_norm": 0.4765625, "learning_rate": 0.00015964912280701756, "loss": 0.4977, "step": 1854 }, { "epoch": 1.0849042817477714, "grad_norm": 0.51171875, "learning_rate": 0.00015955165692007798, "loss": 0.5717, "step": 1856 }, { "epoch": 1.0860733596375858, "grad_norm": 0.51171875, "learning_rate": 0.0001594541910331384, "loss": 0.5352, "step": 1858 }, { "epoch": 1.0872424375274004, "grad_norm": 0.439453125, "learning_rate": 0.00015935672514619885, "loss": 0.5502, "step": 1860 }, { "epoch": 1.0884115154172147, "grad_norm": 0.474609375, "learning_rate": 0.00015925925925925927, "loss": 0.5091, "step": 1862 }, { "epoch": 1.089580593307029, "grad_norm": 0.453125, "learning_rate": 0.0001591617933723197, "loss": 0.5104, "step": 1864 }, { "epoch": 1.0907496711968434, "grad_norm": 0.439453125, "learning_rate": 0.00015906432748538012, "loss": 0.5178, "step": 1866 }, { "epoch": 1.091918749086658, "grad_norm": 0.44921875, "learning_rate": 0.00015896686159844057, "loss": 0.5414, "step": 1868 }, { "epoch": 1.0930878269764723, "grad_norm": 0.46875, "learning_rate": 0.000158869395711501, "loss": 0.4606, "step": 1870 }, { "epoch": 1.0942569048662867, "grad_norm": 0.5, "learning_rate": 0.0001587719298245614, "loss": 0.5489, "step": 1872 }, { "epoch": 1.095425982756101, "grad_norm": 0.447265625, "learning_rate": 0.00015867446393762183, "loss": 0.4745, "step": 1874 }, { "epoch": 1.0965950606459156, "grad_norm": 0.52734375, "learning_rate": 0.00015857699805068228, "loss": 0.5481, "step": 1876 }, { "epoch": 1.09776413853573, "grad_norm": 0.52734375, "learning_rate": 0.0001584795321637427, "loss": 0.5632, "step": 1878 }, { "epoch": 1.0989332164255443, "grad_norm": 0.47265625, "learning_rate": 0.00015838206627680313, "loss": 0.5136, "step": 1880 }, { "epoch": 1.1001022943153587, "grad_norm": 0.61328125, "learning_rate": 0.00015828460038986355, "loss": 0.51, "step": 1882 }, { "epoch": 1.1012713722051732, "grad_norm": 0.474609375, "learning_rate": 0.000158187134502924, "loss": 0.5066, "step": 1884 }, { "epoch": 1.1024404500949876, "grad_norm": 0.48046875, "learning_rate": 0.00015808966861598442, "loss": 0.6068, "step": 1886 }, { "epoch": 1.103609527984802, "grad_norm": 0.478515625, "learning_rate": 0.00015799220272904484, "loss": 0.5732, "step": 1888 }, { "epoch": 1.1047786058746163, "grad_norm": 0.55078125, "learning_rate": 0.00015789473684210527, "loss": 0.4733, "step": 1890 }, { "epoch": 1.1059476837644309, "grad_norm": 0.482421875, "learning_rate": 0.0001577972709551657, "loss": 0.5106, "step": 1892 }, { "epoch": 1.1071167616542452, "grad_norm": 0.50390625, "learning_rate": 0.00015769980506822614, "loss": 0.545, "step": 1894 }, { "epoch": 1.1082858395440596, "grad_norm": 0.73046875, "learning_rate": 0.00015760233918128656, "loss": 0.5188, "step": 1896 }, { "epoch": 1.1094549174338741, "grad_norm": 0.51953125, "learning_rate": 0.00015750487329434698, "loss": 0.5393, "step": 1898 }, { "epoch": 1.1106239953236885, "grad_norm": 0.4453125, "learning_rate": 0.00015740740740740743, "loss": 0.5315, "step": 1900 }, { "epoch": 1.1117930732135028, "grad_norm": 0.48828125, "learning_rate": 0.00015730994152046785, "loss": 0.5042, "step": 1902 }, { "epoch": 1.1129621511033172, "grad_norm": 0.65625, "learning_rate": 0.00015721247563352828, "loss": 0.5491, "step": 1904 }, { "epoch": 1.1141312289931316, "grad_norm": 0.51953125, "learning_rate": 0.0001571150097465887, "loss": 0.5612, "step": 1906 }, { "epoch": 1.1153003068829461, "grad_norm": 0.50390625, "learning_rate": 0.00015701754385964912, "loss": 0.5288, "step": 1908 }, { "epoch": 1.1164693847727605, "grad_norm": 0.4765625, "learning_rate": 0.00015692007797270957, "loss": 0.5249, "step": 1910 }, { "epoch": 1.1176384626625748, "grad_norm": 0.478515625, "learning_rate": 0.00015682261208577, "loss": 0.5217, "step": 1912 }, { "epoch": 1.1188075405523894, "grad_norm": 0.53515625, "learning_rate": 0.00015672514619883041, "loss": 0.5379, "step": 1914 }, { "epoch": 1.1199766184422038, "grad_norm": 0.462890625, "learning_rate": 0.00015662768031189086, "loss": 0.5196, "step": 1916 }, { "epoch": 1.121145696332018, "grad_norm": 0.5078125, "learning_rate": 0.00015653021442495129, "loss": 0.5316, "step": 1918 }, { "epoch": 1.1223147742218325, "grad_norm": 0.455078125, "learning_rate": 0.0001564327485380117, "loss": 0.5255, "step": 1920 }, { "epoch": 1.123483852111647, "grad_norm": 0.4453125, "learning_rate": 0.00015633528265107213, "loss": 0.5311, "step": 1922 }, { "epoch": 1.1246529300014614, "grad_norm": 0.458984375, "learning_rate": 0.00015623781676413255, "loss": 0.5729, "step": 1924 }, { "epoch": 1.1258220078912757, "grad_norm": 0.55078125, "learning_rate": 0.00015614035087719297, "loss": 0.495, "step": 1926 }, { "epoch": 1.12699108578109, "grad_norm": 0.51953125, "learning_rate": 0.00015604288499025342, "loss": 0.5984, "step": 1928 }, { "epoch": 1.1281601636709047, "grad_norm": 0.482421875, "learning_rate": 0.00015594541910331385, "loss": 0.5055, "step": 1930 }, { "epoch": 1.129329241560719, "grad_norm": 0.59765625, "learning_rate": 0.0001558479532163743, "loss": 0.5671, "step": 1932 }, { "epoch": 1.1304983194505334, "grad_norm": 0.5625, "learning_rate": 0.00015575048732943472, "loss": 0.5577, "step": 1934 }, { "epoch": 1.1316673973403477, "grad_norm": 0.4765625, "learning_rate": 0.00015565302144249514, "loss": 0.5637, "step": 1936 }, { "epoch": 1.1328364752301623, "grad_norm": 0.46484375, "learning_rate": 0.00015555555555555556, "loss": 0.5337, "step": 1938 }, { "epoch": 1.1340055531199766, "grad_norm": 0.466796875, "learning_rate": 0.00015545808966861599, "loss": 0.5213, "step": 1940 }, { "epoch": 1.135174631009791, "grad_norm": 0.44921875, "learning_rate": 0.0001553606237816764, "loss": 0.493, "step": 1942 }, { "epoch": 1.1363437088996053, "grad_norm": 0.5078125, "learning_rate": 0.00015526315789473686, "loss": 0.5246, "step": 1944 }, { "epoch": 1.13751278678942, "grad_norm": 0.44921875, "learning_rate": 0.00015516569200779728, "loss": 0.4862, "step": 1946 }, { "epoch": 1.1386818646792343, "grad_norm": 0.50390625, "learning_rate": 0.00015506822612085773, "loss": 0.6714, "step": 1948 }, { "epoch": 1.1398509425690486, "grad_norm": 0.498046875, "learning_rate": 0.00015497076023391815, "loss": 0.57, "step": 1950 }, { "epoch": 1.1410200204588632, "grad_norm": 0.474609375, "learning_rate": 0.00015487329434697857, "loss": 0.488, "step": 1952 }, { "epoch": 1.1421890983486775, "grad_norm": 0.5234375, "learning_rate": 0.000154775828460039, "loss": 0.5453, "step": 1954 }, { "epoch": 1.143358176238492, "grad_norm": 0.51953125, "learning_rate": 0.00015467836257309942, "loss": 0.5405, "step": 1956 }, { "epoch": 1.1445272541283062, "grad_norm": 0.455078125, "learning_rate": 0.00015458089668615984, "loss": 0.4999, "step": 1958 }, { "epoch": 1.1456963320181206, "grad_norm": 0.6015625, "learning_rate": 0.00015448343079922026, "loss": 0.5213, "step": 1960 }, { "epoch": 1.1468654099079352, "grad_norm": 0.51953125, "learning_rate": 0.0001543859649122807, "loss": 0.5285, "step": 1962 }, { "epoch": 1.1480344877977495, "grad_norm": 3.71875, "learning_rate": 0.00015428849902534116, "loss": 0.6403, "step": 1964 }, { "epoch": 1.1492035656875639, "grad_norm": 0.5546875, "learning_rate": 0.00015419103313840158, "loss": 0.5461, "step": 1966 }, { "epoch": 1.1503726435773785, "grad_norm": 0.58203125, "learning_rate": 0.000154093567251462, "loss": 0.581, "step": 1968 }, { "epoch": 1.1515417214671928, "grad_norm": 0.48046875, "learning_rate": 0.00015399610136452243, "loss": 0.5225, "step": 1970 }, { "epoch": 1.1527107993570072, "grad_norm": 0.51171875, "learning_rate": 0.00015389863547758285, "loss": 0.5275, "step": 1972 }, { "epoch": 1.1538798772468215, "grad_norm": 0.478515625, "learning_rate": 0.00015380116959064327, "loss": 0.5758, "step": 1974 }, { "epoch": 1.1550489551366359, "grad_norm": 1.3984375, "learning_rate": 0.0001537037037037037, "loss": 0.7684, "step": 1976 }, { "epoch": 1.1562180330264504, "grad_norm": 1.0234375, "learning_rate": 0.00015360623781676414, "loss": 0.6212, "step": 1978 }, { "epoch": 1.1573871109162648, "grad_norm": 1.09375, "learning_rate": 0.00015350877192982457, "loss": 0.5519, "step": 1980 }, { "epoch": 1.1585561888060791, "grad_norm": 2.21875, "learning_rate": 0.00015341130604288502, "loss": 0.693, "step": 1982 }, { "epoch": 1.1597252666958937, "grad_norm": 0.78515625, "learning_rate": 0.00015331384015594544, "loss": 0.5857, "step": 1984 }, { "epoch": 1.160894344585708, "grad_norm": 0.65625, "learning_rate": 0.00015321637426900586, "loss": 0.5599, "step": 1986 }, { "epoch": 1.1620634224755224, "grad_norm": 0.84375, "learning_rate": 0.00015311890838206628, "loss": 0.5946, "step": 1988 }, { "epoch": 1.1632325003653368, "grad_norm": 0.54296875, "learning_rate": 0.0001530214424951267, "loss": 0.5712, "step": 1990 }, { "epoch": 1.1644015782551513, "grad_norm": 0.6484375, "learning_rate": 0.00015292397660818713, "loss": 0.5825, "step": 1992 }, { "epoch": 1.1655706561449657, "grad_norm": 0.89453125, "learning_rate": 0.00015282651072124755, "loss": 0.5646, "step": 1994 }, { "epoch": 1.16673973403478, "grad_norm": 0.5625, "learning_rate": 0.000152729044834308, "loss": 0.5377, "step": 1996 }, { "epoch": 1.1679088119245944, "grad_norm": 0.5859375, "learning_rate": 0.00015263157894736845, "loss": 0.5101, "step": 1998 }, { "epoch": 1.169077889814409, "grad_norm": 0.51953125, "learning_rate": 0.00015253411306042887, "loss": 0.4811, "step": 2000 }, { "epoch": 1.1702469677042233, "grad_norm": 0.55078125, "learning_rate": 0.0001524366471734893, "loss": 0.5085, "step": 2002 }, { "epoch": 1.1714160455940377, "grad_norm": 0.54296875, "learning_rate": 0.00015233918128654972, "loss": 0.5708, "step": 2004 }, { "epoch": 1.172585123483852, "grad_norm": 0.546875, "learning_rate": 0.00015224171539961014, "loss": 0.5335, "step": 2006 }, { "epoch": 1.1737542013736666, "grad_norm": 0.546875, "learning_rate": 0.00015214424951267056, "loss": 0.6306, "step": 2008 }, { "epoch": 1.174923279263481, "grad_norm": 0.50390625, "learning_rate": 0.00015204678362573098, "loss": 0.5656, "step": 2010 }, { "epoch": 1.1760923571532953, "grad_norm": 0.470703125, "learning_rate": 0.00015194931773879143, "loss": 0.5031, "step": 2012 }, { "epoch": 1.1772614350431096, "grad_norm": 0.50390625, "learning_rate": 0.00015185185185185185, "loss": 0.5342, "step": 2014 }, { "epoch": 1.1784305129329242, "grad_norm": 0.52734375, "learning_rate": 0.0001517543859649123, "loss": 0.5409, "step": 2016 }, { "epoch": 1.1795995908227386, "grad_norm": 0.52734375, "learning_rate": 0.00015165692007797273, "loss": 0.5312, "step": 2018 }, { "epoch": 1.180768668712553, "grad_norm": 0.486328125, "learning_rate": 0.00015155945419103315, "loss": 0.5754, "step": 2020 }, { "epoch": 1.1819377466023675, "grad_norm": 1.015625, "learning_rate": 0.00015146198830409357, "loss": 0.648, "step": 2022 }, { "epoch": 1.1831068244921819, "grad_norm": 0.82421875, "learning_rate": 0.000151364522417154, "loss": 0.6954, "step": 2024 }, { "epoch": 1.1842759023819962, "grad_norm": 0.78515625, "learning_rate": 0.00015126705653021442, "loss": 0.5343, "step": 2026 }, { "epoch": 1.1854449802718106, "grad_norm": 0.62109375, "learning_rate": 0.00015116959064327486, "loss": 0.4973, "step": 2028 }, { "epoch": 1.186614058161625, "grad_norm": 0.73046875, "learning_rate": 0.0001510721247563353, "loss": 0.5654, "step": 2030 }, { "epoch": 1.1877831360514395, "grad_norm": 0.5234375, "learning_rate": 0.00015097465886939574, "loss": 0.557, "step": 2032 }, { "epoch": 1.1889522139412538, "grad_norm": 0.5546875, "learning_rate": 0.00015087719298245616, "loss": 0.5554, "step": 2034 }, { "epoch": 1.1901212918310682, "grad_norm": 0.51171875, "learning_rate": 0.00015077972709551658, "loss": 0.5727, "step": 2036 }, { "epoch": 1.1912903697208828, "grad_norm": 0.49609375, "learning_rate": 0.000150682261208577, "loss": 0.5447, "step": 2038 }, { "epoch": 1.192459447610697, "grad_norm": 0.494140625, "learning_rate": 0.00015058479532163743, "loss": 0.5653, "step": 2040 }, { "epoch": 1.1936285255005115, "grad_norm": 0.458984375, "learning_rate": 0.00015048732943469785, "loss": 0.5057, "step": 2042 }, { "epoch": 1.1947976033903258, "grad_norm": 0.53125, "learning_rate": 0.0001503898635477583, "loss": 0.5383, "step": 2044 }, { "epoch": 1.1959666812801402, "grad_norm": 0.498046875, "learning_rate": 0.00015029239766081872, "loss": 0.5763, "step": 2046 }, { "epoch": 1.1971357591699547, "grad_norm": 0.4921875, "learning_rate": 0.00015019493177387914, "loss": 0.5312, "step": 2048 }, { "epoch": 1.198304837059769, "grad_norm": 0.45703125, "learning_rate": 0.0001500974658869396, "loss": 0.5504, "step": 2050 }, { "epoch": 1.1994739149495834, "grad_norm": 0.44921875, "learning_rate": 0.00015000000000000001, "loss": 0.5234, "step": 2052 }, { "epoch": 1.200642992839398, "grad_norm": 0.474609375, "learning_rate": 0.00014990253411306044, "loss": 0.5458, "step": 2054 }, { "epoch": 1.2018120707292124, "grad_norm": 0.46875, "learning_rate": 0.00014980506822612086, "loss": 0.5399, "step": 2056 }, { "epoch": 1.2029811486190267, "grad_norm": 0.48046875, "learning_rate": 0.00014970760233918128, "loss": 0.5323, "step": 2058 }, { "epoch": 1.204150226508841, "grad_norm": 0.51953125, "learning_rate": 0.0001496101364522417, "loss": 0.5314, "step": 2060 }, { "epoch": 1.2053193043986556, "grad_norm": 0.453125, "learning_rate": 0.00014951267056530215, "loss": 0.4901, "step": 2062 }, { "epoch": 1.20648838228847, "grad_norm": 0.46875, "learning_rate": 0.00014941520467836257, "loss": 0.5305, "step": 2064 }, { "epoch": 1.2076574601782843, "grad_norm": 0.453125, "learning_rate": 0.00014931773879142302, "loss": 0.4642, "step": 2066 }, { "epoch": 1.2088265380680987, "grad_norm": 0.515625, "learning_rate": 0.00014922027290448345, "loss": 0.508, "step": 2068 }, { "epoch": 1.2099956159579133, "grad_norm": 0.46484375, "learning_rate": 0.00014912280701754387, "loss": 0.5329, "step": 2070 }, { "epoch": 1.2111646938477276, "grad_norm": 0.482421875, "learning_rate": 0.0001490253411306043, "loss": 0.5502, "step": 2072 }, { "epoch": 1.212333771737542, "grad_norm": 0.451171875, "learning_rate": 0.0001489278752436647, "loss": 0.5335, "step": 2074 }, { "epoch": 1.2135028496273563, "grad_norm": 0.5078125, "learning_rate": 0.00014883040935672514, "loss": 0.5298, "step": 2076 }, { "epoch": 1.214671927517171, "grad_norm": 0.5078125, "learning_rate": 0.00014873294346978559, "loss": 0.5271, "step": 2078 }, { "epoch": 1.2158410054069853, "grad_norm": 0.57421875, "learning_rate": 0.000148635477582846, "loss": 0.5326, "step": 2080 }, { "epoch": 1.2170100832967996, "grad_norm": 0.515625, "learning_rate": 0.00014853801169590643, "loss": 0.5556, "step": 2082 }, { "epoch": 1.218179161186614, "grad_norm": 0.46484375, "learning_rate": 0.00014844054580896688, "loss": 0.4727, "step": 2084 }, { "epoch": 1.2193482390764285, "grad_norm": 0.470703125, "learning_rate": 0.0001483430799220273, "loss": 0.5172, "step": 2086 }, { "epoch": 1.2205173169662429, "grad_norm": 0.54296875, "learning_rate": 0.00014824561403508772, "loss": 0.5698, "step": 2088 }, { "epoch": 1.2216863948560572, "grad_norm": 0.5078125, "learning_rate": 0.00014814814814814815, "loss": 0.6561, "step": 2090 }, { "epoch": 1.2228554727458718, "grad_norm": 0.515625, "learning_rate": 0.00014805068226120857, "loss": 0.5546, "step": 2092 }, { "epoch": 1.2240245506356862, "grad_norm": 0.53515625, "learning_rate": 0.00014795321637426902, "loss": 0.5504, "step": 2094 }, { "epoch": 1.2251936285255005, "grad_norm": 0.5546875, "learning_rate": 0.00014785575048732944, "loss": 0.5061, "step": 2096 }, { "epoch": 1.2263627064153149, "grad_norm": 0.48828125, "learning_rate": 0.00014775828460038986, "loss": 0.5257, "step": 2098 }, { "epoch": 1.2275317843051292, "grad_norm": 0.49609375, "learning_rate": 0.0001476608187134503, "loss": 0.5323, "step": 2100 }, { "epoch": 1.2287008621949438, "grad_norm": 0.46875, "learning_rate": 0.00014756335282651073, "loss": 0.5322, "step": 2102 }, { "epoch": 1.2298699400847581, "grad_norm": 0.51953125, "learning_rate": 0.00014746588693957116, "loss": 0.5758, "step": 2104 }, { "epoch": 1.2310390179745725, "grad_norm": 0.482421875, "learning_rate": 0.00014736842105263158, "loss": 0.5396, "step": 2106 }, { "epoch": 1.232208095864387, "grad_norm": 0.486328125, "learning_rate": 0.000147270955165692, "loss": 0.5596, "step": 2108 }, { "epoch": 1.2333771737542014, "grad_norm": 0.5078125, "learning_rate": 0.00014717348927875245, "loss": 0.5493, "step": 2110 }, { "epoch": 1.2345462516440158, "grad_norm": 0.478515625, "learning_rate": 0.00014707602339181287, "loss": 0.5155, "step": 2112 }, { "epoch": 1.2357153295338301, "grad_norm": 0.57421875, "learning_rate": 0.0001469785575048733, "loss": 0.5305, "step": 2114 }, { "epoch": 1.2368844074236447, "grad_norm": 0.5078125, "learning_rate": 0.00014688109161793372, "loss": 0.6194, "step": 2116 }, { "epoch": 1.238053485313459, "grad_norm": 0.77734375, "learning_rate": 0.00014678362573099417, "loss": 0.5856, "step": 2118 }, { "epoch": 1.2392225632032734, "grad_norm": 0.546875, "learning_rate": 0.0001466861598440546, "loss": 0.5392, "step": 2120 }, { "epoch": 1.2403916410930877, "grad_norm": 0.5234375, "learning_rate": 0.000146588693957115, "loss": 0.5132, "step": 2122 }, { "epoch": 1.2415607189829023, "grad_norm": 0.5234375, "learning_rate": 0.00014649122807017543, "loss": 0.5524, "step": 2124 }, { "epoch": 1.2427297968727167, "grad_norm": 0.58984375, "learning_rate": 0.00014639376218323588, "loss": 0.5383, "step": 2126 }, { "epoch": 1.243898874762531, "grad_norm": 0.55859375, "learning_rate": 0.0001462962962962963, "loss": 0.548, "step": 2128 }, { "epoch": 1.2450679526523454, "grad_norm": 0.51171875, "learning_rate": 0.00014619883040935673, "loss": 0.5194, "step": 2130 }, { "epoch": 1.24623703054216, "grad_norm": 0.45703125, "learning_rate": 0.00014610136452241715, "loss": 0.5007, "step": 2132 }, { "epoch": 1.2474061084319743, "grad_norm": 0.453125, "learning_rate": 0.0001460038986354776, "loss": 0.5092, "step": 2134 }, { "epoch": 1.2485751863217887, "grad_norm": 0.94921875, "learning_rate": 0.00014590643274853802, "loss": 0.5948, "step": 2136 }, { "epoch": 1.249744264211603, "grad_norm": 0.58203125, "learning_rate": 0.00014580896686159844, "loss": 0.5545, "step": 2138 }, { "epoch": 1.2509133421014176, "grad_norm": 0.56640625, "learning_rate": 0.00014571150097465887, "loss": 0.5342, "step": 2140 }, { "epoch": 1.252082419991232, "grad_norm": 1.2890625, "learning_rate": 0.00014561403508771932, "loss": 0.616, "step": 2142 }, { "epoch": 1.2532514978810463, "grad_norm": 0.5546875, "learning_rate": 0.00014551656920077974, "loss": 0.5408, "step": 2144 }, { "epoch": 1.2544205757708609, "grad_norm": 0.625, "learning_rate": 0.00014541910331384016, "loss": 0.5604, "step": 2146 }, { "epoch": 1.2555896536606752, "grad_norm": 0.484375, "learning_rate": 0.00014532163742690058, "loss": 0.5168, "step": 2148 }, { "epoch": 1.2567587315504896, "grad_norm": 0.59375, "learning_rate": 0.000145224171539961, "loss": 0.5451, "step": 2150 }, { "epoch": 1.257927809440304, "grad_norm": 0.49609375, "learning_rate": 0.00014512670565302145, "loss": 0.5365, "step": 2152 }, { "epoch": 1.2590968873301183, "grad_norm": 0.55859375, "learning_rate": 0.00014502923976608188, "loss": 0.5516, "step": 2154 }, { "epoch": 1.2602659652199328, "grad_norm": 0.5390625, "learning_rate": 0.0001449317738791423, "loss": 0.6553, "step": 2156 }, { "epoch": 1.2614350431097472, "grad_norm": 0.55859375, "learning_rate": 0.00014483430799220275, "loss": 0.5779, "step": 2158 }, { "epoch": 1.2626041209995615, "grad_norm": 0.60546875, "learning_rate": 0.00014473684210526317, "loss": 0.5469, "step": 2160 }, { "epoch": 1.2637731988893761, "grad_norm": 0.515625, "learning_rate": 0.0001446393762183236, "loss": 0.5338, "step": 2162 }, { "epoch": 1.2649422767791905, "grad_norm": 0.48828125, "learning_rate": 0.00014454191033138402, "loss": 0.4948, "step": 2164 }, { "epoch": 1.2661113546690048, "grad_norm": 0.53125, "learning_rate": 0.00014444444444444444, "loss": 0.5273, "step": 2166 }, { "epoch": 1.2672804325588192, "grad_norm": 0.484375, "learning_rate": 0.0001443469785575049, "loss": 0.5314, "step": 2168 }, { "epoch": 1.2684495104486335, "grad_norm": 0.482421875, "learning_rate": 0.0001442495126705653, "loss": 0.5253, "step": 2170 }, { "epoch": 1.269618588338448, "grad_norm": 0.49609375, "learning_rate": 0.00014415204678362573, "loss": 0.6898, "step": 2172 }, { "epoch": 1.2707876662282624, "grad_norm": 0.5234375, "learning_rate": 0.00014405458089668618, "loss": 0.4904, "step": 2174 }, { "epoch": 1.2719567441180768, "grad_norm": 0.515625, "learning_rate": 0.0001439571150097466, "loss": 0.5021, "step": 2176 }, { "epoch": 1.2731258220078914, "grad_norm": 0.50390625, "learning_rate": 0.00014385964912280703, "loss": 0.5574, "step": 2178 }, { "epoch": 1.2742948998977057, "grad_norm": 0.5234375, "learning_rate": 0.00014376218323586745, "loss": 0.563, "step": 2180 }, { "epoch": 1.27546397778752, "grad_norm": 0.47265625, "learning_rate": 0.00014366471734892787, "loss": 0.5243, "step": 2182 }, { "epoch": 1.2766330556773344, "grad_norm": 0.4609375, "learning_rate": 0.0001435672514619883, "loss": 0.596, "step": 2184 }, { "epoch": 1.2778021335671488, "grad_norm": 0.6796875, "learning_rate": 0.00014346978557504874, "loss": 0.5895, "step": 2186 }, { "epoch": 1.2789712114569634, "grad_norm": 0.63671875, "learning_rate": 0.00014337231968810916, "loss": 0.4975, "step": 2188 }, { "epoch": 1.2801402893467777, "grad_norm": 0.55859375, "learning_rate": 0.00014327485380116961, "loss": 0.5065, "step": 2190 }, { "epoch": 1.281309367236592, "grad_norm": 0.6328125, "learning_rate": 0.00014317738791423004, "loss": 0.4677, "step": 2192 }, { "epoch": 1.2824784451264066, "grad_norm": 0.609375, "learning_rate": 0.00014307992202729046, "loss": 0.5402, "step": 2194 }, { "epoch": 1.283647523016221, "grad_norm": 0.49609375, "learning_rate": 0.00014298245614035088, "loss": 0.5598, "step": 2196 }, { "epoch": 1.2848166009060353, "grad_norm": 0.53125, "learning_rate": 0.0001428849902534113, "loss": 0.5707, "step": 2198 }, { "epoch": 1.28598567879585, "grad_norm": 0.57421875, "learning_rate": 0.00014278752436647173, "loss": 0.5603, "step": 2200 }, { "epoch": 1.2871547566856643, "grad_norm": 0.51171875, "learning_rate": 0.00014269005847953217, "loss": 0.494, "step": 2202 }, { "epoch": 1.2883238345754786, "grad_norm": 0.5234375, "learning_rate": 0.0001425925925925926, "loss": 0.5525, "step": 2204 }, { "epoch": 1.289492912465293, "grad_norm": 0.49609375, "learning_rate": 0.00014249512670565305, "loss": 0.5143, "step": 2206 }, { "epoch": 1.2906619903551073, "grad_norm": 0.5390625, "learning_rate": 0.00014239766081871347, "loss": 0.5176, "step": 2208 }, { "epoch": 1.2918310682449219, "grad_norm": 0.5234375, "learning_rate": 0.0001423001949317739, "loss": 0.5043, "step": 2210 }, { "epoch": 1.2930001461347362, "grad_norm": 0.6953125, "learning_rate": 0.0001422027290448343, "loss": 0.571, "step": 2212 }, { "epoch": 1.2941692240245506, "grad_norm": 0.51171875, "learning_rate": 0.00014210526315789474, "loss": 0.5369, "step": 2214 }, { "epoch": 1.2953383019143652, "grad_norm": 0.458984375, "learning_rate": 0.00014200779727095516, "loss": 0.5103, "step": 2216 }, { "epoch": 1.2965073798041795, "grad_norm": 0.46875, "learning_rate": 0.00014191033138401558, "loss": 0.4751, "step": 2218 }, { "epoch": 1.2976764576939939, "grad_norm": 0.48828125, "learning_rate": 0.00014181286549707603, "loss": 0.5009, "step": 2220 }, { "epoch": 1.2988455355838082, "grad_norm": 0.435546875, "learning_rate": 0.00014171539961013648, "loss": 0.5003, "step": 2222 }, { "epoch": 1.3000146134736226, "grad_norm": 0.53125, "learning_rate": 0.0001416179337231969, "loss": 0.4975, "step": 2224 }, { "epoch": 1.3011836913634371, "grad_norm": 0.416015625, "learning_rate": 0.00014152046783625732, "loss": 0.4553, "step": 2226 }, { "epoch": 1.3023527692532515, "grad_norm": 0.49609375, "learning_rate": 0.00014142300194931775, "loss": 0.5384, "step": 2228 }, { "epoch": 1.3035218471430658, "grad_norm": 0.439453125, "learning_rate": 0.00014132553606237817, "loss": 0.4887, "step": 2230 }, { "epoch": 1.3046909250328804, "grad_norm": 0.490234375, "learning_rate": 0.0001412280701754386, "loss": 0.4916, "step": 2232 }, { "epoch": 1.3058600029226948, "grad_norm": 0.50390625, "learning_rate": 0.000141130604288499, "loss": 0.5218, "step": 2234 }, { "epoch": 1.3070290808125091, "grad_norm": 0.51953125, "learning_rate": 0.00014103313840155946, "loss": 0.5185, "step": 2236 }, { "epoch": 1.3081981587023235, "grad_norm": 0.47265625, "learning_rate": 0.0001409356725146199, "loss": 0.5197, "step": 2238 }, { "epoch": 1.3093672365921378, "grad_norm": 0.447265625, "learning_rate": 0.00014083820662768033, "loss": 0.4974, "step": 2240 }, { "epoch": 1.3105363144819524, "grad_norm": 0.486328125, "learning_rate": 0.00014074074074074076, "loss": 0.5262, "step": 2242 }, { "epoch": 1.3117053923717668, "grad_norm": 0.478515625, "learning_rate": 0.00014064327485380118, "loss": 0.5682, "step": 2244 }, { "epoch": 1.312874470261581, "grad_norm": 1.21875, "learning_rate": 0.0001405458089668616, "loss": 0.5945, "step": 2246 }, { "epoch": 1.3140435481513957, "grad_norm": 1.125, "learning_rate": 0.00014044834307992202, "loss": 0.5376, "step": 2248 }, { "epoch": 1.31521262604121, "grad_norm": 0.58203125, "learning_rate": 0.00014035087719298245, "loss": 0.5386, "step": 2250 }, { "epoch": 1.3163817039310244, "grad_norm": 0.51171875, "learning_rate": 0.00014025341130604287, "loss": 0.5174, "step": 2252 }, { "epoch": 1.3175507818208387, "grad_norm": 0.671875, "learning_rate": 0.00014015594541910332, "loss": 0.5459, "step": 2254 }, { "epoch": 1.318719859710653, "grad_norm": 0.474609375, "learning_rate": 0.00014005847953216377, "loss": 0.5424, "step": 2256 }, { "epoch": 1.3198889376004677, "grad_norm": 0.50390625, "learning_rate": 0.0001399610136452242, "loss": 0.5088, "step": 2258 }, { "epoch": 1.321058015490282, "grad_norm": 0.455078125, "learning_rate": 0.0001398635477582846, "loss": 0.5042, "step": 2260 }, { "epoch": 1.3222270933800964, "grad_norm": 0.51953125, "learning_rate": 0.00013976608187134503, "loss": 0.5704, "step": 2262 }, { "epoch": 1.323396171269911, "grad_norm": 0.494140625, "learning_rate": 0.00013966861598440546, "loss": 0.5154, "step": 2264 }, { "epoch": 1.3245652491597253, "grad_norm": 0.55859375, "learning_rate": 0.00013957115009746588, "loss": 0.5377, "step": 2266 }, { "epoch": 1.3257343270495396, "grad_norm": 0.67578125, "learning_rate": 0.0001394736842105263, "loss": 0.5435, "step": 2268 }, { "epoch": 1.3269034049393542, "grad_norm": 0.474609375, "learning_rate": 0.00013937621832358675, "loss": 0.4831, "step": 2270 }, { "epoch": 1.3280724828291686, "grad_norm": 0.72265625, "learning_rate": 0.0001392787524366472, "loss": 0.4896, "step": 2272 }, { "epoch": 1.329241560718983, "grad_norm": 2.96875, "learning_rate": 0.00013918128654970762, "loss": 0.6227, "step": 2274 }, { "epoch": 1.3304106386087973, "grad_norm": 0.4609375, "learning_rate": 0.00013908382066276804, "loss": 0.5034, "step": 2276 }, { "epoch": 1.3315797164986116, "grad_norm": 0.484375, "learning_rate": 0.00013898635477582847, "loss": 0.5445, "step": 2278 }, { "epoch": 1.3327487943884262, "grad_norm": 0.5625, "learning_rate": 0.0001388888888888889, "loss": 0.5254, "step": 2280 }, { "epoch": 1.3339178722782405, "grad_norm": 0.44140625, "learning_rate": 0.0001387914230019493, "loss": 0.5563, "step": 2282 }, { "epoch": 1.335086950168055, "grad_norm": 0.4921875, "learning_rate": 0.00013869395711500973, "loss": 0.5262, "step": 2284 }, { "epoch": 1.3362560280578695, "grad_norm": 0.431640625, "learning_rate": 0.00013859649122807018, "loss": 0.4994, "step": 2286 }, { "epoch": 1.3374251059476838, "grad_norm": 0.484375, "learning_rate": 0.0001384990253411306, "loss": 0.5218, "step": 2288 }, { "epoch": 1.3385941838374982, "grad_norm": 0.5234375, "learning_rate": 0.00013840155945419105, "loss": 0.5622, "step": 2290 }, { "epoch": 1.3397632617273125, "grad_norm": 0.4453125, "learning_rate": 0.00013830409356725148, "loss": 0.5173, "step": 2292 }, { "epoch": 1.3409323396171269, "grad_norm": 0.458984375, "learning_rate": 0.0001382066276803119, "loss": 0.496, "step": 2294 }, { "epoch": 1.3421014175069415, "grad_norm": 0.51171875, "learning_rate": 0.00013810916179337232, "loss": 0.5524, "step": 2296 }, { "epoch": 1.3432704953967558, "grad_norm": 0.51171875, "learning_rate": 0.00013801169590643274, "loss": 0.4955, "step": 2298 }, { "epoch": 1.3444395732865702, "grad_norm": 0.435546875, "learning_rate": 0.00013791423001949317, "loss": 0.5226, "step": 2300 }, { "epoch": 1.3456086511763847, "grad_norm": 0.490234375, "learning_rate": 0.00013781676413255362, "loss": 0.5202, "step": 2302 }, { "epoch": 1.346777729066199, "grad_norm": 0.96484375, "learning_rate": 0.00013771929824561404, "loss": 0.801, "step": 2304 }, { "epoch": 1.3479468069560134, "grad_norm": 0.7265625, "learning_rate": 0.0001376218323586745, "loss": 0.5402, "step": 2306 }, { "epoch": 1.3491158848458278, "grad_norm": 0.5078125, "learning_rate": 0.0001375243664717349, "loss": 0.5445, "step": 2308 }, { "epoch": 1.3502849627356421, "grad_norm": 0.6484375, "learning_rate": 0.00013742690058479533, "loss": 0.5235, "step": 2310 }, { "epoch": 1.3514540406254567, "grad_norm": 0.458984375, "learning_rate": 0.00013732943469785575, "loss": 0.5099, "step": 2312 }, { "epoch": 1.352623118515271, "grad_norm": 0.5078125, "learning_rate": 0.00013723196881091618, "loss": 0.4943, "step": 2314 }, { "epoch": 1.3537921964050854, "grad_norm": 0.5078125, "learning_rate": 0.0001371345029239766, "loss": 0.5101, "step": 2316 }, { "epoch": 1.3549612742949, "grad_norm": 0.48828125, "learning_rate": 0.00013703703703703705, "loss": 0.5452, "step": 2318 }, { "epoch": 1.3561303521847143, "grad_norm": 0.5, "learning_rate": 0.00013693957115009747, "loss": 0.532, "step": 2320 }, { "epoch": 1.3572994300745287, "grad_norm": 0.53125, "learning_rate": 0.0001368421052631579, "loss": 0.5339, "step": 2322 }, { "epoch": 1.358468507964343, "grad_norm": 0.86328125, "learning_rate": 0.00013674463937621834, "loss": 0.6001, "step": 2324 }, { "epoch": 1.3596375858541576, "grad_norm": 0.59765625, "learning_rate": 0.00013664717348927876, "loss": 0.5809, "step": 2326 }, { "epoch": 1.360806663743972, "grad_norm": 0.59765625, "learning_rate": 0.0001365497076023392, "loss": 0.4819, "step": 2328 }, { "epoch": 1.3619757416337863, "grad_norm": 0.62890625, "learning_rate": 0.0001364522417153996, "loss": 0.5462, "step": 2330 }, { "epoch": 1.3631448195236007, "grad_norm": 0.515625, "learning_rate": 0.00013635477582846003, "loss": 0.5619, "step": 2332 }, { "epoch": 1.3643138974134152, "grad_norm": 0.52734375, "learning_rate": 0.00013625730994152048, "loss": 0.514, "step": 2334 }, { "epoch": 1.3654829753032296, "grad_norm": 0.46484375, "learning_rate": 0.0001361598440545809, "loss": 0.5458, "step": 2336 }, { "epoch": 1.366652053193044, "grad_norm": 0.53515625, "learning_rate": 0.00013606237816764133, "loss": 0.5545, "step": 2338 }, { "epoch": 1.3678211310828585, "grad_norm": 0.5234375, "learning_rate": 0.00013596491228070177, "loss": 0.5931, "step": 2340 }, { "epoch": 1.3689902089726729, "grad_norm": 0.4453125, "learning_rate": 0.0001358674463937622, "loss": 0.553, "step": 2342 }, { "epoch": 1.3701592868624872, "grad_norm": 0.72265625, "learning_rate": 0.00013576998050682262, "loss": 0.5348, "step": 2344 }, { "epoch": 1.3713283647523016, "grad_norm": 0.53515625, "learning_rate": 0.00013567251461988304, "loss": 0.584, "step": 2346 }, { "epoch": 1.372497442642116, "grad_norm": 0.423828125, "learning_rate": 0.00013557504873294346, "loss": 0.523, "step": 2348 }, { "epoch": 1.3736665205319305, "grad_norm": 0.68359375, "learning_rate": 0.0001354775828460039, "loss": 0.5684, "step": 2350 }, { "epoch": 1.3748355984217449, "grad_norm": 0.9609375, "learning_rate": 0.00013538011695906434, "loss": 0.5319, "step": 2352 }, { "epoch": 1.3760046763115592, "grad_norm": 0.58984375, "learning_rate": 0.00013528265107212476, "loss": 0.5758, "step": 2354 }, { "epoch": 1.3771737542013738, "grad_norm": 0.458984375, "learning_rate": 0.00013518518518518518, "loss": 0.4881, "step": 2356 }, { "epoch": 1.3783428320911881, "grad_norm": 0.57421875, "learning_rate": 0.00013508771929824563, "loss": 0.5436, "step": 2358 }, { "epoch": 1.3795119099810025, "grad_norm": 0.6328125, "learning_rate": 0.00013499025341130605, "loss": 0.6471, "step": 2360 }, { "epoch": 1.3806809878708168, "grad_norm": 0.5, "learning_rate": 0.00013489278752436647, "loss": 0.5255, "step": 2362 }, { "epoch": 1.3818500657606312, "grad_norm": 0.51171875, "learning_rate": 0.0001347953216374269, "loss": 0.5235, "step": 2364 }, { "epoch": 1.3830191436504458, "grad_norm": 0.48046875, "learning_rate": 0.00013469785575048735, "loss": 0.4981, "step": 2366 }, { "epoch": 1.38418822154026, "grad_norm": 0.51171875, "learning_rate": 0.00013460038986354777, "loss": 0.4869, "step": 2368 }, { "epoch": 1.3853572994300745, "grad_norm": 0.5234375, "learning_rate": 0.0001345029239766082, "loss": 0.4993, "step": 2370 }, { "epoch": 1.386526377319889, "grad_norm": 0.4609375, "learning_rate": 0.0001344054580896686, "loss": 0.4922, "step": 2372 }, { "epoch": 1.3876954552097034, "grad_norm": 1.171875, "learning_rate": 0.00013430799220272906, "loss": 0.653, "step": 2374 }, { "epoch": 1.3888645330995177, "grad_norm": 0.458984375, "learning_rate": 0.00013421052631578948, "loss": 0.5303, "step": 2376 }, { "epoch": 1.390033610989332, "grad_norm": 0.5078125, "learning_rate": 0.0001341130604288499, "loss": 0.5234, "step": 2378 }, { "epoch": 1.3912026888791464, "grad_norm": 0.515625, "learning_rate": 0.00013401559454191033, "loss": 0.5408, "step": 2380 }, { "epoch": 1.392371766768961, "grad_norm": 0.625, "learning_rate": 0.00013391812865497078, "loss": 0.5112, "step": 2382 }, { "epoch": 1.3935408446587754, "grad_norm": 0.478515625, "learning_rate": 0.0001338206627680312, "loss": 0.541, "step": 2384 }, { "epoch": 1.3947099225485897, "grad_norm": 0.4296875, "learning_rate": 0.00013372319688109162, "loss": 0.5173, "step": 2386 }, { "epoch": 1.3958790004384043, "grad_norm": 0.49609375, "learning_rate": 0.00013362573099415205, "loss": 0.4972, "step": 2388 }, { "epoch": 1.3970480783282186, "grad_norm": 0.60546875, "learning_rate": 0.00013352826510721247, "loss": 0.5849, "step": 2390 }, { "epoch": 1.398217156218033, "grad_norm": 0.439453125, "learning_rate": 0.00013343079922027292, "loss": 0.4755, "step": 2392 }, { "epoch": 1.3993862341078476, "grad_norm": 0.48828125, "learning_rate": 0.00013333333333333334, "loss": 0.4992, "step": 2394 }, { "epoch": 1.400555311997662, "grad_norm": 0.474609375, "learning_rate": 0.00013323586744639376, "loss": 0.4972, "step": 2396 }, { "epoch": 1.4017243898874763, "grad_norm": 0.44921875, "learning_rate": 0.0001331384015594542, "loss": 0.517, "step": 2398 }, { "epoch": 1.4028934677772906, "grad_norm": 0.439453125, "learning_rate": 0.00013304093567251463, "loss": 0.5337, "step": 2400 }, { "epoch": 1.404062545667105, "grad_norm": 0.55859375, "learning_rate": 0.00013294346978557506, "loss": 0.6145, "step": 2402 }, { "epoch": 1.4052316235569196, "grad_norm": 0.5546875, "learning_rate": 0.00013284600389863548, "loss": 0.529, "step": 2404 }, { "epoch": 1.406400701446734, "grad_norm": 0.56640625, "learning_rate": 0.0001327485380116959, "loss": 0.5566, "step": 2406 }, { "epoch": 1.4075697793365483, "grad_norm": 0.486328125, "learning_rate": 0.00013265107212475635, "loss": 0.5198, "step": 2408 }, { "epoch": 1.4087388572263628, "grad_norm": 0.48828125, "learning_rate": 0.00013255360623781677, "loss": 0.4804, "step": 2410 }, { "epoch": 1.4099079351161772, "grad_norm": 0.5078125, "learning_rate": 0.0001324561403508772, "loss": 0.4942, "step": 2412 }, { "epoch": 1.4110770130059915, "grad_norm": 0.50390625, "learning_rate": 0.00013235867446393764, "loss": 0.5272, "step": 2414 }, { "epoch": 1.4122460908958059, "grad_norm": 0.55859375, "learning_rate": 0.00013226120857699807, "loss": 0.5051, "step": 2416 }, { "epoch": 1.4134151687856202, "grad_norm": 0.50390625, "learning_rate": 0.0001321637426900585, "loss": 0.5318, "step": 2418 }, { "epoch": 1.4145842466754348, "grad_norm": 0.462890625, "learning_rate": 0.0001320662768031189, "loss": 0.5298, "step": 2420 }, { "epoch": 1.4157533245652492, "grad_norm": 0.5234375, "learning_rate": 0.00013196881091617933, "loss": 0.5054, "step": 2422 }, { "epoch": 1.4169224024550635, "grad_norm": 0.8359375, "learning_rate": 0.00013187134502923976, "loss": 0.5184, "step": 2424 }, { "epoch": 1.418091480344878, "grad_norm": 0.51171875, "learning_rate": 0.0001317738791423002, "loss": 0.5057, "step": 2426 }, { "epoch": 1.4192605582346924, "grad_norm": 0.466796875, "learning_rate": 0.00013167641325536063, "loss": 0.5926, "step": 2428 }, { "epoch": 1.4204296361245068, "grad_norm": 0.5234375, "learning_rate": 0.00013157894736842108, "loss": 0.6182, "step": 2430 }, { "epoch": 1.4215987140143211, "grad_norm": 0.53125, "learning_rate": 0.0001314814814814815, "loss": 0.5316, "step": 2432 }, { "epoch": 1.4227677919041355, "grad_norm": 0.51171875, "learning_rate": 0.00013138401559454192, "loss": 0.5063, "step": 2434 }, { "epoch": 1.42393686979395, "grad_norm": 0.7265625, "learning_rate": 0.00013128654970760234, "loss": 0.4917, "step": 2436 }, { "epoch": 1.4251059476837644, "grad_norm": 0.5546875, "learning_rate": 0.00013118908382066277, "loss": 0.5486, "step": 2438 }, { "epoch": 1.4262750255735788, "grad_norm": 0.5234375, "learning_rate": 0.0001310916179337232, "loss": 0.4851, "step": 2440 }, { "epoch": 1.4274441034633933, "grad_norm": 0.46484375, "learning_rate": 0.00013099415204678364, "loss": 0.5135, "step": 2442 }, { "epoch": 1.4286131813532077, "grad_norm": 0.482421875, "learning_rate": 0.00013089668615984406, "loss": 0.483, "step": 2444 }, { "epoch": 1.429782259243022, "grad_norm": 0.5703125, "learning_rate": 0.0001307992202729045, "loss": 0.5652, "step": 2446 }, { "epoch": 1.4309513371328364, "grad_norm": 0.4921875, "learning_rate": 0.00013070175438596493, "loss": 0.5252, "step": 2448 }, { "epoch": 1.4321204150226508, "grad_norm": 0.447265625, "learning_rate": 0.00013060428849902535, "loss": 0.5489, "step": 2450 }, { "epoch": 1.4332894929124653, "grad_norm": 0.484375, "learning_rate": 0.00013050682261208578, "loss": 0.5033, "step": 2452 }, { "epoch": 1.4344585708022797, "grad_norm": 0.490234375, "learning_rate": 0.0001304093567251462, "loss": 0.5016, "step": 2454 }, { "epoch": 1.435627648692094, "grad_norm": 0.4921875, "learning_rate": 0.00013031189083820662, "loss": 0.4923, "step": 2456 }, { "epoch": 1.4367967265819086, "grad_norm": 0.52734375, "learning_rate": 0.00013021442495126704, "loss": 0.52, "step": 2458 }, { "epoch": 1.437965804471723, "grad_norm": 0.451171875, "learning_rate": 0.0001301169590643275, "loss": 0.513, "step": 2460 }, { "epoch": 1.4391348823615373, "grad_norm": 0.4296875, "learning_rate": 0.00013001949317738794, "loss": 0.4825, "step": 2462 }, { "epoch": 1.4403039602513519, "grad_norm": 0.46484375, "learning_rate": 0.00012992202729044836, "loss": 0.54, "step": 2464 }, { "epoch": 1.4414730381411662, "grad_norm": 0.443359375, "learning_rate": 0.0001298245614035088, "loss": 0.5213, "step": 2466 }, { "epoch": 1.4426421160309806, "grad_norm": 0.52734375, "learning_rate": 0.0001297270955165692, "loss": 0.4893, "step": 2468 }, { "epoch": 1.443811193920795, "grad_norm": 0.455078125, "learning_rate": 0.00012962962962962963, "loss": 0.501, "step": 2470 }, { "epoch": 1.4449802718106093, "grad_norm": 0.515625, "learning_rate": 0.00012953216374269005, "loss": 0.5218, "step": 2472 }, { "epoch": 1.4461493497004239, "grad_norm": 1.0625, "learning_rate": 0.00012943469785575048, "loss": 0.5996, "step": 2474 }, { "epoch": 1.4473184275902382, "grad_norm": 0.4609375, "learning_rate": 0.00012933723196881093, "loss": 0.5197, "step": 2476 }, { "epoch": 1.4484875054800526, "grad_norm": 0.474609375, "learning_rate": 0.00012923976608187135, "loss": 0.5161, "step": 2478 }, { "epoch": 1.4496565833698671, "grad_norm": 0.447265625, "learning_rate": 0.0001291423001949318, "loss": 0.5125, "step": 2480 }, { "epoch": 1.4508256612596815, "grad_norm": 0.455078125, "learning_rate": 0.00012904483430799222, "loss": 0.4591, "step": 2482 }, { "epoch": 1.4519947391494958, "grad_norm": 0.5703125, "learning_rate": 0.00012894736842105264, "loss": 0.5067, "step": 2484 }, { "epoch": 1.4531638170393102, "grad_norm": 0.466796875, "learning_rate": 0.00012884990253411306, "loss": 0.5034, "step": 2486 }, { "epoch": 1.4543328949291245, "grad_norm": 0.9765625, "learning_rate": 0.00012875243664717349, "loss": 0.6573, "step": 2488 }, { "epoch": 1.4555019728189391, "grad_norm": 0.53515625, "learning_rate": 0.0001286549707602339, "loss": 0.5001, "step": 2490 }, { "epoch": 1.4566710507087535, "grad_norm": 0.47265625, "learning_rate": 0.00012855750487329433, "loss": 0.4992, "step": 2492 }, { "epoch": 1.4578401285985678, "grad_norm": 0.6875, "learning_rate": 0.00012846003898635478, "loss": 0.5203, "step": 2494 }, { "epoch": 1.4590092064883824, "grad_norm": 0.46484375, "learning_rate": 0.00012836257309941523, "loss": 0.5263, "step": 2496 }, { "epoch": 1.4601782843781967, "grad_norm": 0.4453125, "learning_rate": 0.00012826510721247565, "loss": 0.5118, "step": 2498 }, { "epoch": 1.461347362268011, "grad_norm": 0.458984375, "learning_rate": 0.00012816764132553607, "loss": 0.4759, "step": 2500 }, { "epoch": 1.4625164401578254, "grad_norm": 0.458984375, "learning_rate": 0.0001280701754385965, "loss": 0.5321, "step": 2502 }, { "epoch": 1.4636855180476398, "grad_norm": 0.4375, "learning_rate": 0.00012797270955165692, "loss": 0.4693, "step": 2504 }, { "epoch": 1.4648545959374544, "grad_norm": 0.427734375, "learning_rate": 0.00012787524366471734, "loss": 0.4409, "step": 2506 }, { "epoch": 1.4660236738272687, "grad_norm": 0.4296875, "learning_rate": 0.00012777777777777776, "loss": 0.4841, "step": 2508 }, { "epoch": 1.467192751717083, "grad_norm": 0.45703125, "learning_rate": 0.0001276803118908382, "loss": 0.482, "step": 2510 }, { "epoch": 1.4683618296068977, "grad_norm": 0.4609375, "learning_rate": 0.00012758284600389864, "loss": 0.5053, "step": 2512 }, { "epoch": 1.469530907496712, "grad_norm": 0.45703125, "learning_rate": 0.00012748538011695908, "loss": 0.4906, "step": 2514 }, { "epoch": 1.4706999853865264, "grad_norm": 0.455078125, "learning_rate": 0.0001273879142300195, "loss": 0.5435, "step": 2516 }, { "epoch": 1.4718690632763407, "grad_norm": 0.455078125, "learning_rate": 0.00012729044834307993, "loss": 0.5278, "step": 2518 }, { "epoch": 1.473038141166155, "grad_norm": 0.67578125, "learning_rate": 0.00012719298245614035, "loss": 0.4676, "step": 2520 }, { "epoch": 1.4742072190559696, "grad_norm": 0.50390625, "learning_rate": 0.00012709551656920077, "loss": 0.4983, "step": 2522 }, { "epoch": 1.475376296945784, "grad_norm": 0.609375, "learning_rate": 0.0001269980506822612, "loss": 0.5437, "step": 2524 }, { "epoch": 1.4765453748355983, "grad_norm": 0.49609375, "learning_rate": 0.00012690058479532165, "loss": 0.4985, "step": 2526 }, { "epoch": 1.477714452725413, "grad_norm": 0.498046875, "learning_rate": 0.00012680311890838207, "loss": 0.5014, "step": 2528 }, { "epoch": 1.4788835306152273, "grad_norm": 0.455078125, "learning_rate": 0.00012670565302144252, "loss": 0.4793, "step": 2530 }, { "epoch": 1.4800526085050416, "grad_norm": 0.5, "learning_rate": 0.00012660818713450294, "loss": 0.5015, "step": 2532 }, { "epoch": 1.4812216863948562, "grad_norm": 0.5078125, "learning_rate": 0.00012651072124756336, "loss": 0.5777, "step": 2534 }, { "epoch": 1.4823907642846705, "grad_norm": 0.52734375, "learning_rate": 0.00012641325536062378, "loss": 0.5194, "step": 2536 }, { "epoch": 1.4835598421744849, "grad_norm": 0.451171875, "learning_rate": 0.0001263157894736842, "loss": 0.5067, "step": 2538 }, { "epoch": 1.4847289200642992, "grad_norm": 0.478515625, "learning_rate": 0.00012621832358674463, "loss": 0.4664, "step": 2540 }, { "epoch": 1.4858979979541136, "grad_norm": 0.4296875, "learning_rate": 0.00012612085769980508, "loss": 0.5486, "step": 2542 }, { "epoch": 1.4870670758439282, "grad_norm": 0.97265625, "learning_rate": 0.0001260233918128655, "loss": 0.4842, "step": 2544 }, { "epoch": 1.4882361537337425, "grad_norm": 0.486328125, "learning_rate": 0.00012592592592592592, "loss": 0.4884, "step": 2546 }, { "epoch": 1.4894052316235569, "grad_norm": 0.5, "learning_rate": 0.00012582846003898637, "loss": 0.5559, "step": 2548 }, { "epoch": 1.4905743095133714, "grad_norm": 0.470703125, "learning_rate": 0.0001257309941520468, "loss": 0.5301, "step": 2550 }, { "epoch": 1.4917433874031858, "grad_norm": 0.50390625, "learning_rate": 0.00012563352826510722, "loss": 0.5174, "step": 2552 }, { "epoch": 1.4929124652930001, "grad_norm": 0.466796875, "learning_rate": 0.00012553606237816764, "loss": 0.4971, "step": 2554 }, { "epoch": 1.4940815431828145, "grad_norm": 0.54296875, "learning_rate": 0.00012543859649122806, "loss": 0.4853, "step": 2556 }, { "epoch": 1.4952506210726288, "grad_norm": 0.578125, "learning_rate": 0.0001253411306042885, "loss": 0.4976, "step": 2558 }, { "epoch": 1.4964196989624434, "grad_norm": 0.47265625, "learning_rate": 0.00012524366471734893, "loss": 0.4784, "step": 2560 }, { "epoch": 1.4975887768522578, "grad_norm": 0.470703125, "learning_rate": 0.00012514619883040936, "loss": 0.4954, "step": 2562 }, { "epoch": 1.4987578547420721, "grad_norm": 0.447265625, "learning_rate": 0.0001250487329434698, "loss": 0.4958, "step": 2564 }, { "epoch": 1.4999269326318867, "grad_norm": 0.4296875, "learning_rate": 0.00012495126705653023, "loss": 0.4747, "step": 2566 }, { "epoch": 1.501096010521701, "grad_norm": 0.5, "learning_rate": 0.00012485380116959065, "loss": 0.4837, "step": 2568 }, { "epoch": 1.5022650884115154, "grad_norm": 0.515625, "learning_rate": 0.00012475633528265107, "loss": 0.5141, "step": 2570 }, { "epoch": 1.50343416630133, "grad_norm": 0.44140625, "learning_rate": 0.0001246588693957115, "loss": 0.497, "step": 2572 }, { "epoch": 1.504603244191144, "grad_norm": 0.5390625, "learning_rate": 0.00012456140350877194, "loss": 0.5424, "step": 2574 }, { "epoch": 1.5057723220809587, "grad_norm": 0.484375, "learning_rate": 0.00012446393762183237, "loss": 0.4895, "step": 2576 }, { "epoch": 1.506941399970773, "grad_norm": 0.44921875, "learning_rate": 0.0001243664717348928, "loss": 0.4274, "step": 2578 }, { "epoch": 1.5081104778605874, "grad_norm": 0.474609375, "learning_rate": 0.0001242690058479532, "loss": 0.474, "step": 2580 }, { "epoch": 1.509279555750402, "grad_norm": 0.5, "learning_rate": 0.00012417153996101366, "loss": 0.4974, "step": 2582 }, { "epoch": 1.5104486336402163, "grad_norm": 0.462890625, "learning_rate": 0.00012407407407407408, "loss": 0.487, "step": 2584 }, { "epoch": 1.5116177115300307, "grad_norm": 0.478515625, "learning_rate": 0.0001239766081871345, "loss": 0.5199, "step": 2586 }, { "epoch": 1.5127867894198452, "grad_norm": 0.4296875, "learning_rate": 0.00012387914230019493, "loss": 0.4695, "step": 2588 }, { "epoch": 1.5139558673096594, "grad_norm": 0.478515625, "learning_rate": 0.00012378167641325538, "loss": 0.5281, "step": 2590 }, { "epoch": 1.515124945199474, "grad_norm": 0.49609375, "learning_rate": 0.0001236842105263158, "loss": 0.5104, "step": 2592 }, { "epoch": 1.5162940230892883, "grad_norm": 0.435546875, "learning_rate": 0.00012358674463937622, "loss": 0.4797, "step": 2594 }, { "epoch": 1.5174631009791026, "grad_norm": 0.51953125, "learning_rate": 0.00012348927875243664, "loss": 0.5521, "step": 2596 }, { "epoch": 1.5186321788689172, "grad_norm": 0.4375, "learning_rate": 0.0001233918128654971, "loss": 0.5292, "step": 2598 }, { "epoch": 1.5198012567587316, "grad_norm": 0.466796875, "learning_rate": 0.00012329434697855751, "loss": 0.5144, "step": 2600 }, { "epoch": 1.520970334648546, "grad_norm": 0.58984375, "learning_rate": 0.00012319688109161794, "loss": 0.5575, "step": 2602 }, { "epoch": 1.5221394125383605, "grad_norm": 0.498046875, "learning_rate": 0.00012309941520467836, "loss": 0.5626, "step": 2604 }, { "epoch": 1.5233084904281746, "grad_norm": 0.58984375, "learning_rate": 0.0001230019493177388, "loss": 0.5235, "step": 2606 }, { "epoch": 1.5244775683179892, "grad_norm": 0.455078125, "learning_rate": 0.00012290448343079923, "loss": 0.508, "step": 2608 }, { "epoch": 1.5256466462078035, "grad_norm": 0.49609375, "learning_rate": 0.00012280701754385965, "loss": 0.4921, "step": 2610 }, { "epoch": 1.526815724097618, "grad_norm": 0.48046875, "learning_rate": 0.00012270955165692008, "loss": 0.4954, "step": 2612 }, { "epoch": 1.5279848019874325, "grad_norm": 0.455078125, "learning_rate": 0.0001226120857699805, "loss": 0.5225, "step": 2614 }, { "epoch": 1.5291538798772468, "grad_norm": 1.6875, "learning_rate": 0.00012251461988304095, "loss": 0.4836, "step": 2616 }, { "epoch": 1.5303229577670612, "grad_norm": 0.47265625, "learning_rate": 0.00012241715399610137, "loss": 0.4991, "step": 2618 }, { "epoch": 1.5314920356568757, "grad_norm": 0.462890625, "learning_rate": 0.0001223196881091618, "loss": 0.4973, "step": 2620 }, { "epoch": 1.53266111354669, "grad_norm": 0.474609375, "learning_rate": 0.00012222222222222224, "loss": 0.5018, "step": 2622 }, { "epoch": 1.5338301914365045, "grad_norm": 0.609375, "learning_rate": 0.00012212475633528266, "loss": 0.5728, "step": 2624 }, { "epoch": 1.534999269326319, "grad_norm": 0.474609375, "learning_rate": 0.00012202729044834309, "loss": 0.5437, "step": 2626 }, { "epoch": 1.5361683472161332, "grad_norm": 0.466796875, "learning_rate": 0.00012192982456140352, "loss": 0.533, "step": 2628 }, { "epoch": 1.5373374251059477, "grad_norm": 0.462890625, "learning_rate": 0.00012183235867446394, "loss": 0.4743, "step": 2630 }, { "epoch": 1.538506502995762, "grad_norm": 0.41015625, "learning_rate": 0.00012173489278752437, "loss": 0.4799, "step": 2632 }, { "epoch": 1.5396755808855764, "grad_norm": 0.5859375, "learning_rate": 0.00012163742690058479, "loss": 0.5407, "step": 2634 }, { "epoch": 1.540844658775391, "grad_norm": 0.4296875, "learning_rate": 0.00012153996101364522, "loss": 0.4884, "step": 2636 }, { "epoch": 1.5420137366652054, "grad_norm": 0.4765625, "learning_rate": 0.00012144249512670567, "loss": 0.5327, "step": 2638 }, { "epoch": 1.5431828145550197, "grad_norm": 1.0703125, "learning_rate": 0.0001213450292397661, "loss": 0.6054, "step": 2640 }, { "epoch": 1.5443518924448343, "grad_norm": 0.55078125, "learning_rate": 0.00012124756335282652, "loss": 0.5225, "step": 2642 }, { "epoch": 1.5455209703346484, "grad_norm": 0.478515625, "learning_rate": 0.00012115009746588694, "loss": 0.4628, "step": 2644 }, { "epoch": 1.546690048224463, "grad_norm": 0.5390625, "learning_rate": 0.00012105263157894738, "loss": 0.5348, "step": 2646 }, { "epoch": 1.5478591261142773, "grad_norm": 0.5078125, "learning_rate": 0.0001209551656920078, "loss": 0.5142, "step": 2648 }, { "epoch": 1.5490282040040917, "grad_norm": 0.53125, "learning_rate": 0.00012085769980506822, "loss": 0.4994, "step": 2650 }, { "epoch": 1.5501972818939063, "grad_norm": 0.515625, "learning_rate": 0.00012076023391812866, "loss": 0.5443, "step": 2652 }, { "epoch": 1.5513663597837206, "grad_norm": 0.458984375, "learning_rate": 0.0001206627680311891, "loss": 0.4998, "step": 2654 }, { "epoch": 1.552535437673535, "grad_norm": 0.462890625, "learning_rate": 0.00012056530214424953, "loss": 0.5259, "step": 2656 }, { "epoch": 1.5537045155633495, "grad_norm": 0.48828125, "learning_rate": 0.00012046783625730995, "loss": 0.5091, "step": 2658 }, { "epoch": 1.5548735934531637, "grad_norm": 0.4765625, "learning_rate": 0.00012037037037037037, "loss": 0.6077, "step": 2660 }, { "epoch": 1.5560426713429782, "grad_norm": 0.859375, "learning_rate": 0.00012027290448343081, "loss": 0.4897, "step": 2662 }, { "epoch": 1.5572117492327926, "grad_norm": 0.4765625, "learning_rate": 0.00012017543859649123, "loss": 0.53, "step": 2664 }, { "epoch": 1.558380827122607, "grad_norm": 0.67578125, "learning_rate": 0.00012007797270955165, "loss": 0.5213, "step": 2666 }, { "epoch": 1.5595499050124215, "grad_norm": 0.73046875, "learning_rate": 0.00011998050682261208, "loss": 0.4866, "step": 2668 }, { "epoch": 1.5607189829022359, "grad_norm": 0.546875, "learning_rate": 0.00011988304093567253, "loss": 0.5091, "step": 2670 }, { "epoch": 1.5618880607920502, "grad_norm": 0.50390625, "learning_rate": 0.00011978557504873296, "loss": 0.4714, "step": 2672 }, { "epoch": 1.5630571386818648, "grad_norm": 0.74609375, "learning_rate": 0.00011968810916179338, "loss": 0.56, "step": 2674 }, { "epoch": 1.564226216571679, "grad_norm": 0.453125, "learning_rate": 0.0001195906432748538, "loss": 0.5047, "step": 2676 }, { "epoch": 1.5653952944614935, "grad_norm": 0.52734375, "learning_rate": 0.00011949317738791423, "loss": 0.5009, "step": 2678 }, { "epoch": 1.5665643723513079, "grad_norm": 0.44140625, "learning_rate": 0.00011939571150097466, "loss": 0.4999, "step": 2680 }, { "epoch": 1.5677334502411222, "grad_norm": 0.49609375, "learning_rate": 0.00011929824561403509, "loss": 0.5216, "step": 2682 }, { "epoch": 1.5689025281309368, "grad_norm": 0.52734375, "learning_rate": 0.00011920077972709551, "loss": 0.4842, "step": 2684 }, { "epoch": 1.5700716060207511, "grad_norm": 0.46875, "learning_rate": 0.00011910331384015596, "loss": 0.5055, "step": 2686 }, { "epoch": 1.5712406839105655, "grad_norm": 0.50390625, "learning_rate": 0.00011900584795321638, "loss": 0.4664, "step": 2688 }, { "epoch": 1.57240976180038, "grad_norm": 0.56640625, "learning_rate": 0.00011890838206627682, "loss": 0.5221, "step": 2690 }, { "epoch": 1.5735788396901944, "grad_norm": 0.439453125, "learning_rate": 0.00011881091617933724, "loss": 0.481, "step": 2692 }, { "epoch": 1.5747479175800088, "grad_norm": 0.478515625, "learning_rate": 0.00011871345029239766, "loss": 0.6087, "step": 2694 }, { "epoch": 1.5759169954698233, "grad_norm": 0.55078125, "learning_rate": 0.0001186159844054581, "loss": 0.5085, "step": 2696 }, { "epoch": 1.5770860733596375, "grad_norm": 0.40625, "learning_rate": 0.00011851851851851852, "loss": 0.4165, "step": 2698 }, { "epoch": 1.578255151249452, "grad_norm": 0.49609375, "learning_rate": 0.00011842105263157894, "loss": 0.5322, "step": 2700 }, { "epoch": 1.5794242291392664, "grad_norm": 0.4765625, "learning_rate": 0.00011832358674463939, "loss": 0.5114, "step": 2702 }, { "epoch": 1.5805933070290807, "grad_norm": 0.4921875, "learning_rate": 0.00011822612085769981, "loss": 0.4954, "step": 2704 }, { "epoch": 1.5817623849188953, "grad_norm": 0.5703125, "learning_rate": 0.00011812865497076025, "loss": 0.5638, "step": 2706 }, { "epoch": 1.5829314628087097, "grad_norm": 0.4375, "learning_rate": 0.00011803118908382067, "loss": 0.5842, "step": 2708 }, { "epoch": 1.584100540698524, "grad_norm": 0.84765625, "learning_rate": 0.0001179337231968811, "loss": 0.5528, "step": 2710 }, { "epoch": 1.5852696185883386, "grad_norm": 0.53515625, "learning_rate": 0.00011783625730994152, "loss": 0.5358, "step": 2712 }, { "epoch": 1.5864386964781527, "grad_norm": 0.6328125, "learning_rate": 0.00011773879142300195, "loss": 0.6037, "step": 2714 }, { "epoch": 1.5876077743679673, "grad_norm": 0.44921875, "learning_rate": 0.00011764132553606237, "loss": 0.4884, "step": 2716 }, { "epoch": 1.5887768522577816, "grad_norm": 0.546875, "learning_rate": 0.00011754385964912282, "loss": 0.5316, "step": 2718 }, { "epoch": 1.589945930147596, "grad_norm": 0.49609375, "learning_rate": 0.00011744639376218325, "loss": 0.5015, "step": 2720 }, { "epoch": 1.5911150080374106, "grad_norm": 0.46484375, "learning_rate": 0.00011734892787524367, "loss": 0.4958, "step": 2722 }, { "epoch": 1.592284085927225, "grad_norm": 0.45703125, "learning_rate": 0.0001172514619883041, "loss": 0.4921, "step": 2724 }, { "epoch": 1.5934531638170393, "grad_norm": 0.421875, "learning_rate": 0.00011715399610136453, "loss": 0.4645, "step": 2726 }, { "epoch": 1.5946222417068538, "grad_norm": 0.462890625, "learning_rate": 0.00011705653021442495, "loss": 0.4816, "step": 2728 }, { "epoch": 1.595791319596668, "grad_norm": 0.494140625, "learning_rate": 0.00011695906432748539, "loss": 0.5128, "step": 2730 }, { "epoch": 1.5969603974864826, "grad_norm": 0.451171875, "learning_rate": 0.00011686159844054581, "loss": 0.4852, "step": 2732 }, { "epoch": 1.598129475376297, "grad_norm": 0.55078125, "learning_rate": 0.00011676413255360626, "loss": 0.5084, "step": 2734 }, { "epoch": 1.5992985532661113, "grad_norm": 0.54296875, "learning_rate": 0.00011666666666666668, "loss": 0.4967, "step": 2736 }, { "epoch": 1.6004676311559258, "grad_norm": 0.408203125, "learning_rate": 0.0001165692007797271, "loss": 0.4603, "step": 2738 }, { "epoch": 1.6016367090457402, "grad_norm": 0.45703125, "learning_rate": 0.00011647173489278754, "loss": 0.4917, "step": 2740 }, { "epoch": 1.6028057869355545, "grad_norm": 0.53515625, "learning_rate": 0.00011637426900584796, "loss": 0.5217, "step": 2742 }, { "epoch": 1.603974864825369, "grad_norm": 0.4140625, "learning_rate": 0.00011627680311890838, "loss": 0.4649, "step": 2744 }, { "epoch": 1.6051439427151832, "grad_norm": 0.51171875, "learning_rate": 0.0001161793372319688, "loss": 0.505, "step": 2746 }, { "epoch": 1.6063130206049978, "grad_norm": 0.419921875, "learning_rate": 0.00011608187134502924, "loss": 0.4309, "step": 2748 }, { "epoch": 1.6074820984948122, "grad_norm": 0.47265625, "learning_rate": 0.00011598440545808969, "loss": 0.5105, "step": 2750 }, { "epoch": 1.6086511763846265, "grad_norm": 0.5625, "learning_rate": 0.00011588693957115011, "loss": 0.5273, "step": 2752 }, { "epoch": 1.609820254274441, "grad_norm": 0.63671875, "learning_rate": 0.00011578947368421053, "loss": 0.5164, "step": 2754 }, { "epoch": 1.6109893321642554, "grad_norm": 0.435546875, "learning_rate": 0.00011569200779727096, "loss": 0.4782, "step": 2756 }, { "epoch": 1.6121584100540698, "grad_norm": 0.4921875, "learning_rate": 0.00011559454191033139, "loss": 0.4889, "step": 2758 }, { "epoch": 1.6133274879438844, "grad_norm": 0.63671875, "learning_rate": 0.00011549707602339181, "loss": 0.5186, "step": 2760 }, { "epoch": 1.6144965658336987, "grad_norm": 0.47265625, "learning_rate": 0.00011539961013645224, "loss": 0.4711, "step": 2762 }, { "epoch": 1.615665643723513, "grad_norm": 0.484375, "learning_rate": 0.00011530214424951267, "loss": 0.4893, "step": 2764 }, { "epoch": 1.6168347216133276, "grad_norm": 0.48828125, "learning_rate": 0.00011520467836257311, "loss": 0.5227, "step": 2766 }, { "epoch": 1.6180037995031418, "grad_norm": 0.455078125, "learning_rate": 0.00011510721247563354, "loss": 0.4831, "step": 2768 }, { "epoch": 1.6191728773929563, "grad_norm": 0.4453125, "learning_rate": 0.00011500974658869397, "loss": 0.5047, "step": 2770 }, { "epoch": 1.6203419552827707, "grad_norm": 0.451171875, "learning_rate": 0.00011491228070175439, "loss": 0.512, "step": 2772 }, { "epoch": 1.621511033172585, "grad_norm": 0.44140625, "learning_rate": 0.00011481481481481482, "loss": 0.4838, "step": 2774 }, { "epoch": 1.6226801110623996, "grad_norm": 1.2265625, "learning_rate": 0.00011471734892787525, "loss": 0.5778, "step": 2776 }, { "epoch": 1.623849188952214, "grad_norm": 0.48046875, "learning_rate": 0.00011461988304093567, "loss": 0.4849, "step": 2778 }, { "epoch": 1.6250182668420283, "grad_norm": 0.53125, "learning_rate": 0.00011452241715399609, "loss": 0.5256, "step": 2780 }, { "epoch": 1.626187344731843, "grad_norm": 0.455078125, "learning_rate": 0.00011442495126705654, "loss": 0.5054, "step": 2782 }, { "epoch": 1.627356422621657, "grad_norm": 0.57421875, "learning_rate": 0.00011432748538011698, "loss": 0.4739, "step": 2784 }, { "epoch": 1.6285255005114716, "grad_norm": 0.51171875, "learning_rate": 0.0001142300194931774, "loss": 0.6099, "step": 2786 }, { "epoch": 1.629694578401286, "grad_norm": 0.466796875, "learning_rate": 0.00011413255360623782, "loss": 0.5187, "step": 2788 }, { "epoch": 1.6308636562911003, "grad_norm": 0.50390625, "learning_rate": 0.00011403508771929824, "loss": 0.46, "step": 2790 }, { "epoch": 1.6320327341809149, "grad_norm": 0.494140625, "learning_rate": 0.00011393762183235868, "loss": 0.5229, "step": 2792 }, { "epoch": 1.6332018120707292, "grad_norm": 0.5390625, "learning_rate": 0.0001138401559454191, "loss": 0.4822, "step": 2794 }, { "epoch": 1.6343708899605436, "grad_norm": 0.51953125, "learning_rate": 0.00011374269005847952, "loss": 0.5022, "step": 2796 }, { "epoch": 1.6355399678503582, "grad_norm": 0.451171875, "learning_rate": 0.00011364522417153997, "loss": 0.4819, "step": 2798 }, { "epoch": 1.6367090457401723, "grad_norm": 0.70703125, "learning_rate": 0.0001135477582846004, "loss": 0.4742, "step": 2800 }, { "epoch": 1.6378781236299869, "grad_norm": 0.48046875, "learning_rate": 0.00011345029239766083, "loss": 0.4819, "step": 2802 }, { "epoch": 1.6390472015198012, "grad_norm": 0.62890625, "learning_rate": 0.00011335282651072125, "loss": 0.4933, "step": 2804 }, { "epoch": 1.6402162794096156, "grad_norm": 0.54296875, "learning_rate": 0.00011325536062378168, "loss": 0.4929, "step": 2806 }, { "epoch": 1.6413853572994301, "grad_norm": 0.455078125, "learning_rate": 0.00011315789473684211, "loss": 0.4854, "step": 2808 }, { "epoch": 1.6425544351892445, "grad_norm": 0.498046875, "learning_rate": 0.00011306042884990253, "loss": 0.5641, "step": 2810 }, { "epoch": 1.6437235130790588, "grad_norm": 0.52734375, "learning_rate": 0.00011296296296296296, "loss": 0.5033, "step": 2812 }, { "epoch": 1.6448925909688734, "grad_norm": 0.484375, "learning_rate": 0.0001128654970760234, "loss": 0.487, "step": 2814 }, { "epoch": 1.6460616688586878, "grad_norm": 0.478515625, "learning_rate": 0.00011276803118908383, "loss": 0.4896, "step": 2816 }, { "epoch": 1.6472307467485021, "grad_norm": 0.5390625, "learning_rate": 0.00011267056530214426, "loss": 0.4961, "step": 2818 }, { "epoch": 1.6483998246383167, "grad_norm": 0.44921875, "learning_rate": 0.00011257309941520469, "loss": 0.4786, "step": 2820 }, { "epoch": 1.6495689025281308, "grad_norm": 0.8984375, "learning_rate": 0.00011247563352826511, "loss": 0.55, "step": 2822 }, { "epoch": 1.6507379804179454, "grad_norm": 0.5625, "learning_rate": 0.00011237816764132553, "loss": 0.5174, "step": 2824 }, { "epoch": 1.6519070583077597, "grad_norm": 0.4609375, "learning_rate": 0.00011228070175438597, "loss": 0.4896, "step": 2826 }, { "epoch": 1.653076136197574, "grad_norm": 0.5234375, "learning_rate": 0.00011218323586744639, "loss": 0.5106, "step": 2828 }, { "epoch": 1.6542452140873887, "grad_norm": 0.68359375, "learning_rate": 0.00011208576998050684, "loss": 0.6019, "step": 2830 }, { "epoch": 1.655414291977203, "grad_norm": 0.76171875, "learning_rate": 0.00011198830409356726, "loss": 0.5032, "step": 2832 }, { "epoch": 1.6565833698670174, "grad_norm": 0.6015625, "learning_rate": 0.0001118908382066277, "loss": 0.56, "step": 2834 }, { "epoch": 1.657752447756832, "grad_norm": 0.58203125, "learning_rate": 0.00011179337231968812, "loss": 0.4863, "step": 2836 }, { "epoch": 1.658921525646646, "grad_norm": 0.44140625, "learning_rate": 0.00011169590643274854, "loss": 0.499, "step": 2838 }, { "epoch": 1.6600906035364607, "grad_norm": 1.125, "learning_rate": 0.00011159844054580896, "loss": 0.5816, "step": 2840 }, { "epoch": 1.661259681426275, "grad_norm": 0.6796875, "learning_rate": 0.0001115009746588694, "loss": 0.4982, "step": 2842 }, { "epoch": 1.6624287593160894, "grad_norm": 0.48828125, "learning_rate": 0.00011140350877192982, "loss": 0.5131, "step": 2844 }, { "epoch": 1.663597837205904, "grad_norm": 0.625, "learning_rate": 0.00011130604288499027, "loss": 0.4955, "step": 2846 }, { "epoch": 1.6647669150957183, "grad_norm": 0.498046875, "learning_rate": 0.0001112085769980507, "loss": 0.5469, "step": 2848 }, { "epoch": 1.6659359929855326, "grad_norm": 0.447265625, "learning_rate": 0.00011111111111111112, "loss": 0.5019, "step": 2850 }, { "epoch": 1.6671050708753472, "grad_norm": 0.53515625, "learning_rate": 0.00011101364522417155, "loss": 0.4795, "step": 2852 }, { "epoch": 1.6682741487651613, "grad_norm": 0.64453125, "learning_rate": 0.00011091617933723197, "loss": 0.5395, "step": 2854 }, { "epoch": 1.669443226654976, "grad_norm": 0.51953125, "learning_rate": 0.0001108187134502924, "loss": 0.4709, "step": 2856 }, { "epoch": 1.6706123045447903, "grad_norm": 0.50390625, "learning_rate": 0.00011072124756335282, "loss": 0.4927, "step": 2858 }, { "epoch": 1.6717813824346046, "grad_norm": 0.462890625, "learning_rate": 0.00011062378167641326, "loss": 0.4645, "step": 2860 }, { "epoch": 1.6729504603244192, "grad_norm": 0.46875, "learning_rate": 0.0001105263157894737, "loss": 0.5206, "step": 2862 }, { "epoch": 1.6741195382142335, "grad_norm": 0.4453125, "learning_rate": 0.00011042884990253413, "loss": 0.4863, "step": 2864 }, { "epoch": 1.675288616104048, "grad_norm": 0.4921875, "learning_rate": 0.00011033138401559455, "loss": 0.4934, "step": 2866 }, { "epoch": 1.6764576939938625, "grad_norm": 0.482421875, "learning_rate": 0.00011023391812865498, "loss": 0.494, "step": 2868 }, { "epoch": 1.6776267718836766, "grad_norm": 0.45703125, "learning_rate": 0.00011013645224171541, "loss": 0.4683, "step": 2870 }, { "epoch": 1.6787958497734912, "grad_norm": 0.48046875, "learning_rate": 0.00011003898635477583, "loss": 0.4944, "step": 2872 }, { "epoch": 1.6799649276633055, "grad_norm": 0.46875, "learning_rate": 0.00010994152046783625, "loss": 0.4921, "step": 2874 }, { "epoch": 1.6811340055531199, "grad_norm": 0.44140625, "learning_rate": 0.00010984405458089669, "loss": 0.4924, "step": 2876 }, { "epoch": 1.6823030834429344, "grad_norm": 0.470703125, "learning_rate": 0.00010974658869395714, "loss": 0.4903, "step": 2878 }, { "epoch": 1.6834721613327488, "grad_norm": 0.47265625, "learning_rate": 0.00010964912280701756, "loss": 0.4925, "step": 2880 }, { "epoch": 1.6846412392225631, "grad_norm": 0.427734375, "learning_rate": 0.00010955165692007798, "loss": 0.509, "step": 2882 }, { "epoch": 1.6858103171123777, "grad_norm": 0.443359375, "learning_rate": 0.0001094541910331384, "loss": 0.5393, "step": 2884 }, { "epoch": 1.686979395002192, "grad_norm": 0.42578125, "learning_rate": 0.00010935672514619884, "loss": 0.4775, "step": 2886 }, { "epoch": 1.6881484728920064, "grad_norm": 0.46875, "learning_rate": 0.00010925925925925926, "loss": 0.5132, "step": 2888 }, { "epoch": 1.689317550781821, "grad_norm": 0.439453125, "learning_rate": 0.00010916179337231968, "loss": 0.4972, "step": 2890 }, { "epoch": 1.6904866286716351, "grad_norm": 0.447265625, "learning_rate": 0.00010906432748538011, "loss": 0.482, "step": 2892 }, { "epoch": 1.6916557065614497, "grad_norm": 0.4296875, "learning_rate": 0.00010896686159844056, "loss": 0.4649, "step": 2894 }, { "epoch": 1.692824784451264, "grad_norm": 1.0625, "learning_rate": 0.00010886939571150099, "loss": 0.5683, "step": 2896 }, { "epoch": 1.6939938623410784, "grad_norm": 0.44921875, "learning_rate": 0.00010877192982456141, "loss": 0.4917, "step": 2898 }, { "epoch": 1.695162940230893, "grad_norm": 0.46875, "learning_rate": 0.00010867446393762184, "loss": 0.5137, "step": 2900 }, { "epoch": 1.6963320181207073, "grad_norm": 0.46484375, "learning_rate": 0.00010857699805068227, "loss": 0.5205, "step": 2902 }, { "epoch": 1.6975010960105217, "grad_norm": 0.443359375, "learning_rate": 0.0001084795321637427, "loss": 0.4467, "step": 2904 }, { "epoch": 1.6986701739003363, "grad_norm": 0.453125, "learning_rate": 0.00010838206627680312, "loss": 0.5017, "step": 2906 }, { "epoch": 1.6998392517901504, "grad_norm": 0.455078125, "learning_rate": 0.00010828460038986354, "loss": 0.4427, "step": 2908 }, { "epoch": 1.701008329679965, "grad_norm": 0.447265625, "learning_rate": 0.00010818713450292399, "loss": 0.486, "step": 2910 }, { "epoch": 1.7021774075697793, "grad_norm": 0.62109375, "learning_rate": 0.00010808966861598442, "loss": 0.5212, "step": 2912 }, { "epoch": 1.7033464854595937, "grad_norm": 0.45703125, "learning_rate": 0.00010799220272904485, "loss": 0.449, "step": 2914 }, { "epoch": 1.7045155633494082, "grad_norm": 0.43359375, "learning_rate": 0.00010789473684210527, "loss": 0.4504, "step": 2916 }, { "epoch": 1.7056846412392226, "grad_norm": 0.416015625, "learning_rate": 0.00010779727095516569, "loss": 0.4477, "step": 2918 }, { "epoch": 1.706853719129037, "grad_norm": 0.66796875, "learning_rate": 0.00010769980506822613, "loss": 0.5277, "step": 2920 }, { "epoch": 1.7080227970188515, "grad_norm": 0.48046875, "learning_rate": 0.00010760233918128655, "loss": 0.5377, "step": 2922 }, { "epoch": 1.7091918749086656, "grad_norm": 0.484375, "learning_rate": 0.00010750487329434697, "loss": 0.4533, "step": 2924 }, { "epoch": 1.7103609527984802, "grad_norm": 0.466796875, "learning_rate": 0.00010740740740740742, "loss": 0.514, "step": 2926 }, { "epoch": 1.7115300306882946, "grad_norm": 0.50390625, "learning_rate": 0.00010730994152046784, "loss": 0.4668, "step": 2928 }, { "epoch": 1.712699108578109, "grad_norm": 0.482421875, "learning_rate": 0.00010721247563352828, "loss": 0.4949, "step": 2930 }, { "epoch": 1.7138681864679235, "grad_norm": 0.462890625, "learning_rate": 0.0001071150097465887, "loss": 0.4852, "step": 2932 }, { "epoch": 1.7150372643577378, "grad_norm": 0.447265625, "learning_rate": 0.00010701754385964912, "loss": 0.4805, "step": 2934 }, { "epoch": 1.7162063422475522, "grad_norm": 0.44140625, "learning_rate": 0.00010692007797270956, "loss": 0.4607, "step": 2936 }, { "epoch": 1.7173754201373668, "grad_norm": 0.44921875, "learning_rate": 0.00010682261208576998, "loss": 0.4716, "step": 2938 }, { "epoch": 1.718544498027181, "grad_norm": 0.451171875, "learning_rate": 0.0001067251461988304, "loss": 0.4457, "step": 2940 }, { "epoch": 1.7197135759169955, "grad_norm": 0.447265625, "learning_rate": 0.00010662768031189085, "loss": 0.481, "step": 2942 }, { "epoch": 1.7208826538068098, "grad_norm": 0.4375, "learning_rate": 0.00010653021442495128, "loss": 0.4695, "step": 2944 }, { "epoch": 1.7220517316966242, "grad_norm": 0.416015625, "learning_rate": 0.00010643274853801171, "loss": 0.4423, "step": 2946 }, { "epoch": 1.7232208095864388, "grad_norm": 0.447265625, "learning_rate": 0.00010633528265107213, "loss": 0.4777, "step": 2948 }, { "epoch": 1.724389887476253, "grad_norm": 0.55859375, "learning_rate": 0.00010623781676413256, "loss": 0.5063, "step": 2950 }, { "epoch": 1.7255589653660675, "grad_norm": 0.494140625, "learning_rate": 0.00010614035087719298, "loss": 0.487, "step": 2952 }, { "epoch": 1.726728043255882, "grad_norm": 0.484375, "learning_rate": 0.00010604288499025342, "loss": 0.4766, "step": 2954 }, { "epoch": 1.7278971211456964, "grad_norm": 0.482421875, "learning_rate": 0.00010594541910331384, "loss": 0.4855, "step": 2956 }, { "epoch": 1.7290661990355107, "grad_norm": 0.51171875, "learning_rate": 0.00010584795321637429, "loss": 0.5181, "step": 2958 }, { "epoch": 1.7302352769253253, "grad_norm": 0.49609375, "learning_rate": 0.00010575048732943471, "loss": 0.513, "step": 2960 }, { "epoch": 1.7314043548151394, "grad_norm": 0.451171875, "learning_rate": 0.00010565302144249513, "loss": 0.4913, "step": 2962 }, { "epoch": 1.732573432704954, "grad_norm": 0.466796875, "learning_rate": 0.00010555555555555557, "loss": 0.4949, "step": 2964 }, { "epoch": 1.7337425105947684, "grad_norm": 0.494140625, "learning_rate": 0.00010545808966861599, "loss": 0.4734, "step": 2966 }, { "epoch": 1.7349115884845827, "grad_norm": 0.53125, "learning_rate": 0.00010536062378167641, "loss": 0.4935, "step": 2968 }, { "epoch": 1.7360806663743973, "grad_norm": 0.447265625, "learning_rate": 0.00010526315789473685, "loss": 0.4807, "step": 2970 }, { "epoch": 1.7372497442642116, "grad_norm": 0.4609375, "learning_rate": 0.00010516569200779727, "loss": 0.4963, "step": 2972 }, { "epoch": 1.738418822154026, "grad_norm": 0.4765625, "learning_rate": 0.00010506822612085772, "loss": 0.4904, "step": 2974 }, { "epoch": 1.7395879000438406, "grad_norm": 0.462890625, "learning_rate": 0.00010497076023391814, "loss": 0.4773, "step": 2976 }, { "epoch": 1.7407569779336547, "grad_norm": 0.443359375, "learning_rate": 0.00010487329434697856, "loss": 0.5176, "step": 2978 }, { "epoch": 1.7419260558234693, "grad_norm": 0.466796875, "learning_rate": 0.000104775828460039, "loss": 0.4819, "step": 2980 }, { "epoch": 1.7430951337132836, "grad_norm": 0.46875, "learning_rate": 0.00010467836257309942, "loss": 0.4813, "step": 2982 }, { "epoch": 1.744264211603098, "grad_norm": 0.44921875, "learning_rate": 0.00010458089668615984, "loss": 0.4732, "step": 2984 }, { "epoch": 1.7454332894929125, "grad_norm": 0.41796875, "learning_rate": 0.00010448343079922027, "loss": 0.4551, "step": 2986 }, { "epoch": 1.746602367382727, "grad_norm": 0.44140625, "learning_rate": 0.0001043859649122807, "loss": 0.4832, "step": 2988 }, { "epoch": 1.7477714452725412, "grad_norm": 0.462890625, "learning_rate": 0.00010428849902534115, "loss": 0.5009, "step": 2990 }, { "epoch": 1.7489405231623558, "grad_norm": 0.43359375, "learning_rate": 0.00010419103313840157, "loss": 0.4791, "step": 2992 }, { "epoch": 1.75010960105217, "grad_norm": 0.5, "learning_rate": 0.000104093567251462, "loss": 0.4453, "step": 2994 }, { "epoch": 1.7512786789419845, "grad_norm": 0.462890625, "learning_rate": 0.00010399610136452242, "loss": 0.4768, "step": 2996 }, { "epoch": 1.7524477568317989, "grad_norm": 0.494140625, "learning_rate": 0.00010389863547758286, "loss": 0.5192, "step": 2998 }, { "epoch": 1.7536168347216132, "grad_norm": 0.45703125, "learning_rate": 0.00010380116959064328, "loss": 0.5391, "step": 3000 } ], "logging_steps": 2, "max_steps": 5130, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.72870338772992e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }