{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9995664614584236, "eval_steps": 500, "global_step": 1441, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006936616665221538, "grad_norm": 43.466719125161504, "learning_rate": 0.0, "loss": 2.7095, "step": 1 }, { "epoch": 0.0013873233330443076, "grad_norm": 6.96466711942635, "learning_rate": 2.5595802480981545e-06, "loss": 2.0266, "step": 2 }, { "epoch": 0.0020809849995664614, "grad_norm": 5.728387691489433, "learning_rate": 4.056838710822129e-06, "loss": 2.0082, "step": 3 }, { "epoch": 0.0027746466660886152, "grad_norm": 4.686811256036109, "learning_rate": 5.119160496196309e-06, "loss": 1.9589, "step": 4 }, { "epoch": 0.003468308332610769, "grad_norm": 3.3529589095838874, "learning_rate": 5.943161289177871e-06, "loss": 1.9014, "step": 5 }, { "epoch": 0.004161969999132923, "grad_norm": 3.5493873116106522, "learning_rate": 6.616418958920285e-06, "loss": 1.8478, "step": 6 }, { "epoch": 0.004855631665655077, "grad_norm": 3.2248318636630753, "learning_rate": 7.185650207899778e-06, "loss": 1.8631, "step": 7 }, { "epoch": 0.0055492933321772304, "grad_norm": 2.5559675536274558, "learning_rate": 7.678740744294464e-06, "loss": 1.8771, "step": 8 }, { "epoch": 0.006242954998699385, "grad_norm": 2.852602324925445, "learning_rate": 8.113677421644258e-06, "loss": 1.8517, "step": 9 }, { "epoch": 0.006936616665221538, "grad_norm": 2.8475514523205687, "learning_rate": 8.502741537276027e-06, "loss": 1.8444, "step": 10 }, { "epoch": 0.007630278331743692, "grad_norm": 2.4445595331823116, "learning_rate": 8.854692840710254e-06, "loss": 1.8173, "step": 11 }, { "epoch": 0.008323939998265846, "grad_norm": 2.6236339242092606, "learning_rate": 9.175999207018439e-06, "loss": 1.797, "step": 12 }, { "epoch": 0.009017601664787999, "grad_norm": 2.7480536817929178, "learning_rate": 9.471572411831843e-06, "loss": 1.7525, "step": 13 }, { "epoch": 0.009711263331310154, "grad_norm": 2.8195069251274267, "learning_rate": 9.745230455997932e-06, "loss": 1.7421, "step": 14 }, { "epoch": 0.010404924997832308, "grad_norm": 2.574781912417627, "learning_rate": 9.999999999999999e-06, "loss": 1.7563, "step": 15 }, { "epoch": 0.011098586664354461, "grad_norm": 3.018787395539229, "learning_rate": 1e-05, "loss": 1.7235, "step": 16 }, { "epoch": 0.011792248330876614, "grad_norm": 2.7321515111278085, "learning_rate": 1e-05, "loss": 1.7372, "step": 17 }, { "epoch": 0.01248590999739877, "grad_norm": 2.535876566743008, "learning_rate": 1e-05, "loss": 1.7513, "step": 18 }, { "epoch": 0.013179571663920923, "grad_norm": 2.503169241059693, "learning_rate": 1e-05, "loss": 1.7598, "step": 19 }, { "epoch": 0.013873233330443076, "grad_norm": 2.5731073684349752, "learning_rate": 1e-05, "loss": 1.7947, "step": 20 }, { "epoch": 0.01456689499696523, "grad_norm": 2.5424918030748245, "learning_rate": 1e-05, "loss": 1.7304, "step": 21 }, { "epoch": 0.015260556663487385, "grad_norm": 2.280575429076331, "learning_rate": 1e-05, "loss": 1.6995, "step": 22 }, { "epoch": 0.015954218330009536, "grad_norm": 2.493960760059663, "learning_rate": 1e-05, "loss": 1.6744, "step": 23 }, { "epoch": 0.01664787999653169, "grad_norm": 2.238029250091602, "learning_rate": 1e-05, "loss": 1.7026, "step": 24 }, { "epoch": 0.017341541663053846, "grad_norm": 2.2414764681865362, "learning_rate": 1e-05, "loss": 1.6437, "step": 25 }, { "epoch": 0.018035203329575998, "grad_norm": 2.3762425127817135, "learning_rate": 1e-05, "loss": 1.7258, "step": 26 }, { "epoch": 0.018728864996098153, "grad_norm": 2.471220653637295, "learning_rate": 1e-05, "loss": 1.7413, "step": 27 }, { "epoch": 0.01942252666262031, "grad_norm": 2.236115513685021, "learning_rate": 1e-05, "loss": 1.7788, "step": 28 }, { "epoch": 0.02011618832914246, "grad_norm": 2.1158621654066128, "learning_rate": 1e-05, "loss": 1.6955, "step": 29 }, { "epoch": 0.020809849995664615, "grad_norm": 2.438210038803027, "learning_rate": 1e-05, "loss": 1.7435, "step": 30 }, { "epoch": 0.021503511662186767, "grad_norm": 2.2049092103165484, "learning_rate": 1e-05, "loss": 1.7103, "step": 31 }, { "epoch": 0.022197173328708922, "grad_norm": 1.992834120231878, "learning_rate": 1e-05, "loss": 1.6941, "step": 32 }, { "epoch": 0.022890834995231077, "grad_norm": 2.3021272374064767, "learning_rate": 1e-05, "loss": 1.667, "step": 33 }, { "epoch": 0.02358449666175323, "grad_norm": 2.1119253371859426, "learning_rate": 1e-05, "loss": 1.6849, "step": 34 }, { "epoch": 0.024278158328275384, "grad_norm": 2.3243314568193445, "learning_rate": 1e-05, "loss": 1.6466, "step": 35 }, { "epoch": 0.02497181999479754, "grad_norm": 2.2531985009302176, "learning_rate": 1e-05, "loss": 1.6459, "step": 36 }, { "epoch": 0.02566548166131969, "grad_norm": 2.1844804086893777, "learning_rate": 1e-05, "loss": 1.6794, "step": 37 }, { "epoch": 0.026359143327841845, "grad_norm": 2.2530966998426245, "learning_rate": 1e-05, "loss": 1.6601, "step": 38 }, { "epoch": 0.027052804994364, "grad_norm": 2.1421735569008686, "learning_rate": 1e-05, "loss": 1.6335, "step": 39 }, { "epoch": 0.027746466660886152, "grad_norm": 2.181725062044003, "learning_rate": 1e-05, "loss": 1.6319, "step": 40 }, { "epoch": 0.028440128327408307, "grad_norm": 2.193692520161033, "learning_rate": 1e-05, "loss": 1.6363, "step": 41 }, { "epoch": 0.02913378999393046, "grad_norm": 2.1086393378123818, "learning_rate": 1e-05, "loss": 1.6735, "step": 42 }, { "epoch": 0.029827451660452614, "grad_norm": 2.0131860874790175, "learning_rate": 1e-05, "loss": 1.5967, "step": 43 }, { "epoch": 0.03052111332697477, "grad_norm": 2.1758701969638, "learning_rate": 1e-05, "loss": 1.667, "step": 44 }, { "epoch": 0.03121477499349692, "grad_norm": 2.10881747915129, "learning_rate": 1e-05, "loss": 1.6002, "step": 45 }, { "epoch": 0.03190843666001907, "grad_norm": 2.1929235643249707, "learning_rate": 1e-05, "loss": 1.6109, "step": 46 }, { "epoch": 0.03260209832654123, "grad_norm": 2.2382019559100503, "learning_rate": 1e-05, "loss": 1.6309, "step": 47 }, { "epoch": 0.03329575999306338, "grad_norm": 2.204542258967164, "learning_rate": 1e-05, "loss": 1.6123, "step": 48 }, { "epoch": 0.03398942165958554, "grad_norm": 2.1983829138121194, "learning_rate": 1e-05, "loss": 1.6371, "step": 49 }, { "epoch": 0.03468308332610769, "grad_norm": 2.010616851796152, "learning_rate": 1e-05, "loss": 1.648, "step": 50 }, { "epoch": 0.03537674499262985, "grad_norm": 2.109337350234141, "learning_rate": 1e-05, "loss": 1.5939, "step": 51 }, { "epoch": 0.036070406659151996, "grad_norm": 2.5304099594684657, "learning_rate": 1e-05, "loss": 1.5806, "step": 52 }, { "epoch": 0.03676406832567415, "grad_norm": 2.159949455039585, "learning_rate": 1e-05, "loss": 1.6898, "step": 53 }, { "epoch": 0.037457729992196306, "grad_norm": 2.3028114991190227, "learning_rate": 1e-05, "loss": 1.6102, "step": 54 }, { "epoch": 0.03815139165871846, "grad_norm": 2.137637210948916, "learning_rate": 1e-05, "loss": 1.5976, "step": 55 }, { "epoch": 0.03884505332524062, "grad_norm": 2.283279122337232, "learning_rate": 1e-05, "loss": 1.6182, "step": 56 }, { "epoch": 0.039538714991762765, "grad_norm": 2.352555898736503, "learning_rate": 1e-05, "loss": 1.5764, "step": 57 }, { "epoch": 0.04023237665828492, "grad_norm": 2.2640721232017578, "learning_rate": 1e-05, "loss": 1.606, "step": 58 }, { "epoch": 0.040926038324807075, "grad_norm": 2.2046947158269052, "learning_rate": 1e-05, "loss": 1.6621, "step": 59 }, { "epoch": 0.04161969999132923, "grad_norm": 2.2356515653560955, "learning_rate": 1e-05, "loss": 1.5783, "step": 60 }, { "epoch": 0.042313361657851385, "grad_norm": 2.0914853090761656, "learning_rate": 1e-05, "loss": 1.6386, "step": 61 }, { "epoch": 0.04300702332437353, "grad_norm": 2.0219313193549335, "learning_rate": 1e-05, "loss": 1.5409, "step": 62 }, { "epoch": 0.04370068499089569, "grad_norm": 2.0579915512686675, "learning_rate": 1e-05, "loss": 1.6024, "step": 63 }, { "epoch": 0.044394346657417844, "grad_norm": 2.1350362922236563, "learning_rate": 1e-05, "loss": 1.5979, "step": 64 }, { "epoch": 0.04508800832394, "grad_norm": 2.1574547660781493, "learning_rate": 1e-05, "loss": 1.5968, "step": 65 }, { "epoch": 0.045781669990462154, "grad_norm": 2.032927708489895, "learning_rate": 1e-05, "loss": 1.623, "step": 66 }, { "epoch": 0.04647533165698431, "grad_norm": 2.310697906396082, "learning_rate": 1e-05, "loss": 1.58, "step": 67 }, { "epoch": 0.04716899332350646, "grad_norm": 2.082166726223734, "learning_rate": 1e-05, "loss": 1.5855, "step": 68 }, { "epoch": 0.04786265499002861, "grad_norm": 2.1805698151273205, "learning_rate": 1e-05, "loss": 1.5406, "step": 69 }, { "epoch": 0.04855631665655077, "grad_norm": 2.1517290677451366, "learning_rate": 1e-05, "loss": 1.5794, "step": 70 }, { "epoch": 0.04924997832307292, "grad_norm": 2.2324944455303286, "learning_rate": 1e-05, "loss": 1.5897, "step": 71 }, { "epoch": 0.04994363998959508, "grad_norm": 2.117363841822313, "learning_rate": 1e-05, "loss": 1.6344, "step": 72 }, { "epoch": 0.050637301656117226, "grad_norm": 2.0840589709094055, "learning_rate": 1e-05, "loss": 1.5593, "step": 73 }, { "epoch": 0.05133096332263938, "grad_norm": 2.139635042522081, "learning_rate": 1e-05, "loss": 1.5604, "step": 74 }, { "epoch": 0.052024624989161536, "grad_norm": 1.9773852269088779, "learning_rate": 1e-05, "loss": 1.5132, "step": 75 }, { "epoch": 0.05271828665568369, "grad_norm": 2.2497342140467227, "learning_rate": 1e-05, "loss": 1.5689, "step": 76 }, { "epoch": 0.053411948322205846, "grad_norm": 2.1828363324950515, "learning_rate": 1e-05, "loss": 1.5775, "step": 77 }, { "epoch": 0.054105609988728, "grad_norm": 2.397270079677964, "learning_rate": 1e-05, "loss": 1.5876, "step": 78 }, { "epoch": 0.05479927165525015, "grad_norm": 2.132037167524842, "learning_rate": 1e-05, "loss": 1.5388, "step": 79 }, { "epoch": 0.055492933321772304, "grad_norm": 2.105011522745257, "learning_rate": 1e-05, "loss": 1.5878, "step": 80 }, { "epoch": 0.05618659498829446, "grad_norm": 2.331199900780172, "learning_rate": 1e-05, "loss": 1.6095, "step": 81 }, { "epoch": 0.056880256654816615, "grad_norm": 2.0679915620822915, "learning_rate": 1e-05, "loss": 1.5489, "step": 82 }, { "epoch": 0.05757391832133877, "grad_norm": 2.1615822894325154, "learning_rate": 1e-05, "loss": 1.6244, "step": 83 }, { "epoch": 0.05826757998786092, "grad_norm": 1.979313705736586, "learning_rate": 1e-05, "loss": 1.5881, "step": 84 }, { "epoch": 0.05896124165438307, "grad_norm": 2.1629527862643063, "learning_rate": 1e-05, "loss": 1.5775, "step": 85 }, { "epoch": 0.05965490332090523, "grad_norm": 2.3681176161050956, "learning_rate": 1e-05, "loss": 1.5714, "step": 86 }, { "epoch": 0.06034856498742738, "grad_norm": 1.9456880551179383, "learning_rate": 1e-05, "loss": 1.5516, "step": 87 }, { "epoch": 0.06104222665394954, "grad_norm": 2.1286899349051343, "learning_rate": 1e-05, "loss": 1.5368, "step": 88 }, { "epoch": 0.06173588832047169, "grad_norm": 2.124774583311733, "learning_rate": 1e-05, "loss": 1.5244, "step": 89 }, { "epoch": 0.06242954998699384, "grad_norm": 2.255844207300473, "learning_rate": 1e-05, "loss": 1.6468, "step": 90 }, { "epoch": 0.063123211653516, "grad_norm": 2.253221543014536, "learning_rate": 1e-05, "loss": 1.5459, "step": 91 }, { "epoch": 0.06381687332003814, "grad_norm": 2.161784717901113, "learning_rate": 1e-05, "loss": 1.5295, "step": 92 }, { "epoch": 0.0645105349865603, "grad_norm": 2.100171711959203, "learning_rate": 1e-05, "loss": 1.5432, "step": 93 }, { "epoch": 0.06520419665308246, "grad_norm": 2.0472421305517776, "learning_rate": 1e-05, "loss": 1.5722, "step": 94 }, { "epoch": 0.06589785831960461, "grad_norm": 2.174127269142891, "learning_rate": 1e-05, "loss": 1.5301, "step": 95 }, { "epoch": 0.06659151998612677, "grad_norm": 2.302981491215993, "learning_rate": 1e-05, "loss": 1.5495, "step": 96 }, { "epoch": 0.06728518165264892, "grad_norm": 2.093167418727749, "learning_rate": 1e-05, "loss": 1.5604, "step": 97 }, { "epoch": 0.06797884331917108, "grad_norm": 2.0951541746401263, "learning_rate": 1e-05, "loss": 1.5769, "step": 98 }, { "epoch": 0.06867250498569323, "grad_norm": 1.9576369412503245, "learning_rate": 1e-05, "loss": 1.6003, "step": 99 }, { "epoch": 0.06936616665221539, "grad_norm": 2.3419839237311266, "learning_rate": 1e-05, "loss": 1.4925, "step": 100 }, { "epoch": 0.07005982831873754, "grad_norm": 2.0633427358605774, "learning_rate": 1e-05, "loss": 1.5698, "step": 101 }, { "epoch": 0.0707534899852597, "grad_norm": 2.1113080337620795, "learning_rate": 1e-05, "loss": 1.593, "step": 102 }, { "epoch": 0.07144715165178184, "grad_norm": 2.200709143893939, "learning_rate": 1e-05, "loss": 1.5487, "step": 103 }, { "epoch": 0.07214081331830399, "grad_norm": 2.115437173154835, "learning_rate": 1e-05, "loss": 1.5184, "step": 104 }, { "epoch": 0.07283447498482615, "grad_norm": 1.9592926067021565, "learning_rate": 1e-05, "loss": 1.5824, "step": 105 }, { "epoch": 0.0735281366513483, "grad_norm": 2.0541106318447366, "learning_rate": 1e-05, "loss": 1.5265, "step": 106 }, { "epoch": 0.07422179831787046, "grad_norm": 1.9381187865210794, "learning_rate": 1e-05, "loss": 1.5628, "step": 107 }, { "epoch": 0.07491545998439261, "grad_norm": 2.1035504116541084, "learning_rate": 1e-05, "loss": 1.6001, "step": 108 }, { "epoch": 0.07560912165091477, "grad_norm": 2.143235938125612, "learning_rate": 1e-05, "loss": 1.5153, "step": 109 }, { "epoch": 0.07630278331743692, "grad_norm": 2.0039392778273357, "learning_rate": 1e-05, "loss": 1.5603, "step": 110 }, { "epoch": 0.07699644498395908, "grad_norm": 2.0158397924903233, "learning_rate": 1e-05, "loss": 1.5216, "step": 111 }, { "epoch": 0.07769010665048123, "grad_norm": 1.894029714001099, "learning_rate": 1e-05, "loss": 1.5587, "step": 112 }, { "epoch": 0.07838376831700339, "grad_norm": 2.432706928463119, "learning_rate": 1e-05, "loss": 1.5707, "step": 113 }, { "epoch": 0.07907742998352553, "grad_norm": 2.036785190887291, "learning_rate": 1e-05, "loss": 1.4926, "step": 114 }, { "epoch": 0.07977109165004768, "grad_norm": 1.9400189433970951, "learning_rate": 1e-05, "loss": 1.6043, "step": 115 }, { "epoch": 0.08046475331656984, "grad_norm": 1.919202983721404, "learning_rate": 1e-05, "loss": 1.537, "step": 116 }, { "epoch": 0.081158414983092, "grad_norm": 2.0977172125022707, "learning_rate": 1e-05, "loss": 1.5141, "step": 117 }, { "epoch": 0.08185207664961415, "grad_norm": 2.06407983728045, "learning_rate": 1e-05, "loss": 1.5672, "step": 118 }, { "epoch": 0.0825457383161363, "grad_norm": 2.078141165316271, "learning_rate": 1e-05, "loss": 1.5411, "step": 119 }, { "epoch": 0.08323939998265846, "grad_norm": 2.128515674818184, "learning_rate": 1e-05, "loss": 1.574, "step": 120 }, { "epoch": 0.08393306164918062, "grad_norm": 2.204639065305811, "learning_rate": 1e-05, "loss": 1.5411, "step": 121 }, { "epoch": 0.08462672331570277, "grad_norm": 2.028431294158661, "learning_rate": 1e-05, "loss": 1.4471, "step": 122 }, { "epoch": 0.08532038498222493, "grad_norm": 1.9634386333120701, "learning_rate": 1e-05, "loss": 1.5498, "step": 123 }, { "epoch": 0.08601404664874707, "grad_norm": 2.029215017631285, "learning_rate": 1e-05, "loss": 1.5568, "step": 124 }, { "epoch": 0.08670770831526922, "grad_norm": 2.0663413738174397, "learning_rate": 1e-05, "loss": 1.5426, "step": 125 }, { "epoch": 0.08740136998179138, "grad_norm": 1.9619014687764207, "learning_rate": 1e-05, "loss": 1.5133, "step": 126 }, { "epoch": 0.08809503164831353, "grad_norm": 1.9162219051787666, "learning_rate": 1e-05, "loss": 1.5464, "step": 127 }, { "epoch": 0.08878869331483569, "grad_norm": 2.261542371230024, "learning_rate": 1e-05, "loss": 1.4646, "step": 128 }, { "epoch": 0.08948235498135784, "grad_norm": 1.9715291734543514, "learning_rate": 1e-05, "loss": 1.5132, "step": 129 }, { "epoch": 0.09017601664788, "grad_norm": 2.0203777348766834, "learning_rate": 1e-05, "loss": 1.5354, "step": 130 }, { "epoch": 0.09086967831440215, "grad_norm": 1.9138585711617677, "learning_rate": 1e-05, "loss": 1.5407, "step": 131 }, { "epoch": 0.09156333998092431, "grad_norm": 2.0173322994738596, "learning_rate": 1e-05, "loss": 1.5439, "step": 132 }, { "epoch": 0.09225700164744646, "grad_norm": 2.107269356193457, "learning_rate": 1e-05, "loss": 1.5155, "step": 133 }, { "epoch": 0.09295066331396862, "grad_norm": 1.8484907386064835, "learning_rate": 1e-05, "loss": 1.5765, "step": 134 }, { "epoch": 0.09364432498049076, "grad_norm": 1.9245702524265067, "learning_rate": 1e-05, "loss": 1.5584, "step": 135 }, { "epoch": 0.09433798664701291, "grad_norm": 2.01491446609071, "learning_rate": 1e-05, "loss": 1.5813, "step": 136 }, { "epoch": 0.09503164831353507, "grad_norm": 2.139603451563103, "learning_rate": 1e-05, "loss": 1.5156, "step": 137 }, { "epoch": 0.09572530998005722, "grad_norm": 2.02926570002898, "learning_rate": 1e-05, "loss": 1.5331, "step": 138 }, { "epoch": 0.09641897164657938, "grad_norm": 2.0788419828866314, "learning_rate": 1e-05, "loss": 1.5504, "step": 139 }, { "epoch": 0.09711263331310153, "grad_norm": 2.1919839551775016, "learning_rate": 1e-05, "loss": 1.519, "step": 140 }, { "epoch": 0.09780629497962369, "grad_norm": 1.905930415266768, "learning_rate": 1e-05, "loss": 1.5084, "step": 141 }, { "epoch": 0.09849995664614584, "grad_norm": 2.107652533544824, "learning_rate": 1e-05, "loss": 1.5004, "step": 142 }, { "epoch": 0.099193618312668, "grad_norm": 1.9142412206590709, "learning_rate": 1e-05, "loss": 1.5488, "step": 143 }, { "epoch": 0.09988727997919015, "grad_norm": 1.8910378379533608, "learning_rate": 1e-05, "loss": 1.4912, "step": 144 }, { "epoch": 0.10058094164571231, "grad_norm": 2.0531944494577385, "learning_rate": 1e-05, "loss": 1.5011, "step": 145 }, { "epoch": 0.10127460331223445, "grad_norm": 1.9561470527096887, "learning_rate": 1e-05, "loss": 1.483, "step": 146 }, { "epoch": 0.1019682649787566, "grad_norm": 2.0182745726837186, "learning_rate": 1e-05, "loss": 1.4969, "step": 147 }, { "epoch": 0.10266192664527876, "grad_norm": 2.1655704406766305, "learning_rate": 1e-05, "loss": 1.5151, "step": 148 }, { "epoch": 0.10335558831180092, "grad_norm": 2.067383515526746, "learning_rate": 1e-05, "loss": 1.4932, "step": 149 }, { "epoch": 0.10404924997832307, "grad_norm": 2.16565372834465, "learning_rate": 1e-05, "loss": 1.5211, "step": 150 }, { "epoch": 0.10474291164484523, "grad_norm": 2.067199972931285, "learning_rate": 1e-05, "loss": 1.5182, "step": 151 }, { "epoch": 0.10543657331136738, "grad_norm": 2.021560871683723, "learning_rate": 1e-05, "loss": 1.5216, "step": 152 }, { "epoch": 0.10613023497788954, "grad_norm": 2.0837036440044914, "learning_rate": 1e-05, "loss": 1.4789, "step": 153 }, { "epoch": 0.10682389664441169, "grad_norm": 2.105326579649833, "learning_rate": 1e-05, "loss": 1.5532, "step": 154 }, { "epoch": 0.10751755831093385, "grad_norm": 1.9531358624572481, "learning_rate": 1e-05, "loss": 1.5709, "step": 155 }, { "epoch": 0.108211219977456, "grad_norm": 2.064713519765923, "learning_rate": 1e-05, "loss": 1.5378, "step": 156 }, { "epoch": 0.10890488164397814, "grad_norm": 2.1618499671142226, "learning_rate": 1e-05, "loss": 1.5176, "step": 157 }, { "epoch": 0.1095985433105003, "grad_norm": 1.9086208868139072, "learning_rate": 1e-05, "loss": 1.5546, "step": 158 }, { "epoch": 0.11029220497702245, "grad_norm": 2.02664550739396, "learning_rate": 1e-05, "loss": 1.4997, "step": 159 }, { "epoch": 0.11098586664354461, "grad_norm": 1.999544656281972, "learning_rate": 1e-05, "loss": 1.5003, "step": 160 }, { "epoch": 0.11167952831006676, "grad_norm": 2.053756926485644, "learning_rate": 1e-05, "loss": 1.5292, "step": 161 }, { "epoch": 0.11237318997658892, "grad_norm": 2.094168834382519, "learning_rate": 1e-05, "loss": 1.4676, "step": 162 }, { "epoch": 0.11306685164311107, "grad_norm": 2.002754028656303, "learning_rate": 1e-05, "loss": 1.5105, "step": 163 }, { "epoch": 0.11376051330963323, "grad_norm": 1.9434708295091558, "learning_rate": 1e-05, "loss": 1.5053, "step": 164 }, { "epoch": 0.11445417497615538, "grad_norm": 2.0330157669650815, "learning_rate": 1e-05, "loss": 1.4888, "step": 165 }, { "epoch": 0.11514783664267754, "grad_norm": 1.9289370322291217, "learning_rate": 1e-05, "loss": 1.5277, "step": 166 }, { "epoch": 0.1158414983091997, "grad_norm": 1.959995924753835, "learning_rate": 1e-05, "loss": 1.4825, "step": 167 }, { "epoch": 0.11653515997572184, "grad_norm": 1.7991848893944669, "learning_rate": 1e-05, "loss": 1.5005, "step": 168 }, { "epoch": 0.11722882164224399, "grad_norm": 1.8734306548761093, "learning_rate": 1e-05, "loss": 1.515, "step": 169 }, { "epoch": 0.11792248330876615, "grad_norm": 2.0015044635043218, "learning_rate": 1e-05, "loss": 1.5252, "step": 170 }, { "epoch": 0.1186161449752883, "grad_norm": 2.2665936004262273, "learning_rate": 1e-05, "loss": 1.4838, "step": 171 }, { "epoch": 0.11930980664181046, "grad_norm": 1.9544216259159037, "learning_rate": 1e-05, "loss": 1.4925, "step": 172 }, { "epoch": 0.12000346830833261, "grad_norm": 2.1154532935532715, "learning_rate": 1e-05, "loss": 1.5146, "step": 173 }, { "epoch": 0.12069712997485477, "grad_norm": 2.1021548296617927, "learning_rate": 1e-05, "loss": 1.4915, "step": 174 }, { "epoch": 0.12139079164137692, "grad_norm": 2.0834121176679634, "learning_rate": 1e-05, "loss": 1.5075, "step": 175 }, { "epoch": 0.12208445330789908, "grad_norm": 2.0016972202996213, "learning_rate": 1e-05, "loss": 1.5255, "step": 176 }, { "epoch": 0.12277811497442123, "grad_norm": 2.0612598236678523, "learning_rate": 1e-05, "loss": 1.5149, "step": 177 }, { "epoch": 0.12347177664094337, "grad_norm": 2.0229183358200484, "learning_rate": 1e-05, "loss": 1.5249, "step": 178 }, { "epoch": 0.12416543830746553, "grad_norm": 2.263550407359551, "learning_rate": 1e-05, "loss": 1.5244, "step": 179 }, { "epoch": 0.12485909997398768, "grad_norm": 2.08456086134308, "learning_rate": 1e-05, "loss": 1.4981, "step": 180 }, { "epoch": 0.12555276164050985, "grad_norm": 2.0774621427372386, "learning_rate": 1e-05, "loss": 1.5143, "step": 181 }, { "epoch": 0.126246423307032, "grad_norm": 2.1010285933058626, "learning_rate": 1e-05, "loss": 1.4504, "step": 182 }, { "epoch": 0.12694008497355416, "grad_norm": 2.0102176748558405, "learning_rate": 1e-05, "loss": 1.4146, "step": 183 }, { "epoch": 0.1276337466400763, "grad_norm": 2.095717278113951, "learning_rate": 1e-05, "loss": 1.5395, "step": 184 }, { "epoch": 0.12832740830659844, "grad_norm": 2.193298827450061, "learning_rate": 1e-05, "loss": 1.4769, "step": 185 }, { "epoch": 0.1290210699731206, "grad_norm": 1.9388355574681662, "learning_rate": 1e-05, "loss": 1.5375, "step": 186 }, { "epoch": 0.12971473163964276, "grad_norm": 2.0877632629967913, "learning_rate": 1e-05, "loss": 1.5471, "step": 187 }, { "epoch": 0.1304083933061649, "grad_norm": 2.22367106492369, "learning_rate": 1e-05, "loss": 1.4583, "step": 188 }, { "epoch": 0.13110205497268707, "grad_norm": 1.9943605205254191, "learning_rate": 1e-05, "loss": 1.5008, "step": 189 }, { "epoch": 0.13179571663920922, "grad_norm": 2.096604510058919, "learning_rate": 1e-05, "loss": 1.5449, "step": 190 }, { "epoch": 0.13248937830573138, "grad_norm": 1.9264619546423505, "learning_rate": 1e-05, "loss": 1.5068, "step": 191 }, { "epoch": 0.13318303997225353, "grad_norm": 2.250824707812072, "learning_rate": 1e-05, "loss": 1.4505, "step": 192 }, { "epoch": 0.13387670163877569, "grad_norm": 2.166944357294215, "learning_rate": 1e-05, "loss": 1.5341, "step": 193 }, { "epoch": 0.13457036330529784, "grad_norm": 2.0250424027409673, "learning_rate": 1e-05, "loss": 1.4852, "step": 194 }, { "epoch": 0.13526402497182, "grad_norm": 2.122488575543949, "learning_rate": 1e-05, "loss": 1.4973, "step": 195 }, { "epoch": 0.13595768663834215, "grad_norm": 2.1098754788199647, "learning_rate": 1e-05, "loss": 1.5069, "step": 196 }, { "epoch": 0.1366513483048643, "grad_norm": 2.088689323771004, "learning_rate": 1e-05, "loss": 1.4936, "step": 197 }, { "epoch": 0.13734500997138646, "grad_norm": 1.88466378266683, "learning_rate": 1e-05, "loss": 1.5068, "step": 198 }, { "epoch": 0.13803867163790862, "grad_norm": 2.0160265524905845, "learning_rate": 1e-05, "loss": 1.5368, "step": 199 }, { "epoch": 0.13873233330443077, "grad_norm": 1.8858982582224784, "learning_rate": 1e-05, "loss": 1.4676, "step": 200 }, { "epoch": 0.13942599497095293, "grad_norm": 1.8632802765837246, "learning_rate": 1e-05, "loss": 1.5151, "step": 201 }, { "epoch": 0.14011965663747508, "grad_norm": 1.9258461198592782, "learning_rate": 1e-05, "loss": 1.501, "step": 202 }, { "epoch": 0.14081331830399724, "grad_norm": 1.9872609586669983, "learning_rate": 1e-05, "loss": 1.4688, "step": 203 }, { "epoch": 0.1415069799705194, "grad_norm": 1.8678557794614834, "learning_rate": 1e-05, "loss": 1.5344, "step": 204 }, { "epoch": 0.14220064163704152, "grad_norm": 1.8914203459451417, "learning_rate": 1e-05, "loss": 1.4892, "step": 205 }, { "epoch": 0.14289430330356367, "grad_norm": 1.9911065628954558, "learning_rate": 1e-05, "loss": 1.5092, "step": 206 }, { "epoch": 0.14358796497008583, "grad_norm": 2.0485687976317966, "learning_rate": 1e-05, "loss": 1.5167, "step": 207 }, { "epoch": 0.14428162663660798, "grad_norm": 1.9496266310348234, "learning_rate": 1e-05, "loss": 1.4602, "step": 208 }, { "epoch": 0.14497528830313014, "grad_norm": 2.2871818269739754, "learning_rate": 1e-05, "loss": 1.4685, "step": 209 }, { "epoch": 0.1456689499696523, "grad_norm": 1.984619214022057, "learning_rate": 1e-05, "loss": 1.4409, "step": 210 }, { "epoch": 0.14636261163617445, "grad_norm": 1.9696667123704634, "learning_rate": 1e-05, "loss": 1.5176, "step": 211 }, { "epoch": 0.1470562733026966, "grad_norm": 1.933399558736106, "learning_rate": 1e-05, "loss": 1.5327, "step": 212 }, { "epoch": 0.14774993496921876, "grad_norm": 2.1861585612655885, "learning_rate": 1e-05, "loss": 1.4685, "step": 213 }, { "epoch": 0.14844359663574092, "grad_norm": 2.0747570404469804, "learning_rate": 1e-05, "loss": 1.4925, "step": 214 }, { "epoch": 0.14913725830226307, "grad_norm": 2.0747837540555096, "learning_rate": 1e-05, "loss": 1.5019, "step": 215 }, { "epoch": 0.14983091996878523, "grad_norm": 2.0326660466120297, "learning_rate": 1e-05, "loss": 1.4597, "step": 216 }, { "epoch": 0.15052458163530738, "grad_norm": 1.7823779998232254, "learning_rate": 1e-05, "loss": 1.4539, "step": 217 }, { "epoch": 0.15121824330182954, "grad_norm": 2.157188291473536, "learning_rate": 1e-05, "loss": 1.5397, "step": 218 }, { "epoch": 0.1519119049683517, "grad_norm": 2.0183890562420905, "learning_rate": 1e-05, "loss": 1.4753, "step": 219 }, { "epoch": 0.15260556663487385, "grad_norm": 1.9531135223586058, "learning_rate": 1e-05, "loss": 1.4315, "step": 220 }, { "epoch": 0.153299228301396, "grad_norm": 2.038347402060851, "learning_rate": 1e-05, "loss": 1.4246, "step": 221 }, { "epoch": 0.15399288996791816, "grad_norm": 2.0488940765067967, "learning_rate": 1e-05, "loss": 1.4461, "step": 222 }, { "epoch": 0.1546865516344403, "grad_norm": 2.231661501840028, "learning_rate": 1e-05, "loss": 1.472, "step": 223 }, { "epoch": 0.15538021330096247, "grad_norm": 1.9308119943251083, "learning_rate": 1e-05, "loss": 1.5128, "step": 224 }, { "epoch": 0.15607387496748462, "grad_norm": 2.229149512340474, "learning_rate": 1e-05, "loss": 1.4479, "step": 225 }, { "epoch": 0.15676753663400678, "grad_norm": 2.0928578049018483, "learning_rate": 1e-05, "loss": 1.5009, "step": 226 }, { "epoch": 0.1574611983005289, "grad_norm": 1.9663752374718868, "learning_rate": 1e-05, "loss": 1.4818, "step": 227 }, { "epoch": 0.15815485996705106, "grad_norm": 2.1034138941034786, "learning_rate": 1e-05, "loss": 1.4859, "step": 228 }, { "epoch": 0.15884852163357321, "grad_norm": 2.3560738411841626, "learning_rate": 1e-05, "loss": 1.4779, "step": 229 }, { "epoch": 0.15954218330009537, "grad_norm": 2.4332643261654403, "learning_rate": 1e-05, "loss": 1.4809, "step": 230 }, { "epoch": 0.16023584496661752, "grad_norm": 2.0278596830202757, "learning_rate": 1e-05, "loss": 1.4559, "step": 231 }, { "epoch": 0.16092950663313968, "grad_norm": 2.3341820259604984, "learning_rate": 1e-05, "loss": 1.479, "step": 232 }, { "epoch": 0.16162316829966183, "grad_norm": 1.9872750396163408, "learning_rate": 1e-05, "loss": 1.4715, "step": 233 }, { "epoch": 0.162316829966184, "grad_norm": 1.9999402472599845, "learning_rate": 1e-05, "loss": 1.5459, "step": 234 }, { "epoch": 0.16301049163270614, "grad_norm": 2.128103039977492, "learning_rate": 1e-05, "loss": 1.5096, "step": 235 }, { "epoch": 0.1637041532992283, "grad_norm": 1.9768244155184505, "learning_rate": 1e-05, "loss": 1.5127, "step": 236 }, { "epoch": 0.16439781496575046, "grad_norm": 1.788660125975532, "learning_rate": 1e-05, "loss": 1.4742, "step": 237 }, { "epoch": 0.1650914766322726, "grad_norm": 2.091066870692497, "learning_rate": 1e-05, "loss": 1.4978, "step": 238 }, { "epoch": 0.16578513829879477, "grad_norm": 1.9576004965176508, "learning_rate": 1e-05, "loss": 1.45, "step": 239 }, { "epoch": 0.16647879996531692, "grad_norm": 2.0698620965680528, "learning_rate": 1e-05, "loss": 1.4502, "step": 240 }, { "epoch": 0.16717246163183908, "grad_norm": 2.07030748449516, "learning_rate": 1e-05, "loss": 1.5028, "step": 241 }, { "epoch": 0.16786612329836123, "grad_norm": 2.0470194343255455, "learning_rate": 1e-05, "loss": 1.4714, "step": 242 }, { "epoch": 0.16855978496488339, "grad_norm": 2.0084641094309794, "learning_rate": 1e-05, "loss": 1.4819, "step": 243 }, { "epoch": 0.16925344663140554, "grad_norm": 2.0146863278209106, "learning_rate": 1e-05, "loss": 1.4604, "step": 244 }, { "epoch": 0.1699471082979277, "grad_norm": 2.068474055525701, "learning_rate": 1e-05, "loss": 1.4494, "step": 245 }, { "epoch": 0.17064076996444985, "grad_norm": 1.9451803439502662, "learning_rate": 1e-05, "loss": 1.4424, "step": 246 }, { "epoch": 0.171334431630972, "grad_norm": 2.0197781186907835, "learning_rate": 1e-05, "loss": 1.4614, "step": 247 }, { "epoch": 0.17202809329749413, "grad_norm": 1.9032139101620693, "learning_rate": 1e-05, "loss": 1.4731, "step": 248 }, { "epoch": 0.1727217549640163, "grad_norm": 2.115774791694279, "learning_rate": 1e-05, "loss": 1.416, "step": 249 }, { "epoch": 0.17341541663053844, "grad_norm": 2.0399470329500047, "learning_rate": 1e-05, "loss": 1.4481, "step": 250 }, { "epoch": 0.1741090782970606, "grad_norm": 2.2047084098275027, "learning_rate": 1e-05, "loss": 1.4822, "step": 251 }, { "epoch": 0.17480273996358275, "grad_norm": 2.0604716275395374, "learning_rate": 1e-05, "loss": 1.4698, "step": 252 }, { "epoch": 0.1754964016301049, "grad_norm": 2.0292160669190107, "learning_rate": 1e-05, "loss": 1.4292, "step": 253 }, { "epoch": 0.17619006329662706, "grad_norm": 1.8814965372884114, "learning_rate": 1e-05, "loss": 1.448, "step": 254 }, { "epoch": 0.17688372496314922, "grad_norm": 2.1057756496393805, "learning_rate": 1e-05, "loss": 1.4335, "step": 255 }, { "epoch": 0.17757738662967137, "grad_norm": 2.3132846447554227, "learning_rate": 1e-05, "loss": 1.4856, "step": 256 }, { "epoch": 0.17827104829619353, "grad_norm": 2.0736038589358734, "learning_rate": 1e-05, "loss": 1.4483, "step": 257 }, { "epoch": 0.17896470996271568, "grad_norm": 2.1910317153749985, "learning_rate": 1e-05, "loss": 1.4896, "step": 258 }, { "epoch": 0.17965837162923784, "grad_norm": 2.054248882735396, "learning_rate": 1e-05, "loss": 1.4202, "step": 259 }, { "epoch": 0.18035203329576, "grad_norm": 1.9735586840820536, "learning_rate": 1e-05, "loss": 1.492, "step": 260 }, { "epoch": 0.18104569496228215, "grad_norm": 2.0486128891661357, "learning_rate": 1e-05, "loss": 1.4842, "step": 261 }, { "epoch": 0.1817393566288043, "grad_norm": 2.0487679336561535, "learning_rate": 1e-05, "loss": 1.3914, "step": 262 }, { "epoch": 0.18243301829532646, "grad_norm": 2.122214099406227, "learning_rate": 1e-05, "loss": 1.5043, "step": 263 }, { "epoch": 0.18312667996184862, "grad_norm": 1.9819401665450367, "learning_rate": 1e-05, "loss": 1.461, "step": 264 }, { "epoch": 0.18382034162837077, "grad_norm": 1.8894079827925012, "learning_rate": 1e-05, "loss": 1.5248, "step": 265 }, { "epoch": 0.18451400329489293, "grad_norm": 1.9721789820827966, "learning_rate": 1e-05, "loss": 1.4289, "step": 266 }, { "epoch": 0.18520766496141508, "grad_norm": 1.9579977527698131, "learning_rate": 1e-05, "loss": 1.4313, "step": 267 }, { "epoch": 0.18590132662793724, "grad_norm": 2.0444647256611885, "learning_rate": 1e-05, "loss": 1.4645, "step": 268 }, { "epoch": 0.1865949882944594, "grad_norm": 2.1214502305742275, "learning_rate": 1e-05, "loss": 1.4712, "step": 269 }, { "epoch": 0.18728864996098152, "grad_norm": 1.9384086787313417, "learning_rate": 1e-05, "loss": 1.4641, "step": 270 }, { "epoch": 0.18798231162750367, "grad_norm": 1.9903323950267982, "learning_rate": 1e-05, "loss": 1.4438, "step": 271 }, { "epoch": 0.18867597329402583, "grad_norm": 2.2552584575632264, "learning_rate": 1e-05, "loss": 1.4954, "step": 272 }, { "epoch": 0.18936963496054798, "grad_norm": 2.0400461455788075, "learning_rate": 1e-05, "loss": 1.4574, "step": 273 }, { "epoch": 0.19006329662707014, "grad_norm": 1.985049295968063, "learning_rate": 1e-05, "loss": 1.4714, "step": 274 }, { "epoch": 0.1907569582935923, "grad_norm": 2.0084753961869173, "learning_rate": 1e-05, "loss": 1.4557, "step": 275 }, { "epoch": 0.19145061996011445, "grad_norm": 2.0212614192536473, "learning_rate": 1e-05, "loss": 1.4841, "step": 276 }, { "epoch": 0.1921442816266366, "grad_norm": 2.1484850741149035, "learning_rate": 1e-05, "loss": 1.5103, "step": 277 }, { "epoch": 0.19283794329315876, "grad_norm": 1.9196012631959583, "learning_rate": 1e-05, "loss": 1.4313, "step": 278 }, { "epoch": 0.19353160495968091, "grad_norm": 1.77676382629001, "learning_rate": 1e-05, "loss": 1.4468, "step": 279 }, { "epoch": 0.19422526662620307, "grad_norm": 1.9938658500301698, "learning_rate": 1e-05, "loss": 1.4, "step": 280 }, { "epoch": 0.19491892829272522, "grad_norm": 2.00675288394433, "learning_rate": 1e-05, "loss": 1.393, "step": 281 }, { "epoch": 0.19561258995924738, "grad_norm": 1.9133179166179877, "learning_rate": 1e-05, "loss": 1.4856, "step": 282 }, { "epoch": 0.19630625162576953, "grad_norm": 2.059048555946398, "learning_rate": 1e-05, "loss": 1.4583, "step": 283 }, { "epoch": 0.1969999132922917, "grad_norm": 1.9633303202331707, "learning_rate": 1e-05, "loss": 1.4713, "step": 284 }, { "epoch": 0.19769357495881384, "grad_norm": 2.062438630313021, "learning_rate": 1e-05, "loss": 1.4695, "step": 285 }, { "epoch": 0.198387236625336, "grad_norm": 2.0881034716487017, "learning_rate": 1e-05, "loss": 1.4088, "step": 286 }, { "epoch": 0.19908089829185815, "grad_norm": 2.1703550326161416, "learning_rate": 1e-05, "loss": 1.5003, "step": 287 }, { "epoch": 0.1997745599583803, "grad_norm": 2.074897325155734, "learning_rate": 1e-05, "loss": 1.4233, "step": 288 }, { "epoch": 0.20046822162490247, "grad_norm": 2.1455316048865667, "learning_rate": 1e-05, "loss": 1.499, "step": 289 }, { "epoch": 0.20116188329142462, "grad_norm": 2.101826626794352, "learning_rate": 1e-05, "loss": 1.423, "step": 290 }, { "epoch": 0.20185554495794678, "grad_norm": 1.9433232261964963, "learning_rate": 1e-05, "loss": 1.4258, "step": 291 }, { "epoch": 0.2025492066244689, "grad_norm": 2.0443499477866878, "learning_rate": 1e-05, "loss": 1.4755, "step": 292 }, { "epoch": 0.20324286829099106, "grad_norm": 2.0522507133871795, "learning_rate": 1e-05, "loss": 1.4472, "step": 293 }, { "epoch": 0.2039365299575132, "grad_norm": 2.138230646253678, "learning_rate": 1e-05, "loss": 1.49, "step": 294 }, { "epoch": 0.20463019162403537, "grad_norm": 1.8226561062473423, "learning_rate": 1e-05, "loss": 1.4556, "step": 295 }, { "epoch": 0.20532385329055752, "grad_norm": 2.038783861493903, "learning_rate": 1e-05, "loss": 1.4549, "step": 296 }, { "epoch": 0.20601751495707968, "grad_norm": 2.0515757165248054, "learning_rate": 1e-05, "loss": 1.4662, "step": 297 }, { "epoch": 0.20671117662360183, "grad_norm": 2.0235486233300564, "learning_rate": 1e-05, "loss": 1.4422, "step": 298 }, { "epoch": 0.207404838290124, "grad_norm": 1.9481327456484323, "learning_rate": 1e-05, "loss": 1.4871, "step": 299 }, { "epoch": 0.20809849995664614, "grad_norm": 2.1490389171940962, "learning_rate": 1e-05, "loss": 1.4438, "step": 300 }, { "epoch": 0.2087921616231683, "grad_norm": 1.8636272783701093, "learning_rate": 1e-05, "loss": 1.3833, "step": 301 }, { "epoch": 0.20948582328969045, "grad_norm": 2.0132879418102525, "learning_rate": 1e-05, "loss": 1.5079, "step": 302 }, { "epoch": 0.2101794849562126, "grad_norm": 1.982488095685725, "learning_rate": 1e-05, "loss": 1.4497, "step": 303 }, { "epoch": 0.21087314662273476, "grad_norm": 1.8238922950218728, "learning_rate": 1e-05, "loss": 1.4535, "step": 304 }, { "epoch": 0.21156680828925692, "grad_norm": 1.93657599411593, "learning_rate": 1e-05, "loss": 1.413, "step": 305 }, { "epoch": 0.21226046995577907, "grad_norm": 1.8702247940122898, "learning_rate": 1e-05, "loss": 1.4419, "step": 306 }, { "epoch": 0.21295413162230123, "grad_norm": 1.888122691714043, "learning_rate": 1e-05, "loss": 1.4674, "step": 307 }, { "epoch": 0.21364779328882338, "grad_norm": 1.937294530390563, "learning_rate": 1e-05, "loss": 1.5099, "step": 308 }, { "epoch": 0.21434145495534554, "grad_norm": 1.7678460926752717, "learning_rate": 1e-05, "loss": 1.4653, "step": 309 }, { "epoch": 0.2150351166218677, "grad_norm": 1.9151083060534004, "learning_rate": 1e-05, "loss": 1.4605, "step": 310 }, { "epoch": 0.21572877828838985, "grad_norm": 1.7787361825304597, "learning_rate": 1e-05, "loss": 1.4693, "step": 311 }, { "epoch": 0.216422439954912, "grad_norm": 1.9120724944581113, "learning_rate": 1e-05, "loss": 1.4881, "step": 312 }, { "epoch": 0.21711610162143413, "grad_norm": 1.779761445996506, "learning_rate": 1e-05, "loss": 1.4091, "step": 313 }, { "epoch": 0.2178097632879563, "grad_norm": 2.0485121226808527, "learning_rate": 1e-05, "loss": 1.434, "step": 314 }, { "epoch": 0.21850342495447844, "grad_norm": 1.8510715617371953, "learning_rate": 1e-05, "loss": 1.4502, "step": 315 }, { "epoch": 0.2191970866210006, "grad_norm": 2.0639411090008717, "learning_rate": 1e-05, "loss": 1.4663, "step": 316 }, { "epoch": 0.21989074828752275, "grad_norm": 1.8671225612100188, "learning_rate": 1e-05, "loss": 1.4679, "step": 317 }, { "epoch": 0.2205844099540449, "grad_norm": 2.089367239670572, "learning_rate": 1e-05, "loss": 1.4195, "step": 318 }, { "epoch": 0.22127807162056706, "grad_norm": 1.989736808079142, "learning_rate": 1e-05, "loss": 1.4397, "step": 319 }, { "epoch": 0.22197173328708922, "grad_norm": 1.9101605077692294, "learning_rate": 1e-05, "loss": 1.3985, "step": 320 }, { "epoch": 0.22266539495361137, "grad_norm": 1.9384620738328688, "learning_rate": 1e-05, "loss": 1.4223, "step": 321 }, { "epoch": 0.22335905662013353, "grad_norm": 2.1584155527633198, "learning_rate": 1e-05, "loss": 1.4836, "step": 322 }, { "epoch": 0.22405271828665568, "grad_norm": 2.0224370906370694, "learning_rate": 1e-05, "loss": 1.4454, "step": 323 }, { "epoch": 0.22474637995317784, "grad_norm": 2.0788397833125765, "learning_rate": 1e-05, "loss": 1.5232, "step": 324 }, { "epoch": 0.2254400416197, "grad_norm": 2.0341494419793427, "learning_rate": 1e-05, "loss": 1.4617, "step": 325 }, { "epoch": 0.22613370328622215, "grad_norm": 2.008192362429119, "learning_rate": 1e-05, "loss": 1.4375, "step": 326 }, { "epoch": 0.2268273649527443, "grad_norm": 1.7720050916534584, "learning_rate": 1e-05, "loss": 1.4517, "step": 327 }, { "epoch": 0.22752102661926646, "grad_norm": 1.9703464005250477, "learning_rate": 1e-05, "loss": 1.4201, "step": 328 }, { "epoch": 0.2282146882857886, "grad_norm": 1.8046683048934846, "learning_rate": 1e-05, "loss": 1.4441, "step": 329 }, { "epoch": 0.22890834995231077, "grad_norm": 1.996448453239083, "learning_rate": 1e-05, "loss": 1.411, "step": 330 }, { "epoch": 0.22960201161883292, "grad_norm": 2.202209909202205, "learning_rate": 1e-05, "loss": 1.4443, "step": 331 }, { "epoch": 0.23029567328535508, "grad_norm": 1.9288611183096158, "learning_rate": 1e-05, "loss": 1.4261, "step": 332 }, { "epoch": 0.23098933495187723, "grad_norm": 1.9931802186149232, "learning_rate": 1e-05, "loss": 1.4228, "step": 333 }, { "epoch": 0.2316829966183994, "grad_norm": 1.9300657348149677, "learning_rate": 1e-05, "loss": 1.4791, "step": 334 }, { "epoch": 0.23237665828492152, "grad_norm": 1.952656400463476, "learning_rate": 1e-05, "loss": 1.4682, "step": 335 }, { "epoch": 0.23307031995144367, "grad_norm": 1.920902486845839, "learning_rate": 1e-05, "loss": 1.4286, "step": 336 }, { "epoch": 0.23376398161796583, "grad_norm": 1.9689037515231558, "learning_rate": 1e-05, "loss": 1.439, "step": 337 }, { "epoch": 0.23445764328448798, "grad_norm": 1.928911178735969, "learning_rate": 1e-05, "loss": 1.4464, "step": 338 }, { "epoch": 0.23515130495101014, "grad_norm": 2.1288190614130134, "learning_rate": 1e-05, "loss": 1.4386, "step": 339 }, { "epoch": 0.2358449666175323, "grad_norm": 2.076072187290826, "learning_rate": 1e-05, "loss": 1.4928, "step": 340 }, { "epoch": 0.23653862828405445, "grad_norm": 1.8705047682268778, "learning_rate": 1e-05, "loss": 1.4536, "step": 341 }, { "epoch": 0.2372322899505766, "grad_norm": 1.9565604386223752, "learning_rate": 1e-05, "loss": 1.4783, "step": 342 }, { "epoch": 0.23792595161709876, "grad_norm": 2.028892379215102, "learning_rate": 1e-05, "loss": 1.4272, "step": 343 }, { "epoch": 0.2386196132836209, "grad_norm": 1.8650321597118786, "learning_rate": 1e-05, "loss": 1.4742, "step": 344 }, { "epoch": 0.23931327495014307, "grad_norm": 2.1612306768116083, "learning_rate": 1e-05, "loss": 1.4405, "step": 345 }, { "epoch": 0.24000693661666522, "grad_norm": 2.0151472915975748, "learning_rate": 1e-05, "loss": 1.4246, "step": 346 }, { "epoch": 0.24070059828318738, "grad_norm": 2.1329263531073472, "learning_rate": 1e-05, "loss": 1.4483, "step": 347 }, { "epoch": 0.24139425994970953, "grad_norm": 1.8503952674647635, "learning_rate": 1e-05, "loss": 1.3852, "step": 348 }, { "epoch": 0.2420879216162317, "grad_norm": 1.9232118601702997, "learning_rate": 1e-05, "loss": 1.4704, "step": 349 }, { "epoch": 0.24278158328275384, "grad_norm": 1.8955902858682858, "learning_rate": 1e-05, "loss": 1.4472, "step": 350 }, { "epoch": 0.243475244949276, "grad_norm": 2.0209412893438503, "learning_rate": 1e-05, "loss": 1.453, "step": 351 }, { "epoch": 0.24416890661579815, "grad_norm": 2.0049406945887736, "learning_rate": 1e-05, "loss": 1.4429, "step": 352 }, { "epoch": 0.2448625682823203, "grad_norm": 1.8970813615917181, "learning_rate": 1e-05, "loss": 1.4145, "step": 353 }, { "epoch": 0.24555622994884246, "grad_norm": 1.998898929474657, "learning_rate": 1e-05, "loss": 1.3887, "step": 354 }, { "epoch": 0.24624989161536462, "grad_norm": 2.0390603685379944, "learning_rate": 1e-05, "loss": 1.477, "step": 355 }, { "epoch": 0.24694355328188675, "grad_norm": 1.9473625966599237, "learning_rate": 1e-05, "loss": 1.422, "step": 356 }, { "epoch": 0.2476372149484089, "grad_norm": 1.8621154509930413, "learning_rate": 1e-05, "loss": 1.4551, "step": 357 }, { "epoch": 0.24833087661493106, "grad_norm": 2.079525202308318, "learning_rate": 1e-05, "loss": 1.496, "step": 358 }, { "epoch": 0.2490245382814532, "grad_norm": 2.0606010137774162, "learning_rate": 1e-05, "loss": 1.3982, "step": 359 }, { "epoch": 0.24971819994797537, "grad_norm": 2.4565497491600015, "learning_rate": 1e-05, "loss": 1.4833, "step": 360 }, { "epoch": 0.2504118616144975, "grad_norm": 1.9825019010439706, "learning_rate": 1e-05, "loss": 1.4271, "step": 361 }, { "epoch": 0.2511055232810197, "grad_norm": 1.9086652448767583, "learning_rate": 1e-05, "loss": 1.4125, "step": 362 }, { "epoch": 0.25179918494754183, "grad_norm": 2.0999318583687625, "learning_rate": 1e-05, "loss": 1.4235, "step": 363 }, { "epoch": 0.252492846614064, "grad_norm": 1.985190014569069, "learning_rate": 1e-05, "loss": 1.4501, "step": 364 }, { "epoch": 0.25318650828058614, "grad_norm": 1.910068963847788, "learning_rate": 1e-05, "loss": 1.4658, "step": 365 }, { "epoch": 0.2538801699471083, "grad_norm": 2.1692090204353764, "learning_rate": 1e-05, "loss": 1.4418, "step": 366 }, { "epoch": 0.25457383161363045, "grad_norm": 1.9609194594229094, "learning_rate": 1e-05, "loss": 1.4975, "step": 367 }, { "epoch": 0.2552674932801526, "grad_norm": 1.9921583707202606, "learning_rate": 1e-05, "loss": 1.46, "step": 368 }, { "epoch": 0.25596115494667476, "grad_norm": 1.946762435726228, "learning_rate": 1e-05, "loss": 1.4346, "step": 369 }, { "epoch": 0.2566548166131969, "grad_norm": 1.9131156156498506, "learning_rate": 1e-05, "loss": 1.4145, "step": 370 }, { "epoch": 0.2573484782797191, "grad_norm": 2.0667230542895885, "learning_rate": 1e-05, "loss": 1.4428, "step": 371 }, { "epoch": 0.2580421399462412, "grad_norm": 1.8834877477820704, "learning_rate": 1e-05, "loss": 1.4198, "step": 372 }, { "epoch": 0.2587358016127634, "grad_norm": 2.192277081033706, "learning_rate": 1e-05, "loss": 1.4732, "step": 373 }, { "epoch": 0.2594294632792855, "grad_norm": 2.0923048016266406, "learning_rate": 1e-05, "loss": 1.4748, "step": 374 }, { "epoch": 0.2601231249458077, "grad_norm": 2.086280677890939, "learning_rate": 1e-05, "loss": 1.4405, "step": 375 }, { "epoch": 0.2608167866123298, "grad_norm": 1.905489769808509, "learning_rate": 1e-05, "loss": 1.4445, "step": 376 }, { "epoch": 0.261510448278852, "grad_norm": 2.0450091978264466, "learning_rate": 1e-05, "loss": 1.4186, "step": 377 }, { "epoch": 0.26220410994537413, "grad_norm": 2.133929617755548, "learning_rate": 1e-05, "loss": 1.4572, "step": 378 }, { "epoch": 0.2628977716118963, "grad_norm": 2.040274726964717, "learning_rate": 1e-05, "loss": 1.4238, "step": 379 }, { "epoch": 0.26359143327841844, "grad_norm": 2.031802523503596, "learning_rate": 1e-05, "loss": 1.485, "step": 380 }, { "epoch": 0.2642850949449406, "grad_norm": 1.8546387757797609, "learning_rate": 1e-05, "loss": 1.4579, "step": 381 }, { "epoch": 0.26497875661146275, "grad_norm": 2.0419344295588893, "learning_rate": 1e-05, "loss": 1.4663, "step": 382 }, { "epoch": 0.26567241827798493, "grad_norm": 1.9135093773488887, "learning_rate": 1e-05, "loss": 1.3924, "step": 383 }, { "epoch": 0.26636607994450706, "grad_norm": 1.9016139415831943, "learning_rate": 1e-05, "loss": 1.4308, "step": 384 }, { "epoch": 0.26705974161102924, "grad_norm": 1.9082693062876364, "learning_rate": 1e-05, "loss": 1.3594, "step": 385 }, { "epoch": 0.26775340327755137, "grad_norm": 1.9587042651997673, "learning_rate": 1e-05, "loss": 1.4879, "step": 386 }, { "epoch": 0.26844706494407355, "grad_norm": 1.8806230175431726, "learning_rate": 1e-05, "loss": 1.4101, "step": 387 }, { "epoch": 0.2691407266105957, "grad_norm": 1.7730356951796558, "learning_rate": 1e-05, "loss": 1.4196, "step": 388 }, { "epoch": 0.2698343882771178, "grad_norm": 2.0320726341653255, "learning_rate": 1e-05, "loss": 1.4207, "step": 389 }, { "epoch": 0.27052804994364, "grad_norm": 1.9758337171264837, "learning_rate": 1e-05, "loss": 1.3527, "step": 390 }, { "epoch": 0.2712217116101621, "grad_norm": 1.8958218154268665, "learning_rate": 1e-05, "loss": 1.4802, "step": 391 }, { "epoch": 0.2719153732766843, "grad_norm": 2.0207609598652, "learning_rate": 1e-05, "loss": 1.4304, "step": 392 }, { "epoch": 0.27260903494320643, "grad_norm": 1.9475691858544313, "learning_rate": 1e-05, "loss": 1.4018, "step": 393 }, { "epoch": 0.2733026966097286, "grad_norm": 1.9389041388735107, "learning_rate": 1e-05, "loss": 1.4343, "step": 394 }, { "epoch": 0.27399635827625074, "grad_norm": 1.905336185660673, "learning_rate": 1e-05, "loss": 1.4283, "step": 395 }, { "epoch": 0.2746900199427729, "grad_norm": 1.729440202579088, "learning_rate": 1e-05, "loss": 1.4066, "step": 396 }, { "epoch": 0.27538368160929505, "grad_norm": 1.7537140656073078, "learning_rate": 1e-05, "loss": 1.3722, "step": 397 }, { "epoch": 0.27607734327581723, "grad_norm": 1.8075317974729948, "learning_rate": 1e-05, "loss": 1.3875, "step": 398 }, { "epoch": 0.27677100494233936, "grad_norm": 1.8284022154266728, "learning_rate": 1e-05, "loss": 1.401, "step": 399 }, { "epoch": 0.27746466660886154, "grad_norm": 2.0107348857071563, "learning_rate": 1e-05, "loss": 1.4018, "step": 400 }, { "epoch": 0.27815832827538367, "grad_norm": 2.117911173802508, "learning_rate": 1e-05, "loss": 1.4517, "step": 401 }, { "epoch": 0.27885198994190585, "grad_norm": 1.8033673093925715, "learning_rate": 1e-05, "loss": 1.3847, "step": 402 }, { "epoch": 0.279545651608428, "grad_norm": 1.7714377429685437, "learning_rate": 1e-05, "loss": 1.4638, "step": 403 }, { "epoch": 0.28023931327495016, "grad_norm": 1.8489883317533833, "learning_rate": 1e-05, "loss": 1.4183, "step": 404 }, { "epoch": 0.2809329749414723, "grad_norm": 2.0970756827183625, "learning_rate": 1e-05, "loss": 1.4531, "step": 405 }, { "epoch": 0.2816266366079945, "grad_norm": 2.0300925713263247, "learning_rate": 1e-05, "loss": 1.4129, "step": 406 }, { "epoch": 0.2823202982745166, "grad_norm": 2.0694538236815365, "learning_rate": 1e-05, "loss": 1.4346, "step": 407 }, { "epoch": 0.2830139599410388, "grad_norm": 2.125152836002329, "learning_rate": 1e-05, "loss": 1.4636, "step": 408 }, { "epoch": 0.2837076216075609, "grad_norm": 1.783820266730962, "learning_rate": 1e-05, "loss": 1.4327, "step": 409 }, { "epoch": 0.28440128327408304, "grad_norm": 2.0534367028415943, "learning_rate": 1e-05, "loss": 1.4691, "step": 410 }, { "epoch": 0.2850949449406052, "grad_norm": 1.876558701718411, "learning_rate": 1e-05, "loss": 1.383, "step": 411 }, { "epoch": 0.28578860660712735, "grad_norm": 2.0521158454990855, "learning_rate": 1e-05, "loss": 1.4213, "step": 412 }, { "epoch": 0.28648226827364953, "grad_norm": 2.0220971535707877, "learning_rate": 1e-05, "loss": 1.4658, "step": 413 }, { "epoch": 0.28717592994017166, "grad_norm": 2.038148661267878, "learning_rate": 1e-05, "loss": 1.4389, "step": 414 }, { "epoch": 0.28786959160669384, "grad_norm": 1.9405516561969085, "learning_rate": 1e-05, "loss": 1.4077, "step": 415 }, { "epoch": 0.28856325327321597, "grad_norm": 1.7499414272985196, "learning_rate": 1e-05, "loss": 1.4215, "step": 416 }, { "epoch": 0.28925691493973815, "grad_norm": 1.8287074775541738, "learning_rate": 1e-05, "loss": 1.4734, "step": 417 }, { "epoch": 0.2899505766062603, "grad_norm": 2.029489052381792, "learning_rate": 1e-05, "loss": 1.3857, "step": 418 }, { "epoch": 0.29064423827278246, "grad_norm": 1.9688645010655113, "learning_rate": 1e-05, "loss": 1.377, "step": 419 }, { "epoch": 0.2913378999393046, "grad_norm": 1.925094261068192, "learning_rate": 1e-05, "loss": 1.4246, "step": 420 }, { "epoch": 0.2920315616058268, "grad_norm": 2.117190679742464, "learning_rate": 1e-05, "loss": 1.4595, "step": 421 }, { "epoch": 0.2927252232723489, "grad_norm": 2.076005784675454, "learning_rate": 1e-05, "loss": 1.4242, "step": 422 }, { "epoch": 0.2934188849388711, "grad_norm": 1.9173678536303644, "learning_rate": 1e-05, "loss": 1.4209, "step": 423 }, { "epoch": 0.2941125466053932, "grad_norm": 1.9453593461528418, "learning_rate": 1e-05, "loss": 1.4067, "step": 424 }, { "epoch": 0.2948062082719154, "grad_norm": 2.067561860466064, "learning_rate": 1e-05, "loss": 1.4637, "step": 425 }, { "epoch": 0.2954998699384375, "grad_norm": 1.899343422880033, "learning_rate": 1e-05, "loss": 1.4355, "step": 426 }, { "epoch": 0.2961935316049597, "grad_norm": 1.9847242096071747, "learning_rate": 1e-05, "loss": 1.4292, "step": 427 }, { "epoch": 0.29688719327148183, "grad_norm": 1.854907399156512, "learning_rate": 1e-05, "loss": 1.443, "step": 428 }, { "epoch": 0.297580854938004, "grad_norm": 1.955263267566865, "learning_rate": 1e-05, "loss": 1.3928, "step": 429 }, { "epoch": 0.29827451660452614, "grad_norm": 2.053105452037698, "learning_rate": 1e-05, "loss": 1.378, "step": 430 }, { "epoch": 0.2989681782710483, "grad_norm": 1.9109805438113896, "learning_rate": 1e-05, "loss": 1.3645, "step": 431 }, { "epoch": 0.29966183993757045, "grad_norm": 2.043933400997838, "learning_rate": 1e-05, "loss": 1.435, "step": 432 }, { "epoch": 0.3003555016040926, "grad_norm": 2.1971134552732785, "learning_rate": 1e-05, "loss": 1.4391, "step": 433 }, { "epoch": 0.30104916327061476, "grad_norm": 1.9910023447928422, "learning_rate": 1e-05, "loss": 1.4084, "step": 434 }, { "epoch": 0.3017428249371369, "grad_norm": 2.0432125795374607, "learning_rate": 1e-05, "loss": 1.4423, "step": 435 }, { "epoch": 0.30243648660365907, "grad_norm": 1.9609905480132668, "learning_rate": 1e-05, "loss": 1.4511, "step": 436 }, { "epoch": 0.3031301482701812, "grad_norm": 1.8755027815135972, "learning_rate": 1e-05, "loss": 1.3937, "step": 437 }, { "epoch": 0.3038238099367034, "grad_norm": 1.9069584207191346, "learning_rate": 1e-05, "loss": 1.38, "step": 438 }, { "epoch": 0.3045174716032255, "grad_norm": 2.174136818483425, "learning_rate": 1e-05, "loss": 1.4535, "step": 439 }, { "epoch": 0.3052111332697477, "grad_norm": 1.8297620514780848, "learning_rate": 1e-05, "loss": 1.4285, "step": 440 }, { "epoch": 0.3059047949362698, "grad_norm": 1.8949261520180545, "learning_rate": 1e-05, "loss": 1.4336, "step": 441 }, { "epoch": 0.306598456602792, "grad_norm": 1.9549110409659416, "learning_rate": 1e-05, "loss": 1.373, "step": 442 }, { "epoch": 0.30729211826931413, "grad_norm": 1.9504761362373537, "learning_rate": 1e-05, "loss": 1.3915, "step": 443 }, { "epoch": 0.3079857799358363, "grad_norm": 1.8994399368824695, "learning_rate": 1e-05, "loss": 1.3887, "step": 444 }, { "epoch": 0.30867944160235844, "grad_norm": 1.8796327717601744, "learning_rate": 1e-05, "loss": 1.4075, "step": 445 }, { "epoch": 0.3093731032688806, "grad_norm": 1.9681787692068275, "learning_rate": 1e-05, "loss": 1.4476, "step": 446 }, { "epoch": 0.31006676493540275, "grad_norm": 2.0373492235623285, "learning_rate": 1e-05, "loss": 1.3909, "step": 447 }, { "epoch": 0.31076042660192493, "grad_norm": 1.8891756577788732, "learning_rate": 1e-05, "loss": 1.4804, "step": 448 }, { "epoch": 0.31145408826844706, "grad_norm": 1.833019021556368, "learning_rate": 1e-05, "loss": 1.4207, "step": 449 }, { "epoch": 0.31214774993496924, "grad_norm": 1.836006077142845, "learning_rate": 1e-05, "loss": 1.3933, "step": 450 }, { "epoch": 0.31284141160149137, "grad_norm": 1.8756183495188887, "learning_rate": 1e-05, "loss": 1.4081, "step": 451 }, { "epoch": 0.31353507326801355, "grad_norm": 1.8832247399343314, "learning_rate": 1e-05, "loss": 1.3978, "step": 452 }, { "epoch": 0.3142287349345357, "grad_norm": 1.9411565103780286, "learning_rate": 1e-05, "loss": 1.4346, "step": 453 }, { "epoch": 0.3149223966010578, "grad_norm": 1.9226021145167342, "learning_rate": 1e-05, "loss": 1.4192, "step": 454 }, { "epoch": 0.31561605826758, "grad_norm": 1.864849847975618, "learning_rate": 1e-05, "loss": 1.4023, "step": 455 }, { "epoch": 0.3163097199341021, "grad_norm": 1.841130556375779, "learning_rate": 1e-05, "loss": 1.3843, "step": 456 }, { "epoch": 0.3170033816006243, "grad_norm": 1.9910619133587744, "learning_rate": 1e-05, "loss": 1.4196, "step": 457 }, { "epoch": 0.31769704326714643, "grad_norm": 1.871071007237103, "learning_rate": 1e-05, "loss": 1.4398, "step": 458 }, { "epoch": 0.3183907049336686, "grad_norm": 1.9525027065397538, "learning_rate": 1e-05, "loss": 1.4221, "step": 459 }, { "epoch": 0.31908436660019074, "grad_norm": 1.875959289372641, "learning_rate": 1e-05, "loss": 1.4568, "step": 460 }, { "epoch": 0.3197780282667129, "grad_norm": 1.896488018271282, "learning_rate": 1e-05, "loss": 1.4443, "step": 461 }, { "epoch": 0.32047168993323505, "grad_norm": 1.819017143897324, "learning_rate": 1e-05, "loss": 1.3804, "step": 462 }, { "epoch": 0.32116535159975723, "grad_norm": 1.904368280434257, "learning_rate": 1e-05, "loss": 1.3803, "step": 463 }, { "epoch": 0.32185901326627936, "grad_norm": 1.9630080846645839, "learning_rate": 1e-05, "loss": 1.4099, "step": 464 }, { "epoch": 0.32255267493280154, "grad_norm": 1.8933783131596658, "learning_rate": 1e-05, "loss": 1.4426, "step": 465 }, { "epoch": 0.32324633659932367, "grad_norm": 1.9919230915473398, "learning_rate": 1e-05, "loss": 1.4099, "step": 466 }, { "epoch": 0.32393999826584585, "grad_norm": 2.0057958567267606, "learning_rate": 1e-05, "loss": 1.4409, "step": 467 }, { "epoch": 0.324633659932368, "grad_norm": 2.0237082745853088, "learning_rate": 1e-05, "loss": 1.4331, "step": 468 }, { "epoch": 0.32532732159889016, "grad_norm": 1.9072494572387, "learning_rate": 1e-05, "loss": 1.4298, "step": 469 }, { "epoch": 0.3260209832654123, "grad_norm": 2.1235561783712718, "learning_rate": 1e-05, "loss": 1.4125, "step": 470 }, { "epoch": 0.3267146449319345, "grad_norm": 1.844374731285956, "learning_rate": 1e-05, "loss": 1.3672, "step": 471 }, { "epoch": 0.3274083065984566, "grad_norm": 2.1204902498109326, "learning_rate": 1e-05, "loss": 1.3909, "step": 472 }, { "epoch": 0.3281019682649788, "grad_norm": 2.084912862868831, "learning_rate": 1e-05, "loss": 1.4521, "step": 473 }, { "epoch": 0.3287956299315009, "grad_norm": 1.9363382366089963, "learning_rate": 1e-05, "loss": 1.4166, "step": 474 }, { "epoch": 0.32948929159802304, "grad_norm": 2.043516332073307, "learning_rate": 1e-05, "loss": 1.4221, "step": 475 }, { "epoch": 0.3301829532645452, "grad_norm": 2.125873099222709, "learning_rate": 1e-05, "loss": 1.4894, "step": 476 }, { "epoch": 0.33087661493106735, "grad_norm": 2.042031475330511, "learning_rate": 1e-05, "loss": 1.4025, "step": 477 }, { "epoch": 0.33157027659758953, "grad_norm": 1.7849937086119454, "learning_rate": 1e-05, "loss": 1.4132, "step": 478 }, { "epoch": 0.33226393826411166, "grad_norm": 1.7931574592397888, "learning_rate": 1e-05, "loss": 1.4273, "step": 479 }, { "epoch": 0.33295759993063384, "grad_norm": 2.826745637249205, "learning_rate": 1e-05, "loss": 1.3552, "step": 480 }, { "epoch": 0.33365126159715597, "grad_norm": 1.9644473691572903, "learning_rate": 1e-05, "loss": 1.3728, "step": 481 }, { "epoch": 0.33434492326367815, "grad_norm": 1.8815785092027932, "learning_rate": 1e-05, "loss": 1.3836, "step": 482 }, { "epoch": 0.3350385849302003, "grad_norm": 2.1707103250407265, "learning_rate": 1e-05, "loss": 1.3872, "step": 483 }, { "epoch": 0.33573224659672246, "grad_norm": 1.805124910060426, "learning_rate": 1e-05, "loss": 1.4269, "step": 484 }, { "epoch": 0.3364259082632446, "grad_norm": 2.0566289424785165, "learning_rate": 1e-05, "loss": 1.4353, "step": 485 }, { "epoch": 0.33711956992976677, "grad_norm": 1.9392776766197952, "learning_rate": 1e-05, "loss": 1.42, "step": 486 }, { "epoch": 0.3378132315962889, "grad_norm": 1.7990589474342267, "learning_rate": 1e-05, "loss": 1.4232, "step": 487 }, { "epoch": 0.3385068932628111, "grad_norm": 2.042243634171937, "learning_rate": 1e-05, "loss": 1.4094, "step": 488 }, { "epoch": 0.3392005549293332, "grad_norm": 1.9990437442788238, "learning_rate": 1e-05, "loss": 1.4349, "step": 489 }, { "epoch": 0.3398942165958554, "grad_norm": 2.1051212689157777, "learning_rate": 1e-05, "loss": 1.3838, "step": 490 }, { "epoch": 0.3405878782623775, "grad_norm": 2.1229600713115238, "learning_rate": 1e-05, "loss": 1.3403, "step": 491 }, { "epoch": 0.3412815399288997, "grad_norm": 1.8779948313961718, "learning_rate": 1e-05, "loss": 1.3938, "step": 492 }, { "epoch": 0.34197520159542183, "grad_norm": 1.9140202797636157, "learning_rate": 1e-05, "loss": 1.3917, "step": 493 }, { "epoch": 0.342668863261944, "grad_norm": 1.974739574014709, "learning_rate": 1e-05, "loss": 1.4054, "step": 494 }, { "epoch": 0.34336252492846614, "grad_norm": 1.9199035763474215, "learning_rate": 1e-05, "loss": 1.3922, "step": 495 }, { "epoch": 0.34405618659498827, "grad_norm": 1.9177941872465927, "learning_rate": 1e-05, "loss": 1.4168, "step": 496 }, { "epoch": 0.34474984826151045, "grad_norm": 2.00337000954214, "learning_rate": 1e-05, "loss": 1.3949, "step": 497 }, { "epoch": 0.3454435099280326, "grad_norm": 1.998945535857092, "learning_rate": 1e-05, "loss": 1.3929, "step": 498 }, { "epoch": 0.34613717159455476, "grad_norm": 2.0182449229484636, "learning_rate": 1e-05, "loss": 1.422, "step": 499 }, { "epoch": 0.3468308332610769, "grad_norm": 1.9361126702422482, "learning_rate": 1e-05, "loss": 1.3964, "step": 500 }, { "epoch": 0.34752449492759907, "grad_norm": 2.0248490486946316, "learning_rate": 1e-05, "loss": 1.4293, "step": 501 }, { "epoch": 0.3482181565941212, "grad_norm": 1.8982822396985153, "learning_rate": 1e-05, "loss": 1.3778, "step": 502 }, { "epoch": 0.3489118182606434, "grad_norm": 2.059259100223746, "learning_rate": 1e-05, "loss": 1.4509, "step": 503 }, { "epoch": 0.3496054799271655, "grad_norm": 2.00329460831796, "learning_rate": 1e-05, "loss": 1.3878, "step": 504 }, { "epoch": 0.3502991415936877, "grad_norm": 1.9719718567850673, "learning_rate": 1e-05, "loss": 1.3829, "step": 505 }, { "epoch": 0.3509928032602098, "grad_norm": 1.9623427987164561, "learning_rate": 1e-05, "loss": 1.4036, "step": 506 }, { "epoch": 0.351686464926732, "grad_norm": 2.022806123443883, "learning_rate": 1e-05, "loss": 1.4068, "step": 507 }, { "epoch": 0.35238012659325413, "grad_norm": 2.00015503601285, "learning_rate": 1e-05, "loss": 1.4248, "step": 508 }, { "epoch": 0.3530737882597763, "grad_norm": 1.904741684713311, "learning_rate": 1e-05, "loss": 1.3519, "step": 509 }, { "epoch": 0.35376744992629844, "grad_norm": 1.9455166972699935, "learning_rate": 1e-05, "loss": 1.4152, "step": 510 }, { "epoch": 0.3544611115928206, "grad_norm": 1.9409989480916887, "learning_rate": 1e-05, "loss": 1.3916, "step": 511 }, { "epoch": 0.35515477325934275, "grad_norm": 1.8707742492583141, "learning_rate": 1e-05, "loss": 1.4362, "step": 512 }, { "epoch": 0.35584843492586493, "grad_norm": 1.8314588678913148, "learning_rate": 1e-05, "loss": 1.4279, "step": 513 }, { "epoch": 0.35654209659238706, "grad_norm": 1.913107053098294, "learning_rate": 1e-05, "loss": 1.4686, "step": 514 }, { "epoch": 0.35723575825890924, "grad_norm": 1.8781008328794606, "learning_rate": 1e-05, "loss": 1.4257, "step": 515 }, { "epoch": 0.35792941992543137, "grad_norm": 1.858897618474902, "learning_rate": 1e-05, "loss": 1.3299, "step": 516 }, { "epoch": 0.35862308159195355, "grad_norm": 1.7902966802183116, "learning_rate": 1e-05, "loss": 1.3806, "step": 517 }, { "epoch": 0.3593167432584757, "grad_norm": 1.9406550842391148, "learning_rate": 1e-05, "loss": 1.4069, "step": 518 }, { "epoch": 0.3600104049249978, "grad_norm": 2.032374189763469, "learning_rate": 1e-05, "loss": 1.3893, "step": 519 }, { "epoch": 0.36070406659152, "grad_norm": 2.0622944109637946, "learning_rate": 1e-05, "loss": 1.4044, "step": 520 }, { "epoch": 0.3613977282580421, "grad_norm": 1.7694621919787776, "learning_rate": 1e-05, "loss": 1.3781, "step": 521 }, { "epoch": 0.3620913899245643, "grad_norm": 1.9876645276332312, "learning_rate": 1e-05, "loss": 1.3658, "step": 522 }, { "epoch": 0.3627850515910864, "grad_norm": 1.8286541300136883, "learning_rate": 1e-05, "loss": 1.3954, "step": 523 }, { "epoch": 0.3634787132576086, "grad_norm": 2.1598158167597785, "learning_rate": 1e-05, "loss": 1.4233, "step": 524 }, { "epoch": 0.36417237492413074, "grad_norm": 2.0125775815004308, "learning_rate": 1e-05, "loss": 1.425, "step": 525 }, { "epoch": 0.3648660365906529, "grad_norm": 1.8220999673991007, "learning_rate": 1e-05, "loss": 1.4689, "step": 526 }, { "epoch": 0.36555969825717505, "grad_norm": 1.9067444062678136, "learning_rate": 1e-05, "loss": 1.4184, "step": 527 }, { "epoch": 0.36625335992369723, "grad_norm": 1.8244010981958079, "learning_rate": 1e-05, "loss": 1.4307, "step": 528 }, { "epoch": 0.36694702159021936, "grad_norm": 1.9491392457779768, "learning_rate": 1e-05, "loss": 1.4046, "step": 529 }, { "epoch": 0.36764068325674154, "grad_norm": 1.8922109612389026, "learning_rate": 1e-05, "loss": 1.4312, "step": 530 }, { "epoch": 0.36833434492326367, "grad_norm": 1.8416689512552131, "learning_rate": 1e-05, "loss": 1.3785, "step": 531 }, { "epoch": 0.36902800658978585, "grad_norm": 1.8783492413280818, "learning_rate": 1e-05, "loss": 1.4109, "step": 532 }, { "epoch": 0.369721668256308, "grad_norm": 1.849885383573315, "learning_rate": 1e-05, "loss": 1.378, "step": 533 }, { "epoch": 0.37041532992283016, "grad_norm": 2.071142129836207, "learning_rate": 1e-05, "loss": 1.3897, "step": 534 }, { "epoch": 0.3711089915893523, "grad_norm": 1.955558194146127, "learning_rate": 1e-05, "loss": 1.3781, "step": 535 }, { "epoch": 0.37180265325587447, "grad_norm": 1.8539973789480575, "learning_rate": 1e-05, "loss": 1.3616, "step": 536 }, { "epoch": 0.3724963149223966, "grad_norm": 1.9837890849438713, "learning_rate": 1e-05, "loss": 1.3713, "step": 537 }, { "epoch": 0.3731899765889188, "grad_norm": 1.9070579057354713, "learning_rate": 1e-05, "loss": 1.3479, "step": 538 }, { "epoch": 0.3738836382554409, "grad_norm": 1.9425946819156692, "learning_rate": 1e-05, "loss": 1.376, "step": 539 }, { "epoch": 0.37457729992196304, "grad_norm": 1.8771696216581655, "learning_rate": 1e-05, "loss": 1.3827, "step": 540 }, { "epoch": 0.3752709615884852, "grad_norm": 1.9186911492371799, "learning_rate": 1e-05, "loss": 1.376, "step": 541 }, { "epoch": 0.37596462325500735, "grad_norm": 1.7294504716898789, "learning_rate": 1e-05, "loss": 1.3706, "step": 542 }, { "epoch": 0.37665828492152953, "grad_norm": 1.7712039048530457, "learning_rate": 1e-05, "loss": 1.4438, "step": 543 }, { "epoch": 0.37735194658805166, "grad_norm": 1.83847300072347, "learning_rate": 1e-05, "loss": 1.4161, "step": 544 }, { "epoch": 0.37804560825457384, "grad_norm": 1.825146012889819, "learning_rate": 1e-05, "loss": 1.4118, "step": 545 }, { "epoch": 0.37873926992109597, "grad_norm": 1.7989485574696917, "learning_rate": 1e-05, "loss": 1.3689, "step": 546 }, { "epoch": 0.37943293158761815, "grad_norm": 2.1484169058996816, "learning_rate": 1e-05, "loss": 1.3477, "step": 547 }, { "epoch": 0.3801265932541403, "grad_norm": 2.0781089621700777, "learning_rate": 1e-05, "loss": 1.4372, "step": 548 }, { "epoch": 0.38082025492066246, "grad_norm": 1.8367820389064522, "learning_rate": 1e-05, "loss": 1.4556, "step": 549 }, { "epoch": 0.3815139165871846, "grad_norm": 1.8748672667110173, "learning_rate": 1e-05, "loss": 1.4059, "step": 550 }, { "epoch": 0.38220757825370677, "grad_norm": 1.8956025199638242, "learning_rate": 1e-05, "loss": 1.4102, "step": 551 }, { "epoch": 0.3829012399202289, "grad_norm": 1.9601967006624061, "learning_rate": 1e-05, "loss": 1.3879, "step": 552 }, { "epoch": 0.3835949015867511, "grad_norm": 2.0406526524881707, "learning_rate": 1e-05, "loss": 1.3885, "step": 553 }, { "epoch": 0.3842885632532732, "grad_norm": 2.091444436822791, "learning_rate": 1e-05, "loss": 1.3999, "step": 554 }, { "epoch": 0.3849822249197954, "grad_norm": 1.7714596927341815, "learning_rate": 1e-05, "loss": 1.3452, "step": 555 }, { "epoch": 0.3856758865863175, "grad_norm": 1.825942444423705, "learning_rate": 1e-05, "loss": 1.4142, "step": 556 }, { "epoch": 0.3863695482528397, "grad_norm": 1.871025102921204, "learning_rate": 1e-05, "loss": 1.3421, "step": 557 }, { "epoch": 0.38706320991936183, "grad_norm": 1.9540951936439066, "learning_rate": 1e-05, "loss": 1.384, "step": 558 }, { "epoch": 0.387756871585884, "grad_norm": 1.8147562505586048, "learning_rate": 1e-05, "loss": 1.3794, "step": 559 }, { "epoch": 0.38845053325240614, "grad_norm": 1.8138847080198641, "learning_rate": 1e-05, "loss": 1.4251, "step": 560 }, { "epoch": 0.38914419491892827, "grad_norm": 1.8462886417351692, "learning_rate": 1e-05, "loss": 1.398, "step": 561 }, { "epoch": 0.38983785658545045, "grad_norm": 2.076297894089803, "learning_rate": 1e-05, "loss": 1.3529, "step": 562 }, { "epoch": 0.3905315182519726, "grad_norm": 1.8009871240759965, "learning_rate": 1e-05, "loss": 1.3739, "step": 563 }, { "epoch": 0.39122517991849476, "grad_norm": 1.872935136959733, "learning_rate": 1e-05, "loss": 1.3842, "step": 564 }, { "epoch": 0.3919188415850169, "grad_norm": 1.8957537468315788, "learning_rate": 1e-05, "loss": 1.3633, "step": 565 }, { "epoch": 0.39261250325153907, "grad_norm": 1.991998761167588, "learning_rate": 1e-05, "loss": 1.3742, "step": 566 }, { "epoch": 0.3933061649180612, "grad_norm": 1.7016217238658489, "learning_rate": 1e-05, "loss": 1.4253, "step": 567 }, { "epoch": 0.3939998265845834, "grad_norm": 1.8800648961699629, "learning_rate": 1e-05, "loss": 1.4451, "step": 568 }, { "epoch": 0.3946934882511055, "grad_norm": 2.0019042611698774, "learning_rate": 1e-05, "loss": 1.3997, "step": 569 }, { "epoch": 0.3953871499176277, "grad_norm": 2.215323946324701, "learning_rate": 1e-05, "loss": 1.3577, "step": 570 }, { "epoch": 0.3960808115841498, "grad_norm": 2.0407346613187016, "learning_rate": 1e-05, "loss": 1.3808, "step": 571 }, { "epoch": 0.396774473250672, "grad_norm": 1.874235605467884, "learning_rate": 1e-05, "loss": 1.3883, "step": 572 }, { "epoch": 0.3974681349171941, "grad_norm": 1.9076714486170196, "learning_rate": 1e-05, "loss": 1.3785, "step": 573 }, { "epoch": 0.3981617965837163, "grad_norm": 1.9438453292084767, "learning_rate": 1e-05, "loss": 1.3888, "step": 574 }, { "epoch": 0.39885545825023844, "grad_norm": 1.9010804186629797, "learning_rate": 1e-05, "loss": 1.4044, "step": 575 }, { "epoch": 0.3995491199167606, "grad_norm": 2.1173409005527213, "learning_rate": 1e-05, "loss": 1.4043, "step": 576 }, { "epoch": 0.40024278158328275, "grad_norm": 1.924471913084561, "learning_rate": 1e-05, "loss": 1.3636, "step": 577 }, { "epoch": 0.40093644324980493, "grad_norm": 2.129867365540973, "learning_rate": 1e-05, "loss": 1.4308, "step": 578 }, { "epoch": 0.40163010491632706, "grad_norm": 1.9302914340159718, "learning_rate": 1e-05, "loss": 1.3937, "step": 579 }, { "epoch": 0.40232376658284924, "grad_norm": 1.978021194477141, "learning_rate": 1e-05, "loss": 1.4076, "step": 580 }, { "epoch": 0.40301742824937137, "grad_norm": 1.9801917793272694, "learning_rate": 1e-05, "loss": 1.3756, "step": 581 }, { "epoch": 0.40371108991589355, "grad_norm": 1.8510216673051632, "learning_rate": 1e-05, "loss": 1.3643, "step": 582 }, { "epoch": 0.4044047515824157, "grad_norm": 2.183109714099149, "learning_rate": 1e-05, "loss": 1.4537, "step": 583 }, { "epoch": 0.4050984132489378, "grad_norm": 2.349408814867446, "learning_rate": 1e-05, "loss": 1.4192, "step": 584 }, { "epoch": 0.40579207491546, "grad_norm": 2.2006000600981106, "learning_rate": 1e-05, "loss": 1.4079, "step": 585 }, { "epoch": 0.4064857365819821, "grad_norm": 1.8981822287744043, "learning_rate": 1e-05, "loss": 1.4069, "step": 586 }, { "epoch": 0.4071793982485043, "grad_norm": 1.760320500086242, "learning_rate": 1e-05, "loss": 1.3738, "step": 587 }, { "epoch": 0.4078730599150264, "grad_norm": 1.8312063491211514, "learning_rate": 1e-05, "loss": 1.3887, "step": 588 }, { "epoch": 0.4085667215815486, "grad_norm": 1.760204313519952, "learning_rate": 1e-05, "loss": 1.3332, "step": 589 }, { "epoch": 0.40926038324807074, "grad_norm": 1.7343834044934423, "learning_rate": 1e-05, "loss": 1.3435, "step": 590 }, { "epoch": 0.4099540449145929, "grad_norm": 2.2088224043427687, "learning_rate": 1e-05, "loss": 1.4162, "step": 591 }, { "epoch": 0.41064770658111505, "grad_norm": 1.8200743654853602, "learning_rate": 1e-05, "loss": 1.3931, "step": 592 }, { "epoch": 0.41134136824763723, "grad_norm": 1.8435586030279356, "learning_rate": 1e-05, "loss": 1.4093, "step": 593 }, { "epoch": 0.41203502991415936, "grad_norm": 1.916392749662975, "learning_rate": 1e-05, "loss": 1.3894, "step": 594 }, { "epoch": 0.41272869158068154, "grad_norm": 1.9161401196518564, "learning_rate": 1e-05, "loss": 1.3726, "step": 595 }, { "epoch": 0.41342235324720367, "grad_norm": 1.8756544580848145, "learning_rate": 1e-05, "loss": 1.3697, "step": 596 }, { "epoch": 0.41411601491372585, "grad_norm": 1.898787166818739, "learning_rate": 1e-05, "loss": 1.3941, "step": 597 }, { "epoch": 0.414809676580248, "grad_norm": 1.8044701136277606, "learning_rate": 1e-05, "loss": 1.3999, "step": 598 }, { "epoch": 0.41550333824677016, "grad_norm": 1.8310418953275842, "learning_rate": 1e-05, "loss": 1.3379, "step": 599 }, { "epoch": 0.4161969999132923, "grad_norm": 1.9227996432956476, "learning_rate": 1e-05, "loss": 1.362, "step": 600 }, { "epoch": 0.41689066157981447, "grad_norm": 1.8826360241413953, "learning_rate": 1e-05, "loss": 1.413, "step": 601 }, { "epoch": 0.4175843232463366, "grad_norm": 1.6984816355671049, "learning_rate": 1e-05, "loss": 1.3703, "step": 602 }, { "epoch": 0.4182779849128588, "grad_norm": 1.8568350133313958, "learning_rate": 1e-05, "loss": 1.3909, "step": 603 }, { "epoch": 0.4189716465793809, "grad_norm": 1.917415585499782, "learning_rate": 1e-05, "loss": 1.3698, "step": 604 }, { "epoch": 0.41966530824590303, "grad_norm": 1.8859603577321011, "learning_rate": 1e-05, "loss": 1.4034, "step": 605 }, { "epoch": 0.4203589699124252, "grad_norm": 1.8889156912270977, "learning_rate": 1e-05, "loss": 1.4125, "step": 606 }, { "epoch": 0.42105263157894735, "grad_norm": 1.911197022952282, "learning_rate": 1e-05, "loss": 1.3655, "step": 607 }, { "epoch": 0.42174629324546953, "grad_norm": 1.9782065265119402, "learning_rate": 1e-05, "loss": 1.4201, "step": 608 }, { "epoch": 0.42243995491199166, "grad_norm": 1.8747816354205493, "learning_rate": 1e-05, "loss": 1.3664, "step": 609 }, { "epoch": 0.42313361657851384, "grad_norm": 1.9012852889769853, "learning_rate": 1e-05, "loss": 1.3259, "step": 610 }, { "epoch": 0.42382727824503597, "grad_norm": 1.8979118524613148, "learning_rate": 1e-05, "loss": 1.3971, "step": 611 }, { "epoch": 0.42452093991155815, "grad_norm": 1.8520667588064368, "learning_rate": 1e-05, "loss": 1.4135, "step": 612 }, { "epoch": 0.4252146015780803, "grad_norm": 1.8342529385215973, "learning_rate": 1e-05, "loss": 1.3863, "step": 613 }, { "epoch": 0.42590826324460246, "grad_norm": 1.8295048253836184, "learning_rate": 1e-05, "loss": 1.3087, "step": 614 }, { "epoch": 0.4266019249111246, "grad_norm": 1.7695063050348687, "learning_rate": 1e-05, "loss": 1.4162, "step": 615 }, { "epoch": 0.42729558657764677, "grad_norm": 2.082777123696616, "learning_rate": 1e-05, "loss": 1.4278, "step": 616 }, { "epoch": 0.4279892482441689, "grad_norm": 1.7051067845957866, "learning_rate": 1e-05, "loss": 1.4095, "step": 617 }, { "epoch": 0.4286829099106911, "grad_norm": 1.8695270037035383, "learning_rate": 1e-05, "loss": 1.4503, "step": 618 }, { "epoch": 0.4293765715772132, "grad_norm": 1.9281893067578681, "learning_rate": 1e-05, "loss": 1.38, "step": 619 }, { "epoch": 0.4300702332437354, "grad_norm": 1.777014355958319, "learning_rate": 1e-05, "loss": 1.3984, "step": 620 }, { "epoch": 0.4307638949102575, "grad_norm": 1.842917760375874, "learning_rate": 1e-05, "loss": 1.3853, "step": 621 }, { "epoch": 0.4314575565767797, "grad_norm": 1.8267047634185034, "learning_rate": 1e-05, "loss": 1.3524, "step": 622 }, { "epoch": 0.4321512182433018, "grad_norm": 2.0773027372115394, "learning_rate": 1e-05, "loss": 1.3901, "step": 623 }, { "epoch": 0.432844879909824, "grad_norm": 1.8770974629057278, "learning_rate": 1e-05, "loss": 1.376, "step": 624 }, { "epoch": 0.43353854157634614, "grad_norm": 1.8429345410006643, "learning_rate": 1e-05, "loss": 1.3877, "step": 625 }, { "epoch": 0.43423220324286826, "grad_norm": 1.7713210328211393, "learning_rate": 1e-05, "loss": 1.3976, "step": 626 }, { "epoch": 0.43492586490939045, "grad_norm": 1.8286568674310186, "learning_rate": 1e-05, "loss": 1.3465, "step": 627 }, { "epoch": 0.4356195265759126, "grad_norm": 1.8690248103614588, "learning_rate": 1e-05, "loss": 1.3826, "step": 628 }, { "epoch": 0.43631318824243476, "grad_norm": 2.025467675895459, "learning_rate": 1e-05, "loss": 1.3671, "step": 629 }, { "epoch": 0.4370068499089569, "grad_norm": 1.8903327597767532, "learning_rate": 1e-05, "loss": 1.3548, "step": 630 }, { "epoch": 0.43770051157547907, "grad_norm": 1.7825184036448585, "learning_rate": 1e-05, "loss": 1.3978, "step": 631 }, { "epoch": 0.4383941732420012, "grad_norm": 1.9191594921321666, "learning_rate": 1e-05, "loss": 1.4214, "step": 632 }, { "epoch": 0.4390878349085234, "grad_norm": 1.8710466935982961, "learning_rate": 1e-05, "loss": 1.4685, "step": 633 }, { "epoch": 0.4397814965750455, "grad_norm": 1.9252588373442325, "learning_rate": 1e-05, "loss": 1.3529, "step": 634 }, { "epoch": 0.4404751582415677, "grad_norm": 1.8307810199559515, "learning_rate": 1e-05, "loss": 1.379, "step": 635 }, { "epoch": 0.4411688199080898, "grad_norm": 1.9458457674285412, "learning_rate": 1e-05, "loss": 1.3734, "step": 636 }, { "epoch": 0.441862481574612, "grad_norm": 1.8137288582058262, "learning_rate": 1e-05, "loss": 1.3309, "step": 637 }, { "epoch": 0.4425561432411341, "grad_norm": 1.972371773276373, "learning_rate": 1e-05, "loss": 1.4303, "step": 638 }, { "epoch": 0.4432498049076563, "grad_norm": 1.9163578588020878, "learning_rate": 1e-05, "loss": 1.3191, "step": 639 }, { "epoch": 0.44394346657417844, "grad_norm": 1.8861727336930172, "learning_rate": 1e-05, "loss": 1.374, "step": 640 }, { "epoch": 0.4446371282407006, "grad_norm": 1.7493814291306995, "learning_rate": 1e-05, "loss": 1.3992, "step": 641 }, { "epoch": 0.44533078990722275, "grad_norm": 1.8289477836679608, "learning_rate": 1e-05, "loss": 1.3472, "step": 642 }, { "epoch": 0.44602445157374493, "grad_norm": 1.9323084827259125, "learning_rate": 1e-05, "loss": 1.3966, "step": 643 }, { "epoch": 0.44671811324026706, "grad_norm": 1.866327691734681, "learning_rate": 1e-05, "loss": 1.3793, "step": 644 }, { "epoch": 0.44741177490678924, "grad_norm": 1.8217937158784923, "learning_rate": 1e-05, "loss": 1.4244, "step": 645 }, { "epoch": 0.44810543657331137, "grad_norm": 1.8364372691697852, "learning_rate": 1e-05, "loss": 1.3755, "step": 646 }, { "epoch": 0.4487990982398335, "grad_norm": 2.013774748739019, "learning_rate": 1e-05, "loss": 1.3807, "step": 647 }, { "epoch": 0.4494927599063557, "grad_norm": 1.805702981754065, "learning_rate": 1e-05, "loss": 1.3942, "step": 648 }, { "epoch": 0.4501864215728778, "grad_norm": 1.9948282199322953, "learning_rate": 1e-05, "loss": 1.4366, "step": 649 }, { "epoch": 0.4508800832394, "grad_norm": 1.978073570947312, "learning_rate": 1e-05, "loss": 1.4254, "step": 650 }, { "epoch": 0.4515737449059221, "grad_norm": 1.9232595674423625, "learning_rate": 1e-05, "loss": 1.404, "step": 651 }, { "epoch": 0.4522674065724443, "grad_norm": 1.855166028756208, "learning_rate": 1e-05, "loss": 1.3791, "step": 652 }, { "epoch": 0.4529610682389664, "grad_norm": 1.7786637265688048, "learning_rate": 1e-05, "loss": 1.3936, "step": 653 }, { "epoch": 0.4536547299054886, "grad_norm": 1.6650125213330889, "learning_rate": 1e-05, "loss": 1.38, "step": 654 }, { "epoch": 0.45434839157201073, "grad_norm": 1.9212024581799898, "learning_rate": 1e-05, "loss": 1.3951, "step": 655 }, { "epoch": 0.4550420532385329, "grad_norm": 1.855795867809702, "learning_rate": 1e-05, "loss": 1.3563, "step": 656 }, { "epoch": 0.45573571490505504, "grad_norm": 1.8769724131475898, "learning_rate": 1e-05, "loss": 1.3974, "step": 657 }, { "epoch": 0.4564293765715772, "grad_norm": 1.7926551468562113, "learning_rate": 1e-05, "loss": 1.3974, "step": 658 }, { "epoch": 0.45712303823809936, "grad_norm": 1.9178863763832097, "learning_rate": 1e-05, "loss": 1.4146, "step": 659 }, { "epoch": 0.45781669990462154, "grad_norm": 1.9353453542577745, "learning_rate": 1e-05, "loss": 1.4187, "step": 660 }, { "epoch": 0.45851036157114367, "grad_norm": 1.952403141808128, "learning_rate": 1e-05, "loss": 1.4008, "step": 661 }, { "epoch": 0.45920402323766585, "grad_norm": 1.757904596732749, "learning_rate": 1e-05, "loss": 1.4264, "step": 662 }, { "epoch": 0.459897684904188, "grad_norm": 1.9395774430498347, "learning_rate": 1e-05, "loss": 1.3671, "step": 663 }, { "epoch": 0.46059134657071016, "grad_norm": 1.8890203107447723, "learning_rate": 1e-05, "loss": 1.3715, "step": 664 }, { "epoch": 0.4612850082372323, "grad_norm": 1.8286704559781526, "learning_rate": 1e-05, "loss": 1.3941, "step": 665 }, { "epoch": 0.46197866990375447, "grad_norm": 1.810001367913328, "learning_rate": 1e-05, "loss": 1.3883, "step": 666 }, { "epoch": 0.4626723315702766, "grad_norm": 1.9152160013729407, "learning_rate": 1e-05, "loss": 1.3813, "step": 667 }, { "epoch": 0.4633659932367988, "grad_norm": 2.0094067199071546, "learning_rate": 1e-05, "loss": 1.3176, "step": 668 }, { "epoch": 0.4640596549033209, "grad_norm": 1.8734070756332013, "learning_rate": 1e-05, "loss": 1.4233, "step": 669 }, { "epoch": 0.46475331656984303, "grad_norm": 1.7612782018456719, "learning_rate": 1e-05, "loss": 1.3438, "step": 670 }, { "epoch": 0.4654469782363652, "grad_norm": 1.8683597363973128, "learning_rate": 1e-05, "loss": 1.355, "step": 671 }, { "epoch": 0.46614063990288734, "grad_norm": 2.0012042847679914, "learning_rate": 1e-05, "loss": 1.4169, "step": 672 }, { "epoch": 0.4668343015694095, "grad_norm": 1.9416276311234992, "learning_rate": 1e-05, "loss": 1.3742, "step": 673 }, { "epoch": 0.46752796323593165, "grad_norm": 1.8922567854833168, "learning_rate": 1e-05, "loss": 1.3779, "step": 674 }, { "epoch": 0.46822162490245384, "grad_norm": 1.7793620481094168, "learning_rate": 1e-05, "loss": 1.3915, "step": 675 }, { "epoch": 0.46891528656897596, "grad_norm": 2.167051856102186, "learning_rate": 1e-05, "loss": 1.3989, "step": 676 }, { "epoch": 0.46960894823549815, "grad_norm": 1.871191468625491, "learning_rate": 1e-05, "loss": 1.3547, "step": 677 }, { "epoch": 0.4703026099020203, "grad_norm": 1.7304516085856814, "learning_rate": 1e-05, "loss": 1.3808, "step": 678 }, { "epoch": 0.47099627156854246, "grad_norm": 1.745983949797038, "learning_rate": 1e-05, "loss": 1.3425, "step": 679 }, { "epoch": 0.4716899332350646, "grad_norm": 1.8779422010813842, "learning_rate": 1e-05, "loss": 1.3592, "step": 680 }, { "epoch": 0.47238359490158677, "grad_norm": 1.9934611197767778, "learning_rate": 1e-05, "loss": 1.3538, "step": 681 }, { "epoch": 0.4730772565681089, "grad_norm": 1.9623233741400323, "learning_rate": 1e-05, "loss": 1.346, "step": 682 }, { "epoch": 0.4737709182346311, "grad_norm": 1.796679910714133, "learning_rate": 1e-05, "loss": 1.355, "step": 683 }, { "epoch": 0.4744645799011532, "grad_norm": 1.7722903042578, "learning_rate": 1e-05, "loss": 1.4071, "step": 684 }, { "epoch": 0.4751582415676754, "grad_norm": 1.8052196535090335, "learning_rate": 1e-05, "loss": 1.3814, "step": 685 }, { "epoch": 0.4758519032341975, "grad_norm": 1.7948553541752041, "learning_rate": 1e-05, "loss": 1.3729, "step": 686 }, { "epoch": 0.4765455649007197, "grad_norm": 2.007593236556849, "learning_rate": 1e-05, "loss": 1.3673, "step": 687 }, { "epoch": 0.4772392265672418, "grad_norm": 1.9797683213286288, "learning_rate": 1e-05, "loss": 1.3629, "step": 688 }, { "epoch": 0.477932888233764, "grad_norm": 1.9548220267121557, "learning_rate": 1e-05, "loss": 1.383, "step": 689 }, { "epoch": 0.47862654990028614, "grad_norm": 1.929009442664775, "learning_rate": 1e-05, "loss": 1.3921, "step": 690 }, { "epoch": 0.47932021156680826, "grad_norm": 1.8421018984754722, "learning_rate": 1e-05, "loss": 1.3672, "step": 691 }, { "epoch": 0.48001387323333045, "grad_norm": 1.8234045076467948, "learning_rate": 1e-05, "loss": 1.398, "step": 692 }, { "epoch": 0.4807075348998526, "grad_norm": 2.066405568328085, "learning_rate": 1e-05, "loss": 1.3348, "step": 693 }, { "epoch": 0.48140119656637476, "grad_norm": 1.9963663982661528, "learning_rate": 1e-05, "loss": 1.3662, "step": 694 }, { "epoch": 0.4820948582328969, "grad_norm": 1.883319503377311, "learning_rate": 1e-05, "loss": 1.3835, "step": 695 }, { "epoch": 0.48278851989941907, "grad_norm": 1.56186845290527, "learning_rate": 1e-05, "loss": 1.3386, "step": 696 }, { "epoch": 0.4834821815659412, "grad_norm": 1.8687876803046082, "learning_rate": 1e-05, "loss": 1.3501, "step": 697 }, { "epoch": 0.4841758432324634, "grad_norm": 2.0432055137936036, "learning_rate": 1e-05, "loss": 1.366, "step": 698 }, { "epoch": 0.4848695048989855, "grad_norm": 1.7989863129663937, "learning_rate": 1e-05, "loss": 1.387, "step": 699 }, { "epoch": 0.4855631665655077, "grad_norm": 1.9153016999718706, "learning_rate": 1e-05, "loss": 1.4391, "step": 700 }, { "epoch": 0.4862568282320298, "grad_norm": 1.8256041578604432, "learning_rate": 1e-05, "loss": 1.385, "step": 701 }, { "epoch": 0.486950489898552, "grad_norm": 1.8391212949021127, "learning_rate": 1e-05, "loss": 1.388, "step": 702 }, { "epoch": 0.4876441515650741, "grad_norm": 1.9978982769762854, "learning_rate": 1e-05, "loss": 1.3841, "step": 703 }, { "epoch": 0.4883378132315963, "grad_norm": 1.9247875553761613, "learning_rate": 1e-05, "loss": 1.3514, "step": 704 }, { "epoch": 0.48903147489811843, "grad_norm": 1.8114868285875125, "learning_rate": 1e-05, "loss": 1.3659, "step": 705 }, { "epoch": 0.4897251365646406, "grad_norm": 1.8506801571462923, "learning_rate": 1e-05, "loss": 1.4349, "step": 706 }, { "epoch": 0.49041879823116274, "grad_norm": 1.8480926190821445, "learning_rate": 1e-05, "loss": 1.3768, "step": 707 }, { "epoch": 0.4911124598976849, "grad_norm": 1.9547558985555982, "learning_rate": 1e-05, "loss": 1.3767, "step": 708 }, { "epoch": 0.49180612156420705, "grad_norm": 1.7241605811559706, "learning_rate": 1e-05, "loss": 1.4277, "step": 709 }, { "epoch": 0.49249978323072924, "grad_norm": 1.9597827919247104, "learning_rate": 1e-05, "loss": 1.2981, "step": 710 }, { "epoch": 0.49319344489725137, "grad_norm": 2.241085697740784, "learning_rate": 1e-05, "loss": 1.3738, "step": 711 }, { "epoch": 0.4938871065637735, "grad_norm": 2.1107944914928023, "learning_rate": 1e-05, "loss": 1.3788, "step": 712 }, { "epoch": 0.4945807682302957, "grad_norm": 1.9819856716802244, "learning_rate": 1e-05, "loss": 1.3805, "step": 713 }, { "epoch": 0.4952744298968178, "grad_norm": 1.7818814238988587, "learning_rate": 1e-05, "loss": 1.3864, "step": 714 }, { "epoch": 0.49596809156334, "grad_norm": 1.8028425771808247, "learning_rate": 1e-05, "loss": 1.4047, "step": 715 }, { "epoch": 0.4966617532298621, "grad_norm": 1.810766896499592, "learning_rate": 1e-05, "loss": 1.3899, "step": 716 }, { "epoch": 0.4973554148963843, "grad_norm": 1.8166160127003885, "learning_rate": 1e-05, "loss": 1.3603, "step": 717 }, { "epoch": 0.4980490765629064, "grad_norm": 2.13361803754948, "learning_rate": 1e-05, "loss": 1.4022, "step": 718 }, { "epoch": 0.4987427382294286, "grad_norm": 2.016997065796454, "learning_rate": 1e-05, "loss": 1.4034, "step": 719 }, { "epoch": 0.49943639989595073, "grad_norm": 2.0748032742705793, "learning_rate": 1e-05, "loss": 1.4105, "step": 720 }, { "epoch": 0.5001300615624729, "grad_norm": 1.8523948307863285, "learning_rate": 1e-05, "loss": 1.3697, "step": 721 }, { "epoch": 0.500823723228995, "grad_norm": 1.8122879610484914, "learning_rate": 1e-05, "loss": 1.404, "step": 722 }, { "epoch": 0.5015173848955172, "grad_norm": 1.8832267152617044, "learning_rate": 1e-05, "loss": 1.412, "step": 723 }, { "epoch": 0.5022110465620394, "grad_norm": 1.8257618111117269, "learning_rate": 1e-05, "loss": 1.4016, "step": 724 }, { "epoch": 0.5029047082285615, "grad_norm": 1.7942051170674276, "learning_rate": 1e-05, "loss": 1.3686, "step": 725 }, { "epoch": 0.5035983698950837, "grad_norm": 1.97042982772308, "learning_rate": 1e-05, "loss": 1.3575, "step": 726 }, { "epoch": 0.5042920315616058, "grad_norm": 2.053740555710897, "learning_rate": 1e-05, "loss": 1.379, "step": 727 }, { "epoch": 0.504985693228128, "grad_norm": 1.977833812394873, "learning_rate": 1e-05, "loss": 1.3174, "step": 728 }, { "epoch": 0.5056793548946501, "grad_norm": 1.9322288272504544, "learning_rate": 1e-05, "loss": 1.4038, "step": 729 }, { "epoch": 0.5063730165611723, "grad_norm": 2.0400420303295275, "learning_rate": 1e-05, "loss": 1.3953, "step": 730 }, { "epoch": 0.5070666782276945, "grad_norm": 1.976799409278783, "learning_rate": 1e-05, "loss": 1.3895, "step": 731 }, { "epoch": 0.5077603398942167, "grad_norm": 1.9306400432028785, "learning_rate": 1e-05, "loss": 1.4228, "step": 732 }, { "epoch": 0.5084540015607387, "grad_norm": 1.7347175465813025, "learning_rate": 1e-05, "loss": 1.3704, "step": 733 }, { "epoch": 0.5091476632272609, "grad_norm": 1.7180782433090167, "learning_rate": 1e-05, "loss": 1.3547, "step": 734 }, { "epoch": 0.5098413248937831, "grad_norm": 1.852662073693016, "learning_rate": 1e-05, "loss": 1.4228, "step": 735 }, { "epoch": 0.5105349865603052, "grad_norm": 1.8075356657945834, "learning_rate": 1e-05, "loss": 1.3887, "step": 736 }, { "epoch": 0.5112286482268273, "grad_norm": 1.7367287190868546, "learning_rate": 1e-05, "loss": 1.3752, "step": 737 }, { "epoch": 0.5119223098933495, "grad_norm": 1.8286379353529887, "learning_rate": 1e-05, "loss": 1.3963, "step": 738 }, { "epoch": 0.5126159715598717, "grad_norm": 1.7825873791247597, "learning_rate": 1e-05, "loss": 1.397, "step": 739 }, { "epoch": 0.5133096332263938, "grad_norm": 2.0430860649068396, "learning_rate": 1e-05, "loss": 1.368, "step": 740 }, { "epoch": 0.514003294892916, "grad_norm": 1.82138787437088, "learning_rate": 1e-05, "loss": 1.3888, "step": 741 }, { "epoch": 0.5146969565594381, "grad_norm": 1.9949710877946953, "learning_rate": 1e-05, "loss": 1.3221, "step": 742 }, { "epoch": 0.5153906182259603, "grad_norm": 1.7999499915500219, "learning_rate": 1e-05, "loss": 1.3214, "step": 743 }, { "epoch": 0.5160842798924824, "grad_norm": 1.944101461453592, "learning_rate": 1e-05, "loss": 1.3641, "step": 744 }, { "epoch": 0.5167779415590046, "grad_norm": 1.9484142602052452, "learning_rate": 1e-05, "loss": 1.4183, "step": 745 }, { "epoch": 0.5174716032255268, "grad_norm": 1.9261361736136446, "learning_rate": 1e-05, "loss": 1.39, "step": 746 }, { "epoch": 0.518165264892049, "grad_norm": 1.873564918154059, "learning_rate": 1e-05, "loss": 1.3511, "step": 747 }, { "epoch": 0.518858926558571, "grad_norm": 1.975733457802349, "learning_rate": 1e-05, "loss": 1.3786, "step": 748 }, { "epoch": 0.5195525882250932, "grad_norm": 1.9396312909048352, "learning_rate": 1e-05, "loss": 1.3161, "step": 749 }, { "epoch": 0.5202462498916154, "grad_norm": 1.736368973000078, "learning_rate": 1e-05, "loss": 1.3704, "step": 750 }, { "epoch": 0.5209399115581376, "grad_norm": 1.7911382730465684, "learning_rate": 1e-05, "loss": 1.3379, "step": 751 }, { "epoch": 0.5216335732246596, "grad_norm": 1.8590253300206483, "learning_rate": 1e-05, "loss": 1.3531, "step": 752 }, { "epoch": 0.5223272348911818, "grad_norm": 1.9765690927121422, "learning_rate": 1e-05, "loss": 1.3728, "step": 753 }, { "epoch": 0.523020896557704, "grad_norm": 2.0025745088354148, "learning_rate": 1e-05, "loss": 1.3939, "step": 754 }, { "epoch": 0.5237145582242261, "grad_norm": 1.9192921819712603, "learning_rate": 1e-05, "loss": 1.3914, "step": 755 }, { "epoch": 0.5244082198907483, "grad_norm": 1.8600023632893699, "learning_rate": 1e-05, "loss": 1.3499, "step": 756 }, { "epoch": 0.5251018815572704, "grad_norm": 1.9355189855324415, "learning_rate": 1e-05, "loss": 1.3471, "step": 757 }, { "epoch": 0.5257955432237926, "grad_norm": 2.0191814963451495, "learning_rate": 1e-05, "loss": 1.4244, "step": 758 }, { "epoch": 0.5264892048903147, "grad_norm": 1.883726645686763, "learning_rate": 1e-05, "loss": 1.3619, "step": 759 }, { "epoch": 0.5271828665568369, "grad_norm": 1.7947607621987598, "learning_rate": 1e-05, "loss": 1.3404, "step": 760 }, { "epoch": 0.5278765282233591, "grad_norm": 1.8889256123493143, "learning_rate": 1e-05, "loss": 1.4012, "step": 761 }, { "epoch": 0.5285701898898812, "grad_norm": 1.9230513498083945, "learning_rate": 1e-05, "loss": 1.3475, "step": 762 }, { "epoch": 0.5292638515564033, "grad_norm": 1.8067019812009846, "learning_rate": 1e-05, "loss": 1.3911, "step": 763 }, { "epoch": 0.5299575132229255, "grad_norm": 1.8380178891570427, "learning_rate": 1e-05, "loss": 1.3614, "step": 764 }, { "epoch": 0.5306511748894477, "grad_norm": 1.8235814437026021, "learning_rate": 1e-05, "loss": 1.3556, "step": 765 }, { "epoch": 0.5313448365559699, "grad_norm": 1.830617817421823, "learning_rate": 1e-05, "loss": 1.3844, "step": 766 }, { "epoch": 0.5320384982224919, "grad_norm": 1.9168498268351752, "learning_rate": 1e-05, "loss": 1.3669, "step": 767 }, { "epoch": 0.5327321598890141, "grad_norm": 1.945357782268639, "learning_rate": 1e-05, "loss": 1.3595, "step": 768 }, { "epoch": 0.5334258215555363, "grad_norm": 2.237537090117557, "learning_rate": 1e-05, "loss": 1.3683, "step": 769 }, { "epoch": 0.5341194832220585, "grad_norm": 1.8612240335487669, "learning_rate": 1e-05, "loss": 1.3937, "step": 770 }, { "epoch": 0.5348131448885806, "grad_norm": 2.0804814565765266, "learning_rate": 1e-05, "loss": 1.3706, "step": 771 }, { "epoch": 0.5355068065551027, "grad_norm": 1.9951415692701584, "learning_rate": 1e-05, "loss": 1.3696, "step": 772 }, { "epoch": 0.5362004682216249, "grad_norm": 1.7248104660180126, "learning_rate": 1e-05, "loss": 1.3747, "step": 773 }, { "epoch": 0.5368941298881471, "grad_norm": 1.7711620914058126, "learning_rate": 1e-05, "loss": 1.3776, "step": 774 }, { "epoch": 0.5375877915546692, "grad_norm": 1.7502844737194139, "learning_rate": 1e-05, "loss": 1.3719, "step": 775 }, { "epoch": 0.5382814532211914, "grad_norm": 1.975837716180595, "learning_rate": 1e-05, "loss": 1.3392, "step": 776 }, { "epoch": 0.5389751148877135, "grad_norm": 2.079903617254406, "learning_rate": 1e-05, "loss": 1.3789, "step": 777 }, { "epoch": 0.5396687765542356, "grad_norm": 1.8776752694265728, "learning_rate": 1e-05, "loss": 1.3968, "step": 778 }, { "epoch": 0.5403624382207578, "grad_norm": 1.7524287384136423, "learning_rate": 1e-05, "loss": 1.3538, "step": 779 }, { "epoch": 0.54105609988728, "grad_norm": 2.034871602096054, "learning_rate": 1e-05, "loss": 1.3546, "step": 780 }, { "epoch": 0.5417497615538022, "grad_norm": 2.0262231787193117, "learning_rate": 1e-05, "loss": 1.351, "step": 781 }, { "epoch": 0.5424434232203242, "grad_norm": 1.8800790132697096, "learning_rate": 1e-05, "loss": 1.4063, "step": 782 }, { "epoch": 0.5431370848868464, "grad_norm": 1.8397602468821888, "learning_rate": 1e-05, "loss": 1.3284, "step": 783 }, { "epoch": 0.5438307465533686, "grad_norm": 1.7812165456068076, "learning_rate": 1e-05, "loss": 1.3116, "step": 784 }, { "epoch": 0.5445244082198908, "grad_norm": 1.7794765249417972, "learning_rate": 1e-05, "loss": 1.3796, "step": 785 }, { "epoch": 0.5452180698864129, "grad_norm": 1.9715381587594996, "learning_rate": 1e-05, "loss": 1.3771, "step": 786 }, { "epoch": 0.545911731552935, "grad_norm": 1.7931337412184942, "learning_rate": 1e-05, "loss": 1.4108, "step": 787 }, { "epoch": 0.5466053932194572, "grad_norm": 1.8048639309386612, "learning_rate": 1e-05, "loss": 1.3689, "step": 788 }, { "epoch": 0.5472990548859794, "grad_norm": 1.8622884245084281, "learning_rate": 1e-05, "loss": 1.3513, "step": 789 }, { "epoch": 0.5479927165525015, "grad_norm": 1.816207257972127, "learning_rate": 1e-05, "loss": 1.3331, "step": 790 }, { "epoch": 0.5486863782190237, "grad_norm": 1.7508374499546198, "learning_rate": 1e-05, "loss": 1.3594, "step": 791 }, { "epoch": 0.5493800398855458, "grad_norm": 1.7086134745690817, "learning_rate": 1e-05, "loss": 1.343, "step": 792 }, { "epoch": 0.550073701552068, "grad_norm": 1.8826776634822224, "learning_rate": 1e-05, "loss": 1.4312, "step": 793 }, { "epoch": 0.5507673632185901, "grad_norm": 1.85161413953065, "learning_rate": 1e-05, "loss": 1.403, "step": 794 }, { "epoch": 0.5514610248851123, "grad_norm": 2.169026140246512, "learning_rate": 1e-05, "loss": 1.3514, "step": 795 }, { "epoch": 0.5521546865516345, "grad_norm": 1.8974393651417674, "learning_rate": 1e-05, "loss": 1.376, "step": 796 }, { "epoch": 0.5528483482181566, "grad_norm": 1.7818922230701455, "learning_rate": 1e-05, "loss": 1.3365, "step": 797 }, { "epoch": 0.5535420098846787, "grad_norm": 1.8268621071764144, "learning_rate": 1e-05, "loss": 1.3267, "step": 798 }, { "epoch": 0.5542356715512009, "grad_norm": 1.741645703166344, "learning_rate": 1e-05, "loss": 1.3401, "step": 799 }, { "epoch": 0.5549293332177231, "grad_norm": 1.7684014664169725, "learning_rate": 1e-05, "loss": 1.3453, "step": 800 }, { "epoch": 0.5556229948842452, "grad_norm": 1.872082247187476, "learning_rate": 1e-05, "loss": 1.3849, "step": 801 }, { "epoch": 0.5563166565507673, "grad_norm": 1.843965544279036, "learning_rate": 1e-05, "loss": 1.3515, "step": 802 }, { "epoch": 0.5570103182172895, "grad_norm": 1.7503097471047913, "learning_rate": 1e-05, "loss": 1.3772, "step": 803 }, { "epoch": 0.5577039798838117, "grad_norm": 2.152027612117084, "learning_rate": 1e-05, "loss": 1.3369, "step": 804 }, { "epoch": 0.5583976415503338, "grad_norm": 1.9678830918650134, "learning_rate": 1e-05, "loss": 1.3363, "step": 805 }, { "epoch": 0.559091303216856, "grad_norm": 1.7813746189463364, "learning_rate": 1e-05, "loss": 1.3268, "step": 806 }, { "epoch": 0.5597849648833781, "grad_norm": 1.835262281512248, "learning_rate": 1e-05, "loss": 1.3155, "step": 807 }, { "epoch": 0.5604786265499003, "grad_norm": 1.778985993964162, "learning_rate": 1e-05, "loss": 1.3913, "step": 808 }, { "epoch": 0.5611722882164224, "grad_norm": 2.07464881345254, "learning_rate": 1e-05, "loss": 1.35, "step": 809 }, { "epoch": 0.5618659498829446, "grad_norm": 2.186610515953725, "learning_rate": 1e-05, "loss": 1.3954, "step": 810 }, { "epoch": 0.5625596115494668, "grad_norm": 2.129937538785471, "learning_rate": 1e-05, "loss": 1.3644, "step": 811 }, { "epoch": 0.563253273215989, "grad_norm": 2.113811860287663, "learning_rate": 1e-05, "loss": 1.3397, "step": 812 }, { "epoch": 0.563946934882511, "grad_norm": 2.0550884583810487, "learning_rate": 1e-05, "loss": 1.3316, "step": 813 }, { "epoch": 0.5646405965490332, "grad_norm": 1.85558729084722, "learning_rate": 1e-05, "loss": 1.3586, "step": 814 }, { "epoch": 0.5653342582155554, "grad_norm": 1.7940205357205186, "learning_rate": 1e-05, "loss": 1.3328, "step": 815 }, { "epoch": 0.5660279198820776, "grad_norm": 1.8867105415827423, "learning_rate": 1e-05, "loss": 1.4099, "step": 816 }, { "epoch": 0.5667215815485996, "grad_norm": 1.9561860862534044, "learning_rate": 1e-05, "loss": 1.4491, "step": 817 }, { "epoch": 0.5674152432151218, "grad_norm": 2.0876714983856313, "learning_rate": 1e-05, "loss": 1.364, "step": 818 }, { "epoch": 0.568108904881644, "grad_norm": 1.99399680820796, "learning_rate": 1e-05, "loss": 1.3033, "step": 819 }, { "epoch": 0.5688025665481661, "grad_norm": 1.976493240480479, "learning_rate": 1e-05, "loss": 1.333, "step": 820 }, { "epoch": 0.5694962282146883, "grad_norm": 1.95662043100926, "learning_rate": 1e-05, "loss": 1.3136, "step": 821 }, { "epoch": 0.5701898898812104, "grad_norm": 1.6497733053601713, "learning_rate": 1e-05, "loss": 1.3861, "step": 822 }, { "epoch": 0.5708835515477326, "grad_norm": 1.818935374314111, "learning_rate": 1e-05, "loss": 1.3857, "step": 823 }, { "epoch": 0.5715772132142547, "grad_norm": 1.747967079872631, "learning_rate": 1e-05, "loss": 1.349, "step": 824 }, { "epoch": 0.5722708748807769, "grad_norm": 1.8908128266598878, "learning_rate": 1e-05, "loss": 1.3616, "step": 825 }, { "epoch": 0.5729645365472991, "grad_norm": 1.8720455934356435, "learning_rate": 1e-05, "loss": 1.3238, "step": 826 }, { "epoch": 0.5736581982138212, "grad_norm": 1.7833794579975666, "learning_rate": 1e-05, "loss": 1.3956, "step": 827 }, { "epoch": 0.5743518598803433, "grad_norm": 2.1032994082238203, "learning_rate": 1e-05, "loss": 1.3575, "step": 828 }, { "epoch": 0.5750455215468655, "grad_norm": 2.0134774144141487, "learning_rate": 1e-05, "loss": 1.3464, "step": 829 }, { "epoch": 0.5757391832133877, "grad_norm": 1.8711603789528202, "learning_rate": 1e-05, "loss": 1.4115, "step": 830 }, { "epoch": 0.5764328448799099, "grad_norm": 1.7819151783946192, "learning_rate": 1e-05, "loss": 1.3615, "step": 831 }, { "epoch": 0.5771265065464319, "grad_norm": 1.9690584032628007, "learning_rate": 1e-05, "loss": 1.3901, "step": 832 }, { "epoch": 0.5778201682129541, "grad_norm": 1.9642954115887026, "learning_rate": 1e-05, "loss": 1.3518, "step": 833 }, { "epoch": 0.5785138298794763, "grad_norm": 1.744497422769626, "learning_rate": 1e-05, "loss": 1.3469, "step": 834 }, { "epoch": 0.5792074915459985, "grad_norm": 2.0755337336634767, "learning_rate": 1e-05, "loss": 1.3823, "step": 835 }, { "epoch": 0.5799011532125206, "grad_norm": 1.8340533083393944, "learning_rate": 1e-05, "loss": 1.374, "step": 836 }, { "epoch": 0.5805948148790427, "grad_norm": 1.7560881859736863, "learning_rate": 1e-05, "loss": 1.4051, "step": 837 }, { "epoch": 0.5812884765455649, "grad_norm": 1.7470644935895128, "learning_rate": 1e-05, "loss": 1.3751, "step": 838 }, { "epoch": 0.5819821382120871, "grad_norm": 1.6498957999043187, "learning_rate": 1e-05, "loss": 1.3425, "step": 839 }, { "epoch": 0.5826757998786092, "grad_norm": 1.708284581084529, "learning_rate": 1e-05, "loss": 1.3772, "step": 840 }, { "epoch": 0.5833694615451314, "grad_norm": 1.7341358375022322, "learning_rate": 1e-05, "loss": 1.4271, "step": 841 }, { "epoch": 0.5840631232116535, "grad_norm": 1.96090088265045, "learning_rate": 1e-05, "loss": 1.3811, "step": 842 }, { "epoch": 0.5847567848781756, "grad_norm": 1.8309156039915615, "learning_rate": 1e-05, "loss": 1.3518, "step": 843 }, { "epoch": 0.5854504465446978, "grad_norm": 1.766087750027872, "learning_rate": 1e-05, "loss": 1.2791, "step": 844 }, { "epoch": 0.58614410821122, "grad_norm": 1.840859949308765, "learning_rate": 1e-05, "loss": 1.3364, "step": 845 }, { "epoch": 0.5868377698777422, "grad_norm": 1.8811612648079954, "learning_rate": 1e-05, "loss": 1.3866, "step": 846 }, { "epoch": 0.5875314315442642, "grad_norm": 1.7882571243218808, "learning_rate": 1e-05, "loss": 1.3818, "step": 847 }, { "epoch": 0.5882250932107864, "grad_norm": 1.7751021276239416, "learning_rate": 1e-05, "loss": 1.3877, "step": 848 }, { "epoch": 0.5889187548773086, "grad_norm": 1.863598117479808, "learning_rate": 1e-05, "loss": 1.3283, "step": 849 }, { "epoch": 0.5896124165438308, "grad_norm": 1.7934939156618281, "learning_rate": 1e-05, "loss": 1.3909, "step": 850 }, { "epoch": 0.5903060782103529, "grad_norm": 1.8701501570430885, "learning_rate": 1e-05, "loss": 1.3911, "step": 851 }, { "epoch": 0.590999739876875, "grad_norm": 1.8075702135316054, "learning_rate": 1e-05, "loss": 1.3267, "step": 852 }, { "epoch": 0.5916934015433972, "grad_norm": 1.7844322891888966, "learning_rate": 1e-05, "loss": 1.3749, "step": 853 }, { "epoch": 0.5923870632099194, "grad_norm": 1.6364137503955962, "learning_rate": 1e-05, "loss": 1.3903, "step": 854 }, { "epoch": 0.5930807248764415, "grad_norm": 1.845018525439836, "learning_rate": 1e-05, "loss": 1.3638, "step": 855 }, { "epoch": 0.5937743865429637, "grad_norm": 1.6438686492795926, "learning_rate": 1e-05, "loss": 1.3356, "step": 856 }, { "epoch": 0.5944680482094858, "grad_norm": 1.8461296262824984, "learning_rate": 1e-05, "loss": 1.318, "step": 857 }, { "epoch": 0.595161709876008, "grad_norm": 1.7523581116541502, "learning_rate": 1e-05, "loss": 1.3278, "step": 858 }, { "epoch": 0.5958553715425301, "grad_norm": 1.7821675206089143, "learning_rate": 1e-05, "loss": 1.3257, "step": 859 }, { "epoch": 0.5965490332090523, "grad_norm": 1.8952855386403753, "learning_rate": 1e-05, "loss": 1.352, "step": 860 }, { "epoch": 0.5972426948755745, "grad_norm": 1.9061495251459173, "learning_rate": 1e-05, "loss": 1.361, "step": 861 }, { "epoch": 0.5979363565420966, "grad_norm": 1.807002944323855, "learning_rate": 1e-05, "loss": 1.3498, "step": 862 }, { "epoch": 0.5986300182086187, "grad_norm": 1.8427398989259318, "learning_rate": 1e-05, "loss": 1.3442, "step": 863 }, { "epoch": 0.5993236798751409, "grad_norm": 2.159582705739885, "learning_rate": 1e-05, "loss": 1.4137, "step": 864 }, { "epoch": 0.6000173415416631, "grad_norm": 1.8186698531077894, "learning_rate": 1e-05, "loss": 1.3743, "step": 865 }, { "epoch": 0.6007110032081852, "grad_norm": 1.7754924011854336, "learning_rate": 1e-05, "loss": 1.3758, "step": 866 }, { "epoch": 0.6014046648747073, "grad_norm": 1.7146434615480912, "learning_rate": 1e-05, "loss": 1.3655, "step": 867 }, { "epoch": 0.6020983265412295, "grad_norm": 1.6864187498892416, "learning_rate": 1e-05, "loss": 1.3842, "step": 868 }, { "epoch": 0.6027919882077517, "grad_norm": 1.7625940182372917, "learning_rate": 1e-05, "loss": 1.3563, "step": 869 }, { "epoch": 0.6034856498742738, "grad_norm": 1.8385401005573274, "learning_rate": 1e-05, "loss": 1.317, "step": 870 }, { "epoch": 0.604179311540796, "grad_norm": 1.8249865396470937, "learning_rate": 1e-05, "loss": 1.3666, "step": 871 }, { "epoch": 0.6048729732073181, "grad_norm": 1.774024094473835, "learning_rate": 1e-05, "loss": 1.3677, "step": 872 }, { "epoch": 0.6055666348738403, "grad_norm": 1.9300061255791403, "learning_rate": 1e-05, "loss": 1.3668, "step": 873 }, { "epoch": 0.6062602965403624, "grad_norm": 1.8538055069719046, "learning_rate": 1e-05, "loss": 1.3639, "step": 874 }, { "epoch": 0.6069539582068846, "grad_norm": 1.8234801513872547, "learning_rate": 1e-05, "loss": 1.3183, "step": 875 }, { "epoch": 0.6076476198734068, "grad_norm": 1.8887104952070137, "learning_rate": 1e-05, "loss": 1.3798, "step": 876 }, { "epoch": 0.608341281539929, "grad_norm": 1.967685873156897, "learning_rate": 1e-05, "loss": 1.3733, "step": 877 }, { "epoch": 0.609034943206451, "grad_norm": 1.8930452243619467, "learning_rate": 1e-05, "loss": 1.3627, "step": 878 }, { "epoch": 0.6097286048729732, "grad_norm": 1.9917808148339706, "learning_rate": 1e-05, "loss": 1.3637, "step": 879 }, { "epoch": 0.6104222665394954, "grad_norm": 1.7953716355637714, "learning_rate": 1e-05, "loss": 1.3552, "step": 880 }, { "epoch": 0.6111159282060176, "grad_norm": 2.054090380165633, "learning_rate": 1e-05, "loss": 1.3309, "step": 881 }, { "epoch": 0.6118095898725396, "grad_norm": 1.774125479388038, "learning_rate": 1e-05, "loss": 1.3346, "step": 882 }, { "epoch": 0.6125032515390618, "grad_norm": 1.788110427730304, "learning_rate": 1e-05, "loss": 1.3907, "step": 883 }, { "epoch": 0.613196913205584, "grad_norm": 1.9117105836931287, "learning_rate": 1e-05, "loss": 1.3662, "step": 884 }, { "epoch": 0.6138905748721061, "grad_norm": 1.9761438291924842, "learning_rate": 1e-05, "loss": 1.3487, "step": 885 }, { "epoch": 0.6145842365386283, "grad_norm": 1.7939497044726074, "learning_rate": 1e-05, "loss": 1.3617, "step": 886 }, { "epoch": 0.6152778982051504, "grad_norm": 1.8137308254099254, "learning_rate": 1e-05, "loss": 1.4017, "step": 887 }, { "epoch": 0.6159715598716726, "grad_norm": 1.9358335601206476, "learning_rate": 1e-05, "loss": 1.3221, "step": 888 }, { "epoch": 0.6166652215381947, "grad_norm": 1.8023857684374447, "learning_rate": 1e-05, "loss": 1.3645, "step": 889 }, { "epoch": 0.6173588832047169, "grad_norm": 1.7858631263519855, "learning_rate": 1e-05, "loss": 1.3915, "step": 890 }, { "epoch": 0.6180525448712391, "grad_norm": 1.8447027974665438, "learning_rate": 1e-05, "loss": 1.3331, "step": 891 }, { "epoch": 0.6187462065377612, "grad_norm": 1.8973843442097094, "learning_rate": 1e-05, "loss": 1.3053, "step": 892 }, { "epoch": 0.6194398682042833, "grad_norm": 1.7623453790808643, "learning_rate": 1e-05, "loss": 1.2926, "step": 893 }, { "epoch": 0.6201335298708055, "grad_norm": 1.8017523463560905, "learning_rate": 1e-05, "loss": 1.3647, "step": 894 }, { "epoch": 0.6208271915373277, "grad_norm": 1.7450544611778176, "learning_rate": 1e-05, "loss": 1.3658, "step": 895 }, { "epoch": 0.6215208532038499, "grad_norm": 1.823024100294463, "learning_rate": 1e-05, "loss": 1.3508, "step": 896 }, { "epoch": 0.6222145148703719, "grad_norm": 1.8524166096635302, "learning_rate": 1e-05, "loss": 1.3446, "step": 897 }, { "epoch": 0.6229081765368941, "grad_norm": 1.7203042262118677, "learning_rate": 1e-05, "loss": 1.3838, "step": 898 }, { "epoch": 0.6236018382034163, "grad_norm": 1.7896042667571013, "learning_rate": 1e-05, "loss": 1.3401, "step": 899 }, { "epoch": 0.6242954998699385, "grad_norm": 1.8566869792112495, "learning_rate": 1e-05, "loss": 1.3856, "step": 900 }, { "epoch": 0.6249891615364606, "grad_norm": 1.7236631672011284, "learning_rate": 1e-05, "loss": 1.3455, "step": 901 }, { "epoch": 0.6256828232029827, "grad_norm": 1.9857281332079058, "learning_rate": 1e-05, "loss": 1.3347, "step": 902 }, { "epoch": 0.6263764848695049, "grad_norm": 1.916490049064551, "learning_rate": 1e-05, "loss": 1.2964, "step": 903 }, { "epoch": 0.6270701465360271, "grad_norm": 1.9713572657543152, "learning_rate": 1e-05, "loss": 1.3302, "step": 904 }, { "epoch": 0.6277638082025492, "grad_norm": 1.7610441348646735, "learning_rate": 1e-05, "loss": 1.3344, "step": 905 }, { "epoch": 0.6284574698690714, "grad_norm": 1.799617843727853, "learning_rate": 1e-05, "loss": 1.3284, "step": 906 }, { "epoch": 0.6291511315355935, "grad_norm": 1.7832143262655586, "learning_rate": 1e-05, "loss": 1.3356, "step": 907 }, { "epoch": 0.6298447932021156, "grad_norm": 1.7971828284576865, "learning_rate": 1e-05, "loss": 1.3207, "step": 908 }, { "epoch": 0.6305384548686378, "grad_norm": 1.7536423641131738, "learning_rate": 1e-05, "loss": 1.3499, "step": 909 }, { "epoch": 0.63123211653516, "grad_norm": 1.7401477745871217, "learning_rate": 1e-05, "loss": 1.3398, "step": 910 }, { "epoch": 0.6319257782016822, "grad_norm": 1.7218362910164169, "learning_rate": 1e-05, "loss": 1.4193, "step": 911 }, { "epoch": 0.6326194398682042, "grad_norm": 1.9808528889101304, "learning_rate": 1e-05, "loss": 1.392, "step": 912 }, { "epoch": 0.6333131015347264, "grad_norm": 1.9064408882431807, "learning_rate": 1e-05, "loss": 1.3615, "step": 913 }, { "epoch": 0.6340067632012486, "grad_norm": 1.7248366549906144, "learning_rate": 1e-05, "loss": 1.3231, "step": 914 }, { "epoch": 0.6347004248677708, "grad_norm": 1.795395048611617, "learning_rate": 1e-05, "loss": 1.3851, "step": 915 }, { "epoch": 0.6353940865342929, "grad_norm": 1.7344884424253888, "learning_rate": 1e-05, "loss": 1.3528, "step": 916 }, { "epoch": 0.636087748200815, "grad_norm": 1.700061571947052, "learning_rate": 1e-05, "loss": 1.3348, "step": 917 }, { "epoch": 0.6367814098673372, "grad_norm": 1.8646566885856952, "learning_rate": 1e-05, "loss": 1.357, "step": 918 }, { "epoch": 0.6374750715338594, "grad_norm": 1.8652265942281396, "learning_rate": 1e-05, "loss": 1.3343, "step": 919 }, { "epoch": 0.6381687332003815, "grad_norm": 1.7925656802981118, "learning_rate": 1e-05, "loss": 1.3308, "step": 920 }, { "epoch": 0.6388623948669037, "grad_norm": 1.7896253293176538, "learning_rate": 1e-05, "loss": 1.3813, "step": 921 }, { "epoch": 0.6395560565334258, "grad_norm": 1.6534489942501098, "learning_rate": 1e-05, "loss": 1.3233, "step": 922 }, { "epoch": 0.640249718199948, "grad_norm": 1.8033026812852484, "learning_rate": 1e-05, "loss": 1.3074, "step": 923 }, { "epoch": 0.6409433798664701, "grad_norm": 1.822451647210804, "learning_rate": 1e-05, "loss": 1.3217, "step": 924 }, { "epoch": 0.6416370415329923, "grad_norm": 1.7821886850533442, "learning_rate": 1e-05, "loss": 1.3531, "step": 925 }, { "epoch": 0.6423307031995145, "grad_norm": 1.8446137766083273, "learning_rate": 1e-05, "loss": 1.3552, "step": 926 }, { "epoch": 0.6430243648660365, "grad_norm": 1.7365041612964318, "learning_rate": 1e-05, "loss": 1.3472, "step": 927 }, { "epoch": 0.6437180265325587, "grad_norm": 1.9116647427342783, "learning_rate": 1e-05, "loss": 1.3356, "step": 928 }, { "epoch": 0.6444116881990809, "grad_norm": 1.6818488752250975, "learning_rate": 1e-05, "loss": 1.35, "step": 929 }, { "epoch": 0.6451053498656031, "grad_norm": 1.8566305164008303, "learning_rate": 1e-05, "loss": 1.3643, "step": 930 }, { "epoch": 0.6457990115321252, "grad_norm": 1.7420953544687154, "learning_rate": 1e-05, "loss": 1.3148, "step": 931 }, { "epoch": 0.6464926731986473, "grad_norm": 1.885744413844102, "learning_rate": 1e-05, "loss": 1.3492, "step": 932 }, { "epoch": 0.6471863348651695, "grad_norm": 1.7944270298154161, "learning_rate": 1e-05, "loss": 1.3513, "step": 933 }, { "epoch": 0.6478799965316917, "grad_norm": 1.7535218523742484, "learning_rate": 1e-05, "loss": 1.4101, "step": 934 }, { "epoch": 0.6485736581982138, "grad_norm": 1.7229852034518358, "learning_rate": 1e-05, "loss": 1.3437, "step": 935 }, { "epoch": 0.649267319864736, "grad_norm": 1.896304422647214, "learning_rate": 1e-05, "loss": 1.3156, "step": 936 }, { "epoch": 0.6499609815312581, "grad_norm": 1.8055244846850502, "learning_rate": 1e-05, "loss": 1.3691, "step": 937 }, { "epoch": 0.6506546431977803, "grad_norm": 1.684941036557295, "learning_rate": 1e-05, "loss": 1.3579, "step": 938 }, { "epoch": 0.6513483048643024, "grad_norm": 1.8888571415510795, "learning_rate": 1e-05, "loss": 1.3632, "step": 939 }, { "epoch": 0.6520419665308246, "grad_norm": 1.8160274262290288, "learning_rate": 1e-05, "loss": 1.3099, "step": 940 }, { "epoch": 0.6527356281973468, "grad_norm": 1.755049632438486, "learning_rate": 1e-05, "loss": 1.3519, "step": 941 }, { "epoch": 0.653429289863869, "grad_norm": 1.89712944315266, "learning_rate": 1e-05, "loss": 1.3494, "step": 942 }, { "epoch": 0.654122951530391, "grad_norm": 1.765188366032801, "learning_rate": 1e-05, "loss": 1.3856, "step": 943 }, { "epoch": 0.6548166131969132, "grad_norm": 1.9497117765562002, "learning_rate": 1e-05, "loss": 1.3665, "step": 944 }, { "epoch": 0.6555102748634354, "grad_norm": 1.799101172711031, "learning_rate": 1e-05, "loss": 1.3445, "step": 945 }, { "epoch": 0.6562039365299576, "grad_norm": 1.712616408878392, "learning_rate": 1e-05, "loss": 1.3491, "step": 946 }, { "epoch": 0.6568975981964796, "grad_norm": 1.7946208261432808, "learning_rate": 1e-05, "loss": 1.3731, "step": 947 }, { "epoch": 0.6575912598630018, "grad_norm": 1.7262699314904466, "learning_rate": 1e-05, "loss": 1.3585, "step": 948 }, { "epoch": 0.658284921529524, "grad_norm": 1.9628450564778277, "learning_rate": 1e-05, "loss": 1.301, "step": 949 }, { "epoch": 0.6589785831960461, "grad_norm": 1.8061202922829884, "learning_rate": 1e-05, "loss": 1.3783, "step": 950 }, { "epoch": 0.6596722448625683, "grad_norm": 1.7121738551781767, "learning_rate": 1e-05, "loss": 1.357, "step": 951 }, { "epoch": 0.6603659065290904, "grad_norm": 1.8850562141039617, "learning_rate": 1e-05, "loss": 1.2819, "step": 952 }, { "epoch": 0.6610595681956126, "grad_norm": 1.867193802881424, "learning_rate": 1e-05, "loss": 1.3156, "step": 953 }, { "epoch": 0.6617532298621347, "grad_norm": 1.7443179431377005, "learning_rate": 1e-05, "loss": 1.2751, "step": 954 }, { "epoch": 0.6624468915286569, "grad_norm": 1.8733131607506688, "learning_rate": 1e-05, "loss": 1.3534, "step": 955 }, { "epoch": 0.6631405531951791, "grad_norm": 1.9784306105729255, "learning_rate": 1e-05, "loss": 1.2742, "step": 956 }, { "epoch": 0.6638342148617012, "grad_norm": 1.8959702823237385, "learning_rate": 1e-05, "loss": 1.4094, "step": 957 }, { "epoch": 0.6645278765282233, "grad_norm": 1.664080974193892, "learning_rate": 1e-05, "loss": 1.3658, "step": 958 }, { "epoch": 0.6652215381947455, "grad_norm": 1.782008443874851, "learning_rate": 1e-05, "loss": 1.347, "step": 959 }, { "epoch": 0.6659151998612677, "grad_norm": 1.8460350587229146, "learning_rate": 1e-05, "loss": 1.3639, "step": 960 }, { "epoch": 0.6666088615277899, "grad_norm": 1.6425972064330443, "learning_rate": 1e-05, "loss": 1.3693, "step": 961 }, { "epoch": 0.6673025231943119, "grad_norm": 1.8565662897573758, "learning_rate": 1e-05, "loss": 1.3636, "step": 962 }, { "epoch": 0.6679961848608341, "grad_norm": 1.7302118661778385, "learning_rate": 1e-05, "loss": 1.3566, "step": 963 }, { "epoch": 0.6686898465273563, "grad_norm": 1.9616014623863918, "learning_rate": 1e-05, "loss": 1.374, "step": 964 }, { "epoch": 0.6693835081938785, "grad_norm": 1.84734239559959, "learning_rate": 1e-05, "loss": 1.2604, "step": 965 }, { "epoch": 0.6700771698604006, "grad_norm": 1.6679061078794732, "learning_rate": 1e-05, "loss": 1.344, "step": 966 }, { "epoch": 0.6707708315269227, "grad_norm": 1.9249982922421873, "learning_rate": 1e-05, "loss": 1.3773, "step": 967 }, { "epoch": 0.6714644931934449, "grad_norm": 1.9130662887594385, "learning_rate": 1e-05, "loss": 1.3389, "step": 968 }, { "epoch": 0.6721581548599671, "grad_norm": 2.0490408445129553, "learning_rate": 1e-05, "loss": 1.3271, "step": 969 }, { "epoch": 0.6728518165264892, "grad_norm": 1.7966765927847634, "learning_rate": 1e-05, "loss": 1.3082, "step": 970 }, { "epoch": 0.6735454781930114, "grad_norm": 1.8365578777762348, "learning_rate": 1e-05, "loss": 1.3813, "step": 971 }, { "epoch": 0.6742391398595335, "grad_norm": 1.7856135676874048, "learning_rate": 1e-05, "loss": 1.3791, "step": 972 }, { "epoch": 0.6749328015260556, "grad_norm": 1.7189842050358264, "learning_rate": 1e-05, "loss": 1.3859, "step": 973 }, { "epoch": 0.6756264631925778, "grad_norm": 1.697736604178591, "learning_rate": 1e-05, "loss": 1.3531, "step": 974 }, { "epoch": 0.6763201248591, "grad_norm": 1.85446433118358, "learning_rate": 1e-05, "loss": 1.3377, "step": 975 }, { "epoch": 0.6770137865256222, "grad_norm": 1.817734345502182, "learning_rate": 1e-05, "loss": 1.3316, "step": 976 }, { "epoch": 0.6777074481921442, "grad_norm": 1.6362921639652548, "learning_rate": 1e-05, "loss": 1.2879, "step": 977 }, { "epoch": 0.6784011098586664, "grad_norm": 1.7236748146694982, "learning_rate": 1e-05, "loss": 1.3023, "step": 978 }, { "epoch": 0.6790947715251886, "grad_norm": 1.874808634354032, "learning_rate": 1e-05, "loss": 1.3291, "step": 979 }, { "epoch": 0.6797884331917108, "grad_norm": 2.143605571091092, "learning_rate": 1e-05, "loss": 1.3305, "step": 980 }, { "epoch": 0.6804820948582329, "grad_norm": 1.7140845838212821, "learning_rate": 1e-05, "loss": 1.2762, "step": 981 }, { "epoch": 0.681175756524755, "grad_norm": 1.7627892247793258, "learning_rate": 1e-05, "loss": 1.3793, "step": 982 }, { "epoch": 0.6818694181912772, "grad_norm": 1.9327806260640557, "learning_rate": 1e-05, "loss": 1.3314, "step": 983 }, { "epoch": 0.6825630798577994, "grad_norm": 1.9570006600139125, "learning_rate": 1e-05, "loss": 1.3092, "step": 984 }, { "epoch": 0.6832567415243215, "grad_norm": 1.857815246395867, "learning_rate": 1e-05, "loss": 1.328, "step": 985 }, { "epoch": 0.6839504031908437, "grad_norm": 1.6521408115836034, "learning_rate": 1e-05, "loss": 1.3649, "step": 986 }, { "epoch": 0.6846440648573658, "grad_norm": 1.6883217532733774, "learning_rate": 1e-05, "loss": 1.3742, "step": 987 }, { "epoch": 0.685337726523888, "grad_norm": 1.7657705434647315, "learning_rate": 1e-05, "loss": 1.4009, "step": 988 }, { "epoch": 0.6860313881904101, "grad_norm": 1.8388470770976078, "learning_rate": 1e-05, "loss": 1.3385, "step": 989 }, { "epoch": 0.6867250498569323, "grad_norm": 1.6446394362553027, "learning_rate": 1e-05, "loss": 1.3747, "step": 990 }, { "epoch": 0.6874187115234545, "grad_norm": 1.76332053954708, "learning_rate": 1e-05, "loss": 1.3744, "step": 991 }, { "epoch": 0.6881123731899765, "grad_norm": 1.8551504155963352, "learning_rate": 1e-05, "loss": 1.3461, "step": 992 }, { "epoch": 0.6888060348564987, "grad_norm": 1.8242720423216203, "learning_rate": 1e-05, "loss": 1.3949, "step": 993 }, { "epoch": 0.6894996965230209, "grad_norm": 1.8728688560334699, "learning_rate": 1e-05, "loss": 1.4202, "step": 994 }, { "epoch": 0.6901933581895431, "grad_norm": 1.8128080132317514, "learning_rate": 1e-05, "loss": 1.3528, "step": 995 }, { "epoch": 0.6908870198560652, "grad_norm": 1.7706897683233593, "learning_rate": 1e-05, "loss": 1.3772, "step": 996 }, { "epoch": 0.6915806815225873, "grad_norm": 1.867842253838222, "learning_rate": 1e-05, "loss": 1.3735, "step": 997 }, { "epoch": 0.6922743431891095, "grad_norm": 1.908519372133083, "learning_rate": 1e-05, "loss": 1.3093, "step": 998 }, { "epoch": 0.6929680048556317, "grad_norm": 1.7680738442803956, "learning_rate": 1e-05, "loss": 1.3205, "step": 999 }, { "epoch": 0.6936616665221538, "grad_norm": 1.752432098173259, "learning_rate": 1e-05, "loss": 1.3451, "step": 1000 }, { "epoch": 0.694355328188676, "grad_norm": 1.7866959543838525, "learning_rate": 1e-05, "loss": 1.3288, "step": 1001 }, { "epoch": 0.6950489898551981, "grad_norm": 1.6320555911640122, "learning_rate": 1e-05, "loss": 1.3157, "step": 1002 }, { "epoch": 0.6957426515217203, "grad_norm": 1.7766379486245896, "learning_rate": 1e-05, "loss": 1.3559, "step": 1003 }, { "epoch": 0.6964363131882424, "grad_norm": 1.7801227267982318, "learning_rate": 1e-05, "loss": 1.389, "step": 1004 }, { "epoch": 0.6971299748547646, "grad_norm": 1.7763939485733111, "learning_rate": 1e-05, "loss": 1.4214, "step": 1005 }, { "epoch": 0.6978236365212868, "grad_norm": 1.7466154961438336, "learning_rate": 1e-05, "loss": 1.3974, "step": 1006 }, { "epoch": 0.6985172981878089, "grad_norm": 1.8018054751465553, "learning_rate": 1e-05, "loss": 1.3146, "step": 1007 }, { "epoch": 0.699210959854331, "grad_norm": 1.7015362135443022, "learning_rate": 1e-05, "loss": 1.3521, "step": 1008 }, { "epoch": 0.6999046215208532, "grad_norm": 1.8044732359887248, "learning_rate": 1e-05, "loss": 1.3298, "step": 1009 }, { "epoch": 0.7005982831873754, "grad_norm": 1.7248926110752036, "learning_rate": 1e-05, "loss": 1.311, "step": 1010 }, { "epoch": 0.7012919448538976, "grad_norm": 1.7408526444267358, "learning_rate": 1e-05, "loss": 1.3346, "step": 1011 }, { "epoch": 0.7019856065204196, "grad_norm": 1.756432069962424, "learning_rate": 1e-05, "loss": 1.3608, "step": 1012 }, { "epoch": 0.7026792681869418, "grad_norm": 1.9004070384049725, "learning_rate": 1e-05, "loss": 1.3061, "step": 1013 }, { "epoch": 0.703372929853464, "grad_norm": 1.7674772646104595, "learning_rate": 1e-05, "loss": 1.3365, "step": 1014 }, { "epoch": 0.7040665915199861, "grad_norm": 1.815286596926447, "learning_rate": 1e-05, "loss": 1.3112, "step": 1015 }, { "epoch": 0.7047602531865083, "grad_norm": 1.9200058514873535, "learning_rate": 1e-05, "loss": 1.3702, "step": 1016 }, { "epoch": 0.7054539148530304, "grad_norm": 1.7499367861528972, "learning_rate": 1e-05, "loss": 1.3707, "step": 1017 }, { "epoch": 0.7061475765195526, "grad_norm": 1.925251587075512, "learning_rate": 1e-05, "loss": 1.3208, "step": 1018 }, { "epoch": 0.7068412381860747, "grad_norm": 1.7154198796482498, "learning_rate": 1e-05, "loss": 1.3336, "step": 1019 }, { "epoch": 0.7075348998525969, "grad_norm": 1.837360393002266, "learning_rate": 1e-05, "loss": 1.3328, "step": 1020 }, { "epoch": 0.7082285615191191, "grad_norm": 1.6211349139215232, "learning_rate": 1e-05, "loss": 1.3284, "step": 1021 }, { "epoch": 0.7089222231856412, "grad_norm": 1.866016563395064, "learning_rate": 1e-05, "loss": 1.3198, "step": 1022 }, { "epoch": 0.7096158848521633, "grad_norm": 1.6839566806665383, "learning_rate": 1e-05, "loss": 1.3651, "step": 1023 }, { "epoch": 0.7103095465186855, "grad_norm": 1.7159632620855965, "learning_rate": 1e-05, "loss": 1.3268, "step": 1024 }, { "epoch": 0.7110032081852077, "grad_norm": 1.806422188485046, "learning_rate": 1e-05, "loss": 1.2901, "step": 1025 }, { "epoch": 0.7116968698517299, "grad_norm": 1.7329697047767731, "learning_rate": 1e-05, "loss": 1.278, "step": 1026 }, { "epoch": 0.7123905315182519, "grad_norm": 1.7358387057504157, "learning_rate": 1e-05, "loss": 1.3959, "step": 1027 }, { "epoch": 0.7130841931847741, "grad_norm": 1.7843805164975317, "learning_rate": 1e-05, "loss": 1.3158, "step": 1028 }, { "epoch": 0.7137778548512963, "grad_norm": 1.8034487032951743, "learning_rate": 1e-05, "loss": 1.3239, "step": 1029 }, { "epoch": 0.7144715165178185, "grad_norm": 1.7806257674138806, "learning_rate": 1e-05, "loss": 1.348, "step": 1030 }, { "epoch": 0.7151651781843406, "grad_norm": 1.8562808148693768, "learning_rate": 1e-05, "loss": 1.3675, "step": 1031 }, { "epoch": 0.7158588398508627, "grad_norm": 1.9354803733254098, "learning_rate": 1e-05, "loss": 1.3331, "step": 1032 }, { "epoch": 0.7165525015173849, "grad_norm": 1.7581325487197559, "learning_rate": 1e-05, "loss": 1.3468, "step": 1033 }, { "epoch": 0.7172461631839071, "grad_norm": 1.802989571777425, "learning_rate": 1e-05, "loss": 1.3429, "step": 1034 }, { "epoch": 0.7179398248504292, "grad_norm": 1.7427548981299272, "learning_rate": 1e-05, "loss": 1.3555, "step": 1035 }, { "epoch": 0.7186334865169514, "grad_norm": 1.9113368723599877, "learning_rate": 1e-05, "loss": 1.3458, "step": 1036 }, { "epoch": 0.7193271481834735, "grad_norm": 1.746074482528233, "learning_rate": 1e-05, "loss": 1.256, "step": 1037 }, { "epoch": 0.7200208098499956, "grad_norm": 1.8354380987152568, "learning_rate": 1e-05, "loss": 1.336, "step": 1038 }, { "epoch": 0.7207144715165178, "grad_norm": 1.945273024066637, "learning_rate": 1e-05, "loss": 1.3985, "step": 1039 }, { "epoch": 0.72140813318304, "grad_norm": 1.843452934712379, "learning_rate": 1e-05, "loss": 1.3193, "step": 1040 }, { "epoch": 0.7221017948495622, "grad_norm": 1.9504440192269308, "learning_rate": 1e-05, "loss": 1.3296, "step": 1041 }, { "epoch": 0.7227954565160842, "grad_norm": 1.8706540071171702, "learning_rate": 1e-05, "loss": 1.3403, "step": 1042 }, { "epoch": 0.7234891181826064, "grad_norm": 1.9488283001981537, "learning_rate": 1e-05, "loss": 1.3245, "step": 1043 }, { "epoch": 0.7241827798491286, "grad_norm": 1.6969025009984984, "learning_rate": 1e-05, "loss": 1.3041, "step": 1044 }, { "epoch": 0.7248764415156508, "grad_norm": 2.0072497667921443, "learning_rate": 1e-05, "loss": 1.3447, "step": 1045 }, { "epoch": 0.7255701031821729, "grad_norm": 1.8741053070227567, "learning_rate": 1e-05, "loss": 1.3814, "step": 1046 }, { "epoch": 0.726263764848695, "grad_norm": 1.8874635982516978, "learning_rate": 1e-05, "loss": 1.328, "step": 1047 }, { "epoch": 0.7269574265152172, "grad_norm": 1.8011724514294678, "learning_rate": 1e-05, "loss": 1.3165, "step": 1048 }, { "epoch": 0.7276510881817394, "grad_norm": 1.7615037862154284, "learning_rate": 1e-05, "loss": 1.3405, "step": 1049 }, { "epoch": 0.7283447498482615, "grad_norm": 1.8936590686337285, "learning_rate": 1e-05, "loss": 1.3304, "step": 1050 }, { "epoch": 0.7290384115147837, "grad_norm": 1.7561903582794731, "learning_rate": 1e-05, "loss": 1.3135, "step": 1051 }, { "epoch": 0.7297320731813058, "grad_norm": 1.9910837825376402, "learning_rate": 1e-05, "loss": 1.3328, "step": 1052 }, { "epoch": 0.730425734847828, "grad_norm": 1.7932757018881984, "learning_rate": 1e-05, "loss": 1.3244, "step": 1053 }, { "epoch": 0.7311193965143501, "grad_norm": 1.8115170033234536, "learning_rate": 1e-05, "loss": 1.3256, "step": 1054 }, { "epoch": 0.7318130581808723, "grad_norm": 1.846025824397134, "learning_rate": 1e-05, "loss": 1.2897, "step": 1055 }, { "epoch": 0.7325067198473945, "grad_norm": 1.8145673138145253, "learning_rate": 1e-05, "loss": 1.3333, "step": 1056 }, { "epoch": 0.7332003815139165, "grad_norm": 1.6430330564527496, "learning_rate": 1e-05, "loss": 1.3569, "step": 1057 }, { "epoch": 0.7338940431804387, "grad_norm": 1.7140027949989802, "learning_rate": 1e-05, "loss": 1.3096, "step": 1058 }, { "epoch": 0.7345877048469609, "grad_norm": 1.9676800557132401, "learning_rate": 1e-05, "loss": 1.3327, "step": 1059 }, { "epoch": 0.7352813665134831, "grad_norm": 1.76220089185509, "learning_rate": 1e-05, "loss": 1.3837, "step": 1060 }, { "epoch": 0.7359750281800052, "grad_norm": 1.8650314152543974, "learning_rate": 1e-05, "loss": 1.31, "step": 1061 }, { "epoch": 0.7366686898465273, "grad_norm": 2.1105903583633054, "learning_rate": 1e-05, "loss": 1.3189, "step": 1062 }, { "epoch": 0.7373623515130495, "grad_norm": 1.9330298633028662, "learning_rate": 1e-05, "loss": 1.3116, "step": 1063 }, { "epoch": 0.7380560131795717, "grad_norm": 1.8697841477985335, "learning_rate": 1e-05, "loss": 1.3059, "step": 1064 }, { "epoch": 0.7387496748460938, "grad_norm": 1.8167855532087722, "learning_rate": 1e-05, "loss": 1.3725, "step": 1065 }, { "epoch": 0.739443336512616, "grad_norm": 1.9237163410520044, "learning_rate": 1e-05, "loss": 1.3216, "step": 1066 }, { "epoch": 0.7401369981791381, "grad_norm": 1.6945312309384704, "learning_rate": 1e-05, "loss": 1.3205, "step": 1067 }, { "epoch": 0.7408306598456603, "grad_norm": 1.8403636035002944, "learning_rate": 1e-05, "loss": 1.323, "step": 1068 }, { "epoch": 0.7415243215121824, "grad_norm": 1.8463721460630584, "learning_rate": 1e-05, "loss": 1.3215, "step": 1069 }, { "epoch": 0.7422179831787046, "grad_norm": 1.9058089022783489, "learning_rate": 1e-05, "loss": 1.3117, "step": 1070 }, { "epoch": 0.7429116448452268, "grad_norm": 2.1180556320179935, "learning_rate": 1e-05, "loss": 1.3762, "step": 1071 }, { "epoch": 0.7436053065117489, "grad_norm": 1.8334895855944495, "learning_rate": 1e-05, "loss": 1.3443, "step": 1072 }, { "epoch": 0.744298968178271, "grad_norm": 2.1232912269113573, "learning_rate": 1e-05, "loss": 1.3149, "step": 1073 }, { "epoch": 0.7449926298447932, "grad_norm": 1.933335743238272, "learning_rate": 1e-05, "loss": 1.3036, "step": 1074 }, { "epoch": 0.7456862915113154, "grad_norm": 1.7161884340751359, "learning_rate": 1e-05, "loss": 1.337, "step": 1075 }, { "epoch": 0.7463799531778376, "grad_norm": 1.8821904641272937, "learning_rate": 1e-05, "loss": 1.3484, "step": 1076 }, { "epoch": 0.7470736148443596, "grad_norm": 1.8420209136091366, "learning_rate": 1e-05, "loss": 1.3255, "step": 1077 }, { "epoch": 0.7477672765108818, "grad_norm": 1.6844962997358441, "learning_rate": 1e-05, "loss": 1.3763, "step": 1078 }, { "epoch": 0.748460938177404, "grad_norm": 1.6853533211643357, "learning_rate": 1e-05, "loss": 1.334, "step": 1079 }, { "epoch": 0.7491545998439261, "grad_norm": 1.7019512963652867, "learning_rate": 1e-05, "loss": 1.2932, "step": 1080 }, { "epoch": 0.7498482615104483, "grad_norm": 1.811962898514401, "learning_rate": 1e-05, "loss": 1.3587, "step": 1081 }, { "epoch": 0.7505419231769704, "grad_norm": 1.7755093149292678, "learning_rate": 1e-05, "loss": 1.2892, "step": 1082 }, { "epoch": 0.7512355848434926, "grad_norm": 1.8629164456603942, "learning_rate": 1e-05, "loss": 1.3037, "step": 1083 }, { "epoch": 0.7519292465100147, "grad_norm": 1.7886710463953541, "learning_rate": 1e-05, "loss": 1.2803, "step": 1084 }, { "epoch": 0.7526229081765369, "grad_norm": 1.695860963172146, "learning_rate": 1e-05, "loss": 1.4153, "step": 1085 }, { "epoch": 0.7533165698430591, "grad_norm": 1.9667102987846012, "learning_rate": 1e-05, "loss": 1.3436, "step": 1086 }, { "epoch": 0.7540102315095812, "grad_norm": 1.7749867238254802, "learning_rate": 1e-05, "loss": 1.3543, "step": 1087 }, { "epoch": 0.7547038931761033, "grad_norm": 1.8601842356859748, "learning_rate": 1e-05, "loss": 1.3261, "step": 1088 }, { "epoch": 0.7553975548426255, "grad_norm": 1.7933714883604228, "learning_rate": 1e-05, "loss": 1.375, "step": 1089 }, { "epoch": 0.7560912165091477, "grad_norm": 1.9108510451623624, "learning_rate": 1e-05, "loss": 1.3306, "step": 1090 }, { "epoch": 0.7567848781756699, "grad_norm": 1.7031275623765503, "learning_rate": 1e-05, "loss": 1.3776, "step": 1091 }, { "epoch": 0.7574785398421919, "grad_norm": 1.723756241244783, "learning_rate": 1e-05, "loss": 1.3069, "step": 1092 }, { "epoch": 0.7581722015087141, "grad_norm": 1.790196818419589, "learning_rate": 1e-05, "loss": 1.2887, "step": 1093 }, { "epoch": 0.7588658631752363, "grad_norm": 1.6853767916962972, "learning_rate": 1e-05, "loss": 1.3328, "step": 1094 }, { "epoch": 0.7595595248417585, "grad_norm": 1.8283907475641092, "learning_rate": 1e-05, "loss": 1.3394, "step": 1095 }, { "epoch": 0.7602531865082806, "grad_norm": 1.7686661299844872, "learning_rate": 1e-05, "loss": 1.3213, "step": 1096 }, { "epoch": 0.7609468481748027, "grad_norm": 2.0255015885014678, "learning_rate": 1e-05, "loss": 1.3553, "step": 1097 }, { "epoch": 0.7616405098413249, "grad_norm": 1.887116897314805, "learning_rate": 1e-05, "loss": 1.2652, "step": 1098 }, { "epoch": 0.7623341715078471, "grad_norm": 2.0827192956718816, "learning_rate": 1e-05, "loss": 1.2931, "step": 1099 }, { "epoch": 0.7630278331743692, "grad_norm": 1.6696475938133672, "learning_rate": 1e-05, "loss": 1.344, "step": 1100 }, { "epoch": 0.7637214948408914, "grad_norm": 1.8696319148514944, "learning_rate": 1e-05, "loss": 1.2875, "step": 1101 }, { "epoch": 0.7644151565074135, "grad_norm": 1.9161318005356125, "learning_rate": 1e-05, "loss": 1.3039, "step": 1102 }, { "epoch": 0.7651088181739356, "grad_norm": 1.7357326775365562, "learning_rate": 1e-05, "loss": 1.327, "step": 1103 }, { "epoch": 0.7658024798404578, "grad_norm": 1.9374495795431432, "learning_rate": 1e-05, "loss": 1.3118, "step": 1104 }, { "epoch": 0.76649614150698, "grad_norm": 1.8970035705740447, "learning_rate": 1e-05, "loss": 1.3518, "step": 1105 }, { "epoch": 0.7671898031735022, "grad_norm": 1.7310701480408488, "learning_rate": 1e-05, "loss": 1.3011, "step": 1106 }, { "epoch": 0.7678834648400242, "grad_norm": 1.6347319795022848, "learning_rate": 1e-05, "loss": 1.3885, "step": 1107 }, { "epoch": 0.7685771265065464, "grad_norm": 1.7389954204005627, "learning_rate": 1e-05, "loss": 1.3407, "step": 1108 }, { "epoch": 0.7692707881730686, "grad_norm": 1.8545763638618038, "learning_rate": 1e-05, "loss": 1.3475, "step": 1109 }, { "epoch": 0.7699644498395908, "grad_norm": 1.830290361513446, "learning_rate": 1e-05, "loss": 1.3372, "step": 1110 }, { "epoch": 0.7706581115061129, "grad_norm": 1.7554211037949135, "learning_rate": 1e-05, "loss": 1.3382, "step": 1111 }, { "epoch": 0.771351773172635, "grad_norm": 1.791182051330342, "learning_rate": 1e-05, "loss": 1.2865, "step": 1112 }, { "epoch": 0.7720454348391572, "grad_norm": 1.7944697955298015, "learning_rate": 1e-05, "loss": 1.3521, "step": 1113 }, { "epoch": 0.7727390965056794, "grad_norm": 1.7280244820641575, "learning_rate": 1e-05, "loss": 1.2855, "step": 1114 }, { "epoch": 0.7734327581722015, "grad_norm": 1.646134246431486, "learning_rate": 1e-05, "loss": 1.3443, "step": 1115 }, { "epoch": 0.7741264198387237, "grad_norm": 1.9614831482482202, "learning_rate": 1e-05, "loss": 1.3511, "step": 1116 }, { "epoch": 0.7748200815052458, "grad_norm": 1.8931116828060321, "learning_rate": 1e-05, "loss": 1.3014, "step": 1117 }, { "epoch": 0.775513743171768, "grad_norm": 1.766409312274434, "learning_rate": 1e-05, "loss": 1.3368, "step": 1118 }, { "epoch": 0.7762074048382901, "grad_norm": 1.7681164259480207, "learning_rate": 1e-05, "loss": 1.3277, "step": 1119 }, { "epoch": 0.7769010665048123, "grad_norm": 1.8795478880896714, "learning_rate": 1e-05, "loss": 1.358, "step": 1120 }, { "epoch": 0.7775947281713345, "grad_norm": 1.7552042041280342, "learning_rate": 1e-05, "loss": 1.338, "step": 1121 }, { "epoch": 0.7782883898378565, "grad_norm": 1.6825666179566667, "learning_rate": 1e-05, "loss": 1.3468, "step": 1122 }, { "epoch": 0.7789820515043787, "grad_norm": 2.069237558777561, "learning_rate": 1e-05, "loss": 1.3285, "step": 1123 }, { "epoch": 0.7796757131709009, "grad_norm": 1.8945277365950668, "learning_rate": 1e-05, "loss": 1.3457, "step": 1124 }, { "epoch": 0.7803693748374231, "grad_norm": 1.804604515643157, "learning_rate": 1e-05, "loss": 1.3198, "step": 1125 }, { "epoch": 0.7810630365039452, "grad_norm": 1.6687410426653992, "learning_rate": 1e-05, "loss": 1.3571, "step": 1126 }, { "epoch": 0.7817566981704673, "grad_norm": 1.8408255487647456, "learning_rate": 1e-05, "loss": 1.348, "step": 1127 }, { "epoch": 0.7824503598369895, "grad_norm": 1.7473823693404393, "learning_rate": 1e-05, "loss": 1.3449, "step": 1128 }, { "epoch": 0.7831440215035117, "grad_norm": 1.762367934431706, "learning_rate": 1e-05, "loss": 1.3296, "step": 1129 }, { "epoch": 0.7838376831700338, "grad_norm": 1.8065045373977573, "learning_rate": 1e-05, "loss": 1.3088, "step": 1130 }, { "epoch": 0.784531344836556, "grad_norm": 1.7702834178449094, "learning_rate": 1e-05, "loss": 1.3251, "step": 1131 }, { "epoch": 0.7852250065030781, "grad_norm": 1.684651956430982, "learning_rate": 1e-05, "loss": 1.3221, "step": 1132 }, { "epoch": 0.7859186681696003, "grad_norm": 1.718401813004295, "learning_rate": 1e-05, "loss": 1.3583, "step": 1133 }, { "epoch": 0.7866123298361224, "grad_norm": 1.7593076318939966, "learning_rate": 1e-05, "loss": 1.3481, "step": 1134 }, { "epoch": 0.7873059915026446, "grad_norm": 1.9066726353857328, "learning_rate": 1e-05, "loss": 1.298, "step": 1135 }, { "epoch": 0.7879996531691668, "grad_norm": 1.666639956360219, "learning_rate": 1e-05, "loss": 1.3321, "step": 1136 }, { "epoch": 0.7886933148356889, "grad_norm": 1.722161549449429, "learning_rate": 1e-05, "loss": 1.3081, "step": 1137 }, { "epoch": 0.789386976502211, "grad_norm": 1.8556050317918311, "learning_rate": 1e-05, "loss": 1.3155, "step": 1138 }, { "epoch": 0.7900806381687332, "grad_norm": 1.906163646274444, "learning_rate": 1e-05, "loss": 1.3127, "step": 1139 }, { "epoch": 0.7907742998352554, "grad_norm": 1.91777805112317, "learning_rate": 1e-05, "loss": 1.3267, "step": 1140 }, { "epoch": 0.7914679615017776, "grad_norm": 1.758622480343078, "learning_rate": 1e-05, "loss": 1.3604, "step": 1141 }, { "epoch": 0.7921616231682996, "grad_norm": 1.818304243791579, "learning_rate": 1e-05, "loss": 1.3744, "step": 1142 }, { "epoch": 0.7928552848348218, "grad_norm": 1.7096779670968696, "learning_rate": 1e-05, "loss": 1.2967, "step": 1143 }, { "epoch": 0.793548946501344, "grad_norm": 1.8288634798052434, "learning_rate": 1e-05, "loss": 1.3836, "step": 1144 }, { "epoch": 0.7942426081678661, "grad_norm": 1.837569087516797, "learning_rate": 1e-05, "loss": 1.3306, "step": 1145 }, { "epoch": 0.7949362698343883, "grad_norm": 1.757663683708377, "learning_rate": 1e-05, "loss": 1.3289, "step": 1146 }, { "epoch": 0.7956299315009104, "grad_norm": 1.5733908350037635, "learning_rate": 1e-05, "loss": 1.3293, "step": 1147 }, { "epoch": 0.7963235931674326, "grad_norm": 1.7006097263550368, "learning_rate": 1e-05, "loss": 1.3544, "step": 1148 }, { "epoch": 0.7970172548339547, "grad_norm": 1.852767855910649, "learning_rate": 1e-05, "loss": 1.367, "step": 1149 }, { "epoch": 0.7977109165004769, "grad_norm": 1.8236237376393378, "learning_rate": 1e-05, "loss": 1.3298, "step": 1150 }, { "epoch": 0.7984045781669991, "grad_norm": 1.7542840584318944, "learning_rate": 1e-05, "loss": 1.3305, "step": 1151 }, { "epoch": 0.7990982398335212, "grad_norm": 1.7363578108911824, "learning_rate": 1e-05, "loss": 1.312, "step": 1152 }, { "epoch": 0.7997919015000433, "grad_norm": 1.7447796524158012, "learning_rate": 1e-05, "loss": 1.3146, "step": 1153 }, { "epoch": 0.8004855631665655, "grad_norm": 1.7998354010324689, "learning_rate": 1e-05, "loss": 1.2651, "step": 1154 }, { "epoch": 0.8011792248330877, "grad_norm": 1.72370764035216, "learning_rate": 1e-05, "loss": 1.3313, "step": 1155 }, { "epoch": 0.8018728864996099, "grad_norm": 1.9027302235993584, "learning_rate": 1e-05, "loss": 1.3275, "step": 1156 }, { "epoch": 0.8025665481661319, "grad_norm": 1.767361577842723, "learning_rate": 1e-05, "loss": 1.281, "step": 1157 }, { "epoch": 0.8032602098326541, "grad_norm": 1.6957352595013901, "learning_rate": 1e-05, "loss": 1.339, "step": 1158 }, { "epoch": 0.8039538714991763, "grad_norm": 1.8811200117431954, "learning_rate": 1e-05, "loss": 1.3694, "step": 1159 }, { "epoch": 0.8046475331656985, "grad_norm": 1.8732110913975033, "learning_rate": 1e-05, "loss": 1.3828, "step": 1160 }, { "epoch": 0.8053411948322206, "grad_norm": 1.806929633499466, "learning_rate": 1e-05, "loss": 1.293, "step": 1161 }, { "epoch": 0.8060348564987427, "grad_norm": 1.9234019762346322, "learning_rate": 1e-05, "loss": 1.356, "step": 1162 }, { "epoch": 0.8067285181652649, "grad_norm": 1.8612299857621706, "learning_rate": 1e-05, "loss": 1.2944, "step": 1163 }, { "epoch": 0.8074221798317871, "grad_norm": 1.8949079530850956, "learning_rate": 1e-05, "loss": 1.2759, "step": 1164 }, { "epoch": 0.8081158414983092, "grad_norm": 1.835747159847785, "learning_rate": 1e-05, "loss": 1.2959, "step": 1165 }, { "epoch": 0.8088095031648314, "grad_norm": 1.6994420208999754, "learning_rate": 1e-05, "loss": 1.3546, "step": 1166 }, { "epoch": 0.8095031648313535, "grad_norm": 1.7058565978885374, "learning_rate": 1e-05, "loss": 1.3016, "step": 1167 }, { "epoch": 0.8101968264978756, "grad_norm": 1.7202328196552465, "learning_rate": 1e-05, "loss": 1.3244, "step": 1168 }, { "epoch": 0.8108904881643978, "grad_norm": 1.5969257126998517, "learning_rate": 1e-05, "loss": 1.2941, "step": 1169 }, { "epoch": 0.81158414983092, "grad_norm": 2.033122314469678, "learning_rate": 1e-05, "loss": 1.3112, "step": 1170 }, { "epoch": 0.8122778114974422, "grad_norm": 1.818607135695885, "learning_rate": 1e-05, "loss": 1.3372, "step": 1171 }, { "epoch": 0.8129714731639642, "grad_norm": 1.7332347851944176, "learning_rate": 1e-05, "loss": 1.2963, "step": 1172 }, { "epoch": 0.8136651348304864, "grad_norm": 1.79912386221675, "learning_rate": 1e-05, "loss": 1.3192, "step": 1173 }, { "epoch": 0.8143587964970086, "grad_norm": 1.759247644611021, "learning_rate": 1e-05, "loss": 1.3087, "step": 1174 }, { "epoch": 0.8150524581635308, "grad_norm": 1.81196721719679, "learning_rate": 1e-05, "loss": 1.3347, "step": 1175 }, { "epoch": 0.8157461198300529, "grad_norm": 1.8329350171829653, "learning_rate": 1e-05, "loss": 1.347, "step": 1176 }, { "epoch": 0.816439781496575, "grad_norm": 2.097683322497983, "learning_rate": 1e-05, "loss": 1.3187, "step": 1177 }, { "epoch": 0.8171334431630972, "grad_norm": 1.682730633916439, "learning_rate": 1e-05, "loss": 1.326, "step": 1178 }, { "epoch": 0.8178271048296194, "grad_norm": 1.7967406525079352, "learning_rate": 1e-05, "loss": 1.3229, "step": 1179 }, { "epoch": 0.8185207664961415, "grad_norm": 1.7155776890591734, "learning_rate": 1e-05, "loss": 1.3143, "step": 1180 }, { "epoch": 0.8192144281626637, "grad_norm": 1.731948891871397, "learning_rate": 1e-05, "loss": 1.3044, "step": 1181 }, { "epoch": 0.8199080898291858, "grad_norm": 1.6258267218121343, "learning_rate": 1e-05, "loss": 1.2966, "step": 1182 }, { "epoch": 0.820601751495708, "grad_norm": 1.8947923997247118, "learning_rate": 1e-05, "loss": 1.25, "step": 1183 }, { "epoch": 0.8212954131622301, "grad_norm": 1.9268221611589265, "learning_rate": 1e-05, "loss": 1.363, "step": 1184 }, { "epoch": 0.8219890748287523, "grad_norm": 1.6151381049081766, "learning_rate": 1e-05, "loss": 1.3323, "step": 1185 }, { "epoch": 0.8226827364952745, "grad_norm": 1.6754745238527986, "learning_rate": 1e-05, "loss": 1.2736, "step": 1186 }, { "epoch": 0.8233763981617965, "grad_norm": 1.6566597340456615, "learning_rate": 1e-05, "loss": 1.3292, "step": 1187 }, { "epoch": 0.8240700598283187, "grad_norm": 1.7868911126447826, "learning_rate": 1e-05, "loss": 1.3146, "step": 1188 }, { "epoch": 0.8247637214948409, "grad_norm": 1.7395269678979228, "learning_rate": 1e-05, "loss": 1.3172, "step": 1189 }, { "epoch": 0.8254573831613631, "grad_norm": 1.846976443991522, "learning_rate": 1e-05, "loss": 1.2845, "step": 1190 }, { "epoch": 0.8261510448278852, "grad_norm": 1.9263142545199925, "learning_rate": 1e-05, "loss": 1.3218, "step": 1191 }, { "epoch": 0.8268447064944073, "grad_norm": 1.6753572265693735, "learning_rate": 1e-05, "loss": 1.3347, "step": 1192 }, { "epoch": 0.8275383681609295, "grad_norm": 1.8402922863638769, "learning_rate": 1e-05, "loss": 1.3389, "step": 1193 }, { "epoch": 0.8282320298274517, "grad_norm": 1.7003805951507205, "learning_rate": 1e-05, "loss": 1.3016, "step": 1194 }, { "epoch": 0.8289256914939738, "grad_norm": 1.7016887764707231, "learning_rate": 1e-05, "loss": 1.3312, "step": 1195 }, { "epoch": 0.829619353160496, "grad_norm": 1.7450143183218212, "learning_rate": 1e-05, "loss": 1.303, "step": 1196 }, { "epoch": 0.8303130148270181, "grad_norm": 1.7676264585702774, "learning_rate": 1e-05, "loss": 1.317, "step": 1197 }, { "epoch": 0.8310066764935403, "grad_norm": 2.0267151482602177, "learning_rate": 1e-05, "loss": 1.2687, "step": 1198 }, { "epoch": 0.8317003381600624, "grad_norm": 1.6648828369964448, "learning_rate": 1e-05, "loss": 1.3436, "step": 1199 }, { "epoch": 0.8323939998265846, "grad_norm": 1.7437721633894316, "learning_rate": 1e-05, "loss": 1.2719, "step": 1200 }, { "epoch": 0.8330876614931068, "grad_norm": 1.76080267027574, "learning_rate": 1e-05, "loss": 1.3364, "step": 1201 }, { "epoch": 0.8337813231596289, "grad_norm": 1.6023127093441858, "learning_rate": 1e-05, "loss": 1.3567, "step": 1202 }, { "epoch": 0.834474984826151, "grad_norm": 1.807984298617728, "learning_rate": 1e-05, "loss": 1.3243, "step": 1203 }, { "epoch": 0.8351686464926732, "grad_norm": 1.7507073872441543, "learning_rate": 1e-05, "loss": 1.2748, "step": 1204 }, { "epoch": 0.8358623081591954, "grad_norm": 1.7340226078649954, "learning_rate": 1e-05, "loss": 1.3236, "step": 1205 }, { "epoch": 0.8365559698257176, "grad_norm": 1.8128370813992278, "learning_rate": 1e-05, "loss": 1.3398, "step": 1206 }, { "epoch": 0.8372496314922396, "grad_norm": 1.8385313973464554, "learning_rate": 1e-05, "loss": 1.3244, "step": 1207 }, { "epoch": 0.8379432931587618, "grad_norm": 1.8202487072236857, "learning_rate": 1e-05, "loss": 1.2841, "step": 1208 }, { "epoch": 0.838636954825284, "grad_norm": 1.8226882804903315, "learning_rate": 1e-05, "loss": 1.3251, "step": 1209 }, { "epoch": 0.8393306164918061, "grad_norm": 1.7796315284012894, "learning_rate": 1e-05, "loss": 1.3503, "step": 1210 }, { "epoch": 0.8400242781583283, "grad_norm": 1.770507934678552, "learning_rate": 1e-05, "loss": 1.2873, "step": 1211 }, { "epoch": 0.8407179398248504, "grad_norm": 1.7983739916772261, "learning_rate": 1e-05, "loss": 1.3309, "step": 1212 }, { "epoch": 0.8414116014913726, "grad_norm": 1.8581265442393458, "learning_rate": 1e-05, "loss": 1.2976, "step": 1213 }, { "epoch": 0.8421052631578947, "grad_norm": 1.9709414587672909, "learning_rate": 1e-05, "loss": 1.3244, "step": 1214 }, { "epoch": 0.8427989248244169, "grad_norm": 1.934195079670126, "learning_rate": 1e-05, "loss": 1.3197, "step": 1215 }, { "epoch": 0.8434925864909391, "grad_norm": 1.5905446332751805, "learning_rate": 1e-05, "loss": 1.3144, "step": 1216 }, { "epoch": 0.8441862481574612, "grad_norm": 1.8012514681125382, "learning_rate": 1e-05, "loss": 1.2984, "step": 1217 }, { "epoch": 0.8448799098239833, "grad_norm": 1.8322906369289444, "learning_rate": 1e-05, "loss": 1.3365, "step": 1218 }, { "epoch": 0.8455735714905055, "grad_norm": 1.7098942463292028, "learning_rate": 1e-05, "loss": 1.2994, "step": 1219 }, { "epoch": 0.8462672331570277, "grad_norm": 1.806262096500155, "learning_rate": 1e-05, "loss": 1.3135, "step": 1220 }, { "epoch": 0.8469608948235499, "grad_norm": 1.8088243417346301, "learning_rate": 1e-05, "loss": 1.2548, "step": 1221 }, { "epoch": 0.8476545564900719, "grad_norm": 1.7615794277621235, "learning_rate": 1e-05, "loss": 1.369, "step": 1222 }, { "epoch": 0.8483482181565941, "grad_norm": 1.7446487987023735, "learning_rate": 1e-05, "loss": 1.3417, "step": 1223 }, { "epoch": 0.8490418798231163, "grad_norm": 1.9292547174186359, "learning_rate": 1e-05, "loss": 1.2619, "step": 1224 }, { "epoch": 0.8497355414896385, "grad_norm": 1.7385855689942564, "learning_rate": 1e-05, "loss": 1.3367, "step": 1225 }, { "epoch": 0.8504292031561606, "grad_norm": 1.7893857799911939, "learning_rate": 1e-05, "loss": 1.3231, "step": 1226 }, { "epoch": 0.8511228648226827, "grad_norm": 1.7904188745955463, "learning_rate": 1e-05, "loss": 1.273, "step": 1227 }, { "epoch": 0.8518165264892049, "grad_norm": 1.7311471203311328, "learning_rate": 1e-05, "loss": 1.3036, "step": 1228 }, { "epoch": 0.852510188155727, "grad_norm": 1.8405281776102504, "learning_rate": 1e-05, "loss": 1.3266, "step": 1229 }, { "epoch": 0.8532038498222492, "grad_norm": 1.7910111933959387, "learning_rate": 1e-05, "loss": 1.332, "step": 1230 }, { "epoch": 0.8538975114887714, "grad_norm": 1.692055737327334, "learning_rate": 1e-05, "loss": 1.2594, "step": 1231 }, { "epoch": 0.8545911731552935, "grad_norm": 1.715888470034707, "learning_rate": 1e-05, "loss": 1.2974, "step": 1232 }, { "epoch": 0.8552848348218156, "grad_norm": 1.6920575665727629, "learning_rate": 1e-05, "loss": 1.3081, "step": 1233 }, { "epoch": 0.8559784964883378, "grad_norm": 1.8820212499017275, "learning_rate": 1e-05, "loss": 1.2579, "step": 1234 }, { "epoch": 0.85667215815486, "grad_norm": 1.7590796982626669, "learning_rate": 1e-05, "loss": 1.346, "step": 1235 }, { "epoch": 0.8573658198213822, "grad_norm": 1.7175897993182907, "learning_rate": 1e-05, "loss": 1.2952, "step": 1236 }, { "epoch": 0.8580594814879042, "grad_norm": 1.8872407973916834, "learning_rate": 1e-05, "loss": 1.2678, "step": 1237 }, { "epoch": 0.8587531431544264, "grad_norm": 1.820685981576064, "learning_rate": 1e-05, "loss": 1.3161, "step": 1238 }, { "epoch": 0.8594468048209486, "grad_norm": 1.7106563905509156, "learning_rate": 1e-05, "loss": 1.2805, "step": 1239 }, { "epoch": 0.8601404664874708, "grad_norm": 1.835121012618051, "learning_rate": 1e-05, "loss": 1.3411, "step": 1240 }, { "epoch": 0.8608341281539929, "grad_norm": 1.759648204282831, "learning_rate": 1e-05, "loss": 1.3169, "step": 1241 }, { "epoch": 0.861527789820515, "grad_norm": 1.8108439058590786, "learning_rate": 1e-05, "loss": 1.3365, "step": 1242 }, { "epoch": 0.8622214514870372, "grad_norm": 1.8264838456429158, "learning_rate": 1e-05, "loss": 1.3122, "step": 1243 }, { "epoch": 0.8629151131535594, "grad_norm": 1.8561243343760983, "learning_rate": 1e-05, "loss": 1.3126, "step": 1244 }, { "epoch": 0.8636087748200815, "grad_norm": 1.7744496574339597, "learning_rate": 1e-05, "loss": 1.3419, "step": 1245 }, { "epoch": 0.8643024364866037, "grad_norm": 1.7102771712690799, "learning_rate": 1e-05, "loss": 1.3913, "step": 1246 }, { "epoch": 0.8649960981531258, "grad_norm": 1.637011563504676, "learning_rate": 1e-05, "loss": 1.3294, "step": 1247 }, { "epoch": 0.865689759819648, "grad_norm": 1.7378852441605857, "learning_rate": 1e-05, "loss": 1.3604, "step": 1248 }, { "epoch": 0.8663834214861701, "grad_norm": 1.892145137680064, "learning_rate": 1e-05, "loss": 1.3839, "step": 1249 }, { "epoch": 0.8670770831526923, "grad_norm": 1.7114640198993971, "learning_rate": 1e-05, "loss": 1.3164, "step": 1250 }, { "epoch": 0.8677707448192145, "grad_norm": 1.857472545380348, "learning_rate": 1e-05, "loss": 1.3409, "step": 1251 }, { "epoch": 0.8684644064857365, "grad_norm": 1.6636749531242045, "learning_rate": 1e-05, "loss": 1.3103, "step": 1252 }, { "epoch": 0.8691580681522587, "grad_norm": 1.853753890533686, "learning_rate": 1e-05, "loss": 1.3372, "step": 1253 }, { "epoch": 0.8698517298187809, "grad_norm": 1.7274626500933439, "learning_rate": 1e-05, "loss": 1.3198, "step": 1254 }, { "epoch": 0.8705453914853031, "grad_norm": 1.7681696772117461, "learning_rate": 1e-05, "loss": 1.3451, "step": 1255 }, { "epoch": 0.8712390531518251, "grad_norm": 1.7074991789276823, "learning_rate": 1e-05, "loss": 1.306, "step": 1256 }, { "epoch": 0.8719327148183473, "grad_norm": 1.8022859354848255, "learning_rate": 1e-05, "loss": 1.307, "step": 1257 }, { "epoch": 0.8726263764848695, "grad_norm": 1.722826958334196, "learning_rate": 1e-05, "loss": 1.348, "step": 1258 }, { "epoch": 0.8733200381513917, "grad_norm": 1.6278820667115972, "learning_rate": 1e-05, "loss": 1.3156, "step": 1259 }, { "epoch": 0.8740136998179138, "grad_norm": 1.8615627558866032, "learning_rate": 1e-05, "loss": 1.3642, "step": 1260 }, { "epoch": 0.874707361484436, "grad_norm": 1.6977360433618884, "learning_rate": 1e-05, "loss": 1.3661, "step": 1261 }, { "epoch": 0.8754010231509581, "grad_norm": 1.7119935578912018, "learning_rate": 1e-05, "loss": 1.343, "step": 1262 }, { "epoch": 0.8760946848174803, "grad_norm": 1.8828273441618772, "learning_rate": 1e-05, "loss": 1.3286, "step": 1263 }, { "epoch": 0.8767883464840024, "grad_norm": 1.7991745556104566, "learning_rate": 1e-05, "loss": 1.2877, "step": 1264 }, { "epoch": 0.8774820081505246, "grad_norm": 1.7692125953469466, "learning_rate": 1e-05, "loss": 1.3132, "step": 1265 }, { "epoch": 0.8781756698170468, "grad_norm": 1.763539004235656, "learning_rate": 1e-05, "loss": 1.2516, "step": 1266 }, { "epoch": 0.8788693314835689, "grad_norm": 1.8097604071898914, "learning_rate": 1e-05, "loss": 1.322, "step": 1267 }, { "epoch": 0.879562993150091, "grad_norm": 1.8504952339835505, "learning_rate": 1e-05, "loss": 1.3257, "step": 1268 }, { "epoch": 0.8802566548166132, "grad_norm": 1.8228681067260823, "learning_rate": 1e-05, "loss": 1.3207, "step": 1269 }, { "epoch": 0.8809503164831354, "grad_norm": 1.8569141776272553, "learning_rate": 1e-05, "loss": 1.3415, "step": 1270 }, { "epoch": 0.8816439781496576, "grad_norm": 1.8653037577154865, "learning_rate": 1e-05, "loss": 1.2882, "step": 1271 }, { "epoch": 0.8823376398161796, "grad_norm": 1.7842091027351248, "learning_rate": 1e-05, "loss": 1.3011, "step": 1272 }, { "epoch": 0.8830313014827018, "grad_norm": 1.865231093038103, "learning_rate": 1e-05, "loss": 1.3421, "step": 1273 }, { "epoch": 0.883724963149224, "grad_norm": 1.8445484018508556, "learning_rate": 1e-05, "loss": 1.2994, "step": 1274 }, { "epoch": 0.8844186248157461, "grad_norm": 2.0112180595207585, "learning_rate": 1e-05, "loss": 1.3491, "step": 1275 }, { "epoch": 0.8851122864822683, "grad_norm": 1.9313320335876165, "learning_rate": 1e-05, "loss": 1.2972, "step": 1276 }, { "epoch": 0.8858059481487904, "grad_norm": 1.766314463801755, "learning_rate": 1e-05, "loss": 1.3551, "step": 1277 }, { "epoch": 0.8864996098153126, "grad_norm": 1.6939200547787165, "learning_rate": 1e-05, "loss": 1.3346, "step": 1278 }, { "epoch": 0.8871932714818347, "grad_norm": 1.6553779025228499, "learning_rate": 1e-05, "loss": 1.3331, "step": 1279 }, { "epoch": 0.8878869331483569, "grad_norm": 1.672522216377223, "learning_rate": 1e-05, "loss": 1.3143, "step": 1280 }, { "epoch": 0.888580594814879, "grad_norm": 1.6477930929287763, "learning_rate": 1e-05, "loss": 1.2655, "step": 1281 }, { "epoch": 0.8892742564814012, "grad_norm": 1.9082634726577539, "learning_rate": 1e-05, "loss": 1.3101, "step": 1282 }, { "epoch": 0.8899679181479233, "grad_norm": 1.6619545725705775, "learning_rate": 1e-05, "loss": 1.3134, "step": 1283 }, { "epoch": 0.8906615798144455, "grad_norm": 1.8186223504046093, "learning_rate": 1e-05, "loss": 1.3356, "step": 1284 }, { "epoch": 0.8913552414809677, "grad_norm": 1.6755642327768199, "learning_rate": 1e-05, "loss": 1.3005, "step": 1285 }, { "epoch": 0.8920489031474899, "grad_norm": 1.7174416264813488, "learning_rate": 1e-05, "loss": 1.3347, "step": 1286 }, { "epoch": 0.8927425648140119, "grad_norm": 1.7186710255881854, "learning_rate": 1e-05, "loss": 1.2908, "step": 1287 }, { "epoch": 0.8934362264805341, "grad_norm": 1.6346319252027368, "learning_rate": 1e-05, "loss": 1.2927, "step": 1288 }, { "epoch": 0.8941298881470563, "grad_norm": 1.8474495055475482, "learning_rate": 1e-05, "loss": 1.3255, "step": 1289 }, { "epoch": 0.8948235498135785, "grad_norm": 1.8788390378390694, "learning_rate": 1e-05, "loss": 1.3385, "step": 1290 }, { "epoch": 0.8955172114801006, "grad_norm": 1.629566356590922, "learning_rate": 1e-05, "loss": 1.3482, "step": 1291 }, { "epoch": 0.8962108731466227, "grad_norm": 1.7412289424540435, "learning_rate": 1e-05, "loss": 1.267, "step": 1292 }, { "epoch": 0.8969045348131449, "grad_norm": 1.5904673607353297, "learning_rate": 1e-05, "loss": 1.3107, "step": 1293 }, { "epoch": 0.897598196479667, "grad_norm": 1.7668320875825854, "learning_rate": 1e-05, "loss": 1.2968, "step": 1294 }, { "epoch": 0.8982918581461892, "grad_norm": 1.6434271863603802, "learning_rate": 1e-05, "loss": 1.3531, "step": 1295 }, { "epoch": 0.8989855198127114, "grad_norm": 1.6446049917777663, "learning_rate": 1e-05, "loss": 1.3067, "step": 1296 }, { "epoch": 0.8996791814792335, "grad_norm": 1.7114850493139724, "learning_rate": 1e-05, "loss": 1.2526, "step": 1297 }, { "epoch": 0.9003728431457556, "grad_norm": 1.7703812278053084, "learning_rate": 1e-05, "loss": 1.3284, "step": 1298 }, { "epoch": 0.9010665048122778, "grad_norm": 1.783598858952647, "learning_rate": 1e-05, "loss": 1.3449, "step": 1299 }, { "epoch": 0.9017601664788, "grad_norm": 1.7594643222023651, "learning_rate": 1e-05, "loss": 1.2865, "step": 1300 }, { "epoch": 0.9024538281453222, "grad_norm": 1.7979500686841217, "learning_rate": 1e-05, "loss": 1.2826, "step": 1301 }, { "epoch": 0.9031474898118442, "grad_norm": 1.7054221381249888, "learning_rate": 1e-05, "loss": 1.3525, "step": 1302 }, { "epoch": 0.9038411514783664, "grad_norm": 1.7423889164937596, "learning_rate": 1e-05, "loss": 1.3044, "step": 1303 }, { "epoch": 0.9045348131448886, "grad_norm": 1.8080520927021586, "learning_rate": 1e-05, "loss": 1.331, "step": 1304 }, { "epoch": 0.9052284748114108, "grad_norm": 1.9903383637954946, "learning_rate": 1e-05, "loss": 1.3213, "step": 1305 }, { "epoch": 0.9059221364779328, "grad_norm": 1.622554431537696, "learning_rate": 1e-05, "loss": 1.3307, "step": 1306 }, { "epoch": 0.906615798144455, "grad_norm": 1.891392187663753, "learning_rate": 1e-05, "loss": 1.3484, "step": 1307 }, { "epoch": 0.9073094598109772, "grad_norm": 1.5686215128270367, "learning_rate": 1e-05, "loss": 1.3101, "step": 1308 }, { "epoch": 0.9080031214774994, "grad_norm": 1.7879215692958745, "learning_rate": 1e-05, "loss": 1.3499, "step": 1309 }, { "epoch": 0.9086967831440215, "grad_norm": 1.7583743340224038, "learning_rate": 1e-05, "loss": 1.2757, "step": 1310 }, { "epoch": 0.9093904448105437, "grad_norm": 1.7396925828139032, "learning_rate": 1e-05, "loss": 1.2925, "step": 1311 }, { "epoch": 0.9100841064770658, "grad_norm": 1.7958710900831494, "learning_rate": 1e-05, "loss": 1.3058, "step": 1312 }, { "epoch": 0.910777768143588, "grad_norm": 1.764507973805378, "learning_rate": 1e-05, "loss": 1.297, "step": 1313 }, { "epoch": 0.9114714298101101, "grad_norm": 1.8550793576750044, "learning_rate": 1e-05, "loss": 1.3146, "step": 1314 }, { "epoch": 0.9121650914766323, "grad_norm": 1.6803054339754033, "learning_rate": 1e-05, "loss": 1.2966, "step": 1315 }, { "epoch": 0.9128587531431545, "grad_norm": 1.885929787117982, "learning_rate": 1e-05, "loss": 1.3201, "step": 1316 }, { "epoch": 0.9135524148096765, "grad_norm": 1.6131928750808537, "learning_rate": 1e-05, "loss": 1.2899, "step": 1317 }, { "epoch": 0.9142460764761987, "grad_norm": 1.8283426626966652, "learning_rate": 1e-05, "loss": 1.3149, "step": 1318 }, { "epoch": 0.9149397381427209, "grad_norm": 1.7895417190819627, "learning_rate": 1e-05, "loss": 1.3135, "step": 1319 }, { "epoch": 0.9156333998092431, "grad_norm": 1.6655898144804122, "learning_rate": 1e-05, "loss": 1.3506, "step": 1320 }, { "epoch": 0.9163270614757651, "grad_norm": 1.789395903422411, "learning_rate": 1e-05, "loss": 1.2944, "step": 1321 }, { "epoch": 0.9170207231422873, "grad_norm": 1.5554671892032146, "learning_rate": 1e-05, "loss": 1.3035, "step": 1322 }, { "epoch": 0.9177143848088095, "grad_norm": 1.7462183316659554, "learning_rate": 1e-05, "loss": 1.2871, "step": 1323 }, { "epoch": 0.9184080464753317, "grad_norm": 1.6407955316695355, "learning_rate": 1e-05, "loss": 1.2842, "step": 1324 }, { "epoch": 0.9191017081418538, "grad_norm": 1.8054786932821483, "learning_rate": 1e-05, "loss": 1.2801, "step": 1325 }, { "epoch": 0.919795369808376, "grad_norm": 1.7774882584688303, "learning_rate": 1e-05, "loss": 1.3111, "step": 1326 }, { "epoch": 0.9204890314748981, "grad_norm": 1.7487308168423374, "learning_rate": 1e-05, "loss": 1.3086, "step": 1327 }, { "epoch": 0.9211826931414203, "grad_norm": 1.830348412513128, "learning_rate": 1e-05, "loss": 1.2398, "step": 1328 }, { "epoch": 0.9218763548079424, "grad_norm": 1.8745249129153505, "learning_rate": 1e-05, "loss": 1.3174, "step": 1329 }, { "epoch": 0.9225700164744646, "grad_norm": 1.6997745347291242, "learning_rate": 1e-05, "loss": 1.2795, "step": 1330 }, { "epoch": 0.9232636781409868, "grad_norm": 1.6461004692498915, "learning_rate": 1e-05, "loss": 1.369, "step": 1331 }, { "epoch": 0.9239573398075089, "grad_norm": 1.867846237084916, "learning_rate": 1e-05, "loss": 1.3492, "step": 1332 }, { "epoch": 0.924651001474031, "grad_norm": 1.7338725633875975, "learning_rate": 1e-05, "loss": 1.3425, "step": 1333 }, { "epoch": 0.9253446631405532, "grad_norm": 1.7058365907688622, "learning_rate": 1e-05, "loss": 1.3289, "step": 1334 }, { "epoch": 0.9260383248070754, "grad_norm": 1.6812689204725293, "learning_rate": 1e-05, "loss": 1.3166, "step": 1335 }, { "epoch": 0.9267319864735976, "grad_norm": 1.7854594207936143, "learning_rate": 1e-05, "loss": 1.3357, "step": 1336 }, { "epoch": 0.9274256481401196, "grad_norm": 1.8632837097505413, "learning_rate": 1e-05, "loss": 1.2385, "step": 1337 }, { "epoch": 0.9281193098066418, "grad_norm": 1.838541328017032, "learning_rate": 1e-05, "loss": 1.2703, "step": 1338 }, { "epoch": 0.928812971473164, "grad_norm": 1.8476109239446323, "learning_rate": 1e-05, "loss": 1.2934, "step": 1339 }, { "epoch": 0.9295066331396861, "grad_norm": 1.735630365745877, "learning_rate": 1e-05, "loss": 1.2978, "step": 1340 }, { "epoch": 0.9302002948062082, "grad_norm": 1.6182767450065856, "learning_rate": 1e-05, "loss": 1.2999, "step": 1341 }, { "epoch": 0.9308939564727304, "grad_norm": 1.7314432313872, "learning_rate": 1e-05, "loss": 1.291, "step": 1342 }, { "epoch": 0.9315876181392526, "grad_norm": 1.9298812546673758, "learning_rate": 1e-05, "loss": 1.3352, "step": 1343 }, { "epoch": 0.9322812798057747, "grad_norm": 1.890305436960625, "learning_rate": 1e-05, "loss": 1.3603, "step": 1344 }, { "epoch": 0.9329749414722969, "grad_norm": 1.7012847869156689, "learning_rate": 1e-05, "loss": 1.2705, "step": 1345 }, { "epoch": 0.933668603138819, "grad_norm": 1.7624548099246455, "learning_rate": 1e-05, "loss": 1.3229, "step": 1346 }, { "epoch": 0.9343622648053412, "grad_norm": 1.8174643785961295, "learning_rate": 1e-05, "loss": 1.2613, "step": 1347 }, { "epoch": 0.9350559264718633, "grad_norm": 1.766018762161775, "learning_rate": 1e-05, "loss": 1.3907, "step": 1348 }, { "epoch": 0.9357495881383855, "grad_norm": 1.8138067691928326, "learning_rate": 1e-05, "loss": 1.2878, "step": 1349 }, { "epoch": 0.9364432498049077, "grad_norm": 1.8378030129613956, "learning_rate": 1e-05, "loss": 1.3192, "step": 1350 }, { "epoch": 0.9371369114714299, "grad_norm": 2.012247327683597, "learning_rate": 1e-05, "loss": 1.3096, "step": 1351 }, { "epoch": 0.9378305731379519, "grad_norm": 1.704402821997915, "learning_rate": 1e-05, "loss": 1.3304, "step": 1352 }, { "epoch": 0.9385242348044741, "grad_norm": 1.7885693471410014, "learning_rate": 1e-05, "loss": 1.3065, "step": 1353 }, { "epoch": 0.9392178964709963, "grad_norm": 1.6661434949017349, "learning_rate": 1e-05, "loss": 1.3307, "step": 1354 }, { "epoch": 0.9399115581375185, "grad_norm": 1.7962358503874996, "learning_rate": 1e-05, "loss": 1.3129, "step": 1355 }, { "epoch": 0.9406052198040405, "grad_norm": 1.718189556083174, "learning_rate": 1e-05, "loss": 1.2994, "step": 1356 }, { "epoch": 0.9412988814705627, "grad_norm": 1.9092876605935403, "learning_rate": 1e-05, "loss": 1.3859, "step": 1357 }, { "epoch": 0.9419925431370849, "grad_norm": 1.8331634920945954, "learning_rate": 1e-05, "loss": 1.2668, "step": 1358 }, { "epoch": 0.942686204803607, "grad_norm": 1.7795600411027266, "learning_rate": 1e-05, "loss": 1.3158, "step": 1359 }, { "epoch": 0.9433798664701292, "grad_norm": 1.7718647937619816, "learning_rate": 1e-05, "loss": 1.2766, "step": 1360 }, { "epoch": 0.9440735281366514, "grad_norm": 1.7833385647199957, "learning_rate": 1e-05, "loss": 1.3188, "step": 1361 }, { "epoch": 0.9447671898031735, "grad_norm": 1.7094537156535095, "learning_rate": 1e-05, "loss": 1.2916, "step": 1362 }, { "epoch": 0.9454608514696956, "grad_norm": 1.7218575319930665, "learning_rate": 1e-05, "loss": 1.363, "step": 1363 }, { "epoch": 0.9461545131362178, "grad_norm": 1.8291690131158997, "learning_rate": 1e-05, "loss": 1.3067, "step": 1364 }, { "epoch": 0.94684817480274, "grad_norm": 1.5981127580277035, "learning_rate": 1e-05, "loss": 1.316, "step": 1365 }, { "epoch": 0.9475418364692622, "grad_norm": 1.7097133890389782, "learning_rate": 1e-05, "loss": 1.31, "step": 1366 }, { "epoch": 0.9482354981357842, "grad_norm": 1.7177869042959357, "learning_rate": 1e-05, "loss": 1.3261, "step": 1367 }, { "epoch": 0.9489291598023064, "grad_norm": 1.7391018887085676, "learning_rate": 1e-05, "loss": 1.3476, "step": 1368 }, { "epoch": 0.9496228214688286, "grad_norm": 1.835174110490517, "learning_rate": 1e-05, "loss": 1.3198, "step": 1369 }, { "epoch": 0.9503164831353508, "grad_norm": 1.8268669014129535, "learning_rate": 1e-05, "loss": 1.3338, "step": 1370 }, { "epoch": 0.9510101448018728, "grad_norm": 1.6581115729537674, "learning_rate": 1e-05, "loss": 1.2378, "step": 1371 }, { "epoch": 0.951703806468395, "grad_norm": 1.809305155351, "learning_rate": 1e-05, "loss": 1.3432, "step": 1372 }, { "epoch": 0.9523974681349172, "grad_norm": 1.7869170912103947, "learning_rate": 1e-05, "loss": 1.3173, "step": 1373 }, { "epoch": 0.9530911298014394, "grad_norm": 1.5946853342076313, "learning_rate": 1e-05, "loss": 1.3202, "step": 1374 }, { "epoch": 0.9537847914679615, "grad_norm": 1.7449681923450435, "learning_rate": 1e-05, "loss": 1.3368, "step": 1375 }, { "epoch": 0.9544784531344837, "grad_norm": 1.7867785161300096, "learning_rate": 1e-05, "loss": 1.2523, "step": 1376 }, { "epoch": 0.9551721148010058, "grad_norm": 1.7493793708109926, "learning_rate": 1e-05, "loss": 1.3143, "step": 1377 }, { "epoch": 0.955865776467528, "grad_norm": 1.6708212386781847, "learning_rate": 1e-05, "loss": 1.3197, "step": 1378 }, { "epoch": 0.9565594381340501, "grad_norm": 1.7624923123146528, "learning_rate": 1e-05, "loss": 1.2976, "step": 1379 }, { "epoch": 0.9572530998005723, "grad_norm": 1.8444740226606935, "learning_rate": 1e-05, "loss": 1.3251, "step": 1380 }, { "epoch": 0.9579467614670945, "grad_norm": 1.5953334218837774, "learning_rate": 1e-05, "loss": 1.2664, "step": 1381 }, { "epoch": 0.9586404231336165, "grad_norm": 1.9455556499425986, "learning_rate": 1e-05, "loss": 1.2883, "step": 1382 }, { "epoch": 0.9593340848001387, "grad_norm": 1.7447825684583413, "learning_rate": 1e-05, "loss": 1.3101, "step": 1383 }, { "epoch": 0.9600277464666609, "grad_norm": 1.7815515049059591, "learning_rate": 1e-05, "loss": 1.2931, "step": 1384 }, { "epoch": 0.9607214081331831, "grad_norm": 1.8473821407663016, "learning_rate": 1e-05, "loss": 1.2916, "step": 1385 }, { "epoch": 0.9614150697997051, "grad_norm": 1.8287625516543624, "learning_rate": 1e-05, "loss": 1.3302, "step": 1386 }, { "epoch": 0.9621087314662273, "grad_norm": 1.9131178436318876, "learning_rate": 1e-05, "loss": 1.292, "step": 1387 }, { "epoch": 0.9628023931327495, "grad_norm": 1.6731211620160438, "learning_rate": 1e-05, "loss": 1.2662, "step": 1388 }, { "epoch": 0.9634960547992717, "grad_norm": 1.8100000986534872, "learning_rate": 1e-05, "loss": 1.3462, "step": 1389 }, { "epoch": 0.9641897164657938, "grad_norm": 1.915036526264764, "learning_rate": 1e-05, "loss": 1.2819, "step": 1390 }, { "epoch": 0.964883378132316, "grad_norm": 1.9069098292199362, "learning_rate": 1e-05, "loss": 1.3002, "step": 1391 }, { "epoch": 0.9655770397988381, "grad_norm": 1.7930073332304715, "learning_rate": 1e-05, "loss": 1.2816, "step": 1392 }, { "epoch": 0.9662707014653603, "grad_norm": 1.8530639734903125, "learning_rate": 1e-05, "loss": 1.3521, "step": 1393 }, { "epoch": 0.9669643631318824, "grad_norm": 1.7315815210793186, "learning_rate": 1e-05, "loss": 1.299, "step": 1394 }, { "epoch": 0.9676580247984046, "grad_norm": 1.8034499383080111, "learning_rate": 1e-05, "loss": 1.2712, "step": 1395 }, { "epoch": 0.9683516864649268, "grad_norm": 1.8019607332463998, "learning_rate": 1e-05, "loss": 1.3431, "step": 1396 }, { "epoch": 0.9690453481314489, "grad_norm": 1.7026349253270283, "learning_rate": 1e-05, "loss": 1.3306, "step": 1397 }, { "epoch": 0.969739009797971, "grad_norm": 1.8029112479233105, "learning_rate": 1e-05, "loss": 1.2791, "step": 1398 }, { "epoch": 0.9704326714644932, "grad_norm": 1.6640138276094931, "learning_rate": 1e-05, "loss": 1.3551, "step": 1399 }, { "epoch": 0.9711263331310154, "grad_norm": 1.8773693641133087, "learning_rate": 1e-05, "loss": 1.3214, "step": 1400 }, { "epoch": 0.9718199947975376, "grad_norm": 1.6189625859175014, "learning_rate": 1e-05, "loss": 1.3045, "step": 1401 }, { "epoch": 0.9725136564640596, "grad_norm": 1.8655984733394448, "learning_rate": 1e-05, "loss": 1.295, "step": 1402 }, { "epoch": 0.9732073181305818, "grad_norm": 1.6495459791796045, "learning_rate": 1e-05, "loss": 1.3209, "step": 1403 }, { "epoch": 0.973900979797104, "grad_norm": 1.7262446380853802, "learning_rate": 1e-05, "loss": 1.3179, "step": 1404 }, { "epoch": 0.9745946414636261, "grad_norm": 1.8782015233926619, "learning_rate": 1e-05, "loss": 1.2991, "step": 1405 }, { "epoch": 0.9752883031301482, "grad_norm": 1.6294972087965263, "learning_rate": 1e-05, "loss": 1.3022, "step": 1406 }, { "epoch": 0.9759819647966704, "grad_norm": 1.79967261492696, "learning_rate": 1e-05, "loss": 1.3019, "step": 1407 }, { "epoch": 0.9766756264631926, "grad_norm": 1.8301257668132722, "learning_rate": 1e-05, "loss": 1.2952, "step": 1408 }, { "epoch": 0.9773692881297147, "grad_norm": 1.7858886287008595, "learning_rate": 1e-05, "loss": 1.3455, "step": 1409 }, { "epoch": 0.9780629497962369, "grad_norm": 1.645627337737642, "learning_rate": 1e-05, "loss": 1.2788, "step": 1410 }, { "epoch": 0.978756611462759, "grad_norm": 1.8493218045117754, "learning_rate": 1e-05, "loss": 1.2946, "step": 1411 }, { "epoch": 0.9794502731292812, "grad_norm": 1.9503637295771707, "learning_rate": 1e-05, "loss": 1.2623, "step": 1412 }, { "epoch": 0.9801439347958033, "grad_norm": 1.854587002167059, "learning_rate": 1e-05, "loss": 1.2966, "step": 1413 }, { "epoch": 0.9808375964623255, "grad_norm": 1.7795692734633988, "learning_rate": 1e-05, "loss": 1.3538, "step": 1414 }, { "epoch": 0.9815312581288477, "grad_norm": 1.8026744879009458, "learning_rate": 1e-05, "loss": 1.3534, "step": 1415 }, { "epoch": 0.9822249197953699, "grad_norm": 1.7902484770921805, "learning_rate": 1e-05, "loss": 1.3155, "step": 1416 }, { "epoch": 0.9829185814618919, "grad_norm": 1.7212800748245163, "learning_rate": 1e-05, "loss": 1.2941, "step": 1417 }, { "epoch": 0.9836122431284141, "grad_norm": 1.801323328800465, "learning_rate": 1e-05, "loss": 1.3823, "step": 1418 }, { "epoch": 0.9843059047949363, "grad_norm": 1.7021881473197502, "learning_rate": 1e-05, "loss": 1.3371, "step": 1419 }, { "epoch": 0.9849995664614585, "grad_norm": 1.8046118812459044, "learning_rate": 1e-05, "loss": 1.3433, "step": 1420 }, { "epoch": 0.9856932281279805, "grad_norm": 1.6685271100241381, "learning_rate": 1e-05, "loss": 1.2897, "step": 1421 }, { "epoch": 0.9863868897945027, "grad_norm": 1.6454889789915157, "learning_rate": 1e-05, "loss": 1.3029, "step": 1422 }, { "epoch": 0.9870805514610249, "grad_norm": 1.8167751934930396, "learning_rate": 1e-05, "loss": 1.278, "step": 1423 }, { "epoch": 0.987774213127547, "grad_norm": 1.7280323726187787, "learning_rate": 1e-05, "loss": 1.3137, "step": 1424 }, { "epoch": 0.9884678747940692, "grad_norm": 1.670222942611059, "learning_rate": 1e-05, "loss": 1.3248, "step": 1425 }, { "epoch": 0.9891615364605914, "grad_norm": 1.799342131947682, "learning_rate": 1e-05, "loss": 1.3215, "step": 1426 }, { "epoch": 0.9898551981271135, "grad_norm": 1.8123264473720575, "learning_rate": 1e-05, "loss": 1.3548, "step": 1427 }, { "epoch": 0.9905488597936356, "grad_norm": 1.7747475128663022, "learning_rate": 1e-05, "loss": 1.2765, "step": 1428 }, { "epoch": 0.9912425214601578, "grad_norm": 1.8041680654218195, "learning_rate": 1e-05, "loss": 1.31, "step": 1429 }, { "epoch": 0.99193618312668, "grad_norm": 1.7105917151107914, "learning_rate": 1e-05, "loss": 1.2958, "step": 1430 }, { "epoch": 0.9926298447932022, "grad_norm": 1.6626340057182631, "learning_rate": 1e-05, "loss": 1.3484, "step": 1431 }, { "epoch": 0.9933235064597242, "grad_norm": 1.6610746706958375, "learning_rate": 1e-05, "loss": 1.3256, "step": 1432 }, { "epoch": 0.9940171681262464, "grad_norm": 1.7615692816320323, "learning_rate": 1e-05, "loss": 1.3303, "step": 1433 }, { "epoch": 0.9947108297927686, "grad_norm": 1.7533916584851055, "learning_rate": 1e-05, "loss": 1.2955, "step": 1434 }, { "epoch": 0.9954044914592908, "grad_norm": 1.7232538432339657, "learning_rate": 1e-05, "loss": 1.3177, "step": 1435 }, { "epoch": 0.9960981531258128, "grad_norm": 1.7441612385283174, "learning_rate": 1e-05, "loss": 1.3013, "step": 1436 }, { "epoch": 0.996791814792335, "grad_norm": 1.8597160448645227, "learning_rate": 1e-05, "loss": 1.2923, "step": 1437 }, { "epoch": 0.9974854764588572, "grad_norm": 1.742932722373845, "learning_rate": 1e-05, "loss": 1.3469, "step": 1438 }, { "epoch": 0.9981791381253794, "grad_norm": 1.8961285251801105, "learning_rate": 1e-05, "loss": 1.3101, "step": 1439 }, { "epoch": 0.9988727997919015, "grad_norm": 1.6967076327288442, "learning_rate": 1e-05, "loss": 1.3157, "step": 1440 }, { "epoch": 0.9995664614584236, "grad_norm": 1.669534255209671, "learning_rate": 1e-05, "loss": 1.3203, "step": 1441 }, { "epoch": 0.9995664614584236, "step": 1441, "total_flos": 2332396250726400.0, "train_loss": 1.4015557253881927, "train_runtime": 204310.3582, "train_samples_per_second": 0.903, "train_steps_per_second": 0.007 } ], "logging_steps": 1.0, "max_steps": 1441, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 10, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2332396250726400.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }