{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9982608695652173, "eval_steps": 500, "global_step": 574, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0036036036036036037, "grad_norm": 0.1164047870616515, "learning_rate": 3.5714285714285714e-06, "loss": 0.1263, "step": 1 }, { "epoch": 0.007207207207207207, "grad_norm": 0.1627219461416066, "learning_rate": 7.142857142857143e-06, "loss": 0.1446, "step": 2 }, { "epoch": 0.010810810810810811, "grad_norm": 0.1157756817304506, "learning_rate": 1.0714285714285714e-05, "loss": 0.1441, "step": 3 }, { "epoch": 0.014414414414414415, "grad_norm": 0.14566785288918435, "learning_rate": 1.4285714285714285e-05, "loss": 0.1466, "step": 4 }, { "epoch": 0.018018018018018018, "grad_norm": 0.13068033224281192, "learning_rate": 1.785714285714286e-05, "loss": 0.1342, "step": 5 }, { "epoch": 0.021621621621621623, "grad_norm": 0.15128910055561917, "learning_rate": 2.1428571428571428e-05, "loss": 0.1263, "step": 6 }, { "epoch": 0.025225225225225224, "grad_norm": 0.12625301643275005, "learning_rate": 2.5e-05, "loss": 0.1306, "step": 7 }, { "epoch": 0.02882882882882883, "grad_norm": 0.1341542973939784, "learning_rate": 2.857142857142857e-05, "loss": 0.1132, "step": 8 }, { "epoch": 0.032432432432432434, "grad_norm": 0.09503727827074428, "learning_rate": 3.2142857142857144e-05, "loss": 0.0921, "step": 9 }, { "epoch": 0.036036036036036036, "grad_norm": 0.13724411508071346, "learning_rate": 3.571428571428572e-05, "loss": 0.123, "step": 10 }, { "epoch": 0.03963963963963964, "grad_norm": 0.1249162520121657, "learning_rate": 3.928571428571429e-05, "loss": 0.1183, "step": 11 }, { "epoch": 0.043243243243243246, "grad_norm": 0.14002577339626954, "learning_rate": 4.2857142857142856e-05, "loss": 0.1297, "step": 12 }, { "epoch": 0.04684684684684685, "grad_norm": 0.12032689516166056, "learning_rate": 4.642857142857143e-05, "loss": 0.1144, "step": 13 }, { "epoch": 0.05045045045045045, "grad_norm": 0.11822508923100593, "learning_rate": 5e-05, "loss": 0.127, "step": 14 }, { "epoch": 0.05405405405405406, "grad_norm": 0.12626847998511856, "learning_rate": 5.3571428571428575e-05, "loss": 0.1246, "step": 15 }, { "epoch": 0.05765765765765766, "grad_norm": 0.1394135180306787, "learning_rate": 5.714285714285714e-05, "loss": 0.1249, "step": 16 }, { "epoch": 0.06126126126126126, "grad_norm": 0.1486627737985617, "learning_rate": 6.0714285714285715e-05, "loss": 0.1156, "step": 17 }, { "epoch": 0.06486486486486487, "grad_norm": 0.08348858837628631, "learning_rate": 6.428571428571429e-05, "loss": 0.0972, "step": 18 }, { "epoch": 0.06846846846846846, "grad_norm": 0.19888777350730014, "learning_rate": 6.785714285714286e-05, "loss": 0.146, "step": 19 }, { "epoch": 0.07207207207207207, "grad_norm": 0.12757635434343284, "learning_rate": 7.142857142857143e-05, "loss": 0.109, "step": 20 }, { "epoch": 0.07567567567567568, "grad_norm": 0.17261365946211904, "learning_rate": 7.500000000000001e-05, "loss": 0.139, "step": 21 }, { "epoch": 0.07927927927927927, "grad_norm": 0.19568066543467844, "learning_rate": 7.857142857142858e-05, "loss": 0.1366, "step": 22 }, { "epoch": 0.08288288288288288, "grad_norm": 0.20224174296046235, "learning_rate": 8.214285714285714e-05, "loss": 0.1437, "step": 23 }, { "epoch": 0.08648648648648649, "grad_norm": 0.24075864691751864, "learning_rate": 8.571428571428571e-05, "loss": 0.16, "step": 24 }, { "epoch": 0.09009009009009009, "grad_norm": 0.1506076228406242, "learning_rate": 8.92857142857143e-05, "loss": 0.1074, "step": 25 }, { "epoch": 0.0936936936936937, "grad_norm": 0.16087708567600026, "learning_rate": 9.285714285714286e-05, "loss": 0.1071, "step": 26 }, { "epoch": 0.0972972972972973, "grad_norm": 0.15632293632911032, "learning_rate": 9.642857142857143e-05, "loss": 0.115, "step": 27 }, { "epoch": 0.1009009009009009, "grad_norm": 0.15351274455794925, "learning_rate": 0.0001, "loss": 0.1083, "step": 28 }, { "epoch": 0.1045045045045045, "grad_norm": 0.17410060629498864, "learning_rate": 0.00010357142857142859, "loss": 0.1143, "step": 29 }, { "epoch": 0.10810810810810811, "grad_norm": 0.11003201267949979, "learning_rate": 0.00010714285714285715, "loss": 0.0822, "step": 30 }, { "epoch": 0.11171171171171171, "grad_norm": 0.14850855070731758, "learning_rate": 0.00011071428571428572, "loss": 0.1422, "step": 31 }, { "epoch": 0.11531531531531532, "grad_norm": 0.11574892051893418, "learning_rate": 0.00011428571428571428, "loss": 0.0917, "step": 32 }, { "epoch": 0.11891891891891893, "grad_norm": 0.12172342587299105, "learning_rate": 0.00011785714285714287, "loss": 0.1125, "step": 33 }, { "epoch": 0.12252252252252252, "grad_norm": 0.10233939138594608, "learning_rate": 0.00012142857142857143, "loss": 0.0916, "step": 34 }, { "epoch": 0.12612612612612611, "grad_norm": 0.1277262526433454, "learning_rate": 0.000125, "loss": 0.1168, "step": 35 }, { "epoch": 0.12972972972972974, "grad_norm": 0.1510932624260595, "learning_rate": 0.00012857142857142858, "loss": 0.1366, "step": 36 }, { "epoch": 0.13333333333333333, "grad_norm": 0.1381015918157766, "learning_rate": 0.00013214285714285715, "loss": 0.1112, "step": 37 }, { "epoch": 0.13693693693693693, "grad_norm": 0.12744142332679428, "learning_rate": 0.00013571428571428572, "loss": 0.1169, "step": 38 }, { "epoch": 0.14054054054054055, "grad_norm": 0.12605036861900049, "learning_rate": 0.0001392857142857143, "loss": 0.1265, "step": 39 }, { "epoch": 0.14414414414414414, "grad_norm": 0.09494597572244792, "learning_rate": 0.00014285714285714287, "loss": 0.0856, "step": 40 }, { "epoch": 0.14774774774774774, "grad_norm": 0.11563858827548382, "learning_rate": 0.00014642857142857141, "loss": 0.091, "step": 41 }, { "epoch": 0.15135135135135136, "grad_norm": 0.10978464408087514, "learning_rate": 0.00015000000000000001, "loss": 0.0934, "step": 42 }, { "epoch": 0.15495495495495495, "grad_norm": 0.14748905325763195, "learning_rate": 0.0001535714285714286, "loss": 0.1042, "step": 43 }, { "epoch": 0.15855855855855855, "grad_norm": 0.15028736539177057, "learning_rate": 0.00015714285714285716, "loss": 0.1232, "step": 44 }, { "epoch": 0.16216216216216217, "grad_norm": 0.16834971174993532, "learning_rate": 0.00016071428571428573, "loss": 0.1388, "step": 45 }, { "epoch": 0.16576576576576577, "grad_norm": 0.1364758084150375, "learning_rate": 0.00016428571428571428, "loss": 0.1245, "step": 46 }, { "epoch": 0.16936936936936936, "grad_norm": 0.1906244918264085, "learning_rate": 0.00016785714285714288, "loss": 0.1419, "step": 47 }, { "epoch": 0.17297297297297298, "grad_norm": 0.15582113106280285, "learning_rate": 0.00017142857142857143, "loss": 0.1284, "step": 48 }, { "epoch": 0.17657657657657658, "grad_norm": 0.1653708513472312, "learning_rate": 0.000175, "loss": 0.1265, "step": 49 }, { "epoch": 0.18018018018018017, "grad_norm": 0.11080370021143991, "learning_rate": 0.0001785714285714286, "loss": 0.1136, "step": 50 }, { "epoch": 0.1837837837837838, "grad_norm": 0.14497196744014715, "learning_rate": 0.00018214285714285714, "loss": 0.1336, "step": 51 }, { "epoch": 0.1873873873873874, "grad_norm": 0.11471991362976224, "learning_rate": 0.00018571428571428572, "loss": 0.1009, "step": 52 }, { "epoch": 0.19099099099099098, "grad_norm": 0.1518492774928798, "learning_rate": 0.0001892857142857143, "loss": 0.1445, "step": 53 }, { "epoch": 0.1945945945945946, "grad_norm": 0.15638927885876117, "learning_rate": 0.00019285714285714286, "loss": 0.1095, "step": 54 }, { "epoch": 0.1981981981981982, "grad_norm": 0.12245693248057901, "learning_rate": 0.00019642857142857144, "loss": 0.099, "step": 55 }, { "epoch": 0.2018018018018018, "grad_norm": 0.13146029758520172, "learning_rate": 0.0002, "loss": 0.1215, "step": 56 }, { "epoch": 0.20540540540540542, "grad_norm": 0.1449923810118862, "learning_rate": 0.00019999801019909556, "loss": 0.1376, "step": 57 }, { "epoch": 0.209009009009009, "grad_norm": 0.12275432649506118, "learning_rate": 0.0001999920408755684, "loss": 0.111, "step": 58 }, { "epoch": 0.2126126126126126, "grad_norm": 0.13294914056261917, "learning_rate": 0.00019998209226697376, "loss": 0.1184, "step": 59 }, { "epoch": 0.21621621621621623, "grad_norm": 0.11587311682416103, "learning_rate": 0.00019996816476922677, "loss": 0.1029, "step": 60 }, { "epoch": 0.21981981981981982, "grad_norm": 0.211706197616785, "learning_rate": 0.00019995025893658627, "loss": 0.1323, "step": 61 }, { "epoch": 0.22342342342342342, "grad_norm": 0.135734632583536, "learning_rate": 0.00019992837548163316, "loss": 0.1073, "step": 62 }, { "epoch": 0.22702702702702704, "grad_norm": 0.1916821730614324, "learning_rate": 0.00019990251527524178, "loss": 0.1287, "step": 63 }, { "epoch": 0.23063063063063063, "grad_norm": 0.1696786518231171, "learning_rate": 0.00019987267934654538, "loss": 0.1467, "step": 64 }, { "epoch": 0.23423423423423423, "grad_norm": 0.1556597769170162, "learning_rate": 0.00019983886888289514, "loss": 0.1074, "step": 65 }, { "epoch": 0.23783783783783785, "grad_norm": 0.11435037299616506, "learning_rate": 0.00019980108522981284, "loss": 0.1054, "step": 66 }, { "epoch": 0.24144144144144145, "grad_norm": 0.1377527772698083, "learning_rate": 0.00019975932989093747, "loss": 0.1167, "step": 67 }, { "epoch": 0.24504504504504504, "grad_norm": 0.13089085083126692, "learning_rate": 0.00019971360452796522, "loss": 0.1268, "step": 68 }, { "epoch": 0.24864864864864866, "grad_norm": 0.1461221542311374, "learning_rate": 0.00019966391096058346, "loss": 0.1353, "step": 69 }, { "epoch": 0.25225225225225223, "grad_norm": 0.10972604571483792, "learning_rate": 0.0001996102511663983, "loss": 0.0839, "step": 70 }, { "epoch": 0.25585585585585585, "grad_norm": 0.14694411558687645, "learning_rate": 0.0001995526272808559, "loss": 0.1266, "step": 71 }, { "epoch": 0.2594594594594595, "grad_norm": 0.2623900472581046, "learning_rate": 0.00019949104159715743, "loss": 0.1192, "step": 72 }, { "epoch": 0.26306306306306304, "grad_norm": 0.14253202316127417, "learning_rate": 0.0001994254965661679, "loss": 0.1268, "step": 73 }, { "epoch": 0.26666666666666666, "grad_norm": 0.22775504622269988, "learning_rate": 0.0001993559947963185, "loss": 0.1624, "step": 74 }, { "epoch": 0.2702702702702703, "grad_norm": 0.18974052313619846, "learning_rate": 0.00019928253905350296, "loss": 0.1656, "step": 75 }, { "epoch": 0.27387387387387385, "grad_norm": 0.18281811162027828, "learning_rate": 0.00019920513226096733, "loss": 0.1512, "step": 76 }, { "epoch": 0.2774774774774775, "grad_norm": 0.15981121539784604, "learning_rate": 0.00019912377749919374, "loss": 0.1414, "step": 77 }, { "epoch": 0.2810810810810811, "grad_norm": 0.11107015310290616, "learning_rate": 0.00019903847800577777, "loss": 0.0732, "step": 78 }, { "epoch": 0.28468468468468466, "grad_norm": 0.11807284847655806, "learning_rate": 0.00019894923717529955, "loss": 0.1158, "step": 79 }, { "epoch": 0.2882882882882883, "grad_norm": 0.1674132871400004, "learning_rate": 0.00019885605855918885, "loss": 0.1363, "step": 80 }, { "epoch": 0.2918918918918919, "grad_norm": 0.09521837972620555, "learning_rate": 0.00019875894586558355, "loss": 0.0761, "step": 81 }, { "epoch": 0.2954954954954955, "grad_norm": 0.13313059706266978, "learning_rate": 0.00019865790295918212, "loss": 0.114, "step": 82 }, { "epoch": 0.2990990990990991, "grad_norm": 0.15752991450823575, "learning_rate": 0.00019855293386108992, "loss": 0.1143, "step": 83 }, { "epoch": 0.3027027027027027, "grad_norm": 0.11681361221271575, "learning_rate": 0.0001984440427486591, "loss": 0.0955, "step": 84 }, { "epoch": 0.3063063063063063, "grad_norm": 0.1435158350817726, "learning_rate": 0.00019833123395532226, "loss": 0.1292, "step": 85 }, { "epoch": 0.3099099099099099, "grad_norm": 0.1174821097766054, "learning_rate": 0.00019821451197042026, "loss": 0.119, "step": 86 }, { "epoch": 0.31351351351351353, "grad_norm": 0.14421204301690782, "learning_rate": 0.00019809388143902332, "loss": 0.1313, "step": 87 }, { "epoch": 0.3171171171171171, "grad_norm": 0.13517965622709482, "learning_rate": 0.0001979693471617462, "loss": 0.1297, "step": 88 }, { "epoch": 0.3207207207207207, "grad_norm": 0.17145867539050777, "learning_rate": 0.00019784091409455728, "loss": 0.1359, "step": 89 }, { "epoch": 0.32432432432432434, "grad_norm": 0.11953352813577937, "learning_rate": 0.00019770858734858126, "loss": 0.0878, "step": 90 }, { "epoch": 0.3279279279279279, "grad_norm": 0.13673174823647213, "learning_rate": 0.00019757237218989563, "loss": 0.1265, "step": 91 }, { "epoch": 0.33153153153153153, "grad_norm": 0.16490479257041854, "learning_rate": 0.00019743227403932134, "loss": 0.1393, "step": 92 }, { "epoch": 0.33513513513513515, "grad_norm": 0.12343228690652325, "learning_rate": 0.000197288298472207, "loss": 0.1174, "step": 93 }, { "epoch": 0.3387387387387387, "grad_norm": 0.16745916159569352, "learning_rate": 0.00019714045121820676, "loss": 0.1235, "step": 94 }, { "epoch": 0.34234234234234234, "grad_norm": 0.16922526895888806, "learning_rate": 0.00019698873816105273, "loss": 0.0975, "step": 95 }, { "epoch": 0.34594594594594597, "grad_norm": 0.129484224418453, "learning_rate": 0.00019683316533832042, "loss": 0.0928, "step": 96 }, { "epoch": 0.34954954954954953, "grad_norm": 0.1514002748369919, "learning_rate": 0.0001966737389411887, "loss": 0.1341, "step": 97 }, { "epoch": 0.35315315315315315, "grad_norm": 0.11053077601153272, "learning_rate": 0.00019651046531419332, "loss": 0.09, "step": 98 }, { "epoch": 0.3567567567567568, "grad_norm": 0.11955256535981768, "learning_rate": 0.00019634335095497458, "loss": 0.0978, "step": 99 }, { "epoch": 0.36036036036036034, "grad_norm": 0.12318848470518083, "learning_rate": 0.0001961724025140185, "loss": 0.1123, "step": 100 }, { "epoch": 0.36396396396396397, "grad_norm": 0.19878288570661823, "learning_rate": 0.0001959976267943923, "loss": 0.1449, "step": 101 }, { "epoch": 0.3675675675675676, "grad_norm": 0.11498348089609609, "learning_rate": 0.0001958190307514737, "loss": 0.101, "step": 102 }, { "epoch": 0.37117117117117115, "grad_norm": 0.12807480490548945, "learning_rate": 0.00019563662149267406, "loss": 0.1115, "step": 103 }, { "epoch": 0.3747747747747748, "grad_norm": 0.1537951698344796, "learning_rate": 0.0001954504062771555, "loss": 0.1099, "step": 104 }, { "epoch": 0.3783783783783784, "grad_norm": 0.13376774584465406, "learning_rate": 0.0001952603925155422, "loss": 0.0945, "step": 105 }, { "epoch": 0.38198198198198197, "grad_norm": 0.11095795904499461, "learning_rate": 0.0001950665877696252, "loss": 0.1001, "step": 106 }, { "epoch": 0.3855855855855856, "grad_norm": 0.1176293890483276, "learning_rate": 0.00019486899975206166, "loss": 0.1114, "step": 107 }, { "epoch": 0.3891891891891892, "grad_norm": 0.16600471258328028, "learning_rate": 0.0001946676363260679, "loss": 0.1565, "step": 108 }, { "epoch": 0.3927927927927928, "grad_norm": 0.12969105825015786, "learning_rate": 0.0001944625055051065, "loss": 0.0942, "step": 109 }, { "epoch": 0.3963963963963964, "grad_norm": 0.1260399594140325, "learning_rate": 0.00019425361545256727, "loss": 0.1151, "step": 110 }, { "epoch": 0.4, "grad_norm": 0.12141152738985596, "learning_rate": 0.00019404097448144257, "loss": 0.0953, "step": 111 }, { "epoch": 0.4036036036036036, "grad_norm": 0.16522441528864815, "learning_rate": 0.00019382459105399632, "loss": 0.1483, "step": 112 }, { "epoch": 0.4072072072072072, "grad_norm": 0.16464941562962845, "learning_rate": 0.00019360447378142728, "loss": 0.1145, "step": 113 }, { "epoch": 0.41081081081081083, "grad_norm": 0.1301041115410939, "learning_rate": 0.00019338063142352644, "loss": 0.109, "step": 114 }, { "epoch": 0.4144144144144144, "grad_norm": 0.15394069789981274, "learning_rate": 0.00019315307288832835, "loss": 0.1484, "step": 115 }, { "epoch": 0.418018018018018, "grad_norm": 0.1337025315682798, "learning_rate": 0.00019292180723175654, "loss": 0.1083, "step": 116 }, { "epoch": 0.42162162162162165, "grad_norm": 0.11697191221298965, "learning_rate": 0.00019268684365726326, "loss": 0.1104, "step": 117 }, { "epoch": 0.4252252252252252, "grad_norm": 0.14885108765057334, "learning_rate": 0.00019244819151546322, "loss": 0.1349, "step": 118 }, { "epoch": 0.42882882882882883, "grad_norm": 0.16748465670739565, "learning_rate": 0.00019220586030376134, "loss": 0.1375, "step": 119 }, { "epoch": 0.43243243243243246, "grad_norm": 0.16065529576883042, "learning_rate": 0.00019195985966597494, "loss": 0.1158, "step": 120 }, { "epoch": 0.436036036036036, "grad_norm": 0.14710431466862364, "learning_rate": 0.0001917101993919498, "loss": 0.1123, "step": 121 }, { "epoch": 0.43963963963963965, "grad_norm": 0.1535583565878682, "learning_rate": 0.00019145688941717075, "loss": 0.1244, "step": 122 }, { "epoch": 0.44324324324324327, "grad_norm": 0.15887496082691002, "learning_rate": 0.00019119993982236606, "loss": 0.1099, "step": 123 }, { "epoch": 0.44684684684684683, "grad_norm": 0.17132720394894463, "learning_rate": 0.00019093936083310653, "loss": 0.1366, "step": 124 }, { "epoch": 0.45045045045045046, "grad_norm": 0.1304195997449305, "learning_rate": 0.00019067516281939825, "loss": 0.1042, "step": 125 }, { "epoch": 0.4540540540540541, "grad_norm": 0.13720183539624425, "learning_rate": 0.00019040735629527027, "loss": 0.0939, "step": 126 }, { "epoch": 0.45765765765765765, "grad_norm": 0.1878348429175824, "learning_rate": 0.00019013595191835574, "loss": 0.1421, "step": 127 }, { "epoch": 0.46126126126126127, "grad_norm": 0.15221296411188612, "learning_rate": 0.00018986096048946824, "loss": 0.1207, "step": 128 }, { "epoch": 0.4648648648648649, "grad_norm": 0.12530318604533355, "learning_rate": 0.0001895823929521716, "loss": 0.1101, "step": 129 }, { "epoch": 0.46846846846846846, "grad_norm": 0.11753990553496706, "learning_rate": 0.0001893002603923446, "loss": 0.0814, "step": 130 }, { "epoch": 0.4720720720720721, "grad_norm": 0.15143122574748422, "learning_rate": 0.00018901457403773967, "loss": 0.1259, "step": 131 }, { "epoch": 0.4756756756756757, "grad_norm": 0.12436714806981373, "learning_rate": 0.00018872534525753615, "loss": 0.1148, "step": 132 }, { "epoch": 0.47927927927927927, "grad_norm": 0.13100199539485474, "learning_rate": 0.00018843258556188787, "loss": 0.1189, "step": 133 }, { "epoch": 0.4828828828828829, "grad_norm": 0.1667053146851425, "learning_rate": 0.00018813630660146488, "loss": 0.1494, "step": 134 }, { "epoch": 0.4864864864864865, "grad_norm": 0.09073673518826318, "learning_rate": 0.00018783652016699014, "loss": 0.0799, "step": 135 }, { "epoch": 0.4900900900900901, "grad_norm": 0.13343062211829884, "learning_rate": 0.0001875332381887699, "loss": 0.1241, "step": 136 }, { "epoch": 0.4936936936936937, "grad_norm": 0.12750970397329575, "learning_rate": 0.0001872264727362194, "loss": 0.1386, "step": 137 }, { "epoch": 0.4972972972972973, "grad_norm": 0.10441877909622974, "learning_rate": 0.00018691623601738199, "loss": 0.0888, "step": 138 }, { "epoch": 0.5009009009009009, "grad_norm": 0.14478179850573814, "learning_rate": 0.00018660254037844388, "loss": 0.1056, "step": 139 }, { "epoch": 0.5045045045045045, "grad_norm": 0.13712755003139512, "learning_rate": 0.00018628539830324229, "loss": 0.1489, "step": 140 }, { "epoch": 0.5081081081081081, "grad_norm": 0.12379021926600628, "learning_rate": 0.000185964822412769, "loss": 0.1071, "step": 141 }, { "epoch": 0.5117117117117117, "grad_norm": 0.12266406139545731, "learning_rate": 0.00018564082546466805, "loss": 0.1141, "step": 142 }, { "epoch": 0.5153153153153153, "grad_norm": 0.09332411107267007, "learning_rate": 0.00018531342035272766, "loss": 0.0876, "step": 143 }, { "epoch": 0.518918918918919, "grad_norm": 0.12722229275266542, "learning_rate": 0.00018498262010636774, "loss": 0.123, "step": 144 }, { "epoch": 0.5225225225225225, "grad_norm": 0.1934624210241968, "learning_rate": 0.00018464843789012085, "loss": 0.1891, "step": 145 }, { "epoch": 0.5261261261261261, "grad_norm": 0.1202206919464269, "learning_rate": 0.00018431088700310844, "loss": 0.1157, "step": 146 }, { "epoch": 0.5297297297297298, "grad_norm": 0.11855534138749764, "learning_rate": 0.0001839699808785118, "loss": 0.1126, "step": 147 }, { "epoch": 0.5333333333333333, "grad_norm": 0.10719514027165045, "learning_rate": 0.00018362573308303718, "loss": 0.0907, "step": 148 }, { "epoch": 0.5369369369369369, "grad_norm": 0.11210467216409752, "learning_rate": 0.00018327815731637612, "loss": 0.1007, "step": 149 }, { "epoch": 0.5405405405405406, "grad_norm": 0.12526617885340885, "learning_rate": 0.00018292726741066007, "loss": 0.1049, "step": 150 }, { "epoch": 0.5441441441441441, "grad_norm": 0.1477303393799172, "learning_rate": 0.00018257307732991008, "loss": 0.1516, "step": 151 }, { "epoch": 0.5477477477477477, "grad_norm": 0.14857702506705278, "learning_rate": 0.00018221560116948103, "loss": 0.1453, "step": 152 }, { "epoch": 0.5513513513513514, "grad_norm": 0.14008518634545825, "learning_rate": 0.0001818548531555006, "loss": 0.1297, "step": 153 }, { "epoch": 0.554954954954955, "grad_norm": 0.12658212522638404, "learning_rate": 0.0001814908476443034, "loss": 0.1155, "step": 154 }, { "epoch": 0.5585585585585585, "grad_norm": 0.149670716923037, "learning_rate": 0.00018112359912185924, "loss": 0.1211, "step": 155 }, { "epoch": 0.5621621621621622, "grad_norm": 0.11342605203968036, "learning_rate": 0.000180753122203197, "loss": 0.0899, "step": 156 }, { "epoch": 0.5657657657657658, "grad_norm": 0.15888593819383173, "learning_rate": 0.00018037943163182283, "loss": 0.1445, "step": 157 }, { "epoch": 0.5693693693693693, "grad_norm": 0.12437893978089608, "learning_rate": 0.00018000254227913348, "loss": 0.1152, "step": 158 }, { "epoch": 0.572972972972973, "grad_norm": 0.11638937373238138, "learning_rate": 0.0001796224691438244, "loss": 0.1123, "step": 159 }, { "epoch": 0.5765765765765766, "grad_norm": 0.14812854362945038, "learning_rate": 0.00017923922735129302, "loss": 0.1263, "step": 160 }, { "epoch": 0.5801801801801801, "grad_norm": 0.10770071386782099, "learning_rate": 0.0001788528321530366, "loss": 0.0955, "step": 161 }, { "epoch": 0.5837837837837838, "grad_norm": 0.1870539683925041, "learning_rate": 0.00017846329892604547, "loss": 0.1124, "step": 162 }, { "epoch": 0.5873873873873874, "grad_norm": 0.1560374478952629, "learning_rate": 0.00017807064317219094, "loss": 0.122, "step": 163 }, { "epoch": 0.590990990990991, "grad_norm": 0.14789972168680796, "learning_rate": 0.00017767488051760857, "loss": 0.0955, "step": 164 }, { "epoch": 0.5945945945945946, "grad_norm": 0.17954009944461283, "learning_rate": 0.00017727602671207605, "loss": 0.1326, "step": 165 }, { "epoch": 0.5981981981981982, "grad_norm": 0.12473531577026101, "learning_rate": 0.00017687409762838664, "loss": 0.139, "step": 166 }, { "epoch": 0.6018018018018018, "grad_norm": 0.18890214448118112, "learning_rate": 0.00017646910926171747, "loss": 0.158, "step": 167 }, { "epoch": 0.6054054054054054, "grad_norm": 0.1158510197827391, "learning_rate": 0.00017606107772899287, "loss": 0.124, "step": 168 }, { "epoch": 0.609009009009009, "grad_norm": 0.1513359972404607, "learning_rate": 0.00017565001926824313, "loss": 0.1535, "step": 169 }, { "epoch": 0.6126126126126126, "grad_norm": 0.11561240472832256, "learning_rate": 0.00017523595023795813, "loss": 0.097, "step": 170 }, { "epoch": 0.6162162162162163, "grad_norm": 0.14453378759822266, "learning_rate": 0.00017481888711643655, "loss": 0.1369, "step": 171 }, { "epoch": 0.6198198198198198, "grad_norm": 0.10823698221755142, "learning_rate": 0.00017439884650112989, "loss": 0.0854, "step": 172 }, { "epoch": 0.6234234234234234, "grad_norm": 0.16461158555393735, "learning_rate": 0.0001739758451079821, "loss": 0.1327, "step": 173 }, { "epoch": 0.6270270270270271, "grad_norm": 0.13330810816894179, "learning_rate": 0.00017354989977076422, "loss": 0.0988, "step": 174 }, { "epoch": 0.6306306306306306, "grad_norm": 0.1603897957937655, "learning_rate": 0.00017312102744040467, "loss": 0.1517, "step": 175 }, { "epoch": 0.6342342342342342, "grad_norm": 0.1387499574229483, "learning_rate": 0.00017268924518431438, "loss": 0.1159, "step": 176 }, { "epoch": 0.6378378378378379, "grad_norm": 0.15123399261590567, "learning_rate": 0.0001722545701857079, "loss": 0.135, "step": 177 }, { "epoch": 0.6414414414414414, "grad_norm": 0.201686818845506, "learning_rate": 0.0001718170197429193, "loss": 0.1601, "step": 178 }, { "epoch": 0.645045045045045, "grad_norm": 0.16050791333444517, "learning_rate": 0.0001713766112687139, "loss": 0.1376, "step": 179 }, { "epoch": 0.6486486486486487, "grad_norm": 0.13004224853328716, "learning_rate": 0.00017093336228959536, "loss": 0.1191, "step": 180 }, { "epoch": 0.6522522522522523, "grad_norm": 0.10061992398695434, "learning_rate": 0.000170487290445108, "loss": 0.0958, "step": 181 }, { "epoch": 0.6558558558558558, "grad_norm": 0.09779721051938423, "learning_rate": 0.0001700384134871351, "loss": 0.098, "step": 182 }, { "epoch": 0.6594594594594595, "grad_norm": 0.12862092154540355, "learning_rate": 0.0001695867492791921, "loss": 0.1083, "step": 183 }, { "epoch": 0.6630630630630631, "grad_norm": 0.13476322854527875, "learning_rate": 0.00016913231579571608, "loss": 0.1466, "step": 184 }, { "epoch": 0.6666666666666666, "grad_norm": 0.09595530640274692, "learning_rate": 0.00016867513112135013, "loss": 0.0842, "step": 185 }, { "epoch": 0.6702702702702703, "grad_norm": 0.15679543098949758, "learning_rate": 0.00016821521345022377, "loss": 0.1338, "step": 186 }, { "epoch": 0.6738738738738739, "grad_norm": 0.14388550615027906, "learning_rate": 0.00016775258108522908, "loss": 0.1125, "step": 187 }, { "epoch": 0.6774774774774774, "grad_norm": 0.14073204006731552, "learning_rate": 0.0001672872524372919, "loss": 0.139, "step": 188 }, { "epoch": 0.6810810810810811, "grad_norm": 0.09327785295917886, "learning_rate": 0.00016681924602463962, "loss": 0.0876, "step": 189 }, { "epoch": 0.6846846846846847, "grad_norm": 0.0966354577674113, "learning_rate": 0.00016634858047206378, "loss": 0.0817, "step": 190 }, { "epoch": 0.6882882882882883, "grad_norm": 0.1298212529485729, "learning_rate": 0.00016587527451017938, "loss": 0.1248, "step": 191 }, { "epoch": 0.6918918918918919, "grad_norm": 0.15190505228456444, "learning_rate": 0.00016539934697467894, "loss": 0.1346, "step": 192 }, { "epoch": 0.6954954954954955, "grad_norm": 0.12074435445615049, "learning_rate": 0.0001649208168055833, "loss": 0.1218, "step": 193 }, { "epoch": 0.6990990990990991, "grad_norm": 0.11339361129121636, "learning_rate": 0.0001644397030464877, "loss": 0.0945, "step": 194 }, { "epoch": 0.7027027027027027, "grad_norm": 0.1480633681266718, "learning_rate": 0.00016395602484380406, "loss": 0.143, "step": 195 }, { "epoch": 0.7063063063063063, "grad_norm": 0.13202765755871132, "learning_rate": 0.0001634698014459988, "loss": 0.1256, "step": 196 }, { "epoch": 0.7099099099099099, "grad_norm": 0.10905065599283695, "learning_rate": 0.00016298105220282713, "loss": 0.1024, "step": 197 }, { "epoch": 0.7135135135135136, "grad_norm": 0.10616436723037755, "learning_rate": 0.00016248979656456275, "loss": 0.1066, "step": 198 }, { "epoch": 0.7171171171171171, "grad_norm": 0.1063733952868901, "learning_rate": 0.0001619960540812239, "loss": 0.1065, "step": 199 }, { "epoch": 0.7207207207207207, "grad_norm": 0.1648449550913926, "learning_rate": 0.00016149984440179537, "loss": 0.1416, "step": 200 }, { "epoch": 0.7243243243243244, "grad_norm": 0.14504142427358913, "learning_rate": 0.00016100118727344659, "loss": 0.1323, "step": 201 }, { "epoch": 0.7279279279279279, "grad_norm": 0.15511743070098452, "learning_rate": 0.00016050010254074564, "loss": 0.1259, "step": 202 }, { "epoch": 0.7315315315315315, "grad_norm": 0.12280785450706579, "learning_rate": 0.00015999661014486956, "loss": 0.1165, "step": 203 }, { "epoch": 0.7351351351351352, "grad_norm": 0.13888404263902684, "learning_rate": 0.00015949073012281093, "loss": 0.1047, "step": 204 }, { "epoch": 0.7387387387387387, "grad_norm": 0.10651036692593711, "learning_rate": 0.00015898248260658016, "loss": 0.1181, "step": 205 }, { "epoch": 0.7423423423423423, "grad_norm": 0.10861974936989245, "learning_rate": 0.0001584718878224047, "loss": 0.1064, "step": 206 }, { "epoch": 0.745945945945946, "grad_norm": 0.10231866176721904, "learning_rate": 0.00015795896608992378, "loss": 0.0988, "step": 207 }, { "epoch": 0.7495495495495496, "grad_norm": 0.15024568241023914, "learning_rate": 0.00015744373782137992, "loss": 0.1489, "step": 208 }, { "epoch": 0.7531531531531531, "grad_norm": 0.12371588452286458, "learning_rate": 0.00015692622352080662, "loss": 0.116, "step": 209 }, { "epoch": 0.7567567567567568, "grad_norm": 0.11392721432010788, "learning_rate": 0.00015640644378321235, "loss": 0.1015, "step": 210 }, { "epoch": 0.7603603603603604, "grad_norm": 0.11201427932233406, "learning_rate": 0.00015588441929376097, "loss": 0.0863, "step": 211 }, { "epoch": 0.7639639639639639, "grad_norm": 0.159849063390471, "learning_rate": 0.00015536017082694846, "loss": 0.1651, "step": 212 }, { "epoch": 0.7675675675675676, "grad_norm": 0.13935698141384686, "learning_rate": 0.00015483371924577635, "loss": 0.1262, "step": 213 }, { "epoch": 0.7711711711711712, "grad_norm": 0.15388913192797118, "learning_rate": 0.00015430508550092124, "loss": 0.1602, "step": 214 }, { "epoch": 0.7747747747747747, "grad_norm": 0.11744911276482749, "learning_rate": 0.00015377429062990122, "loss": 0.1082, "step": 215 }, { "epoch": 0.7783783783783784, "grad_norm": 0.14669529425537173, "learning_rate": 0.00015324135575623857, "loss": 0.1329, "step": 216 }, { "epoch": 0.781981981981982, "grad_norm": 0.09725689202217797, "learning_rate": 0.00015270630208861916, "loss": 0.1001, "step": 217 }, { "epoch": 0.7855855855855856, "grad_norm": 0.09066648478601479, "learning_rate": 0.00015216915092004847, "loss": 0.1005, "step": 218 }, { "epoch": 0.7891891891891892, "grad_norm": 0.10556590806339675, "learning_rate": 0.00015162992362700406, "loss": 0.104, "step": 219 }, { "epoch": 0.7927927927927928, "grad_norm": 0.10568504388848617, "learning_rate": 0.00015108864166858506, "loss": 0.1079, "step": 220 }, { "epoch": 0.7963963963963964, "grad_norm": 0.13168798693648778, "learning_rate": 0.0001505453265856581, "loss": 0.1319, "step": 221 }, { "epoch": 0.8, "grad_norm": 0.11471998852906086, "learning_rate": 0.00015000000000000001, "loss": 0.1181, "step": 222 }, { "epoch": 0.8036036036036036, "grad_norm": 0.11715811264986671, "learning_rate": 0.00014945268361343748, "loss": 0.1053, "step": 223 }, { "epoch": 0.8072072072072072, "grad_norm": 0.10491568472945026, "learning_rate": 0.00014890339920698334, "loss": 0.0931, "step": 224 }, { "epoch": 0.8108108108108109, "grad_norm": 0.1381588520705617, "learning_rate": 0.00014835216863996975, "loss": 0.1417, "step": 225 }, { "epoch": 0.8144144144144144, "grad_norm": 0.15507649891779268, "learning_rate": 0.0001477990138491783, "loss": 0.1418, "step": 226 }, { "epoch": 0.818018018018018, "grad_norm": 0.11750139732263555, "learning_rate": 0.0001472439568479671, "loss": 0.1207, "step": 227 }, { "epoch": 0.8216216216216217, "grad_norm": 0.12736893690378323, "learning_rate": 0.00014668701972539458, "loss": 0.1277, "step": 228 }, { "epoch": 0.8252252252252252, "grad_norm": 0.11333975714941213, "learning_rate": 0.00014612822464534059, "loss": 0.1113, "step": 229 }, { "epoch": 0.8288288288288288, "grad_norm": 0.11811846805876995, "learning_rate": 0.00014556759384562416, "loss": 0.1174, "step": 230 }, { "epoch": 0.8324324324324325, "grad_norm": 0.12819157247369997, "learning_rate": 0.00014500514963711883, "loss": 0.1143, "step": 231 }, { "epoch": 0.836036036036036, "grad_norm": 0.11727175144557134, "learning_rate": 0.0001444409144028644, "loss": 0.1153, "step": 232 }, { "epoch": 0.8396396396396396, "grad_norm": 0.13457703292067713, "learning_rate": 0.00014387491059717652, "loss": 0.1199, "step": 233 }, { "epoch": 0.8432432432432433, "grad_norm": 0.11901299124274167, "learning_rate": 0.00014330716074475286, "loss": 0.1147, "step": 234 }, { "epoch": 0.8468468468468469, "grad_norm": 0.10353994887251415, "learning_rate": 0.00014273768743977685, "loss": 0.1026, "step": 235 }, { "epoch": 0.8504504504504504, "grad_norm": 0.10419191980690304, "learning_rate": 0.0001421665133450184, "loss": 0.1063, "step": 236 }, { "epoch": 0.8540540540540541, "grad_norm": 0.12748698891225302, "learning_rate": 0.00014159366119093214, "loss": 0.1079, "step": 237 }, { "epoch": 0.8576576576576577, "grad_norm": 0.16200721887310557, "learning_rate": 0.00014101915377475274, "loss": 0.1152, "step": 238 }, { "epoch": 0.8612612612612612, "grad_norm": 0.15795975577284813, "learning_rate": 0.0001404430139595877, "loss": 0.1542, "step": 239 }, { "epoch": 0.8648648648648649, "grad_norm": 0.14933463930244448, "learning_rate": 0.0001398652646735076, "loss": 0.1236, "step": 240 }, { "epoch": 0.8684684684684685, "grad_norm": 0.16198753222835588, "learning_rate": 0.0001392859289086334, "loss": 0.1375, "step": 241 }, { "epoch": 0.872072072072072, "grad_norm": 0.13433467388254222, "learning_rate": 0.00013870502972022173, "loss": 0.1323, "step": 242 }, { "epoch": 0.8756756756756757, "grad_norm": 0.12593674925296103, "learning_rate": 0.00013812259022574717, "loss": 0.1216, "step": 243 }, { "epoch": 0.8792792792792793, "grad_norm": 0.13013719230493928, "learning_rate": 0.00013753863360398241, "loss": 0.1247, "step": 244 }, { "epoch": 0.8828828828828829, "grad_norm": 0.11799267349520824, "learning_rate": 0.0001369531830940757, "loss": 0.1086, "step": 245 }, { "epoch": 0.8864864864864865, "grad_norm": 0.08312084262618047, "learning_rate": 0.00013636626199462615, "loss": 0.0813, "step": 246 }, { "epoch": 0.8900900900900901, "grad_norm": 0.1338651554767216, "learning_rate": 0.00013577789366275644, "loss": 0.137, "step": 247 }, { "epoch": 0.8936936936936937, "grad_norm": 0.10150227632820087, "learning_rate": 0.0001351881015131833, "loss": 0.0975, "step": 248 }, { "epoch": 0.8972972972972973, "grad_norm": 0.10189929428402296, "learning_rate": 0.00013459690901728588, "loss": 0.0923, "step": 249 }, { "epoch": 0.9009009009009009, "grad_norm": 0.1408210936693087, "learning_rate": 0.00013400433970217135, "loss": 0.1378, "step": 250 }, { "epoch": 0.9045045045045045, "grad_norm": 0.11765895193363322, "learning_rate": 0.000133410417149739, "loss": 0.1096, "step": 251 }, { "epoch": 0.9081081081081082, "grad_norm": 0.1413792560787727, "learning_rate": 0.00013281516499574135, "loss": 0.1401, "step": 252 }, { "epoch": 0.9117117117117117, "grad_norm": 0.08054406846656884, "learning_rate": 0.00013221860692884396, "loss": 0.0835, "step": 253 }, { "epoch": 0.9153153153153153, "grad_norm": 0.12127761773938303, "learning_rate": 0.0001316207666896824, "loss": 0.118, "step": 254 }, { "epoch": 0.918918918918919, "grad_norm": 0.10139113989817501, "learning_rate": 0.00013102166806991768, "loss": 0.0966, "step": 255 }, { "epoch": 0.9225225225225225, "grad_norm": 0.10511129293269068, "learning_rate": 0.00013042133491128935, "loss": 0.0846, "step": 256 }, { "epoch": 0.9261261261261261, "grad_norm": 0.13928639672942275, "learning_rate": 0.00012981979110466654, "loss": 0.1106, "step": 257 }, { "epoch": 0.9297297297297298, "grad_norm": 0.1575504268549112, "learning_rate": 0.00012921706058909756, "loss": 0.1022, "step": 258 }, { "epoch": 0.9333333333333333, "grad_norm": 0.09456528837585412, "learning_rate": 0.00012861316735085686, "loss": 0.0943, "step": 259 }, { "epoch": 0.9369369369369369, "grad_norm": 0.11421875251828266, "learning_rate": 0.00012800813542249072, "loss": 0.0988, "step": 260 }, { "epoch": 0.9405405405405406, "grad_norm": 0.11985070545179864, "learning_rate": 0.00012740198888186064, "loss": 0.1238, "step": 261 }, { "epoch": 0.9441441441441442, "grad_norm": 0.09679571111756961, "learning_rate": 0.00012679475185118535, "loss": 0.1063, "step": 262 }, { "epoch": 0.9477477477477477, "grad_norm": 0.09782919038732428, "learning_rate": 0.0001261864484960807, "loss": 0.1039, "step": 263 }, { "epoch": 0.9513513513513514, "grad_norm": 0.17653628828090737, "learning_rate": 0.00012557710302459803, "loss": 0.1354, "step": 264 }, { "epoch": 0.954954954954955, "grad_norm": 0.1409157686607275, "learning_rate": 0.00012496673968626068, "loss": 0.1181, "step": 265 }, { "epoch": 0.9585585585585585, "grad_norm": 0.16396955244736236, "learning_rate": 0.0001243553827710992, "loss": 0.1352, "step": 266 }, { "epoch": 0.9621621621621622, "grad_norm": 0.11706567681214818, "learning_rate": 0.0001237430566086844, "loss": 0.1103, "step": 267 }, { "epoch": 0.9657657657657658, "grad_norm": 0.11951814155751256, "learning_rate": 0.00012312978556715932, "loss": 0.1182, "step": 268 }, { "epoch": 0.9693693693693693, "grad_norm": 0.1098976660754676, "learning_rate": 0.00012251559405226941, "loss": 0.0981, "step": 269 }, { "epoch": 0.972972972972973, "grad_norm": 0.13497038508376635, "learning_rate": 0.00012190050650639131, "loss": 0.139, "step": 270 }, { "epoch": 0.9765765765765766, "grad_norm": 0.10505221561748224, "learning_rate": 0.00012128454740756014, "loss": 0.0968, "step": 271 }, { "epoch": 0.9801801801801802, "grad_norm": 0.09400827669331373, "learning_rate": 0.00012066774126849529, "loss": 0.091, "step": 272 }, { "epoch": 0.9837837837837838, "grad_norm": 0.13553635299634834, "learning_rate": 0.00012005011263562513, "loss": 0.1269, "step": 273 }, { "epoch": 0.9873873873873874, "grad_norm": 0.12708467697016343, "learning_rate": 0.00011943168608810978, "loss": 0.1393, "step": 274 }, { "epoch": 0.990990990990991, "grad_norm": 0.12975117728566488, "learning_rate": 0.00011881248623686338, "loss": 0.1305, "step": 275 }, { "epoch": 0.9945945945945946, "grad_norm": 0.13065574753229398, "learning_rate": 0.00011819253772357442, "loss": 0.1236, "step": 276 }, { "epoch": 0.9981981981981982, "grad_norm": 0.11060359555949814, "learning_rate": 0.00011757186521972512, "loss": 0.1018, "step": 277 }, { "epoch": 0.9981981981981982, "eval_loss": 0.12383058667182922, "eval_runtime": 52.8086, "eval_samples_per_second": 4.431, "eval_steps_per_second": 0.568, "step": 277 }, { "epoch": 1.0018018018018018, "grad_norm": 0.11956545731254414, "learning_rate": 0.00011695049342560968, "loss": 0.0926, "step": 278 }, { "epoch": 1.0054054054054054, "grad_norm": 0.07533035620811855, "learning_rate": 0.00011632844706935124, "loss": 0.0797, "step": 279 }, { "epoch": 1.009009009009009, "grad_norm": 0.07020760288792346, "learning_rate": 0.00011570575090591791, "loss": 0.0607, "step": 280 }, { "epoch": 1.0126126126126127, "grad_norm": 0.07819045444088978, "learning_rate": 0.00011508242971613741, "loss": 0.0735, "step": 281 }, { "epoch": 1.0162162162162163, "grad_norm": 0.10053168911518578, "learning_rate": 0.0001144585083057111, "loss": 0.0835, "step": 282 }, { "epoch": 1.0198198198198198, "grad_norm": 0.10526070984024917, "learning_rate": 0.0001138340115042267, "loss": 0.0951, "step": 283 }, { "epoch": 1.0234234234234234, "grad_norm": 0.09945638649949284, "learning_rate": 0.00011320896416417026, "loss": 0.0767, "step": 284 }, { "epoch": 1.027027027027027, "grad_norm": 0.07761913145188672, "learning_rate": 0.00011258339115993696, "loss": 0.0683, "step": 285 }, { "epoch": 1.0306306306306305, "grad_norm": 0.09704735378738133, "learning_rate": 0.0001119573173868415, "loss": 0.0743, "step": 286 }, { "epoch": 1.0342342342342343, "grad_norm": 0.07516525486775329, "learning_rate": 0.000111330767760127, "loss": 0.055, "step": 287 }, { "epoch": 1.037837837837838, "grad_norm": 0.12817568478073565, "learning_rate": 0.00011070376721397373, "loss": 0.0812, "step": 288 }, { "epoch": 1.0414414414414415, "grad_norm": 0.14184653764167465, "learning_rate": 0.00011007634070050684, "loss": 0.1011, "step": 289 }, { "epoch": 1.045045045045045, "grad_norm": 0.12176639431416836, "learning_rate": 0.00010944851318880314, "loss": 0.0658, "step": 290 }, { "epoch": 1.0486486486486486, "grad_norm": 0.12307205920891376, "learning_rate": 0.00010882030966389766, "loss": 0.0681, "step": 291 }, { "epoch": 1.0522522522522522, "grad_norm": 0.10538765068004156, "learning_rate": 0.00010819175512578926, "loss": 0.0641, "step": 292 }, { "epoch": 1.055855855855856, "grad_norm": 0.1835426273917669, "learning_rate": 0.00010756287458844569, "loss": 0.0741, "step": 293 }, { "epoch": 1.0594594594594595, "grad_norm": 0.15004556066173333, "learning_rate": 0.00010693369307880816, "loss": 0.0697, "step": 294 }, { "epoch": 1.063063063063063, "grad_norm": 0.17087142751095086, "learning_rate": 0.00010630423563579551, "loss": 0.0908, "step": 295 }, { "epoch": 1.0666666666666667, "grad_norm": 0.11365528787847244, "learning_rate": 0.00010567452730930743, "loss": 0.0618, "step": 296 }, { "epoch": 1.0702702702702702, "grad_norm": 0.1164578811980725, "learning_rate": 0.00010504459315922788, "loss": 0.0622, "step": 297 }, { "epoch": 1.0738738738738738, "grad_norm": 0.188496193123827, "learning_rate": 0.00010441445825442772, "loss": 0.1077, "step": 298 }, { "epoch": 1.0774774774774776, "grad_norm": 0.10295065488210545, "learning_rate": 0.00010378414767176705, "loss": 0.0735, "step": 299 }, { "epoch": 1.0810810810810811, "grad_norm": 0.15374433915948252, "learning_rate": 0.00010315368649509716, "loss": 0.085, "step": 300 }, { "epoch": 1.0846846846846847, "grad_norm": 0.09815844567459187, "learning_rate": 0.00010252309981426244, "loss": 0.054, "step": 301 }, { "epoch": 1.0882882882882883, "grad_norm": 0.09380489686860213, "learning_rate": 0.0001018924127241019, "loss": 0.058, "step": 302 }, { "epoch": 1.0918918918918918, "grad_norm": 0.1284280573750672, "learning_rate": 0.00010126165032345038, "loss": 0.0769, "step": 303 }, { "epoch": 1.0954954954954954, "grad_norm": 0.1229186112985099, "learning_rate": 0.00010063083771413975, "loss": 0.0859, "step": 304 }, { "epoch": 1.0990990990990992, "grad_norm": 0.09913923706830541, "learning_rate": 0.0001, "loss": 0.077, "step": 305 }, { "epoch": 1.1027027027027028, "grad_norm": 0.10294321295155191, "learning_rate": 9.936916228586028e-05, "loss": 0.0756, "step": 306 }, { "epoch": 1.1063063063063063, "grad_norm": 0.09262673197538004, "learning_rate": 9.873834967654964e-05, "loss": 0.0583, "step": 307 }, { "epoch": 1.10990990990991, "grad_norm": 0.09476725567348886, "learning_rate": 9.810758727589813e-05, "loss": 0.0659, "step": 308 }, { "epoch": 1.1135135135135135, "grad_norm": 0.10386525977756432, "learning_rate": 9.747690018573757e-05, "loss": 0.0604, "step": 309 }, { "epoch": 1.117117117117117, "grad_norm": 0.1368700701066548, "learning_rate": 9.684631350490287e-05, "loss": 0.0728, "step": 310 }, { "epoch": 1.1207207207207208, "grad_norm": 0.1432741421446136, "learning_rate": 9.621585232823298e-05, "loss": 0.0922, "step": 311 }, { "epoch": 1.1243243243243244, "grad_norm": 0.11931342929926518, "learning_rate": 9.55855417455723e-05, "loss": 0.0763, "step": 312 }, { "epoch": 1.127927927927928, "grad_norm": 0.10803701150902781, "learning_rate": 9.495540684077216e-05, "loss": 0.0661, "step": 313 }, { "epoch": 1.1315315315315315, "grad_norm": 0.12884484782558658, "learning_rate": 9.432547269069261e-05, "loss": 0.0606, "step": 314 }, { "epoch": 1.135135135135135, "grad_norm": 0.1344643405568192, "learning_rate": 9.36957643642045e-05, "loss": 0.0779, "step": 315 }, { "epoch": 1.1387387387387387, "grad_norm": 0.12760304961299287, "learning_rate": 9.306630692119182e-05, "loss": 0.0603, "step": 316 }, { "epoch": 1.1423423423423422, "grad_norm": 0.11600505233354397, "learning_rate": 9.243712541155436e-05, "loss": 0.0731, "step": 317 }, { "epoch": 1.145945945945946, "grad_norm": 0.12470577794958661, "learning_rate": 9.180824487421077e-05, "loss": 0.0712, "step": 318 }, { "epoch": 1.1495495495495496, "grad_norm": 0.1612422337136671, "learning_rate": 9.117969033610236e-05, "loss": 0.0683, "step": 319 }, { "epoch": 1.1531531531531531, "grad_norm": 0.14102273355899492, "learning_rate": 9.055148681119688e-05, "loss": 0.0674, "step": 320 }, { "epoch": 1.1567567567567567, "grad_norm": 0.14126790004481535, "learning_rate": 8.992365929949319e-05, "loss": 0.0812, "step": 321 }, { "epoch": 1.1603603603603603, "grad_norm": 0.15229918606873402, "learning_rate": 8.929623278602627e-05, "loss": 0.0701, "step": 322 }, { "epoch": 1.163963963963964, "grad_norm": 0.15483466842409133, "learning_rate": 8.866923223987302e-05, "loss": 0.0736, "step": 323 }, { "epoch": 1.1675675675675676, "grad_norm": 0.14180233144913557, "learning_rate": 8.80426826131585e-05, "loss": 0.0783, "step": 324 }, { "epoch": 1.1711711711711712, "grad_norm": 0.13636974208873606, "learning_rate": 8.741660884006303e-05, "loss": 0.0694, "step": 325 }, { "epoch": 1.1747747747747748, "grad_norm": 0.09259441173399019, "learning_rate": 8.679103583582979e-05, "loss": 0.0524, "step": 326 }, { "epoch": 1.1783783783783783, "grad_norm": 0.13715824142259708, "learning_rate": 8.616598849577333e-05, "loss": 0.08, "step": 327 }, { "epoch": 1.181981981981982, "grad_norm": 0.12376655099748206, "learning_rate": 8.554149169428894e-05, "loss": 0.0784, "step": 328 }, { "epoch": 1.1855855855855855, "grad_norm": 0.09319112288968635, "learning_rate": 8.491757028386263e-05, "loss": 0.0586, "step": 329 }, { "epoch": 1.1891891891891893, "grad_norm": 0.10475390780150441, "learning_rate": 8.429424909408214e-05, "loss": 0.0563, "step": 330 }, { "epoch": 1.1927927927927928, "grad_norm": 0.138568291584437, "learning_rate": 8.367155293064878e-05, "loss": 0.0894, "step": 331 }, { "epoch": 1.1963963963963964, "grad_norm": 0.10228738965627485, "learning_rate": 8.304950657439033e-05, "loss": 0.0571, "step": 332 }, { "epoch": 1.2, "grad_norm": 0.09804366197069557, "learning_rate": 8.242813478027492e-05, "loss": 0.0632, "step": 333 }, { "epoch": 1.2036036036036035, "grad_norm": 0.09408876005476795, "learning_rate": 8.180746227642562e-05, "loss": 0.0553, "step": 334 }, { "epoch": 1.2072072072072073, "grad_norm": 0.13733518876900813, "learning_rate": 8.118751376313664e-05, "loss": 0.074, "step": 335 }, { "epoch": 1.2108108108108109, "grad_norm": 0.11537452385210425, "learning_rate": 8.056831391189023e-05, "loss": 0.0686, "step": 336 }, { "epoch": 1.2144144144144144, "grad_norm": 0.1072867298305809, "learning_rate": 7.99498873643749e-05, "loss": 0.0628, "step": 337 }, { "epoch": 1.218018018018018, "grad_norm": 0.10612605300014923, "learning_rate": 7.93322587315047e-05, "loss": 0.0678, "step": 338 }, { "epoch": 1.2216216216216216, "grad_norm": 0.09775413101898157, "learning_rate": 7.87154525924399e-05, "loss": 0.0577, "step": 339 }, { "epoch": 1.2252252252252251, "grad_norm": 0.11955490707565056, "learning_rate": 7.809949349360872e-05, "loss": 0.0576, "step": 340 }, { "epoch": 1.2288288288288287, "grad_norm": 0.10380886011793584, "learning_rate": 7.74844059477306e-05, "loss": 0.0603, "step": 341 }, { "epoch": 1.2324324324324325, "grad_norm": 0.13589511087320075, "learning_rate": 7.687021443284071e-05, "loss": 0.0773, "step": 342 }, { "epoch": 1.236036036036036, "grad_norm": 0.1167184661521862, "learning_rate": 7.625694339131564e-05, "loss": 0.0677, "step": 343 }, { "epoch": 1.2396396396396396, "grad_norm": 0.12624459810290067, "learning_rate": 7.564461722890081e-05, "loss": 0.0802, "step": 344 }, { "epoch": 1.2432432432432432, "grad_norm": 0.12261184267145957, "learning_rate": 7.503326031373931e-05, "loss": 0.0649, "step": 345 }, { "epoch": 1.2468468468468468, "grad_norm": 0.16140905294131228, "learning_rate": 7.442289697540201e-05, "loss": 0.0648, "step": 346 }, { "epoch": 1.2504504504504506, "grad_norm": 0.1369989260957558, "learning_rate": 7.381355150391933e-05, "loss": 0.074, "step": 347 }, { "epoch": 1.2540540540540541, "grad_norm": 0.10405503701690619, "learning_rate": 7.32052481488147e-05, "loss": 0.0683, "step": 348 }, { "epoch": 1.2576576576576577, "grad_norm": 0.11234589174920957, "learning_rate": 7.25980111181394e-05, "loss": 0.0643, "step": 349 }, { "epoch": 1.2612612612612613, "grad_norm": 0.09321884974416474, "learning_rate": 7.19918645775093e-05, "loss": 0.0571, "step": 350 }, { "epoch": 1.2648648648648648, "grad_norm": 0.12641606453495435, "learning_rate": 7.138683264914314e-05, "loss": 0.0702, "step": 351 }, { "epoch": 1.2684684684684684, "grad_norm": 0.09916971519672783, "learning_rate": 7.078293941090249e-05, "loss": 0.0669, "step": 352 }, { "epoch": 1.272072072072072, "grad_norm": 0.11635429968669815, "learning_rate": 7.018020889533348e-05, "loss": 0.071, "step": 353 }, { "epoch": 1.2756756756756757, "grad_norm": 0.1634329754196539, "learning_rate": 6.957866508871068e-05, "loss": 0.0956, "step": 354 }, { "epoch": 1.2792792792792793, "grad_norm": 0.10617306865400682, "learning_rate": 6.897833193008231e-05, "loss": 0.0601, "step": 355 }, { "epoch": 1.2828828828828829, "grad_norm": 0.09334802201378282, "learning_rate": 6.83792333103176e-05, "loss": 0.0633, "step": 356 }, { "epoch": 1.2864864864864864, "grad_norm": 0.09803158330755328, "learning_rate": 6.77813930711561e-05, "loss": 0.059, "step": 357 }, { "epoch": 1.29009009009009, "grad_norm": 0.1431245474215182, "learning_rate": 6.718483500425867e-05, "loss": 0.0942, "step": 358 }, { "epoch": 1.2936936936936938, "grad_norm": 0.093782557814536, "learning_rate": 6.658958285026102e-05, "loss": 0.0606, "step": 359 }, { "epoch": 1.2972972972972974, "grad_norm": 0.11393416329015467, "learning_rate": 6.599566029782863e-05, "loss": 0.0717, "step": 360 }, { "epoch": 1.300900900900901, "grad_norm": 0.13923517110417213, "learning_rate": 6.540309098271416e-05, "loss": 0.0702, "step": 361 }, { "epoch": 1.3045045045045045, "grad_norm": 0.11247902438190675, "learning_rate": 6.48118984868167e-05, "loss": 0.0671, "step": 362 }, { "epoch": 1.308108108108108, "grad_norm": 0.08843201816576483, "learning_rate": 6.42221063372436e-05, "loss": 0.0528, "step": 363 }, { "epoch": 1.3117117117117116, "grad_norm": 0.09716724428964281, "learning_rate": 6.363373800537387e-05, "loss": 0.064, "step": 364 }, { "epoch": 1.3153153153153152, "grad_norm": 0.15158524613568697, "learning_rate": 6.304681690592431e-05, "loss": 0.0704, "step": 365 }, { "epoch": 1.318918918918919, "grad_norm": 0.16465157701305033, "learning_rate": 6.246136639601764e-05, "loss": 0.0834, "step": 366 }, { "epoch": 1.3225225225225226, "grad_norm": 0.12163370539770614, "learning_rate": 6.187740977425285e-05, "loss": 0.0734, "step": 367 }, { "epoch": 1.3261261261261261, "grad_norm": 0.10264467748684972, "learning_rate": 6.129497027977829e-05, "loss": 0.0688, "step": 368 }, { "epoch": 1.3297297297297297, "grad_norm": 0.0908877614772114, "learning_rate": 6.071407109136662e-05, "loss": 0.056, "step": 369 }, { "epoch": 1.3333333333333333, "grad_norm": 0.13375636202891336, "learning_rate": 6.0134735326492456e-05, "loss": 0.0775, "step": 370 }, { "epoch": 1.336936936936937, "grad_norm": 0.11306366286295132, "learning_rate": 5.955698604041231e-05, "loss": 0.0609, "step": 371 }, { "epoch": 1.3405405405405406, "grad_norm": 0.1273835553046169, "learning_rate": 5.8980846225247286e-05, "loss": 0.0653, "step": 372 }, { "epoch": 1.3441441441441442, "grad_norm": 0.10437088258018255, "learning_rate": 5.8406338809067874e-05, "loss": 0.0639, "step": 373 }, { "epoch": 1.3477477477477477, "grad_norm": 0.10184725937214306, "learning_rate": 5.7833486654981606e-05, "loss": 0.0531, "step": 374 }, { "epoch": 1.3513513513513513, "grad_norm": 0.1439045913318642, "learning_rate": 5.726231256022316e-05, "loss": 0.0756, "step": 375 }, { "epoch": 1.3549549549549549, "grad_norm": 0.1161759648717375, "learning_rate": 5.669283925524715e-05, "loss": 0.0564, "step": 376 }, { "epoch": 1.3585585585585584, "grad_norm": 0.14605792072752058, "learning_rate": 5.6125089402823485e-05, "loss": 0.0676, "step": 377 }, { "epoch": 1.3621621621621622, "grad_norm": 0.11801235296837954, "learning_rate": 5.555908559713561e-05, "loss": 0.0922, "step": 378 }, { "epoch": 1.3657657657657658, "grad_norm": 0.0986919703756885, "learning_rate": 5.4994850362881214e-05, "loss": 0.0624, "step": 379 }, { "epoch": 1.3693693693693694, "grad_norm": 0.10287784532358656, "learning_rate": 5.443240615437586e-05, "loss": 0.0692, "step": 380 }, { "epoch": 1.372972972972973, "grad_norm": 0.1340740211739596, "learning_rate": 5.387177535465945e-05, "loss": 0.0835, "step": 381 }, { "epoch": 1.3765765765765765, "grad_norm": 0.138304807951036, "learning_rate": 5.331298027460539e-05, "loss": 0.0749, "step": 382 }, { "epoch": 1.3801801801801803, "grad_norm": 0.11145520456541595, "learning_rate": 5.275604315203293e-05, "loss": 0.062, "step": 383 }, { "epoch": 1.3837837837837839, "grad_norm": 0.09247928908715367, "learning_rate": 5.2200986150821696e-05, "loss": 0.0571, "step": 384 }, { "epoch": 1.3873873873873874, "grad_norm": 0.1101460184702829, "learning_rate": 5.164783136003027e-05, "loss": 0.076, "step": 385 }, { "epoch": 1.390990990990991, "grad_norm": 0.09455088802615286, "learning_rate": 5.109660079301668e-05, "loss": 0.0708, "step": 386 }, { "epoch": 1.3945945945945946, "grad_norm": 0.12915569506877214, "learning_rate": 5.0547316386562507e-05, "loss": 0.0683, "step": 387 }, { "epoch": 1.3981981981981981, "grad_norm": 0.14298967602118384, "learning_rate": 5.000000000000002e-05, "loss": 0.0802, "step": 388 }, { "epoch": 1.4018018018018017, "grad_norm": 0.13870069410611544, "learning_rate": 4.945467341434195e-05, "loss": 0.0849, "step": 389 }, { "epoch": 1.4054054054054055, "grad_norm": 0.1542562430256231, "learning_rate": 4.891135833141495e-05, "loss": 0.0875, "step": 390 }, { "epoch": 1.409009009009009, "grad_norm": 0.11965550746042822, "learning_rate": 4.837007637299595e-05, "loss": 0.0599, "step": 391 }, { "epoch": 1.4126126126126126, "grad_norm": 0.13624840902702007, "learning_rate": 4.783084907995156e-05, "loss": 0.0805, "step": 392 }, { "epoch": 1.4162162162162162, "grad_norm": 0.0916114816428032, "learning_rate": 4.729369791138085e-05, "loss": 0.0523, "step": 393 }, { "epoch": 1.4198198198198198, "grad_norm": 0.10008527767691701, "learning_rate": 4.675864424376146e-05, "loss": 0.0623, "step": 394 }, { "epoch": 1.4234234234234235, "grad_norm": 0.1329413030801314, "learning_rate": 4.622570937009879e-05, "loss": 0.0768, "step": 395 }, { "epoch": 1.427027027027027, "grad_norm": 0.09990109338082896, "learning_rate": 4.569491449907878e-05, "loss": 0.0624, "step": 396 }, { "epoch": 1.4306306306306307, "grad_norm": 0.10994747606781914, "learning_rate": 4.5166280754223676e-05, "loss": 0.0554, "step": 397 }, { "epoch": 1.4342342342342342, "grad_norm": 0.11716673038318011, "learning_rate": 4.4639829173051554e-05, "loss": 0.0487, "step": 398 }, { "epoch": 1.4378378378378378, "grad_norm": 0.14740947528399398, "learning_rate": 4.411558070623907e-05, "loss": 0.0705, "step": 399 }, { "epoch": 1.4414414414414414, "grad_norm": 0.10905802122912368, "learning_rate": 4.359355621678764e-05, "loss": 0.0669, "step": 400 }, { "epoch": 1.445045045045045, "grad_norm": 0.1372304462996132, "learning_rate": 4.307377647919343e-05, "loss": 0.0754, "step": 401 }, { "epoch": 1.4486486486486487, "grad_norm": 0.1220267603792114, "learning_rate": 4.255626217862013e-05, "loss": 0.061, "step": 402 }, { "epoch": 1.4522522522522523, "grad_norm": 0.10857496168656595, "learning_rate": 4.204103391007623e-05, "loss": 0.0666, "step": 403 }, { "epoch": 1.4558558558558559, "grad_norm": 0.12862098370827468, "learning_rate": 4.152811217759529e-05, "loss": 0.0702, "step": 404 }, { "epoch": 1.4594594594594594, "grad_norm": 0.12121566089700932, "learning_rate": 4.1017517393419826e-05, "loss": 0.0701, "step": 405 }, { "epoch": 1.463063063063063, "grad_norm": 0.12880561806718493, "learning_rate": 4.0509269877189106e-05, "loss": 0.0855, "step": 406 }, { "epoch": 1.4666666666666668, "grad_norm": 0.1103245174300563, "learning_rate": 4.000338985513046e-05, "loss": 0.0645, "step": 407 }, { "epoch": 1.4702702702702704, "grad_norm": 0.11471559496567096, "learning_rate": 3.9499897459254375e-05, "loss": 0.069, "step": 408 }, { "epoch": 1.473873873873874, "grad_norm": 0.1068585326163203, "learning_rate": 3.899881272655342e-05, "loss": 0.0584, "step": 409 }, { "epoch": 1.4774774774774775, "grad_norm": 0.12180507516800447, "learning_rate": 3.8500155598204644e-05, "loss": 0.0767, "step": 410 }, { "epoch": 1.481081081081081, "grad_norm": 0.10787694069789268, "learning_rate": 3.8003945918776143e-05, "loss": 0.0652, "step": 411 }, { "epoch": 1.4846846846846846, "grad_norm": 0.11726705088810202, "learning_rate": 3.75102034354373e-05, "loss": 0.0629, "step": 412 }, { "epoch": 1.4882882882882882, "grad_norm": 0.10762739211432558, "learning_rate": 3.701894779717286e-05, "loss": 0.0555, "step": 413 }, { "epoch": 1.491891891891892, "grad_norm": 0.09308134521067353, "learning_rate": 3.653019855400123e-05, "loss": 0.0564, "step": 414 }, { "epoch": 1.4954954954954955, "grad_norm": 0.1524049565519016, "learning_rate": 3.6043975156195987e-05, "loss": 0.0809, "step": 415 }, { "epoch": 1.499099099099099, "grad_norm": 0.13279511071977756, "learning_rate": 3.5560296953512295e-05, "loss": 0.069, "step": 416 }, { "epoch": 1.5027027027027027, "grad_norm": 0.13286192000673344, "learning_rate": 3.507918319441672e-05, "loss": 0.0748, "step": 417 }, { "epoch": 1.5063063063063065, "grad_norm": 0.17493415887546465, "learning_rate": 3.460065302532108e-05, "loss": 0.0828, "step": 418 }, { "epoch": 1.50990990990991, "grad_norm": 0.11057059837297634, "learning_rate": 3.4124725489820645e-05, "loss": 0.063, "step": 419 }, { "epoch": 1.5135135135135136, "grad_norm": 0.12735355112497881, "learning_rate": 3.365141952793622e-05, "loss": 0.0732, "step": 420 }, { "epoch": 1.5171171171171172, "grad_norm": 0.1632337270590702, "learning_rate": 3.3180753975360415e-05, "loss": 0.0775, "step": 421 }, { "epoch": 1.5207207207207207, "grad_norm": 0.14697481590979153, "learning_rate": 3.2712747562708115e-05, "loss": 0.0863, "step": 422 }, { "epoch": 1.5243243243243243, "grad_norm": 0.1660001616377742, "learning_rate": 3.224741891477095e-05, "loss": 0.082, "step": 423 }, { "epoch": 1.5279279279279279, "grad_norm": 0.13048897912831442, "learning_rate": 3.178478654977624e-05, "loss": 0.0794, "step": 424 }, { "epoch": 1.5315315315315314, "grad_norm": 0.11489435507746414, "learning_rate": 3.132486887864992e-05, "loss": 0.0694, "step": 425 }, { "epoch": 1.535135135135135, "grad_norm": 0.07639223925814809, "learning_rate": 3.086768420428392e-05, "loss": 0.0413, "step": 426 }, { "epoch": 1.5387387387387388, "grad_norm": 0.11453408038112112, "learning_rate": 3.0413250720807883e-05, "loss": 0.0658, "step": 427 }, { "epoch": 1.5423423423423424, "grad_norm": 0.1667771876584119, "learning_rate": 2.9961586512864947e-05, "loss": 0.0798, "step": 428 }, { "epoch": 1.545945945945946, "grad_norm": 0.11714414917895159, "learning_rate": 2.9512709554892003e-05, "loss": 0.0693, "step": 429 }, { "epoch": 1.5495495495495497, "grad_norm": 0.1229110888950789, "learning_rate": 2.9066637710404675e-05, "loss": 0.0747, "step": 430 }, { "epoch": 1.5531531531531533, "grad_norm": 0.1759221433787884, "learning_rate": 2.8623388731286093e-05, "loss": 0.0795, "step": 431 }, { "epoch": 1.5567567567567568, "grad_norm": 0.09594983270629338, "learning_rate": 2.818298025708075e-05, "loss": 0.059, "step": 432 }, { "epoch": 1.5603603603603604, "grad_norm": 0.090367685537466, "learning_rate": 2.7745429814292145e-05, "loss": 0.0531, "step": 433 }, { "epoch": 1.563963963963964, "grad_norm": 0.13195671170326462, "learning_rate": 2.7310754815685624e-05, "loss": 0.075, "step": 434 }, { "epoch": 1.5675675675675675, "grad_norm": 0.09101802024714302, "learning_rate": 2.687897255959536e-05, "loss": 0.0541, "step": 435 }, { "epoch": 1.571171171171171, "grad_norm": 0.12030212380166792, "learning_rate": 2.6450100229235795e-05, "loss": 0.0739, "step": 436 }, { "epoch": 1.5747747747747747, "grad_norm": 0.09282633668248956, "learning_rate": 2.6024154892017937e-05, "loss": 0.0585, "step": 437 }, { "epoch": 1.5783783783783782, "grad_norm": 0.1437367048181832, "learning_rate": 2.5601153498870134e-05, "loss": 0.0744, "step": 438 }, { "epoch": 1.581981981981982, "grad_norm": 0.12980503831327814, "learning_rate": 2.518111288356345e-05, "loss": 0.0741, "step": 439 }, { "epoch": 1.5855855855855856, "grad_norm": 0.12335416962202174, "learning_rate": 2.4764049762041874e-05, "loss": 0.072, "step": 440 }, { "epoch": 1.5891891891891892, "grad_norm": 0.10492372446642, "learning_rate": 2.4349980731756894e-05, "loss": 0.0716, "step": 441 }, { "epoch": 1.592792792792793, "grad_norm": 0.14899587708147966, "learning_rate": 2.3938922271007147e-05, "loss": 0.0925, "step": 442 }, { "epoch": 1.5963963963963965, "grad_norm": 0.13084935137363646, "learning_rate": 2.353089073828255e-05, "loss": 0.0823, "step": 443 }, { "epoch": 1.6, "grad_norm": 0.13030296264089844, "learning_rate": 2.312590237161335e-05, "loss": 0.0725, "step": 444 }, { "epoch": 1.6036036036036037, "grad_norm": 0.09700200874554347, "learning_rate": 2.2723973287923962e-05, "loss": 0.0664, "step": 445 }, { "epoch": 1.6072072072072072, "grad_norm": 0.10633986810502474, "learning_rate": 2.2325119482391467e-05, "loss": 0.0679, "step": 446 }, { "epoch": 1.6108108108108108, "grad_norm": 0.11142544471141212, "learning_rate": 2.1929356827809057e-05, "loss": 0.0614, "step": 447 }, { "epoch": 1.6144144144144144, "grad_norm": 0.09250950825140586, "learning_rate": 2.1536701073954558e-05, "loss": 0.0588, "step": 448 }, { "epoch": 1.618018018018018, "grad_norm": 0.1332480809778604, "learning_rate": 2.1147167846963422e-05, "loss": 0.0803, "step": 449 }, { "epoch": 1.6216216216216215, "grad_norm": 0.14121311299673173, "learning_rate": 2.0760772648707016e-05, "loss": 0.0947, "step": 450 }, { "epoch": 1.6252252252252253, "grad_norm": 0.10826628920408898, "learning_rate": 2.037753085617563e-05, "loss": 0.0704, "step": 451 }, { "epoch": 1.6288288288288288, "grad_norm": 0.08498951848877008, "learning_rate": 1.999745772086655e-05, "loss": 0.0416, "step": 452 }, { "epoch": 1.6324324324324324, "grad_norm": 0.15866638731251254, "learning_rate": 1.9620568368177184e-05, "loss": 0.0836, "step": 453 }, { "epoch": 1.6360360360360362, "grad_norm": 0.11879058663162903, "learning_rate": 1.924687779680302e-05, "loss": 0.0721, "step": 454 }, { "epoch": 1.6396396396396398, "grad_norm": 0.16212624210699175, "learning_rate": 1.8876400878140775e-05, "loss": 0.0788, "step": 455 }, { "epoch": 1.6432432432432433, "grad_norm": 0.1205599470750737, "learning_rate": 1.8509152355696623e-05, "loss": 0.0891, "step": 456 }, { "epoch": 1.646846846846847, "grad_norm": 0.10723522883909493, "learning_rate": 1.8145146844499383e-05, "loss": 0.0702, "step": 457 }, { "epoch": 1.6504504504504505, "grad_norm": 0.10243191710751155, "learning_rate": 1.7784398830519e-05, "loss": 0.0558, "step": 458 }, { "epoch": 1.654054054054054, "grad_norm": 0.08896982101996669, "learning_rate": 1.742692267008996e-05, "loss": 0.0603, "step": 459 }, { "epoch": 1.6576576576576576, "grad_norm": 0.13618088732046185, "learning_rate": 1.7072732589339955e-05, "loss": 0.0744, "step": 460 }, { "epoch": 1.6612612612612612, "grad_norm": 0.14798363521421304, "learning_rate": 1.672184268362391e-05, "loss": 0.0875, "step": 461 }, { "epoch": 1.6648648648648647, "grad_norm": 0.10230547180028401, "learning_rate": 1.6374266916962832e-05, "loss": 0.0497, "step": 462 }, { "epoch": 1.6684684684684683, "grad_norm": 0.10092039325506573, "learning_rate": 1.6030019121488227e-05, "loss": 0.0611, "step": 463 }, { "epoch": 1.672072072072072, "grad_norm": 0.10528958039623297, "learning_rate": 1.5689112996891576e-05, "loss": 0.0719, "step": 464 }, { "epoch": 1.6756756756756757, "grad_norm": 0.1357872146845403, "learning_rate": 1.535156210987917e-05, "loss": 0.0792, "step": 465 }, { "epoch": 1.6792792792792792, "grad_norm": 0.16092347024456755, "learning_rate": 1.5017379893632255e-05, "loss": 0.0984, "step": 466 }, { "epoch": 1.682882882882883, "grad_norm": 0.152681018425233, "learning_rate": 1.4686579647272336e-05, "loss": 0.0665, "step": 467 }, { "epoch": 1.6864864864864866, "grad_norm": 0.10940364333158531, "learning_rate": 1.4359174535331999e-05, "loss": 0.0678, "step": 468 }, { "epoch": 1.6900900900900901, "grad_norm": 0.10224571124448578, "learning_rate": 1.4035177587230996e-05, "loss": 0.0681, "step": 469 }, { "epoch": 1.6936936936936937, "grad_norm": 0.1117508851307017, "learning_rate": 1.3714601696757712e-05, "loss": 0.0705, "step": 470 }, { "epoch": 1.6972972972972973, "grad_norm": 0.11380708386619733, "learning_rate": 1.339745962155613e-05, "loss": 0.0787, "step": 471 }, { "epoch": 1.7009009009009008, "grad_norm": 0.13014691697383945, "learning_rate": 1.3083763982618025e-05, "loss": 0.0746, "step": 472 }, { "epoch": 1.7045045045045044, "grad_norm": 0.14178781462677711, "learning_rate": 1.2773527263780626e-05, "loss": 0.0802, "step": 473 }, { "epoch": 1.708108108108108, "grad_norm": 0.13903277880623008, "learning_rate": 1.2466761811230098e-05, "loss": 0.0727, "step": 474 }, { "epoch": 1.7117117117117115, "grad_norm": 0.14354542440047877, "learning_rate": 1.2163479833009894e-05, "loss": 0.0698, "step": 475 }, { "epoch": 1.7153153153153153, "grad_norm": 0.12299158086851347, "learning_rate": 1.1863693398535114e-05, "loss": 0.0659, "step": 476 }, { "epoch": 1.718918918918919, "grad_norm": 0.09882489616646957, "learning_rate": 1.1567414438112156e-05, "loss": 0.0626, "step": 477 }, { "epoch": 1.7225225225225225, "grad_norm": 0.11570669617801674, "learning_rate": 1.1274654742463841e-05, "loss": 0.0646, "step": 478 }, { "epoch": 1.7261261261261263, "grad_norm": 0.09504741769642676, "learning_rate": 1.0985425962260343e-05, "loss": 0.0587, "step": 479 }, { "epoch": 1.7297297297297298, "grad_norm": 0.12981804877027894, "learning_rate": 1.0699739607655435e-05, "loss": 0.0646, "step": 480 }, { "epoch": 1.7333333333333334, "grad_norm": 0.10868480851483209, "learning_rate": 1.0417607047828426e-05, "loss": 0.0671, "step": 481 }, { "epoch": 1.736936936936937, "grad_norm": 0.13696903097667856, "learning_rate": 1.01390395105318e-05, "loss": 0.0786, "step": 482 }, { "epoch": 1.7405405405405405, "grad_norm": 0.1157828731078967, "learning_rate": 9.864048081644261e-06, "loss": 0.0714, "step": 483 }, { "epoch": 1.744144144144144, "grad_norm": 0.08900480095300936, "learning_rate": 9.592643704729753e-06, "loss": 0.0544, "step": 484 }, { "epoch": 1.7477477477477477, "grad_norm": 0.10882498606926169, "learning_rate": 9.324837180601741e-06, "loss": 0.0645, "step": 485 }, { "epoch": 1.7513513513513512, "grad_norm": 0.1879024889958637, "learning_rate": 9.060639166893493e-06, "loss": 0.0682, "step": 486 }, { "epoch": 1.7549549549549548, "grad_norm": 0.1072698067292599, "learning_rate": 8.80006017763395e-06, "loss": 0.0558, "step": 487 }, { "epoch": 1.7585585585585586, "grad_norm": 0.1177990807822665, "learning_rate": 8.543110582829272e-06, "loss": 0.0592, "step": 488 }, { "epoch": 1.7621621621621621, "grad_norm": 0.1144052781708329, "learning_rate": 8.289800608050202e-06, "loss": 0.0685, "step": 489 }, { "epoch": 1.7657657657657657, "grad_norm": 0.13290723111377578, "learning_rate": 8.040140334025082e-06, "loss": 0.0787, "step": 490 }, { "epoch": 1.7693693693693695, "grad_norm": 0.14325581569648474, "learning_rate": 7.794139696238645e-06, "loss": 0.0767, "step": 491 }, { "epoch": 1.772972972972973, "grad_norm": 0.13940373462189817, "learning_rate": 7.551808484536782e-06, "loss": 0.0713, "step": 492 }, { "epoch": 1.7765765765765766, "grad_norm": 0.13465478912453277, "learning_rate": 7.313156342736738e-06, "loss": 0.0838, "step": 493 }, { "epoch": 1.7801801801801802, "grad_norm": 0.10122791673864097, "learning_rate": 7.078192768243486e-06, "loss": 0.0577, "step": 494 }, { "epoch": 1.7837837837837838, "grad_norm": 0.1571652297573745, "learning_rate": 6.846927111671686e-06, "loss": 0.0905, "step": 495 }, { "epoch": 1.7873873873873873, "grad_norm": 0.12718467808306833, "learning_rate": 6.61936857647355e-06, "loss": 0.0566, "step": 496 }, { "epoch": 1.790990990990991, "grad_norm": 0.15255418113431185, "learning_rate": 6.395526218572723e-06, "loss": 0.0646, "step": 497 }, { "epoch": 1.7945945945945945, "grad_norm": 0.13721542773233897, "learning_rate": 6.175408946003703e-06, "loss": 0.0752, "step": 498 }, { "epoch": 1.798198198198198, "grad_norm": 0.11503986327868332, "learning_rate": 5.959025518557437e-06, "loss": 0.0753, "step": 499 }, { "epoch": 1.8018018018018018, "grad_norm": 0.09966839333280018, "learning_rate": 5.746384547432737e-06, "loss": 0.0658, "step": 500 }, { "epoch": 1.8054054054054054, "grad_norm": 0.13097929394232063, "learning_rate": 5.5374944948935135e-06, "loss": 0.0647, "step": 501 }, { "epoch": 1.809009009009009, "grad_norm": 0.11916565098885482, "learning_rate": 5.332363673932106e-06, "loss": 0.0683, "step": 502 }, { "epoch": 1.8126126126126128, "grad_norm": 0.14740354567299807, "learning_rate": 5.131000247938367e-06, "loss": 0.0855, "step": 503 }, { "epoch": 1.8162162162162163, "grad_norm": 0.12344512852482942, "learning_rate": 4.933412230374812e-06, "loss": 0.0689, "step": 504 }, { "epoch": 1.8198198198198199, "grad_norm": 0.15079351397270196, "learning_rate": 4.7396074844577975e-06, "loss": 0.0865, "step": 505 }, { "epoch": 1.8234234234234235, "grad_norm": 0.13387613481208308, "learning_rate": 4.549593722844492e-06, "loss": 0.0761, "step": 506 }, { "epoch": 1.827027027027027, "grad_norm": 0.07611010028345803, "learning_rate": 4.363378507325955e-06, "loss": 0.0431, "step": 507 }, { "epoch": 1.8306306306306306, "grad_norm": 0.13798397287470085, "learning_rate": 4.180969248526334e-06, "loss": 0.0806, "step": 508 }, { "epoch": 1.8342342342342342, "grad_norm": 0.17673529491591075, "learning_rate": 4.002373205607723e-06, "loss": 0.0974, "step": 509 }, { "epoch": 1.8378378378378377, "grad_norm": 0.09469787111672756, "learning_rate": 3.827597485981527e-06, "loss": 0.0592, "step": 510 }, { "epoch": 1.8414414414414413, "grad_norm": 0.10244301409759372, "learning_rate": 3.6566490450254286e-06, "loss": 0.0629, "step": 511 }, { "epoch": 1.845045045045045, "grad_norm": 0.09239653678543874, "learning_rate": 3.4895346858066724e-06, "loss": 0.0637, "step": 512 }, { "epoch": 1.8486486486486486, "grad_norm": 0.136056185401784, "learning_rate": 3.3262610588113307e-06, "loss": 0.0784, "step": 513 }, { "epoch": 1.8522522522522522, "grad_norm": 0.12238600483310333, "learning_rate": 3.1668346616795963e-06, "loss": 0.0703, "step": 514 }, { "epoch": 1.855855855855856, "grad_norm": 0.15777844339693373, "learning_rate": 3.011261838947277e-06, "loss": 0.0861, "step": 515 }, { "epoch": 1.8594594594594596, "grad_norm": 0.149256801647933, "learning_rate": 2.859548781793242e-06, "loss": 0.0816, "step": 516 }, { "epoch": 1.8630630630630631, "grad_norm": 0.1158625228168308, "learning_rate": 2.711701527793031e-06, "loss": 0.0757, "step": 517 }, { "epoch": 1.8666666666666667, "grad_norm": 0.127467190710186, "learning_rate": 2.5677259606786684e-06, "loss": 0.0718, "step": 518 }, { "epoch": 1.8702702702702703, "grad_norm": 0.11178985379767997, "learning_rate": 2.4276278101044046e-06, "loss": 0.0729, "step": 519 }, { "epoch": 1.8738738738738738, "grad_norm": 0.13318873038147186, "learning_rate": 2.291412651418778e-06, "loss": 0.0874, "step": 520 }, { "epoch": 1.8774774774774774, "grad_norm": 0.11095794613520046, "learning_rate": 2.159085905442737e-06, "loss": 0.0632, "step": 521 }, { "epoch": 1.881081081081081, "grad_norm": 0.13466988936312552, "learning_rate": 2.03065283825381e-06, "loss": 0.0696, "step": 522 }, { "epoch": 1.8846846846846845, "grad_norm": 0.1501784745699571, "learning_rate": 1.9061185609766995e-06, "loss": 0.0781, "step": 523 }, { "epoch": 1.8882882882882883, "grad_norm": 0.12439813775960844, "learning_rate": 1.7854880295797405e-06, "loss": 0.0778, "step": 524 }, { "epoch": 1.8918918918918919, "grad_norm": 0.10543026463333441, "learning_rate": 1.6687660446777277e-06, "loss": 0.0658, "step": 525 }, { "epoch": 1.8954954954954955, "grad_norm": 0.1493387910647338, "learning_rate": 1.5559572513409338e-06, "loss": 0.0733, "step": 526 }, { "epoch": 1.8990990990990992, "grad_norm": 0.11530728589499968, "learning_rate": 1.4470661389100804e-06, "loss": 0.066, "step": 527 }, { "epoch": 1.9027027027027028, "grad_norm": 0.0937287880770091, "learning_rate": 1.3420970408178913e-06, "loss": 0.0508, "step": 528 }, { "epoch": 1.9063063063063064, "grad_norm": 0.1140049075150384, "learning_rate": 1.241054134416464e-06, "loss": 0.0722, "step": 529 }, { "epoch": 1.90990990990991, "grad_norm": 0.10812439212587106, "learning_rate": 1.143941440811147e-06, "loss": 0.0607, "step": 530 }, { "epoch": 1.9135135135135135, "grad_norm": 0.10082812889823659, "learning_rate": 1.0507628247004465e-06, "loss": 0.0611, "step": 531 }, { "epoch": 1.917117117117117, "grad_norm": 0.1646769037018901, "learning_rate": 9.615219942222474e-07, "loss": 0.0907, "step": 532 }, { "epoch": 1.9207207207207206, "grad_norm": 0.13342387053057897, "learning_rate": 8.762225008062674e-07, "loss": 0.0705, "step": 533 }, { "epoch": 1.9243243243243242, "grad_norm": 0.10099025782756708, "learning_rate": 7.948677390326786e-07, "loss": 0.0594, "step": 534 }, { "epoch": 1.9279279279279278, "grad_norm": 0.10984097417427825, "learning_rate": 7.174609464970505e-07, "loss": 0.064, "step": 535 }, { "epoch": 1.9315315315315316, "grad_norm": 0.09250058769796181, "learning_rate": 6.440052036815081e-07, "loss": 0.0488, "step": 536 }, { "epoch": 1.9351351351351351, "grad_norm": 0.11346138285376553, "learning_rate": 5.745034338321187e-07, "loss": 0.0709, "step": 537 }, { "epoch": 1.9387387387387387, "grad_norm": 0.12145882785931153, "learning_rate": 5.089584028425743e-07, "loss": 0.0628, "step": 538 }, { "epoch": 1.9423423423423425, "grad_norm": 0.09764192386128882, "learning_rate": 4.4737271914411236e-07, "loss": 0.0581, "step": 539 }, { "epoch": 1.945945945945946, "grad_norm": 0.11336911485782475, "learning_rate": 3.8974883360169966e-07, "loss": 0.0652, "step": 540 }, { "epoch": 1.9495495495495496, "grad_norm": 0.1197688852133187, "learning_rate": 3.360890394165539e-07, "loss": 0.0797, "step": 541 }, { "epoch": 1.9531531531531532, "grad_norm": 0.12068524258623581, "learning_rate": 2.86395472034795e-07, "loss": 0.0796, "step": 542 }, { "epoch": 1.9567567567567568, "grad_norm": 0.10022614951335301, "learning_rate": 2.4067010906254626e-07, "loss": 0.0651, "step": 543 }, { "epoch": 1.9603603603603603, "grad_norm": 0.1030646452744324, "learning_rate": 1.989147701871641e-07, "loss": 0.0536, "step": 544 }, { "epoch": 1.9639639639639639, "grad_norm": 0.07714530191454869, "learning_rate": 1.611311171048735e-07, "loss": 0.0494, "step": 545 }, { "epoch": 1.9675675675675675, "grad_norm": 0.1157905922358179, "learning_rate": 1.2732065345462118e-07, "loss": 0.075, "step": 546 }, { "epoch": 1.971171171171171, "grad_norm": 0.07486976400470267, "learning_rate": 9.748472475823444e-08, "loss": 0.0454, "step": 547 }, { "epoch": 1.9747747747747748, "grad_norm": 0.09262953733370366, "learning_rate": 7.162451836685291e-08, "loss": 0.0556, "step": 548 }, { "epoch": 1.9783783783783784, "grad_norm": 0.11616787129919331, "learning_rate": 4.974106341374407e-08, "loss": 0.0767, "step": 549 }, { "epoch": 1.981981981981982, "grad_norm": 0.11058833058189024, "learning_rate": 3.183523077324724e-08, "loss": 0.0618, "step": 550 }, { "epoch": 1.9855855855855857, "grad_norm": 0.10561626473162405, "learning_rate": 1.7907733026223394e-08, "loss": 0.0719, "step": 551 }, { "epoch": 1.9891891891891893, "grad_norm": 0.1403058364172833, "learning_rate": 7.959124431622389e-09, "loss": 0.092, "step": 552 }, { "epoch": 1.9927927927927929, "grad_norm": 0.1326365804299446, "learning_rate": 1.989800904445005e-09, "loss": 0.0678, "step": 553 }, { "epoch": 1.9963963963963964, "grad_norm": 0.14065192361832085, "learning_rate": 0.0, "loss": 0.0822, "step": 554 }, { "epoch": 1.9963963963963964, "eval_loss": 0.13327383995056152, "eval_runtime": 50.2794, "eval_samples_per_second": 4.654, "eval_steps_per_second": 0.597, "step": 554 }, { "epoch": 1.9321739130434783, "grad_norm": 0.2707502844732912, "learning_rate": 6.683334509453465e-07, "loss": 0.1107, "step": 555 }, { "epoch": 1.9356521739130435, "grad_norm": 0.3664545335936956, "learning_rate": 5.999024760054095e-07, "loss": 0.1484, "step": 556 }, { "epoch": 1.9391304347826086, "grad_norm": 0.2879184291998937, "learning_rate": 5.351560625760254e-07, "loss": 0.1073, "step": 557 }, { "epoch": 1.942608695652174, "grad_norm": 0.32527132334089365, "learning_rate": 4.7409661067642217e-07, "loss": 0.1341, "step": 558 }, { "epoch": 1.9460869565217391, "grad_norm": 0.27961707371818556, "learning_rate": 4.167263836575286e-07, "loss": 0.0999, "step": 559 }, { "epoch": 1.9495652173913043, "grad_norm": 0.2705336974198295, "learning_rate": 3.630475081181861e-07, "loss": 0.0858, "step": 560 }, { "epoch": 1.9530434782608697, "grad_norm": 0.2630558680757462, "learning_rate": 3.1306197382624526e-07, "loss": 0.1232, "step": 561 }, { "epoch": 1.9565217391304348, "grad_norm": 0.2875514596758492, "learning_rate": 2.667716336448356e-07, "loss": 0.1385, "step": 562 }, { "epoch": 1.96, "grad_norm": 0.3682978167407908, "learning_rate": 2.2417820346367635e-07, "loss": 0.1103, "step": 563 }, { "epoch": 1.9634782608695653, "grad_norm": 0.26822875711798694, "learning_rate": 1.8528326213548274e-07, "loss": 0.1138, "step": 564 }, { "epoch": 1.9669565217391303, "grad_norm": 0.2418187498704634, "learning_rate": 1.50088251417424e-07, "loss": 0.1105, "step": 565 }, { "epoch": 1.9704347826086956, "grad_norm": 0.3401484224136178, "learning_rate": 1.1859447591769934e-07, "loss": 0.1377, "step": 566 }, { "epoch": 1.973913043478261, "grad_norm": 0.23339563722483295, "learning_rate": 9.080310304716567e-08, "loss": 0.1081, "step": 567 }, { "epoch": 1.977391304347826, "grad_norm": 0.33523600802230064, "learning_rate": 6.671516297606095e-08, "loss": 0.1395, "step": 568 }, { "epoch": 1.9808695652173913, "grad_norm": 0.22230937288937844, "learning_rate": 4.6331548595845984e-08, "loss": 0.0859, "step": 569 }, { "epoch": 1.9843478260869565, "grad_norm": 0.37789798217710224, "learning_rate": 2.965301548606414e-08, "loss": 0.1846, "step": 570 }, { "epoch": 1.9878260869565216, "grad_norm": 0.29502175807159936, "learning_rate": 1.6680181886352676e-08, "loss": 0.1208, "step": 571 }, { "epoch": 1.991304347826087, "grad_norm": 0.3456726449106421, "learning_rate": 7.413528673549941e-09, "loss": 0.1142, "step": 572 }, { "epoch": 1.9947826086956522, "grad_norm": 0.34376689538870997, "learning_rate": 1.8533993438318852e-09, "loss": 0.1442, "step": 573 }, { "epoch": 1.9982608695652173, "grad_norm": 0.3072940536392916, "learning_rate": 0.0, "loss": 0.1067, "step": 574 }, { "epoch": 1.9982608695652173, "step": 574, "total_flos": 466200922914816.0, "train_loss": 0.0, "train_runtime": 1.1199, "train_samples_per_second": 8208.056, "train_steps_per_second": 512.557 } ], "logging_steps": 1, "max_steps": 574, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 466200922914816.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }